/net/netfilter/ |
D | nf_synproxy_core.c | 496 nth->ack_seq = htonl(ntohl(th->seq) + 1); in synproxy_send_client_synack() 542 nth->ack_seq = htonl(ntohl(th->ack_seq) - 1); in synproxy_send_server_syn() 583 nth->seq = htonl(ntohl(th->ack_seq)); in synproxy_send_server_ack() 584 nth->ack_seq = htonl(ntohl(th->seq) + 1); in synproxy_send_server_ack() 622 nth->ack_seq = th->ack_seq; in synproxy_send_client_ack() 643 mss = __cookie_v4_check(ip_hdr(skb), th, ntohl(th->ack_seq) - 1); in synproxy_recv_client_ack() 733 synproxy->isn = ntohl(th->ack_seq); in ipv4_synproxy_hook() 909 nth->ack_seq = htonl(ntohl(th->seq) + 1); in synproxy_send_client_synack_ipv6() 956 nth->ack_seq = htonl(ntohl(th->ack_seq) - 1); in synproxy_send_server_syn_ipv6() 997 nth->seq = htonl(ntohl(th->ack_seq)); in synproxy_send_server_ack_ipv6() [all …]
|
D | nf_conntrack_seqadj.c | 198 if (after(ntohl(tcph->ack_seq) - other_way->offset_before, in nf_ct_seq_adjust() 204 newack = htonl(ntohl(tcph->ack_seq) - ackoff); in nf_ct_seq_adjust() 205 inet_proto_csum_replace4(&tcph->check, skb, tcph->ack_seq, newack, in nf_ct_seq_adjust() 208 ntohl(tcph->seq), ntohl(newseq), ntohl(tcph->ack_seq), in nf_ct_seq_adjust() 210 tcph->ack_seq = newack; in nf_ct_seq_adjust()
|
D | nf_log_common.c | 79 ntohl(th->seq), ntohl(th->ack_seq)); in nf_log_dump_tcp_header()
|
D | nf_conntrack_proto_tcp.c | 483 ack = sack = ntohl(tcph->ack_seq); in tcp_in_window() 938 && ntohl(th->ack_seq) == ct->proto.tcp.last_end) { in nf_conntrack_tcp_packet() 1089 && ntohl(th->ack_seq) == ct->proto.tcp.last_end) { in nf_conntrack_tcp_packet()
|
/net/netfilter/ipvs/ |
D | ip_vs_app.c | 317 __u32 ack_seq = ntohl(th->ack_seq); in vs_fix_ack_seq() local 327 if(after(ack_seq, vseq->init_seq+vseq->delta)) { in vs_fix_ack_seq() 328 th->ack_seq = htonl(ack_seq - vseq->delta); in vs_fix_ack_seq() 333 th->ack_seq = htonl(ack_seq - vseq->previous_delta); in vs_fix_ack_seq()
|
/net/mptcp/ |
D | protocol.c | 165 max_seq = space > 0 ? space + msk->ack_seq : msk->ack_seq; in mptcp_data_queue_ofo() 296 if (MPTCP_SKB_CB(skb)->map_seq == msk->ack_seq) { in __mptcp_move_skb() 298 WRITE_ONCE(msk->ack_seq, msk->ack_seq + copy_len); in __mptcp_move_skb() 306 } else if (after64(MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq)) { in __mptcp_move_skb() 372 if (msk->ack_seq == rcv_data_fin_seq) { in mptcp_pending_data_fin() 417 WRITE_ONCE(msk->ack_seq, msk->ack_seq + 1); in mptcp_check_data_fin() 558 if (after64(MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq)) in mptcp_ofo_queue() 565 msk->ack_seq))) { in mptcp_ofo_queue() 574 int delta = msk->ack_seq - MPTCP_SKB_CB(skb)->map_seq; in mptcp_ofo_queue() 578 MPTCP_SKB_CB(skb)->map_seq, msk->ack_seq, in mptcp_ofo_queue() [all …]
|
D | subflow.c | 895 u64 ack_seq; in subflow_check_data_avail() local 908 subflow->map_seq = READ_ONCE(msk->ack_seq); in subflow_check_data_avail() 932 WRITE_ONCE(msk->ack_seq, subflow->map_seq); in subflow_check_data_avail() 936 old_ack = READ_ONCE(msk->ack_seq); in subflow_check_data_avail() 937 ack_seq = mptcp_subflow_get_mapped_dsn(subflow); in subflow_check_data_avail() 939 ack_seq); in subflow_check_data_avail() 940 if (ack_seq == old_ack) { in subflow_check_data_avail() 943 } else if (after64(ack_seq, old_ack)) { in subflow_check_data_avail() 951 mptcp_subflow_discard_data(ssk, skb, old_ack - ack_seq); in subflow_check_data_avail()
|
D | mptcp_diag.c | 144 info->mptcpi_rcv_nxt = READ_ONCE(msk->ack_seq); in mptcp_diag_get_info()
|
D | options.c | 526 opts->ext_copy.data_ack = READ_ONCE(msk->ack_seq); in mptcp_established_options_dss() 530 opts->ext_copy.data_ack32 = (uint32_t)READ_ONCE(msk->ack_seq); in mptcp_established_options_dss() 853 expand_ack(READ_ONCE(msk->ack_seq), data_fin_seq, use_64bit)); in mptcp_update_rcv_data_fin()
|
D | protocol.h | 203 u64 ack_seq; member
|
/net/ipv6/netfilter/ |
D | nf_reject_ipv6.c | 103 tcph->seq = oth->ack_seq; in nf_reject_ip6_tcphdr_put() 104 tcph->ack_seq = 0; in nf_reject_ip6_tcphdr_put() 107 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip6_tcphdr_put()
|
/net/ipv4/netfilter/ |
D | nf_reject_ipv4.c | 82 tcph->seq = oth->ack_seq; in nf_reject_ip_tcphdr_put() 84 tcph->ack_seq = htonl(ntohl(oth->seq) + oth->syn + oth->fin + in nf_reject_ip_tcphdr_put()
|
/net/ipv4/ |
D | tcp_input.c | 121 void (*cad)(struct sock *sk, u32 ack_seq)) in clean_acked_data_enable() argument 1254 if (before(start_seq_0, TCP_SKB_CB(ack_skb)->ack_seq)) { in tcp_check_dsack() 1834 if (before(TCP_SKB_CB(ack_skb)->ack_seq, prior_snd_una - tp->max_window)) in tcp_sacktag_write_queue() 1860 if ((TCP_SKB_CB(ack_skb)->ack_seq != tp->snd_una) && in tcp_sacktag_write_queue() 3503 const u32 ack, const u32 ack_seq, in tcp_may_update_window() argument 3507 after(ack_seq, tp->snd_wl1) || in tcp_may_update_window() 3508 (ack_seq == tp->snd_wl1 && nwin > tp->snd_wnd); in tcp_may_update_window() 3537 u32 ack_seq) in tcp_ack_update_window() argument 3546 if (tcp_may_update_window(tp, ack, ack_seq, nwin)) { in tcp_ack_update_window() 3548 tcp_update_wl(tp, ack_seq); in tcp_ack_update_window() [all …]
|
D | tcp_offload.c | 242 flush |= (__force int)(th->ack_seq ^ th2->ack_seq); in tcp_gro_receive()
|
D | tcp_ipv4.c | 699 rep.th.seq = th->ack_seq; in tcp_v4_send_reset() 702 rep.th.ack_seq = htonl(ntohl(th->seq) + th->syn + th->fin + in tcp_v4_send_reset() 865 rep.th.ack_seq = htonl(ack); in tcp_v4_send_ack() 1844 if (likely(!before(TCP_SKB_CB(skb)->ack_seq, TCP_SKB_CB(tail)->ack_seq))) { in tcp_add_backlog() 1845 TCP_SKB_CB(tail)->ack_seq = TCP_SKB_CB(skb)->ack_seq; in tcp_add_backlog() 1923 TCP_SKB_CB(skb)->ack_seq = ntohl(th->ack_seq); in tcp_v4_fill_cb()
|
D | syncookies.c | 329 __u32 cookie = ntohl(th->ack_seq) - 1; in cookie_v4_check()
|
D | tcp_minisocks.c | 703 (TCP_SKB_CB(skb)->ack_seq != in tcp_check_req()
|
D | tcp_output.c | 1330 th->ack_seq = htonl(rcv_nxt); in __tcp_transmit_skb() 3624 th->ack_seq = htonl(tcp_rsk(req)->rcv_nxt); in tcp_make_synack()
|
/net/rds/ |
D | ib_cm.c | 104 __be64 ack_seq = 0; in rds_ib_cm_connect_complete() local 122 ack_seq = get_unaligned(&dp->ricp_v6.dp_ack_seq); in rds_ib_cm_connect_complete() 129 ack_seq = get_unaligned(&dp->ricp_v4.dp_ack_seq); in rds_ib_cm_connect_complete() 188 if (ack_seq) in rds_ib_cm_connect_complete() 189 rds_send_drop_acked(conn, be64_to_cpu(ack_seq), in rds_ib_cm_connect_complete()
|
/net/ipv6/ |
D | tcp_ipv6.c | 922 t1->ack_seq = htonl(ack); in tcp_v6_send_response() 1005 u32 seq = 0, ack_seq = 0; in tcp_v6_send_reset() local 1075 seq = ntohl(th->ack_seq); in tcp_v6_send_reset() 1077 ack_seq = ntohl(th->seq) + th->syn + th->fin + skb->len - in tcp_v6_send_reset() 1099 tcp_v6_send_response(sk, skb, seq, ack_seq, 0, 0, 0, oif, key, 1, in tcp_v6_send_reset() 1591 TCP_SKB_CB(skb)->ack_seq = ntohl(th->ack_seq); in tcp_v6_fill_cb()
|
D | syncookies.c | 136 __u32 cookie = ntohl(th->ack_seq) - 1; in cookie_v6_check()
|
/net/sched/ |
D | sch_cake.c | 999 u32 ack_seq_a = ntohl(tcph_a->ack_seq); in cake_tcph_sack_compare() 1247 after(ntohl(tcph_check->ack_seq), ntohl(tcph->ack_seq))) in cake_ack_filter() 1259 (ntohl(tcph_check->ack_seq) == ntohl(tcph->ack_seq) && in cake_ack_filter()
|
/net/core/ |
D | filter.c | 6534 cookie = ntohl(th->ack_seq) - 1; in BPF_CALL_5()
|