/net/tipc/ |
D | link.c | 197 u16 rcv_nxt; member 428 if (more(l->snd_nxt, l->rcv_nxt)) in link_bc_rcv_gap() 429 gap = l->snd_nxt - l->rcv_nxt; in link_bc_rcv_gap() 431 gap = buf_seqno(skb) - l->rcv_nxt; in link_bc_rcv_gap() 456 return l->rcv_nxt; in tipc_link_rcv_nxt() 1012 l->rcv_nxt = 1; in tipc_link_reset() 1042 u16 bc_ack = l->bc_rcvlink->rcv_nxt - 1; in tipc_link_xmit() 1043 u16 ack = l->rcv_nxt - 1; in tipc_link_xmit() 1173 u16 bc_ack = l->bc_rcvlink->rcv_nxt - 1; in tipc_link_advance_backlog() 1176 u16 ack = l->rcv_nxt - 1; in tipc_link_advance_backlog() [all …]
|
D | name_distr.c | 322 u16 *rcv_nxt, bool *open) in tipc_named_dequeue() argument 338 *rcv_nxt = seqno; in tipc_named_dequeue() 348 if (*open && (*rcv_nxt == seqno)) { in tipc_named_dequeue() 349 (*rcv_nxt)++; in tipc_named_dequeue() 355 if (less(seqno, *rcv_nxt)) { in tipc_named_dequeue() 373 u16 *rcv_nxt, bool *open) in tipc_named_rcv() argument 382 while ((skb = tipc_named_dequeue(namedq, rcv_nxt, open))) { in tipc_named_rcv()
|
D | name_distr.h | 75 u16 *rcv_nxt, bool *open);
|
D | node.c | 1940 u16 rcv_nxt, syncpt, dlv_nxt, inputq_len; in tipc_node_check_state() local 1953 rcv_nxt = tipc_link_rcv_nxt(l); in tipc_node_check_state() 1993 if ((usr != LINK_PROTOCOL) && less(oseqno, rcv_nxt)) in tipc_node_check_state() 2023 if (!more(rcv_nxt, n->sync_point)) in tipc_node_check_state()
|
/net/ipv4/ |
D | tcp_fastopen.c | 179 if (TCP_SKB_CB(skb)->end_seq == tp->rcv_nxt) in tcp_fastopen_add_skb() 202 tp->rcv_nxt = TCP_SKB_CB(skb)->end_seq; in tcp_fastopen_add_skb() 289 tp->rcv_nxt = TCP_SKB_CB(skb)->seq + 1; in tcp_fastopen_create_child() 293 tcp_rsk(req)->rcv_nxt = tp->rcv_nxt; in tcp_fastopen_create_child() 294 tp->rcv_wup = tp->rcv_nxt; in tcp_fastopen_create_child()
|
D | tcp_input.c | 667 if (before(tp->rcv_nxt, tp->rcv_rtt_est.seq)) in tcp_rcv_rtt_measure() 675 tp->rcv_rtt_est.seq = tp->rcv_nxt + tp->rcv_wnd; in tcp_rcv_rtt_measure() 3548 u32 delta = seq - tp->rcv_nxt; in tcp_rcv_nxt_update() 3552 WRITE_ONCE(tp->rcv_nxt, seq); in tcp_rcv_nxt_update() 4286 (th->ack && seq == TCP_SKB_CB(skb)->end_seq && seq == tp->rcv_nxt) && in tcp_disordered_ack() 4323 !after(seq, tp->rcv_nxt + tcp_receive_window(tp)); in tcp_sequence() 4463 if (before(seq, tp->rcv_nxt)) in tcp_dsack_set() 4503 before(TCP_SKB_CB(skb)->seq, tp->rcv_nxt)) { in tcp_send_dupack() 4511 if (after(TCP_SKB_CB(skb)->end_seq, tp->rcv_nxt)) in tcp_send_dupack() 4512 end_seq = tp->rcv_nxt; in tcp_send_dupack() [all …]
|
D | tcp_minisocks.c | 271 tcptw->tw_rcv_nxt = tp->rcv_nxt; in tcp_time_wait() 482 WRITE_ONCE(newtp->rcv_nxt, seq); in tcp_create_openreq_child() 715 tcp_rsk(req)->rcv_nxt, tcp_rsk(req)->rcv_nxt + req->rsk_rcv_wnd)) { in tcp_check_req() 732 if (tmp_opt.saw_tstamp && !after(TCP_SKB_CB(skb)->seq, tcp_rsk(req)->rcv_nxt)) in tcp_check_req()
|
D | tcp_dctcp.h | 35 *prior_rcv_nxt = tcp_sk(sk)->rcv_nxt; in dctcp_ece_ack_update()
|
D | bpf_tcp_ca.c | 152 BPF_CALL_2(bpf_tcp_send_ack, struct tcp_sock *, tp, u32, rcv_nxt) in BPF_CALL_2() argument 155 __tcp_send_ack((struct sock *)tp, rcv_nxt); in BPF_CALL_2()
|
D | tcp_output.c | 180 static inline void tcp_event_ack_sent(struct sock *sk, u32 rcv_nxt) in tcp_event_ack_sent() argument 192 if (unlikely(rcv_nxt != tp->rcv_nxt)) in tcp_event_ack_sent() 278 tp->rcv_wup = tp->rcv_nxt; in tcp_select_window() 1236 int clone_it, gfp_t gfp_mask, u32 rcv_nxt) in __tcp_transmit_skb() argument 1328 th->ack_seq = htonl(rcv_nxt); in __tcp_transmit_skb() 1376 tcp_event_ack_sent(sk, rcv_nxt); in __tcp_transmit_skb() 1421 tcp_sk(sk)->rcv_nxt); in tcp_transmit_skb() 3625 th->ack_seq = htonl(tcp_rsk(req)->rcv_nxt); in tcp_make_synack() 3736 tp->rcv_nxt = 0; in tcp_connect_init() 3739 tp->rcv_wup = tp->rcv_nxt; in tcp_connect_init() [all …]
|
D | tcp.c | 1552 tp->copied_seq, TCP_SKB_CB(skb)->end_seq, tp->rcv_nxt); in tcp_cleanup_rbuf() 1558 tp->rcv_nxt - tp->rcv_wup > icsk->icsk_ack.rcv_mss || in tcp_cleanup_rbuf() 2280 u32 rcv_nxt = READ_ONCE(tp->rcv_nxt); in tcp_inq_hint() local 2283 inq = rcv_nxt - copied_seq; in tcp_inq_hint() 2286 inq = tp->rcv_nxt - tp->copied_seq; in tcp_inq_hint() 2379 *seq, TCP_SKB_CB(skb)->seq, tp->rcv_nxt, in tcp_recvmsg_locked() 2394 *seq, TCP_SKB_CB(skb)->seq, tp->rcv_nxt, flags); in tcp_recvmsg_locked() 3004 WRITE_ONCE(tp->copied_seq, tp->rcv_nxt); in tcp_disconnect() 3130 if (after(opt.snd_wl1, tp->rcv_nxt + opt.rcv_wnd)) in tcp_repair_set_window() 3133 if (after(opt.rcv_wup, tp->rcv_nxt)) in tcp_repair_set_window() [all …]
|
D | tcp_dctcp.c | 84 ca->prior_rcv_nxt = tp->rcv_nxt; in dctcp_init()
|
D | tcp_diag.c | 29 r->idiag_rqueue = max_t(int, READ_ONCE(tp->rcv_nxt) - in tcp_diag_get_info()
|
D | tcp_ipv4.c | 970 tcp_rsk(req)->rcv_nxt, in tcp_v4_reqsk_send_ack() 2659 rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) - in get_tcp4_sock()
|
/net/ipv6/ |
D | tcp_ipv6.c | 1171 tcp_rsk(req)->rcv_nxt, in tcp_v6_reqsk_send_ack() 1580 if (TCP_SKB_CB(opt_skb)->end_seq == tp->rcv_nxt && in tcp_v6_do_rcv() 2050 rx_queue = max_t(int, READ_ONCE(tp->rcv_nxt) - in get_tcp6_sock()
|
/net/core/ |
D | filter.c | 6652 case offsetof(struct bpf_tcp_sock, rcv_nxt): in bpf_tcp_sock_convert_ctx_access() 6653 BPF_TCP_SOCK_GET_COMMON(rcv_nxt); in bpf_tcp_sock_convert_ctx_access() 9780 case offsetof(struct bpf_sock_ops, rcv_nxt): in sock_ops_convert_ctx_access() 9781 SOCK_OPS_GET_TCP_SOCK_FIELD(rcv_nxt); in sock_ops_convert_ctx_access()
|
/net/mptcp/ |
D | protocol.c | 480 ((READ_ONCE(tp->rcv_nxt) - READ_ONCE(tp->rcv_wup) > in mptcp_subflow_could_cleanup()
|