/kernel/linux/linux-5.10/samples/bpf/ |
D | tcp_dumpstats_kern.c | 28 struct bpf_tcp_sock *tcp_sk; in _sockops() local 56 tcp_sk = bpf_tcp_sock(sk); in _sockops() 57 if (!tcp_sk) in _sockops() 63 tcp_sk->dsack_dups, tcp_sk->delivered); in _sockops() 65 tcp_sk->delivered_ce, tcp_sk->icsk_retransmits); in _sockops()
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/progs/ |
D | tcp_rtt.c | 27 struct bpf_tcp_sock *tcp_sk; in _sockops() local 48 tcp_sk = bpf_tcp_sock(sk); in _sockops() 49 if (!tcp_sk) in _sockops() 54 storage->dsack_dups = tcp_sk->dsack_dups; in _sockops() 55 storage->delivered = tcp_sk->delivered; in _sockops() 56 storage->delivered_ce = tcp_sk->delivered_ce; in _sockops() 57 storage->icsk_retransmits = tcp_sk->icsk_retransmits; in _sockops()
|
D | bpf_dctcp.c | 56 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 77 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 86 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG() 116 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss() 135 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ece_ack_cwr() 165 *prior_rcv_nxt = tcp_sk(sk)->rcv_nxt; in dctcp_ece_ack_update() 194 return max(tcp_sk(sk)->snd_cwnd, ca->loss_cwnd); in BPF_PROG() 200 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
|
D | bpf_cubic.c | 158 return tcp_sk(sk)->tcp_mstamp; in bictcp_clock_us() 163 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 188 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in BPF_PROG() 202 delta = now - tcp_sk(sk)->lsndtime; in BPF_PROG() 389 struct tcp_sock *tp = tcp_sk(sk); in BPF_STRUCT_OPS() 408 const struct tcp_sock *tp = tcp_sk(sk); in BPF_STRUCT_OPS() 455 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 502 const struct tcp_sock *tp = tcp_sk(sk); in BPF_STRUCT_OPS() 530 const struct tcp_sock *tp = tcp_sk(sk); in BPF_STRUCT_OPS()
|
/kernel/linux/linux-5.10/net/ipv4/ |
D | tcp_timer.c | 32 start_ts = tcp_sk(sk)->retrans_stamp; in tcp_clamp_rto_to_user_timeout() 35 elapsed = tcp_time_stamp(tcp_sk(sk)) - start_ts; in tcp_clamp_rto_to_user_timeout() 104 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources() 218 start_ts = tcp_sk(sk)->retrans_stamp; in retransmits_timed_out() 227 return (s32)(tcp_time_stamp(tcp_sk(sk)) - start_ts - timeout) >= 0; in retransmits_timed_out() 234 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout() 316 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_delack_timer_handler() 359 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer() 411 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer() 450 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer() [all …]
|
D | tcp_bbr.c | 244 unsigned int mss = tcp_sk(sk)->mss_cache; in bbr_rate_bytes_per_sec() 266 struct tcp_sock *tp = tcp_sk(sk); in bbr_init_pacing_rate_from_rtt() 285 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_pacing_rate() 303 struct tcp_sock *tp = tcp_sk(sk); in bbr_tso_segs_goal() 320 struct tcp_sock *tp = tcp_sk(sk); in bbr_save_cwnd() 331 struct tcp_sock *tp = tcp_sk(sk); in bbr_cwnd_event() 436 struct tcp_sock *tp = tcp_sk(sk); in bbr_packets_in_net_at_edt() 480 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd_to_recover_or_restore() 519 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd() 554 struct tcp_sock *tp = tcp_sk(sk); in bbr_is_next_cycle_phase() [all …]
|
D | tcp_dctcp.c | 77 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_init() 105 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ssthresh() 113 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_update_alpha() 147 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss() 185 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_get_info() 213 return max(tcp_sk(sk)->snd_cwnd, ca->loss_cwnd); in dctcp_cwnd_undo()
|
D | tcp_recovery.c | 12 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd() 65 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss() 102 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost() 156 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout() 192 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd() 223 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost()
|
D | tcp_input.c | 144 bool unknown_opt = tcp_sk(sk)->rx_opt.saw_unknown && in bpf_skops_parse_hdr() 145 BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr() 147 bool parse_all_opt = BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr() 240 tcp_sk(sk)->advmss); in tcp_measure_rcv_mss() 264 len -= tcp_sk(sk)->tcp_header_len; in tcp_measure_rcv_mss() 280 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack() 321 tcp_sk(sk)->ecn_flags &= ~TCP_ECN_DEMAND_CWR; in tcp_ecn_accept_cwr() 339 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce() 371 if (tcp_sk(sk)->ecn_flags & TCP_ECN_OK) in tcp_ecn_check_ce() 401 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand() [all …]
|
D | tcp_westwood.c | 73 w->snd_una = tcp_sk(sk)->snd_una; in tcp_westwood_init() 127 w->snd_una = tcp_sk(sk)->snd_una; in westwood_update_window() 165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw() 182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count() 219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin() 242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
|
D | tcp_output.c | 68 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_new_data_sent() 97 const struct tcp_sock *tp = tcp_sk(sk); in tcp_acceptable_seq() 123 struct tcp_sock *tp = tcp_sk(sk); in tcp_advertise_mss() 144 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_restart() 183 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_ack_sent() 259 struct tcp_sock *tp = tcp_sk(sk); in tcp_select_window() 309 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_synack() 322 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_syn() 366 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send() 478 if (likely(!BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_hdr_opt_len() [all …]
|
D | tcp_cubic.c | 122 return tcp_sk(sk)->tcp_mstamp; in bictcp_clock_us() 127 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 146 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in bictcp_init() 156 delta = now - tcp_sk(sk)->lsndtime; in bictcp_cwnd_event() 333 struct tcp_sock *tp = tcp_sk(sk); in bictcp_cong_avoid() 350 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_recalc_ssthresh() 395 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 457 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_acked()
|
D | tcp_rate.c | 42 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_sent() 82 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_delivered() 118 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_gen() 190 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited()
|
D | tcp_highspeed.c | 102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init() 114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid() 153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
|
D | tcp_cdg.c | 143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update() 244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff() 265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid() 302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked() 331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh() 348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event() 376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
|
D | tcp_fastopen.c | 188 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_add_skb() 277 tp = tcp_sk(child); in tcp_fastopen_create_child() 353 tcp_sk(sk)->fastopen_no_cookie || in tcp_fastopen_no_cookie() 449 tcp_sk(sk)->fastopen_client_fail = TFO_COOKIE_UNAVAILABLE; in tcp_fastopen_cookie_check() 463 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_defer_connect() 558 tcp_sk(sk)->syn_fastopen_ch = 1; in tcp_fastopen_active_should_disable() 570 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_disable_ofo_check() 597 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_detect_blackhole()
|
D | tcp_dctcp.h | 6 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ece_ack_cwr() 35 *prior_rcv_nxt = tcp_sk(sk)->rcv_nxt; in dctcp_ece_ack_update()
|
D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid() 36 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
|
D | tcp_hybla.c | 39 tcp_sk(sk)->srtt_us / (rtt0 * USEC_PER_MSEC), in hybla_recalc_param() 48 struct tcp_sock *tp = tcp_sk(sk); in hybla_init() 92 struct tcp_sock *tp = tcp_sk(sk); in hybla_cong_avoid()
|
D | tcp_illinois.c | 59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset() 224 struct tcp_sock *tp = tcp_sk(sk); in update_params() 262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid() 297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
|
/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
D | tcp_nip_output.c | 84 const struct tcp_sock *tp = tcp_sk(sk); in __tcp_nip_mtu_to_mss() 118 (tcp_sk(sk)->tcp_header_len - sizeof(struct tcphdr)); in tcp_nip_mtu_to_mss() 132 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_event_new_data_sent() 148 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) { in tcp_nip_check_probe_timer() 155 tcp_sk(sk)->packets_out, inet_csk(sk)->icsk_pending); in tcp_nip_check_probe_timer() 172 struct tcp_sock *tp = tcp_sk(sk); in __nip_tcp_select_window() 232 struct tcp_sock *tp = tcp_sk(sk); in nip_tcp_select_window() 292 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_connect_init() 366 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_connect_queue_skb() 380 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_advertise_mss() [all …]
|
D | tcp_nip_timer.c | 68 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_nip_delack_timer_handler() 149 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_retransmit_timer() 219 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_probe_timer() 277 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_nip_write_timer_handler() 315 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_keepalive_is_timeout() 343 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_keepalive_timer()
|
D | tcp_nip_input.c | 311 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_ofo_queue() 363 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_data_queue_ofo() 537 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_data_queue() 548 tp->snd_up = (TCP_SKB_CB(skb)->seq - tcp_sk(sk)->rcv_nxt) / mss; in tcp_nip_data_queue() 631 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_push_pending_frames() 689 struct tcp_sock *tp = tcp_sk(sk); in __tcp_nip_ack_snd_check() 740 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_rearm_rto() 789 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_clean_rtx_queue() 1000 struct tcp_sock *newtp = tcp_sk(newsk); in tcp_nip_create_openreq_child() 1104 const struct tcp_sock *tp = tcp_sk(sk_listener); in tcp_nip_openreq_init_rwin() [all …]
|
D | tcp_nip.c | 631 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_connect() 890 newtp = tcp_sk(newsk); in tcp_nip_syn_recv_sock() 906 if (tcp_sk(sk)->rx_opt.user_mss && in tcp_nip_syn_recv_sock() 907 tcp_sk(sk)->rx_opt.user_mss < newtp->advmss) in tcp_nip_syn_recv_sock() 908 newtp->advmss = tcp_sk(sk)->rx_opt.user_mss; in tcp_nip_syn_recv_sock() 965 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_keepalive_para_update() 1020 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_keepalive_enable() 1075 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_keepalive_disable() 1119 struct tcp_sock *tp = tcp_sk(sk); in _tcp_sock_priv_init() 1148 struct tcp_sock *tp = tcp_sk(sk); in tcp_nip_init_sock() [all …]
|
/kernel/linux/linux-5.10/include/net/ |
D | tcp.h | 502 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_overflow() 504 WRITE_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp, now); in tcp_synq_overflow() 525 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_no_recent_overflow() 610 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers() 613 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers() 688 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check() 1232 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh() 1277 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited() 1305 s64 delay = tcp_sk(sk)->tcp_wstamp_ns - tcp_sk(sk)->tcp_clock_cache; in tcp_pacing_delay() 1341 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) in tcp_check_probe_timer() [all …]
|