/kernel/linux/linux-5.10/net/ipv4/ |
D | tcp_westwood.c | 191 w->accounted += tp->mss_cache; in westwood_acked_count() 192 w->cumul_ack = tp->mss_cache; in westwood_acked_count() 195 if (w->cumul_ack > tp->mss_cache) { in westwood_acked_count() 199 w->cumul_ack = tp->mss_cache; in westwood_acked_count() 222 return max_t(u32, (w->bw_est * w->rtt_min) / tp->mss_cache, 2); in tcp_westwood_bw_rttmin()
|
D | tcp_nv.c | 350 bytes_acked >= (ca->nv_min_cwnd - 1) * tp->mss_cache && in tcpnv_acked() 356 ca->nv_min_cwnd * tp->mss_cache; in tcpnv_acked() 368 80000ULL * tp->mss_cache); in tcpnv_acked()
|
D | tcp_dctcp.c | 197 info->dctcp.dctcp_ab_ecn = tp->mss_cache * in dctcp_get_info() 199 info->dctcp.dctcp_ab_tot = tp->mss_cache * in dctcp_get_info()
|
D | tcp_rate.c | 193 tp->write_seq - tp->snd_nxt < tp->mss_cache && in tcp_rate_check_app_limited()
|
D | tcp_bbr.c | 244 unsigned int mss = tcp_sk(sk)->mss_cache; in bbr_rate_bytes_per_sec() 312 segs = max_t(u32, bytes / tp->mss_cache, bbr_min_tso_segs(sk)); in bbr_tso_segs_goal() 1113 bw = bw * tp->mss_cache * USEC_PER_SEC >> BW_SCALE; in bbr_get_info()
|
D | tcp_output.c | 1812 tp->mss_cache = mss_now; in tcp_sync_mss() 1830 mss_now = tp->mss_cache; in tcp_current_mss() 2205 cong_win = (tp->snd_cwnd - in_flight) * tp->mss_cache; in tcp_tso_should_defer() 2210 if (limit >= max_segs * tp->mss_cache) in tcp_tso_should_defer() 2219 u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); in tcp_tso_should_defer() 2233 if (limit > tcp_max_tso_deferred_mss(tp) * tp->mss_cache) in tcp_tso_should_defer() 2358 size_needed = probe_size + (tp->reordering + 1) * tp->mss_cache; in tcp_mtu_probe()
|
D | tcp_input.c | 409 per_mss = max_t(u32, tp->rx_opt.mss_clamp, tp->mss_cache) + in tcp_sndbuf_expand() 594 unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache); in tcp_initialize_rcv_mss() 894 rate = (u64)tp->mss_cache * ((USEC_PER_SEC / 100) << 3); in tcp_update_pacing_rate() 993 if (seq_len > tp->mss_cache) in tcp_dsack_seen() 994 dup_segs = DIV_ROUND_UP(seq_len, tp->mss_cache); in tcp_dsack_seen() 1019 const u32 mss = tp->mss_cache; in tcp_check_sack_reordering() 2713 val = (u64)tp->snd_cwnd * tcp_mss_to_mtu(sk, tp->mss_cache); in tcp_mtup_probe_success() 2854 tp->snd_una + tp->reordering * tp->mss_cache); in tcp_force_fast_retransmit() 3324 if (pkts_acked == 1 && last_in_flight < tp->mss_cache && in tcp_clean_rtx_queue()
|
D | tcp.c | 398 rate64 = (u64)rate * tp->mss_cache * USEC_PER_SEC; in tcp_compute_delivery_rate() 441 tp->mss_cache = TCP_MSS_DEFAULT; in tcp_init_sock() 3542 info->tcpi_snd_mss = tp->mss_cache; in tcp_get_info() 3707 val = tp->mss_cache; in do_tcp_getsockopt()
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/progs/ |
D | test_sock_fields.c | 95 dst->mss_cache = src->mss_cache; in tpcpy()
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/siw/ |
D | siw_qp_tx.c | 677 c_tx->tcp_seglen = tp->mss_cache * tp->gso_segs; in siw_update_tcpseg() 680 tp->mss_cache * in siw_update_tcpseg() 683 c_tx->tcp_seglen = tp->mss_cache; in siw_update_tcpseg()
|
/kernel/linux/linux-5.10/include/linux/ |
D | tcp.h | 207 u32 mss_cache; /* Cached effective mss, not including SACKS */ member
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/prog_tests/ |
D | sock_fields.c | 94 tp->rcv_nxt, tp->snd_nxt, tp->snd_una, tp->mss_cache, in print_tp()
|
/kernel/linux/linux-5.10/include/uapi/linux/ |
D | bpf.h | 4199 __u32 mss_cache; /* Cached effective mss, not including SACKS */ member 4519 __u32 mss_cache; member
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/linux/ |
D | bpf.h | 647 __u32 mss_cache; member 871 __u32 mss_cache; member
|
/kernel/linux/linux-5.10/tools/include/uapi/linux/ |
D | bpf.h | 4199 __u32 mss_cache; /* Cached effective mss, not including SACKS */ member 4519 __u32 mss_cache; member
|
/kernel/linux/patches/linux-4.19/prebuilts/usr/include/linux/ |
D | bpf.h | 606 __u32 mss_cache; member
|
/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
D | tcp_nip_output.c | 678 tp->mss_cache = mss_now; in tcp_nip_sync_mss() 693 mss_now = tp->mss_cache; in tcp_nip_current_mss()
|
D | tcp_nip.c | 949 newtp->mss_cache = TCP_BASE_MSS; in tcp_nip_syn_recv_sock() 1386 tp->mss_cache = TCP_MSS_DEFAULT; in tcp_nip_init_sock()
|
D | tcp_nip_input.c | 2180 unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache); in tcp_nip_initialize_rcv_mss()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_io.c | 192 FLOWC_PARAM(MSS, tp->mss_cache); in send_tx_flowc_wr()
|
D | chtls_cm.c | 153 tp->mss_cache = tp->rx_opt.mss_clamp; in assign_rxopt()
|
/kernel/linux/linux-5.10/net/ipv6/ |
D | tcp_ipv6.c | 365 if (tcp_mtu_to_mss(sk, mtu) >= tcp_sk(sk)->mss_cache) in tcp_v6_mtu_reduced()
|
/kernel/linux/linux-5.10/net/core/ |
D | filter.c | 6344 case offsetof(struct bpf_tcp_sock, mss_cache): in bpf_tcp_sock_convert_ctx_access() 6345 BPF_TCP_SOCK_GET_COMMON(mss_cache); in bpf_tcp_sock_convert_ctx_access() 9440 case offsetof(struct bpf_sock_ops, mss_cache): in sock_ops_convert_ctx_access() 9441 SOCK_OPS_GET_TCP_SOCK_FIELD(mss_cache); in sock_ops_convert_ctx_access()
|
/kernel/linux/patches/linux-5.10/hispark_taurus_patch/ |
D | hispark_taurus.patch | 45332 + * if tcp_mtu_probe() use (2 * tp->mss_cache) as probe_size, 48557 + * if tcp_mtu_probe() use (2 * tp->mss_cache) as probe_size,
|
/kernel/linux/patches/linux-4.19/hispark_taurus_patch/ |
D | hispark_taurus.patch | 338304 + * if tcp_mtu_probe() use (2 * tp->mss_cache) as probe_size, 341522 + * if tcp_mtu_probe() use (2 * tp->mss_cache) as probe_size,
|