Searched refs:tcp_sk (Results 1 – 5 of 5) sorted by relevance
510 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_overflow()512 WRITE_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp, now); in tcp_synq_overflow()533 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_no_recent_overflow()618 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers()621 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers()693 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check()1222 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh()1267 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited()1296 s64 pacing_delay = skb ? skb->tstamp : tcp_sk(sk)->tcp_wstamp_ns; in tcp_pacing_delay()1298 pacing_delay -= tcp_sk(sk)->tcp_clock_cache; in tcp_pacing_delay()[all …]
93 struct sock *tcp_sk; member
74 struct sock * __percpu *tcp_sk; member
429 static inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_sk() function460 rcu_access_pointer(tcp_sk(sk)->fastopen_rsk) != NULL; in tcp_passive_fastopen()
256 const struct tcp_sock *tp = tcp_sk(sk);