Lines Matching refs:sk
34 static void tcp_write_err(struct sock *sk) in tcp_write_err() argument
36 sk->sk_err = sk->sk_err_soft ? : ETIMEDOUT; in tcp_write_err()
37 sk->sk_error_report(sk); in tcp_write_err()
39 tcp_write_queue_purge(sk); in tcp_write_err()
40 tcp_done(sk); in tcp_write_err()
41 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err()
68 static int tcp_out_of_resources(struct sock *sk, bool do_reset) in tcp_out_of_resources() argument
70 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources()
79 if (sk->sk_err_soft) in tcp_out_of_resources()
82 if (tcp_check_oom(sk, shift)) { in tcp_out_of_resources()
90 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_out_of_resources()
91 tcp_done(sk); in tcp_out_of_resources()
92 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources()
96 if (!check_net(sock_net(sk))) { in tcp_out_of_resources()
98 tcp_done(sk); in tcp_out_of_resources()
110 static int tcp_orphan_retries(struct sock *sk, bool alive) in tcp_orphan_retries() argument
112 int retries = sock_net(sk)->ipv4.sysctl_tcp_orphan_retries; /* May be zero. */ in tcp_orphan_retries()
115 if (sk->sk_err_soft && !alive) in tcp_orphan_retries()
126 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk) in tcp_mtu_probing() argument
128 struct net *net = sock_net(sk); in tcp_mtu_probing()
135 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtu_probing()
137 struct net *net = sock_net(sk); in tcp_mtu_probing()
138 struct tcp_sock *tp = tcp_sk(sk); in tcp_mtu_probing()
141 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1; in tcp_mtu_probing()
145 icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss); in tcp_mtu_probing()
146 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtu_probing()
165 static bool retransmits_timed_out(struct sock *sk, in retransmits_timed_out() argument
172 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out()
175 start_ts = tcp_sk(sk)->retrans_stamp; in retransmits_timed_out()
177 start_ts = tcp_skb_timestamp(tcp_write_queue_head(sk)); in retransmits_timed_out()
188 return (tcp_time_stamp(tcp_sk(sk)) - start_ts) >= jiffies_to_msecs(timeout); in retransmits_timed_out()
192 static int tcp_write_timeout(struct sock *sk) in tcp_write_timeout() argument
194 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout()
195 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout()
196 struct net *net = sock_net(sk); in tcp_write_timeout()
200 if ((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV)) { in tcp_write_timeout()
202 dst_negative_advice(sk); in tcp_write_timeout()
204 tcp_fastopen_cache_set(sk, 0, NULL, true, 0); in tcp_write_timeout()
206 NET_INC_STATS(sock_net(sk), in tcp_write_timeout()
209 sk_rethink_txhash(sk); in tcp_write_timeout()
214 if (retransmits_timed_out(sk, net->ipv4.sysctl_tcp_retries1, 0)) { in tcp_write_timeout()
221 tcp_fastopen_cache_set(sk, 0, NULL, true, 0); in tcp_write_timeout()
223 NET_INC_STATS(sock_net(sk), in tcp_write_timeout()
227 tcp_mtu_probing(icsk, sk); in tcp_write_timeout()
229 dst_negative_advice(sk); in tcp_write_timeout()
231 sk_rethink_txhash(sk); in tcp_write_timeout()
235 if (sock_flag(sk, SOCK_DEAD)) { in tcp_write_timeout()
238 retry_until = tcp_orphan_retries(sk, alive); in tcp_write_timeout()
240 !retransmits_timed_out(sk, retry_until, 0); in tcp_write_timeout()
242 if (tcp_out_of_resources(sk, do_reset)) in tcp_write_timeout()
245 expired = retransmits_timed_out(sk, retry_until, in tcp_write_timeout()
250 tcp_write_err(sk); in tcp_write_timeout()
257 void tcp_delack_timer_handler(struct sock *sk) in tcp_delack_timer_handler() argument
259 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler()
261 sk_mem_reclaim_partial(sk); in tcp_delack_timer_handler()
263 if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) || in tcp_delack_timer_handler()
268 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in tcp_delack_timer_handler()
273 if (inet_csk_ack_scheduled(sk)) { in tcp_delack_timer_handler()
284 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_delack_timer_handler()
285 tcp_send_ack(sk); in tcp_delack_timer_handler()
286 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler()
290 if (tcp_under_memory_pressure(sk)) in tcp_delack_timer_handler()
291 sk_mem_reclaim(sk); in tcp_delack_timer_handler()
306 struct sock *sk = (struct sock *)data; in tcp_delack_timer() local
308 bh_lock_sock(sk); in tcp_delack_timer()
309 if (!sock_owned_by_user(sk)) { in tcp_delack_timer()
310 tcp_delack_timer_handler(sk); in tcp_delack_timer()
312 inet_csk(sk)->icsk_ack.blocked = 1; in tcp_delack_timer()
313 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer()
315 if (!test_and_set_bit(TCP_DELACK_TIMER_DEFERRED, &sk->sk_tsq_flags)) in tcp_delack_timer()
316 sock_hold(sk); in tcp_delack_timer()
318 bh_unlock_sock(sk); in tcp_delack_timer()
319 sock_put(sk); in tcp_delack_timer()
322 static void tcp_probe_timer(struct sock *sk) in tcp_probe_timer() argument
324 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer()
325 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer()
329 if (tp->packets_out || !tcp_send_head(sk)) { in tcp_probe_timer()
342 start_ts = tcp_skb_timestamp(tcp_send_head(sk)); in tcp_probe_timer()
344 tcp_send_head(sk)->skb_mstamp = tp->tcp_mstamp; in tcp_probe_timer()
350 max_probes = sock_net(sk)->ipv4.sysctl_tcp_retries2; in tcp_probe_timer()
351 if (sock_flag(sk, SOCK_DEAD)) { in tcp_probe_timer()
354 max_probes = tcp_orphan_retries(sk, alive); in tcp_probe_timer()
357 if (tcp_out_of_resources(sk, true)) in tcp_probe_timer()
362 abort: tcp_write_err(sk); in tcp_probe_timer()
365 tcp_send_probe0(sk); in tcp_probe_timer()
373 static void tcp_fastopen_synack_timer(struct sock *sk) in tcp_fastopen_synack_timer() argument
375 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer()
377 sock_net(sk)->ipv4.sysctl_tcp_synack_retries + 1; /* add one more retry for fastopen */ in tcp_fastopen_synack_timer()
380 req = tcp_sk(sk)->fastopen_rsk; in tcp_fastopen_synack_timer()
384 tcp_write_err(sk); in tcp_fastopen_synack_timer()
392 inet_rtx_syn_ack(sk, req); in tcp_fastopen_synack_timer()
395 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_fastopen_synack_timer()
411 void tcp_retransmit_timer(struct sock *sk) in tcp_retransmit_timer() argument
413 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer()
414 struct net *net = sock_net(sk); in tcp_retransmit_timer()
415 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer()
419 WARN_ON_ONCE(sk->sk_state != TCP_SYN_RECV && in tcp_retransmit_timer()
420 sk->sk_state != TCP_FIN_WAIT1); in tcp_retransmit_timer()
421 tcp_fastopen_synack_timer(sk); in tcp_retransmit_timer()
431 skb = tcp_rtx_queue_head(sk); in tcp_retransmit_timer()
437 if (!tp->snd_wnd && !sock_flag(sk, SOCK_DEAD) && in tcp_retransmit_timer()
438 !((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV))) { in tcp_retransmit_timer()
444 struct inet_sock *inet = inet_sk(sk); in tcp_retransmit_timer()
445 if (sk->sk_family == AF_INET) { in tcp_retransmit_timer()
453 else if (sk->sk_family == AF_INET6) { in tcp_retransmit_timer()
455 &sk->sk_v6_daddr, in tcp_retransmit_timer()
462 tcp_write_err(sk); in tcp_retransmit_timer()
465 tcp_enter_loss(sk); in tcp_retransmit_timer()
466 tcp_retransmit_skb(sk, skb, 1); in tcp_retransmit_timer()
467 __sk_dst_reset(sk); in tcp_retransmit_timer()
471 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTS); in tcp_retransmit_timer()
472 if (tcp_write_timeout(sk)) in tcp_retransmit_timer()
493 __NET_INC_STATS(sock_net(sk), mib_idx); in tcp_retransmit_timer()
496 tcp_enter_loss(sk); in tcp_retransmit_timer()
498 if (tcp_retransmit_skb(sk, tcp_write_queue_head(sk), 1) > 0) { in tcp_retransmit_timer()
504 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_retransmit_timer()
538 if (sk->sk_state == TCP_ESTABLISHED && in tcp_retransmit_timer()
548 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, icsk->icsk_rto, TCP_RTO_MAX); in tcp_retransmit_timer()
549 if (retransmits_timed_out(sk, net->ipv4.sysctl_tcp_retries1 + 1, 0)) in tcp_retransmit_timer()
550 __sk_dst_reset(sk); in tcp_retransmit_timer()
557 void tcp_write_timer_handler(struct sock *sk) in tcp_write_timer_handler() argument
559 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timer_handler()
562 if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) || in tcp_write_timer_handler()
567 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in tcp_write_timer_handler()
571 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_write_timer_handler()
576 tcp_rack_reo_timeout(sk); in tcp_write_timer_handler()
579 tcp_send_loss_probe(sk); in tcp_write_timer_handler()
583 tcp_retransmit_timer(sk); in tcp_write_timer_handler()
587 tcp_probe_timer(sk); in tcp_write_timer_handler()
592 sk_mem_reclaim(sk); in tcp_write_timer_handler()
597 struct sock *sk = (struct sock *)data; in tcp_write_timer() local
599 bh_lock_sock(sk); in tcp_write_timer()
600 if (!sock_owned_by_user(sk)) { in tcp_write_timer()
601 tcp_write_timer_handler(sk); in tcp_write_timer()
604 if (!test_and_set_bit(TCP_WRITE_TIMER_DEFERRED, &sk->sk_tsq_flags)) in tcp_write_timer()
605 sock_hold(sk); in tcp_write_timer()
607 bh_unlock_sock(sk); in tcp_write_timer()
608 sock_put(sk); in tcp_write_timer()
619 void tcp_set_keepalive(struct sock *sk, int val) in tcp_set_keepalive() argument
621 if ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) in tcp_set_keepalive()
624 if (val && !sock_flag(sk, SOCK_KEEPOPEN)) in tcp_set_keepalive()
625 inet_csk_reset_keepalive_timer(sk, keepalive_time_when(tcp_sk(sk))); in tcp_set_keepalive()
627 inet_csk_delete_keepalive_timer(sk); in tcp_set_keepalive()
634 struct sock *sk = (struct sock *) data; in tcp_keepalive_timer() local
635 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_keepalive_timer()
636 struct tcp_sock *tp = tcp_sk(sk); in tcp_keepalive_timer()
640 bh_lock_sock(sk); in tcp_keepalive_timer()
641 if (sock_owned_by_user(sk)) { in tcp_keepalive_timer()
643 inet_csk_reset_keepalive_timer (sk, HZ/20); in tcp_keepalive_timer()
647 if (sk->sk_state == TCP_LISTEN) { in tcp_keepalive_timer()
653 if (sk->sk_state == TCP_FIN_WAIT2 && sock_flag(sk, SOCK_DEAD)) { in tcp_keepalive_timer()
655 const int tmo = tcp_fin_time(sk) - TCP_TIMEWAIT_LEN; in tcp_keepalive_timer()
658 tcp_time_wait(sk, TCP_FIN_WAIT2, tmo); in tcp_keepalive_timer()
662 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_keepalive_timer()
666 if (!sock_flag(sk, SOCK_KEEPOPEN) || in tcp_keepalive_timer()
667 ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_SYN_SENT))) in tcp_keepalive_timer()
673 if (tp->packets_out || tcp_send_head(sk)) in tcp_keepalive_timer()
687 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_keepalive_timer()
688 tcp_write_err(sk); in tcp_keepalive_timer()
691 if (tcp_write_wakeup(sk, LINUX_MIB_TCPKEEPALIVE) <= 0) { in tcp_keepalive_timer()
705 sk_mem_reclaim(sk); in tcp_keepalive_timer()
708 inet_csk_reset_keepalive_timer (sk, elapsed); in tcp_keepalive_timer()
712 tcp_done(sk); in tcp_keepalive_timer()
715 bh_unlock_sock(sk); in tcp_keepalive_timer()
716 sock_put(sk); in tcp_keepalive_timer()
719 void tcp_init_xmit_timers(struct sock *sk) in tcp_init_xmit_timers() argument
721 inet_csk_init_xmit_timers(sk, &tcp_write_timer, &tcp_delack_timer, in tcp_init_xmit_timers()
723 hrtimer_init(&tcp_sk(sk)->pacing_timer, CLOCK_MONOTONIC, in tcp_init_xmit_timers()
725 tcp_sk(sk)->pacing_timer.function = tcp_pace_kick; in tcp_init_xmit_timers()