Lines Matching +full:packet +full:- +full:oriented
1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * NET Generic infrastructure for INET connection oriented protocols.
60 /** inet_connection_sock - INET connection oriented sock
83 * @icsk_probes_tstamp: Probe timestamp (cleared by non-zero window ack)
130 __u32 lrcvtime; /* timestamp of last received data packet */
167 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca()
191 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack()
196 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled()
201 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init()
212 icsk->icsk_pending = 0; in inet_csk_clear_xmit_timer()
214 sk_stop_timer(sk, &icsk->icsk_retransmit_timer); in inet_csk_clear_xmit_timer()
217 icsk->icsk_ack.pending = 0; in inet_csk_clear_xmit_timer()
218 icsk->icsk_ack.retry = 0; in inet_csk_clear_xmit_timer()
220 sk_stop_timer(sk, &icsk->icsk_delack_timer); in inet_csk_clear_xmit_timer()
235 if (!icsk->nata_retries_enabled) in get_nata_rto()
248 if (icsk->nata_retries_type == NATA_STL) in get_nata_rto()
249 return sk->sk_state == TCP_SYN_SENT ? in get_nata_rto()
250 icsk->nata_syn_rto : icsk->nata_data_rto; in get_nata_rto()
255 when_nata = icsk->nata_data_rto; in get_nata_rto()
256 if (icsk->icsk_retransmits > icsk->nata_data_retries) { in get_nata_rto()
257 shift = icsk->icsk_retransmits - icsk->nata_data_retries; in get_nata_rto()
290 icsk->icsk_pending = what; in inet_csk_reset_xmit_timer()
291 icsk->icsk_timeout = jiffies + when; in inet_csk_reset_xmit_timer()
292 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in inet_csk_reset_xmit_timer()
294 icsk->icsk_ack.pending |= ICSK_ACK_TIMER; in inet_csk_reset_xmit_timer()
295 icsk->icsk_ack.timeout = jiffies + when; in inet_csk_reset_xmit_timer()
296 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in inet_csk_reset_xmit_timer()
306 u64 when = (u64)icsk->icsk_rto << icsk->icsk_backoff; in inet_csk_rto_backoff()
332 reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_added()
337 return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_len()
342 return inet_csk_reqsk_queue_len(sk) >= sk->sk_max_ack_backlog; in inet_csk_reqsk_queue_is_full()
352 this_cpu_inc(*sk->sk_prot->orphan_count); in inet_csk_prepare_for_destroy_sock()
363 return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ? in inet_csk_listen_poll()
382 inet_csk(sk)->icsk_ack.pingpong = TCP_PINGPONG_THRESH; in inet_csk_enter_pingpong_mode()
387 inet_csk(sk)->icsk_ack.pingpong = 0; in inet_csk_exit_pingpong_mode()
392 return inet_csk(sk)->icsk_ack.pingpong >= TCP_PINGPONG_THRESH; in inet_csk_in_pingpong_mode()
397 return inet_sk(sk)->is_icsk && !!inet_csk(sk)->icsk_ulp_ops; in inet_csk_has_ulp()
404 spin_lock_init(&icsk->icsk_accept_queue.rskq_lock); in inet_init_csk_locks()
405 spin_lock_init(&icsk->icsk_accept_queue.fastopenq.lock); in inet_init_csk_locks()