Lines Matching refs:tp
12 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd() local
14 if (!tp->reord_seen) { in tcp_rack_reo_wnd()
21 if (tp->sacked_out >= tp->reordering && in tcp_rack_reo_wnd()
33 return min((tcp_min_rtt(tp) >> 2) * tp->rack.reo_wnd_steps, in tcp_rack_reo_wnd()
34 tp->srtt_us >> 3); in tcp_rack_reo_wnd()
37 s32 tcp_rack_skb_timeout(struct tcp_sock *tp, struct sk_buff *skb, u32 reo_wnd) in tcp_rack_skb_timeout() argument
39 return tp->rack.rtt_us + reo_wnd - in tcp_rack_skb_timeout()
40 tcp_stamp_us_delta(tp->tcp_mstamp, tcp_skb_timestamp_us(skb)); in tcp_rack_skb_timeout()
65 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss() local
71 list_for_each_entry_safe(skb, n, &tp->tsorted_sent_queue, in tcp_rack_detect_loss()
81 if (!tcp_rack_sent_after(tp->rack.mstamp, in tcp_rack_detect_loss()
83 tp->rack.end_seq, scb->end_seq)) in tcp_rack_detect_loss()
89 remaining = tcp_rack_skb_timeout(tp, skb, reo_wnd); in tcp_rack_detect_loss()
102 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost() local
105 if (!tp->rack.advanced) in tcp_rack_mark_lost()
109 tp->rack.advanced = 0; in tcp_rack_mark_lost()
123 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance() argument
128 rtt_us = tcp_stamp_us_delta(tp->tcp_mstamp, xmit_time); in tcp_rack_advance()
129 if (rtt_us < tcp_min_rtt(tp) && (sacked & TCPCB_RETRANS)) { in tcp_rack_advance()
142 tp->rack.advanced = 1; in tcp_rack_advance()
143 tp->rack.rtt_us = rtt_us; in tcp_rack_advance()
144 if (tcp_rack_sent_after(xmit_time, tp->rack.mstamp, in tcp_rack_advance()
145 end_seq, tp->rack.end_seq)) { in tcp_rack_advance()
146 tp->rack.mstamp = xmit_time; in tcp_rack_advance()
147 tp->rack.end_seq = end_seq; in tcp_rack_advance()
156 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout() local
159 prior_inflight = tcp_packets_in_flight(tp); in tcp_rack_reo_timeout()
161 if (prior_inflight != tcp_packets_in_flight(tp)) { in tcp_rack_reo_timeout()
192 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd() local
200 if (before(rs->prior_delivered, tp->rack.last_delivered)) in tcp_rack_update_reo_wnd()
201 tp->rack.dsack_seen = 0; in tcp_rack_update_reo_wnd()
204 if (tp->rack.dsack_seen) { in tcp_rack_update_reo_wnd()
205 tp->rack.reo_wnd_steps = min_t(u32, 0xFF, in tcp_rack_update_reo_wnd()
206 tp->rack.reo_wnd_steps + 1); in tcp_rack_update_reo_wnd()
207 tp->rack.dsack_seen = 0; in tcp_rack_update_reo_wnd()
208 tp->rack.last_delivered = tp->delivered; in tcp_rack_update_reo_wnd()
209 tp->rack.reo_wnd_persist = TCP_RACK_RECOVERY_THRESH; in tcp_rack_update_reo_wnd()
210 } else if (!tp->rack.reo_wnd_persist) { in tcp_rack_update_reo_wnd()
211 tp->rack.reo_wnd_steps = 1; in tcp_rack_update_reo_wnd()
223 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost() local
225 if ((state < TCP_CA_Recovery && tp->sacked_out >= tp->reordering) || in tcp_newreno_mark_lost()