• Home
  • Raw
  • Download

Lines Matching refs:tp

12 	struct tcp_sock *tp = tcp_sk(sk);  in tcp_rack_reo_wnd()  local
14 if (!tp->reord_seen) { in tcp_rack_reo_wnd()
21 if (tp->sacked_out >= tp->reordering && in tcp_rack_reo_wnd()
33 return min((tcp_min_rtt(tp) >> 2) * tp->rack.reo_wnd_steps, in tcp_rack_reo_wnd()
34 tp->srtt_us >> 3); in tcp_rack_reo_wnd()
37 s32 tcp_rack_skb_timeout(struct tcp_sock *tp, struct sk_buff *skb, u32 reo_wnd) in tcp_rack_skb_timeout() argument
39 return tp->rack.rtt_us + reo_wnd - in tcp_rack_skb_timeout()
40 tcp_stamp_us_delta(tp->tcp_mstamp, tcp_skb_timestamp_us(skb)); in tcp_rack_skb_timeout()
65 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss() local
71 list_for_each_entry_safe(skb, n, &tp->tsorted_sent_queue, in tcp_rack_detect_loss()
81 if (!tcp_rack_sent_after(tp->rack.mstamp, in tcp_rack_detect_loss()
83 tp->rack.end_seq, scb->end_seq)) in tcp_rack_detect_loss()
89 remaining = tcp_rack_skb_timeout(tp, skb, reo_wnd); in tcp_rack_detect_loss()
102 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost() local
105 if (!tp->rack.advanced) in tcp_rack_mark_lost()
109 tp->rack.advanced = 0; in tcp_rack_mark_lost()
123 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance() argument
128 rtt_us = tcp_stamp_us_delta(tp->tcp_mstamp, xmit_time); in tcp_rack_advance()
129 if (rtt_us < tcp_min_rtt(tp) && (sacked & TCPCB_RETRANS)) { in tcp_rack_advance()
142 tp->rack.advanced = 1; in tcp_rack_advance()
143 tp->rack.rtt_us = rtt_us; in tcp_rack_advance()
144 if (tcp_rack_sent_after(xmit_time, tp->rack.mstamp, in tcp_rack_advance()
145 end_seq, tp->rack.end_seq)) { in tcp_rack_advance()
146 tp->rack.mstamp = xmit_time; in tcp_rack_advance()
147 tp->rack.end_seq = end_seq; in tcp_rack_advance()
156 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout() local
158 u32 lost = tp->lost; in tcp_rack_reo_timeout()
160 prior_inflight = tcp_packets_in_flight(tp); in tcp_rack_reo_timeout()
162 if (prior_inflight != tcp_packets_in_flight(tp)) { in tcp_rack_reo_timeout()
166 tcp_cwnd_reduction(sk, 1, tp->lost - lost, 0); in tcp_rack_reo_timeout()
194 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd() local
202 if (before(rs->prior_delivered, tp->rack.last_delivered)) in tcp_rack_update_reo_wnd()
203 tp->rack.dsack_seen = 0; in tcp_rack_update_reo_wnd()
206 if (tp->rack.dsack_seen) { in tcp_rack_update_reo_wnd()
207 tp->rack.reo_wnd_steps = min_t(u32, 0xFF, in tcp_rack_update_reo_wnd()
208 tp->rack.reo_wnd_steps + 1); in tcp_rack_update_reo_wnd()
209 tp->rack.dsack_seen = 0; in tcp_rack_update_reo_wnd()
210 tp->rack.last_delivered = tp->delivered; in tcp_rack_update_reo_wnd()
211 tp->rack.reo_wnd_persist = TCP_RACK_RECOVERY_THRESH; in tcp_rack_update_reo_wnd()
212 } else if (!tp->rack.reo_wnd_persist) { in tcp_rack_update_reo_wnd()
213 tp->rack.reo_wnd_steps = 1; in tcp_rack_update_reo_wnd()
225 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost() local
227 if ((state < TCP_CA_Recovery && tp->sacked_out >= tp->reordering) || in tcp_newreno_mark_lost()