Lines Matching refs:icsk

41 	struct inet_connection_sock *icsk = inet_csk(sk);  in tcp_clamp_rto_to_user_timeout()  local
45 if (!icsk->icsk_user_timeout || !start_ts) in tcp_clamp_rto_to_user_timeout()
46 return icsk->icsk_rto; in tcp_clamp_rto_to_user_timeout()
48 if (elapsed >= icsk->icsk_user_timeout) in tcp_clamp_rto_to_user_timeout()
51 return min_t(u32, icsk->icsk_rto, msecs_to_jiffies(icsk->icsk_user_timeout - elapsed)); in tcp_clamp_rto_to_user_timeout()
153 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk) in tcp_mtu_probing() argument
162 if (!icsk->icsk_mtup.enabled) { in tcp_mtu_probing()
163 icsk->icsk_mtup.enabled = 1; in tcp_mtu_probing()
164 icsk->icsk_mtup.probe_timestamp = tcp_jiffies32; in tcp_mtu_probing()
166 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1; in tcp_mtu_probing()
169 icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss); in tcp_mtu_probing()
171 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtu_probing()
218 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout() local
225 if (icsk->icsk_retransmits) { in tcp_write_timeout()
230 retry_until = icsk->icsk_syn_retries ? : net->ipv4.sysctl_tcp_syn_retries; in tcp_write_timeout()
231 expired = icsk->icsk_retransmits >= retry_until; in tcp_write_timeout()
235 tcp_mtu_probing(icsk, sk); in tcp_write_timeout()
244 const bool alive = icsk->icsk_rto < TCP_RTO_MAX; in tcp_write_timeout()
254 icsk->icsk_user_timeout); in tcp_write_timeout()
260 icsk->icsk_retransmits, in tcp_write_timeout()
261 icsk->icsk_rto, (int)expired); in tcp_write_timeout()
275 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler() local
280 !(icsk->icsk_ack.pending & ICSK_ACK_TIMER)) in tcp_delack_timer_handler()
283 if (time_after(icsk->icsk_ack.timeout, jiffies)) { in tcp_delack_timer_handler()
284 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in tcp_delack_timer_handler()
287 icsk->icsk_ack.pending &= ~ICSK_ACK_TIMER; in tcp_delack_timer_handler()
290 if (!icsk->icsk_ack.pingpong) { in tcp_delack_timer_handler()
292 icsk->icsk_ack.ato = min(icsk->icsk_ack.ato << 1, icsk->icsk_rto); in tcp_delack_timer_handler()
297 icsk->icsk_ack.pingpong = 0; in tcp_delack_timer_handler()
298 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_delack_timer_handler()
322 struct inet_connection_sock *icsk = in tcp_delack_timer() local
323 from_timer(icsk, t, icsk_delack_timer); in tcp_delack_timer()
324 struct sock *sk = &icsk->icsk_inet.sk; in tcp_delack_timer()
330 icsk->icsk_ack.blocked = 1; in tcp_delack_timer()
342 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer() local
349 icsk->icsk_probes_out = 0; in tcp_probe_timer()
364 else if (icsk->icsk_user_timeout && in tcp_probe_timer()
365 (s32)(tcp_time_stamp(tp) - start_ts) > icsk->icsk_user_timeout) in tcp_probe_timer()
370 const bool alive = inet_csk_rto_backoff(icsk, TCP_RTO_MAX) < TCP_RTO_MAX; in tcp_probe_timer()
373 if (!alive && icsk->icsk_backoff >= max_probes) in tcp_probe_timer()
379 if (icsk->icsk_probes_out > max_probes) { in tcp_probe_timer()
393 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer() local
394 int max_retries = icsk->icsk_syn_retries ? : in tcp_fastopen_synack_timer()
412 icsk->icsk_retransmits++; in tcp_fastopen_synack_timer()
433 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer() local
488 if (icsk->icsk_retransmits == 0) { in tcp_retransmit_timer()
491 if (icsk->icsk_ca_state == TCP_CA_Recovery) { in tcp_retransmit_timer()
496 } else if (icsk->icsk_ca_state == TCP_CA_Loss) { in tcp_retransmit_timer()
498 } else if ((icsk->icsk_ca_state == TCP_CA_Disorder) || in tcp_retransmit_timer()
516 if (!icsk->icsk_retransmits) in tcp_retransmit_timer()
517 icsk->icsk_retransmits = 1; in tcp_retransmit_timer()
519 min(icsk->icsk_rto, TCP_RESOURCE_PROBE_INTERVAL), in tcp_retransmit_timer()
539 icsk->icsk_backoff++; in tcp_retransmit_timer()
540 icsk->icsk_retransmits++; in tcp_retransmit_timer()
555 icsk->icsk_retransmits <= TCP_THIN_LINEAR_RETRIES) { in tcp_retransmit_timer()
556 icsk->icsk_backoff = 0; in tcp_retransmit_timer()
557 icsk->icsk_rto = min(__tcp_set_rto(tp), TCP_RTO_MAX); in tcp_retransmit_timer()
560 icsk->icsk_rto = min(icsk->icsk_rto << 1, TCP_RTO_MAX); in tcp_retransmit_timer()
574 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timer_handler() local
578 !icsk->icsk_pending) in tcp_write_timer_handler()
581 if (time_after(icsk->icsk_timeout, jiffies)) { in tcp_write_timer_handler()
582 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in tcp_write_timer_handler()
587 event = icsk->icsk_pending; in tcp_write_timer_handler()
597 icsk->icsk_pending = 0; in tcp_write_timer_handler()
601 icsk->icsk_pending = 0; in tcp_write_timer_handler()
612 struct inet_connection_sock *icsk = in tcp_write_timer() local
613 from_timer(icsk, t, icsk_retransmit_timer); in tcp_write_timer()
614 struct sock *sk = &icsk->icsk_inet.sk; in tcp_write_timer()
652 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_keepalive_timer() local
699 if ((icsk->icsk_user_timeout != 0 && in tcp_keepalive_timer()
700 elapsed >= msecs_to_jiffies(icsk->icsk_user_timeout) && in tcp_keepalive_timer()
701 icsk->icsk_probes_out > 0) || in tcp_keepalive_timer()
702 (icsk->icsk_user_timeout == 0 && in tcp_keepalive_timer()
703 icsk->icsk_probes_out >= keepalive_probes(tp))) { in tcp_keepalive_timer()
709 icsk->icsk_probes_out++; in tcp_keepalive_timer()