Lines Matching refs:sk

25 static u32 tcp_retransmit_stamp(const struct sock *sk)  in tcp_retransmit_stamp()  argument
27 u32 start_ts = tcp_sk(sk)->retrans_stamp; in tcp_retransmit_stamp()
30 struct sk_buff *head = tcp_rtx_queue_head(sk); in tcp_retransmit_stamp()
39 static u32 tcp_clamp_rto_to_user_timeout(const struct sock *sk) in tcp_clamp_rto_to_user_timeout() argument
41 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout()
44 start_ts = tcp_retransmit_stamp(sk); in tcp_clamp_rto_to_user_timeout()
47 elapsed = tcp_time_stamp(tcp_sk(sk)) - start_ts; in tcp_clamp_rto_to_user_timeout()
61 static void tcp_write_err(struct sock *sk) in tcp_write_err() argument
63 sk->sk_err = sk->sk_err_soft ? : ETIMEDOUT; in tcp_write_err()
64 sk->sk_error_report(sk); in tcp_write_err()
66 tcp_write_queue_purge(sk); in tcp_write_err()
67 tcp_done(sk); in tcp_write_err()
68 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err()
95 static int tcp_out_of_resources(struct sock *sk, bool do_reset) in tcp_out_of_resources() argument
97 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources()
106 if (sk->sk_err_soft) in tcp_out_of_resources()
109 if (tcp_check_oom(sk, shift)) { in tcp_out_of_resources()
117 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_out_of_resources()
118 tcp_done(sk); in tcp_out_of_resources()
119 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources()
123 if (!check_net(sock_net(sk))) { in tcp_out_of_resources()
125 tcp_done(sk); in tcp_out_of_resources()
137 static int tcp_orphan_retries(struct sock *sk, bool alive) in tcp_orphan_retries() argument
139 int retries = sock_net(sk)->ipv4.sysctl_tcp_orphan_retries; /* May be zero. */ in tcp_orphan_retries()
142 if (sk->sk_err_soft && !alive) in tcp_orphan_retries()
153 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk) in tcp_mtu_probing() argument
155 const struct net *net = sock_net(sk); in tcp_mtu_probing()
166 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1; in tcp_mtu_probing()
168 mss = max(mss, 68 - tcp_sk(sk)->tcp_header_len); in tcp_mtu_probing()
169 icsk->icsk_mtup.search_low = tcp_mss_to_mtu(sk, mss); in tcp_mtu_probing()
171 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtu_probing()
188 static bool retransmits_timed_out(struct sock *sk, in retransmits_timed_out() argument
195 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out()
198 start_ts = tcp_retransmit_stamp(sk); in retransmits_timed_out()
212 return (tcp_time_stamp(tcp_sk(sk)) - start_ts) >= timeout; in retransmits_timed_out()
216 static int tcp_write_timeout(struct sock *sk) in tcp_write_timeout() argument
218 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout()
219 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout()
220 struct net *net = sock_net(sk); in tcp_write_timeout()
224 if ((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV)) { in tcp_write_timeout()
226 dst_negative_advice(sk); in tcp_write_timeout()
228 sk_rethink_txhash(sk); in tcp_write_timeout()
233 if (retransmits_timed_out(sk, net->ipv4.sysctl_tcp_retries1, 0)) { in tcp_write_timeout()
235 tcp_mtu_probing(icsk, sk); in tcp_write_timeout()
237 dst_negative_advice(sk); in tcp_write_timeout()
239 sk_rethink_txhash(sk); in tcp_write_timeout()
243 if (sock_flag(sk, SOCK_DEAD)) { in tcp_write_timeout()
246 retry_until = tcp_orphan_retries(sk, alive); in tcp_write_timeout()
248 !retransmits_timed_out(sk, retry_until, 0); in tcp_write_timeout()
250 if (tcp_out_of_resources(sk, do_reset)) in tcp_write_timeout()
253 expired = retransmits_timed_out(sk, retry_until, in tcp_write_timeout()
256 tcp_fastopen_active_detect_blackhole(sk, expired); in tcp_write_timeout()
259 tcp_call_bpf_3arg(sk, BPF_SOCK_OPS_RTO_CB, in tcp_write_timeout()
265 tcp_write_err(sk); in tcp_write_timeout()
273 void tcp_delack_timer_handler(struct sock *sk) in tcp_delack_timer_handler() argument
275 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler()
277 sk_mem_reclaim_partial(sk); in tcp_delack_timer_handler()
279 if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) || in tcp_delack_timer_handler()
284 sk_reset_timer(sk, &icsk->icsk_delack_timer, icsk->icsk_ack.timeout); in tcp_delack_timer_handler()
289 if (inet_csk_ack_scheduled(sk)) { in tcp_delack_timer_handler()
300 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_delack_timer_handler()
301 tcp_send_ack(sk); in tcp_delack_timer_handler()
302 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler()
306 if (tcp_under_memory_pressure(sk)) in tcp_delack_timer_handler()
307 sk_mem_reclaim(sk); in tcp_delack_timer_handler()
324 struct sock *sk = &icsk->icsk_inet.sk; in tcp_delack_timer() local
326 bh_lock_sock(sk); in tcp_delack_timer()
327 if (!sock_owned_by_user(sk)) { in tcp_delack_timer()
328 tcp_delack_timer_handler(sk); in tcp_delack_timer()
331 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer()
333 if (!test_and_set_bit(TCP_DELACK_TIMER_DEFERRED, &sk->sk_tsq_flags)) in tcp_delack_timer()
334 sock_hold(sk); in tcp_delack_timer()
336 bh_unlock_sock(sk); in tcp_delack_timer()
337 sock_put(sk); in tcp_delack_timer()
340 static void tcp_probe_timer(struct sock *sk) in tcp_probe_timer() argument
342 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer()
343 struct sk_buff *skb = tcp_send_head(sk); in tcp_probe_timer()
344 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer()
368 max_probes = sock_net(sk)->ipv4.sysctl_tcp_retries2; in tcp_probe_timer()
369 if (sock_flag(sk, SOCK_DEAD)) { in tcp_probe_timer()
372 max_probes = tcp_orphan_retries(sk, alive); in tcp_probe_timer()
375 if (tcp_out_of_resources(sk, true)) in tcp_probe_timer()
380 abort: tcp_write_err(sk); in tcp_probe_timer()
383 tcp_send_probe0(sk); in tcp_probe_timer()
391 static void tcp_fastopen_synack_timer(struct sock *sk) in tcp_fastopen_synack_timer() argument
393 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer()
395 sock_net(sk)->ipv4.sysctl_tcp_synack_retries + 1; /* add one more retry for fastopen */ in tcp_fastopen_synack_timer()
398 req = tcp_sk(sk)->fastopen_rsk; in tcp_fastopen_synack_timer()
402 tcp_write_err(sk); in tcp_fastopen_synack_timer()
410 inet_rtx_syn_ack(sk, req); in tcp_fastopen_synack_timer()
413 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_fastopen_synack_timer()
429 void tcp_retransmit_timer(struct sock *sk) in tcp_retransmit_timer() argument
431 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer()
432 struct net *net = sock_net(sk); in tcp_retransmit_timer()
433 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer()
436 WARN_ON_ONCE(sk->sk_state != TCP_SYN_RECV && in tcp_retransmit_timer()
437 sk->sk_state != TCP_FIN_WAIT1); in tcp_retransmit_timer()
438 tcp_fastopen_synack_timer(sk); in tcp_retransmit_timer()
447 WARN_ON(tcp_rtx_queue_empty(sk)); in tcp_retransmit_timer()
451 if (!tp->snd_wnd && !sock_flag(sk, SOCK_DEAD) && in tcp_retransmit_timer()
452 !((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV))) { in tcp_retransmit_timer()
458 struct inet_sock *inet = inet_sk(sk); in tcp_retransmit_timer()
459 if (sk->sk_family == AF_INET) { in tcp_retransmit_timer()
467 else if (sk->sk_family == AF_INET6) { in tcp_retransmit_timer()
469 &sk->sk_v6_daddr, in tcp_retransmit_timer()
476 tcp_write_err(sk); in tcp_retransmit_timer()
479 tcp_enter_loss(sk); in tcp_retransmit_timer()
480 tcp_retransmit_skb(sk, tcp_rtx_queue_head(sk), 1); in tcp_retransmit_timer()
481 __sk_dst_reset(sk); in tcp_retransmit_timer()
485 if (tcp_write_timeout(sk)) in tcp_retransmit_timer()
507 __NET_INC_STATS(sock_net(sk), mib_idx); in tcp_retransmit_timer()
510 tcp_enter_loss(sk); in tcp_retransmit_timer()
512 if (tcp_retransmit_skb(sk, tcp_rtx_queue_head(sk), 1) > 0) { in tcp_retransmit_timer()
518 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_retransmit_timer()
552 if (sk->sk_state == TCP_ESTABLISHED && in tcp_retransmit_timer()
562 inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS, in tcp_retransmit_timer()
563 tcp_clamp_rto_to_user_timeout(sk), TCP_RTO_MAX); in tcp_retransmit_timer()
564 if (retransmits_timed_out(sk, net->ipv4.sysctl_tcp_retries1 + 1, 0)) in tcp_retransmit_timer()
565 __sk_dst_reset(sk); in tcp_retransmit_timer()
572 void tcp_write_timer_handler(struct sock *sk) in tcp_write_timer_handler() argument
574 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timer_handler()
577 if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) || in tcp_write_timer_handler()
582 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in tcp_write_timer_handler()
586 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_write_timer_handler()
591 tcp_rack_reo_timeout(sk); in tcp_write_timer_handler()
594 tcp_send_loss_probe(sk); in tcp_write_timer_handler()
598 tcp_retransmit_timer(sk); in tcp_write_timer_handler()
602 tcp_probe_timer(sk); in tcp_write_timer_handler()
607 sk_mem_reclaim(sk); in tcp_write_timer_handler()
614 struct sock *sk = &icsk->icsk_inet.sk; in tcp_write_timer() local
616 bh_lock_sock(sk); in tcp_write_timer()
617 if (!sock_owned_by_user(sk)) { in tcp_write_timer()
618 tcp_write_timer_handler(sk); in tcp_write_timer()
621 if (!test_and_set_bit(TCP_WRITE_TIMER_DEFERRED, &sk->sk_tsq_flags)) in tcp_write_timer()
622 sock_hold(sk); in tcp_write_timer()
624 bh_unlock_sock(sk); in tcp_write_timer()
625 sock_put(sk); in tcp_write_timer()
636 void tcp_set_keepalive(struct sock *sk, int val) in tcp_set_keepalive() argument
638 if ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) in tcp_set_keepalive()
641 if (val && !sock_flag(sk, SOCK_KEEPOPEN)) in tcp_set_keepalive()
642 inet_csk_reset_keepalive_timer(sk, keepalive_time_when(tcp_sk(sk))); in tcp_set_keepalive()
644 inet_csk_delete_keepalive_timer(sk); in tcp_set_keepalive()
651 struct sock *sk = from_timer(sk, t, sk_timer); in tcp_keepalive_timer() local
652 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_keepalive_timer()
653 struct tcp_sock *tp = tcp_sk(sk); in tcp_keepalive_timer()
657 bh_lock_sock(sk); in tcp_keepalive_timer()
658 if (sock_owned_by_user(sk)) { in tcp_keepalive_timer()
660 inet_csk_reset_keepalive_timer (sk, HZ/20); in tcp_keepalive_timer()
664 if (sk->sk_state == TCP_LISTEN) { in tcp_keepalive_timer()
670 if (sk->sk_state == TCP_FIN_WAIT2 && sock_flag(sk, SOCK_DEAD)) { in tcp_keepalive_timer()
672 const int tmo = tcp_fin_time(sk) - TCP_TIMEWAIT_LEN; in tcp_keepalive_timer()
675 tcp_time_wait(sk, TCP_FIN_WAIT2, tmo); in tcp_keepalive_timer()
679 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_keepalive_timer()
683 if (!sock_flag(sk, SOCK_KEEPOPEN) || in tcp_keepalive_timer()
684 ((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_SYN_SENT))) in tcp_keepalive_timer()
690 if (tp->packets_out || !tcp_write_queue_empty(sk)) in tcp_keepalive_timer()
704 tcp_send_active_reset(sk, GFP_ATOMIC); in tcp_keepalive_timer()
705 tcp_write_err(sk); in tcp_keepalive_timer()
708 if (tcp_write_wakeup(sk, LINUX_MIB_TCPKEEPALIVE) <= 0) { in tcp_keepalive_timer()
722 sk_mem_reclaim(sk); in tcp_keepalive_timer()
725 inet_csk_reset_keepalive_timer (sk, elapsed); in tcp_keepalive_timer()
729 tcp_done(sk); in tcp_keepalive_timer()
732 bh_unlock_sock(sk); in tcp_keepalive_timer()
733 sock_put(sk); in tcp_keepalive_timer()
739 struct sock *sk = (struct sock *)tp; in tcp_compressed_ack_kick() local
741 bh_lock_sock(sk); in tcp_compressed_ack_kick()
742 if (!sock_owned_by_user(sk)) { in tcp_compressed_ack_kick()
744 tcp_send_ack(sk); in tcp_compressed_ack_kick()
747 &sk->sk_tsq_flags)) in tcp_compressed_ack_kick()
748 sock_hold(sk); in tcp_compressed_ack_kick()
750 bh_unlock_sock(sk); in tcp_compressed_ack_kick()
752 sock_put(sk); in tcp_compressed_ack_kick()
757 void tcp_init_xmit_timers(struct sock *sk) in tcp_init_xmit_timers() argument
759 inet_csk_init_xmit_timers(sk, &tcp_write_timer, &tcp_delack_timer, in tcp_init_xmit_timers()
761 hrtimer_init(&tcp_sk(sk)->pacing_timer, CLOCK_MONOTONIC, in tcp_init_xmit_timers()
763 tcp_sk(sk)->pacing_timer.function = tcp_pace_kick; in tcp_init_xmit_timers()
765 hrtimer_init(&tcp_sk(sk)->compressed_ack_timer, CLOCK_MONOTONIC, in tcp_init_xmit_timers()
767 tcp_sk(sk)->compressed_ack_timer.function = tcp_compressed_ack_kick; in tcp_init_xmit_timers()