Lines Matching refs:tcp_sk

520 	last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp);  in tcp_synq_overflow()
522 WRITE_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp, now); in tcp_synq_overflow()
543 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_no_recent_overflow()
627 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers()
630 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers()
705 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check()
1227 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh()
1272 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited()
1297 s64 delay = tcp_sk(sk)->tcp_wstamp_ns - tcp_sk(sk)->tcp_clock_cache; in tcp_pacing_delay()
1335 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) in tcp_check_probe_timer()
1382 struct tcp_sock *tp = tcp_sk(sk); in tcp_slow_start_after_idle_check()
1443 const struct tcp_sock *tp = tcp_sk(sk); in tcp_epollin_ready()
1491 int fin_timeout = tcp_sk(sk)->linger2 ? : sock_net(sk)->ipv4.sysctl_tcp_fin_timeout; in tcp_fin_time()
1823 const struct tcp_sock *tp = tcp_sk(sk); in tcp_write_queue_empty()
1879 struct tcp_sock *tp = tcp_sk(sk); in tcp_push_pending_frames()
1902 tcp_sk(sk)->highest_sack = skb_rb_next(skb); in tcp_advance_highest_sack()
1907 return tcp_sk(sk)->highest_sack; in tcp_highest_sack()
1912 tcp_sk(sk)->highest_sack = tcp_rtx_queue_head(sk); in tcp_highest_sack_reset()
1921 tcp_sk(sk)->highest_sack = new; in tcp_highest_sack_replace()
2089 return rto_time_stamp_us - tcp_sk(sk)->tcp_mstamp; in tcp_rto_delta_us()
2130 struct tcp_sock *tp = tcp_sk(sk); in tcp_inq()
2348 if (BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), BPF_SOCK_OPS_RTT_CB_FLAG)) in tcp_bpf_rtt()
2378 tcp_twsk(sk)->tw_tx_delay : tcp_sk(sk)->tcp_tx_delay; in tcp_transmit_time()