Lines Matching refs:tcp_sk

510 	last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp);  in tcp_synq_overflow()
512 WRITE_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp, now); in tcp_synq_overflow()
533 last_overflow = READ_ONCE(tcp_sk(sk)->rx_opt.ts_recent_stamp); in tcp_synq_no_recent_overflow()
617 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers()
620 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers()
690 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check()
1225 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh()
1270 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited()
1295 s64 delay = tcp_sk(sk)->tcp_wstamp_ns - tcp_sk(sk)->tcp_clock_cache; in tcp_pacing_delay()
1331 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) in tcp_check_probe_timer()
1378 struct tcp_sock *tp = tcp_sk(sk); in tcp_slow_start_after_idle_check()
1470 int fin_timeout = tcp_sk(sk)->linger2 ? : sock_net(sk)->ipv4.sysctl_tcp_fin_timeout; in tcp_fin_time()
1802 const struct tcp_sock *tp = tcp_sk(sk); in tcp_write_queue_empty()
1858 struct tcp_sock *tp = tcp_sk(sk); in tcp_push_pending_frames()
1881 tcp_sk(sk)->highest_sack = skb_rb_next(skb); in tcp_advance_highest_sack()
1886 return tcp_sk(sk)->highest_sack; in tcp_highest_sack()
1891 tcp_sk(sk)->highest_sack = tcp_rtx_queue_head(sk); in tcp_highest_sack_reset()
1900 tcp_sk(sk)->highest_sack = new; in tcp_highest_sack_replace()
1974 const struct tcp_sock *tp = tcp_sk(sk); in tcp_stream_memory_free()
2081 return rto_time_stamp_us - tcp_sk(sk)->tcp_mstamp; in tcp_rto_delta_us()
2122 struct tcp_sock *tp = tcp_sk(sk); in tcp_inq()
2339 if (BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), BPF_SOCK_OPS_RTT_CB_FLAG)) in tcp_bpf_rtt()
2369 tcp_twsk(sk)->tw_tx_delay : tcp_sk(sk)->tcp_tx_delay; in tcp_transmit_time()