Home
last modified time | relevance | path

Searched refs:tcp_sk (Results 1 – 25 of 54) sorted by relevance

123

/Linux-v5.4/samples/bpf/
Dtcp_dumpstats_kern.c28 struct bpf_tcp_sock *tcp_sk; in _sockops() local
56 tcp_sk = bpf_tcp_sock(sk); in _sockops()
57 if (!tcp_sk) in _sockops()
63 tcp_sk->dsack_dups, tcp_sk->delivered); in _sockops()
65 tcp_sk->delivered_ce, tcp_sk->icsk_retransmits); in _sockops()
/Linux-v5.4/tools/testing/selftests/bpf/progs/
Dtcp_rtt.c28 struct bpf_tcp_sock *tcp_sk; in _sockops() local
49 tcp_sk = bpf_tcp_sock(sk); in _sockops()
50 if (!tcp_sk) in _sockops()
55 storage->dsack_dups = tcp_sk->dsack_dups; in _sockops()
56 storage->delivered = tcp_sk->delivered; in _sockops()
57 storage->delivered_ce = tcp_sk->delivered_ce; in _sockops()
58 storage->icsk_retransmits = tcp_sk->icsk_retransmits; in _sockops()
/Linux-v5.4/net/ipv4/
Dtcp_timer.c32 start_ts = tcp_sk(sk)->retrans_stamp; in tcp_clamp_rto_to_user_timeout()
35 elapsed = tcp_time_stamp(tcp_sk(sk)) - start_ts; in tcp_clamp_rto_to_user_timeout()
86 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources()
200 start_ts = tcp_sk(sk)->retrans_stamp; in retransmits_timed_out()
209 return (s32)(tcp_time_stamp(tcp_sk(sk)) - start_ts - timeout) >= 0; in retransmits_timed_out()
216 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout()
298 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_delack_timer_handler()
342 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer()
394 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer()
433 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer()
[all …]
Dtcp_bbr.c244 unsigned int mss = tcp_sk(sk)->mss_cache; in bbr_rate_bytes_per_sec()
266 struct tcp_sock *tp = tcp_sk(sk); in bbr_init_pacing_rate_from_rtt()
285 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_pacing_rate()
303 struct tcp_sock *tp = tcp_sk(sk); in bbr_tso_segs_goal()
319 struct tcp_sock *tp = tcp_sk(sk); in bbr_save_cwnd()
330 struct tcp_sock *tp = tcp_sk(sk); in bbr_cwnd_event()
435 struct tcp_sock *tp = tcp_sk(sk); in bbr_packets_in_net_at_edt()
479 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd_to_recover_or_restore()
518 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd()
553 struct tcp_sock *tp = tcp_sk(sk); in bbr_is_next_cycle_phase()
[all …]
Dtcp_recovery.c7 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_skb_lost()
26 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd()
78 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss()
115 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost()
168 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout()
204 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd()
234 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost()
Dtcp_input.c176 tcp_sk(sk)->advmss); in tcp_measure_rcv_mss()
200 len -= tcp_sk(sk)->tcp_header_len; in tcp_measure_rcv_mss()
216 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack()
257 tcp_sk(sk)->ecn_flags &= ~TCP_ECN_DEMAND_CWR; in tcp_ecn_accept_cwr()
274 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce()
306 if (tcp_sk(sk)->ecn_flags & TCP_ECN_OK) in tcp_ecn_check_ce()
336 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand()
394 struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window()
411 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window()
442 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space()
[all …]
Dtcp_dctcp.c77 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_init()
105 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ssthresh()
113 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_update_alpha()
147 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss()
185 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_get_info()
213 return max(tcp_sk(sk)->snd_cwnd, ca->loss_cwnd); in dctcp_cwnd_undo()
Dtcp_output.c67 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_new_data_sent()
92 const struct tcp_sock *tp = tcp_sk(sk); in tcp_acceptable_seq()
118 struct tcp_sock *tp = tcp_sk(sk); in tcp_advertise_mss()
139 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_restart()
181 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_ack_sent()
257 struct tcp_sock *tp = tcp_sk(sk); in tcp_select_window()
307 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_synack()
320 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_syn()
364 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send()
591 struct tcp_sock *tp = tcp_sk(sk); in tcp_syn_options()
[all …]
Dtcp_westwood.c73 w->snd_una = tcp_sk(sk)->snd_una; in tcp_westwood_init()
127 w->snd_una = tcp_sk(sk)->snd_una; in westwood_update_window()
165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw()
182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count()
219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin()
242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
Dtcp_rate.c42 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_sent()
81 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_delivered()
113 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_gen()
185 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited()
Dtcp_highspeed.c102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init()
114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid()
153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
Dtcp_cdg.c143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update()
244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff()
265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid()
302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked()
331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh()
348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event()
376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
Dtcp_cubic.c131 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset()
150 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in bictcp_init()
160 delta = now - tcp_sk(sk)->lsndtime; in bictcp_cwnd_event()
337 struct tcp_sock *tp = tcp_sk(sk); in bictcp_cong_avoid()
356 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_recalc_ssthresh()
381 struct tcp_sock *tp = tcp_sk(sk); in hystart_update()
432 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_acked()
Dtcp_dctcp.h6 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ece_ack_cwr()
35 *prior_rcv_nxt = tcp_sk(sk)->rcv_nxt; in dctcp_ece_ack_update()
Dtcp_scalable.c21 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid()
35 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
Dtcp_fastopen.c165 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_add_skb()
254 tp = tcp_sk(child); in tcp_fastopen_create_child()
330 tcp_sk(sk)->fastopen_no_cookie || in tcp_fastopen_no_cookie()
438 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_defer_connect()
512 tcp_sk(sk)->syn_fastopen_ch = 1; in tcp_fastopen_active_should_disable()
524 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_disable_ofo_check()
551 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_detect_blackhole()
Dtcp_hybla.c39 tcp_sk(sk)->srtt_us / (rtt0 * USEC_PER_MSEC), in hybla_recalc_param()
48 struct tcp_sock *tp = tcp_sk(sk); in hybla_init()
92 struct tcp_sock *tp = tcp_sk(sk); in hybla_cong_avoid()
Dtcp_illinois.c59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset()
224 struct tcp_sock *tp = tcp_sk(sk); in update_params()
262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid()
297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
Dtcp_bic.c77 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in bictcp_init()
142 struct tcp_sock *tp = tcp_sk(sk); in bictcp_cong_avoid()
162 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_recalc_ssthresh()
Dtcp_yeah.c45 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_init()
74 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_cong_avoid()
208 const struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_ssthresh()
Dtcp_diag.c27 const struct tcp_sock *tp = tcp_sk(sk); in tcp_diag_get_info()
123 md5sig = rcu_dereference(tcp_sk(sk)->md5sig_info); in tcp_diag_get_aux()
156 md5sig = rcu_dereference(tcp_sk(sk)->md5sig_info); in tcp_diag_get_aux_size()
Dtcp.c411 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock()
496 const struct tcp_sock *tp = tcp_sk(sk); in tcp_poll()
596 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl()
653 struct tcp_sock *tp = tcp_sk(sk); in skb_entail()
698 struct tcp_sock *tp = tcp_sk(sk); in tcp_push()
908 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_size_goal()
959 struct tcp_sock *tp = tcp_sk(sk); in do_tcp_sendpages()
1136 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_fastopen()
1181 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_locked()
1450 struct tcp_sock *tp = tcp_sk(sk); in tcp_recv_urg()
[all …]
Dtcp_lp.c136 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_remote_hz_estimator()
189 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_owd_calculator()
266 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_pkts_acked()
Dtcp_cong.c181 tcp_sk(sk)->prior_ssthresh = 0; in tcp_init_congestion_control()
436 struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_cong_avoid()
455 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_ssthresh()
463 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_undo_cwnd()
/Linux-v5.4/include/net/
Dtcp.h503 last_overflow = tcp_sk(sk)->rx_opt.ts_recent_stamp; in tcp_synq_overflow()
505 tcp_sk(sk)->rx_opt.ts_recent_stamp = now; in tcp_synq_overflow()
525 last_overflow = tcp_sk(sk)->rx_opt.ts_recent_stamp; in tcp_synq_no_recent_overflow()
600 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers()
603 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers()
674 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check()
1195 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh()
1240 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited()
1266 s64 pacing_delay = skb ? skb->tstamp : tcp_sk(sk)->tcp_wstamp_ns; in tcp_pacing_delay()
1268 pacing_delay -= tcp_sk(sk)->tcp_clock_cache; in tcp_pacing_delay()
[all …]

123