/Linux-v4.19/net/ipv4/ |
D | tcp_bbr.c | 209 unsigned int mss = tcp_sk(sk)->mss_cache; in bbr_rate_bytes_per_sec() 233 struct tcp_sock *tp = tcp_sk(sk); in bbr_init_pacing_rate_from_rtt() 258 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_pacing_rate() 276 struct tcp_sock *tp = tcp_sk(sk); in bbr_tso_segs_goal() 292 struct tcp_sock *tp = tcp_sk(sk); in bbr_save_cwnd() 303 struct tcp_sock *tp = tcp_sk(sk); in bbr_cwnd_event() 382 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd_to_recover_or_restore() 421 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd() 449 struct tcp_sock *tp = tcp_sk(sk); in bbr_is_next_cycle_phase() 485 struct tcp_sock *tp = tcp_sk(sk); in bbr_advance_cycle_phase() [all …]
|
D | tcp_timer.c | 27 u32 start_ts = tcp_sk(sk)->retrans_stamp; in tcp_retransmit_stamp() 47 elapsed = tcp_time_stamp(tcp_sk(sk)) - start_ts; in tcp_clamp_rto_to_user_timeout() 97 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources() 168 mss = max(mss, 68 - tcp_sk(sk)->tcp_header_len); in tcp_mtu_probing() 212 return (tcp_time_stamp(tcp_sk(sk)) - start_ts) >= timeout; in retransmits_timed_out() 219 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout() 300 tcp_mstamp_refresh(tcp_sk(sk)); in tcp_delack_timer_handler() 344 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer() 398 req = tcp_sk(sk)->fastopen_rsk; in tcp_fastopen_synack_timer() 431 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer() [all …]
|
D | tcp_recovery.c | 7 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_skb_lost() 26 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd() 78 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss() 114 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost() 167 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout() 203 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd() 233 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost()
|
D | tcp_input.c | 170 tcp_sk(sk)->advmss); in tcp_measure_rcv_mss() 194 len -= tcp_sk(sk)->tcp_header_len; in tcp_measure_rcv_mss() 210 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack() 251 tcp_sk(sk)->ecn_flags &= ~TCP_ECN_DEMAND_CWR; in tcp_ecn_accept_cwr() 268 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce() 300 if (tcp_sk(sk)->ecn_flags & TCP_ECN_OK) in tcp_ecn_check_ce() 330 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand() 387 struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window() 404 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window() 432 u32 mss = tcp_sk(sk)->advmss; in tcp_fixup_rcvbuf() [all …]
|
D | tcp_westwood.c | 72 w->snd_una = tcp_sk(sk)->snd_una; in tcp_westwood_init() 126 w->snd_una = tcp_sk(sk)->snd_una; in westwood_update_window() 164 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw() 181 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count() 218 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin() 241 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
|
D | tcp_output.c | 55 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_new_data_sent() 80 const struct tcp_sock *tp = tcp_sk(sk); in tcp_acceptable_seq() 106 struct tcp_sock *tp = tcp_sk(sk); in tcp_advertise_mss() 127 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_restart() 166 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_ack_sent() 261 struct tcp_sock *tp = tcp_sk(sk); in tcp_select_window() 311 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_synack() 324 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_syn() 368 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send() 595 struct tcp_sock *tp = tcp_sk(sk); in tcp_syn_options() [all …]
|
D | tcp_dctcp.c | 86 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_init() 115 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ssthresh() 130 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ce_state_0_to_1() 151 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ce_state_1_to_0() 171 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_update_alpha() 275 return max(tcp_sk(sk)->snd_cwnd, ca->loss_cwnd); in dctcp_cwnd_undo()
|
D | tcp_rate.c | 41 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_sent() 78 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_delivered() 111 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_gen() 183 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited()
|
D | tcp_highspeed.c | 101 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init() 113 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid() 152 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
|
D | tcp_cdg.c | 142 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update() 243 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff() 264 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid() 301 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked() 330 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh() 347 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event() 375 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
|
D | tcp_cubic.c | 130 struct tcp_sock *tp = tcp_sk(sk); in bictcp_hystart_reset() 149 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in bictcp_init() 159 delta = now - tcp_sk(sk)->lsndtime; in bictcp_cwnd_event() 336 struct tcp_sock *tp = tcp_sk(sk); in bictcp_cong_avoid() 355 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_recalc_ssthresh() 380 struct tcp_sock *tp = tcp_sk(sk); in hystart_update() 431 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_acked()
|
D | tcp_scalable.c | 20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid() 34 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
|
D | tcp_fastopen.c | 177 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_add_skb() 241 tp = tcp_sk(child); in tcp_fastopen_create_child() 317 tcp_sk(sk)->fastopen_no_cookie || in tcp_fastopen_no_cookie() 411 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_defer_connect() 485 tcp_sk(sk)->syn_fastopen_ch = 1; in tcp_fastopen_active_should_disable() 497 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_disable_ofo_check() 524 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_active_detect_blackhole()
|
D | tcp_hybla.c | 38 tcp_sk(sk)->srtt_us / (rtt0 * USEC_PER_MSEC), in hybla_recalc_param() 47 struct tcp_sock *tp = tcp_sk(sk); in hybla_init() 91 struct tcp_sock *tp = tcp_sk(sk); in hybla_cong_avoid()
|
D | tcp_illinois.c | 58 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset() 223 struct tcp_sock *tp = tcp_sk(sk); in update_params() 261 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid() 296 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
|
D | tcp_diag.c | 31 const struct tcp_sock *tp = tcp_sk(sk); in tcp_diag_get_info() 97 md5sig = rcu_dereference(tcp_sk(sk)->md5sig_info); in tcp_diag_get_aux() 120 md5sig = rcu_dereference(tcp_sk(sk)->md5sig_info); in tcp_diag_get_aux_size()
|
D | tcp_yeah.c | 44 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_init() 73 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_cong_avoid() 207 const struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_ssthresh()
|
D | tcp_bic.c | 76 tcp_sk(sk)->snd_ssthresh = initial_ssthresh; in bictcp_init() 141 struct tcp_sock *tp = tcp_sk(sk); in bictcp_cong_avoid() 161 const struct tcp_sock *tp = tcp_sk(sk); in bictcp_recalc_ssthresh()
|
D | tcp_lp.c | 135 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_remote_hz_estimator() 188 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_owd_calculator() 265 struct tcp_sock *tp = tcp_sk(sk); in tcp_lp_pkts_acked()
|
D | tcp.c | 410 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_sock() 507 const struct tcp_sock *tp = tcp_sk(sk); in tcp_poll() 607 struct tcp_sock *tp = tcp_sk(sk); in tcp_ioctl() 662 struct tcp_sock *tp = tcp_sk(sk); in skb_entail() 707 struct tcp_sock *tp = tcp_sk(sk); in tcp_push() 905 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_size_goal() 940 struct tcp_sock *tp = tcp_sk(sk); in do_tcp_sendpages() 1132 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_fastopen() 1176 struct tcp_sock *tp = tcp_sk(sk); in tcp_sendmsg_locked() 1457 struct tcp_sock *tp = tcp_sk(sk); in tcp_recv_urg() [all …]
|
D | tcp_ipv4.c | 115 struct tcp_sock *tp = tcp_sk(sk); in tcp_twsk_unique() 206 struct tcp_sock *tp = tcp_sk(sk); in tcp_v4_connect() 348 mtu = tcp_sk(sk)->mtu_info; in tcp_v4_mtu_reduced() 483 tp = tcp_sk(sk); in tcp_v4_err() 770 ctl_sk = *this_cpu_ptr(net->ipv4.tcp_sk); in tcp_v4_send_reset() 862 ctl_sk = *this_cpu_ptr(net->ipv4.tcp_sk); in tcp_v4_send_ack() 902 tcp_sk(sk)->snd_nxt; in tcp_v4_reqsk_send_ack() 977 const struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_lookup() 1019 const struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_lookup_exact() 1060 struct tcp_sock *tp = tcp_sk(sk); in tcp_md5_do_add() [all …]
|
D | tcp_cong.c | 180 tcp_sk(sk)->prior_ssthresh = 0; in tcp_init_congestion_control() 435 struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_cong_avoid() 454 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_ssthresh() 462 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reno_undo_cwnd()
|
D | tcp_nv.c | 125 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_reset() 182 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_cong_avoid() 210 const struct tcp_sock *tp = tcp_sk(sk); in tcpnv_recalc_ssthresh() 242 struct tcp_sock *tp = tcp_sk(sk); in tcpnv_acked()
|
/Linux-v4.19/include/net/ |
D | tcp.h | 492 last_overflow = tcp_sk(sk)->rx_opt.ts_recent_stamp; in tcp_synq_overflow() 494 tcp_sk(sk)->rx_opt.ts_recent_stamp = now; in tcp_synq_overflow() 514 last_overflow = tcp_sk(sk)->rx_opt.ts_recent_stamp; in tcp_synq_no_recent_overflow() 589 if (hrtimer_try_to_cancel(&tcp_sk(sk)->pacing_timer) == 1) in tcp_clear_xmit_timers() 592 if (hrtimer_try_to_cancel(&tcp_sk(sk)->compressed_ack_timer) == 1) in tcp_clear_xmit_timers() 663 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check() 1172 const struct tcp_sock *tp = tcp_sk(sk); in tcp_current_ssthresh() 1217 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited() 1259 if (!tcp_sk(sk)->packets_out && !inet_csk(sk)->icsk_pending) in tcp_check_probe_timer() 1324 struct tcp_sock *tp = tcp_sk(sk); in tcp_slow_start_after_idle_check() [all …]
|
/Linux-v4.19/drivers/crypto/chelsio/chtls/ |
D | chtls_io.c | 113 struct tcp_sock *tp = tcp_sk(sk); in send_flowc_wr() 175 tp = tcp_sk(sk); in send_tx_flowc_wr() 395 ((tcp_sk(sk)->nonagle & TCP_NAGLE_OFF) ? 0 : in tls_tx_data_wr() 533 ((tcp_sk(sk)->nonagle & TCP_NAGLE_OFF) ? 0 : in make_tx_data_wr() 611 tp = tcp_sk(sk); in chtls_push_frames() 736 struct tcp_sock *tp = tcp_sk(sk); in should_push() 766 return should_push(sk) && !corked(tcp_sk(sk), flags); in send_should_push() 776 struct tcp_sock *tp = tcp_sk(sk); in chtls_tcp_push() 826 struct tcp_sock *tp = tcp_sk(sk); in skb_entail() 993 struct tcp_sock *tp = tcp_sk(sk); in chtls_sendmsg() [all …]
|