Searched refs:mss_cache (Results 1 – 13 of 13) sorted by relevance
/Linux-v4.19/net/ipv4/ |
D | tcp_westwood.c | 190 w->accounted += tp->mss_cache; in westwood_acked_count() 191 w->cumul_ack = tp->mss_cache; in westwood_acked_count() 194 if (w->cumul_ack > tp->mss_cache) { in westwood_acked_count() 198 w->cumul_ack = tp->mss_cache; in westwood_acked_count() 221 return max_t(u32, (w->bw_est * w->rtt_min) / tp->mss_cache, 2); in tcp_westwood_bw_rttmin()
|
D | tcp_nv.c | 349 bytes_acked >= (ca->nv_min_cwnd - 1) * tp->mss_cache && in tcpnv_acked() 355 ca->nv_min_cwnd * tp->mss_cache; in tcpnv_acked() 367 80000ULL * tp->mss_cache); in tcpnv_acked()
|
D | tcp_rate.c | 186 tp->write_seq - tp->snd_nxt < tp->mss_cache && in tcp_rate_check_app_limited()
|
D | tcp_bbr.c | 209 unsigned int mss = tcp_sk(sk)->mss_cache; in bbr_rate_bytes_per_sec() 284 segs = max_t(u32, bytes / tp->mss_cache, bbr_min_tso_segs(sk)); in bbr_tso_segs_goal() 920 bw = bw * tp->mss_cache * USEC_PER_SEC >> BW_SCALE; in bbr_get_info()
|
D | tcp_output.c | 1551 tp->mss_cache = mss_now; in tcp_sync_mss() 1569 mss_now = tp->mss_cache; in tcp_current_mss() 1933 cong_win = (tp->snd_cwnd - in_flight) * tp->mss_cache; in tcp_tso_should_defer() 1938 if (limit >= max_segs * tp->mss_cache) in tcp_tso_should_defer() 1947 u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); in tcp_tso_should_defer() 1961 if (limit > tcp_max_tso_deferred_mss(tp) * tp->mss_cache) in tcp_tso_should_defer() 2068 size_needed = probe_size + (tp->reordering + 1) * tp->mss_cache; in tcp_mtu_probe()
|
D | tcp_input.c | 338 per_mss = max_t(u32, tp->rx_opt.mss_clamp, tp->mss_cache) + in tcp_sndbuf_expand() 518 unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache); in tcp_initialize_rcv_mss() 812 rate = (u64)tp->mss_cache * ((USEC_PER_SEC / 100) << 3); in tcp_update_pacing_rate() 895 const u32 mss = tp->mss_cache; in tcp_check_sack_reordering() 2572 tcp_mss_to_mtu(sk, tp->mss_cache) / in tcp_mtup_probe_success() 2765 tp->snd_una + tp->reordering * tp->mss_cache); in tcp_force_fast_retransmit() 3171 if (pkts_acked == 1 && last_in_flight < tp->mss_cache && in tcp_clean_rtx_queue()
|
D | tcp.c | 396 rate64 = (u64)rate * tp->mss_cache * USEC_PER_SEC; in tcp_compute_delivery_rate() 437 tp->mss_cache = TCP_MSS_DEFAULT; in tcp_init_sock() 3163 info->tcpi_snd_mss = tp->mss_cache; in tcp_get_info() 3314 val = tp->mss_cache; in do_tcp_getsockopt()
|
/Linux-v4.19/include/linux/ |
D | tcp.h | 208 u32 mss_cache; /* Cached effective mss, not including SACKS */ member
|
/Linux-v4.19/include/uapi/linux/ |
D | bpf.h | 2577 __u32 mss_cache; member
|
/Linux-v4.19/tools/include/uapi/linux/ |
D | bpf.h | 2577 __u32 mss_cache; member
|
/Linux-v4.19/drivers/crypto/chelsio/chtls/ |
D | chtls_io.c | 195 FLOWC_PARAM(MSS, tp->mss_cache); in send_tx_flowc_wr()
|
D | chtls_cm.c | 115 tp->mss_cache = tp->rx_opt.mss_clamp; in assign_rxopt()
|
/Linux-v4.19/net/core/ |
D | filter.c | 6640 case offsetof(struct bpf_sock_ops, mss_cache): in sock_ops_convert_ctx_access() 6641 SOCK_OPS_GET_FIELD(mss_cache, mss_cache, struct tcp_sock); in sock_ops_convert_ctx_access()
|