Home
last modified time | relevance | path

Searched refs:tcp_sock (Results 1 – 25 of 74) sorted by relevance

123

/Linux-v6.6/net/ipv4/
Dbpf_tcp_ca.c98 case offsetof(struct tcp_sock, snd_cwnd): in bpf_tcp_ca_btf_struct_access()
99 end = offsetofend(struct tcp_sock, snd_cwnd); in bpf_tcp_ca_btf_struct_access()
101 case offsetof(struct tcp_sock, snd_cwnd_cnt): in bpf_tcp_ca_btf_struct_access()
102 end = offsetofend(struct tcp_sock, snd_cwnd_cnt); in bpf_tcp_ca_btf_struct_access()
104 case offsetof(struct tcp_sock, snd_ssthresh): in bpf_tcp_ca_btf_struct_access()
105 end = offsetofend(struct tcp_sock, snd_ssthresh); in bpf_tcp_ca_btf_struct_access()
107 case offsetof(struct tcp_sock, ecn_flags): in bpf_tcp_ca_btf_struct_access()
108 end = offsetofend(struct tcp_sock, ecn_flags); in bpf_tcp_ca_btf_struct_access()
110 case offsetof(struct tcp_sock, app_limited): in bpf_tcp_ca_btf_struct_access()
111 end = offsetofend(struct tcp_sock, app_limited); in bpf_tcp_ca_btf_struct_access()
[all …]
Dtcp_recovery.c7 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_wnd()
32 s32 tcp_rack_skb_timeout(struct tcp_sock *tp, struct sk_buff *skb, u32 reo_wnd) in tcp_rack_skb_timeout()
60 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_detect_loss()
97 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_mark_lost()
118 void tcp_rack_advance(struct tcp_sock *tp, u8 sacked, u32 end_seq, in tcp_rack_advance()
151 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_reo_timeout()
189 struct tcp_sock *tp = tcp_sk(sk); in tcp_rack_update_reo_wnd()
220 struct tcp_sock *tp = tcp_sk(sk); in tcp_newreno_mark_lost()
Dtcp_input.c335 static void tcp_ecn_queue_cwr(struct tcp_sock *tp) in tcp_ecn_queue_cwr()
355 static void tcp_ecn_withdraw_cwr(struct tcp_sock *tp) in tcp_ecn_withdraw_cwr()
362 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce()
398 static void tcp_ecn_rcv_synack(struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_synack()
404 static void tcp_ecn_rcv_syn(struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_syn()
410 static bool tcp_ecn_rcv_ecn_echo(const struct tcp_sock *tp, const struct tcphdr *th) in tcp_ecn_rcv_ecn_echo()
424 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand()
483 const struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window()
520 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window()
560 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space()
[all …]
Dtcp_dctcp.c70 static void dctcp_reset(const struct tcp_sock *tp, struct dctcp *ca) in dctcp_reset()
80 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_init()
110 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ssthresh()
118 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_update_alpha()
166 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss()
208 const struct tcp_sock *tp = tcp_sk(sk); in dctcp_get_info()
235 struct tcp_sock *tp = tcp_sk(sk); in dctcp_cwnd_undo()
Dtcp_output.c53 void tcp_mstamp_refresh(struct tcp_sock *tp) in tcp_mstamp_refresh()
68 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_new_data_sent()
97 const struct tcp_sock *tp = tcp_sk(sk); in tcp_acceptable_seq()
123 struct tcp_sock *tp = tcp_sk(sk); in tcp_advertise_mss()
144 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_restart()
161 static void tcp_event_data_sent(struct tcp_sock *tp, in tcp_event_data_sent()
182 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_ack_sent()
258 struct tcp_sock *tp = tcp_sk(sk); in tcp_select_window()
318 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_synack()
331 struct tcp_sock *tp = tcp_sk(sk); in tcp_ecn_send_syn()
[all …]
Dtcp_bbr.c268 struct tcp_sock *tp = tcp_sk(sk); in bbr_init_pacing_rate_from_rtt()
287 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_pacing_rate()
305 struct tcp_sock *tp = tcp_sk(sk); in bbr_tso_segs_goal()
322 struct tcp_sock *tp = tcp_sk(sk); in bbr_save_cwnd()
333 struct tcp_sock *tp = tcp_sk(sk); in bbr_cwnd_event()
438 struct tcp_sock *tp = tcp_sk(sk); in bbr_packets_in_net_at_edt()
482 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd_to_recover_or_restore()
521 struct tcp_sock *tp = tcp_sk(sk); in bbr_set_cwnd()
556 struct tcp_sock *tp = tcp_sk(sk); in bbr_is_next_cycle_phase()
592 struct tcp_sock *tp = tcp_sk(sk); in bbr_advance_cycle_phase()
[all …]
Dtcp_highspeed.c102 struct tcp_sock *tp = tcp_sk(sk); in hstcp_init()
114 struct tcp_sock *tp = tcp_sk(sk); in hstcp_cong_avoid()
153 const struct tcp_sock *tp = tcp_sk(sk); in hstcp_ssthresh()
Dtcp_rate.c42 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_sent()
83 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_skb_delivered()
120 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_gen()
196 struct tcp_sock *tp = tcp_sk(sk); in tcp_rate_check_app_limited()
Dtcp_cdg.c143 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_hystart_update()
244 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_backoff()
265 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cong_avoid()
302 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_acked()
331 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_ssthresh()
348 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_cwnd_event()
376 struct tcp_sock *tp = tcp_sk(sk); in tcp_cdg_init()
Dtcp_timer.c106 struct tcp_sock *tp = tcp_sk(sk); in tcp_out_of_resources()
236 struct tcp_sock *tp = tcp_sk(sk); in tcp_write_timeout()
301 struct tcp_sock *tp = tcp_sk(sk); in tcp_delack_timer_handler()
372 struct tcp_sock *tp = tcp_sk(sk); in tcp_probe_timer()
425 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastopen_synack_timer()
460 const struct tcp_sock *tp = tcp_sk(sk); in tcp_rtx_probe0_timed_out()
487 struct tcp_sock *tp = tcp_sk(sk); in tcp_retransmit_timer()
720 struct tcp_sock *tp = tcp_sk(sk); in tcp_keepalive_timer()
805 struct tcp_sock *tp = container_of(timer, struct tcp_sock, compressed_ack_timer); in tcp_compressed_ack_kick()
Dtcp_scalable.c20 struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_cong_avoid()
36 const struct tcp_sock *tp = tcp_sk(sk); in tcp_scalable_ssthresh()
Dtcp_westwood.c165 const struct tcp_sock *tp = tcp_sk(sk); in westwood_fast_bw()
182 const struct tcp_sock *tp = tcp_sk(sk); in westwood_acked_count()
219 const struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_bw_rttmin()
242 struct tcp_sock *tp = tcp_sk(sk); in tcp_westwood_event()
Dtcp_illinois.c59 struct tcp_sock *tp = tcp_sk(sk); in rtt_reset()
224 struct tcp_sock *tp = tcp_sk(sk); in update_params()
262 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_cong_avoid()
297 struct tcp_sock *tp = tcp_sk(sk); in tcp_illinois_ssthresh()
Dtcp_yeah.c43 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_init()
60 struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_cong_avoid()
189 const struct tcp_sock *tp = tcp_sk(sk); in tcp_yeah_ssthresh()
Dtcp_vegas.c73 const struct tcp_sock *tp = tcp_sk(sk); in vegas_enable()
160 static inline u32 tcp_vegas_ssthresh(struct tcp_sock *tp) in tcp_vegas_ssthresh()
167 struct tcp_sock *tp = tcp_sk(sk); in tcp_vegas_cong_avoid()
/Linux-v6.6/tools/testing/selftests/bpf/progs/
Dtcp_ca_write_sk_pacing.c14 static inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_sk()
16 return (struct tcp_sock *)sk; in tcp_sk()
19 static inline unsigned int tcp_left_out(const struct tcp_sock *tp) in tcp_left_out()
24 static inline unsigned int tcp_packets_in_flight(const struct tcp_sock *tp) in tcp_packets_in_flight()
44 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
Dbpf_dctcp.c51 static __always_inline void dctcp_reset(const struct tcp_sock *tp, in dctcp_reset()
63 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
111 struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
120 const struct tcp_sock *tp = tcp_sk(sk); in BPF_PROG()
150 struct tcp_sock *tp = tcp_sk(sk); in dctcp_react_to_loss()
169 struct tcp_sock *tp = tcp_sk(sk); in dctcp_ece_ack_cwr()
Dtcp_ca_incompl_cong_ops.c10 static inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_sk()
12 return (struct tcp_sock *)sk; in tcp_sk()
Dtcp_ca_update.c13 static inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_sk()
15 return (struct tcp_sock *)sk; in tcp_sk()
Dcgrp_ls_attach_cgroup.c27 struct tcp_sock *tcp_sk; in set_cookie()
56 struct tcp_sock *tcp_sk; in update_cookie_sockops()
/Linux-v6.6/tools/testing/selftests/bpf/
Dbpf_tcp_helpers.h62 struct tcp_sock { struct
101 static __always_inline struct tcp_sock *tcp_sk(const struct sock *sk) in tcp_sk() argument
103 return (struct tcp_sock *)sk; in tcp_sk()
200 static __always_inline bool tcp_in_slow_start(const struct tcp_sock *tp) in tcp_in_slow_start()
207 const struct tcp_sock *tp = tcp_sk(sk); in tcp_is_cwnd_limited()
230 extern __u32 tcp_slow_start(struct tcp_sock *tp, __u32 acked) __ksym;
231 extern void tcp_cong_avoid_ai(struct tcp_sock *tp, __u32 w, __u32 acked) __ksym;
/Linux-v6.6/include/net/
Dtcp.h394 void tcp_clear_retrans(struct tcp_sock *tp);
592 void tcp_mark_push(struct tcp_sock *tp, struct sk_buff *skb);
650 static inline int tcp_bound_to_half_wnd(struct tcp_sock *tp, int pktsize) in tcp_bound_to_half_wnd()
694 static inline u32 __tcp_set_rto(const struct tcp_sock *tp) in __tcp_set_rto()
699 static inline void __tcp_fast_path_on(struct tcp_sock *tp, u32 snd_wnd) in __tcp_fast_path_on()
710 static inline void tcp_fast_path_on(struct tcp_sock *tp) in tcp_fast_path_on()
717 struct tcp_sock *tp = tcp_sk(sk); in tcp_fast_path_check()
748 static inline u32 tcp_min_rtt(const struct tcp_sock *tp) in tcp_min_rtt()
757 static inline u32 tcp_receive_window(const struct tcp_sock *tp) in tcp_receive_window()
798 static inline u32 tcp_time_stamp(const struct tcp_sock *tp) in tcp_time_stamp()
[all …]
/Linux-v6.6/include/linux/
Dtcp.h177 struct tcp_sock { struct
477 #define tcp_sk(ptr) container_of_const(ptr, struct tcp_sock, inet_conn.icsk_inet.sk) argument
482 #define tcp_sk_rw(ptr) container_of(ptr, struct tcp_sock, inet_conn.icsk_inet.sk)
521 static inline void tcp_move_syn(struct tcp_sock *tp, in tcp_move_syn()
528 static inline void tcp_saved_syn_free(struct tcp_sock *tp) in tcp_saved_syn_free()
544 static inline u16 tcp_mss_clamp(const struct tcp_sock *tp, u16 mss) in tcp_mss_clamp()
/Linux-v6.6/net/mptcp/
Dfastopen.c14 struct tcp_sock *tp; in mptcp_fastopen_subflow_synack_set_params()
22 ssk = subflow->tcp_sock; in mptcp_fastopen_subflow_synack_set_params()
/Linux-v6.6/include/trace/events/
Dmptcp.h39 if (subflow->tcp_sock && sk_fullsock(subflow->tcp_sock))
40 __entry->free = sk_stream_memory_free(subflow->tcp_sock);

123