| /Linux-v5.15/include/net/ |
| D | inet_connection_sock.h | 147 static inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function 154 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca() 177 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_SCHED; in inet_csk_schedule_ack() 182 return inet_csk(sk)->icsk_ack.pending & ICSK_ACK_SCHED; in inet_csk_ack_scheduled() 187 memset(&inet_csk(sk)->icsk_ack, 0, sizeof(inet_csk(sk)->icsk_ack)); in inet_csk_delack_init() 195 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() 220 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() 272 reqsk_queue_added(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_added() 277 return reqsk_queue_len(&inet_csk(sk)->icsk_accept_queue); in inet_csk_reqsk_queue_len() 303 return !reqsk_queue_empty(&inet_csk(sk)->icsk_accept_queue) ? in inet_csk_listen_poll() [all …]
|
| D | tcp.h | 358 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode() 678 if (inet_csk(sk)->icsk_rto > TCP_RTO_MAX) in tcp_bound_rto() 679 inet_csk(sk)->icsk_rto = TCP_RTO_MAX; in tcp_bound_rto() 718 u32 rto_min = inet_csk(sk)->icsk_rto_min; in tcp_rto_min() 1132 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn() 1139 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state() 1148 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_event() 1218 (1 << inet_csk(sk)->icsk_ca_state); in tcp_in_cwnd_reduction() 1319 return max_t(unsigned long, inet_csk(sk)->icsk_rto, TCP_RTO_MIN); in tcp_probe0_base() 1327 inet_csk(sk)->icsk_backoff); in tcp_probe0_when() [all …]
|
| D | espintcp.h | 35 struct inet_connection_sock *icsk = inet_csk(sk); in espintcp_getctx()
|
| /Linux-v5.15/net/ipv4/ |
| D | tcp_recovery.c | 18 if (inet_csk(sk)->icsk_ca_state >= TCP_CA_Recovery) in tcp_rack_reo_wnd() 113 timeout, inet_csk(sk)->icsk_rto); in tcp_rack_mark_lost() 162 if (inet_csk(sk)->icsk_ca_state != TCP_CA_Recovery) { in tcp_rack_reo_timeout() 164 if (!inet_csk(sk)->icsk_ca_ops->cong_control) in tcp_rack_reo_timeout() 169 if (inet_csk(sk)->icsk_pending != ICSK_TIME_RETRANS) in tcp_rack_reo_timeout() 222 const u8 state = inet_csk(sk)->icsk_ca_state; in tcp_newreno_mark_lost()
|
| D | tcp_timer.c | 28 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout() 45 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_probe0_to_user_timeout() 215 if (!inet_csk(sk)->icsk_retransmits) in retransmits_timed_out() 233 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timeout() 291 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_delack_timer_handler() 357 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_probe_timer() 408 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastopen_synack_timer() 452 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_retransmit_timer() 598 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_write_timer_handler() 676 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_keepalive_timer()
|
| D | inet_connection_sock.c | 408 if (!inet_csk(sk)->icsk_bind_hash) in inet_csk_get_port() 410 WARN_ON(inet_csk(sk)->icsk_bind_hash != tb); in inet_csk_get_port() 425 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() 473 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() 563 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() 574 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() 782 reqsk_queue_removed(&inet_csk(sk)->icsk_accept_queue, req); in inet_csk_reqsk_queue_drop() 824 reqsk_queue_migrated(&inet_csk(nsk)->icsk_accept_queue, req); in reqsk_timer_handler() 830 icsk = inet_csk(sk_listener); in reqsk_timer_handler() 887 reqsk_queue_removed(&inet_csk(oreq->rsk_listener)->icsk_accept_queue, oreq); in reqsk_timer_handler() [all …]
|
| D | tcp_ulp.c | 106 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() 132 struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_set_ulp()
|
| D | tcp_dctcp.h | 29 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 33 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
| D | tcp_output.c | 67 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() 152 while ((delta -= inet_csk(sk)->icsk_rto) > 0 && cwnd > restart_cwnd) in tcp_cwnd_restart() 163 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() 1115 inet_csk(sk)->icsk_af_ops->mtu_reduced(sk); in tcp_release_cb() 1241 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() 1700 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() 1741 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mss_to_mtu() 1764 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_init() 1802 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_sync_mss() 1837 if (mtu != inet_csk(sk)->icsk_pmtu_cookie) in tcp_current_mss() [all …]
|
| D | inet_hashtables.c | 100 inet_csk(sk)->icsk_bind_hash = tb; in inet_bind_hash() 115 tb = inet_csk(sk)->icsk_bind_hash; in __inet_put_port() 117 inet_csk(sk)->icsk_bind_hash = NULL; in __inet_put_port() 142 tb = inet_csk(sk)->icsk_bind_hash; in __inet_inherit_port() 207 hlist_add_tail_rcu(&inet_csk(sk)->icsk_listen_portaddr_node, in inet_hash2() 210 hlist_add_head_rcu(&inet_csk(sk)->icsk_listen_portaddr_node, in inet_hash2() 221 WARN_ON_ONCE(hlist_unhashed(&inet_csk(sk)->icsk_listen_portaddr_node))) in inet_unhash2() 227 hlist_del_init_rcu(&inet_csk(sk)->icsk_listen_portaddr_node); in inet_unhash2() 613 struct inet_bind_bucket *tb = inet_csk(sk)->icsk_bind_hash; in inet_reuseport_add_sock() 623 inet_csk(sk2)->icsk_bind_hash == tb && in inet_reuseport_add_sock() [all …]
|
| D | tcp_cong.c | 160 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control() 179 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control() 194 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_reinit_congestion_control() 213 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_congestion_control() 355 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_congestion_control()
|
| D | tcp_fastopen.c | 49 inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); in tcp_fastopen_destroy_cipher() 88 q = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_reset_cipher() 249 struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; in tcp_fastopen_create_child() 253 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_fastopen_create_child() 315 fastopenq = &inet_csk(sk)->icsk_accept_queue.fastopenq; in tcp_fastopen_queue_check() 584 u32 timeouts = inet_csk(sk)->icsk_retransmits; in tcp_fastopen_active_detect_blackhole()
|
| D | tcp_input.c | 229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() 280 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack() 292 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode() 306 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode() 329 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in tcp_ecn_accept_cwr() 403 const struct tcp_congestion_ops *ca_ops = inet_csk(sk)->icsk_ca_ops; in tcp_sndbuf_expand() 468 return 2 * inet_csk(sk)->icsk_ack.rcv_mss; in __tcp_grow_window() 519 inet_csk(sk)->icsk_ack.quick |= 1; in tcp_grow_window() 567 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window() 600 inet_csk(sk)->icsk_ack.rcv_mss = hint; in tcp_initialize_rcv_mss() [all …]
|
| D | tcp_dctcp.c | 98 inet_csk(sk)->icsk_ca_ops = &dctcp_reno; in dctcp_init() 156 new_state != inet_csk(sk)->icsk_ca_state) in dctcp_state() 193 if (inet_csk(sk)->icsk_ca_ops != &dctcp_reno) { in dctcp_get_info()
|
| D | tcp_minisocks.c | 255 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() 408 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() 470 newicsk = inet_csk(newsk); in tcp_create_openreq_child() 759 if (req->num_timeout < inet_csk(sk)->icsk_accept_queue.rskq_defer_accept && in tcp_check_req() 772 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in tcp_check_req() 778 reqsk_queue_removed(&inet_csk(req->rsk_listener)->icsk_accept_queue, req); in tcp_check_req()
|
| D | tcp_diag.c | 115 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux() 146 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux_size()
|
| /Linux-v5.15/net/dccp/ |
| D | output.c | 46 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_transmit_skb() 163 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_sync_mss() 272 inet_csk(sk)->icsk_rto, in dccp_xmit_packet() 390 if (inet_csk(sk)->icsk_af_ops->rebuild_header(sk) != 0) in dccp_retransmit_skb() 394 inet_csk(sk)->icsk_retransmits++; in dccp_retransmit_skb() 521 int err = inet_csk(sk)->icsk_af_ops->rebuild_header(sk); in dccp_send_reset() 546 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_connect() 590 inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN; in dccp_send_ack() 610 struct inet_connection_sock *icsk = inet_csk(sk);
|
| D | minisocks.c | 38 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_time_wait() 91 struct inet_connection_sock *newicsk = inet_csk(newsk); in dccp_create_openreq_child() 194 child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL, in dccp_check_req()
|
| D | timer.c | 33 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_write_timeout() 85 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_retransmit_timer()
|
| D | ipv6.c | 156 if (inet_csk(sk)->icsk_pmtu_cookie > dst_mtu(dst)) in dccp_v6_err() 444 inet_csk(newsk)->icsk_af_ops = &dccp_ipv6_mapped; in dccp_v6_request_recv_sock() 464 dccp_sync_mss(newsk, inet_csk(newsk)->icsk_pmtu_cookie); in dccp_v6_request_recv_sock() 536 inet_csk(newsk)->icsk_ext_hdr_len = 0; in dccp_v6_request_recv_sock() 538 inet_csk(newsk)->icsk_ext_hdr_len = opt->opt_nflen + in dccp_v6_request_recv_sock() 822 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_v6_connect() 1017 inet_csk(sk)->icsk_af_ops = &dccp_ipv6_af_ops; in dccp_v6_init_sock()
|
| D | proto.c | 101 if (inet_csk(sk)->icsk_bind_hash != NULL && in dccp_set_state() 186 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_init_sock() 222 if (inet_csk(sk)->icsk_bind_hash != NULL) in dccp_destroy_sock() 260 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_disconnect() 578 return inet_csk(sk)->icsk_af_ops->setsockopt(sk, level, in dccp_setsockopt() 686 return inet_csk(sk)->icsk_af_ops->getsockopt(sk, level, in dccp_getsockopt()
|
| /Linux-v5.15/tools/testing/selftests/bpf/progs/ |
| D | bpf_dctcp.c | 151 new_state != BPF_CORE_READ_BITFIELD(inet_csk(sk), icsk_ca_state)) in BPF_PROG() 184 if (inet_csk(sk)->icsk_ack.pending & ICSK_ACK_TIMER) { in dctcp_ece_ack_update() 188 inet_csk(sk)->icsk_ack.pending |= ICSK_ACK_NOW; in dctcp_ece_ack_update()
|
| /Linux-v5.15/net/mptcp/ |
| D | mptcp_diag.c | 24 return inet_sk_diag_fill(sk, inet_csk(sk), skb, cb, req, NLM_F_MULTI, in sk_diag_dump() 53 err = inet_sk_diag_fill(sk, inet_csk(sk), rep, cb, req, 0, in mptcp_diag_dump_one()
|
| /Linux-v5.15/tools/testing/selftests/bpf/ |
| D | bpf_tcp_helpers.h | 86 static __always_inline struct inet_connection_sock *inet_csk(const struct sock *sk) in inet_csk() function 93 return (void *)inet_csk(sk)->icsk_ca_priv; in inet_csk_ca()
|
| /Linux-v5.15/net/core/ |
| D | request_sock.c | 97 fastopenq = &inet_csk(lsk)->icsk_accept_queue.fastopenq; in reqsk_fastopen_remove()
|