/Linux-v5.4/net/dccp/ |
D | timer.c | 33 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_write_timeout() local 37 if (icsk->icsk_retransmits != 0) in dccp_write_timeout() 39 retry_until = icsk->icsk_syn_retries ? in dccp_write_timeout() 42 if (icsk->icsk_retransmits >= sysctl_dccp_retries1) { in dccp_write_timeout() 72 if (icsk->icsk_retransmits >= retry_until) { in dccp_write_timeout() 85 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_retransmit_timer() local 98 if (icsk->icsk_retransmits == 0) in dccp_retransmit_timer() 106 if (--icsk->icsk_retransmits == 0) in dccp_retransmit_timer() 107 icsk->icsk_retransmits = 1; in dccp_retransmit_timer() 109 min(icsk->icsk_rto, in dccp_retransmit_timer() [all …]
|
D | output.c | 46 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_transmit_skb() local 73 if (icsk->icsk_retransmits == 0) in dccp_transmit_skb() 131 icsk->icsk_af_ops->send_check(sk, skb); in dccp_transmit_skb() 138 err = icsk->icsk_af_ops->queue_xmit(sk, skb, &inet->cork.fl); in dccp_transmit_skb() 161 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_sync_mss() local 167 cur_mps -= (icsk->icsk_af_ops->net_header_len + icsk->icsk_ext_hdr_len + in dccp_sync_mss() 187 icsk->icsk_pmtu_cookie = pmtu; in dccp_sync_mss() 537 struct inet_connection_sock *icsk = inet_csk(sk); in dccp_connect() local 564 icsk->icsk_retransmits = 0; in dccp_connect() 566 icsk->icsk_rto, DCCP_RTO_MAX); in dccp_connect() [all …]
|
D | diag.c | 19 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_get_info() local 24 info->tcpi_retransmits = icsk->icsk_retransmits; in dccp_get_info() 25 info->tcpi_probes = icsk->icsk_probes_out; in dccp_get_info() 26 info->tcpi_backoff = icsk->icsk_backoff; in dccp_get_info() 27 info->tcpi_pmtu = icsk->icsk_pmtu_cookie; in dccp_get_info()
|
D | minisocks.c | 38 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_time_wait() local 39 const int rto = (icsk->icsk_rto << 2) - (icsk->icsk_rto >> 1); in dccp_time_wait()
|
D | input.c | 401 const struct inet_connection_sock *icsk = inet_csk(sk); in dccp_rcv_request_sent_state_process() local 443 dccp_sync_mss(sk, icsk->icsk_pmtu_cookie); in dccp_rcv_request_sent_state_process() 472 icsk->icsk_af_ops->rebuild_header(sk); in dccp_rcv_request_sent_state_process() 480 icsk->icsk_accept_queue.rskq_defer_accept) { in dccp_rcv_request_sent_state_process()
|
/Linux-v5.4/net/ipv4/ |
D | tcp_timer.c | 28 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_rto_to_user_timeout() local 33 if (!icsk->icsk_user_timeout) in tcp_clamp_rto_to_user_timeout() 34 return icsk->icsk_rto; in tcp_clamp_rto_to_user_timeout() 36 remaining = icsk->icsk_user_timeout - elapsed; in tcp_clamp_rto_to_user_timeout() 40 return min_t(u32, icsk->icsk_rto, msecs_to_jiffies(remaining)); in tcp_clamp_rto_to_user_timeout() 142 static void tcp_mtu_probing(struct inet_connection_sock *icsk, struct sock *sk) in tcp_mtu_probing() argument 151 if (!icsk->icsk_mtup.enabled) { in tcp_mtu_probing() 152 icsk->icsk_mtup.enabled = 1; in tcp_mtu_probing() 153 icsk->icsk_mtup.probe_timestamp = tcp_jiffies32; in tcp_mtu_probing() 155 mss = tcp_mtu_to_mss(sk, icsk->icsk_mtup.search_low) >> 1; in tcp_mtu_probing() [all …]
|
D | tcp_ulp.c | 101 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_update_ulp() local 103 if (!icsk->icsk_ulp_ops) { in tcp_update_ulp() 108 if (icsk->icsk_ulp_ops->update) in tcp_update_ulp() 109 icsk->icsk_ulp_ops->update(sk, proto); in tcp_update_ulp() 114 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_ulp() local 120 if (!icsk->icsk_ulp_ops) in tcp_cleanup_ulp() 123 if (icsk->icsk_ulp_ops->release) in tcp_cleanup_ulp() 124 icsk->icsk_ulp_ops->release(sk); in tcp_cleanup_ulp() 125 module_put(icsk->icsk_ulp_ops->owner); in tcp_cleanup_ulp() 127 icsk->icsk_ulp_ops = NULL; in tcp_cleanup_ulp() [all …]
|
D | tcp_cong.c | 160 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_assign_congestion_control() local 167 icsk->icsk_ca_ops = ca; in tcp_assign_congestion_control() 170 memset(icsk->icsk_ca_priv, 0, sizeof(icsk->icsk_ca_priv)); in tcp_assign_congestion_control() 179 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_congestion_control() local 182 if (icsk->icsk_ca_ops->init) in tcp_init_congestion_control() 183 icsk->icsk_ca_ops->init(sk); in tcp_init_congestion_control() 193 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_reinit_congestion_control() local 196 icsk->icsk_ca_ops = ca; in tcp_reinit_congestion_control() 197 icsk->icsk_ca_setsockopt = 1; in tcp_reinit_congestion_control() 198 memset(icsk->icsk_ca_priv, 0, sizeof(icsk->icsk_ca_priv)); in tcp_reinit_congestion_control() [all …]
|
D | inet_connection_sock.c | 391 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_wait_for_connect() local 413 if (reqsk_queue_empty(&icsk->icsk_accept_queue)) in inet_csk_wait_for_connect() 418 if (!reqsk_queue_empty(&icsk->icsk_accept_queue)) in inet_csk_wait_for_connect() 439 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_accept() local 440 struct request_sock_queue *queue = &icsk->icsk_accept_queue; in inet_csk_accept() 508 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_init_xmit_timers() local 510 timer_setup(&icsk->icsk_retransmit_timer, retransmit_handler, 0); in inet_csk_init_xmit_timers() 511 timer_setup(&icsk->icsk_delack_timer, delack_handler, 0); in inet_csk_init_xmit_timers() 513 icsk->icsk_pending = icsk->icsk_ack.pending = 0; in inet_csk_init_xmit_timers() 519 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timers() local [all …]
|
D | tcp_output.c | 66 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_new_data_sent() local 76 if (!prior_packets || icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) in tcp_event_new_data_sent() 159 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_sent() local 170 if (before(tp->lsndtime, icsk->icsk_ack.lrcvtime) && in tcp_event_data_sent() 171 (u32)(now - icsk->icsk_ack.lrcvtime) < icsk->icsk_ack.ato) in tcp_event_data_sent() 1016 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_transmit_skb() local 1141 icsk->icsk_af_ops->send_check(sk, skb); in __tcp_transmit_skb() 1169 err = icsk->icsk_af_ops->queue_xmit(sk, skb, &inet->cork.fl); in __tcp_transmit_skb() 1462 const struct inet_connection_sock *icsk = inet_csk(sk); in __tcp_mtu_to_mss() local 1468 mss_now = pmtu - icsk->icsk_af_ops->net_header_len - sizeof(struct tcphdr); in __tcp_mtu_to_mss() [all …]
|
D | tcp_input.c | 118 void clean_acked_data_enable(struct inet_connection_sock *icsk, in clean_acked_data_enable() argument 121 icsk->icsk_clean_acked = cad; in clean_acked_data_enable() 126 void clean_acked_data_disable(struct inet_connection_sock *icsk) in clean_acked_data_disable() argument 129 icsk->icsk_clean_acked = NULL; in clean_acked_data_disable() 164 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() local 165 const unsigned int lss = icsk->icsk_ack.last_seg_size; in tcp_measure_rcv_mss() 168 icsk->icsk_ack.last_seg_size = 0; in tcp_measure_rcv_mss() 174 if (len >= icsk->icsk_ack.rcv_mss) { in tcp_measure_rcv_mss() 175 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len, in tcp_measure_rcv_mss() 178 if (unlikely(len > icsk->icsk_ack.rcv_mss + in tcp_measure_rcv_mss() [all …]
|
D | tcp.c | 410 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_sock() local 419 icsk->icsk_rto = TCP_TIMEOUT_INIT; in tcp_init_sock() 451 icsk->icsk_sync_mss = tcp_sync_mss; in tcp_init_sock() 1535 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cleanup_rbuf() local 1538 if (icsk->icsk_ack.blocked || in tcp_cleanup_rbuf() 1540 tp->rcv_nxt - tp->rcv_wup > icsk->icsk_ack.rcv_mss || in tcp_cleanup_rbuf() 1548 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED2) || in tcp_cleanup_rbuf() 1549 ((icsk->icsk_ack.pending & ICSK_ACK_PUSHED) && in tcp_cleanup_rbuf() 2564 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_disconnect() local 2616 icsk->icsk_backoff = 0; in tcp_disconnect() [all …]
|
D | tcp_diag.c | 115 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux() local 135 ulp_ops = icsk->icsk_ulp_ops; in tcp_diag_get_aux() 146 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_diag_get_aux_size() local 170 ulp_ops = icsk->icsk_ulp_ops; in tcp_diag_get_aux_size()
|
D | tcp_htcp.c | 84 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_rtt() local 92 if (icsk->icsk_ca_state == TCP_CA_Open) { in measure_rtt() 104 const struct inet_connection_sock *icsk = inet_csk(sk); in measure_achieved_throughput() local 109 if (icsk->icsk_ca_state == TCP_CA_Open) in measure_achieved_throughput() 119 if (!((1 << icsk->icsk_ca_state) & (TCPF_CA_Open | TCPF_CA_Disorder))) { in measure_achieved_throughput()
|
D | tcp_minisocks.c | 255 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_time_wait() local 264 const int rto = (icsk->icsk_rto << 2) - (icsk->icsk_rto >> 1); in tcp_time_wait() 408 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_openreq_child() local 418 icsk->icsk_ca_dst_locked = tcp_ca_dst_locked(dst); in tcp_ca_openreq_child() 419 icsk->icsk_ca_ops = ca; in tcp_ca_openreq_child() 427 (!icsk->icsk_ca_setsockopt || in tcp_ca_openreq_child() 428 !try_module_get(icsk->icsk_ca_ops->owner))) in tcp_ca_openreq_child()
|
D | inet_diag.c | 159 int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk, in inet_sk_diag_fill() argument 218 if (!icsk) { in inet_sk_diag_fill() 223 if (icsk->icsk_pending == ICSK_TIME_RETRANS || in inet_sk_diag_fill() 224 icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in inet_sk_diag_fill() 225 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in inet_sk_diag_fill() 227 r->idiag_retrans = icsk->icsk_retransmits; in inet_sk_diag_fill() 229 jiffies_to_msecs(icsk->icsk_timeout - jiffies); in inet_sk_diag_fill() 230 } else if (icsk->icsk_pending == ICSK_TIME_PROBE0) { in inet_sk_diag_fill() 232 r->idiag_retrans = icsk->icsk_probes_out; in inet_sk_diag_fill() 234 jiffies_to_msecs(icsk->icsk_timeout - jiffies); in inet_sk_diag_fill() [all …]
|
D | tcp_ipv4.c | 428 struct inet_connection_sock *icsk; in tcp_v4_err() local 481 icsk = inet_csk(sk); in tcp_v4_err() 530 if (seq != tp->snd_una || !icsk->icsk_retransmits || in tcp_v4_err() 531 !icsk->icsk_backoff || fastopen) in tcp_v4_err() 541 icsk->icsk_backoff--; in tcp_v4_err() 542 icsk->icsk_rto = tp->srtt_us ? __tcp_set_rto(tp) : in tcp_v4_err() 544 icsk->icsk_rto = inet_csk_rto_backoff(icsk, TCP_RTO_MAX); in tcp_v4_err() 549 remaining = icsk->icsk_rto - in tcp_v4_err() 2080 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v4_init_sock() local 2084 icsk->icsk_af_ops = &ipv4_specific; in tcp_v4_init_sock() [all …]
|
D | inet_timewait_sock.c | 105 const struct inet_connection_sock *icsk = inet_csk(sk); in inet_twsk_hashdance() local 116 tw->tw_tb = icsk->icsk_bind_hash; in inet_twsk_hashdance() 117 WARN_ON(!icsk->icsk_bind_hash); in inet_twsk_hashdance()
|
D | tcp_yeah.c | 63 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_yeah_pkts_acked() local 66 if (icsk->icsk_ca_state == TCP_CA_Open) in tcp_yeah_pkts_acked()
|
D | tcp_bic.c | 191 const struct inet_connection_sock *icsk = inet_csk(sk); in bictcp_acked() local 193 if (icsk->icsk_ca_state == TCP_CA_Open) { in bictcp_acked()
|
D | tcp_nv.c | 242 const struct inet_connection_sock *icsk = inet_csk(sk); in tcpnv_acked() local 256 if (icsk->icsk_ca_state != TCP_CA_Open && in tcpnv_acked() 257 icsk->icsk_ca_state != TCP_CA_Disorder) in tcpnv_acked()
|
/Linux-v5.4/include/net/ |
D | inet_connection_sock.h | 197 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_clear_xmit_timer() local 200 icsk->icsk_pending = 0; in inet_csk_clear_xmit_timer() 202 sk_stop_timer(sk, &icsk->icsk_retransmit_timer); in inet_csk_clear_xmit_timer() 205 icsk->icsk_ack.blocked = icsk->icsk_ack.pending = 0; in inet_csk_clear_xmit_timer() 207 sk_stop_timer(sk, &icsk->icsk_delack_timer); in inet_csk_clear_xmit_timer() 221 struct inet_connection_sock *icsk = inet_csk(sk); in inet_csk_reset_xmit_timer() local 232 icsk->icsk_pending = what; in inet_csk_reset_xmit_timer() 233 icsk->icsk_timeout = jiffies + when; in inet_csk_reset_xmit_timer() 234 sk_reset_timer(sk, &icsk->icsk_retransmit_timer, icsk->icsk_timeout); in inet_csk_reset_xmit_timer() 236 icsk->icsk_ack.pending |= ICSK_ACK_TIMER; in inet_csk_reset_xmit_timer() [all …]
|
D | tcp.h | 349 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_dec_quickack_mode() local 351 if (icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 352 if (pkts >= icsk->icsk_ack.quick) { in tcp_dec_quickack_mode() 353 icsk->icsk_ack.quick = 0; in tcp_dec_quickack_mode() 355 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_dec_quickack_mode() 357 icsk->icsk_ack.quick -= pkts; in tcp_dec_quickack_mode() 1100 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ca_needs_ecn() local 1102 return icsk->icsk_ca_ops->flags & TCP_CONG_NEEDS_ECN; in tcp_ca_needs_ecn() 1107 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_set_ca_state() local 1109 if (icsk->icsk_ca_ops->set_state) in tcp_set_ca_state() [all …]
|
/Linux-v5.4/net/ipv6/ |
D | tcp_ipv6.c | 149 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v6_connect() local 229 u32 exthdrlen = icsk->icsk_ext_hdr_len; in tcp_v6_connect() 239 icsk->icsk_af_ops = &ipv6_mapped; in tcp_v6_connect() 248 icsk->icsk_ext_hdr_len = exthdrlen; in tcp_v6_connect() 249 icsk->icsk_af_ops = &ipv6_specific; in tcp_v6_connect() 296 icsk->icsk_ext_hdr_len = 0; in tcp_v6_connect() 298 icsk->icsk_ext_hdr_len = opt->opt_flen + in tcp_v6_connect() 1805 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_v6_init_sock() local 1809 icsk->icsk_af_ops = &ipv6_specific; in tcp_v6_init_sock() 1866 const struct inet_connection_sock *icsk = inet_csk(sp); in get_tcp6_sock() local [all …]
|
D | inet6_hashtables.c | 123 struct inet_connection_sock *icsk; in inet6_lhash2_lookup() local 128 inet_lhash2_for_each_icsk_rcu(icsk, &ilb2->head) { in inet6_lhash2_lookup() 129 sk = (struct sock *)icsk; in inet6_lhash2_lookup()
|