Lines Matching refs:icsk
120 void clean_acked_data_enable(struct inet_connection_sock *icsk, in clean_acked_data_enable() argument
123 icsk->icsk_clean_acked = cad; in clean_acked_data_enable()
128 void clean_acked_data_disable(struct inet_connection_sock *icsk) in clean_acked_data_disable() argument
131 icsk->icsk_clean_acked = NULL; in clean_acked_data_disable()
229 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_measure_rcv_mss() local
230 const unsigned int lss = icsk->icsk_ack.last_seg_size; in tcp_measure_rcv_mss()
233 icsk->icsk_ack.last_seg_size = 0; in tcp_measure_rcv_mss()
239 if (len >= icsk->icsk_ack.rcv_mss) { in tcp_measure_rcv_mss()
240 icsk->icsk_ack.rcv_mss = min_t(unsigned int, len, in tcp_measure_rcv_mss()
243 if (unlikely(len > icsk->icsk_ack.rcv_mss + in tcp_measure_rcv_mss()
266 icsk->icsk_ack.last_seg_size = len; in tcp_measure_rcv_mss()
268 icsk->icsk_ack.rcv_mss = len; in tcp_measure_rcv_mss()
272 if (icsk->icsk_ack.pending & ICSK_ACK_PUSHED) in tcp_measure_rcv_mss()
273 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED2; in tcp_measure_rcv_mss()
274 icsk->icsk_ack.pending |= ICSK_ACK_PUSHED; in tcp_measure_rcv_mss()
280 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_incr_quickack() local
281 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack()
286 if (quickacks > icsk->icsk_ack.quick) in tcp_incr_quickack()
287 icsk->icsk_ack.quick = quickacks; in tcp_incr_quickack()
292 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_quickack_mode() local
296 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_enter_quickack_mode()
306 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_quickack_mode() local
310 (icsk->icsk_ack.quick && !inet_csk_in_pingpong_mode(sk)); in tcp_in_quickack_mode()
567 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clamp_window() local
570 icsk->icsk_ack.quick = 0; in tcp_clamp_window()
768 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_event_data_recv() local
779 if (!icsk->icsk_ack.ato) { in tcp_event_data_recv()
784 icsk->icsk_ack.ato = TCP_ATO_MIN; in tcp_event_data_recv()
786 int m = now - icsk->icsk_ack.lrcvtime; in tcp_event_data_recv()
790 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + TCP_ATO_MIN / 2; in tcp_event_data_recv()
791 } else if (m < icsk->icsk_ack.ato) { in tcp_event_data_recv()
792 icsk->icsk_ack.ato = (icsk->icsk_ack.ato >> 1) + m; in tcp_event_data_recv()
793 if (icsk->icsk_ack.ato > icsk->icsk_rto) in tcp_event_data_recv()
794 icsk->icsk_ack.ato = icsk->icsk_rto; in tcp_event_data_recv()
795 } else if (m > icsk->icsk_rto) { in tcp_event_data_recv()
803 icsk->icsk_ack.lrcvtime = now; in tcp_event_data_recv()
2128 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_enter_loss() local
2131 bool new_recovery = icsk->icsk_ca_state < TCP_CA_Recovery; in tcp_enter_loss()
2136 if (icsk->icsk_ca_state <= TCP_CA_Disorder || in tcp_enter_loss()
2138 (icsk->icsk_ca_state == TCP_CA_Loss && !icsk->icsk_retransmits)) { in tcp_enter_loss()
2141 tp->snd_ssthresh = icsk->icsk_ca_ops->ssthresh(sk); in tcp_enter_loss()
2152 if (icsk->icsk_ca_state <= TCP_CA_Disorder && in tcp_enter_loss()
2165 (new_recovery || icsk->icsk_retransmits) && in tcp_enter_loss()
2483 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_undo_cwnd_reduction() local
2485 tp->snd_cwnd = icsk->icsk_ca_ops->undo_cwnd(sk); in tcp_undo_cwnd_reduction()
2688 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_failed() local
2690 icsk->icsk_mtup.search_high = icsk->icsk_mtup.probe_size - 1; in tcp_mtup_probe_failed()
2691 icsk->icsk_mtup.probe_size = 0; in tcp_mtup_probe_failed()
2698 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_mtup_probe_success() local
2704 icsk->icsk_mtup.probe_size; in tcp_mtup_probe_success()
2709 icsk->icsk_mtup.search_low = icsk->icsk_mtup.probe_size; in tcp_mtup_probe_success()
2710 icsk->icsk_mtup.probe_size = 0; in tcp_mtup_probe_success()
2711 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_mtup_probe_success()
2721 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_simple_retransmit() local
2761 if (icsk->icsk_ca_state != TCP_CA_Loss) { in tcp_simple_retransmit()
2927 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_fastretrans_alert() local
2951 if (icsk->icsk_ca_state == TCP_CA_Open) { in tcp_fastretrans_alert()
2955 switch (icsk->icsk_ca_state) { in tcp_fastretrans_alert()
2976 switch (icsk->icsk_ca_state) { in tcp_fastretrans_alert()
2988 if (icsk->icsk_ca_state != TCP_CA_Recovery) { in tcp_fastretrans_alert()
3000 if (!(icsk->icsk_ca_state == TCP_CA_Open || in tcp_fastretrans_alert()
3012 if (icsk->icsk_ca_state <= TCP_CA_Disorder) in tcp_fastretrans_alert()
3022 if (icsk->icsk_ca_state < TCP_CA_CWR && in tcp_fastretrans_alert()
3023 icsk->icsk_mtup.probe_size && in tcp_fastretrans_alert()
3121 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_avoid() local
3123 icsk->icsk_ca_ops->cong_avoid(sk, ack, acked); in tcp_cong_avoid()
3132 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rearm_rto() local
3146 if (icsk->icsk_pending == ICSK_TIME_REO_TIMEOUT || in tcp_rearm_rto()
3147 icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) { in tcp_rearm_rto()
3213 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_clean_rtx_queue() local
3346 if (unlikely(icsk->icsk_mtup.probe_size && in tcp_clean_rtx_queue()
3382 if (icsk->icsk_ca_ops->pkts_acked) { in tcp_clean_rtx_queue()
3387 icsk->icsk_ca_ops->pkts_acked(sk, &sample); in tcp_clean_rtx_queue()
3395 icsk = inet_csk(sk); in tcp_clean_rtx_queue()
3398 tp->lost_out, icsk->icsk_ca_state); in tcp_clean_rtx_queue()
3403 tp->sacked_out, icsk->icsk_ca_state); in tcp_clean_rtx_queue()
3408 tp->retrans_out, icsk->icsk_ca_state); in tcp_clean_rtx_queue()
3418 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack_probe() local
3426 icsk->icsk_backoff = 0; in tcp_ack_probe()
3427 icsk->icsk_probes_tstamp = 0; in tcp_ack_probe()
3469 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_cong_control() local
3471 if (icsk->icsk_ca_ops->cong_control) { in tcp_cong_control()
3472 icsk->icsk_ca_ops->cong_control(sk, rs); in tcp_cong_control()
3685 const struct inet_connection_sock *icsk = inet_csk(sk); in tcp_in_ack_event() local
3687 if (icsk->icsk_ca_ops->in_ack_event) in tcp_in_ack_event()
3688 icsk->icsk_ca_ops->in_ack_event(sk, flags); in tcp_in_ack_event()
3730 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_ack() local
3773 icsk->icsk_retransmits = 0; in tcp_ack()
3777 if (icsk->icsk_clean_acked) in tcp_ack()
3778 icsk->icsk_clean_acked(sk, ack); in tcp_ack()
3846 icsk->icsk_probes_out = 0; in tcp_ack()
5946 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_init_transfer() local
5950 icsk->icsk_af_ops->rebuild_header(sk); in tcp_init_transfer()
5967 if (!icsk->icsk_ca_initialized) in tcp_init_transfer()
5975 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_finish_connect() local
5978 icsk->icsk_ack.lrcvtime = tcp_jiffies32; in tcp_finish_connect()
5981 icsk->icsk_af_ops->sk_rx_dst_set(sk, skb); in tcp_finish_connect()
6093 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_synsent_state_process() local
6115 if (icsk->icsk_retransmits == 0) in tcp_rcv_synsent_state_process()
6192 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_rcv_synsent_state_process()
6216 icsk->icsk_accept_queue.rskq_defer_accept || in tcp_rcv_synsent_state_process()
6286 tcp_sync_mss(sk, icsk->icsk_pmtu_cookie); in tcp_rcv_synsent_state_process()
6364 struct inet_connection_sock *icsk = inet_csk(sk); in tcp_rcv_state_process() local
6389 acceptable = icsk->icsk_af_ops->conn_request(sk, skb) >= 0; in tcp_rcv_state_process()