Lines Matching refs:tcp_sk

145 	bool unknown_opt = tcp_sk(sk)->rx_opt.saw_unknown &&  in bpf_skops_parse_hdr()
146 BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr()
148 bool parse_all_opt = BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr()
248 tcp_sk(sk)->scaling_ratio = val ? val : 1; in tcp_measure_rcv_mss()
251 tcp_sk(sk)->advmss); in tcp_measure_rcv_mss()
288 len -= tcp_sk(sk)->tcp_header_len; in tcp_measure_rcv_mss()
304 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack()
344 tcp_sk(sk)->ecn_flags &= ~TCP_ECN_DEMAND_CWR; in tcp_ecn_accept_cwr()
362 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce()
394 if (tcp_sk(sk)->ecn_flags & TCP_ECN_OK) in tcp_ecn_check_ce()
424 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand()
483 const struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window()
520 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window()
560 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space()
596 struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_window()
624 const struct tcp_sock *tp = tcp_sk(sk); in tcp_initialize_rcv_mss()
699 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_rtt_measure_ts()
725 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_space_adjust()
793 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_data_recv()
847 struct tcp_sock *tp = tcp_sk(sk); in tcp_rtt_estimator()
915 const struct tcp_sock *tp = tcp_sk(sk); in tcp_update_pacing_rate()
952 const struct tcp_sock *tp = tcp_sk(sk); in tcp_set_rto()
1052 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reordering()
1105 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_skb_lost()
1270 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_dsack()
1368 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_one()
1453 struct tcp_sock *tp = tcp_sk(sk); in tcp_shifted_skb()
1566 struct tcp_sock *tp = tcp_sk(sk); in tcp_shift_skb_data()
1705 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_walk()
1829 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_write_queue()
2044 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_reno_reordering()
2060 struct tcp_sock *tp = tcp_sk(sk); in tcp_add_reno_sack()
2077 struct tcp_sock *tp = tcp_sk(sk); in tcp_remove_reno_sacks()
2125 struct tcp_sock *tp = tcp_sk(sk); in tcp_timeout_mark_lost()
2157 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_loss()
2214 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reneging()
2341 struct tcp_sock *tp = tcp_sk(sk); in tcp_time_to_recover()
2361 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_head_lost()
2407 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_scoreboard()
2461 const struct tcp_sock *tp = tcp_sk(sk); in tcp_any_retrans_done()
2477 struct tcp_sock *tp = tcp_sk(sk); in DBGUNDO()
2503 struct tcp_sock *tp = tcp_sk(sk); in tcp_undo_cwnd_reduction()
2537 struct tcp_sock *tp = tcp_sk(sk); in tcp_is_non_sack_preventing_reopen()
2553 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_recovery()
2582 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_dsack()
2598 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_loss()
2631 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_cwnd_reduction()
2645 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_reduction()
2671 struct tcp_sock *tp = tcp_sk(sk); in tcp_end_cwnd_reduction()
2688 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_cwr()
2701 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_keep_open()
2715 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_to_open()
2741 struct tcp_sock *tp = tcp_sk(sk); in tcp_mtup_probe_success()
2769 struct tcp_sock *tp = tcp_sk(sk); in tcp_simple_retransmit()
2821 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_recovery()
2848 struct tcp_sock *tp = tcp_sk(sk); in tcp_process_loss()
2900 struct tcp_sock *tp = tcp_sk(sk); in tcp_force_fast_retransmit()
2910 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_partial()
2942 struct tcp_sock *tp = tcp_sk(sk); in tcp_identify_packet_loss()
2975 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastretrans_alert()
3092 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rtt_min()
3109 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_update_rtt()
3171 tcp_sk(sk)->snd_cwnd_stamp = tcp_jiffies32; in tcp_cong_avoid()
3180 struct tcp_sock *tp = tcp_sk(sk); in tcp_rearm_rto()
3216 struct tcp_sock *tp = tcp_sk(sk); in tcp_tso_acked()
3245 before(shinfo->tskey, tcp_sk(sk)->snd_una)) { in tcp_ack_tstamp()
3262 struct tcp_sock *tp = tcp_sk(sk); in tcp_clean_rtx_queue()
3466 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_probe()
3501 if (tcp_sk(sk)->reordering > in tcp_may_raise_cwnd()
3573 struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_update_window()
3653 struct tcp_sock *tp = tcp_sk(sk); in tcp_send_challenge_ack()
3711 struct tcp_sock *tp = tcp_sk(sk); in tcp_process_tlp_ack()
3753 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_recovery()
3772 struct tcp_sock *tp = tcp_sk(sk); in tcp_newly_delivered()
3787 struct tcp_sock *tp = tcp_sk(sk); in tcp_ack()
4288 const struct tcp_sock *tp = tcp_sk(sk); in tcp_disordered_ack()
4309 const struct tcp_sock *tp = tcp_sk(sk); in tcp_paws_discard()
4390 struct tcp_sock *tp = tcp_sk(sk); in tcp_fin()
4471 struct tcp_sock *tp = tcp_sk(sk); in tcp_dsack_set()
4491 struct tcp_sock *tp = tcp_sk(sk); in tcp_dsack_extend()
4506 if (TCP_SKB_CB(skb)->seq == tcp_sk(sk)->duplicate_sack[0].start_seq && in tcp_rcv_spurious_retrans()
4513 struct tcp_sock *tp = tcp_sk(sk); in tcp_send_dupack()
4564 struct tcp_sock *tp = tcp_sk(sk); in tcp_sack_compress_send_ack()
4591 struct tcp_sock *tp = tcp_sk(sk); in tcp_sack_new_ofo_skb()
4752 struct tcp_sock *tp = tcp_sk(sk); in tcp_ofo_queue()
4819 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_queue_ofo()
4972 tcp_rcv_nxt_update(tcp_sk(sk), TCP_SKB_CB(skb)->end_seq); in tcp_queue_rcv()
5015 TCP_SKB_CB(skb)->seq = tcp_sk(sk)->rcv_nxt; in tcp_send_rcvq()
5017 TCP_SKB_CB(skb)->ack_seq = tcp_sk(sk)->snd_una - 1; in tcp_send_rcvq()
5040 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_queue()
5319 struct tcp_sock *tp = tcp_sk(sk); in tcp_collapse_ofo_queue()
5379 struct tcp_sock *tp = tcp_sk(sk); in tcp_prune_ofo_queue()
5433 struct tcp_sock *tp = tcp_sk(sk); in tcp_prune_queue()
5476 const struct tcp_sock *tp = tcp_sk(sk); in tcp_should_expand_sndbuf()
5511 struct tcp_sock *tp = tcp_sk(sk); in tcp_new_space()
5554 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ack_snd_check()
5632 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_urg()
5698 struct tcp_sock *tp = tcp_sk(sk); in tcp_urg()
5731 const struct tcp_sock *tp = tcp_sk(sk); in tcp_reset_check()
5744 struct tcp_sock *tp = tcp_sk(sk); in tcp_validate_incoming()
5886 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_established()
6074 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_transfer()
6101 struct tcp_sock *tp = tcp_sk(sk); in tcp_finish_connect()
6132 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_fastopen_synack()
6204 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_spurious_syn()
6221 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_synsent_state_process()
6462 tcp_sk(sk)->retrans_stamp = 0; in tcp_rcv_synrecv_state_fastopen()
6468 req = rcu_dereference_protected(tcp_sk(sk)->fastopen_rsk, in tcp_rcv_synrecv_state_fastopen()
6492 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_state_process()
6830 ireq->smc_ok = rx_opt->smc_ok && !(tcp_sk(sk)->smc_hs_congested && in tcp_openreq_init()
6831 tcp_sk(sk)->smc_hs_congested(sk)); in tcp_openreq_init()
6902 if (tcp_sk(sk)->save_syn) { in tcp_reqsk_record_syn()
6908 if (tcp_sk(sk)->save_syn == 2) { /* Save full header. */ in tcp_reqsk_record_syn()
6936 struct tcp_sock *tp = tcp_sk(sk); in tcp_get_syncookie_mss()
6966 struct tcp_sock *tp = tcp_sk(sk); in tcp_conn_request()