Lines Matching refs:tcp_sk

145 	bool unknown_opt = tcp_sk(sk)->rx_opt.saw_unknown &&  in bpf_skops_parse_hdr()
146 BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr()
148 bool parse_all_opt = BPF_SOCK_OPS_TEST_FLAG(tcp_sk(sk), in bpf_skops_parse_hdr()
241 tcp_sk(sk)->advmss); in tcp_measure_rcv_mss()
265 len -= tcp_sk(sk)->tcp_header_len; in tcp_measure_rcv_mss()
281 unsigned int quickacks = tcp_sk(sk)->rcv_wnd / (2 * icsk->icsk_ack.rcv_mss); in tcp_incr_quickack()
322 tcp_sk(sk)->ecn_flags &= ~TCP_ECN_DEMAND_CWR; in tcp_ecn_accept_cwr()
340 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ecn_check_ce()
372 if (tcp_sk(sk)->ecn_flags & TCP_ECN_OK) in tcp_ecn_check_ce()
402 const struct tcp_sock *tp = tcp_sk(sk); in tcp_sndbuf_expand()
461 struct tcp_sock *tp = tcp_sk(sk); in __tcp_grow_window()
498 struct tcp_sock *tp = tcp_sk(sk); in tcp_grow_window()
538 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_buffer_space()
574 struct tcp_sock *tp = tcp_sk(sk); in tcp_clamp_window()
602 const struct tcp_sock *tp = tcp_sk(sk); in tcp_initialize_rcv_mss()
677 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_rtt_measure_ts()
703 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_space_adjust()
776 struct tcp_sock *tp = tcp_sk(sk); in tcp_event_data_recv()
830 struct tcp_sock *tp = tcp_sk(sk); in tcp_rtt_estimator()
898 const struct tcp_sock *tp = tcp_sk(sk); in tcp_update_pacing_rate()
935 const struct tcp_sock *tp = tcp_sk(sk); in tcp_set_rto()
1035 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reordering()
1088 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_skb_lost()
1253 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_dsack()
1351 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_one()
1436 struct tcp_sock *tp = tcp_sk(sk); in tcp_shifted_skb()
1549 struct tcp_sock *tp = tcp_sk(sk); in tcp_shift_skb_data()
1688 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_walk()
1812 struct tcp_sock *tp = tcp_sk(sk); in tcp_sacktag_write_queue()
2027 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_reno_reordering()
2043 struct tcp_sock *tp = tcp_sk(sk); in tcp_add_reno_sack()
2060 struct tcp_sock *tp = tcp_sk(sk); in tcp_remove_reno_sacks()
2108 struct tcp_sock *tp = tcp_sk(sk); in tcp_timeout_mark_lost()
2140 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_loss()
2197 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_sack_reneging()
2323 struct tcp_sock *tp = tcp_sk(sk); in tcp_time_to_recover()
2343 struct tcp_sock *tp = tcp_sk(sk); in tcp_mark_head_lost()
2389 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_scoreboard()
2443 const struct tcp_sock *tp = tcp_sk(sk); in tcp_any_retrans_done()
2459 struct tcp_sock *tp = tcp_sk(sk); in DBGUNDO()
2485 struct tcp_sock *tp = tcp_sk(sk); in tcp_undo_cwnd_reduction()
2519 struct tcp_sock *tp = tcp_sk(sk); in tcp_is_non_sack_preventing_reopen()
2535 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_recovery()
2564 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_dsack()
2580 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_loss()
2613 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_cwnd_reduction()
2627 struct tcp_sock *tp = tcp_sk(sk); in tcp_cwnd_reduction()
2653 struct tcp_sock *tp = tcp_sk(sk); in tcp_end_cwnd_reduction()
2670 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_cwr()
2683 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_keep_open()
2697 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_to_open()
2723 struct tcp_sock *tp = tcp_sk(sk); in tcp_mtup_probe_success()
2751 struct tcp_sock *tp = tcp_sk(sk); in tcp_simple_retransmit()
2803 struct tcp_sock *tp = tcp_sk(sk); in tcp_enter_recovery()
2830 struct tcp_sock *tp = tcp_sk(sk); in tcp_process_loss()
2882 struct tcp_sock *tp = tcp_sk(sk); in tcp_force_fast_retransmit()
2892 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_partial()
2924 struct tcp_sock *tp = tcp_sk(sk); in tcp_identify_packet_loss()
2957 struct tcp_sock *tp = tcp_sk(sk); in tcp_fastretrans_alert()
3074 struct tcp_sock *tp = tcp_sk(sk); in tcp_update_rtt_min()
3091 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_update_rtt()
3153 tcp_sk(sk)->snd_cwnd_stamp = tcp_jiffies32; in tcp_cong_avoid()
3162 struct tcp_sock *tp = tcp_sk(sk); in tcp_rearm_rto()
3198 struct tcp_sock *tp = tcp_sk(sk); in tcp_tso_acked()
3227 before(shinfo->tskey, tcp_sk(sk)->snd_una)) { in tcp_ack_tstamp()
3244 struct tcp_sock *tp = tcp_sk(sk); in tcp_clean_rtx_queue()
3448 const struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_probe()
3483 if (tcp_sk(sk)->reordering > in tcp_may_raise_cwnd()
3555 struct tcp_sock *tp = tcp_sk(sk); in tcp_ack_update_window()
3629 struct tcp_sock *tp = tcp_sk(sk); in tcp_send_challenge_ack()
3686 struct tcp_sock *tp = tcp_sk(sk); in tcp_process_tlp_ack()
3728 struct tcp_sock *tp = tcp_sk(sk); in tcp_xmit_recovery()
3747 struct tcp_sock *tp = tcp_sk(sk); in tcp_newly_delivered()
3762 struct tcp_sock *tp = tcp_sk(sk); in tcp_ack()
4264 const struct tcp_sock *tp = tcp_sk(sk); in tcp_disordered_ack()
4285 const struct tcp_sock *tp = tcp_sk(sk); in tcp_paws_discard()
4360 struct tcp_sock *tp = tcp_sk(sk); in tcp_fin()
4441 struct tcp_sock *tp = tcp_sk(sk); in tcp_dsack_set()
4461 struct tcp_sock *tp = tcp_sk(sk); in tcp_dsack_extend()
4476 if (TCP_SKB_CB(skb)->seq == tcp_sk(sk)->duplicate_sack[0].start_seq && in tcp_rcv_spurious_retrans()
4483 struct tcp_sock *tp = tcp_sk(sk); in tcp_send_dupack()
4534 struct tcp_sock *tp = tcp_sk(sk); in tcp_sack_compress_send_ack()
4561 struct tcp_sock *tp = tcp_sk(sk); in tcp_sack_new_ofo_skb()
4722 struct tcp_sock *tp = tcp_sk(sk); in tcp_ofo_queue()
4789 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_queue_ofo()
4942 tcp_rcv_nxt_update(tcp_sk(sk), TCP_SKB_CB(skb)->end_seq); in tcp_queue_rcv()
4985 TCP_SKB_CB(skb)->seq = tcp_sk(sk)->rcv_nxt; in tcp_send_rcvq()
4987 TCP_SKB_CB(skb)->ack_seq = tcp_sk(sk)->snd_una - 1; in tcp_send_rcvq()
5010 struct tcp_sock *tp = tcp_sk(sk); in tcp_data_queue()
5283 struct tcp_sock *tp = tcp_sk(sk); in tcp_collapse_ofo_queue()
5341 struct tcp_sock *tp = tcp_sk(sk); in tcp_prune_ofo_queue()
5386 struct tcp_sock *tp = tcp_sk(sk); in tcp_prune_queue()
5429 const struct tcp_sock *tp = tcp_sk(sk); in tcp_should_expand_sndbuf()
5464 struct tcp_sock *tp = tcp_sk(sk); in tcp_new_space()
5507 struct tcp_sock *tp = tcp_sk(sk); in __tcp_ack_snd_check()
5585 struct tcp_sock *tp = tcp_sk(sk); in tcp_check_urg()
5651 struct tcp_sock *tp = tcp_sk(sk); in tcp_urg()
5684 struct tcp_sock *tp = tcp_sk(sk); in tcp_reset_check()
5697 struct tcp_sock *tp = tcp_sk(sk); in tcp_validate_incoming()
5837 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_established()
6025 struct tcp_sock *tp = tcp_sk(sk); in tcp_init_transfer()
6052 struct tcp_sock *tp = tcp_sk(sk); in tcp_finish_connect()
6083 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_fastopen_synack()
6155 struct tcp_sock *tp = tcp_sk(sk); in tcp_try_undo_spurious_syn()
6172 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_synsent_state_process()
6413 tcp_sk(sk)->retrans_stamp = 0; in tcp_rcv_synrecv_state_fastopen()
6419 req = rcu_dereference_protected(tcp_sk(sk)->fastopen_rsk, in tcp_rcv_synrecv_state_fastopen()
6443 struct tcp_sock *tp = tcp_sk(sk); in tcp_rcv_state_process()
6781 ireq->smc_ok = rx_opt->smc_ok && !(tcp_sk(sk)->smc_hs_congested && in tcp_openreq_init()
6782 tcp_sk(sk)->smc_hs_congested(sk)); in tcp_openreq_init()
6845 if (tcp_sk(sk)->save_syn) { in tcp_reqsk_record_syn()
6851 if (tcp_sk(sk)->save_syn == 2) { /* Save full header. */ in tcp_reqsk_record_syn()
6879 struct tcp_sock *tp = tcp_sk(sk); in tcp_get_syncookie_mss()
6909 struct tcp_sock *tp = tcp_sk(sk); in tcp_conn_request()