Lines Matching refs:sock_net
216 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn()
451 min(sndmem, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[2]))); in tcp_sndbuf_expand()
486 int window = tcp_win_from_space(sk, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])) >> 1; in __tcp_grow_window()
559 int tcp_app_win = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_app_win); in tcp_init_buffer_space()
598 struct net *net = sock_net(sk); in tcp_clamp_window()
750 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf) && in tcp_rcv_space_adjust()
766 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_rcv_space_adjust()
930 rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio); in tcp_update_pacing_rate()
932 rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio); in tcp_update_pacing_rate()
1071 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering)); in tcp_check_sack_reordering()
1076 NET_INC_STATS(sock_net(sk), in tcp_check_sack_reordering()
1116 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPLOSTRETRANSMIT, in tcp_mark_skb_lost()
1276 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECV); in tcp_check_dsack()
1283 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKOFORECV); in tcp_check_dsack()
1290 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKIGNOREDDUBIOUS); in tcp_check_dsack()
1294 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPDSACKRECVSEGS, dup_segs); in tcp_check_dsack()
1497 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTED); in tcp_shifted_skb()
1524 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKMERGED); in tcp_shifted_skb()
1695 NET_INC_STATS(sock_net(sk), LINUX_MIB_SACKSHIFTFALLBACK); in tcp_shift_skb_data()
1887 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_sacktag_write_queue()
2050 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering)); in tcp_check_reno_reordering()
2052 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRENOREORDER); in tcp_check_reno_reordering()
2115 return READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_recovery) & in tcp_is_rack()
2132 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSACKRENEGING); in tcp_timeout_mark_lost()
2158 struct net *net = sock_net(sk); in tcp_enter_loss()
2568 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_try_undo_recovery()
2589 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDSACKUNDO); in tcp_try_undo_dsack()
2604 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPLOSSUNDO); in tcp_try_undo_loss()
2606 NET_INC_STATS(sock_net(sk), in tcp_try_undo_loss()
2736 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPFAIL); in tcp_mtup_probe_failed()
2759 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMTUPSUCCESS); in tcp_mtup_probe_success()
2829 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_enter_recovery()
2931 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPARTIALUNDO); in tcp_try_undo_partial()
3091 u32 wlen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_min_rtt_wlen) * HZ; in tcp_update_rtt_min()
3502 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_reordering)) in tcp_may_raise_cwnd()
3654 struct net *net = sock_net(sk); in tcp_send_challenge_ack()
3730 NET_INC_STATS(sock_net(sk), in tcp_process_tlp_ack()
3771 const struct net *net = sock_net(sk); in tcp_newly_delivered()
3859 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPACKS); in tcp_ack()
3866 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPUREACKS); in tcp_ack()
4473 if (tcp_is_sack(tp) && READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_dsack)) { in tcp_dsack_set()
4481 NET_INC_STATS(sock_net(sk), mib_idx); in tcp_dsack_set()
4508 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDUPLICATEDATAREHASH); in tcp_rcv_spurious_retrans()
4517 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_send_dupack()
4520 if (tcp_is_sack(tp) && READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_dsack)) { in tcp_send_dupack()
4576 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_sack_compress_send_ack()
4709 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOALESCE); in tcp_try_coalesce()
4828 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFODROP); in tcp_data_queue_ofo()
4839 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOQUEUE); in tcp_data_queue_ofo()
4891 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4911 NET_INC_STATS(sock_net(sk), in tcp_data_queue_ofo()
4941 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOMERGE); in tcp_data_queue_ofo()
5007 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_send_rcvq()
5070 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
5085 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVQDROP); in tcp_data_queue()
5123 NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOST); in tcp_data_queue()
5150 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPZEROWINDOWDROP); in tcp_data_queue()
5179 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPRCVCOLLAPSED); in tcp_collapse_one()
5412 NET_INC_STATS(sock_net(sk), LINUX_MIB_OFOPRUNED); in tcp_prune_ofo_queue()
5435 NET_INC_STATS(sock_net(sk), LINUX_MIB_PRUNECALLED); in tcp_prune_queue()
5467 NET_INC_STATS(sock_net(sk), LINUX_MIB_RCVPRUNED); in tcp_prune_queue()
5581 tp->compressed_ack >= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_comp_sack_nr)) in __tcp_ack_snd_check()
5603 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_comp_sack_delay_ns), in __tcp_ack_snd_check()
5607 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_comp_sack_slack_ns), in __tcp_ack_snd_check()
5635 if (ptr && !READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_stdurg)) in tcp_check_urg()
5748 if (tcp_fast_parse_options(sock_net(sk), skb, th, tp) && in tcp_validate_incoming()
5754 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_validate_incoming()
5755 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5777 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_validate_incoming()
5838 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_validate_incoming()
5839 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSYNCHALLENGE); in tcp_validate_incoming()
5974 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
5998 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPHPHITS); in tcp_rcv_established()
6063 TCP_INC_STATS(sock_net(sk), TCP_MIB_CSUMERRORS); in tcp_rcv_established()
6064 TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS); in tcp_rcv_established()
6143 tcp_parse_options(sock_net(sk), synack, &opt, 0, NULL); in tcp_rcv_fastopen_synack()
6175 NET_INC_STATS(sock_net(sk), in tcp_rcv_fastopen_synack()
6181 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENACTIVE); in tcp_rcv_fastopen_synack()
6227 tcp_parse_options(sock_net(sk), skb, &tp->rx_opt, 0, &foc); in tcp_rcv_synsent_state_process()
6253 NET_INC_STATS(sock_net(sk), in tcp_rcv_synsent_state_process()
6649 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6658 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6722 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONDATA); in tcp_rcv_state_process()
6789 const struct net *net = sock_net(listen_sk); in tcp_ecn_create_request()
6851 write_pnet(&ireq->ireq_net, sock_net(sk_listener)); in inet_reqsk_alloc()
6867 struct net *net = sock_net(sk); in tcp_syn_flood_action()
6877 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDOCOOKIES); in tcp_syn_flood_action()
6880 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPREQQFULLDROP); in tcp_syn_flood_action()
6939 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) != 2 && in tcp_get_syncookie_mss()
6947 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_get_syncookie_mss()
6967 struct net *net = sock_net(sk); in tcp_conn_request()
6988 NET_INC_STATS(sock_net(sk), LINUX_MIB_LISTENOVERFLOWS); in tcp_conn_request()
7006 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, in tcp_conn_request()