Lines Matching refs:sock_net

83 	NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPORIGDATASENT,  in tcp_event_new_data_sent()
185 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_event_ack_sent()
229 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows)) in tcp_select_initial_window()
240 space = max_t(u32, space, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_select_initial_window()
259 struct net *net = sock_net(sk); in tcp_select_window()
333 bool use_ecn = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn) == 1 || in tcp_ecn_send_syn()
355 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback)) in tcp_ecn_clear_syn()
800 if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_timestamps) && !*md5)) { in tcp_syn_options()
806 if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_window_scaling))) { in tcp_syn_options()
811 if (likely(READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_sack))) { in tcp_syn_options()
1398 TCP_ADD_STATS(sock_net(sk), TCP_MIB_OUTSEGS, in __tcp_transmit_skb()
1569 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPWQUEUETOOBIG); in tcp_fragment()
1723 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_min_snd_mss)); in __tcp_mtu_to_mss()
1764 struct net *net = sock_net(sk); in tcp_mtup_init()
1906 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_slow_start_after_idle) && in tcp_cwnd_validate()
1984 r = tcp_min_rtt(tcp_sk(sk)) >> READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tso_rtt_log); in tcp_tso_autosize()
2003 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_min_tso_segs); in tcp_tso_segs()
2223 win_divisor = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_tso_win_divisor); in tcp_tso_should_defer()
2285 struct net *net = sock_net(sk); in tcp_mtu_check_reprobe()
2391 struct net *net = sock_net(sk); in tcp_mtu_probe()
2578 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_limit_output_bytes)); in tcp_small_queue_check()
2812 early_retrans = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_early_retrans); in tcp_schedule_loss_probe()
2859 NET_INC_STATS(sock_net(sk), in skb_still_in_host_queue()
2926 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPLOSSPROBES); in tcp_send_loss_probe()
3020 struct net *net = sock_net(sk); in __tcp_select_window()
3217 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_retrans_collapse)) in tcp_retrans_try_collapse()
3322 TCP_ADD_STATS(sock_net(sk), TCP_MIB_RETRANSSEGS, segs); in __tcp_retransmit_skb()
3324 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSYNRETRANS); in __tcp_retransmit_skb()
3366 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPRETRANSFAIL, segs); in __tcp_retransmit_skb()
3459 NET_ADD_STATS(sock_net(sk), mib_idx, tcp_skb_pcount(skb)); in tcp_xmit_retransmit_queue()
3554 TCP_INC_STATS(sock_net(sk), TCP_MIB_OUTRSTS); in tcp_send_active_reset()
3559 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTFAILED); in tcp_send_active_reset()
3570 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTFAILED); in tcp_send_active_reset()
3718 TCP_INC_STATS(sock_net(sk), TCP_MIB_OUTSEGS); in tcp_make_synack()
3769 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_timestamps)) in tcp_connect_init()
3800 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_window_scaling), in tcp_connect_init()
3925 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPORIGDATASENT); in tcp_send_syn_data()
3991 TCP_INC_STATS(sock_net(sk), TCP_MIB_ACTIVEOPENS); in tcp_connect()
4133 NET_INC_STATS(sock_net(sk), mib); in tcp_xmit_probe_skb()
4198 struct net *net = sock_net(sk); in tcp_send_probe0()
4240 TCP_INC_STATS(sock_net(sk), TCP_MIB_RETRANSSEGS); in tcp_rtx_synack()
4241 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPSYNRETRANS); in tcp_rtx_synack()