Lines Matching refs:sock_net
319 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check()
336 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie()
351 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen()
357 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen()
376 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
394 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
399 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
403 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
491 struct net *net = sock_net(sk); in tcp_fastopen_active_disable()
493 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout)) in tcp_fastopen_active_disable()
515 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout); in tcp_fastopen_active_should_disable()
523 tfo_da_times = atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times); in tcp_fastopen_active_should_disable()
534 timeout = READ_ONCE(sock_net(sk)->ipv4.tfo_active_disable_stamp) + in tcp_fastopen_active_should_disable()
570 atomic_read(&sock_net(sk)->ipv4.tfo_active_disable_times)) { in tcp_fastopen_active_disable_ofo_check()
573 atomic_set(&sock_net(sk)->ipv4.tfo_active_disable_times, 0); in tcp_fastopen_active_disable_ofo_check()
591 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENACTIVEFAIL); in tcp_fastopen_active_detect_blackhole()