/Linux-v4.19/net/ipv4/ |
D | tcp_fastopen.c | 124 ctx = rcu_dereference(sock_net(sk)->ipv4.tcp_fastopen_ctx); in __tcp_fastopen_cookie_gen() 299 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check() 316 return (sock_net(sk)->ipv4.sysctl_tcp_fastopen & flag) || in tcp_fastopen_no_cookie() 331 int tcp_fastopen = sock_net(sk)->ipv4.sysctl_tcp_fastopen; in tcp_try_fastopen() 336 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen() 366 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 370 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENPASSIVEFAIL); in tcp_try_fastopen() 372 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENPASSIVEFAIL); in tcp_try_fastopen() 455 struct net *net = sock_net(sk); in tcp_fastopen_active_disable() 468 unsigned int tfo_bh_timeout = sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout; in tcp_fastopen_active_should_disable() [all …]
|
D | syncookies.c | 300 if (!sock_net(sk)->ipv4.sysctl_tcp_syncookies || !th->ack || th->rst) in cookie_v4_check() 308 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v4_check() 312 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v4_check() 316 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v4_check() 319 tsoff = secure_tcp_ts_off(sock_net(sk), in cookie_v4_check() 325 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v4_check() 360 RCU_INIT_POINTER(ireq->ireq_opt, tcp_v4_save_options(sock_net(sk), skb)); in cookie_v4_check() 381 rt = ip_route_output_key(sock_net(sk), &fl4); in cookie_v4_check() 396 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), &rt->dst); in cookie_v4_check()
|
D | tcp_timer.c | 68 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err() 119 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources() 123 if (!check_net(sock_net(sk))) { in tcp_out_of_resources() 139 int retries = sock_net(sk)->ipv4.sysctl_tcp_orphan_retries; /* May be zero. */ in tcp_orphan_retries() 155 const struct net *net = sock_net(sk); in tcp_mtu_probing() 220 struct net *net = sock_net(sk); in tcp_write_timeout() 302 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler() 331 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer() 368 max_probes = sock_net(sk)->ipv4.sysctl_tcp_retries2; in tcp_probe_timer() 395 sock_net(sk)->ipv4.sysctl_tcp_synack_retries + 1; /* add one more retry for fastopen */ in tcp_fastopen_synack_timer() [all …]
|
D | tcp_input.c | 145 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn() 356 sk->sk_sndbuf = min(sndmem, sock_net(sk)->ipv4.sysctl_tcp_wmem[2]); in tcp_sndbuf_expand() 390 int window = tcp_win_from_space(sk, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]) >> 1; in __tcp_grow_window() 441 if (sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf) in tcp_fixup_rcvbuf() 445 sk->sk_rcvbuf = min(rcvmem, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_fixup_rcvbuf() 453 int tcp_app_win = sock_net(sk)->ipv4.sysctl_tcp_app_win; in tcp_init_buffer_space() 493 struct net *net = sock_net(sk); in tcp_clamp_window() 641 if (sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf && in tcp_rcv_space_adjust() 662 sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_rcv_space_adjust() 823 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio; in tcp_update_pacing_rate() [all …]
|
D | inet_hashtables.c | 52 return inet6_ehashfn(sock_net(sk), in sk_ehashfn() 56 return inet_ehashfn(sock_net(sk), in sk_ehashfn() 108 const int bhash = inet_bhashfn(sock_net(sk), inet_sk(sk)->inet_num, in __inet_put_port() 134 const int bhash = inet_bhashfn(sock_net(sk), port, in __inet_inherit_port() 152 if (net_eq(ib_net(tb), sock_net(sk)) && in __inet_inherit_port() 158 sock_net(sk), head, port); in __inet_inherit_port() 179 hash = ipv6_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 184 hash = ipv4_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 233 if (net_eq(sock_net(sk), net) && inet->inet_num == hnum && in compute_score() 448 struct net *net = sock_net(sk); in __inet_check_established() [all …]
|
D | tcp_minisocks.c | 257 struct inet_timewait_death_row *tcp_death_row = &sock_net(sk)->ipv4.tcp_death_row; in tcp_time_wait() 330 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEWAITOVERFLOW); in tcp_time_wait() 567 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child() 597 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req() 639 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 729 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 734 __NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req() 753 __TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req() 776 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req() 797 if (!sock_net(sk)->ipv4.sysctl_tcp_abort_on_overflow) { in tcp_check_req() [all …]
|
D | tcp_output.c | 67 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPORIGDATASENT, in tcp_event_new_data_sent() 169 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_event_ack_sent() 228 if (sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows) in tcp_select_initial_window() 236 space = max_t(u32, space, sock_net(sk)->ipv4.sysctl_tcp_rmem[2]); in tcp_select_initial_window() 276 NET_INC_STATS(sock_net(sk), in tcp_select_window() 287 sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows) in tcp_select_window() 299 NET_INC_STATS(sock_net(sk), in tcp_select_window() 302 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFROMZEROWINDOWADV); in tcp_select_window() 326 bool use_ecn = sock_net(sk)->ipv4.sysctl_tcp_ecn == 1 || in tcp_ecn_send_syn() 348 if (sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback) in tcp_ecn_clear_syn() [all …]
|
D | udp.c | 151 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse() 188 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse2() 211 struct net *net = sock_net(sk); in udp_reuseport_add_sock() 216 if (net_eq(sock_net(sk2), net) && in udp_reuseport_add_sock() 246 struct net *net = sock_net(sk); in udp_lib_get_port() 332 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in udp_lib_get_port() 358 ipv4_portaddr_hash(sock_net(sk), htonl(INADDR_ANY), snum); in udp_v4_get_port() 360 ipv4_portaddr_hash(sock_net(sk), inet_sk(sk)->inet_rcv_saddr, 0); in udp_v4_get_port() 375 if (!net_eq(sock_net(sk), net) || in compute_score() 574 if (!net_eq(sock_net(sk), net) || in __udp_is_mcast_sock() [all …]
|
D | raw_diag.c | 96 struct net *net = sock_net(in_skb->sk); in raw_diag_dump_one() 153 struct net *net = sock_net(skb->sk); in raw_diag_dump() 170 if (!net_eq(sock_net(sk), net)) in raw_diag_dump() 207 struct net *net = sock_net(in_skb->sk); in raw_diag_destroy()
|
D | udp_diag.c | 41 struct net *net = sock_net(in_skb->sk); in udp_dump_one() 102 struct net *net = sock_net(skb->sk); in udp_dump() 121 if (!net_eq(sock_net(sk), net)) in udp_dump() 175 struct net *net = sock_net(in_skb->sk); in __udp_diag_destroy()
|
D | tcp_recovery.c | 14 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPLOSTRETRANSMIT, in tcp_mark_skb_lost() 36 !(sock_net(sk)->ipv4.sysctl_tcp_recovery & TCP_RACK_NO_DUPTHRESH)) in tcp_rack_reo_wnd() 205 if (sock_net(sk)->ipv4.sysctl_tcp_recovery & TCP_RACK_STATIC_REO_WND || in tcp_rack_update_reo_wnd()
|
D | raw.c | 106 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash_sk() 119 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash_sk() 131 if (net_eq(sock_net(sk), net) && inet->inet_num == num && in __raw_v4_lookup() 354 struct net *net = sock_net(sk); in raw_send_hdrinc() 506 struct net *net = sock_net(sk); in raw_sendmsg() 625 ipc.oif == l3mdev_master_ifindex_by_index(sock_net(sk), in raw_sendmsg() 732 tb_id = l3mdev_fib_table_by_index(sock_net(sk), in raw_bind() 735 chk_addr_ret = inet_addr_type_table(sock_net(sk), addr->sin_addr.s_addr, in raw_bind() 1005 if (sock_net(sk) == seq_file_net(seq)) in raw_get_first() 1022 } while (sk && sock_net(sk) != seq_file_net(seq)); in raw_get_next()
|
/Linux-v4.19/net/can/ |
D | raw.c | 276 if (!net_eq(dev_net(dev), sock_net(sk))) in raw_notifier() 367 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_release() 373 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_release() 415 dev = dev_get_by_index(sock_net(sk), addr->can_ifindex); in raw_bind() 431 err = raw_enable_allfilters(sock_net(sk), dev, sk); in raw_bind() 437 err = raw_enable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 446 dev = dev_get_by_index(sock_net(sk), in raw_bind() 454 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 528 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_setsockopt() 533 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt() [all …]
|
/Linux-v4.19/net/ipv6/ |
D | syncookies.c | 149 if (!sock_net(sk)->ipv4.sysctl_tcp_syncookies || !th->ack || th->rst) in cookie_v6_check() 157 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v6_check() 161 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v6_check() 165 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v6_check() 168 tsoff = secure_tcpv6_ts_off(sock_net(sk), in cookie_v6_check() 174 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v6_check() 255 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), dst); in cookie_v6_check()
|
D | udp.c | 95 ipv6_portaddr_hash(sock_net(sk), &in6addr_any, snum); in udp_v6_get_port() 97 ipv6_portaddr_hash(sock_net(sk), &sk->sk_v6_rcv_saddr, 0); in udp_v6_get_port() 106 u16 new_hash = ipv6_portaddr_hash(sock_net(sk), in udp_v6_rehash() 121 if (!net_eq(sock_net(sk), net) || in compute_score() 384 UDP_INC_STATS(sock_net(sk), UDP_MIB_INERRORS, in udpv6_recvmsg() 387 UDP6_INC_STATS(sock_net(sk), UDP_MIB_INERRORS, in udpv6_recvmsg() 395 UDP_INC_STATS(sock_net(sk), UDP_MIB_INDATAGRAMS, in udpv6_recvmsg() 398 UDP6_INC_STATS(sock_net(sk), UDP_MIB_INDATAGRAMS, in udpv6_recvmsg() 447 UDP_INC_STATS(sock_net(sk), in udpv6_recvmsg() 449 UDP_INC_STATS(sock_net(sk), in udpv6_recvmsg() [all …]
|
/Linux-v4.19/net/vmw_vsock/ |
D | diag.c | 67 net = sock_net(skb->sk); in vsock_diag_dump() 87 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 121 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 152 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
|
/Linux-v4.19/net/ieee802154/ |
D | socket.c | 145 dev_load(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 146 dev = dev_get_by_name(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 190 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash() 200 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash() 226 dev = ieee802154_get_dev(sock_net(sk), &addr); in raw_bind() 268 dev = dev_getfirstbyhwtype(sock_net(sk), ARPHRD_IEEE802154); in raw_sendmsg() 270 dev = dev_get_by_index(sock_net(sk), sk->sk_bound_dev_if); in raw_sendmsg() 472 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in dgram_hash() 482 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in dgram_unhash() 519 dev = ieee802154_get_dev(sock_net(sk), &haddr); in dgram_bind() [all …]
|
/Linux-v4.19/net/sctp/ |
D | endpointola.c | 60 struct net *net = sock_net(sk); in sctp_endpoint_init() 308 net_eq(sock_net(ep->base.sk), net)) { in sctp_endpoint_is_match() 357 struct net *net = sock_net(ep->base.sk); in sctp_endpoint_is_peeled_off() 396 net = sock_net(sk); in sctp_endpoint_bh_rcv() 448 SCTP_INC_STATS(sock_net(ep->base.sk), SCTP_MIB_INCTRLCHUNKS); in sctp_endpoint_bh_rcv()
|
/Linux-v4.19/net/unix/ |
D | diag.c | 185 struct net *net = sock_net(skb->sk); in unix_diag_dump() 200 if (!net_eq(sock_net(sk), net)) in unix_diag_dump() 251 struct net *net = sock_net(in_skb->sk); in unix_diag_get_exact() 260 if (!net_eq(sock_net(sk), net)) in unix_diag_get_exact() 298 struct net *net = sock_net(skb->sk); in unix_diag_handler_dump()
|
/Linux-v4.19/net/phonet/ |
D | socket.c | 92 if (!net_eq(sock_net(sknode), net)) in pn_find_sock_by_sa() 129 if (!net_eq(sock_net(sknode), net)) in pn_deliver_sock_broadcast() 188 if (saddr && phonet_address_lookup(sock_net(sk), saddr)) in pn_socket_bind() 385 dev = dev_get_by_index(sock_net(sk), in pn_socket_ioctl() 388 dev = phonet_device_get(sock_net(sk)); in pn_socket_ioctl() 497 struct net *net = sock_net(sk); in pn_sock_get_port() 552 if (!net_eq(net, sock_net(sknode))) in pn_sock_get_idx() 569 while (sk && !net_eq(net, sock_net(sk))); in pn_sock_get_next() 659 if (!net_eq(sock_net(sk), &init_net)) in pn_sock_bind_res()
|
D | pn_netlink.c | 67 struct net *net = sock_net(skb->sk); in addr_doit() 141 pndevs = phonet_device_list(sock_net(skb->sk)); in getaddr_dumpit() 234 struct net *net = sock_net(skb->sk); in route_doit() 278 struct net *net = sock_net(skb->sk); in route_dumpit()
|
/Linux-v4.19/net/netlink/ |
D | diag.c | 93 struct net *net = sock_net(skb->sk); in __netlink_diag_dump() 132 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump() 158 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump() 226 struct net *net = sock_net(skb->sk); in netlink_diag_handler_dump()
|
D | af_netlink.c | 288 if (!net_eq(dev_net(dev), sock_net(sk))) in __netlink_deliver_tap_skb() 343 netlink_deliver_tap(sock_net(dst), skb); in netlink_deliver_tap_kernel() 495 !net_eq(sock_net(&nlk->sk), read_pnet(&x->pnet)); in netlink_compare() 520 netlink_compare_arg_init(&arg, sock_net(sk), nlk_sk(sk)->portid); in __netlink_insert() 771 nlk->netlink_unbind(sock_net(sk), i + 1); in netlink_release() 784 .net = sock_net(sk), in netlink_release() 813 sock_prot_inuse_add(sock_net(sk), &netlink_proto, -1); in netlink_release() 822 struct net *net = sock_net(sk); in netlink_autobind() 919 return netlink_ns_capable(skb, sock_net(skb->sk)->user_ns, cap); in netlink_net_capable() 926 ns_capable(sock_net(sock->sk)->user_ns, CAP_NET_ADMIN); in netlink_allowed() [all …]
|
/Linux-v4.19/net/tipc/ |
D | socket.c | 246 u32 onode = tipc_own_addr(sock_net(sk)); in tipc_sk_respond() 253 tipc_node_xmit_skb(sock_net(sk), skb, dnode, selector); in tipc_sk_respond() 292 u32 self = tipc_own_addr(sock_net(sk)); in tsk_peer_msg() 501 struct net *net = sock_net(sk); in __tipc_shutdown() 682 addr->addr.id.node = tipc_own_addr(sock_net(sk)); in tipc_getname() 768 struct net *net = sock_net(sk); in tipc_sendmcast() 882 struct net *net = sock_net(sk); in tipc_send_group_unicast() 929 struct net *net = sock_net(sk); in tipc_send_group_anycast() 1005 struct net *net = sock_net(sk); in tipc_send_group_bcast() 1081 struct net *net = sock_net(sk); in tipc_send_group_mcast() [all …]
|
/Linux-v4.19/net/openvswitch/ |
D | datapath.c | 530 struct net *net = sock_net(skb->sk); in ovs_packet_cmd_execute() 884 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_new() 1122 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_set() 1226 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_get() 1252 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex); in ovs_flow_cmd_get() 1285 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_del() 1307 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex); in ovs_flow_cmd_del() 1345 netlink_set_err(sock_net(skb->sk)->genl_sock, 0, 0, PTR_ERR(reply)); in ovs_flow_cmd_del() 1372 dp = get_dp_rcu(sock_net(skb->sk), ovs_header->dp_ifindex); in ovs_flow_cmd_dump() 1531 dp = lookup_datapath(sock_net(skb->sk), info->userhdr, info->attrs); in ovs_dp_reset_user_features() [all …]
|