Lines Matching refs:vn
1474 struct vxlan_net *vn; in __vxlan_sock_release_prep() local
1481 vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); in __vxlan_sock_release_prep()
1482 spin_lock(&vn->sock_lock); in __vxlan_sock_release_prep()
1488 spin_unlock(&vn->sock_lock); in __vxlan_sock_release_prep()
2888 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_del_dev() local
2890 spin_lock(&vn->sock_lock); in vxlan_vs_del_dev()
2895 spin_unlock(&vn->sock_lock); in vxlan_vs_del_dev()
2901 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_dev() local
2905 spin_lock(&vn->sock_lock); in vxlan_vs_add_dev()
2907 spin_unlock(&vn->sock_lock); in vxlan_vs_add_dev()
3131 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vxlan_offload_rx_ports() local
3134 spin_lock(&vn->sock_lock); in vxlan_offload_rx_ports()
3136 hlist_for_each_entry_rcu(vs, &vn->sock_list[i], hlist) { in vxlan_offload_rx_ports()
3150 spin_unlock(&vn->sock_lock); in vxlan_offload_rx_ports()
3377 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vxlan_socket_create() local
3400 spin_lock(&vn->sock_lock); in vxlan_socket_create()
3406 spin_unlock(&vn->sock_lock); in vxlan_socket_create()
3425 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in __vxlan_sock_add() local
3436 spin_lock(&vn->sock_lock); in __vxlan_sock_add()
3441 spin_unlock(&vn->sock_lock); in __vxlan_sock_add()
3444 spin_unlock(&vn->sock_lock); in __vxlan_sock_add()
3497 struct vxlan_net *vn = net_generic(src_net, vxlan_net_id); in vxlan_vni_in_use() local
3500 list_for_each_entry(tmp, &vn->vxlan_list, next) { in vxlan_vni_in_use()
3751 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in __vxlan_dev_create() local
3815 list_add(&vxlan->next, &vn->vxlan_list); in __vxlan_dev_create()
4432 static void vxlan_handle_lowerdev_unregister(struct vxlan_net *vn, in vxlan_handle_lowerdev_unregister() argument
4438 list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) { in vxlan_handle_lowerdev_unregister()
4458 struct vxlan_net *vn = net_generic(dev_net(dev), vxlan_net_id); in vxlan_netdevice_event() local
4461 vxlan_handle_lowerdev_unregister(vn, dev); in vxlan_netdevice_event()
4640 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vxlan_init_net() local
4643 INIT_LIST_HEAD(&vn->vxlan_list); in vxlan_init_net()
4644 spin_lock_init(&vn->sock_lock); in vxlan_init_net()
4645 vn->nexthop_notifier_block.notifier_call = vxlan_nexthop_event; in vxlan_init_net()
4648 INIT_HLIST_HEAD(&vn->sock_list[h]); in vxlan_init_net()
4650 return register_nexthop_notifier(net, &vn->nexthop_notifier_block, in vxlan_init_net()
4656 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vxlan_destroy_tunnels() local
4664 list_for_each_entry_safe(vxlan, next, &vn->vxlan_list, next) { in vxlan_destroy_tunnels()
4681 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vxlan_exit_batch_net() local
4683 unregister_nexthop_notifier(net, &vn->nexthop_notifier_block); in vxlan_exit_batch_net()
4693 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vxlan_exit_batch_net() local
4696 WARN_ON_ONCE(!hlist_empty(&vn->sock_list[h])); in vxlan_exit_batch_net()