Lines Matching refs:vs
88 static inline bool vxlan_collect_metadata(struct vxlan_sock *vs) in vxlan_collect_metadata() argument
90 return vs->flags & VXLAN_F_COLLECT_METADATA || in vxlan_collect_metadata()
159 static inline struct hlist_head *vni_head(struct vxlan_sock *vs, __be32 vni) in vni_head() argument
161 return &vs->vni_list[hash_32((__force u32)vni, VNI_HASH_BITS)]; in vni_head()
191 struct vxlan_sock *vs; in vxlan_find_sock() local
195 hlist_for_each_entry_rcu(vs, vs_head(net, port), hlist) { in vxlan_find_sock()
196 if (inet_sk(vs->sock->sk)->inet_sport == port && in vxlan_find_sock()
197 vxlan_get_sk_family(vs) == family && in vxlan_find_sock()
198 vs->flags == flags && in vxlan_find_sock()
199 vs->sock->sk->sk_bound_dev_if == ifindex) in vxlan_find_sock()
200 return vs; in vxlan_find_sock()
205 static struct vxlan_dev *vxlan_vs_find_vni(struct vxlan_sock *vs, int ifindex, in vxlan_vs_find_vni() argument
211 if (vs->flags & VXLAN_F_COLLECT_METADATA) in vxlan_vs_find_vni()
214 hlist_for_each_entry_rcu(node, vni_head(vs, vni), hlist) { in vxlan_vs_find_vni()
237 struct vxlan_sock *vs; in vxlan_find_vni() local
239 vs = vxlan_find_sock(net, family, port, flags, ifindex); in vxlan_find_vni()
240 if (!vs) in vxlan_find_vni()
243 return vxlan_vs_find_vni(vs, ifindex, vni); in vxlan_find_vni()
736 struct vxlan_sock *vs = rcu_dereference_sk_user_data(sk); in vxlan_gro_receive() local
755 if ((flags & VXLAN_HF_RCO) && (vs->flags & VXLAN_F_REMCSUM_RX)) { in vxlan_gro_receive()
758 !!(vs->flags & in vxlan_gro_receive()
1393 static bool __vxlan_sock_release_prep(struct vxlan_sock *vs) in __vxlan_sock_release_prep() argument
1397 if (!vs) in __vxlan_sock_release_prep()
1399 if (!refcount_dec_and_test(&vs->refcnt)) in __vxlan_sock_release_prep()
1402 vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); in __vxlan_sock_release_prep()
1404 hlist_del_rcu(&vs->hlist); in __vxlan_sock_release_prep()
1405 udp_tunnel_notify_del_rx_port(vs->sock, in __vxlan_sock_release_prep()
1406 (vs->flags & VXLAN_F_GPE) ? in __vxlan_sock_release_prep()
1593 struct vxlan_sock *vs, in vxlan_set_mac() argument
1608 if (vxlan_get_sk_family(vs) == AF_INET) { in vxlan_set_mac()
1625 static bool vxlan_ecn_decapsulate(struct vxlan_sock *vs, void *oiph, in vxlan_ecn_decapsulate() argument
1630 if (vxlan_get_sk_family(vs) == AF_INET) in vxlan_ecn_decapsulate()
1638 if (vxlan_get_sk_family(vs) == AF_INET) in vxlan_ecn_decapsulate()
1654 struct vxlan_sock *vs; in vxlan_rcv() local
1679 vs = rcu_dereference_sk_user_data(sk); in vxlan_rcv()
1680 if (!vs) in vxlan_rcv()
1685 vxlan = vxlan_vs_find_vni(vs, skb->dev->ifindex, vni); in vxlan_rcv()
1692 if (vs->flags & VXLAN_F_GPE) { in vxlan_rcv()
1693 if (!vxlan_parse_gpe_hdr(&unparsed, &protocol, skb, vs->flags)) in vxlan_rcv()
1702 if (vxlan_collect_metadata(vs)) { in vxlan_rcv()
1705 tun_dst = udp_tun_rx_dst(skb, vxlan_get_sk_family(vs), TUNNEL_KEY, in vxlan_rcv()
1718 if (vs->flags & VXLAN_F_REMCSUM_RX) in vxlan_rcv()
1719 if (!vxlan_remcsum(&unparsed, skb, vs->flags)) in vxlan_rcv()
1721 if (vs->flags & VXLAN_F_GBP) in vxlan_rcv()
1722 vxlan_parse_gbp_hdr(&unparsed, skb, vs->flags, md); in vxlan_rcv()
1740 if (!vxlan_set_mac(vxlan, vs, skb, vni)) in vxlan_rcv()
1751 if (!vxlan_ecn_decapsulate(vs, oiph, skb)) { in vxlan_rcv()
1787 struct vxlan_sock *vs; in vxlan_err_lookup() local
1799 vs = rcu_dereference_sk_user_data(sk); in vxlan_err_lookup()
1800 if (!vs) in vxlan_err_lookup()
1804 vxlan = vxlan_vs_find_vni(vs, skb->dev->ifindex, vni); in vxlan_err_lookup()
2769 static void vxlan_vs_add_dev(struct vxlan_sock *vs, struct vxlan_dev *vxlan, in vxlan_vs_add_dev() argument
2777 hlist_add_head_rcu(&node->hlist, vni_head(vs, vni)); in vxlan_vs_add_dev()
2991 struct vxlan_sock *vs; in vxlan_offload_rx_ports() local
2998 hlist_for_each_entry_rcu(vs, &vn->sock_list[i], hlist) { in vxlan_offload_rx_ports()
3001 if (vs->flags & VXLAN_F_GPE) in vxlan_offload_rx_ports()
3007 udp_tunnel_push_rx_port(dev, vs->sock, type); in vxlan_offload_rx_ports()
3009 udp_tunnel_drop_rx_port(dev, vs->sock, type); in vxlan_offload_rx_ports()
3219 struct vxlan_sock *vs; in vxlan_socket_create() local
3224 vs = kzalloc(sizeof(*vs), GFP_KERNEL); in vxlan_socket_create()
3225 if (!vs) in vxlan_socket_create()
3229 INIT_HLIST_HEAD(&vs->vni_list[h]); in vxlan_socket_create()
3233 kfree(vs); in vxlan_socket_create()
3237 vs->sock = sock; in vxlan_socket_create()
3238 refcount_set(&vs->refcnt, 1); in vxlan_socket_create()
3239 vs->flags = (flags & VXLAN_F_RCV_FLAGS); in vxlan_socket_create()
3242 hlist_add_head_rcu(&vs->hlist, vs_head(net, port)); in vxlan_socket_create()
3244 (vs->flags & VXLAN_F_GPE) ? in vxlan_socket_create()
3251 tunnel_cfg.sk_user_data = vs; in vxlan_socket_create()
3261 return vs; in vxlan_socket_create()
3267 struct vxlan_sock *vs = NULL; in __vxlan_sock_add() local
3277 vs = vxlan_find_sock(vxlan->net, ipv6 ? AF_INET6 : AF_INET, in __vxlan_sock_add()
3280 if (vs && !refcount_inc_not_zero(&vs->refcnt)) { in __vxlan_sock_add()
3286 if (!vs) in __vxlan_sock_add()
3287 vs = vxlan_socket_create(vxlan->net, ipv6, in __vxlan_sock_add()
3290 if (IS_ERR(vs)) in __vxlan_sock_add()
3291 return PTR_ERR(vs); in __vxlan_sock_add()
3294 rcu_assign_pointer(vxlan->vn6_sock, vs); in __vxlan_sock_add()
3299 rcu_assign_pointer(vxlan->vn4_sock, vs); in __vxlan_sock_add()
3302 vxlan_vs_add_dev(vs, vxlan, node); in __vxlan_sock_add()