Lines Matching refs:vs
93 static inline bool vxlan_collect_metadata(struct vxlan_sock *vs) in vxlan_collect_metadata() argument
95 return vs->flags & VXLAN_F_COLLECT_METADATA || in vxlan_collect_metadata()
164 static inline struct hlist_head *vni_head(struct vxlan_sock *vs, __be32 vni) in vni_head() argument
166 return &vs->vni_list[hash_32((__force u32)vni, VNI_HASH_BITS)]; in vni_head()
201 struct vxlan_sock *vs; in vxlan_find_sock() local
205 hlist_for_each_entry_rcu(vs, vs_head(net, port), hlist) { in vxlan_find_sock()
206 if (inet_sk(vs->sock->sk)->inet_sport == port && in vxlan_find_sock()
207 vxlan_get_sk_family(vs) == family && in vxlan_find_sock()
208 vs->flags == flags && in vxlan_find_sock()
209 vs->sock->sk->sk_bound_dev_if == ifindex) in vxlan_find_sock()
210 return vs; in vxlan_find_sock()
215 static struct vxlan_dev *vxlan_vs_find_vni(struct vxlan_sock *vs, int ifindex, in vxlan_vs_find_vni() argument
221 if (vs->flags & VXLAN_F_COLLECT_METADATA) in vxlan_vs_find_vni()
224 hlist_for_each_entry_rcu(node, vni_head(vs, vni), hlist) { in vxlan_vs_find_vni()
247 struct vxlan_sock *vs; in vxlan_find_vni() local
249 vs = vxlan_find_sock(net, family, port, flags, ifindex); in vxlan_find_vni()
250 if (!vs) in vxlan_find_vni()
253 return vxlan_vs_find_vni(vs, ifindex, vni); in vxlan_find_vni()
768 struct vxlan_sock *vs = rcu_dereference_sk_user_data(sk); in vxlan_gro_receive() local
787 if ((flags & VXLAN_HF_RCO) && (vs->flags & VXLAN_F_REMCSUM_RX)) { in vxlan_gro_receive()
790 !!(vs->flags & in vxlan_gro_receive()
1571 static bool __vxlan_sock_release_prep(struct vxlan_sock *vs) in __vxlan_sock_release_prep() argument
1575 if (!vs) in __vxlan_sock_release_prep()
1577 if (!refcount_dec_and_test(&vs->refcnt)) in __vxlan_sock_release_prep()
1580 vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); in __vxlan_sock_release_prep()
1582 hlist_del_rcu(&vs->hlist); in __vxlan_sock_release_prep()
1583 udp_tunnel_notify_del_rx_port(vs->sock, in __vxlan_sock_release_prep()
1584 (vs->flags & VXLAN_F_GPE) ? in __vxlan_sock_release_prep()
1771 struct vxlan_sock *vs, in vxlan_set_mac() argument
1786 if (vxlan_get_sk_family(vs) == AF_INET) { in vxlan_set_mac()
1803 static bool vxlan_ecn_decapsulate(struct vxlan_sock *vs, void *oiph, in vxlan_ecn_decapsulate() argument
1808 if (vxlan_get_sk_family(vs) == AF_INET) in vxlan_ecn_decapsulate()
1816 if (vxlan_get_sk_family(vs) == AF_INET) in vxlan_ecn_decapsulate()
1831 struct vxlan_sock *vs; in vxlan_rcv() local
1856 vs = rcu_dereference_sk_user_data(sk); in vxlan_rcv()
1857 if (!vs) in vxlan_rcv()
1862 vxlan = vxlan_vs_find_vni(vs, skb->dev->ifindex, vni); in vxlan_rcv()
1869 if (vs->flags & VXLAN_F_GPE) { in vxlan_rcv()
1870 if (!vxlan_parse_gpe_hdr(&unparsed, &protocol, skb, vs->flags)) in vxlan_rcv()
1879 if (vs->flags & VXLAN_F_REMCSUM_RX) in vxlan_rcv()
1880 if (unlikely(!vxlan_remcsum(&unparsed, skb, vs->flags))) in vxlan_rcv()
1883 if (vxlan_collect_metadata(vs)) { in vxlan_rcv()
1886 tun_dst = udp_tun_rx_dst(skb, vxlan_get_sk_family(vs), TUNNEL_KEY, in vxlan_rcv()
1899 if (vs->flags & VXLAN_F_GBP) in vxlan_rcv()
1900 vxlan_parse_gbp_hdr(&unparsed, skb, vs->flags, md); in vxlan_rcv()
1918 if (!vxlan_set_mac(vxlan, vs, skb, vni)) in vxlan_rcv()
1929 if (!vxlan_ecn_decapsulate(vs, oiph, skb)) { in vxlan_rcv()
1960 struct vxlan_sock *vs; in vxlan_err_lookup() local
1972 vs = rcu_dereference_sk_user_data(sk); in vxlan_err_lookup()
1973 if (!vs) in vxlan_err_lookup()
1977 vxlan = vxlan_vs_find_vni(vs, skb->dev->ifindex, vni); in vxlan_err_lookup()
3028 static void vxlan_vs_add_dev(struct vxlan_sock *vs, struct vxlan_dev *vxlan, in vxlan_vs_add_dev() argument
3036 hlist_add_head_rcu(&node->hlist, vni_head(vs, vni)); in vxlan_vs_add_dev()
3261 struct vxlan_sock *vs; in vxlan_offload_rx_ports() local
3268 hlist_for_each_entry_rcu(vs, &vn->sock_list[i], hlist) { in vxlan_offload_rx_ports()
3271 if (vs->flags & VXLAN_F_GPE) in vxlan_offload_rx_ports()
3277 udp_tunnel_push_rx_port(dev, vs->sock, type); in vxlan_offload_rx_ports()
3279 udp_tunnel_drop_rx_port(dev, vs->sock, type); in vxlan_offload_rx_ports()
3509 struct vxlan_sock *vs; in vxlan_socket_create() local
3514 vs = kzalloc(sizeof(*vs), GFP_KERNEL); in vxlan_socket_create()
3515 if (!vs) in vxlan_socket_create()
3519 INIT_HLIST_HEAD(&vs->vni_list[h]); in vxlan_socket_create()
3523 kfree(vs); in vxlan_socket_create()
3527 vs->sock = sock; in vxlan_socket_create()
3528 refcount_set(&vs->refcnt, 1); in vxlan_socket_create()
3529 vs->flags = (flags & VXLAN_F_RCV_FLAGS); in vxlan_socket_create()
3532 hlist_add_head_rcu(&vs->hlist, vs_head(net, port)); in vxlan_socket_create()
3534 (vs->flags & VXLAN_F_GPE) ? in vxlan_socket_create()
3541 tunnel_cfg.sk_user_data = vs; in vxlan_socket_create()
3551 return vs; in vxlan_socket_create()
3557 struct vxlan_sock *vs = NULL; in __vxlan_sock_add() local
3567 vs = vxlan_find_sock(vxlan->net, ipv6 ? AF_INET6 : AF_INET, in __vxlan_sock_add()
3570 if (vs && !refcount_inc_not_zero(&vs->refcnt)) { in __vxlan_sock_add()
3576 if (!vs) in __vxlan_sock_add()
3577 vs = vxlan_socket_create(vxlan->net, ipv6, in __vxlan_sock_add()
3580 if (IS_ERR(vs)) in __vxlan_sock_add()
3581 return PTR_ERR(vs); in __vxlan_sock_add()
3584 rcu_assign_pointer(vxlan->vn6_sock, vs); in __vxlan_sock_add()
3589 rcu_assign_pointer(vxlan->vn4_sock, vs); in __vxlan_sock_add()
3592 vxlan_vs_add_dev(vs, vxlan, node); in __vxlan_sock_add()