|
@@ -2105,6 +2105,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
|
|
src_port = udp_flow_src_port(dev_net(dev), skb, vxlan->cfg.port_min,
|
|
src_port = udp_flow_src_port(dev_net(dev), skb, vxlan->cfg.port_min,
|
|
vxlan->cfg.port_max, true);
|
|
vxlan->cfg.port_max, true);
|
|
|
|
|
|
|
|
+ rcu_read_lock();
|
|
if (dst->sa.sa_family == AF_INET) {
|
|
if (dst->sa.sa_family == AF_INET) {
|
|
struct vxlan_sock *sock4 = rcu_dereference(vxlan->vn4_sock);
|
|
struct vxlan_sock *sock4 = rcu_dereference(vxlan->vn4_sock);
|
|
struct rtable *rt;
|
|
struct rtable *rt;
|
|
@@ -2127,7 +2128,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
|
|
dst_port, vni, &rt->dst,
|
|
dst_port, vni, &rt->dst,
|
|
rt->rt_flags);
|
|
rt->rt_flags);
|
|
if (err)
|
|
if (err)
|
|
- return;
|
|
|
|
|
|
+ goto out_unlock;
|
|
} else if (info->key.tun_flags & TUNNEL_DONT_FRAGMENT) {
|
|
} else if (info->key.tun_flags & TUNNEL_DONT_FRAGMENT) {
|
|
df = htons(IP_DF);
|
|
df = htons(IP_DF);
|
|
}
|
|
}
|
|
@@ -2166,7 +2167,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
|
|
dst_port, vni, ndst,
|
|
dst_port, vni, ndst,
|
|
rt6i_flags);
|
|
rt6i_flags);
|
|
if (err)
|
|
if (err)
|
|
- return;
|
|
|
|
|
|
+ goto out_unlock;
|
|
}
|
|
}
|
|
|
|
|
|
tos = ip_tunnel_ecn_encap(tos, old_iph, skb);
|
|
tos = ip_tunnel_ecn_encap(tos, old_iph, skb);
|
|
@@ -2183,6 +2184,8 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
|
|
label, src_port, dst_port, !udp_sum);
|
|
label, src_port, dst_port, !udp_sum);
|
|
#endif
|
|
#endif
|
|
}
|
|
}
|
|
|
|
+out_unlock:
|
|
|
|
+ rcu_read_unlock();
|
|
return;
|
|
return;
|
|
|
|
|
|
drop:
|
|
drop:
|
|
@@ -2191,6 +2194,7 @@ drop:
|
|
return;
|
|
return;
|
|
|
|
|
|
tx_error:
|
|
tx_error:
|
|
|
|
+ rcu_read_unlock();
|
|
if (err == -ELOOP)
|
|
if (err == -ELOOP)
|
|
dev->stats.collisions++;
|
|
dev->stats.collisions++;
|
|
else if (err == -ENETUNREACH)
|
|
else if (err == -ENETUNREACH)
|