|
@@ -32,42 +32,56 @@ static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb,
|
|
netdev_features_t features),
|
|
netdev_features_t features),
|
|
__be16 new_protocol, bool is_ipv6)
|
|
__be16 new_protocol, bool is_ipv6)
|
|
{
|
|
{
|
|
|
|
+ int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb);
|
|
struct sk_buff *segs = ERR_PTR(-EINVAL);
|
|
struct sk_buff *segs = ERR_PTR(-EINVAL);
|
|
|
|
+ bool remcsum, need_csum, offload_csum;
|
|
|
|
+ struct udphdr *uh = udp_hdr(skb);
|
|
u16 mac_offset = skb->mac_header;
|
|
u16 mac_offset = skb->mac_header;
|
|
- int mac_len = skb->mac_len;
|
|
|
|
- int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb);
|
|
|
|
__be16 protocol = skb->protocol;
|
|
__be16 protocol = skb->protocol;
|
|
- netdev_features_t enc_features;
|
|
|
|
|
|
+ u16 mac_len = skb->mac_len;
|
|
int udp_offset, outer_hlen;
|
|
int udp_offset, outer_hlen;
|
|
- unsigned int oldlen;
|
|
|
|
- bool need_csum = !!(skb_shinfo(skb)->gso_type &
|
|
|
|
- SKB_GSO_UDP_TUNNEL_CSUM);
|
|
|
|
- bool remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM);
|
|
|
|
- bool offload_csum = false, dont_encap = (need_csum || remcsum);
|
|
|
|
-
|
|
|
|
- oldlen = (u16)~skb->len;
|
|
|
|
|
|
+ u32 partial;
|
|
|
|
|
|
if (unlikely(!pskb_may_pull(skb, tnl_hlen)))
|
|
if (unlikely(!pskb_may_pull(skb, tnl_hlen)))
|
|
goto out;
|
|
goto out;
|
|
|
|
|
|
|
|
+ /* adjust partial header checksum to negate old length */
|
|
|
|
+ partial = (__force u32)uh->check + (__force u16)~uh->len;
|
|
|
|
+
|
|
|
|
+ /* setup inner skb. */
|
|
skb->encapsulation = 0;
|
|
skb->encapsulation = 0;
|
|
__skb_pull(skb, tnl_hlen);
|
|
__skb_pull(skb, tnl_hlen);
|
|
skb_reset_mac_header(skb);
|
|
skb_reset_mac_header(skb);
|
|
skb_set_network_header(skb, skb_inner_network_offset(skb));
|
|
skb_set_network_header(skb, skb_inner_network_offset(skb));
|
|
skb->mac_len = skb_inner_network_offset(skb);
|
|
skb->mac_len = skb_inner_network_offset(skb);
|
|
skb->protocol = new_protocol;
|
|
skb->protocol = new_protocol;
|
|
|
|
+
|
|
|
|
+ need_csum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM);
|
|
skb->encap_hdr_csum = need_csum;
|
|
skb->encap_hdr_csum = need_csum;
|
|
|
|
+
|
|
|
|
+ remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM);
|
|
skb->remcsum_offload = remcsum;
|
|
skb->remcsum_offload = remcsum;
|
|
|
|
|
|
/* Try to offload checksum if possible */
|
|
/* Try to offload checksum if possible */
|
|
offload_csum = !!(need_csum &&
|
|
offload_csum = !!(need_csum &&
|
|
- ((skb->dev->features & NETIF_F_HW_CSUM) ||
|
|
|
|
- (skb->dev->features & (is_ipv6 ?
|
|
|
|
- NETIF_F_IPV6_CSUM : NETIF_F_IP_CSUM))));
|
|
|
|
|
|
+ (skb->dev->features &
|
|
|
|
+ (is_ipv6 ? (NETIF_F_HW_CSUM | NETIF_F_IPV6_CSUM) :
|
|
|
|
+ (NETIF_F_HW_CSUM | NETIF_F_IP_CSUM))));
|
|
|
|
+
|
|
|
|
+ features &= skb->dev->hw_enc_features;
|
|
|
|
+
|
|
|
|
+ /* The only checksum offload we care about from here on out is the
|
|
|
|
+ * outer one so strip the existing checksum feature flags and
|
|
|
|
+ * instead set the flag based on our outer checksum offload value.
|
|
|
|
+ */
|
|
|
|
+ if (remcsum) {
|
|
|
|
+ features &= ~NETIF_F_CSUM_MASK;
|
|
|
|
+ if (offload_csum)
|
|
|
|
+ features |= NETIF_F_HW_CSUM;
|
|
|
|
+ }
|
|
|
|
|
|
/* segment inner packet. */
|
|
/* segment inner packet. */
|
|
- enc_features = skb->dev->hw_enc_features & features;
|
|
|
|
- segs = gso_inner_segment(skb, enc_features);
|
|
|
|
|
|
+ segs = gso_inner_segment(skb, features);
|
|
if (IS_ERR_OR_NULL(segs)) {
|
|
if (IS_ERR_OR_NULL(segs)) {
|
|
skb_gso_error_unwind(skb, protocol, tnl_hlen, mac_offset,
|
|
skb_gso_error_unwind(skb, protocol, tnl_hlen, mac_offset,
|
|
mac_len);
|
|
mac_len);
|
|
@@ -78,17 +92,13 @@ static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb,
|
|
udp_offset = outer_hlen - tnl_hlen;
|
|
udp_offset = outer_hlen - tnl_hlen;
|
|
skb = segs;
|
|
skb = segs;
|
|
do {
|
|
do {
|
|
- struct udphdr *uh;
|
|
|
|
- int len;
|
|
|
|
- __be32 delta;
|
|
|
|
|
|
+ __be16 len;
|
|
|
|
|
|
- if (dont_encap) {
|
|
|
|
- skb->encapsulation = 0;
|
|
|
|
|
|
+ if (remcsum)
|
|
skb->ip_summed = CHECKSUM_NONE;
|
|
skb->ip_summed = CHECKSUM_NONE;
|
|
- } else {
|
|
|
|
- /* Only set up inner headers if we might be offloading
|
|
|
|
- * inner checksum.
|
|
|
|
- */
|
|
|
|
|
|
+
|
|
|
|
+ /* Set up inner headers if we are offloading inner checksum */
|
|
|
|
+ if (skb->ip_summed == CHECKSUM_PARTIAL) {
|
|
skb_reset_inner_headers(skb);
|
|
skb_reset_inner_headers(skb);
|
|
skb->encapsulation = 1;
|
|
skb->encapsulation = 1;
|
|
}
|
|
}
|
|
@@ -96,43 +106,28 @@ static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb,
|
|
skb->mac_len = mac_len;
|
|
skb->mac_len = mac_len;
|
|
skb->protocol = protocol;
|
|
skb->protocol = protocol;
|
|
|
|
|
|
- skb_push(skb, outer_hlen);
|
|
|
|
|
|
+ __skb_push(skb, outer_hlen);
|
|
skb_reset_mac_header(skb);
|
|
skb_reset_mac_header(skb);
|
|
skb_set_network_header(skb, mac_len);
|
|
skb_set_network_header(skb, mac_len);
|
|
skb_set_transport_header(skb, udp_offset);
|
|
skb_set_transport_header(skb, udp_offset);
|
|
- len = skb->len - udp_offset;
|
|
|
|
|
|
+ len = htons(skb->len - udp_offset);
|
|
uh = udp_hdr(skb);
|
|
uh = udp_hdr(skb);
|
|
- uh->len = htons(len);
|
|
|
|
|
|
+ uh->len = len;
|
|
|
|
|
|
if (!need_csum)
|
|
if (!need_csum)
|
|
continue;
|
|
continue;
|
|
|
|
|
|
- delta = htonl(oldlen + len);
|
|
|
|
-
|
|
|
|
uh->check = ~csum_fold((__force __wsum)
|
|
uh->check = ~csum_fold((__force __wsum)
|
|
- ((__force u32)uh->check +
|
|
|
|
- (__force u32)delta));
|
|
|
|
- if (offload_csum) {
|
|
|
|
- skb->ip_summed = CHECKSUM_PARTIAL;
|
|
|
|
- skb->csum_start = skb_transport_header(skb) - skb->head;
|
|
|
|
- skb->csum_offset = offsetof(struct udphdr, check);
|
|
|
|
- } else if (remcsum) {
|
|
|
|
- /* Need to calculate checksum from scratch,
|
|
|
|
- * inner checksums are never when doing
|
|
|
|
- * remote_checksum_offload.
|
|
|
|
- */
|
|
|
|
-
|
|
|
|
- skb->csum = skb_checksum(skb, udp_offset,
|
|
|
|
- skb->len - udp_offset,
|
|
|
|
- 0);
|
|
|
|
- uh->check = csum_fold(skb->csum);
|
|
|
|
- if (uh->check == 0)
|
|
|
|
- uh->check = CSUM_MANGLED_0;
|
|
|
|
- } else {
|
|
|
|
- uh->check = gso_make_checksum(skb, ~uh->check);
|
|
|
|
|
|
+ ((__force u32)len + partial));
|
|
|
|
|
|
|
|
+ if (skb->encapsulation || !offload_csum) {
|
|
|
|
+ uh->check = gso_make_checksum(skb, ~uh->check);
|
|
if (uh->check == 0)
|
|
if (uh->check == 0)
|
|
uh->check = CSUM_MANGLED_0;
|
|
uh->check = CSUM_MANGLED_0;
|
|
|
|
+ } else {
|
|
|
|
+ skb->ip_summed = CHECKSUM_PARTIAL;
|
|
|
|
+ skb->csum_start = skb_transport_header(skb) - skb->head;
|
|
|
|
+ skb->csum_offset = offsetof(struct udphdr, check);
|
|
}
|
|
}
|
|
} while ((skb = skb->next));
|
|
} while ((skb = skb->next));
|
|
out:
|
|
out:
|