gre_offload.c 6.2 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13
/*
 *	IPV4 GSO/GRO offload support
 *	Linux INET implementation
 *
 *	This program is free software; you can redistribute it and/or
 *	modify it under the terms of the GNU General Public License
 *	as published by the Free Software Foundation; either version
 *	2 of the License, or (at your option) any later version.
 *
 *	GRE GSO support
 */

#include <linux/skbuff.h>
14
#include <linux/init.h>
15 16 17 18 19 20 21 22
#include <net/protocol.h>
#include <net/gre.h>

static struct sk_buff *gre_gso_segment(struct sk_buff *skb,
				       netdev_features_t features)
{
	struct sk_buff *segs = ERR_PTR(-EINVAL);
	netdev_features_t enc_features;
23
	int ghl;
24
	struct gre_base_hdr *greh;
25
	u16 mac_offset = skb->mac_header;
26 27 28 29 30 31 32 33 34 35 36
	int mac_len = skb->mac_len;
	__be16 protocol = skb->protocol;
	int tnl_hlen;
	bool csum;

	if (unlikely(skb_shinfo(skb)->gso_type &
				~(SKB_GSO_TCPV4 |
				  SKB_GSO_TCPV6 |
				  SKB_GSO_UDP |
				  SKB_GSO_DODGY |
				  SKB_GSO_TCP_ECN |
E
Eric Dumazet 已提交
37
				  SKB_GSO_GRE |
T
Tom Herbert 已提交
38
				  SKB_GSO_GRE_CSUM |
E
Eric Dumazet 已提交
39
				  SKB_GSO_IPIP)))
40 41
		goto out;

42 43 44
	if (!skb->encapsulation)
		goto out;

45 46 47 48 49
	if (unlikely(!pskb_may_pull(skb, sizeof(*greh))))
		goto out;

	greh = (struct gre_base_hdr *)skb_transport_header(skb);

50
	ghl = skb_inner_mac_header(skb) - skb_transport_header(skb);
51 52 53 54
	if (unlikely(ghl < sizeof(*greh)))
		goto out;

	csum = !!(greh->flags & GRE_CSUM);
T
Tom Herbert 已提交
55 56
	if (csum)
		skb->encap_hdr_csum = 1;
57 58 59 60 61

	/* setup inner skb. */
	skb->protocol = greh->protocol;
	skb->encapsulation = 0;

62 63 64
	if (unlikely(!pskb_may_pull(skb, ghl)))
		goto out;

65 66 67 68 69 70
	__skb_pull(skb, ghl);
	skb_reset_mac_header(skb);
	skb_set_network_header(skb, skb_inner_network_offset(skb));
	skb->mac_len = skb_inner_network_offset(skb);

	/* segment inner packet. */
71
	enc_features = skb->dev->hw_enc_features & features;
72
	segs = skb_mac_gso_segment(skb, enc_features);
H
Himangi Saraogi 已提交
73
	if (IS_ERR_OR_NULL(segs)) {
74
		skb_gso_error_unwind(skb, protocol, ghl, mac_offset, mac_len);
75
		goto out;
76
	}
77 78 79 80 81 82 83 84 85 86 87 88 89

	skb = segs;
	tnl_hlen = skb_tnl_header_len(skb);
	do {
		__skb_push(skb, ghl);
		if (csum) {
			__be32 *pcsum;

			if (skb_has_shared_frag(skb)) {
				int err;

				err = __skb_linearize(skb);
				if (err) {
90
					kfree_skb_list(segs);
91 92 93 94 95
					segs = ERR_PTR(err);
					goto out;
				}
			}

T
Tom Herbert 已提交
96 97 98 99
			skb_reset_transport_header(skb);

			greh = (struct gre_base_hdr *)
			    skb_transport_header(skb);
100 101
			pcsum = (__be32 *)(greh + 1);
			*pcsum = 0;
T
Tom Herbert 已提交
102
			*(__sum16 *)pcsum = gso_make_checksum(skb, 0);
103 104 105
		}
		__skb_push(skb, tnl_hlen - ghl);

106 107 108
		skb_reset_inner_headers(skb);
		skb->encapsulation = 1;

109 110 111 112 113 114 115 116 117
		skb_reset_mac_header(skb);
		skb_set_network_header(skb, mac_len);
		skb->mac_len = mac_len;
		skb->protocol = protocol;
	} while ((skb = skb->next));
out:
	return segs;
}

118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168
static struct sk_buff **gre_gro_receive(struct sk_buff **head,
					struct sk_buff *skb)
{
	struct sk_buff **pp = NULL;
	struct sk_buff *p;
	const struct gre_base_hdr *greh;
	unsigned int hlen, grehlen;
	unsigned int off;
	int flush = 1;
	struct packet_offload *ptype;
	__be16 type;

	off = skb_gro_offset(skb);
	hlen = off + sizeof(*greh);
	greh = skb_gro_header_fast(skb, off);
	if (skb_gro_header_hard(skb, hlen)) {
		greh = skb_gro_header_slow(skb, hlen, off);
		if (unlikely(!greh))
			goto out;
	}

	/* Only support version 0 and K (key), C (csum) flags. Note that
	 * although the support for the S (seq#) flag can be added easily
	 * for GRO, this is problematic for GSO hence can not be enabled
	 * here because a GRO pkt may end up in the forwarding path, thus
	 * requiring GSO support to break it up correctly.
	 */
	if ((greh->flags & ~(GRE_KEY|GRE_CSUM)) != 0)
		goto out;

	type = greh->protocol;

	rcu_read_lock();
	ptype = gro_find_receive_by_type(type);
	if (ptype == NULL)
		goto out_unlock;

	grehlen = GRE_HEADER_SECTION;

	if (greh->flags & GRE_KEY)
		grehlen += GRE_HEADER_SECTION;

	if (greh->flags & GRE_CSUM)
		grehlen += GRE_HEADER_SECTION;

	hlen = off + grehlen;
	if (skb_gro_header_hard(skb, hlen)) {
		greh = skb_gro_header_slow(skb, hlen, off);
		if (unlikely(!greh))
			goto out_unlock;
	}
169 170

	/* Don't bother verifying checksum if we're going to flush anyway. */
171 172
	if ((greh->flags & GRE_CSUM) && !NAPI_GRO_CB(skb)->flush) {
		if (skb_gro_checksum_simple_validate(skb))
173
			goto out_unlock;
174

175 176 177 178
		skb_gro_checksum_try_convert(skb, IPPROTO_GRE, 0,
					     null_compute_pseudo);
	}

179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225
	flush = 0;

	for (p = *head; p; p = p->next) {
		const struct gre_base_hdr *greh2;

		if (!NAPI_GRO_CB(p)->same_flow)
			continue;

		/* The following checks are needed to ensure only pkts
		 * from the same tunnel are considered for aggregation.
		 * The criteria for "the same tunnel" includes:
		 * 1) same version (we only support version 0 here)
		 * 2) same protocol (we only support ETH_P_IP for now)
		 * 3) same set of flags
		 * 4) same key if the key field is present.
		 */
		greh2 = (struct gre_base_hdr *)(p->data + off);

		if (greh2->flags != greh->flags ||
		    greh2->protocol != greh->protocol) {
			NAPI_GRO_CB(p)->same_flow = 0;
			continue;
		}
		if (greh->flags & GRE_KEY) {
			/* compare keys */
			if (*(__be32 *)(greh2+1) != *(__be32 *)(greh+1)) {
				NAPI_GRO_CB(p)->same_flow = 0;
				continue;
			}
		}
	}

	skb_gro_pull(skb, grehlen);

	/* Adjusted NAPI_GRO_CB(skb)->csum after skb_gro_pull()*/
	skb_gro_postpull_rcsum(skb, greh, grehlen);

	pp = ptype->callbacks.gro_receive(head, skb);

out_unlock:
	rcu_read_unlock();
out:
	NAPI_GRO_CB(skb)->flush |= flush;

	return pp;
}

226
static int gre_gro_complete(struct sk_buff *skb, int nhoff)
227 228 229 230 231 232 233
{
	struct gre_base_hdr *greh = (struct gre_base_hdr *)(skb->data + nhoff);
	struct packet_offload *ptype;
	unsigned int grehlen = sizeof(*greh);
	int err = -ENOENT;
	__be16 type;

234 235 236
	skb->encapsulation = 1;
	skb_shinfo(skb)->gso_type = SKB_GSO_GRE;

237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252
	type = greh->protocol;
	if (greh->flags & GRE_KEY)
		grehlen += GRE_HEADER_SECTION;

	if (greh->flags & GRE_CSUM)
		grehlen += GRE_HEADER_SECTION;

	rcu_read_lock();
	ptype = gro_find_complete_by_type(type);
	if (ptype != NULL)
		err = ptype->callbacks.gro_complete(skb, nhoff + grehlen);

	rcu_read_unlock();
	return err;
}

253 254 255
static const struct net_offload gre_offload = {
	.callbacks = {
		.gso_segment = gre_gso_segment,
256 257
		.gro_receive = gre_gro_receive,
		.gro_complete = gre_gro_complete,
258 259 260
	},
};

261
static int __init gre_offload_init(void)
262 263 264
{
	return inet_add_offload(&gre_offload, IPPROTO_GRE);
}
265
device_initcall(gre_offload_init);