act_skbedit.c 9.8 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0-only
2 3 4 5 6 7 8 9 10 11 12 13 14
/*
 * Copyright (c) 2008, Intel Corporation.
 *
 * Author: Alexander Duyck <alexander.h.duyck@intel.com>
 */

#include <linux/module.h>
#include <linux/init.h>
#include <linux/kernel.h>
#include <linux/skbuff.h>
#include <linux/rtnetlink.h>
#include <net/netlink.h>
#include <net/pkt_sched.h>
15 16 17
#include <net/ip.h>
#include <net/ipv6.h>
#include <net/dsfield.h>
18
#include <net/pkt_cls.h>
19 20 21 22

#include <linux/tc_act/tc_skbedit.h>
#include <net/tc_act/tc_skbedit.h>

23
static unsigned int skbedit_net_id;
24
static struct tc_action_ops act_skbedit_ops;
25

26 27
static int tcf_skbedit_act(struct sk_buff *skb, const struct tc_action *a,
			   struct tcf_result *res)
28
{
29
	struct tcf_skbedit *d = to_skbedit(a);
30 31
	struct tcf_skbedit_params *params;
	int action;
32

33
	tcf_lastuse_update(&d->tcf_tm);
34
	bstats_cpu_update(this_cpu_ptr(d->common.cpu_bstats), skb);
35

36
	params = rcu_dereference_bh(d->params);
37 38 39 40 41
	action = READ_ONCE(d->tcf_action);

	if (params->flags & SKBEDIT_F_PRIORITY)
		skb->priority = params->priority;
	if (params->flags & SKBEDIT_F_INHERITDSFIELD) {
42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59
		int wlen = skb_network_offset(skb);

		switch (tc_skb_protocol(skb)) {
		case htons(ETH_P_IP):
			wlen += sizeof(struct iphdr);
			if (!pskb_may_pull(skb, wlen))
				goto err;
			skb->priority = ipv4_get_dsfield(ip_hdr(skb)) >> 2;
			break;

		case htons(ETH_P_IPV6):
			wlen += sizeof(struct ipv6hdr);
			if (!pskb_may_pull(skb, wlen))
				goto err;
			skb->priority = ipv6_get_dsfield(ipv6_hdr(skb)) >> 2;
			break;
		}
	}
60 61 62 63 64 65
	if (params->flags & SKBEDIT_F_QUEUE_MAPPING &&
	    skb->dev->real_num_tx_queues > params->queue_mapping)
		skb_set_queue_mapping(skb, params->queue_mapping);
	if (params->flags & SKBEDIT_F_MARK) {
		skb->mark &= ~params->mask;
		skb->mark |= params->mark & params->mask;
66
	}
67 68 69
	if (params->flags & SKBEDIT_F_PTYPE)
		skb->pkt_type = params->ptype;
	return action;
70

71
err:
72
	qstats_drop_inc(this_cpu_ptr(d->common.cpu_qstats));
73
	return TC_ACT_SHOT;
74 75 76 77 78 79
}

static const struct nla_policy skbedit_policy[TCA_SKBEDIT_MAX + 1] = {
	[TCA_SKBEDIT_PARMS]		= { .len = sizeof(struct tc_skbedit) },
	[TCA_SKBEDIT_PRIORITY]		= { .len = sizeof(u32) },
	[TCA_SKBEDIT_QUEUE_MAPPING]	= { .len = sizeof(u16) },
80
	[TCA_SKBEDIT_MARK]		= { .len = sizeof(u32) },
81
	[TCA_SKBEDIT_PTYPE]		= { .len = sizeof(u16) },
82
	[TCA_SKBEDIT_MASK]		= { .len = sizeof(u32) },
83
	[TCA_SKBEDIT_FLAGS]		= { .len = sizeof(u64) },
84 85
};

86
static int tcf_skbedit_init(struct net *net, struct nlattr *nla,
87
			    struct nlattr *est, struct tc_action **a,
88
			    int ovr, int bind, bool rtnl_held,
89
			    struct tcf_proto *tp,
90
			    struct netlink_ext_ack *extack)
91
{
92
	struct tc_action_net *tn = net_generic(net, skbedit_net_id);
93
	struct tcf_skbedit_params *params_new;
94
	struct nlattr *tb[TCA_SKBEDIT_MAX + 1];
95
	struct tcf_chain *goto_ch = NULL;
96 97
	struct tc_skbedit *parm;
	struct tcf_skbedit *d;
98
	u32 flags = 0, *priority = NULL, *mark = NULL, *mask = NULL;
99
	u16 *queue_mapping = NULL, *ptype = NULL;
100 101
	bool exists = false;
	int ret = 0, err;
102
	u32 index;
103 104 105 106

	if (nla == NULL)
		return -EINVAL;

107 108
	err = nla_parse_nested_deprecated(tb, TCA_SKBEDIT_MAX, nla,
					  skbedit_policy, NULL);
109 110 111 112 113 114 115 116 117 118 119 120 121 122 123
	if (err < 0)
		return err;

	if (tb[TCA_SKBEDIT_PARMS] == NULL)
		return -EINVAL;

	if (tb[TCA_SKBEDIT_PRIORITY] != NULL) {
		flags |= SKBEDIT_F_PRIORITY;
		priority = nla_data(tb[TCA_SKBEDIT_PRIORITY]);
	}

	if (tb[TCA_SKBEDIT_QUEUE_MAPPING] != NULL) {
		flags |= SKBEDIT_F_QUEUE_MAPPING;
		queue_mapping = nla_data(tb[TCA_SKBEDIT_QUEUE_MAPPING]);
	}
124

125 126 127 128 129 130 131
	if (tb[TCA_SKBEDIT_PTYPE] != NULL) {
		ptype = nla_data(tb[TCA_SKBEDIT_PTYPE]);
		if (!skb_pkt_type_ok(*ptype))
			return -EINVAL;
		flags |= SKBEDIT_F_PTYPE;
	}

132 133 134 135 136
	if (tb[TCA_SKBEDIT_MARK] != NULL) {
		flags |= SKBEDIT_F_MARK;
		mark = nla_data(tb[TCA_SKBEDIT_MARK]);
	}

137 138 139 140 141
	if (tb[TCA_SKBEDIT_MASK] != NULL) {
		flags |= SKBEDIT_F_MASK;
		mask = nla_data(tb[TCA_SKBEDIT_MASK]);
	}

142 143 144 145 146 147 148
	if (tb[TCA_SKBEDIT_FLAGS] != NULL) {
		u64 *pure_flags = nla_data(tb[TCA_SKBEDIT_FLAGS]);

		if (*pure_flags & SKBEDIT_F_INHERITDSFIELD)
			flags |= SKBEDIT_F_INHERITDSFIELD;
	}

149
	parm = nla_data(tb[TCA_SKBEDIT_PARMS]);
150 151
	index = parm->index;
	err = tcf_idr_check_alloc(tn, &index, a, bind);
152 153 154
	if (err < 0)
		return err;
	exists = err;
155 156 157 158
	if (exists && bind)
		return 0;

	if (!flags) {
159 160
		if (exists)
			tcf_idr_release(*a, bind);
161
		else
162
			tcf_idr_cleanup(tn, index);
163 164 165 166
		return -EINVAL;
	}

	if (!exists) {
167
		ret = tcf_idr_create(tn, index, est, a,
168
				     &act_skbedit_ops, bind, true);
169
		if (ret) {
170
			tcf_idr_cleanup(tn, index);
171
			return ret;
172
		}
173

174
		d = to_skbedit(*a);
175 176
		ret = ACT_P_CREATED;
	} else {
177
		d = to_skbedit(*a);
178 179
		if (!ovr) {
			tcf_idr_release(*a, bind);
180
			return -EEXIST;
181
		}
182
	}
183 184 185
	err = tcf_action_check_ctrlact(parm->action, tp, &goto_ch, extack);
	if (err < 0)
		goto release_idr;
186

187 188
	params_new = kzalloc(sizeof(*params_new), GFP_KERNEL);
	if (unlikely(!params_new)) {
189 190
		err = -ENOMEM;
		goto put_chain;
191 192 193
	}

	params_new->flags = flags;
194
	if (flags & SKBEDIT_F_PRIORITY)
195
		params_new->priority = *priority;
196
	if (flags & SKBEDIT_F_QUEUE_MAPPING)
197
		params_new->queue_mapping = *queue_mapping;
198
	if (flags & SKBEDIT_F_MARK)
199
		params_new->mark = *mark;
200
	if (flags & SKBEDIT_F_PTYPE)
201
		params_new->ptype = *ptype;
202
	/* default behaviour is to use all the bits */
203
	params_new->mask = 0xffffffff;
204
	if (flags & SKBEDIT_F_MASK)
205
		params_new->mask = *mask;
206

207
	spin_lock_bh(&d->tcf_lock);
208
	goto_ch = tcf_action_set_ctrlact(*a, parm->action, goto_ch);
209 210
	params_new = rcu_replace_pointer(d->params, params_new,
					 lockdep_is_held(&d->tcf_lock));
211 212 213
	spin_unlock_bh(&d->tcf_lock);
	if (params_new)
		kfree_rcu(params_new, rcu);
214 215
	if (goto_ch)
		tcf_chain_put_by_act(goto_ch);
216 217

	if (ret == ACT_P_CREATED)
218
		tcf_idr_insert(tn, *a);
219
	return ret;
220 221 222 223 224 225
put_chain:
	if (goto_ch)
		tcf_chain_put_by_act(goto_ch);
release_idr:
	tcf_idr_release(*a, bind);
	return err;
226 227
}

E
Eric Dumazet 已提交
228 229
static int tcf_skbedit_dump(struct sk_buff *skb, struct tc_action *a,
			    int bind, int ref)
230 231
{
	unsigned char *b = skb_tail_pointer(skb);
232
	struct tcf_skbedit *d = to_skbedit(a);
233
	struct tcf_skbedit_params *params;
234 235
	struct tc_skbedit opt = {
		.index   = d->tcf_index,
236 237
		.refcnt  = refcount_read(&d->tcf_refcnt) - ref,
		.bindcnt = atomic_read(&d->tcf_bindcnt) - bind,
238
	};
239
	u64 pure_flags = 0;
240 241
	struct tcf_t t;

242 243 244 245
	spin_lock_bh(&d->tcf_lock);
	params = rcu_dereference_protected(d->params,
					   lockdep_is_held(&d->tcf_lock));
	opt.action = d->tcf_action;
246

247 248
	if (nla_put(skb, TCA_SKBEDIT_PARMS, sizeof(opt), &opt))
		goto nla_put_failure;
249 250
	if ((params->flags & SKBEDIT_F_PRIORITY) &&
	    nla_put_u32(skb, TCA_SKBEDIT_PRIORITY, params->priority))
251
		goto nla_put_failure;
252 253
	if ((params->flags & SKBEDIT_F_QUEUE_MAPPING) &&
	    nla_put_u16(skb, TCA_SKBEDIT_QUEUE_MAPPING, params->queue_mapping))
254
		goto nla_put_failure;
255 256
	if ((params->flags & SKBEDIT_F_MARK) &&
	    nla_put_u32(skb, TCA_SKBEDIT_MARK, params->mark))
257
		goto nla_put_failure;
258 259
	if ((params->flags & SKBEDIT_F_PTYPE) &&
	    nla_put_u16(skb, TCA_SKBEDIT_PTYPE, params->ptype))
260
		goto nla_put_failure;
261 262
	if ((params->flags & SKBEDIT_F_MASK) &&
	    nla_put_u32(skb, TCA_SKBEDIT_MASK, params->mask))
263
		goto nla_put_failure;
264
	if (params->flags & SKBEDIT_F_INHERITDSFIELD)
265 266 267 268
		pure_flags |= SKBEDIT_F_INHERITDSFIELD;
	if (pure_flags != 0 &&
	    nla_put(skb, TCA_SKBEDIT_FLAGS, sizeof(pure_flags), &pure_flags))
		goto nla_put_failure;
269 270

	tcf_tm_dump(&t, &d->tcf_tm);
271
	if (nla_put_64bit(skb, TCA_SKBEDIT_TM, sizeof(t), &t, TCA_SKBEDIT_PAD))
272
		goto nla_put_failure;
273 274
	spin_unlock_bh(&d->tcf_lock);

275 276 277
	return skb->len;

nla_put_failure:
278
	spin_unlock_bh(&d->tcf_lock);
279 280 281 282
	nlmsg_trim(skb, b);
	return -1;
}

283 284 285 286 287 288 289 290 291 292
static void tcf_skbedit_cleanup(struct tc_action *a)
{
	struct tcf_skbedit *d = to_skbedit(a);
	struct tcf_skbedit_params *params;

	params = rcu_dereference_protected(d->params, 1);
	if (params)
		kfree_rcu(params, rcu);
}

293 294
static int tcf_skbedit_walker(struct net *net, struct sk_buff *skb,
			      struct netlink_callback *cb, int type,
295 296
			      const struct tc_action_ops *ops,
			      struct netlink_ext_ack *extack)
297 298 299
{
	struct tc_action_net *tn = net_generic(net, skbedit_net_id);

300
	return tcf_generic_walker(tn, skb, cb, type, ops, extack);
301 302
}

303
static int tcf_skbedit_search(struct net *net, struct tc_action **a, u32 index)
304 305 306
{
	struct tc_action_net *tn = net_generic(net, skbedit_net_id);

307
	return tcf_idr_search(tn, a, index);
308 309
}

310 311 312 313 314 315 316 317 318 319 320
static size_t tcf_skbedit_get_fill_size(const struct tc_action *act)
{
	return nla_total_size(sizeof(struct tc_skbedit))
		+ nla_total_size(sizeof(u32)) /* TCA_SKBEDIT_PRIORITY */
		+ nla_total_size(sizeof(u16)) /* TCA_SKBEDIT_QUEUE_MAPPING */
		+ nla_total_size(sizeof(u32)) /* TCA_SKBEDIT_MARK */
		+ nla_total_size(sizeof(u16)) /* TCA_SKBEDIT_PTYPE */
		+ nla_total_size(sizeof(u32)) /* TCA_SKBEDIT_MASK */
		+ nla_total_size_64bit(sizeof(u64)); /* TCA_SKBEDIT_FLAGS */
}

321 322
static struct tc_action_ops act_skbedit_ops = {
	.kind		=	"skbedit",
323
	.id		=	TCA_ID_SKBEDIT,
324
	.owner		=	THIS_MODULE,
325
	.act		=	tcf_skbedit_act,
326 327
	.dump		=	tcf_skbedit_dump,
	.init		=	tcf_skbedit_init,
328
	.cleanup	=	tcf_skbedit_cleanup,
329
	.walk		=	tcf_skbedit_walker,
330
	.get_fill_size	=	tcf_skbedit_get_fill_size,
331
	.lookup		=	tcf_skbedit_search,
332
	.size		=	sizeof(struct tcf_skbedit),
333 334 335 336 337 338
};

static __net_init int skbedit_init_net(struct net *net)
{
	struct tc_action_net *tn = net_generic(net, skbedit_net_id);

339
	return tc_action_net_init(net, tn, &act_skbedit_ops);
340 341
}

342
static void __net_exit skbedit_exit_net(struct list_head *net_list)
343
{
344
	tc_action_net_exit(net_list, skbedit_net_id);
345 346 347 348
}

static struct pernet_operations skbedit_net_ops = {
	.init = skbedit_init_net,
349
	.exit_batch = skbedit_exit_net,
350 351
	.id   = &skbedit_net_id,
	.size = sizeof(struct tc_action_net),
352 353 354 355 356 357 358 359
};

MODULE_AUTHOR("Alexander Duyck, <alexander.h.duyck@intel.com>");
MODULE_DESCRIPTION("SKB Editing");
MODULE_LICENSE("GPL");

static int __init skbedit_init_module(void)
{
360
	return tcf_register_action(&act_skbedit_ops, &skbedit_net_ops);
361 362 363 364
}

static void __exit skbedit_cleanup_module(void)
{
365
	tcf_unregister_action(&act_skbedit_ops, &skbedit_net_ops);
366 367 368 369
}

module_init(skbedit_init_module);
module_exit(skbedit_cleanup_module);