x_tables.h 17.3 KB
Newer Older
1 2
#ifndef _X_TABLES_H
#define _X_TABLES_H
3
#include <linux/kernel.h>
4 5
#include <linux/types.h>

6
#define XT_FUNCTION_MAXNAMELEN 30
7
#define XT_EXTENSION_MAXNAMELEN 29
8 9
#define XT_TABLE_MAXNAMELEN 32

E
Eric Dumazet 已提交
10
struct xt_entry_match {
11 12
	union {
		struct {
13
			__u16 match_size;
14 15

			/* Used by userspace */
16
			char name[XT_EXTENSION_MAXNAMELEN];
17
			__u8 revision;
18 19
		} user;
		struct {
20
			__u16 match_size;
21 22 23 24 25 26

			/* Used inside the kernel */
			struct xt_match *match;
		} kernel;

		/* Total length */
27
		__u16 match_size;
28 29 30 31 32
	} u;

	unsigned char data[0];
};

E
Eric Dumazet 已提交
33
struct xt_entry_target {
34 35
	union {
		struct {
36
			__u16 target_size;
37 38

			/* Used by userspace */
39
			char name[XT_EXTENSION_MAXNAMELEN];
40
			__u8 revision;
41 42
		} user;
		struct {
43
			__u16 target_size;
44 45 46 47 48 49

			/* Used inside the kernel */
			struct xt_target *target;
		} kernel;

		/* Total length */
50
		__u16 target_size;
51 52 53 54 55
	} u;

	unsigned char data[0];
};

56 57 58 59 60 61 62 63
#define XT_TARGET_INIT(__name, __size)					       \
{									       \
	.target.u.user = {						       \
		.target_size	= XT_ALIGN(__size),			       \
		.name		= __name,				       \
	},								       \
}

E
Eric Dumazet 已提交
64
struct xt_standard_target {
65 66 67 68
	struct xt_entry_target target;
	int verdict;
};

69 70 71 72 73
struct xt_error_target {
	struct xt_entry_target target;
	char errorname[XT_FUNCTION_MAXNAMELEN];
};

74 75
/* The argument to IPT_SO_GET_REVISION_*.  Returns highest revision
 * kernel supports, if >= revision. */
E
Eric Dumazet 已提交
76
struct xt_get_revision {
77
	char name[XT_EXTENSION_MAXNAMELEN];
78
	__u8 revision;
79 80 81 82 83 84 85 86
};

/* CONTINUE verdict for targets */
#define XT_CONTINUE 0xFFFFFFFF

/* For standard target */
#define XT_RETURN (-NF_REPEAT - 1)

87 88 89 90 91
/* this is a dummy structure to find out the alignment requirement for a struct
 * containing all the fundamental data types that are used in ipt_entry,
 * ip6t_entry and arpt_entry.  This sucks, and it is a hack.  It will be my
 * personal pleasure to remove it -HW
 */
E
Eric Dumazet 已提交
92
struct _xt_align {
93 94 95 96
	__u8 u8;
	__u16 u16;
	__u32 u32;
	__u64 u64;
97 98
};

99
#define XT_ALIGN(s) __ALIGN_KERNEL((s), __alignof__(struct _xt_align))
100 101 102 103 104 105 106 107 108

/* Standard return verdict, or do jump. */
#define XT_STANDARD_TARGET ""
/* Error verdict. */
#define XT_ERROR_TARGET "ERROR"

#define SET_COUNTER(c,b,p) do { (c).bcnt = (b); (c).pcnt = (p); } while(0)
#define ADD_COUNTER(c,b,p) do { (c).bcnt += (b); (c).pcnt += (p); } while(0)

E
Eric Dumazet 已提交
109
struct xt_counters {
110
	__u64 pcnt, bcnt;			/* Packet and byte counters */
111 112 113
};

/* The argument to IPT_SO_ADD_COUNTERS. */
E
Eric Dumazet 已提交
114
struct xt_counters_info {
115 116 117 118 119 120 121 122 123 124 125
	/* Which table. */
	char name[XT_TABLE_MAXNAMELEN];

	unsigned int num_counters;

	/* The counters (actually `number' of these). */
	struct xt_counters counters[0];
};

#define XT_INV_PROTO		0x40	/* Invert the sense of PROTO. */

126
#ifndef __KERNEL__
127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169
/* fn returns 0 to continue iteration */
#define XT_MATCH_ITERATE(type, e, fn, args...)			\
({								\
	unsigned int __i;					\
	int __ret = 0;						\
	struct xt_entry_match *__m;				\
								\
	for (__i = sizeof(type);				\
	     __i < (e)->target_offset;				\
	     __i += __m->u.match_size) {			\
		__m = (void *)e + __i;				\
								\
		__ret = fn(__m , ## args);			\
		if (__ret != 0)					\
			break;					\
	}							\
	__ret;							\
})

/* fn returns 0 to continue iteration */
#define XT_ENTRY_ITERATE_CONTINUE(type, entries, size, n, fn, args...) \
({								\
	unsigned int __i, __n;					\
	int __ret = 0;						\
	type *__entry;						\
								\
	for (__i = 0, __n = 0; __i < (size);			\
	     __i += __entry->next_offset, __n++) { 		\
		__entry = (void *)(entries) + __i;		\
		if (__n < n)					\
			continue;				\
								\
		__ret = fn(__entry , ## args);			\
		if (__ret != 0)					\
			break;					\
	}							\
	__ret;							\
})

/* fn returns 0 to continue iteration */
#define XT_ENTRY_ITERATE(type, entries, size, fn, args...) \
	XT_ENTRY_ITERATE_CONTINUE(type, entries, size, 0, fn, args)

170 171 172 173 174 175 176 177
#endif /* !__KERNEL__ */

/* pos is normally a struct ipt_entry/ip6t_entry/etc. */
#define xt_entry_foreach(pos, ehead, esize) \
	for ((pos) = (typeof(pos))(ehead); \
	     (pos) < (typeof(pos))((char *)(ehead) + (esize)); \
	     (pos) = (typeof(pos))((char *)(pos) + (pos)->next_offset))

178 179 180 181 182 183 184 185
/* can only be xt_entry_match, so no use of typeof here */
#define xt_ematch_foreach(pos, entry) \
	for ((pos) = (struct xt_entry_match *)entry->elems; \
	     (pos) < (struct xt_entry_match *)((char *)(entry) + \
	             (entry)->target_offset); \
	     (pos) = (struct xt_entry_match *)((char *)(pos) + \
	             (pos)->u.match_size))

186 187 188 189
#ifdef __KERNEL__

#include <linux/netdevice.h>

190
/**
191
 * struct xt_action_param - parameters for matches/targets
192
 *
193 194 195 196
 * @match:	the match extension
 * @target:	the target extension
 * @matchinfo:	per-match data
 * @targetinfo:	per-target data
197 198 199 200
 * @in:		input netdevice
 * @out:	output netdevice
 * @fragoff:	packet is a fragment, this is the data offset
 * @thoff:	position of transport header relative to skb->data
201
 * @hook:	hook number given packet came from
202 203
 * @family:	Actual NFPROTO_* through which the function is invoked
 * 		(helpful when match->family == NFPROTO_UNSPEC)
204 205 206
 *
 * Fields written to by extensions:
 *
207
 * @hotdrop:	drop packet if we had inspection problems
208
 * Network namespace obtainable using dev_net(in/out)
209
 */
210 211 212 213 214 215 216 217
struct xt_action_param {
	union {
		const struct xt_match *match;
		const struct xt_target *target;
	};
	union {
		const void *matchinfo, *targinfo;
	};
218 219 220
	const struct net_device *in, *out;
	int fragoff;
	unsigned int thoff;
221
	unsigned int hooknum;
222
	u_int8_t family;
223
	bool hotdrop;
224 225
};

226 227 228 229
/**
 * struct xt_mtchk_param - parameters for match extensions'
 * checkentry functions
 *
230
 * @net:	network namespace through which the check was invoked
231 232
 * @table:	table the rule is tried to be inserted into
 * @entryinfo:	the family-specific rule data
233
 * 		(struct ipt_ip, ip6t_ip, arpt_arp or (note) ebt_entry)
234 235 236
 * @match:	struct xt_match through which this function was invoked
 * @matchinfo:	per-match data
 * @hook_mask:	via which hooks the new rule is reachable
237
 * Other fields as above.
238 239
 */
struct xt_mtchk_param {
240
	struct net *net;
241 242 243 244 245
	const char *table;
	const void *entryinfo;
	const struct xt_match *match;
	void *matchinfo;
	unsigned int hook_mask;
246
	u_int8_t family;
247 248
};

249 250 251 252
/**
 * struct xt_mdtor_param - match destructor parameters
 * Fields as above.
 */
253
struct xt_mtdtor_param {
254
	struct net *net;
255 256
	const struct xt_match *match;
	void *matchinfo;
257
	u_int8_t family;
258 259
};

260 261 262 263 264 265 266 267 268 269
/**
 * struct xt_tgchk_param - parameters for target extensions'
 * checkentry functions
 *
 * @entryinfo:	the family-specific rule data
 * 		(struct ipt_entry, ip6t_entry, arpt_entry, ebt_entry)
 *
 * Other fields see above.
 */
struct xt_tgchk_param {
270
	struct net *net;
271
	const char *table;
272
	const void *entryinfo;
273 274 275
	const struct xt_target *target;
	void *targinfo;
	unsigned int hook_mask;
276
	u_int8_t family;
277 278
};

279 280
/* Target destructor parameters */
struct xt_tgdtor_param {
281
	struct net *net;
282 283
	const struct xt_target *target;
	void *targinfo;
284
	u_int8_t family;
285 286
};

E
Eric Dumazet 已提交
287
struct xt_match {
288 289
	struct list_head list;

290
	const char name[XT_EXTENSION_MAXNAMELEN];
291
	u_int8_t revision;
292 293 294 295 296 297

	/* Return true or false: return FALSE and set *hotdrop = 1 to
           force immediate packet drop. */
	/* Arguments changed since 2.6.9, as this must now handle
	   non-linear skb, using skb_header_pointer and
	   skb_ip_make_writable. */
298
	bool (*match)(const struct sk_buff *skb,
299
		      struct xt_action_param *);
300 301

	/* Called when user tries to insert an entry of this type. */
302
	int (*checkentry)(const struct xt_mtchk_param *);
303 304

	/* Called when entry of this type deleted. */
305
	void (*destroy)(const struct xt_mtdtor_param *);
306
#ifdef CONFIG_COMPAT
307
	/* Called when userspace align differs from kernel space one */
308 309
	void (*compat_from_user)(void *dst, const void *src);
	int (*compat_to_user)(void __user *dst, const void *src);
310
#endif
311 312
	/* Set this to THIS_MODULE if you are a module, otherwise NULL */
	struct module *me;
313

314
	const char *table;
315
	unsigned int matchsize;
316
#ifdef CONFIG_COMPAT
317
	unsigned int compatsize;
318
#endif
319 320
	unsigned int hooks;
	unsigned short proto;
321 322

	unsigned short family;
323 324 325
};

/* Registration hooks for targets. */
E
Eric Dumazet 已提交
326
struct xt_target {
327 328
	struct list_head list;

329
	const char name[XT_EXTENSION_MAXNAMELEN];
330
	u_int8_t revision;
331 332 333 334

	/* Returns verdict. Argument order changed since 2.6.9, as this
	   must now handle non-linear skbs, using skb_copy_bits and
	   skb_ip_make_writable. */
335
	unsigned int (*target)(struct sk_buff *skb,
336
			       const struct xt_action_param *);
337 338 339 340

	/* Called when user tries to insert an entry of this type:
           hook_mask is a bitmask of hooks from which it can be
           called. */
341
	/* Should return 0 on success or an error code otherwise (-Exxxx). */
342
	int (*checkentry)(const struct xt_tgchk_param *);
343 344

	/* Called when entry of this type deleted. */
345
	void (*destroy)(const struct xt_tgdtor_param *);
346
#ifdef CONFIG_COMPAT
347
	/* Called when userspace align differs from kernel space one */
348 349
	void (*compat_from_user)(void *dst, const void *src);
	int (*compat_to_user)(void __user *dst, const void *src);
350
#endif
351 352
	/* Set this to THIS_MODULE if you are a module, otherwise NULL */
	struct module *me;
353

354
	const char *table;
355
	unsigned int targetsize;
356
#ifdef CONFIG_COMPAT
357
	unsigned int compatsize;
358
#endif
359 360
	unsigned int hooks;
	unsigned short proto;
361 362

	unsigned short family;
363 364 365
};

/* Furniture shopping... */
E
Eric Dumazet 已提交
366
struct xt_table {
367 368 369 370 371 372
	struct list_head list;

	/* What hooks you will enter on */
	unsigned int valid_hooks;

	/* Man behind the curtain... */
373
	struct xt_table_info *private;
374 375 376 377

	/* Set this to THIS_MODULE if you are a module, otherwise NULL */
	struct module *me;

378
	u_int8_t af;		/* address/protocol family */
379
	int priority;		/* hook order */
380 381 382

	/* A unique name... */
	const char name[XT_TABLE_MAXNAMELEN];
383 384 385 386 387
};

#include <linux/netfilter_ipv4.h>

/* The table itself */
E
Eric Dumazet 已提交
388
struct xt_table_info {
389 390 391 392 393 394 395 396
	/* Size per table */
	unsigned int size;
	/* Number of entries: FIXME. --RR */
	unsigned int number;
	/* Initial number of entries. Needed for module usage count */
	unsigned int initial_entries;

	/* Entry points and underflows */
397 398
	unsigned int hook_entry[NF_INET_NUMHOOKS];
	unsigned int underflow[NF_INET_NUMHOOKS];
399

400 401 402 403 404
	/*
	 * Number of user chains. Since tables cannot have loops, at most
	 * @stacksize jumps (number of user chains) can possibly be made.
	 */
	unsigned int stacksize;
405
	unsigned int __percpu *stackptr;
406
	void ***jumpstack;
407
	/* ipt_entry tables: one per CPU */
408
	/* Note : this field MUST be the last one, see XT_TABLE_INFO_SZ */
409
	void *entries[1];
410 411
};

412 413
#define XT_TABLE_INFO_SZ (offsetof(struct xt_table_info, entries) \
			  + nr_cpu_ids * sizeof(char *))
414 415
extern int xt_register_target(struct xt_target *target);
extern void xt_unregister_target(struct xt_target *target);
416 417 418
extern int xt_register_targets(struct xt_target *target, unsigned int n);
extern void xt_unregister_targets(struct xt_target *target, unsigned int n);

419 420
extern int xt_register_match(struct xt_match *target);
extern void xt_unregister_match(struct xt_match *target);
421 422
extern int xt_register_matches(struct xt_match *match, unsigned int n);
extern void xt_unregister_matches(struct xt_match *match, unsigned int n);
423

424
extern int xt_check_match(struct xt_mtchk_param *,
425
			  unsigned int size, u_int8_t proto, bool inv_proto);
426
extern int xt_check_target(struct xt_tgchk_param *,
427
			   unsigned int size, u_int8_t proto, bool inv_proto);
428

429
extern struct xt_table *xt_register_table(struct net *net,
430
					  const struct xt_table *table,
431 432
					  struct xt_table_info *bootstrap,
					  struct xt_table_info *newinfo);
433 434 435 436 437 438 439
extern void *xt_unregister_table(struct xt_table *table);

extern struct xt_table_info *xt_replace_table(struct xt_table *table,
					      unsigned int num_counters,
					      struct xt_table_info *newinfo,
					      int *error);

440 441
extern struct xt_match *xt_find_match(u8 af, const char *name, u8 revision);
extern struct xt_target *xt_find_target(u8 af, const char *name, u8 revision);
442 443
extern struct xt_match *xt_request_find_match(u8 af, const char *name,
					      u8 revision);
444
extern struct xt_target *xt_request_find_target(u8 af, const char *name,
445
						u8 revision);
446 447
extern int xt_find_revision(u8 af, const char *name, u8 revision,
			    int target, int *err);
448

449
extern struct xt_table *xt_find_table_lock(struct net *net, u_int8_t af,
450
					   const char *name);
451 452
extern void xt_table_unlock(struct xt_table *t);

453 454
extern int xt_proto_init(struct net *net, u_int8_t af);
extern void xt_proto_fini(struct net *net, u_int8_t af);
455 456 457

extern struct xt_table_info *xt_alloc_table_info(unsigned int size);
extern void xt_free_table_info(struct xt_table_info *info);
458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474

/*
 * Per-CPU spinlock associated with per-cpu table entries, and
 * with a counter for the "reading" side that allows a recursive
 * reader to avoid taking the lock and deadlocking.
 *
 * "reading" is used by ip/arp/ip6 tables rule processing which runs per-cpu.
 * It needs to ensure that the rules are not being changed while the packet
 * is being processed. In some cases, the read lock will be acquired
 * twice on the same CPU; this is okay because of the count.
 *
 * "writing" is used when reading counters.
 *  During replace any readers that are using the old tables have to complete
 *  before freeing the old table. This is handled by the write locking
 *  necessary for reading the counters.
 */
struct xt_info_lock {
475
	seqlock_t lock;
476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498
	unsigned char readers;
};
DECLARE_PER_CPU(struct xt_info_lock, xt_info_locks);

/*
 * Note: we need to ensure that preemption is disabled before acquiring
 * the per-cpu-variable, so we do it as a two step process rather than
 * using "spin_lock_bh()".
 *
 * We _also_ need to disable bottom half processing before updating our
 * nesting count, to make sure that the only kind of re-entrancy is this
 * code being called by itself: since the count+lock is not an atomic
 * operation, we can allow no races.
 *
 * _Only_ that special combination of being per-cpu and never getting
 * re-entered asynchronously means that the count is safe.
 */
static inline void xt_info_rdlock_bh(void)
{
	struct xt_info_lock *lock;

	local_bh_disable();
	lock = &__get_cpu_var(xt_info_locks);
499
	if (likely(!lock->readers++))
500
		write_seqlock(&lock->lock);
501 502 503 504 505 506
}

static inline void xt_info_rdunlock_bh(void)
{
	struct xt_info_lock *lock = &__get_cpu_var(xt_info_locks);

507
	if (likely(!--lock->readers))
508
		write_sequnlock(&lock->lock);
509 510 511 512 513 514 515 516 517 518
	local_bh_enable();
}

/*
 * The "writer" side needs to get exclusive access to the lock,
 * regardless of readers.  This must be called with bottom half
 * processing (and thus also preemption) disabled.
 */
static inline void xt_info_wrlock(unsigned int cpu)
{
519
	write_seqlock(&per_cpu(xt_info_locks, cpu).lock);
520 521 522 523
}

static inline void xt_info_wrunlock(unsigned int cpu)
{
524
	write_sequnlock(&per_cpu(xt_info_locks, cpu).lock);
525
}
526

527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549
/*
 * This helper is performance critical and must be inlined
 */
static inline unsigned long ifname_compare_aligned(const char *_a,
						   const char *_b,
						   const char *_mask)
{
	const unsigned long *a = (const unsigned long *)_a;
	const unsigned long *b = (const unsigned long *)_b;
	const unsigned long *mask = (const unsigned long *)_mask;
	unsigned long ret;

	ret = (a[0] ^ b[0]) & mask[0];
	if (IFNAMSIZ > sizeof(unsigned long))
		ret |= (a[1] ^ b[1]) & mask[1];
	if (IFNAMSIZ > 2 * sizeof(unsigned long))
		ret |= (a[2] ^ b[2]) & mask[2];
	if (IFNAMSIZ > 3 * sizeof(unsigned long))
		ret |= (a[3] ^ b[3]) & mask[3];
	BUILD_BUG_ON(IFNAMSIZ > 4 * sizeof(unsigned long));
	return ret;
}

550 551 552
extern struct nf_hook_ops *xt_hook_link(const struct xt_table *, nf_hookfn *);
extern void xt_hook_unlink(const struct xt_table *, struct nf_hook_ops *);

553 554 555
#ifdef CONFIG_COMPAT
#include <net/compat.h>

E
Eric Dumazet 已提交
556
struct compat_xt_entry_match {
557 558 559 560 561 562
	union {
		struct {
			u_int16_t match_size;
			char name[XT_FUNCTION_MAXNAMELEN - 1];
			u_int8_t revision;
		} user;
563 564 565 566
		struct {
			u_int16_t match_size;
			compat_uptr_t match;
		} kernel;
567 568 569 570 571
		u_int16_t match_size;
	} u;
	unsigned char data[0];
};

E
Eric Dumazet 已提交
572
struct compat_xt_entry_target {
573 574 575 576 577 578
	union {
		struct {
			u_int16_t target_size;
			char name[XT_FUNCTION_MAXNAMELEN - 1];
			u_int8_t revision;
		} user;
579 580 581 582
		struct {
			u_int16_t target_size;
			compat_uptr_t target;
		} kernel;
583 584 585 586 587 588 589 590 591
		u_int16_t target_size;
	} u;
	unsigned char data[0];
};

/* FIXME: this works only on 32 bit tasks
 * need to change whole approach in order to calculate align as function of
 * current task alignment */

E
Eric Dumazet 已提交
592
struct compat_xt_counters {
593
	compat_u64 pcnt, bcnt;			/* Packet and byte counters */
594 595
};

E
Eric Dumazet 已提交
596
struct compat_xt_counters_info {
597 598 599 600 601
	char name[XT_TABLE_MAXNAMELEN];
	compat_uint_t num_counters;
	struct compat_xt_counters counters[0];
};

602 603 604 605 606 607 608
struct _compat_xt_align {
	__u8 u8;
	__u16 u16;
	__u32 u32;
	compat_u64 u64;
};

609
#define COMPAT_XT_ALIGN(s) __ALIGN_KERNEL((s), __alignof__(struct _compat_xt_align))
610

611 612
extern void xt_compat_lock(u_int8_t af);
extern void xt_compat_unlock(u_int8_t af);
613

614
extern int xt_compat_add_offset(u_int8_t af, unsigned int offset, int delta);
615
extern void xt_compat_flush_offsets(u_int8_t af);
616
extern void xt_compat_init_offsets(u_int8_t af, unsigned int number);
617
extern int xt_compat_calc_jump(u_int8_t af, unsigned int offset);
618

619
extern int xt_compat_match_offset(const struct xt_match *match);
620
extern int xt_compat_match_from_user(struct xt_entry_match *m,
621
				     void **dstptr, unsigned int *size);
622
extern int xt_compat_match_to_user(const struct xt_entry_match *m,
623
				   void __user **dstptr, unsigned int *size);
624

625
extern int xt_compat_target_offset(const struct xt_target *target);
626
extern void xt_compat_target_from_user(struct xt_entry_target *t,
627
				       void **dstptr, unsigned int *size);
628
extern int xt_compat_target_to_user(const struct xt_entry_target *t,
629
				    void __user **dstptr, unsigned int *size);
630 631

#endif /* CONFIG_COMPAT */
632 633 634
#endif /* __KERNEL__ */

#endif /* _X_TABLES_H */