x_tables.h 15.8 KB
Newer Older
1 2 3
#ifndef _X_TABLES_H
#define _X_TABLES_H

4 5
#include <linux/types.h>

6 7 8
#define XT_FUNCTION_MAXNAMELEN 30
#define XT_TABLE_MAXNAMELEN 32

9 10 11 12
struct xt_entry_match
{
	union {
		struct {
13
			__u16 match_size;
14 15 16 17

			/* Used by userspace */
			char name[XT_FUNCTION_MAXNAMELEN-1];

18
			__u8 revision;
19 20
		} user;
		struct {
21
			__u16 match_size;
22 23 24 25 26 27

			/* Used inside the kernel */
			struct xt_match *match;
		} kernel;

		/* Total length */
28
		__u16 match_size;
29 30 31 32 33 34 35 36 37
	} u;

	unsigned char data[0];
};

struct xt_entry_target
{
	union {
		struct {
38
			__u16 target_size;
39 40 41 42

			/* Used by userspace */
			char name[XT_FUNCTION_MAXNAMELEN-1];

43
			__u8 revision;
44 45
		} user;
		struct {
46
			__u16 target_size;
47 48 49 50 51 52

			/* Used inside the kernel */
			struct xt_target *target;
		} kernel;

		/* Total length */
53
		__u16 target_size;
54 55 56 57 58
	} u;

	unsigned char data[0];
};

59 60 61 62 63 64 65 66
#define XT_TARGET_INIT(__name, __size)					       \
{									       \
	.target.u.user = {						       \
		.target_size	= XT_ALIGN(__size),			       \
		.name		= __name,				       \
	},								       \
}

67 68 69 70 71 72
struct xt_standard_target
{
	struct xt_entry_target target;
	int verdict;
};

73 74 75 76 77 78
/* The argument to IPT_SO_GET_REVISION_*.  Returns highest revision
 * kernel supports, if >= revision. */
struct xt_get_revision
{
	char name[XT_FUNCTION_MAXNAMELEN-1];

79
	__u8 revision;
80 81 82 83 84 85 86 87
};

/* CONTINUE verdict for targets */
#define XT_CONTINUE 0xFFFFFFFF

/* For standard target */
#define XT_RETURN (-NF_REPEAT - 1)

88 89 90 91 92 93 94
/* this is a dummy structure to find out the alignment requirement for a struct
 * containing all the fundamental data types that are used in ipt_entry,
 * ip6t_entry and arpt_entry.  This sucks, and it is a hack.  It will be my
 * personal pleasure to remove it -HW
 */
struct _xt_align
{
95 96 97 98
	__u8 u8;
	__u16 u16;
	__u32 u32;
	__u64 u64;
99 100 101 102
};

#define XT_ALIGN(s) (((s) + (__alignof__(struct _xt_align)-1)) 	\
			& ~(__alignof__(struct _xt_align)-1))
103 104 105 106 107 108 109 110 111 112 113

/* Standard return verdict, or do jump. */
#define XT_STANDARD_TARGET ""
/* Error verdict. */
#define XT_ERROR_TARGET "ERROR"

#define SET_COUNTER(c,b,p) do { (c).bcnt = (b); (c).pcnt = (p); } while(0)
#define ADD_COUNTER(c,b,p) do { (c).bcnt += (b); (c).pcnt += (p); } while(0)

struct xt_counters
{
114
	__u64 pcnt, bcnt;			/* Packet and byte counters */
115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130
};

/* The argument to IPT_SO_ADD_COUNTERS. */
struct xt_counters_info
{
	/* Which table. */
	char name[XT_TABLE_MAXNAMELEN];

	unsigned int num_counters;

	/* The counters (actually `number' of these). */
	struct xt_counters counters[0];
};

#define XT_INV_PROTO		0x40	/* Invert the sense of PROTO. */

131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173
/* fn returns 0 to continue iteration */
#define XT_MATCH_ITERATE(type, e, fn, args...)			\
({								\
	unsigned int __i;					\
	int __ret = 0;						\
	struct xt_entry_match *__m;				\
								\
	for (__i = sizeof(type);				\
	     __i < (e)->target_offset;				\
	     __i += __m->u.match_size) {			\
		__m = (void *)e + __i;				\
								\
		__ret = fn(__m , ## args);			\
		if (__ret != 0)					\
			break;					\
	}							\
	__ret;							\
})

/* fn returns 0 to continue iteration */
#define XT_ENTRY_ITERATE_CONTINUE(type, entries, size, n, fn, args...) \
({								\
	unsigned int __i, __n;					\
	int __ret = 0;						\
	type *__entry;						\
								\
	for (__i = 0, __n = 0; __i < (size);			\
	     __i += __entry->next_offset, __n++) { 		\
		__entry = (void *)(entries) + __i;		\
		if (__n < n)					\
			continue;				\
								\
		__ret = fn(__entry , ## args);			\
		if (__ret != 0)					\
			break;					\
	}							\
	__ret;							\
})

/* fn returns 0 to continue iteration */
#define XT_ENTRY_ITERATE(type, entries, size, fn, args...) \
	XT_ENTRY_ITERATE_CONTINUE(type, entries, size, 0, fn, args)

174 175 176 177
#ifdef __KERNEL__

#include <linux/netdevice.h>

178 179 180 181 182 183 184 185 186 187
/**
 * struct xt_match_param - parameters for match extensions' match functions
 *
 * @in:		input netdevice
 * @out:	output netdevice
 * @match:	struct xt_match through which this function was invoked
 * @matchinfo:	per-match data
 * @fragoff:	packet is a fragment, this is the data offset
 * @thoff:	position of transport header relative to skb->data
 * @hotdrop:	drop packet if we had inspection problems
188 189
 * @family:	Actual NFPROTO_* through which the function is invoked
 * 		(helpful when match->family == NFPROTO_UNSPEC)
190 191 192 193 194 195 196 197
 */
struct xt_match_param {
	const struct net_device *in, *out;
	const struct xt_match *match;
	const void *matchinfo;
	int fragoff;
	unsigned int thoff;
	bool *hotdrop;
198
	u_int8_t family;
199 200
};

201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217
/**
 * struct xt_mtchk_param - parameters for match extensions'
 * checkentry functions
 *
 * @table:	table the rule is tried to be inserted into
 * @entryinfo:	the family-specific rule data
 * 		(struct ipt_ip, ip6t_ip, ebt_entry)
 * @match:	struct xt_match through which this function was invoked
 * @matchinfo:	per-match data
 * @hook_mask:	via which hooks the new rule is reachable
 */
struct xt_mtchk_param {
	const char *table;
	const void *entryinfo;
	const struct xt_match *match;
	void *matchinfo;
	unsigned int hook_mask;
218
	u_int8_t family;
219 220
};

221 222 223 224
/* Match destructor parameters */
struct xt_mtdtor_param {
	const struct xt_match *match;
	void *matchinfo;
225
	u_int8_t family;
226 227
};

228 229 230 231 232 233 234 235 236 237 238 239 240 241
/**
 * struct xt_target_param - parameters for target extensions' target functions
 *
 * @hooknum:	hook through which this target was invoked
 * @target:	struct xt_target through which this function was invoked
 * @targinfo:	per-target data
 *
 * Other fields see above.
 */
struct xt_target_param {
	const struct net_device *in, *out;
	unsigned int hooknum;
	const struct xt_target *target;
	const void *targinfo;
242
	u_int8_t family;
243 244
};

245 246 247 248 249 250 251 252 253 254 255
/**
 * struct xt_tgchk_param - parameters for target extensions'
 * checkentry functions
 *
 * @entryinfo:	the family-specific rule data
 * 		(struct ipt_entry, ip6t_entry, arpt_entry, ebt_entry)
 *
 * Other fields see above.
 */
struct xt_tgchk_param {
	const char *table;
256
	const void *entryinfo;
257 258 259
	const struct xt_target *target;
	void *targinfo;
	unsigned int hook_mask;
260
	u_int8_t family;
261 262
};

263 264 265 266
/* Target destructor parameters */
struct xt_tgdtor_param {
	const struct xt_target *target;
	void *targinfo;
267
	u_int8_t family;
268 269
};

270 271 272 273 274
struct xt_match
{
	struct list_head list;

	const char name[XT_FUNCTION_MAXNAMELEN-1];
275
	u_int8_t revision;
276 277 278 279 280 281

	/* Return true or false: return FALSE and set *hotdrop = 1 to
           force immediate packet drop. */
	/* Arguments changed since 2.6.9, as this must now handle
	   non-linear skb, using skb_header_pointer and
	   skb_ip_make_writable. */
282
	bool (*match)(const struct sk_buff *skb,
283
		      const struct xt_match_param *);
284 285

	/* Called when user tries to insert an entry of this type. */
286
	bool (*checkentry)(const struct xt_mtchk_param *);
287 288

	/* Called when entry of this type deleted. */
289
	void (*destroy)(const struct xt_mtdtor_param *);
290

291
	/* Called when userspace align differs from kernel space one */
292 293
	void (*compat_from_user)(void *dst, void *src);
	int (*compat_to_user)(void __user *dst, void *src);
294

295 296
	/* Set this to THIS_MODULE if you are a module, otherwise NULL */
	struct module *me;
297

298 299 300
	/* Free to use by each match */
	unsigned long data;

301
	const char *table;
302
	unsigned int matchsize;
303
	unsigned int compatsize;
304 305
	unsigned int hooks;
	unsigned short proto;
306 307

	unsigned short family;
308 309 310 311 312 313 314 315 316 317 318 319
};

/* Registration hooks for targets. */
struct xt_target
{
	struct list_head list;

	const char name[XT_FUNCTION_MAXNAMELEN-1];

	/* Returns verdict. Argument order changed since 2.6.9, as this
	   must now handle non-linear skbs, using skb_copy_bits and
	   skb_ip_make_writable. */
320
	unsigned int (*target)(struct sk_buff *skb,
321
			       const struct xt_target_param *);
322 323 324 325 326

	/* Called when user tries to insert an entry of this type:
           hook_mask is a bitmask of hooks from which it can be
           called. */
	/* Should return true or false. */
327
	bool (*checkentry)(const struct xt_tgchk_param *);
328 329

	/* Called when entry of this type deleted. */
330
	void (*destroy)(const struct xt_tgdtor_param *);
331

332
	/* Called when userspace align differs from kernel space one */
333 334
	void (*compat_from_user)(void *dst, void *src);
	int (*compat_to_user)(void __user *dst, void *src);
335

336 337
	/* Set this to THIS_MODULE if you are a module, otherwise NULL */
	struct module *me;
338

339
	const char *table;
340
	unsigned int targetsize;
341
	unsigned int compatsize;
342 343
	unsigned int hooks;
	unsigned short proto;
344 345

	unsigned short family;
346
	u_int8_t revision;
347 348 349 350 351 352 353 354 355 356 357
};

/* Furniture shopping... */
struct xt_table
{
	struct list_head list;

	/* What hooks you will enter on */
	unsigned int valid_hooks;

	/* Man behind the curtain... */
358
	struct xt_table_info *private;
359 360 361 362

	/* Set this to THIS_MODULE if you are a module, otherwise NULL */
	struct module *me;

363
	u_int8_t af;		/* address/protocol family */
364 365 366

	/* A unique name... */
	const char name[XT_TABLE_MAXNAMELEN];
367 368 369 370 371 372 373 374 375 376 377 378 379 380 381
};

#include <linux/netfilter_ipv4.h>

/* The table itself */
struct xt_table_info
{
	/* Size per table */
	unsigned int size;
	/* Number of entries: FIXME. --RR */
	unsigned int number;
	/* Initial number of entries. Needed for module usage count */
	unsigned int initial_entries;

	/* Entry points and underflows */
382 383
	unsigned int hook_entry[NF_INET_NUMHOOKS];
	unsigned int underflow[NF_INET_NUMHOOKS];
384 385

	/* ipt_entry tables: one per CPU */
386
	/* Note : this field MUST be the last one, see XT_TABLE_INFO_SZ */
387
	void *entries[1];
388 389
};

390 391
#define XT_TABLE_INFO_SZ (offsetof(struct xt_table_info, entries) \
			  + nr_cpu_ids * sizeof(char *))
392 393
extern int xt_register_target(struct xt_target *target);
extern void xt_unregister_target(struct xt_target *target);
394 395 396
extern int xt_register_targets(struct xt_target *target, unsigned int n);
extern void xt_unregister_targets(struct xt_target *target, unsigned int n);

397 398
extern int xt_register_match(struct xt_match *target);
extern void xt_unregister_match(struct xt_match *target);
399 400
extern int xt_register_matches(struct xt_match *match, unsigned int n);
extern void xt_unregister_matches(struct xt_match *match, unsigned int n);
401

402
extern int xt_check_match(struct xt_mtchk_param *,
403
			  unsigned int size, u_int8_t proto, bool inv_proto);
404
extern int xt_check_target(struct xt_tgchk_param *,
405
			   unsigned int size, u_int8_t proto, bool inv_proto);
406

407 408
extern struct xt_table *xt_register_table(struct net *net,
					  struct xt_table *table,
409 410
					  struct xt_table_info *bootstrap,
					  struct xt_table_info *newinfo);
411 412 413 414 415 416 417
extern void *xt_unregister_table(struct xt_table *table);

extern struct xt_table_info *xt_replace_table(struct xt_table *table,
					      unsigned int num_counters,
					      struct xt_table_info *newinfo,
					      int *error);

418 419 420
extern struct xt_match *xt_find_match(u8 af, const char *name, u8 revision);
extern struct xt_target *xt_find_target(u8 af, const char *name, u8 revision);
extern struct xt_target *xt_request_find_target(u8 af, const char *name,
421
						u8 revision);
422 423
extern int xt_find_revision(u8 af, const char *name, u8 revision,
			    int target, int *err);
424

425
extern struct xt_table *xt_find_table_lock(struct net *net, u_int8_t af,
426
					   const char *name);
427 428
extern void xt_table_unlock(struct xt_table *t);

429 430
extern int xt_proto_init(struct net *net, u_int8_t af);
extern void xt_proto_fini(struct net *net, u_int8_t af);
431 432 433

extern struct xt_table_info *xt_alloc_table_info(unsigned int size);
extern void xt_free_table_info(struct xt_table_info *info);
434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474

/*
 * Per-CPU spinlock associated with per-cpu table entries, and
 * with a counter for the "reading" side that allows a recursive
 * reader to avoid taking the lock and deadlocking.
 *
 * "reading" is used by ip/arp/ip6 tables rule processing which runs per-cpu.
 * It needs to ensure that the rules are not being changed while the packet
 * is being processed. In some cases, the read lock will be acquired
 * twice on the same CPU; this is okay because of the count.
 *
 * "writing" is used when reading counters.
 *  During replace any readers that are using the old tables have to complete
 *  before freeing the old table. This is handled by the write locking
 *  necessary for reading the counters.
 */
struct xt_info_lock {
	spinlock_t lock;
	unsigned char readers;
};
DECLARE_PER_CPU(struct xt_info_lock, xt_info_locks);

/*
 * Note: we need to ensure that preemption is disabled before acquiring
 * the per-cpu-variable, so we do it as a two step process rather than
 * using "spin_lock_bh()".
 *
 * We _also_ need to disable bottom half processing before updating our
 * nesting count, to make sure that the only kind of re-entrancy is this
 * code being called by itself: since the count+lock is not an atomic
 * operation, we can allow no races.
 *
 * _Only_ that special combination of being per-cpu and never getting
 * re-entered asynchronously means that the count is safe.
 */
static inline void xt_info_rdlock_bh(void)
{
	struct xt_info_lock *lock;

	local_bh_disable();
	lock = &__get_cpu_var(xt_info_locks);
475
	if (likely(!lock->readers++))
476 477 478 479 480 481 482
		spin_lock(&lock->lock);
}

static inline void xt_info_rdunlock_bh(void)
{
	struct xt_info_lock *lock = &__get_cpu_var(xt_info_locks);

483
	if (likely(!--lock->readers))
484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501
		spin_unlock(&lock->lock);
	local_bh_enable();
}

/*
 * The "writer" side needs to get exclusive access to the lock,
 * regardless of readers.  This must be called with bottom half
 * processing (and thus also preemption) disabled.
 */
static inline void xt_info_wrlock(unsigned int cpu)
{
	spin_lock(&per_cpu(xt_info_locks, cpu).lock);
}

static inline void xt_info_wrunlock(unsigned int cpu)
{
	spin_unlock(&per_cpu(xt_info_locks, cpu).lock);
}
502

503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525
/*
 * This helper is performance critical and must be inlined
 */
static inline unsigned long ifname_compare_aligned(const char *_a,
						   const char *_b,
						   const char *_mask)
{
	const unsigned long *a = (const unsigned long *)_a;
	const unsigned long *b = (const unsigned long *)_b;
	const unsigned long *mask = (const unsigned long *)_mask;
	unsigned long ret;

	ret = (a[0] ^ b[0]) & mask[0];
	if (IFNAMSIZ > sizeof(unsigned long))
		ret |= (a[1] ^ b[1]) & mask[1];
	if (IFNAMSIZ > 2 * sizeof(unsigned long))
		ret |= (a[2] ^ b[2]) & mask[2];
	if (IFNAMSIZ > 3 * sizeof(unsigned long))
		ret |= (a[3] ^ b[3]) & mask[3];
	BUILD_BUG_ON(IFNAMSIZ > 4 * sizeof(unsigned long));
	return ret;
}

526 527 528 529 530 531 532 533 534 535 536
#ifdef CONFIG_COMPAT
#include <net/compat.h>

struct compat_xt_entry_match
{
	union {
		struct {
			u_int16_t match_size;
			char name[XT_FUNCTION_MAXNAMELEN - 1];
			u_int8_t revision;
		} user;
537 538 539 540
		struct {
			u_int16_t match_size;
			compat_uptr_t match;
		} kernel;
541 542 543 544 545 546 547 548 549 550 551 552 553
		u_int16_t match_size;
	} u;
	unsigned char data[0];
};

struct compat_xt_entry_target
{
	union {
		struct {
			u_int16_t target_size;
			char name[XT_FUNCTION_MAXNAMELEN - 1];
			u_int8_t revision;
		} user;
554 555 556 557
		struct {
			u_int16_t target_size;
			compat_uptr_t target;
		} kernel;
558 559 560 561 562 563 564 565 566 567 568
		u_int16_t target_size;
	} u;
	unsigned char data[0];
};

/* FIXME: this works only on 32 bit tasks
 * need to change whole approach in order to calculate align as function of
 * current task alignment */

struct compat_xt_counters
{
569
#if defined(CONFIG_X86_64) || defined(CONFIG_IA64)
570
	u_int32_t cnt[4];
571 572 573
#else
	u_int64_t cnt[2];
#endif
574 575 576 577 578 579 580 581 582 583 584 585
};

struct compat_xt_counters_info
{
	char name[XT_TABLE_MAXNAMELEN];
	compat_uint_t num_counters;
	struct compat_xt_counters counters[0];
};

#define COMPAT_XT_ALIGN(s) (((s) + (__alignof__(struct compat_xt_counters)-1)) \
		& ~(__alignof__(struct compat_xt_counters)-1))

586 587
extern void xt_compat_lock(u_int8_t af);
extern void xt_compat_unlock(u_int8_t af);
588

589 590 591
extern int xt_compat_add_offset(u_int8_t af, unsigned int offset, short delta);
extern void xt_compat_flush_offsets(u_int8_t af);
extern short xt_compat_calc_jump(u_int8_t af, unsigned int offset);
592

593
extern int xt_compat_match_offset(const struct xt_match *match);
594
extern int xt_compat_match_from_user(struct xt_entry_match *m,
595
				     void **dstptr, unsigned int *size);
596
extern int xt_compat_match_to_user(struct xt_entry_match *m,
597
				   void __user **dstptr, unsigned int *size);
598

599
extern int xt_compat_target_offset(const struct xt_target *target);
600
extern void xt_compat_target_from_user(struct xt_entry_target *t,
601
				       void **dstptr, unsigned int *size);
602
extern int xt_compat_target_to_user(struct xt_entry_target *t,
603
				    void __user **dstptr, unsigned int *size);
604 605

#endif /* CONFIG_COMPAT */
606 607 608
#endif /* __KERNEL__ */

#endif /* _X_TABLES_H */