nf_conntrack_netlink.c 47.4 KB
Newer Older
1 2 3 4
/* Connection tracking via netlink socket. Allows for user space
 * protocol helpers and general trouble making from userspace.
 *
 * (C) 2001 by Jay Schulist <jschlst@samba.org>
5
 * (C) 2002-2006 by Harald Welte <laforge@gnumonks.org>
6
 * (C) 2003 by Patrick Mchardy <kaber@trash.net>
7
 * (C) 2005-2008 by Pablo Neira Ayuso <pablo@netfilter.org>
8
 *
9
 * Initial connection tracking via netlink development funded and
10 11 12 13 14 15 16 17 18 19 20
 * generally made possible by Network Robots, Inc. (www.networkrobots.com)
 *
 * Further development of this code funded by Astaro AG (http://www.astaro.com)
 *
 * This software may be used and distributed according to the terms
 * of the GNU General Public License, incorporated herein by reference.
 */

#include <linux/init.h>
#include <linux/module.h>
#include <linux/kernel.h>
21
#include <linux/rculist.h>
22
#include <linux/rculist_nulls.h>
23 24 25 26 27 28
#include <linux/types.h>
#include <linux/timer.h>
#include <linux/skbuff.h>
#include <linux/errno.h>
#include <linux/netlink.h>
#include <linux/spinlock.h>
29
#include <linux/interrupt.h>
30 31

#include <linux/netfilter.h>
32
#include <net/netlink.h>
33
#include <net/sock.h>
34 35
#include <net/netfilter/nf_conntrack.h>
#include <net/netfilter/nf_conntrack_core.h>
36
#include <net/netfilter/nf_conntrack_expect.h>
37 38
#include <net/netfilter/nf_conntrack_helper.h>
#include <net/netfilter/nf_conntrack_l3proto.h>
39
#include <net/netfilter/nf_conntrack_l4proto.h>
40
#include <net/netfilter/nf_conntrack_tuple.h>
41
#include <net/netfilter/nf_conntrack_acct.h>
42 43 44 45
#ifdef CONFIG_NF_NAT_NEEDED
#include <net/netfilter/nf_nat_core.h>
#include <net/netfilter/nf_nat_protocol.h>
#endif
46 47 48 49 50 51

#include <linux/netfilter/nfnetlink.h>
#include <linux/netfilter/nfnetlink_conntrack.h>

MODULE_LICENSE("GPL");

52
static char __initdata version[] = "0.93";
53 54

static inline int
55
ctnetlink_dump_tuples_proto(struct sk_buff *skb,
56
			    const struct nf_conntrack_tuple *tuple,
57
			    struct nf_conntrack_l4proto *l4proto)
58 59
{
	int ret = 0;
60
	struct nlattr *nest_parms;
61

62 63 64
	nest_parms = nla_nest_start(skb, CTA_TUPLE_PROTO | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
65
	NLA_PUT_U8(skb, CTA_PROTO_NUM, tuple->dst.protonum);
66

67 68
	if (likely(l4proto->tuple_to_nlattr))
		ret = l4proto->tuple_to_nlattr(skb, tuple);
69

70
	nla_nest_end(skb, nest_parms);
71 72 73

	return ret;

74
nla_put_failure:
75 76 77 78
	return -1;
}

static inline int
79 80 81
ctnetlink_dump_tuples_ip(struct sk_buff *skb,
			 const struct nf_conntrack_tuple *tuple,
			 struct nf_conntrack_l3proto *l3proto)
82 83
{
	int ret = 0;
84 85 86 87 88
	struct nlattr *nest_parms;

	nest_parms = nla_nest_start(skb, CTA_TUPLE_IP | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
89

90 91
	if (likely(l3proto->tuple_to_nlattr))
		ret = l3proto->tuple_to_nlattr(skb, tuple);
92

93
	nla_nest_end(skb, nest_parms);
94

95 96
	return ret;

97
nla_put_failure:
98 99 100
	return -1;
}

101
static int
102 103 104 105 106
ctnetlink_dump_tuples(struct sk_buff *skb,
		      const struct nf_conntrack_tuple *tuple)
{
	int ret;
	struct nf_conntrack_l3proto *l3proto;
107
	struct nf_conntrack_l4proto *l4proto;
108

109
	l3proto = __nf_ct_l3proto_find(tuple->src.l3num);
110
	ret = ctnetlink_dump_tuples_ip(skb, tuple, l3proto);
111 112 113 114

	if (unlikely(ret < 0))
		return ret;

115
	l4proto = __nf_ct_l4proto_find(tuple->src.l3num, tuple->dst.protonum);
116
	ret = ctnetlink_dump_tuples_proto(skb, tuple, l4proto);
117 118 119 120 121 122 123

	return ret;
}

static inline int
ctnetlink_dump_status(struct sk_buff *skb, const struct nf_conn *ct)
{
124
	NLA_PUT_BE32(skb, CTA_STATUS, htonl(ct->status));
125 126
	return 0;

127
nla_put_failure:
128 129 130 131 132 133
	return -1;
}

static inline int
ctnetlink_dump_timeout(struct sk_buff *skb, const struct nf_conn *ct)
{
134
	long timeout = (ct->timeout.expires - jiffies) / HZ;
135

136
	if (timeout < 0)
137
		timeout = 0;
138

139
	NLA_PUT_BE32(skb, CTA_TIMEOUT, htonl(timeout));
140 141
	return 0;

142
nla_put_failure:
143 144 145 146
	return -1;
}

static inline int
147
ctnetlink_dump_protoinfo(struct sk_buff *skb, struct nf_conn *ct)
148
{
149
	struct nf_conntrack_l4proto *l4proto;
150
	struct nlattr *nest_proto;
151 152
	int ret;

153 154
	l4proto = __nf_ct_l4proto_find(nf_ct_l3num(ct), nf_ct_protonum(ct));
	if (!l4proto->to_nlattr)
155
		return 0;
156

157 158 159
	nest_proto = nla_nest_start(skb, CTA_PROTOINFO | NLA_F_NESTED);
	if (!nest_proto)
		goto nla_put_failure;
160

161
	ret = l4proto->to_nlattr(skb, nest_proto, ct);
162

163
	nla_nest_end(skb, nest_proto);
164 165 166

	return ret;

167
nla_put_failure:
168 169 170 171 172 173
	return -1;
}

static inline int
ctnetlink_dump_helpinfo(struct sk_buff *skb, const struct nf_conn *ct)
{
174
	struct nlattr *nest_helper;
175
	const struct nf_conn_help *help = nfct_help(ct);
176
	struct nf_conntrack_helper *helper;
177

178
	if (!help)
179
		return 0;
180

181 182 183 184
	helper = rcu_dereference(help->helper);
	if (!helper)
		goto out;

185 186 187
	nest_helper = nla_nest_start(skb, CTA_HELP | NLA_F_NESTED);
	if (!nest_helper)
		goto nla_put_failure;
188
	NLA_PUT_STRING(skb, CTA_HELP_NAME, helper->name);
189

190 191
	if (helper->to_nlattr)
		helper->to_nlattr(skb, ct);
192

193
	nla_nest_end(skb, nest_helper);
194
out:
195 196
	return 0;

197
nla_put_failure:
198 199 200
	return -1;
}

201
static int
202 203 204 205
ctnetlink_dump_counters(struct sk_buff *skb, const struct nf_conn *ct,
			enum ip_conntrack_dir dir)
{
	enum ctattr_type type = dir ? CTA_COUNTERS_REPLY: CTA_COUNTERS_ORIG;
206
	struct nlattr *nest_count;
207 208 209 210 211
	const struct nf_conn_counter *acct;

	acct = nf_conn_acct_find(ct);
	if (!acct)
		return 0;
212

213 214 215 216
	nest_count = nla_nest_start(skb, type | NLA_F_NESTED);
	if (!nest_count)
		goto nla_put_failure;

217 218 219 220
	NLA_PUT_BE64(skb, CTA_COUNTERS_PACKETS,
		     cpu_to_be64(acct[dir].packets));
	NLA_PUT_BE64(skb, CTA_COUNTERS_BYTES,
		     cpu_to_be64(acct[dir].bytes));
221

222
	nla_nest_end(skb, nest_count);
223 224 225

	return 0;

226
nla_put_failure:
227 228 229 230 231 232 233
	return -1;
}

#ifdef CONFIG_NF_CONNTRACK_MARK
static inline int
ctnetlink_dump_mark(struct sk_buff *skb, const struct nf_conn *ct)
{
234
	NLA_PUT_BE32(skb, CTA_MARK, htonl(ct->mark));
235 236
	return 0;

237
nla_put_failure:
238 239 240 241 242 243
	return -1;
}
#else
#define ctnetlink_dump_mark(a, b) (0)
#endif

244 245 246 247
#ifdef CONFIG_NF_CONNTRACK_SECMARK
static inline int
ctnetlink_dump_secmark(struct sk_buff *skb, const struct nf_conn *ct)
{
248
	NLA_PUT_BE32(skb, CTA_SECMARK, htonl(ct->secmark));
249 250 251 252 253 254 255 256 257
	return 0;

nla_put_failure:
	return -1;
}
#else
#define ctnetlink_dump_secmark(a, b) (0)
#endif

258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280
#define master_tuple(ct) &(ct->master->tuplehash[IP_CT_DIR_ORIGINAL].tuple)

static inline int
ctnetlink_dump_master(struct sk_buff *skb, const struct nf_conn *ct)
{
	struct nlattr *nest_parms;

	if (!(ct->status & IPS_EXPECTED))
		return 0;

	nest_parms = nla_nest_start(skb, CTA_TUPLE_MASTER | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
	if (ctnetlink_dump_tuples(skb, master_tuple(ct)) < 0)
		goto nla_put_failure;
	nla_nest_end(skb, nest_parms);

	return 0;

nla_put_failure:
	return -1;
}

281
#ifdef CONFIG_NF_NAT_NEEDED
282
static int
283 284 285 286 287 288 289 290
dump_nat_seq_adj(struct sk_buff *skb, const struct nf_nat_seq *natseq, int type)
{
	struct nlattr *nest_parms;

	nest_parms = nla_nest_start(skb, type | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;

291 292 293 294 295 296
	NLA_PUT_BE32(skb, CTA_NAT_SEQ_CORRECTION_POS,
		     htonl(natseq->correction_pos));
	NLA_PUT_BE32(skb, CTA_NAT_SEQ_OFFSET_BEFORE,
		     htonl(natseq->offset_before));
	NLA_PUT_BE32(skb, CTA_NAT_SEQ_OFFSET_AFTER,
		     htonl(natseq->offset_after));
297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328

	nla_nest_end(skb, nest_parms);

	return 0;

nla_put_failure:
	return -1;
}

static inline int
ctnetlink_dump_nat_seq_adj(struct sk_buff *skb, const struct nf_conn *ct)
{
	struct nf_nat_seq *natseq;
	struct nf_conn_nat *nat = nfct_nat(ct);

	if (!(ct->status & IPS_SEQ_ADJUST) || !nat)
		return 0;

	natseq = &nat->seq[IP_CT_DIR_ORIGINAL];
	if (dump_nat_seq_adj(skb, natseq, CTA_NAT_SEQ_ADJ_ORIG) == -1)
		return -1;

	natseq = &nat->seq[IP_CT_DIR_REPLY];
	if (dump_nat_seq_adj(skb, natseq, CTA_NAT_SEQ_ADJ_REPLY) == -1)
		return -1;

	return 0;
}
#else
#define ctnetlink_dump_nat_seq_adj(a, b) (0)
#endif

329 330 331
static inline int
ctnetlink_dump_id(struct sk_buff *skb, const struct nf_conn *ct)
{
332
	NLA_PUT_BE32(skb, CTA_ID, htonl((unsigned long)ct));
333 334
	return 0;

335
nla_put_failure:
336 337 338 339 340 341
	return -1;
}

static inline int
ctnetlink_dump_use(struct sk_buff *skb, const struct nf_conn *ct)
{
342
	NLA_PUT_BE32(skb, CTA_USE, htonl(atomic_read(&ct->ct_general.use)));
343 344
	return 0;

345
nla_put_failure:
346 347 348 349 350
	return -1;
}

static int
ctnetlink_fill_info(struct sk_buff *skb, u32 pid, u32 seq,
351
		    int event, struct nf_conn *ct)
352 353 354
{
	struct nlmsghdr *nlh;
	struct nfgenmsg *nfmsg;
355
	struct nlattr *nest_parms;
356
	unsigned int flags = pid ? NLM_F_MULTI : 0;
357 358

	event |= NFNL_SUBSYS_CTNETLINK << 8;
359 360 361
	nlh = nlmsg_put(skb, pid, seq, event, sizeof(*nfmsg), flags);
	if (nlh == NULL)
		goto nlmsg_failure;
362

363
	nfmsg = nlmsg_data(nlh);
364
	nfmsg->nfgen_family = nf_ct_l3num(ct);
365 366 367
	nfmsg->version      = NFNETLINK_V0;
	nfmsg->res_id	    = 0;

368 369 370
	nest_parms = nla_nest_start(skb, CTA_TUPLE_ORIG | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
371
	if (ctnetlink_dump_tuples(skb, nf_ct_tuple(ct, IP_CT_DIR_ORIGINAL)) < 0)
372 373
		goto nla_put_failure;
	nla_nest_end(skb, nest_parms);
374

375 376 377
	nest_parms = nla_nest_start(skb, CTA_TUPLE_REPLY | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
378
	if (ctnetlink_dump_tuples(skb, nf_ct_tuple(ct, IP_CT_DIR_REPLY)) < 0)
379 380
		goto nla_put_failure;
	nla_nest_end(skb, nest_parms);
381 382 383 384 385 386 387 388

	if (ctnetlink_dump_status(skb, ct) < 0 ||
	    ctnetlink_dump_timeout(skb, ct) < 0 ||
	    ctnetlink_dump_counters(skb, ct, IP_CT_DIR_ORIGINAL) < 0 ||
	    ctnetlink_dump_counters(skb, ct, IP_CT_DIR_REPLY) < 0 ||
	    ctnetlink_dump_protoinfo(skb, ct) < 0 ||
	    ctnetlink_dump_helpinfo(skb, ct) < 0 ||
	    ctnetlink_dump_mark(skb, ct) < 0 ||
389
	    ctnetlink_dump_secmark(skb, ct) < 0 ||
390
	    ctnetlink_dump_id(skb, ct) < 0 ||
391
	    ctnetlink_dump_use(skb, ct) < 0 ||
392
	    ctnetlink_dump_master(skb, ct) < 0 ||
393
	    ctnetlink_dump_nat_seq_adj(skb, ct) < 0)
394
		goto nla_put_failure;
395

396
	nlmsg_end(skb, nlh);
397 398 399
	return skb->len;

nlmsg_failure:
400
nla_put_failure:
401
	nlmsg_cancel(skb, nlh);
402 403 404 405
	return -1;
}

#ifdef CONFIG_NF_CONNTRACK_EVENTS
406 407
static inline size_t
ctnetlink_proto_size(const struct nf_conn *ct)
408 409 410
{
	struct nf_conntrack_l3proto *l3proto;
	struct nf_conntrack_l4proto *l4proto;
411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433
	size_t len = 0;

	rcu_read_lock();
	l3proto = __nf_ct_l3proto_find(nf_ct_l3num(ct));
	len += l3proto->nla_size;

	l4proto = __nf_ct_l4proto_find(nf_ct_l3num(ct), nf_ct_protonum(ct));
	len += l4proto->nla_size;
	rcu_read_unlock();

	return len;
}

static inline size_t
ctnetlink_nlmsg_size(const struct nf_conn *ct)
{
	return NLMSG_ALIGN(sizeof(struct nfgenmsg))
	       + 3 * nla_total_size(0) /* CTA_TUPLE_ORIG|REPL|MASTER */
	       + 3 * nla_total_size(0) /* CTA_TUPLE_IP */
	       + 3 * nla_total_size(0) /* CTA_TUPLE_PROTO */
	       + 3 * nla_total_size(sizeof(u_int8_t)) /* CTA_PROTO_NUM */
	       + nla_total_size(sizeof(u_int32_t)) /* CTA_ID */
	       + nla_total_size(sizeof(u_int32_t)) /* CTA_STATUS */
434
#ifdef CONFIG_NF_CT_ACCT
435 436 437
	       + 2 * nla_total_size(0) /* CTA_COUNTERS_ORIG|REPL */
	       + 2 * nla_total_size(sizeof(uint64_t)) /* CTA_COUNTERS_PACKETS */
	       + 2 * nla_total_size(sizeof(uint64_t)) /* CTA_COUNTERS_BYTES */
438
#endif
439 440 441 442
	       + nla_total_size(sizeof(u_int32_t)) /* CTA_TIMEOUT */
	       + nla_total_size(0) /* CTA_PROTOINFO */
	       + nla_total_size(0) /* CTA_HELP */
	       + nla_total_size(NF_CT_HELPER_NAME_LEN) /* CTA_HELP_NAME */
443
#ifdef CONFIG_NF_CONNTRACK_SECMARK
444
	       + nla_total_size(sizeof(u_int32_t)) /* CTA_SECMARK */
445 446
#endif
#ifdef CONFIG_NF_NAT_NEEDED
447 448
	       + 2 * nla_total_size(0) /* CTA_NAT_SEQ_ADJ_ORIG|REPL */
	       + 6 * nla_total_size(sizeof(u_int32_t)) /* CTA_NAT_SEQ_OFFSET */
449 450
#endif
#ifdef CONFIG_NF_CONNTRACK_MARK
451
	       + nla_total_size(sizeof(u_int32_t)) /* CTA_MARK */
452
#endif
453 454
	       + ctnetlink_proto_size(ct)
	       ;
455 456
}

457 458
static int
ctnetlink_conntrack_event(unsigned int events, struct nf_ct_event *item)
459
{
460
	struct net *net;
461 462
	struct nlmsghdr *nlh;
	struct nfgenmsg *nfmsg;
463
	struct nlattr *nest_parms;
464
	struct nf_conn *ct = item->ct;
465 466 467
	struct sk_buff *skb;
	unsigned int type;
	unsigned int flags = 0, group;
468
	int err;
469 470 471

	/* ignore our fake conntrack entry */
	if (ct == &nf_conntrack_untracked)
472
		return 0;
473

474
	if (events & (1 << IPCT_DESTROY)) {
475 476
		type = IPCTNL_MSG_CT_DELETE;
		group = NFNLGRP_CONNTRACK_DESTROY;
477
	} else  if (events & ((1 << IPCT_NEW) | (1 << IPCT_RELATED))) {
478 479 480
		type = IPCTNL_MSG_CT_NEW;
		flags = NLM_F_CREATE|NLM_F_EXCL;
		group = NFNLGRP_CONNTRACK_NEW;
481
	} else  if (events) {
482 483 484
		type = IPCTNL_MSG_CT_NEW;
		group = NFNLGRP_CONNTRACK_UPDATE;
	} else
485
		return 0;
486

487 488
	net = nf_ct_net(ct);
	if (!item->report && !nfnetlink_has_listeners(net, group))
489
		return 0;
490

491 492
	skb = nlmsg_new(ctnetlink_nlmsg_size(ct), GFP_ATOMIC);
	if (skb == NULL)
493
		goto errout;
494 495

	type |= NFNL_SUBSYS_CTNETLINK << 8;
496 497 498
	nlh = nlmsg_put(skb, item->pid, 0, type, sizeof(*nfmsg), flags);
	if (nlh == NULL)
		goto nlmsg_failure;
499

500
	nfmsg = nlmsg_data(nlh);
501
	nfmsg->nfgen_family = nf_ct_l3num(ct);
502 503 504
	nfmsg->version	= NFNETLINK_V0;
	nfmsg->res_id	= 0;

505
	rcu_read_lock();
506 507 508
	nest_parms = nla_nest_start(skb, CTA_TUPLE_ORIG | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
509
	if (ctnetlink_dump_tuples(skb, nf_ct_tuple(ct, IP_CT_DIR_ORIGINAL)) < 0)
510 511
		goto nla_put_failure;
	nla_nest_end(skb, nest_parms);
512

513 514 515
	nest_parms = nla_nest_start(skb, CTA_TUPLE_REPLY | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
516
	if (ctnetlink_dump_tuples(skb, nf_ct_tuple(ct, IP_CT_DIR_REPLY)) < 0)
517 518
		goto nla_put_failure;
	nla_nest_end(skb, nest_parms);
519

520 521 522
	if (ctnetlink_dump_id(skb, ct) < 0)
		goto nla_put_failure;

523 524 525
	if (ctnetlink_dump_status(skb, ct) < 0)
		goto nla_put_failure;

526
	if (events & (1 << IPCT_DESTROY)) {
527 528
		if (ctnetlink_dump_counters(skb, ct, IP_CT_DIR_ORIGINAL) < 0 ||
		    ctnetlink_dump_counters(skb, ct, IP_CT_DIR_REPLY) < 0)
529
			goto nla_put_failure;
530 531
	} else {
		if (ctnetlink_dump_timeout(skb, ct) < 0)
532
			goto nla_put_failure;
533

534
		if (events & (1 << IPCT_PROTOINFO)
535
		    && ctnetlink_dump_protoinfo(skb, ct) < 0)
536
			goto nla_put_failure;
537

538
		if ((events & (1 << IPCT_HELPER) || nfct_help(ct))
539
		    && ctnetlink_dump_helpinfo(skb, ct) < 0)
540
			goto nla_put_failure;
541

542
#ifdef CONFIG_NF_CONNTRACK_SECMARK
543
		if ((events & (1 << IPCT_SECMARK) || ct->secmark)
544 545 546
		    && ctnetlink_dump_secmark(skb, ct) < 0)
			goto nla_put_failure;
#endif
547

548
		if (events & (1 << IPCT_RELATED) &&
549 550 551
		    ctnetlink_dump_master(skb, ct) < 0)
			goto nla_put_failure;

552
		if (events & (1 << IPCT_NATSEQADJ) &&
553 554
		    ctnetlink_dump_nat_seq_adj(skb, ct) < 0)
			goto nla_put_failure;
555
	}
556

557
#ifdef CONFIG_NF_CONNTRACK_MARK
558
	if ((events & (1 << IPCT_MARK) || ct->mark)
559 560 561
	    && ctnetlink_dump_mark(skb, ct) < 0)
		goto nla_put_failure;
#endif
562
	rcu_read_unlock();
563

564
	nlmsg_end(skb, nlh);
565
	err = nfnetlink_send(skb, net, item->pid, group, item->report,
566
			     GFP_ATOMIC);
567 568 569
	if (err == -ENOBUFS || err == -EAGAIN)
		return -ENOBUFS;

570
	return 0;
571

572
nla_put_failure:
573
	rcu_read_unlock();
574
	nlmsg_cancel(skb, nlh);
575
nlmsg_failure:
576
	kfree_skb(skb);
577
errout:
578
	nfnetlink_set_err(net, 0, group, -ENOBUFS);
579
	return 0;
580 581 582 583 584
}
#endif /* CONFIG_NF_CONNTRACK_EVENTS */

static int ctnetlink_done(struct netlink_callback *cb)
{
585 586
	if (cb->args[1])
		nf_ct_put((struct nf_conn *)cb->args[1]);
587 588 589 590 591 592
	return 0;
}

static int
ctnetlink_dump_table(struct sk_buff *skb, struct netlink_callback *cb)
{
593
	struct net *net = sock_net(skb->sk);
594
	struct nf_conn *ct, *last;
595
	struct nf_conntrack_tuple_hash *h;
596
	struct hlist_nulls_node *n;
597
	struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
598
	u_int8_t l3proto = nfmsg->nfgen_family;
599

600
	rcu_read_lock();
601
	last = (struct nf_conn *)cb->args[1];
602 603
	for (; cb->args[0] < nf_conntrack_htable_size; cb->args[0]++) {
restart:
604
		hlist_nulls_for_each_entry_rcu(h, n, &net->ct.hash[cb->args[0]],
605
					 hnnode) {
606
			if (NF_CT_DIRECTION(h) != IP_CT_DIR_ORIGINAL)
607 608
				continue;
			ct = nf_ct_tuplehash_to_ctrack(h);
609 610
			if (!atomic_inc_not_zero(&ct->ct_general.use))
				continue;
611 612 613
			/* Dump entries of a given L3 protocol number.
			 * If it is not specified, ie. l3proto == 0,
			 * then dump everything. */
614
			if (l3proto && nf_ct_l3num(ct) != l3proto)
615
				goto releasect;
616 617
			if (cb->args[1]) {
				if (ct != last)
618
					goto releasect;
619
				cb->args[1] = 0;
620
			}
621
			if (ctnetlink_fill_info(skb, NETLINK_CB(cb->skb).pid,
622
						cb->nlh->nlmsg_seq,
623
						IPCTNL_MSG_CT_NEW, ct) < 0) {
624
				cb->args[1] = (unsigned long)ct;
625
				goto out;
626
			}
627

628
			if (NFNL_MSG_TYPE(cb->nlh->nlmsg_type) ==
629 630 631 632 633 634 635
						IPCTNL_MSG_CT_GET_CTRZERO) {
				struct nf_conn_counter *acct;

				acct = nf_conn_acct_find(ct);
				if (acct)
					memset(acct, 0, sizeof(struct nf_conn_counter[IP_CT_DIR_MAX]));
			}
636 637
releasect:
		nf_ct_put(ct);
638
		}
639
		if (cb->args[1]) {
640 641
			cb->args[1] = 0;
			goto restart;
642 643
		}
	}
644
out:
645
	rcu_read_unlock();
646 647
	if (last)
		nf_ct_put(last);
648 649 650 651 652

	return skb->len;
}

static inline int
653
ctnetlink_parse_tuple_ip(struct nlattr *attr, struct nf_conntrack_tuple *tuple)
654
{
655
	struct nlattr *tb[CTA_IP_MAX+1];
656 657 658
	struct nf_conntrack_l3proto *l3proto;
	int ret = 0;

659
	nla_parse_nested(tb, CTA_IP_MAX, attr, NULL);
660

661 662
	rcu_read_lock();
	l3proto = __nf_ct_l3proto_find(tuple->src.l3num);
663

664 665 666 667 668 669
	if (likely(l3proto->nlattr_to_tuple)) {
		ret = nla_validate_nested(attr, CTA_IP_MAX,
					  l3proto->nla_policy);
		if (ret == 0)
			ret = l3proto->nlattr_to_tuple(tb, tuple);
	}
670

671
	rcu_read_unlock();
672 673 674 675

	return ret;
}

676 677
static const struct nla_policy proto_nla_policy[CTA_PROTO_MAX+1] = {
	[CTA_PROTO_NUM]	= { .type = NLA_U8 },
678 679 680
};

static inline int
681
ctnetlink_parse_tuple_proto(struct nlattr *attr,
682 683
			    struct nf_conntrack_tuple *tuple)
{
684
	struct nlattr *tb[CTA_PROTO_MAX+1];
685
	struct nf_conntrack_l4proto *l4proto;
686 687
	int ret = 0;

688 689 690
	ret = nla_parse_nested(tb, CTA_PROTO_MAX, attr, proto_nla_policy);
	if (ret < 0)
		return ret;
691

692
	if (!tb[CTA_PROTO_NUM])
693
		return -EINVAL;
694
	tuple->dst.protonum = nla_get_u8(tb[CTA_PROTO_NUM]);
695

696 697
	rcu_read_lock();
	l4proto = __nf_ct_l4proto_find(tuple->src.l3num, tuple->dst.protonum);
698

699 700 701 702 703 704
	if (likely(l4proto->nlattr_to_tuple)) {
		ret = nla_validate_nested(attr, CTA_PROTO_MAX,
					  l4proto->nla_policy);
		if (ret == 0)
			ret = l4proto->nlattr_to_tuple(tb, tuple);
	}
705

706
	rcu_read_unlock();
707

708 709 710
	return ret;
}

711
static int
712 713
ctnetlink_parse_tuple(const struct nlattr * const cda[],
		      struct nf_conntrack_tuple *tuple,
714 715
		      enum ctattr_tuple type, u_int8_t l3num)
{
716
	struct nlattr *tb[CTA_TUPLE_MAX+1];
717 718 719 720
	int err;

	memset(tuple, 0, sizeof(*tuple));

721
	nla_parse_nested(tb, CTA_TUPLE_MAX, cda[type], NULL);
722

723
	if (!tb[CTA_TUPLE_IP])
724 725 726 727
		return -EINVAL;

	tuple->src.l3num = l3num;

728
	err = ctnetlink_parse_tuple_ip(tb[CTA_TUPLE_IP], tuple);
729 730 731
	if (err < 0)
		return err;

732
	if (!tb[CTA_TUPLE_PROTO])
733 734
		return -EINVAL;

735
	err = ctnetlink_parse_tuple_proto(tb[CTA_TUPLE_PROTO], tuple);
736 737 738 739 740 741 742 743 744 745 746 747 748
	if (err < 0)
		return err;

	/* orig and expect tuples get DIR_ORIGINAL */
	if (type == CTA_TUPLE_REPLY)
		tuple->dst.dir = IP_CT_DIR_REPLY;
	else
		tuple->dst.dir = IP_CT_DIR_ORIGINAL;

	return 0;
}

static inline int
749
ctnetlink_parse_help(const struct nlattr *attr, char **helper_name)
750
{
751
	struct nlattr *tb[CTA_HELP_MAX+1];
752

753
	nla_parse_nested(tb, CTA_HELP_MAX, attr, NULL);
754

755
	if (!tb[CTA_HELP_NAME])
756 757
		return -EINVAL;

758
	*helper_name = nla_data(tb[CTA_HELP_NAME]);
759 760 761 762

	return 0;
}

763 764 765 766 767 768
static const struct nla_policy ct_nla_policy[CTA_MAX+1] = {
	[CTA_STATUS] 		= { .type = NLA_U32 },
	[CTA_TIMEOUT] 		= { .type = NLA_U32 },
	[CTA_MARK]		= { .type = NLA_U32 },
	[CTA_USE]		= { .type = NLA_U32 },
	[CTA_ID]		= { .type = NLA_U32 },
769 770 771
};

static int
772
ctnetlink_del_conntrack(struct sock *ctnl, struct sk_buff *skb,
773 774
			const struct nlmsghdr *nlh,
			const struct nlattr * const cda[])
775
{
776
	struct net *net = sock_net(ctnl);
777 778 779
	struct nf_conntrack_tuple_hash *h;
	struct nf_conntrack_tuple tuple;
	struct nf_conn *ct;
780
	struct nfgenmsg *nfmsg = nlmsg_data(nlh);
781 782 783
	u_int8_t u3 = nfmsg->nfgen_family;
	int err = 0;

784
	if (cda[CTA_TUPLE_ORIG])
785
		err = ctnetlink_parse_tuple(cda, &tuple, CTA_TUPLE_ORIG, u3);
786
	else if (cda[CTA_TUPLE_REPLY])
787 788 789
		err = ctnetlink_parse_tuple(cda, &tuple, CTA_TUPLE_REPLY, u3);
	else {
		/* Flush the whole table */
790
		nf_conntrack_flush_report(net,
791 792
					 NETLINK_CB(skb).pid,
					 nlmsg_report(nlh));
793 794 795 796 797 798
		return 0;
	}

	if (err < 0)
		return err;

799
	h = nf_conntrack_find_get(net, &tuple);
800
	if (!h)
801 802 803
		return -ENOENT;

	ct = nf_ct_tuplehash_to_ctrack(h);
804

805
	if (cda[CTA_ID]) {
806
		u_int32_t id = ntohl(nla_get_be32(cda[CTA_ID]));
807
		if (id != (u32)(unsigned long)ct) {
808 809 810
			nf_ct_put(ct);
			return -ENOENT;
		}
811
	}
812

813 814 815 816 817 818 819 820 821
	if (nf_conntrack_event_report(IPCT_DESTROY, ct,
				      NETLINK_CB(skb).pid,
				      nlmsg_report(nlh)) < 0) {
		nf_ct_delete_from_lists(ct);
		/* we failed to report the event, try later */
		nf_ct_insert_dying_list(ct);
		nf_ct_put(ct);
		return 0;
	}
822 823 824 825

	/* death_by_timeout would report the event again */
	set_bit(IPS_DYING_BIT, &ct->status);

826
	nf_ct_kill(ct);
827 828 829 830 831 832
	nf_ct_put(ct);

	return 0;
}

static int
833
ctnetlink_get_conntrack(struct sock *ctnl, struct sk_buff *skb,
834 835
			const struct nlmsghdr *nlh,
			const struct nlattr * const cda[])
836
{
837
	struct net *net = sock_net(ctnl);
838 839 840 841
	struct nf_conntrack_tuple_hash *h;
	struct nf_conntrack_tuple tuple;
	struct nf_conn *ct;
	struct sk_buff *skb2 = NULL;
842
	struct nfgenmsg *nfmsg = nlmsg_data(nlh);
843 844 845
	u_int8_t u3 = nfmsg->nfgen_family;
	int err = 0;

846
	if (nlh->nlmsg_flags & NLM_F_DUMP)
847 848
		return netlink_dump_start(ctnl, skb, nlh, ctnetlink_dump_table,
					  ctnetlink_done);
849

850
	if (cda[CTA_TUPLE_ORIG])
851
		err = ctnetlink_parse_tuple(cda, &tuple, CTA_TUPLE_ORIG, u3);
852
	else if (cda[CTA_TUPLE_REPLY])
853 854 855 856 857 858 859
		err = ctnetlink_parse_tuple(cda, &tuple, CTA_TUPLE_REPLY, u3);
	else
		return -EINVAL;

	if (err < 0)
		return err;

860
	h = nf_conntrack_find_get(net, &tuple);
861
	if (!h)
862
		return -ENOENT;
863

864 865 866
	ct = nf_ct_tuplehash_to_ctrack(h);

	err = -ENOMEM;
867 868
	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
	if (skb2 == NULL) {
869 870 871 872
		nf_ct_put(ct);
		return -ENOMEM;
	}

873
	rcu_read_lock();
874
	err = ctnetlink_fill_info(skb2, NETLINK_CB(skb).pid, nlh->nlmsg_seq,
875
				  IPCTNL_MSG_CT_NEW, ct);
876
	rcu_read_unlock();
877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892
	nf_ct_put(ct);
	if (err <= 0)
		goto free;

	err = netlink_unicast(ctnl, skb2, NETLINK_CB(skb).pid, MSG_DONTWAIT);
	if (err < 0)
		goto out;

	return 0;

free:
	kfree_skb(skb2);
out:
	return err;
}

893
#ifdef CONFIG_NF_NAT_NEEDED
894 895 896
static int
ctnetlink_parse_nat_setup(struct nf_conn *ct,
			  enum nf_nat_manip_type manip,
897
			  const struct nlattr *attr)
898 899 900 901 902
{
	typeof(nfnetlink_parse_nat_setup_hook) parse_nat_setup;

	parse_nat_setup = rcu_dereference(nfnetlink_parse_nat_setup_hook);
	if (!parse_nat_setup) {
903
#ifdef CONFIG_MODULES
904
		rcu_read_unlock();
905
		spin_unlock_bh(&nf_conntrack_lock);
906 907 908
		nfnl_unlock();
		if (request_module("nf-nat-ipv4") < 0) {
			nfnl_lock();
909
			spin_lock_bh(&nf_conntrack_lock);
910 911 912 913
			rcu_read_lock();
			return -EOPNOTSUPP;
		}
		nfnl_lock();
914
		spin_lock_bh(&nf_conntrack_lock);
915 916 917 918 919 920 921 922 923
		rcu_read_lock();
		if (nfnetlink_parse_nat_setup_hook)
			return -EAGAIN;
#endif
		return -EOPNOTSUPP;
	}

	return parse_nat_setup(ct, manip, attr);
}
924
#endif
925

926
static int
927
ctnetlink_change_status(struct nf_conn *ct, const struct nlattr * const cda[])
928 929
{
	unsigned long d;
930
	unsigned int status = ntohl(nla_get_be32(cda[CTA_STATUS]));
931 932 933 934
	d = ct->status ^ status;

	if (d & (IPS_EXPECTED|IPS_CONFIRMED|IPS_DYING))
		/* unchangeable */
935
		return -EBUSY;
936

937 938
	if (d & IPS_SEEN_REPLY && !(status & IPS_SEEN_REPLY))
		/* SEEN_REPLY bit can only be set */
939
		return -EBUSY;
940

941 942
	if (d & IPS_ASSURED && !(status & IPS_ASSURED))
		/* ASSURED bit can only be set */
943
		return -EBUSY;
944 945 946

	/* Be careful here, modifying NAT bits can screw up things,
	 * so don't let users modify them directly if they don't pass
947
	 * nf_nat_range. */
948 949 950 951
	ct->status |= status & ~(IPS_NAT_DONE_MASK | IPS_NAT_MASK);
	return 0;
}

952
static int
953
ctnetlink_change_nat(struct nf_conn *ct, const struct nlattr * const cda[])
954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976
{
#ifdef CONFIG_NF_NAT_NEEDED
	int ret;

	if (cda[CTA_NAT_DST]) {
		ret = ctnetlink_parse_nat_setup(ct,
						IP_NAT_MANIP_DST,
						cda[CTA_NAT_DST]);
		if (ret < 0)
			return ret;
	}
	if (cda[CTA_NAT_SRC]) {
		ret = ctnetlink_parse_nat_setup(ct,
						IP_NAT_MANIP_SRC,
						cda[CTA_NAT_SRC]);
		if (ret < 0)
			return ret;
	}
	return 0;
#else
	return -EOPNOTSUPP;
#endif
}
977 978

static inline int
979
ctnetlink_change_helper(struct nf_conn *ct, const struct nlattr * const cda[])
980 981
{
	struct nf_conntrack_helper *helper;
982
	struct nf_conn_help *help = nfct_help(ct);
983
	char *helpname = NULL;
984 985 986 987
	int err;

	/* don't change helper of sibling connections */
	if (ct->master)
988
		return -EBUSY;
989

990
	err = ctnetlink_parse_help(cda[CTA_HELP], &helpname);
991 992 993
	if (err < 0)
		return err;

994 995
	if (!strcmp(helpname, "")) {
		if (help && help->helper) {
996 997
			/* we had a helper before ... */
			nf_ct_remove_expectations(ct);
998
			rcu_assign_pointer(help->helper, NULL);
999
		}
1000 1001

		return 0;
1002
	}
1003

1004 1005
	helper = __nf_conntrack_helper_find(helpname, nf_ct_l3num(ct),
					    nf_ct_protonum(ct));
1006 1007 1008 1009 1010 1011 1012 1013 1014 1015
	if (helper == NULL) {
#ifdef CONFIG_MODULES
		spin_unlock_bh(&nf_conntrack_lock);

		if (request_module("nfct-helper-%s", helpname) < 0) {
			spin_lock_bh(&nf_conntrack_lock);
			return -EOPNOTSUPP;
		}

		spin_lock_bh(&nf_conntrack_lock);
1016 1017
		helper = __nf_conntrack_helper_find(helpname, nf_ct_l3num(ct),
						    nf_ct_protonum(ct));
1018 1019 1020
		if (helper)
			return -EAGAIN;
#endif
1021
		return -EOPNOTSUPP;
1022
	}
1023

1024 1025 1026 1027 1028 1029 1030 1031
	if (help) {
		if (help->helper == helper)
			return 0;
		if (help->helper)
			return -EBUSY;
		/* need to zero data of old helper */
		memset(&help->help, 0, sizeof(help->help));
	} else {
1032
		help = nf_ct_helper_ext_add(ct, GFP_ATOMIC);
1033 1034 1035
		if (help == NULL)
			return -ENOMEM;
	}
1036

1037
	rcu_assign_pointer(help->helper, helper);
1038 1039 1040 1041 1042

	return 0;
}

static inline int
1043
ctnetlink_change_timeout(struct nf_conn *ct, const struct nlattr * const cda[])
1044
{
1045
	u_int32_t timeout = ntohl(nla_get_be32(cda[CTA_TIMEOUT]));
1046

1047 1048 1049 1050 1051 1052 1053 1054 1055 1056
	if (!del_timer(&ct->timeout))
		return -ETIME;

	ct->timeout.expires = jiffies + timeout * HZ;
	add_timer(&ct->timeout);

	return 0;
}

static inline int
1057
ctnetlink_change_protoinfo(struct nf_conn *ct, const struct nlattr * const cda[])
1058
{
1059 1060
	const struct nlattr *attr = cda[CTA_PROTOINFO];
	struct nlattr *tb[CTA_PROTOINFO_MAX+1];
1061
	struct nf_conntrack_l4proto *l4proto;
1062 1063
	int err = 0;

1064
	nla_parse_nested(tb, CTA_PROTOINFO_MAX, attr, NULL);
1065

1066 1067
	rcu_read_lock();
	l4proto = __nf_ct_l4proto_find(nf_ct_l3num(ct), nf_ct_protonum(ct));
1068 1069
	if (l4proto->from_nlattr)
		err = l4proto->from_nlattr(tb, ct);
1070
	rcu_read_unlock();
1071 1072 1073 1074

	return err;
}

1075 1076
#ifdef CONFIG_NF_NAT_NEEDED
static inline int
1077
change_nat_seq_adj(struct nf_nat_seq *natseq, const struct nlattr * const attr)
1078 1079 1080 1081 1082 1083 1084 1085 1086
{
	struct nlattr *cda[CTA_NAT_SEQ_MAX+1];

	nla_parse_nested(cda, CTA_NAT_SEQ_MAX, attr, NULL);

	if (!cda[CTA_NAT_SEQ_CORRECTION_POS])
		return -EINVAL;

	natseq->correction_pos =
1087
		ntohl(nla_get_be32(cda[CTA_NAT_SEQ_CORRECTION_POS]));
1088 1089 1090 1091 1092

	if (!cda[CTA_NAT_SEQ_OFFSET_BEFORE])
		return -EINVAL;

	natseq->offset_before =
1093
		ntohl(nla_get_be32(cda[CTA_NAT_SEQ_OFFSET_BEFORE]));
1094 1095 1096 1097 1098

	if (!cda[CTA_NAT_SEQ_OFFSET_AFTER])
		return -EINVAL;

	natseq->offset_after =
1099
		ntohl(nla_get_be32(cda[CTA_NAT_SEQ_OFFSET_AFTER]));
1100 1101 1102 1103 1104

	return 0;
}

static int
1105 1106
ctnetlink_change_nat_seq_adj(struct nf_conn *ct,
			     const struct nlattr * const cda[])
1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135
{
	int ret = 0;
	struct nf_conn_nat *nat = nfct_nat(ct);

	if (!nat)
		return 0;

	if (cda[CTA_NAT_SEQ_ADJ_ORIG]) {
		ret = change_nat_seq_adj(&nat->seq[IP_CT_DIR_ORIGINAL],
					 cda[CTA_NAT_SEQ_ADJ_ORIG]);
		if (ret < 0)
			return ret;

		ct->status |= IPS_SEQ_ADJUST;
	}

	if (cda[CTA_NAT_SEQ_ADJ_REPLY]) {
		ret = change_nat_seq_adj(&nat->seq[IP_CT_DIR_REPLY],
					 cda[CTA_NAT_SEQ_ADJ_REPLY]);
		if (ret < 0)
			return ret;

		ct->status |= IPS_SEQ_ADJUST;
	}

	return 0;
}
#endif

1136
static int
1137 1138
ctnetlink_change_conntrack(struct nf_conn *ct,
			   const struct nlattr * const cda[])
1139 1140 1141
{
	int err;

1142 1143 1144 1145
	/* only allow NAT changes and master assignation for new conntracks */
	if (cda[CTA_NAT_SRC] || cda[CTA_NAT_DST] || cda[CTA_TUPLE_MASTER])
		return -EOPNOTSUPP;

1146
	if (cda[CTA_HELP]) {
1147 1148 1149 1150 1151
		err = ctnetlink_change_helper(ct, cda);
		if (err < 0)
			return err;
	}

1152
	if (cda[CTA_TIMEOUT]) {
1153 1154 1155 1156 1157
		err = ctnetlink_change_timeout(ct, cda);
		if (err < 0)
			return err;
	}

1158
	if (cda[CTA_STATUS]) {
1159 1160 1161 1162 1163
		err = ctnetlink_change_status(ct, cda);
		if (err < 0)
			return err;
	}

1164
	if (cda[CTA_PROTOINFO]) {
1165 1166 1167 1168 1169
		err = ctnetlink_change_protoinfo(ct, cda);
		if (err < 0)
			return err;
	}

1170
#if defined(CONFIG_NF_CONNTRACK_MARK)
1171
	if (cda[CTA_MARK])
1172
		ct->mark = ntohl(nla_get_be32(cda[CTA_MARK]));
1173 1174
#endif

1175 1176 1177 1178 1179 1180 1181 1182
#ifdef CONFIG_NF_NAT_NEEDED
	if (cda[CTA_NAT_SEQ_ADJ_ORIG] || cda[CTA_NAT_SEQ_ADJ_REPLY]) {
		err = ctnetlink_change_nat_seq_adj(ct, cda);
		if (err < 0)
			return err;
	}
#endif

1183 1184 1185
	return 0;
}

1186
static struct nf_conn *
1187 1188
ctnetlink_create_conntrack(struct net *net,
			   const struct nlattr * const cda[],
1189
			   struct nf_conntrack_tuple *otuple,
1190
			   struct nf_conntrack_tuple *rtuple,
1191
			   u8 u3)
1192 1193 1194
{
	struct nf_conn *ct;
	int err = -EINVAL;
1195
	struct nf_conntrack_helper *helper;
1196

1197
	ct = nf_conntrack_alloc(net, otuple, rtuple, GFP_ATOMIC);
1198
	if (IS_ERR(ct))
1199
		return ERR_PTR(-ENOMEM);
1200

1201
	if (!cda[CTA_TIMEOUT])
1202
		goto err1;
1203
	ct->timeout.expires = ntohl(nla_get_be32(cda[CTA_TIMEOUT]));
1204 1205 1206 1207

	ct->timeout.expires = jiffies + ct->timeout.expires * HZ;
	ct->status |= IPS_CONFIRMED;

1208
	rcu_read_lock();
1209
 	if (cda[CTA_HELP]) {
1210
		char *helpname = NULL;
1211 1212
 
 		err = ctnetlink_parse_help(cda[CTA_HELP], &helpname);
1213 1214
 		if (err < 0)
			goto err2;
1215

1216 1217
		helper = __nf_conntrack_helper_find(helpname, nf_ct_l3num(ct),
						    nf_ct_protonum(ct));
1218 1219 1220 1221 1222
		if (helper == NULL) {
			rcu_read_unlock();
#ifdef CONFIG_MODULES
			if (request_module("nfct-helper-%s", helpname) < 0) {
				err = -EOPNOTSUPP;
1223
				goto err1;
1224 1225 1226
			}

			rcu_read_lock();
1227 1228 1229
			helper = __nf_conntrack_helper_find(helpname,
							    nf_ct_l3num(ct),
							    nf_ct_protonum(ct));
1230 1231
			if (helper) {
				err = -EAGAIN;
1232
				goto err2;
1233 1234 1235 1236
			}
			rcu_read_unlock();
#endif
			err = -EOPNOTSUPP;
1237
			goto err1;
1238 1239 1240 1241 1242 1243
		} else {
			struct nf_conn_help *help;

			help = nf_ct_helper_ext_add(ct, GFP_ATOMIC);
			if (help == NULL) {
				err = -ENOMEM;
1244
				goto err2;
1245 1246 1247 1248 1249 1250 1251
			}

			/* not in hash table yet so not strictly necessary */
			rcu_assign_pointer(help->helper, helper);
		}
	} else {
		/* try an implicit helper assignation */
1252
		err = __nf_ct_try_assign_helper(ct, NULL, GFP_ATOMIC);
1253 1254
		if (err < 0)
			goto err2;
1255 1256
	}

1257
	if (cda[CTA_STATUS]) {
1258
		err = ctnetlink_change_status(ct, cda);
1259 1260
		if (err < 0)
			goto err2;
1261 1262 1263 1264
	}

	if (cda[CTA_NAT_SRC] || cda[CTA_NAT_DST]) {
		err = ctnetlink_change_nat(ct, cda);
1265 1266
		if (err < 0)
			goto err2;
1267
	}
1268

1269 1270 1271
#ifdef CONFIG_NF_NAT_NEEDED
	if (cda[CTA_NAT_SEQ_ADJ_ORIG] || cda[CTA_NAT_SEQ_ADJ_REPLY]) {
		err = ctnetlink_change_nat_seq_adj(ct, cda);
1272 1273
		if (err < 0)
			goto err2;
1274 1275 1276
	}
#endif

1277
	if (cda[CTA_PROTOINFO]) {
1278
		err = ctnetlink_change_protoinfo(ct, cda);
1279 1280
		if (err < 0)
			goto err2;
1281 1282
	}

1283
	nf_ct_acct_ext_add(ct, GFP_ATOMIC);
1284
	nf_ct_ecache_ext_add(ct, 0, 0, GFP_ATOMIC);
1285

1286
#if defined(CONFIG_NF_CONNTRACK_MARK)
1287
	if (cda[CTA_MARK])
1288
		ct->mark = ntohl(nla_get_be32(cda[CTA_MARK]));
1289 1290
#endif

1291
	/* setup master conntrack: this is a confirmed expectation */
1292 1293 1294 1295 1296 1297 1298
	if (cda[CTA_TUPLE_MASTER]) {
		struct nf_conntrack_tuple master;
		struct nf_conntrack_tuple_hash *master_h;
		struct nf_conn *master_ct;

		err = ctnetlink_parse_tuple(cda, &master, CTA_TUPLE_MASTER, u3);
		if (err < 0)
1299
			goto err2;
1300

1301
		master_h = nf_conntrack_find_get(net, &master);
1302 1303
		if (master_h == NULL) {
			err = -ENOENT;
1304
			goto err2;
1305 1306
		}
		master_ct = nf_ct_tuplehash_to_ctrack(master_h);
1307
		__set_bit(IPS_EXPECTED_BIT, &ct->status);
1308
		ct->master = master_ct;
1309
	}
1310

1311 1312
	add_timer(&ct->timeout);
	nf_conntrack_hash_insert(ct);
1313
	rcu_read_unlock();
1314

1315
	return ct;
1316

1317 1318 1319
err2:
	rcu_read_unlock();
err1:
1320
	nf_conntrack_free(ct);
1321
	return ERR_PTR(err);
1322 1323
}

1324 1325
static int
ctnetlink_new_conntrack(struct sock *ctnl, struct sk_buff *skb,
1326 1327
			const struct nlmsghdr *nlh,
			const struct nlattr * const cda[])
1328
{
1329
	struct net *net = sock_net(ctnl);
1330 1331
	struct nf_conntrack_tuple otuple, rtuple;
	struct nf_conntrack_tuple_hash *h = NULL;
1332
	struct nfgenmsg *nfmsg = nlmsg_data(nlh);
1333 1334 1335
	u_int8_t u3 = nfmsg->nfgen_family;
	int err = 0;

1336
	if (cda[CTA_TUPLE_ORIG]) {
1337 1338 1339 1340 1341
		err = ctnetlink_parse_tuple(cda, &otuple, CTA_TUPLE_ORIG, u3);
		if (err < 0)
			return err;
	}

1342
	if (cda[CTA_TUPLE_REPLY]) {
1343 1344 1345 1346 1347
		err = ctnetlink_parse_tuple(cda, &rtuple, CTA_TUPLE_REPLY, u3);
		if (err < 0)
			return err;
	}

1348
	spin_lock_bh(&nf_conntrack_lock);
1349
	if (cda[CTA_TUPLE_ORIG])
1350
		h = __nf_conntrack_find(net, &otuple);
1351
	else if (cda[CTA_TUPLE_REPLY])
1352
		h = __nf_conntrack_find(net, &rtuple);
1353 1354 1355

	if (h == NULL) {
		err = -ENOENT;
1356 1357
		if (nlh->nlmsg_flags & NLM_F_CREATE) {
			struct nf_conn *ct;
1358
			enum ip_conntrack_events events;
1359

1360
			ct = ctnetlink_create_conntrack(net, cda, &otuple,
1361 1362 1363
							&rtuple, u3);
			if (IS_ERR(ct)) {
				err = PTR_ERR(ct);
1364 1365
				goto out_unlock;
			}
1366 1367 1368
			err = 0;
			nf_conntrack_get(&ct->ct_general);
			spin_unlock_bh(&nf_conntrack_lock);
1369 1370 1371 1372 1373
			if (test_bit(IPS_EXPECTED_BIT, &ct->status))
				events = IPCT_RELATED;
			else
				events = IPCT_NEW;

1374 1375
			nf_conntrack_eventmask_report((1 << IPCT_REPLY) |
						      (1 << IPCT_ASSURED) |
1376 1377 1378 1379 1380 1381
						      (1 << IPCT_HELPER) |
						      (1 << IPCT_PROTOINFO) |
						      (1 << IPCT_NATSEQADJ) |
						      (1 << IPCT_MARK) | events,
						      ct, NETLINK_CB(skb).pid,
						      nlmsg_report(nlh));
1382 1383 1384
			nf_ct_put(ct);
		} else
			spin_unlock_bh(&nf_conntrack_lock);
1385

1386 1387 1388 1389 1390 1391 1392
		return err;
	}
	/* implicit 'else' */

	/* We manipulate the conntrack inside the global conntrack table lock,
	 * so there's no need to increase the refcount */
	err = -EEXIST;
1393
	if (!(nlh->nlmsg_flags & NLM_F_EXCL)) {
1394 1395 1396 1397 1398 1399
		struct nf_conn *ct = nf_ct_tuplehash_to_ctrack(h);

		err = ctnetlink_change_conntrack(ct, cda);
		if (err == 0) {
			nf_conntrack_get(&ct->ct_general);
			spin_unlock_bh(&nf_conntrack_lock);
1400 1401
			nf_conntrack_eventmask_report((1 << IPCT_REPLY) |
						      (1 << IPCT_ASSURED) |
1402 1403 1404 1405 1406 1407
						      (1 << IPCT_HELPER) |
						      (1 << IPCT_PROTOINFO) |
						      (1 << IPCT_NATSEQADJ) |
						      (1 << IPCT_MARK),
						      ct, NETLINK_CB(skb).pid,
						      nlmsg_report(nlh));
1408 1409 1410 1411 1412
			nf_ct_put(ct);
		} else
			spin_unlock_bh(&nf_conntrack_lock);

		return err;
1413
	}
1414 1415

out_unlock:
1416
	spin_unlock_bh(&nf_conntrack_lock);
1417 1418 1419
	return err;
}

1420 1421 1422
/***********************************************************************
 * EXPECT
 ***********************************************************************/
1423 1424 1425 1426 1427 1428

static inline int
ctnetlink_exp_dump_tuple(struct sk_buff *skb,
			 const struct nf_conntrack_tuple *tuple,
			 enum ctattr_expect type)
{
1429
	struct nlattr *nest_parms;
1430

1431 1432 1433
	nest_parms = nla_nest_start(skb, type | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
1434
	if (ctnetlink_dump_tuples(skb, tuple) < 0)
1435 1436
		goto nla_put_failure;
	nla_nest_end(skb, nest_parms);
1437 1438 1439

	return 0;

1440
nla_put_failure:
1441
	return -1;
1442
}
1443

1444 1445 1446
static inline int
ctnetlink_exp_dump_mask(struct sk_buff *skb,
			const struct nf_conntrack_tuple *tuple,
1447
			const struct nf_conntrack_tuple_mask *mask)
1448 1449 1450
{
	int ret;
	struct nf_conntrack_l3proto *l3proto;
1451
	struct nf_conntrack_l4proto *l4proto;
1452
	struct nf_conntrack_tuple m;
1453
	struct nlattr *nest_parms;
1454 1455 1456 1457 1458

	memset(&m, 0xFF, sizeof(m));
	m.src.u.all = mask->src.u.all;
	memcpy(&m.src.u3, &mask->src.u3, sizeof(m.src.u3));

1459 1460 1461
	nest_parms = nla_nest_start(skb, CTA_EXPECT_MASK | NLA_F_NESTED);
	if (!nest_parms)
		goto nla_put_failure;
1462

1463
	l3proto = __nf_ct_l3proto_find(tuple->src.l3num);
1464
	ret = ctnetlink_dump_tuples_ip(skb, &m, l3proto);
1465 1466

	if (unlikely(ret < 0))
1467
		goto nla_put_failure;
1468

1469
	l4proto = __nf_ct_l4proto_find(tuple->src.l3num, tuple->dst.protonum);
1470
	ret = ctnetlink_dump_tuples_proto(skb, &m, l4proto);
1471
	if (unlikely(ret < 0))
1472
		goto nla_put_failure;
1473

1474
	nla_nest_end(skb, nest_parms);
1475 1476 1477

	return 0;

1478
nla_put_failure:
1479 1480 1481
	return -1;
}

1482
static int
1483
ctnetlink_exp_dump_expect(struct sk_buff *skb,
1484
			  const struct nf_conntrack_expect *exp)
1485 1486
{
	struct nf_conn *master = exp->master;
1487 1488 1489 1490
	long timeout = (exp->timeout.expires - jiffies) / HZ;

	if (timeout < 0)
		timeout = 0;
1491 1492

	if (ctnetlink_exp_dump_tuple(skb, &exp->tuple, CTA_EXPECT_TUPLE) < 0)
1493
		goto nla_put_failure;
1494
	if (ctnetlink_exp_dump_mask(skb, &exp->tuple, &exp->mask) < 0)
1495
		goto nla_put_failure;
1496 1497 1498
	if (ctnetlink_exp_dump_tuple(skb,
				 &master->tuplehash[IP_CT_DIR_ORIGINAL].tuple,
				 CTA_EXPECT_MASTER) < 0)
1499
		goto nla_put_failure;
1500

1501
	NLA_PUT_BE32(skb, CTA_EXPECT_TIMEOUT, htonl(timeout));
1502
	NLA_PUT_BE32(skb, CTA_EXPECT_ID, htonl((unsigned long)exp));
1503 1504

	return 0;
1505

1506
nla_put_failure:
1507 1508 1509 1510 1511
	return -1;
}

static int
ctnetlink_exp_fill_info(struct sk_buff *skb, u32 pid, u32 seq,
1512
			int event, const struct nf_conntrack_expect *exp)
1513 1514 1515
{
	struct nlmsghdr *nlh;
	struct nfgenmsg *nfmsg;
1516
	unsigned int flags = pid ? NLM_F_MULTI : 0;
1517 1518

	event |= NFNL_SUBSYS_CTNETLINK_EXP << 8;
1519 1520 1521
	nlh = nlmsg_put(skb, pid, seq, event, sizeof(*nfmsg), flags);
	if (nlh == NULL)
		goto nlmsg_failure;
1522

1523
	nfmsg = nlmsg_data(nlh);
1524 1525 1526 1527 1528
	nfmsg->nfgen_family = exp->tuple.src.l3num;
	nfmsg->version	    = NFNETLINK_V0;
	nfmsg->res_id	    = 0;

	if (ctnetlink_exp_dump_expect(skb, exp) < 0)
1529
		goto nla_put_failure;
1530

1531
	nlmsg_end(skb, nlh);
1532 1533 1534
	return skb->len;

nlmsg_failure:
1535
nla_put_failure:
1536
	nlmsg_cancel(skb, nlh);
1537 1538 1539 1540
	return -1;
}

#ifdef CONFIG_NF_CONNTRACK_EVENTS
1541 1542
static int
ctnetlink_expect_event(unsigned int events, struct nf_exp_event *item)
1543
{
1544 1545
	struct nf_conntrack_expect *exp = item->exp;
	struct net *net = nf_ct_exp_net(exp);
1546 1547 1548 1549 1550 1551
	struct nlmsghdr *nlh;
	struct nfgenmsg *nfmsg;
	struct sk_buff *skb;
	unsigned int type;
	int flags = 0;

1552
	if (events & (1 << IPEXP_NEW)) {
1553 1554 1555
		type = IPCTNL_MSG_EXP_NEW;
		flags = NLM_F_CREATE|NLM_F_EXCL;
	} else
1556
		return 0;
1557

1558
	if (!item->report &&
1559
	    !nfnetlink_has_listeners(net, NFNLGRP_CONNTRACK_EXP_NEW))
1560
		return 0;
1561

1562 1563
	skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC);
	if (skb == NULL)
1564
		goto errout;
1565

1566
	type |= NFNL_SUBSYS_CTNETLINK_EXP << 8;
1567 1568 1569
	nlh = nlmsg_put(skb, item->pid, 0, type, sizeof(*nfmsg), flags);
	if (nlh == NULL)
		goto nlmsg_failure;
1570

1571
	nfmsg = nlmsg_data(nlh);
1572 1573 1574 1575
	nfmsg->nfgen_family = exp->tuple.src.l3num;
	nfmsg->version	    = NFNETLINK_V0;
	nfmsg->res_id	    = 0;

1576
	rcu_read_lock();
1577
	if (ctnetlink_exp_dump_expect(skb, exp) < 0)
1578
		goto nla_put_failure;
1579
	rcu_read_unlock();
1580

1581
	nlmsg_end(skb, nlh);
1582
	nfnetlink_send(skb, net, item->pid, NFNLGRP_CONNTRACK_EXP_NEW,
1583 1584
		       item->report, GFP_ATOMIC);
	return 0;
1585

1586
nla_put_failure:
1587
	rcu_read_unlock();
1588
	nlmsg_cancel(skb, nlh);
1589
nlmsg_failure:
1590
	kfree_skb(skb);
1591
errout:
1592
	nfnetlink_set_err(net, 0, 0, -ENOBUFS);
1593
	return 0;
1594 1595
}
#endif
1596 1597
static int ctnetlink_exp_done(struct netlink_callback *cb)
{
1598 1599
	if (cb->args[1])
		nf_ct_expect_put((struct nf_conntrack_expect *)cb->args[1]);
1600 1601
	return 0;
}
1602 1603 1604 1605

static int
ctnetlink_exp_dump_table(struct sk_buff *skb, struct netlink_callback *cb)
{
1606
	struct net *net = sock_net(skb->sk);
1607
	struct nf_conntrack_expect *exp, *last;
1608
	struct nfgenmsg *nfmsg = nlmsg_data(cb->nlh);
1609
	struct hlist_node *n;
1610
	u_int8_t l3proto = nfmsg->nfgen_family;
1611

1612
	rcu_read_lock();
1613 1614
	last = (struct nf_conntrack_expect *)cb->args[1];
	for (; cb->args[0] < nf_ct_expect_hsize; cb->args[0]++) {
1615
restart:
1616
		hlist_for_each_entry(exp, n, &net->ct.expect_hash[cb->args[0]],
1617 1618
				     hnode) {
			if (l3proto && exp->tuple.src.l3num != l3proto)
1619
				continue;
1620 1621 1622 1623 1624
			if (cb->args[1]) {
				if (exp != last)
					continue;
				cb->args[1] = 0;
			}
1625 1626
			if (ctnetlink_exp_fill_info(skb,
						    NETLINK_CB(cb->skb).pid,
1627 1628
						    cb->nlh->nlmsg_seq,
						    IPCTNL_MSG_EXP_NEW,
1629
						    exp) < 0) {
1630 1631
				if (!atomic_inc_not_zero(&exp->use))
					continue;
1632 1633 1634
				cb->args[1] = (unsigned long)exp;
				goto out;
			}
1635
		}
1636 1637 1638
		if (cb->args[1]) {
			cb->args[1] = 0;
			goto restart;
1639 1640
		}
	}
1641
out:
1642
	rcu_read_unlock();
1643 1644
	if (last)
		nf_ct_expect_put(last);
1645 1646 1647 1648

	return skb->len;
}

1649 1650 1651
static const struct nla_policy exp_nla_policy[CTA_EXPECT_MAX+1] = {
	[CTA_EXPECT_TIMEOUT]	= { .type = NLA_U32 },
	[CTA_EXPECT_ID]		= { .type = NLA_U32 },
1652 1653 1654
};

static int
1655
ctnetlink_get_expect(struct sock *ctnl, struct sk_buff *skb,
1656 1657
		     const struct nlmsghdr *nlh,
		     const struct nlattr * const cda[])
1658
{
1659
	struct net *net = sock_net(ctnl);
1660 1661 1662
	struct nf_conntrack_tuple tuple;
	struct nf_conntrack_expect *exp;
	struct sk_buff *skb2;
1663
	struct nfgenmsg *nfmsg = nlmsg_data(nlh);
1664 1665 1666 1667
	u_int8_t u3 = nfmsg->nfgen_family;
	int err = 0;

	if (nlh->nlmsg_flags & NLM_F_DUMP) {
1668 1669
		return netlink_dump_start(ctnl, skb, nlh,
					  ctnetlink_exp_dump_table,
1670
					  ctnetlink_exp_done);
1671 1672
	}

1673
	if (cda[CTA_EXPECT_MASTER])
1674 1675 1676 1677 1678 1679 1680
		err = ctnetlink_parse_tuple(cda, &tuple, CTA_EXPECT_MASTER, u3);
	else
		return -EINVAL;

	if (err < 0)
		return err;

1681
	exp = nf_ct_expect_find_get(net, &tuple);
1682 1683 1684
	if (!exp)
		return -ENOENT;

1685
	if (cda[CTA_EXPECT_ID]) {
1686
		__be32 id = nla_get_be32(cda[CTA_EXPECT_ID]);
1687
		if (ntohl(id) != (u32)(unsigned long)exp) {
1688
			nf_ct_expect_put(exp);
1689 1690
			return -ENOENT;
		}
1691
	}
1692 1693

	err = -ENOMEM;
1694 1695
	skb2 = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
	if (skb2 == NULL)
1696
		goto out;
1697

1698
	rcu_read_lock();
1699
	err = ctnetlink_exp_fill_info(skb2, NETLINK_CB(skb).pid,
1700
				      nlh->nlmsg_seq, IPCTNL_MSG_EXP_NEW, exp);
1701
	rcu_read_unlock();
1702 1703 1704
	if (err <= 0)
		goto free;

1705
	nf_ct_expect_put(exp);
1706 1707 1708 1709 1710 1711

	return netlink_unicast(ctnl, skb2, NETLINK_CB(skb).pid, MSG_DONTWAIT);

free:
	kfree_skb(skb2);
out:
1712
	nf_ct_expect_put(exp);
1713 1714 1715 1716
	return err;
}

static int
1717
ctnetlink_del_expect(struct sock *ctnl, struct sk_buff *skb,
1718 1719
		     const struct nlmsghdr *nlh,
		     const struct nlattr * const cda[])
1720
{
1721
	struct net *net = sock_net(ctnl);
1722
	struct nf_conntrack_expect *exp;
1723
	struct nf_conntrack_tuple tuple;
1724
	struct nfgenmsg *nfmsg = nlmsg_data(nlh);
1725
	struct hlist_node *n, *next;
1726
	u_int8_t u3 = nfmsg->nfgen_family;
1727
	unsigned int i;
1728 1729
	int err;

1730
	if (cda[CTA_EXPECT_TUPLE]) {
1731 1732 1733 1734 1735 1736
		/* delete a single expect by tuple */
		err = ctnetlink_parse_tuple(cda, &tuple, CTA_EXPECT_TUPLE, u3);
		if (err < 0)
			return err;

		/* bump usage count to 2 */
1737
		exp = nf_ct_expect_find_get(net, &tuple);
1738 1739 1740
		if (!exp)
			return -ENOENT;

1741
		if (cda[CTA_EXPECT_ID]) {
1742
			__be32 id = nla_get_be32(cda[CTA_EXPECT_ID]);
1743
			if (ntohl(id) != (u32)(unsigned long)exp) {
1744
				nf_ct_expect_put(exp);
1745 1746 1747 1748 1749
				return -ENOENT;
			}
		}

		/* after list removal, usage count == 1 */
1750
		nf_ct_unexpect_related(exp);
1751
		/* have to put what we 'get' above.
1752
		 * after this line usage count == 0 */
1753
		nf_ct_expect_put(exp);
1754 1755
	} else if (cda[CTA_EXPECT_HELP_NAME]) {
		char *name = nla_data(cda[CTA_EXPECT_HELP_NAME]);
1756
		struct nf_conn_help *m_help;
1757 1758

		/* delete all expectations for this helper */
1759
		spin_lock_bh(&nf_conntrack_lock);
1760 1761
		for (i = 0; i < nf_ct_expect_hsize; i++) {
			hlist_for_each_entry_safe(exp, n, next,
1762
						  &net->ct.expect_hash[i],
1763 1764
						  hnode) {
				m_help = nfct_help(exp->master);
1765 1766
				if (!strcmp(m_help->helper->name, name) &&
				    del_timer(&exp->timeout)) {
1767 1768 1769
					nf_ct_unlink_expect(exp);
					nf_ct_expect_put(exp);
				}
1770 1771
			}
		}
1772
		spin_unlock_bh(&nf_conntrack_lock);
1773 1774
	} else {
		/* This basically means we have to flush everything*/
1775
		spin_lock_bh(&nf_conntrack_lock);
1776 1777
		for (i = 0; i < nf_ct_expect_hsize; i++) {
			hlist_for_each_entry_safe(exp, n, next,
1778
						  &net->ct.expect_hash[i],
1779 1780 1781 1782 1783
						  hnode) {
				if (del_timer(&exp->timeout)) {
					nf_ct_unlink_expect(exp);
					nf_ct_expect_put(exp);
				}
1784 1785
			}
		}
1786
		spin_unlock_bh(&nf_conntrack_lock);
1787 1788 1789 1790 1791
	}

	return 0;
}
static int
1792 1793
ctnetlink_change_expect(struct nf_conntrack_expect *x,
			const struct nlattr * const cda[])
1794 1795 1796 1797 1798
{
	return -EOPNOTSUPP;
}

static int
1799 1800
ctnetlink_create_expect(struct net *net, const struct nlattr * const cda[],
			u_int8_t u3,
1801
			u32 pid, int report)
1802 1803 1804 1805 1806
{
	struct nf_conntrack_tuple tuple, mask, master_tuple;
	struct nf_conntrack_tuple_hash *h = NULL;
	struct nf_conntrack_expect *exp;
	struct nf_conn *ct;
1807
	struct nf_conn_help *help;
1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821
	int err = 0;

	/* caller guarantees that those three CTA_EXPECT_* exist */
	err = ctnetlink_parse_tuple(cda, &tuple, CTA_EXPECT_TUPLE, u3);
	if (err < 0)
		return err;
	err = ctnetlink_parse_tuple(cda, &mask, CTA_EXPECT_MASK, u3);
	if (err < 0)
		return err;
	err = ctnetlink_parse_tuple(cda, &master_tuple, CTA_EXPECT_MASTER, u3);
	if (err < 0)
		return err;

	/* Look for master conntrack of this expectation */
1822
	h = nf_conntrack_find_get(net, &master_tuple);
1823 1824 1825
	if (!h)
		return -ENOENT;
	ct = nf_ct_tuplehash_to_ctrack(h);
1826
	help = nfct_help(ct);
1827

1828
	if (!help || !help->helper) {
1829
		/* such conntrack hasn't got any helper, abort */
1830
		err = -EOPNOTSUPP;
1831 1832 1833
		goto out;
	}

1834
	exp = nf_ct_expect_alloc(ct);
1835 1836 1837 1838
	if (!exp) {
		err = -ENOMEM;
		goto out;
	}
1839

1840
	exp->class = 0;
1841 1842 1843
	exp->expectfn = NULL;
	exp->flags = 0;
	exp->master = ct;
1844
	exp->helper = NULL;
1845
	memcpy(&exp->tuple, &tuple, sizeof(struct nf_conntrack_tuple));
1846 1847
	memcpy(&exp->mask.src.u3, &mask.src.u3, sizeof(exp->mask.src.u3));
	exp->mask.src.u.all = mask.src.u.all;
1848

1849
	err = nf_ct_expect_related_report(exp, pid, report);
1850
	nf_ct_expect_put(exp);
1851

1852
out:
1853 1854 1855 1856 1857 1858
	nf_ct_put(nf_ct_tuplehash_to_ctrack(h));
	return err;
}

static int
ctnetlink_new_expect(struct sock *ctnl, struct sk_buff *skb,
1859 1860
		     const struct nlmsghdr *nlh,
		     const struct nlattr * const cda[])
1861
{
1862
	struct net *net = sock_net(ctnl);
1863 1864
	struct nf_conntrack_tuple tuple;
	struct nf_conntrack_expect *exp;
1865
	struct nfgenmsg *nfmsg = nlmsg_data(nlh);
1866 1867 1868
	u_int8_t u3 = nfmsg->nfgen_family;
	int err = 0;

1869 1870 1871
	if (!cda[CTA_EXPECT_TUPLE]
	    || !cda[CTA_EXPECT_MASK]
	    || !cda[CTA_EXPECT_MASTER])
1872 1873 1874 1875 1876 1877
		return -EINVAL;

	err = ctnetlink_parse_tuple(cda, &tuple, CTA_EXPECT_TUPLE, u3);
	if (err < 0)
		return err;

1878
	spin_lock_bh(&nf_conntrack_lock);
1879
	exp = __nf_ct_expect_find(net, &tuple);
1880 1881

	if (!exp) {
1882
		spin_unlock_bh(&nf_conntrack_lock);
1883
		err = -ENOENT;
1884
		if (nlh->nlmsg_flags & NLM_F_CREATE) {
1885
			err = ctnetlink_create_expect(net, cda,
1886 1887 1888 1889
						      u3,
						      NETLINK_CB(skb).pid,
						      nlmsg_report(nlh));
		}
1890 1891 1892 1893 1894 1895
		return err;
	}

	err = -EEXIST;
	if (!(nlh->nlmsg_flags & NLM_F_EXCL))
		err = ctnetlink_change_expect(exp, cda);
1896
	spin_unlock_bh(&nf_conntrack_lock);
1897 1898 1899 1900 1901

	return err;
}

#ifdef CONFIG_NF_CONNTRACK_EVENTS
1902 1903
static struct nf_ct_event_notifier ctnl_notifier = {
	.fcn = ctnetlink_conntrack_event,
1904 1905
};

1906 1907
static struct nf_exp_event_notifier ctnl_notifier_exp = {
	.fcn = ctnetlink_expect_event,
1908 1909 1910
};
#endif

1911
static const struct nfnl_callback ctnl_cb[IPCTNL_MSG_MAX] = {
1912
	[IPCTNL_MSG_CT_NEW]		= { .call = ctnetlink_new_conntrack,
1913 1914
					    .attr_count = CTA_MAX,
					    .policy = ct_nla_policy },
1915
	[IPCTNL_MSG_CT_GET] 		= { .call = ctnetlink_get_conntrack,
1916 1917
					    .attr_count = CTA_MAX,
					    .policy = ct_nla_policy },
1918
	[IPCTNL_MSG_CT_DELETE]  	= { .call = ctnetlink_del_conntrack,
1919 1920
					    .attr_count = CTA_MAX,
					    .policy = ct_nla_policy },
1921
	[IPCTNL_MSG_CT_GET_CTRZERO] 	= { .call = ctnetlink_get_conntrack,
1922 1923
					    .attr_count = CTA_MAX,
					    .policy = ct_nla_policy },
1924 1925
};

1926
static const struct nfnl_callback ctnl_exp_cb[IPCTNL_MSG_EXP_MAX] = {
1927
	[IPCTNL_MSG_EXP_GET]		= { .call = ctnetlink_get_expect,
1928 1929
					    .attr_count = CTA_EXPECT_MAX,
					    .policy = exp_nla_policy },
1930
	[IPCTNL_MSG_EXP_NEW]		= { .call = ctnetlink_new_expect,
1931 1932
					    .attr_count = CTA_EXPECT_MAX,
					    .policy = exp_nla_policy },
1933
	[IPCTNL_MSG_EXP_DELETE]		= { .call = ctnetlink_del_expect,
1934 1935
					    .attr_count = CTA_EXPECT_MAX,
					    .policy = exp_nla_policy },
1936 1937
};

1938
static const struct nfnetlink_subsystem ctnl_subsys = {
1939 1940 1941 1942 1943 1944
	.name				= "conntrack",
	.subsys_id			= NFNL_SUBSYS_CTNETLINK,
	.cb_count			= IPCTNL_MSG_MAX,
	.cb				= ctnl_cb,
};

1945
static const struct nfnetlink_subsystem ctnl_exp_subsys = {
1946 1947 1948 1949 1950 1951
	.name				= "conntrack_expect",
	.subsys_id			= NFNL_SUBSYS_CTNETLINK_EXP,
	.cb_count			= IPCTNL_MSG_EXP_MAX,
	.cb				= ctnl_exp_cb,
};

1952
MODULE_ALIAS("ip_conntrack_netlink");
1953
MODULE_ALIAS_NFNL_SUBSYS(NFNL_SUBSYS_CTNETLINK);
1954
MODULE_ALIAS_NFNL_SUBSYS(NFNL_SUBSYS_CTNETLINK_EXP);
1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979

static int __init ctnetlink_init(void)
{
	int ret;

	printk("ctnetlink v%s: registering with nfnetlink.\n", version);
	ret = nfnetlink_subsys_register(&ctnl_subsys);
	if (ret < 0) {
		printk("ctnetlink_init: cannot register with nfnetlink.\n");
		goto err_out;
	}

	ret = nfnetlink_subsys_register(&ctnl_exp_subsys);
	if (ret < 0) {
		printk("ctnetlink_init: cannot register exp with nfnetlink.\n");
		goto err_unreg_subsys;
	}

#ifdef CONFIG_NF_CONNTRACK_EVENTS
	ret = nf_conntrack_register_notifier(&ctnl_notifier);
	if (ret < 0) {
		printk("ctnetlink_init: cannot register notifier.\n");
		goto err_unreg_exp_subsys;
	}

1980
	ret = nf_ct_expect_register_notifier(&ctnl_notifier_exp);
1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005
	if (ret < 0) {
		printk("ctnetlink_init: cannot expect register notifier.\n");
		goto err_unreg_notifier;
	}
#endif

	return 0;

#ifdef CONFIG_NF_CONNTRACK_EVENTS
err_unreg_notifier:
	nf_conntrack_unregister_notifier(&ctnl_notifier);
err_unreg_exp_subsys:
	nfnetlink_subsys_unregister(&ctnl_exp_subsys);
#endif
err_unreg_subsys:
	nfnetlink_subsys_unregister(&ctnl_subsys);
err_out:
	return ret;
}

static void __exit ctnetlink_exit(void)
{
	printk("ctnetlink: unregistering from nfnetlink.\n");

#ifdef CONFIG_NF_CONNTRACK_EVENTS
2006
	nf_ct_expect_unregister_notifier(&ctnl_notifier_exp);
2007 2008 2009 2010 2011 2012 2013 2014 2015 2016
	nf_conntrack_unregister_notifier(&ctnl_notifier);
#endif

	nfnetlink_subsys_unregister(&ctnl_exp_subsys);
	nfnetlink_subsys_unregister(&ctnl_subsys);
	return;
}

module_init(ctnetlink_init);
module_exit(ctnetlink_exit);