l2tp_core.c 51.4 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
/*
 * L2TP core.
 *
 * Copyright (c) 2008,2009,2010 Katalix Systems Ltd
 *
 * This file contains some code of the original L2TPv2 pppol2tp
 * driver, which has the following copyright:
 *
 * Authors:	Martijn van Oosterhout <kleptog@svana.org>
 *		James Chapman (jchapman@katalix.com)
 * Contributors:
 *		Michal Ostrowski <mostrows@speakeasy.net>
 *		Arnaldo Carvalho de Melo <acme@xconectiva.com.br>
 *		David S. Miller (davem@redhat.com)
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */

21 22
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

23 24 25
#include <linux/module.h>
#include <linux/string.h>
#include <linux/list.h>
J
James Chapman 已提交
26
#include <linux/rculist.h>
27 28 29 30 31 32 33 34 35 36 37 38 39 40 41
#include <linux/uaccess.h>

#include <linux/kernel.h>
#include <linux/spinlock.h>
#include <linux/kthread.h>
#include <linux/sched.h>
#include <linux/slab.h>
#include <linux/errno.h>
#include <linux/jiffies.h>

#include <linux/netdevice.h>
#include <linux/net.h>
#include <linux/inetdevice.h>
#include <linux/skbuff.h>
#include <linux/init.h>
42
#include <linux/in.h>
43 44
#include <linux/ip.h>
#include <linux/udp.h>
45
#include <linux/l2tp.h>
46 47 48 49 50 51 52 53 54
#include <linux/hash.h>
#include <linux/sort.h>
#include <linux/file.h>
#include <linux/nsproxy.h>
#include <net/net_namespace.h>
#include <net/netns/generic.h>
#include <net/dst.h>
#include <net/ip.h>
#include <net/udp.h>
T
Tom Herbert 已提交
55
#include <net/udp_tunnel.h>
56
#include <net/inet_common.h>
57
#include <net/xfrm.h>
58
#include <net/protocol.h>
59 60 61
#include <net/inet6_connection_sock.h>
#include <net/inet_ecn.h>
#include <net/ip6_route.h>
62
#include <net/ip6_checksum.h>
63 64

#include <asm/byteorder.h>
A
Arun Sharma 已提交
65
#include <linux/atomic.h>
66 67 68 69 70 71 72 73 74 75 76 77 78 79

#include "l2tp_core.h"

#define L2TP_DRV_VERSION	"V2.0"

/* L2TP header constants */
#define L2TP_HDRFLAG_T	   0x8000
#define L2TP_HDRFLAG_L	   0x4000
#define L2TP_HDRFLAG_S	   0x0800
#define L2TP_HDRFLAG_O	   0x0200
#define L2TP_HDRFLAG_P	   0x0100

#define L2TP_HDR_VER_MASK  0x000F
#define L2TP_HDR_VER_2	   0x0002
80
#define L2TP_HDR_VER_3	   0x0003
81 82 83 84 85 86 87 88 89 90 91 92 93 94

/* L2TPv3 default L2-specific sublayer */
#define L2TP_SLFLAG_S	   0x40000000
#define L2TP_SL_SEQ_MASK   0x00ffffff

#define L2TP_HDR_SIZE_SEQ		10
#define L2TP_HDR_SIZE_NOSEQ		6

/* Default trace flags */
#define L2TP_DEFAULT_DEBUG_FLAGS	0

/* Private data stored for received packets in the skb.
 */
struct l2tp_skb_cb {
95
	u32			ns;
96 97 98 99 100 101 102 103 104
	u16			has_seq;
	u16			length;
	unsigned long		expires;
};

#define L2TP_SKB_CB(skb)	((struct l2tp_skb_cb *) &skb->cb[sizeof(struct inet_skb_parm)])

static atomic_t l2tp_tunnel_count;
static atomic_t l2tp_session_count;
105
static struct workqueue_struct *l2tp_wq;
106 107 108 109 110

/* per-net private data for this module */
static unsigned int l2tp_net_id;
struct l2tp_net {
	struct list_head l2tp_tunnel_list;
J
James Chapman 已提交
111
	spinlock_t l2tp_tunnel_list_lock;
112
	struct hlist_head l2tp_session_hlist[L2TP_HASH_SIZE_2];
J
James Chapman 已提交
113
	spinlock_t l2tp_session_hlist_lock;
114 115
};

116

117 118 119 120 121
static inline struct l2tp_tunnel *l2tp_tunnel(struct sock *sk)
{
	return sk->sk_user_data;
}

122
static inline struct l2tp_net *l2tp_pernet(const struct net *net)
123 124 125 126 127 128
{
	BUG_ON(!net);

	return net_generic(net, l2tp_net_id);
}

129 130 131 132 133 134 135 136 137 138 139 140
/* Session hash global list for L2TPv3.
 * The session_id SHOULD be random according to RFC3931, but several
 * L2TP implementations use incrementing session_ids.  So we do a real
 * hash on the session_id, rather than a simple bitmask.
 */
static inline struct hlist_head *
l2tp_session_id_hash_2(struct l2tp_net *pn, u32 session_id)
{
	return &pn->l2tp_session_hlist[hash_32(session_id, L2TP_HASH_BITS_2)];

}

141 142 143 144
/* Lookup the tunnel socket, possibly involving the fs code if the socket is
 * owned by userspace.  A struct sock returned from this function must be
 * released using l2tp_tunnel_sock_put once you're done with it.
 */
145
static struct sock *l2tp_tunnel_sock_lookup(struct l2tp_tunnel *tunnel)
146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164
{
	int err = 0;
	struct socket *sock = NULL;
	struct sock *sk = NULL;

	if (!tunnel)
		goto out;

	if (tunnel->fd >= 0) {
		/* Socket is owned by userspace, who might be in the process
		 * of closing it.  Look the socket up using the fd to ensure
		 * consistency.
		 */
		sock = sockfd_lookup(tunnel->fd, &err);
		if (sock)
			sk = sock->sk;
	} else {
		/* Socket is owned by kernelspace */
		sk = tunnel->sock;
165
		sock_hold(sk);
166 167 168 169 170 171 172
	}

out:
	return sk;
}

/* Drop a reference to a tunnel socket obtained via. l2tp_tunnel_sock_put */
173
static void l2tp_tunnel_sock_put(struct sock *sk)
174 175 176 177 178 179 180 181 182
{
	struct l2tp_tunnel *tunnel = l2tp_sock_to_tunnel(sk);
	if (tunnel) {
		if (tunnel->fd >= 0) {
			/* Socket is owned by userspace */
			sockfd_put(sk->sk_socket);
		}
		sock_put(sk);
	}
183
	sock_put(sk);
184 185
}

186 187 188 189 190 191 192 193 194 195 196 197
/* Session hash list.
 * The session_id SHOULD be random according to RFC2661, but several
 * L2TP implementations (Cisco and Microsoft) use incrementing
 * session_ids.  So we do a real hash on the session_id, rather than a
 * simple bitmask.
 */
static inline struct hlist_head *
l2tp_session_id_hash(struct l2tp_tunnel *tunnel, u32 session_id)
{
	return &tunnel->session_hlist[hash_32(session_id, L2TP_HASH_BITS)];
}

198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218
/* Lookup a tunnel. A new reference is held on the returned tunnel. */
struct l2tp_tunnel *l2tp_tunnel_get(const struct net *net, u32 tunnel_id)
{
	const struct l2tp_net *pn = l2tp_pernet(net);
	struct l2tp_tunnel *tunnel;

	rcu_read_lock_bh();
	list_for_each_entry_rcu(tunnel, &pn->l2tp_tunnel_list, list) {
		if (tunnel->tunnel_id == tunnel_id) {
			l2tp_tunnel_inc_refcount(tunnel);
			rcu_read_unlock_bh();

			return tunnel;
		}
	}
	rcu_read_unlock_bh();

	return NULL;
}
EXPORT_SYMBOL_GPL(l2tp_tunnel_get);

219
/* Lookup a session. A new reference is held on the returned session.
220 221
 * Optionally calls session->ref() too if do_ref is true.
 */
222
struct l2tp_session *l2tp_session_get(const struct net *net,
223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267
				      struct l2tp_tunnel *tunnel,
				      u32 session_id, bool do_ref)
{
	struct hlist_head *session_list;
	struct l2tp_session *session;

	if (!tunnel) {
		struct l2tp_net *pn = l2tp_pernet(net);

		session_list = l2tp_session_id_hash_2(pn, session_id);

		rcu_read_lock_bh();
		hlist_for_each_entry_rcu(session, session_list, global_hlist) {
			if (session->session_id == session_id) {
				l2tp_session_inc_refcount(session);
				if (do_ref && session->ref)
					session->ref(session);
				rcu_read_unlock_bh();

				return session;
			}
		}
		rcu_read_unlock_bh();

		return NULL;
	}

	session_list = l2tp_session_id_hash(tunnel, session_id);
	read_lock_bh(&tunnel->hlist_lock);
	hlist_for_each_entry(session, session_list, hlist) {
		if (session->session_id == session_id) {
			l2tp_session_inc_refcount(session);
			if (do_ref && session->ref)
				session->ref(session);
			read_unlock_bh(&tunnel->hlist_lock);

			return session;
		}
	}
	read_unlock_bh(&tunnel->hlist_lock);

	return NULL;
}
EXPORT_SYMBOL_GPL(l2tp_session_get);

268 269
struct l2tp_session *l2tp_session_get_nth(struct l2tp_tunnel *tunnel, int nth,
					  bool do_ref)
270 271 272 273 274 275 276
{
	int hash;
	struct l2tp_session *session;
	int count = 0;

	read_lock_bh(&tunnel->hlist_lock);
	for (hash = 0; hash < L2TP_HASH_SIZE; hash++) {
277
		hlist_for_each_entry(session, &tunnel->session_hlist[hash], hlist) {
278
			if (++count > nth) {
279 280 281
				l2tp_session_inc_refcount(session);
				if (do_ref && session->ref)
					session->ref(session);
282 283 284 285 286 287 288 289 290 291
				read_unlock_bh(&tunnel->hlist_lock);
				return session;
			}
		}
	}

	read_unlock_bh(&tunnel->hlist_lock);

	return NULL;
}
292
EXPORT_SYMBOL_GPL(l2tp_session_get_nth);
293

294 295 296
/* Lookup a session by interface name.
 * This is very inefficient but is only used by management interfaces.
 */
297 298
struct l2tp_session *l2tp_session_get_by_ifname(const struct net *net,
						const char *ifname,
299
						bool do_ref)
300 301 302 303 304
{
	struct l2tp_net *pn = l2tp_pernet(net);
	int hash;
	struct l2tp_session *session;

J
James Chapman 已提交
305
	rcu_read_lock_bh();
306
	for (hash = 0; hash < L2TP_HASH_SIZE_2; hash++) {
307
		hlist_for_each_entry_rcu(session, &pn->l2tp_session_hlist[hash], global_hlist) {
308
			if (!strcmp(session->ifname, ifname)) {
309 310 311
				l2tp_session_inc_refcount(session);
				if (do_ref && session->ref)
					session->ref(session);
J
James Chapman 已提交
312
				rcu_read_unlock_bh();
313

314 315 316 317 318
				return session;
			}
		}
	}

J
James Chapman 已提交
319
	rcu_read_unlock_bh();
320 321 322

	return NULL;
}
323
EXPORT_SYMBOL_GPL(l2tp_session_get_by_ifname);
324

325 326 327 328 329 330 331
static int l2tp_session_add_to_tunnel(struct l2tp_tunnel *tunnel,
				      struct l2tp_session *session)
{
	struct l2tp_session *session_walk;
	struct hlist_head *g_head;
	struct hlist_head *head;
	struct l2tp_net *pn;
332
	int err;
333 334 335 336

	head = l2tp_session_id_hash(tunnel, session->session_id);

	write_lock_bh(&tunnel->hlist_lock);
337 338 339 340 341
	if (!tunnel->acpt_newsess) {
		err = -ENODEV;
		goto err_tlock;
	}

342
	hlist_for_each_entry(session_walk, head, hlist)
343 344 345 346
		if (session_walk->session_id == session->session_id) {
			err = -EEXIST;
			goto err_tlock;
		}
347 348 349 350 351 352 353

	if (tunnel->version == L2TP_HDR_VER_3) {
		pn = l2tp_pernet(tunnel->l2tp_net);
		g_head = l2tp_session_id_hash_2(l2tp_pernet(tunnel->l2tp_net),
						session->session_id);

		spin_lock_bh(&pn->l2tp_session_hlist_lock);
354

355
		hlist_for_each_entry(session_walk, g_head, global_hlist)
356 357 358 359
			if (session_walk->session_id == session->session_id) {
				err = -EEXIST;
				goto err_tlock_pnlock;
			}
360

361 362
		l2tp_tunnel_inc_refcount(tunnel);
		sock_hold(tunnel->sock);
363
		hlist_add_head_rcu(&session->global_hlist, g_head);
364

365
		spin_unlock_bh(&pn->l2tp_session_hlist_lock);
366 367 368
	} else {
		l2tp_tunnel_inc_refcount(tunnel);
		sock_hold(tunnel->sock);
369 370 371 372 373 374 375
	}

	hlist_add_head(&session->hlist, head);
	write_unlock_bh(&tunnel->hlist_lock);

	return 0;

376
err_tlock_pnlock:
377
	spin_unlock_bh(&pn->l2tp_session_hlist_lock);
378
err_tlock:
379 380
	write_unlock_bh(&tunnel->hlist_lock);

381
	return err;
382 383
}

384 385
/* Lookup a tunnel by id
 */
386
struct l2tp_tunnel *l2tp_tunnel_find(const struct net *net, u32 tunnel_id)
387 388 389 390
{
	struct l2tp_tunnel *tunnel;
	struct l2tp_net *pn = l2tp_pernet(net);

J
James Chapman 已提交
391 392
	rcu_read_lock_bh();
	list_for_each_entry_rcu(tunnel, &pn->l2tp_tunnel_list, list) {
393
		if (tunnel->tunnel_id == tunnel_id) {
J
James Chapman 已提交
394
			rcu_read_unlock_bh();
395 396 397
			return tunnel;
		}
	}
J
James Chapman 已提交
398
	rcu_read_unlock_bh();
399 400 401 402 403

	return NULL;
}
EXPORT_SYMBOL_GPL(l2tp_tunnel_find);

404
struct l2tp_tunnel *l2tp_tunnel_find_nth(const struct net *net, int nth)
405 406 407 408 409
{
	struct l2tp_net *pn = l2tp_pernet(net);
	struct l2tp_tunnel *tunnel;
	int count = 0;

J
James Chapman 已提交
410 411
	rcu_read_lock_bh();
	list_for_each_entry_rcu(tunnel, &pn->l2tp_tunnel_list, list) {
412
		if (++count > nth) {
J
James Chapman 已提交
413
			rcu_read_unlock_bh();
414 415 416 417
			return tunnel;
		}
	}

J
James Chapman 已提交
418
	rcu_read_unlock_bh();
419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434

	return NULL;
}
EXPORT_SYMBOL_GPL(l2tp_tunnel_find_nth);

/*****************************************************************************
 * Receive data handling
 *****************************************************************************/

/* Queue a skb in order. We come here only if the skb has an L2TP sequence
 * number.
 */
static void l2tp_recv_queue_skb(struct l2tp_session *session, struct sk_buff *skb)
{
	struct sk_buff *skbp;
	struct sk_buff *tmp;
435
	u32 ns = L2TP_SKB_CB(skb)->ns;
436 437 438 439 440

	spin_lock_bh(&session->reorder_q.lock);
	skb_queue_walk_safe(&session->reorder_q, skbp, tmp) {
		if (L2TP_SKB_CB(skbp)->ns > ns) {
			__skb_queue_before(&session->reorder_q, skbp, skb);
441 442 443 444
			l2tp_dbg(session, L2TP_MSG_SEQ,
				 "%s: pkt %hu, inserted before %hu, reorder_q len=%d\n",
				 session->name, ns, L2TP_SKB_CB(skbp)->ns,
				 skb_queue_len(&session->reorder_q));
445
			atomic_long_inc(&session->stats.rx_oos_packets);
446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467
			goto out;
		}
	}

	__skb_queue_tail(&session->reorder_q, skb);

out:
	spin_unlock_bh(&session->reorder_q.lock);
}

/* Dequeue a single skb.
 */
static void l2tp_recv_dequeue_skb(struct l2tp_session *session, struct sk_buff *skb)
{
	struct l2tp_tunnel *tunnel = session->tunnel;
	int length = L2TP_SKB_CB(skb)->length;

	/* We're about to requeue the skb, so return resources
	 * to its current owner (a socket receive buffer).
	 */
	skb_orphan(skb);

468 469 470 471
	atomic_long_inc(&tunnel->stats.rx_packets);
	atomic_long_add(length, &tunnel->stats.rx_bytes);
	atomic_long_inc(&session->stats.rx_packets);
	atomic_long_add(length, &session->stats.rx_bytes);
472 473 474 475

	if (L2TP_SKB_CB(skb)->has_seq) {
		/* Bump our Nr */
		session->nr++;
476
		session->nr &= session->nr_max;
477

478 479
		l2tp_dbg(session, L2TP_MSG_SEQ, "%s: updated nr to %hu\n",
			 session->name, session->nr);
480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503
	}

	/* call private receive handler */
	if (session->recv_skb != NULL)
		(*session->recv_skb)(session, skb, L2TP_SKB_CB(skb)->length);
	else
		kfree_skb(skb);

	if (session->deref)
		(*session->deref)(session);
}

/* Dequeue skbs from the session's reorder_q, subject to packet order.
 * Skbs that have been in the queue for too long are simply discarded.
 */
static void l2tp_recv_dequeue(struct l2tp_session *session)
{
	struct sk_buff *skb;
	struct sk_buff *tmp;

	/* If the pkt at the head of the queue has the nr that we
	 * expect to send up next, dequeue it and any other
	 * in-sequence packets behind it.
	 */
504
start:
505 506 507
	spin_lock_bh(&session->reorder_q.lock);
	skb_queue_walk_safe(&session->reorder_q, skb, tmp) {
		if (time_after(jiffies, L2TP_SKB_CB(skb)->expires)) {
508 509
			atomic_long_inc(&session->stats.rx_seq_discards);
			atomic_long_inc(&session->stats.rx_errors);
510 511 512 513 514
			l2tp_dbg(session, L2TP_MSG_SEQ,
				 "%s: oos pkt %u len %d discarded (too old), waiting for %u, reorder_q_len=%d\n",
				 session->name, L2TP_SKB_CB(skb)->ns,
				 L2TP_SKB_CB(skb)->length, session->nr,
				 skb_queue_len(&session->reorder_q));
515
			session->reorder_skip = 1;
516 517 518 519 520 521 522 523
			__skb_unlink(skb, &session->reorder_q);
			kfree_skb(skb);
			if (session->deref)
				(*session->deref)(session);
			continue;
		}

		if (L2TP_SKB_CB(skb)->has_seq) {
524
			if (session->reorder_skip) {
525 526 527 528
				l2tp_dbg(session, L2TP_MSG_SEQ,
					 "%s: advancing nr to next pkt: %u -> %u",
					 session->name, session->nr,
					 L2TP_SKB_CB(skb)->ns);
529 530 531
				session->reorder_skip = 0;
				session->nr = L2TP_SKB_CB(skb)->ns;
			}
532
			if (L2TP_SKB_CB(skb)->ns != session->nr) {
533 534 535 536 537
				l2tp_dbg(session, L2TP_MSG_SEQ,
					 "%s: holding oos pkt %u len %d, waiting for %u, reorder_q_len=%d\n",
					 session->name, L2TP_SKB_CB(skb)->ns,
					 L2TP_SKB_CB(skb)->length, session->nr,
					 skb_queue_len(&session->reorder_q));
538 539 540 541 542 543 544 545 546 547
				goto out;
			}
		}
		__skb_unlink(skb, &session->reorder_q);

		/* Process the skb. We release the queue lock while we
		 * do so to let other contexts process the queue.
		 */
		spin_unlock_bh(&session->reorder_q.lock);
		l2tp_recv_dequeue_skb(session, skb);
548
		goto start;
549 550 551 552 553 554
	}

out:
	spin_unlock_bh(&session->reorder_q.lock);
}

555 556 557 558 559 560 561 562 563 564 565 566
static int l2tp_seq_check_rx_window(struct l2tp_session *session, u32 nr)
{
	u32 nws;

	if (nr >= session->nr)
		nws = nr - session->nr;
	else
		nws = (session->nr_max + 1) - (session->nr - nr);

	return nws < session->nr_window_size;
}

567 568 569 570 571
/* If packet has sequence numbers, queue it if acceptable. Returns 0 if
 * acceptable, else non-zero.
 */
static int l2tp_recv_data_seq(struct l2tp_session *session, struct sk_buff *skb)
{
572 573 574 575 576 577 578 579 580 581 582
	if (!l2tp_seq_check_rx_window(session, L2TP_SKB_CB(skb)->ns)) {
		/* Packet sequence number is outside allowed window.
		 * Discard it.
		 */
		l2tp_dbg(session, L2TP_MSG_SEQ,
			 "%s: pkt %u len %d discarded, outside window, nr=%u\n",
			 session->name, L2TP_SKB_CB(skb)->ns,
			 L2TP_SKB_CB(skb)->length, session->nr);
		goto discard;
	}

583 584 585 586 587
	if (session->reorder_timeout != 0) {
		/* Packet reordering enabled. Add skb to session's
		 * reorder queue, in order of ns.
		 */
		l2tp_recv_queue_skb(session, skb);
588 589 590 591 592 593 594 595 596 597
		goto out;
	}

	/* Packet reordering disabled. Discard out-of-sequence packets, while
	 * tracking the number if in-sequence packets after the first OOS packet
	 * is seen. After nr_oos_count_max in-sequence packets, reset the
	 * sequence number to re-enable packet reception.
	 */
	if (L2TP_SKB_CB(skb)->ns == session->nr) {
		skb_queue_tail(&session->reorder_q, skb);
598
	} else {
599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614
		u32 nr_oos = L2TP_SKB_CB(skb)->ns;
		u32 nr_next = (session->nr_oos + 1) & session->nr_max;

		if (nr_oos == nr_next)
			session->nr_oos_count++;
		else
			session->nr_oos_count = 0;

		session->nr_oos = nr_oos;
		if (session->nr_oos_count > session->nr_oos_count_max) {
			session->reorder_skip = 1;
			l2tp_dbg(session, L2TP_MSG_SEQ,
				 "%s: %d oos packets received. Resetting sequence numbers\n",
				 session->name, session->nr_oos_count);
		}
		if (!session->reorder_skip) {
615 616 617 618 619 620 621 622 623 624 625
			atomic_long_inc(&session->stats.rx_seq_discards);
			l2tp_dbg(session, L2TP_MSG_SEQ,
				 "%s: oos pkt %u len %d discarded, waiting for %u, reorder_q_len=%d\n",
				 session->name, L2TP_SKB_CB(skb)->ns,
				 L2TP_SKB_CB(skb)->length, session->nr,
				 skb_queue_len(&session->reorder_q));
			goto discard;
		}
		skb_queue_tail(&session->reorder_q, skb);
	}

626
out:
627 628 629 630 631 632
	return 0;

discard:
	return 1;
}

633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691
/* Do receive processing of L2TP data frames. We handle both L2TPv2
 * and L2TPv3 data frames here.
 *
 * L2TPv2 Data Message Header
 *
 *  0                   1                   2                   3
 *  0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |T|L|x|x|S|x|O|P|x|x|x|x|  Ver  |          Length (opt)         |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |           Tunnel ID           |           Session ID          |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |             Ns (opt)          |             Nr (opt)          |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |      Offset Size (opt)        |    Offset pad... (opt)
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 *
 * Data frames are marked by T=0. All other fields are the same as
 * those in L2TP control frames.
 *
 * L2TPv3 Data Message Header
 *
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                      L2TP Session Header                      |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                      L2-Specific Sublayer                     |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                        Tunnel Payload                      ...
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 *
 * L2TPv3 Session Header Over IP
 *
 *  0                   1                   2                   3
 *  0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                           Session ID                          |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |               Cookie (optional, maximum 64 bits)...
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 *                                                                 |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 *
 * L2TPv3 L2-Specific Sublayer Format
 *
 *  0                   1                   2                   3
 *  0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |x|S|x|x|x|x|x|x|              Sequence Number                  |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 *
 * Cookie value, sublayer format and offset (pad) are negotiated with
 * the peer when the session is set up. Unlike L2TPv2, we do not need
 * to parse the packet header to determine if optional fields are
 * present.
 *
 * Caller must already have parsed the frame and determined that it is
 * a data (not control) frame before coming here. Fields up to the
 * session-id have already been parsed and ptr points to the data
 * after the session-id.
692 693 694
 *
 * session->ref() must have been called prior to l2tp_recv_common().
 * session->deref() will be called automatically after skb is processed.
695
 */
696 697 698
void l2tp_recv_common(struct l2tp_session *session, struct sk_buff *skb,
		      unsigned char *ptr, unsigned char *optr, u16 hdrflags,
		      int length, int (*payload_hook)(struct sk_buff *skb))
699
{
700
	struct l2tp_tunnel *tunnel = session->tunnel;
701
	int offset;
702
	u32 ns, nr;
703

704 705 706
	/* Parse and check optional cookie */
	if (session->peer_cookie_len > 0) {
		if (memcmp(ptr, &session->peer_cookie[0], session->peer_cookie_len)) {
707 708 709 710
			l2tp_info(tunnel, L2TP_MSG_DATA,
				  "%s: cookie mismatch (%u/%u). Discarding.\n",
				  tunnel->name, tunnel->tunnel_id,
				  session->session_id);
711
			atomic_long_inc(&session->stats.rx_cookie_discards);
712 713 714 715 716
			goto discard;
		}
		ptr += session->peer_cookie_len;
	}

717 718 719 720 721 722 723 724 725
	/* Handle the optional sequence numbers. Sequence numbers are
	 * in different places for L2TPv2 and L2TPv3.
	 *
	 * If we are the LAC, enable/disable sequence numbers under
	 * the control of the LNS.  If no sequence numbers present but
	 * we were expecting them, discard frame.
	 */
	ns = nr = 0;
	L2TP_SKB_CB(skb)->has_seq = 0;
726 727 728 729 730 731
	if (tunnel->version == L2TP_HDR_VER_2) {
		if (hdrflags & L2TP_HDRFLAG_S) {
			ns = ntohs(*(__be16 *) ptr);
			ptr += 2;
			nr = ntohs(*(__be16 *) ptr);
			ptr += 2;
732

733 734 735
			/* Store L2TP info in the skb */
			L2TP_SKB_CB(skb)->ns = ns;
			L2TP_SKB_CB(skb)->has_seq = 1;
736

737 738 739
			l2tp_dbg(session, L2TP_MSG_SEQ,
				 "%s: recv data ns=%u, nr=%u, session nr=%u\n",
				 session->name, ns, nr, session->nr);
740 741 742 743 744 745 746 747 748 749 750
		}
	} else if (session->l2specific_type == L2TP_L2SPECTYPE_DEFAULT) {
		u32 l2h = ntohl(*(__be32 *) ptr);

		if (l2h & 0x40000000) {
			ns = l2h & 0x00ffffff;

			/* Store L2TP info in the skb */
			L2TP_SKB_CB(skb)->ns = ns;
			L2TP_SKB_CB(skb)->has_seq = 1;

751 752 753
			l2tp_dbg(session, L2TP_MSG_SEQ,
				 "%s: recv data ns=%u, session nr=%u\n",
				 session->name, ns, session->nr);
754
		}
755 756
	}

757 758 759
	/* Advance past L2-specific header, if present */
	ptr += session->l2specific_len;

760 761 762 763 764 765
	if (L2TP_SKB_CB(skb)->has_seq) {
		/* Received a packet with sequence numbers. If we're the LNS,
		 * check if we sre sending sequence numbers and if not,
		 * configure it so.
		 */
		if ((!session->lns_mode) && (!session->send_seq)) {
766 767 768
			l2tp_info(session, L2TP_MSG_SEQ,
				  "%s: requested to enable seq numbers by LNS\n",
				  session->name);
769
			session->send_seq = 1;
770
			l2tp_session_set_header_len(session, tunnel->version);
771 772 773 774 775 776
		}
	} else {
		/* No sequence numbers.
		 * If user has configured mandatory sequence numbers, discard.
		 */
		if (session->recv_seq) {
777 778 779
			l2tp_warn(session, L2TP_MSG_SEQ,
				  "%s: recv data has no seq numbers when required. Discarding.\n",
				  session->name);
780
			atomic_long_inc(&session->stats.rx_seq_discards);
781 782 783 784 785 786 787 788 789
			goto discard;
		}

		/* If we're the LAC and we're sending sequence numbers, the
		 * LNS has requested that we no longer send sequence numbers.
		 * If we're the LNS and we're sending sequence numbers, the
		 * LAC is broken. Discard the frame.
		 */
		if ((!session->lns_mode) && (session->send_seq)) {
790 791 792
			l2tp_info(session, L2TP_MSG_SEQ,
				  "%s: requested to disable seq numbers by LNS\n",
				  session->name);
793
			session->send_seq = 0;
794
			l2tp_session_set_header_len(session, tunnel->version);
795
		} else if (session->send_seq) {
796 797 798
			l2tp_warn(session, L2TP_MSG_SEQ,
				  "%s: recv data has no seq numbers when required. Discarding.\n",
				  session->name);
799
			atomic_long_inc(&session->stats.rx_seq_discards);
800 801 802 803
			goto discard;
		}
	}

804 805 806 807 808 809 810 811 812 813 814 815 816
	/* Session data offset is handled differently for L2TPv2 and
	 * L2TPv3. For L2TPv2, there is an optional 16-bit value in
	 * the header. For L2TPv3, the offset is negotiated using AVPs
	 * in the session setup control protocol.
	 */
	if (tunnel->version == L2TP_HDR_VER_2) {
		/* If offset bit set, skip it. */
		if (hdrflags & L2TP_HDRFLAG_O) {
			offset = ntohs(*(__be16 *)ptr);
			ptr += 2 + offset;
		}
	} else
		ptr += session->offset;
817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842

	offset = ptr - optr;
	if (!pskb_may_pull(skb, offset))
		goto discard;

	__skb_pull(skb, offset);

	/* If caller wants to process the payload before we queue the
	 * packet, do so now.
	 */
	if (payload_hook)
		if ((*payload_hook)(skb))
			goto discard;

	/* Prepare skb for adding to the session's reorder_q.  Hold
	 * packets for max reorder_timeout or 1 second if not
	 * reordering.
	 */
	L2TP_SKB_CB(skb)->length = length;
	L2TP_SKB_CB(skb)->expires = jiffies +
		(session->reorder_timeout ? session->reorder_timeout : HZ);

	/* Add packet to the session's receive queue. Reordering is done here, if
	 * enabled. Saved L2TP protocol info is stored in skb->sb[].
	 */
	if (L2TP_SKB_CB(skb)->has_seq) {
843 844
		if (l2tp_recv_data_seq(session, skb))
			goto discard;
845 846 847 848 849 850 851 852 853 854 855
	} else {
		/* No sequence numbers. Add the skb to the tail of the
		 * reorder queue. This ensures that it will be
		 * delivered after all previous sequenced skbs.
		 */
		skb_queue_tail(&session->reorder_q, skb);
	}

	/* Try to dequeue as many skbs from reorder_q as we can. */
	l2tp_recv_dequeue(session);

856
	return;
857 858

discard:
859
	atomic_long_inc(&session->stats.rx_errors);
860 861 862 863
	kfree_skb(skb);

	if (session->deref)
		(*session->deref)(session);
864 865 866
}
EXPORT_SYMBOL(l2tp_recv_common);

867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883
/* Drop skbs from the session's reorder_q
 */
int l2tp_session_queue_purge(struct l2tp_session *session)
{
	struct sk_buff *skb = NULL;
	BUG_ON(!session);
	BUG_ON(session->magic != L2TP_SESSION_MAGIC);
	while ((skb = skb_dequeue(&session->reorder_q))) {
		atomic_long_inc(&session->stats.rx_errors);
		kfree_skb(skb);
		if (session->deref)
			(*session->deref)(session);
	}
	return 0;
}
EXPORT_SYMBOL_GPL(l2tp_session_queue_purge);

884 885 886 887 888 889
/* Internal UDP receive frame. Do the real work of receiving an L2TP data frame
 * here. The skb is not on a list when we get here.
 * Returns 0 if the packet was a data packet and was successfully passed on.
 * Returns 1 if the packet was not a good data packet and could not be
 * forwarded.  All such packets are passed up to userspace to deal with.
 */
890 891
static int l2tp_udp_recv_core(struct l2tp_tunnel *tunnel, struct sk_buff *skb,
			      int (*payload_hook)(struct sk_buff *skb))
892 893 894 895 896 897 898 899
{
	struct l2tp_session *session = NULL;
	unsigned char *ptr, *optr;
	u16 hdrflags;
	u32 tunnel_id, session_id;
	u16 version;
	int length;

900
	/* UDP has verifed checksum */
901 902 903 904 905 906

	/* UDP always verifies the packet length. */
	__skb_pull(skb, sizeof(struct udphdr));

	/* Short packet? */
	if (!pskb_may_pull(skb, L2TP_HDR_SIZE_SEQ)) {
907 908 909
		l2tp_info(tunnel, L2TP_MSG_DATA,
			  "%s: recv short packet (len=%d)\n",
			  tunnel->name, skb->len);
910 911 912 913 914 915 916 917 918
		goto error;
	}

	/* Trace packet contents, if enabled */
	if (tunnel->debug & L2TP_MSG_DATA) {
		length = min(32u, skb->len);
		if (!pskb_may_pull(skb, length))
			goto error;

919 920
		pr_debug("%s: recv\n", tunnel->name);
		print_hex_dump_bytes("", DUMP_PREFIX_OFFSET, skb->data, length);
921 922
	}

E
Eric Dumazet 已提交
923 924 925
	/* Point to L2TP header */
	optr = ptr = skb->data;

926 927 928 929 930 931
	/* Get L2TP header flags */
	hdrflags = ntohs(*(__be16 *) ptr);

	/* Check protocol version */
	version = hdrflags & L2TP_HDR_VER_MASK;
	if (version != tunnel->version) {
932 933 934
		l2tp_info(tunnel, L2TP_MSG_DATA,
			  "%s: recv protocol version mismatch: got %d expected %d\n",
			  tunnel->name, version, tunnel->version);
935 936 937 938 939 940 941 942
		goto error;
	}

	/* Get length of L2TP packet */
	length = skb->len;

	/* If type is control packet, it is handled by userspace. */
	if (hdrflags & L2TP_HDRFLAG_T) {
943 944 945
		l2tp_dbg(tunnel, L2TP_MSG_DATA,
			 "%s: recv control packet, len=%d\n",
			 tunnel->name, length);
946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969
		goto error;
	}

	/* Skip flags */
	ptr += 2;

	if (tunnel->version == L2TP_HDR_VER_2) {
		/* If length is present, skip it */
		if (hdrflags & L2TP_HDRFLAG_L)
			ptr += 2;

		/* Extract tunnel and session ID */
		tunnel_id = ntohs(*(__be16 *) ptr);
		ptr += 2;
		session_id = ntohs(*(__be16 *) ptr);
		ptr += 2;
	} else {
		ptr += 2;	/* skip reserved bits */
		tunnel_id = tunnel->tunnel_id;
		session_id = ntohl(*(__be32 *) ptr);
		ptr += 4;
	}

	/* Find the session context */
970
	session = l2tp_session_get(tunnel->l2tp_net, tunnel, session_id, true);
971
	if (!session || !session->recv_skb) {
972 973 974 975 976 977
		if (session) {
			if (session->deref)
				session->deref(session);
			l2tp_session_dec_refcount(session);
		}

978
		/* Not found? Pass to userspace to deal with */
979 980 981
		l2tp_info(tunnel, L2TP_MSG_DATA,
			  "%s: no session found (%u/%u). Passing up.\n",
			  tunnel->name, tunnel_id, session_id);
982 983 984 985
		goto error;
	}

	l2tp_recv_common(session, skb, ptr, optr, hdrflags, length, payload_hook);
986
	l2tp_session_dec_refcount(session);
987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010

	return 0;

error:
	/* Put UDP header back */
	__skb_push(skb, sizeof(struct udphdr));

	return 1;
}

/* UDP encapsulation receive handler. See net/ipv4/udp.c.
 * Return codes:
 * 0 : success.
 * <0: error
 * >0: skb should be passed up to userspace as UDP.
 */
int l2tp_udp_encap_recv(struct sock *sk, struct sk_buff *skb)
{
	struct l2tp_tunnel *tunnel;

	tunnel = l2tp_sock_to_tunnel(sk);
	if (tunnel == NULL)
		goto pass_up;

1011 1012
	l2tp_dbg(tunnel, L2TP_MSG_DATA, "%s: received %d bytes\n",
		 tunnel->name, skb->len);
1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032

	if (l2tp_udp_recv_core(tunnel, skb, tunnel->recv_payload_hook))
		goto pass_up_put;

	sock_put(sk);
	return 0;

pass_up_put:
	sock_put(sk);
pass_up:
	return 1;
}
EXPORT_SYMBOL_GPL(l2tp_udp_encap_recv);

/************************************************************************
 * Transmit handling
 ***********************************************************************/

/* Build an L2TP header for the session into the buffer provided.
 */
1033
static int l2tp_build_l2tpv2_header(struct l2tp_session *session, void *buf)
1034
{
1035
	struct l2tp_tunnel *tunnel = session->tunnel;
1036
	__be16 *bufp = buf;
1037
	__be16 *optr = buf;
1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052
	u16 flags = L2TP_HDR_VER_2;
	u32 tunnel_id = tunnel->peer_tunnel_id;
	u32 session_id = session->peer_session_id;

	if (session->send_seq)
		flags |= L2TP_HDRFLAG_S;

	/* Setup L2TP header. */
	*bufp++ = htons(flags);
	*bufp++ = htons(tunnel_id);
	*bufp++ = htons(session_id);
	if (session->send_seq) {
		*bufp++ = htons(session->ns);
		*bufp++ = 0;
		session->ns++;
1053
		session->ns &= 0xffff;
1054 1055
		l2tp_dbg(session, L2TP_MSG_SEQ, "%s: updated ns to %u\n",
			 session->name, session->ns);
1056
	}
1057 1058

	return bufp - optr;
1059 1060
}

1061
static int l2tp_build_l2tpv3_header(struct l2tp_session *session, void *buf)
1062
{
1063
	struct l2tp_tunnel *tunnel = session->tunnel;
1064 1065 1066
	char *bufp = buf;
	char *optr = bufp;

1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077
	/* Setup L2TP header. The header differs slightly for UDP and
	 * IP encapsulations. For UDP, there is 4 bytes of flags.
	 */
	if (tunnel->encap == L2TP_ENCAPTYPE_UDP) {
		u16 flags = L2TP_HDR_VER_3;
		*((__be16 *) bufp) = htons(flags);
		bufp += 2;
		*((__be16 *) bufp) = 0;
		bufp += 2;
	}

1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090
	*((__be32 *) bufp) = htonl(session->peer_session_id);
	bufp += 4;
	if (session->cookie_len) {
		memcpy(bufp, &session->cookie[0], session->cookie_len);
		bufp += session->cookie_len;
	}
	if (session->l2specific_len) {
		if (session->l2specific_type == L2TP_L2SPECTYPE_DEFAULT) {
			u32 l2h = 0;
			if (session->send_seq) {
				l2h = 0x40000000 | session->ns;
				session->ns++;
				session->ns &= 0xffffff;
1091 1092 1093
				l2tp_dbg(session, L2TP_MSG_SEQ,
					 "%s: updated ns to %u\n",
					 session->name, session->ns);
1094 1095 1096 1097 1098 1099 1100 1101
			}

			*((__be32 *) bufp) = htonl(l2h);
		}
		bufp += session->l2specific_len;
	}
	if (session->offset)
		bufp += session->offset;
1102

1103
	return bufp - optr;
1104 1105
}

1106
static int l2tp_xmit_core(struct l2tp_session *session, struct sk_buff *skb,
1107
			  struct flowi *fl, size_t data_len)
1108 1109 1110 1111 1112 1113 1114
{
	struct l2tp_tunnel *tunnel = session->tunnel;
	unsigned int len = skb->len;
	int error;

	/* Debug */
	if (session->send_seq)
1115
		l2tp_dbg(session, L2TP_MSG_DATA, "%s: send %zd bytes, ns=%u\n",
1116
			 session->name, data_len, session->ns - 1);
1117
	else
1118
		l2tp_dbg(session, L2TP_MSG_DATA, "%s: send %zd bytes\n",
1119
			 session->name, data_len);
1120 1121

	if (session->debug & L2TP_MSG_DATA) {
1122 1123
		int uhlen = (tunnel->encap == L2TP_ENCAPTYPE_UDP) ? sizeof(struct udphdr) : 0;
		unsigned char *datap = skb->data + uhlen;
1124

1125 1126 1127
		pr_debug("%s: xmit\n", session->name);
		print_hex_dump_bytes("", DUMP_PREFIX_OFFSET,
				     datap, min_t(size_t, 32, len - uhlen));
1128 1129 1130
	}

	/* Queue the packet to IP for output */
W
WANG Cong 已提交
1131
	skb->ignore_df = 1;
1132
#if IS_ENABLED(CONFIG_IPV6)
E
Eric Dumazet 已提交
1133
	if (tunnel->sock->sk_family == PF_INET6 && !tunnel->v4mapped)
1134
		error = inet6_csk_xmit(tunnel->sock, skb, NULL);
1135 1136
	else
#endif
1137
		error = ip_queue_xmit(tunnel->sock, skb, fl);
1138 1139 1140

	/* Update stats */
	if (error >= 0) {
1141 1142 1143 1144
		atomic_long_inc(&tunnel->stats.tx_packets);
		atomic_long_add(len, &tunnel->stats.tx_bytes);
		atomic_long_inc(&session->stats.tx_packets);
		atomic_long_add(len, &session->stats.tx_bytes);
1145
	} else {
1146 1147
		atomic_long_inc(&tunnel->stats.tx_errors);
		atomic_long_inc(&session->stats.tx_errors);
1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158
	}

	return 0;
}

/* If caller requires the skb to have a ppp header, the header must be
 * inserted in the skb data before calling this function.
 */
int l2tp_xmit_skb(struct l2tp_session *session, struct sk_buff *skb, int hdr_len)
{
	int data_len = skb->len;
1159 1160
	struct l2tp_tunnel *tunnel = session->tunnel;
	struct sock *sk = tunnel->sock;
1161
	struct flowi *fl;
1162 1163 1164
	struct udphdr *uh;
	struct inet_sock *inet;
	int headroom;
1165 1166
	int uhlen = (tunnel->encap == L2TP_ENCAPTYPE_UDP) ? sizeof(struct udphdr) : 0;
	int udp_len;
1167
	int ret = NET_XMIT_SUCCESS;
1168 1169 1170 1171 1172 1173

	/* Check that there's enough headroom in the skb to insert IP,
	 * UDP and L2TP headers. If not enough, expand it to
	 * make room. Adjust truesize.
	 */
	headroom = NET_SKB_PAD + sizeof(struct iphdr) +
1174
		uhlen + hdr_len;
1175
	if (skb_cow_head(skb, headroom)) {
1176 1177
		kfree_skb(skb);
		return NET_XMIT_DROP;
1178
	}
1179 1180

	/* Setup L2TP header */
1181
	session->build_header(session, __skb_push(skb, hdr_len));
1182

1183
	/* Reset skb netfilter state */
1184 1185 1186 1187 1188
	memset(&(IPCB(skb)->opt), 0, sizeof(IPCB(skb)->opt));
	IPCB(skb)->flags &= ~(IPSKB_XFRM_TUNNEL_SIZE | IPSKB_XFRM_TRANSFORMED |
			      IPSKB_REROUTED);
	nf_reset(skb);

1189 1190
	bh_lock_sock(sk);
	if (sock_owned_by_user(sk)) {
1191 1192
		kfree_skb(skb);
		ret = NET_XMIT_DROP;
1193 1194 1195
		goto out_unlock;
	}

1196 1197
	/* Get routing info from the tunnel socket */
	skb_dst_drop(skb);
1198
	skb_dst_set(skb, dst_clone(__sk_dst_check(sk, 0)));
1199

1200 1201
	inet = inet_sk(sk);
	fl = &inet->cork.fl;
1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213
	switch (tunnel->encap) {
	case L2TP_ENCAPTYPE_UDP:
		/* Setup UDP header */
		__skb_push(skb, sizeof(*uh));
		skb_reset_transport_header(skb);
		uh = udp_hdr(skb);
		uh->source = inet->inet_sport;
		uh->dest = inet->inet_dport;
		udp_len = uhlen + hdr_len + data_len;
		uh->len = htons(udp_len);

		/* Calculate UDP checksum if configured to do so */
1214
#if IS_ENABLED(CONFIG_IPV6)
1215
		if (sk->sk_family == PF_INET6 && !tunnel->v4mapped)
T
Tom Herbert 已提交
1216 1217 1218
			udp6_set_csum(udp_get_no_check6_tx(sk),
				      skb, &inet6_sk(sk)->saddr,
				      &sk->sk_v6_daddr, udp_len);
1219 1220
		else
#endif
T
Tom Herbert 已提交
1221 1222
		udp_set_csum(sk->sk_no_check_tx, skb, inet->inet_saddr,
			     inet->inet_daddr, udp_len);
1223 1224 1225 1226
		break;

	case L2TP_ENCAPTYPE_IP:
		break;
1227 1228
	}

1229
	l2tp_xmit_core(session, skb, fl, data_len);
1230 1231
out_unlock:
	bh_unlock_sock(sk);
1232

1233
	return ret;
1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244
}
EXPORT_SYMBOL_GPL(l2tp_xmit_skb);

/*****************************************************************************
 * Tinnel and session create/destroy.
 *****************************************************************************/

/* Tunnel socket destruct hook.
 * The tunnel context is deleted only when all session sockets have been
 * closed.
 */
1245
static void l2tp_tunnel_destruct(struct sock *sk)
1246
{
1247
	struct l2tp_tunnel *tunnel = l2tp_tunnel(sk);
1248
	struct l2tp_net *pn;
1249 1250 1251 1252

	if (tunnel == NULL)
		goto end;

1253
	l2tp_info(tunnel, L2TP_MSG_CONTROL, "%s: closing...\n", tunnel->name);
1254 1255


1256
	/* Disable udp encapsulation */
1257 1258 1259 1260 1261
	switch (tunnel->encap) {
	case L2TP_ENCAPTYPE_UDP:
		/* No longer an encapsulation socket. See net/ipv4/udp.c */
		(udp_sk(sk))->encap_type = 0;
		(udp_sk(sk))->encap_rcv = NULL;
1262
		(udp_sk(sk))->encap_destroy = NULL;
1263 1264 1265 1266
		break;
	case L2TP_ENCAPTYPE_IP:
		break;
	}
1267 1268 1269 1270 1271

	/* Remove hooks into tunnel socket */
	sk->sk_destruct = tunnel->old_sk_destruct;
	sk->sk_user_data = NULL;

1272 1273 1274 1275 1276 1277
	/* Remove the tunnel struct from the tunnel list */
	pn = l2tp_pernet(tunnel->l2tp_net);
	spin_lock_bh(&pn->l2tp_tunnel_list_lock);
	list_del_rcu(&tunnel->list);
	spin_unlock_bh(&pn->l2tp_tunnel_list_lock);
	atomic_dec(&l2tp_tunnel_count);
1278

1279
	l2tp_tunnel_closeall(tunnel);
1280 1281

	tunnel->sock = NULL;
1282 1283
	l2tp_tunnel_dec_refcount(tunnel);

1284 1285 1286
	/* Call the original destructor */
	if (sk->sk_destruct)
		(*sk->sk_destruct)(sk);
1287 1288 1289 1290 1291 1292
end:
	return;
}

/* When the tunnel is closed, all the attached sessions need to go too.
 */
T
Tom Parkin 已提交
1293
void l2tp_tunnel_closeall(struct l2tp_tunnel *tunnel)
1294 1295 1296 1297 1298 1299 1300 1301
{
	int hash;
	struct hlist_node *walk;
	struct hlist_node *tmp;
	struct l2tp_session *session;

	BUG_ON(tunnel == NULL);

1302 1303
	l2tp_info(tunnel, L2TP_MSG_CONTROL, "%s: closing all sessions...\n",
		  tunnel->name);
1304 1305

	write_lock_bh(&tunnel->hlist_lock);
1306
	tunnel->acpt_newsess = false;
1307 1308 1309 1310 1311
	for (hash = 0; hash < L2TP_HASH_SIZE; hash++) {
again:
		hlist_for_each_safe(walk, tmp, &tunnel->session_hlist[hash]) {
			session = hlist_entry(walk, struct l2tp_session, hlist);

1312 1313
			l2tp_info(session, L2TP_MSG_CONTROL,
				  "%s: closing session\n", session->name);
1314 1315 1316 1317 1318 1319 1320 1321

			hlist_del_init(&session->hlist);

			if (session->ref != NULL)
				(*session->ref)(session);

			write_unlock_bh(&tunnel->hlist_lock);

1322
			__l2tp_session_unhash(session);
1323 1324
			l2tp_session_queue_purge(session);

1325 1326 1327 1328 1329 1330
			if (session->session_close != NULL)
				(*session->session_close)(session);

			if (session->deref != NULL)
				(*session->deref)(session);

1331 1332
			l2tp_session_dec_refcount(session);

1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344
			write_lock_bh(&tunnel->hlist_lock);

			/* Now restart from the beginning of this hash
			 * chain.  We always remove a session from the
			 * list so we are guaranteed to make forward
			 * progress.
			 */
			goto again;
		}
	}
	write_unlock_bh(&tunnel->hlist_lock);
}
T
Tom Parkin 已提交
1345
EXPORT_SYMBOL_GPL(l2tp_tunnel_closeall);
1346

1347 1348 1349 1350 1351 1352 1353 1354 1355 1356
/* Tunnel socket destroy hook for UDP encapsulation */
static void l2tp_udp_encap_destroy(struct sock *sk)
{
	struct l2tp_tunnel *tunnel = l2tp_sock_to_tunnel(sk);
	if (tunnel) {
		l2tp_tunnel_closeall(tunnel);
		sock_put(sk);
	}
}

1357 1358 1359 1360 1361 1362 1363 1364
/* Workqueue tunnel deletion function */
static void l2tp_tunnel_del_work(struct work_struct *work)
{
	struct l2tp_tunnel *tunnel = NULL;
	struct socket *sock = NULL;
	struct sock *sk = NULL;

	tunnel = container_of(work, struct l2tp_tunnel, del_work);
1365 1366 1367

	l2tp_tunnel_closeall(tunnel);

1368 1369
	sk = l2tp_tunnel_sock_lookup(tunnel);
	if (!sk)
1370
		goto out;
1371 1372 1373

	sock = sk->sk_socket;

1374 1375 1376 1377
	/* If the tunnel socket was created by userspace, then go through the
	 * inet layer to shut the socket down, and let userspace close it.
	 * Otherwise, if we created the socket directly within the kernel, use
	 * the sk API to release it here.
1378 1379
	 * In either case the tunnel resources are freed in the socket
	 * destructor when the tunnel socket goes away.
1380
	 */
1381 1382 1383
	if (tunnel->fd >= 0) {
		if (sock)
			inet_shutdown(sock, 2);
1384
	} else {
1385
		if (sock) {
1386
			kernel_sock_shutdown(sock, SHUT_RDWR);
1387 1388
			sock_release(sock);
		}
1389
	}
1390 1391

	l2tp_tunnel_sock_put(sk);
1392 1393
out:
	l2tp_tunnel_dec_refcount(tunnel);
1394 1395
}

1396 1397 1398
/* Create a socket for the tunnel, if one isn't set up by
 * userspace. This is used for static tunnels where there is no
 * managing L2TP daemon.
1399 1400 1401 1402 1403
 *
 * Since we don't want these sockets to keep a namespace alive by
 * themselves, we drop the socket's namespace refcount after creation.
 * These sockets are freed when the namespace exits using the pernet
 * exit hook.
1404
 */
1405 1406 1407 1408 1409
static int l2tp_tunnel_sock_create(struct net *net,
				u32 tunnel_id,
				u32 peer_tunnel_id,
				struct l2tp_tunnel_cfg *cfg,
				struct socket **sockp)
1410 1411
{
	int err = -EINVAL;
1412
	struct socket *sock = NULL;
T
Tom Herbert 已提交
1413
	struct udp_port_cfg udp_conf;
1414 1415 1416

	switch (cfg->encap) {
	case L2TP_ENCAPTYPE_UDP:
T
Tom Herbert 已提交
1417 1418
		memset(&udp_conf, 0, sizeof(udp_conf));

1419 1420
#if IS_ENABLED(CONFIG_IPV6)
		if (cfg->local_ip6 && cfg->peer_ip6) {
T
Tom Herbert 已提交
1421 1422 1423 1424 1425 1426
			udp_conf.family = AF_INET6;
			memcpy(&udp_conf.local_ip6, cfg->local_ip6,
			       sizeof(udp_conf.local_ip6));
			memcpy(&udp_conf.peer_ip6, cfg->peer_ip6,
			       sizeof(udp_conf.peer_ip6));
			udp_conf.use_udp6_tx_checksums =
1427
			  ! cfg->udp6_zero_tx_checksums;
T
Tom Herbert 已提交
1428
			udp_conf.use_udp6_rx_checksums =
1429
			  ! cfg->udp6_zero_rx_checksums;
1430 1431 1432
		} else
#endif
		{
T
Tom Herbert 已提交
1433 1434 1435 1436
			udp_conf.family = AF_INET;
			udp_conf.local_ip = cfg->local_ip;
			udp_conf.peer_ip = cfg->peer_ip;
			udp_conf.use_udp_checksums = cfg->use_udp_checksums;
1437
		}
1438

T
Tom Herbert 已提交
1439 1440 1441 1442 1443 1444
		udp_conf.local_udp_port = htons(cfg->local_udp_port);
		udp_conf.peer_udp_port = htons(cfg->peer_udp_port);

		err = udp_sock_create(net, &udp_conf, &sock);
		if (err < 0)
			goto out;
1445 1446 1447 1448

		break;

	case L2TP_ENCAPTYPE_IP:
1449 1450
#if IS_ENABLED(CONFIG_IPV6)
		if (cfg->local_ip6 && cfg->peer_ip6) {
T
Tom Herbert 已提交
1451 1452
			struct sockaddr_l2tpip6 ip6_addr = {0};

1453
			err = sock_create_kern(net, AF_INET6, SOCK_DGRAM,
1454
					  IPPROTO_L2TP, &sock);
1455 1456
			if (err < 0)
				goto out;
1457

1458 1459 1460 1461 1462 1463 1464 1465
			ip6_addr.l2tp_family = AF_INET6;
			memcpy(&ip6_addr.l2tp_addr, cfg->local_ip6,
			       sizeof(ip6_addr.l2tp_addr));
			ip6_addr.l2tp_conn_id = tunnel_id;
			err = kernel_bind(sock, (struct sockaddr *) &ip6_addr,
					  sizeof(ip6_addr));
			if (err < 0)
				goto out;
1466

1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478
			ip6_addr.l2tp_family = AF_INET6;
			memcpy(&ip6_addr.l2tp_addr, cfg->peer_ip6,
			       sizeof(ip6_addr.l2tp_addr));
			ip6_addr.l2tp_conn_id = peer_tunnel_id;
			err = kernel_connect(sock,
					     (struct sockaddr *) &ip6_addr,
					     sizeof(ip6_addr), 0);
			if (err < 0)
				goto out;
		} else
#endif
		{
T
Tom Herbert 已提交
1479 1480
			struct sockaddr_l2tpip ip_addr = {0};

1481
			err = sock_create_kern(net, AF_INET, SOCK_DGRAM,
1482
					  IPPROTO_L2TP, &sock);
1483 1484
			if (err < 0)
				goto out;
1485

1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501
			ip_addr.l2tp_family = AF_INET;
			ip_addr.l2tp_addr = cfg->local_ip;
			ip_addr.l2tp_conn_id = tunnel_id;
			err = kernel_bind(sock, (struct sockaddr *) &ip_addr,
					  sizeof(ip_addr));
			if (err < 0)
				goto out;

			ip_addr.l2tp_family = AF_INET;
			ip_addr.l2tp_addr = cfg->peer_ip;
			ip_addr.l2tp_conn_id = peer_tunnel_id;
			err = kernel_connect(sock, (struct sockaddr *) &ip_addr,
					     sizeof(ip_addr), 0);
			if (err < 0)
				goto out;
		}
1502 1503 1504 1505 1506 1507 1508
		break;

	default:
		goto out;
	}

out:
1509
	*sockp = sock;
1510
	if ((err < 0) && sock) {
1511
		kernel_sock_shutdown(sock, SHUT_RDWR);
1512
		sock_release(sock);
1513 1514 1515 1516 1517 1518
		*sockp = NULL;
	}

	return err;
}

E
Eric Dumazet 已提交
1519 1520
static struct lock_class_key l2tp_socket_class;

1521 1522 1523 1524 1525 1526 1527
int l2tp_tunnel_create(struct net *net, int fd, int version, u32 tunnel_id, u32 peer_tunnel_id, struct l2tp_tunnel_cfg *cfg, struct l2tp_tunnel **tunnelp)
{
	struct l2tp_tunnel *tunnel = NULL;
	int err;
	struct socket *sock = NULL;
	struct sock *sk = NULL;
	struct l2tp_net *pn;
1528
	enum l2tp_encap_type encap = L2TP_ENCAPTYPE_UDP;
1529 1530

	/* Get the tunnel socket from the fd, which was opened by
1531 1532
	 * the userspace L2TP daemon. If not specified, create a
	 * kernel socket.
1533
	 */
1534
	if (fd < 0) {
1535 1536
		err = l2tp_tunnel_sock_create(net, tunnel_id, peer_tunnel_id,
				cfg, &sock);
1537 1538 1539 1540 1541
		if (err < 0)
			goto err;
	} else {
		sock = sockfd_lookup(fd, &err);
		if (!sock) {
1542
			pr_err("tunl %u: sockfd_lookup(fd=%d) returned %d\n",
1543
			       tunnel_id, fd, err);
1544 1545 1546 1547 1548 1549 1550 1551
			err = -EBADF;
			goto err;
		}

		/* Reject namespace mismatches */
		if (!net_eq(sock_net(sock->sk), net)) {
			pr_err("tunl %u: netns mismatch\n", tunnel_id);
			err = -EINVAL;
1552 1553
			goto err;
		}
1554 1555 1556 1557
	}

	sk = sock->sk;

1558 1559 1560
	if (cfg != NULL)
		encap = cfg->encap;

1561
	/* Quick sanity checks */
1562 1563 1564 1565
	switch (encap) {
	case L2TP_ENCAPTYPE_UDP:
		err = -EPROTONOSUPPORT;
		if (sk->sk_protocol != IPPROTO_UDP) {
1566
			pr_err("tunl %hu: fd %d wrong protocol, got %d, expected %d\n",
1567 1568 1569 1570 1571 1572 1573
			       tunnel_id, fd, sk->sk_protocol, IPPROTO_UDP);
			goto err;
		}
		break;
	case L2TP_ENCAPTYPE_IP:
		err = -EPROTONOSUPPORT;
		if (sk->sk_protocol != IPPROTO_L2TP) {
1574
			pr_err("tunl %hu: fd %d wrong protocol, got %d, expected %d\n",
1575 1576 1577 1578
			       tunnel_id, fd, sk->sk_protocol, IPPROTO_L2TP);
			goto err;
		}
		break;
1579 1580 1581
	}

	/* Check if this socket has already been prepped */
1582
	tunnel = l2tp_tunnel(sk);
1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602
	if (tunnel != NULL) {
		/* This socket has already been prepped */
		err = -EBUSY;
		goto err;
	}

	tunnel = kzalloc(sizeof(struct l2tp_tunnel), GFP_KERNEL);
	if (tunnel == NULL) {
		err = -ENOMEM;
		goto err;
	}

	tunnel->version = version;
	tunnel->tunnel_id = tunnel_id;
	tunnel->peer_tunnel_id = peer_tunnel_id;
	tunnel->debug = L2TP_DEFAULT_DEBUG_FLAGS;

	tunnel->magic = L2TP_TUNNEL_MAGIC;
	sprintf(&tunnel->name[0], "tunl %u", tunnel_id);
	rwlock_init(&tunnel->hlist_lock);
1603
	tunnel->acpt_newsess = true;
1604 1605 1606 1607 1608

	/* The net we belong to */
	tunnel->l2tp_net = net;
	pn = l2tp_pernet(net);

1609
	if (cfg != NULL)
1610 1611
		tunnel->debug = cfg->debug;

1612 1613 1614 1615 1616
#if IS_ENABLED(CONFIG_IPV6)
	if (sk->sk_family == PF_INET6) {
		struct ipv6_pinfo *np = inet6_sk(sk);

		if (ipv6_addr_v4mapped(&np->saddr) &&
1617
		    ipv6_addr_v4mapped(&sk->sk_v6_daddr)) {
1618 1619 1620 1621
			struct inet_sock *inet = inet_sk(sk);

			tunnel->v4mapped = true;
			inet->inet_saddr = np->saddr.s6_addr32[3];
1622 1623
			inet->inet_rcv_saddr = sk->sk_v6_rcv_saddr.s6_addr32[3];
			inet->inet_daddr = sk->sk_v6_daddr.s6_addr32[3];
1624 1625 1626 1627 1628 1629
		} else {
			tunnel->v4mapped = false;
		}
	}
#endif

1630
	/* Mark socket as an encapsulation socket. See net/ipv4/udp.c */
1631 1632
	tunnel->encap = encap;
	if (encap == L2TP_ENCAPTYPE_UDP) {
1633
		struct udp_tunnel_sock_cfg udp_cfg = { };
1634 1635 1636 1637 1638

		udp_cfg.sk_user_data = tunnel;
		udp_cfg.encap_type = UDP_ENCAP_L2TPINUDP;
		udp_cfg.encap_rcv = l2tp_udp_encap_recv;
		udp_cfg.encap_destroy = l2tp_udp_encap_destroy;
1639

1640 1641 1642 1643
		setup_udp_tunnel_sock(net, sock, &udp_cfg);
	} else {
		sk->sk_user_data = tunnel;
	}
1644 1645 1646 1647 1648 1649 1650

	/* Hook on the tunnel socket destructor so that we can cleanup
	 * if the tunnel socket goes away.
	 */
	tunnel->old_sk_destruct = sk->sk_destruct;
	sk->sk_destruct = &l2tp_tunnel_destruct;
	tunnel->sock = sk;
1651
	tunnel->fd = fd;
E
Eric Dumazet 已提交
1652 1653
	lockdep_set_class_and_name(&sk->sk_lock.slock, &l2tp_socket_class, "l2tp_sock");

1654 1655
	sk->sk_allocation = GFP_ATOMIC;

1656 1657 1658
	/* Init delete workqueue struct */
	INIT_WORK(&tunnel->del_work, l2tp_tunnel_del_work);

1659 1660 1661 1662 1663
	/* Add tunnel to our list */
	INIT_LIST_HEAD(&tunnel->list);
	atomic_inc(&l2tp_tunnel_count);

	/* Bump the reference count. The tunnel context is deleted
E
Eric Dumazet 已提交
1664
	 * only when this drops to zero. Must be done before list insertion
1665
	 */
1666
	refcount_set(&tunnel->ref_count, 1);
E
Eric Dumazet 已提交
1667 1668 1669
	spin_lock_bh(&pn->l2tp_tunnel_list_lock);
	list_add_rcu(&tunnel->list, &pn->l2tp_tunnel_list);
	spin_unlock_bh(&pn->l2tp_tunnel_list_lock);
1670 1671 1672 1673 1674 1675

	err = 0;
err:
	if (tunnelp)
		*tunnelp = tunnel;

1676 1677 1678 1679
	/* If tunnel's socket was created by the kernel, it doesn't
	 *  have a file.
	 */
	if (sock && sock->file)
1680 1681 1682 1683 1684 1685
		sockfd_put(sock);

	return err;
}
EXPORT_SYMBOL_GPL(l2tp_tunnel_create);

1686 1687 1688 1689
/* This function is used by the netlink TUNNEL_DELETE command.
 */
int l2tp_tunnel_delete(struct l2tp_tunnel *tunnel)
{
1690 1691 1692 1693 1694 1695
	l2tp_tunnel_inc_refcount(tunnel);
	if (false == queue_work(l2tp_wq, &tunnel->del_work)) {
		l2tp_tunnel_dec_refcount(tunnel);
		return 1;
	}
	return 0;
1696 1697 1698
}
EXPORT_SYMBOL_GPL(l2tp_tunnel_delete);

1699 1700 1701 1702
/* Really kill the session.
 */
void l2tp_session_free(struct l2tp_session *session)
{
1703
	struct l2tp_tunnel *tunnel = session->tunnel;
1704

1705
	BUG_ON(refcount_read(&session->ref_count) != 0);
1706

1707
	if (tunnel) {
1708
		BUG_ON(tunnel->magic != L2TP_TUNNEL_MAGIC);
1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727
		if (session->session_id != 0)
			atomic_dec(&l2tp_session_count);
		sock_put(tunnel->sock);
		session->tunnel = NULL;
		l2tp_tunnel_dec_refcount(tunnel);
	}

	kfree(session);
}
EXPORT_SYMBOL_GPL(l2tp_session_free);

/* Remove an l2tp session from l2tp_core's hash lists.
 * Provides a tidyup interface for pseudowire code which can't just route all
 * shutdown via. l2tp_session_delete and a pseudowire-specific session_close
 * callback.
 */
void __l2tp_session_unhash(struct l2tp_session *session)
{
	struct l2tp_tunnel *tunnel = session->tunnel;
1728

1729 1730 1731
	/* Remove the session from core hashes */
	if (tunnel) {
		/* Remove from the per-tunnel hash */
1732 1733 1734 1735
		write_lock_bh(&tunnel->hlist_lock);
		hlist_del_init(&session->hlist);
		write_unlock_bh(&tunnel->hlist_lock);

1736
		/* For L2TPv3 we have a per-net hash: remove from there, too */
1737 1738
		if (tunnel->version != L2TP_HDR_VER_2) {
			struct l2tp_net *pn = l2tp_pernet(tunnel->l2tp_net);
J
James Chapman 已提交
1739 1740 1741 1742
			spin_lock_bh(&pn->l2tp_session_hlist_lock);
			hlist_del_init_rcu(&session->global_hlist);
			spin_unlock_bh(&pn->l2tp_session_hlist_lock);
			synchronize_rcu();
1743
		}
1744 1745
	}
}
1746
EXPORT_SYMBOL_GPL(__l2tp_session_unhash);
1747

1748 1749 1750 1751 1752
/* This function is used by the netlink SESSION_DELETE command and by
   pseudowire modules.
 */
int l2tp_session_delete(struct l2tp_session *session)
{
1753 1754 1755
	if (session->ref)
		(*session->ref)(session);
	__l2tp_session_unhash(session);
1756
	l2tp_session_queue_purge(session);
1757 1758
	if (session->session_close != NULL)
		(*session->session_close)(session);
1759
	if (session->deref)
1760
		(*session->deref)(session);
1761 1762 1763 1764 1765
	l2tp_session_dec_refcount(session);
	return 0;
}
EXPORT_SYMBOL_GPL(l2tp_session_delete);

1766 1767 1768
/* We come here whenever a session's send_seq, cookie_len or
 * l2specific_len parameters are set.
 */
1769
void l2tp_session_set_header_len(struct l2tp_session *session, int version)
1770 1771 1772 1773 1774 1775
{
	if (version == L2TP_HDR_VER_2) {
		session->hdr_len = 6;
		if (session->send_seq)
			session->hdr_len += 4;
	} else {
1776 1777 1778
		session->hdr_len = 4 + session->cookie_len + session->l2specific_len + session->offset;
		if (session->tunnel->encap == L2TP_ENCAPTYPE_UDP)
			session->hdr_len += 4;
1779 1780 1781
	}

}
1782
EXPORT_SYMBOL_GPL(l2tp_session_set_header_len);
1783

1784 1785 1786
struct l2tp_session *l2tp_session_create(int priv_size, struct l2tp_tunnel *tunnel, u32 session_id, u32 peer_session_id, struct l2tp_session_cfg *cfg)
{
	struct l2tp_session *session;
1787
	int err;
1788 1789 1790 1791 1792 1793 1794 1795

	session = kzalloc(sizeof(struct l2tp_session) + priv_size, GFP_KERNEL);
	if (session != NULL) {
		session->magic = L2TP_SESSION_MAGIC;
		session->tunnel = tunnel;

		session->session_id = session_id;
		session->peer_session_id = peer_session_id;
1796
		session->nr = 0;
1797 1798 1799 1800 1801
		if (tunnel->version == L2TP_HDR_VER_2)
			session->nr_max = 0xffff;
		else
			session->nr_max = 0xffffff;
		session->nr_window_size = session->nr_max / 2;
1802 1803 1804 1805
		session->nr_oos_count_max = 4;

		/* Use NR of first received packet */
		session->reorder_skip = 1;
1806 1807 1808 1809 1810 1811 1812

		sprintf(&session->name[0], "sess %u/%u",
			tunnel->tunnel_id, session->session_id);

		skb_queue_head_init(&session->reorder_q);

		INIT_HLIST_NODE(&session->hlist);
1813
		INIT_HLIST_NODE(&session->global_hlist);
1814 1815 1816 1817 1818

		/* Inherit debug options from tunnel */
		session->debug = tunnel->debug;

		if (cfg) {
1819
			session->pwtype = cfg->pw_type;
1820 1821 1822 1823 1824 1825
			session->debug = cfg->debug;
			session->mtu = cfg->mtu;
			session->mru = cfg->mru;
			session->send_seq = cfg->send_seq;
			session->recv_seq = cfg->recv_seq;
			session->lns_mode = cfg->lns_mode;
1826 1827 1828 1829 1830 1831 1832 1833
			session->reorder_timeout = cfg->reorder_timeout;
			session->offset = cfg->offset;
			session->l2specific_type = cfg->l2specific_type;
			session->l2specific_len = cfg->l2specific_len;
			session->cookie_len = cfg->cookie_len;
			memcpy(&session->cookie[0], &cfg->cookie[0], cfg->cookie_len);
			session->peer_cookie_len = cfg->peer_cookie_len;
			memcpy(&session->peer_cookie[0], &cfg->peer_cookie[0], cfg->peer_cookie_len);
1834 1835
		}

1836 1837 1838 1839 1840 1841 1842
		if (tunnel->version == L2TP_HDR_VER_2)
			session->build_header = l2tp_build_l2tpv2_header;
		else
			session->build_header = l2tp_build_l2tpv3_header;

		l2tp_session_set_header_len(session, tunnel->version);

1843 1844
		refcount_set(&session->ref_count, 1);

1845 1846 1847 1848 1849 1850 1851
		err = l2tp_session_add_to_tunnel(tunnel, session);
		if (err) {
			kfree(session);

			return ERR_PTR(err);
		}

1852 1853 1854
		/* Ignore management session in session count value */
		if (session->session_id != 0)
			atomic_inc(&l2tp_session_count);
1855 1856

		return session;
1857 1858
	}

1859
	return ERR_PTR(-ENOMEM);
1860 1861 1862 1863 1864 1865 1866 1867 1868
}
EXPORT_SYMBOL_GPL(l2tp_session_create);

/*****************************************************************************
 * Init and cleanup
 *****************************************************************************/

static __net_init int l2tp_init_net(struct net *net)
{
J
Jiri Pirko 已提交
1869
	struct l2tp_net *pn = net_generic(net, l2tp_net_id);
1870
	int hash;
1871 1872

	INIT_LIST_HEAD(&pn->l2tp_tunnel_list);
J
James Chapman 已提交
1873
	spin_lock_init(&pn->l2tp_tunnel_list_lock);
1874

1875 1876 1877
	for (hash = 0; hash < L2TP_HASH_SIZE_2; hash++)
		INIT_HLIST_HEAD(&pn->l2tp_session_hlist[hash]);

J
James Chapman 已提交
1878
	spin_lock_init(&pn->l2tp_session_hlist_lock);
1879

1880 1881 1882
	return 0;
}

1883 1884 1885 1886 1887 1888 1889 1890 1891 1892
static __net_exit void l2tp_exit_net(struct net *net)
{
	struct l2tp_net *pn = l2tp_pernet(net);
	struct l2tp_tunnel *tunnel = NULL;

	rcu_read_lock_bh();
	list_for_each_entry_rcu(tunnel, &pn->l2tp_tunnel_list, list) {
		(void)l2tp_tunnel_delete(tunnel);
	}
	rcu_read_unlock_bh();
1893 1894 1895

	flush_workqueue(l2tp_wq);
	rcu_barrier();
1896 1897
}

1898 1899
static struct pernet_operations l2tp_net_ops = {
	.init = l2tp_init_net,
1900
	.exit = l2tp_exit_net,
1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912
	.id   = &l2tp_net_id,
	.size = sizeof(struct l2tp_net),
};

static int __init l2tp_init(void)
{
	int rc = 0;

	rc = register_pernet_device(&l2tp_net_ops);
	if (rc)
		goto out;

1913
	l2tp_wq = alloc_workqueue("l2tp", WQ_UNBOUND, 0);
1914 1915
	if (!l2tp_wq) {
		pr_err("alloc_workqueue failed\n");
1916
		unregister_pernet_device(&l2tp_net_ops);
1917 1918 1919 1920
		rc = -ENOMEM;
		goto out;
	}

1921
	pr_info("L2TP core driver, %s\n", L2TP_DRV_VERSION);
1922 1923 1924 1925 1926 1927 1928 1929

out:
	return rc;
}

static void __exit l2tp_exit(void)
{
	unregister_pernet_device(&l2tp_net_ops);
1930 1931 1932 1933
	if (l2tp_wq) {
		destroy_workqueue(l2tp_wq);
		l2tp_wq = NULL;
	}
1934 1935 1936 1937 1938 1939 1940 1941 1942 1943
}

module_init(l2tp_init);
module_exit(l2tp_exit);

MODULE_AUTHOR("James Chapman <jchapman@katalix.com>");
MODULE_DESCRIPTION("L2TP core");
MODULE_LICENSE("GPL");
MODULE_VERSION(L2TP_DRV_VERSION);