xenbus.c 28.2 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0-or-later
I
Ian Campbell 已提交
2 3 4 5 6 7 8 9
/*
 * Xenbus code for netif backend
 *
 * Copyright (C) 2005 Rusty Russell <rusty@rustcorp.com.au>
 * Copyright (C) 2005 XenSource Ltd
*/

#include "common.h"
10 11
#include <linux/vmalloc.h>
#include <linux/rtnetlink.h>
I
Ian Campbell 已提交
12

13 14
static int connect_data_rings(struct backend_info *be,
			      struct xenvif_queue *queue);
15 16
static void connect(struct backend_info *be);
static int read_xenbus_vif_flags(struct backend_info *be);
17
static int backend_create_xenvif(struct backend_info *be);
I
Ian Campbell 已提交
18
static void unregister_hotplug_status_watch(struct backend_info *be);
19
static void xen_unregister_watchers(struct xenvif *vif);
20 21
static void set_backend_state(struct backend_info *be,
			      enum xenbus_state state);
I
Ian Campbell 已提交
22

23 24 25 26 27 28 29 30
#ifdef CONFIG_DEBUG_FS
struct dentry *xen_netback_dbg_root = NULL;

static int xenvif_read_io_ring(struct seq_file *m, void *v)
{
	struct xenvif_queue *queue = m->private;
	struct xen_netif_tx_back_ring *tx_ring = &queue->tx;
	struct xen_netif_rx_back_ring *rx_ring = &queue->rx;
31
	struct netdev_queue *dev_queue;
32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91

	if (tx_ring->sring) {
		struct xen_netif_tx_sring *sring = tx_ring->sring;

		seq_printf(m, "Queue %d\nTX: nr_ents %u\n", queue->id,
			   tx_ring->nr_ents);
		seq_printf(m, "req prod %u (%d) cons %u (%d) event %u (%d)\n",
			   sring->req_prod,
			   sring->req_prod - sring->rsp_prod,
			   tx_ring->req_cons,
			   tx_ring->req_cons - sring->rsp_prod,
			   sring->req_event,
			   sring->req_event - sring->rsp_prod);
		seq_printf(m, "rsp prod %u (base) pvt %u (%d) event %u (%d)\n",
			   sring->rsp_prod,
			   tx_ring->rsp_prod_pvt,
			   tx_ring->rsp_prod_pvt - sring->rsp_prod,
			   sring->rsp_event,
			   sring->rsp_event - sring->rsp_prod);
		seq_printf(m, "pending prod %u pending cons %u nr_pending_reqs %u\n",
			   queue->pending_prod,
			   queue->pending_cons,
			   nr_pending_reqs(queue));
		seq_printf(m, "dealloc prod %u dealloc cons %u dealloc_queue %u\n\n",
			   queue->dealloc_prod,
			   queue->dealloc_cons,
			   queue->dealloc_prod - queue->dealloc_cons);
	}

	if (rx_ring->sring) {
		struct xen_netif_rx_sring *sring = rx_ring->sring;

		seq_printf(m, "RX: nr_ents %u\n", rx_ring->nr_ents);
		seq_printf(m, "req prod %u (%d) cons %u (%d) event %u (%d)\n",
			   sring->req_prod,
			   sring->req_prod - sring->rsp_prod,
			   rx_ring->req_cons,
			   rx_ring->req_cons - sring->rsp_prod,
			   sring->req_event,
			   sring->req_event - sring->rsp_prod);
		seq_printf(m, "rsp prod %u (base) pvt %u (%d) event %u (%d)\n\n",
			   sring->rsp_prod,
			   rx_ring->rsp_prod_pvt,
			   rx_ring->rsp_prod_pvt - sring->rsp_prod,
			   sring->rsp_event,
			   sring->rsp_event - sring->rsp_prod);
	}

	seq_printf(m, "NAPI state: %lx NAPI weight: %d TX queue len %u\n"
		   "Credit timer_pending: %d, credit: %lu, usec: %lu\n"
		   "remaining: %lu, expires: %lu, now: %lu\n",
		   queue->napi.state, queue->napi.weight,
		   skb_queue_len(&queue->tx_queue),
		   timer_pending(&queue->credit_timeout),
		   queue->credit_bytes,
		   queue->credit_usec,
		   queue->remaining_credit,
		   queue->credit_timeout.expires,
		   jiffies);

92 93 94 95 96 97 98
	dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id);

	seq_printf(m, "\nRx internal queue: len %u max %u pkts %u %s\n",
		   queue->rx_queue_len, queue->rx_queue_max,
		   skb_queue_len(&queue->rx_queue),
		   netif_tx_queue_stopped(dev_queue) ? "stopped" : "running");

99 100 101 102
	return 0;
}

#define XENVIF_KICK_STR "kick"
103
#define BUFFER_SIZE     32
104 105 106 107 108 109 110 111

static ssize_t
xenvif_write_io_ring(struct file *filp, const char __user *buf, size_t count,
		     loff_t *ppos)
{
	struct xenvif_queue *queue =
		((struct seq_file *)filp->private_data)->private;
	int len;
112
	char write[BUFFER_SIZE];
113 114 115 116

	/* don't allow partial writes and check the length */
	if (*ppos != 0)
		return 0;
117
	if (count >= sizeof(write))
118 119 120
		return -ENOSPC;

	len = simple_write_to_buffer(write,
121
				     sizeof(write) - 1,
122 123 124 125 126 127
				     ppos,
				     buf,
				     count);
	if (len < 0)
		return len;

128 129
	write[len] = '\0';

130 131 132 133 134 135 136 137 138 139
	if (!strncmp(write, XENVIF_KICK_STR, sizeof(XENVIF_KICK_STR) - 1))
		xenvif_interrupt(0, (void *)queue);
	else {
		pr_warn("Unknown command to io_ring_q%d. Available: kick\n",
			queue->id);
		count = -EINVAL;
	}
	return count;
}

140
static int xenvif_io_ring_open(struct inode *inode, struct file *filp)
141 142 143 144 145 146 147 148 149 150 151 152 153
{
	int ret;
	void *queue = NULL;

	if (inode->i_private)
		queue = inode->i_private;
	ret = single_open(filp, xenvif_read_io_ring, queue);
	filp->f_mode |= FMODE_PWRITE;
	return ret;
}

static const struct file_operations xenvif_dbg_io_ring_ops_fops = {
	.owner = THIS_MODULE,
154
	.open = xenvif_io_ring_open,
155 156 157 158 159 160
	.read = seq_read,
	.llseek = seq_lseek,
	.release = single_release,
	.write = xenvif_write_io_ring,
};

161
static int xenvif_ctrl_show(struct seq_file *m, void *v)
162 163 164 165 166 167 168
{
	struct xenvif *vif = m->private;

	xenvif_dump_hash_info(vif, m);

	return 0;
}
169
DEFINE_SHOW_ATTRIBUTE(xenvif_ctrl);
170

171
static void xenvif_debugfs_addif(struct xenvif *vif)
172 173 174 175 176
{
	int i;

	vif->xenvif_dbg_root = debugfs_create_dir(vif->dev->name,
						  xen_netback_dbg_root);
177 178 179 180 181 182 183 184 185 186 187 188
	for (i = 0; i < vif->num_queues; ++i) {
		char filename[sizeof("io_ring_q") + 4];

		snprintf(filename, sizeof(filename), "io_ring_q%d", i);
		debugfs_create_file(filename, 0600, vif->xenvif_dbg_root,
				    &vif->queues[i],
				    &xenvif_dbg_io_ring_ops_fops);
	}

	if (vif->ctrl_irq)
		debugfs_create_file("ctrl", 0400, vif->xenvif_dbg_root, vif,
				    &xenvif_ctrl_fops);
189 190 191 192
}

static void xenvif_debugfs_delif(struct xenvif *vif)
{
193
	debugfs_remove_recursive(vif->xenvif_dbg_root);
194 195 196 197
	vif->xenvif_dbg_root = NULL;
}
#endif /* CONFIG_DEBUG_FS */

I
Ian Campbell 已提交
198 199 200 201 202 203 204 205 206 207
/*
 * Handle the creation of the hotplug script environment.  We add the script
 * and vif variables to the environment, for the benefit of the vif-* hotplug
 * scripts.
 */
static int netback_uevent(struct xenbus_device *xdev,
			  struct kobj_uevent_env *env)
{
	struct backend_info *be = dev_get_drvdata(&xdev->dev);

208 209 210 211 212 213 214
	if (!be)
		return 0;

	if (add_uevent_var(env, "script=%s", be->hotplug_script))
		return -ENOMEM;

	if (!be->vif)
I
Ian Campbell 已提交
215 216 217 218 219 220
		return 0;

	return add_uevent_var(env, "vif=%s", be->vif->dev->name);
}


221
static int backend_create_xenvif(struct backend_info *be)
I
Ian Campbell 已提交
222 223 224 225
{
	int err;
	long handle;
	struct xenbus_device *dev = be->dev;
226
	struct xenvif *vif;
I
Ian Campbell 已提交
227 228

	if (be->vif != NULL)
229
		return 0;
I
Ian Campbell 已提交
230 231 232 233

	err = xenbus_scanf(XBT_NIL, dev->nodename, "handle", "%li", &handle);
	if (err != 1) {
		xenbus_dev_fatal(dev, err, "reading handle");
234
		return (err < 0) ? err : -EINVAL;
I
Ian Campbell 已提交
235 236
	}

237 238 239
	vif = xenvif_alloc(&dev->dev, dev->otherend_id, handle);
	if (IS_ERR(vif)) {
		err = PTR_ERR(vif);
I
Ian Campbell 已提交
240
		xenbus_dev_fatal(dev, err, "creating interface");
241
		return err;
I
Ian Campbell 已提交
242
	}
243
	be->vif = vif;
244
	vif->be = be;
I
Ian Campbell 已提交
245 246

	kobject_uevent(&dev->dev.kobj, KOBJ_ONLINE);
247
	return 0;
I
Ian Campbell 已提交
248 249
}

250
static void backend_disconnect(struct backend_info *be)
I
Ian Campbell 已提交
251
{
252 253 254
	struct xenvif *vif = be->vif;

	if (vif) {
255
		unsigned int num_queues = vif->num_queues;
256 257
		unsigned int queue_index;

258
		xen_unregister_watchers(vif);
259
#ifdef CONFIG_DEBUG_FS
260
		xenvif_debugfs_delif(vif);
261
#endif /* CONFIG_DEBUG_FS */
262 263
		xenvif_disconnect_data(vif);

264 265 266
		/* At this point some of the handlers may still be active
		 * so we need to have additional synchronization here.
		 */
267
		vif->num_queues = 0;
268
		synchronize_net();
269

270 271 272 273 274
		for (queue_index = 0; queue_index < num_queues; ++queue_index)
			xenvif_deinit_queue(&vif->queues[queue_index]);

		vfree(vif->queues);
		vif->queues = NULL;
275

276
		xenvif_disconnect_ctrl(vif);
277
	}
278 279
}

280
static void backend_connect(struct backend_info *be)
281
{
282 283 284
	if (be->vif)
		connect(be);
}
285

286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302
static inline void backend_switch_state(struct backend_info *be,
					enum xenbus_state state)
{
	struct xenbus_device *dev = be->dev;

	pr_debug("%s -> %s\n", dev->nodename, xenbus_strstate(state));
	be->state = state;

	/* If we are waiting for a hotplug script then defer the
	 * actual xenbus state change.
	 */
	if (!be->have_hotplug_status_watch)
		xenbus_switch_state(dev, state);
}

/* Handle backend state transitions:
 *
303
 * The backend state starts in Initialising and the following transitions are
304 305
 * allowed.
 *
306 307 308 309 310 311 312 313 314
 * Initialising -> InitWait -> Connected
 *          \
 *           \        ^    \         |
 *            \       |     \        |
 *             \      |      \       |
 *              \     |       \      |
 *               \    |        \     |
 *                \   |         \    |
 *                 V  |          V   V
315
 *
316
 *                  Closed  <-> Closing
317 318 319 320 321 322 323 324 325
 *
 * The state argument specifies the eventual state of the backend and the
 * function transitions to that state via the shortest path.
 */
static void set_backend_state(struct backend_info *be,
			      enum xenbus_state state)
{
	while (be->state != state) {
		switch (be->state) {
326 327 328 329 330 331 332 333 334 335 336 337 338 339
		case XenbusStateInitialising:
			switch (state) {
			case XenbusStateInitWait:
			case XenbusStateConnected:
			case XenbusStateClosing:
				backend_switch_state(be, XenbusStateInitWait);
				break;
			case XenbusStateClosed:
				backend_switch_state(be, XenbusStateClosed);
				break;
			default:
				BUG();
			}
			break;
340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392
		case XenbusStateClosed:
			switch (state) {
			case XenbusStateInitWait:
			case XenbusStateConnected:
				backend_switch_state(be, XenbusStateInitWait);
				break;
			case XenbusStateClosing:
				backend_switch_state(be, XenbusStateClosing);
				break;
			default:
				BUG();
			}
			break;
		case XenbusStateInitWait:
			switch (state) {
			case XenbusStateConnected:
				backend_connect(be);
				backend_switch_state(be, XenbusStateConnected);
				break;
			case XenbusStateClosing:
			case XenbusStateClosed:
				backend_switch_state(be, XenbusStateClosing);
				break;
			default:
				BUG();
			}
			break;
		case XenbusStateConnected:
			switch (state) {
			case XenbusStateInitWait:
			case XenbusStateClosing:
			case XenbusStateClosed:
				backend_disconnect(be);
				backend_switch_state(be, XenbusStateClosing);
				break;
			default:
				BUG();
			}
			break;
		case XenbusStateClosing:
			switch (state) {
			case XenbusStateInitWait:
			case XenbusStateConnected:
			case XenbusStateClosed:
				backend_switch_state(be, XenbusStateClosed);
				break;
			default:
				BUG();
			}
			break;
		default:
			BUG();
		}
I
Ian Campbell 已提交
393 394 395 396 397 398 399 400 401 402 403
	}
}

/**
 * Callback received when the frontend's state changes.
 */
static void frontend_changed(struct xenbus_device *dev,
			     enum xenbus_state frontend_state)
{
	struct backend_info *be = dev_get_drvdata(&dev->dev);

404
	pr_debug("%s -> %s\n", dev->otherend, xenbus_strstate(frontend_state));
I
Ian Campbell 已提交
405 406 407 408 409

	be->frontend_state = frontend_state;

	switch (frontend_state) {
	case XenbusStateInitialising:
410
		set_backend_state(be, XenbusStateInitWait);
I
Ian Campbell 已提交
411 412 413 414 415 416
		break;

	case XenbusStateInitialised:
		break;

	case XenbusStateConnected:
417
		set_backend_state(be, XenbusStateConnected);
I
Ian Campbell 已提交
418 419 420
		break;

	case XenbusStateClosing:
421
		set_backend_state(be, XenbusStateClosing);
I
Ian Campbell 已提交
422 423 424
		break;

	case XenbusStateClosed:
425
		set_backend_state(be, XenbusStateClosed);
I
Ian Campbell 已提交
426 427
		if (xenbus_dev_is_online(dev))
			break;
428
		/* fall through - if not online */
I
Ian Campbell 已提交
429
	case XenbusStateUnknown:
430
		set_backend_state(be, XenbusStateClosed);
I
Ian Campbell 已提交
431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499
		device_unregister(&dev->dev);
		break;

	default:
		xenbus_dev_fatal(dev, -EINVAL, "saw state %d at frontend",
				 frontend_state);
		break;
	}
}


static void xen_net_read_rate(struct xenbus_device *dev,
			      unsigned long *bytes, unsigned long *usec)
{
	char *s, *e;
	unsigned long b, u;
	char *ratestr;

	/* Default to unlimited bandwidth. */
	*bytes = ~0UL;
	*usec = 0;

	ratestr = xenbus_read(XBT_NIL, dev->nodename, "rate", NULL);
	if (IS_ERR(ratestr))
		return;

	s = ratestr;
	b = simple_strtoul(s, &e, 10);
	if ((s == e) || (*e != ','))
		goto fail;

	s = e + 1;
	u = simple_strtoul(s, &e, 10);
	if ((s == e) || (*e != '\0'))
		goto fail;

	*bytes = b;
	*usec = u;

	kfree(ratestr);
	return;

 fail:
	pr_warn("Failed to parse network rate limit. Traffic unlimited.\n");
	kfree(ratestr);
}

static int xen_net_read_mac(struct xenbus_device *dev, u8 mac[])
{
	char *s, *e, *macstr;
	int i;

	macstr = s = xenbus_read(XBT_NIL, dev->nodename, "mac", NULL);
	if (IS_ERR(macstr))
		return PTR_ERR(macstr);

	for (i = 0; i < ETH_ALEN; i++) {
		mac[i] = simple_strtoul(s, &e, 16);
		if ((s == e) || (*e != ((i == ETH_ALEN-1) ? '\0' : ':'))) {
			kfree(macstr);
			return -ENOENT;
		}
		s = e+1;
	}

	kfree(macstr);
	return 0;
}

500
static void xen_net_rate_changed(struct xenbus_watch *watch,
501
				 const char *path, const char *token)
502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521
{
	struct xenvif *vif = container_of(watch, struct xenvif, credit_watch);
	struct xenbus_device *dev = xenvif_to_xenbus_device(vif);
	unsigned long   credit_bytes;
	unsigned long   credit_usec;
	unsigned int queue_index;

	xen_net_read_rate(dev, &credit_bytes, &credit_usec);
	for (queue_index = 0; queue_index < vif->num_queues; queue_index++) {
		struct xenvif_queue *queue = &vif->queues[queue_index];

		queue->credit_bytes = credit_bytes;
		queue->credit_usec = credit_usec;
		if (!mod_timer_pending(&queue->credit_timeout, jiffies) &&
			queue->remaining_credit > queue->credit_bytes) {
			queue->remaining_credit = queue->credit_bytes;
		}
	}
}

522 523
static int xen_register_credit_watch(struct xenbus_device *dev,
				     struct xenvif *vif)
524 525 526 527 528
{
	int err = 0;
	char *node;
	unsigned maxlen = strlen(dev->nodename) + sizeof("/rate");

529 530 531
	if (vif->credit_watch.node)
		return -EADDRINUSE;

532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547
	node = kmalloc(maxlen, GFP_KERNEL);
	if (!node)
		return -ENOMEM;
	snprintf(node, maxlen, "%s/rate", dev->nodename);
	vif->credit_watch.node = node;
	vif->credit_watch.callback = xen_net_rate_changed;
	err = register_xenbus_watch(&vif->credit_watch);
	if (err) {
		pr_err("Failed to set watcher %s\n", vif->credit_watch.node);
		kfree(node);
		vif->credit_watch.node = NULL;
		vif->credit_watch.callback = NULL;
	}
	return err;
}

548
static void xen_unregister_credit_watch(struct xenvif *vif)
549 550 551 552 553 554 555 556
{
	if (vif->credit_watch.node) {
		unregister_xenbus_watch(&vif->credit_watch);
		kfree(vif->credit_watch.node);
		vif->credit_watch.node = NULL;
	}
}

557
static void xen_mcast_ctrl_changed(struct xenbus_watch *watch,
558
				   const char *path, const char *token)
559 560 561 562 563
{
	struct xenvif *vif = container_of(watch, struct xenvif,
					  mcast_ctrl_watch);
	struct xenbus_device *dev = xenvif_to_xenbus_device(vif);

564 565
	vif->multicast_control = !!xenbus_read_unsigned(dev->otherend,
					"request-multicast-control", 0);
566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622
}

static int xen_register_mcast_ctrl_watch(struct xenbus_device *dev,
					 struct xenvif *vif)
{
	int err = 0;
	char *node;
	unsigned maxlen = strlen(dev->otherend) +
		sizeof("/request-multicast-control");

	if (vif->mcast_ctrl_watch.node) {
		pr_err_ratelimited("Watch is already registered\n");
		return -EADDRINUSE;
	}

	node = kmalloc(maxlen, GFP_KERNEL);
	if (!node) {
		pr_err("Failed to allocate memory for watch\n");
		return -ENOMEM;
	}
	snprintf(node, maxlen, "%s/request-multicast-control",
		 dev->otherend);
	vif->mcast_ctrl_watch.node = node;
	vif->mcast_ctrl_watch.callback = xen_mcast_ctrl_changed;
	err = register_xenbus_watch(&vif->mcast_ctrl_watch);
	if (err) {
		pr_err("Failed to set watcher %s\n",
		       vif->mcast_ctrl_watch.node);
		kfree(node);
		vif->mcast_ctrl_watch.node = NULL;
		vif->mcast_ctrl_watch.callback = NULL;
	}
	return err;
}

static void xen_unregister_mcast_ctrl_watch(struct xenvif *vif)
{
	if (vif->mcast_ctrl_watch.node) {
		unregister_xenbus_watch(&vif->mcast_ctrl_watch);
		kfree(vif->mcast_ctrl_watch.node);
		vif->mcast_ctrl_watch.node = NULL;
	}
}

static void xen_register_watchers(struct xenbus_device *dev,
				  struct xenvif *vif)
{
	xen_register_credit_watch(dev, vif);
	xen_register_mcast_ctrl_watch(dev, vif);
}

static void xen_unregister_watchers(struct xenvif *vif)
{
	xen_unregister_mcast_ctrl_watch(vif);
	xen_unregister_credit_watch(vif);
}

I
Ian Campbell 已提交
623 624 625 626 627 628 629 630 631 632
static void unregister_hotplug_status_watch(struct backend_info *be)
{
	if (be->have_hotplug_status_watch) {
		unregister_xenbus_watch(&be->hotplug_status_watch);
		kfree(be->hotplug_status_watch.node);
	}
	be->have_hotplug_status_watch = 0;
}

static void hotplug_status_changed(struct xenbus_watch *watch,
633 634
				   const char *path,
				   const char *token)
I
Ian Campbell 已提交
635 636 637 638 639 640 641 642 643 644 645
{
	struct backend_info *be = container_of(watch,
					       struct backend_info,
					       hotplug_status_watch);
	char *str;
	unsigned int len;

	str = xenbus_read(XBT_NIL, be->dev->nodename, "hotplug-status", &len);
	if (IS_ERR(str))
		return;
	if (len == sizeof("connected")-1 && !memcmp(str, "connected", len)) {
646 647 648
		/* Complete any pending state change */
		xenbus_switch_state(be->dev, be->state);

I
Ian Campbell 已提交
649 650 651 652 653 654
		/* Not interested in this watch anymore. */
		unregister_hotplug_status_watch(be);
	}
	kfree(str);
}

655 656 657 658 659 660 661 662 663
static int connect_ctrl_ring(struct backend_info *be)
{
	struct xenbus_device *dev = be->dev;
	struct xenvif *vif = be->vif;
	unsigned int val;
	grant_ref_t ring_ref;
	unsigned int evtchn;
	int err;

664 665 666
	err = xenbus_scanf(XBT_NIL, dev->otherend,
			   "ctrl-ring-ref", "%u", &val);
	if (err < 0)
667 668 669 670
		goto done; /* The frontend does not have a control ring */

	ring_ref = val;

671 672 673
	err = xenbus_scanf(XBT_NIL, dev->otherend,
			   "event-channel-ctrl", "%u", &val);
	if (err < 0) {
674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696
		xenbus_dev_fatal(dev, err,
				 "reading %s/event-channel-ctrl",
				 dev->otherend);
		goto fail;
	}

	evtchn = val;

	err = xenvif_connect_ctrl(vif, ring_ref, evtchn);
	if (err) {
		xenbus_dev_fatal(dev, err,
				 "mapping shared-frame %u port %u",
				 ring_ref, evtchn);
		goto fail;
	}

done:
	return 0;

fail:
	return err;
}

I
Ian Campbell 已提交
697 698 699 700
static void connect(struct backend_info *be)
{
	int err;
	struct xenbus_device *dev = be->dev;
701 702
	unsigned long credit_bytes, credit_usec;
	unsigned int queue_index;
703
	unsigned int requested_num_queues;
704
	struct xenvif_queue *queue;
I
Ian Campbell 已提交
705

706 707 708
	/* Check whether the frontend requested multiple queues
	 * and read the number requested.
	 */
709 710 711
	requested_num_queues = xenbus_read_unsigned(dev->otherend,
					"multi-queue-num-queues", 1);
	if (requested_num_queues > xenvif_max_queues) {
712
		/* buggy or malicious guest */
713
		xenbus_dev_fatal(dev, -EINVAL,
714 715 716 717 718
				 "guest requested %u queues, exceeding the maximum of %u.",
				 requested_num_queues, xenvif_max_queues);
		return;
	}

I
Ian Campbell 已提交
719 720 721 722 723 724
	err = xen_net_read_mac(dev, be->vif->fe_dev_addr);
	if (err) {
		xenbus_dev_fatal(dev, err, "parsing %s/mac", dev->nodename);
		return;
	}

725
	xen_net_read_rate(dev, &credit_bytes, &credit_usec);
726
	xen_unregister_watchers(be->vif);
727
	xen_register_watchers(dev, be->vif);
728 729
	read_xenbus_vif_flags(be);

730 731 732 733 734 735
	err = connect_ctrl_ring(be);
	if (err) {
		xenbus_dev_fatal(dev, err, "connecting control ring");
		return;
	}

736
	/* Use the number of queues requested by the frontend */
737 738
	be->vif->queues = vzalloc(array_size(requested_num_queues,
					     sizeof(struct xenvif_queue)));
739 740 741 742 743 744
	if (!be->vif->queues) {
		xenbus_dev_fatal(dev, -ENOMEM,
				 "allocating queues");
		return;
	}

745
	be->vif->num_queues = requested_num_queues;
746
	be->vif->stalled_queues = requested_num_queues;
747 748 749 750 751 752 753 754 755

	for (queue_index = 0; queue_index < requested_num_queues; ++queue_index) {
		queue = &be->vif->queues[queue_index];
		queue->vif = be->vif;
		queue->id = queue_index;
		snprintf(queue->name, sizeof(queue->name), "%s-q%u",
				be->vif->dev->name, queue->id);

		err = xenvif_init_queue(queue);
756 757 758 759 760 761 762
		if (err) {
			/* xenvif_init_queue() cleans up after itself on
			 * failure, but we need to clean up any previously
			 * initialised queues. Set num_queues to i so that
			 * earlier queues can be destroyed using the regular
			 * disconnect logic.
			 */
763
			be->vif->num_queues = queue_index;
764
			goto err;
765
		}
766

767
		queue->credit_bytes = credit_bytes;
768
		queue->remaining_credit = credit_bytes;
769
		queue->credit_usec = credit_usec;
770

771
		err = connect_data_rings(be, queue);
772
		if (err) {
773 774 775 776
			/* connect_data_rings() cleans up after itself on
			 * failure, but we need to clean up after
			 * xenvif_init_queue() here, and also clean up any
			 * previously initialised queues.
777 778
			 */
			xenvif_deinit_queue(queue);
779
			be->vif->num_queues = queue_index;
780
			goto err;
781
		}
782 783
	}

784
#ifdef CONFIG_DEBUG_FS
785
	xenvif_debugfs_addif(be->vif);
786
#endif /* CONFIG_DEBUG_FS */
787

788 789 790 791 792 793 794 795
	/* Initialisation completed, tell core driver the number of
	 * active queues.
	 */
	rtnl_lock();
	netif_set_real_num_tx_queues(be->vif->dev, requested_num_queues);
	netif_set_real_num_rx_queues(be->vif->dev, requested_num_queues);
	rtnl_unlock();

796
	xenvif_carrier_on(be->vif);
I
Ian Campbell 已提交
797 798 799 800 801

	unregister_hotplug_status_watch(be);
	err = xenbus_watch_pathfmt(dev, &be->hotplug_status_watch,
				   hotplug_status_changed,
				   "%s/%s", dev->nodename, "hotplug-status");
802
	if (!err)
I
Ian Campbell 已提交
803 804
		be->have_hotplug_status_watch = 1;

805 806 807 808 809
	netif_tx_wake_all_queues(be->vif->dev);

	return;

err:
810
	if (be->vif->num_queues > 0)
811
		xenvif_disconnect_data(be->vif); /* Clean up existing queues */
812 813
	for (queue_index = 0; queue_index < be->vif->num_queues; ++queue_index)
		xenvif_deinit_queue(&be->vif->queues[queue_index]);
814 815
	vfree(be->vif->queues);
	be->vif->queues = NULL;
816
	be->vif->num_queues = 0;
817
	xenvif_disconnect_ctrl(be->vif);
818
	return;
I
Ian Campbell 已提交
819 820 821
}


822 823
static int connect_data_rings(struct backend_info *be,
			      struct xenvif_queue *queue)
I
Ian Campbell 已提交
824 825
{
	struct xenbus_device *dev = be->dev;
826
	unsigned int num_queues = queue->vif->num_queues;
I
Ian Campbell 已提交
827
	unsigned long tx_ring_ref, rx_ring_ref;
828
	unsigned int tx_evtchn, rx_evtchn;
I
Ian Campbell 已提交
829
	int err;
830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858
	char *xspath;
	size_t xspathsize;
	const size_t xenstore_path_ext_size = 11; /* sufficient for "/queue-NNN" */

	/* If the frontend requested 1 queue, or we have fallen back
	 * to single queue due to lack of frontend support for multi-
	 * queue, expect the remaining XenStore keys in the toplevel
	 * directory. Otherwise, expect them in a subdirectory called
	 * queue-N.
	 */
	if (num_queues == 1) {
		xspath = kzalloc(strlen(dev->otherend) + 1, GFP_KERNEL);
		if (!xspath) {
			xenbus_dev_fatal(dev, -ENOMEM,
					 "reading ring references");
			return -ENOMEM;
		}
		strcpy(xspath, dev->otherend);
	} else {
		xspathsize = strlen(dev->otherend) + xenstore_path_ext_size;
		xspath = kzalloc(xspathsize, GFP_KERNEL);
		if (!xspath) {
			xenbus_dev_fatal(dev, -ENOMEM,
					 "reading ring references");
			return -ENOMEM;
		}
		snprintf(xspath, xspathsize, "%s/queue-%u", dev->otherend,
			 queue->id);
	}
I
Ian Campbell 已提交
859

860
	err = xenbus_gather(XBT_NIL, xspath,
I
Ian Campbell 已提交
861
			    "tx-ring-ref", "%lu", &tx_ring_ref,
862
			    "rx-ring-ref", "%lu", &rx_ring_ref, NULL);
I
Ian Campbell 已提交
863 864
	if (err) {
		xenbus_dev_fatal(dev, err,
865
				 "reading %s/ring-ref",
866 867
				 xspath);
		goto err;
I
Ian Campbell 已提交
868 869
	}

870
	/* Try split event channels first, then single event channel. */
871
	err = xenbus_gather(XBT_NIL, xspath,
872 873 874
			    "event-channel-tx", "%u", &tx_evtchn,
			    "event-channel-rx", "%u", &rx_evtchn, NULL);
	if (err < 0) {
875
		err = xenbus_scanf(XBT_NIL, xspath,
876 877 878 879
				   "event-channel", "%u", &tx_evtchn);
		if (err < 0) {
			xenbus_dev_fatal(dev, err,
					 "reading %s/event-channel(-tx/rx)",
880 881
					 xspath);
			goto err;
882 883 884 885
		}
		rx_evtchn = tx_evtchn;
	}

886
	/* Map the shared frame, irq etc. */
887 888
	err = xenvif_connect_data(queue, tx_ring_ref, rx_ring_ref,
				  tx_evtchn, rx_evtchn);
889 890 891 892 893
	if (err) {
		xenbus_dev_fatal(dev, err,
				 "mapping shared-frames %lu/%lu port tx %u rx %u",
				 tx_ring_ref, rx_ring_ref,
				 tx_evtchn, rx_evtchn);
894
		goto err;
895 896
	}

897 898 899 900
	err = 0;
err: /* Regular return falls through with err == 0 */
	kfree(xspath);
	return err;
901 902 903 904 905 906 907
}

static int read_xenbus_vif_flags(struct backend_info *be)
{
	struct xenvif *vif = be->vif;
	struct xenbus_device *dev = be->dev;
	unsigned int rx_copy;
908
	int err;
909

I
Ian Campbell 已提交
910 911 912 913 914 915 916 917 918 919 920 921 922 923
	err = xenbus_scanf(XBT_NIL, dev->otherend, "request-rx-copy", "%u",
			   &rx_copy);
	if (err == -ENOENT) {
		err = 0;
		rx_copy = 0;
	}
	if (err < 0) {
		xenbus_dev_fatal(dev, err, "reading %s/request-rx-copy",
				 dev->otherend);
		return err;
	}
	if (!rx_copy)
		return -EOPNOTSUPP;

924
	if (!xenbus_read_unsigned(dev->otherend, "feature-rx-notify", 0)) {
925 926 927 928 929 930
		/* - Reduce drain timeout to poll more frequently for
		 *   Rx requests.
		 * - Disable Rx stall detection.
		 */
		be->vif->drain_timeout = msecs_to_jiffies(30);
		be->vif->stall_timeout = 0;
I
Ian Campbell 已提交
931 932
	}

933
	vif->can_sg = !!xenbus_read_unsigned(dev->otherend, "feature-sg", 0);
I
Ian Campbell 已提交
934

935 936
	vif->gso_mask = 0;

937
	if (xenbus_read_unsigned(dev->otherend, "feature-gso-tcpv4", 0))
938
		vif->gso_mask |= GSO_BIT(TCPV4);
I
Ian Campbell 已提交
939

940
	if (xenbus_read_unsigned(dev->otherend, "feature-gso-tcpv6", 0))
941 942
		vif->gso_mask |= GSO_BIT(TCPV6);

943 944
	vif->ip_csum = !xenbus_read_unsigned(dev->otherend,
					     "feature-no-csum-offload", 0);
945

946 947
	vif->ipv6_csum = !!xenbus_read_unsigned(dev->otherend,
						"feature-ipv6-csum-offload", 0);
I
Ian Campbell 已提交
948 949 950 951

	return 0;
}

952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095
static int netback_remove(struct xenbus_device *dev)
{
	struct backend_info *be = dev_get_drvdata(&dev->dev);

	set_backend_state(be, XenbusStateClosed);

	unregister_hotplug_status_watch(be);
	if (be->vif) {
		kobject_uevent(&dev->dev.kobj, KOBJ_OFFLINE);
		xen_unregister_watchers(be->vif);
		xenbus_rm(XBT_NIL, dev->nodename, "hotplug-status");
		xenvif_free(be->vif);
		be->vif = NULL;
	}
	kfree(be->hotplug_script);
	kfree(be);
	dev_set_drvdata(&dev->dev, NULL);
	return 0;
}

/**
 * Entry point to this code when a new device is created.  Allocate the basic
 * structures and switch to InitWait.
 */
static int netback_probe(struct xenbus_device *dev,
			 const struct xenbus_device_id *id)
{
	const char *message;
	struct xenbus_transaction xbt;
	int err;
	int sg;
	const char *script;
	struct backend_info *be = kzalloc(sizeof(*be), GFP_KERNEL);

	if (!be) {
		xenbus_dev_fatal(dev, -ENOMEM,
				 "allocating backend structure");
		return -ENOMEM;
	}

	be->dev = dev;
	dev_set_drvdata(&dev->dev, be);

	sg = 1;

	do {
		err = xenbus_transaction_start(&xbt);
		if (err) {
			xenbus_dev_fatal(dev, err, "starting transaction");
			goto fail;
		}

		err = xenbus_printf(xbt, dev->nodename, "feature-sg", "%d", sg);
		if (err) {
			message = "writing feature-sg";
			goto abort_transaction;
		}

		err = xenbus_printf(xbt, dev->nodename, "feature-gso-tcpv4",
				    "%d", sg);
		if (err) {
			message = "writing feature-gso-tcpv4";
			goto abort_transaction;
		}

		err = xenbus_printf(xbt, dev->nodename, "feature-gso-tcpv6",
				    "%d", sg);
		if (err) {
			message = "writing feature-gso-tcpv6";
			goto abort_transaction;
		}

		/* We support partial checksum setup for IPv6 packets */
		err = xenbus_printf(xbt, dev->nodename,
				    "feature-ipv6-csum-offload",
				    "%d", 1);
		if (err) {
			message = "writing feature-ipv6-csum-offload";
			goto abort_transaction;
		}

		/* We support rx-copy path. */
		err = xenbus_printf(xbt, dev->nodename,
				    "feature-rx-copy", "%d", 1);
		if (err) {
			message = "writing feature-rx-copy";
			goto abort_transaction;
		}

		/* We don't support rx-flip path (except old guests who
		 * don't grok this feature flag).
		 */
		err = xenbus_printf(xbt, dev->nodename,
				    "feature-rx-flip", "%d", 0);
		if (err) {
			message = "writing feature-rx-flip";
			goto abort_transaction;
		}

		/* We support dynamic multicast-control. */
		err = xenbus_printf(xbt, dev->nodename,
				    "feature-multicast-control", "%d", 1);
		if (err) {
			message = "writing feature-multicast-control";
			goto abort_transaction;
		}

		err = xenbus_printf(xbt, dev->nodename,
				    "feature-dynamic-multicast-control",
				    "%d", 1);
		if (err) {
			message = "writing feature-dynamic-multicast-control";
			goto abort_transaction;
		}

		err = xenbus_transaction_end(xbt, 0);
	} while (err == -EAGAIN);

	if (err) {
		xenbus_dev_fatal(dev, err, "completing transaction");
		goto fail;
	}

	/* Split event channels support, this is optional so it is not
	 * put inside the above loop.
	 */
	err = xenbus_printf(XBT_NIL, dev->nodename,
			    "feature-split-event-channels",
			    "%u", separate_tx_rx_irq);
	if (err)
		pr_debug("Error writing feature-split-event-channels\n");

	/* Multi-queue support: This is an optional feature. */
	err = xenbus_printf(XBT_NIL, dev->nodename,
			    "multi-queue-max-queues", "%u", xenvif_max_queues);
	if (err)
		pr_debug("Error writing multi-queue-max-queues\n");

	err = xenbus_printf(XBT_NIL, dev->nodename,
			    "feature-ctrl-ring",
			    "%u", true);
	if (err)
		pr_debug("Error writing feature-ctrl-ring\n");

1096 1097
	backend_switch_state(be, XenbusStateInitWait);

1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122
	script = xenbus_read(XBT_NIL, dev->nodename, "script", NULL);
	if (IS_ERR(script)) {
		err = PTR_ERR(script);
		xenbus_dev_fatal(dev, err, "reading script");
		goto fail;
	}

	be->hotplug_script = script;

	/* This kicks hotplug scripts, so do it immediately. */
	err = backend_create_xenvif(be);
	if (err)
		goto fail;

	return 0;

abort_transaction:
	xenbus_transaction_end(xbt, 1);
	xenbus_dev_fatal(dev, err, "%s", message);
fail:
	pr_debug("failed\n");
	netback_remove(dev);
	return err;
}

I
Ian Campbell 已提交
1123 1124 1125 1126 1127
static const struct xenbus_device_id netback_ids[] = {
	{ "vif" },
	{ "" }
};

1128 1129
static struct xenbus_driver netback_driver = {
	.ids = netback_ids,
I
Ian Campbell 已提交
1130 1131 1132 1133
	.probe = netback_probe,
	.remove = netback_remove,
	.uevent = netback_uevent,
	.otherend_changed = frontend_changed,
1134
};
I
Ian Campbell 已提交
1135 1136 1137

int xenvif_xenbus_init(void)
{
1138
	return xenbus_register_backend(&netback_driver);
I
Ian Campbell 已提交
1139
}
1140 1141 1142 1143 1144

void xenvif_xenbus_fini(void)
{
	return xenbus_unregister_driver(&netback_driver);
}