fcoe.c 74.2 KB
Newer Older
1
/*
2
 * Copyright(c) 2007 - 2009 Intel Corporation. All rights reserved.
3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27
 *
 * This program is free software; you can redistribute it and/or modify it
 * under the terms and conditions of the GNU General Public License,
 * version 2, as published by the Free Software Foundation.
 *
 * This program is distributed in the hope it will be useful, but WITHOUT
 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
 * more details.
 *
 * You should have received a copy of the GNU General Public License along with
 * this program; if not, write to the Free Software Foundation, Inc.,
 * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
 *
 * Maintained at www.Open-FCoE.org
 */

#include <linux/module.h>
#include <linux/spinlock.h>
#include <linux/netdevice.h>
#include <linux/etherdevice.h>
#include <linux/ethtool.h>
#include <linux/if_ether.h>
#include <linux/if_vlan.h>
#include <linux/crc32.h>
28
#include <linux/slab.h>
29 30 31 32
#include <linux/cpu.h>
#include <linux/fs.h>
#include <linux/sysfs.h>
#include <linux/ctype.h>
33
#include <linux/workqueue.h>
34 35
#include <net/dcbnl.h>
#include <net/dcbevent.h>
36 37 38 39 40 41 42
#include <scsi/scsi_tcq.h>
#include <scsi/scsicam.h>
#include <scsi/scsi_transport.h>
#include <scsi/scsi_transport_fc.h>
#include <net/rtnetlink.h>

#include <scsi/fc/fc_encaps.h>
43
#include <scsi/fc/fc_fip.h>
44
#include <scsi/fc/fc_fcoe.h>
45 46 47 48 49

#include <scsi/libfc.h>
#include <scsi/fc_frame.h>
#include <scsi/libfcoe.h>

50
#include "fcoe.h"
51

52 53
MODULE_AUTHOR("Open-FCoE.org");
MODULE_DESCRIPTION("FCoE");
54
MODULE_LICENSE("GPL v2");
55

56
/* Performance tuning parameters for fcoe */
57
static unsigned int fcoe_ddp_min = 4096;
58 59 60 61
module_param_named(ddp_min, fcoe_ddp_min, uint, S_IRUGO | S_IWUSR);
MODULE_PARM_DESC(ddp_min, "Minimum I/O size in bytes for "	\
		 "Direct Data Placement (DDP).");

62 63 64 65 66
unsigned int fcoe_debug_logging;
module_param_named(debug_logging, fcoe_debug_logging, int, S_IRUGO|S_IWUSR);
MODULE_PARM_DESC(debug_logging, "a bit mask of logging levels");

static DEFINE_MUTEX(fcoe_config_mutex);
67

68 69
static struct workqueue_struct *fcoe_wq;

70 71 72
/* fcoe_percpu_clean completion.  Waiter protected by fcoe_create_mutex */
static DECLARE_COMPLETION(fcoe_flush_completion);

73
/* fcoe host list */
74
/* must only by accessed under the RTNL mutex */
75 76
static LIST_HEAD(fcoe_hostlist);
static DEFINE_PER_CPU(struct fcoe_percpu_s, fcoe_percpu);
77

78
/* Function Prototypes */
79
static int fcoe_reset(struct Scsi_Host *);
80 81 82
static int fcoe_xmit(struct fc_lport *, struct fc_frame *);
static int fcoe_rcv(struct sk_buff *, struct net_device *,
		    struct packet_type *, struct net_device *);
83 84
static int fcoe_percpu_receive_thread(void *);
static void fcoe_percpu_clean(struct fc_lport *);
85
static int fcoe_link_speed_update(struct fc_lport *);
86
static int fcoe_link_ok(struct fc_lport *);
87 88 89 90

static struct fc_lport *fcoe_hostlist_lookup(const struct net_device *);
static int fcoe_hostlist_add(const struct fc_lport *);

91 92 93
static int fcoe_device_notification(struct notifier_block *, ulong, void *);
static void fcoe_dev_setup(void);
static void fcoe_dev_cleanup(void);
94 95 96 97 98 99 100 101 102 103 104 105 106 107
static struct fcoe_interface
*fcoe_hostlist_lookup_port(const struct net_device *);

static int fcoe_fip_recv(struct sk_buff *, struct net_device *,
			 struct packet_type *, struct net_device *);

static void fcoe_fip_send(struct fcoe_ctlr *, struct sk_buff *);
static void fcoe_update_src_mac(struct fc_lport *, u8 *);
static u8 *fcoe_get_src_mac(struct fc_lport *);
static void fcoe_destroy_work(struct work_struct *);

static int fcoe_ddp_setup(struct fc_lport *, u16, struct scatterlist *,
			  unsigned int);
static int fcoe_ddp_done(struct fc_lport *, u16);
108 109
static int fcoe_ddp_target(struct fc_lport *, u16, struct scatterlist *,
			   unsigned int);
110
static int fcoe_cpu_callback(struct notifier_block *, unsigned long, void *);
111 112
static int fcoe_dcb_app_notification(struct notifier_block *notifier,
				     ulong event, void *ptr);
113

114 115 116 117 118
static bool fcoe_match(struct net_device *netdev);
static int fcoe_create(struct net_device *netdev, enum fip_state fip_mode);
static int fcoe_destroy(struct net_device *netdev);
static int fcoe_enable(struct net_device *netdev);
static int fcoe_disable(struct net_device *netdev);
119 120 121 122 123 124 125 126

static struct fc_seq *fcoe_elsct_send(struct fc_lport *,
				      u32 did, struct fc_frame *,
				      unsigned int op,
				      void (*resp)(struct fc_seq *,
						   struct fc_frame *,
						   void *),
				      void *, u32 timeout);
127
static void fcoe_recv_frame(struct sk_buff *skb);
128

129 130
static void fcoe_get_lesb(struct fc_lport *, struct fc_els_lesb *);

131
/* notification function for packets from net device */
132 133 134 135
static struct notifier_block fcoe_notifier = {
	.notifier_call = fcoe_device_notification,
};

136 137 138 139 140
/* notification function for CPU hotplug events */
static struct notifier_block fcoe_cpu_notifier = {
	.notifier_call = fcoe_cpu_callback,
};

141 142 143 144 145
/* notification function for DCB events */
static struct notifier_block dcb_notifier = {
	.notifier_call = fcoe_dcb_app_notification,
};

146 147
static struct scsi_transport_template *fcoe_nport_scsi_transport;
static struct scsi_transport_template *fcoe_vport_scsi_transport;
148

149 150 151 152
static int fcoe_vport_destroy(struct fc_vport *);
static int fcoe_vport_create(struct fc_vport *, bool disabled);
static int fcoe_vport_disable(struct fc_vport *, bool disable);
static void fcoe_set_vport_symbolic_name(struct fc_vport *);
153
static void fcoe_set_port_id(struct fc_lport *, u32, struct fc_frame *);
154 155 156 157 158 159 160 161 162 163 164 165 166 167 168
static void fcoe_ctlr_get_lesb(struct fcoe_ctlr_device *);
static void fcoe_fcf_get_vlan_id(struct fcoe_fcf_device *);

static struct fcoe_sysfs_function_template fcoe_sysfs_templ = {
	.get_fcoe_ctlr_mode = fcoe_ctlr_get_fip_mode,
	.get_fcoe_ctlr_link_fail = fcoe_ctlr_get_lesb,
	.get_fcoe_ctlr_vlink_fail = fcoe_ctlr_get_lesb,
	.get_fcoe_ctlr_miss_fka = fcoe_ctlr_get_lesb,
	.get_fcoe_ctlr_symb_err = fcoe_ctlr_get_lesb,
	.get_fcoe_ctlr_err_block = fcoe_ctlr_get_lesb,
	.get_fcoe_ctlr_fcs_error = fcoe_ctlr_get_lesb,

	.get_fcoe_fcf_selected = fcoe_fcf_get_selected,
	.get_fcoe_fcf_vlan_id = fcoe_fcf_get_vlan_id,
};
169 170 171 172 173

static struct libfc_function_template fcoe_libfc_fcn_templ = {
	.frame_send = fcoe_xmit,
	.ddp_setup = fcoe_ddp_setup,
	.ddp_done = fcoe_ddp_done,
174
	.ddp_target = fcoe_ddp_target,
175
	.elsct_send = fcoe_elsct_send,
176
	.get_lesb = fcoe_get_lesb,
177
	.lport_set_port_id = fcoe_set_port_id,
178
};
179

180
static struct fc_function_template fcoe_nport_fc_functions = {
181 182 183 184 185 186
	.show_host_node_name = 1,
	.show_host_port_name = 1,
	.show_host_supported_classes = 1,
	.show_host_supported_fc4s = 1,
	.show_host_active_fc4s = 1,
	.show_host_maxframe_size = 1,
187 188 189 190 191 192 193 194
	.show_host_serial_number = 1,
	.show_host_manufacturer = 1,
	.show_host_model = 1,
	.show_host_model_description = 1,
	.show_host_hardware_version = 1,
	.show_host_driver_version = 1,
	.show_host_firmware_version = 1,
	.show_host_optionrom_version = 1,
195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218

	.show_host_port_id = 1,
	.show_host_supported_speeds = 1,
	.get_host_speed = fc_get_host_speed,
	.show_host_speed = 1,
	.show_host_port_type = 1,
	.get_host_port_state = fc_get_host_port_state,
	.show_host_port_state = 1,
	.show_host_symbolic_name = 1,

	.dd_fcrport_size = sizeof(struct fc_rport_libfc_priv),
	.show_rport_maxframe_size = 1,
	.show_rport_supported_classes = 1,

	.show_host_fabric_name = 1,
	.show_starget_node_name = 1,
	.show_starget_port_name = 1,
	.show_starget_port_id = 1,
	.set_rport_dev_loss_tmo = fc_set_rport_loss_tmo,
	.show_rport_dev_loss_tmo = 1,
	.get_fc_host_stats = fc_get_host_stats,
	.issue_fc_host_lip = fcoe_reset,

	.terminate_rport_io = fc_rport_terminate_io,
219 220 221 222

	.vport_create = fcoe_vport_create,
	.vport_delete = fcoe_vport_destroy,
	.vport_disable = fcoe_vport_disable,
223
	.set_vport_symbolic_name = fcoe_set_vport_symbolic_name,
224 225

	.bsg_request = fc_lport_bsg_request,
226 227
};

228
static struct fc_function_template fcoe_vport_fc_functions = {
229 230 231 232 233 234
	.show_host_node_name = 1,
	.show_host_port_name = 1,
	.show_host_supported_classes = 1,
	.show_host_supported_fc4s = 1,
	.show_host_active_fc4s = 1,
	.show_host_maxframe_size = 1,
235 236 237 238 239 240 241 242
	.show_host_serial_number = 1,
	.show_host_manufacturer = 1,
	.show_host_model = 1,
	.show_host_model_description = 1,
	.show_host_hardware_version = 1,
	.show_host_driver_version = 1,
	.show_host_firmware_version = 1,
	.show_host_optionrom_version = 1,
243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266

	.show_host_port_id = 1,
	.show_host_supported_speeds = 1,
	.get_host_speed = fc_get_host_speed,
	.show_host_speed = 1,
	.show_host_port_type = 1,
	.get_host_port_state = fc_get_host_port_state,
	.show_host_port_state = 1,
	.show_host_symbolic_name = 1,

	.dd_fcrport_size = sizeof(struct fc_rport_libfc_priv),
	.show_rport_maxframe_size = 1,
	.show_rport_supported_classes = 1,

	.show_host_fabric_name = 1,
	.show_starget_node_name = 1,
	.show_starget_port_name = 1,
	.show_starget_port_id = 1,
	.set_rport_dev_loss_tmo = fc_set_rport_loss_tmo,
	.show_rport_dev_loss_tmo = 1,
	.get_fc_host_stats = fc_get_host_stats,
	.issue_fc_host_lip = fcoe_reset,

	.terminate_rport_io = fc_rport_terminate_io,
267 268

	.bsg_request = fc_lport_bsg_request,
269 270
};

271 272 273 274 275 276 277 278 279 280 281 282
static struct scsi_host_template fcoe_shost_template = {
	.module = THIS_MODULE,
	.name = "FCoE Driver",
	.proc_name = FCOE_NAME,
	.queuecommand = fc_queuecommand,
	.eh_abort_handler = fc_eh_abort,
	.eh_device_reset_handler = fc_eh_device_reset,
	.eh_host_reset_handler = fc_eh_host_reset,
	.slave_alloc = fc_slave_alloc,
	.change_queue_depth = fc_change_queue_depth,
	.change_queue_type = fc_change_queue_type,
	.this_id = -1,
283
	.cmd_per_lun = 3,
284 285 286 287 288 289
	.can_queue = FCOE_MAX_OUTSTANDING_COMMANDS,
	.use_clustering = ENABLE_CLUSTERING,
	.sg_tablesize = SG_ALL,
	.max_sectors = 0xffff,
};

290
/**
291 292 293
 * fcoe_interface_setup() - Setup a FCoE interface
 * @fcoe:   The new FCoE interface
 * @netdev: The net device that the fcoe interface is on
294 295
 *
 * Returns : 0 for success
296
 * Locking: must be called with the RTNL mutex held
297 298 299 300
 */
static int fcoe_interface_setup(struct fcoe_interface *fcoe,
				struct net_device *netdev)
{
301
	struct fcoe_ctlr *fip = fcoe_to_ctlr(fcoe);
302
	struct netdev_hw_addr *ha;
303
	struct net_device *real_dev;
304
	u8 flogi_maddr[ETH_ALEN];
305
	const struct net_device_ops *ops;
306 307 308

	fcoe->netdev = netdev;

309 310 311 312 313 314 315 316
	/* Let LLD initialize for FCoE */
	ops = netdev->netdev_ops;
	if (ops->ndo_fcoe_enable) {
		if (ops->ndo_fcoe_enable(netdev))
			FCOE_NETDEV_DBG(netdev, "Failed to enable FCoE"
					" specific feature for LLD.\n");
	}

317
	/* Do not support for bonding device */
J
Jiri Pirko 已提交
318
	if (netdev->priv_flags & IFF_BONDING && netdev->flags & IFF_MASTER) {
319
		FCOE_NETDEV_DBG(netdev, "Bonded interfaces not supported\n");
320 321 322 323 324
		return -EOPNOTSUPP;
	}

	/* look for SAN MAC address, if multiple SAN MACs exist, only
	 * use the first one for SPMA */
325 326
	real_dev = (netdev->priv_flags & IFF_802_1Q_VLAN) ?
		vlan_dev_real_dev(netdev) : netdev;
327
	fcoe->realdev = real_dev;
328
	rcu_read_lock();
329
	for_each_dev_addr(real_dev, ha) {
330
		if ((ha->type == NETDEV_HW_ADDR_T_SAN) &&
Y
Yi Zou 已提交
331
		    (is_valid_ether_addr(ha->addr))) {
332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348
			memcpy(fip->ctl_src_addr, ha->addr, ETH_ALEN);
			fip->spma = 1;
			break;
		}
	}
	rcu_read_unlock();

	/* setup Source Mac Address */
	if (!fip->spma)
		memcpy(fip->ctl_src_addr, netdev->dev_addr, netdev->addr_len);

	/*
	 * Add FCoE MAC address as second unicast MAC address
	 * or enter promiscuous mode if not capable of listening
	 * for multiple unicast MACs.
	 */
	memcpy(flogi_maddr, (u8[6]) FC_FCOE_FLOGI_MAC, ETH_ALEN);
349
	dev_uc_add(netdev, flogi_maddr);
350
	if (fip->spma)
351
		dev_uc_add(netdev, fip->ctl_src_addr);
352 353 354 355 356
	if (fip->mode == FIP_MODE_VN2VN) {
		dev_mc_add(netdev, FIP_ALL_VN2VN_MACS);
		dev_mc_add(netdev, FIP_ALL_P2P_MACS);
	} else
		dev_mc_add(netdev, FIP_ALL_ENODE_MACS);
357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374

	/*
	 * setup the receive function from ethernet driver
	 * on the ethertype for the given device
	 */
	fcoe->fcoe_packet_type.func = fcoe_rcv;
	fcoe->fcoe_packet_type.type = __constant_htons(ETH_P_FCOE);
	fcoe->fcoe_packet_type.dev = netdev;
	dev_add_pack(&fcoe->fcoe_packet_type);

	fcoe->fip_packet_type.func = fcoe_fip_recv;
	fcoe->fip_packet_type.type = htons(ETH_P_FIP);
	fcoe->fip_packet_type.dev = netdev;
	dev_add_pack(&fcoe->fip_packet_type);

	return 0;
}

375
/**
376 377
 * fcoe_interface_create() - Create a FCoE interface on a net device
 * @netdev: The net device to create the FCoE interface on
378
 * @fip_mode: The mode to use for FIP
379 380 381
 *
 * Returns: pointer to a struct fcoe_interface or NULL on error
 */
382 383
static struct fcoe_interface *fcoe_interface_create(struct net_device *netdev,
						    enum fip_state fip_mode)
384
{
385
	struct fcoe_ctlr_device *ctlr_dev;
386
	struct fcoe_ctlr *ctlr;
387
	struct fcoe_interface *fcoe;
388
	int size;
389
	int err;
390

391 392 393 394 395 396 397
	if (!try_module_get(THIS_MODULE)) {
		FCOE_NETDEV_DBG(netdev,
				"Could not get a reference to the module\n");
		fcoe = ERR_PTR(-EBUSY);
		goto out;
	}

398
	size = sizeof(struct fcoe_ctlr) + sizeof(struct fcoe_interface);
399 400 401 402
	ctlr_dev = fcoe_ctlr_device_add(&netdev->dev, &fcoe_sysfs_templ,
					size);
	if (!ctlr_dev) {
		FCOE_DBG("Failed to add fcoe_ctlr_device\n");
403
		fcoe = ERR_PTR(-ENOMEM);
404
		goto out_putmod;
405 406
	}

407 408 409
	ctlr = fcoe_ctlr_device_priv(ctlr_dev);
	fcoe = fcoe_ctlr_priv(ctlr);

410
	dev_hold(netdev);
411 412 413 414

	/*
	 * Initialize FIP.
	 */
415 416 417 418
	fcoe_ctlr_init(ctlr, fip_mode);
	ctlr->send = fcoe_fip_send;
	ctlr->update_mac = fcoe_update_src_mac;
	ctlr->get_src_addr = fcoe_get_src_mac;
419

420 421
	err = fcoe_interface_setup(fcoe, netdev);
	if (err) {
422
		fcoe_ctlr_destroy(ctlr);
423
		fcoe_ctlr_device_delete(ctlr_dev);
424
		dev_put(netdev);
425
		fcoe = ERR_PTR(err);
426
		goto out_putmod;
427
	}
428

429 430
	goto out;

431
out_putmod:
432 433
	module_put(THIS_MODULE);
out:
434 435 436
	return fcoe;
}

437
/**
438
 * fcoe_interface_remove() - remove FCoE interface from netdev
439
 * @fcoe: The FCoE interface to be cleaned up
440 441
 *
 * Caller must be holding the RTNL mutex
442
 */
443
static void fcoe_interface_remove(struct fcoe_interface *fcoe)
444 445
{
	struct net_device *netdev = fcoe->netdev;
446
	struct fcoe_ctlr *fip = fcoe_to_ctlr(fcoe);
447
	u8 flogi_maddr[ETH_ALEN];
448
	const struct net_device_ops *ops;
449 450 451 452 453 454 455 456 457 458 459 460 461

	/*
	 * Don't listen for Ethernet packets anymore.
	 * synchronize_net() ensures that the packet handlers are not running
	 * on another CPU. dev_remove_pack() would do that, this calls the
	 * unsyncronized version __dev_remove_pack() to avoid multiple delays.
	 */
	__dev_remove_pack(&fcoe->fcoe_packet_type);
	__dev_remove_pack(&fcoe->fip_packet_type);
	synchronize_net();

	/* Delete secondary MAC addresses */
	memcpy(flogi_maddr, (u8[6]) FC_FCOE_FLOGI_MAC, ETH_ALEN);
462
	dev_uc_del(netdev, flogi_maddr);
463
	if (fip->spma)
464
		dev_uc_del(netdev, fip->ctl_src_addr);
465 466 467 468 469
	if (fip->mode == FIP_MODE_VN2VN) {
		dev_mc_del(netdev, FIP_ALL_VN2VN_MACS);
		dev_mc_del(netdev, FIP_ALL_P2P_MACS);
	} else
		dev_mc_del(netdev, FIP_ALL_ENODE_MACS);
470 471 472 473 474 475 476 477

	/* Tell the LLD we are done w/ FCoE */
	ops = netdev->netdev_ops;
	if (ops->ndo_fcoe_disable) {
		if (ops->ndo_fcoe_disable(netdev))
			FCOE_NETDEV_DBG(netdev, "Failed to disable FCoE"
					" specific feature for LLD.\n");
	}
478 479 480 481 482 483 484 485 486 487 488
	fcoe->removed = 1;
}


/**
 * fcoe_interface_cleanup() - Clean up a FCoE interface
 * @fcoe: The FCoE interface to be cleaned up
 */
static void fcoe_interface_cleanup(struct fcoe_interface *fcoe)
{
	struct net_device *netdev = fcoe->netdev;
489
	struct fcoe_ctlr *fip = fcoe_to_ctlr(fcoe);
490
	struct fcoe_ctlr_device *ctlr_dev = fcoe_ctlr_to_ctlr_dev(fip);
491

492 493 494
	rtnl_lock();
	if (!fcoe->removed)
		fcoe_interface_remove(fcoe);
495 496
	rtnl_unlock();

497
	/* Release the self-reference taken during fcoe_interface_create() */
498 499
	/* tear-down the FCoE controller */
	fcoe_ctlr_destroy(fip);
500
	scsi_host_put(fip->lp->host);
501
	fcoe_ctlr_device_delete(ctlr_dev);
502 503
	dev_put(netdev);
	module_put(THIS_MODULE);
504 505
}

506
/**
507 508 509 510 511 512
 * fcoe_fip_recv() - Handler for received FIP frames
 * @skb:      The receive skb
 * @netdev:   The associated net device
 * @ptype:    The packet_type structure which was used to register this handler
 * @orig_dev: The original net_device the the skb was received on.
 *	      (in case dev is a bond)
513 514 515
 *
 * Returns: 0 for success
 */
516
static int fcoe_fip_recv(struct sk_buff *skb, struct net_device *netdev,
517 518 519
			 struct packet_type *ptype,
			 struct net_device *orig_dev)
{
520
	struct fcoe_interface *fcoe;
521
	struct fcoe_ctlr *ctlr;
522

523
	fcoe = container_of(ptype, struct fcoe_interface, fip_packet_type);
524 525
	ctlr = fcoe_to_ctlr(fcoe);
	fcoe_ctlr_recv(ctlr, skb);
526 527 528
	return 0;
}

529 530 531 532 533 534 535 536 537 538 539 540 541
/**
 * fcoe_port_send() - Send an Ethernet-encapsulated FIP/FCoE frame
 * @port: The FCoE port
 * @skb: The FIP/FCoE packet to be sent
 */
static void fcoe_port_send(struct fcoe_port *port, struct sk_buff *skb)
{
	if (port->fcoe_pending_queue.qlen)
		fcoe_check_wait_queue(port->lport, skb);
	else if (fcoe_start_io(skb))
		fcoe_check_wait_queue(port->lport, skb);
}

542
/**
543 544 545
 * fcoe_fip_send() - Send an Ethernet-encapsulated FIP frame
 * @fip: The FCoE controller
 * @skb: The FIP packet to be sent
546 547 548
 */
static void fcoe_fip_send(struct fcoe_ctlr *fip, struct sk_buff *skb)
{
549
	skb->dev = fcoe_from_ctlr(fip)->netdev;
550
	fcoe_port_send(lport_priv(fip->lp), skb);
551 552 553
}

/**
554 555 556
 * fcoe_update_src_mac() - Update the Ethernet MAC filters
 * @lport: The local port to update the source MAC on
 * @addr:  Unicast MAC address to add
557 558 559 560
 *
 * Remove any previously-set unicast MAC filter.
 * Add secondary FCoE MAC address filter for our OUI.
 */
561
static void fcoe_update_src_mac(struct fc_lport *lport, u8 *addr)
562
{
563
	struct fcoe_port *port = lport_priv(lport);
564
	struct fcoe_interface *fcoe = port->priv;
565

566
	if (!is_zero_ether_addr(port->data_src_addr))
567
		dev_uc_del(fcoe->netdev, port->data_src_addr);
568
	if (!is_zero_ether_addr(addr))
569
		dev_uc_add(fcoe->netdev, addr);
570
	memcpy(port->data_src_addr, addr, ETH_ALEN);
571 572
}

573 574 575 576 577 578 579 580 581 582 583
/**
 * fcoe_get_src_mac() - return the Ethernet source address for an lport
 * @lport: libfc lport
 */
static u8 *fcoe_get_src_mac(struct fc_lport *lport)
{
	struct fcoe_port *port = lport_priv(lport);

	return port->data_src_addr;
}

584
/**
585 586
 * fcoe_lport_config() - Set up a local port
 * @lport: The local port to be setup
587 588 589
 *
 * Returns: 0 for success
 */
590
static int fcoe_lport_config(struct fc_lport *lport)
591
{
592 593 594 595 596 597 598 599 600 601 602
	lport->link_up = 0;
	lport->qfull = 0;
	lport->max_retry_count = 3;
	lport->max_rport_retry_count = 3;
	lport->e_d_tov = 2 * 1000;	/* FC-FS default */
	lport->r_a_tov = 2 * 2 * 1000;
	lport->service_params = (FCP_SPPF_INIT_FCN | FCP_SPPF_RD_XRDY_DIS |
				 FCP_SPPF_RETRY | FCP_SPPF_CONF_COMPL);
	lport->does_npiv = 1;

	fc_lport_init_stats(lport);
603 604

	/* lport fc_lport related configuration */
605
	fc_lport_config(lport);
606 607

	/* offload related configuration */
608 609 610 611 612
	lport->crc_offload = 0;
	lport->seq_offload = 0;
	lport->lro_enabled = 0;
	lport->lro_xid = 0;
	lport->lso_max = 0;
613 614 615 616

	return 0;
}

617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660
/**
 * fcoe_netdev_features_change - Updates the lport's offload flags based
 * on the LLD netdev's FCoE feature flags
 */
static void fcoe_netdev_features_change(struct fc_lport *lport,
					struct net_device *netdev)
{
	mutex_lock(&lport->lp_mutex);

	if (netdev->features & NETIF_F_SG)
		lport->sg_supp = 1;
	else
		lport->sg_supp = 0;

	if (netdev->features & NETIF_F_FCOE_CRC) {
		lport->crc_offload = 1;
		FCOE_NETDEV_DBG(netdev, "Supports FCCRC offload\n");
	} else {
		lport->crc_offload = 0;
	}

	if (netdev->features & NETIF_F_FSO) {
		lport->seq_offload = 1;
		lport->lso_max = netdev->gso_max_size;
		FCOE_NETDEV_DBG(netdev, "Supports LSO for max len 0x%x\n",
				lport->lso_max);
	} else {
		lport->seq_offload = 0;
		lport->lso_max = 0;
	}

	if (netdev->fcoe_ddp_xid) {
		lport->lro_enabled = 1;
		lport->lro_xid = netdev->fcoe_ddp_xid;
		FCOE_NETDEV_DBG(netdev, "Supports LRO for max xid 0x%x\n",
				lport->lro_xid);
	} else {
		lport->lro_enabled = 0;
		lport->lro_xid = 0;
	}

	mutex_unlock(&lport->lp_mutex);
}

661
/**
662 663 664
 * fcoe_netdev_config() - Set up net devive for SW FCoE
 * @lport:  The local port that is associated with the net device
 * @netdev: The associated net device
665
 *
666
 * Must be called after fcoe_lport_config() as it will use local port mutex
667
 *
668
 * Returns: 0 for success
669
 */
670
static int fcoe_netdev_config(struct fc_lport *lport, struct net_device *netdev)
671 672 673
{
	u32 mfs;
	u64 wwnn, wwpn;
674
	struct fcoe_interface *fcoe;
675
	struct fcoe_ctlr *ctlr;
676
	struct fcoe_port *port;
677 678

	/* Setup lport private data to point to fcoe softc */
679
	port = lport_priv(lport);
680
	fcoe = port->priv;
681
	ctlr = fcoe_to_ctlr(fcoe);
682 683 684 685 686 687

	/*
	 * Determine max frame size based on underlying device and optional
	 * user-configured limit.  If the MFS is too low, fcoe_link_ok()
	 * will return 0, so do this first.
	 */
688 689 690 691 692 693
	mfs = netdev->mtu;
	if (netdev->features & NETIF_F_FCOE_MTU) {
		mfs = FCOE_MTU;
		FCOE_NETDEV_DBG(netdev, "Supports FCOE_MTU of %d bytes\n", mfs);
	}
	mfs -= (sizeof(struct fcoe_hdr) + sizeof(struct fcoe_crc_eof));
694
	if (fc_set_mfs(lport, mfs))
695 696 697
		return -EINVAL;

	/* offload features support */
698
	fcoe_netdev_features_change(lport, netdev);
699

700 701
	skb_queue_head_init(&port->fcoe_pending_queue);
	port->fcoe_pending_queue_active = 0;
702
	setup_timer(&port->timer, fcoe_queue_timer, (unsigned long)lport);
703

704 705
	fcoe_link_speed_update(lport);

706
	if (!lport->vport) {
707
		if (fcoe_get_wwn(netdev, &wwnn, NETDEV_FCOE_WWNN))
708
			wwnn = fcoe_wwn_from_mac(ctlr->ctl_src_addr, 1, 0);
709
		fc_set_wwnn(lport, wwnn);
710
		if (fcoe_get_wwn(netdev, &wwpn, NETDEV_FCOE_WWPN))
711
			wwpn = fcoe_wwn_from_mac(ctlr->ctl_src_addr,
712
						 2, 0);
713
		fc_set_wwpn(lport, wwpn);
714
	}
715 716 717 718 719

	return 0;
}

/**
720 721 722
 * fcoe_shost_config() - Set up the SCSI host associated with a local port
 * @lport: The local port
 * @dev:   The device associated with the SCSI host
723 724 725
 *
 * Must be called after fcoe_lport_config() and fcoe_netdev_config()
 *
726
 * Returns: 0 for success
727
 */
728
static int fcoe_shost_config(struct fc_lport *lport, struct device *dev)
729 730 731 732
{
	int rc = 0;

	/* lport scsi host config */
733 734 735
	lport->host->max_lun = FCOE_MAX_LUN;
	lport->host->max_id = FCOE_MAX_FCP_TARGET;
	lport->host->max_channel = 0;
736 737
	lport->host->max_cmd_len = FCOE_MAX_CMD_LEN;

738
	if (lport->vport)
739
		lport->host->transportt = fcoe_vport_scsi_transport;
740
	else
741
		lport->host->transportt = fcoe_nport_scsi_transport;
742 743

	/* add the new host to the SCSI-ml */
744
	rc = scsi_add_host(lport->host, dev);
745
	if (rc) {
746
		FCOE_NETDEV_DBG(fcoe_netdev(lport), "fcoe_shost_config: "
747
				"error on scsi_add_host\n");
748 749
		return rc;
	}
750

751
	if (!lport->vport)
752
		fc_host_max_npiv_vports(lport->host) = USHRT_MAX;
753

754
	snprintf(fc_host_symbolic_name(lport->host), FC_SYMBOLIC_NAME_SIZE,
755
		 "%s v%s over %s", FCOE_NAME, FCOE_VERSION,
756
		 fcoe_netdev(lport)->name);
757 758 759 760

	return 0;
}

761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839

/**
 * fcoe_fdmi_info() - Get FDMI related info from net devive for SW FCoE
 * @lport:  The local port that is associated with the net device
 * @netdev: The associated net device
 *
 * Must be called after fcoe_shost_config() as it will use local port mutex
 *
 */
static void fcoe_fdmi_info(struct fc_lport *lport, struct net_device *netdev)
{
	struct fcoe_interface *fcoe;
	struct fcoe_port *port;
	struct net_device *realdev;
	int rc;
	struct netdev_fcoe_hbainfo fdmi;

	port = lport_priv(lport);
	fcoe = port->priv;
	realdev = fcoe->realdev;

	if (!realdev)
		return;

	/* No FDMI state m/c for NPIV ports */
	if (lport->vport)
		return;

	if (realdev->netdev_ops->ndo_fcoe_get_hbainfo) {
		memset(&fdmi, 0, sizeof(fdmi));
		rc = realdev->netdev_ops->ndo_fcoe_get_hbainfo(realdev,
							       &fdmi);
		if (rc) {
			printk(KERN_INFO "fcoe: Failed to retrieve FDMI "
					"information from netdev.\n");
			return;
		}

		snprintf(fc_host_serial_number(lport->host),
			 FC_SERIAL_NUMBER_SIZE,
			 "%s",
			 fdmi.serial_number);
		snprintf(fc_host_manufacturer(lport->host),
			 FC_SERIAL_NUMBER_SIZE,
			 "%s",
			 fdmi.manufacturer);
		snprintf(fc_host_model(lport->host),
			 FC_SYMBOLIC_NAME_SIZE,
			 "%s",
			 fdmi.model);
		snprintf(fc_host_model_description(lport->host),
			 FC_SYMBOLIC_NAME_SIZE,
			 "%s",
			 fdmi.model_description);
		snprintf(fc_host_hardware_version(lport->host),
			 FC_VERSION_STRING_SIZE,
			 "%s",
			 fdmi.hardware_version);
		snprintf(fc_host_driver_version(lport->host),
			 FC_VERSION_STRING_SIZE,
			 "%s",
			 fdmi.driver_version);
		snprintf(fc_host_optionrom_version(lport->host),
			 FC_VERSION_STRING_SIZE,
			 "%s",
			 fdmi.optionrom_version);
		snprintf(fc_host_firmware_version(lport->host),
			 FC_VERSION_STRING_SIZE,
			 "%s",
			 fdmi.firmware_version);

		/* Enable FDMI lport states */
		lport->fdmi_enabled = 1;
	} else {
		lport->fdmi_enabled = 0;
		printk(KERN_INFO "fcoe: No FDMI support.\n");
	}
}

840 841 842
/**
 * fcoe_oem_match() - The match routine for the offloaded exchange manager
 * @fp: The I/O frame
843
 *
844 845 846 847 848 849 850 851 852
 * This routine will be associated with an exchange manager (EM). When
 * the libfc exchange handling code is looking for an EM to use it will
 * call this routine and pass it the frame that it wishes to send. This
 * routine will return True if the associated EM is to be used and False
 * if the echange code should continue looking for an EM.
 *
 * The offload EM that this routine is associated with will handle any
 * packets that are for SCSI read requests.
 *
853 854 855
 * This has been enhanced to work when FCoE stack is operating in target
 * mode.
 *
856
 * Returns: True for read types I/O, otherwise returns false.
857
 */
858
static bool fcoe_oem_match(struct fc_frame *fp)
859
{
860 861 862 863 864 865
	struct fc_frame_header *fh = fc_frame_header_get(fp);
	struct fcp_cmnd *fcp;

	if (fc_fcp_is_read(fr_fsp(fp)) &&
	    (fr_fsp(fp)->data_len > fcoe_ddp_min))
		return true;
866 867 868
	else if ((fr_fsp(fp) == NULL) &&
		 (fh->fh_r_ctl == FC_RCTL_DD_UNSOL_CMD) &&
		 (ntohs(fh->fh_rx_id) == FC_XID_UNKNOWN)) {
869
		fcp = fc_frame_payload_get(fp, sizeof(*fcp));
870 871
		if ((fcp->fc_flags & FCP_CFL_WRDATA) &&
		    (ntohl(fcp->fc_dl) > fcoe_ddp_min))
872 873 874
			return true;
	}
	return false;
875 876
}

877
/**
878 879
 * fcoe_em_config() - Allocate and configure an exchange manager
 * @lport: The local port that the new EM will be associated with
880
 *
881
 * Returns: 0 on success
882
 */
883
static inline int fcoe_em_config(struct fc_lport *lport)
884
{
885
	struct fcoe_port *port = lport_priv(lport);
886
	struct fcoe_interface *fcoe = port->priv;
887
	struct fcoe_interface *oldfcoe = NULL;
888
	struct net_device *old_real_dev, *cur_real_dev;
889 890 891 892 893 894 895
	u16 min_xid = FCOE_MIN_XID;
	u16 max_xid = FCOE_MAX_XID;

	/*
	 * Check if need to allocate an em instance for
	 * offload exchange ids to be shared across all VN_PORTs/lport.
	 */
896 897 898
	if (!lport->lro_enabled || !lport->lro_xid ||
	    (lport->lro_xid >= max_xid)) {
		lport->lro_xid = 0;
899 900 901 902 903
		goto skip_oem;
	}

	/*
	 * Reuse existing offload em instance in case
904
	 * it is already allocated on real eth device
905
	 */
906 907
	if (fcoe->netdev->priv_flags & IFF_802_1Q_VLAN)
		cur_real_dev = vlan_dev_real_dev(fcoe->netdev);
908
	else
909
		cur_real_dev = fcoe->netdev;
910

911 912 913
	list_for_each_entry(oldfcoe, &fcoe_hostlist, list) {
		if (oldfcoe->netdev->priv_flags & IFF_802_1Q_VLAN)
			old_real_dev = vlan_dev_real_dev(oldfcoe->netdev);
914
		else
915
			old_real_dev = oldfcoe->netdev;
916 917

		if (cur_real_dev == old_real_dev) {
918
			fcoe->oem = oldfcoe->oem;
919 920 921 922
			break;
		}
	}

923
	if (fcoe->oem) {
924
		if (!fc_exch_mgr_add(lport, fcoe->oem, fcoe_oem_match)) {
925 926
			printk(KERN_ERR "fcoe_em_config: failed to add "
			       "offload em:%p on interface:%s\n",
927
			       fcoe->oem, fcoe->netdev->name);
928 929 930
			return -ENOMEM;
		}
	} else {
931 932 933
		fcoe->oem = fc_exch_mgr_alloc(lport, FC_CLASS_3,
					      FCOE_MIN_XID, lport->lro_xid,
					      fcoe_oem_match);
934
		if (!fcoe->oem) {
935 936
			printk(KERN_ERR "fcoe_em_config: failed to allocate "
			       "em for offload exches on interface:%s\n",
937
			       fcoe->netdev->name);
938 939 940 941 942 943 944
			return -ENOMEM;
		}
	}

	/*
	 * Exclude offload EM xid range from next EM xid range.
	 */
945
	min_xid += lport->lro_xid + 1;
946 947

skip_oem:
948
	if (!fc_exch_mgr_alloc(lport, FC_CLASS_3, min_xid, max_xid, NULL)) {
949
		printk(KERN_ERR "fcoe_em_config: failed to "
950
		       "allocate em on interface %s\n", fcoe->netdev->name);
951
		return -ENOMEM;
952
	}
953 954 955 956 957

	return 0;
}

/**
958 959
 * fcoe_if_destroy() - Tear down a SW FCoE instance
 * @lport: The local port to be destroyed
960
 *
961
 */
962
static void fcoe_if_destroy(struct fc_lport *lport)
963
{
964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984
	struct fcoe_port *port = lport_priv(lport);
	struct fcoe_interface *fcoe = port->priv;
	struct net_device *netdev = fcoe->netdev;

	FCOE_NETDEV_DBG(netdev, "Destroying interface\n");

	/* Logout of the fabric */
	fc_fabric_logoff(lport);

	/* Cleanup the fc_lport */
	fc_lport_destroy(lport);

	/* Stop the transmit retry timer */
	del_timer_sync(&port->timer);

	/* Free existing transmit skbs */
	fcoe_clean_pending_queue(lport);

	rtnl_lock();
	if (!is_zero_ether_addr(port->data_src_addr))
		dev_uc_del(netdev, port->data_src_addr);
985 986 987 988
	if (lport->vport)
		synchronize_net();
	else
		fcoe_interface_remove(fcoe);
989 990
	rtnl_unlock();

991
	/* Free queued packets for the per-CPU receive threads */
992
	fcoe_percpu_clean(lport);
993

994
	/* Detach from the scsi-ml */
995 996
	fc_remove_host(lport->host);
	scsi_remove_host(lport->host);
997

998 999 1000
	/* Destroy lport scsi_priv */
	fc_fcp_destroy(lport);

1001
	/* There are no more rports or I/O, free the EM */
1002
	fc_exch_mgr_free(lport);
1003 1004

	/* Free memory used by statistical counters */
1005
	fc_lport_free_stats(lport);
1006

1007 1008 1009 1010 1011 1012
	/*
	 * Release the Scsi_Host for vport but hold on to
	 * master lport until it fcoe interface fully cleaned-up.
	 */
	if (lport->vport)
		scsi_host_put(lport->host);
1013 1014
}

1015 1016 1017 1018 1019 1020
/**
 * fcoe_ddp_setup() - Call a LLD's ddp_setup through the net device
 * @lport: The local port to setup DDP for
 * @xid:   The exchange ID for this DDP transfer
 * @sgl:   The scatterlist describing this transfer
 * @sgc:   The number of sg items
1021
 *
1022
 * Returns: 0 if the DDP context was not configured
1023
 */
1024 1025
static int fcoe_ddp_setup(struct fc_lport *lport, u16 xid,
			  struct scatterlist *sgl, unsigned int sgc)
1026
{
1027
	struct net_device *netdev = fcoe_netdev(lport);
1028

1029 1030 1031 1032
	if (netdev->netdev_ops->ndo_fcoe_ddp_setup)
		return netdev->netdev_ops->ndo_fcoe_ddp_setup(netdev,
							      xid, sgl,
							      sgc);
1033 1034 1035 1036

	return 0;
}

1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058
/**
 * fcoe_ddp_target() - Call a LLD's ddp_target through the net device
 * @lport: The local port to setup DDP for
 * @xid:   The exchange ID for this DDP transfer
 * @sgl:   The scatterlist describing this transfer
 * @sgc:   The number of sg items
 *
 * Returns: 0 if the DDP context was not configured
 */
static int fcoe_ddp_target(struct fc_lport *lport, u16 xid,
			   struct scatterlist *sgl, unsigned int sgc)
{
	struct net_device *netdev = fcoe_netdev(lport);

	if (netdev->netdev_ops->ndo_fcoe_ddp_target)
		return netdev->netdev_ops->ndo_fcoe_ddp_target(netdev, xid,
							       sgl, sgc);

	return 0;
}


1059 1060 1061 1062
/**
 * fcoe_ddp_done() - Call a LLD's ddp_done through the net device
 * @lport: The local port to complete DDP on
 * @xid:   The exchange ID for this DDP transfer
1063
 *
1064
 * Returns: the length of data that have been completed by DDP
1065
 */
1066
static int fcoe_ddp_done(struct fc_lport *lport, u16 xid)
1067
{
1068
	struct net_device *netdev = fcoe_netdev(lport);
1069

1070 1071
	if (netdev->netdev_ops->ndo_fcoe_ddp_done)
		return netdev->netdev_ops->ndo_fcoe_ddp_done(netdev, xid);
1072 1073 1074 1075
	return 0;
}

/**
1076 1077 1078 1079
 * fcoe_if_create() - Create a FCoE instance on an interface
 * @fcoe:   The FCoE interface to create a local port on
 * @parent: The device pointer to be the parent in sysfs for the SCSI host
 * @npiv:   Indicates if the port is a vport or not
1080
 *
1081
 * Creates a fc_lport instance and a Scsi_Host instance and configure them.
1082
 *
1083
 * Returns: The allocated fc_lport or an error pointer
1084
 */
1085
static struct fc_lport *fcoe_if_create(struct fcoe_interface *fcoe,
1086
				       struct device *parent, int npiv)
1087
{
1088
	struct fcoe_ctlr *ctlr = fcoe_to_ctlr(fcoe);
1089
	struct net_device *netdev = fcoe->netdev;
1090
	struct fc_lport *lport, *n_port;
1091
	struct fcoe_port *port;
1092
	struct Scsi_Host *shost;
1093
	int rc;
1094 1095 1096 1097 1098
	/*
	 * parent is only a vport if npiv is 1,
	 * but we'll only use vport in that case so go ahead and set it
	 */
	struct fc_vport *vport = dev_to_vport(parent);
1099

1100
	FCOE_NETDEV_DBG(netdev, "Create Interface\n");
1101

1102 1103 1104 1105 1106
	if (!npiv)
		lport = libfc_host_alloc(&fcoe_shost_template, sizeof(*port));
	else
		lport = libfc_vport_create(vport, sizeof(*port));

1107
	if (!lport) {
1108
		FCOE_NETDEV_DBG(netdev, "Could not allocate host structure\n");
1109
		rc = -ENOMEM;
1110
		goto out;
1111
	}
1112
	port = lport_priv(lport);
1113
	port->lport = lport;
1114 1115 1116
	port->priv = fcoe;
	port->max_queue_depth = FCOE_MAX_QUEUE_DEPTH;
	port->min_queue_depth = FCOE_MIN_QUEUE_DEPTH;
1117
	INIT_WORK(&port->destroy_work, fcoe_destroy_work);
1118

1119
	/* configure a fc_lport including the exchange manager */
1120
	rc = fcoe_lport_config(lport);
1121
	if (rc) {
1122 1123
		FCOE_NETDEV_DBG(netdev, "Could not configure lport for the "
				"interface\n");
1124 1125 1126
		goto out_host_put;
	}

1127
	if (npiv) {
1128 1129
		FCOE_NETDEV_DBG(netdev, "Setting vport names, "
				"%16.16llx %16.16llx\n",
1130
				vport->node_name, vport->port_name);
1131 1132 1133 1134
		fc_set_wwnn(lport, vport->node_name);
		fc_set_wwpn(lport, vport->port_name);
	}

1135
	/* configure lport network properties */
1136
	rc = fcoe_netdev_config(lport, netdev);
1137
	if (rc) {
1138 1139
		FCOE_NETDEV_DBG(netdev, "Could not configure netdev for the "
				"interface\n");
1140
		goto out_lp_destroy;
1141
	}
1142

1143
	/* configure lport scsi host properties */
1144
	rc = fcoe_shost_config(lport, parent);
1145
	if (rc) {
1146 1147
		FCOE_NETDEV_DBG(netdev, "Could not configure shost for the "
				"interface\n");
1148
		goto out_lp_destroy;
1149 1150
	}

1151
	/* Initialize the library */
1152
	rc = fcoe_libfc_config(lport, ctlr, &fcoe_libfc_fcn_templ, 1);
1153
	if (rc) {
1154
		FCOE_NETDEV_DBG(netdev, "Could not configure libfc for the "
1155
				"interface\n");
1156
		goto out_lp_destroy;
1157 1158
	}

1159 1160 1161
	/* Initialized FDMI information */
	fcoe_fdmi_info(lport, netdev);

1162 1163 1164 1165 1166 1167 1168 1169 1170 1171
	/*
	 * fcoe_em_alloc() and fcoe_hostlist_add() both
	 * need to be atomic with respect to other changes to the
	 * hostlist since fcoe_em_alloc() looks for an existing EM
	 * instance on host list updated by fcoe_hostlist_add().
	 *
	 * This is currently handled through the fcoe_config_mutex
	 * begin held.
	 */
	if (!npiv)
1172 1173
		/* lport exch manager allocation */
		rc = fcoe_em_config(lport);
1174 1175 1176 1177 1178 1179 1180 1181 1182
	else {
		shost = vport_to_shost(vport);
		n_port = shost_priv(shost);
		rc = fc_exch_mgr_list_clone(n_port, lport);
	}

	if (rc) {
		FCOE_NETDEV_DBG(netdev, "Could not configure the EM\n");
		goto out_lp_destroy;
1183 1184
	}

1185
	return lport;
1186 1187

out_lp_destroy:
1188
	fc_exch_mgr_free(lport);
1189
out_host_put:
1190 1191 1192
	scsi_host_put(lport->host);
out:
	return ERR_PTR(rc);
1193 1194 1195
}

/**
1196
 * fcoe_if_init() - Initialization routine for fcoe.ko
1197
 *
1198 1199 1200
 * Attaches the SW FCoE transport to the FC transport
 *
 * Returns: 0 on success
1201 1202 1203 1204
 */
static int __init fcoe_if_init(void)
{
	/* attach to scsi transport */
1205 1206 1207 1208
	fcoe_nport_scsi_transport =
		fc_attach_transport(&fcoe_nport_fc_functions);
	fcoe_vport_scsi_transport =
		fc_attach_transport(&fcoe_vport_fc_functions);
1209

1210
	if (!fcoe_nport_scsi_transport) {
1211
		printk(KERN_ERR "fcoe: Failed to attach to the FC transport\n");
1212 1213 1214 1215 1216 1217 1218
		return -ENODEV;
	}

	return 0;
}

/**
1219 1220 1221
 * fcoe_if_exit() - Tear down fcoe.ko
 *
 * Detaches the SW FCoE transport from the FC transport
1222
 *
1223
 * Returns: 0 on success
1224
 */
1225
static int __exit fcoe_if_exit(void)
1226
{
1227 1228 1229 1230
	fc_release_transport(fcoe_nport_scsi_transport);
	fc_release_transport(fcoe_vport_scsi_transport);
	fcoe_nport_scsi_transport = NULL;
	fcoe_vport_scsi_transport = NULL;
1231 1232 1233
	return 0;
}

1234
/**
1235 1236
 * fcoe_percpu_thread_create() - Create a receive thread for an online CPU
 * @cpu: The CPU index of the CPU to create a receive thread for
1237 1238 1239 1240 1241 1242 1243 1244
 */
static void fcoe_percpu_thread_create(unsigned int cpu)
{
	struct fcoe_percpu_s *p;
	struct task_struct *thread;

	p = &per_cpu(fcoe_percpu, cpu);

1245 1246 1247
	thread = kthread_create_on_node(fcoe_percpu_receive_thread,
					(void *)p, cpu_to_node(cpu),
					"fcoethread/%d", cpu);
1248

1249
	if (likely(!IS_ERR(thread))) {
1250 1251 1252 1253 1254 1255 1256 1257 1258 1259
		kthread_bind(thread, cpu);
		wake_up_process(thread);

		spin_lock_bh(&p->fcoe_rx_list.lock);
		p->thread = thread;
		spin_unlock_bh(&p->fcoe_rx_list.lock);
	}
}

/**
1260 1261
 * fcoe_percpu_thread_destroy() - Remove the receive thread of a CPU
 * @cpu: The CPU index of the CPU whose receive thread is to be destroyed
1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274
 *
 * Destroys a per-CPU Rx thread. Any pending skbs are moved to the
 * current CPU's Rx thread. If the thread being destroyed is bound to
 * the CPU processing this context the skbs will be freed.
 */
static void fcoe_percpu_thread_destroy(unsigned int cpu)
{
	struct fcoe_percpu_s *p;
	struct task_struct *thread;
	struct page *crc_eof;
	struct sk_buff *skb;
#ifdef CONFIG_SMP
	struct fcoe_percpu_s *p0;
1275
	unsigned targ_cpu = get_cpu();
1276 1277
#endif /* CONFIG_SMP */

1278
	FCOE_DBG("Destroying receive thread for CPU %d\n", cpu);
1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299

	/* Prevent any new skbs from being queued for this CPU. */
	p = &per_cpu(fcoe_percpu, cpu);
	spin_lock_bh(&p->fcoe_rx_list.lock);
	thread = p->thread;
	p->thread = NULL;
	crc_eof = p->crc_eof_page;
	p->crc_eof_page = NULL;
	p->crc_eof_offset = 0;
	spin_unlock_bh(&p->fcoe_rx_list.lock);

#ifdef CONFIG_SMP
	/*
	 * Don't bother moving the skb's if this context is running
	 * on the same CPU that is having its thread destroyed. This
	 * can easily happen when the module is removed.
	 */
	if (cpu != targ_cpu) {
		p0 = &per_cpu(fcoe_percpu, targ_cpu);
		spin_lock_bh(&p0->fcoe_rx_list.lock);
		if (p0->thread) {
1300 1301
			FCOE_DBG("Moving frames from CPU %d to CPU %d\n",
				 cpu, targ_cpu);
1302 1303 1304 1305 1306 1307 1308

			while ((skb = __skb_dequeue(&p->fcoe_rx_list)) != NULL)
				__skb_queue_tail(&p0->fcoe_rx_list, skb);
			spin_unlock_bh(&p0->fcoe_rx_list.lock);
		} else {
			/*
			 * The targeted CPU is not initialized and cannot accept
1309
			 * new	skbs. Unlock the targeted CPU and drop the skbs
1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330
			 * on the CPU that is going offline.
			 */
			while ((skb = __skb_dequeue(&p->fcoe_rx_list)) != NULL)
				kfree_skb(skb);
			spin_unlock_bh(&p0->fcoe_rx_list.lock);
		}
	} else {
		/*
		 * This scenario occurs when the module is being removed
		 * and all threads are being destroyed. skbs will continue
		 * to be shifted from the CPU thread that is being removed
		 * to the CPU thread associated with the CPU that is processing
		 * the module removal. Once there is only one CPU Rx thread it
		 * will reach this case and we will drop all skbs and later
		 * stop the thread.
		 */
		spin_lock_bh(&p->fcoe_rx_list.lock);
		while ((skb = __skb_dequeue(&p->fcoe_rx_list)) != NULL)
			kfree_skb(skb);
		spin_unlock_bh(&p->fcoe_rx_list.lock);
	}
1331
	put_cpu();
1332 1333
#else
	/*
1334
	 * This a non-SMP scenario where the singular Rx thread is
1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350
	 * being removed. Free all skbs and stop the thread.
	 */
	spin_lock_bh(&p->fcoe_rx_list.lock);
	while ((skb = __skb_dequeue(&p->fcoe_rx_list)) != NULL)
		kfree_skb(skb);
	spin_unlock_bh(&p->fcoe_rx_list.lock);
#endif

	if (thread)
		kthread_stop(thread);

	if (crc_eof)
		put_page(crc_eof);
}

/**
1351 1352 1353 1354
 * fcoe_cpu_callback() - Handler for CPU hotplug events
 * @nfb:    The callback data block
 * @action: The event triggering the callback
 * @hcpu:   The index of the CPU that the event is for
1355
 *
1356
 * This creates or destroys per-CPU data for fcoe
1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367
 *
 * Returns NOTIFY_OK always.
 */
static int fcoe_cpu_callback(struct notifier_block *nfb,
			     unsigned long action, void *hcpu)
{
	unsigned cpu = (unsigned long)hcpu;

	switch (action) {
	case CPU_ONLINE:
	case CPU_ONLINE_FROZEN:
1368
		FCOE_DBG("CPU %x online: Create Rx thread\n", cpu);
1369 1370 1371 1372
		fcoe_percpu_thread_create(cpu);
		break;
	case CPU_DEAD:
	case CPU_DEAD_FROZEN:
1373
		FCOE_DBG("CPU %x offline: Remove Rx thread\n", cpu);
1374 1375 1376 1377 1378 1379 1380 1381
		fcoe_percpu_thread_destroy(cpu);
		break;
	default:
		break;
	}
	return NOTIFY_OK;
}

1382 1383 1384 1385
/**
 * fcoe_select_cpu() - Selects CPU to handle post-processing of incoming
 *			command.
 *
1386 1387
 * This routine selects next CPU based on cpumask to distribute
 * incoming requests in round robin.
1388
 *
1389
 * Returns: int CPU number
1390
 */
1391
static inline unsigned int fcoe_select_cpu(void)
1392 1393 1394
{
	static unsigned int selected_cpu;

1395 1396 1397 1398
	selected_cpu = cpumask_next(selected_cpu, cpu_online_mask);
	if (selected_cpu >= nr_cpu_ids)
		selected_cpu = cpumask_first(cpu_online_mask);

1399 1400 1401
	return selected_cpu;
}

1402
/**
1403 1404 1405 1406 1407
 * fcoe_rcv() - Receive packets from a net device
 * @skb:    The received packet
 * @netdev: The net device that the packet was received on
 * @ptype:  The packet type context
 * @olddev: The last device net device
1408
 *
1409 1410
 * This routine is called by NET_RX_SOFTIRQ. It receives a packet, builds a
 * FC frame and passes the frame to libfc.
1411 1412
 *
 * Returns: 0 for success
1413
 */
1414
static int fcoe_rcv(struct sk_buff *skb, struct net_device *netdev,
1415 1416
	     struct packet_type *ptype, struct net_device *olddev)
{
1417
	struct fc_lport *lport;
1418
	struct fcoe_rcv_info *fr;
1419
	struct fcoe_ctlr *ctlr;
1420
	struct fcoe_interface *fcoe;
1421 1422
	struct fc_frame_header *fh;
	struct fcoe_percpu_s *fps;
1423
	struct ethhdr *eh;
1424
	unsigned int cpu;
1425

1426
	fcoe = container_of(ptype, struct fcoe_interface, fcoe_packet_type);
1427 1428
	ctlr = fcoe_to_ctlr(fcoe);
	lport = ctlr->lp;
1429 1430
	if (unlikely(!lport)) {
		FCOE_NETDEV_DBG(netdev, "Cannot find hba structure");
1431 1432
		goto err2;
	}
1433
	if (!lport->link_up)
1434
		goto err2;
1435

1436
	FCOE_NETDEV_DBG(netdev, "skb_info: len:%d data_len:%d head:%p "
1437 1438 1439 1440
			"data:%p tail:%p end:%p sum:%d dev:%s",
			skb->len, skb->data_len, skb->head, skb->data,
			skb_tail_pointer(skb), skb_end_pointer(skb),
			skb->csum, skb->dev ? skb->dev->name : "<NULL>");
1441

1442 1443
	eh = eth_hdr(skb);

1444 1445
	if (is_fip_mode(ctlr) &&
	    compare_ether_addr(eh->h_source, ctlr->dest_addr)) {
1446 1447
		FCOE_NETDEV_DBG(netdev, "wrong source mac address:%pM\n",
				eh->h_source);
1448 1449 1450 1451 1452 1453 1454 1455
		goto err;
	}

	/*
	 * Check for minimum frame length, and make sure required FCoE
	 * and FC headers are pulled into the linear data area.
	 */
	if (unlikely((skb->len < FCOE_MIN_FRAME) ||
1456
		     !pskb_may_pull(skb, FCOE_HEADER_LEN)))
1457 1458 1459 1460 1461
		goto err;

	skb_set_transport_header(skb, sizeof(struct fcoe_hdr));
	fh = (struct fc_frame_header *) skb_transport_header(skb);

1462 1463 1464 1465 1466 1467
	if (ntoh24(&eh->h_dest[3]) != ntoh24(fh->fh_d_id)) {
		FCOE_NETDEV_DBG(netdev, "FC frame d_id mismatch with MAC:%pM\n",
				eh->h_dest);
		goto err;
	}

1468
	fr = fcoe_dev_from_skb(skb);
1469
	fr->fr_dev = lport;
1470

1471
	/*
1472 1473 1474
	 * In case the incoming frame's exchange is originated from
	 * the initiator, then received frame's exchange id is ANDed
	 * with fc_cpu_mask bits to get the same cpu on which exchange
1475 1476
	 * was originated, otherwise select cpu using rx exchange id
	 * or fcoe_select_cpu().
1477
	 */
1478 1479
	if (ntoh24(fh->fh_f_ctl) & FC_FC_EX_CTX)
		cpu = ntohs(fh->fh_ox_id) & fc_cpu_mask;
1480
	else {
1481 1482 1483
		if (ntohs(fh->fh_rx_id) == FC_XID_UNKNOWN)
			cpu = fcoe_select_cpu();
		else
1484
			cpu = ntohs(fh->fh_rx_id) & fc_cpu_mask;
1485
	}
1486 1487 1488 1489

	if (cpu >= nr_cpu_ids)
		goto err;

1490
	fps = &per_cpu(fcoe_percpu, cpu);
1491
	spin_lock(&fps->fcoe_rx_list.lock);
1492 1493 1494 1495 1496 1497
	if (unlikely(!fps->thread)) {
		/*
		 * The targeted CPU is not ready, let's target
		 * the first CPU now. For non-SMP systems this
		 * will check the same CPU twice.
		 */
1498
		FCOE_NETDEV_DBG(netdev, "CPU is online, but no receive thread "
1499 1500
				"ready for incoming skb- using first online "
				"CPU.\n");
1501

1502
		spin_unlock(&fps->fcoe_rx_list.lock);
1503
		cpu = cpumask_first(cpu_online_mask);
1504
		fps = &per_cpu(fcoe_percpu, cpu);
1505
		spin_lock(&fps->fcoe_rx_list.lock);
1506
		if (!fps->thread) {
1507
			spin_unlock(&fps->fcoe_rx_list.lock);
1508 1509 1510 1511 1512 1513 1514 1515 1516
			goto err;
		}
	}

	/*
	 * We now have a valid CPU that we're targeting for
	 * this skb. We also have this receive thread locked,
	 * so we're free to queue skbs into it's queue.
	 */
1517

1518 1519 1520 1521 1522 1523
	/*
	 * Note: We used to have a set of conditions under which we would
	 * call fcoe_recv_frame directly, rather than queuing to the rx list
	 * as it could save a few cycles, but doing so is prohibited, as
	 * fcoe_recv_frame has several paths that may sleep, which is forbidden
	 * in softirq context.
1524
	 */
1525
	__skb_queue_tail(&fps->fcoe_rx_list, skb);
1526
	if (fps->thread->state == TASK_INTERRUPTIBLE)
1527
		wake_up_process(fps->thread);
1528
	spin_unlock(&fps->fcoe_rx_list.lock);
1529 1530 1531

	return 0;
err:
1532
	per_cpu_ptr(lport->stats, get_cpu())->ErrorFrames++;
1533
	put_cpu();
1534 1535 1536 1537 1538 1539
err2:
	kfree_skb(skb);
	return -1;
}

/**
1540
 * fcoe_alloc_paged_crc_eof() - Allocate a page to be used for the trailer CRC
1541 1542 1543
 * @skb:  The packet to be transmitted
 * @tlen: The total length of the trailer
 *
1544
 * Returns: 0 for success
1545
 */
1546
static int fcoe_alloc_paged_crc_eof(struct sk_buff *skb, int tlen)
1547 1548
{
	struct fcoe_percpu_s *fps;
1549
	int rc;
1550

1551
	fps = &get_cpu_var(fcoe_percpu);
1552
	rc = fcoe_get_paged_crc_eof(skb, tlen, fps);
1553
	put_cpu_var(fcoe_percpu);
1554

1555
	return rc;
1556 1557 1558
}

/**
1559 1560 1561
 * fcoe_xmit() - Transmit a FCoE frame
 * @lport: The local port that the frame is to be transmitted for
 * @fp:	   The frame to be transmitted
1562
 *
1563
 * Return: 0 for success
1564
 */
1565
static int fcoe_xmit(struct fc_lport *lport, struct fc_frame *fp)
1566
{
1567
	int wlen;
1568 1569 1570 1571
	u32 crc;
	struct ethhdr *eh;
	struct fcoe_crc_eof *cp;
	struct sk_buff *skb;
1572
	struct fc_stats *stats;
1573 1574 1575 1576
	struct fc_frame_header *fh;
	unsigned int hlen;		/* header length implies the version */
	unsigned int tlen;		/* trailer length */
	unsigned int elen;		/* eth header, may include vlan */
1577
	struct fcoe_port *port = lport_priv(lport);
1578
	struct fcoe_interface *fcoe = port->priv;
1579
	struct fcoe_ctlr *ctlr = fcoe_to_ctlr(fcoe);
1580 1581 1582 1583 1584 1585
	u8 sof, eof;
	struct fcoe_hdr *hp;

	WARN_ON((fr_len(fp) % sizeof(u32)) != 0);

	fh = fc_frame_header_get(fp);
1586 1587 1588
	skb = fp_skb(fp);
	wlen = skb->len / FCOE_WORD_TO_BYTE;

1589
	if (!lport->link_up) {
1590
		kfree_skb(skb);
1591
		return 0;
1592 1593
	}

1594
	if (unlikely(fh->fh_type == FC_TYPE_ELS) &&
1595
	    fcoe_ctlr_els_send(ctlr, lport, skb))
1596 1597
		return 0;

1598 1599 1600
	sof = fr_sof(fp);
	eof = fr_eof(fp);

1601
	elen = sizeof(struct ethhdr);
1602 1603 1604 1605 1606
	hlen = sizeof(struct fcoe_hdr);
	tlen = sizeof(struct fcoe_crc_eof);
	wlen = (skb->len - tlen + sizeof(crc)) / FCOE_WORD_TO_BYTE;

	/* crc offload */
1607
	if (likely(lport->crc_offload)) {
1608
		skb->ip_summed = CHECKSUM_UNNECESSARY;
1609 1610 1611 1612 1613 1614 1615 1616
		skb->csum_start = skb_headroom(skb);
		skb->csum_offset = skb->len;
		crc = 0;
	} else {
		skb->ip_summed = CHECKSUM_NONE;
		crc = fcoe_fc_crc(fp);
	}

1617
	/* copy port crc and eof to the skb buff */
1618 1619
	if (skb_is_nonlinear(skb)) {
		skb_frag_t *frag;
1620
		if (fcoe_alloc_paged_crc_eof(skb, tlen)) {
R
Roel Kluin 已提交
1621
			kfree_skb(skb);
1622 1623 1624
			return -ENOMEM;
		}
		frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1];
1625
		cp = kmap_atomic(skb_frag_page(frag))
1626 1627 1628 1629 1630 1631 1632 1633 1634 1635
			+ frag->page_offset;
	} else {
		cp = (struct fcoe_crc_eof *)skb_put(skb, tlen);
	}

	memset(cp, 0, sizeof(*cp));
	cp->fcoe_eof = eof;
	cp->fcoe_crc32 = cpu_to_le32(~crc);

	if (skb_is_nonlinear(skb)) {
1636
		kunmap_atomic(cp);
1637 1638 1639
		cp = NULL;
	}

1640
	/* adjust skb network/transport offsets to match mac/fcoe/port */
1641 1642 1643 1644
	skb_push(skb, elen + hlen);
	skb_reset_mac_header(skb);
	skb_reset_network_header(skb);
	skb->mac_len = elen;
1645
	skb->protocol = htons(ETH_P_FCOE);
1646 1647
	skb->priority = port->priority;

1648 1649 1650 1651 1652 1653 1654
	if (fcoe->netdev->priv_flags & IFF_802_1Q_VLAN &&
	    fcoe->realdev->features & NETIF_F_HW_VLAN_TX) {
		skb->vlan_tci = VLAN_TAG_PRESENT |
				vlan_dev_vlan_id(fcoe->netdev);
		skb->dev = fcoe->realdev;
	} else
		skb->dev = fcoe->netdev;
1655 1656 1657 1658

	/* fill up mac and fcoe headers */
	eh = eth_hdr(skb);
	eh->h_proto = htons(ETH_P_FCOE);
1659 1660
	memcpy(eh->h_dest, ctlr->dest_addr, ETH_ALEN);
	if (ctlr->map_dest)
1661
		memcpy(eh->h_dest + 3, fh->fh_d_id, 3);
1662

1663 1664
	if (unlikely(ctlr->flogi_oxid != FC_XID_UNKNOWN))
		memcpy(eh->h_source, ctlr->ctl_src_addr, ETH_ALEN);
1665
	else
1666
		memcpy(eh->h_source, port->data_src_addr, ETH_ALEN);
1667 1668 1669 1670 1671 1672 1673

	hp = (struct fcoe_hdr *)(eh + 1);
	memset(hp, 0, sizeof(*hp));
	if (FC_FCOE_VER)
		FC_FCOE_ENCAPS_VER(hp, FC_FCOE_VER);
	hp->fcoe_sof = sof;

1674
	/* fcoe lso, mss is in max_payload which is non-zero for FCP data */
1675
	if (lport->seq_offload && fr_max_payload(fp)) {
1676 1677 1678 1679 1680 1681
		skb_shinfo(skb)->gso_type = SKB_GSO_FCOE;
		skb_shinfo(skb)->gso_size = fr_max_payload(fp);
	} else {
		skb_shinfo(skb)->gso_type = 0;
		skb_shinfo(skb)->gso_size = 0;
	}
1682
	/* update tx stats: regardless if LLD fails */
1683
	stats = per_cpu_ptr(lport->stats, get_cpu());
1684 1685
	stats->TxFrames++;
	stats->TxWords += wlen;
1686
	put_cpu();
1687 1688

	/* send down to lld */
1689
	fr_dev(fp) = lport;
1690
	fcoe_port_send(port, skb);
1691 1692 1693
	return 0;
}

1694
/**
1695 1696
 * fcoe_percpu_flush_done() - Indicate per-CPU queue flush completion
 * @skb: The completed skb (argument required by destructor)
1697 1698 1699 1700 1701 1702
 */
static void fcoe_percpu_flush_done(struct sk_buff *skb)
{
	complete(&fcoe_flush_completion);
}

1703 1704 1705 1706 1707 1708 1709 1710 1711 1712
/**
 * fcoe_filter_frames() - filter out bad fcoe frames, i.e. bad CRC
 * @lport: The local port the frame was received on
 * @fp:	   The received frame
 *
 * Return: 0 on passing filtering checks
 */
static inline int fcoe_filter_frames(struct fc_lport *lport,
				     struct fc_frame *fp)
{
1713
	struct fcoe_ctlr *ctlr;
1714 1715 1716
	struct fcoe_interface *fcoe;
	struct fc_frame_header *fh;
	struct sk_buff *skb = (struct sk_buff *)fp;
1717
	struct fc_stats *stats;
1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733

	/*
	 * We only check CRC if no offload is available and if it is
	 * it's solicited data, in which case, the FCP layer would
	 * check it during the copy.
	 */
	if (lport->crc_offload && skb->ip_summed == CHECKSUM_UNNECESSARY)
		fr_flags(fp) &= ~FCPHF_CRC_UNCHECKED;
	else
		fr_flags(fp) |= FCPHF_CRC_UNCHECKED;

	fh = (struct fc_frame_header *) skb_transport_header(skb);
	fh = fc_frame_header_get(fp);
	if (fh->fh_r_ctl == FC_RCTL_DD_SOL_DATA && fh->fh_type == FC_TYPE_FCP)
		return 0;

1734
	fcoe = ((struct fcoe_port *)lport_priv(lport))->priv;
1735 1736
	ctlr = fcoe_to_ctlr(fcoe);
	if (is_fip_mode(ctlr) && fc_frame_payload_op(fp) == ELS_LOGO &&
1737 1738 1739 1740 1741
	    ntoh24(fh->fh_s_id) == FC_FID_FLOGI) {
		FCOE_DBG("fcoe: dropping FCoE lport LOGO in fip mode\n");
		return -EINVAL;
	}

1742
	if (!(fr_flags(fp) & FCPHF_CRC_UNCHECKED) ||
1743 1744 1745 1746 1747
	    le32_to_cpu(fr_crc(fp)) == ~crc32(~0, skb->data, skb->len)) {
		fr_flags(fp) &= ~FCPHF_CRC_UNCHECKED;
		return 0;
	}

1748
	stats = per_cpu_ptr(lport->stats, get_cpu());
1749 1750 1751
	stats->InvalidCRCCount++;
	if (stats->InvalidCRCCount < 5)
		printk(KERN_WARNING "fcoe: dropping frame with CRC error\n");
1752
	put_cpu();
1753 1754 1755
	return -EINVAL;
}

1756
/**
1757 1758
 * fcoe_recv_frame() - process a single received frame
 * @skb: frame to process
1759
 */
1760
static void fcoe_recv_frame(struct sk_buff *skb)
1761 1762
{
	u32 fr_len;
1763
	struct fc_lport *lport;
1764
	struct fcoe_rcv_info *fr;
1765
	struct fc_stats *stats;
1766 1767
	struct fcoe_crc_eof crc_eof;
	struct fc_frame *fp;
1768
	struct fcoe_port *port;
1769 1770
	struct fcoe_hdr *hp;

1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787
	fr = fcoe_dev_from_skb(skb);
	lport = fr->fr_dev;
	if (unlikely(!lport)) {
		if (skb->destructor != fcoe_percpu_flush_done)
			FCOE_NETDEV_DBG(skb->dev, "NULL lport in skb");
		kfree_skb(skb);
		return;
	}

	FCOE_NETDEV_DBG(skb->dev, "skb_info: len:%d data_len:%d "
			"head:%p data:%p tail:%p end:%p sum:%d dev:%s",
			skb->len, skb->data_len,
			skb->head, skb->data, skb_tail_pointer(skb),
			skb_end_pointer(skb), skb->csum,
			skb->dev ? skb->dev->name : "<NULL>");

	port = lport_priv(lport);
1788
	skb_linearize(skb); /* check for skb_is_nonlinear is within skb_linearize */
1789 1790 1791 1792 1793 1794 1795

	/*
	 * Frame length checks and setting up the header pointers
	 * was done in fcoe_rcv already.
	 */
	hp = (struct fcoe_hdr *) skb_network_header(skb);

1796
	stats = per_cpu_ptr(lport->stats, get_cpu());
1797 1798 1799 1800 1801 1802 1803 1804
	if (unlikely(FC_FCOE_DECAPS_VER(hp) != FC_FCOE_VER)) {
		if (stats->ErrorFrames < 5)
			printk(KERN_WARNING "fcoe: FCoE version "
			       "mismatch: The frame has "
			       "version %x, but the "
			       "initiator supports version "
			       "%x\n", FC_FCOE_DECAPS_VER(hp),
			       FC_FCOE_VER);
1805
		goto drop;
1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819
	}

	skb_pull(skb, sizeof(struct fcoe_hdr));
	fr_len = skb->len - sizeof(struct fcoe_crc_eof);

	stats->RxFrames++;
	stats->RxWords += fr_len / FCOE_WORD_TO_BYTE;

	fp = (struct fc_frame *)skb;
	fc_frame_init(fp);
	fr_dev(fp) = lport;
	fr_sof(fp) = hp->fcoe_sof;

	/* Copy out the CRC and EOF trailer for access */
1820 1821
	if (skb_copy_bits(skb, fr_len, &crc_eof, sizeof(crc_eof)))
		goto drop;
1822 1823
	fr_eof(fp) = crc_eof.fcoe_eof;
	fr_crc(fp) = crc_eof.fcoe_crc32;
1824 1825
	if (pskb_trim(skb, fr_len))
		goto drop;
1826

1827 1828 1829 1830
	if (!fcoe_filter_frames(lport, fp)) {
		put_cpu();
		fc_exch_recv(lport, fp);
		return;
1831
	}
1832 1833 1834 1835
drop:
	stats->ErrorFrames++;
	put_cpu();
	kfree_skb(skb);
1836 1837 1838 1839 1840 1841 1842 1843
}

/**
 * fcoe_percpu_receive_thread() - The per-CPU packet receive thread
 * @arg: The per-CPU context
 *
 * Return: 0 for success
 */
1844
static int fcoe_percpu_receive_thread(void *arg)
1845 1846 1847
{
	struct fcoe_percpu_s *p = arg;
	struct sk_buff *skb;
1848 1849 1850
	struct sk_buff_head tmp;

	skb_queue_head_init(&tmp);
1851

1852
	set_user_nice(current, -20);
1853 1854 1855 1856

	while (!kthread_should_stop()) {

		spin_lock_bh(&p->fcoe_rx_list.lock);
1857 1858 1859 1860 1861 1862 1863 1864
		skb_queue_splice_init(&p->fcoe_rx_list, &tmp);
		spin_unlock_bh(&p->fcoe_rx_list.lock);

		while ((skb = __skb_dequeue(&tmp)) != NULL)
			fcoe_recv_frame(skb);

		spin_lock_bh(&p->fcoe_rx_list.lock);
		if (!skb_queue_len(&p->fcoe_rx_list)) {
1865 1866 1867 1868
			set_current_state(TASK_INTERRUPTIBLE);
			spin_unlock_bh(&p->fcoe_rx_list.lock);
			schedule();
			set_current_state(TASK_RUNNING);
1869 1870
		} else
			spin_unlock_bh(&p->fcoe_rx_list.lock);
1871 1872 1873 1874 1875
	}
	return 0;
}

/**
1876
 * fcoe_dev_setup() - Setup the link change notification interface
1877
 */
1878
static void fcoe_dev_setup(void)
1879
{
1880
	register_dcbevent_notifier(&dcb_notifier);
1881 1882 1883 1884
	register_netdevice_notifier(&fcoe_notifier);
}

/**
1885
 * fcoe_dev_cleanup() - Cleanup the link change notification interface
1886
 */
1887 1888
static void fcoe_dev_cleanup(void)
{
1889
	unregister_dcbevent_notifier(&dcb_notifier);
1890 1891 1892
	unregister_netdevice_notifier(&fcoe_notifier);
}

1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914
static struct fcoe_interface *
fcoe_hostlist_lookup_realdev_port(struct net_device *netdev)
{
	struct fcoe_interface *fcoe;
	struct net_device *real_dev;

	list_for_each_entry(fcoe, &fcoe_hostlist, list) {
		if (fcoe->netdev->priv_flags & IFF_802_1Q_VLAN)
			real_dev = vlan_dev_real_dev(fcoe->netdev);
		else
			real_dev = fcoe->netdev;

		if (netdev == real_dev)
			return fcoe;
	}
	return NULL;
}

static int fcoe_dcb_app_notification(struct notifier_block *notifier,
				     ulong event, void *ptr)
{
	struct dcb_app_type *entry = ptr;
1915
	struct fcoe_ctlr *ctlr;
1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932
	struct fcoe_interface *fcoe;
	struct net_device *netdev;
	struct fcoe_port *port;
	int prio;

	if (entry->app.selector != DCB_APP_IDTYPE_ETHTYPE)
		return NOTIFY_OK;

	netdev = dev_get_by_index(&init_net, entry->ifindex);
	if (!netdev)
		return NOTIFY_OK;

	fcoe = fcoe_hostlist_lookup_realdev_port(netdev);
	dev_put(netdev);
	if (!fcoe)
		return NOTIFY_OK;

1933 1934
	ctlr = fcoe_to_ctlr(fcoe);

1935 1936 1937 1938 1939 1940 1941 1942 1943 1944
	if (entry->dcbx & DCB_CAP_DCBX_VER_CEE)
		prio = ffs(entry->app.priority) - 1;
	else
		prio = entry->app.priority;

	if (prio < 0)
		return NOTIFY_OK;

	if (entry->app.protocol == ETH_P_FIP ||
	    entry->app.protocol == ETH_P_FCOE)
1945
		ctlr->priority = prio;
1946 1947

	if (entry->app.protocol == ETH_P_FCOE) {
1948
		port = lport_priv(ctlr->lp);
1949 1950 1951 1952 1953 1954
		port->priority = prio;
	}

	return NOTIFY_OK;
}

1955
/**
1956 1957 1958 1959
 * fcoe_device_notification() - Handler for net device events
 * @notifier: The context of the notification
 * @event:    The type of event
 * @ptr:      The net device that the event was on
1960
 *
1961
 * This function is called by the Ethernet driver in case of link change event.
1962 1963
 *
 * Returns: 0 for success
1964
 */
1965 1966 1967
static int fcoe_device_notification(struct notifier_block *notifier,
				    ulong event, void *ptr)
{
1968
	struct fc_lport *lport = NULL;
1969
	struct net_device *netdev = ptr;
1970
	struct fcoe_ctlr *ctlr;
1971
	struct fcoe_interface *fcoe;
1972
	struct fcoe_port *port;
1973
	struct fc_stats *stats;
1974
	u32 link_possible = 1;
1975 1976 1977
	u32 mfs;
	int rc = NOTIFY_OK;

1978
	list_for_each_entry(fcoe, &fcoe_hostlist, list) {
1979
		if (fcoe->netdev == netdev) {
1980 1981
			ctlr = fcoe_to_ctlr(fcoe);
			lport = ctlr->lp;
1982 1983 1984
			break;
		}
	}
1985
	if (!lport) {
1986 1987 1988 1989 1990 1991 1992
		rc = NOTIFY_DONE;
		goto out;
	}

	switch (event) {
	case NETDEV_DOWN:
	case NETDEV_GOING_DOWN:
1993
		link_possible = 0;
1994 1995 1996 1997 1998
		break;
	case NETDEV_UP:
	case NETDEV_CHANGE:
		break;
	case NETDEV_CHANGEMTU:
1999 2000
		if (netdev->features & NETIF_F_FCOE_MTU)
			break;
2001 2002
		mfs = netdev->mtu - (sizeof(struct fcoe_hdr) +
				     sizeof(struct fcoe_crc_eof));
2003
		if (mfs >= FC_MIN_MAX_FRAME)
2004
			fc_set_mfs(lport, mfs);
2005 2006 2007
		break;
	case NETDEV_REGISTER:
		break;
2008 2009
	case NETDEV_UNREGISTER:
		list_del(&fcoe->list);
2010
		port = lport_priv(ctlr->lp);
2011
		queue_work(fcoe_wq, &port->destroy_work);
2012 2013
		goto out;
		break;
2014 2015 2016
	case NETDEV_FEAT_CHANGE:
		fcoe_netdev_features_change(lport, netdev);
		break;
2017
	default:
2018
		FCOE_NETDEV_DBG(netdev, "Unknown event %ld "
2019
				"from netdev netlink\n", event);
2020
	}
2021 2022 2023

	fcoe_link_speed_update(lport);

2024
	if (link_possible && !fcoe_link_ok(lport))
2025 2026
		fcoe_ctlr_link_up(ctlr);
	else if (fcoe_ctlr_link_down(ctlr)) {
2027
		stats = per_cpu_ptr(lport->stats, get_cpu());
2028
		stats->LinkFailureCount++;
2029
		put_cpu();
2030
		fcoe_clean_pending_queue(lport);
2031 2032 2033 2034 2035
	}
out:
	return rc;
}

2036 2037
/**
 * fcoe_disable() - Disables a FCoE interface
2038
 * @netdev  : The net_device object the Ethernet interface to create on
2039
 *
2040
 * Called from fcoe transport.
2041 2042 2043
 *
 * Returns: 0 for success
 */
2044
static int fcoe_disable(struct net_device *netdev)
2045
{
2046
	struct fcoe_ctlr *ctlr;
2047 2048 2049 2050 2051
	struct fcoe_interface *fcoe;
	int rc = 0;

	mutex_lock(&fcoe_config_mutex);

2052
	rtnl_lock();
2053 2054 2055
	fcoe = fcoe_hostlist_lookup_port(netdev);
	rtnl_unlock();

2056
	if (fcoe) {
2057 2058 2059
		ctlr = fcoe_to_ctlr(fcoe);
		fcoe_ctlr_link_down(ctlr);
		fcoe_clean_pending_queue(ctlr->lp);
2060
	} else
2061 2062 2063 2064 2065 2066 2067 2068
		rc = -ENODEV;

	mutex_unlock(&fcoe_config_mutex);
	return rc;
}

/**
 * fcoe_enable() - Enables a FCoE interface
2069
 * @netdev  : The net_device object the Ethernet interface to create on
2070
 *
2071
 * Called from fcoe transport.
2072 2073 2074
 *
 * Returns: 0 for success
 */
2075
static int fcoe_enable(struct net_device *netdev)
2076
{
2077
	struct fcoe_ctlr *ctlr;
2078 2079 2080 2081
	struct fcoe_interface *fcoe;
	int rc = 0;

	mutex_lock(&fcoe_config_mutex);
2082
	rtnl_lock();
2083 2084 2085
	fcoe = fcoe_hostlist_lookup_port(netdev);
	rtnl_unlock();

2086
	if (!fcoe) {
2087
		rc = -ENODEV;
2088 2089
		goto out;
	}
2090

2091 2092 2093 2094 2095 2096
	ctlr = fcoe_to_ctlr(fcoe);

	if (!fcoe_link_ok(ctlr->lp))
		fcoe_ctlr_link_up(ctlr);

out:
2097 2098 2099 2100
	mutex_unlock(&fcoe_config_mutex);
	return rc;
}

2101
/**
2102
 * fcoe_destroy() - Destroy a FCoE interface
2103
 * @netdev  : The net_device object the Ethernet interface to create on
2104
 *
2105
 * Called from fcoe transport
2106 2107
 *
 * Returns: 0 for success
2108
 */
2109
static int fcoe_destroy(struct net_device *netdev)
2110
{
2111
	struct fcoe_ctlr *ctlr;
2112
	struct fcoe_interface *fcoe;
2113
	struct fc_lport *lport;
2114
	struct fcoe_port *port;
2115
	int rc = 0;
2116

2117
	mutex_lock(&fcoe_config_mutex);
2118
	rtnl_lock();
2119 2120
	fcoe = fcoe_hostlist_lookup_port(netdev);
	if (!fcoe) {
2121
		rc = -ENODEV;
2122
		goto out_nodev;
2123
	}
2124 2125
	ctlr = fcoe_to_ctlr(fcoe);
	lport = ctlr->lp;
2126
	port = lport_priv(lport);
2127
	list_del(&fcoe->list);
2128
	queue_work(fcoe_wq, &port->destroy_work);
2129
out_nodev:
2130
	rtnl_unlock();
2131
	mutex_unlock(&fcoe_config_mutex);
2132 2133 2134
	return rc;
}

2135 2136 2137 2138
/**
 * fcoe_destroy_work() - Destroy a FCoE port in a deferred work context
 * @work: Handle to the FCoE port to be destroyed
 */
2139 2140 2141
static void fcoe_destroy_work(struct work_struct *work)
{
	struct fcoe_port *port;
2142
	struct fcoe_interface *fcoe;
2143 2144 2145

	port = container_of(work, struct fcoe_port, destroy_work);
	mutex_lock(&fcoe_config_mutex);
2146 2147

	fcoe = port->priv;
2148
	fcoe_if_destroy(port->lport);
2149
	fcoe_interface_cleanup(fcoe);
2150

2151 2152 2153
	mutex_unlock(&fcoe_config_mutex);
}

2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167
/**
 * fcoe_match() - Check if the FCoE is supported on the given netdevice
 * @netdev  : The net_device object the Ethernet interface to create on
 *
 * Called from fcoe transport.
 *
 * Returns: always returns true as this is the default FCoE transport,
 * i.e., support all netdevs.
 */
static bool fcoe_match(struct net_device *netdev)
{
	return true;
}

2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179
/**
 * fcoe_dcb_create() - Initialize DCB attributes and hooks
 * @netdev: The net_device object of the L2 link that should be queried
 * @port: The fcoe_port to bind FCoE APP priority with
 * @
 */
static void fcoe_dcb_create(struct fcoe_interface *fcoe)
{
#ifdef CONFIG_DCB
	int dcbx;
	u8 fup, up;
	struct net_device *netdev = fcoe->realdev;
2180 2181
	struct fcoe_ctlr *ctlr = fcoe_to_ctlr(fcoe);
	struct fcoe_port *port = lport_priv(ctlr->lp);
2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203
	struct dcb_app app = {
				.priority = 0,
				.protocol = ETH_P_FCOE
			     };

	/* setup DCB priority attributes. */
	if (netdev && netdev->dcbnl_ops && netdev->dcbnl_ops->getdcbx) {
		dcbx = netdev->dcbnl_ops->getdcbx(netdev);

		if (dcbx & DCB_CAP_DCBX_VER_IEEE) {
			app.selector = IEEE_8021QAZ_APP_SEL_ETHERTYPE;
			up = dcb_ieee_getapp_mask(netdev, &app);
			app.protocol = ETH_P_FIP;
			fup = dcb_ieee_getapp_mask(netdev, &app);
		} else {
			app.selector = DCB_APP_IDTYPE_ETHTYPE;
			up = dcb_getapp(netdev, &app);
			app.protocol = ETH_P_FIP;
			fup = dcb_getapp(netdev, &app);
		}

		port->priority = ffs(up) ? ffs(up) - 1 : 0;
2204
		ctlr->priority = ffs(fup) ? ffs(fup) - 1 : port->priority;
2205 2206 2207 2208
	}
#endif
}

2209
/**
2210
 * fcoe_create() - Create a fcoe interface
2211 2212
 * @netdev  : The net_device object the Ethernet interface to create on
 * @fip_mode: The FIP mode for this creation
2213
 *
2214
 * Called from fcoe transport
2215 2216
 *
 * Returns: 0 for success
2217
 */
2218
static int fcoe_create(struct net_device *netdev, enum fip_state fip_mode)
2219
{
2220
	int rc = 0;
2221
	struct fcoe_ctlr_device *ctlr_dev;
2222
	struct fcoe_ctlr *ctlr;
2223
	struct fcoe_interface *fcoe;
2224
	struct fc_lport *lport;
2225

2226
	mutex_lock(&fcoe_config_mutex);
2227
	rtnl_lock();
2228

2229 2230 2231
	/* look for existing lport */
	if (fcoe_hostlist_lookup(netdev)) {
		rc = -EEXIST;
2232
		goto out_nodev;
2233 2234
	}

2235
	fcoe = fcoe_interface_create(netdev, fip_mode);
2236 2237
	if (IS_ERR(fcoe)) {
		rc = PTR_ERR(fcoe);
2238
		goto out_nodev;
2239 2240
	}

2241
	ctlr = fcoe_to_ctlr(fcoe);
2242 2243
	ctlr_dev = fcoe_ctlr_to_ctlr_dev(ctlr);
	lport = fcoe_if_create(fcoe, &ctlr_dev->dev, 0);
2244
	if (IS_ERR(lport)) {
2245
		printk(KERN_ERR "fcoe: Failed to create interface (%s)\n",
2246 2247
		       netdev->name);
		rc = -EIO;
2248
		rtnl_unlock();
2249
		fcoe_interface_cleanup(fcoe);
2250
		goto out_nortnl;
2251
	}
2252

2253
	/* Make this the "master" N_Port */
2254
	ctlr->lp = lport;
2255

2256 2257 2258
	/* setup DCB priority attributes. */
	fcoe_dcb_create(fcoe);

2259 2260 2261
	/* add to lports list */
	fcoe_hostlist_add(lport);

2262 2263 2264
	/* start FIP Discovery and FLOGI */
	lport->boot_time = jiffies;
	fc_fabric_login(lport);
2265 2266
	if (!fcoe_link_ok(lport)) {
		rtnl_unlock();
2267
		fcoe_ctlr_link_up(ctlr);
2268 2269 2270
		mutex_unlock(&fcoe_config_mutex);
		return rc;
	}
2271

2272
out_nodev:
2273
	rtnl_unlock();
2274
out_nortnl:
2275
	mutex_unlock(&fcoe_config_mutex);
2276 2277 2278
	return rc;
}

2279
/**
2280 2281
 * fcoe_link_speed_update() - Update the supported and actual link speeds
 * @lport: The local port to update speeds for
2282
 *
2283 2284
 * Returns: 0 if the ethtool query was successful
 *          -1 if the ethtool query failed
2285
 */
2286
static int fcoe_link_speed_update(struct fc_lport *lport)
2287
{
2288
	struct net_device *netdev = fcoe_netdev(lport);
2289
	struct ethtool_cmd ecmd;
2290

2291
	if (!__ethtool_get_settings(netdev, &ecmd)) {
2292
		lport->link_supported_speeds &=
2293 2294 2295
			~(FC_PORTSPEED_1GBIT | FC_PORTSPEED_10GBIT);
		if (ecmd.supported & (SUPPORTED_1000baseT_Half |
				      SUPPORTED_1000baseT_Full))
2296
			lport->link_supported_speeds |= FC_PORTSPEED_1GBIT;
2297
		if (ecmd.supported & SUPPORTED_10000baseT_Full)
2298
			lport->link_supported_speeds |=
2299
				FC_PORTSPEED_10GBIT;
2300 2301
		switch (ethtool_cmd_speed(&ecmd)) {
		case SPEED_1000:
2302
			lport->link_speed = FC_PORTSPEED_1GBIT;
2303 2304
			break;
		case SPEED_10000:
2305
			lport->link_speed = FC_PORTSPEED_10GBIT;
2306 2307
			break;
		}
2308 2309 2310
		return 0;
	}
	return -1;
2311 2312
}

2313 2314 2315 2316 2317 2318 2319
/**
 * fcoe_link_ok() - Check if the link is OK for a local port
 * @lport: The local port to check link on
 *
 * Returns: 0 if link is UP and OK, -1 if not
 *
 */
2320
static int fcoe_link_ok(struct fc_lport *lport)
2321
{
2322
	struct net_device *netdev = fcoe_netdev(lport);
2323 2324 2325 2326 2327 2328

	if (netif_oper_up(netdev))
		return 0;
	return -1;
}

2329
/**
2330 2331
 * fcoe_percpu_clean() - Clear all pending skbs for an local port
 * @lport: The local port whose skbs are to be cleared
2332 2333 2334 2335 2336 2337 2338
 *
 * Must be called with fcoe_create_mutex held to single-thread completion.
 *
 * This flushes the pending skbs by adding a new skb to each queue and
 * waiting until they are all freed.  This assures us that not only are
 * there no packets that will be handled by the lport, but also that any
 * threads already handling packet have returned.
2339
 */
2340
static void fcoe_percpu_clean(struct fc_lport *lport)
2341 2342
{
	struct fcoe_percpu_s *pp;
2343
	struct sk_buff *skb;
2344
	unsigned int cpu;
2345

2346 2347
	for_each_possible_cpu(cpu) {
		pp = &per_cpu(fcoe_percpu, cpu);
2348

2349
		if (!pp->thread || !cpu_online(cpu))
2350 2351 2352
			continue;

		skb = dev_alloc_skb(0);
2353
		if (!skb)
2354
			continue;
2355

2356 2357
		skb->destructor = fcoe_percpu_flush_done;

2358
		spin_lock_bh(&pp->fcoe_rx_list.lock);
2359 2360 2361
		__skb_queue_tail(&pp->fcoe_rx_list, skb);
		if (pp->fcoe_rx_list.qlen == 1)
			wake_up_process(pp->thread);
2362
		spin_unlock_bh(&pp->fcoe_rx_list.lock);
2363 2364

		wait_for_completion(&fcoe_flush_completion);
2365 2366 2367
	}
}

2368
/**
2369 2370
 * fcoe_reset() - Reset a local port
 * @shost: The SCSI host associated with the local port to be reset
2371
 *
2372
 * Returns: Always 0 (return value required by FC transport template)
2373
 */
2374
static int fcoe_reset(struct Scsi_Host *shost)
2375 2376
{
	struct fc_lport *lport = shost_priv(shost);
2377 2378
	struct fcoe_port *port = lport_priv(lport);
	struct fcoe_interface *fcoe = port->priv;
2379
	struct fcoe_ctlr *ctlr = fcoe_to_ctlr(fcoe);
2380

2381 2382 2383 2384
	fcoe_ctlr_link_down(ctlr);
	fcoe_clean_pending_queue(ctlr->lp);
	if (!fcoe_link_ok(ctlr->lp))
		fcoe_ctlr_link_up(ctlr);
2385 2386 2387
	return 0;
}

2388
/**
2389 2390
 * fcoe_hostlist_lookup_port() - Find the FCoE interface associated with a net device
 * @netdev: The net device used as a key
2391
 *
2392 2393 2394
 * Locking: Must be called with the RNL mutex held.
 *
 * Returns: NULL or the FCoE interface
2395
 */
2396
static struct fcoe_interface *
2397
fcoe_hostlist_lookup_port(const struct net_device *netdev)
2398
{
2399
	struct fcoe_interface *fcoe;
2400

2401
	list_for_each_entry(fcoe, &fcoe_hostlist, list) {
2402
		if (fcoe->netdev == netdev)
2403
			return fcoe;
2404 2405 2406 2407
	}
	return NULL;
}

2408
/**
2409 2410 2411
 * fcoe_hostlist_lookup() - Find the local port associated with a
 *			    given net device
 * @netdev: The netdevice used as a key
2412
 *
2413 2414 2415
 * Locking: Must be called with the RTNL mutex held
 *
 * Returns: NULL or the local port
2416
 */
2417
static struct fc_lport *fcoe_hostlist_lookup(const struct net_device *netdev)
2418
{
2419
	struct fcoe_ctlr *ctlr;
2420
	struct fcoe_interface *fcoe;
2421

2422
	fcoe = fcoe_hostlist_lookup_port(netdev);
2423 2424
	ctlr = fcoe_to_ctlr(fcoe);
	return (fcoe) ? ctlr->lp : NULL;
2425 2426
}

2427
/**
2428 2429 2430
 * fcoe_hostlist_add() - Add the FCoE interface identified by a local
 *			 port to the hostlist
 * @lport: The local port that identifies the FCoE interface to be added
2431
 *
2432
 * Locking: must be called with the RTNL mutex held
2433 2434
 *
 * Returns: 0 for success
2435
 */
2436
static int fcoe_hostlist_add(const struct fc_lport *lport)
2437
{
2438 2439 2440 2441 2442 2443
	struct fcoe_interface *fcoe;
	struct fcoe_port *port;

	fcoe = fcoe_hostlist_lookup_port(fcoe_netdev(lport));
	if (!fcoe) {
		port = lport_priv(lport);
2444
		fcoe = port->priv;
2445
		list_add_tail(&fcoe->list, &fcoe_hostlist);
2446 2447 2448 2449
	}
	return 0;
}

2450 2451 2452 2453 2454 2455 2456 2457 2458 2459 2460 2461

static struct fcoe_transport fcoe_sw_transport = {
	.name = {FCOE_TRANSPORT_DEFAULT},
	.attached = false,
	.list = LIST_HEAD_INIT(fcoe_sw_transport.list),
	.match = fcoe_match,
	.create = fcoe_create,
	.destroy = fcoe_destroy,
	.enable = fcoe_enable,
	.disable = fcoe_disable,
};

2462
/**
2463
 * fcoe_init() - Initialize fcoe.ko
2464
 *
2465
 * Returns: 0 on success, or a negative value on failure
2466
 */
2467 2468
static int __init fcoe_init(void)
{
2469
	struct fcoe_percpu_s *p;
2470
	unsigned int cpu;
2471
	int rc = 0;
2472

2473 2474 2475 2476
	fcoe_wq = alloc_workqueue("fcoe", 0, 0);
	if (!fcoe_wq)
		return -ENOMEM;

2477 2478 2479 2480 2481 2482 2483 2484
	/* register as a fcoe transport */
	rc = fcoe_transport_attach(&fcoe_sw_transport);
	if (rc) {
		printk(KERN_ERR "failed to register an fcoe transport, check "
			"if libfcoe is loaded\n");
		return rc;
	}

2485 2486
	mutex_lock(&fcoe_config_mutex);

2487
	for_each_possible_cpu(cpu) {
2488
		p = &per_cpu(fcoe_percpu, cpu);
2489 2490 2491
		skb_queue_head_init(&p->fcoe_rx_list);
	}

2492 2493
	for_each_online_cpu(cpu)
		fcoe_percpu_thread_create(cpu);
2494

2495 2496 2497 2498
	/* Initialize per CPU interrupt thread */
	rc = register_hotcpu_notifier(&fcoe_cpu_notifier);
	if (rc)
		goto out_free;
2499

2500
	/* Setup link change notification */
2501 2502
	fcoe_dev_setup();

2503 2504 2505
	rc = fcoe_if_init();
	if (rc)
		goto out_free;
2506

2507
	mutex_unlock(&fcoe_config_mutex);
2508
	return 0;
2509 2510 2511 2512 2513

out_free:
	for_each_online_cpu(cpu) {
		fcoe_percpu_thread_destroy(cpu);
	}
2514
	mutex_unlock(&fcoe_config_mutex);
2515
	destroy_workqueue(fcoe_wq);
2516
	return rc;
2517 2518 2519 2520
}
module_init(fcoe_init);

/**
2521
 * fcoe_exit() - Clean up fcoe.ko
2522
 *
2523
 * Returns: 0 on success or a  negative value on failure
2524
 */
2525 2526
static void __exit fcoe_exit(void)
{
2527
	struct fcoe_interface *fcoe, *tmp;
2528
	struct fcoe_ctlr *ctlr;
2529
	struct fcoe_port *port;
2530
	unsigned int cpu;
2531

2532 2533
	mutex_lock(&fcoe_config_mutex);

2534 2535
	fcoe_dev_cleanup();

2536
	/* releases the associated fcoe hosts */
2537 2538
	rtnl_lock();
	list_for_each_entry_safe(fcoe, tmp, &fcoe_hostlist, list) {
2539
		list_del(&fcoe->list);
2540 2541
		ctlr = fcoe_to_ctlr(fcoe);
		port = lport_priv(ctlr->lp);
2542
		queue_work(fcoe_wq, &port->destroy_work);
2543
	}
2544
	rtnl_unlock();
2545

2546 2547
	unregister_hotcpu_notifier(&fcoe_cpu_notifier);

2548
	for_each_online_cpu(cpu)
2549
		fcoe_percpu_thread_destroy(cpu);
2550

2551
	mutex_unlock(&fcoe_config_mutex);
2552

2553 2554 2555 2556 2557
	/*
	 * destroy_work's may be chained but destroy_workqueue()
	 * can take care of them. Just kill the fcoe_wq.
	 */
	destroy_workqueue(fcoe_wq);
2558

2559 2560 2561 2562 2563
	/*
	 * Detaching from the scsi transport must happen after all
	 * destroys are done on the fcoe_wq. destroy_workqueue will
	 * enusre the fcoe_wq is flushed.
	 */
2564
	fcoe_if_exit();
2565 2566 2567

	/* detach from fcoe transport */
	fcoe_transport_detach(&fcoe_sw_transport);
2568 2569
}
module_exit(fcoe_exit);
2570 2571 2572 2573 2574 2575 2576

/**
 * fcoe_flogi_resp() - FCoE specific FLOGI and FDISC response handler
 * @seq: active sequence in the FLOGI or FDISC exchange
 * @fp: response frame, or error encoded in a pointer (timeout)
 * @arg: pointer the the fcoe_ctlr structure
 *
2577
 * This handles MAC address management for FCoE, then passes control on to
2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590
 * the libfc FLOGI response handler.
 */
static void fcoe_flogi_resp(struct fc_seq *seq, struct fc_frame *fp, void *arg)
{
	struct fcoe_ctlr *fip = arg;
	struct fc_exch *exch = fc_seq_exch(seq);
	struct fc_lport *lport = exch->lp;
	u8 *mac;

	if (IS_ERR(fp))
		goto done;

	mac = fr_cb(fp)->granted_mac;
2591 2592 2593 2594 2595
	/* pre-FIP */
	if (is_zero_ether_addr(mac))
		fcoe_ctlr_recv_flogi(fip, lport, fp);
	if (!is_zero_ether_addr(mac))
		fcoe_update_src_mac(lport, mac);
2596 2597 2598 2599 2600 2601 2602 2603 2604 2605
done:
	fc_lport_flogi_resp(seq, fp, lport);
}

/**
 * fcoe_logo_resp() - FCoE specific LOGO response handler
 * @seq: active sequence in the LOGO exchange
 * @fp: response frame, or error encoded in a pointer (timeout)
 * @arg: pointer the the fcoe_ctlr structure
 *
2606
 * This handles MAC address management for FCoE, then passes control on to
2607 2608 2609 2610
 * the libfc LOGO response handler.
 */
static void fcoe_logo_resp(struct fc_seq *seq, struct fc_frame *fp, void *arg)
{
2611
	struct fc_lport *lport = arg;
2612 2613 2614
	static u8 zero_mac[ETH_ALEN] = { 0 };

	if (!IS_ERR(fp))
2615
		fcoe_update_src_mac(lport, zero_mac);
2616 2617 2618 2619 2620 2621 2622 2623 2624 2625 2626 2627
	fc_lport_logo_resp(seq, fp, lport);
}

/**
 * fcoe_elsct_send - FCoE specific ELS handler
 *
 * This does special case handling of FIP encapsualted ELS exchanges for FCoE,
 * using FCoE specific response handlers and passing the FIP controller as
 * the argument (the lport is still available from the exchange).
 *
 * Most of the work here is just handed off to the libfc routine.
 */
2628 2629 2630 2631 2632 2633
static struct fc_seq *fcoe_elsct_send(struct fc_lport *lport, u32 did,
				      struct fc_frame *fp, unsigned int op,
				      void (*resp)(struct fc_seq *,
						   struct fc_frame *,
						   void *),
				      void *arg, u32 timeout)
2634 2635
{
	struct fcoe_port *port = lport_priv(lport);
2636
	struct fcoe_interface *fcoe = port->priv;
2637
	struct fcoe_ctlr *fip = fcoe_to_ctlr(fcoe);
2638 2639 2640 2641 2642
	struct fc_frame_header *fh = fc_frame_header_get(fp);

	switch (op) {
	case ELS_FLOGI:
	case ELS_FDISC:
2643 2644
		if (lport->point_to_multipoint)
			break;
2645 2646 2647 2648 2649 2650 2651
		return fc_elsct_send(lport, did, fp, op, fcoe_flogi_resp,
				     fip, timeout);
	case ELS_LOGO:
		/* only hook onto fabric logouts, not port logouts */
		if (ntoh24(fh->fh_d_id) != FC_FID_FLOGI)
			break;
		return fc_elsct_send(lport, did, fp, op, fcoe_logo_resp,
2652
				     lport, timeout);
2653 2654 2655 2656
	}
	return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
}

2657 2658 2659 2660 2661 2662 2663 2664 2665 2666 2667 2668
/**
 * fcoe_vport_create() - create an fc_host/scsi_host for a vport
 * @vport: fc_vport object to create a new fc_host for
 * @disabled: start the new fc_host in a disabled state by default?
 *
 * Returns: 0 for success
 */
static int fcoe_vport_create(struct fc_vport *vport, bool disabled)
{
	struct Scsi_Host *shost = vport_to_shost(vport);
	struct fc_lport *n_port = shost_priv(shost);
	struct fcoe_port *port = lport_priv(n_port);
2669
	struct fcoe_interface *fcoe = port->priv;
2670 2671
	struct net_device *netdev = fcoe->netdev;
	struct fc_lport *vn_port;
2672 2673 2674 2675 2676
	int rc;
	char buf[32];

	rc = fcoe_validate_vport_create(vport);
	if (rc) {
2677
		fcoe_wwn_to_str(vport->port_name, buf, sizeof(buf));
2678 2679 2680 2681 2682
		printk(KERN_ERR "fcoe: Failed to create vport, "
			"WWPN (0x%s) already exists\n",
			buf);
		return rc;
	}
2683 2684

	mutex_lock(&fcoe_config_mutex);
2685
	rtnl_lock();
2686
	vn_port = fcoe_if_create(fcoe, &vport->dev, 1);
2687
	rtnl_unlock();
2688 2689 2690 2691 2692 2693 2694 2695 2696 2697 2698 2699 2700 2701 2702 2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719 2720
	mutex_unlock(&fcoe_config_mutex);

	if (IS_ERR(vn_port)) {
		printk(KERN_ERR "fcoe: fcoe_vport_create(%s) failed\n",
		       netdev->name);
		return -EIO;
	}

	if (disabled) {
		fc_vport_set_state(vport, FC_VPORT_DISABLED);
	} else {
		vn_port->boot_time = jiffies;
		fc_fabric_login(vn_port);
		fc_vport_setlink(vn_port);
	}
	return 0;
}

/**
 * fcoe_vport_destroy() - destroy the fc_host/scsi_host for a vport
 * @vport: fc_vport object that is being destroyed
 *
 * Returns: 0 for success
 */
static int fcoe_vport_destroy(struct fc_vport *vport)
{
	struct Scsi_Host *shost = vport_to_shost(vport);
	struct fc_lport *n_port = shost_priv(shost);
	struct fc_lport *vn_port = vport->dd_data;

	mutex_lock(&n_port->lp_mutex);
	list_del(&vn_port->list);
	mutex_unlock(&n_port->lp_mutex);
2721 2722 2723 2724 2725

	mutex_lock(&fcoe_config_mutex);
	fcoe_if_destroy(vn_port);
	mutex_unlock(&fcoe_config_mutex);

2726 2727 2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747 2748 2749
	return 0;
}

/**
 * fcoe_vport_disable() - change vport state
 * @vport: vport to bring online/offline
 * @disable: should the vport be disabled?
 */
static int fcoe_vport_disable(struct fc_vport *vport, bool disable)
{
	struct fc_lport *lport = vport->dd_data;

	if (disable) {
		fc_vport_set_state(vport, FC_VPORT_DISABLED);
		fc_fabric_logoff(lport);
	} else {
		lport->boot_time = jiffies;
		fc_fabric_login(lport);
		fc_vport_setlink(lport);
	}

	return 0;
}

2750 2751 2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767 2768 2769 2770 2771 2772 2773 2774 2775 2776 2777
/**
 * fcoe_vport_set_symbolic_name() - append vport string to symbolic name
 * @vport: fc_vport with a new symbolic name string
 *
 * After generating a new symbolic name string, a new RSPN_ID request is
 * sent to the name server.  There is no response handler, so if it fails
 * for some reason it will not be retried.
 */
static void fcoe_set_vport_symbolic_name(struct fc_vport *vport)
{
	struct fc_lport *lport = vport->dd_data;
	struct fc_frame *fp;
	size_t len;

	snprintf(fc_host_symbolic_name(lport->host), FC_SYMBOLIC_NAME_SIZE,
		 "%s v%s over %s : %s", FCOE_NAME, FCOE_VERSION,
		 fcoe_netdev(lport)->name, vport->symbolic_name);

	if (lport->state != LPORT_ST_READY)
		return;

	len = strnlen(fc_host_symbolic_name(lport->host), 255);
	fp = fc_frame_alloc(lport,
			    sizeof(struct fc_ct_hdr) +
			    sizeof(struct fc_ns_rspn) + len);
	if (!fp)
		return;
	lport->tt.elsct_send(lport, FC_FID_DIR_SERV, fp, FC_NS_RSPN_ID,
2778
			     NULL, NULL, 3 * lport->r_a_tov);
2779
}
2780 2781 2782 2783 2784 2785 2786 2787 2788 2789 2790

/**
 * fcoe_get_lesb() - Fill the FCoE Link Error Status Block
 * @lport: the local port
 * @fc_lesb: the link error status block
 */
static void fcoe_get_lesb(struct fc_lport *lport,
			 struct fc_els_lesb *fc_lesb)
{
	struct net_device *netdev = fcoe_netdev(lport);

2791
	__fcoe_get_lesb(lport, fc_lesb, netdev);
2792
}
2793

2794 2795 2796 2797 2798 2799 2800 2801 2802 2803 2804 2805 2806 2807 2808 2809 2810 2811 2812 2813 2814 2815 2816 2817 2818 2819 2820 2821 2822 2823 2824 2825 2826 2827
static void fcoe_ctlr_get_lesb(struct fcoe_ctlr_device *ctlr_dev)
{
	struct fcoe_ctlr *fip = fcoe_ctlr_device_priv(ctlr_dev);
	struct net_device *netdev = fcoe_netdev(fip->lp);
	struct fcoe_fc_els_lesb *fcoe_lesb;
	struct fc_els_lesb fc_lesb;

	__fcoe_get_lesb(fip->lp, &fc_lesb, netdev);
	fcoe_lesb = (struct fcoe_fc_els_lesb *)(&fc_lesb);

	ctlr_dev->lesb.lesb_link_fail =
		ntohl(fcoe_lesb->lesb_link_fail);
	ctlr_dev->lesb.lesb_vlink_fail =
		ntohl(fcoe_lesb->lesb_vlink_fail);
	ctlr_dev->lesb.lesb_miss_fka =
		ntohl(fcoe_lesb->lesb_miss_fka);
	ctlr_dev->lesb.lesb_symb_err =
		ntohl(fcoe_lesb->lesb_symb_err);
	ctlr_dev->lesb.lesb_err_block =
		ntohl(fcoe_lesb->lesb_err_block);
	ctlr_dev->lesb.lesb_fcs_error =
		ntohl(fcoe_lesb->lesb_fcs_error);
}

static void fcoe_fcf_get_vlan_id(struct fcoe_fcf_device *fcf_dev)
{
	struct fcoe_ctlr_device *ctlr_dev =
		fcoe_fcf_dev_to_ctlr_dev(fcf_dev);
	struct fcoe_ctlr *ctlr = fcoe_ctlr_device_priv(ctlr_dev);
	struct fcoe_interface *fcoe = fcoe_ctlr_priv(ctlr);

	fcf_dev->vlan_id = vlan_dev_vlan_id(fcoe->netdev);
}

2828 2829 2830 2831 2832 2833 2834 2835 2836 2837 2838 2839 2840 2841 2842 2843
/**
 * fcoe_set_port_id() - Callback from libfc when Port_ID is set.
 * @lport: the local port
 * @port_id: the port ID
 * @fp: the received frame, if any, that caused the port_id to be set.
 *
 * This routine handles the case where we received a FLOGI and are
 * entering point-to-point mode.  We need to call fcoe_ctlr_recv_flogi()
 * so it can set the non-mapped mode and gateway address.
 *
 * The FLOGI LS_ACC is handled by fcoe_flogi_resp().
 */
static void fcoe_set_port_id(struct fc_lport *lport,
			     u32 port_id, struct fc_frame *fp)
{
	struct fcoe_port *port = lport_priv(lport);
2844
	struct fcoe_interface *fcoe = port->priv;
2845
	struct fcoe_ctlr *ctlr = fcoe_to_ctlr(fcoe);
2846 2847

	if (fp && fc_frame_payload_op(fp) == ELS_FLOGI)
2848
		fcoe_ctlr_recv_flogi(ctlr, lport, fp);
2849
}