fec_main.c 91.5 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4
/*
 * Fast Ethernet Controller (FEC) driver for Motorola MPC8xx.
 * Copyright (c) 1997 Dan Malek (dmalek@jlc.net)
 *
5
 * Right now, I am very wasteful with the buffers.  I allocate memory
L
Linus Torvalds 已提交
6 7 8 9 10 11 12 13 14
 * pages and then divide them into 2K frame buffers.  This way I know I
 * have buffers large enough to hold one frame within one buffer descriptor.
 * Once I get this working, I will use 64 or 128 byte CPM buffers, which
 * will be much more memory efficient and will easily handle lots of
 * small packets.
 *
 * Much better multiple PHY support by Magnus Damm.
 * Copyright (c) 2000 Ericsson Radio Systems AB.
 *
15 16
 * Support for FEC controller of ColdFire processors.
 * Copyright (c) 2001-2005 Greg Ungerer (gerg@snapgear.com)
17 18
 *
 * Bug fixes and cleanup by Philippe De Muyter (phdm@macqel.be)
19
 * Copyright (c) 2004-2006 Macq Electronique SA.
20
 *
S
Shawn Guo 已提交
21
 * Copyright (C) 2010-2011 Freescale Semiconductor, Inc.
L
Linus Torvalds 已提交
22 23 24 25 26 27 28 29 30 31 32 33 34 35
 */

#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/string.h>
#include <linux/ptrace.h>
#include <linux/errno.h>
#include <linux/ioport.h>
#include <linux/slab.h>
#include <linux/interrupt.h>
#include <linux/delay.h>
#include <linux/netdevice.h>
#include <linux/etherdevice.h>
#include <linux/skbuff.h>
36 37 38
#include <linux/in.h>
#include <linux/ip.h>
#include <net/ip.h>
N
Nimrod Andy 已提交
39
#include <net/tso.h>
40 41 42
#include <linux/tcp.h>
#include <linux/udp.h>
#include <linux/icmp.h>
L
Linus Torvalds 已提交
43 44 45
#include <linux/spinlock.h>
#include <linux/workqueue.h>
#include <linux/bitops.h>
46 47
#include <linux/io.h>
#include <linux/irq.h>
48
#include <linux/clk.h>
49
#include <linux/platform_device.h>
50
#include <linux/phy.h>
51
#include <linux/fec.h>
52 53 54
#include <linux/of.h>
#include <linux/of_device.h>
#include <linux/of_gpio.h>
55
#include <linux/of_mdio.h>
56
#include <linux/of_net.h>
57
#include <linux/regulator/consumer.h>
58
#include <linux/if_vlan.h>
F
Fabio Estevam 已提交
59
#include <linux/pinctrl/consumer.h>
60
#include <linux/prefetch.h>
L
Linus Torvalds 已提交
61

62
#include <asm/cacheflush.h>
63

L
Linus Torvalds 已提交
64 65
#include "fec.h"

66
static void set_multicast_list(struct net_device *ndev);
67
static void fec_enet_itr_coal_init(struct net_device *ndev);
68

69 70
#define DRIVER_NAME	"fec"

71 72
#define FEC_ENET_GET_QUQUE(_x) ((_x == 0) ? 1 : ((_x == 1) ? 2 : 0))

73 74 75 76 77 78 79 80
/* Pause frame feild and FIFO threshold */
#define FEC_ENET_FCE	(1 << 5)
#define FEC_ENET_RSEM_V	0x84
#define FEC_ENET_RSFL_V	16
#define FEC_ENET_RAEM_V	0x8
#define FEC_ENET_RAFL_V	0x8
#define FEC_ENET_OPD_V	0xFFF0

81 82
static struct platform_device_id fec_devtype[] = {
	{
83
		/* keep it for coldfire */
84 85
		.name = DRIVER_NAME,
		.driver_data = 0,
86 87
	}, {
		.name = "imx25-fec",
88
		.driver_data = FEC_QUIRK_USE_GASKET | FEC_QUIRK_HAS_RACC,
89 90
	}, {
		.name = "imx27-fec",
91
		.driver_data = FEC_QUIRK_HAS_RACC,
92 93
	}, {
		.name = "imx28-fec",
94
		.driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_SWAP_FRAME |
95
				FEC_QUIRK_SINGLE_MDIO | FEC_QUIRK_HAS_RACC,
S
Shawn Guo 已提交
96 97
	}, {
		.name = "imx6q-fec",
98
		.driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
99
				FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
100 101
				FEC_QUIRK_HAS_VLAN | FEC_QUIRK_ERR006358 |
				FEC_QUIRK_HAS_RACC,
102
	}, {
103
		.name = "mvf600-fec",
104
		.driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_RACC,
105 106 107 108
	}, {
		.name = "imx6sx-fec",
		.driver_data = FEC_QUIRK_ENET_MAC | FEC_QUIRK_HAS_GBIT |
				FEC_QUIRK_HAS_BUFDESC_EX | FEC_QUIRK_HAS_CSUM |
109
				FEC_QUIRK_HAS_VLAN | FEC_QUIRK_HAS_AVB |
110 111
				FEC_QUIRK_ERR007885 | FEC_QUIRK_BUG_CAPTURE |
				FEC_QUIRK_HAS_RACC,
112 113 114
	}, {
		/* sentinel */
	}
115
};
116
MODULE_DEVICE_TABLE(platform, fec_devtype);
117

118
enum imx_fec_type {
L
Lothar Waßmann 已提交
119
	IMX25_FEC = 1,	/* runs on i.mx25/50/53 */
120 121
	IMX27_FEC,	/* runs on i.mx27/35/51 */
	IMX28_FEC,
S
Shawn Guo 已提交
122
	IMX6Q_FEC,
123
	MVF600_FEC,
124
	IMX6SX_FEC,
125 126 127 128 129 130
};

static const struct of_device_id fec_dt_ids[] = {
	{ .compatible = "fsl,imx25-fec", .data = &fec_devtype[IMX25_FEC], },
	{ .compatible = "fsl,imx27-fec", .data = &fec_devtype[IMX27_FEC], },
	{ .compatible = "fsl,imx28-fec", .data = &fec_devtype[IMX28_FEC], },
S
Shawn Guo 已提交
131
	{ .compatible = "fsl,imx6q-fec", .data = &fec_devtype[IMX6Q_FEC], },
132
	{ .compatible = "fsl,mvf600-fec", .data = &fec_devtype[MVF600_FEC], },
133
	{ .compatible = "fsl,imx6sx-fec", .data = &fec_devtype[IMX6SX_FEC], },
134 135 136 137
	{ /* sentinel */ }
};
MODULE_DEVICE_TABLE(of, fec_dt_ids);

138 139 140
static unsigned char macaddr[ETH_ALEN];
module_param_array(macaddr, byte, NULL, 0);
MODULE_PARM_DESC(macaddr, "FEC Ethernet MAC address");
L
Linus Torvalds 已提交
141

142
#if defined(CONFIG_M5272)
L
Linus Torvalds 已提交
143 144 145 146 147 148 149 150 151 152
/*
 * Some hardware gets it MAC address out of local flash memory.
 * if this is non-zero then assume it is the address to get MAC from.
 */
#if defined(CONFIG_NETtel)
#define	FEC_FLASHMAC	0xf0006006
#elif defined(CONFIG_GILBARCONAP) || defined(CONFIG_SCALES)
#define	FEC_FLASHMAC	0xf0006000
#elif defined(CONFIG_CANCam)
#define	FEC_FLASHMAC	0xf0020000
153 154 155
#elif defined (CONFIG_M5272C3)
#define	FEC_FLASHMAC	(0xffe04000 + 4)
#elif defined(CONFIG_MOD5272)
L
Lothar Waßmann 已提交
156
#define FEC_FLASHMAC	0xffc0406b
L
Linus Torvalds 已提交
157 158 159
#else
#define	FEC_FLASHMAC	0
#endif
160
#endif /* CONFIG_M5272 */
161

162
/* The FEC stores dest/src/type/vlan, data, and checksum for receive packets.
L
Linus Torvalds 已提交
163
 */
164
#define PKT_MAXBUF_SIZE		1522
L
Linus Torvalds 已提交
165
#define PKT_MINBUF_SIZE		64
166
#define PKT_MAXBLR_SIZE		1536
L
Linus Torvalds 已提交
167

168 169 170 171 172
/* FEC receive acceleration */
#define FEC_RACC_IPDIS		(1 << 1)
#define FEC_RACC_PRODIS		(1 << 2)
#define FEC_RACC_OPTIONS	(FEC_RACC_IPDIS | FEC_RACC_PRODIS)

L
Linus Torvalds 已提交
173
/*
174
 * The 5270/5271/5280/5282/532x RX control register also contains maximum frame
L
Linus Torvalds 已提交
175 176 177
 * size bits. Other FEC hardware does not, so we need to take that into
 * account when setting it.
 */
178
#if defined(CONFIG_M523x) || defined(CONFIG_M527x) || defined(CONFIG_M528x) || \
179
    defined(CONFIG_M520x) || defined(CONFIG_M532x) || defined(CONFIG_ARM)
L
Linus Torvalds 已提交
180 181 182 183 184
#define	OPT_FRAME_SIZE	(PKT_MAXBUF_SIZE << 16)
#else
#define	OPT_FRAME_SIZE	0
#endif

185 186 187 188 189 190 191 192
/* FEC MII MMFR bits definition */
#define FEC_MMFR_ST		(1 << 30)
#define FEC_MMFR_OP_READ	(2 << 28)
#define FEC_MMFR_OP_WRITE	(1 << 28)
#define FEC_MMFR_PA(v)		((v & 0x1f) << 23)
#define FEC_MMFR_RA(v)		((v & 0x1f) << 18)
#define FEC_MMFR_TA		(2 << 16)
#define FEC_MMFR_DATA(v)	(v & 0xffff)
N
Nimrod Andy 已提交
193 194 195
/* FEC ECR bits definition */
#define FEC_ECR_MAGICEN		(1 << 2)
#define FEC_ECR_SLEEP		(1 << 3)
L
Linus Torvalds 已提交
196

197
#define FEC_MII_TIMEOUT		30000 /* us */
L
Linus Torvalds 已提交
198

S
Sascha Hauer 已提交
199 200
/* Transmitter timeout */
#define TX_TIMEOUT (2 * HZ)
L
Linus Torvalds 已提交
201

202 203
#define FEC_PAUSE_FLAG_AUTONEG	0x1
#define FEC_PAUSE_FLAG_ENABLE	0x2
N
Nimrod Andy 已提交
204 205 206
#define FEC_WOL_HAS_MAGIC_PACKET	(0x1 << 0)
#define FEC_WOL_FLAG_ENABLE		(0x1 << 1)
#define FEC_WOL_FLAG_SLEEP_ON		(0x1 << 2)
207

208 209
#define COPYBREAK_DEFAULT	256

N
Nimrod Andy 已提交
210 211 212 213 214 215 216 217 218
#define TSO_HEADER_SIZE		128
/* Max number of allowed TCP segments for software TSO */
#define FEC_MAX_TSO_SEGS	100
#define FEC_MAX_SKB_DESCS	(FEC_MAX_TSO_SEGS * 2 + MAX_SKB_FRAGS)

#define IS_TSO_HEADER(txq, addr) \
	((addr >= txq->tso_hdrs_dma) && \
	(addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE))

L
Lothar Waßmann 已提交
219 220
static int mii_cnt;

221
static inline
222 223 224
struct bufdesc *fec_enet_get_nextdesc(struct bufdesc *bdp,
				      struct fec_enet_private *fep,
				      int queue_id)
225
{
226 227
	struct bufdesc *new_bd = bdp + 1;
	struct bufdesc_ex *ex_new_bd = (struct bufdesc_ex *)bdp + 1;
228 229
	struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue_id];
	struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id];
230 231 232 233
	struct bufdesc_ex *ex_base;
	struct bufdesc *base;
	int ring_size;

234 235 236 237
	if (bdp >= txq->tx_bd_base) {
		base = txq->tx_bd_base;
		ring_size = txq->tx_ring_size;
		ex_base = (struct bufdesc_ex *)txq->tx_bd_base;
238
	} else {
239 240 241
		base = rxq->rx_bd_base;
		ring_size = rxq->rx_ring_size;
		ex_base = (struct bufdesc_ex *)rxq->rx_bd_base;
242 243 244 245 246
	}

	if (fep->bufdesc_ex)
		return (struct bufdesc *)((ex_new_bd >= (ex_base + ring_size)) ?
			ex_base : ex_new_bd);
247
	else
248 249
		return (new_bd >= (base + ring_size)) ?
			base : new_bd;
250 251
}

252
static inline
253 254 255
struct bufdesc *fec_enet_get_prevdesc(struct bufdesc *bdp,
				      struct fec_enet_private *fep,
				      int queue_id)
256
{
257 258
	struct bufdesc *new_bd = bdp - 1;
	struct bufdesc_ex *ex_new_bd = (struct bufdesc_ex *)bdp - 1;
259 260
	struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue_id];
	struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue_id];
261 262 263 264
	struct bufdesc_ex *ex_base;
	struct bufdesc *base;
	int ring_size;

265 266 267 268
	if (bdp >= txq->tx_bd_base) {
		base = txq->tx_bd_base;
		ring_size = txq->tx_ring_size;
		ex_base = (struct bufdesc_ex *)txq->tx_bd_base;
269
	} else {
270 271 272
		base = rxq->rx_bd_base;
		ring_size = rxq->rx_ring_size;
		ex_base = (struct bufdesc_ex *)rxq->rx_bd_base;
273 274 275 276 277
	}

	if (fep->bufdesc_ex)
		return (struct bufdesc *)((ex_new_bd < ex_base) ?
			(ex_new_bd + ring_size) : ex_new_bd);
278
	else
279
		return (new_bd < base) ? (new_bd + ring_size) : new_bd;
280 281
}

282 283 284 285 286 287
static int fec_enet_get_bd_index(struct bufdesc *base, struct bufdesc *bdp,
				struct fec_enet_private *fep)
{
	return ((const char *)bdp - (const char *)base) / fep->bufdesc_size;
}

288 289
static int fec_enet_get_free_txdesc_num(struct fec_enet_private *fep,
					struct fec_enet_priv_tx_q *txq)
290 291 292
{
	int entries;

293 294
	entries = ((const char *)txq->dirty_tx -
			(const char *)txq->cur_tx) / fep->bufdesc_size - 1;
295

296
	return entries > 0 ? entries : entries + txq->tx_ring_size;
297 298
}

299
static void swap_buffer(void *bufaddr, int len)
300 301 302 303
{
	int i;
	unsigned int *buf = bufaddr;

304
	for (i = 0; i < len; i += 4, buf++)
305
		swab32s(buf);
306 307
}

308 309 310 311 312 313 314 315 316 317
static void swap_buffer2(void *dst_buf, void *src_buf, int len)
{
	int i;
	unsigned int *src = src_buf;
	unsigned int *dst = dst_buf;

	for (i = 0; i < len; i += 4, src++, dst++)
		*dst = swab32p(src);
}

318 319 320
static void fec_dump(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
321 322 323
	struct bufdesc *bdp;
	struct fec_enet_priv_tx_q *txq;
	int index = 0;
324 325 326 327

	netdev_info(ndev, "TX ring dump\n");
	pr_info("Nr     SC     addr       len  SKB\n");

328 329 330
	txq = fep->tx_queue[0];
	bdp = txq->tx_bd_base;

331 332 333
	do {
		pr_info("%3u %c%c 0x%04x 0x%08lx %4u %p\n",
			index,
334 335
			bdp == txq->cur_tx ? 'S' : ' ',
			bdp == txq->dirty_tx ? 'H' : ' ',
336
			bdp->cbd_sc, bdp->cbd_bufaddr, bdp->cbd_datlen,
337 338
			txq->tx_skbuff[index]);
		bdp = fec_enet_get_nextdesc(bdp, fep, 0);
339
		index++;
340
	} while (bdp != txq->tx_bd_base);
341 342
}

343 344 345 346 347
static inline bool is_ipv4_pkt(struct sk_buff *skb)
{
	return skb->protocol == htons(ETH_P_IP) && ip_hdr(skb)->version == 4;
}

348 349 350 351 352 353 354 355 356 357
static int
fec_enet_clear_csum(struct sk_buff *skb, struct net_device *ndev)
{
	/* Only run for packets requiring a checksum. */
	if (skb->ip_summed != CHECKSUM_PARTIAL)
		return 0;

	if (unlikely(skb_cow_head(skb, 0)))
		return -1;

358 359
	if (is_ipv4_pkt(skb))
		ip_hdr(skb)->check = 0;
360 361 362 363 364
	*(__sum16 *)(skb->head + skb->csum_start + skb->csum_offset) = 0;

	return 0;
}

365
static int
366 367 368
fec_enet_txq_submit_frag_skb(struct fec_enet_priv_tx_q *txq,
			     struct sk_buff *skb,
			     struct net_device *ndev)
L
Linus Torvalds 已提交
369
{
370
	struct fec_enet_private *fep = netdev_priv(ndev);
371
	struct bufdesc *bdp = txq->cur_tx;
372 373
	struct bufdesc_ex *ebdp;
	int nr_frags = skb_shinfo(skb)->nr_frags;
374
	unsigned short queue = skb_get_queue_mapping(skb);
375 376 377 378
	int frag, frag_len;
	unsigned short status;
	unsigned int estatus = 0;
	skb_frag_t *this_frag;
379
	unsigned int index;
380
	void *bufaddr;
381
	dma_addr_t addr;
382
	int i;
L
Linus Torvalds 已提交
383

384 385
	for (frag = 0; frag < nr_frags; frag++) {
		this_frag = &skb_shinfo(skb)->frags[frag];
386
		bdp = fec_enet_get_nextdesc(bdp, fep, queue);
387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405
		ebdp = (struct bufdesc_ex *)bdp;

		status = bdp->cbd_sc;
		status &= ~BD_ENET_TX_STATS;
		status |= (BD_ENET_TX_TC | BD_ENET_TX_READY);
		frag_len = skb_shinfo(skb)->frags[frag].size;

		/* Handle the last BD specially */
		if (frag == nr_frags - 1) {
			status |= (BD_ENET_TX_INTR | BD_ENET_TX_LAST);
			if (fep->bufdesc_ex) {
				estatus |= BD_ENET_TX_INT;
				if (unlikely(skb_shinfo(skb)->tx_flags &
					SKBTX_HW_TSTAMP && fep->hwts_tx_en))
					estatus |= BD_ENET_TX_TS;
			}
		}

		if (fep->bufdesc_ex) {
406
			if (fep->quirks & FEC_QUIRK_HAS_AVB)
407
				estatus |= FEC_TX_BD_FTYPE(queue);
408 409 410 411 412 413 414 415
			if (skb->ip_summed == CHECKSUM_PARTIAL)
				estatus |= BD_ENET_TX_PINS | BD_ENET_TX_IINS;
			ebdp->cbd_bdu = 0;
			ebdp->cbd_esc = estatus;
		}

		bufaddr = page_address(this_frag->page.p) + this_frag->page_offset;

416
		index = fec_enet_get_bd_index(txq->tx_bd_base, bdp, fep);
417
		if (((unsigned long) bufaddr) & fep->tx_align ||
418
			fep->quirks & FEC_QUIRK_SWAP_FRAME) {
419 420
			memcpy(txq->tx_bounce[index], bufaddr, frag_len);
			bufaddr = txq->tx_bounce[index];
421

422
			if (fep->quirks & FEC_QUIRK_SWAP_FRAME)
423 424 425
				swap_buffer(bufaddr, frag_len);
		}

426 427 428
		addr = dma_map_single(&fep->pdev->dev, bufaddr, frag_len,
				      DMA_TO_DEVICE);
		if (dma_mapping_error(&fep->pdev->dev, addr)) {
429 430 431 432 433 434
			dev_kfree_skb_any(skb);
			if (net_ratelimit())
				netdev_err(ndev, "Tx DMA memory map failed\n");
			goto dma_mapping_error;
		}

435
		bdp->cbd_bufaddr = addr;
436 437 438 439
		bdp->cbd_datlen = frag_len;
		bdp->cbd_sc = status;
	}

440
	txq->cur_tx = bdp;
441 442 443 444

	return 0;

dma_mapping_error:
445
	bdp = txq->cur_tx;
446
	for (i = 0; i < frag; i++) {
447
		bdp = fec_enet_get_nextdesc(bdp, fep, queue);
448 449 450 451 452
		dma_unmap_single(&fep->pdev->dev, bdp->cbd_bufaddr,
				bdp->cbd_datlen, DMA_TO_DEVICE);
	}
	return NETDEV_TX_OK;
}
L
Linus Torvalds 已提交
453

454 455
static int fec_enet_txq_submit_skb(struct fec_enet_priv_tx_q *txq,
				   struct sk_buff *skb, struct net_device *ndev)
456 457 458 459 460
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int nr_frags = skb_shinfo(skb)->nr_frags;
	struct bufdesc *bdp, *last_bdp;
	void *bufaddr;
461
	dma_addr_t addr;
462 463
	unsigned short status;
	unsigned short buflen;
464
	unsigned short queue;
465 466
	unsigned int estatus = 0;
	unsigned int index;
N
Nimrod Andy 已提交
467
	int entries_free;
468
	int ret;
S
Sascha Hauer 已提交
469

470
	entries_free = fec_enet_get_free_txdesc_num(fep, txq);
N
Nimrod Andy 已提交
471 472 473 474 475 476 477
	if (entries_free < MAX_SKB_FRAGS + 1) {
		dev_kfree_skb_any(skb);
		if (net_ratelimit())
			netdev_err(ndev, "NOT enough BD for SG!\n");
		return NETDEV_TX_OK;
	}

478 479
	/* Protocol checksum off-load for TCP and UDP. */
	if (fec_enet_clear_csum(skb, ndev)) {
480
		dev_kfree_skb_any(skb);
481 482 483
		return NETDEV_TX_OK;
	}

484
	/* Fill in a Tx ring entry */
485
	bdp = txq->cur_tx;
486
	status = bdp->cbd_sc;
487
	status &= ~BD_ENET_TX_STATS;
L
Linus Torvalds 已提交
488

S
Sascha Hauer 已提交
489
	/* Set buffer length and buffer pointer */
490
	bufaddr = skb->data;
491
	buflen = skb_headlen(skb);
L
Linus Torvalds 已提交
492

493 494
	queue = skb_get_queue_mapping(skb);
	index = fec_enet_get_bd_index(txq->tx_bd_base, bdp, fep);
495
	if (((unsigned long) bufaddr) & fep->tx_align ||
496
		fep->quirks & FEC_QUIRK_SWAP_FRAME) {
497 498
		memcpy(txq->tx_bounce[index], skb->data, buflen);
		bufaddr = txq->tx_bounce[index];
L
Linus Torvalds 已提交
499

500
		if (fep->quirks & FEC_QUIRK_SWAP_FRAME)
501 502
			swap_buffer(bufaddr, buflen);
	}
503

504 505 506
	/* Push the data cache so the CPM does not get stale memory data. */
	addr = dma_map_single(&fep->pdev->dev, bufaddr, buflen, DMA_TO_DEVICE);
	if (dma_mapping_error(&fep->pdev->dev, addr)) {
507 508 509 510 511
		dev_kfree_skb_any(skb);
		if (net_ratelimit())
			netdev_err(ndev, "Tx DMA memory map failed\n");
		return NETDEV_TX_OK;
	}
L
Linus Torvalds 已提交
512

513
	if (nr_frags) {
514
		ret = fec_enet_txq_submit_frag_skb(txq, skb, ndev);
515 516 517 518 519 520 521 522 523 524 525 526
		if (ret)
			return ret;
	} else {
		status |= (BD_ENET_TX_INTR | BD_ENET_TX_LAST);
		if (fep->bufdesc_ex) {
			estatus = BD_ENET_TX_INT;
			if (unlikely(skb_shinfo(skb)->tx_flags &
				SKBTX_HW_TSTAMP && fep->hwts_tx_en))
				estatus |= BD_ENET_TX_TS;
		}
	}

527 528 529
	if (fep->bufdesc_ex) {

		struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;
530

531
		if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP &&
532
			fep->hwts_tx_en))
533
			skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
534

535
		if (fep->quirks & FEC_QUIRK_HAS_AVB)
536 537
			estatus |= FEC_TX_BD_FTYPE(queue);

538 539 540 541 542
		if (skb->ip_summed == CHECKSUM_PARTIAL)
			estatus |= BD_ENET_TX_PINS | BD_ENET_TX_IINS;

		ebdp->cbd_bdu = 0;
		ebdp->cbd_esc = estatus;
543
	}
544

545 546
	last_bdp = txq->cur_tx;
	index = fec_enet_get_bd_index(txq->tx_bd_base, last_bdp, fep);
547
	/* Save skb pointer */
548
	txq->tx_skbuff[index] = skb;
549 550

	bdp->cbd_datlen = buflen;
551
	bdp->cbd_bufaddr = addr;
552

553 554 555
	/* Send it on its way.  Tell FEC it's ready, interrupt when done,
	 * it's the last BD of the frame, and to put the CRC on the end.
	 */
556
	status |= (BD_ENET_TX_READY | BD_ENET_TX_TC);
557 558
	bdp->cbd_sc = status;

S
Sascha Hauer 已提交
559
	/* If this was the last BD in the ring, start at the beginning again. */
560
	bdp = fec_enet_get_nextdesc(last_bdp, fep, queue);
L
Linus Torvalds 已提交
561

562 563
	skb_tx_timestamp(skb);

564
	txq->cur_tx = bdp;
565 566

	/* Trigger transmission start */
567
	writel(0, fep->hwp + FEC_X_DES_ACTIVE(queue));
L
Linus Torvalds 已提交
568

569
	return 0;
L
Linus Torvalds 已提交
570 571
}

N
Nimrod Andy 已提交
572
static int
573 574 575 576
fec_enet_txq_put_data_tso(struct fec_enet_priv_tx_q *txq, struct sk_buff *skb,
			  struct net_device *ndev,
			  struct bufdesc *bdp, int index, char *data,
			  int size, bool last_tcp, bool is_last)
577 578
{
	struct fec_enet_private *fep = netdev_priv(ndev);
579
	struct bufdesc_ex *ebdp = container_of(bdp, struct bufdesc_ex, desc);
580
	unsigned short queue = skb_get_queue_mapping(skb);
N
Nimrod Andy 已提交
581 582
	unsigned short status;
	unsigned int estatus = 0;
583
	dma_addr_t addr;
584 585

	status = bdp->cbd_sc;
N
Nimrod Andy 已提交
586
	status &= ~BD_ENET_TX_STATS;
587

N
Nimrod Andy 已提交
588 589
	status |= (BD_ENET_TX_TC | BD_ENET_TX_READY);

590
	if (((unsigned long) data) & fep->tx_align ||
591
		fep->quirks & FEC_QUIRK_SWAP_FRAME) {
592 593
		memcpy(txq->tx_bounce[index], data, size);
		data = txq->tx_bounce[index];
N
Nimrod Andy 已提交
594

595
		if (fep->quirks & FEC_QUIRK_SWAP_FRAME)
N
Nimrod Andy 已提交
596 597 598
			swap_buffer(data, size);
	}

599 600
	addr = dma_map_single(&fep->pdev->dev, data, size, DMA_TO_DEVICE);
	if (dma_mapping_error(&fep->pdev->dev, addr)) {
N
Nimrod Andy 已提交
601
		dev_kfree_skb_any(skb);
602
		if (net_ratelimit())
N
Nimrod Andy 已提交
603
			netdev_err(ndev, "Tx DMA memory map failed\n");
604 605 606
		return NETDEV_TX_BUSY;
	}

607 608 609
	bdp->cbd_datlen = size;
	bdp->cbd_bufaddr = addr;

N
Nimrod Andy 已提交
610
	if (fep->bufdesc_ex) {
611
		if (fep->quirks & FEC_QUIRK_HAS_AVB)
612
			estatus |= FEC_TX_BD_FTYPE(queue);
N
Nimrod Andy 已提交
613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633
		if (skb->ip_summed == CHECKSUM_PARTIAL)
			estatus |= BD_ENET_TX_PINS | BD_ENET_TX_IINS;
		ebdp->cbd_bdu = 0;
		ebdp->cbd_esc = estatus;
	}

	/* Handle the last BD specially */
	if (last_tcp)
		status |= (BD_ENET_TX_LAST | BD_ENET_TX_TC);
	if (is_last) {
		status |= BD_ENET_TX_INTR;
		if (fep->bufdesc_ex)
			ebdp->cbd_esc |= BD_ENET_TX_INT;
	}

	bdp->cbd_sc = status;

	return 0;
}

static int
634 635 636
fec_enet_txq_put_hdr_tso(struct fec_enet_priv_tx_q *txq,
			 struct sk_buff *skb, struct net_device *ndev,
			 struct bufdesc *bdp, int index)
N
Nimrod Andy 已提交
637 638 639
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int hdr_len = skb_transport_offset(skb) + tcp_hdrlen(skb);
640
	struct bufdesc_ex *ebdp = container_of(bdp, struct bufdesc_ex, desc);
641
	unsigned short queue = skb_get_queue_mapping(skb);
N
Nimrod Andy 已提交
642 643 644 645 646 647 648 649 650
	void *bufaddr;
	unsigned long dmabuf;
	unsigned short status;
	unsigned int estatus = 0;

	status = bdp->cbd_sc;
	status &= ~BD_ENET_TX_STATS;
	status |= (BD_ENET_TX_TC | BD_ENET_TX_READY);

651 652
	bufaddr = txq->tso_hdrs + index * TSO_HEADER_SIZE;
	dmabuf = txq->tso_hdrs_dma + index * TSO_HEADER_SIZE;
653
	if (((unsigned long)bufaddr) & fep->tx_align ||
654
		fep->quirks & FEC_QUIRK_SWAP_FRAME) {
655 656
		memcpy(txq->tx_bounce[index], skb->data, hdr_len);
		bufaddr = txq->tx_bounce[index];
N
Nimrod Andy 已提交
657

658
		if (fep->quirks & FEC_QUIRK_SWAP_FRAME)
N
Nimrod Andy 已提交
659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674
			swap_buffer(bufaddr, hdr_len);

		dmabuf = dma_map_single(&fep->pdev->dev, bufaddr,
					hdr_len, DMA_TO_DEVICE);
		if (dma_mapping_error(&fep->pdev->dev, dmabuf)) {
			dev_kfree_skb_any(skb);
			if (net_ratelimit())
				netdev_err(ndev, "Tx DMA memory map failed\n");
			return NETDEV_TX_BUSY;
		}
	}

	bdp->cbd_bufaddr = dmabuf;
	bdp->cbd_datlen = hdr_len;

	if (fep->bufdesc_ex) {
675
		if (fep->quirks & FEC_QUIRK_HAS_AVB)
676
			estatus |= FEC_TX_BD_FTYPE(queue);
N
Nimrod Andy 已提交
677 678 679 680 681 682 683 684 685 686 687
		if (skb->ip_summed == CHECKSUM_PARTIAL)
			estatus |= BD_ENET_TX_PINS | BD_ENET_TX_IINS;
		ebdp->cbd_bdu = 0;
		ebdp->cbd_esc = estatus;
	}

	bdp->cbd_sc = status;

	return 0;
}

688 689 690
static int fec_enet_txq_submit_tso(struct fec_enet_priv_tx_q *txq,
				   struct sk_buff *skb,
				   struct net_device *ndev)
N
Nimrod Andy 已提交
691 692 693 694
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int hdr_len = skb_transport_offset(skb) + tcp_hdrlen(skb);
	int total_len, data_left;
695 696
	struct bufdesc *bdp = txq->cur_tx;
	unsigned short queue = skb_get_queue_mapping(skb);
N
Nimrod Andy 已提交
697 698 699 700
	struct tso_t tso;
	unsigned int index = 0;
	int ret;

701
	if (tso_count_descs(skb) >= fec_enet_get_free_txdesc_num(fep, txq)) {
N
Nimrod Andy 已提交
702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720
		dev_kfree_skb_any(skb);
		if (net_ratelimit())
			netdev_err(ndev, "NOT enough BD for TSO!\n");
		return NETDEV_TX_OK;
	}

	/* Protocol checksum off-load for TCP and UDP. */
	if (fec_enet_clear_csum(skb, ndev)) {
		dev_kfree_skb_any(skb);
		return NETDEV_TX_OK;
	}

	/* Initialize the TSO handler, and prepare the first payload */
	tso_start(skb, &tso);

	total_len = skb->len - hdr_len;
	while (total_len > 0) {
		char *hdr;

721
		index = fec_enet_get_bd_index(txq->tx_bd_base, bdp, fep);
N
Nimrod Andy 已提交
722 723 724 725
		data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len);
		total_len -= data_left;

		/* prepare packet headers: MAC + IP + TCP */
726
		hdr = txq->tso_hdrs + index * TSO_HEADER_SIZE;
N
Nimrod Andy 已提交
727
		tso_build_hdr(skb, hdr, &tso, data_left, total_len == 0);
728
		ret = fec_enet_txq_put_hdr_tso(txq, skb, ndev, bdp, index);
N
Nimrod Andy 已提交
729 730 731 732 733 734 735
		if (ret)
			goto err_release;

		while (data_left > 0) {
			int size;

			size = min_t(int, tso.size, data_left);
736 737 738 739 740 741 742
			bdp = fec_enet_get_nextdesc(bdp, fep, queue);
			index = fec_enet_get_bd_index(txq->tx_bd_base,
						      bdp, fep);
			ret = fec_enet_txq_put_data_tso(txq, skb, ndev,
							bdp, index,
							tso.data, size,
							size == data_left,
N
Nimrod Andy 已提交
743 744 745 746 747 748 749 750
							total_len == 0);
			if (ret)
				goto err_release;

			data_left -= size;
			tso_build_data(skb, &tso, size);
		}

751
		bdp = fec_enet_get_nextdesc(bdp, fep, queue);
N
Nimrod Andy 已提交
752 753 754
	}

	/* Save skb pointer */
755
	txq->tx_skbuff[index] = skb;
N
Nimrod Andy 已提交
756 757

	skb_tx_timestamp(skb);
758
	txq->cur_tx = bdp;
N
Nimrod Andy 已提交
759 760

	/* Trigger transmission start */
761
	if (!(fep->quirks & FEC_QUIRK_ERR007885) ||
762 763 764 765 766
	    !readl(fep->hwp + FEC_X_DES_ACTIVE(queue)) ||
	    !readl(fep->hwp + FEC_X_DES_ACTIVE(queue)) ||
	    !readl(fep->hwp + FEC_X_DES_ACTIVE(queue)) ||
	    !readl(fep->hwp + FEC_X_DES_ACTIVE(queue)))
		writel(0, fep->hwp + FEC_X_DES_ACTIVE(queue));
N
Nimrod Andy 已提交
767 768 769 770 771 772 773 774 775 776 777 778 779

	return 0;

err_release:
	/* TODO: Release all used data descriptors for TSO */
	return ret;
}

static netdev_tx_t
fec_enet_start_xmit(struct sk_buff *skb, struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int entries_free;
780 781 782
	unsigned short queue;
	struct fec_enet_priv_tx_q *txq;
	struct netdev_queue *nq;
N
Nimrod Andy 已提交
783 784
	int ret;

785 786 787 788
	queue = skb_get_queue_mapping(skb);
	txq = fep->tx_queue[queue];
	nq = netdev_get_tx_queue(ndev, queue);

N
Nimrod Andy 已提交
789
	if (skb_is_gso(skb))
790
		ret = fec_enet_txq_submit_tso(txq, skb, ndev);
N
Nimrod Andy 已提交
791
	else
792
		ret = fec_enet_txq_submit_skb(txq, skb, ndev);
793 794
	if (ret)
		return ret;
795

796 797 798
	entries_free = fec_enet_get_free_txdesc_num(fep, txq);
	if (entries_free <= txq->tx_stop_threshold)
		netif_tx_stop_queue(nq);
799 800 801 802

	return NETDEV_TX_OK;
}

803 804 805 806 807
/* Init RX & TX buffer descriptors
 */
static void fec_enet_bd_init(struct net_device *dev)
{
	struct fec_enet_private *fep = netdev_priv(dev);
808 809
	struct fec_enet_priv_tx_q *txq;
	struct fec_enet_priv_rx_q *rxq;
810 811
	struct bufdesc *bdp;
	unsigned int i;
812
	unsigned int q;
813

814 815 816 817
	for (q = 0; q < fep->num_rx_queues; q++) {
		/* Initialize the receive buffer descriptors. */
		rxq = fep->rx_queue[q];
		bdp = rxq->rx_bd_base;
818

819
		for (i = 0; i < rxq->rx_ring_size; i++) {
820

821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843
			/* Initialize the BD for every fragment in the page. */
			if (bdp->cbd_bufaddr)
				bdp->cbd_sc = BD_ENET_RX_EMPTY;
			else
				bdp->cbd_sc = 0;
			bdp = fec_enet_get_nextdesc(bdp, fep, q);
		}

		/* Set the last buffer to wrap */
		bdp = fec_enet_get_prevdesc(bdp, fep, q);
		bdp->cbd_sc |= BD_SC_WRAP;

		rxq->cur_rx = rxq->rx_bd_base;
	}

	for (q = 0; q < fep->num_tx_queues; q++) {
		/* ...and the same for transmit */
		txq = fep->tx_queue[q];
		bdp = txq->tx_bd_base;
		txq->cur_tx = bdp;

		for (i = 0; i < txq->tx_ring_size; i++) {
			/* Initialize the BD for every fragment in the page. */
844
			bdp->cbd_sc = 0;
845 846 847 848 849 850 851 852 853 854 855 856
			if (txq->tx_skbuff[i]) {
				dev_kfree_skb_any(txq->tx_skbuff[i]);
				txq->tx_skbuff[i] = NULL;
			}
			bdp->cbd_bufaddr = 0;
			bdp = fec_enet_get_nextdesc(bdp, fep, q);
		}

		/* Set the last buffer to wrap */
		bdp = fec_enet_get_prevdesc(bdp, fep, q);
		bdp->cbd_sc |= BD_SC_WRAP;
		txq->dirty_tx = bdp;
857
	}
858
}
859

F
Frank Li 已提交
860 861 862 863 864 865 866 867 868
static void fec_enet_active_rxring(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int i;

	for (i = 0; i < fep->num_rx_queues; i++)
		writel(0, fep->hwp + FEC_R_DES_ACTIVE(i));
}

869 870 871 872 873 874
static void fec_enet_enable_ring(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	struct fec_enet_priv_tx_q *txq;
	struct fec_enet_priv_rx_q *rxq;
	int i;
875

876 877 878
	for (i = 0; i < fep->num_rx_queues; i++) {
		rxq = fep->rx_queue[i];
		writel(rxq->bd_dma, fep->hwp + FEC_R_DES_START(i));
879
		writel(PKT_MAXBLR_SIZE, fep->hwp + FEC_R_BUFF_SIZE(i));
880

881 882 883 884 885
		/* enable DMA1/2 */
		if (i)
			writel(RCMR_MATCHEN | RCMR_CMP(i),
			       fep->hwp + FEC_RCMR(i));
	}
886

887 888 889 890 891 892 893 894
	for (i = 0; i < fep->num_tx_queues; i++) {
		txq = fep->tx_queue[i];
		writel(txq->bd_dma, fep->hwp + FEC_X_DES_START(i));

		/* enable DMA1/2 */
		if (i)
			writel(DMA_CLASS_EN | IDLE_SLOPE(i),
			       fep->hwp + FEC_DMA_CFG(i));
895
	}
896
}
897

898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913
static void fec_enet_reset_skb(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	struct fec_enet_priv_tx_q *txq;
	int i, j;

	for (i = 0; i < fep->num_tx_queues; i++) {
		txq = fep->tx_queue[i];

		for (j = 0; j < txq->tx_ring_size; j++) {
			if (txq->tx_skbuff[j]) {
				dev_kfree_skb_any(txq->tx_skbuff[j]);
				txq->tx_skbuff[j] = NULL;
			}
		}
	}
914 915
}

916 917 918 919
/*
 * This function is called to start or restart the FEC during a link
 * change, transmit timeout, or to reconfigure the FEC.  The network
 * packet processing for this device must be stopped before this call.
920
 */
L
Linus Torvalds 已提交
921
static void
922
fec_restart(struct net_device *ndev)
L
Linus Torvalds 已提交
923
{
924
	struct fec_enet_private *fep = netdev_priv(ndev);
925
	u32 val;
926 927
	u32 temp_mac[2];
	u32 rcntl = OPT_FRAME_SIZE | 0x04;
S
Shawn Guo 已提交
928
	u32 ecntl = 0x2; /* ETHEREN */
L
Linus Torvalds 已提交
929

930 931 932 933
	/* Whack a reset.  We should wait for this.
	 * For i.MX6SX SOC, enet use AXI bus, we use disable MAC
	 * instead of reset MAC itself.
	 */
934
	if (fep->quirks & FEC_QUIRK_HAS_AVB) {
935 936 937 938 939
		writel(0, fep->hwp + FEC_ECNTRL);
	} else {
		writel(1, fep->hwp + FEC_ECNTRL);
		udelay(10);
	}
L
Linus Torvalds 已提交
940

941 942 943 944
	/*
	 * enet-mac reset will reset mac address registers too,
	 * so need to reconfigure it.
	 */
945
	if (fep->quirks & FEC_QUIRK_ENET_MAC) {
946 947 948 949
		memcpy(&temp_mac, ndev->dev_addr, ETH_ALEN);
		writel(cpu_to_be32(temp_mac[0]), fep->hwp + FEC_ADDR_LOW);
		writel(cpu_to_be32(temp_mac[1]), fep->hwp + FEC_ADDR_HIGH);
	}
L
Linus Torvalds 已提交
950

951
	/* Clear any outstanding interrupt. */
952
	writel(0xffffffff, fep->hwp + FEC_IEVENT);
L
Linus Torvalds 已提交
953

954 955
	fec_enet_bd_init(ndev);

956
	fec_enet_enable_ring(ndev);
957

958 959
	/* Reset tx SKB buffers. */
	fec_enet_reset_skb(ndev);
960

961
	/* Enable MII mode */
962
	if (fep->full_duplex == DUPLEX_FULL) {
963
		/* FD enable */
964 965
		writel(0x04, fep->hwp + FEC_X_CNTRL);
	} else {
966 967
		/* No Rcv on Xmit */
		rcntl |= 0x02;
968 969
		writel(0x0, fep->hwp + FEC_X_CNTRL);
	}
970

971 972 973
	/* Set MII speed */
	writel(fep->phy_speed, fep->hwp + FEC_MII_SPEED);

G
Guenter Roeck 已提交
974
#if !defined(CONFIG_M5272)
975 976 977 978 979 980 981 982 983
	if (fep->quirks & FEC_QUIRK_HAS_RACC) {
		/* set RX checksum */
		val = readl(fep->hwp + FEC_RACC);
		if (fep->csum_flags & FLAG_RX_CSUM_ENABLED)
			val |= FEC_RACC_OPTIONS;
		else
			val &= ~FEC_RACC_OPTIONS;
		writel(val, fep->hwp + FEC_RACC);
	}
G
Guenter Roeck 已提交
984
#endif
985

986 987 988 989
	/*
	 * The phy interface and speed need to get configured
	 * differently on enet-mac.
	 */
990
	if (fep->quirks & FEC_QUIRK_ENET_MAC) {
991 992
		/* Enable flow control and length check */
		rcntl |= 0x40000000 | 0x00000020;
993

S
Shawn Guo 已提交
994
		/* RGMII, RMII or MII */
M
Markus Pargmann 已提交
995 996 997 998
		if (fep->phy_interface == PHY_INTERFACE_MODE_RGMII ||
		    fep->phy_interface == PHY_INTERFACE_MODE_RGMII_ID ||
		    fep->phy_interface == PHY_INTERFACE_MODE_RGMII_RXID ||
		    fep->phy_interface == PHY_INTERFACE_MODE_RGMII_TXID)
S
Shawn Guo 已提交
999 1000
			rcntl |= (1 << 6);
		else if (fep->phy_interface == PHY_INTERFACE_MODE_RMII)
1001
			rcntl |= (1 << 8);
1002
		else
1003
			rcntl &= ~(1 << 8);
1004

S
Shawn Guo 已提交
1005 1006 1007 1008 1009 1010 1011 1012 1013
		/* 1G, 100M or 10M */
		if (fep->phy_dev) {
			if (fep->phy_dev->speed == SPEED_1000)
				ecntl |= (1 << 5);
			else if (fep->phy_dev->speed == SPEED_100)
				rcntl &= ~(1 << 9);
			else
				rcntl |= (1 << 9);
		}
1014 1015
	} else {
#ifdef FEC_MIIGSK_ENR
1016
		if (fep->quirks & FEC_QUIRK_USE_GASKET) {
1017
			u32 cfgr;
1018 1019 1020 1021 1022 1023 1024 1025
			/* disable the gasket and wait */
			writel(0, fep->hwp + FEC_MIIGSK_ENR);
			while (readl(fep->hwp + FEC_MIIGSK_ENR) & 4)
				udelay(1);

			/*
			 * configure the gasket:
			 *   RMII, 50 MHz, no loopback, no echo
1026
			 *   MII, 25 MHz, no loopback, no echo
1027
			 */
1028 1029 1030 1031 1032
			cfgr = (fep->phy_interface == PHY_INTERFACE_MODE_RMII)
				? BM_MIIGSK_CFGR_RMII : BM_MIIGSK_CFGR_MII;
			if (fep->phy_dev && fep->phy_dev->speed == SPEED_10)
				cfgr |= BM_MIIGSK_CFGR_FRCONT_10M;
			writel(cfgr, fep->hwp + FEC_MIIGSK_CFGR);
1033 1034 1035

			/* re-enable the gasket */
			writel(2, fep->hwp + FEC_MIIGSK_ENR);
1036
		}
1037 1038
#endif
	}
1039

G
Guenter Roeck 已提交
1040
#if !defined(CONFIG_M5272)
1041 1042 1043 1044 1045 1046
	/* enable pause frame*/
	if ((fep->pause_flag & FEC_PAUSE_FLAG_ENABLE) ||
	    ((fep->pause_flag & FEC_PAUSE_FLAG_AUTONEG) &&
	     fep->phy_dev && fep->phy_dev->pause)) {
		rcntl |= FEC_ENET_FCE;

1047
		/* set FIFO threshold parameter to reduce overrun */
1048 1049 1050 1051 1052 1053 1054 1055 1056 1057
		writel(FEC_ENET_RSEM_V, fep->hwp + FEC_R_FIFO_RSEM);
		writel(FEC_ENET_RSFL_V, fep->hwp + FEC_R_FIFO_RSFL);
		writel(FEC_ENET_RAEM_V, fep->hwp + FEC_R_FIFO_RAEM);
		writel(FEC_ENET_RAFL_V, fep->hwp + FEC_R_FIFO_RAFL);

		/* OPD */
		writel(FEC_ENET_OPD_V, fep->hwp + FEC_OPD);
	} else {
		rcntl &= ~FEC_ENET_FCE;
	}
G
Guenter Roeck 已提交
1058
#endif /* !defined(CONFIG_M5272) */
1059

1060
	writel(rcntl, fep->hwp + FEC_R_CNTRL);
1061

1062 1063 1064 1065 1066 1067 1068
	/* Setup multicast filter. */
	set_multicast_list(ndev);
#ifndef CONFIG_M5272
	writel(0, fep->hwp + FEC_HASH_TABLE_HIGH);
	writel(0, fep->hwp + FEC_HASH_TABLE_LOW);
#endif

1069
	if (fep->quirks & FEC_QUIRK_ENET_MAC) {
S
Shawn Guo 已提交
1070 1071 1072 1073 1074 1075
		/* enable ENET endian swap */
		ecntl |= (1 << 8);
		/* enable ENET store and forward mode */
		writel(1 << 8, fep->hwp + FEC_X_WMRK);
	}

1076 1077
	if (fep->bufdesc_ex)
		ecntl |= (1 << 4);
1078

1079
#ifndef CONFIG_M5272
1080 1081
	/* Enable the MIB statistic event counters */
	writel(0 << 31, fep->hwp + FEC_MIB_CTRLSTAT);
1082 1083
#endif

1084
	/* And last, enable the transmit and receive processing */
S
Shawn Guo 已提交
1085
	writel(ecntl, fep->hwp + FEC_ECNTRL);
F
Frank Li 已提交
1086
	fec_enet_active_rxring(ndev);
1087

1088 1089 1090
	if (fep->bufdesc_ex)
		fec_ptp_start_cyclecounter(ndev);

1091
	/* Enable interrupts we wish to service */
1092 1093 1094 1095
	if (fep->link)
		writel(FEC_DEFAULT_IMASK, fep->hwp + FEC_IMASK);
	else
		writel(FEC_ENET_MII, fep->hwp + FEC_IMASK);
1096 1097 1098 1099

	/* Init the interrupt coalescing */
	fec_enet_itr_coal_init(ndev);

1100 1101 1102 1103 1104 1105
}

static void
fec_stop(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
N
Nimrod Andy 已提交
1106
	struct fec_platform_data *pdata = fep->pdev->dev.platform_data;
1107
	u32 rmii_mode = readl(fep->hwp + FEC_R_CNTRL) & (1 << 8);
N
Nimrod Andy 已提交
1108
	u32 val;
1109 1110 1111 1112 1113 1114

	/* We cannot expect a graceful transmit stop without link !!! */
	if (fep->link) {
		writel(1, fep->hwp + FEC_X_CNTRL); /* Graceful transmit stop */
		udelay(10);
		if (!(readl(fep->hwp + FEC_IEVENT) & FEC_ENET_GRA))
1115
			netdev_err(ndev, "Graceful transmit stop did not complete!\n");
1116 1117
	}

1118 1119 1120 1121
	/* Whack a reset.  We should wait for this.
	 * For i.MX6SX SOC, enet use AXI bus, we use disable MAC
	 * instead of reset MAC itself.
	 */
N
Nimrod Andy 已提交
1122 1123 1124 1125 1126 1127 1128 1129
	if (!(fep->wol_flag & FEC_WOL_FLAG_SLEEP_ON)) {
		if (fep->quirks & FEC_QUIRK_HAS_AVB) {
			writel(0, fep->hwp + FEC_ECNTRL);
		} else {
			writel(1, fep->hwp + FEC_ECNTRL);
			udelay(10);
		}
		writel(FEC_DEFAULT_IMASK, fep->hwp + FEC_IMASK);
1130
	} else {
N
Nimrod Andy 已提交
1131 1132 1133 1134 1135 1136 1137
		writel(FEC_DEFAULT_IMASK | FEC_ENET_WAKEUP, fep->hwp + FEC_IMASK);
		val = readl(fep->hwp + FEC_ECNTRL);
		val |= (FEC_ECR_MAGICEN | FEC_ECR_SLEEP);
		writel(val, fep->hwp + FEC_ECNTRL);

		if (pdata && pdata->sleep_mode_enable)
			pdata->sleep_mode_enable(true);
1138
	}
1139
	writel(fep->phy_speed, fep->hwp + FEC_MII_SPEED);
S
Shawn Guo 已提交
1140 1141

	/* We have to keep ENET enabled to have MII interrupt stay working */
N
Nimrod Andy 已提交
1142 1143
	if (fep->quirks & FEC_QUIRK_ENET_MAC &&
		!(fep->wol_flag & FEC_WOL_FLAG_SLEEP_ON)) {
S
Shawn Guo 已提交
1144
		writel(2, fep->hwp + FEC_ECNTRL);
1145 1146
		writel(rmii_mode, fep->hwp + FEC_R_CNTRL);
	}
L
Linus Torvalds 已提交
1147 1148 1149
}


1150 1151 1152 1153 1154
static void
fec_timeout(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

1155 1156
	fec_dump(ndev);

1157 1158
	ndev->stats.tx_errors++;

1159
	schedule_work(&fep->tx_timeout_work);
1160 1161
}

1162
static void fec_enet_timeout_work(struct work_struct *work)
1163 1164
{
	struct fec_enet_private *fep =
1165
		container_of(work, struct fec_enet_private, tx_timeout_work);
1166
	struct net_device *ndev = fep->netdev;
1167

1168 1169 1170 1171 1172 1173 1174 1175
	rtnl_lock();
	if (netif_device_present(ndev) || netif_running(ndev)) {
		napi_disable(&fep->napi);
		netif_tx_lock_bh(ndev);
		fec_restart(ndev);
		netif_wake_queue(ndev);
		netif_tx_unlock_bh(ndev);
		napi_enable(&fep->napi);
1176
	}
1177
	rtnl_unlock();
1178 1179
}

1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194
static void
fec_enet_hwtstamp(struct fec_enet_private *fep, unsigned ts,
	struct skb_shared_hwtstamps *hwtstamps)
{
	unsigned long flags;
	u64 ns;

	spin_lock_irqsave(&fep->tmreg_lock, flags);
	ns = timecounter_cyc2time(&fep->tc, ts);
	spin_unlock_irqrestore(&fep->tmreg_lock, flags);

	memset(hwtstamps, 0, sizeof(*hwtstamps));
	hwtstamps->hwtstamp = ns_to_ktime(ns);
}

L
Linus Torvalds 已提交
1195
static void
1196
fec_enet_tx_queue(struct net_device *ndev, u16 queue_id)
L
Linus Torvalds 已提交
1197 1198
{
	struct	fec_enet_private *fep;
1199
	struct bufdesc *bdp;
1200
	unsigned short status;
L
Linus Torvalds 已提交
1201
	struct	sk_buff	*skb;
1202 1203
	struct fec_enet_priv_tx_q *txq;
	struct netdev_queue *nq;
1204
	int	index = 0;
N
Nimrod Andy 已提交
1205
	int	entries_free;
L
Linus Torvalds 已提交
1206

1207
	fep = netdev_priv(ndev);
1208 1209 1210 1211 1212 1213 1214

	queue_id = FEC_ENET_GET_QUQUE(queue_id);

	txq = fep->tx_queue[queue_id];
	/* get next bdp of dirty_tx */
	nq = netdev_get_tx_queue(ndev, queue_id);
	bdp = txq->dirty_tx;
L
Linus Torvalds 已提交
1215

1216
	/* get next bdp of dirty_tx */
1217
	bdp = fec_enet_get_nextdesc(bdp, fep, queue_id);
1218

1219
	while (((status = bdp->cbd_sc) & BD_ENET_TX_READY) == 0) {
1220 1221

		/* current queue is empty */
1222
		if (bdp == txq->cur_tx)
S
Sascha Hauer 已提交
1223 1224
			break;

1225
		index = fec_enet_get_bd_index(txq->tx_bd_base, bdp, fep);
1226

1227
		skb = txq->tx_skbuff[index];
1228
		txq->tx_skbuff[index] = NULL;
1229 1230 1231 1232 1233 1234 1235 1236
		if (!IS_TSO_HEADER(txq, bdp->cbd_bufaddr))
			dma_unmap_single(&fep->pdev->dev, bdp->cbd_bufaddr,
					bdp->cbd_datlen, DMA_TO_DEVICE);
		bdp->cbd_bufaddr = 0;
		if (!skb) {
			bdp = fec_enet_get_nextdesc(bdp, fep, queue_id);
			continue;
		}
1237

L
Linus Torvalds 已提交
1238
		/* Check for errors. */
1239
		if (status & (BD_ENET_TX_HB | BD_ENET_TX_LC |
L
Linus Torvalds 已提交
1240 1241
				   BD_ENET_TX_RL | BD_ENET_TX_UN |
				   BD_ENET_TX_CSL)) {
1242
			ndev->stats.tx_errors++;
1243
			if (status & BD_ENET_TX_HB)  /* No heartbeat */
1244
				ndev->stats.tx_heartbeat_errors++;
1245
			if (status & BD_ENET_TX_LC)  /* Late collision */
1246
				ndev->stats.tx_window_errors++;
1247
			if (status & BD_ENET_TX_RL)  /* Retrans limit */
1248
				ndev->stats.tx_aborted_errors++;
1249
			if (status & BD_ENET_TX_UN)  /* Underrun */
1250
				ndev->stats.tx_fifo_errors++;
1251
			if (status & BD_ENET_TX_CSL) /* Carrier lost */
1252
				ndev->stats.tx_carrier_errors++;
L
Linus Torvalds 已提交
1253
		} else {
1254
			ndev->stats.tx_packets++;
1255
			ndev->stats.tx_bytes += skb->len;
L
Linus Torvalds 已提交
1256 1257
		}

1258 1259
		if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS) &&
			fep->bufdesc_ex) {
1260
			struct skb_shared_hwtstamps shhwtstamps;
1261
			struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;
1262

1263
			fec_enet_hwtstamp(fep, ebdp->ts, &shhwtstamps);
1264 1265
			skb_tstamp_tx(skb, &shhwtstamps);
		}
1266

L
Linus Torvalds 已提交
1267 1268 1269
		/* Deferred means some collisions occurred during transmit,
		 * but we eventually sent the packet OK.
		 */
1270
		if (status & BD_ENET_TX_DEF)
1271
			ndev->stats.collisions++;
1272

S
Sascha Hauer 已提交
1273
		/* Free the sk buffer associated with this last transmit */
L
Linus Torvalds 已提交
1274
		dev_kfree_skb_any(skb);
1275

1276
		txq->dirty_tx = bdp;
1277

S
Sascha Hauer 已提交
1278
		/* Update pointer to next buffer descriptor to be transmitted */
1279
		bdp = fec_enet_get_nextdesc(bdp, fep, queue_id);
1280

S
Sascha Hauer 已提交
1281
		/* Since we have freed up a buffer, the ring is no longer full
L
Linus Torvalds 已提交
1282
		 */
N
Nimrod Andy 已提交
1283
		if (netif_queue_stopped(ndev)) {
1284 1285 1286
			entries_free = fec_enet_get_free_txdesc_num(fep, txq);
			if (entries_free >= txq->tx_wake_threshold)
				netif_tx_wake_queue(nq);
N
Nimrod Andy 已提交
1287
		}
L
Linus Torvalds 已提交
1288
	}
1289 1290

	/* ERR006538: Keep the transmitter going */
1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306
	if (bdp != txq->cur_tx &&
	    readl(fep->hwp + FEC_X_DES_ACTIVE(queue_id)) == 0)
		writel(0, fep->hwp + FEC_X_DES_ACTIVE(queue_id));
}

static void
fec_enet_tx(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	u16 queue_id;
	/* First process class A queue, then Class B and Best Effort queue */
	for_each_set_bit(queue_id, &fep->work_tx, FEC_ENET_MAX_TX_QS) {
		clear_bit(queue_id, &fep->work_tx);
		fec_enet_tx_queue(ndev, queue_id);
	}
	return;
L
Linus Torvalds 已提交
1307 1308
}

1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331
static int
fec_enet_new_rxbdp(struct net_device *ndev, struct bufdesc *bdp, struct sk_buff *skb)
{
	struct  fec_enet_private *fep = netdev_priv(ndev);
	int off;

	off = ((unsigned long)skb->data) & fep->rx_align;
	if (off)
		skb_reserve(skb, fep->rx_align + 1 - off);

	bdp->cbd_bufaddr = dma_map_single(&fep->pdev->dev, skb->data,
					  FEC_ENET_RX_FRSIZE - fep->rx_align,
					  DMA_FROM_DEVICE);
	if (dma_mapping_error(&fep->pdev->dev, bdp->cbd_bufaddr)) {
		if (net_ratelimit())
			netdev_err(ndev, "Rx DMA memory map failed\n");
		return -ENOMEM;
	}

	return 0;
}

static bool fec_enet_copybreak(struct net_device *ndev, struct sk_buff **skb,
1332
			       struct bufdesc *bdp, u32 length, bool swap)
1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346
{
	struct  fec_enet_private *fep = netdev_priv(ndev);
	struct sk_buff *new_skb;

	if (length > fep->rx_copybreak)
		return false;

	new_skb = netdev_alloc_skb(ndev, length);
	if (!new_skb)
		return false;

	dma_sync_single_for_cpu(&fep->pdev->dev, bdp->cbd_bufaddr,
				FEC_ENET_RX_FRSIZE - fep->rx_align,
				DMA_FROM_DEVICE);
1347 1348 1349 1350
	if (!swap)
		memcpy(new_skb->data, (*skb)->data, length);
	else
		swap_buffer2(new_skb->data, (*skb)->data, length);
1351 1352 1353 1354 1355
	*skb = new_skb;

	return true;
}

L
Linus Torvalds 已提交
1356 1357 1358 1359 1360
/* During a receive, the cur_rx points to the current incoming buffer.
 * When we update through the ring, if the next incoming buffer has
 * not been given to the system, we just set the empty indicator,
 * effectively tossing the packet.
 */
1361
static int
1362
fec_enet_rx_queue(struct net_device *ndev, int budget, u16 queue_id)
L
Linus Torvalds 已提交
1363
{
1364
	struct fec_enet_private *fep = netdev_priv(ndev);
1365
	struct fec_enet_priv_rx_q *rxq;
S
Sascha Hauer 已提交
1366
	struct bufdesc *bdp;
1367
	unsigned short status;
1368 1369
	struct  sk_buff *skb_new = NULL;
	struct  sk_buff *skb;
L
Linus Torvalds 已提交
1370 1371
	ushort	pkt_len;
	__u8 *data;
1372
	int	pkt_received = 0;
1373 1374 1375
	struct	bufdesc_ex *ebdp = NULL;
	bool	vlan_packet_rcvd = false;
	u16	vlan_tag;
1376
	int	index = 0;
1377
	bool	is_copybreak;
1378
	bool	need_swap = fep->quirks & FEC_QUIRK_SWAP_FRAME;
1379

1380 1381
#ifdef CONFIG_M532x
	flush_cache_all();
1382
#endif
1383 1384
	queue_id = FEC_ENET_GET_QUQUE(queue_id);
	rxq = fep->rx_queue[queue_id];
L
Linus Torvalds 已提交
1385 1386 1387 1388

	/* First, grab all of the stats for the incoming packet.
	 * These get messed up if we get called due to a busy condition.
	 */
1389
	bdp = rxq->cur_rx;
L
Linus Torvalds 已提交
1390

S
Sascha Hauer 已提交
1391
	while (!((status = bdp->cbd_sc) & BD_ENET_RX_EMPTY)) {
L
Linus Torvalds 已提交
1392

1393 1394 1395 1396
		if (pkt_received >= budget)
			break;
		pkt_received++;

S
Sascha Hauer 已提交
1397 1398 1399 1400
		/* Since we have allocated space to hold a complete frame,
		 * the last indicator should be set.
		 */
		if ((status & BD_ENET_RX_LAST) == 0)
1401
			netdev_err(ndev, "rcv is not +last\n");
L
Linus Torvalds 已提交
1402

1403

S
Sascha Hauer 已提交
1404 1405
		/* Check for errors. */
		if (status & (BD_ENET_RX_LG | BD_ENET_RX_SH | BD_ENET_RX_NO |
L
Linus Torvalds 已提交
1406
			   BD_ENET_RX_CR | BD_ENET_RX_OV)) {
1407
			ndev->stats.rx_errors++;
S
Sascha Hauer 已提交
1408 1409
			if (status & (BD_ENET_RX_LG | BD_ENET_RX_SH)) {
				/* Frame too long or too short. */
1410
				ndev->stats.rx_length_errors++;
S
Sascha Hauer 已提交
1411 1412
			}
			if (status & BD_ENET_RX_NO)	/* Frame alignment */
1413
				ndev->stats.rx_frame_errors++;
S
Sascha Hauer 已提交
1414
			if (status & BD_ENET_RX_CR)	/* CRC Error */
1415
				ndev->stats.rx_crc_errors++;
S
Sascha Hauer 已提交
1416
			if (status & BD_ENET_RX_OV)	/* FIFO overrun */
1417
				ndev->stats.rx_fifo_errors++;
L
Linus Torvalds 已提交
1418 1419
		}

S
Sascha Hauer 已提交
1420 1421 1422 1423 1424
		/* Report late collisions as a frame error.
		 * On this error, the BD is closed, but we don't know what we
		 * have in the buffer.  So, just drop this frame on the floor.
		 */
		if (status & BD_ENET_RX_CL) {
1425 1426
			ndev->stats.rx_errors++;
			ndev->stats.rx_frame_errors++;
S
Sascha Hauer 已提交
1427 1428
			goto rx_processing_done;
		}
L
Linus Torvalds 已提交
1429

S
Sascha Hauer 已提交
1430
		/* Process the incoming frame. */
1431
		ndev->stats.rx_packets++;
S
Sascha Hauer 已提交
1432
		pkt_len = bdp->cbd_datlen;
1433
		ndev->stats.rx_bytes += pkt_len;
L
Linus Torvalds 已提交
1434

1435
		index = fec_enet_get_bd_index(rxq->rx_bd_base, bdp, fep);
1436
		skb = rxq->rx_skbuff[index];
1437

1438 1439 1440 1441
		/* The packet length includes FCS, but we don't want to
		 * include that when passing upstream as it messes up
		 * bridging applications.
		 */
1442 1443
		is_copybreak = fec_enet_copybreak(ndev, &skb, bdp, pkt_len - 4,
						  need_swap);
1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457
		if (!is_copybreak) {
			skb_new = netdev_alloc_skb(ndev, FEC_ENET_RX_FRSIZE);
			if (unlikely(!skb_new)) {
				ndev->stats.rx_dropped++;
				goto rx_processing_done;
			}
			dma_unmap_single(&fep->pdev->dev, bdp->cbd_bufaddr,
					 FEC_ENET_RX_FRSIZE - fep->rx_align,
					 DMA_FROM_DEVICE);
		}

		prefetch(skb->data - NET_IP_ALIGN);
		skb_put(skb, pkt_len - 4);
		data = skb->data;
1458
		if (!is_copybreak && need_swap)
1459 1460
			swap_buffer(data, pkt_len);

1461 1462 1463 1464 1465 1466 1467 1468
		/* Extract the enhanced buffer descriptor */
		ebdp = NULL;
		if (fep->bufdesc_ex)
			ebdp = (struct bufdesc_ex *)bdp;

		/* If this is a VLAN packet remove the VLAN Tag */
		vlan_packet_rcvd = false;
		if ((ndev->features & NETIF_F_HW_VLAN_CTAG_RX) &&
1469
			fep->bufdesc_ex && (ebdp->cbd_esc & BD_ENET_RX_VLAN)) {
1470 1471 1472 1473 1474 1475
			/* Push and remove the vlan tag */
			struct vlan_hdr *vlan_header =
					(struct vlan_hdr *) (data + ETH_HLEN);
			vlan_tag = ntohs(vlan_header->h_vlan_TCI);

			vlan_packet_rcvd = true;
1476

1477
			memmove(skb->data + VLAN_HLEN, data, ETH_ALEN * 2);
1478
			skb_pull(skb, VLAN_HLEN);
1479 1480
		}

1481
		skb->protocol = eth_type_trans(skb, ndev);
L
Linus Torvalds 已提交
1482

1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494
		/* Get receive timestamp from the skb */
		if (fep->hwts_rx_en && fep->bufdesc_ex)
			fec_enet_hwtstamp(fep, ebdp->ts,
					  skb_hwtstamps(skb));

		if (fep->bufdesc_ex &&
		    (fep->csum_flags & FLAG_RX_CSUM_ENABLED)) {
			if (!(ebdp->cbd_esc & FLAG_RX_CSUM_ERROR)) {
				/* don't check it */
				skb->ip_summed = CHECKSUM_UNNECESSARY;
			} else {
				skb_checksum_none_assert(skb);
1495
			}
1496
		}
1497

1498 1499 1500 1501 1502
		/* Handle received VLAN packets */
		if (vlan_packet_rcvd)
			__vlan_hwaccel_put_tag(skb,
					       htons(ETH_P_8021Q),
					       vlan_tag);
1503

1504 1505 1506 1507 1508 1509 1510 1511 1512
		napi_gro_receive(&fep->napi, skb);

		if (is_copybreak) {
			dma_sync_single_for_device(&fep->pdev->dev, bdp->cbd_bufaddr,
						   FEC_ENET_RX_FRSIZE - fep->rx_align,
						   DMA_FROM_DEVICE);
		} else {
			rxq->rx_skbuff[index] = skb_new;
			fec_enet_new_rxbdp(ndev, bdp, skb_new);
S
Sascha Hauer 已提交
1513
		}
S
Sascha Hauer 已提交
1514

S
Sascha Hauer 已提交
1515 1516 1517
rx_processing_done:
		/* Clear the status flags for this buffer */
		status &= ~BD_ENET_RX_STATS;
L
Linus Torvalds 已提交
1518

S
Sascha Hauer 已提交
1519 1520 1521
		/* Mark the buffer empty */
		status |= BD_ENET_RX_EMPTY;
		bdp->cbd_sc = status;
1522

1523 1524 1525 1526 1527 1528 1529
		if (fep->bufdesc_ex) {
			struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;

			ebdp->cbd_esc = BD_ENET_RX_INT;
			ebdp->cbd_prot = 0;
			ebdp->cbd_bdu = 0;
		}
1530

S
Sascha Hauer 已提交
1531
		/* Update BD pointer to next entry */
1532
		bdp = fec_enet_get_nextdesc(bdp, fep, queue_id);
1533

S
Sascha Hauer 已提交
1534 1535 1536 1537
		/* Doing this here will keep the FEC running while we process
		 * incoming frames.  On a heavily loaded network, we should be
		 * able to keep up at the expense of system resources.
		 */
1538
		writel(0, fep->hwp + FEC_R_DES_ACTIVE(queue_id));
S
Sascha Hauer 已提交
1539
	}
1540 1541 1542
	rxq->cur_rx = bdp;
	return pkt_received;
}
L
Linus Torvalds 已提交
1543

1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555
static int
fec_enet_rx(struct net_device *ndev, int budget)
{
	int     pkt_received = 0;
	u16	queue_id;
	struct fec_enet_private *fep = netdev_priv(ndev);

	for_each_set_bit(queue_id, &fep->work_rx, FEC_ENET_MAX_RX_QS) {
		clear_bit(queue_id, &fep->work_rx);
		pkt_received += fec_enet_rx_queue(ndev,
					budget - pkt_received, queue_id);
	}
1556
	return pkt_received;
L
Linus Torvalds 已提交
1557 1558
}

1559 1560 1561 1562 1563 1564 1565 1566
static bool
fec_enet_collect_events(struct fec_enet_private *fep, uint int_events)
{
	if (int_events == 0)
		return false;

	if (int_events & FEC_ENET_RXF)
		fep->work_rx |= (1 << 2);
F
Frank Li 已提交
1567 1568 1569 1570
	if (int_events & FEC_ENET_RXF_1)
		fep->work_rx |= (1 << 0);
	if (int_events & FEC_ENET_RXF_2)
		fep->work_rx |= (1 << 1);
1571 1572 1573

	if (int_events & FEC_ENET_TXF)
		fep->work_tx |= (1 << 2);
F
Frank Li 已提交
1574 1575 1576 1577
	if (int_events & FEC_ENET_TXF_1)
		fep->work_tx |= (1 << 0);
	if (int_events & FEC_ENET_TXF_2)
		fep->work_tx |= (1 << 1);
1578 1579 1580 1581

	return true;
}

1582 1583 1584 1585 1586 1587 1588 1589
static irqreturn_t
fec_enet_interrupt(int irq, void *dev_id)
{
	struct net_device *ndev = dev_id;
	struct fec_enet_private *fep = netdev_priv(ndev);
	uint int_events;
	irqreturn_t ret = IRQ_NONE;

1590
	int_events = readl(fep->hwp + FEC_IEVENT);
N
Nimrod Andy 已提交
1591
	writel(int_events, fep->hwp + FEC_IEVENT);
1592
	fec_enet_collect_events(fep, int_events);
1593

1594
	if ((fep->work_tx || fep->work_rx) && fep->link) {
1595
		ret = IRQ_HANDLED;
1596

N
Nimrod Andy 已提交
1597 1598 1599 1600 1601
		if (napi_schedule_prep(&fep->napi)) {
			/* Disable the NAPI interrupts */
			writel(FEC_ENET_MII, fep->hwp + FEC_IMASK);
			__napi_schedule(&fep->napi);
		}
1602
	}
1603

1604 1605 1606 1607
	if (int_events & FEC_ENET_MII) {
		ret = IRQ_HANDLED;
		complete(&fep->mdio_done);
	}
1608

1609 1610
	if (fep->ptp_clock)
		fec_ptp_check_pps_event(fep);
1611

1612 1613 1614
	return ret;
}

1615 1616 1617 1618
static int fec_enet_rx_napi(struct napi_struct *napi, int budget)
{
	struct net_device *ndev = napi->dev;
	struct fec_enet_private *fep = netdev_priv(ndev);
1619 1620 1621
	int pkts;

	pkts = fec_enet_rx(ndev, budget);
1622

1623 1624
	fec_enet_tx(ndev);

1625 1626 1627 1628 1629 1630
	if (pkts < budget) {
		napi_complete(napi);
		writel(FEC_DEFAULT_IMASK, fep->hwp + FEC_IMASK);
	}
	return pkts;
}
1631

1632
/* ------------------------------------------------------------------------- */
1633
static void fec_get_mac(struct net_device *ndev)
L
Linus Torvalds 已提交
1634
{
1635
	struct fec_enet_private *fep = netdev_priv(ndev);
J
Jingoo Han 已提交
1636
	struct fec_platform_data *pdata = dev_get_platdata(&fep->pdev->dev);
1637
	unsigned char *iap, tmpaddr[ETH_ALEN];
L
Linus Torvalds 已提交
1638

1639 1640 1641 1642 1643 1644 1645 1646
	/*
	 * try to get mac address in following order:
	 *
	 * 1) module parameter via kernel command line in form
	 *    fec.macaddr=0x00,0x04,0x9f,0x01,0x30,0xe0
	 */
	iap = macaddr;

1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658
	/*
	 * 2) from device tree data
	 */
	if (!is_valid_ether_addr(iap)) {
		struct device_node *np = fep->pdev->dev.of_node;
		if (np) {
			const char *mac = of_get_mac_address(np);
			if (mac)
				iap = (unsigned char *) mac;
		}
	}

1659
	/*
1660
	 * 3) from flash or fuse (via platform data)
1661 1662 1663 1664 1665 1666 1667
	 */
	if (!is_valid_ether_addr(iap)) {
#ifdef CONFIG_M5272
		if (FEC_FLASHMAC)
			iap = (unsigned char *)FEC_FLASHMAC;
#else
		if (pdata)
1668
			iap = (unsigned char *)&pdata->mac;
1669 1670 1671 1672
#endif
	}

	/*
1673
	 * 4) FEC mac registers set by bootloader
1674 1675
	 */
	if (!is_valid_ether_addr(iap)) {
1676 1677 1678 1679
		*((__be32 *) &tmpaddr[0]) =
			cpu_to_be32(readl(fep->hwp + FEC_ADDR_LOW));
		*((__be16 *) &tmpaddr[4]) =
			cpu_to_be16(readl(fep->hwp + FEC_ADDR_HIGH) >> 16);
1680
		iap = &tmpaddr[0];
L
Linus Torvalds 已提交
1681 1682
	}

1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694
	/*
	 * 5) random mac address
	 */
	if (!is_valid_ether_addr(iap)) {
		/* Report it and use a random ethernet address instead */
		netdev_err(ndev, "Invalid MAC address: %pM\n", iap);
		eth_hw_addr_random(ndev);
		netdev_info(ndev, "Using random MAC address: %pM\n",
			    ndev->dev_addr);
		return;
	}

1695
	memcpy(ndev->dev_addr, iap, ETH_ALEN);
L
Linus Torvalds 已提交
1696

1697 1698
	/* Adjust MAC if using macaddr */
	if (iap == macaddr)
S
Shawn Guo 已提交
1699
		 ndev->dev_addr[ETH_ALEN-1] = macaddr[ETH_ALEN-1] + fep->dev_id;
L
Linus Torvalds 已提交
1700 1701
}

1702
/* ------------------------------------------------------------------------- */
L
Linus Torvalds 已提交
1703

1704 1705 1706
/*
 * Phy section
 */
1707
static void fec_enet_adjust_link(struct net_device *ndev)
L
Linus Torvalds 已提交
1708
{
1709
	struct fec_enet_private *fep = netdev_priv(ndev);
1710 1711
	struct phy_device *phy_dev = fep->phy_dev;
	int status_change = 0;
L
Linus Torvalds 已提交
1712

1713 1714 1715
	/* Prevent a state halted on mii error */
	if (fep->mii_timeout && phy_dev->state == PHY_HALTED) {
		phy_dev->state = PHY_RESUMING;
1716
		return;
1717
	}
L
Linus Torvalds 已提交
1718

1719 1720 1721 1722 1723 1724 1725 1726
	/*
	 * If the netdev is down, or is going down, we're not interested
	 * in link state events, so just mark our idea of the link as down
	 * and ignore the event.
	 */
	if (!netif_running(ndev) || !netif_device_present(ndev)) {
		fep->link = 0;
	} else if (phy_dev->link) {
1727
		if (!fep->link) {
1728
			fep->link = phy_dev->link;
1729 1730
			status_change = 1;
		}
L
Linus Torvalds 已提交
1731

1732 1733
		if (fep->full_duplex != phy_dev->duplex) {
			fep->full_duplex = phy_dev->duplex;
1734
			status_change = 1;
1735
		}
1736 1737 1738 1739 1740 1741 1742

		if (phy_dev->speed != fep->speed) {
			fep->speed = phy_dev->speed;
			status_change = 1;
		}

		/* if any of the above changed restart the FEC */
1743 1744 1745
		if (status_change) {
			napi_disable(&fep->napi);
			netif_tx_lock_bh(ndev);
1746
			fec_restart(ndev);
1747
			netif_wake_queue(ndev);
1748
			netif_tx_unlock_bh(ndev);
1749 1750
			napi_enable(&fep->napi);
		}
1751 1752
	} else {
		if (fep->link) {
1753 1754
			napi_disable(&fep->napi);
			netif_tx_lock_bh(ndev);
1755
			fec_stop(ndev);
1756 1757
			netif_tx_unlock_bh(ndev);
			napi_enable(&fep->napi);
1758
			fep->link = phy_dev->link;
1759 1760
			status_change = 1;
		}
L
Linus Torvalds 已提交
1761
	}
1762

1763 1764 1765
	if (status_change)
		phy_print_status(phy_dev);
}
L
Linus Torvalds 已提交
1766

1767
static int fec_enet_mdio_read(struct mii_bus *bus, int mii_id, int regnum)
L
Linus Torvalds 已提交
1768
{
1769
	struct fec_enet_private *fep = bus->priv;
1770
	unsigned long time_left;
L
Linus Torvalds 已提交
1771

1772
	fep->mii_timeout = 0;
1773
	init_completion(&fep->mdio_done);
1774 1775 1776 1777 1778 1779 1780

	/* start a read op */
	writel(FEC_MMFR_ST | FEC_MMFR_OP_READ |
		FEC_MMFR_PA(mii_id) | FEC_MMFR_RA(regnum) |
		FEC_MMFR_TA, fep->hwp + FEC_MII_DATA);

	/* wait for end of transfer */
1781 1782 1783 1784
	time_left = wait_for_completion_timeout(&fep->mdio_done,
			usecs_to_jiffies(FEC_MII_TIMEOUT));
	if (time_left == 0) {
		fep->mii_timeout = 1;
1785
		netdev_err(fep->netdev, "MDIO read timeout\n");
1786
		return -ETIMEDOUT;
L
Linus Torvalds 已提交
1787 1788
	}

1789 1790
	/* return value */
	return FEC_MMFR_DATA(readl(fep->hwp + FEC_MII_DATA));
1791
}
1792

1793 1794
static int fec_enet_mdio_write(struct mii_bus *bus, int mii_id, int regnum,
			   u16 value)
L
Linus Torvalds 已提交
1795
{
1796
	struct fec_enet_private *fep = bus->priv;
1797
	unsigned long time_left;
L
Linus Torvalds 已提交
1798

1799
	fep->mii_timeout = 0;
1800
	init_completion(&fep->mdio_done);
L
Linus Torvalds 已提交
1801

1802 1803
	/* start a write op */
	writel(FEC_MMFR_ST | FEC_MMFR_OP_WRITE |
1804 1805 1806 1807 1808
		FEC_MMFR_PA(mii_id) | FEC_MMFR_RA(regnum) |
		FEC_MMFR_TA | FEC_MMFR_DATA(value),
		fep->hwp + FEC_MII_DATA);

	/* wait for end of transfer */
1809 1810 1811 1812
	time_left = wait_for_completion_timeout(&fep->mdio_done,
			usecs_to_jiffies(FEC_MII_TIMEOUT));
	if (time_left == 0) {
		fep->mii_timeout = 1;
1813
		netdev_err(fep->netdev, "MDIO write timeout\n");
1814
		return -ETIMEDOUT;
1815
	}
L
Linus Torvalds 已提交
1816

1817
	return 0;
1818
}
L
Linus Torvalds 已提交
1819

1820 1821 1822 1823 1824 1825 1826 1827 1828
static int fec_enet_clk_enable(struct net_device *ndev, bool enable)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int ret;

	if (enable) {
		ret = clk_prepare_enable(fep->clk_ahb);
		if (ret)
			return ret;
1829 1830 1831
		ret = clk_prepare_enable(fep->clk_ipg);
		if (ret)
			goto failed_clk_ipg;
1832 1833 1834 1835 1836 1837
		if (fep->clk_enet_out) {
			ret = clk_prepare_enable(fep->clk_enet_out);
			if (ret)
				goto failed_clk_enet_out;
		}
		if (fep->clk_ptp) {
1838
			mutex_lock(&fep->ptp_clk_mutex);
1839
			ret = clk_prepare_enable(fep->clk_ptp);
1840 1841
			if (ret) {
				mutex_unlock(&fep->ptp_clk_mutex);
1842
				goto failed_clk_ptp;
1843 1844 1845 1846
			} else {
				fep->ptp_clk_on = true;
			}
			mutex_unlock(&fep->ptp_clk_mutex);
1847
		}
1848 1849 1850 1851 1852
		if (fep->clk_ref) {
			ret = clk_prepare_enable(fep->clk_ref);
			if (ret)
				goto failed_clk_ref;
		}
1853 1854
	} else {
		clk_disable_unprepare(fep->clk_ahb);
1855
		clk_disable_unprepare(fep->clk_ipg);
1856 1857
		if (fep->clk_enet_out)
			clk_disable_unprepare(fep->clk_enet_out);
1858 1859
		if (fep->clk_ptp) {
			mutex_lock(&fep->ptp_clk_mutex);
1860
			clk_disable_unprepare(fep->clk_ptp);
1861 1862 1863
			fep->ptp_clk_on = false;
			mutex_unlock(&fep->ptp_clk_mutex);
		}
1864 1865
		if (fep->clk_ref)
			clk_disable_unprepare(fep->clk_ref);
1866 1867 1868
	}

	return 0;
1869 1870 1871 1872

failed_clk_ref:
	if (fep->clk_ref)
		clk_disable_unprepare(fep->clk_ref);
1873 1874 1875 1876
failed_clk_ptp:
	if (fep->clk_enet_out)
		clk_disable_unprepare(fep->clk_enet_out);
failed_clk_enet_out:
1877 1878
		clk_disable_unprepare(fep->clk_ipg);
failed_clk_ipg:
1879 1880 1881 1882 1883
		clk_disable_unprepare(fep->clk_ahb);

	return ret;
}

1884
static int fec_enet_mii_probe(struct net_device *ndev)
1885
{
1886
	struct fec_enet_private *fep = netdev_priv(ndev);
1887
	struct phy_device *phy_dev = NULL;
1888 1889 1890
	char mdio_bus_id[MII_BUS_ID_SIZE];
	char phy_name[MII_BUS_ID_SIZE + 3];
	int phy_id;
S
Shawn Guo 已提交
1891
	int dev_id = fep->dev_id;
1892

1893 1894
	fep->phy_dev = NULL;

1895 1896 1897 1898
	if (fep->phy_node) {
		phy_dev = of_phy_connect(ndev, fep->phy_node,
					 &fec_enet_adjust_link, 0,
					 fep->phy_interface);
1899 1900
		if (!phy_dev)
			return -ENODEV;
1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911
	} else {
		/* check for attached phy */
		for (phy_id = 0; (phy_id < PHY_MAX_ADDR); phy_id++) {
			if ((fep->mii_bus->phy_mask & (1 << phy_id)))
				continue;
			if (fep->mii_bus->phy_map[phy_id] == NULL)
				continue;
			if (fep->mii_bus->phy_map[phy_id]->phy_id == 0)
				continue;
			if (dev_id--)
				continue;
1912
			strlcpy(mdio_bus_id, fep->mii_bus->id, MII_BUS_ID_SIZE);
1913 1914
			break;
		}
L
Linus Torvalds 已提交
1915

1916 1917
		if (phy_id >= PHY_MAX_ADDR) {
			netdev_info(ndev, "no PHY, assuming direct connection to switch\n");
1918
			strlcpy(mdio_bus_id, "fixed-0", MII_BUS_ID_SIZE);
1919 1920 1921 1922 1923 1924 1925
			phy_id = 0;
		}

		snprintf(phy_name, sizeof(phy_name),
			 PHY_ID_FMT, mdio_bus_id, phy_id);
		phy_dev = phy_connect(ndev, phy_name, &fec_enet_adjust_link,
				      fep->phy_interface);
1926 1927 1928
	}

	if (IS_ERR(phy_dev)) {
1929
		netdev_err(ndev, "could not attach to PHY\n");
1930
		return PTR_ERR(phy_dev);
1931
	}
L
Linus Torvalds 已提交
1932

1933
	/* mask with MAC supported features */
1934
	if (fep->quirks & FEC_QUIRK_HAS_GBIT) {
S
Shawn Guo 已提交
1935
		phy_dev->supported &= PHY_GBIT_FEATURES;
1936
		phy_dev->supported &= ~SUPPORTED_1000baseT_Half;
G
Guenter Roeck 已提交
1937
#if !defined(CONFIG_M5272)
1938
		phy_dev->supported |= SUPPORTED_Pause;
G
Guenter Roeck 已提交
1939
#endif
1940
	}
S
Shawn Guo 已提交
1941 1942 1943
	else
		phy_dev->supported &= PHY_BASIC_FEATURES;

1944
	phy_dev->advertising = phy_dev->supported;
L
Linus Torvalds 已提交
1945

1946 1947 1948
	fep->phy_dev = phy_dev;
	fep->link = 0;
	fep->full_duplex = 0;
L
Linus Torvalds 已提交
1949

1950 1951 1952
	netdev_info(ndev, "Freescale FEC PHY driver [%s] (mii_bus:phy_addr=%s, irq=%d)\n",
		    fep->phy_dev->drv->name, dev_name(&fep->phy_dev->dev),
		    fep->phy_dev->irq);
1953

1954
	return 0;
L
Linus Torvalds 已提交
1955 1956
}

1957
static int fec_enet_mii_init(struct platform_device *pdev)
1958
{
1959
	static struct mii_bus *fec0_mii_bus;
1960 1961
	struct net_device *ndev = platform_get_drvdata(pdev);
	struct fec_enet_private *fep = netdev_priv(ndev);
1962
	struct device_node *node;
1963
	int err = -ENXIO, i;
1964
	u32 mii_speed, holdtime;
1965

1966
	/*
1967
	 * The i.MX28 dual fec interfaces are not equal.
1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981
	 * Here are the differences:
	 *
	 *  - fec0 supports MII & RMII modes while fec1 only supports RMII
	 *  - fec0 acts as the 1588 time master while fec1 is slave
	 *  - external phys can only be configured by fec0
	 *
	 * That is to say fec1 can not work independently. It only works
	 * when fec0 is working. The reason behind this design is that the
	 * second interface is added primarily for Switch mode.
	 *
	 * Because of the last point above, both phys are attached on fec0
	 * mdio interface in board design, and need to be configured by
	 * fec0 mii_bus.
	 */
1982
	if ((fep->quirks & FEC_QUIRK_SINGLE_MDIO) && fep->dev_id > 0) {
1983
		/* fec1 uses fec0 mii_bus */
L
Lothar Waßmann 已提交
1984 1985 1986 1987 1988 1989
		if (mii_cnt && fec0_mii_bus) {
			fep->mii_bus = fec0_mii_bus;
			mii_cnt++;
			return 0;
		}
		return -ENOENT;
1990 1991
	}

1992
	fep->mii_timeout = 0;
L
Linus Torvalds 已提交
1993

1994 1995
	/*
	 * Set MII speed to 2.5 MHz (= clk_get_rate() / 2 * phy_speed)
S
Shawn Guo 已提交
1996 1997 1998 1999 2000
	 *
	 * The formula for FEC MDC is 'ref_freq / (MII_SPEED x 2)' while
	 * for ENET-MAC is 'ref_freq / ((MII_SPEED + 1) x 2)'.  The i.MX28
	 * Reference Manual has an error on this, and gets fixed on i.MX6Q
	 * document.
2001
	 */
2002
	mii_speed = DIV_ROUND_UP(clk_get_rate(fep->clk_ipg), 5000000);
2003
	if (fep->quirks & FEC_QUIRK_ENET_MAC)
2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028
		mii_speed--;
	if (mii_speed > 63) {
		dev_err(&pdev->dev,
			"fec clock (%lu) to fast to get right mii speed\n",
			clk_get_rate(fep->clk_ipg));
		err = -EINVAL;
		goto err_out;
	}

	/*
	 * The i.MX28 and i.MX6 types have another filed in the MSCR (aka
	 * MII_SPEED) register that defines the MDIO output hold time. Earlier
	 * versions are RAZ there, so just ignore the difference and write the
	 * register always.
	 * The minimal hold time according to IEE802.3 (clause 22) is 10 ns.
	 * HOLDTIME + 1 is the number of clk cycles the fec is holding the
	 * output.
	 * The HOLDTIME bitfield takes values between 0 and 7 (inclusive).
	 * Given that ceil(clkrate / 5000000) <= 64, the calculation for
	 * holdtime cannot result in a value greater than 3.
	 */
	holdtime = DIV_ROUND_UP(clk_get_rate(fep->clk_ipg), 100000000) - 1;

	fep->phy_speed = mii_speed << 1 | holdtime << 8;

2029
	writel(fep->phy_speed, fep->hwp + FEC_MII_SPEED);
L
Linus Torvalds 已提交
2030

2031 2032 2033 2034
	fep->mii_bus = mdiobus_alloc();
	if (fep->mii_bus == NULL) {
		err = -ENOMEM;
		goto err_out;
L
Linus Torvalds 已提交
2035 2036
	}

2037 2038 2039
	fep->mii_bus->name = "fec_enet_mii_bus";
	fep->mii_bus->read = fec_enet_mdio_read;
	fep->mii_bus->write = fec_enet_mdio_write;
2040 2041
	snprintf(fep->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
		pdev->name, fep->dev_id + 1);
2042 2043 2044 2045 2046 2047 2048
	fep->mii_bus->priv = fep;
	fep->mii_bus->parent = &pdev->dev;

	fep->mii_bus->irq = kmalloc(sizeof(int) * PHY_MAX_ADDR, GFP_KERNEL);
	if (!fep->mii_bus->irq) {
		err = -ENOMEM;
		goto err_out_free_mdiobus;
L
Linus Torvalds 已提交
2049 2050
	}

2051 2052
	for (i = 0; i < PHY_MAX_ADDR; i++)
		fep->mii_bus->irq[i] = PHY_POLL;
L
Linus Torvalds 已提交
2053

2054 2055 2056 2057 2058 2059 2060 2061 2062
	node = of_get_child_by_name(pdev->dev.of_node, "mdio");
	if (node) {
		err = of_mdiobus_register(fep->mii_bus, node);
		of_node_put(node);
	} else {
		err = mdiobus_register(fep->mii_bus);
	}

	if (err)
2063
		goto err_out_free_mdio_irq;
L
Linus Torvalds 已提交
2064

L
Lothar Waßmann 已提交
2065 2066
	mii_cnt++;

2067
	/* save fec0 mii_bus */
2068
	if (fep->quirks & FEC_QUIRK_SINGLE_MDIO)
2069 2070
		fec0_mii_bus = fep->mii_bus;

2071
	return 0;
L
Linus Torvalds 已提交
2072

2073 2074 2075 2076 2077 2078
err_out_free_mdio_irq:
	kfree(fep->mii_bus->irq);
err_out_free_mdiobus:
	mdiobus_free(fep->mii_bus);
err_out:
	return err;
L
Linus Torvalds 已提交
2079 2080
}

2081
static void fec_enet_mii_remove(struct fec_enet_private *fep)
L
Linus Torvalds 已提交
2082
{
L
Lothar Waßmann 已提交
2083 2084 2085 2086 2087
	if (--mii_cnt == 0) {
		mdiobus_unregister(fep->mii_bus);
		kfree(fep->mii_bus->irq);
		mdiobus_free(fep->mii_bus);
	}
L
Linus Torvalds 已提交
2088 2089
}

2090
static int fec_enet_get_settings(struct net_device *ndev,
2091
				  struct ethtool_cmd *cmd)
L
Linus Torvalds 已提交
2092
{
2093
	struct fec_enet_private *fep = netdev_priv(ndev);
2094
	struct phy_device *phydev = fep->phy_dev;
L
Linus Torvalds 已提交
2095

2096 2097
	if (!phydev)
		return -ENODEV;
L
Linus Torvalds 已提交
2098

2099
	return phy_ethtool_gset(phydev, cmd);
L
Linus Torvalds 已提交
2100 2101
}

2102
static int fec_enet_set_settings(struct net_device *ndev,
2103
				 struct ethtool_cmd *cmd)
L
Linus Torvalds 已提交
2104
{
2105
	struct fec_enet_private *fep = netdev_priv(ndev);
2106
	struct phy_device *phydev = fep->phy_dev;
L
Linus Torvalds 已提交
2107

2108 2109
	if (!phydev)
		return -ENODEV;
L
Linus Torvalds 已提交
2110

2111
	return phy_ethtool_sset(phydev, cmd);
L
Linus Torvalds 已提交
2112 2113
}

2114
static void fec_enet_get_drvinfo(struct net_device *ndev,
2115
				 struct ethtool_drvinfo *info)
L
Linus Torvalds 已提交
2116
{
2117
	struct fec_enet_private *fep = netdev_priv(ndev);
2118

2119 2120 2121 2122
	strlcpy(info->driver, fep->pdev->dev.driver->name,
		sizeof(info->driver));
	strlcpy(info->version, "Revision: 1.0", sizeof(info->version));
	strlcpy(info->bus_info, dev_name(&ndev->dev), sizeof(info->bus_info));
L
Linus Torvalds 已提交
2123 2124
}

2125 2126 2127 2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200
static int fec_enet_get_regs_len(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	struct resource *r;
	int s = 0;

	r = platform_get_resource(fep->pdev, IORESOURCE_MEM, 0);
	if (r)
		s = resource_size(r);

	return s;
}

/* List of registers that can be safety be read to dump them with ethtool */
#if defined(CONFIG_M523x) || defined(CONFIG_M527x) || defined(CONFIG_M528x) || \
	defined(CONFIG_M520x) || defined(CONFIG_M532x) ||		\
	defined(CONFIG_ARCH_MXC) || defined(CONFIG_SOC_IMX28)
static u32 fec_enet_register_offset[] = {
	FEC_IEVENT, FEC_IMASK, FEC_R_DES_ACTIVE_0, FEC_X_DES_ACTIVE_0,
	FEC_ECNTRL, FEC_MII_DATA, FEC_MII_SPEED, FEC_MIB_CTRLSTAT, FEC_R_CNTRL,
	FEC_X_CNTRL, FEC_ADDR_LOW, FEC_ADDR_HIGH, FEC_OPD, FEC_TXIC0, FEC_TXIC1,
	FEC_TXIC2, FEC_RXIC0, FEC_RXIC1, FEC_RXIC2, FEC_HASH_TABLE_HIGH,
	FEC_HASH_TABLE_LOW, FEC_GRP_HASH_TABLE_HIGH, FEC_GRP_HASH_TABLE_LOW,
	FEC_X_WMRK, FEC_R_BOUND, FEC_R_FSTART, FEC_R_DES_START_1,
	FEC_X_DES_START_1, FEC_R_BUFF_SIZE_1, FEC_R_DES_START_2,
	FEC_X_DES_START_2, FEC_R_BUFF_SIZE_2, FEC_R_DES_START_0,
	FEC_X_DES_START_0, FEC_R_BUFF_SIZE_0, FEC_R_FIFO_RSFL, FEC_R_FIFO_RSEM,
	FEC_R_FIFO_RAEM, FEC_R_FIFO_RAFL, FEC_RACC, FEC_RCMR_1, FEC_RCMR_2,
	FEC_DMA_CFG_1, FEC_DMA_CFG_2, FEC_R_DES_ACTIVE_1, FEC_X_DES_ACTIVE_1,
	FEC_R_DES_ACTIVE_2, FEC_X_DES_ACTIVE_2, FEC_QOS_SCHEME,
	RMON_T_DROP, RMON_T_PACKETS, RMON_T_BC_PKT, RMON_T_MC_PKT,
	RMON_T_CRC_ALIGN, RMON_T_UNDERSIZE, RMON_T_OVERSIZE, RMON_T_FRAG,
	RMON_T_JAB, RMON_T_COL, RMON_T_P64, RMON_T_P65TO127, RMON_T_P128TO255,
	RMON_T_P256TO511, RMON_T_P512TO1023, RMON_T_P1024TO2047,
	RMON_T_P_GTE2048, RMON_T_OCTETS,
	IEEE_T_DROP, IEEE_T_FRAME_OK, IEEE_T_1COL, IEEE_T_MCOL, IEEE_T_DEF,
	IEEE_T_LCOL, IEEE_T_EXCOL, IEEE_T_MACERR, IEEE_T_CSERR, IEEE_T_SQE,
	IEEE_T_FDXFC, IEEE_T_OCTETS_OK,
	RMON_R_PACKETS, RMON_R_BC_PKT, RMON_R_MC_PKT, RMON_R_CRC_ALIGN,
	RMON_R_UNDERSIZE, RMON_R_OVERSIZE, RMON_R_FRAG, RMON_R_JAB,
	RMON_R_RESVD_O, RMON_R_P64, RMON_R_P65TO127, RMON_R_P128TO255,
	RMON_R_P256TO511, RMON_R_P512TO1023, RMON_R_P1024TO2047,
	RMON_R_P_GTE2048, RMON_R_OCTETS,
	IEEE_R_DROP, IEEE_R_FRAME_OK, IEEE_R_CRC, IEEE_R_ALIGN, IEEE_R_MACERR,
	IEEE_R_FDXFC, IEEE_R_OCTETS_OK
};
#else
static u32 fec_enet_register_offset[] = {
	FEC_ECNTRL, FEC_IEVENT, FEC_IMASK, FEC_IVEC, FEC_R_DES_ACTIVE_0,
	FEC_R_DES_ACTIVE_1, FEC_R_DES_ACTIVE_2, FEC_X_DES_ACTIVE_0,
	FEC_X_DES_ACTIVE_1, FEC_X_DES_ACTIVE_2, FEC_MII_DATA, FEC_MII_SPEED,
	FEC_R_BOUND, FEC_R_FSTART, FEC_X_WMRK, FEC_X_FSTART, FEC_R_CNTRL,
	FEC_MAX_FRM_LEN, FEC_X_CNTRL, FEC_ADDR_LOW, FEC_ADDR_HIGH,
	FEC_GRP_HASH_TABLE_HIGH, FEC_GRP_HASH_TABLE_LOW, FEC_R_DES_START_0,
	FEC_R_DES_START_1, FEC_R_DES_START_2, FEC_X_DES_START_0,
	FEC_X_DES_START_1, FEC_X_DES_START_2, FEC_R_BUFF_SIZE_0,
	FEC_R_BUFF_SIZE_1, FEC_R_BUFF_SIZE_2
};
#endif

static void fec_enet_get_regs(struct net_device *ndev,
			      struct ethtool_regs *regs, void *regbuf)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	u32 __iomem *theregs = (u32 __iomem *)fep->hwp;
	u32 *buf = (u32 *)regbuf;
	u32 i, off;

	memset(buf, 0, regs->len);

	for (i = 0; i < ARRAY_SIZE(fec_enet_register_offset); i++) {
		off = fec_enet_register_offset[i] / 4;
		buf[off] = readl(&theregs[off]);
	}
}

2201 2202 2203 2204 2205 2206 2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226 2227 2228 2229
static int fec_enet_get_ts_info(struct net_device *ndev,
				struct ethtool_ts_info *info)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

	if (fep->bufdesc_ex) {

		info->so_timestamping = SOF_TIMESTAMPING_TX_SOFTWARE |
					SOF_TIMESTAMPING_RX_SOFTWARE |
					SOF_TIMESTAMPING_SOFTWARE |
					SOF_TIMESTAMPING_TX_HARDWARE |
					SOF_TIMESTAMPING_RX_HARDWARE |
					SOF_TIMESTAMPING_RAW_HARDWARE;
		if (fep->ptp_clock)
			info->phc_index = ptp_clock_index(fep->ptp_clock);
		else
			info->phc_index = -1;

		info->tx_types = (1 << HWTSTAMP_TX_OFF) |
				 (1 << HWTSTAMP_TX_ON);

		info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) |
				   (1 << HWTSTAMP_FILTER_ALL);
		return 0;
	} else {
		return ethtool_op_get_ts_info(ndev, info);
	}
}

G
Guenter Roeck 已提交
2230 2231
#if !defined(CONFIG_M5272)

2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245 2246
static void fec_enet_get_pauseparam(struct net_device *ndev,
				    struct ethtool_pauseparam *pause)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

	pause->autoneg = (fep->pause_flag & FEC_PAUSE_FLAG_AUTONEG) != 0;
	pause->tx_pause = (fep->pause_flag & FEC_PAUSE_FLAG_ENABLE) != 0;
	pause->rx_pause = pause->tx_pause;
}

static int fec_enet_set_pauseparam(struct net_device *ndev,
				   struct ethtool_pauseparam *pause)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

2247 2248 2249
	if (!fep->phy_dev)
		return -ENODEV;

2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274
	if (pause->tx_pause != pause->rx_pause) {
		netdev_info(ndev,
			"hardware only support enable/disable both tx and rx");
		return -EINVAL;
	}

	fep->pause_flag = 0;

	/* tx pause must be same as rx pause */
	fep->pause_flag |= pause->rx_pause ? FEC_PAUSE_FLAG_ENABLE : 0;
	fep->pause_flag |= pause->autoneg ? FEC_PAUSE_FLAG_AUTONEG : 0;

	if (pause->rx_pause || pause->autoneg) {
		fep->phy_dev->supported |= ADVERTISED_Pause;
		fep->phy_dev->advertising |= ADVERTISED_Pause;
	} else {
		fep->phy_dev->supported &= ~ADVERTISED_Pause;
		fep->phy_dev->advertising &= ~ADVERTISED_Pause;
	}

	if (pause->autoneg) {
		if (netif_running(ndev))
			fec_stop(ndev);
		phy_start_aneg(fep->phy_dev);
	}
2275 2276 2277
	if (netif_running(ndev)) {
		napi_disable(&fep->napi);
		netif_tx_lock_bh(ndev);
2278
		fec_restart(ndev);
2279
		netif_wake_queue(ndev);
2280
		netif_tx_unlock_bh(ndev);
2281 2282
		napi_enable(&fep->napi);
	}
2283 2284 2285 2286

	return 0;
}

2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333 2334 2335 2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371 2372 2373 2374 2375 2376 2377 2378 2379 2380 2381 2382 2383 2384
static const struct fec_stat {
	char name[ETH_GSTRING_LEN];
	u16 offset;
} fec_stats[] = {
	/* RMON TX */
	{ "tx_dropped", RMON_T_DROP },
	{ "tx_packets", RMON_T_PACKETS },
	{ "tx_broadcast", RMON_T_BC_PKT },
	{ "tx_multicast", RMON_T_MC_PKT },
	{ "tx_crc_errors", RMON_T_CRC_ALIGN },
	{ "tx_undersize", RMON_T_UNDERSIZE },
	{ "tx_oversize", RMON_T_OVERSIZE },
	{ "tx_fragment", RMON_T_FRAG },
	{ "tx_jabber", RMON_T_JAB },
	{ "tx_collision", RMON_T_COL },
	{ "tx_64byte", RMON_T_P64 },
	{ "tx_65to127byte", RMON_T_P65TO127 },
	{ "tx_128to255byte", RMON_T_P128TO255 },
	{ "tx_256to511byte", RMON_T_P256TO511 },
	{ "tx_512to1023byte", RMON_T_P512TO1023 },
	{ "tx_1024to2047byte", RMON_T_P1024TO2047 },
	{ "tx_GTE2048byte", RMON_T_P_GTE2048 },
	{ "tx_octets", RMON_T_OCTETS },

	/* IEEE TX */
	{ "IEEE_tx_drop", IEEE_T_DROP },
	{ "IEEE_tx_frame_ok", IEEE_T_FRAME_OK },
	{ "IEEE_tx_1col", IEEE_T_1COL },
	{ "IEEE_tx_mcol", IEEE_T_MCOL },
	{ "IEEE_tx_def", IEEE_T_DEF },
	{ "IEEE_tx_lcol", IEEE_T_LCOL },
	{ "IEEE_tx_excol", IEEE_T_EXCOL },
	{ "IEEE_tx_macerr", IEEE_T_MACERR },
	{ "IEEE_tx_cserr", IEEE_T_CSERR },
	{ "IEEE_tx_sqe", IEEE_T_SQE },
	{ "IEEE_tx_fdxfc", IEEE_T_FDXFC },
	{ "IEEE_tx_octets_ok", IEEE_T_OCTETS_OK },

	/* RMON RX */
	{ "rx_packets", RMON_R_PACKETS },
	{ "rx_broadcast", RMON_R_BC_PKT },
	{ "rx_multicast", RMON_R_MC_PKT },
	{ "rx_crc_errors", RMON_R_CRC_ALIGN },
	{ "rx_undersize", RMON_R_UNDERSIZE },
	{ "rx_oversize", RMON_R_OVERSIZE },
	{ "rx_fragment", RMON_R_FRAG },
	{ "rx_jabber", RMON_R_JAB },
	{ "rx_64byte", RMON_R_P64 },
	{ "rx_65to127byte", RMON_R_P65TO127 },
	{ "rx_128to255byte", RMON_R_P128TO255 },
	{ "rx_256to511byte", RMON_R_P256TO511 },
	{ "rx_512to1023byte", RMON_R_P512TO1023 },
	{ "rx_1024to2047byte", RMON_R_P1024TO2047 },
	{ "rx_GTE2048byte", RMON_R_P_GTE2048 },
	{ "rx_octets", RMON_R_OCTETS },

	/* IEEE RX */
	{ "IEEE_rx_drop", IEEE_R_DROP },
	{ "IEEE_rx_frame_ok", IEEE_R_FRAME_OK },
	{ "IEEE_rx_crc", IEEE_R_CRC },
	{ "IEEE_rx_align", IEEE_R_ALIGN },
	{ "IEEE_rx_macerr", IEEE_R_MACERR },
	{ "IEEE_rx_fdxfc", IEEE_R_FDXFC },
	{ "IEEE_rx_octets_ok", IEEE_R_OCTETS_OK },
};

static void fec_enet_get_ethtool_stats(struct net_device *dev,
	struct ethtool_stats *stats, u64 *data)
{
	struct fec_enet_private *fep = netdev_priv(dev);
	int i;

	for (i = 0; i < ARRAY_SIZE(fec_stats); i++)
		data[i] = readl(fep->hwp + fec_stats[i].offset);
}

static void fec_enet_get_strings(struct net_device *netdev,
	u32 stringset, u8 *data)
{
	int i;
	switch (stringset) {
	case ETH_SS_STATS:
		for (i = 0; i < ARRAY_SIZE(fec_stats); i++)
			memcpy(data + i * ETH_GSTRING_LEN,
				fec_stats[i].name, ETH_GSTRING_LEN);
		break;
	}
}

static int fec_enet_get_sset_count(struct net_device *dev, int sset)
{
	switch (sset) {
	case ETH_SS_STATS:
		return ARRAY_SIZE(fec_stats);
	default:
		return -EOPNOTSUPP;
	}
}
G
Guenter Roeck 已提交
2385
#endif /* !defined(CONFIG_M5272) */
2386

2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397
static int fec_enet_nway_reset(struct net_device *dev)
{
	struct fec_enet_private *fep = netdev_priv(dev);
	struct phy_device *phydev = fep->phy_dev;

	if (!phydev)
		return -ENODEV;

	return genphy_restart_aneg(phydev);
}

2398 2399 2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413 2414
/* ITR clock source is enet system clock (clk_ahb).
 * TCTT unit is cycle_ns * 64 cycle
 * So, the ICTT value = X us / (cycle_ns * 64)
 */
static int fec_enet_us_to_itr_clock(struct net_device *ndev, int us)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

	return us * (fep->itr_clk_rate / 64000) / 1000;
}

/* Set threshold for interrupt coalescing */
static void fec_enet_itr_coal_set(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int rx_itr, tx_itr;

2415
	if (!(fep->quirks & FEC_QUIRK_HAS_AVB))
2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431 2432 2433 2434 2435 2436 2437 2438 2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450
		return;

	/* Must be greater than zero to avoid unpredictable behavior */
	if (!fep->rx_time_itr || !fep->rx_pkts_itr ||
	    !fep->tx_time_itr || !fep->tx_pkts_itr)
		return;

	/* Select enet system clock as Interrupt Coalescing
	 * timer Clock Source
	 */
	rx_itr = FEC_ITR_CLK_SEL;
	tx_itr = FEC_ITR_CLK_SEL;

	/* set ICFT and ICTT */
	rx_itr |= FEC_ITR_ICFT(fep->rx_pkts_itr);
	rx_itr |= FEC_ITR_ICTT(fec_enet_us_to_itr_clock(ndev, fep->rx_time_itr));
	tx_itr |= FEC_ITR_ICFT(fep->tx_pkts_itr);
	tx_itr |= FEC_ITR_ICTT(fec_enet_us_to_itr_clock(ndev, fep->tx_time_itr));

	rx_itr |= FEC_ITR_EN;
	tx_itr |= FEC_ITR_EN;

	writel(tx_itr, fep->hwp + FEC_TXIC0);
	writel(rx_itr, fep->hwp + FEC_RXIC0);
	writel(tx_itr, fep->hwp + FEC_TXIC1);
	writel(rx_itr, fep->hwp + FEC_RXIC1);
	writel(tx_itr, fep->hwp + FEC_TXIC2);
	writel(rx_itr, fep->hwp + FEC_RXIC2);
}

static int
fec_enet_get_coalesce(struct net_device *ndev, struct ethtool_coalesce *ec)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

2451
	if (!(fep->quirks & FEC_QUIRK_HAS_AVB))
2452 2453 2454 2455 2456 2457 2458 2459 2460 2461 2462 2463 2464 2465 2466 2467 2468
		return -EOPNOTSUPP;

	ec->rx_coalesce_usecs = fep->rx_time_itr;
	ec->rx_max_coalesced_frames = fep->rx_pkts_itr;

	ec->tx_coalesce_usecs = fep->tx_time_itr;
	ec->tx_max_coalesced_frames = fep->tx_pkts_itr;

	return 0;
}

static int
fec_enet_set_coalesce(struct net_device *ndev, struct ethtool_coalesce *ec)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	unsigned int cycle;

2469
	if (!(fep->quirks & FEC_QUIRK_HAS_AVB))
2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484 2485 2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505 2506 2507 2508 2509 2510 2511 2512 2513 2514 2515 2516 2517
		return -EOPNOTSUPP;

	if (ec->rx_max_coalesced_frames > 255) {
		pr_err("Rx coalesced frames exceed hardware limiation");
		return -EINVAL;
	}

	if (ec->tx_max_coalesced_frames > 255) {
		pr_err("Tx coalesced frame exceed hardware limiation");
		return -EINVAL;
	}

	cycle = fec_enet_us_to_itr_clock(ndev, fep->rx_time_itr);
	if (cycle > 0xFFFF) {
		pr_err("Rx coalesed usec exceeed hardware limiation");
		return -EINVAL;
	}

	cycle = fec_enet_us_to_itr_clock(ndev, fep->tx_time_itr);
	if (cycle > 0xFFFF) {
		pr_err("Rx coalesed usec exceeed hardware limiation");
		return -EINVAL;
	}

	fep->rx_time_itr = ec->rx_coalesce_usecs;
	fep->rx_pkts_itr = ec->rx_max_coalesced_frames;

	fep->tx_time_itr = ec->tx_coalesce_usecs;
	fep->tx_pkts_itr = ec->tx_max_coalesced_frames;

	fec_enet_itr_coal_set(ndev);

	return 0;
}

static void fec_enet_itr_coal_init(struct net_device *ndev)
{
	struct ethtool_coalesce ec;

	ec.rx_coalesce_usecs = FEC_ITR_ICTT_DEFAULT;
	ec.rx_max_coalesced_frames = FEC_ITR_ICFT_DEFAULT;

	ec.tx_coalesce_usecs = FEC_ITR_ICTT_DEFAULT;
	ec.tx_max_coalesced_frames = FEC_ITR_ICFT_DEFAULT;

	fec_enet_set_coalesce(ndev, &ec);
}

2518 2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537 2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555
static int fec_enet_get_tunable(struct net_device *netdev,
				const struct ethtool_tunable *tuna,
				void *data)
{
	struct fec_enet_private *fep = netdev_priv(netdev);
	int ret = 0;

	switch (tuna->id) {
	case ETHTOOL_RX_COPYBREAK:
		*(u32 *)data = fep->rx_copybreak;
		break;
	default:
		ret = -EINVAL;
		break;
	}

	return ret;
}

static int fec_enet_set_tunable(struct net_device *netdev,
				const struct ethtool_tunable *tuna,
				const void *data)
{
	struct fec_enet_private *fep = netdev_priv(netdev);
	int ret = 0;

	switch (tuna->id) {
	case ETHTOOL_RX_COPYBREAK:
		fep->rx_copybreak = *(u32 *)data;
		break;
	default:
		ret = -EINVAL;
		break;
	}

	return ret;
}

N
Nimrod Andy 已提交
2556 2557 2558 2559 2560 2561 2562 2563 2564 2565 2566 2567 2568 2569 2570 2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590 2591 2592 2593
static void
fec_enet_get_wol(struct net_device *ndev, struct ethtool_wolinfo *wol)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

	if (fep->wol_flag & FEC_WOL_HAS_MAGIC_PACKET) {
		wol->supported = WAKE_MAGIC;
		wol->wolopts = fep->wol_flag & FEC_WOL_FLAG_ENABLE ? WAKE_MAGIC : 0;
	} else {
		wol->supported = wol->wolopts = 0;
	}
}

static int
fec_enet_set_wol(struct net_device *ndev, struct ethtool_wolinfo *wol)
{
	struct fec_enet_private *fep = netdev_priv(ndev);

	if (!(fep->wol_flag & FEC_WOL_HAS_MAGIC_PACKET))
		return -EINVAL;

	if (wol->wolopts & ~WAKE_MAGIC)
		return -EINVAL;

	device_set_wakeup_enable(&ndev->dev, wol->wolopts & WAKE_MAGIC);
	if (device_may_wakeup(&ndev->dev)) {
		fep->wol_flag |= FEC_WOL_FLAG_ENABLE;
		if (fep->irq[0] > 0)
			enable_irq_wake(fep->irq[0]);
	} else {
		fep->wol_flag &= (~FEC_WOL_FLAG_ENABLE);
		if (fep->irq[0] > 0)
			disable_irq_wake(fep->irq[0]);
	}

	return 0;
}

S
stephen hemminger 已提交
2594
static const struct ethtool_ops fec_enet_ethtool_ops = {
2595 2596 2597
	.get_settings		= fec_enet_get_settings,
	.set_settings		= fec_enet_set_settings,
	.get_drvinfo		= fec_enet_get_drvinfo,
2598 2599
	.get_regs_len		= fec_enet_get_regs_len,
	.get_regs		= fec_enet_get_regs,
2600
	.nway_reset		= fec_enet_nway_reset,
2601
	.get_link		= ethtool_op_get_link,
2602 2603
	.get_coalesce		= fec_enet_get_coalesce,
	.set_coalesce		= fec_enet_set_coalesce,
2604
#ifndef CONFIG_M5272
2605 2606
	.get_pauseparam		= fec_enet_get_pauseparam,
	.set_pauseparam		= fec_enet_set_pauseparam,
2607
	.get_strings		= fec_enet_get_strings,
2608
	.get_ethtool_stats	= fec_enet_get_ethtool_stats,
2609 2610
	.get_sset_count		= fec_enet_get_sset_count,
#endif
2611
	.get_ts_info		= fec_enet_get_ts_info,
2612 2613
	.get_tunable		= fec_enet_get_tunable,
	.set_tunable		= fec_enet_set_tunable,
N
Nimrod Andy 已提交
2614 2615
	.get_wol		= fec_enet_get_wol,
	.set_wol		= fec_enet_set_wol,
2616
};
L
Linus Torvalds 已提交
2617

2618
static int fec_enet_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
L
Linus Torvalds 已提交
2619
{
2620
	struct fec_enet_private *fep = netdev_priv(ndev);
2621
	struct phy_device *phydev = fep->phy_dev;
L
Linus Torvalds 已提交
2622

2623
	if (!netif_running(ndev))
2624
		return -EINVAL;
L
Linus Torvalds 已提交
2625

2626 2627 2628
	if (!phydev)
		return -ENODEV;

2629 2630 2631 2632 2633 2634
	if (fep->bufdesc_ex) {
		if (cmd == SIOCSHWTSTAMP)
			return fec_ptp_set(ndev, rq);
		if (cmd == SIOCGHWTSTAMP)
			return fec_ptp_get(ndev, rq);
	}
2635

2636
	return phy_mii_ioctl(phydev, rq, cmd);
L
Linus Torvalds 已提交
2637 2638
}

2639
static void fec_enet_free_buffers(struct net_device *ndev)
S
Sascha Hauer 已提交
2640
{
2641
	struct fec_enet_private *fep = netdev_priv(ndev);
2642
	unsigned int i;
S
Sascha Hauer 已提交
2643 2644
	struct sk_buff *skb;
	struct bufdesc	*bdp;
2645 2646
	struct fec_enet_priv_tx_q *txq;
	struct fec_enet_priv_rx_q *rxq;
2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657
	unsigned int q;

	for (q = 0; q < fep->num_rx_queues; q++) {
		rxq = fep->rx_queue[q];
		bdp = rxq->rx_bd_base;
		for (i = 0; i < rxq->rx_ring_size; i++) {
			skb = rxq->rx_skbuff[i];
			rxq->rx_skbuff[i] = NULL;
			if (skb) {
				dma_unmap_single(&fep->pdev->dev,
						 bdp->cbd_bufaddr,
2658
						 FEC_ENET_RX_FRSIZE - fep->rx_align,
2659 2660 2661 2662 2663 2664
						 DMA_FROM_DEVICE);
				dev_kfree_skb(skb);
			}
			bdp = fec_enet_get_nextdesc(bdp, fep, q);
		}
	}
2665

2666 2667 2668 2669 2670 2671 2672 2673
	for (q = 0; q < fep->num_tx_queues; q++) {
		txq = fep->tx_queue[q];
		bdp = txq->tx_bd_base;
		for (i = 0; i < txq->tx_ring_size; i++) {
			kfree(txq->tx_bounce[i]);
			txq->tx_bounce[i] = NULL;
			skb = txq->tx_skbuff[i];
			txq->tx_skbuff[i] = NULL;
S
Sascha Hauer 已提交
2674
			dev_kfree_skb(skb);
2675
		}
S
Sascha Hauer 已提交
2676
	}
2677
}
S
Sascha Hauer 已提交
2678

2679 2680 2681 2682 2683 2684 2685 2686 2687 2688 2689 2690 2691 2692 2693 2694
static void fec_enet_free_queue(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int i;
	struct fec_enet_priv_tx_q *txq;

	for (i = 0; i < fep->num_tx_queues; i++)
		if (fep->tx_queue[i] && fep->tx_queue[i]->tso_hdrs) {
			txq = fep->tx_queue[i];
			dma_free_coherent(NULL,
					  txq->tx_ring_size * TSO_HEADER_SIZE,
					  txq->tso_hdrs,
					  txq->tso_hdrs_dma);
		}

	for (i = 0; i < fep->num_rx_queues; i++)
2695
		kfree(fep->rx_queue[i]);
2696
	for (i = 0; i < fep->num_tx_queues; i++)
2697
		kfree(fep->tx_queue[i]);
2698 2699 2700 2701 2702 2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719 2720 2721 2722 2723 2724 2725 2726 2727 2728 2729
}

static int fec_enet_alloc_queue(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	int i;
	int ret = 0;
	struct fec_enet_priv_tx_q *txq;

	for (i = 0; i < fep->num_tx_queues; i++) {
		txq = kzalloc(sizeof(*txq), GFP_KERNEL);
		if (!txq) {
			ret = -ENOMEM;
			goto alloc_failed;
		}

		fep->tx_queue[i] = txq;
		txq->tx_ring_size = TX_RING_SIZE;
		fep->total_tx_ring_size += fep->tx_queue[i]->tx_ring_size;

		txq->tx_stop_threshold = FEC_MAX_SKB_DESCS;
		txq->tx_wake_threshold =
				(txq->tx_ring_size - txq->tx_stop_threshold) / 2;

		txq->tso_hdrs = dma_alloc_coherent(NULL,
					txq->tx_ring_size * TSO_HEADER_SIZE,
					&txq->tso_hdrs_dma,
					GFP_KERNEL);
		if (!txq->tso_hdrs) {
			ret = -ENOMEM;
			goto alloc_failed;
		}
2730
	}
2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747

	for (i = 0; i < fep->num_rx_queues; i++) {
		fep->rx_queue[i] = kzalloc(sizeof(*fep->rx_queue[i]),
					   GFP_KERNEL);
		if (!fep->rx_queue[i]) {
			ret = -ENOMEM;
			goto alloc_failed;
		}

		fep->rx_queue[i]->rx_ring_size = RX_RING_SIZE;
		fep->total_rx_ring_size += fep->rx_queue[i]->rx_ring_size;
	}
	return ret;

alloc_failed:
	fec_enet_free_queue(ndev);
	return ret;
S
Sascha Hauer 已提交
2748 2749
}

2750 2751
static int
fec_enet_alloc_rxq_buffers(struct net_device *ndev, unsigned int queue)
S
Sascha Hauer 已提交
2752
{
2753
	struct fec_enet_private *fep = netdev_priv(ndev);
2754
	unsigned int i;
S
Sascha Hauer 已提交
2755 2756
	struct sk_buff *skb;
	struct bufdesc	*bdp;
2757
	struct fec_enet_priv_rx_q *rxq;
S
Sascha Hauer 已提交
2758

2759
	rxq = fep->rx_queue[queue];
2760 2761
	bdp = rxq->rx_bd_base;
	for (i = 0; i < rxq->rx_ring_size; i++) {
2762
		skb = netdev_alloc_skb(ndev, FEC_ENET_RX_FRSIZE);
2763 2764
		if (!skb)
			goto err_alloc;
S
Sascha Hauer 已提交
2765

2766
		if (fec_enet_new_rxbdp(ndev, bdp, skb)) {
2767
			dev_kfree_skb(skb);
2768
			goto err_alloc;
2769
		}
2770

2771
		rxq->rx_skbuff[i] = skb;
S
Sascha Hauer 已提交
2772
		bdp->cbd_sc = BD_ENET_RX_EMPTY;
2773 2774 2775 2776 2777 2778

		if (fep->bufdesc_ex) {
			struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;
			ebdp->cbd_esc = BD_ENET_RX_INT;
		}

2779
		bdp = fec_enet_get_nextdesc(bdp, fep, queue);
S
Sascha Hauer 已提交
2780 2781 2782
	}

	/* Set the last buffer to wrap. */
2783
	bdp = fec_enet_get_prevdesc(bdp, fep, queue);
S
Sascha Hauer 已提交
2784
	bdp->cbd_sc |= BD_SC_WRAP;
2785
	return 0;
S
Sascha Hauer 已提交
2786

2787 2788 2789 2790 2791 2792 2793 2794 2795 2796 2797 2798 2799 2800
 err_alloc:
	fec_enet_free_buffers(ndev);
	return -ENOMEM;
}

static int
fec_enet_alloc_txq_buffers(struct net_device *ndev, unsigned int queue)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	unsigned int i;
	struct bufdesc  *bdp;
	struct fec_enet_priv_tx_q *txq;

	txq = fep->tx_queue[queue];
2801 2802 2803 2804
	bdp = txq->tx_bd_base;
	for (i = 0; i < txq->tx_ring_size; i++) {
		txq->tx_bounce[i] = kmalloc(FEC_ENET_TX_FRSIZE, GFP_KERNEL);
		if (!txq->tx_bounce[i])
2805
			goto err_alloc;
S
Sascha Hauer 已提交
2806 2807 2808

		bdp->cbd_sc = 0;
		bdp->cbd_bufaddr = 0;
2809

2810 2811
		if (fep->bufdesc_ex) {
			struct bufdesc_ex *ebdp = (struct bufdesc_ex *)bdp;
2812
			ebdp->cbd_esc = BD_ENET_TX_INT;
2813 2814
		}

2815
		bdp = fec_enet_get_nextdesc(bdp, fep, queue);
S
Sascha Hauer 已提交
2816 2817 2818
	}

	/* Set the last buffer to wrap. */
2819
	bdp = fec_enet_get_prevdesc(bdp, fep, queue);
S
Sascha Hauer 已提交
2820 2821 2822
	bdp->cbd_sc |= BD_SC_WRAP;

	return 0;
2823 2824 2825 2826

 err_alloc:
	fec_enet_free_buffers(ndev);
	return -ENOMEM;
S
Sascha Hauer 已提交
2827 2828
}

2829 2830 2831 2832 2833 2834 2835 2836 2837 2838 2839 2840 2841 2842 2843
static int fec_enet_alloc_buffers(struct net_device *ndev)
{
	struct fec_enet_private *fep = netdev_priv(ndev);
	unsigned int i;

	for (i = 0; i < fep->num_rx_queues; i++)
		if (fec_enet_alloc_rxq_buffers(ndev, i))
			return -ENOMEM;

	for (i = 0; i < fep->num_tx_queues; i++)
		if (fec_enet_alloc_txq_buffers(ndev, i))
			return -ENOMEM;
	return 0;
}

L
Linus Torvalds 已提交
2844
static int
2845
fec_enet_open(struct net_device *ndev)
L
Linus Torvalds 已提交
2846
{
2847
	struct fec_enet_private *fep = netdev_priv(ndev);
S
Sascha Hauer 已提交
2848
	int ret;
L
Linus Torvalds 已提交
2849

N
Nimrod Andy 已提交
2850
	pinctrl_pm_select_default_state(&fep->pdev->dev);
2851 2852
	ret = fec_enet_clk_enable(ndev, true);
	if (ret)
2853
		return ret;
2854

L
Linus Torvalds 已提交
2855 2856 2857 2858
	/* I should reset the ring buffers here, but I don't yet know
	 * a simple way to do that.
	 */

2859
	ret = fec_enet_alloc_buffers(ndev);
S
Sascha Hauer 已提交
2860
	if (ret)
2861
		goto err_enet_alloc;
S
Sascha Hauer 已提交
2862

2863 2864 2865
	/* Init MAC prior to mii bus probe */
	fec_restart(ndev);

2866
	/* Probe and connect to PHY when open the interface */
2867
	ret = fec_enet_mii_probe(ndev);
2868 2869
	if (ret)
		goto err_enet_mii_probe;
2870 2871

	napi_enable(&fep->napi);
2872
	phy_start(fep->phy_dev);
2873 2874
	netif_tx_start_all_queues(ndev);

N
Nimrod Andy 已提交
2875 2876 2877
	device_set_wakeup_enable(&ndev->dev, fep->wol_flag &
				 FEC_WOL_FLAG_ENABLE);

S
Sascha Hauer 已提交
2878
	return 0;
2879 2880 2881 2882 2883 2884 2885

err_enet_mii_probe:
	fec_enet_free_buffers(ndev);
err_enet_alloc:
	fec_enet_clk_enable(ndev, false);
	pinctrl_pm_select_sleep_state(&fep->pdev->dev);
	return ret;
L
Linus Torvalds 已提交
2886 2887 2888
}

static int
2889
fec_enet_close(struct net_device *ndev)
L
Linus Torvalds 已提交
2890
{
2891
	struct fec_enet_private *fep = netdev_priv(ndev);
L
Linus Torvalds 已提交
2892

2893 2894
	phy_stop(fep->phy_dev);

2895 2896 2897
	if (netif_device_present(ndev)) {
		napi_disable(&fep->napi);
		netif_tx_disable(ndev);
2898
		fec_stop(ndev);
2899
	}
L
Linus Torvalds 已提交
2900

2901
	phy_disconnect(fep->phy_dev);
2902
	fep->phy_dev = NULL;
2903

2904
	fec_enet_clk_enable(ndev, false);
N
Nimrod Andy 已提交
2905
	pinctrl_pm_select_sleep_state(&fep->pdev->dev);
2906
	fec_enet_free_buffers(ndev);
S
Sascha Hauer 已提交
2907

L
Linus Torvalds 已提交
2908 2909 2910 2911 2912 2913 2914 2915 2916 2917 2918 2919 2920 2921 2922 2923
	return 0;
}

/* Set or clear the multicast filter for this adaptor.
 * Skeleton taken from sunlance driver.
 * The CPM Ethernet implementation allows Multicast as well as individual
 * MAC address filtering.  Some of the drivers check to make sure it is
 * a group multicast address, and discard those that are not.  I guess I
 * will do the same for now, but just remove the test if you want
 * individual filtering as well (do the upper net layers want or support
 * this kind of feature?).
 */

#define HASH_BITS	6		/* #bits in hash */
#define CRC32_POLY	0xEDB88320

2924
static void set_multicast_list(struct net_device *ndev)
L
Linus Torvalds 已提交
2925
{
2926
	struct fec_enet_private *fep = netdev_priv(ndev);
2927
	struct netdev_hw_addr *ha;
2928
	unsigned int i, bit, data, crc, tmp;
L
Linus Torvalds 已提交
2929 2930
	unsigned char hash;

2931
	if (ndev->flags & IFF_PROMISC) {
S
Sascha Hauer 已提交
2932 2933 2934
		tmp = readl(fep->hwp + FEC_R_CNTRL);
		tmp |= 0x8;
		writel(tmp, fep->hwp + FEC_R_CNTRL);
2935 2936
		return;
	}
L
Linus Torvalds 已提交
2937

2938 2939 2940 2941
	tmp = readl(fep->hwp + FEC_R_CNTRL);
	tmp &= ~0x8;
	writel(tmp, fep->hwp + FEC_R_CNTRL);

2942
	if (ndev->flags & IFF_ALLMULTI) {
2943 2944 2945 2946 2947 2948 2949 2950 2951 2952 2953 2954 2955 2956
		/* Catch all multicast addresses, so set the
		 * filter to all 1's
		 */
		writel(0xffffffff, fep->hwp + FEC_GRP_HASH_TABLE_HIGH);
		writel(0xffffffff, fep->hwp + FEC_GRP_HASH_TABLE_LOW);

		return;
	}

	/* Clear filter and add the addresses in hash register
	 */
	writel(0, fep->hwp + FEC_GRP_HASH_TABLE_HIGH);
	writel(0, fep->hwp + FEC_GRP_HASH_TABLE_LOW);

2957
	netdev_for_each_mc_addr(ha, ndev) {
2958 2959 2960
		/* calculate crc32 value of mac address */
		crc = 0xffffffff;

2961
		for (i = 0; i < ndev->addr_len; i++) {
2962
			data = ha->addr[i];
2963 2964 2965
			for (bit = 0; bit < 8; bit++, data >>= 1) {
				crc = (crc >> 1) ^
				(((crc ^ data) & 1) ? CRC32_POLY : 0);
L
Linus Torvalds 已提交
2966 2967
			}
		}
2968 2969 2970 2971 2972 2973 2974 2975 2976 2977 2978 2979 2980 2981 2982

		/* only upper 6 bits (HASH_BITS) are used
		 * which point to specific bit in he hash registers
		 */
		hash = (crc >> (32 - HASH_BITS)) & 0x3f;

		if (hash > 31) {
			tmp = readl(fep->hwp + FEC_GRP_HASH_TABLE_HIGH);
			tmp |= 1 << (hash - 32);
			writel(tmp, fep->hwp + FEC_GRP_HASH_TABLE_HIGH);
		} else {
			tmp = readl(fep->hwp + FEC_GRP_HASH_TABLE_LOW);
			tmp |= 1 << hash;
			writel(tmp, fep->hwp + FEC_GRP_HASH_TABLE_LOW);
		}
L
Linus Torvalds 已提交
2983 2984 2985
	}
}

S
Sascha Hauer 已提交
2986
/* Set a MAC change in hardware. */
S
Sascha Hauer 已提交
2987
static int
2988
fec_set_mac_address(struct net_device *ndev, void *p)
L
Linus Torvalds 已提交
2989
{
2990
	struct fec_enet_private *fep = netdev_priv(ndev);
S
Sascha Hauer 已提交
2991 2992
	struct sockaddr *addr = p;

2993 2994 2995 2996 2997
	if (addr) {
		if (!is_valid_ether_addr(addr->sa_data))
			return -EADDRNOTAVAIL;
		memcpy(ndev->dev_addr, addr->sa_data, ndev->addr_len);
	}
L
Linus Torvalds 已提交
2998

2999 3000
	writel(ndev->dev_addr[3] | (ndev->dev_addr[2] << 8) |
		(ndev->dev_addr[1] << 16) | (ndev->dev_addr[0] << 24),
S
Sascha Hauer 已提交
3001
		fep->hwp + FEC_ADDR_LOW);
3002
	writel((ndev->dev_addr[5] << 16) | (ndev->dev_addr[4] << 24),
3003
		fep->hwp + FEC_ADDR_HIGH);
S
Sascha Hauer 已提交
3004
	return 0;
L
Linus Torvalds 已提交
3005 3006
}

3007
#ifdef CONFIG_NET_POLL_CONTROLLER
3008 3009
/**
 * fec_poll_controller - FEC Poll controller function
3010 3011 3012 3013 3014
 * @dev: The FEC network adapter
 *
 * Polled functionality used by netconsole and others in non interrupt mode
 *
 */
3015
static void fec_poll_controller(struct net_device *dev)
3016 3017 3018 3019 3020 3021 3022 3023 3024 3025 3026 3027 3028 3029
{
	int i;
	struct fec_enet_private *fep = netdev_priv(dev);

	for (i = 0; i < FEC_IRQ_NUM; i++) {
		if (fep->irq[i] > 0) {
			disable_irq(fep->irq[i]);
			fec_enet_interrupt(fep->irq[i], dev);
			enable_irq(fep->irq[i]);
		}
	}
}
#endif

3030
#define FEATURES_NEED_QUIESCE NETIF_F_RXCSUM
3031
static inline void fec_enet_set_netdev_features(struct net_device *netdev,
3032 3033 3034 3035 3036 3037 3038 3039 3040 3041 3042 3043 3044
	netdev_features_t features)
{
	struct fec_enet_private *fep = netdev_priv(netdev);
	netdev_features_t changed = features ^ netdev->features;

	netdev->features = features;

	/* Receive checksum has been changed */
	if (changed & NETIF_F_RXCSUM) {
		if (features & NETIF_F_RXCSUM)
			fep->csum_flags |= FLAG_RX_CSUM_ENABLED;
		else
			fep->csum_flags &= ~FLAG_RX_CSUM_ENABLED;
3045
	}
3046 3047 3048 3049 3050 3051 3052
}

static int fec_set_features(struct net_device *netdev,
	netdev_features_t features)
{
	struct fec_enet_private *fep = netdev_priv(netdev);
	netdev_features_t changed = features ^ netdev->features;
3053

3054
	if (netif_running(netdev) && changed & FEATURES_NEED_QUIESCE) {
3055 3056 3057 3058
		napi_disable(&fep->napi);
		netif_tx_lock_bh(netdev);
		fec_stop(netdev);
		fec_enet_set_netdev_features(netdev, features);
3059
		fec_restart(netdev);
3060
		netif_tx_wake_all_queues(netdev);
3061 3062
		netif_tx_unlock_bh(netdev);
		napi_enable(&fep->napi);
3063 3064
	} else {
		fec_enet_set_netdev_features(netdev, features);
3065 3066 3067 3068 3069
	}

	return 0;
}

S
Sascha Hauer 已提交
3070 3071 3072 3073
static const struct net_device_ops fec_netdev_ops = {
	.ndo_open		= fec_enet_open,
	.ndo_stop		= fec_enet_close,
	.ndo_start_xmit		= fec_enet_start_xmit,
3074
	.ndo_set_rx_mode	= set_multicast_list,
3075
	.ndo_change_mtu		= eth_change_mtu,
S
Sascha Hauer 已提交
3076 3077 3078
	.ndo_validate_addr	= eth_validate_addr,
	.ndo_tx_timeout		= fec_timeout,
	.ndo_set_mac_address	= fec_set_mac_address,
3079
	.ndo_do_ioctl		= fec_enet_ioctl,
3080 3081 3082
#ifdef CONFIG_NET_POLL_CONTROLLER
	.ndo_poll_controller	= fec_poll_controller,
#endif
3083
	.ndo_set_features	= fec_set_features,
S
Sascha Hauer 已提交
3084 3085
};

L
Linus Torvalds 已提交
3086 3087
 /*
  * XXX:  We need to clean up on failure exits here.
3088
  *
L
Linus Torvalds 已提交
3089
  */
3090
static int fec_enet_init(struct net_device *ndev)
L
Linus Torvalds 已提交
3091
{
3092
	struct fec_enet_private *fep = netdev_priv(ndev);
3093 3094
	struct fec_enet_priv_tx_q *txq;
	struct fec_enet_priv_rx_q *rxq;
S
Sascha Hauer 已提交
3095
	struct bufdesc *cbd_base;
3096
	dma_addr_t bd_dma;
3097
	int bd_size;
3098
	unsigned int i;
3099

3100 3101 3102 3103 3104 3105 3106 3107
#if defined(CONFIG_ARM)
	fep->rx_align = 0xf;
	fep->tx_align = 0xf;
#else
	fep->rx_align = 0x3;
	fep->tx_align = 0x3;
#endif

3108
	fec_enet_alloc_queue(ndev);
N
Nimrod Andy 已提交
3109

3110 3111 3112 3113
	if (fep->bufdesc_ex)
		fep->bufdesc_size = sizeof(struct bufdesc_ex);
	else
		fep->bufdesc_size = sizeof(struct bufdesc);
3114
	bd_size = (fep->total_tx_ring_size + fep->total_rx_ring_size) *
3115
			fep->bufdesc_size;
L
Linus Torvalds 已提交
3116

S
Sascha Hauer 已提交
3117
	/* Allocate memory for buffer descriptors. */
3118 3119
	cbd_base = dmam_alloc_coherent(&fep->pdev->dev, bd_size, &bd_dma,
				       GFP_KERNEL);
3120
	if (!cbd_base) {
N
Nimrod Andy 已提交
3121 3122 3123
		return -ENOMEM;
	}

3124
	memset(cbd_base, 0, bd_size);
L
Linus Torvalds 已提交
3125

3126
	/* Get the Ethernet address */
3127
	fec_get_mac(ndev);
3128 3129
	/* make sure MAC we just acquired is programmed into the hw */
	fec_set_mac_address(ndev, NULL);
L
Linus Torvalds 已提交
3130

S
Sascha Hauer 已提交
3131
	/* Set receive and transmit descriptor base. */
3132 3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156 3157 3158 3159 3160
	for (i = 0; i < fep->num_rx_queues; i++) {
		rxq = fep->rx_queue[i];
		rxq->index = i;
		rxq->rx_bd_base = (struct bufdesc *)cbd_base;
		rxq->bd_dma = bd_dma;
		if (fep->bufdesc_ex) {
			bd_dma += sizeof(struct bufdesc_ex) * rxq->rx_ring_size;
			cbd_base = (struct bufdesc *)
				(((struct bufdesc_ex *)cbd_base) + rxq->rx_ring_size);
		} else {
			bd_dma += sizeof(struct bufdesc) * rxq->rx_ring_size;
			cbd_base += rxq->rx_ring_size;
		}
	}

	for (i = 0; i < fep->num_tx_queues; i++) {
		txq = fep->tx_queue[i];
		txq->index = i;
		txq->tx_bd_base = (struct bufdesc *)cbd_base;
		txq->bd_dma = bd_dma;
		if (fep->bufdesc_ex) {
			bd_dma += sizeof(struct bufdesc_ex) * txq->tx_ring_size;
			cbd_base = (struct bufdesc *)
			 (((struct bufdesc_ex *)cbd_base) + txq->tx_ring_size);
		} else {
			bd_dma += sizeof(struct bufdesc) * txq->tx_ring_size;
			cbd_base += txq->tx_ring_size;
		}
	}
3161

L
Linus Torvalds 已提交
3162

S
Sascha Hauer 已提交
3163
	/* The FEC Ethernet specific entries in the device structure */
3164 3165 3166
	ndev->watchdog_timeo = TX_TIMEOUT;
	ndev->netdev_ops = &fec_netdev_ops;
	ndev->ethtool_ops = &fec_enet_ethtool_ops;
3167

3168
	writel(FEC_RX_DISABLED_IMASK, fep->hwp + FEC_IMASK);
F
Fabio Estevam 已提交
3169
	netif_napi_add(ndev, &fep->napi, fec_enet_rx_napi, NAPI_POLL_WEIGHT);
3170

3171
	if (fep->quirks & FEC_QUIRK_HAS_VLAN)
3172 3173 3174
		/* enable hw VLAN support */
		ndev->features |= NETIF_F_HW_VLAN_CTAG_RX;

3175
	if (fep->quirks & FEC_QUIRK_HAS_CSUM) {
N
Nimrod Andy 已提交
3176 3177
		ndev->gso_max_segs = FEC_MAX_TSO_SEGS;

3178 3179
		/* enable hw accelerator */
		ndev->features |= (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM
N
Nimrod Andy 已提交
3180
				| NETIF_F_RXCSUM | NETIF_F_SG | NETIF_F_TSO);
3181 3182
		fep->csum_flags |= FLAG_RX_CSUM_ENABLED;
	}
3183

3184
	if (fep->quirks & FEC_QUIRK_HAS_AVB) {
3185 3186 3187 3188
		fep->tx_align = 0;
		fep->rx_align = 0x3f;
	}

3189 3190
	ndev->hw_features = ndev->features;

3191
	fec_restart(ndev);
L
Linus Torvalds 已提交
3192 3193 3194 3195

	return 0;
}

3196
#ifdef CONFIG_OF
3197
static void fec_reset_phy(struct platform_device *pdev)
3198 3199
{
	int err, phy_reset;
3200
	int msec = 1;
3201 3202 3203
	struct device_node *np = pdev->dev.of_node;

	if (!np)
3204
		return;
3205

3206 3207 3208 3209 3210
	of_property_read_u32(np, "phy-reset-duration", &msec);
	/* A sane reset duration should not be longer than 1s */
	if (msec > 1000)
		msec = 1;

3211
	phy_reset = of_get_named_gpio(np, "phy-reset-gpios", 0);
3212 3213 3214
	if (!gpio_is_valid(phy_reset))
		return;

3215 3216
	err = devm_gpio_request_one(&pdev->dev, phy_reset,
				    GPIOF_OUT_INIT_LOW, "phy-reset");
3217
	if (err) {
3218
		dev_err(&pdev->dev, "failed to get phy-reset-gpios: %d\n", err);
3219
		return;
3220
	}
3221
	msleep(msec);
3222 3223 3224
	gpio_set_value(phy_reset, 1);
}
#else /* CONFIG_OF */
3225
static void fec_reset_phy(struct platform_device *pdev)
3226 3227 3228 3229 3230 3231 3232 3233
{
	/*
	 * In case of platform probe, the reset has been done
	 * by machine code.
	 */
}
#endif /* CONFIG_OF */

3234 3235 3236 3237 3238 3239 3240 3241 3242 3243 3244 3245 3246
static void
fec_enet_get_queue_num(struct platform_device *pdev, int *num_tx, int *num_rx)
{
	struct device_node *np = pdev->dev.of_node;
	int err;

	*num_tx = *num_rx = 1;

	if (!np || !of_device_is_available(np))
		return;

	/* parse the num of tx and rx queues */
	err = of_property_read_u32(np, "fsl,num-tx-queues", num_tx);
3247
	if (err)
3248
		*num_tx = 1;
3249 3250 3251

	err = of_property_read_u32(np, "fsl,num-rx-queues", num_rx);
	if (err)
3252 3253 3254
		*num_rx = 1;

	if (*num_tx < 1 || *num_tx > FEC_ENET_MAX_TX_QS) {
3255 3256
		dev_warn(&pdev->dev, "Invalid num_tx(=%d), fall back to 1\n",
			 *num_tx);
3257 3258 3259 3260 3261
		*num_tx = 1;
		return;
	}

	if (*num_rx < 1 || *num_rx > FEC_ENET_MAX_RX_QS) {
3262 3263
		dev_warn(&pdev->dev, "Invalid num_rx(=%d), fall back to 1\n",
			 *num_rx);
3264 3265 3266 3267 3268 3269
		*num_rx = 1;
		return;
	}

}

3270
static int
3271 3272 3273
fec_probe(struct platform_device *pdev)
{
	struct fec_enet_private *fep;
3274
	struct fec_platform_data *pdata;
3275 3276 3277
	struct net_device *ndev;
	int i, irq, ret = 0;
	struct resource *r;
3278
	const struct of_device_id *of_id;
S
Shawn Guo 已提交
3279
	static int dev_id;
3280
	struct device_node *np = pdev->dev.of_node, *phy_node;
3281 3282
	int num_tx_qs;
	int num_rx_qs;
3283

3284 3285
	fec_enet_get_queue_num(pdev, &num_tx_qs, &num_rx_qs);

3286
	/* Init network device */
3287 3288
	ndev = alloc_etherdev_mqs(sizeof(struct fec_enet_private),
				  num_tx_qs, num_rx_qs);
3289 3290
	if (!ndev)
		return -ENOMEM;
3291 3292 3293 3294 3295 3296

	SET_NETDEV_DEV(ndev, &pdev->dev);

	/* setup board info structure */
	fep = netdev_priv(ndev);

3297 3298 3299 3300 3301
	of_id = of_match_device(fec_dt_ids, &pdev->dev);
	if (of_id)
		pdev->id_entry = of_id->data;
	fep->quirks = pdev->id_entry->driver_data;

3302
	fep->netdev = ndev;
3303 3304 3305
	fep->num_rx_queues = num_rx_qs;
	fep->num_tx_queues = num_tx_qs;

G
Guenter Roeck 已提交
3306
#if !defined(CONFIG_M5272)
3307
	/* default enable pause frame auto negotiation */
3308
	if (fep->quirks & FEC_QUIRK_HAS_GBIT)
3309
		fep->pause_flag |= FEC_PAUSE_FLAG_AUTONEG;
G
Guenter Roeck 已提交
3310
#endif
3311

N
Nimrod Andy 已提交
3312 3313 3314
	/* Select default pin state */
	pinctrl_pm_select_default_state(&pdev->dev);

3315
	r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
3316 3317 3318 3319 3320 3321
	fep->hwp = devm_ioremap_resource(&pdev->dev, r);
	if (IS_ERR(fep->hwp)) {
		ret = PTR_ERR(fep->hwp);
		goto failed_ioremap;
	}

3322
	fep->pdev = pdev;
S
Shawn Guo 已提交
3323
	fep->dev_id = dev_id++;
3324 3325 3326

	platform_set_drvdata(pdev, ndev);

N
Nimrod Andy 已提交
3327 3328 3329
	if (of_get_property(np, "fsl,magic-packet", NULL))
		fep->wol_flag |= FEC_WOL_HAS_MAGIC_PACKET;

3330 3331 3332 3333 3334 3335 3336 3337 3338 3339 3340 3341
	phy_node = of_parse_phandle(np, "phy-handle", 0);
	if (!phy_node && of_phy_is_fixed_link(np)) {
		ret = of_phy_register_fixed_link(np);
		if (ret < 0) {
			dev_err(&pdev->dev,
				"broken fixed-link specification\n");
			goto failed_phy;
		}
		phy_node = of_node_get(np);
	}
	fep->phy_node = phy_node;

3342
	ret = of_get_phy_mode(pdev->dev.of_node);
3343
	if (ret < 0) {
J
Jingoo Han 已提交
3344
		pdata = dev_get_platdata(&pdev->dev);
3345 3346 3347 3348 3349 3350 3351 3352
		if (pdata)
			fep->phy_interface = pdata->phy;
		else
			fep->phy_interface = PHY_INTERFACE_MODE_MII;
	} else {
		fep->phy_interface = ret;
	}

3353 3354 3355
	fep->clk_ipg = devm_clk_get(&pdev->dev, "ipg");
	if (IS_ERR(fep->clk_ipg)) {
		ret = PTR_ERR(fep->clk_ipg);
3356 3357
		goto failed_clk;
	}
3358 3359 3360 3361 3362 3363 3364

	fep->clk_ahb = devm_clk_get(&pdev->dev, "ahb");
	if (IS_ERR(fep->clk_ahb)) {
		ret = PTR_ERR(fep->clk_ahb);
		goto failed_clk;
	}

3365 3366
	fep->itr_clk_rate = clk_get_rate(fep->clk_ahb);

3367 3368 3369 3370 3371
	/* enet_out is optional, depends on board */
	fep->clk_enet_out = devm_clk_get(&pdev->dev, "enet_out");
	if (IS_ERR(fep->clk_enet_out))
		fep->clk_enet_out = NULL;

3372 3373
	fep->ptp_clk_on = false;
	mutex_init(&fep->ptp_clk_mutex);
3374 3375 3376 3377 3378 3379

	/* clk_ref is optional, depends on board */
	fep->clk_ref = devm_clk_get(&pdev->dev, "enet_clk_ref");
	if (IS_ERR(fep->clk_ref))
		fep->clk_ref = NULL;

3380
	fep->bufdesc_ex = fep->quirks & FEC_QUIRK_HAS_BUFDESC_EX;
3381 3382
	fep->clk_ptp = devm_clk_get(&pdev->dev, "ptp");
	if (IS_ERR(fep->clk_ptp)) {
3383
		fep->clk_ptp = NULL;
3384
		fep->bufdesc_ex = false;
3385 3386
	}

3387
	ret = fec_enet_clk_enable(ndev, true);
3388 3389 3390
	if (ret)
		goto failed_clk;

3391 3392 3393
	fep->reg_phy = devm_regulator_get(&pdev->dev, "phy");
	if (!IS_ERR(fep->reg_phy)) {
		ret = regulator_enable(fep->reg_phy);
3394 3395 3396 3397 3398
		if (ret) {
			dev_err(&pdev->dev,
				"Failed to enable phy regulator: %d\n", ret);
			goto failed_regulator;
		}
3399 3400
	} else {
		fep->reg_phy = NULL;
3401 3402
	}

3403 3404
	fec_reset_phy(pdev);

F
Fabio Estevam 已提交
3405
	if (fep->bufdesc_ex)
3406
		fec_ptp_init(pdev);
F
Fabio Estevam 已提交
3407 3408 3409 3410 3411 3412 3413 3414 3415 3416 3417 3418 3419

	ret = fec_enet_init(ndev);
	if (ret)
		goto failed_init;

	for (i = 0; i < FEC_IRQ_NUM; i++) {
		irq = platform_get_irq(pdev, i);
		if (irq < 0) {
			if (i)
				break;
			ret = irq;
			goto failed_irq;
		}
F
Fabio Estevam 已提交
3420
		ret = devm_request_irq(&pdev->dev, irq, fec_enet_interrupt,
3421
				       0, pdev->name, ndev);
F
Fabio Estevam 已提交
3422
		if (ret)
F
Fabio Estevam 已提交
3423
			goto failed_irq;
N
Nimrod Andy 已提交
3424 3425

		fep->irq[i] = irq;
F
Fabio Estevam 已提交
3426 3427
	}

3428
	init_completion(&fep->mdio_done);
3429 3430 3431 3432
	ret = fec_enet_mii_init(pdev);
	if (ret)
		goto failed_mii_init;

3433 3434
	/* Carrier starts down, phylib will bring it up */
	netif_carrier_off(ndev);
3435
	fec_enet_clk_enable(ndev, false);
N
Nimrod Andy 已提交
3436
	pinctrl_pm_select_sleep_state(&pdev->dev);
3437

3438 3439 3440 3441
	ret = register_netdev(ndev);
	if (ret)
		goto failed_register;

N
Nimrod Andy 已提交
3442 3443 3444
	device_init_wakeup(&ndev->dev, fep->wol_flag &
			   FEC_WOL_HAS_MAGIC_PACKET);

F
Fabio Estevam 已提交
3445 3446 3447
	if (fep->bufdesc_ex && fep->ptp_clock)
		netdev_info(ndev, "registered PHC device %d\n", fep->dev_id);

3448
	fep->rx_copybreak = COPYBREAK_DEFAULT;
3449
	INIT_WORK(&fep->tx_timeout_work, fec_enet_timeout_work);
3450 3451 3452
	return 0;

failed_register:
3453 3454
	fec_enet_mii_remove(fep);
failed_mii_init:
3455 3456
failed_irq:
failed_init:
3457 3458
	if (fep->reg_phy)
		regulator_disable(fep->reg_phy);
3459
failed_regulator:
3460
	fec_enet_clk_enable(ndev, false);
3461
failed_clk:
3462 3463
failed_phy:
	of_node_put(phy_node);
3464 3465 3466 3467 3468 3469
failed_ioremap:
	free_netdev(ndev);

	return ret;
}

3470
static int
3471 3472 3473 3474 3475
fec_drv_remove(struct platform_device *pdev)
{
	struct net_device *ndev = platform_get_drvdata(pdev);
	struct fec_enet_private *fep = netdev_priv(ndev);

3476
	cancel_delayed_work_sync(&fep->time_keep);
3477
	cancel_work_sync(&fep->tx_timeout_work);
L
Lothar Waßmann 已提交
3478
	unregister_netdev(ndev);
3479
	fec_enet_mii_remove(fep);
3480 3481
	if (fep->reg_phy)
		regulator_disable(fep->reg_phy);
3482 3483
	if (fep->ptp_clock)
		ptp_clock_unregister(fep->ptp_clock);
3484
	of_node_put(fep->phy_node);
3485
	free_netdev(ndev);
3486

3487 3488 3489
	return 0;
}

3490
static int __maybe_unused fec_suspend(struct device *dev)
3491
{
E
Eric Benard 已提交
3492
	struct net_device *ndev = dev_get_drvdata(dev);
3493
	struct fec_enet_private *fep = netdev_priv(ndev);
3494

3495
	rtnl_lock();
3496
	if (netif_running(ndev)) {
N
Nimrod Andy 已提交
3497 3498
		if (fep->wol_flag & FEC_WOL_FLAG_ENABLE)
			fep->wol_flag |= FEC_WOL_FLAG_SLEEP_ON;
3499
		phy_stop(fep->phy_dev);
3500 3501
		napi_disable(&fep->napi);
		netif_tx_lock_bh(ndev);
3502
		netif_device_detach(ndev);
3503 3504
		netif_tx_unlock_bh(ndev);
		fec_stop(ndev);
3505
		fec_enet_clk_enable(ndev, false);
N
Nimrod Andy 已提交
3506 3507
		if (!(fep->wol_flag & FEC_WOL_FLAG_ENABLE))
			pinctrl_pm_select_sleep_state(&fep->pdev->dev);
3508
	}
3509 3510
	rtnl_unlock();

N
Nimrod Andy 已提交
3511
	if (fep->reg_phy && !(fep->wol_flag & FEC_WOL_FLAG_ENABLE))
3512 3513
		regulator_disable(fep->reg_phy);

3514 3515 3516 3517 3518 3519
	/* SOC supply clock to phy, when clock is disabled, phy link down
	 * SOC control phy regulator, when regulator is disabled, phy link down
	 */
	if (fep->clk_enet_out || fep->reg_phy)
		fep->link = 0;

3520 3521 3522
	return 0;
}

3523
static int __maybe_unused fec_resume(struct device *dev)
3524
{
E
Eric Benard 已提交
3525
	struct net_device *ndev = dev_get_drvdata(dev);
3526
	struct fec_enet_private *fep = netdev_priv(ndev);
N
Nimrod Andy 已提交
3527
	struct fec_platform_data *pdata = fep->pdev->dev.platform_data;
3528
	int ret;
N
Nimrod Andy 已提交
3529
	int val;
3530

N
Nimrod Andy 已提交
3531
	if (fep->reg_phy && !(fep->wol_flag & FEC_WOL_FLAG_ENABLE)) {
3532 3533 3534 3535
		ret = regulator_enable(fep->reg_phy);
		if (ret)
			return ret;
	}
3536

3537
	rtnl_lock();
3538
	if (netif_running(ndev)) {
3539 3540 3541 3542 3543
		ret = fec_enet_clk_enable(ndev, true);
		if (ret) {
			rtnl_unlock();
			goto failed_clk;
		}
N
Nimrod Andy 已提交
3544 3545 3546 3547 3548 3549 3550 3551 3552 3553
		if (fep->wol_flag & FEC_WOL_FLAG_ENABLE) {
			if (pdata && pdata->sleep_mode_enable)
				pdata->sleep_mode_enable(false);
			val = readl(fep->hwp + FEC_ECNTRL);
			val &= ~(FEC_ECR_MAGICEN | FEC_ECR_SLEEP);
			writel(val, fep->hwp + FEC_ECNTRL);
			fep->wol_flag &= ~FEC_WOL_FLAG_SLEEP_ON;
		} else {
			pinctrl_pm_select_default_state(&fep->pdev->dev);
		}
3554
		fec_restart(ndev);
3555
		netif_tx_lock_bh(ndev);
3556
		netif_device_attach(ndev);
3557
		netif_tx_unlock_bh(ndev);
3558
		napi_enable(&fep->napi);
3559
		phy_start(fep->phy_dev);
3560
	}
3561
	rtnl_unlock();
3562

3563
	return 0;
3564

3565
failed_clk:
3566 3567 3568
	if (fep->reg_phy)
		regulator_disable(fep->reg_phy);
	return ret;
3569 3570
}

3571
static SIMPLE_DEV_PM_OPS(fec_pm_ops, fec_suspend, fec_resume);
3572

3573 3574
static struct platform_driver fec_driver = {
	.driver	= {
3575
		.name	= DRIVER_NAME,
E
Eric Benard 已提交
3576
		.pm	= &fec_pm_ops,
3577
		.of_match_table = fec_dt_ids,
3578
	},
3579
	.id_table = fec_devtype,
E
Eric Benard 已提交
3580
	.probe	= fec_probe,
3581
	.remove	= fec_drv_remove,
3582 3583
};

3584
module_platform_driver(fec_driver);
L
Linus Torvalds 已提交
3585

F
Fabio Estevam 已提交
3586
MODULE_ALIAS("platform:"DRIVER_NAME);
L
Linus Torvalds 已提交
3587
MODULE_LICENSE("GPL");