bfin_mac.c 43.1 KB
Newer Older
1
/*
2
 * Blackfin On-Chip MAC Driver
3
 *
4
 * Copyright 2004-2010 Analog Devices Inc.
5
 *
6
 * Enter bugs at http://blackfin.uclinux.org/
7
 *
8
 * Licensed under the GPL-2 or later.
9 10
 */

11 12 13 14 15
#define DRV_VERSION	"1.1"
#define DRV_DESC	"Blackfin on-chip Ethernet MAC driver"

#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32
#include <linux/init.h>
#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/sched.h>
#include <linux/slab.h>
#include <linux/delay.h>
#include <linux/timer.h>
#include <linux/errno.h>
#include <linux/irq.h>
#include <linux/io.h>
#include <linux/ioport.h>
#include <linux/crc32.h>
#include <linux/device.h>
#include <linux/spinlock.h>
#include <linux/mii.h>
#include <linux/netdevice.h>
#include <linux/etherdevice.h>
33
#include <linux/ethtool.h>
34 35 36 37 38 39
#include <linux/skbuff.h>
#include <linux/platform_device.h>

#include <asm/dma.h>
#include <linux/dma-mapping.h>

40
#include <asm/div64.h>
41
#include <asm/dpmc.h>
42 43 44
#include <asm/blackfin.h>
#include <asm/cacheflush.h>
#include <asm/portmux.h>
45
#include <mach/pll.h>
46 47 48

#include "bfin_mac.h"

49
MODULE_AUTHOR("Bryan Wu, Luke Yang");
50 51
MODULE_LICENSE("GPL");
MODULE_DESCRIPTION(DRV_DESC);
52
MODULE_ALIAS("platform:bfin_mac");
53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189

#if defined(CONFIG_BFIN_MAC_USE_L1)
# define bfin_mac_alloc(dma_handle, size)  l1_data_sram_zalloc(size)
# define bfin_mac_free(dma_handle, ptr)    l1_data_sram_free(ptr)
#else
# define bfin_mac_alloc(dma_handle, size) \
	dma_alloc_coherent(NULL, size, dma_handle, GFP_KERNEL)
# define bfin_mac_free(dma_handle, ptr) \
	dma_free_coherent(NULL, sizeof(*ptr), ptr, dma_handle)
#endif

#define PKT_BUF_SZ 1580

#define MAX_TIMEOUT_CNT	500

/* pointers to maintain transmit list */
static struct net_dma_desc_tx *tx_list_head;
static struct net_dma_desc_tx *tx_list_tail;
static struct net_dma_desc_rx *rx_list_head;
static struct net_dma_desc_rx *rx_list_tail;
static struct net_dma_desc_rx *current_rx_ptr;
static struct net_dma_desc_tx *current_tx_ptr;
static struct net_dma_desc_tx *tx_desc;
static struct net_dma_desc_rx *rx_desc;

static void desc_list_free(void)
{
	struct net_dma_desc_rx *r;
	struct net_dma_desc_tx *t;
	int i;
#if !defined(CONFIG_BFIN_MAC_USE_L1)
	dma_addr_t dma_handle = 0;
#endif

	if (tx_desc) {
		t = tx_list_head;
		for (i = 0; i < CONFIG_BFIN_TX_DESC_NUM; i++) {
			if (t) {
				if (t->skb) {
					dev_kfree_skb(t->skb);
					t->skb = NULL;
				}
				t = t->next;
			}
		}
		bfin_mac_free(dma_handle, tx_desc);
	}

	if (rx_desc) {
		r = rx_list_head;
		for (i = 0; i < CONFIG_BFIN_RX_DESC_NUM; i++) {
			if (r) {
				if (r->skb) {
					dev_kfree_skb(r->skb);
					r->skb = NULL;
				}
				r = r->next;
			}
		}
		bfin_mac_free(dma_handle, rx_desc);
	}
}

static int desc_list_init(void)
{
	int i;
	struct sk_buff *new_skb;
#if !defined(CONFIG_BFIN_MAC_USE_L1)
	/*
	 * This dma_handle is useless in Blackfin dma_alloc_coherent().
	 * The real dma handler is the return value of dma_alloc_coherent().
	 */
	dma_addr_t dma_handle;
#endif

	tx_desc = bfin_mac_alloc(&dma_handle,
				sizeof(struct net_dma_desc_tx) *
				CONFIG_BFIN_TX_DESC_NUM);
	if (tx_desc == NULL)
		goto init_error;

	rx_desc = bfin_mac_alloc(&dma_handle,
				sizeof(struct net_dma_desc_rx) *
				CONFIG_BFIN_RX_DESC_NUM);
	if (rx_desc == NULL)
		goto init_error;

	/* init tx_list */
	tx_list_head = tx_list_tail = tx_desc;

	for (i = 0; i < CONFIG_BFIN_TX_DESC_NUM; i++) {
		struct net_dma_desc_tx *t = tx_desc + i;
		struct dma_descriptor *a = &(t->desc_a);
		struct dma_descriptor *b = &(t->desc_b);

		/*
		 * disable DMA
		 * read from memory WNR = 0
		 * wordsize is 32 bits
		 * 6 half words is desc size
		 * large desc flow
		 */
		a->config = WDSIZE_32 | NDSIZE_6 | DMAFLOW_LARGE;
		a->start_addr = (unsigned long)t->packet;
		a->x_count = 0;
		a->next_dma_desc = b;

		/*
		 * enabled DMA
		 * write to memory WNR = 1
		 * wordsize is 32 bits
		 * disable interrupt
		 * 6 half words is desc size
		 * large desc flow
		 */
		b->config = DMAEN | WNR | WDSIZE_32 | NDSIZE_6 | DMAFLOW_LARGE;
		b->start_addr = (unsigned long)(&(t->status));
		b->x_count = 0;

		t->skb = NULL;
		tx_list_tail->desc_b.next_dma_desc = a;
		tx_list_tail->next = t;
		tx_list_tail = t;
	}
	tx_list_tail->next = tx_list_head;	/* tx_list is a circle */
	tx_list_tail->desc_b.next_dma_desc = &(tx_list_head->desc_a);
	current_tx_ptr = tx_list_head;

	/* init rx_list */
	rx_list_head = rx_list_tail = rx_desc;

	for (i = 0; i < CONFIG_BFIN_RX_DESC_NUM; i++) {
		struct net_dma_desc_rx *r = rx_desc + i;
		struct dma_descriptor *a = &(r->desc_a);
		struct dma_descriptor *b = &(r->desc_b);

		/* allocate a new skb for next time receive */
190
		new_skb = dev_alloc_skb(PKT_BUF_SZ + NET_IP_ALIGN);
191
		if (!new_skb) {
192
			pr_notice("init: low on mem - packet dropped\n");
193 194
			goto init_error;
		}
195
		skb_reserve(new_skb, NET_IP_ALIGN);
196 197 198 199 200
		/* Invidate the data cache of skb->data range when it is write back
		 * cache. It will prevent overwritting the new data from DMA
		 */
		blackfin_dcache_invalidate_range((unsigned long)new_skb->head,
					 (unsigned long)new_skb->end);
201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241
		r->skb = new_skb;

		/*
		 * enabled DMA
		 * write to memory WNR = 1
		 * wordsize is 32 bits
		 * disable interrupt
		 * 6 half words is desc size
		 * large desc flow
		 */
		a->config = DMAEN | WNR | WDSIZE_32 | NDSIZE_6 | DMAFLOW_LARGE;
		/* since RXDWA is enabled */
		a->start_addr = (unsigned long)new_skb->data - 2;
		a->x_count = 0;
		a->next_dma_desc = b;

		/*
		 * enabled DMA
		 * write to memory WNR = 1
		 * wordsize is 32 bits
		 * enable interrupt
		 * 6 half words is desc size
		 * large desc flow
		 */
		b->config = DMAEN | WNR | WDSIZE_32 | DI_EN |
				NDSIZE_6 | DMAFLOW_LARGE;
		b->start_addr = (unsigned long)(&(r->status));
		b->x_count = 0;

		rx_list_tail->desc_b.next_dma_desc = a;
		rx_list_tail->next = r;
		rx_list_tail = r;
	}
	rx_list_tail->next = rx_list_head;	/* rx_list is a circle */
	rx_list_tail->desc_b.next_dma_desc = &(rx_list_head->desc_a);
	current_rx_ptr = rx_list_head;

	return 0;

init_error:
	desc_list_free();
242
	pr_err("kmalloc failed\n");
243 244 245 246 247 248
	return -ENOMEM;
}


/*---PHY CONTROL AND CONFIGURATION-----------------------------------------*/

249 250 251
/*
 * MII operations
 */
252
/* Wait until the previous MDC/MDIO transaction has completed */
253
static int bfin_mdio_poll(void)
254 255 256 257 258
{
	int timeout_cnt = MAX_TIMEOUT_CNT;

	/* poll the STABUSY bit */
	while ((bfin_read_EMAC_STAADD()) & STABUSY) {
259
		udelay(1);
260
		if (timeout_cnt-- < 0) {
261
			pr_err("wait MDC/MDIO transaction to complete timeout\n");
262
			return -ETIMEDOUT;
263 264
		}
	}
265 266

	return 0;
267 268 269
}

/* Read an off-chip register in a PHY through the MDC/MDIO port */
A
Adrian Bunk 已提交
270
static int bfin_mdiobus_read(struct mii_bus *bus, int phy_addr, int regnum)
271
{
272 273 274 275 276
	int ret;

	ret = bfin_mdio_poll();
	if (ret)
		return ret;
277

278
	/* read mode */
279 280
	bfin_write_EMAC_STAADD(SET_PHYAD((u16) phy_addr) |
				SET_REGAD((u16) regnum) |
281 282
				STABUSY);

283 284 285
	ret = bfin_mdio_poll();
	if (ret)
		return ret;
286 287

	return (int) bfin_read_EMAC_STADAT();
288 289 290
}

/* Write an off-chip register in a PHY through the MDC/MDIO port */
A
Adrian Bunk 已提交
291 292
static int bfin_mdiobus_write(struct mii_bus *bus, int phy_addr, int regnum,
			      u16 value)
293
{
294 295 296 297 298
	int ret;

	ret = bfin_mdio_poll();
	if (ret)
		return ret;
299 300

	bfin_write_EMAC_STADAT((u32) value);
301 302

	/* write mode */
303 304
	bfin_write_EMAC_STAADD(SET_PHYAD((u16) phy_addr) |
				SET_REGAD((u16) regnum) |
305 306 307
				STAOP |
				STABUSY);

308
	return bfin_mdio_poll();
309 310
}

A
Adrian Bunk 已提交
311
static int bfin_mdiobus_reset(struct mii_bus *bus)
312
{
313
	return 0;
314 315
}

B
Bryan Wu 已提交
316
static void bfin_mac_adjust_link(struct net_device *dev)
317
{
B
Bryan Wu 已提交
318
	struct bfin_mac_local *lp = netdev_priv(dev);
319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338
	struct phy_device *phydev = lp->phydev;
	unsigned long flags;
	int new_state = 0;

	spin_lock_irqsave(&lp->lock, flags);
	if (phydev->link) {
		/* Now we make sure that we can be in full duplex mode.
		 * If not, we operate in half-duplex mode. */
		if (phydev->duplex != lp->old_duplex) {
			u32 opmode = bfin_read_EMAC_OPMODE();
			new_state = 1;

			if (phydev->duplex)
				opmode |= FDMODE;
			else
				opmode &= ~(FDMODE);

			bfin_write_EMAC_OPMODE(opmode);
			lp->old_duplex = phydev->duplex;
		}
339

340
		if (phydev->speed != lp->old_speed) {
341 342 343 344 345 346 347 348 349 350
			if (phydev->interface == PHY_INTERFACE_MODE_RMII) {
				u32 opmode = bfin_read_EMAC_OPMODE();
				switch (phydev->speed) {
				case 10:
					opmode |= RMII_10;
					break;
				case 100:
					opmode &= ~RMII_10;
					break;
				default:
351 352 353
					netdev_warn(dev,
						"Ack! Speed (%d) is not 10/100!\n",
						phydev->speed);
354 355 356
					break;
				}
				bfin_write_EMAC_OPMODE(opmode);
357
			}
358

359 360 361
			new_state = 1;
			lp->old_speed = phydev->speed;
		}
362

363 364 365 366 367 368 369 370 371
		if (!lp->old_link) {
			new_state = 1;
			lp->old_link = 1;
		}
	} else if (lp->old_link) {
		new_state = 1;
		lp->old_link = 0;
		lp->old_speed = 0;
		lp->old_duplex = -1;
372 373
	}

374 375 376 377
	if (new_state) {
		u32 opmode = bfin_read_EMAC_OPMODE();
		phy_print_status(phydev);
		pr_debug("EMAC_OPMODE = 0x%08x\n", opmode);
378
	}
379 380

	spin_unlock_irqrestore(&lp->lock, flags);
381 382
}

383 384 385
/* MDC  = 2.5 MHz */
#define MDC_CLK 2500000

386
static int mii_probe(struct net_device *dev, int phy_mode)
387
{
B
Bryan Wu 已提交
388
	struct bfin_mac_local *lp = netdev_priv(dev);
389 390 391
	struct phy_device *phydev = NULL;
	unsigned short sysctl;
	int i;
392
	u32 sclk, mdc_div;
393

394
	/* Enable PHY output early */
395 396
	if (!(bfin_read_VR_CTL() & CLKBUFOE))
		bfin_write_VR_CTL(bfin_read_VR_CTL() | CLKBUFOE);
397

398 399 400
	sclk = get_sclk();
	mdc_div = ((sclk / MDC_CLK) / 2) - 1;

401
	sysctl = bfin_read_EMAC_SYSCTL();
402
	sysctl = (sysctl & ~MDCDIV) | SET_MDCDIV(mdc_div);
403 404
	bfin_write_EMAC_SYSCTL(sysctl);

405 406
	/* search for connected PHY device */
	for (i = 0; i < PHY_MAX_ADDR; ++i) {
407
		struct phy_device *const tmp_phydev = lp->mii_bus->phy_map[i];
408

409 410
		if (!tmp_phydev)
			continue; /* no PHY here... */
411

412 413 414 415 416 417
		phydev = tmp_phydev;
		break; /* found it */
	}

	/* now we are supposed to have a proper phydev, to attach to... */
	if (!phydev) {
418
		netdev_err(dev, "no phy device found\n");
419
		return -ENODEV;
420 421
	}

422 423
	if (phy_mode != PHY_INTERFACE_MODE_RMII &&
		phy_mode != PHY_INTERFACE_MODE_MII) {
424
		netdev_err(dev, "invalid phy interface mode\n");
425 426 427
		return -EINVAL;
	}

428
	phydev = phy_connect(dev, dev_name(&phydev->dev), &bfin_mac_adjust_link,
429
			0, phy_mode);
430

431
	if (IS_ERR(phydev)) {
432
		netdev_err(dev, "could not attach PHY\n");
433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452
		return PTR_ERR(phydev);
	}

	/* mask with MAC supported features */
	phydev->supported &= (SUPPORTED_10baseT_Half
			      | SUPPORTED_10baseT_Full
			      | SUPPORTED_100baseT_Half
			      | SUPPORTED_100baseT_Full
			      | SUPPORTED_Autoneg
			      | SUPPORTED_Pause | SUPPORTED_Asym_Pause
			      | SUPPORTED_MII
			      | SUPPORTED_TP);

	phydev->advertising = phydev->supported;

	lp->old_link = 0;
	lp->old_speed = 0;
	lp->old_duplex = -1;
	lp->phydev = phydev;

453 454 455 456
	pr_info("attached PHY driver [%s] "
	        "(mii_bus:phy_addr=%s, irq=%d, mdc_clk=%dHz(mdc_div=%d)@sclk=%dMHz)\n",
	        phydev->drv->name, dev_name(&phydev->dev), phydev->irq,
	        MDC_CLK, mdc_div, sclk/1000000);
457 458 459 460

	return 0;
}

461 462 463 464
/*
 * Ethtool support
 */

465 466 467 468 469 470 471 472
/*
 * interrupt routine for magic packet wakeup
 */
static irqreturn_t bfin_mac_wake_interrupt(int irq, void *dev_id)
{
	return IRQ_HANDLED;
}

473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500
static int
bfin_mac_ethtool_getsettings(struct net_device *dev, struct ethtool_cmd *cmd)
{
	struct bfin_mac_local *lp = netdev_priv(dev);

	if (lp->phydev)
		return phy_ethtool_gset(lp->phydev, cmd);

	return -EINVAL;
}

static int
bfin_mac_ethtool_setsettings(struct net_device *dev, struct ethtool_cmd *cmd)
{
	struct bfin_mac_local *lp = netdev_priv(dev);

	if (!capable(CAP_NET_ADMIN))
		return -EPERM;

	if (lp->phydev)
		return phy_ethtool_sset(lp->phydev, cmd);

	return -EINVAL;
}

static void bfin_mac_ethtool_getdrvinfo(struct net_device *dev,
					struct ethtool_drvinfo *info)
{
501
	strcpy(info->driver, KBUILD_MODNAME);
502 503
	strcpy(info->version, DRV_VERSION);
	strcpy(info->fw_version, "N/A");
504
	strcpy(info->bus_info, dev_name(&dev->dev));
505 506
}

507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550
static void bfin_mac_ethtool_getwol(struct net_device *dev,
	struct ethtool_wolinfo *wolinfo)
{
	struct bfin_mac_local *lp = netdev_priv(dev);

	wolinfo->supported = WAKE_MAGIC;
	wolinfo->wolopts = lp->wol;
}

static int bfin_mac_ethtool_setwol(struct net_device *dev,
	struct ethtool_wolinfo *wolinfo)
{
	struct bfin_mac_local *lp = netdev_priv(dev);
	int rc;

	if (wolinfo->wolopts & (WAKE_MAGICSECURE |
				WAKE_UCAST |
				WAKE_MCAST |
				WAKE_BCAST |
				WAKE_ARP))
		return -EOPNOTSUPP;

	lp->wol = wolinfo->wolopts;

	if (lp->wol && !lp->irq_wake_requested) {
		/* register wake irq handler */
		rc = request_irq(IRQ_MAC_WAKEDET, bfin_mac_wake_interrupt,
				 IRQF_DISABLED, "EMAC_WAKE", dev);
		if (rc)
			return rc;
		lp->irq_wake_requested = true;
	}

	if (!lp->wol && lp->irq_wake_requested) {
		free_irq(IRQ_MAC_WAKEDET, dev);
		lp->irq_wake_requested = false;
	}

	/* Make sure the PHY driver doesn't suspend */
	device_init_wakeup(&dev->dev, lp->wol);

	return 0;
}

551
static const struct ethtool_ops bfin_mac_ethtool_ops = {
552 553 554 555
	.get_settings = bfin_mac_ethtool_getsettings,
	.set_settings = bfin_mac_ethtool_setsettings,
	.get_link = ethtool_op_get_link,
	.get_drvinfo = bfin_mac_ethtool_getdrvinfo,
556 557
	.get_wol = bfin_mac_ethtool_getwol,
	.set_wol = bfin_mac_ethtool_setwol,
558 559
};

560
/**************************************************************************/
561
static void setup_system_regs(struct net_device *dev)
562
{
563 564
	struct bfin_mac_local *lp = netdev_priv(dev);
	int i;
565 566 567 568 569 570 571
	unsigned short sysctl;

	/*
	 * Odd word alignment for Receive Frame DMA word
	 * Configure checksum support and rcve frame word alignment
	 */
	sysctl = bfin_read_EMAC_SYSCTL();
572 573 574 575 576 577 578 579 580
	/*
	 * check if interrupt is requested for any PHY,
	 * enable PHY interrupt only if needed
	 */
	for (i = 0; i < PHY_MAX_ADDR; ++i)
		if (lp->mii_bus->irq[i] != PHY_POLL)
			break;
	if (i < PHY_MAX_ADDR)
		sysctl |= PHYIE;
581
	sysctl |= RXDWA;
582
#if defined(BFIN_MAC_CSUM_OFFLOAD)
583
	sysctl |= RXCKS;
584
#else
585
	sysctl &= ~RXCKS;
586 587
#endif
	bfin_write_EMAC_SYSCTL(sysctl);
588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603

	bfin_write_EMAC_MMC_CTL(RSTC | CROLL);

	/* Initialize the TX DMA channel registers */
	bfin_write_DMA2_X_COUNT(0);
	bfin_write_DMA2_X_MODIFY(4);
	bfin_write_DMA2_Y_COUNT(0);
	bfin_write_DMA2_Y_MODIFY(0);

	/* Initialize the RX DMA channel registers */
	bfin_write_DMA1_X_COUNT(0);
	bfin_write_DMA1_X_MODIFY(4);
	bfin_write_DMA1_Y_COUNT(0);
	bfin_write_DMA1_Y_MODIFY(0);
}

604
static void setup_mac_addr(u8 *mac_addr)
605 606 607 608 609 610 611 612 613
{
	u32 addr_low = le32_to_cpu(*(__le32 *) & mac_addr[0]);
	u16 addr_hi = le16_to_cpu(*(__le16 *) & mac_addr[4]);

	/* this depends on a little-endian machine */
	bfin_write_EMAC_ADDRLO(addr_low);
	bfin_write_EMAC_ADDRHI(addr_hi);
}

B
Bryan Wu 已提交
614
static int bfin_mac_set_mac_address(struct net_device *dev, void *p)
615 616 617 618 619 620 621 622 623
{
	struct sockaddr *addr = p;
	if (netif_running(dev))
		return -EBUSY;
	memcpy(dev->dev_addr, addr->sa_data, dev->addr_len);
	setup_mac_addr(dev->dev_addr);
	return 0;
}

624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808
#ifdef CONFIG_BFIN_MAC_USE_HWSTAMP
#define bfin_mac_hwtstamp_is_none(cfg) ((cfg) == HWTSTAMP_FILTER_NONE)

static int bfin_mac_hwtstamp_ioctl(struct net_device *netdev,
		struct ifreq *ifr, int cmd)
{
	struct hwtstamp_config config;
	struct bfin_mac_local *lp = netdev_priv(netdev);
	u16 ptpctl;
	u32 ptpfv1, ptpfv2, ptpfv3, ptpfoff;

	if (copy_from_user(&config, ifr->ifr_data, sizeof(config)))
		return -EFAULT;

	pr_debug("%s config flag:0x%x, tx_type:0x%x, rx_filter:0x%x\n",
			__func__, config.flags, config.tx_type, config.rx_filter);

	/* reserved for future extensions */
	if (config.flags)
		return -EINVAL;

	if ((config.tx_type != HWTSTAMP_TX_OFF) &&
			(config.tx_type != HWTSTAMP_TX_ON))
		return -ERANGE;

	ptpctl = bfin_read_EMAC_PTP_CTL();

	switch (config.rx_filter) {
	case HWTSTAMP_FILTER_NONE:
		/*
		 * Dont allow any timestamping
		 */
		ptpfv3 = 0xFFFFFFFF;
		bfin_write_EMAC_PTP_FV3(ptpfv3);
		break;
	case HWTSTAMP_FILTER_PTP_V1_L4_EVENT:
	case HWTSTAMP_FILTER_PTP_V1_L4_SYNC:
	case HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ:
		/*
		 * Clear the five comparison mask bits (bits[12:8]) in EMAC_PTP_CTL)
		 * to enable all the field matches.
		 */
		ptpctl &= ~0x1F00;
		bfin_write_EMAC_PTP_CTL(ptpctl);
		/*
		 * Keep the default values of the EMAC_PTP_FOFF register.
		 */
		ptpfoff = 0x4A24170C;
		bfin_write_EMAC_PTP_FOFF(ptpfoff);
		/*
		 * Keep the default values of the EMAC_PTP_FV1 and EMAC_PTP_FV2
		 * registers.
		 */
		ptpfv1 = 0x11040800;
		bfin_write_EMAC_PTP_FV1(ptpfv1);
		ptpfv2 = 0x0140013F;
		bfin_write_EMAC_PTP_FV2(ptpfv2);
		/*
		 * The default value (0xFFFC) allows the timestamping of both
		 * received Sync messages and Delay_Req messages.
		 */
		ptpfv3 = 0xFFFFFFFC;
		bfin_write_EMAC_PTP_FV3(ptpfv3);

		config.rx_filter = HWTSTAMP_FILTER_PTP_V1_L4_EVENT;
		break;
	case HWTSTAMP_FILTER_PTP_V2_L4_EVENT:
	case HWTSTAMP_FILTER_PTP_V2_L4_SYNC:
	case HWTSTAMP_FILTER_PTP_V2_DELAY_REQ:
		/* Clear all five comparison mask bits (bits[12:8]) in the
		 * EMAC_PTP_CTL register to enable all the field matches.
		 */
		ptpctl &= ~0x1F00;
		bfin_write_EMAC_PTP_CTL(ptpctl);
		/*
		 * Keep the default values of the EMAC_PTP_FOFF register, except set
		 * the PTPCOF field to 0x2A.
		 */
		ptpfoff = 0x2A24170C;
		bfin_write_EMAC_PTP_FOFF(ptpfoff);
		/*
		 * Keep the default values of the EMAC_PTP_FV1 and EMAC_PTP_FV2
		 * registers.
		 */
		ptpfv1 = 0x11040800;
		bfin_write_EMAC_PTP_FV1(ptpfv1);
		ptpfv2 = 0x0140013F;
		bfin_write_EMAC_PTP_FV2(ptpfv2);
		/*
		 * To allow the timestamping of Pdelay_Req and Pdelay_Resp, set
		 * the value to 0xFFF0.
		 */
		ptpfv3 = 0xFFFFFFF0;
		bfin_write_EMAC_PTP_FV3(ptpfv3);

		config.rx_filter = HWTSTAMP_FILTER_PTP_V2_L4_EVENT;
		break;
	case HWTSTAMP_FILTER_PTP_V2_L2_EVENT:
	case HWTSTAMP_FILTER_PTP_V2_L2_SYNC:
	case HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ:
		/*
		 * Clear bits 8 and 12 of the EMAC_PTP_CTL register to enable only the
		 * EFTM and PTPCM field comparison.
		 */
		ptpctl &= ~0x1100;
		bfin_write_EMAC_PTP_CTL(ptpctl);
		/*
		 * Keep the default values of all the fields of the EMAC_PTP_FOFF
		 * register, except set the PTPCOF field to 0x0E.
		 */
		ptpfoff = 0x0E24170C;
		bfin_write_EMAC_PTP_FOFF(ptpfoff);
		/*
		 * Program bits [15:0] of the EMAC_PTP_FV1 register to 0x88F7, which
		 * corresponds to PTP messages on the MAC layer.
		 */
		ptpfv1 = 0x110488F7;
		bfin_write_EMAC_PTP_FV1(ptpfv1);
		ptpfv2 = 0x0140013F;
		bfin_write_EMAC_PTP_FV2(ptpfv2);
		/*
		 * To allow the timestamping of Pdelay_Req and Pdelay_Resp
		 * messages, set the value to 0xFFF0.
		 */
		ptpfv3 = 0xFFFFFFF0;
		bfin_write_EMAC_PTP_FV3(ptpfv3);

		config.rx_filter = HWTSTAMP_FILTER_PTP_V2_L2_EVENT;
		break;
	default:
		return -ERANGE;
	}

	if (config.tx_type == HWTSTAMP_TX_OFF &&
	    bfin_mac_hwtstamp_is_none(config.rx_filter)) {
		ptpctl &= ~PTP_EN;
		bfin_write_EMAC_PTP_CTL(ptpctl);

		SSYNC();
	} else {
		ptpctl |= PTP_EN;
		bfin_write_EMAC_PTP_CTL(ptpctl);

		/*
		 * clear any existing timestamp
		 */
		bfin_read_EMAC_PTP_RXSNAPLO();
		bfin_read_EMAC_PTP_RXSNAPHI();

		bfin_read_EMAC_PTP_TXSNAPLO();
		bfin_read_EMAC_PTP_TXSNAPHI();

		/*
		 * Set registers so that rollover occurs soon to test this.
		 */
		bfin_write_EMAC_PTP_TIMELO(0x00000000);
		bfin_write_EMAC_PTP_TIMEHI(0xFF800000);

		SSYNC();

		lp->compare.last_update = 0;
		timecounter_init(&lp->clock,
				&lp->cycles,
				ktime_to_ns(ktime_get_real()));
		timecompare_update(&lp->compare, 0);
	}

	lp->stamp_cfg = config;
	return copy_to_user(ifr->ifr_data, &config, sizeof(config)) ?
		-EFAULT : 0;
}

static void bfin_dump_hwtamp(char *s, ktime_t *hw, ktime_t *ts, struct timecompare *cmp)
{
	ktime_t sys = ktime_get_real();

	pr_debug("%s %s hardware:%d,%d transform system:%d,%d system:%d,%d, cmp:%lld, %lld\n",
			__func__, s, hw->tv.sec, hw->tv.nsec, ts->tv.sec, ts->tv.nsec, sys.tv.sec,
			sys.tv.nsec, cmp->offset, cmp->skew);
}

static void bfin_tx_hwtstamp(struct net_device *netdev, struct sk_buff *skb)
{
	struct bfin_mac_local *lp = netdev_priv(netdev);

809
	if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) {
810 811 812 813 814
		int timeout_cnt = MAX_TIMEOUT_CNT;

		/* When doing time stamping, keep the connection to the socket
		 * a while longer
		 */
815
		skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
816 817 818 819 820 821 822 823 824 825

		/*
		 * The timestamping is done at the EMAC module's MII/RMII interface
		 * when the module sees the Start of Frame of an event message packet. This
		 * interface is the closest possible place to the physical Ethernet transmission
		 * medium, providing the best timing accuracy.
		 */
		while ((!(bfin_read_EMAC_PTP_ISTAT() & TXTL)) && (--timeout_cnt))
			udelay(1);
		if (timeout_cnt == 0)
826
			netdev_err(netdev, "timestamp the TX packet failed\n");
827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924
		else {
			struct skb_shared_hwtstamps shhwtstamps;
			u64 ns;
			u64 regval;

			regval = bfin_read_EMAC_PTP_TXSNAPLO();
			regval |= (u64)bfin_read_EMAC_PTP_TXSNAPHI() << 32;
			memset(&shhwtstamps, 0, sizeof(shhwtstamps));
			ns = timecounter_cyc2time(&lp->clock,
					regval);
			timecompare_update(&lp->compare, ns);
			shhwtstamps.hwtstamp = ns_to_ktime(ns);
			shhwtstamps.syststamp =
				timecompare_transform(&lp->compare, ns);
			skb_tstamp_tx(skb, &shhwtstamps);

			bfin_dump_hwtamp("TX", &shhwtstamps.hwtstamp, &shhwtstamps.syststamp, &lp->compare);
		}
	}
}

static void bfin_rx_hwtstamp(struct net_device *netdev, struct sk_buff *skb)
{
	struct bfin_mac_local *lp = netdev_priv(netdev);
	u32 valid;
	u64 regval, ns;
	struct skb_shared_hwtstamps *shhwtstamps;

	if (bfin_mac_hwtstamp_is_none(lp->stamp_cfg.rx_filter))
		return;

	valid = bfin_read_EMAC_PTP_ISTAT() & RXEL;
	if (!valid)
		return;

	shhwtstamps = skb_hwtstamps(skb);

	regval = bfin_read_EMAC_PTP_RXSNAPLO();
	regval |= (u64)bfin_read_EMAC_PTP_RXSNAPHI() << 32;
	ns = timecounter_cyc2time(&lp->clock, regval);
	timecompare_update(&lp->compare, ns);
	memset(shhwtstamps, 0, sizeof(*shhwtstamps));
	shhwtstamps->hwtstamp = ns_to_ktime(ns);
	shhwtstamps->syststamp = timecompare_transform(&lp->compare, ns);

	bfin_dump_hwtamp("RX", &shhwtstamps->hwtstamp, &shhwtstamps->syststamp, &lp->compare);
}

/*
 * bfin_read_clock - read raw cycle counter (to be used by time counter)
 */
static cycle_t bfin_read_clock(const struct cyclecounter *tc)
{
	u64 stamp;

	stamp =  bfin_read_EMAC_PTP_TIMELO();
	stamp |= (u64)bfin_read_EMAC_PTP_TIMEHI() << 32ULL;

	return stamp;
}

#define PTP_CLK 25000000

static void bfin_mac_hwtstamp_init(struct net_device *netdev)
{
	struct bfin_mac_local *lp = netdev_priv(netdev);
	u64 append;

	/* Initialize hardware timer */
	append = PTP_CLK * (1ULL << 32);
	do_div(append, get_sclk());
	bfin_write_EMAC_PTP_ADDEND((u32)append);

	memset(&lp->cycles, 0, sizeof(lp->cycles));
	lp->cycles.read = bfin_read_clock;
	lp->cycles.mask = CLOCKSOURCE_MASK(64);
	lp->cycles.mult = 1000000000 / PTP_CLK;
	lp->cycles.shift = 0;

	/* Synchronize our NIC clock against system wall clock */
	memset(&lp->compare, 0, sizeof(lp->compare));
	lp->compare.source = &lp->clock;
	lp->compare.target = ktime_get_real;
	lp->compare.num_samples = 10;

	/* Initialize hwstamp config */
	lp->stamp_cfg.rx_filter = HWTSTAMP_FILTER_NONE;
	lp->stamp_cfg.tx_type = HWTSTAMP_TX_OFF;
}

#else
# define bfin_mac_hwtstamp_is_none(cfg) 0
# define bfin_mac_hwtstamp_init(dev)
# define bfin_mac_hwtstamp_ioctl(dev, ifr, cmd) (-EOPNOTSUPP)
# define bfin_rx_hwtstamp(dev, skb)
# define bfin_tx_hwtstamp(dev, skb)
#endif

925 926 927 928 929 930 931 932 933 934 935 936 937 938 939
static inline void _tx_reclaim_skb(void)
{
	do {
		tx_list_head->desc_a.config &= ~DMAEN;
		tx_list_head->status.status_word = 0;
		if (tx_list_head->skb) {
			dev_kfree_skb(tx_list_head->skb);
			tx_list_head->skb = NULL;
		}
		tx_list_head = tx_list_head->next;

	} while (tx_list_head->status.status_word != 0);
}

static void tx_reclaim_skb(struct bfin_mac_local *lp)
940 941 942
{
	int timeout_cnt = MAX_TIMEOUT_CNT;

943 944
	if (tx_list_head->status.status_word != 0)
		_tx_reclaim_skb();
945

946
	if (current_tx_ptr->next == tx_list_head) {
947
		while (tx_list_head->status.status_word == 0) {
948
			/* slow down polling to avoid too many queue stop. */
949
			udelay(10);
950 951 952 953
			/* reclaim skb if DMA is not running. */
			if (!(bfin_read_DMA2_IRQ_STATUS() & DMA_RUN))
				break;
			if (timeout_cnt-- < 0)
954 955
				break;
		}
956 957 958 959 960

		if (timeout_cnt >= 0)
			_tx_reclaim_skb();
		else
			netif_stop_queue(lp->ndev);
961 962
	}

963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978
	if (current_tx_ptr->next != tx_list_head &&
		netif_queue_stopped(lp->ndev))
		netif_wake_queue(lp->ndev);

	if (tx_list_head != current_tx_ptr) {
		/* shorten the timer interval if tx queue is stopped */
		if (netif_queue_stopped(lp->ndev))
			lp->tx_reclaim_timer.expires =
				jiffies + (TX_RECLAIM_JIFFIES >> 4);
		else
			lp->tx_reclaim_timer.expires =
				jiffies + TX_RECLAIM_JIFFIES;

		mod_timer(&lp->tx_reclaim_timer,
			lp->tx_reclaim_timer.expires);
	}
979 980

	return;
981
}
982

983 984 985
static void tx_reclaim_skb_timeout(unsigned long lp)
{
	tx_reclaim_skb((struct bfin_mac_local *)lp);
986 987
}

B
Bryan Wu 已提交
988
static int bfin_mac_hard_start_xmit(struct sk_buff *skb,
989 990
				struct net_device *dev)
{
991
	struct bfin_mac_local *lp = netdev_priv(dev);
992
	u16 *data;
993
	u32 data_align = (unsigned long)(skb->data) & 0x3;
994

995 996
	current_tx_ptr->skb = skb;

997 998 999
	if (data_align == 0x2) {
		/* move skb->data to current_tx_ptr payload */
		data = (u16 *)(skb->data) - 1;
1000 1001 1002 1003 1004 1005 1006
		*data = (u16)(skb->len);
		/*
		 * When transmitting an Ethernet packet, the PTP_TSYNC module requires
		 * a DMA_Length_Word field associated with the packet. The lower 12 bits
		 * of this field are the length of the packet payload in bytes and the higher
		 * 4 bits are the timestamping enable field.
		 */
1007
		if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)
1008 1009
			*data |= 0x1000;

1010 1011 1012 1013
		current_tx_ptr->desc_a.start_addr = (u32)data;
		/* this is important! */
		blackfin_dcache_flush_range((u32)data,
				(u32)((u8 *)data + skb->len + 4));
1014
	} else {
1015
		*((u16 *)(current_tx_ptr->packet)) = (u16)(skb->len);
1016
		/* enable timestamping for the sent packet */
1017
		if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)
1018
			*((u16 *)(current_tx_ptr->packet)) |= 0x1000;
1019 1020 1021 1022 1023 1024 1025
		memcpy((u8 *)(current_tx_ptr->packet + 2), skb->data,
			skb->len);
		current_tx_ptr->desc_a.start_addr =
			(u32)current_tx_ptr->packet;
		blackfin_dcache_flush_range(
			(u32)current_tx_ptr->packet,
			(u32)(current_tx_ptr->packet + skb->len + 2));
1026 1027
	}

1028 1029 1030 1031 1032 1033
	/* make sure the internal data buffers in the core are drained
	 * so that the DMA descriptors are completely written when the
	 * DMA engine goes to fetch them below
	 */
	SSYNC();

1034 1035 1036
	/* always clear status buffer before start tx dma */
	current_tx_ptr->status.status_word = 0;

1037 1038 1039 1040
	/* enable this packet's dma */
	current_tx_ptr->desc_a.config |= DMAEN;

	/* tx dma is running, just return */
1041
	if (bfin_read_DMA2_IRQ_STATUS() & DMA_RUN)
1042 1043 1044 1045 1046 1047 1048 1049 1050 1051
		goto out;

	/* tx dma is not running */
	bfin_write_DMA2_NEXT_DESC_PTR(&(current_tx_ptr->desc_a));
	/* dma enabled, read from memory, size is 6 */
	bfin_write_DMA2_CONFIG(current_tx_ptr->desc_a.config);
	/* Turn on the EMAC tx */
	bfin_write_EMAC_OPMODE(bfin_read_EMAC_OPMODE() | TE);

out:
1052 1053
	bfin_tx_hwtstamp(dev, skb);

1054
	current_tx_ptr = current_tx_ptr->next;
1055 1056
	dev->stats.tx_packets++;
	dev->stats.tx_bytes += (skb->len);
1057 1058 1059

	tx_reclaim_skb(lp);

1060
	return NETDEV_TX_OK;
1061 1062
}

1063
#define IP_HEADER_OFF  0
1064 1065 1066
#define RX_ERROR_MASK (RX_LONG | RX_ALIGN | RX_CRC | RX_LEN | \
	RX_FRAG | RX_ADDR | RX_DMAO | RX_PHY | RX_LATE | RX_RANGE)

B
Bryan Wu 已提交
1067
static void bfin_mac_rx(struct net_device *dev)
1068 1069 1070
{
	struct sk_buff *skb, *new_skb;
	unsigned short len;
1071
	struct bfin_mac_local *lp __maybe_unused = netdev_priv(dev);
1072 1073 1074 1075
#if defined(BFIN_MAC_CSUM_OFFLOAD)
	unsigned int i;
	unsigned char fcs[ETH_FCS_LEN + 1];
#endif
1076

1077 1078 1079 1080
	/* check if frame status word reports an error condition
	 * we which case we simply drop the packet
	 */
	if (current_rx_ptr->status.status_word & RX_ERROR_MASK) {
1081
		netdev_notice(dev, "rx: receive error - packet dropped\n");
1082 1083 1084 1085
		dev->stats.rx_dropped++;
		goto out;
	}

1086 1087
	/* allocate a new skb for next time receive */
	skb = current_rx_ptr->skb;
1088

1089
	new_skb = dev_alloc_skb(PKT_BUF_SZ + NET_IP_ALIGN);
1090
	if (!new_skb) {
1091
		netdev_notice(dev, "rx: low on mem - packet dropped\n");
1092
		dev->stats.rx_dropped++;
1093 1094 1095
		goto out;
	}
	/* reserve 2 bytes for RXDWA padding */
1096
	skb_reserve(new_skb, NET_IP_ALIGN);
1097 1098 1099 1100 1101 1102
	/* Invidate the data cache of skb->data range when it is write back
	 * cache. It will prevent overwritting the new data from DMA
	 */
	blackfin_dcache_invalidate_range((unsigned long)new_skb->head,
					 (unsigned long)new_skb->end);

1103 1104 1105
	current_rx_ptr->skb = new_skb;
	current_rx_ptr->desc_a.start_addr = (unsigned long)new_skb->data - 2;

1106
	len = (unsigned short)((current_rx_ptr->status.status_word) & RX_FRLEN);
1107 1108
	/* Deduce Ethernet FCS length from Ethernet payload length */
	len -= ETH_FCS_LEN;
1109 1110 1111
	skb_put(skb, len);

	skb->protocol = eth_type_trans(skb, dev);
1112 1113 1114

	bfin_rx_hwtstamp(dev, skb);

1115
#if defined(BFIN_MAC_CSUM_OFFLOAD)
1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141
	/* Checksum offloading only works for IPv4 packets with the standard IP header
	 * length of 20 bytes, because the blackfin MAC checksum calculation is
	 * based on that assumption. We must NOT use the calculated checksum if our
	 * IP version or header break that assumption.
	 */
	if (skb->data[IP_HEADER_OFF] == 0x45) {
		skb->csum = current_rx_ptr->status.ip_payload_csum;
		/*
		 * Deduce Ethernet FCS from hardware generated IP payload checksum.
		 * IP checksum is based on 16-bit one's complement algorithm.
		 * To deduce a value from checksum is equal to add its inversion.
		 * If the IP payload len is odd, the inversed FCS should also
		 * begin from odd address and leave first byte zero.
		 */
		if (skb->len % 2) {
			fcs[0] = 0;
			for (i = 0; i < ETH_FCS_LEN; i++)
				fcs[i + 1] = ~skb->data[skb->len + i];
			skb->csum = csum_partial(fcs, ETH_FCS_LEN + 1, skb->csum);
		} else {
			for (i = 0; i < ETH_FCS_LEN; i++)
				fcs[i] = ~skb->data[skb->len + i];
			skb->csum = csum_partial(fcs, ETH_FCS_LEN, skb->csum);
		}
		skb->ip_summed = CHECKSUM_COMPLETE;
	}
1142 1143 1144
#endif

	netif_rx(skb);
1145 1146
	dev->stats.rx_packets++;
	dev->stats.rx_bytes += len;
1147
out:
1148 1149 1150 1151 1152
	current_rx_ptr->status.status_word = 0x00000000;
	current_rx_ptr = current_rx_ptr->next;
}

/* interrupt routine to handle rx and error signal */
B
Bryan Wu 已提交
1153
static irqreturn_t bfin_mac_interrupt(int irq, void *dev_id)
1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172
{
	struct net_device *dev = dev_id;
	int number = 0;

get_one_packet:
	if (current_rx_ptr->status.status_word == 0) {
		/* no more new packet received */
		if (number == 0) {
			if (current_rx_ptr->next->status.status_word != 0) {
				current_rx_ptr = current_rx_ptr->next;
				goto real_rx;
			}
		}
		bfin_write_DMA1_IRQ_STATUS(bfin_read_DMA1_IRQ_STATUS() |
					   DMA_DONE | DMA_ERR);
		return IRQ_HANDLED;
	}

real_rx:
B
Bryan Wu 已提交
1173
	bfin_mac_rx(dev);
1174 1175 1176 1177 1178
	number++;
	goto get_one_packet;
}

#ifdef CONFIG_NET_POLL_CONTROLLER
B
Bryan Wu 已提交
1179
static void bfin_mac_poll(struct net_device *dev)
1180
{
1181 1182
	struct bfin_mac_local *lp = netdev_priv(dev);

1183
	disable_irq(IRQ_MAC_RX);
B
Bryan Wu 已提交
1184
	bfin_mac_interrupt(IRQ_MAC_RX, dev);
1185
	tx_reclaim_skb(lp);
1186 1187 1188 1189
	enable_irq(IRQ_MAC_RX);
}
#endif				/* CONFIG_NET_POLL_CONTROLLER */

B
Bryan Wu 已提交
1190
static void bfin_mac_disable(void)
1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203
{
	unsigned int opmode;

	opmode = bfin_read_EMAC_OPMODE();
	opmode &= (~RE);
	opmode &= (~TE);
	/* Turn off the EMAC */
	bfin_write_EMAC_OPMODE(opmode);
}

/*
 * Enable Interrupts, Receive, and Transmit
 */
1204
static int bfin_mac_enable(struct phy_device *phydev)
1205
{
1206
	int ret;
1207 1208
	u32 opmode;

1209
	pr_debug("%s\n", __func__);
1210 1211 1212 1213 1214 1215

	/* Set RX DMA */
	bfin_write_DMA1_NEXT_DESC_PTR(&(rx_list_head->desc_a));
	bfin_write_DMA1_CONFIG(rx_list_head->desc_a.config);

	/* Wait MII done */
1216 1217 1218
	ret = bfin_mdio_poll();
	if (ret)
		return ret;
1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233

	/* We enable only RX here */
	/* ASTP   : Enable Automatic Pad Stripping
	   PR     : Promiscuous Mode for test
	   PSF    : Receive frames with total length less than 64 bytes.
	   FDMODE : Full Duplex Mode
	   LB     : Internal Loopback for test
	   RE     : Receiver Enable */
	opmode = bfin_read_EMAC_OPMODE();
	if (opmode & FDMODE)
		opmode |= PSF;
	else
		opmode |= DRO | DC | PSF;
	opmode |= RE;

1234 1235
	if (phydev->interface == PHY_INTERFACE_MODE_RMII) {
		opmode |= RMII; /* For Now only 100MBit are supported */
1236
#if (defined(CONFIG_BF537) || defined(CONFIG_BF536)) && CONFIG_BF_REV_0_2
1237
		opmode |= TE;
1238
#endif
1239 1240
	}

1241 1242
	/* Turn on the EMAC rx */
	bfin_write_EMAC_OPMODE(opmode);
1243 1244

	return 0;
1245 1246 1247
}

/* Our watchdog timed out. Called by the networking layer */
B
Bryan Wu 已提交
1248
static void bfin_mac_timeout(struct net_device *dev)
1249
{
1250 1251
	struct bfin_mac_local *lp = netdev_priv(dev);

1252
	pr_debug("%s: %s\n", dev->name, __func__);
1253

B
Bryan Wu 已提交
1254
	bfin_mac_disable();
1255

1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270
	del_timer(&lp->tx_reclaim_timer);

	/* reset tx queue and free skb */
	while (tx_list_head != current_tx_ptr) {
		tx_list_head->desc_a.config &= ~DMAEN;
		tx_list_head->status.status_word = 0;
		if (tx_list_head->skb) {
			dev_kfree_skb(tx_list_head->skb);
			tx_list_head->skb = NULL;
		}
		tx_list_head = tx_list_head->next;
	}

	if (netif_queue_stopped(lp->ndev))
		netif_wake_queue(lp->ndev);
1271

1272
	bfin_mac_enable(lp->phydev);
1273 1274

	/* We can accept TX packets again */
E
Eric Dumazet 已提交
1275
	dev->trans_start = jiffies; /* prevent tx timeout */
1276 1277 1278
	netif_wake_queue(dev);
}

B
Bryan Wu 已提交
1279
static void bfin_mac_multicast_hash(struct net_device *dev)
1280 1281
{
	u32 emac_hashhi, emac_hashlo;
1282
	struct netdev_hw_addr *ha;
1283 1284 1285 1286 1287
	char *addrs;
	u32 crc;

	emac_hashhi = emac_hashlo = 0;

1288 1289
	netdev_for_each_mc_addr(ha, dev) {
		addrs = ha->addr;
1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307

		/* skip non-multicast addresses */
		if (!(*addrs & 1))
			continue;

		crc = ether_crc(ETH_ALEN, addrs);
		crc >>= 26;

		if (crc & 0x20)
			emac_hashhi |= 1 << (crc & 0x1f);
		else
			emac_hashlo |= 1 << (crc & 0x1f);
	}

	bfin_write_EMAC_HASHHI(emac_hashhi);
	bfin_write_EMAC_HASHLO(emac_hashlo);
}

1308 1309 1310 1311 1312 1313
/*
 * This routine will, depending on the values passed to it,
 * either make it accept multicast packets, go into
 * promiscuous mode (for TCPDUMP and cousins) or accept
 * a select set of multicast packets
 */
B
Bryan Wu 已提交
1314
static void bfin_mac_set_multicast_list(struct net_device *dev)
1315 1316 1317 1318
{
	u32 sysctl;

	if (dev->flags & IFF_PROMISC) {
1319
		netdev_info(dev, "set promisc mode\n");
1320
		sysctl = bfin_read_EMAC_OPMODE();
1321
		sysctl |= PR;
1322
		bfin_write_EMAC_OPMODE(sysctl);
1323
	} else if (dev->flags & IFF_ALLMULTI) {
1324 1325 1326 1327
		/* accept all multicast */
		sysctl = bfin_read_EMAC_OPMODE();
		sysctl |= PAM;
		bfin_write_EMAC_OPMODE(sysctl);
1328
	} else if (!netdev_mc_empty(dev)) {
1329 1330 1331 1332
		/* set up multicast hash table */
		sysctl = bfin_read_EMAC_OPMODE();
		sysctl |= HM;
		bfin_write_EMAC_OPMODE(sysctl);
B
Bryan Wu 已提交
1333
		bfin_mac_multicast_hash(dev);
1334 1335 1336 1337 1338 1339 1340 1341
	} else {
		/* clear promisc or multicast mode */
		sysctl = bfin_read_EMAC_OPMODE();
		sysctl &= ~(RAF | PAM);
		bfin_write_EMAC_OPMODE(sysctl);
	}
}

1342 1343
static int bfin_mac_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd)
{
1344 1345 1346 1347 1348
	struct bfin_mac_local *lp = netdev_priv(netdev);

	if (!netif_running(netdev))
		return -EINVAL;

1349 1350 1351 1352
	switch (cmd) {
	case SIOCSHWTSTAMP:
		return bfin_mac_hwtstamp_ioctl(netdev, ifr, cmd);
	default:
1353 1354 1355 1356
		if (lp->phydev)
			return phy_mii_ioctl(lp->phydev, ifr, cmd);
		else
			return -EOPNOTSUPP;
1357 1358 1359
	}
}

1360 1361 1362
/*
 * this puts the device in an inactive state
 */
B
Bryan Wu 已提交
1363
static void bfin_mac_shutdown(struct net_device *dev)
1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376
{
	/* Turn off the EMAC */
	bfin_write_EMAC_OPMODE(0x00000000);
	/* Turn off the EMAC RX DMA */
	bfin_write_DMA1_CONFIG(0x0000);
	bfin_write_DMA2_CONFIG(0x0000);
}

/*
 * Open and Initialize the interface
 *
 * Set up everything, reset the card, etc..
 */
B
Bryan Wu 已提交
1377
static int bfin_mac_open(struct net_device *dev)
1378
{
B
Bryan Wu 已提交
1379
	struct bfin_mac_local *lp = netdev_priv(dev);
1380
	int ret;
1381
	pr_debug("%s: %s\n", dev->name, __func__);
1382 1383 1384 1385 1386 1387 1388

	/*
	 * Check that the address is valid.  If its not, refuse
	 * to bring the device up.  The user must specify an
	 * address using ifconfig eth0 hw ether xx:xx:xx:xx:xx:xx
	 */
	if (!is_valid_ether_addr(dev->dev_addr)) {
1389
		netdev_warn(dev, "no valid ethernet hw addr\n");
1390 1391 1392 1393
		return -EINVAL;
	}

	/* initial rx and tx list */
1394 1395 1396
	ret = desc_list_init();
	if (ret)
		return ret;
1397

1398
	phy_start(lp->phydev);
V
Vitja Makarov 已提交
1399
	phy_write(lp->phydev, MII_BMCR, BMCR_RESET);
1400
	setup_system_regs(dev);
1401
	setup_mac_addr(dev->dev_addr);
1402

B
Bryan Wu 已提交
1403
	bfin_mac_disable();
1404
	ret = bfin_mac_enable(lp->phydev);
1405 1406
	if (ret)
		return ret;
1407
	pr_debug("hardware init finished\n");
1408

1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419
	netif_start_queue(dev);
	netif_carrier_on(dev);

	return 0;
}

/*
 * this makes the board clean up everything that it can
 * and not talk to the outside world.   Caused by
 * an 'ifconfig ethX down'
 */
B
Bryan Wu 已提交
1420
static int bfin_mac_close(struct net_device *dev)
1421
{
B
Bryan Wu 已提交
1422
	struct bfin_mac_local *lp = netdev_priv(dev);
1423
	pr_debug("%s: %s\n", dev->name, __func__);
1424 1425 1426 1427

	netif_stop_queue(dev);
	netif_carrier_off(dev);

1428
	phy_stop(lp->phydev);
V
Vitja Makarov 已提交
1429
	phy_write(lp->phydev, MII_BMCR, BMCR_PDOWN);
1430

1431
	/* clear everything */
B
Bryan Wu 已提交
1432
	bfin_mac_shutdown(dev);
1433 1434 1435 1436 1437 1438 1439

	/* free the rx/tx buffers */
	desc_list_free();

	return 0;
}

1440 1441 1442 1443 1444 1445 1446
static const struct net_device_ops bfin_mac_netdev_ops = {
	.ndo_open		= bfin_mac_open,
	.ndo_stop		= bfin_mac_close,
	.ndo_start_xmit		= bfin_mac_hard_start_xmit,
	.ndo_set_mac_address	= bfin_mac_set_mac_address,
	.ndo_tx_timeout		= bfin_mac_timeout,
	.ndo_set_multicast_list	= bfin_mac_set_multicast_list,
1447
	.ndo_do_ioctl           = bfin_mac_ioctl,
1448 1449 1450 1451 1452 1453 1454
	.ndo_validate_addr	= eth_validate_addr,
	.ndo_change_mtu		= eth_change_mtu,
#ifdef CONFIG_NET_POLL_CONTROLLER
	.ndo_poll_controller	= bfin_mac_poll,
#endif
};

1455
static int __devinit bfin_mac_probe(struct platform_device *pdev)
1456
{
B
Bryan Wu 已提交
1457 1458
	struct net_device *ndev;
	struct bfin_mac_local *lp;
1459
	struct platform_device *pd;
1460
	struct bfin_mii_bus_platform_data *mii_bus_data;
1461
	int rc;
B
Bryan Wu 已提交
1462 1463 1464 1465 1466 1467 1468 1469 1470 1471

	ndev = alloc_etherdev(sizeof(struct bfin_mac_local));
	if (!ndev) {
		dev_err(&pdev->dev, "Cannot allocate net device!\n");
		return -ENOMEM;
	}

	SET_NETDEV_DEV(ndev, &pdev->dev);
	platform_set_drvdata(pdev, ndev);
	lp = netdev_priv(ndev);
1472
	lp->ndev = ndev;
1473 1474

	/* Grab the MAC address in the MAC */
B
Bryan Wu 已提交
1475 1476
	*(__le32 *) (&(ndev->dev_addr[0])) = cpu_to_le32(bfin_read_EMAC_ADDRLO());
	*(__le16 *) (&(ndev->dev_addr[4])) = cpu_to_le16((u16) bfin_read_EMAC_ADDRHI());
1477 1478 1479 1480 1481

	/* probe mac */
	/*todo: how to proble? which is revision_register */
	bfin_write_EMAC_ADDRLO(0x12345678);
	if (bfin_read_EMAC_ADDRLO() != 0x12345678) {
B
Bryan Wu 已提交
1482 1483 1484
		dev_err(&pdev->dev, "Cannot detect Blackfin on-chip ethernet MAC controller!\n");
		rc = -ENODEV;
		goto out_err_probe_mac;
1485 1486 1487
	}


B
Bryan Wu 已提交
1488 1489 1490 1491 1492 1493 1494 1495
	/*
	 * Is it valid? (Did bootloader initialize it?)
	 * Grab the MAC from the board somehow
	 * this is done in the arch/blackfin/mach-bfxxx/boards/eth_mac.c
	 */
	if (!is_valid_ether_addr(ndev->dev_addr))
		bfin_get_ether_addr(ndev->dev_addr);

1496
	/* If still not valid, get a random one */
B
Bryan Wu 已提交
1497 1498
	if (!is_valid_ether_addr(ndev->dev_addr))
		random_ether_addr(ndev->dev_addr);
1499

B
Bryan Wu 已提交
1500
	setup_mac_addr(ndev->dev_addr);
1501

1502 1503 1504 1505
	if (!pdev->dev.platform_data) {
		dev_err(&pdev->dev, "Cannot get platform device bfin_mii_bus!\n");
		rc = -ENODEV;
		goto out_err_probe_mac;
B
Bryan Wu 已提交
1506
	}
1507 1508
	pd = pdev->dev.platform_data;
	lp->mii_bus = platform_get_drvdata(pd);
1509 1510 1511
	if (!lp->mii_bus) {
		dev_err(&pdev->dev, "Cannot get mii_bus!\n");
		rc = -ENODEV;
1512
		goto out_err_probe_mac;
1513
	}
1514
	lp->mii_bus->priv = ndev;
1515
	mii_bus_data = pd->dev.platform_data;
1516

1517
	rc = mii_probe(ndev, mii_bus_data->phy_mode);
B
Bryan Wu 已提交
1518 1519 1520 1521
	if (rc) {
		dev_err(&pdev->dev, "MII Probe failed!\n");
		goto out_err_mii_probe;
	}
1522

1523
	/* Fill in the fields of the device structure with ethernet values. */
B
Bryan Wu 已提交
1524 1525
	ether_setup(ndev);

1526
	ndev->netdev_ops = &bfin_mac_netdev_ops;
1527
	ndev->ethtool_ops = &bfin_mac_ethtool_ops;
1528

1529 1530 1531 1532
	init_timer(&lp->tx_reclaim_timer);
	lp->tx_reclaim_timer.data = (unsigned long)lp;
	lp->tx_reclaim_timer.function = tx_reclaim_skb_timeout;

1533 1534 1535 1536
	spin_lock_init(&lp->lock);

	/* now, enable interrupts */
	/* register irq handler */
B
Bryan Wu 已提交
1537
	rc = request_irq(IRQ_MAC_RX, bfin_mac_interrupt,
1538
			IRQF_DISABLED, "EMAC_RX", ndev);
B
Bryan Wu 已提交
1539 1540 1541 1542
	if (rc) {
		dev_err(&pdev->dev, "Cannot request Blackfin MAC RX IRQ!\n");
		rc = -EBUSY;
		goto out_err_request_irq;
1543 1544
	}

B
Bryan Wu 已提交
1545 1546 1547 1548
	rc = register_netdev(ndev);
	if (rc) {
		dev_err(&pdev->dev, "Cannot register net device!\n");
		goto out_err_reg_ndev;
1549 1550
	}

1551 1552
	bfin_mac_hwtstamp_init(ndev);

B
Bryan Wu 已提交
1553
	/* now, print out the card info, in a short format.. */
1554
	netdev_info(ndev, "%s, Version %s\n", DRV_DESC, DRV_VERSION);
1555

B
Bryan Wu 已提交
1556
	return 0;
1557

B
Bryan Wu 已提交
1558 1559 1560 1561
out_err_reg_ndev:
	free_irq(IRQ_MAC_RX, ndev);
out_err_request_irq:
out_err_mii_probe:
1562 1563
	mdiobus_unregister(lp->mii_bus);
	mdiobus_free(lp->mii_bus);
B
Bryan Wu 已提交
1564 1565 1566
out_err_probe_mac:
	platform_set_drvdata(pdev, NULL);
	free_netdev(ndev);
1567

B
Bryan Wu 已提交
1568
	return rc;
1569 1570
}

1571
static int __devexit bfin_mac_remove(struct platform_device *pdev)
1572 1573
{
	struct net_device *ndev = platform_get_drvdata(pdev);
B
Bryan Wu 已提交
1574
	struct bfin_mac_local *lp = netdev_priv(ndev);
1575 1576 1577

	platform_set_drvdata(pdev, NULL);

1578
	lp->mii_bus->priv = NULL;
B
Bryan Wu 已提交
1579

1580 1581 1582 1583 1584 1585 1586 1587 1588
	unregister_netdev(ndev);

	free_irq(IRQ_MAC_RX, ndev);

	free_netdev(ndev);

	return 0;
}

1589 1590
#ifdef CONFIG_PM
static int bfin_mac_suspend(struct platform_device *pdev, pm_message_t mesg)
1591
{
1592
	struct net_device *net_dev = platform_get_drvdata(pdev);
1593
	struct bfin_mac_local *lp = netdev_priv(net_dev);
1594

1595 1596 1597 1598 1599 1600 1601 1602
	if (lp->wol) {
		bfin_write_EMAC_OPMODE((bfin_read_EMAC_OPMODE() & ~TE) | RE);
		bfin_write_EMAC_WKUP_CTL(MPKE);
		enable_irq_wake(IRQ_MAC_WAKEDET);
	} else {
		if (netif_running(net_dev))
			bfin_mac_close(net_dev);
	}
1603

1604 1605 1606 1607 1608
	return 0;
}

static int bfin_mac_resume(struct platform_device *pdev)
{
1609
	struct net_device *net_dev = platform_get_drvdata(pdev);
1610
	struct bfin_mac_local *lp = netdev_priv(net_dev);
1611

1612 1613 1614 1615 1616 1617 1618 1619
	if (lp->wol) {
		bfin_write_EMAC_OPMODE(bfin_read_EMAC_OPMODE() | TE);
		bfin_write_EMAC_WKUP_CTL(0);
		disable_irq_wake(IRQ_MAC_WAKEDET);
	} else {
		if (netif_running(net_dev))
			bfin_mac_open(net_dev);
	}
1620

1621 1622
	return 0;
}
1623 1624 1625 1626
#else
#define bfin_mac_suspend NULL
#define bfin_mac_resume NULL
#endif	/* CONFIG_PM */
1627

1628 1629 1630
static int __devinit bfin_mii_bus_probe(struct platform_device *pdev)
{
	struct mii_bus *miibus;
1631 1632
	struct bfin_mii_bus_platform_data *mii_bus_pd;
	const unsigned short *pin_req;
1633 1634
	int rc, i;

1635 1636 1637 1638 1639 1640
	mii_bus_pd = dev_get_platdata(&pdev->dev);
	if (!mii_bus_pd) {
		dev_err(&pdev->dev, "No peripherals in platform data!\n");
		return -EINVAL;
	}

1641 1642 1643 1644
	/*
	 * We are setting up a network card,
	 * so set the GPIO pins to Ethernet mode
	 */
1645
	pin_req = mii_bus_pd->mac_peripherals;
1646
	rc = peripheral_request_list(pin_req, KBUILD_MODNAME);
1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661
	if (rc) {
		dev_err(&pdev->dev, "Requesting peripherals failed!\n");
		return rc;
	}

	rc = -ENOMEM;
	miibus = mdiobus_alloc();
	if (miibus == NULL)
		goto out_err_alloc;
	miibus->read = bfin_mdiobus_read;
	miibus->write = bfin_mdiobus_write;
	miibus->reset = bfin_mdiobus_reset;

	miibus->parent = &pdev->dev;
	miibus->name = "bfin_mii_bus";
1662 1663
	miibus->phy_mask = mii_bus_pd->phy_mask;

1664 1665
	snprintf(miibus->id, MII_BUS_ID_SIZE, "0");
	miibus->irq = kmalloc(sizeof(int)*PHY_MAX_ADDR, GFP_KERNEL);
1666 1667 1668 1669
	if (!miibus->irq)
		goto out_err_irq_alloc;

	for (i = rc; i < PHY_MAX_ADDR; ++i)
1670 1671
		miibus->irq[i] = PHY_POLL;

1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685
	rc = clamp(mii_bus_pd->phydev_number, 0, PHY_MAX_ADDR);
	if (rc != mii_bus_pd->phydev_number)
		dev_err(&pdev->dev, "Invalid number (%i) of phydevs\n",
			mii_bus_pd->phydev_number);
	for (i = 0; i < rc; ++i) {
		unsigned short phyaddr = mii_bus_pd->phydev_data[i].addr;
		if (phyaddr < PHY_MAX_ADDR)
			miibus->irq[phyaddr] = mii_bus_pd->phydev_data[i].irq;
		else
			dev_err(&pdev->dev,
				"Invalid PHY address %i for phydev %i\n",
				phyaddr, i);
	}

1686 1687 1688 1689 1690 1691 1692 1693 1694 1695
	rc = mdiobus_register(miibus);
	if (rc) {
		dev_err(&pdev->dev, "Cannot register MDIO bus!\n");
		goto out_err_mdiobus_register;
	}

	platform_set_drvdata(pdev, miibus);
	return 0;

out_err_mdiobus_register:
1696
	kfree(miibus->irq);
1697
out_err_irq_alloc:
1698 1699 1700 1701 1702 1703 1704 1705 1706 1707
	mdiobus_free(miibus);
out_err_alloc:
	peripheral_free_list(pin_req);

	return rc;
}

static int __devexit bfin_mii_bus_remove(struct platform_device *pdev)
{
	struct mii_bus *miibus = platform_get_drvdata(pdev);
1708 1709 1710
	struct bfin_mii_bus_platform_data *mii_bus_pd =
		dev_get_platdata(&pdev->dev);

1711 1712
	platform_set_drvdata(pdev, NULL);
	mdiobus_unregister(miibus);
1713
	kfree(miibus->irq);
1714
	mdiobus_free(miibus);
1715 1716
	peripheral_free_list(mii_bus_pd->mac_peripherals);

1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728
	return 0;
}

static struct platform_driver bfin_mii_bus_driver = {
	.probe = bfin_mii_bus_probe,
	.remove = __devexit_p(bfin_mii_bus_remove),
	.driver = {
		.name = "bfin_mii_bus",
		.owner	= THIS_MODULE,
	},
};

1729 1730
static struct platform_driver bfin_mac_driver = {
	.probe = bfin_mac_probe,
1731
	.remove = __devexit_p(bfin_mac_remove),
1732 1733 1734
	.resume = bfin_mac_resume,
	.suspend = bfin_mac_suspend,
	.driver = {
1735
		.name = KBUILD_MODNAME,
1736 1737
		.owner	= THIS_MODULE,
	},
1738 1739 1740 1741
};

static int __init bfin_mac_init(void)
{
1742 1743 1744 1745 1746
	int ret;
	ret = platform_driver_register(&bfin_mii_bus_driver);
	if (!ret)
		return platform_driver_register(&bfin_mac_driver);
	return -ENODEV;
1747 1748 1749 1750 1751 1752 1753
}

module_init(bfin_mac_init);

static void __exit bfin_mac_cleanup(void)
{
	platform_driver_unregister(&bfin_mac_driver);
1754
	platform_driver_unregister(&bfin_mii_bus_driver);
1755 1756 1757
}

module_exit(bfin_mac_cleanup);
1758