提交 e2aadb4a 编写于 作者: P Petri Gynther 提交者: David S. Miller

net: bcmgenet: rework Tx NAPI code

Introduce new bcmgenet functions to handle the NAPI calls to:
netif_napi_add()
napi_enable()
napi_disable()
netif_napi_del()
Signed-off-by: NPetri Gynther <pgynther@google.com>
Reviewed-by: NFlorian Fainelli <f.fainelli@gmail.com>
Signed-off-by: NDavid S. Miller <davem@davemloft.net>
上级 b2e97eca
...@@ -1733,7 +1733,6 @@ static void bcmgenet_init_tx_ring(struct bcmgenet_priv *priv, ...@@ -1733,7 +1733,6 @@ static void bcmgenet_init_tx_ring(struct bcmgenet_priv *priv,
spin_lock_init(&ring->lock); spin_lock_init(&ring->lock);
ring->priv = priv; ring->priv = priv;
netif_napi_add(priv->dev, &ring->napi, bcmgenet_tx_poll, 64);
ring->index = index; ring->index = index;
if (index == DESC_INDEX) { if (index == DESC_INDEX) {
ring->queue = 0; ring->queue = 0;
...@@ -1777,17 +1776,6 @@ static void bcmgenet_init_tx_ring(struct bcmgenet_priv *priv, ...@@ -1777,17 +1776,6 @@ static void bcmgenet_init_tx_ring(struct bcmgenet_priv *priv,
TDMA_WRITE_PTR); TDMA_WRITE_PTR);
bcmgenet_tdma_ring_writel(priv, index, end_ptr * words_per_bd - 1, bcmgenet_tdma_ring_writel(priv, index, end_ptr * words_per_bd - 1,
DMA_END_ADDR); DMA_END_ADDR);
napi_enable(&ring->napi);
}
static void bcmgenet_fini_tx_ring(struct bcmgenet_priv *priv,
unsigned int index)
{
struct bcmgenet_tx_ring *ring = &priv->tx_rings[index];
napi_disable(&ring->napi);
netif_napi_del(&ring->napi);
} }
/* Initialize a RDMA ring */ /* Initialize a RDMA ring */
...@@ -1835,6 +1823,62 @@ static int bcmgenet_init_rx_ring(struct bcmgenet_priv *priv, ...@@ -1835,6 +1823,62 @@ static int bcmgenet_init_rx_ring(struct bcmgenet_priv *priv,
return ret; return ret;
} }
static void bcmgenet_init_tx_napi(struct bcmgenet_priv *priv)
{
unsigned int i;
struct bcmgenet_tx_ring *ring;
for (i = 0; i < priv->hw_params->tx_queues; ++i) {
ring = &priv->tx_rings[i];
netif_napi_add(priv->dev, &ring->napi, bcmgenet_tx_poll, 64);
}
ring = &priv->tx_rings[DESC_INDEX];
netif_napi_add(priv->dev, &ring->napi, bcmgenet_tx_poll, 64);
}
static void bcmgenet_enable_tx_napi(struct bcmgenet_priv *priv)
{
unsigned int i;
struct bcmgenet_tx_ring *ring;
for (i = 0; i < priv->hw_params->tx_queues; ++i) {
ring = &priv->tx_rings[i];
napi_enable(&ring->napi);
}
ring = &priv->tx_rings[DESC_INDEX];
napi_enable(&ring->napi);
}
static void bcmgenet_disable_tx_napi(struct bcmgenet_priv *priv)
{
unsigned int i;
struct bcmgenet_tx_ring *ring;
for (i = 0; i < priv->hw_params->tx_queues; ++i) {
ring = &priv->tx_rings[i];
napi_disable(&ring->napi);
}
ring = &priv->tx_rings[DESC_INDEX];
napi_disable(&ring->napi);
}
static void bcmgenet_fini_tx_napi(struct bcmgenet_priv *priv)
{
unsigned int i;
struct bcmgenet_tx_ring *ring;
for (i = 0; i < priv->hw_params->tx_queues; ++i) {
ring = &priv->tx_rings[i];
netif_napi_del(&ring->napi);
}
ring = &priv->tx_rings[DESC_INDEX];
netif_napi_del(&ring->napi);
}
/* Initialize Tx queues /* Initialize Tx queues
* *
* Queues 0-3 are priority-based, each one has 32 descriptors, * Queues 0-3 are priority-based, each one has 32 descriptors,
...@@ -1895,6 +1939,9 @@ static void bcmgenet_init_tx_queues(struct net_device *dev) ...@@ -1895,6 +1939,9 @@ static void bcmgenet_init_tx_queues(struct net_device *dev)
bcmgenet_tdma_writel(priv, dma_priority[1], DMA_PRIORITY_1); bcmgenet_tdma_writel(priv, dma_priority[1], DMA_PRIORITY_1);
bcmgenet_tdma_writel(priv, dma_priority[2], DMA_PRIORITY_2); bcmgenet_tdma_writel(priv, dma_priority[2], DMA_PRIORITY_2);
/* Initialize Tx NAPI */
bcmgenet_init_tx_napi(priv);
/* Enable Tx queues */ /* Enable Tx queues */
bcmgenet_tdma_writel(priv, ring_cfg, DMA_RING_CFG); bcmgenet_tdma_writel(priv, ring_cfg, DMA_RING_CFG);
...@@ -2036,12 +2083,7 @@ static void __bcmgenet_fini_dma(struct bcmgenet_priv *priv) ...@@ -2036,12 +2083,7 @@ static void __bcmgenet_fini_dma(struct bcmgenet_priv *priv)
static void bcmgenet_fini_dma(struct bcmgenet_priv *priv) static void bcmgenet_fini_dma(struct bcmgenet_priv *priv)
{ {
int i; bcmgenet_fini_tx_napi(priv);
bcmgenet_fini_tx_ring(priv, DESC_INDEX);
for (i = 0; i < priv->hw_params->tx_queues; i++)
bcmgenet_fini_tx_ring(priv, i);
__bcmgenet_fini_dma(priv); __bcmgenet_fini_dma(priv);
} }
...@@ -2466,6 +2508,7 @@ static void bcmgenet_netif_start(struct net_device *dev) ...@@ -2466,6 +2508,7 @@ static void bcmgenet_netif_start(struct net_device *dev)
/* Start the network engine */ /* Start the network engine */
napi_enable(&priv->napi); napi_enable(&priv->napi);
bcmgenet_enable_tx_napi(priv);
umac_enable_set(priv, CMD_TX_EN | CMD_RX_EN, true); umac_enable_set(priv, CMD_TX_EN | CMD_RX_EN, true);
...@@ -2574,6 +2617,7 @@ static void bcmgenet_netif_stop(struct net_device *dev) ...@@ -2574,6 +2617,7 @@ static void bcmgenet_netif_stop(struct net_device *dev)
phy_stop(priv->phydev); phy_stop(priv->phydev);
bcmgenet_intr_disable(priv); bcmgenet_intr_disable(priv);
bcmgenet_disable_tx_napi(priv);
/* Wait for pending work items to complete. Since interrupts are /* Wait for pending work items to complete. Since interrupts are
* disabled no new work will be scheduled. * disabled no new work will be scheduled.
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册