提交 7598b349 编写于 作者: Z Zhu Yanjun 提交者: David S. Miller

forcedeth: replace pci_map_single with dma_map_single functions

pci_map_single functions are obsolete. So replace them with
dma_map_single functions.
Signed-off-by: NZhu Yanjun <yanjun.zhu@oracle.com>
Signed-off-by: NDavid S. Miller <davem@davemloft.net>
上级 5416953f
...@@ -1813,12 +1813,12 @@ static int nv_alloc_rx(struct net_device *dev) ...@@ -1813,12 +1813,12 @@ static int nv_alloc_rx(struct net_device *dev)
struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD); struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD);
if (skb) { if (skb) {
np->put_rx_ctx->skb = skb; np->put_rx_ctx->skb = skb;
np->put_rx_ctx->dma = pci_map_single(np->pci_dev, np->put_rx_ctx->dma = dma_map_single(&np->pci_dev->dev,
skb->data, skb->data,
skb_tailroom(skb), skb_tailroom(skb),
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
if (pci_dma_mapping_error(np->pci_dev, if (dma_mapping_error(&np->pci_dev->dev,
np->put_rx_ctx->dma)) { np->put_rx_ctx->dma)) {
kfree_skb(skb); kfree_skb(skb);
goto packet_dropped; goto packet_dropped;
} }
...@@ -1854,12 +1854,12 @@ static int nv_alloc_rx_optimized(struct net_device *dev) ...@@ -1854,12 +1854,12 @@ static int nv_alloc_rx_optimized(struct net_device *dev)
struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD); struct sk_buff *skb = netdev_alloc_skb(dev, np->rx_buf_sz + NV_RX_ALLOC_PAD);
if (skb) { if (skb) {
np->put_rx_ctx->skb = skb; np->put_rx_ctx->skb = skb;
np->put_rx_ctx->dma = pci_map_single(np->pci_dev, np->put_rx_ctx->dma = dma_map_single(&np->pci_dev->dev,
skb->data, skb->data,
skb_tailroom(skb), skb_tailroom(skb),
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
if (pci_dma_mapping_error(np->pci_dev, if (dma_mapping_error(&np->pci_dev->dev,
np->put_rx_ctx->dma)) { np->put_rx_ctx->dma)) {
kfree_skb(skb); kfree_skb(skb);
goto packet_dropped; goto packet_dropped;
} }
...@@ -1977,9 +1977,9 @@ static void nv_unmap_txskb(struct fe_priv *np, struct nv_skb_map *tx_skb) ...@@ -1977,9 +1977,9 @@ static void nv_unmap_txskb(struct fe_priv *np, struct nv_skb_map *tx_skb)
{ {
if (tx_skb->dma) { if (tx_skb->dma) {
if (tx_skb->dma_single) if (tx_skb->dma_single)
pci_unmap_single(np->pci_dev, tx_skb->dma, dma_unmap_single(&np->pci_dev->dev, tx_skb->dma,
tx_skb->dma_len, tx_skb->dma_len,
PCI_DMA_TODEVICE); DMA_TO_DEVICE);
else else
pci_unmap_page(np->pci_dev, tx_skb->dma, pci_unmap_page(np->pci_dev, tx_skb->dma,
tx_skb->dma_len, tx_skb->dma_len,
...@@ -2047,10 +2047,10 @@ static void nv_drain_rx(struct net_device *dev) ...@@ -2047,10 +2047,10 @@ static void nv_drain_rx(struct net_device *dev)
} }
wmb(); wmb();
if (np->rx_skb[i].skb) { if (np->rx_skb[i].skb) {
pci_unmap_single(np->pci_dev, np->rx_skb[i].dma, dma_unmap_single(&np->pci_dev->dev, np->rx_skb[i].dma,
(skb_end_pointer(np->rx_skb[i].skb) - (skb_end_pointer(np->rx_skb[i].skb) -
np->rx_skb[i].skb->data), np->rx_skb[i].skb->data),
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
dev_kfree_skb(np->rx_skb[i].skb); dev_kfree_skb(np->rx_skb[i].skb);
np->rx_skb[i].skb = NULL; np->rx_skb[i].skb = NULL;
} }
...@@ -2224,10 +2224,11 @@ static netdev_tx_t nv_start_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -2224,10 +2224,11 @@ static netdev_tx_t nv_start_xmit(struct sk_buff *skb, struct net_device *dev)
prev_tx = put_tx; prev_tx = put_tx;
prev_tx_ctx = np->put_tx_ctx; prev_tx_ctx = np->put_tx_ctx;
bcnt = (size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : size; bcnt = (size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : size;
np->put_tx_ctx->dma = pci_map_single(np->pci_dev, skb->data + offset, bcnt, np->put_tx_ctx->dma = dma_map_single(&np->pci_dev->dev,
PCI_DMA_TODEVICE); skb->data + offset, bcnt,
if (pci_dma_mapping_error(np->pci_dev, DMA_TO_DEVICE);
np->put_tx_ctx->dma)) { if (dma_mapping_error(&np->pci_dev->dev,
np->put_tx_ctx->dma)) {
/* on DMA mapping error - drop the packet */ /* on DMA mapping error - drop the packet */
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
u64_stats_update_begin(&np->swstats_tx_syncp); u64_stats_update_begin(&np->swstats_tx_syncp);
...@@ -2373,10 +2374,11 @@ static netdev_tx_t nv_start_xmit_optimized(struct sk_buff *skb, ...@@ -2373,10 +2374,11 @@ static netdev_tx_t nv_start_xmit_optimized(struct sk_buff *skb,
prev_tx = put_tx; prev_tx = put_tx;
prev_tx_ctx = np->put_tx_ctx; prev_tx_ctx = np->put_tx_ctx;
bcnt = (size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : size; bcnt = (size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : size;
np->put_tx_ctx->dma = pci_map_single(np->pci_dev, skb->data + offset, bcnt, np->put_tx_ctx->dma = dma_map_single(&np->pci_dev->dev,
PCI_DMA_TODEVICE); skb->data + offset, bcnt,
if (pci_dma_mapping_error(np->pci_dev, DMA_TO_DEVICE);
np->put_tx_ctx->dma)) { if (dma_mapping_error(&np->pci_dev->dev,
np->put_tx_ctx->dma)) {
/* on DMA mapping error - drop the packet */ /* on DMA mapping error - drop the packet */
dev_kfree_skb_any(skb); dev_kfree_skb_any(skb);
u64_stats_update_begin(&np->swstats_tx_syncp); u64_stats_update_begin(&np->swstats_tx_syncp);
...@@ -2810,9 +2812,9 @@ static int nv_rx_process(struct net_device *dev, int limit) ...@@ -2810,9 +2812,9 @@ static int nv_rx_process(struct net_device *dev, int limit)
* TODO: check if a prefetch of the first cacheline improves * TODO: check if a prefetch of the first cacheline improves
* the performance. * the performance.
*/ */
pci_unmap_single(np->pci_dev, np->get_rx_ctx->dma, dma_unmap_single(&np->pci_dev->dev, np->get_rx_ctx->dma,
np->get_rx_ctx->dma_len, np->get_rx_ctx->dma_len,
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
skb = np->get_rx_ctx->skb; skb = np->get_rx_ctx->skb;
np->get_rx_ctx->skb = NULL; np->get_rx_ctx->skb = NULL;
...@@ -2916,9 +2918,9 @@ static int nv_rx_process_optimized(struct net_device *dev, int limit) ...@@ -2916,9 +2918,9 @@ static int nv_rx_process_optimized(struct net_device *dev, int limit)
* TODO: check if a prefetch of the first cacheline improves * TODO: check if a prefetch of the first cacheline improves
* the performance. * the performance.
*/ */
pci_unmap_single(np->pci_dev, np->get_rx_ctx->dma, dma_unmap_single(&np->pci_dev->dev, np->get_rx_ctx->dma,
np->get_rx_ctx->dma_len, np->get_rx_ctx->dma_len,
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
skb = np->get_rx_ctx->skb; skb = np->get_rx_ctx->skb;
np->get_rx_ctx->skb = NULL; np->get_rx_ctx->skb = NULL;
...@@ -5070,11 +5072,11 @@ static int nv_loopback_test(struct net_device *dev) ...@@ -5070,11 +5072,11 @@ static int nv_loopback_test(struct net_device *dev)
ret = 0; ret = 0;
goto out; goto out;
} }
test_dma_addr = pci_map_single(np->pci_dev, tx_skb->data, test_dma_addr = dma_map_single(&np->pci_dev->dev, tx_skb->data,
skb_tailroom(tx_skb), skb_tailroom(tx_skb),
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
if (pci_dma_mapping_error(np->pci_dev, if (dma_mapping_error(&np->pci_dev->dev,
test_dma_addr)) { test_dma_addr)) {
dev_kfree_skb_any(tx_skb); dev_kfree_skb_any(tx_skb);
goto out; goto out;
} }
...@@ -5129,9 +5131,9 @@ static int nv_loopback_test(struct net_device *dev) ...@@ -5129,9 +5131,9 @@ static int nv_loopback_test(struct net_device *dev)
} }
} }
pci_unmap_single(np->pci_dev, test_dma_addr, dma_unmap_single(&np->pci_dev->dev, test_dma_addr,
(skb_end_pointer(tx_skb) - tx_skb->data), (skb_end_pointer(tx_skb) - tx_skb->data),
PCI_DMA_TODEVICE); DMA_TO_DEVICE);
dev_kfree_skb_any(tx_skb); dev_kfree_skb_any(tx_skb);
out: out:
/* stop engines */ /* stop engines */
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册