Commit 24dd377a authored by Christophe JAILLET's avatar Christophe JAILLET Committed by David S. Miller

wan: wanxl: switch from 'pci_' to 'dma_' API

The wrappers in include/linux/pci-dma-compat.h should go away.

The patch has been generated with the coccinelle script below and has been
hand modified to replace GFP_ with a correct flag.
It has been compile tested.

When memory is allocated in 'wanxl_pci_init_one()', GFP_KERNEL can be used
because it is a probe function and no lock is acquired.
Moreover, just a few lines above, GFP_KERNEL is already used.

@@
@@
-    PCI_DMA_BIDIRECTIONAL
+    DMA_BIDIRECTIONAL

@@
@@
-    PCI_DMA_TODEVICE
+    DMA_TO_DEVICE

@@
@@
-    PCI_DMA_FROMDEVICE
+    DMA_FROM_DEVICE

@@
@@
-    PCI_DMA_NONE
+    DMA_NONE

@@
expression e1, e2, e3;
@@
-    pci_alloc_consistent(e1, e2, e3)
+    dma_alloc_coherent(&e1->dev, e2, e3, GFP_)

@@
expression e1, e2, e3;
@@
-    pci_zalloc_consistent(e1, e2, e3)
+    dma_alloc_coherent(&e1->dev, e2, e3, GFP_)

@@
expression e1, e2, e3, e4;
@@
-    pci_free_consistent(e1, e2, e3, e4)
+    dma_free_coherent(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_map_single(e1, e2, e3, e4)
+    dma_map_single(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_unmap_single(e1, e2, e3, e4)
+    dma_unmap_single(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4, e5;
@@
-    pci_map_page(e1, e2, e3, e4, e5)
+    dma_map_page(&e1->dev, e2, e3, e4, e5)

@@
expression e1, e2, e3, e4;
@@
-    pci_unmap_page(e1, e2, e3, e4)
+    dma_unmap_page(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_map_sg(e1, e2, e3, e4)
+    dma_map_sg(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_unmap_sg(e1, e2, e3, e4)
+    dma_unmap_sg(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_dma_sync_single_for_cpu(e1, e2, e3, e4)
+    dma_sync_single_for_cpu(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_dma_sync_single_for_device(e1, e2, e3, e4)
+    dma_sync_single_for_device(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_dma_sync_sg_for_cpu(e1, e2, e3, e4)
+    dma_sync_sg_for_cpu(&e1->dev, e2, e3, e4)

@@
expression e1, e2, e3, e4;
@@
-    pci_dma_sync_sg_for_device(e1, e2, e3, e4)
+    dma_sync_sg_for_device(&e1->dev, e2, e3, e4)

@@
expression e1, e2;
@@
-    pci_dma_mapping_error(e1, e2)
+    dma_mapping_error(&e1->dev, e2)

@@
expression e1, e2;
@@
-    pci_set_dma_mask(e1, e2)
+    dma_set_mask(&e1->dev, e2)

@@
expression e1, e2;
@@
-    pci_set_consistent_dma_mask(e1, e2)
+    dma_set_coherent_mask(&e1->dev, e2)
Signed-off-by: default avatarChristophe JAILLET <christophe.jaillet@wanadoo.fr>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 7c9864bb
...@@ -99,7 +99,7 @@ static inline port_status_t *get_status(struct port *port) ...@@ -99,7 +99,7 @@ static inline port_status_t *get_status(struct port *port)
static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr, static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
size_t size, int direction) size_t size, int direction)
{ {
dma_addr_t addr = pci_map_single(pdev, ptr, size, direction); dma_addr_t addr = dma_map_single(&pdev->dev, ptr, size, direction);
if (addr + size > 0x100000000LL) if (addr + size > 0x100000000LL)
pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n", pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
pci_name(pdev), (unsigned long long)addr); pci_name(pdev), (unsigned long long)addr);
...@@ -180,8 +180,8 @@ static inline void wanxl_tx_intr(struct port *port) ...@@ -180,8 +180,8 @@ static inline void wanxl_tx_intr(struct port *port)
dev->stats.tx_bytes += skb->len; dev->stats.tx_bytes += skb->len;
} }
desc->stat = PACKET_EMPTY; /* Free descriptor */ desc->stat = PACKET_EMPTY; /* Free descriptor */
pci_unmap_single(port->card->pdev, desc->address, skb->len, dma_unmap_single(&port->card->pdev->dev, desc->address,
PCI_DMA_TODEVICE); skb->len, DMA_TO_DEVICE);
dev_consume_skb_irq(skb); dev_consume_skb_irq(skb);
port->tx_in = (port->tx_in + 1) % TX_BUFFERS; port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
} }
...@@ -207,9 +207,9 @@ static inline void wanxl_rx_intr(struct card *card) ...@@ -207,9 +207,9 @@ static inline void wanxl_rx_intr(struct card *card)
if (!skb) if (!skb)
dev->stats.rx_dropped++; dev->stats.rx_dropped++;
else { else {
pci_unmap_single(card->pdev, desc->address, dma_unmap_single(&card->pdev->dev,
BUFFER_LENGTH, desc->address, BUFFER_LENGTH,
PCI_DMA_FROMDEVICE); DMA_FROM_DEVICE);
skb_put(skb, desc->length); skb_put(skb, desc->length);
#ifdef DEBUG_PKT #ifdef DEBUG_PKT
...@@ -227,9 +227,10 @@ static inline void wanxl_rx_intr(struct card *card) ...@@ -227,9 +227,10 @@ static inline void wanxl_rx_intr(struct card *card)
if (!skb) { if (!skb) {
skb = dev_alloc_skb(BUFFER_LENGTH); skb = dev_alloc_skb(BUFFER_LENGTH);
desc->address = skb ? desc->address = skb ?
pci_map_single(card->pdev, skb->data, dma_map_single(&card->pdev->dev,
skb->data,
BUFFER_LENGTH, BUFFER_LENGTH,
PCI_DMA_FROMDEVICE) : 0; DMA_FROM_DEVICE) : 0;
card->rx_skbs[card->rx_in] = skb; card->rx_skbs[card->rx_in] = skb;
} }
} }
...@@ -291,8 +292,8 @@ static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev) ...@@ -291,8 +292,8 @@ static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
#endif #endif
port->tx_skbs[port->tx_out] = skb; port->tx_skbs[port->tx_out] = skb;
desc->address = pci_map_single(port->card->pdev, skb->data, skb->len, desc->address = dma_map_single(&port->card->pdev->dev, skb->data,
PCI_DMA_TODEVICE); skb->len, DMA_TO_DEVICE);
desc->length = skb->len; desc->length = skb->len;
desc->stat = PACKET_FULL; desc->stat = PACKET_FULL;
writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node), writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
...@@ -451,9 +452,9 @@ static int wanxl_close(struct net_device *dev) ...@@ -451,9 +452,9 @@ static int wanxl_close(struct net_device *dev)
if (desc->stat != PACKET_EMPTY) { if (desc->stat != PACKET_EMPTY) {
desc->stat = PACKET_EMPTY; desc->stat = PACKET_EMPTY;
pci_unmap_single(port->card->pdev, desc->address, dma_unmap_single(&port->card->pdev->dev,
port->tx_skbs[i]->len, desc->address, port->tx_skbs[i]->len,
PCI_DMA_TODEVICE); DMA_TO_DEVICE);
dev_kfree_skb(port->tx_skbs[i]); dev_kfree_skb(port->tx_skbs[i]);
} }
} }
...@@ -524,9 +525,9 @@ static void wanxl_pci_remove_one(struct pci_dev *pdev) ...@@ -524,9 +525,9 @@ static void wanxl_pci_remove_one(struct pci_dev *pdev)
for (i = 0; i < RX_QUEUE_LENGTH; i++) for (i = 0; i < RX_QUEUE_LENGTH; i++)
if (card->rx_skbs[i]) { if (card->rx_skbs[i]) {
pci_unmap_single(card->pdev, dma_unmap_single(&card->pdev->dev,
card->status->rx_descs[i].address, card->status->rx_descs[i].address,
BUFFER_LENGTH, PCI_DMA_FROMDEVICE); BUFFER_LENGTH, DMA_FROM_DEVICE);
dev_kfree_skb(card->rx_skbs[i]); dev_kfree_skb(card->rx_skbs[i]);
} }
...@@ -534,7 +535,7 @@ static void wanxl_pci_remove_one(struct pci_dev *pdev) ...@@ -534,7 +535,7 @@ static void wanxl_pci_remove_one(struct pci_dev *pdev)
iounmap(card->plx); iounmap(card->plx);
if (card->status) if (card->status)
pci_free_consistent(pdev, sizeof(struct card_status), dma_free_coherent(&pdev->dev, sizeof(struct card_status),
card->status, card->status_address); card->status, card->status_address);
pci_release_regions(pdev); pci_release_regions(pdev);
...@@ -579,8 +580,8 @@ static int wanxl_pci_init_one(struct pci_dev *pdev, ...@@ -579,8 +580,8 @@ static int wanxl_pci_init_one(struct pci_dev *pdev,
We set both dma_mask and consistent_dma_mask to 28 bits We set both dma_mask and consistent_dma_mask to 28 bits
and pray pci_alloc_consistent() will use this info. It should and pray pci_alloc_consistent() will use this info. It should
work on most platforms */ work on most platforms */
if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(28)) || if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(28)) ||
pci_set_dma_mask(pdev, DMA_BIT_MASK(28))) { dma_set_mask(&pdev->dev, DMA_BIT_MASK(28))) {
pr_err("No usable DMA configuration\n"); pr_err("No usable DMA configuration\n");
pci_disable_device(pdev); pci_disable_device(pdev);
return -EIO; return -EIO;
...@@ -608,9 +609,9 @@ static int wanxl_pci_init_one(struct pci_dev *pdev, ...@@ -608,9 +609,9 @@ static int wanxl_pci_init_one(struct pci_dev *pdev,
pci_set_drvdata(pdev, card); pci_set_drvdata(pdev, card);
card->pdev = pdev; card->pdev = pdev;
card->status = pci_alloc_consistent(pdev, card->status = dma_alloc_coherent(&pdev->dev,
sizeof(struct card_status), sizeof(struct card_status),
&card->status_address); &card->status_address, GFP_KERNEL);
if (card->status == NULL) { if (card->status == NULL) {
wanxl_pci_remove_one(pdev); wanxl_pci_remove_one(pdev);
return -ENOBUFS; return -ENOBUFS;
...@@ -625,8 +626,8 @@ static int wanxl_pci_init_one(struct pci_dev *pdev, ...@@ -625,8 +626,8 @@ static int wanxl_pci_init_one(struct pci_dev *pdev,
/* FIXME when PCI/DMA subsystems are fixed. /* FIXME when PCI/DMA subsystems are fixed.
We set both dma_mask and consistent_dma_mask back to 32 bits We set both dma_mask and consistent_dma_mask back to 32 bits
to indicate the card can do 32-bit DMA addressing */ to indicate the card can do 32-bit DMA addressing */
if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) || if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)) ||
pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) { dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
pr_err("No usable DMA configuration\n"); pr_err("No usable DMA configuration\n");
wanxl_pci_remove_one(pdev); wanxl_pci_remove_one(pdev);
return -EIO; return -EIO;
...@@ -699,9 +700,8 @@ static int wanxl_pci_init_one(struct pci_dev *pdev, ...@@ -699,9 +700,8 @@ static int wanxl_pci_init_one(struct pci_dev *pdev,
card->rx_skbs[i] = skb; card->rx_skbs[i] = skb;
if (skb) if (skb)
card->status->rx_descs[i].address = card->status->rx_descs[i].address =
pci_map_single(card->pdev, skb->data, dma_map_single(&card->pdev->dev, skb->data,
BUFFER_LENGTH, BUFFER_LENGTH, DMA_FROM_DEVICE);
PCI_DMA_FROMDEVICE);
} }
mem = ioremap(mem_phy, PDM_OFFSET + sizeof(firmware)); mem = ioremap(mem_phy, PDM_OFFSET + sizeof(firmware));
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment