qlge: use the DMA state API instead of the pci equivalents
This replace the PCI DMA state API (include/linux/pci-dma.h) with the DMA equivalents since the PCI DMA state API will be obsolete. No functional change. For further information about the background: http://marc.info/?l=linux-netdev&m=127037540020276&w=2 Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp> Cc: Ron Mercer <ron.mercer@qlogic.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
committed by
David S. Miller
parent
56e3b9df13
commit
64b9b41de8
@@ -1344,8 +1344,8 @@ struct oal {
|
|||||||
};
|
};
|
||||||
|
|
||||||
struct map_list {
|
struct map_list {
|
||||||
DECLARE_PCI_UNMAP_ADDR(mapaddr);
|
DEFINE_DMA_UNMAP_ADDR(mapaddr);
|
||||||
DECLARE_PCI_UNMAP_LEN(maplen);
|
DEFINE_DMA_UNMAP_LEN(maplen);
|
||||||
};
|
};
|
||||||
|
|
||||||
struct tx_ring_desc {
|
struct tx_ring_desc {
|
||||||
@@ -1373,8 +1373,8 @@ struct bq_desc {
|
|||||||
} p;
|
} p;
|
||||||
__le64 *addr;
|
__le64 *addr;
|
||||||
u32 index;
|
u32 index;
|
||||||
DECLARE_PCI_UNMAP_ADDR(mapaddr);
|
DEFINE_DMA_UNMAP_ADDR(mapaddr);
|
||||||
DECLARE_PCI_UNMAP_LEN(maplen);
|
DEFINE_DMA_UNMAP_LEN(maplen);
|
||||||
};
|
};
|
||||||
|
|
||||||
#define QL_TXQ_IDX(qdev, skb) (smp_processor_id()%(qdev->tx_ring_count))
|
#define QL_TXQ_IDX(qdev, skb) (smp_processor_id()%(qdev->tx_ring_count))
|
||||||
|
@@ -1057,7 +1057,7 @@ static struct bq_desc *ql_get_curr_lchunk(struct ql_adapter *qdev,
|
|||||||
struct bq_desc *lbq_desc = ql_get_curr_lbuf(rx_ring);
|
struct bq_desc *lbq_desc = ql_get_curr_lbuf(rx_ring);
|
||||||
|
|
||||||
pci_dma_sync_single_for_cpu(qdev->pdev,
|
pci_dma_sync_single_for_cpu(qdev->pdev,
|
||||||
pci_unmap_addr(lbq_desc, mapaddr),
|
dma_unmap_addr(lbq_desc, mapaddr),
|
||||||
rx_ring->lbq_buf_size,
|
rx_ring->lbq_buf_size,
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
|
|
||||||
@@ -1170,8 +1170,8 @@ static void ql_update_lbq(struct ql_adapter *qdev, struct rx_ring *rx_ring)
|
|||||||
|
|
||||||
map = lbq_desc->p.pg_chunk.map +
|
map = lbq_desc->p.pg_chunk.map +
|
||||||
lbq_desc->p.pg_chunk.offset;
|
lbq_desc->p.pg_chunk.offset;
|
||||||
pci_unmap_addr_set(lbq_desc, mapaddr, map);
|
dma_unmap_addr_set(lbq_desc, mapaddr, map);
|
||||||
pci_unmap_len_set(lbq_desc, maplen,
|
dma_unmap_len_set(lbq_desc, maplen,
|
||||||
rx_ring->lbq_buf_size);
|
rx_ring->lbq_buf_size);
|
||||||
*lbq_desc->addr = cpu_to_le64(map);
|
*lbq_desc->addr = cpu_to_le64(map);
|
||||||
|
|
||||||
@@ -1241,8 +1241,8 @@ static void ql_update_sbq(struct ql_adapter *qdev, struct rx_ring *rx_ring)
|
|||||||
sbq_desc->p.skb = NULL;
|
sbq_desc->p.skb = NULL;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
pci_unmap_addr_set(sbq_desc, mapaddr, map);
|
dma_unmap_addr_set(sbq_desc, mapaddr, map);
|
||||||
pci_unmap_len_set(sbq_desc, maplen,
|
dma_unmap_len_set(sbq_desc, maplen,
|
||||||
rx_ring->sbq_buf_size);
|
rx_ring->sbq_buf_size);
|
||||||
*sbq_desc->addr = cpu_to_le64(map);
|
*sbq_desc->addr = cpu_to_le64(map);
|
||||||
}
|
}
|
||||||
@@ -1298,18 +1298,18 @@ static void ql_unmap_send(struct ql_adapter *qdev,
|
|||||||
"unmapping OAL area.\n");
|
"unmapping OAL area.\n");
|
||||||
}
|
}
|
||||||
pci_unmap_single(qdev->pdev,
|
pci_unmap_single(qdev->pdev,
|
||||||
pci_unmap_addr(&tx_ring_desc->map[i],
|
dma_unmap_addr(&tx_ring_desc->map[i],
|
||||||
mapaddr),
|
mapaddr),
|
||||||
pci_unmap_len(&tx_ring_desc->map[i],
|
dma_unmap_len(&tx_ring_desc->map[i],
|
||||||
maplen),
|
maplen),
|
||||||
PCI_DMA_TODEVICE);
|
PCI_DMA_TODEVICE);
|
||||||
} else {
|
} else {
|
||||||
netif_printk(qdev, tx_done, KERN_DEBUG, qdev->ndev,
|
netif_printk(qdev, tx_done, KERN_DEBUG, qdev->ndev,
|
||||||
"unmapping frag %d.\n", i);
|
"unmapping frag %d.\n", i);
|
||||||
pci_unmap_page(qdev->pdev,
|
pci_unmap_page(qdev->pdev,
|
||||||
pci_unmap_addr(&tx_ring_desc->map[i],
|
dma_unmap_addr(&tx_ring_desc->map[i],
|
||||||
mapaddr),
|
mapaddr),
|
||||||
pci_unmap_len(&tx_ring_desc->map[i],
|
dma_unmap_len(&tx_ring_desc->map[i],
|
||||||
maplen), PCI_DMA_TODEVICE);
|
maplen), PCI_DMA_TODEVICE);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1348,8 +1348,8 @@ static int ql_map_send(struct ql_adapter *qdev,
|
|||||||
|
|
||||||
tbd->len = cpu_to_le32(len);
|
tbd->len = cpu_to_le32(len);
|
||||||
tbd->addr = cpu_to_le64(map);
|
tbd->addr = cpu_to_le64(map);
|
||||||
pci_unmap_addr_set(&tx_ring_desc->map[map_idx], mapaddr, map);
|
dma_unmap_addr_set(&tx_ring_desc->map[map_idx], mapaddr, map);
|
||||||
pci_unmap_len_set(&tx_ring_desc->map[map_idx], maplen, len);
|
dma_unmap_len_set(&tx_ring_desc->map[map_idx], maplen, len);
|
||||||
map_idx++;
|
map_idx++;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@@ -1402,9 +1402,9 @@ static int ql_map_send(struct ql_adapter *qdev,
|
|||||||
tbd->len =
|
tbd->len =
|
||||||
cpu_to_le32((sizeof(struct tx_buf_desc) *
|
cpu_to_le32((sizeof(struct tx_buf_desc) *
|
||||||
(frag_cnt - frag_idx)) | TX_DESC_C);
|
(frag_cnt - frag_idx)) | TX_DESC_C);
|
||||||
pci_unmap_addr_set(&tx_ring_desc->map[map_idx], mapaddr,
|
dma_unmap_addr_set(&tx_ring_desc->map[map_idx], mapaddr,
|
||||||
map);
|
map);
|
||||||
pci_unmap_len_set(&tx_ring_desc->map[map_idx], maplen,
|
dma_unmap_len_set(&tx_ring_desc->map[map_idx], maplen,
|
||||||
sizeof(struct oal));
|
sizeof(struct oal));
|
||||||
tbd = (struct tx_buf_desc *)&tx_ring_desc->oal;
|
tbd = (struct tx_buf_desc *)&tx_ring_desc->oal;
|
||||||
map_idx++;
|
map_idx++;
|
||||||
@@ -1425,8 +1425,8 @@ static int ql_map_send(struct ql_adapter *qdev,
|
|||||||
|
|
||||||
tbd->addr = cpu_to_le64(map);
|
tbd->addr = cpu_to_le64(map);
|
||||||
tbd->len = cpu_to_le32(frag->size);
|
tbd->len = cpu_to_le32(frag->size);
|
||||||
pci_unmap_addr_set(&tx_ring_desc->map[map_idx], mapaddr, map);
|
dma_unmap_addr_set(&tx_ring_desc->map[map_idx], mapaddr, map);
|
||||||
pci_unmap_len_set(&tx_ring_desc->map[map_idx], maplen,
|
dma_unmap_len_set(&tx_ring_desc->map[map_idx], maplen,
|
||||||
frag->size);
|
frag->size);
|
||||||
|
|
||||||
}
|
}
|
||||||
@@ -1742,8 +1742,8 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
|
|||||||
*/
|
*/
|
||||||
sbq_desc = ql_get_curr_sbuf(rx_ring);
|
sbq_desc = ql_get_curr_sbuf(rx_ring);
|
||||||
pci_unmap_single(qdev->pdev,
|
pci_unmap_single(qdev->pdev,
|
||||||
pci_unmap_addr(sbq_desc, mapaddr),
|
dma_unmap_addr(sbq_desc, mapaddr),
|
||||||
pci_unmap_len(sbq_desc, maplen),
|
dma_unmap_len(sbq_desc, maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
skb = sbq_desc->p.skb;
|
skb = sbq_desc->p.skb;
|
||||||
ql_realign_skb(skb, hdr_len);
|
ql_realign_skb(skb, hdr_len);
|
||||||
@@ -1774,18 +1774,18 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
|
|||||||
*/
|
*/
|
||||||
sbq_desc = ql_get_curr_sbuf(rx_ring);
|
sbq_desc = ql_get_curr_sbuf(rx_ring);
|
||||||
pci_dma_sync_single_for_cpu(qdev->pdev,
|
pci_dma_sync_single_for_cpu(qdev->pdev,
|
||||||
pci_unmap_addr
|
dma_unmap_addr
|
||||||
(sbq_desc, mapaddr),
|
(sbq_desc, mapaddr),
|
||||||
pci_unmap_len
|
dma_unmap_len
|
||||||
(sbq_desc, maplen),
|
(sbq_desc, maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
memcpy(skb_put(skb, length),
|
memcpy(skb_put(skb, length),
|
||||||
sbq_desc->p.skb->data, length);
|
sbq_desc->p.skb->data, length);
|
||||||
pci_dma_sync_single_for_device(qdev->pdev,
|
pci_dma_sync_single_for_device(qdev->pdev,
|
||||||
pci_unmap_addr
|
dma_unmap_addr
|
||||||
(sbq_desc,
|
(sbq_desc,
|
||||||
mapaddr),
|
mapaddr),
|
||||||
pci_unmap_len
|
dma_unmap_len
|
||||||
(sbq_desc,
|
(sbq_desc,
|
||||||
maplen),
|
maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
@@ -1798,9 +1798,9 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
|
|||||||
ql_realign_skb(skb, length);
|
ql_realign_skb(skb, length);
|
||||||
skb_put(skb, length);
|
skb_put(skb, length);
|
||||||
pci_unmap_single(qdev->pdev,
|
pci_unmap_single(qdev->pdev,
|
||||||
pci_unmap_addr(sbq_desc,
|
dma_unmap_addr(sbq_desc,
|
||||||
mapaddr),
|
mapaddr),
|
||||||
pci_unmap_len(sbq_desc,
|
dma_unmap_len(sbq_desc,
|
||||||
maplen),
|
maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
sbq_desc->p.skb = NULL;
|
sbq_desc->p.skb = NULL;
|
||||||
@@ -1839,9 +1839,9 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
pci_unmap_page(qdev->pdev,
|
pci_unmap_page(qdev->pdev,
|
||||||
pci_unmap_addr(lbq_desc,
|
dma_unmap_addr(lbq_desc,
|
||||||
mapaddr),
|
mapaddr),
|
||||||
pci_unmap_len(lbq_desc, maplen),
|
dma_unmap_len(lbq_desc, maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
skb_reserve(skb, NET_IP_ALIGN);
|
skb_reserve(skb, NET_IP_ALIGN);
|
||||||
netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
|
netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
|
||||||
@@ -1874,8 +1874,8 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
|
|||||||
int size, i = 0;
|
int size, i = 0;
|
||||||
sbq_desc = ql_get_curr_sbuf(rx_ring);
|
sbq_desc = ql_get_curr_sbuf(rx_ring);
|
||||||
pci_unmap_single(qdev->pdev,
|
pci_unmap_single(qdev->pdev,
|
||||||
pci_unmap_addr(sbq_desc, mapaddr),
|
dma_unmap_addr(sbq_desc, mapaddr),
|
||||||
pci_unmap_len(sbq_desc, maplen),
|
dma_unmap_len(sbq_desc, maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
if (!(ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HS)) {
|
if (!(ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HS)) {
|
||||||
/*
|
/*
|
||||||
@@ -2737,8 +2737,8 @@ static void ql_free_sbq_buffers(struct ql_adapter *qdev, struct rx_ring *rx_ring
|
|||||||
}
|
}
|
||||||
if (sbq_desc->p.skb) {
|
if (sbq_desc->p.skb) {
|
||||||
pci_unmap_single(qdev->pdev,
|
pci_unmap_single(qdev->pdev,
|
||||||
pci_unmap_addr(sbq_desc, mapaddr),
|
dma_unmap_addr(sbq_desc, mapaddr),
|
||||||
pci_unmap_len(sbq_desc, maplen),
|
dma_unmap_len(sbq_desc, maplen),
|
||||||
PCI_DMA_FROMDEVICE);
|
PCI_DMA_FROMDEVICE);
|
||||||
dev_kfree_skb(sbq_desc->p.skb);
|
dev_kfree_skb(sbq_desc->p.skb);
|
||||||
sbq_desc->p.skb = NULL;
|
sbq_desc->p.skb = NULL;
|
||||||
|
Reference in New Issue
Block a user