Commit 85e8d004 authored by Alexander Duyck's avatar Alexander Duyck Committed by David S. Miller

igb: transition driver to only using advanced descriptors

Currently the driver uses advanced descriptors for its main functionality,
but then uses legacy when testing.  This patch changes this so that
advanced descriptors are used throughout and all mentions of legacy
descriptors are removed.
Signed-off-by: default avatarAlexander Duyck <alexander.h.duyck@intel.com>
Acked-by: default avatarPeter P Waskiewicz Jr <peter.p.waskiewicz.jr@intel.com>
Signed-off-by: default avatarJeff Kirsher <jeffrey.t.kirsher@intel.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent cbd347ad
...@@ -144,144 +144,6 @@ enum e1000_fc_type { ...@@ -144,144 +144,6 @@ enum e1000_fc_type {
e1000_fc_default = 0xFF e1000_fc_default = 0xFF
}; };
/* Receive Descriptor */
struct e1000_rx_desc {
__le64 buffer_addr; /* Address of the descriptor's data buffer */
__le16 length; /* Length of data DMAed into data buffer */
__le16 csum; /* Packet checksum */
u8 status; /* Descriptor status */
u8 errors; /* Descriptor Errors */
__le16 special;
};
/* Receive Descriptor - Extended */
union e1000_rx_desc_extended {
struct {
__le64 buffer_addr;
__le64 reserved;
} read;
struct {
struct {
__le32 mrq; /* Multiple Rx Queues */
union {
__le32 rss; /* RSS Hash */
struct {
__le16 ip_id; /* IP id */
__le16 csum; /* Packet Checksum */
} csum_ip;
} hi_dword;
} lower;
struct {
__le32 status_error; /* ext status/error */
__le16 length;
__le16 vlan; /* VLAN tag */
} upper;
} wb; /* writeback */
};
#define MAX_PS_BUFFERS 4
/* Receive Descriptor - Packet Split */
union e1000_rx_desc_packet_split {
struct {
/* one buffer for protocol header(s), three data buffers */
__le64 buffer_addr[MAX_PS_BUFFERS];
} read;
struct {
struct {
__le32 mrq; /* Multiple Rx Queues */
union {
__le32 rss; /* RSS Hash */
struct {
__le16 ip_id; /* IP id */
__le16 csum; /* Packet Checksum */
} csum_ip;
} hi_dword;
} lower;
struct {
__le32 status_error; /* ext status/error */
__le16 length0; /* length of buffer 0 */
__le16 vlan; /* VLAN tag */
} middle;
struct {
__le16 header_status;
__le16 length[3]; /* length of buffers 1-3 */
} upper;
__le64 reserved;
} wb; /* writeback */
};
/* Transmit Descriptor */
struct e1000_tx_desc {
__le64 buffer_addr; /* Address of the descriptor's data buffer */
union {
__le32 data;
struct {
__le16 length; /* Data buffer length */
u8 cso; /* Checksum offset */
u8 cmd; /* Descriptor control */
} flags;
} lower;
union {
__le32 data;
struct {
u8 status; /* Descriptor status */
u8 css; /* Checksum start */
__le16 special;
} fields;
} upper;
};
/* Offload Context Descriptor */
struct e1000_context_desc {
union {
__le32 ip_config;
struct {
u8 ipcss; /* IP checksum start */
u8 ipcso; /* IP checksum offset */
__le16 ipcse; /* IP checksum end */
} ip_fields;
} lower_setup;
union {
__le32 tcp_config;
struct {
u8 tucss; /* TCP checksum start */
u8 tucso; /* TCP checksum offset */
__le16 tucse; /* TCP checksum end */
} tcp_fields;
} upper_setup;
__le32 cmd_and_length;
union {
__le32 data;
struct {
u8 status; /* Descriptor status */
u8 hdr_len; /* Header length */
__le16 mss; /* Maximum segment size */
} fields;
} tcp_seg_setup;
};
/* Offload data descriptor */
struct e1000_data_desc {
__le64 buffer_addr; /* Address of the descriptor's buffer address */
union {
__le32 data;
struct {
__le16 length; /* Data buffer length */
u8 typ_len_ext;
u8 cmd;
} flags;
} lower;
union {
__le32 data;
struct {
u8 status; /* Descriptor status */
u8 popts; /* Packet Options */
__le16 special;
} fields;
} upper;
};
/* Statistics counters collected by the MAC */ /* Statistics counters collected by the MAC */
struct e1000_hw_stats { struct e1000_hw_stats {
u64 crcerrs; u64 crcerrs;
......
...@@ -180,9 +180,6 @@ struct igb_ring { ...@@ -180,9 +180,6 @@ struct igb_ring {
(&(((union e1000_adv_tx_desc *)((R).desc))[i])) (&(((union e1000_adv_tx_desc *)((R).desc))[i]))
#define E1000_TX_CTXTDESC_ADV(R, i) \ #define E1000_TX_CTXTDESC_ADV(R, i) \
(&(((struct e1000_adv_tx_context_desc *)((R).desc))[i])) (&(((struct e1000_adv_tx_context_desc *)((R).desc))[i]))
#define E1000_GET_DESC(R, i, type) (&(((struct type *)((R).desc))[i]))
#define E1000_TX_DESC(R, i) E1000_GET_DESC(R, i, e1000_tx_desc)
#define E1000_RX_DESC(R, i) E1000_GET_DESC(R, i, e1000_rx_desc)
/* board specific private data structure */ /* board specific private data structure */
......
...@@ -1272,6 +1272,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1272,6 +1272,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
struct igb_ring *tx_ring = &adapter->test_tx_ring; struct igb_ring *tx_ring = &adapter->test_tx_ring;
struct igb_ring *rx_ring = &adapter->test_rx_ring; struct igb_ring *rx_ring = &adapter->test_rx_ring;
struct pci_dev *pdev = adapter->pdev; struct pci_dev *pdev = adapter->pdev;
struct igb_buffer *buffer_info;
u32 rctl; u32 rctl;
int i, ret_val; int i, ret_val;
...@@ -1288,7 +1289,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1288,7 +1289,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
goto err_nomem; goto err_nomem;
} }
tx_ring->size = tx_ring->count * sizeof(struct e1000_tx_desc); tx_ring->size = tx_ring->count * sizeof(union e1000_adv_tx_desc);
tx_ring->size = ALIGN(tx_ring->size, 4096); tx_ring->size = ALIGN(tx_ring->size, 4096);
tx_ring->desc = pci_alloc_consistent(pdev, tx_ring->size, tx_ring->desc = pci_alloc_consistent(pdev, tx_ring->size,
&tx_ring->dma); &tx_ring->dma);
...@@ -1302,7 +1303,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1302,7 +1303,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
((u64) tx_ring->dma & 0x00000000FFFFFFFF)); ((u64) tx_ring->dma & 0x00000000FFFFFFFF));
wr32(E1000_TDBAH(0), ((u64) tx_ring->dma >> 32)); wr32(E1000_TDBAH(0), ((u64) tx_ring->dma >> 32));
wr32(E1000_TDLEN(0), wr32(E1000_TDLEN(0),
tx_ring->count * sizeof(struct e1000_tx_desc)); tx_ring->count * sizeof(union e1000_adv_tx_desc));
wr32(E1000_TDH(0), 0); wr32(E1000_TDH(0), 0);
wr32(E1000_TDT(0), 0); wr32(E1000_TDT(0), 0);
wr32(E1000_TCTL, wr32(E1000_TCTL,
...@@ -1311,27 +1312,31 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1311,27 +1312,31 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
E1000_COLLISION_DISTANCE << E1000_COLD_SHIFT); E1000_COLLISION_DISTANCE << E1000_COLD_SHIFT);
for (i = 0; i < tx_ring->count; i++) { for (i = 0; i < tx_ring->count; i++) {
struct e1000_tx_desc *tx_desc = E1000_TX_DESC(*tx_ring, i); union e1000_adv_tx_desc *tx_desc;
struct sk_buff *skb; struct sk_buff *skb;
unsigned int size = 1024; unsigned int size = 1024;
tx_desc = E1000_TX_DESC_ADV(*tx_ring, i);
skb = alloc_skb(size, GFP_KERNEL); skb = alloc_skb(size, GFP_KERNEL);
if (!skb) { if (!skb) {
ret_val = 3; ret_val = 3;
goto err_nomem; goto err_nomem;
} }
skb_put(skb, size); skb_put(skb, size);
tx_ring->buffer_info[i].skb = skb; buffer_info = &tx_ring->buffer_info[i];
tx_ring->buffer_info[i].length = skb->len; buffer_info->skb = skb;
tx_ring->buffer_info[i].dma = buffer_info->length = skb->len;
pci_map_single(pdev, skb->data, skb->len, buffer_info->dma = pci_map_single(pdev, skb->data, skb->len,
PCI_DMA_TODEVICE); PCI_DMA_TODEVICE);
tx_desc->buffer_addr = cpu_to_le64(tx_ring->buffer_info[i].dma); tx_desc->read.buffer_addr = cpu_to_le64(buffer_info->dma);
tx_desc->lower.data = cpu_to_le32(skb->len); tx_desc->read.olinfo_status = cpu_to_le32(skb->len) <<
tx_desc->lower.data |= cpu_to_le32(E1000_TXD_CMD_EOP | E1000_ADVTXD_PAYLEN_SHIFT;
tx_desc->read.cmd_type_len = cpu_to_le32(skb->len);
tx_desc->read.cmd_type_len |= cpu_to_le32(E1000_TXD_CMD_EOP |
E1000_TXD_CMD_IFCS | E1000_TXD_CMD_IFCS |
E1000_TXD_CMD_RS); E1000_TXD_CMD_RS |
tx_desc->upper.data = 0; E1000_ADVTXD_DTYP_DATA |
E1000_ADVTXD_DCMD_DEXT);
} }
/* Setup Rx descriptor ring and Rx buffers */ /* Setup Rx descriptor ring and Rx buffers */
...@@ -1347,7 +1352,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1347,7 +1352,7 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
goto err_nomem; goto err_nomem;
} }
rx_ring->size = rx_ring->count * sizeof(struct e1000_rx_desc); rx_ring->size = rx_ring->count * sizeof(union e1000_adv_rx_desc);
rx_ring->desc = pci_alloc_consistent(pdev, rx_ring->size, rx_ring->desc = pci_alloc_consistent(pdev, rx_ring->size,
&rx_ring->dma); &rx_ring->dma);
if (!rx_ring->desc) { if (!rx_ring->desc) {
...@@ -1369,12 +1374,14 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1369,12 +1374,14 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
rctl = E1000_RCTL_EN | E1000_RCTL_BAM | E1000_RCTL_RDMTS_HALF | rctl = E1000_RCTL_EN | E1000_RCTL_BAM | E1000_RCTL_RDMTS_HALF |
(adapter->hw.mac.mc_filter_type << E1000_RCTL_MO_SHIFT); (adapter->hw.mac.mc_filter_type << E1000_RCTL_MO_SHIFT);
wr32(E1000_RCTL, rctl); wr32(E1000_RCTL, rctl);
wr32(E1000_SRRCTL(0), 0); wr32(E1000_SRRCTL(0), E1000_SRRCTL_DESCTYPE_ADV_ONEBUF);
for (i = 0; i < rx_ring->count; i++) { for (i = 0; i < rx_ring->count; i++) {
struct e1000_rx_desc *rx_desc = E1000_RX_DESC(*rx_ring, i); union e1000_adv_rx_desc *rx_desc;
struct sk_buff *skb; struct sk_buff *skb;
buffer_info = &rx_ring->buffer_info[i];
rx_desc = E1000_RX_DESC_ADV(*rx_ring, i);
skb = alloc_skb(IGB_RXBUFFER_2048 + NET_IP_ALIGN, skb = alloc_skb(IGB_RXBUFFER_2048 + NET_IP_ALIGN,
GFP_KERNEL); GFP_KERNEL);
if (!skb) { if (!skb) {
...@@ -1382,11 +1389,11 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter) ...@@ -1382,11 +1389,11 @@ static int igb_setup_desc_rings(struct igb_adapter *adapter)
goto err_nomem; goto err_nomem;
} }
skb_reserve(skb, NET_IP_ALIGN); skb_reserve(skb, NET_IP_ALIGN);
rx_ring->buffer_info[i].skb = skb; buffer_info->skb = skb;
rx_ring->buffer_info[i].dma = buffer_info->dma = pci_map_single(pdev, skb->data,
pci_map_single(pdev, skb->data, IGB_RXBUFFER_2048, IGB_RXBUFFER_2048,
PCI_DMA_FROMDEVICE); PCI_DMA_FROMDEVICE);
rx_desc->buffer_addr = cpu_to_le64(rx_ring->buffer_info[i].dma); rx_desc->read.pkt_addr = cpu_to_le64(buffer_info->dma);
memset(skb->data, 0x00, skb->len); memset(skb->data, 0x00, skb->len);
} }
......
...@@ -994,7 +994,7 @@ void igb_reset(struct igb_adapter *adapter) ...@@ -994,7 +994,7 @@ void igb_reset(struct igb_adapter *adapter)
/* the tx fifo also stores 16 bytes of information about the tx /* the tx fifo also stores 16 bytes of information about the tx
* but don't include ethernet FCS because hardware appends it */ * but don't include ethernet FCS because hardware appends it */
min_tx_space = (adapter->max_frame_size + min_tx_space = (adapter->max_frame_size +
sizeof(struct e1000_tx_desc) - sizeof(union e1000_adv_tx_desc) -
ETH_FCS_LEN) * 2; ETH_FCS_LEN) * 2;
min_tx_space = ALIGN(min_tx_space, 1024); min_tx_space = ALIGN(min_tx_space, 1024);
min_tx_space >>= 10; min_tx_space >>= 10;
...@@ -1704,7 +1704,7 @@ int igb_setup_tx_resources(struct igb_adapter *adapter, ...@@ -1704,7 +1704,7 @@ int igb_setup_tx_resources(struct igb_adapter *adapter,
memset(tx_ring->buffer_info, 0, size); memset(tx_ring->buffer_info, 0, size);
/* round up to nearest 4K */ /* round up to nearest 4K */
tx_ring->size = tx_ring->count * sizeof(struct e1000_tx_desc); tx_ring->size = tx_ring->count * sizeof(union e1000_adv_tx_desc);
tx_ring->size = ALIGN(tx_ring->size, 4096); tx_ring->size = ALIGN(tx_ring->size, 4096);
tx_ring->desc = pci_alloc_consistent(pdev, tx_ring->size, tx_ring->desc = pci_alloc_consistent(pdev, tx_ring->size,
...@@ -1773,7 +1773,7 @@ static void igb_configure_tx(struct igb_adapter *adapter) ...@@ -1773,7 +1773,7 @@ static void igb_configure_tx(struct igb_adapter *adapter)
struct igb_ring *ring = &adapter->tx_ring[i]; struct igb_ring *ring = &adapter->tx_ring[i];
j = ring->reg_idx; j = ring->reg_idx;
wr32(E1000_TDLEN(j), wr32(E1000_TDLEN(j),
ring->count * sizeof(struct e1000_tx_desc)); ring->count * sizeof(union e1000_adv_tx_desc));
tdba = ring->dma; tdba = ring->dma;
wr32(E1000_TDBAL(j), wr32(E1000_TDBAL(j),
tdba & 0x00000000ffffffffULL); tdba & 0x00000000ffffffffULL);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment