Commit c9f140eb authored by Lendacky, Thomas's avatar Lendacky, Thomas Committed by David S. Miller

amd-xgbe: Separate Tx/Rx ring data fields into new structs

Move the Tx and Rx related fields within the xgbe_ring_data struct into
their own structs in order to more easily see what fields are used for
each operation.
Signed-off-by: default avatarTom Lendacky <thomas.lendacky@amd.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 7bba35bd
...@@ -335,11 +335,11 @@ static int xgbe_map_rx_buffer(struct xgbe_prv_data *pdata, ...@@ -335,11 +335,11 @@ static int xgbe_map_rx_buffer(struct xgbe_prv_data *pdata,
} }
/* Set up the header page info */ /* Set up the header page info */
xgbe_set_buffer_data(&rdata->rx_hdr, &ring->rx_hdr_pa, xgbe_set_buffer_data(&rdata->rx.hdr, &ring->rx_hdr_pa,
XGBE_SKB_ALLOC_SIZE); XGBE_SKB_ALLOC_SIZE);
/* Set up the buffer page info */ /* Set up the buffer page info */
xgbe_set_buffer_data(&rdata->rx_buf, &ring->rx_buf_pa, xgbe_set_buffer_data(&rdata->rx.buf, &ring->rx_buf_pa,
pdata->rx_buf_size); pdata->rx_buf_size);
return 0; return 0;
...@@ -451,31 +451,29 @@ static void xgbe_unmap_rdata(struct xgbe_prv_data *pdata, ...@@ -451,31 +451,29 @@ static void xgbe_unmap_rdata(struct xgbe_prv_data *pdata,
rdata->skb = NULL; rdata->skb = NULL;
} }
if (rdata->rx_hdr.pa.pages) if (rdata->rx.hdr.pa.pages)
put_page(rdata->rx_hdr.pa.pages); put_page(rdata->rx.hdr.pa.pages);
if (rdata->rx_hdr.pa_unmap.pages) { if (rdata->rx.hdr.pa_unmap.pages) {
dma_unmap_page(pdata->dev, rdata->rx_hdr.pa_unmap.pages_dma, dma_unmap_page(pdata->dev, rdata->rx.hdr.pa_unmap.pages_dma,
rdata->rx_hdr.pa_unmap.pages_len, rdata->rx.hdr.pa_unmap.pages_len,
DMA_FROM_DEVICE); DMA_FROM_DEVICE);
put_page(rdata->rx_hdr.pa_unmap.pages); put_page(rdata->rx.hdr.pa_unmap.pages);
} }
if (rdata->rx_buf.pa.pages) if (rdata->rx.buf.pa.pages)
put_page(rdata->rx_buf.pa.pages); put_page(rdata->rx.buf.pa.pages);
if (rdata->rx_buf.pa_unmap.pages) { if (rdata->rx.buf.pa_unmap.pages) {
dma_unmap_page(pdata->dev, rdata->rx_buf.pa_unmap.pages_dma, dma_unmap_page(pdata->dev, rdata->rx.buf.pa_unmap.pages_dma,
rdata->rx_buf.pa_unmap.pages_len, rdata->rx.buf.pa_unmap.pages_len,
DMA_FROM_DEVICE); DMA_FROM_DEVICE);
put_page(rdata->rx_buf.pa_unmap.pages); put_page(rdata->rx.buf.pa_unmap.pages);
} }
memset(&rdata->rx_hdr, 0, sizeof(rdata->rx_hdr)); memset(&rdata->tx, 0, sizeof(rdata->tx));
memset(&rdata->rx_buf, 0, sizeof(rdata->rx_buf)); memset(&rdata->rx, 0, sizeof(rdata->rx));
rdata->tso_header = 0;
rdata->len = 0;
rdata->interrupt = 0; rdata->interrupt = 0;
rdata->mapped_as_page = 0; rdata->mapped_as_page = 0;
...@@ -534,7 +532,7 @@ static int xgbe_map_tx_skb(struct xgbe_channel *channel, struct sk_buff *skb) ...@@ -534,7 +532,7 @@ static int xgbe_map_tx_skb(struct xgbe_channel *channel, struct sk_buff *skb)
} }
rdata->skb_dma = skb_dma; rdata->skb_dma = skb_dma;
rdata->skb_dma_len = packet->header_len; rdata->skb_dma_len = packet->header_len;
rdata->tso_header = 1; rdata->tx.tso_header = 1;
offset = packet->header_len; offset = packet->header_len;
......
...@@ -1085,10 +1085,10 @@ static void xgbe_rx_desc_reset(struct xgbe_ring_data *rdata) ...@@ -1085,10 +1085,10 @@ static void xgbe_rx_desc_reset(struct xgbe_ring_data *rdata)
* Set buffer 2 (hi) address to buffer dma address (hi) and * Set buffer 2 (hi) address to buffer dma address (hi) and
* set control bits OWN and INTE * set control bits OWN and INTE
*/ */
rdesc->desc0 = cpu_to_le32(lower_32_bits(rdata->rx_hdr.dma)); rdesc->desc0 = cpu_to_le32(lower_32_bits(rdata->rx.hdr.dma));
rdesc->desc1 = cpu_to_le32(upper_32_bits(rdata->rx_hdr.dma)); rdesc->desc1 = cpu_to_le32(upper_32_bits(rdata->rx.hdr.dma));
rdesc->desc2 = cpu_to_le32(lower_32_bits(rdata->rx_buf.dma)); rdesc->desc2 = cpu_to_le32(lower_32_bits(rdata->rx.buf.dma));
rdesc->desc3 = cpu_to_le32(upper_32_bits(rdata->rx_buf.dma)); rdesc->desc3 = cpu_to_le32(upper_32_bits(rdata->rx.buf.dma));
XGMAC_SET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, INTE, XGMAC_SET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, INTE,
rdata->interrupt ? 1 : 0); rdata->interrupt ? 1 : 0);
...@@ -1586,7 +1586,7 @@ static int xgbe_dev_read(struct xgbe_channel *channel) ...@@ -1586,7 +1586,7 @@ static int xgbe_dev_read(struct xgbe_channel *channel)
/* Get the header length */ /* Get the header length */
if (XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, FD)) if (XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, FD))
rdata->hdr_len = XGMAC_GET_BITS_LE(rdesc->desc2, rdata->rx.hdr_len = XGMAC_GET_BITS_LE(rdesc->desc2,
RX_NORMAL_DESC2, HL); RX_NORMAL_DESC2, HL);
/* Get the RSS hash */ /* Get the RSS hash */
...@@ -1610,7 +1610,7 @@ static int xgbe_dev_read(struct xgbe_channel *channel) ...@@ -1610,7 +1610,7 @@ static int xgbe_dev_read(struct xgbe_channel *channel)
} }
/* Get the packet length */ /* Get the packet length */
rdata->len = XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, PL); rdata->rx.len = XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, PL);
if (!XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, LD)) { if (!XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, LD)) {
/* Not all the data has been transferred for this packet */ /* Not all the data has been transferred for this packet */
......
...@@ -1747,14 +1747,14 @@ static struct sk_buff *xgbe_create_skb(struct xgbe_prv_data *pdata, ...@@ -1747,14 +1747,14 @@ static struct sk_buff *xgbe_create_skb(struct xgbe_prv_data *pdata,
u8 *packet; u8 *packet;
unsigned int copy_len; unsigned int copy_len;
skb = netdev_alloc_skb_ip_align(netdev, rdata->rx_hdr.dma_len); skb = netdev_alloc_skb_ip_align(netdev, rdata->rx.hdr.dma_len);
if (!skb) if (!skb)
return NULL; return NULL;
packet = page_address(rdata->rx_hdr.pa.pages) + packet = page_address(rdata->rx.hdr.pa.pages) +
rdata->rx_hdr.pa.pages_offset; rdata->rx.hdr.pa.pages_offset;
copy_len = (rdata->hdr_len) ? rdata->hdr_len : *len; copy_len = (rdata->rx.hdr_len) ? rdata->rx.hdr_len : *len;
copy_len = min(rdata->rx_hdr.dma_len, copy_len); copy_len = min(rdata->rx.hdr.dma_len, copy_len);
skb_copy_to_linear_data(skb, packet, copy_len); skb_copy_to_linear_data(skb, packet, copy_len);
skb_put(skb, copy_len); skb_put(skb, copy_len);
...@@ -1900,13 +1900,13 @@ static int xgbe_rx_poll(struct xgbe_channel *channel, int budget) ...@@ -1900,13 +1900,13 @@ static int xgbe_rx_poll(struct xgbe_channel *channel, int budget)
} }
if (!context) { if (!context) {
put_len = rdata->len - len; put_len = rdata->rx.len - len;
len += put_len; len += put_len;
if (!skb) { if (!skb) {
dma_sync_single_for_cpu(pdata->dev, dma_sync_single_for_cpu(pdata->dev,
rdata->rx_hdr.dma, rdata->rx.hdr.dma,
rdata->rx_hdr.dma_len, rdata->rx.hdr.dma_len,
DMA_FROM_DEVICE); DMA_FROM_DEVICE);
skb = xgbe_create_skb(pdata, rdata, &put_len); skb = xgbe_create_skb(pdata, rdata, &put_len);
...@@ -1918,15 +1918,15 @@ static int xgbe_rx_poll(struct xgbe_channel *channel, int budget) ...@@ -1918,15 +1918,15 @@ static int xgbe_rx_poll(struct xgbe_channel *channel, int budget)
if (put_len) { if (put_len) {
dma_sync_single_for_cpu(pdata->dev, dma_sync_single_for_cpu(pdata->dev,
rdata->rx_buf.dma, rdata->rx.buf.dma,
rdata->rx_buf.dma_len, rdata->rx.buf.dma_len,
DMA_FROM_DEVICE); DMA_FROM_DEVICE);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
rdata->rx_buf.pa.pages, rdata->rx.buf.pa.pages,
rdata->rx_buf.pa.pages_offset, rdata->rx.buf.pa.pages_offset,
put_len, rdata->rx_buf.dma_len); put_len, rdata->rx.buf.dma_len);
rdata->rx_buf.pa.pages = NULL; rdata->rx.buf.pa.pages = NULL;
} }
} }
......
...@@ -271,6 +271,20 @@ struct xgbe_buffer_data { ...@@ -271,6 +271,20 @@ struct xgbe_buffer_data {
unsigned int dma_len; unsigned int dma_len;
}; };
/* Tx-related ring data */
struct xgbe_tx_ring_data {
unsigned int tso_header; /* TSO header indicator */
};
/* Rx-related ring data */
struct xgbe_rx_ring_data {
struct xgbe_buffer_data hdr; /* Header locations */
struct xgbe_buffer_data buf; /* Payload locations */
unsigned short hdr_len; /* Length of received header */
unsigned short len; /* Length of received packet */
};
/* Structure used to hold information related to the descriptor /* Structure used to hold information related to the descriptor
* and the packet associated with the descriptor (always use * and the packet associated with the descriptor (always use
* use the XGBE_GET_DESC_DATA macro to access this data from the ring) * use the XGBE_GET_DESC_DATA macro to access this data from the ring)
...@@ -282,13 +296,9 @@ struct xgbe_ring_data { ...@@ -282,13 +296,9 @@ struct xgbe_ring_data {
struct sk_buff *skb; /* Virtual address of SKB */ struct sk_buff *skb; /* Virtual address of SKB */
dma_addr_t skb_dma; /* DMA address of SKB data */ dma_addr_t skb_dma; /* DMA address of SKB data */
unsigned int skb_dma_len; /* Length of SKB DMA area */ unsigned int skb_dma_len; /* Length of SKB DMA area */
unsigned int tso_header; /* TSO header indicator */
struct xgbe_buffer_data rx_hdr; /* Header locations */ struct xgbe_tx_ring_data tx; /* Tx-related data */
struct xgbe_buffer_data rx_buf; /* Payload locations */ struct xgbe_rx_ring_data rx; /* Rx-related data */
unsigned short hdr_len; /* Length of received header */
unsigned short len; /* Length of received Rx packet */
unsigned int interrupt; /* Interrupt indicator */ unsigned int interrupt; /* Interrupt indicator */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment