Commit e9031f2d authored by Tirthendu Sarkar's avatar Tirthendu Sarkar Committed by Tony Nguyen

i40e: introduce next_to_process to i40e_ring

Add a new field called next_to_process in the i40e_ring that is
advanced for every buffer and change the semantics of next_to_clean to
point to the first buffer of a packet. Driver will use next_to_process
in the same way next_to_clean was used previously.

For the non multi-buffer case, next_to_process and next_to_clean will
always be the same since each packet consists of a single buffer.
Signed-off-by: default avatarTirthendu Sarkar <tirthendu.sarkar@intel.com>
Tested-by: Chandan Kumar Rout <chandanx.rout@intel.com> (A Contingent Worker at Intel)
Signed-off-by: default avatarTony Nguyen <anthony.l.nguyen@intel.com>
parent 2bc0de9a
...@@ -1524,6 +1524,7 @@ void i40e_clean_rx_ring(struct i40e_ring *rx_ring) ...@@ -1524,6 +1524,7 @@ void i40e_clean_rx_ring(struct i40e_ring *rx_ring)
rx_ring->next_to_alloc = 0; rx_ring->next_to_alloc = 0;
rx_ring->next_to_clean = 0; rx_ring->next_to_clean = 0;
rx_ring->next_to_process = 0;
rx_ring->next_to_use = 0; rx_ring->next_to_use = 0;
} }
...@@ -1576,6 +1577,7 @@ int i40e_setup_rx_descriptors(struct i40e_ring *rx_ring) ...@@ -1576,6 +1577,7 @@ int i40e_setup_rx_descriptors(struct i40e_ring *rx_ring)
rx_ring->next_to_alloc = 0; rx_ring->next_to_alloc = 0;
rx_ring->next_to_clean = 0; rx_ring->next_to_clean = 0;
rx_ring->next_to_process = 0;
rx_ring->next_to_use = 0; rx_ring->next_to_use = 0;
/* XDP RX-queue info only needed for RX rings exposed to XDP */ /* XDP RX-queue info only needed for RX rings exposed to XDP */
...@@ -2076,7 +2078,7 @@ static struct i40e_rx_buffer *i40e_get_rx_buffer(struct i40e_ring *rx_ring, ...@@ -2076,7 +2078,7 @@ static struct i40e_rx_buffer *i40e_get_rx_buffer(struct i40e_ring *rx_ring,
{ {
struct i40e_rx_buffer *rx_buffer; struct i40e_rx_buffer *rx_buffer;
rx_buffer = i40e_rx_bi(rx_ring, rx_ring->next_to_clean); rx_buffer = i40e_rx_bi(rx_ring, rx_ring->next_to_process);
rx_buffer->page_count = rx_buffer->page_count =
#if (PAGE_SIZE < 8192) #if (PAGE_SIZE < 8192)
page_count(rx_buffer->page); page_count(rx_buffer->page);
...@@ -2375,16 +2377,16 @@ void i40e_finalize_xdp_rx(struct i40e_ring *rx_ring, unsigned int xdp_res) ...@@ -2375,16 +2377,16 @@ void i40e_finalize_xdp_rx(struct i40e_ring *rx_ring, unsigned int xdp_res)
} }
/** /**
* i40e_inc_ntc: Advance the next_to_clean index * i40e_inc_ntp: Advance the next_to_process index
* @rx_ring: Rx ring * @rx_ring: Rx ring
**/ **/
static void i40e_inc_ntc(struct i40e_ring *rx_ring) static void i40e_inc_ntp(struct i40e_ring *rx_ring)
{ {
u32 ntc = rx_ring->next_to_clean + 1; u32 ntp = rx_ring->next_to_process + 1;
ntc = (ntc < rx_ring->count) ? ntc : 0; ntp = (ntp < rx_ring->count) ? ntp : 0;
rx_ring->next_to_clean = ntc; rx_ring->next_to_process = ntp;
prefetch(I40E_RX_DESC(rx_ring, ntc)); prefetch(I40E_RX_DESC(rx_ring, ntp));
} }
/** /**
...@@ -2421,6 +2423,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget, ...@@ -2421,6 +2423,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget,
xdp_prog = READ_ONCE(rx_ring->xdp_prog); xdp_prog = READ_ONCE(rx_ring->xdp_prog);
while (likely(total_rx_packets < (unsigned int)budget)) { while (likely(total_rx_packets < (unsigned int)budget)) {
u16 ntp = rx_ring->next_to_process;
struct i40e_rx_buffer *rx_buffer; struct i40e_rx_buffer *rx_buffer;
union i40e_rx_desc *rx_desc; union i40e_rx_desc *rx_desc;
unsigned int size; unsigned int size;
...@@ -2433,7 +2436,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget, ...@@ -2433,7 +2436,7 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget,
cleaned_count = 0; cleaned_count = 0;
} }
rx_desc = I40E_RX_DESC(rx_ring, rx_ring->next_to_clean); rx_desc = I40E_RX_DESC(rx_ring, ntp);
/* status_error_len will always be zero for unused descriptors /* status_error_len will always be zero for unused descriptors
* because it's cleared in cleanup, and overlaps with hdr_addr * because it's cleared in cleanup, and overlaps with hdr_addr
...@@ -2452,8 +2455,8 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget, ...@@ -2452,8 +2455,8 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget,
i40e_clean_programming_status(rx_ring, i40e_clean_programming_status(rx_ring,
rx_desc->raw.qword[0], rx_desc->raw.qword[0],
qword); qword);
rx_buffer = i40e_rx_bi(rx_ring, rx_ring->next_to_clean); rx_buffer = i40e_rx_bi(rx_ring, ntp);
i40e_inc_ntc(rx_ring); i40e_inc_ntp(rx_ring);
i40e_reuse_rx_page(rx_ring, rx_buffer); i40e_reuse_rx_page(rx_ring, rx_buffer);
cleaned_count++; cleaned_count++;
continue; continue;
...@@ -2509,7 +2512,8 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget, ...@@ -2509,7 +2512,8 @@ static int i40e_clean_rx_irq(struct i40e_ring *rx_ring, int budget,
i40e_put_rx_buffer(rx_ring, rx_buffer); i40e_put_rx_buffer(rx_ring, rx_buffer);
cleaned_count++; cleaned_count++;
i40e_inc_ntc(rx_ring); i40e_inc_ntp(rx_ring);
rx_ring->next_to_clean = rx_ring->next_to_process;
if (i40e_is_non_eop(rx_ring, rx_desc)) if (i40e_is_non_eop(rx_ring, rx_desc))
continue; continue;
......
...@@ -337,6 +337,10 @@ struct i40e_ring { ...@@ -337,6 +337,10 @@ struct i40e_ring {
u8 dcb_tc; /* Traffic class of ring */ u8 dcb_tc; /* Traffic class of ring */
u8 __iomem *tail; u8 __iomem *tail;
/* Next descriptor to be processed; next_to_clean is updated only on
* processing EOP descriptor
*/
u16 next_to_process;
/* high bit set means dynamic, use accessor routines to read/write. /* high bit set means dynamic, use accessor routines to read/write.
* hardware only supports 2us resolution for the ITR registers. * hardware only supports 2us resolution for the ITR registers.
* these values always store the USER setting, and must be converted * these values always store the USER setting, and must be converted
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment