Commit 65f6ecc9 authored by Hariprasad Shenai's avatar Hariprasad Shenai Committed by David S. Miller

cxgb4vf: Move fl_starv_thres into adapter->sge data structure

Move fl_starv_thres into adapter->sge data structure since it
_could_ be different from adapter to adapter.  Also move other per-adapter
SGE values which had been treated as driver globals into adapter->sge.

Based on original work by Casey Leedom <leedom@chelsio.com>
Signed-off-by: default avatarHariprasad Shenai <hariprasad@chelsio.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 436c2a50
...@@ -299,6 +299,14 @@ struct sge { ...@@ -299,6 +299,14 @@ struct sge {
u16 timer_val[SGE_NTIMERS]; /* interrupt holdoff timer array */ u16 timer_val[SGE_NTIMERS]; /* interrupt holdoff timer array */
u8 counter_val[SGE_NCOUNTERS]; /* interrupt RX threshold array */ u8 counter_val[SGE_NCOUNTERS]; /* interrupt RX threshold array */
/* Decoded Adapter Parameters.
*/
u32 fl_pg_order; /* large page allocation size */
u32 stat_len; /* length of status page at ring end */
u32 pktshift; /* padding between CPL & packet data */
u32 fl_align; /* response queue message alignment */
u32 fl_starve_thres; /* Free List starvation threshold */
/* /*
* Reverse maps from Absolute Queue IDs to associated queue pointers. * Reverse maps from Absolute Queue IDs to associated queue pointers.
* The absolute Queue IDs are in a compact range which start at a * The absolute Queue IDs are in a compact range which start at a
......
...@@ -50,14 +50,6 @@ ...@@ -50,14 +50,6 @@
#include "../cxgb4/t4fw_api.h" #include "../cxgb4/t4fw_api.h"
#include "../cxgb4/t4_msg.h" #include "../cxgb4/t4_msg.h"
/*
* Decoded Adapter Parameters.
*/
static u32 FL_PG_ORDER; /* large page allocation size */
static u32 STAT_LEN; /* length of status page at ring end */
static u32 PKTSHIFT; /* padding between CPL and packet data */
static u32 FL_ALIGN; /* response queue message alignment */
/* /*
* Constants ... * Constants ...
*/ */
...@@ -264,15 +256,19 @@ static inline unsigned int fl_cap(const struct sge_fl *fl) ...@@ -264,15 +256,19 @@ static inline unsigned int fl_cap(const struct sge_fl *fl)
/** /**
* fl_starving - return whether a Free List is starving. * fl_starving - return whether a Free List is starving.
* @adapter: pointer to the adapter
* @fl: the Free List * @fl: the Free List
* *
* Tests specified Free List to see whether the number of buffers * Tests specified Free List to see whether the number of buffers
* available to the hardware has falled below our "starvation" * available to the hardware has falled below our "starvation"
* threshold. * threshold.
*/ */
static inline bool fl_starving(const struct sge_fl *fl) static inline bool fl_starving(const struct adapter *adapter,
const struct sge_fl *fl)
{ {
return fl->avail - fl->pend_cred <= FL_STARVE_THRES; const struct sge *s = &adapter->sge;
return fl->avail - fl->pend_cred <= s->fl_starve_thres;
} }
/** /**
...@@ -457,13 +453,16 @@ static inline void reclaim_completed_tx(struct adapter *adapter, ...@@ -457,13 +453,16 @@ static inline void reclaim_completed_tx(struct adapter *adapter,
/** /**
* get_buf_size - return the size of an RX Free List buffer. * get_buf_size - return the size of an RX Free List buffer.
* @adapter: pointer to the associated adapter
* @sdesc: pointer to the software buffer descriptor * @sdesc: pointer to the software buffer descriptor
*/ */
static inline int get_buf_size(const struct rx_sw_desc *sdesc) static inline int get_buf_size(const struct adapter *adapter,
const struct rx_sw_desc *sdesc)
{ {
return FL_PG_ORDER > 0 && (sdesc->dma_addr & RX_LARGE_BUF) const struct sge *s = &adapter->sge;
? (PAGE_SIZE << FL_PG_ORDER)
: PAGE_SIZE; return (s->fl_pg_order > 0 && (sdesc->dma_addr & RX_LARGE_BUF)
? (PAGE_SIZE << s->fl_pg_order) : PAGE_SIZE);
} }
/** /**
...@@ -483,7 +482,8 @@ static void free_rx_bufs(struct adapter *adapter, struct sge_fl *fl, int n) ...@@ -483,7 +482,8 @@ static void free_rx_bufs(struct adapter *adapter, struct sge_fl *fl, int n)
if (is_buf_mapped(sdesc)) if (is_buf_mapped(sdesc))
dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc), dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc),
get_buf_size(sdesc), PCI_DMA_FROMDEVICE); get_buf_size(adapter, sdesc),
PCI_DMA_FROMDEVICE);
put_page(sdesc->page); put_page(sdesc->page);
sdesc->page = NULL; sdesc->page = NULL;
if (++fl->cidx == fl->size) if (++fl->cidx == fl->size)
...@@ -511,7 +511,8 @@ static void unmap_rx_buf(struct adapter *adapter, struct sge_fl *fl) ...@@ -511,7 +511,8 @@ static void unmap_rx_buf(struct adapter *adapter, struct sge_fl *fl)
if (is_buf_mapped(sdesc)) if (is_buf_mapped(sdesc))
dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc), dma_unmap_page(adapter->pdev_dev, get_buf_addr(sdesc),
get_buf_size(sdesc), PCI_DMA_FROMDEVICE); get_buf_size(adapter, sdesc),
PCI_DMA_FROMDEVICE);
sdesc->page = NULL; sdesc->page = NULL;
if (++fl->cidx == fl->size) if (++fl->cidx == fl->size)
fl->cidx = 0; fl->cidx = 0;
...@@ -589,6 +590,7 @@ static inline void poison_buf(struct page *page, size_t sz) ...@@ -589,6 +590,7 @@ static inline void poison_buf(struct page *page, size_t sz)
static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl, static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl,
int n, gfp_t gfp) int n, gfp_t gfp)
{ {
struct sge *s = &adapter->sge;
struct page *page; struct page *page;
dma_addr_t dma_addr; dma_addr_t dma_addr;
unsigned int cred = fl->avail; unsigned int cred = fl->avail;
...@@ -608,12 +610,12 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl, ...@@ -608,12 +610,12 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl,
* If we don't support large pages, drop directly into the small page * If we don't support large pages, drop directly into the small page
* allocation code. * allocation code.
*/ */
if (FL_PG_ORDER == 0) if (s->fl_pg_order == 0)
goto alloc_small_pages; goto alloc_small_pages;
while (n) { while (n) {
page = alloc_pages(gfp | __GFP_COMP | __GFP_NOWARN, page = alloc_pages(gfp | __GFP_COMP | __GFP_NOWARN,
FL_PG_ORDER); s->fl_pg_order);
if (unlikely(!page)) { if (unlikely(!page)) {
/* /*
* We've failed inour attempt to allocate a "large * We've failed inour attempt to allocate a "large
...@@ -623,10 +625,10 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl, ...@@ -623,10 +625,10 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl,
fl->large_alloc_failed++; fl->large_alloc_failed++;
break; break;
} }
poison_buf(page, PAGE_SIZE << FL_PG_ORDER); poison_buf(page, PAGE_SIZE << s->fl_pg_order);
dma_addr = dma_map_page(adapter->pdev_dev, page, 0, dma_addr = dma_map_page(adapter->pdev_dev, page, 0,
PAGE_SIZE << FL_PG_ORDER, PAGE_SIZE << s->fl_pg_order,
PCI_DMA_FROMDEVICE); PCI_DMA_FROMDEVICE);
if (unlikely(dma_mapping_error(adapter->pdev_dev, dma_addr))) { if (unlikely(dma_mapping_error(adapter->pdev_dev, dma_addr))) {
/* /*
...@@ -637,7 +639,7 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl, ...@@ -637,7 +639,7 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl,
* because DMA mapping resources are typically * because DMA mapping resources are typically
* critical resources once they become scarse. * critical resources once they become scarse.
*/ */
__free_pages(page, FL_PG_ORDER); __free_pages(page, s->fl_pg_order);
goto out; goto out;
} }
dma_addr |= RX_LARGE_BUF; dma_addr |= RX_LARGE_BUF;
...@@ -693,7 +695,7 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl, ...@@ -693,7 +695,7 @@ static unsigned int refill_fl(struct adapter *adapter, struct sge_fl *fl,
fl->pend_cred += cred; fl->pend_cred += cred;
ring_fl_db(adapter, fl); ring_fl_db(adapter, fl);
if (unlikely(fl_starving(fl))) { if (unlikely(fl_starving(adapter, fl))) {
smp_wmb(); smp_wmb();
set_bit(fl->cntxt_id, adapter->sge.starving_fl); set_bit(fl->cntxt_id, adapter->sge.starving_fl);
} }
...@@ -1468,6 +1470,8 @@ static void t4vf_pktgl_free(const struct pkt_gl *gl) ...@@ -1468,6 +1470,8 @@ static void t4vf_pktgl_free(const struct pkt_gl *gl)
static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl,
const struct cpl_rx_pkt *pkt) const struct cpl_rx_pkt *pkt)
{ {
struct adapter *adapter = rxq->rspq.adapter;
struct sge *s = &adapter->sge;
int ret; int ret;
struct sk_buff *skb; struct sk_buff *skb;
...@@ -1478,8 +1482,8 @@ static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl, ...@@ -1478,8 +1482,8 @@ static void do_gro(struct sge_eth_rxq *rxq, const struct pkt_gl *gl,
return; return;
} }
copy_frags(skb, gl, PKTSHIFT); copy_frags(skb, gl, s->pktshift);
skb->len = gl->tot_len - PKTSHIFT; skb->len = gl->tot_len - s->pktshift;
skb->data_len = skb->len; skb->data_len = skb->len;
skb->truesize += skb->data_len; skb->truesize += skb->data_len;
skb->ip_summed = CHECKSUM_UNNECESSARY; skb->ip_summed = CHECKSUM_UNNECESSARY;
...@@ -1516,6 +1520,8 @@ int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, ...@@ -1516,6 +1520,8 @@ int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp,
bool csum_ok = pkt->csum_calc && !pkt->err_vec && bool csum_ok = pkt->csum_calc && !pkt->err_vec &&
(rspq->netdev->features & NETIF_F_RXCSUM); (rspq->netdev->features & NETIF_F_RXCSUM);
struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq);
struct adapter *adapter = rspq->adapter;
struct sge *s = &adapter->sge;
/* /*
* If this is a good TCP packet and we have Generic Receive Offload * If this is a good TCP packet and we have Generic Receive Offload
...@@ -1537,7 +1543,7 @@ int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp, ...@@ -1537,7 +1543,7 @@ int t4vf_ethrx_handler(struct sge_rspq *rspq, const __be64 *rsp,
rxq->stats.rx_drops++; rxq->stats.rx_drops++;
return 0; return 0;
} }
__skb_pull(skb, PKTSHIFT); __skb_pull(skb, s->pktshift);
skb->protocol = eth_type_trans(skb, rspq->netdev); skb->protocol = eth_type_trans(skb, rspq->netdev);
skb_record_rx_queue(skb, rspq->idx); skb_record_rx_queue(skb, rspq->idx);
rxq->stats.pkts++; rxq->stats.pkts++;
...@@ -1648,6 +1654,8 @@ static inline void rspq_next(struct sge_rspq *rspq) ...@@ -1648,6 +1654,8 @@ static inline void rspq_next(struct sge_rspq *rspq)
static int process_responses(struct sge_rspq *rspq, int budget) static int process_responses(struct sge_rspq *rspq, int budget)
{ {
struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq);
struct adapter *adapter = rspq->adapter;
struct sge *s = &adapter->sge;
int budget_left = budget; int budget_left = budget;
while (likely(budget_left)) { while (likely(budget_left)) {
...@@ -1697,7 +1705,7 @@ static int process_responses(struct sge_rspq *rspq, int budget) ...@@ -1697,7 +1705,7 @@ static int process_responses(struct sge_rspq *rspq, int budget)
BUG_ON(frag >= MAX_SKB_FRAGS); BUG_ON(frag >= MAX_SKB_FRAGS);
BUG_ON(rxq->fl.avail == 0); BUG_ON(rxq->fl.avail == 0);
sdesc = &rxq->fl.sdesc[rxq->fl.cidx]; sdesc = &rxq->fl.sdesc[rxq->fl.cidx];
bufsz = get_buf_size(sdesc); bufsz = get_buf_size(adapter, sdesc);
fp->page = sdesc->page; fp->page = sdesc->page;
fp->offset = rspq->offset; fp->offset = rspq->offset;
fp->size = min(bufsz, len); fp->size = min(bufsz, len);
...@@ -1726,7 +1734,7 @@ static int process_responses(struct sge_rspq *rspq, int budget) ...@@ -1726,7 +1734,7 @@ static int process_responses(struct sge_rspq *rspq, int budget)
*/ */
ret = rspq->handler(rspq, rspq->cur_desc, &gl); ret = rspq->handler(rspq, rspq->cur_desc, &gl);
if (likely(ret == 0)) if (likely(ret == 0))
rspq->offset += ALIGN(fp->size, FL_ALIGN); rspq->offset += ALIGN(fp->size, s->fl_align);
else else
restore_rx_bufs(&gl, &rxq->fl, frag); restore_rx_bufs(&gl, &rxq->fl, frag);
} else if (likely(rsp_type == RSP_TYPE_CPL)) { } else if (likely(rsp_type == RSP_TYPE_CPL)) {
...@@ -1963,7 +1971,7 @@ static void sge_rx_timer_cb(unsigned long data) ...@@ -1963,7 +1971,7 @@ static void sge_rx_timer_cb(unsigned long data)
* schedule napi but the FL is no longer starving. * schedule napi but the FL is no longer starving.
* No biggie. * No biggie.
*/ */
if (fl_starving(fl)) { if (fl_starving(adapter, fl)) {
struct sge_eth_rxq *rxq; struct sge_eth_rxq *rxq;
rxq = container_of(fl, struct sge_eth_rxq, fl); rxq = container_of(fl, struct sge_eth_rxq, fl);
...@@ -2047,6 +2055,7 @@ int t4vf_sge_alloc_rxq(struct adapter *adapter, struct sge_rspq *rspq, ...@@ -2047,6 +2055,7 @@ int t4vf_sge_alloc_rxq(struct adapter *adapter, struct sge_rspq *rspq,
int intr_dest, int intr_dest,
struct sge_fl *fl, rspq_handler_t hnd) struct sge_fl *fl, rspq_handler_t hnd)
{ {
struct sge *s = &adapter->sge;
struct port_info *pi = netdev_priv(dev); struct port_info *pi = netdev_priv(dev);
struct fw_iq_cmd cmd, rpl; struct fw_iq_cmd cmd, rpl;
int ret, iqandst, flsz = 0; int ret, iqandst, flsz = 0;
...@@ -2117,7 +2126,7 @@ int t4vf_sge_alloc_rxq(struct adapter *adapter, struct sge_rspq *rspq, ...@@ -2117,7 +2126,7 @@ int t4vf_sge_alloc_rxq(struct adapter *adapter, struct sge_rspq *rspq,
fl->size = roundup(fl->size, FL_PER_EQ_UNIT); fl->size = roundup(fl->size, FL_PER_EQ_UNIT);
fl->desc = alloc_ring(adapter->pdev_dev, fl->size, fl->desc = alloc_ring(adapter->pdev_dev, fl->size,
sizeof(__be64), sizeof(struct rx_sw_desc), sizeof(__be64), sizeof(struct rx_sw_desc),
&fl->addr, &fl->sdesc, STAT_LEN); &fl->addr, &fl->sdesc, s->stat_len);
if (!fl->desc) { if (!fl->desc) {
ret = -ENOMEM; ret = -ENOMEM;
goto err; goto err;
...@@ -2129,7 +2138,7 @@ int t4vf_sge_alloc_rxq(struct adapter *adapter, struct sge_rspq *rspq, ...@@ -2129,7 +2138,7 @@ int t4vf_sge_alloc_rxq(struct adapter *adapter, struct sge_rspq *rspq,
* free list ring) in Egress Queue Units. * free list ring) in Egress Queue Units.
*/ */
flsz = (fl->size / FL_PER_EQ_UNIT + flsz = (fl->size / FL_PER_EQ_UNIT +
STAT_LEN / EQ_UNIT); s->stat_len / EQ_UNIT);
/* /*
* Fill in all the relevant firmware Ingress Queue Command * Fill in all the relevant firmware Ingress Queue Command
...@@ -2217,6 +2226,7 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, ...@@ -2217,6 +2226,7 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq,
struct net_device *dev, struct netdev_queue *devq, struct net_device *dev, struct netdev_queue *devq,
unsigned int iqid) unsigned int iqid)
{ {
struct sge *s = &adapter->sge;
int ret, nentries; int ret, nentries;
struct fw_eq_eth_cmd cmd, rpl; struct fw_eq_eth_cmd cmd, rpl;
struct port_info *pi = netdev_priv(dev); struct port_info *pi = netdev_priv(dev);
...@@ -2225,7 +2235,7 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, ...@@ -2225,7 +2235,7 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq,
* Calculate the size of the hardware TX Queue (including the Status * Calculate the size of the hardware TX Queue (including the Status
* Page on the end of the TX Queue) in units of TX Descriptors. * Page on the end of the TX Queue) in units of TX Descriptors.
*/ */
nentries = txq->q.size + STAT_LEN / sizeof(struct tx_desc); nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc);
/* /*
* Allocate the hardware ring for the TX ring (with space for its * Allocate the hardware ring for the TX ring (with space for its
...@@ -2234,7 +2244,7 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, ...@@ -2234,7 +2244,7 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq,
txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size, txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size,
sizeof(struct tx_desc), sizeof(struct tx_desc),
sizeof(struct tx_sw_desc), sizeof(struct tx_sw_desc),
&txq->q.phys_addr, &txq->q.sdesc, STAT_LEN); &txq->q.phys_addr, &txq->q.sdesc, s->stat_len);
if (!txq->q.desc) if (!txq->q.desc)
return -ENOMEM; return -ENOMEM;
...@@ -2307,8 +2317,10 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, ...@@ -2307,8 +2317,10 @@ int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq,
*/ */
static void free_txq(struct adapter *adapter, struct sge_txq *tq) static void free_txq(struct adapter *adapter, struct sge_txq *tq)
{ {
struct sge *s = &adapter->sge;
dma_free_coherent(adapter->pdev_dev, dma_free_coherent(adapter->pdev_dev,
tq->size * sizeof(*tq->desc) + STAT_LEN, tq->size * sizeof(*tq->desc) + s->stat_len,
tq->desc, tq->phys_addr); tq->desc, tq->phys_addr);
tq->cntxt_id = 0; tq->cntxt_id = 0;
tq->sdesc = NULL; tq->sdesc = NULL;
...@@ -2322,6 +2334,7 @@ static void free_txq(struct adapter *adapter, struct sge_txq *tq) ...@@ -2322,6 +2334,7 @@ static void free_txq(struct adapter *adapter, struct sge_txq *tq)
static void free_rspq_fl(struct adapter *adapter, struct sge_rspq *rspq, static void free_rspq_fl(struct adapter *adapter, struct sge_rspq *rspq,
struct sge_fl *fl) struct sge_fl *fl)
{ {
struct sge *s = &adapter->sge;
unsigned int flid = fl ? fl->cntxt_id : 0xffff; unsigned int flid = fl ? fl->cntxt_id : 0xffff;
t4vf_iq_free(adapter, FW_IQ_TYPE_FL_INT_CAP, t4vf_iq_free(adapter, FW_IQ_TYPE_FL_INT_CAP,
...@@ -2337,7 +2350,7 @@ static void free_rspq_fl(struct adapter *adapter, struct sge_rspq *rspq, ...@@ -2337,7 +2350,7 @@ static void free_rspq_fl(struct adapter *adapter, struct sge_rspq *rspq,
if (fl) { if (fl) {
free_rx_bufs(adapter, fl, fl->avail); free_rx_bufs(adapter, fl, fl->avail);
dma_free_coherent(adapter->pdev_dev, dma_free_coherent(adapter->pdev_dev,
fl->size * sizeof(*fl->desc) + STAT_LEN, fl->size * sizeof(*fl->desc) + s->stat_len,
fl->desc, fl->addr); fl->desc, fl->addr);
kfree(fl->sdesc); kfree(fl->sdesc);
fl->sdesc = NULL; fl->sdesc = NULL;
...@@ -2443,11 +2456,11 @@ int t4vf_sge_init(struct adapter *adapter) ...@@ -2443,11 +2456,11 @@ int t4vf_sge_init(struct adapter *adapter)
* Now translate the adapter parameters into our internal forms. * Now translate the adapter parameters into our internal forms.
*/ */
if (fl1) if (fl1)
FL_PG_ORDER = ilog2(fl1) - PAGE_SHIFT; s->fl_pg_order = ilog2(fl1) - PAGE_SHIFT;
STAT_LEN = ((sge_params->sge_control & EGRSTATUSPAGESIZE_MASK) s->stat_len = ((sge_params->sge_control & EGRSTATUSPAGESIZE_MASK)
? 128 : 64); ? 128 : 64);
PKTSHIFT = PKTSHIFT_GET(sge_params->sge_control); s->pktshift = PKTSHIFT_GET(sge_params->sge_control);
FL_ALIGN = 1 << (INGPADBOUNDARY_GET(sge_params->sge_control) + s->fl_align = 1 << (INGPADBOUNDARY_GET(sge_params->sge_control) +
SGE_INGPADBOUNDARY_SHIFT); SGE_INGPADBOUNDARY_SHIFT);
/* /*
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment