Commit 33912e72 authored by Ayaz Abdulla's avatar Ayaz Abdulla Committed by David S. Miller

forcedeth: add/modify tx done with limit

There are two tx_done routines to handle tx completion processing. Both
these functions now take in a limit value and return the amount of tx
completions. This will be used by a future patch to determine the total
amount of work done.
Signed-off-by: default avatarAyaz Abdulla <aabdulla@nvidia.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 2daac3e8
...@@ -2397,14 +2397,16 @@ static inline void nv_tx_flip_ownership(struct net_device *dev) ...@@ -2397,14 +2397,16 @@ static inline void nv_tx_flip_ownership(struct net_device *dev)
* *
* Caller must own np->lock. * Caller must own np->lock.
*/ */
static void nv_tx_done(struct net_device *dev) static int nv_tx_done(struct net_device *dev, int limit)
{ {
struct fe_priv *np = netdev_priv(dev); struct fe_priv *np = netdev_priv(dev);
u32 flags; u32 flags;
int tx_work = 0;
struct ring_desc* orig_get_tx = np->get_tx.orig; struct ring_desc* orig_get_tx = np->get_tx.orig;
while ((np->get_tx.orig != np->put_tx.orig) && while ((np->get_tx.orig != np->put_tx.orig) &&
!((flags = le32_to_cpu(np->get_tx.orig->flaglen)) & NV_TX_VALID)) { !((flags = le32_to_cpu(np->get_tx.orig->flaglen)) & NV_TX_VALID) &&
(tx_work < limit)) {
dprintk(KERN_DEBUG "%s: nv_tx_done: flags 0x%x.\n", dprintk(KERN_DEBUG "%s: nv_tx_done: flags 0x%x.\n",
dev->name, flags); dev->name, flags);
...@@ -2430,6 +2432,7 @@ static void nv_tx_done(struct net_device *dev) ...@@ -2430,6 +2432,7 @@ static void nv_tx_done(struct net_device *dev)
} }
dev_kfree_skb_any(np->get_tx_ctx->skb); dev_kfree_skb_any(np->get_tx_ctx->skb);
np->get_tx_ctx->skb = NULL; np->get_tx_ctx->skb = NULL;
tx_work++;
} }
} else { } else {
if (flags & NV_TX2_LASTPACKET) { if (flags & NV_TX2_LASTPACKET) {
...@@ -2447,6 +2450,7 @@ static void nv_tx_done(struct net_device *dev) ...@@ -2447,6 +2450,7 @@ static void nv_tx_done(struct net_device *dev)
} }
dev_kfree_skb_any(np->get_tx_ctx->skb); dev_kfree_skb_any(np->get_tx_ctx->skb);
np->get_tx_ctx->skb = NULL; np->get_tx_ctx->skb = NULL;
tx_work++;
} }
} }
if (unlikely(np->get_tx.orig++ == np->last_tx.orig)) if (unlikely(np->get_tx.orig++ == np->last_tx.orig))
...@@ -2458,17 +2462,19 @@ static void nv_tx_done(struct net_device *dev) ...@@ -2458,17 +2462,19 @@ static void nv_tx_done(struct net_device *dev)
np->tx_stop = 0; np->tx_stop = 0;
netif_wake_queue(dev); netif_wake_queue(dev);
} }
return tx_work;
} }
static void nv_tx_done_optimized(struct net_device *dev, int limit) static int nv_tx_done_optimized(struct net_device *dev, int limit)
{ {
struct fe_priv *np = netdev_priv(dev); struct fe_priv *np = netdev_priv(dev);
u32 flags; u32 flags;
int tx_work = 0;
struct ring_desc_ex* orig_get_tx = np->get_tx.ex; struct ring_desc_ex* orig_get_tx = np->get_tx.ex;
while ((np->get_tx.ex != np->put_tx.ex) && while ((np->get_tx.ex != np->put_tx.ex) &&
!((flags = le32_to_cpu(np->get_tx.ex->flaglen)) & NV_TX_VALID) && !((flags = le32_to_cpu(np->get_tx.ex->flaglen)) & NV_TX_VALID) &&
(limit-- > 0)) { (tx_work < limit)) {
dprintk(KERN_DEBUG "%s: nv_tx_done_optimized: flags 0x%x.\n", dprintk(KERN_DEBUG "%s: nv_tx_done_optimized: flags 0x%x.\n",
dev->name, flags); dev->name, flags);
...@@ -2492,6 +2498,7 @@ static void nv_tx_done_optimized(struct net_device *dev, int limit) ...@@ -2492,6 +2498,7 @@ static void nv_tx_done_optimized(struct net_device *dev, int limit)
dev_kfree_skb_any(np->get_tx_ctx->skb); dev_kfree_skb_any(np->get_tx_ctx->skb);
np->get_tx_ctx->skb = NULL; np->get_tx_ctx->skb = NULL;
tx_work++;
if (np->tx_limit) { if (np->tx_limit) {
nv_tx_flip_ownership(dev); nv_tx_flip_ownership(dev);
...@@ -2506,6 +2513,7 @@ static void nv_tx_done_optimized(struct net_device *dev, int limit) ...@@ -2506,6 +2513,7 @@ static void nv_tx_done_optimized(struct net_device *dev, int limit)
np->tx_stop = 0; np->tx_stop = 0;
netif_wake_queue(dev); netif_wake_queue(dev);
} }
return tx_work;
} }
/* /*
...@@ -2578,7 +2586,7 @@ static void nv_tx_timeout(struct net_device *dev) ...@@ -2578,7 +2586,7 @@ static void nv_tx_timeout(struct net_device *dev)
/* 2) check that the packets were not sent already: */ /* 2) check that the packets were not sent already: */
if (!nv_optimized(np)) if (!nv_optimized(np))
nv_tx_done(dev); nv_tx_done(dev, np->tx_ring_size);
else else
nv_tx_done_optimized(dev, np->tx_ring_size); nv_tx_done_optimized(dev, np->tx_ring_size);
...@@ -3433,7 +3441,7 @@ static irqreturn_t nv_nic_irq(int foo, void *data) ...@@ -3433,7 +3441,7 @@ static irqreturn_t nv_nic_irq(int foo, void *data)
nv_msi_workaround(np); nv_msi_workaround(np);
spin_lock(&np->lock); spin_lock(&np->lock);
nv_tx_done(dev); nv_tx_done(dev, np->tx_ring_size);
spin_unlock(&np->lock); spin_unlock(&np->lock);
#ifdef CONFIG_FORCEDETH_NAPI #ifdef CONFIG_FORCEDETH_NAPI
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment