Commit f0417c50 authored by Jakub Kicinski's avatar Jakub Kicinski

Merge branch '100GbE' of git://git.kernel.org/pub/scm/linux/kernel/git/tnguy/net-queue

Tony Nguyen says:

====================
ice: fix synchronization between .ndo_bpf() and reset

Larysa Zaremba says:

PF reset can be triggered asynchronously, by tx_timeout or by a user. With some
unfortunate timings both ice_vsi_rebuild() and .ndo_bpf will try to access and
modify XDP rings at the same time, causing system crash.

The first patch factors out rtnl-locked code from VSI rebuild code to avoid
deadlock. The following changes lock rebuild and .ndo_bpf() critical sections
with an internal mutex as well and provide complementary fixes.

* '100GbE' of git://git.kernel.org/pub/scm/linux/kernel/git/tnguy/net-queue:
  ice: do not bring the VSI up, if it was down before the XDP setup
  ice: remove ICE_CFG_BUSY locking from AF_XDP code
  ice: check ICE_VSI_DOWN under rtnl_lock when preparing for reset
  ice: check for XDP rings instead of bpf program when unconfiguring
  ice: protect XDP configuration with a mutex
  ice: move netif_queue_set_napi to rtnl-protected sections
====================

Link: https://patch.msgid.link/20240903183034.3530411-1-anthony.l.nguyen@intel.comSigned-off-by: default avatarJakub Kicinski <kuba@kernel.org>
parents 2603d315 04c7e14e
...@@ -318,6 +318,7 @@ enum ice_vsi_state { ...@@ -318,6 +318,7 @@ enum ice_vsi_state {
ICE_VSI_UMAC_FLTR_CHANGED, ICE_VSI_UMAC_FLTR_CHANGED,
ICE_VSI_MMAC_FLTR_CHANGED, ICE_VSI_MMAC_FLTR_CHANGED,
ICE_VSI_PROMISC_CHANGED, ICE_VSI_PROMISC_CHANGED,
ICE_VSI_REBUILD_PENDING,
ICE_VSI_STATE_NBITS /* must be last */ ICE_VSI_STATE_NBITS /* must be last */
}; };
...@@ -411,6 +412,7 @@ struct ice_vsi { ...@@ -411,6 +412,7 @@ struct ice_vsi {
struct ice_tx_ring **xdp_rings; /* XDP ring array */ struct ice_tx_ring **xdp_rings; /* XDP ring array */
u16 num_xdp_txq; /* Used XDP queues */ u16 num_xdp_txq; /* Used XDP queues */
u8 xdp_mapping_mode; /* ICE_MAP_MODE_[CONTIG|SCATTER] */ u8 xdp_mapping_mode; /* ICE_MAP_MODE_[CONTIG|SCATTER] */
struct mutex xdp_state_lock;
struct net_device **target_netdevs; struct net_device **target_netdevs;
......
...@@ -190,16 +190,11 @@ static void ice_free_q_vector(struct ice_vsi *vsi, int v_idx) ...@@ -190,16 +190,11 @@ static void ice_free_q_vector(struct ice_vsi *vsi, int v_idx)
} }
q_vector = vsi->q_vectors[v_idx]; q_vector = vsi->q_vectors[v_idx];
ice_for_each_tx_ring(tx_ring, q_vector->tx) { ice_for_each_tx_ring(tx_ring, vsi->q_vectors[v_idx]->tx)
ice_queue_set_napi(vsi, tx_ring->q_index, NETDEV_QUEUE_TYPE_TX,
NULL);
tx_ring->q_vector = NULL; tx_ring->q_vector = NULL;
}
ice_for_each_rx_ring(rx_ring, q_vector->rx) { ice_for_each_rx_ring(rx_ring, vsi->q_vectors[v_idx]->rx)
ice_queue_set_napi(vsi, rx_ring->q_index, NETDEV_QUEUE_TYPE_RX,
NULL);
rx_ring->q_vector = NULL; rx_ring->q_vector = NULL;
}
/* only VSI with an associated netdev is set up with NAPI */ /* only VSI with an associated netdev is set up with NAPI */
if (vsi->netdev) if (vsi->netdev)
......
...@@ -447,6 +447,7 @@ static void ice_vsi_free(struct ice_vsi *vsi) ...@@ -447,6 +447,7 @@ static void ice_vsi_free(struct ice_vsi *vsi)
ice_vsi_free_stats(vsi); ice_vsi_free_stats(vsi);
ice_vsi_free_arrays(vsi); ice_vsi_free_arrays(vsi);
mutex_destroy(&vsi->xdp_state_lock);
mutex_unlock(&pf->sw_mutex); mutex_unlock(&pf->sw_mutex);
devm_kfree(dev, vsi); devm_kfree(dev, vsi);
} }
...@@ -626,6 +627,8 @@ static struct ice_vsi *ice_vsi_alloc(struct ice_pf *pf) ...@@ -626,6 +627,8 @@ static struct ice_vsi *ice_vsi_alloc(struct ice_pf *pf)
pf->next_vsi = ice_get_free_slot(pf->vsi, pf->num_alloc_vsi, pf->next_vsi = ice_get_free_slot(pf->vsi, pf->num_alloc_vsi,
pf->next_vsi); pf->next_vsi);
mutex_init(&vsi->xdp_state_lock);
unlock_pf: unlock_pf:
mutex_unlock(&pf->sw_mutex); mutex_unlock(&pf->sw_mutex);
return vsi; return vsi;
...@@ -2286,9 +2289,6 @@ static int ice_vsi_cfg_def(struct ice_vsi *vsi) ...@@ -2286,9 +2289,6 @@ static int ice_vsi_cfg_def(struct ice_vsi *vsi)
ice_vsi_map_rings_to_vectors(vsi); ice_vsi_map_rings_to_vectors(vsi);
/* Associate q_vector rings to napi */
ice_vsi_set_napi_queues(vsi);
vsi->stat_offsets_loaded = false; vsi->stat_offsets_loaded = false;
/* ICE_VSI_CTRL does not need RSS so skip RSS processing */ /* ICE_VSI_CTRL does not need RSS so skip RSS processing */
...@@ -2426,7 +2426,7 @@ void ice_vsi_decfg(struct ice_vsi *vsi) ...@@ -2426,7 +2426,7 @@ void ice_vsi_decfg(struct ice_vsi *vsi)
dev_err(ice_pf_to_dev(pf), "Failed to remove RDMA scheduler config for VSI %u, err %d\n", dev_err(ice_pf_to_dev(pf), "Failed to remove RDMA scheduler config for VSI %u, err %d\n",
vsi->vsi_num, err); vsi->vsi_num, err);
if (ice_is_xdp_ena_vsi(vsi)) if (vsi->xdp_rings)
/* return value check can be skipped here, it always returns /* return value check can be skipped here, it always returns
* 0 if reset is in progress * 0 if reset is in progress
*/ */
...@@ -2528,7 +2528,7 @@ static void ice_vsi_release_msix(struct ice_vsi *vsi) ...@@ -2528,7 +2528,7 @@ static void ice_vsi_release_msix(struct ice_vsi *vsi)
for (q = 0; q < q_vector->num_ring_tx; q++) { for (q = 0; q < q_vector->num_ring_tx; q++) {
ice_write_itr(&q_vector->tx, 0); ice_write_itr(&q_vector->tx, 0);
wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), 0); wr32(hw, QINT_TQCTL(vsi->txq_map[txq]), 0);
if (ice_is_xdp_ena_vsi(vsi)) { if (vsi->xdp_rings) {
u32 xdp_txq = txq + vsi->num_xdp_txq; u32 xdp_txq = txq + vsi->num_xdp_txq;
wr32(hw, QINT_TQCTL(vsi->txq_map[xdp_txq]), 0); wr32(hw, QINT_TQCTL(vsi->txq_map[xdp_txq]), 0);
...@@ -2628,6 +2628,7 @@ void ice_vsi_close(struct ice_vsi *vsi) ...@@ -2628,6 +2628,7 @@ void ice_vsi_close(struct ice_vsi *vsi)
if (!test_and_set_bit(ICE_VSI_DOWN, vsi->state)) if (!test_and_set_bit(ICE_VSI_DOWN, vsi->state))
ice_down(vsi); ice_down(vsi);
ice_vsi_clear_napi_queues(vsi);
ice_vsi_free_irq(vsi); ice_vsi_free_irq(vsi);
ice_vsi_free_tx_rings(vsi); ice_vsi_free_tx_rings(vsi);
ice_vsi_free_rx_rings(vsi); ice_vsi_free_rx_rings(vsi);
...@@ -2671,8 +2672,7 @@ int ice_ena_vsi(struct ice_vsi *vsi, bool locked) ...@@ -2671,8 +2672,7 @@ int ice_ena_vsi(struct ice_vsi *vsi, bool locked)
*/ */
void ice_dis_vsi(struct ice_vsi *vsi, bool locked) void ice_dis_vsi(struct ice_vsi *vsi, bool locked)
{ {
if (test_bit(ICE_VSI_DOWN, vsi->state)) bool already_down = test_bit(ICE_VSI_DOWN, vsi->state);
return;
set_bit(ICE_VSI_NEEDS_RESTART, vsi->state); set_bit(ICE_VSI_NEEDS_RESTART, vsi->state);
...@@ -2680,134 +2680,70 @@ void ice_dis_vsi(struct ice_vsi *vsi, bool locked) ...@@ -2680,134 +2680,70 @@ void ice_dis_vsi(struct ice_vsi *vsi, bool locked)
if (netif_running(vsi->netdev)) { if (netif_running(vsi->netdev)) {
if (!locked) if (!locked)
rtnl_lock(); rtnl_lock();
already_down = test_bit(ICE_VSI_DOWN, vsi->state);
ice_vsi_close(vsi); if (!already_down)
ice_vsi_close(vsi);
if (!locked) if (!locked)
rtnl_unlock(); rtnl_unlock();
} else { } else if (!already_down) {
ice_vsi_close(vsi); ice_vsi_close(vsi);
} }
} else if (vsi->type == ICE_VSI_CTRL) { } else if (vsi->type == ICE_VSI_CTRL && !already_down) {
ice_vsi_close(vsi); ice_vsi_close(vsi);
} }
} }
/** /**
* __ice_queue_set_napi - Set the napi instance for the queue * ice_vsi_set_napi_queues - associate netdev queues with napi
* @dev: device to which NAPI and queue belong * @vsi: VSI pointer
* @queue_index: Index of queue
* @type: queue type as RX or TX
* @napi: NAPI context
* @locked: is the rtnl_lock already held
*
* Set the napi instance for the queue. Caller indicates the lock status.
*/
static void
__ice_queue_set_napi(struct net_device *dev, unsigned int queue_index,
enum netdev_queue_type type, struct napi_struct *napi,
bool locked)
{
if (!locked)
rtnl_lock();
netif_queue_set_napi(dev, queue_index, type, napi);
if (!locked)
rtnl_unlock();
}
/**
* ice_queue_set_napi - Set the napi instance for the queue
* @vsi: VSI being configured
* @queue_index: Index of queue
* @type: queue type as RX or TX
* @napi: NAPI context
* *
* Set the napi instance for the queue. The rtnl lock state is derived from the * Associate queue[s] with napi for all vectors.
* execution path. * The caller must hold rtnl_lock.
*/ */
void void ice_vsi_set_napi_queues(struct ice_vsi *vsi)
ice_queue_set_napi(struct ice_vsi *vsi, unsigned int queue_index,
enum netdev_queue_type type, struct napi_struct *napi)
{ {
struct ice_pf *pf = vsi->back; struct net_device *netdev = vsi->netdev;
int q_idx, v_idx;
if (!vsi->netdev) if (!netdev)
return; return;
if (current_work() == &pf->serv_task || ice_for_each_rxq(vsi, q_idx)
test_bit(ICE_PREPARED_FOR_RESET, pf->state) || netif_queue_set_napi(netdev, q_idx, NETDEV_QUEUE_TYPE_RX,
test_bit(ICE_DOWN, pf->state) || &vsi->rx_rings[q_idx]->q_vector->napi);
test_bit(ICE_SUSPENDED, pf->state))
__ice_queue_set_napi(vsi->netdev, queue_index, type, napi,
false);
else
__ice_queue_set_napi(vsi->netdev, queue_index, type, napi,
true);
}
/** ice_for_each_txq(vsi, q_idx)
* __ice_q_vector_set_napi_queues - Map queue[s] associated with the napi netif_queue_set_napi(netdev, q_idx, NETDEV_QUEUE_TYPE_TX,
* @q_vector: q_vector pointer &vsi->tx_rings[q_idx]->q_vector->napi);
* @locked: is the rtnl_lock already held
*
* Associate the q_vector napi with all the queue[s] on the vector.
* Caller indicates the lock status.
*/
void __ice_q_vector_set_napi_queues(struct ice_q_vector *q_vector, bool locked)
{
struct ice_rx_ring *rx_ring;
struct ice_tx_ring *tx_ring;
ice_for_each_rx_ring(rx_ring, q_vector->rx)
__ice_queue_set_napi(q_vector->vsi->netdev, rx_ring->q_index,
NETDEV_QUEUE_TYPE_RX, &q_vector->napi,
locked);
ice_for_each_tx_ring(tx_ring, q_vector->tx)
__ice_queue_set_napi(q_vector->vsi->netdev, tx_ring->q_index,
NETDEV_QUEUE_TYPE_TX, &q_vector->napi,
locked);
/* Also set the interrupt number for the NAPI */ /* Also set the interrupt number for the NAPI */
netif_napi_set_irq(&q_vector->napi, q_vector->irq.virq); ice_for_each_q_vector(vsi, v_idx) {
} struct ice_q_vector *q_vector = vsi->q_vectors[v_idx];
/**
* ice_q_vector_set_napi_queues - Map queue[s] associated with the napi
* @q_vector: q_vector pointer
*
* Associate the q_vector napi with all the queue[s] on the vector
*/
void ice_q_vector_set_napi_queues(struct ice_q_vector *q_vector)
{
struct ice_rx_ring *rx_ring;
struct ice_tx_ring *tx_ring;
ice_for_each_rx_ring(rx_ring, q_vector->rx)
ice_queue_set_napi(q_vector->vsi, rx_ring->q_index,
NETDEV_QUEUE_TYPE_RX, &q_vector->napi);
ice_for_each_tx_ring(tx_ring, q_vector->tx) netif_napi_set_irq(&q_vector->napi, q_vector->irq.virq);
ice_queue_set_napi(q_vector->vsi, tx_ring->q_index, }
NETDEV_QUEUE_TYPE_TX, &q_vector->napi);
/* Also set the interrupt number for the NAPI */
netif_napi_set_irq(&q_vector->napi, q_vector->irq.virq);
} }
/** /**
* ice_vsi_set_napi_queues * ice_vsi_clear_napi_queues - dissociate netdev queues from napi
* @vsi: VSI pointer * @vsi: VSI pointer
* *
* Associate queue[s] with napi for all vectors * Clear the association between all VSI queues queue[s] and napi.
* The caller must hold rtnl_lock.
*/ */
void ice_vsi_set_napi_queues(struct ice_vsi *vsi) void ice_vsi_clear_napi_queues(struct ice_vsi *vsi)
{ {
int i; struct net_device *netdev = vsi->netdev;
int q_idx;
if (!vsi->netdev) if (!netdev)
return; return;
ice_for_each_q_vector(vsi, i) ice_for_each_txq(vsi, q_idx)
ice_q_vector_set_napi_queues(vsi->q_vectors[i]); netif_queue_set_napi(netdev, q_idx, NETDEV_QUEUE_TYPE_TX, NULL);
ice_for_each_rxq(vsi, q_idx)
netif_queue_set_napi(netdev, q_idx, NETDEV_QUEUE_TYPE_RX, NULL);
} }
/** /**
...@@ -3039,19 +2975,23 @@ int ice_vsi_rebuild(struct ice_vsi *vsi, u32 vsi_flags) ...@@ -3039,19 +2975,23 @@ int ice_vsi_rebuild(struct ice_vsi *vsi, u32 vsi_flags)
if (WARN_ON(vsi->type == ICE_VSI_VF && !vsi->vf)) if (WARN_ON(vsi->type == ICE_VSI_VF && !vsi->vf))
return -EINVAL; return -EINVAL;
mutex_lock(&vsi->xdp_state_lock);
ret = ice_vsi_realloc_stat_arrays(vsi); ret = ice_vsi_realloc_stat_arrays(vsi);
if (ret) if (ret)
goto err_vsi_cfg; goto unlock;
ice_vsi_decfg(vsi); ice_vsi_decfg(vsi);
ret = ice_vsi_cfg_def(vsi); ret = ice_vsi_cfg_def(vsi);
if (ret) if (ret)
goto err_vsi_cfg; goto unlock;
coalesce = kcalloc(vsi->num_q_vectors, coalesce = kcalloc(vsi->num_q_vectors,
sizeof(struct ice_coalesce_stored), GFP_KERNEL); sizeof(struct ice_coalesce_stored), GFP_KERNEL);
if (!coalesce) if (!coalesce) {
return -ENOMEM; ret = -ENOMEM;
goto decfg;
}
prev_num_q_vectors = ice_vsi_rebuild_get_coalesce(vsi, coalesce); prev_num_q_vectors = ice_vsi_rebuild_get_coalesce(vsi, coalesce);
...@@ -3059,22 +2999,23 @@ int ice_vsi_rebuild(struct ice_vsi *vsi, u32 vsi_flags) ...@@ -3059,22 +2999,23 @@ int ice_vsi_rebuild(struct ice_vsi *vsi, u32 vsi_flags)
if (ret) { if (ret) {
if (vsi_flags & ICE_VSI_FLAG_INIT) { if (vsi_flags & ICE_VSI_FLAG_INIT) {
ret = -EIO; ret = -EIO;
goto err_vsi_cfg_tc_lan; goto free_coalesce;
} }
kfree(coalesce); ret = ice_schedule_reset(pf, ICE_RESET_PFR);
return ice_schedule_reset(pf, ICE_RESET_PFR); goto free_coalesce;
} }
ice_vsi_rebuild_set_coalesce(vsi, coalesce, prev_num_q_vectors); ice_vsi_rebuild_set_coalesce(vsi, coalesce, prev_num_q_vectors);
kfree(coalesce); clear_bit(ICE_VSI_REBUILD_PENDING, vsi->state);
return 0;
err_vsi_cfg_tc_lan: free_coalesce:
ice_vsi_decfg(vsi);
kfree(coalesce); kfree(coalesce);
err_vsi_cfg: decfg:
if (ret)
ice_vsi_decfg(vsi);
unlock:
mutex_unlock(&vsi->xdp_state_lock);
return ret; return ret;
} }
......
...@@ -44,16 +44,10 @@ void ice_vsi_cfg_netdev_tc(struct ice_vsi *vsi, u8 ena_tc); ...@@ -44,16 +44,10 @@ void ice_vsi_cfg_netdev_tc(struct ice_vsi *vsi, u8 ena_tc);
struct ice_vsi * struct ice_vsi *
ice_vsi_setup(struct ice_pf *pf, struct ice_vsi_cfg_params *params); ice_vsi_setup(struct ice_pf *pf, struct ice_vsi_cfg_params *params);
void
ice_queue_set_napi(struct ice_vsi *vsi, unsigned int queue_index,
enum netdev_queue_type type, struct napi_struct *napi);
void __ice_q_vector_set_napi_queues(struct ice_q_vector *q_vector, bool locked);
void ice_q_vector_set_napi_queues(struct ice_q_vector *q_vector);
void ice_vsi_set_napi_queues(struct ice_vsi *vsi); void ice_vsi_set_napi_queues(struct ice_vsi *vsi);
void ice_vsi_clear_napi_queues(struct ice_vsi *vsi);
int ice_vsi_release(struct ice_vsi *vsi); int ice_vsi_release(struct ice_vsi *vsi);
void ice_vsi_close(struct ice_vsi *vsi); void ice_vsi_close(struct ice_vsi *vsi);
......
...@@ -616,6 +616,7 @@ ice_prepare_for_reset(struct ice_pf *pf, enum ice_reset_req reset_type) ...@@ -616,6 +616,7 @@ ice_prepare_for_reset(struct ice_pf *pf, enum ice_reset_req reset_type)
/* clear SW filtering DB */ /* clear SW filtering DB */
ice_clear_hw_tbls(hw); ice_clear_hw_tbls(hw);
/* disable the VSIs and their queues that are not already DOWN */ /* disable the VSIs and their queues that are not already DOWN */
set_bit(ICE_VSI_REBUILD_PENDING, ice_get_main_vsi(pf)->state);
ice_pf_dis_all_vsi(pf, false); ice_pf_dis_all_vsi(pf, false);
if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags)) if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags))
...@@ -3004,8 +3005,8 @@ ice_xdp_setup_prog(struct ice_vsi *vsi, struct bpf_prog *prog, ...@@ -3004,8 +3005,8 @@ ice_xdp_setup_prog(struct ice_vsi *vsi, struct bpf_prog *prog,
struct netlink_ext_ack *extack) struct netlink_ext_ack *extack)
{ {
unsigned int frame_size = vsi->netdev->mtu + ICE_ETH_PKT_HDR_PAD; unsigned int frame_size = vsi->netdev->mtu + ICE_ETH_PKT_HDR_PAD;
bool if_running = netif_running(vsi->netdev);
int ret = 0, xdp_ring_err = 0; int ret = 0, xdp_ring_err = 0;
bool if_running;
if (prog && !prog->aux->xdp_has_frags) { if (prog && !prog->aux->xdp_has_frags) {
if (frame_size > ice_max_xdp_frame_size(vsi)) { if (frame_size > ice_max_xdp_frame_size(vsi)) {
...@@ -3016,13 +3017,17 @@ ice_xdp_setup_prog(struct ice_vsi *vsi, struct bpf_prog *prog, ...@@ -3016,13 +3017,17 @@ ice_xdp_setup_prog(struct ice_vsi *vsi, struct bpf_prog *prog,
} }
/* hot swap progs and avoid toggling link */ /* hot swap progs and avoid toggling link */
if (ice_is_xdp_ena_vsi(vsi) == !!prog) { if (ice_is_xdp_ena_vsi(vsi) == !!prog ||
test_bit(ICE_VSI_REBUILD_PENDING, vsi->state)) {
ice_vsi_assign_bpf_prog(vsi, prog); ice_vsi_assign_bpf_prog(vsi, prog);
return 0; return 0;
} }
if_running = netif_running(vsi->netdev) &&
!test_and_set_bit(ICE_VSI_DOWN, vsi->state);
/* need to stop netdev while setting up the program for Rx rings */ /* need to stop netdev while setting up the program for Rx rings */
if (if_running && !test_and_set_bit(ICE_VSI_DOWN, vsi->state)) { if (if_running) {
ret = ice_down(vsi); ret = ice_down(vsi);
if (ret) { if (ret) {
NL_SET_ERR_MSG_MOD(extack, "Preparing device for XDP attach failed"); NL_SET_ERR_MSG_MOD(extack, "Preparing device for XDP attach failed");
...@@ -3088,21 +3093,28 @@ static int ice_xdp(struct net_device *dev, struct netdev_bpf *xdp) ...@@ -3088,21 +3093,28 @@ static int ice_xdp(struct net_device *dev, struct netdev_bpf *xdp)
{ {
struct ice_netdev_priv *np = netdev_priv(dev); struct ice_netdev_priv *np = netdev_priv(dev);
struct ice_vsi *vsi = np->vsi; struct ice_vsi *vsi = np->vsi;
int ret;
if (vsi->type != ICE_VSI_PF) { if (vsi->type != ICE_VSI_PF) {
NL_SET_ERR_MSG_MOD(xdp->extack, "XDP can be loaded only on PF VSI"); NL_SET_ERR_MSG_MOD(xdp->extack, "XDP can be loaded only on PF VSI");
return -EINVAL; return -EINVAL;
} }
mutex_lock(&vsi->xdp_state_lock);
switch (xdp->command) { switch (xdp->command) {
case XDP_SETUP_PROG: case XDP_SETUP_PROG:
return ice_xdp_setup_prog(vsi, xdp->prog, xdp->extack); ret = ice_xdp_setup_prog(vsi, xdp->prog, xdp->extack);
break;
case XDP_SETUP_XSK_POOL: case XDP_SETUP_XSK_POOL:
return ice_xsk_pool_setup(vsi, xdp->xsk.pool, ret = ice_xsk_pool_setup(vsi, xdp->xsk.pool, xdp->xsk.queue_id);
xdp->xsk.queue_id); break;
default: default:
return -EINVAL; ret = -EINVAL;
} }
mutex_unlock(&vsi->xdp_state_lock);
return ret;
} }
/** /**
...@@ -3558,11 +3570,9 @@ static void ice_napi_add(struct ice_vsi *vsi) ...@@ -3558,11 +3570,9 @@ static void ice_napi_add(struct ice_vsi *vsi)
if (!vsi->netdev) if (!vsi->netdev)
return; return;
ice_for_each_q_vector(vsi, v_idx) { ice_for_each_q_vector(vsi, v_idx)
netif_napi_add(vsi->netdev, &vsi->q_vectors[v_idx]->napi, netif_napi_add(vsi->netdev, &vsi->q_vectors[v_idx]->napi,
ice_napi_poll); ice_napi_poll);
__ice_q_vector_set_napi_queues(vsi->q_vectors[v_idx], false);
}
} }
/** /**
...@@ -5540,7 +5550,9 @@ static int ice_reinit_interrupt_scheme(struct ice_pf *pf) ...@@ -5540,7 +5550,9 @@ static int ice_reinit_interrupt_scheme(struct ice_pf *pf)
if (ret) if (ret)
goto err_reinit; goto err_reinit;
ice_vsi_map_rings_to_vectors(pf->vsi[v]); ice_vsi_map_rings_to_vectors(pf->vsi[v]);
rtnl_lock();
ice_vsi_set_napi_queues(pf->vsi[v]); ice_vsi_set_napi_queues(pf->vsi[v]);
rtnl_unlock();
} }
ret = ice_req_irq_msix_misc(pf); ret = ice_req_irq_msix_misc(pf);
...@@ -5554,8 +5566,12 @@ static int ice_reinit_interrupt_scheme(struct ice_pf *pf) ...@@ -5554,8 +5566,12 @@ static int ice_reinit_interrupt_scheme(struct ice_pf *pf)
err_reinit: err_reinit:
while (v--) while (v--)
if (pf->vsi[v]) if (pf->vsi[v]) {
rtnl_lock();
ice_vsi_clear_napi_queues(pf->vsi[v]);
rtnl_unlock();
ice_vsi_free_q_vectors(pf->vsi[v]); ice_vsi_free_q_vectors(pf->vsi[v]);
}
return ret; return ret;
} }
...@@ -5620,6 +5636,9 @@ static int ice_suspend(struct device *dev) ...@@ -5620,6 +5636,9 @@ static int ice_suspend(struct device *dev)
ice_for_each_vsi(pf, v) { ice_for_each_vsi(pf, v) {
if (!pf->vsi[v]) if (!pf->vsi[v])
continue; continue;
rtnl_lock();
ice_vsi_clear_napi_queues(pf->vsi[v]);
rtnl_unlock();
ice_vsi_free_q_vectors(pf->vsi[v]); ice_vsi_free_q_vectors(pf->vsi[v]);
} }
ice_clear_interrupt_scheme(pf); ice_clear_interrupt_scheme(pf);
...@@ -7233,7 +7252,7 @@ int ice_down(struct ice_vsi *vsi) ...@@ -7233,7 +7252,7 @@ int ice_down(struct ice_vsi *vsi)
if (tx_err) if (tx_err)
netdev_err(vsi->netdev, "Failed stop Tx rings, VSI %d error %d\n", netdev_err(vsi->netdev, "Failed stop Tx rings, VSI %d error %d\n",
vsi->vsi_num, tx_err); vsi->vsi_num, tx_err);
if (!tx_err && ice_is_xdp_ena_vsi(vsi)) { if (!tx_err && vsi->xdp_rings) {
tx_err = ice_vsi_stop_xdp_tx_rings(vsi); tx_err = ice_vsi_stop_xdp_tx_rings(vsi);
if (tx_err) if (tx_err)
netdev_err(vsi->netdev, "Failed stop XDP rings, VSI %d error %d\n", netdev_err(vsi->netdev, "Failed stop XDP rings, VSI %d error %d\n",
...@@ -7250,7 +7269,7 @@ int ice_down(struct ice_vsi *vsi) ...@@ -7250,7 +7269,7 @@ int ice_down(struct ice_vsi *vsi)
ice_for_each_txq(vsi, i) ice_for_each_txq(vsi, i)
ice_clean_tx_ring(vsi->tx_rings[i]); ice_clean_tx_ring(vsi->tx_rings[i]);
if (ice_is_xdp_ena_vsi(vsi)) if (vsi->xdp_rings)
ice_for_each_xdp_txq(vsi, i) ice_for_each_xdp_txq(vsi, i)
ice_clean_tx_ring(vsi->xdp_rings[i]); ice_clean_tx_ring(vsi->xdp_rings[i]);
...@@ -7455,6 +7474,8 @@ int ice_vsi_open(struct ice_vsi *vsi) ...@@ -7455,6 +7474,8 @@ int ice_vsi_open(struct ice_vsi *vsi)
err = netif_set_real_num_rx_queues(vsi->netdev, vsi->num_rxq); err = netif_set_real_num_rx_queues(vsi->netdev, vsi->num_rxq);
if (err) if (err)
goto err_set_qs; goto err_set_qs;
ice_vsi_set_napi_queues(vsi);
} }
err = ice_up_complete(vsi); err = ice_up_complete(vsi);
......
...@@ -39,7 +39,7 @@ static void ice_qp_reset_stats(struct ice_vsi *vsi, u16 q_idx) ...@@ -39,7 +39,7 @@ static void ice_qp_reset_stats(struct ice_vsi *vsi, u16 q_idx)
sizeof(vsi_stat->rx_ring_stats[q_idx]->rx_stats)); sizeof(vsi_stat->rx_ring_stats[q_idx]->rx_stats));
memset(&vsi_stat->tx_ring_stats[q_idx]->stats, 0, memset(&vsi_stat->tx_ring_stats[q_idx]->stats, 0,
sizeof(vsi_stat->tx_ring_stats[q_idx]->stats)); sizeof(vsi_stat->tx_ring_stats[q_idx]->stats));
if (ice_is_xdp_ena_vsi(vsi)) if (vsi->xdp_rings)
memset(&vsi->xdp_rings[q_idx]->ring_stats->stats, 0, memset(&vsi->xdp_rings[q_idx]->ring_stats->stats, 0,
sizeof(vsi->xdp_rings[q_idx]->ring_stats->stats)); sizeof(vsi->xdp_rings[q_idx]->ring_stats->stats));
} }
...@@ -52,7 +52,7 @@ static void ice_qp_reset_stats(struct ice_vsi *vsi, u16 q_idx) ...@@ -52,7 +52,7 @@ static void ice_qp_reset_stats(struct ice_vsi *vsi, u16 q_idx)
static void ice_qp_clean_rings(struct ice_vsi *vsi, u16 q_idx) static void ice_qp_clean_rings(struct ice_vsi *vsi, u16 q_idx)
{ {
ice_clean_tx_ring(vsi->tx_rings[q_idx]); ice_clean_tx_ring(vsi->tx_rings[q_idx]);
if (ice_is_xdp_ena_vsi(vsi)) if (vsi->xdp_rings)
ice_clean_tx_ring(vsi->xdp_rings[q_idx]); ice_clean_tx_ring(vsi->xdp_rings[q_idx]);
ice_clean_rx_ring(vsi->rx_rings[q_idx]); ice_clean_rx_ring(vsi->rx_rings[q_idx]);
} }
...@@ -165,7 +165,6 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx) ...@@ -165,7 +165,6 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx)
struct ice_q_vector *q_vector; struct ice_q_vector *q_vector;
struct ice_tx_ring *tx_ring; struct ice_tx_ring *tx_ring;
struct ice_rx_ring *rx_ring; struct ice_rx_ring *rx_ring;
int timeout = 50;
int fail = 0; int fail = 0;
int err; int err;
...@@ -176,13 +175,6 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx) ...@@ -176,13 +175,6 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx)
rx_ring = vsi->rx_rings[q_idx]; rx_ring = vsi->rx_rings[q_idx];
q_vector = rx_ring->q_vector; q_vector = rx_ring->q_vector;
while (test_and_set_bit(ICE_CFG_BUSY, vsi->state)) {
timeout--;
if (!timeout)
return -EBUSY;
usleep_range(1000, 2000);
}
synchronize_net(); synchronize_net();
netif_carrier_off(vsi->netdev); netif_carrier_off(vsi->netdev);
netif_tx_stop_queue(netdev_get_tx_queue(vsi->netdev, q_idx)); netif_tx_stop_queue(netdev_get_tx_queue(vsi->netdev, q_idx));
...@@ -194,7 +186,7 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx) ...@@ -194,7 +186,7 @@ static int ice_qp_dis(struct ice_vsi *vsi, u16 q_idx)
err = ice_vsi_stop_tx_ring(vsi, ICE_NO_RESET, 0, tx_ring, &txq_meta); err = ice_vsi_stop_tx_ring(vsi, ICE_NO_RESET, 0, tx_ring, &txq_meta);
if (!fail) if (!fail)
fail = err; fail = err;
if (ice_is_xdp_ena_vsi(vsi)) { if (vsi->xdp_rings) {
struct ice_tx_ring *xdp_ring = vsi->xdp_rings[q_idx]; struct ice_tx_ring *xdp_ring = vsi->xdp_rings[q_idx];
memset(&txq_meta, 0, sizeof(txq_meta)); memset(&txq_meta, 0, sizeof(txq_meta));
...@@ -261,7 +253,6 @@ static int ice_qp_ena(struct ice_vsi *vsi, u16 q_idx) ...@@ -261,7 +253,6 @@ static int ice_qp_ena(struct ice_vsi *vsi, u16 q_idx)
netif_tx_start_queue(netdev_get_tx_queue(vsi->netdev, q_idx)); netif_tx_start_queue(netdev_get_tx_queue(vsi->netdev, q_idx));
netif_carrier_on(vsi->netdev); netif_carrier_on(vsi->netdev);
} }
clear_bit(ICE_CFG_BUSY, vsi->state);
return fail; return fail;
} }
...@@ -390,7 +381,8 @@ int ice_xsk_pool_setup(struct ice_vsi *vsi, struct xsk_buff_pool *pool, u16 qid) ...@@ -390,7 +381,8 @@ int ice_xsk_pool_setup(struct ice_vsi *vsi, struct xsk_buff_pool *pool, u16 qid)
goto failure; goto failure;
} }
if_running = netif_running(vsi->netdev) && ice_is_xdp_ena_vsi(vsi); if_running = !test_bit(ICE_VSI_DOWN, vsi->state) &&
ice_is_xdp_ena_vsi(vsi);
if (if_running) { if (if_running) {
struct ice_rx_ring *rx_ring = vsi->rx_rings[qid]; struct ice_rx_ring *rx_ring = vsi->rx_rings[qid];
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment