Commit b1151b74 authored by Mike Marciniszyn's avatar Mike Marciniszyn Committed by Jason Gunthorpe

IB/hfi1: Fix alloc failure with larger txqueuelen

The following allocation with large txqueuelen will result in the
following warning:

  Call Trace:
   __alloc_pages_nodemask+0x283/0x2c0
   kmalloc_large_node+0x3c/0xa0
   __kmalloc_node+0x22a/0x2f0
   hfi1_ipoib_txreq_init+0x19f/0x330 [hfi1]
   hfi1_ipoib_setup_rn+0xd3/0x1a0 [hfi1]
   rdma_init_netdev+0x5a/0x80 [ib_core]
   ipoib_intf_init+0x6c/0x350 [ib_ipoib]
   ipoib_intf_alloc+0x5c/0xc0 [ib_ipoib]
   ipoib_add_one+0xbe/0x300 [ib_ipoib]
   add_client_context+0x12c/0x1a0 [ib_core]
   ib_register_client+0x147/0x190 [ib_core]
   ipoib_init_module+0xdd/0x132 [ib_ipoib]
   do_one_initcall+0x46/0x1c3
   do_init_module+0x5a/0x220
   load_module+0x14c5/0x17f0
   __do_sys_init_module+0x13b/0x180
   do_syscall_64+0x5b/0x1a0
   entry_SYSCALL_64_after_hwframe+0x65/0xca

For ipoib, the txqueuelen is modified with the module parameter
send_queue_size.

Fix by changing to use kv versions of the same allocator to handle the
large allocations.  The allocation embeds a hdr struct that is dma mapped.
Change that struct to a pointer to a kzalloced struct.

Cc: stable@vger.kernel.org
Fixes: d99dc602 ("IB/hfi1: Add functions to transmit datagram ipoib packets")
Link: https://lore.kernel.org/r/1642287756-182313-3-git-send-email-mike.marciniszyn@cornelisnetworks.comReviewed-by: default avatarDennis Dalessandro <dennis.dalessandro@cornelisnetworks.com>
Signed-off-by: default avatarMike Marciniszyn <mike.marciniszyn@cornelisnetworks.com>
Signed-off-by: default avatarJason Gunthorpe <jgg@nvidia.com>
parent 8c83d39c
...@@ -55,7 +55,7 @@ union hfi1_ipoib_flow { ...@@ -55,7 +55,7 @@ union hfi1_ipoib_flow {
*/ */
struct ipoib_txreq { struct ipoib_txreq {
struct sdma_txreq txreq; struct sdma_txreq txreq;
struct hfi1_sdma_header sdma_hdr; struct hfi1_sdma_header *sdma_hdr;
int sdma_status; int sdma_status;
int complete; int complete;
struct hfi1_ipoib_dev_priv *priv; struct hfi1_ipoib_dev_priv *priv;
......
...@@ -122,7 +122,7 @@ static void hfi1_ipoib_free_tx(struct ipoib_txreq *tx, int budget) ...@@ -122,7 +122,7 @@ static void hfi1_ipoib_free_tx(struct ipoib_txreq *tx, int budget)
dd_dev_warn(priv->dd, dd_dev_warn(priv->dd,
"%s: Status = 0x%x pbc 0x%llx txq = %d sde = %d\n", "%s: Status = 0x%x pbc 0x%llx txq = %d sde = %d\n",
__func__, tx->sdma_status, __func__, tx->sdma_status,
le64_to_cpu(tx->sdma_hdr.pbc), tx->txq->q_idx, le64_to_cpu(tx->sdma_hdr->pbc), tx->txq->q_idx,
tx->txq->sde->this_idx); tx->txq->sde->this_idx);
} }
...@@ -231,7 +231,7 @@ static int hfi1_ipoib_build_tx_desc(struct ipoib_txreq *tx, ...@@ -231,7 +231,7 @@ static int hfi1_ipoib_build_tx_desc(struct ipoib_txreq *tx,
{ {
struct hfi1_devdata *dd = txp->dd; struct hfi1_devdata *dd = txp->dd;
struct sdma_txreq *txreq = &tx->txreq; struct sdma_txreq *txreq = &tx->txreq;
struct hfi1_sdma_header *sdma_hdr = &tx->sdma_hdr; struct hfi1_sdma_header *sdma_hdr = tx->sdma_hdr;
u16 pkt_bytes = u16 pkt_bytes =
sizeof(sdma_hdr->pbc) + (txp->hdr_dwords << 2) + tx->skb->len; sizeof(sdma_hdr->pbc) + (txp->hdr_dwords << 2) + tx->skb->len;
int ret; int ret;
...@@ -256,7 +256,7 @@ static void hfi1_ipoib_build_ib_tx_headers(struct ipoib_txreq *tx, ...@@ -256,7 +256,7 @@ static void hfi1_ipoib_build_ib_tx_headers(struct ipoib_txreq *tx,
struct ipoib_txparms *txp) struct ipoib_txparms *txp)
{ {
struct hfi1_ipoib_dev_priv *priv = tx->txq->priv; struct hfi1_ipoib_dev_priv *priv = tx->txq->priv;
struct hfi1_sdma_header *sdma_hdr = &tx->sdma_hdr; struct hfi1_sdma_header *sdma_hdr = tx->sdma_hdr;
struct sk_buff *skb = tx->skb; struct sk_buff *skb = tx->skb;
struct hfi1_pportdata *ppd = ppd_from_ibp(txp->ibp); struct hfi1_pportdata *ppd = ppd_from_ibp(txp->ibp);
struct rdma_ah_attr *ah_attr = txp->ah_attr; struct rdma_ah_attr *ah_attr = txp->ah_attr;
...@@ -483,7 +483,7 @@ static int hfi1_ipoib_send_dma_single(struct net_device *dev, ...@@ -483,7 +483,7 @@ static int hfi1_ipoib_send_dma_single(struct net_device *dev,
if (likely(!ret)) { if (likely(!ret)) {
tx_ok: tx_ok:
trace_sdma_output_ibhdr(txq->priv->dd, trace_sdma_output_ibhdr(txq->priv->dd,
&tx->sdma_hdr.hdr, &tx->sdma_hdr->hdr,
ib_is_sc5(txp->flow.sc5)); ib_is_sc5(txp->flow.sc5));
hfi1_ipoib_check_queue_depth(txq); hfi1_ipoib_check_queue_depth(txq);
return NETDEV_TX_OK; return NETDEV_TX_OK;
...@@ -547,7 +547,7 @@ static int hfi1_ipoib_send_dma_list(struct net_device *dev, ...@@ -547,7 +547,7 @@ static int hfi1_ipoib_send_dma_list(struct net_device *dev,
hfi1_ipoib_check_queue_depth(txq); hfi1_ipoib_check_queue_depth(txq);
trace_sdma_output_ibhdr(txq->priv->dd, trace_sdma_output_ibhdr(txq->priv->dd,
&tx->sdma_hdr.hdr, &tx->sdma_hdr->hdr,
ib_is_sc5(txp->flow.sc5)); ib_is_sc5(txp->flow.sc5));
if (!netdev_xmit_more()) if (!netdev_xmit_more())
...@@ -683,7 +683,8 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv) ...@@ -683,7 +683,8 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv)
{ {
struct net_device *dev = priv->netdev; struct net_device *dev = priv->netdev;
u32 tx_ring_size, tx_item_size; u32 tx_ring_size, tx_item_size;
int i; struct hfi1_ipoib_circ_buf *tx_ring;
int i, j;
/* /*
* Ring holds 1 less than tx_ring_size * Ring holds 1 less than tx_ring_size
...@@ -701,7 +702,9 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv) ...@@ -701,7 +702,9 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv)
for (i = 0; i < dev->num_tx_queues; i++) { for (i = 0; i < dev->num_tx_queues; i++) {
struct hfi1_ipoib_txq *txq = &priv->txqs[i]; struct hfi1_ipoib_txq *txq = &priv->txqs[i];
struct ipoib_txreq *tx;
tx_ring = &txq->tx_ring;
iowait_init(&txq->wait, iowait_init(&txq->wait,
0, 0,
hfi1_ipoib_flush_txq, hfi1_ipoib_flush_txq,
...@@ -725,7 +728,7 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv) ...@@ -725,7 +728,7 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv)
priv->dd->node); priv->dd->node);
txq->tx_ring.items = txq->tx_ring.items =
kcalloc_node(tx_ring_size, tx_item_size, kvzalloc_node(array_size(tx_ring_size, tx_item_size),
GFP_KERNEL, priv->dd->node); GFP_KERNEL, priv->dd->node);
if (!txq->tx_ring.items) if (!txq->tx_ring.items)
goto free_txqs; goto free_txqs;
...@@ -733,6 +736,11 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv) ...@@ -733,6 +736,11 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv)
txq->tx_ring.max_items = tx_ring_size; txq->tx_ring.max_items = tx_ring_size;
txq->tx_ring.shift = ilog2(tx_item_size); txq->tx_ring.shift = ilog2(tx_item_size);
txq->tx_ring.avail = hfi1_ipoib_ring_hwat(txq); txq->tx_ring.avail = hfi1_ipoib_ring_hwat(txq);
tx_ring = &txq->tx_ring;
for (j = 0; j < tx_ring_size; j++)
hfi1_txreq_from_idx(tx_ring, j)->sdma_hdr =
kzalloc_node(sizeof(*tx->sdma_hdr),
GFP_KERNEL, priv->dd->node);
netif_tx_napi_add(dev, &txq->napi, netif_tx_napi_add(dev, &txq->napi,
hfi1_ipoib_poll_tx_ring, hfi1_ipoib_poll_tx_ring,
...@@ -746,7 +754,10 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv) ...@@ -746,7 +754,10 @@ int hfi1_ipoib_txreq_init(struct hfi1_ipoib_dev_priv *priv)
struct hfi1_ipoib_txq *txq = &priv->txqs[i]; struct hfi1_ipoib_txq *txq = &priv->txqs[i];
netif_napi_del(&txq->napi); netif_napi_del(&txq->napi);
kfree(txq->tx_ring.items); tx_ring = &txq->tx_ring;
for (j = 0; j < tx_ring_size; j++)
kfree(hfi1_txreq_from_idx(tx_ring, j)->sdma_hdr);
kvfree(tx_ring->items);
} }
kfree(priv->txqs); kfree(priv->txqs);
...@@ -780,17 +791,20 @@ static void hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq) ...@@ -780,17 +791,20 @@ static void hfi1_ipoib_drain_tx_list(struct hfi1_ipoib_txq *txq)
void hfi1_ipoib_txreq_deinit(struct hfi1_ipoib_dev_priv *priv) void hfi1_ipoib_txreq_deinit(struct hfi1_ipoib_dev_priv *priv)
{ {
int i; int i, j;
for (i = 0; i < priv->netdev->num_tx_queues; i++) { for (i = 0; i < priv->netdev->num_tx_queues; i++) {
struct hfi1_ipoib_txq *txq = &priv->txqs[i]; struct hfi1_ipoib_txq *txq = &priv->txqs[i];
struct hfi1_ipoib_circ_buf *tx_ring = &txq->tx_ring;
iowait_cancel_work(&txq->wait); iowait_cancel_work(&txq->wait);
iowait_sdma_drain(&txq->wait); iowait_sdma_drain(&txq->wait);
hfi1_ipoib_drain_tx_list(txq); hfi1_ipoib_drain_tx_list(txq);
netif_napi_del(&txq->napi); netif_napi_del(&txq->napi);
hfi1_ipoib_drain_tx_ring(txq); hfi1_ipoib_drain_tx_ring(txq);
kfree(txq->tx_ring.items); for (j = 0; j < tx_ring->max_items; j++)
kfree(hfi1_txreq_from_idx(tx_ring, j)->sdma_hdr);
kvfree(tx_ring->items);
} }
kfree(priv->txqs); kfree(priv->txqs);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment