Commit d48b1d29 authored by Chuck Lever's avatar Chuck Lever Committed by Anna Schumaker

xprtrdma: Move init and release helpers

Clean up: Moving these helpers in a separate patch makes later
patches more readable.
Signed-off-by: default avatarChuck Lever <chuck.lever@oracle.com>
Tested-by: default avatarSteve Wise <swise@opengridcomputing.com>
Signed-off-by: default avatarAnna Schumaker <Anna.Schumaker@Netapp.com>
parent 564471d2
...@@ -35,6 +35,12 @@ ...@@ -35,6 +35,12 @@
/* Maximum scatter/gather per FMR */ /* Maximum scatter/gather per FMR */
#define RPCRDMA_MAX_FMR_SGES (64) #define RPCRDMA_MAX_FMR_SGES (64)
/* Access mode of externally registered pages */
enum {
RPCRDMA_FMR_ACCESS_FLAGS = IB_ACCESS_REMOTE_WRITE |
IB_ACCESS_REMOTE_READ,
};
static struct workqueue_struct *fmr_recovery_wq; static struct workqueue_struct *fmr_recovery_wq;
#define FMR_RECOVERY_WQ_FLAGS (WQ_UNBOUND) #define FMR_RECOVERY_WQ_FLAGS (WQ_UNBOUND)
...@@ -59,6 +65,44 @@ fmr_destroy_recovery_wq(void) ...@@ -59,6 +65,44 @@ fmr_destroy_recovery_wq(void)
destroy_workqueue(wq); destroy_workqueue(wq);
} }
static int
__fmr_init(struct rpcrdma_mw *mw, struct ib_pd *pd)
{
static struct ib_fmr_attr fmr_attr = {
.max_pages = RPCRDMA_MAX_FMR_SGES,
.max_maps = 1,
.page_shift = PAGE_SHIFT
};
mw->fmr.physaddrs = kcalloc(RPCRDMA_MAX_FMR_SGES,
sizeof(u64), GFP_KERNEL);
if (!mw->fmr.physaddrs)
goto out_free;
mw->mw_sg = kcalloc(RPCRDMA_MAX_FMR_SGES,
sizeof(*mw->mw_sg), GFP_KERNEL);
if (!mw->mw_sg)
goto out_free;
sg_init_table(mw->mw_sg, RPCRDMA_MAX_FMR_SGES);
mw->fmr.fmr = ib_alloc_fmr(pd, RPCRDMA_FMR_ACCESS_FLAGS,
&fmr_attr);
if (IS_ERR(mw->fmr.fmr))
goto out_fmr_err;
return 0;
out_fmr_err:
dprintk("RPC: %s: ib_alloc_fmr returned %ld\n", __func__,
PTR_ERR(mw->fmr.fmr));
out_free:
kfree(mw->mw_sg);
kfree(mw->fmr.physaddrs);
return -ENOMEM;
}
static int static int
__fmr_unmap(struct rpcrdma_mw *mw) __fmr_unmap(struct rpcrdma_mw *mw)
{ {
...@@ -71,6 +115,30 @@ __fmr_unmap(struct rpcrdma_mw *mw) ...@@ -71,6 +115,30 @@ __fmr_unmap(struct rpcrdma_mw *mw)
return rc; return rc;
} }
static void
__fmr_dma_unmap(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr_seg *seg)
{
struct ib_device *device = r_xprt->rx_ia.ri_device;
int nsegs = seg->mr_nsegs;
while (nsegs--)
rpcrdma_unmap_one(device, seg++);
}
static void
__fmr_release(struct rpcrdma_mw *r)
{
int rc;
kfree(r->fmr.physaddrs);
kfree(r->mw_sg);
rc = ib_dealloc_fmr(r->fmr.fmr);
if (rc)
pr_err("rpcrdma: final ib_dealloc_fmr for %p returned %i\n",
r, rc);
}
/* Deferred reset of a single FMR. Generate a fresh rkey by /* Deferred reset of a single FMR. Generate a fresh rkey by
* replacing the MR. There's no recovery if this fails. * replacing the MR. There's no recovery if this fails.
*/ */
...@@ -119,12 +187,6 @@ static int ...@@ -119,12 +187,6 @@ static int
fmr_op_init(struct rpcrdma_xprt *r_xprt) fmr_op_init(struct rpcrdma_xprt *r_xprt)
{ {
struct rpcrdma_buffer *buf = &r_xprt->rx_buf; struct rpcrdma_buffer *buf = &r_xprt->rx_buf;
int mr_access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_REMOTE_READ;
struct ib_fmr_attr fmr_attr = {
.max_pages = RPCRDMA_MAX_FMR_SGES,
.max_maps = 1,
.page_shift = PAGE_SHIFT
};
struct ib_pd *pd = r_xprt->rx_ia.ri_pd; struct ib_pd *pd = r_xprt->rx_ia.ri_pd;
struct rpcrdma_mw *r; struct rpcrdma_mw *r;
int i, rc; int i, rc;
...@@ -138,35 +200,22 @@ fmr_op_init(struct rpcrdma_xprt *r_xprt) ...@@ -138,35 +200,22 @@ fmr_op_init(struct rpcrdma_xprt *r_xprt)
i *= buf->rb_max_requests; /* one set for each RPC slot */ i *= buf->rb_max_requests; /* one set for each RPC slot */
dprintk("RPC: %s: initalizing %d FMRs\n", __func__, i); dprintk("RPC: %s: initalizing %d FMRs\n", __func__, i);
rc = -ENOMEM;
while (i--) { while (i--) {
r = kzalloc(sizeof(*r), GFP_KERNEL); r = kzalloc(sizeof(*r), GFP_KERNEL);
if (!r) if (!r)
goto out; return -ENOMEM;
r->fmr.physaddrs = kmalloc(RPCRDMA_MAX_FMR_SGES *
sizeof(u64), GFP_KERNEL);
if (!r->fmr.physaddrs)
goto out_free;
r->fmr.fmr = ib_alloc_fmr(pd, mr_access_flags, &fmr_attr); rc = __fmr_init(r, pd);
if (IS_ERR(r->fmr.fmr)) if (rc) {
goto out_fmr_err; kfree(r);
return rc;
}
r->mw_xprt = r_xprt; r->mw_xprt = r_xprt;
list_add(&r->mw_list, &buf->rb_mws); list_add(&r->mw_list, &buf->rb_mws);
list_add(&r->mw_all, &buf->rb_all); list_add(&r->mw_all, &buf->rb_all);
} }
return 0; return 0;
out_fmr_err:
rc = PTR_ERR(r->fmr.fmr);
dprintk("RPC: %s: ib_alloc_fmr status %i\n", __func__, rc);
kfree(r->fmr.physaddrs);
out_free:
kfree(r);
out:
return rc;
} }
/* Use the ib_map_phys_fmr() verb to register a memory region /* Use the ib_map_phys_fmr() verb to register a memory region
...@@ -235,16 +284,6 @@ fmr_op_map(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr_seg *seg, ...@@ -235,16 +284,6 @@ fmr_op_map(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr_seg *seg,
return rc; return rc;
} }
static void
__fmr_dma_unmap(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr_seg *seg)
{
struct ib_device *device = r_xprt->rx_ia.ri_device;
int nsegs = seg->mr_nsegs;
while (nsegs--)
rpcrdma_unmap_one(device, seg++);
}
/* Invalidate all memory regions that were registered for "req". /* Invalidate all memory regions that were registered for "req".
* *
* Sleeps until it is safe for the host CPU to access the * Sleeps until it is safe for the host CPU to access the
...@@ -337,18 +376,11 @@ static void ...@@ -337,18 +376,11 @@ static void
fmr_op_destroy(struct rpcrdma_buffer *buf) fmr_op_destroy(struct rpcrdma_buffer *buf)
{ {
struct rpcrdma_mw *r; struct rpcrdma_mw *r;
int rc;
while (!list_empty(&buf->rb_all)) { while (!list_empty(&buf->rb_all)) {
r = list_entry(buf->rb_all.next, struct rpcrdma_mw, mw_all); r = list_entry(buf->rb_all.next, struct rpcrdma_mw, mw_all);
list_del(&r->mw_all); list_del(&r->mw_all);
kfree(r->fmr.physaddrs); __fmr_release(r);
rc = ib_dealloc_fmr(r->fmr.fmr);
if (rc)
dprintk("RPC: %s: ib_dealloc_fmr failed %i\n",
__func__, rc);
kfree(r); kfree(r);
} }
} }
......
...@@ -98,6 +98,50 @@ frwr_destroy_recovery_wq(void) ...@@ -98,6 +98,50 @@ frwr_destroy_recovery_wq(void)
destroy_workqueue(wq); destroy_workqueue(wq);
} }
static int
__frwr_init(struct rpcrdma_mw *r, struct ib_pd *pd, unsigned int depth)
{
struct rpcrdma_frmr *f = &r->frmr;
int rc;
f->fr_mr = ib_alloc_mr(pd, IB_MR_TYPE_MEM_REG, depth);
if (IS_ERR(f->fr_mr))
goto out_mr_err;
r->mw_sg = kcalloc(depth, sizeof(*r->mw_sg), GFP_KERNEL);
if (!r->mw_sg)
goto out_list_err;
sg_init_table(r->mw_sg, depth);
init_completion(&f->fr_linv_done);
return 0;
out_mr_err:
rc = PTR_ERR(f->fr_mr);
dprintk("RPC: %s: ib_alloc_mr status %i\n",
__func__, rc);
return rc;
out_list_err:
rc = -ENOMEM;
dprintk("RPC: %s: sg allocation failure\n",
__func__);
ib_dereg_mr(f->fr_mr);
return rc;
}
static void
__frwr_release(struct rpcrdma_mw *r)
{
int rc;
rc = ib_dereg_mr(r->frmr.fr_mr);
if (rc)
pr_err("rpcrdma: final ib_dereg_mr for %p returned %i\n",
r, rc);
kfree(r->mw_sg);
}
static int static int
__frwr_reset_mr(struct rpcrdma_ia *ia, struct rpcrdma_mw *r) __frwr_reset_mr(struct rpcrdma_ia *ia, struct rpcrdma_mw *r)
{ {
...@@ -164,52 +208,6 @@ __frwr_queue_recovery(struct rpcrdma_mw *r) ...@@ -164,52 +208,6 @@ __frwr_queue_recovery(struct rpcrdma_mw *r)
queue_work(frwr_recovery_wq, &r->mw_work); queue_work(frwr_recovery_wq, &r->mw_work);
} }
static int
__frwr_init(struct rpcrdma_mw *r, struct ib_pd *pd, unsigned int depth)
{
struct rpcrdma_frmr *f = &r->frmr;
int rc;
f->fr_mr = ib_alloc_mr(pd, IB_MR_TYPE_MEM_REG, depth);
if (IS_ERR(f->fr_mr))
goto out_mr_err;
r->mw_sg = kcalloc(depth, sizeof(*r->mw_sg), GFP_KERNEL);
if (!r->mw_sg)
goto out_list_err;
sg_init_table(r->mw_sg, depth);
init_completion(&f->fr_linv_done);
return 0;
out_mr_err:
rc = PTR_ERR(f->fr_mr);
dprintk("RPC: %s: ib_alloc_mr status %i\n",
__func__, rc);
return rc;
out_list_err:
rc = -ENOMEM;
dprintk("RPC: %s: sg allocation failure\n",
__func__);
ib_dereg_mr(f->fr_mr);
return rc;
}
static void
__frwr_release(struct rpcrdma_mw *r)
{
int rc;
rc = ib_dereg_mr(r->frmr.fr_mr);
if (rc)
dprintk("RPC: %s: ib_dereg_mr status %i\n",
__func__, rc);
kfree(r->mw_sg);
}
static int static int
frwr_op_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep, frwr_op_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep,
struct rpcrdma_create_data_internal *cdata) struct rpcrdma_create_data_internal *cdata)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment