frwr_ops.c 18.9 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0
2
/*
3
 * Copyright (c) 2015, 2017 Oracle.  All rights reserved.
4 5 6 7
 * Copyright (c) 2003-2007 Network Appliance, Inc. All rights reserved.
 */

/* Lightweight memory registration using Fast Registration Work
8
 * Requests (FRWR).
9
 *
10 11
 * FRWR features ordered asynchronous registration and invalidation
 * of arbitrarily-sized memory regions. This is the fastest and safest
12 13 14
 * but most complex memory registration mode.
 */

15 16
/* Normal operation
 *
17
 * A Memory Region is prepared for RDMA Read or Write using a FAST_REG
18
 * Work Request (frwr_map). When the RDMA operation is finished, this
19
 * Memory Region is invalidated using a LOCAL_INV Work Request
20
 * (frwr_unmap_async and frwr_unmap_sync).
21
 *
22 23 24
 * Typically FAST_REG Work Requests are not signaled, and neither are
 * RDMA Send Work Requests (with the exception of signaling occasionally
 * to prevent provider work queue overflows). This greatly reduces HCA
25 26 27 28 29
 * interrupt workload.
 */

/* Transport recovery
 *
30 31 32 33 34 35 36 37 38 39 40
 * frwr_map and frwr_unmap_* cannot run at the same time the transport
 * connect worker is running. The connect worker holds the transport
 * send lock, just as ->send_request does. This prevents frwr_map and
 * the connect worker from running concurrently. When a connection is
 * closed, the Receive completion queue is drained before the allowing
 * the connect worker to get control. This prevents frwr_unmap and the
 * connect worker from running concurrently.
 *
 * When the underlying transport disconnects, MRs that are in flight
 * are flushed and are likely unusable. Thus all flushed MRs are
 * destroyed. New MRs are created on demand.
41 42
 */

43
#include <linux/sunrpc/rpc_rdma.h>
C
Chuck Lever 已提交
44
#include <linux/sunrpc/svc_rdma.h>
45

46
#include "xprt_rdma.h"
47
#include <trace/events/rpcrdma.h>
48 49 50 51 52

#if IS_ENABLED(CONFIG_SUNRPC_DEBUG)
# define RPCDBG_FACILITY	RPCDBG_TRANS
#endif

53 54
/**
 * frwr_is_supported - Check if device supports FRWR
55
 * @device: interface adapter to check
56 57 58
 *
 * Returns true if device supports FRWR, otherwise false
 */
59
bool frwr_is_supported(struct ib_device *device)
60
{
61
	struct ib_device_attr *attrs = &device->attrs;
62 63 64 65 66 67 68 69 70

	if (!(attrs->device_cap_flags & IB_DEVICE_MEM_MGT_EXTENSIONS))
		goto out_not_supported;
	if (attrs->max_fast_reg_page_list_len == 0)
		goto out_not_supported;
	return true;

out_not_supported:
	pr_info("rpcrdma: 'frwr' mode is not supported by device %s\n",
71
		device->name);
72 73 74
	return false;
}

75 76 77 78 79 80
/**
 * frwr_release_mr - Destroy one MR
 * @mr: MR allocated by frwr_init_mr
 *
 */
void frwr_release_mr(struct rpcrdma_mr *mr)
81 82 83 84 85
{
	int rc;

	rc = ib_dereg_mr(mr->frwr.fr_mr);
	if (rc)
86
		trace_xprtrdma_frwr_dereg(mr, rc);
87 88 89 90 91 92 93 94 95 96 97 98 99 100 101
	kfree(mr->mr_sg);
	kfree(mr);
}

/* MRs are dynamically allocated, so simply clean up and release the MR.
 * A replacement MR will subsequently be allocated on demand.
 */
static void
frwr_mr_recycle_worker(struct work_struct *work)
{
	struct rpcrdma_mr *mr = container_of(work, struct rpcrdma_mr, mr_recycle);
	struct rpcrdma_xprt *r_xprt = mr->mr_xprt;

	trace_xprtrdma_mr_recycle(mr);

102
	if (mr->mr_dir != DMA_NONE) {
103
		trace_xprtrdma_mr_unmap(mr);
104
		ib_dma_unmap_sg(r_xprt->rx_ia.ri_id->device,
105
				mr->mr_sg, mr->mr_nents, mr->mr_dir);
106
		mr->mr_dir = DMA_NONE;
107 108 109 110 111 112
	}

	spin_lock(&r_xprt->rx_buf.rb_mrlock);
	list_del(&mr->mr_all);
	r_xprt->rx_stats.mrs_recycled++;
	spin_unlock(&r_xprt->rx_buf.rb_mrlock);
113 114

	frwr_release_mr(mr);
115 116
}

117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136
/* frwr_reset - Place MRs back on the free list
 * @req: request to reset
 *
 * Used after a failed marshal. For FRWR, this means the MRs
 * don't have to be fully released and recreated.
 *
 * NB: This is safe only as long as none of @req's MRs are
 * involved with an ongoing asynchronous FAST_REG or LOCAL_INV
 * Work Request.
 */
void frwr_reset(struct rpcrdma_req *req)
{
	while (!list_empty(&req->rl_registered)) {
		struct rpcrdma_mr *mr;

		mr = rpcrdma_mr_pop(&req->rl_registered);
		rpcrdma_mr_unmap_and_put(mr);
	}
}

137 138 139 140 141 142 143 144 145
/**
 * frwr_init_mr - Initialize one MR
 * @ia: interface adapter
 * @mr: generic MR to prepare for FRWR
 *
 * Returns zero if successful. Otherwise a negative errno
 * is returned.
 */
int frwr_init_mr(struct rpcrdma_ia *ia, struct rpcrdma_mr *mr)
146
{
147
	unsigned int depth = ia->ri_max_frwr_depth;
C
Chuck Lever 已提交
148 149
	struct scatterlist *sg;
	struct ib_mr *frmr;
150 151
	int rc;

C
Chuck Lever 已提交
152 153
	frmr = ib_alloc_mr(ia->ri_pd, ia->ri_mrtype, depth);
	if (IS_ERR(frmr))
154 155
		goto out_mr_err;

C
Chuck Lever 已提交
156 157
	sg = kcalloc(depth, sizeof(*sg), GFP_KERNEL);
	if (!sg)
158 159
		goto out_list_err;

C
Chuck Lever 已提交
160
	mr->frwr.fr_mr = frmr;
161
	mr->mr_dir = DMA_NONE;
162
	INIT_LIST_HEAD(&mr->mr_list);
163
	INIT_WORK(&mr->mr_recycle, frwr_mr_recycle_worker);
C
Chuck Lever 已提交
164 165 166 167
	init_completion(&mr->frwr.fr_linv_done);

	sg_init_table(sg, depth);
	mr->mr_sg = sg;
168 169 170
	return 0;

out_mr_err:
C
Chuck Lever 已提交
171
	rc = PTR_ERR(frmr);
172
	trace_xprtrdma_frwr_alloc(mr, rc);
173 174 175 176 177
	return rc;

out_list_err:
	dprintk("RPC:       %s: sg allocation failure\n",
		__func__);
C
Chuck Lever 已提交
178 179
	ib_dereg_mr(frmr);
	return -ENOMEM;
180 181
}

182 183 184 185 186 187
/**
 * frwr_open - Prepare an endpoint for use with FRWR
 * @ia: interface adapter this endpoint will use
 * @ep: endpoint to prepare
 *
 * On success, sets:
188 189
 *	ep->rep_attr.cap.max_send_wr
 *	ep->rep_attr.cap.max_recv_wr
190
 *	ep->rep_max_requests
191 192 193 194 195
 *	ia->ri_max_segs
 *
 * And these FRWR-related fields:
 *	ia->ri_max_frwr_depth
 *	ia->ri_mrtype
196 197
 *
 * On failure, a negative errno is returned.
198
 */
199
int frwr_open(struct rpcrdma_ia *ia, struct rpcrdma_ep *ep)
C
Chuck Lever 已提交
200
{
201
	struct ib_device_attr *attrs = &ia->ri_id->device->attrs;
202
	int max_qp_wr, depth, delta;
C
Chuck Lever 已提交
203

C
Chuck Lever 已提交
204 205 206 207
	ia->ri_mrtype = IB_MR_TYPE_MEM_REG;
	if (attrs->device_cap_flags & IB_DEVICE_SG_GAPS_REG)
		ia->ri_mrtype = IB_MR_TYPE_SG_GAPS;

C
Chuck Lever 已提交
208 209 210 211 212 213 214 215 216 217 218
	/* Quirk: Some devices advertise a large max_fast_reg_page_list_len
	 * capability, but perform optimally when the MRs are not larger
	 * than a page.
	 */
	if (attrs->max_sge_rd > 1)
		ia->ri_max_frwr_depth = attrs->max_sge_rd;
	else
		ia->ri_max_frwr_depth = attrs->max_fast_reg_page_list_len;
	if (ia->ri_max_frwr_depth > RPCRDMA_MAX_DATA_SEGS)
		ia->ri_max_frwr_depth = RPCRDMA_MAX_DATA_SEGS;
	dprintk("RPC:       %s: max FR page list depth = %u\n",
219 220 221 222 223 224 225 226 227
		__func__, ia->ri_max_frwr_depth);

	/* Add room for frwr register and invalidate WRs.
	 * 1. FRWR reg WR for head
	 * 2. FRWR invalidate WR for head
	 * 3. N FRWR reg WRs for pagelist
	 * 4. N FRWR invalidate WRs for pagelist
	 * 5. FRWR reg WR for tail
	 * 6. FRWR invalidate WR for tail
C
Chuck Lever 已提交
228 229 230 231
	 * 7. The RDMA_SEND WR
	 */
	depth = 7;

232
	/* Calculate N if the device max FRWR depth is smaller than
C
Chuck Lever 已提交
233 234
	 * RPCRDMA_MAX_DATA_SEGS.
	 */
235 236
	if (ia->ri_max_frwr_depth < RPCRDMA_MAX_DATA_SEGS) {
		delta = RPCRDMA_MAX_DATA_SEGS - ia->ri_max_frwr_depth;
C
Chuck Lever 已提交
237
		do {
238 239
			depth += 2; /* FRWR reg + invalidate */
			delta -= ia->ri_max_frwr_depth;
C
Chuck Lever 已提交
240 241 242
		} while (delta > 0);
	}

243
	max_qp_wr = ia->ri_id->device->attrs.max_qp_wr;
244 245 246 247
	max_qp_wr -= RPCRDMA_BACKWARD_WRS;
	max_qp_wr -= 1;
	if (max_qp_wr < RPCRDMA_MIN_SLOT_TABLE)
		return -ENOMEM;
248 249 250
	if (ep->rep_max_requests > max_qp_wr)
		ep->rep_max_requests = max_qp_wr;
	ep->rep_attr.cap.max_send_wr = ep->rep_max_requests * depth;
251
	if (ep->rep_attr.cap.max_send_wr > max_qp_wr) {
252 253
		ep->rep_max_requests = max_qp_wr / depth;
		if (!ep->rep_max_requests)
C
Chuck Lever 已提交
254
			return -EINVAL;
255
		ep->rep_attr.cap.max_send_wr = ep->rep_max_requests * depth;
C
Chuck Lever 已提交
256
	}
257 258
	ep->rep_attr.cap.max_send_wr += RPCRDMA_BACKWARD_WRS;
	ep->rep_attr.cap.max_send_wr += 1; /* for ib_drain_sq */
259
	ep->rep_attr.cap.max_recv_wr = ep->rep_max_requests;
260 261
	ep->rep_attr.cap.max_recv_wr += RPCRDMA_BACKWARD_WRS;
	ep->rep_attr.cap.max_recv_wr += 1; /* for ib_drain_rq */
C
Chuck Lever 已提交
262

263
	ia->ri_max_segs = max_t(unsigned int, 1, RPCRDMA_MAX_DATA_SEGS /
264
				ia->ri_max_frwr_depth);
265 266 267 268
	/* Reply chunks require segments for head and tail buffers */
	ia->ri_max_segs += 2;
	if (ia->ri_max_segs > RPCRDMA_MAX_HDR_SEGS)
		ia->ri_max_segs = RPCRDMA_MAX_HDR_SEGS;
C
Chuck Lever 已提交
269 270 271
	return 0;
}

272 273 274 275 276 277 278
/**
 * frwr_maxpages - Compute size of largest payload
 * @r_xprt: transport
 *
 * Returns maximum size of an RPC message, in pages.
 *
 * FRWR mode conveys a list of pages per chunk segment. The
279 280
 * maximum length of that list is the FRWR page list depth.
 */
281
size_t frwr_maxpages(struct rpcrdma_xprt *r_xprt)
282 283 284 285
{
	struct rpcrdma_ia *ia = &r_xprt->rx_ia;

	return min_t(unsigned int, RPCRDMA_MAX_DATA_SEGS,
286
		     (ia->ri_max_segs - 2) * ia->ri_max_frwr_depth);
287 288
}

289 290 291 292 293 294
/**
 * frwr_map - Register a memory region
 * @r_xprt: controlling transport
 * @seg: memory region co-ordinates
 * @nsegs: number of segments remaining
 * @writing: true when RDMA Write will be used
295
 * @xid: XID of RPC using the registered memory
296 297 298
 * @out: initialized MR
 *
 * Prepare a REG_MR Work Request to register a memory region
299
 * for remote access via RDMA READ or RDMA WRITE.
300 301 302
 *
 * Returns the next segment or a negative errno pointer.
 * On success, the prepared MR is planted in @out.
303
 */
304 305
struct rpcrdma_mr_seg *frwr_map(struct rpcrdma_xprt *r_xprt,
				struct rpcrdma_mr_seg *seg,
C
Chuck Lever 已提交
306
				int nsegs, bool writing, __be32 xid,
307
				struct rpcrdma_mr **out)
308 309
{
	struct rpcrdma_ia *ia = &r_xprt->rx_ia;
C
Chuck Lever 已提交
310
	bool holes_ok = ia->ri_mrtype == IB_MR_TYPE_SG_GAPS;
C
Chuck Lever 已提交
311 312
	struct rpcrdma_mr *mr;
	struct ib_mr *ibmr;
313
	struct ib_reg_wr *reg_wr;
314
	int i, n;
315 316
	u8 key;

C
Chuck Lever 已提交
317 318 319
	mr = rpcrdma_mr_get(r_xprt);
	if (!mr)
		goto out_getmr_err;
320 321 322

	if (nsegs > ia->ri_max_frwr_depth)
		nsegs = ia->ri_max_frwr_depth;
323 324
	for (i = 0; i < nsegs;) {
		if (seg->mr_page)
C
Chuck Lever 已提交
325
			sg_set_page(&mr->mr_sg[i],
326 327 328 329
				    seg->mr_page,
				    seg->mr_len,
				    offset_in_page(seg->mr_offset));
		else
C
Chuck Lever 已提交
330
			sg_set_buf(&mr->mr_sg[i], seg->mr_offset,
331 332
				   seg->mr_len);

333 334
		++seg;
		++i;
C
Chuck Lever 已提交
335 336
		if (holes_ok)
			continue;
337 338 339 340
		if ((i < nsegs && offset_in_page(seg->mr_offset)) ||
		    offset_in_page((seg-1)->mr_offset + (seg-1)->mr_len))
			break;
	}
C
Chuck Lever 已提交
341
	mr->mr_dir = rpcrdma_data_dir(writing);
342

343 344
	mr->mr_nents =
		ib_dma_map_sg(ia->ri_id->device, mr->mr_sg, i, mr->mr_dir);
C
Chuck Lever 已提交
345
	if (!mr->mr_nents)
346 347
		goto out_dmamap_err;

C
Chuck Lever 已提交
348
	ibmr = mr->frwr.fr_mr;
C
Chuck Lever 已提交
349 350
	n = ib_map_mr_sg(ibmr, mr->mr_sg, mr->mr_nents, NULL, PAGE_SIZE);
	if (unlikely(n != mr->mr_nents))
351
		goto out_mapmr_err;
352

353
	ibmr->iova &= 0x00000000ffffffff;
C
Chuck Lever 已提交
354
	ibmr->iova |= ((u64)be32_to_cpu(xid)) << 32;
C
Chuck Lever 已提交
355 356
	key = (u8)(ibmr->rkey & 0x000000FF);
	ib_update_fast_reg_key(ibmr, ++key);
357

C
Chuck Lever 已提交
358
	reg_wr = &mr->frwr.fr_regwr;
C
Chuck Lever 已提交
359 360
	reg_wr->mr = ibmr;
	reg_wr->key = ibmr->rkey;
361 362 363
	reg_wr->access = writing ?
			 IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE :
			 IB_ACCESS_REMOTE_READ;
364

C
Chuck Lever 已提交
365 366 367
	mr->mr_handle = ibmr->rkey;
	mr->mr_length = ibmr->length;
	mr->mr_offset = ibmr->iova;
368
	trace_xprtrdma_mr_map(mr);
369

C
Chuck Lever 已提交
370
	*out = mr;
371
	return seg;
372

373 374 375 376
out_getmr_err:
	xprt_wait_for_buffer_space(&r_xprt->rx_xprt);
	return ERR_PTR(-EAGAIN);

377
out_dmamap_err:
378
	mr->mr_dir = DMA_NONE;
379
	trace_xprtrdma_frwr_sgerr(mr, i);
C
Chuck Lever 已提交
380
	rpcrdma_mr_put(mr);
381
	return ERR_PTR(-EIO);
382 383

out_mapmr_err:
384
	trace_xprtrdma_frwr_maperr(mr, n);
385
	rpcrdma_mr_recycle(mr);
386
	return ERR_PTR(-EIO);
387
}
388

C
Chuck Lever 已提交
389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405
/**
 * frwr_wc_fastreg - Invoked by RDMA provider for a flushed FastReg WC
 * @cq:	completion queue (ignored)
 * @wc:	completed WR
 *
 */
static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc)
{
	struct ib_cqe *cqe = wc->wr_cqe;
	struct rpcrdma_frwr *frwr =
		container_of(cqe, struct rpcrdma_frwr, fr_cqe);

	/* WARNING: Only wr_cqe and status are reliable at this point */
	trace_xprtrdma_wc_fastreg(wc, frwr);
	/* The MR will get recycled when the associated req is retransmitted */
}

406 407 408 409
/**
 * frwr_send - post Send WR containing the RPC Call message
 * @ia: interface adapter
 * @req: Prepared RPC Call
410
 *
411
 * For FRWR, chain any FastReg WRs to the Send WR. Only a
412 413
 * single ib_post_send call is needed to register memory
 * and then post the Send WR.
414 415
 *
 * Returns the result of ib_post_send.
416
 */
417
int frwr_send(struct rpcrdma_ia *ia, struct rpcrdma_req *req)
418
{
419
	struct ib_send_wr *post_wr;
420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438
	struct rpcrdma_mr *mr;

	post_wr = &req->rl_sendctx->sc_wr;
	list_for_each_entry(mr, &req->rl_registered, mr_list) {
		struct rpcrdma_frwr *frwr;

		frwr = &mr->frwr;

		frwr->fr_cqe.done = frwr_wc_fastreg;
		frwr->fr_regwr.wr.next = post_wr;
		frwr->fr_regwr.wr.wr_cqe = &frwr->fr_cqe;
		frwr->fr_regwr.wr.num_sge = 0;
		frwr->fr_regwr.wr.opcode = IB_WR_REG_MR;
		frwr->fr_regwr.wr.send_flags = 0;

		post_wr = &frwr->fr_regwr.wr;
	}

	/* If ib_post_send fails, the next ->send_request for
439
	 * @req will queue these MRs for recovery.
440
	 */
441
	return ib_post_send(ia->ri_id->qp, post_wr, NULL);
442 443
}

444 445 446 447 448
/**
 * frwr_reminv - handle a remotely invalidated mr on the @mrs list
 * @rep: Received reply
 * @mrs: list of MRs to check
 *
449
 */
450
void frwr_reminv(struct rpcrdma_rep *rep, struct list_head *mrs)
451
{
C
Chuck Lever 已提交
452
	struct rpcrdma_mr *mr;
453

C
Chuck Lever 已提交
454 455
	list_for_each_entry(mr, mrs, mr_list)
		if (mr->mr_handle == rep->rr_inv_rkey) {
456
			list_del_init(&mr->mr_list);
457
			trace_xprtrdma_mr_remoteinv(mr);
458
			rpcrdma_mr_unmap_and_put(mr);
459 460 461 462
			break;	/* only one invalidated MR per RPC */
		}
}

C
Chuck Lever 已提交
463 464 465 466 467 468 469 470
static void __frwr_release_mr(struct ib_wc *wc, struct rpcrdma_mr *mr)
{
	if (wc->status != IB_WC_SUCCESS)
		rpcrdma_mr_recycle(mr);
	else
		rpcrdma_mr_unmap_and_put(mr);
}

471
/**
C
Chuck Lever 已提交
472 473 474
 * frwr_wc_localinv - Invoked by RDMA provider for a LOCAL_INV WC
 * @cq:	completion queue (ignored)
 * @wc:	completed WR
475
 *
C
Chuck Lever 已提交
476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492
 */
static void frwr_wc_localinv(struct ib_cq *cq, struct ib_wc *wc)
{
	struct ib_cqe *cqe = wc->wr_cqe;
	struct rpcrdma_frwr *frwr =
		container_of(cqe, struct rpcrdma_frwr, fr_cqe);
	struct rpcrdma_mr *mr = container_of(frwr, struct rpcrdma_mr, frwr);

	/* WARNING: Only wr_cqe and status are reliable at this point */
	trace_xprtrdma_wc_li(wc, frwr);
	__frwr_release_mr(wc, mr);
}

/**
 * frwr_wc_localinv_wake - Invoked by RDMA provider for a LOCAL_INV WC
 * @cq:	completion queue (ignored)
 * @wc:	completed WR
493
 *
C
Chuck Lever 已提交
494
 * Awaken anyone waiting for an MR to finish being fenced.
495
 */
C
Chuck Lever 已提交
496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514
static void frwr_wc_localinv_wake(struct ib_cq *cq, struct ib_wc *wc)
{
	struct ib_cqe *cqe = wc->wr_cqe;
	struct rpcrdma_frwr *frwr =
		container_of(cqe, struct rpcrdma_frwr, fr_cqe);
	struct rpcrdma_mr *mr = container_of(frwr, struct rpcrdma_mr, frwr);

	/* WARNING: Only wr_cqe and status are reliable at this point */
	trace_xprtrdma_wc_li_wake(wc, frwr);
	complete(&frwr->fr_linv_done);
	__frwr_release_mr(wc, mr);
}

/**
 * frwr_unmap_sync - invalidate memory regions that were registered for @req
 * @r_xprt: controlling transport instance
 * @req: rpcrdma_req with a non-empty list of MRs to process
 *
 * Sleeps until it is safe for the host CPU to access the previously mapped
515 516 517 518
 * memory regions. This guarantees that registered MRs are properly fenced
 * from the server before the RPC consumer accesses the data in them. It
 * also ensures proper Send flow control: waking the next RPC waits until
 * this RPC has relinquished all its Send Queue entries.
C
Chuck Lever 已提交
519 520
 */
void frwr_unmap_sync(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
521
{
522 523
	struct ib_send_wr *first, **prev, *last;
	const struct ib_send_wr *bad_wr;
524
	struct rpcrdma_frwr *frwr;
C
Chuck Lever 已提交
525
	struct rpcrdma_mr *mr;
C
Chuck Lever 已提交
526
	int rc;
527

528
	/* ORDER: Invalidate all of the MRs first
529 530 531 532
	 *
	 * Chain the LOCAL_INV Work Requests and post them with
	 * a single ib_post_send() call.
	 */
533
	frwr = NULL;
C
Chuck Lever 已提交
534
	prev = &first;
C
Chuck Lever 已提交
535 536
	while (!list_empty(&req->rl_registered)) {
		mr = rpcrdma_mr_pop(&req->rl_registered);
C
Chuck Lever 已提交
537

538
		trace_xprtrdma_mr_localinv(mr);
C
Chuck Lever 已提交
539
		r_xprt->rx_stats.local_inv_needed++;
C
Chuck Lever 已提交
540

C
Chuck Lever 已提交
541
		frwr = &mr->frwr;
542 543
		frwr->fr_cqe.done = frwr_wc_localinv;
		last = &frwr->fr_invwr;
C
Chuck Lever 已提交
544
		last->next = NULL;
545
		last->wr_cqe = &frwr->fr_cqe;
C
Chuck Lever 已提交
546 547
		last->sg_list = NULL;
		last->num_sge = 0;
C
Chuck Lever 已提交
548
		last->opcode = IB_WR_LOCAL_INV;
C
Chuck Lever 已提交
549
		last->send_flags = IB_SEND_SIGNALED;
C
Chuck Lever 已提交
550
		last->ex.invalidate_rkey = mr->mr_handle;
551

C
Chuck Lever 已提交
552 553
		*prev = last;
		prev = &last->next;
554 555 556 557 558 559
	}

	/* Strong send queue ordering guarantees that when the
	 * last WR in the chain completes, all WRs in the chain
	 * are complete.
	 */
560 561
	frwr->fr_cqe.done = frwr_wc_localinv_wake;
	reinit_completion(&frwr->fr_linv_done);
562

563 564 565 566
	/* Transport disconnect drains the receive CQ before it
	 * replaces the QP. The RPC reply handler won't call us
	 * unless ri_id->qp is a valid pointer.
	 */
567
	bad_wr = NULL;
C
Chuck Lever 已提交
568 569
	rc = ib_post_send(r_xprt->rx_ia.ri_id->qp, first, &bad_wr);
	trace_xprtrdma_post_send(req, rc);
570

C
Chuck Lever 已提交
571 572 573
	/* The final LOCAL_INV WR in the chain is supposed to
	 * do the wake. If it was never posted, the wake will
	 * not happen, so don't wait in that case.
574
	 */
C
Chuck Lever 已提交
575 576 577 578
	if (bad_wr != first)
		wait_for_completion(&frwr->fr_linv_done);
	if (!rc)
		return;
579

C
Chuck Lever 已提交
580
	/* Recycle MRs in the LOCAL_INV chain that did not get posted.
581
	 */
582
	while (bad_wr) {
583 584
		frwr = container_of(bad_wr, struct rpcrdma_frwr,
				    fr_invwr);
C
Chuck Lever 已提交
585
		mr = container_of(frwr, struct rpcrdma_mr, frwr);
586
		bad_wr = bad_wr->next;
587

588 589
		list_del_init(&mr->mr_list);
		rpcrdma_mr_recycle(mr);
590
	}
591
}
592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689

/**
 * frwr_wc_localinv_done - Invoked by RDMA provider for a signaled LOCAL_INV WC
 * @cq:	completion queue (ignored)
 * @wc:	completed WR
 *
 */
static void frwr_wc_localinv_done(struct ib_cq *cq, struct ib_wc *wc)
{
	struct ib_cqe *cqe = wc->wr_cqe;
	struct rpcrdma_frwr *frwr =
		container_of(cqe, struct rpcrdma_frwr, fr_cqe);
	struct rpcrdma_mr *mr = container_of(frwr, struct rpcrdma_mr, frwr);

	/* WARNING: Only wr_cqe and status are reliable at this point */
	trace_xprtrdma_wc_li_done(wc, frwr);
	rpcrdma_complete_rqst(frwr->fr_req->rl_reply);
	__frwr_release_mr(wc, mr);
}

/**
 * frwr_unmap_async - invalidate memory regions that were registered for @req
 * @r_xprt: controlling transport instance
 * @req: rpcrdma_req with a non-empty list of MRs to process
 *
 * This guarantees that registered MRs are properly fenced from the
 * server before the RPC consumer accesses the data in them. It also
 * ensures proper Send flow control: waking the next RPC waits until
 * this RPC has relinquished all its Send Queue entries.
 */
void frwr_unmap_async(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
{
	struct ib_send_wr *first, *last, **prev;
	const struct ib_send_wr *bad_wr;
	struct rpcrdma_frwr *frwr;
	struct rpcrdma_mr *mr;
	int rc;

	/* Chain the LOCAL_INV Work Requests and post them with
	 * a single ib_post_send() call.
	 */
	frwr = NULL;
	prev = &first;
	while (!list_empty(&req->rl_registered)) {
		mr = rpcrdma_mr_pop(&req->rl_registered);

		trace_xprtrdma_mr_localinv(mr);
		r_xprt->rx_stats.local_inv_needed++;

		frwr = &mr->frwr;
		frwr->fr_cqe.done = frwr_wc_localinv;
		frwr->fr_req = req;
		last = &frwr->fr_invwr;
		last->next = NULL;
		last->wr_cqe = &frwr->fr_cqe;
		last->sg_list = NULL;
		last->num_sge = 0;
		last->opcode = IB_WR_LOCAL_INV;
		last->send_flags = IB_SEND_SIGNALED;
		last->ex.invalidate_rkey = mr->mr_handle;

		*prev = last;
		prev = &last->next;
	}

	/* Strong send queue ordering guarantees that when the
	 * last WR in the chain completes, all WRs in the chain
	 * are complete. The last completion will wake up the
	 * RPC waiter.
	 */
	frwr->fr_cqe.done = frwr_wc_localinv_done;

	/* Transport disconnect drains the receive CQ before it
	 * replaces the QP. The RPC reply handler won't call us
	 * unless ri_id->qp is a valid pointer.
	 */
	bad_wr = NULL;
	rc = ib_post_send(r_xprt->rx_ia.ri_id->qp, first, &bad_wr);
	trace_xprtrdma_post_send(req, rc);
	if (!rc)
		return;

	/* Recycle MRs in the LOCAL_INV chain that did not get posted.
	 */
	while (bad_wr) {
		frwr = container_of(bad_wr, struct rpcrdma_frwr, fr_invwr);
		mr = container_of(frwr, struct rpcrdma_mr, frwr);
		bad_wr = bad_wr->next;

		rpcrdma_mr_recycle(mr);
	}

	/* The final LOCAL_INV WR in the chain is supposed to
	 * do the wake. If it was never posted, the wake will
	 * not happen, so wake here in that case.
	 */
	rpcrdma_complete_rqst(req->rl_reply);
}