xref: /freebsd/sys/contrib/rdma/krping/krping.c (revision 63d1fd5970ec814904aa0f4580b10a0d302d08b2)
1 /*
2  * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
3  * Copyright (c) 2006-2009 Open Grid Computing, Inc. All rights reserved.
4  *
5  * This software is available to you under a choice of one of two
6  * licenses.  You may choose to be licensed under the terms of the GNU
7  * General Public License (GPL) Version 2, available from the file
8  * COPYING in the main directory of this source tree, or the
9  * OpenIB.org BSD license below:
10  *
11  *     Redistribution and use in source and binary forms, with or
12  *     without modification, are permitted provided that the following
13  *     conditions are met:
14  *
15  *      - Redistributions of source code must retain the above
16  *        copyright notice, this list of conditions and the following
17  *        disclaimer.
18  *
19  *      - Redistributions in binary form must reproduce the above
20  *        copyright notice, this list of conditions and the following
21  *        disclaimer in the documentation and/or other materials
22  *        provided with the distribution.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31  * SOFTWARE.
32  */
33 
34 #include <sys/cdefs.h>
35 __FBSDID("$FreeBSD$");
36 
37 #include <linux/module.h>
38 #include <linux/moduleparam.h>
39 #include <linux/slab.h>
40 #include <linux/err.h>
41 #include <linux/string.h>
42 #include <linux/list.h>
43 #include <linux/in.h>
44 #include <linux/device.h>
45 #include <linux/pci.h>
46 #include <linux/sched.h>
47 
48 #include <asm/atomic.h>
49 
50 #include <rdma/ib_verbs.h>
51 #include <rdma/rdma_cm.h>
52 
53 #include "krping.h"
54 #include "getopt.h"
55 
56 extern int krping_debug;
57 #define DEBUG_LOG(cb, x...) if (krping_debug) log(LOG_INFO, x)
58 #define PRINTF(cb, x...) log(LOG_INFO, x)
59 #define BIND_INFO 1
60 
61 MODULE_AUTHOR("Steve Wise");
62 MODULE_DESCRIPTION("RDMA ping client/server");
63 MODULE_LICENSE("Dual BSD/GPL");
64 MODULE_VERSION(krping, 1);
65 MODULE_DEPEND(krping, linuxkpi, 1, 1, 1);
66 
67 static __inline uint64_t
68 get_cycles(void)
69 {
70 	uint32_t low, high;
71 	__asm __volatile("rdtsc" : "=a" (low), "=d" (high));
72 	return (low | ((u_int64_t)high << 32));
73 }
74 
75 typedef uint64_t cycles_t;
76 
77 enum mem_type {
78 	DMA = 1,
79 	FASTREG = 2,
80 	MW = 3,
81 	MR = 4
82 };
83 
84 static const struct krping_option krping_opts[] = {
85 	{"count", OPT_INT, 'C'},
86 	{"size", OPT_INT, 'S'},
87 	{"addr", OPT_STRING, 'a'},
88 	{"port", OPT_INT, 'p'},
89 	{"verbose", OPT_NOPARAM, 'v'},
90 	{"validate", OPT_NOPARAM, 'V'},
91 	{"server", OPT_NOPARAM, 's'},
92 	{"client", OPT_NOPARAM, 'c'},
93 	{"mem_mode", OPT_STRING, 'm'},
94 	{"server_inv", OPT_NOPARAM, 'I'},
95  	{"wlat", OPT_NOPARAM, 'l'},
96  	{"rlat", OPT_NOPARAM, 'L'},
97  	{"bw", OPT_NOPARAM, 'B'},
98  	{"duplex", OPT_NOPARAM, 'd'},
99  	{"txdepth", OPT_INT, 'T'},
100  	{"poll", OPT_NOPARAM, 'P'},
101  	{"local_dma_lkey", OPT_NOPARAM, 'Z'},
102  	{"read_inv", OPT_NOPARAM, 'R'},
103  	{"fr", OPT_INT, 'f'},
104 	{NULL, 0, 0}
105 };
106 
107 #define htonll(x) cpu_to_be64((x))
108 #define ntohll(x) cpu_to_be64((x))
109 
110 static struct mutex krping_mutex;
111 
112 /*
113  * List of running krping threads.
114  */
115 static LIST_HEAD(krping_cbs);
116 
117 /*
118  * krping "ping/pong" loop:
119  * 	client sends source rkey/addr/len
120  *	server receives source rkey/add/len
121  *	server rdma reads "ping" data from source
122  * 	server sends "go ahead" on rdma read completion
123  *	client sends sink rkey/addr/len
124  * 	server receives sink rkey/addr/len
125  * 	server rdma writes "pong" data to sink
126  * 	server sends "go ahead" on rdma write completion
127  * 	<repeat loop>
128  */
129 
130 /*
131  * These states are used to signal events between the completion handler
132  * and the main client or server thread.
133  *
134  * Once CONNECTED, they cycle through RDMA_READ_ADV, RDMA_WRITE_ADV,
135  * and RDMA_WRITE_COMPLETE for each ping.
136  */
137 enum test_state {
138 	IDLE = 1,
139 	CONNECT_REQUEST,
140 	ADDR_RESOLVED,
141 	ROUTE_RESOLVED,
142 	CONNECTED,
143 	RDMA_READ_ADV,
144 	RDMA_READ_COMPLETE,
145 	RDMA_WRITE_ADV,
146 	RDMA_WRITE_COMPLETE,
147 	ERROR
148 };
149 
150 struct krping_rdma_info {
151 	uint64_t buf;
152 	uint32_t rkey;
153 	uint32_t size;
154 };
155 
156 /*
157  * Default max buffer size for IO...
158  */
159 #define RPING_BUFSIZE 128*1024
160 #define RPING_SQ_DEPTH 64
161 
162 /*
163  * Control block struct.
164  */
165 struct krping_cb {
166 	void *cookie;
167 	int server;			/* 0 iff client */
168 	struct ib_cq *cq;
169 	struct ib_pd *pd;
170 	struct ib_qp *qp;
171 
172 	enum mem_type mem;
173 	struct ib_mr *dma_mr;
174 
175 	struct ib_fast_reg_page_list *page_list;
176 	int page_list_len;
177 	struct ib_send_wr fastreg_wr;
178 	struct ib_send_wr invalidate_wr;
179 	struct ib_mr *fastreg_mr;
180 	int server_invalidate;
181 	int read_inv;
182 	u8 key;
183 
184 	struct ib_mw *mw;
185 	struct ib_mw_bind bind_attr;
186 
187 	struct ib_recv_wr rq_wr;	/* recv work request record */
188 	struct ib_sge recv_sgl;		/* recv single SGE */
189 	struct krping_rdma_info recv_buf;/* malloc'd buffer */
190 	u64 recv_dma_addr;
191 	DECLARE_PCI_UNMAP_ADDR(recv_mapping)
192 	struct ib_mr *recv_mr;
193 
194 	struct ib_send_wr sq_wr;	/* send work requrest record */
195 	struct ib_sge send_sgl;
196 	struct krping_rdma_info send_buf;/* single send buf */
197 	u64 send_dma_addr;
198 	DECLARE_PCI_UNMAP_ADDR(send_mapping)
199 	struct ib_mr *send_mr;
200 
201 	struct ib_send_wr rdma_sq_wr;	/* rdma work request record */
202 	struct ib_sge rdma_sgl;		/* rdma single SGE */
203 	char *rdma_buf;			/* used as rdma sink */
204 	u64  rdma_dma_addr;
205 	DECLARE_PCI_UNMAP_ADDR(rdma_mapping)
206 	struct ib_mr *rdma_mr;
207 
208 	uint32_t remote_rkey;		/* remote guys RKEY */
209 	uint64_t remote_addr;		/* remote guys TO */
210 	uint32_t remote_len;		/* remote guys LEN */
211 
212 	char *start_buf;		/* rdma read src */
213 	u64  start_dma_addr;
214 	DECLARE_PCI_UNMAP_ADDR(start_mapping)
215 	struct ib_mr *start_mr;
216 
217 	enum test_state state;		/* used for cond/signalling */
218 	wait_queue_head_t sem;
219 	struct krping_stats stats;
220 
221 	uint16_t port;			/* dst port in NBO */
222 	struct in_addr addr;		/* dst addr in NBO */
223 	char *addr_str;			/* dst addr string */
224 	int verbose;			/* verbose logging */
225 	int count;			/* ping count */
226 	int size;			/* ping data size */
227 	int validate;			/* validate ping data */
228 	int wlat;			/* run wlat test */
229 	int rlat;			/* run rlat test */
230 	int bw;				/* run bw test */
231 	int duplex;			/* run bw full duplex test */
232 	int poll;			/* poll or block for rlat test */
233 	int txdepth;			/* SQ depth */
234 	int local_dma_lkey;		/* use 0 for lkey */
235 	int frtest;			/* fastreg test */
236 	int testnum;
237 
238 	/* CM stuff */
239 	struct rdma_cm_id *cm_id;	/* connection on client side,*/
240 					/* listener on server side. */
241 	struct rdma_cm_id *child_cm_id;	/* connection on server side */
242 	struct list_head list;
243 };
244 
245 static int krping_cma_event_handler(struct rdma_cm_id *cma_id,
246 				   struct rdma_cm_event *event)
247 {
248 	int ret;
249 	struct krping_cb *cb = cma_id->context;
250 
251 	DEBUG_LOG(cb, "cma_event type %d cma_id %p (%s)\n", event->event,
252 	    cma_id, (cma_id == cb->cm_id) ? "parent" : "child");
253 
254 	switch (event->event) {
255 	case RDMA_CM_EVENT_ADDR_RESOLVED:
256 		cb->state = ADDR_RESOLVED;
257 		ret = rdma_resolve_route(cma_id, 2000);
258 		if (ret) {
259 			PRINTF(cb, "rdma_resolve_route error %d\n", ret);
260 			wake_up_interruptible(&cb->sem);
261 		}
262 		break;
263 
264 	case RDMA_CM_EVENT_ROUTE_RESOLVED:
265 		cb->state = ROUTE_RESOLVED;
266 		cb->child_cm_id = cma_id;
267 		wake_up_interruptible(&cb->sem);
268 		break;
269 
270 	case RDMA_CM_EVENT_CONNECT_REQUEST:
271 		if (cb->state == IDLE) {
272 			cb->state = CONNECT_REQUEST;
273 			cb->child_cm_id = cma_id;
274 		} else {
275 			PRINTF(cb, "Received connection request in wrong state"
276 			    " (%d)\n", cb->state);
277 		}
278 		DEBUG_LOG(cb, "child cma %p\n", cb->child_cm_id);
279 		wake_up_interruptible(&cb->sem);
280 		break;
281 
282 	case RDMA_CM_EVENT_ESTABLISHED:
283 		DEBUG_LOG(cb, "ESTABLISHED\n");
284 		if (!cb->server) {
285 			cb->state = CONNECTED;
286 		}
287 		wake_up_interruptible(&cb->sem);
288 		break;
289 
290 	case RDMA_CM_EVENT_ADDR_ERROR:
291 	case RDMA_CM_EVENT_ROUTE_ERROR:
292 	case RDMA_CM_EVENT_CONNECT_ERROR:
293 	case RDMA_CM_EVENT_UNREACHABLE:
294 	case RDMA_CM_EVENT_REJECTED:
295 		PRINTF(cb, "cma event %d, error %d\n", event->event,
296 		       event->status);
297 		cb->state = ERROR;
298 		wake_up_interruptible(&cb->sem);
299 		break;
300 
301 	case RDMA_CM_EVENT_DISCONNECTED:
302 		PRINTF(cb, "DISCONNECT EVENT...\n");
303 		cb->state = ERROR;
304 		wake_up_interruptible(&cb->sem);
305 		break;
306 
307 	case RDMA_CM_EVENT_DEVICE_REMOVAL:
308 		PRINTF(cb, "cma detected device removal!!!!\n");
309 		break;
310 
311 	default:
312 		PRINTF(cb, "oof bad type!\n");
313 		wake_up_interruptible(&cb->sem);
314 		break;
315 	}
316 	return 0;
317 }
318 
319 static int server_recv(struct krping_cb *cb, struct ib_wc *wc)
320 {
321 	if (wc->byte_len != sizeof(cb->recv_buf)) {
322 		PRINTF(cb, "Received bogus data, size %d\n",
323 		       wc->byte_len);
324 		return -1;
325 	}
326 
327 	cb->remote_rkey = ntohl(cb->recv_buf.rkey);
328 	cb->remote_addr = ntohll(cb->recv_buf.buf);
329 	cb->remote_len  = ntohl(cb->recv_buf.size);
330 	DEBUG_LOG(cb, "Received rkey %x addr %llx len %d from peer\n",
331 		  cb->remote_rkey, (unsigned long long)cb->remote_addr,
332 		  cb->remote_len);
333 
334 	if (cb->state <= CONNECTED || cb->state == RDMA_WRITE_COMPLETE)
335 		cb->state = RDMA_READ_ADV;
336 	else
337 		cb->state = RDMA_WRITE_ADV;
338 
339 	return 0;
340 }
341 
342 static int client_recv(struct krping_cb *cb, struct ib_wc *wc)
343 {
344 	if (wc->byte_len != sizeof(cb->recv_buf)) {
345 		PRINTF(cb, "Received bogus data, size %d\n",
346 		       wc->byte_len);
347 		return -1;
348 	}
349 
350 	if (cb->state == RDMA_READ_ADV)
351 		cb->state = RDMA_WRITE_ADV;
352 	else
353 		cb->state = RDMA_WRITE_COMPLETE;
354 
355 	return 0;
356 }
357 
358 static void krping_cq_event_handler(struct ib_cq *cq, void *ctx)
359 {
360 	struct krping_cb *cb = ctx;
361 	struct ib_wc wc;
362 	struct ib_recv_wr *bad_wr;
363 	int ret;
364 
365 	BUG_ON(cb->cq != cq);
366 	if (cb->state == ERROR) {
367 		PRINTF(cb, "cq completion in ERROR state\n");
368 		return;
369 	}
370 	if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest)
371 		ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
372 	while ((ret = ib_poll_cq(cb->cq, 1, &wc)) == 1) {
373 		if (wc.status) {
374 			if (wc.status == IB_WC_WR_FLUSH_ERR) {
375 				DEBUG_LOG(cb, "cq flushed\n");
376 				continue;
377 			} else {
378 				PRINTF(cb, "cq completion failed with "
379 				       "wr_id %jx status %d opcode %d vender_err %x\n",
380 					(uintmax_t)wc.wr_id, wc.status, wc.opcode, wc.vendor_err);
381 				goto error;
382 			}
383 		}
384 
385 		switch (wc.opcode) {
386 		case IB_WC_SEND:
387 			DEBUG_LOG(cb, "send completion\n");
388 			cb->stats.send_bytes += cb->send_sgl.length;
389 			cb->stats.send_msgs++;
390 			break;
391 
392 		case IB_WC_RDMA_WRITE:
393 			DEBUG_LOG(cb, "rdma write completion\n");
394 			cb->stats.write_bytes += cb->rdma_sq_wr.sg_list->length;
395 			cb->stats.write_msgs++;
396 			cb->state = RDMA_WRITE_COMPLETE;
397 			wake_up_interruptible(&cb->sem);
398 			break;
399 
400 		case IB_WC_RDMA_READ:
401 			DEBUG_LOG(cb, "rdma read completion\n");
402 			cb->stats.read_bytes += cb->rdma_sq_wr.sg_list->length;
403 			cb->stats.read_msgs++;
404 			cb->state = RDMA_READ_COMPLETE;
405 			wake_up_interruptible(&cb->sem);
406 			break;
407 
408 		case IB_WC_RECV:
409 			DEBUG_LOG(cb, "recv completion\n");
410 			cb->stats.recv_bytes += sizeof(cb->recv_buf);
411 			cb->stats.recv_msgs++;
412 			if (cb->wlat || cb->rlat || cb->bw || cb->frtest)
413 				ret = server_recv(cb, &wc);
414 			else
415 				ret = cb->server ? server_recv(cb, &wc) :
416 						   client_recv(cb, &wc);
417 			if (ret) {
418 				PRINTF(cb, "recv wc error: %d\n", ret);
419 				goto error;
420 			}
421 
422 			ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
423 			if (ret) {
424 				PRINTF(cb, "post recv error: %d\n",
425 				       ret);
426 				goto error;
427 			}
428 			wake_up_interruptible(&cb->sem);
429 			break;
430 
431 		default:
432 			PRINTF(cb,
433 			       "%s:%d Unexpected opcode %d, Shutting down\n",
434 			       __func__, __LINE__, wc.opcode);
435 			goto error;
436 		}
437 	}
438 	if (ret) {
439 		PRINTF(cb, "poll error %d\n", ret);
440 		goto error;
441 	}
442 	return;
443 error:
444 	cb->state = ERROR;
445 	wake_up_interruptible(&cb->sem);
446 }
447 
448 static int krping_accept(struct krping_cb *cb)
449 {
450 	struct rdma_conn_param conn_param;
451 	int ret;
452 
453 	DEBUG_LOG(cb, "accepting client connection request\n");
454 
455 	memset(&conn_param, 0, sizeof conn_param);
456 	conn_param.responder_resources = 1;
457 	conn_param.initiator_depth = 1;
458 
459 	ret = rdma_accept(cb->child_cm_id, &conn_param);
460 	if (ret) {
461 		PRINTF(cb, "rdma_accept error: %d\n", ret);
462 		return ret;
463 	}
464 
465 	if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest) {
466 		wait_event_interruptible(cb->sem, cb->state >= CONNECTED);
467 		if (cb->state == ERROR) {
468 			PRINTF(cb, "wait for CONNECTED state %d\n",
469 				cb->state);
470 			return -1;
471 		}
472 	}
473 	return 0;
474 }
475 
476 static void krping_setup_wr(struct krping_cb *cb)
477 {
478 	cb->recv_sgl.addr = cb->recv_dma_addr;
479 	cb->recv_sgl.length = sizeof cb->recv_buf;
480 	if (cb->local_dma_lkey)
481 		cb->recv_sgl.lkey = cb->qp->device->local_dma_lkey;
482 	else if (cb->mem == DMA)
483 		cb->recv_sgl.lkey = cb->dma_mr->lkey;
484 	else
485 		cb->recv_sgl.lkey = cb->recv_mr->lkey;
486 	cb->rq_wr.sg_list = &cb->recv_sgl;
487 	cb->rq_wr.num_sge = 1;
488 
489 	cb->send_sgl.addr = cb->send_dma_addr;
490 	cb->send_sgl.length = sizeof cb->send_buf;
491 	if (cb->local_dma_lkey)
492 		cb->send_sgl.lkey = cb->qp->device->local_dma_lkey;
493 	else if (cb->mem == DMA)
494 		cb->send_sgl.lkey = cb->dma_mr->lkey;
495 	else
496 		cb->send_sgl.lkey = cb->send_mr->lkey;
497 
498 	cb->sq_wr.opcode = IB_WR_SEND;
499 	cb->sq_wr.send_flags = IB_SEND_SIGNALED;
500 	cb->sq_wr.sg_list = &cb->send_sgl;
501 	cb->sq_wr.num_sge = 1;
502 
503 	if (cb->server || cb->wlat || cb->rlat || cb->bw || cb->frtest) {
504 		cb->rdma_sgl.addr = cb->rdma_dma_addr;
505 		if (cb->mem == MR)
506 			cb->rdma_sgl.lkey = cb->rdma_mr->lkey;
507 		cb->rdma_sq_wr.send_flags = IB_SEND_SIGNALED;
508 		cb->rdma_sq_wr.sg_list = &cb->rdma_sgl;
509 		cb->rdma_sq_wr.num_sge = 1;
510 	}
511 
512 	switch(cb->mem) {
513 	case FASTREG:
514 
515 		/*
516 		 * A chain of 2 WRs, INVALDATE_MR + FAST_REG_MR.
517 		 * both unsignaled.  The client uses them to reregister
518 		 * the rdma buffers with a new key each iteration.
519 		 */
520 		cb->fastreg_wr.opcode = IB_WR_FAST_REG_MR;
521 		cb->fastreg_wr.wr.fast_reg.page_shift = PAGE_SHIFT;
522 		cb->fastreg_wr.wr.fast_reg.length = cb->size;
523 		cb->fastreg_wr.wr.fast_reg.page_list = cb->page_list;
524 		cb->fastreg_wr.wr.fast_reg.page_list_len = cb->page_list_len;
525 
526 		cb->invalidate_wr.next = &cb->fastreg_wr;
527 		cb->invalidate_wr.opcode = IB_WR_LOCAL_INV;
528 		break;
529 	case MW:
530 		cb->bind_attr.wr_id = 0xabbaabba;
531 		cb->bind_attr.send_flags = 0; /* unsignaled */
532 #ifdef BIND_INFO
533 		cb->bind_attr.bind_info.length = cb->size;
534 #else
535 		cb->bind_attr.length = cb->size;
536 #endif
537 		break;
538 	default:
539 		break;
540 	}
541 }
542 
543 static int krping_setup_buffers(struct krping_cb *cb)
544 {
545 	int ret;
546 	struct ib_phys_buf buf;
547 	u64 iovbase;
548 
549 	DEBUG_LOG(cb, "krping_setup_buffers called on cb %p\n", cb);
550 
551 	cb->recv_dma_addr = ib_dma_map_single(cb->pd->device,
552 				   &cb->recv_buf,
553 				   sizeof(cb->recv_buf), DMA_BIDIRECTIONAL);
554 	pci_unmap_addr_set(cb, recv_mapping, cb->recv_dma_addr);
555 	cb->send_dma_addr = ib_dma_map_single(cb->pd->device,
556 					   &cb->send_buf, sizeof(cb->send_buf),
557 					   DMA_BIDIRECTIONAL);
558 	pci_unmap_addr_set(cb, send_mapping, cb->send_dma_addr);
559 
560 	if (cb->mem == DMA) {
561 		cb->dma_mr = ib_get_dma_mr(cb->pd, IB_ACCESS_LOCAL_WRITE|
562 					   IB_ACCESS_REMOTE_READ|
563 				           IB_ACCESS_REMOTE_WRITE);
564 		if (IS_ERR(cb->dma_mr)) {
565 			DEBUG_LOG(cb, "reg_dmamr failed\n");
566 			ret = PTR_ERR(cb->dma_mr);
567 			goto bail;
568 		}
569 	} else {
570 		if (!cb->local_dma_lkey) {
571 			buf.addr = cb->recv_dma_addr;
572 			buf.size = sizeof cb->recv_buf;
573 			DEBUG_LOG(cb, "recv buf dma_addr %jx size %d\n",
574 			    (uintmax_t)buf.addr, (int)buf.size);
575 			iovbase = cb->recv_dma_addr;
576 			cb->recv_mr = ib_reg_phys_mr(cb->pd, &buf, 1,
577 						     IB_ACCESS_LOCAL_WRITE,
578 						     &iovbase);
579 
580 			if (IS_ERR(cb->recv_mr)) {
581 				DEBUG_LOG(cb, "recv_buf reg_mr failed\n");
582 				ret = PTR_ERR(cb->recv_mr);
583 				goto bail;
584 			}
585 
586 			buf.addr = cb->send_dma_addr;
587 			buf.size = sizeof cb->send_buf;
588 			DEBUG_LOG(cb, "send buf dma_addr %jx size %d\n",
589 			    (uintmax_t)buf.addr, (int)buf.size);
590 			iovbase = cb->send_dma_addr;
591 			cb->send_mr = ib_reg_phys_mr(cb->pd, &buf, 1,
592 						     0, &iovbase);
593 
594 			if (IS_ERR(cb->send_mr)) {
595 				DEBUG_LOG(cb, "send_buf reg_mr failed\n");
596 				ret = PTR_ERR(cb->send_mr);
597 				goto bail;
598 			}
599 		}
600 	}
601 
602 	cb->rdma_buf = kmalloc(cb->size, GFP_KERNEL);
603 	if (!cb->rdma_buf) {
604 		DEBUG_LOG(cb, "rdma_buf malloc failed\n");
605 		ret = -ENOMEM;
606 		goto bail;
607 	}
608 
609 	cb->rdma_dma_addr = ib_dma_map_single(cb->pd->device,
610 			       cb->rdma_buf, cb->size,
611 			       DMA_BIDIRECTIONAL);
612 	pci_unmap_addr_set(cb, rdma_mapping, cb->rdma_dma_addr);
613 	if (cb->mem != DMA) {
614 		switch (cb->mem) {
615 		case FASTREG:
616 			cb->page_list_len = (((cb->size - 1) & PAGE_MASK) +
617 				PAGE_SIZE) >> PAGE_SHIFT;
618 			cb->page_list = ib_alloc_fast_reg_page_list(
619 						cb->pd->device,
620 						cb->page_list_len);
621 			if (IS_ERR(cb->page_list)) {
622 				DEBUG_LOG(cb, "recv_buf reg_mr failed\n");
623 				ret = PTR_ERR(cb->page_list);
624 				goto bail;
625 			}
626 			cb->fastreg_mr = ib_alloc_fast_reg_mr(cb->pd,
627 					cb->page_list->max_page_list_len);
628 			if (IS_ERR(cb->fastreg_mr)) {
629 				DEBUG_LOG(cb, "recv_buf reg_mr failed\n");
630 				ret = PTR_ERR(cb->fastreg_mr);
631 				goto bail;
632 			}
633 			DEBUG_LOG(cb, "fastreg rkey 0x%x page_list %p"
634 				" page_list_len %u\n", cb->fastreg_mr->rkey,
635 				cb->page_list, cb->page_list_len);
636 			break;
637 		case MW:
638 			cb->mw = ib_alloc_mw(cb->pd,IB_MW_TYPE_1);
639 			if (IS_ERR(cb->mw)) {
640 				DEBUG_LOG(cb, "recv_buf alloc_mw failed\n");
641 				ret = PTR_ERR(cb->mw);
642 				goto bail;
643 			}
644 			DEBUG_LOG(cb, "mw rkey 0x%x\n", cb->mw->rkey);
645 			/*FALLTHROUGH*/
646 		case MR:
647 			buf.addr = cb->rdma_dma_addr;
648 			buf.size = cb->size;
649 			iovbase = cb->rdma_dma_addr;
650 			cb->rdma_mr = ib_reg_phys_mr(cb->pd, &buf, 1,
651 						IB_ACCESS_LOCAL_WRITE|
652 					     IB_ACCESS_REMOTE_READ|
653 					     IB_ACCESS_REMOTE_WRITE,
654 					     &iovbase);
655 			if (IS_ERR(cb->rdma_mr)) {
656 				DEBUG_LOG(cb, "rdma_buf reg_mr failed\n");
657 				ret = PTR_ERR(cb->rdma_mr);
658 				goto bail;
659 			}
660 			DEBUG_LOG(cb, "rdma buf dma_addr %jx size %d mr rkey 0x%x\n",
661 				(uintmax_t)buf.addr, (int)buf.size, cb->rdma_mr->rkey);
662 			break;
663 		default:
664 			ret = -EINVAL;
665 			goto bail;
666 			break;
667 		}
668 	}
669 
670 	if (!cb->server || cb->wlat || cb->rlat || cb->bw || cb->frtest) {
671 
672 		cb->start_buf = kmalloc(cb->size, GFP_KERNEL);
673 		if (!cb->start_buf) {
674 			DEBUG_LOG(cb, "start_buf malloc failed\n");
675 			ret = -ENOMEM;
676 			goto bail;
677 		}
678 
679 		cb->start_dma_addr = ib_dma_map_single(cb->pd->device,
680 						   cb->start_buf, cb->size,
681 						   DMA_BIDIRECTIONAL);
682 		pci_unmap_addr_set(cb, start_mapping, cb->start_dma_addr);
683 
684 		if (cb->mem == MR || cb->mem == MW) {
685 			unsigned flags = IB_ACCESS_REMOTE_READ;
686 
687 			if (cb->wlat || cb->rlat || cb->bw || cb->frtest) {
688 				flags |= IB_ACCESS_LOCAL_WRITE |
689 					IB_ACCESS_REMOTE_WRITE;
690 			}
691 
692 			buf.addr = cb->start_dma_addr;
693 			buf.size = cb->size;
694 			DEBUG_LOG(cb, "start buf dma_addr %jx size %d\n",
695 				(uintmax_t)buf.addr, (int)buf.size);
696 			iovbase = cb->start_dma_addr;
697 			cb->start_mr = ib_reg_phys_mr(cb->pd, &buf, 1,
698 					     flags,
699 					     &iovbase);
700 
701 			if (IS_ERR(cb->start_mr)) {
702 				DEBUG_LOG(cb, "start_buf reg_mr failed\n");
703 				ret = PTR_ERR(cb->start_mr);
704 				goto bail;
705 			}
706 		}
707 	}
708 
709 	krping_setup_wr(cb);
710 	DEBUG_LOG(cb, "allocated & registered buffers...\n");
711 	return 0;
712 bail:
713 	if (cb->fastreg_mr && !IS_ERR(cb->fastreg_mr))
714 		ib_dereg_mr(cb->fastreg_mr);
715 	if (cb->mw && !IS_ERR(cb->mw))
716 		ib_dealloc_mw(cb->mw);
717 	if (cb->rdma_mr && !IS_ERR(cb->rdma_mr))
718 		ib_dereg_mr(cb->rdma_mr);
719 	if (cb->page_list && !IS_ERR(cb->page_list))
720 		ib_free_fast_reg_page_list(cb->page_list);
721 	if (cb->dma_mr && !IS_ERR(cb->dma_mr))
722 		ib_dereg_mr(cb->dma_mr);
723 	if (cb->recv_mr && !IS_ERR(cb->recv_mr))
724 		ib_dereg_mr(cb->recv_mr);
725 	if (cb->send_mr && !IS_ERR(cb->send_mr))
726 		ib_dereg_mr(cb->send_mr);
727 	if (cb->rdma_buf)
728 		kfree(cb->rdma_buf);
729 	if (cb->start_buf)
730 		kfree(cb->start_buf);
731 	return ret;
732 }
733 
734 static void krping_free_buffers(struct krping_cb *cb)
735 {
736 	DEBUG_LOG(cb, "krping_free_buffers called on cb %p\n", cb);
737 
738 	if (cb->dma_mr)
739 		ib_dereg_mr(cb->dma_mr);
740 	if (cb->send_mr)
741 		ib_dereg_mr(cb->send_mr);
742 	if (cb->recv_mr)
743 		ib_dereg_mr(cb->recv_mr);
744 	if (cb->rdma_mr)
745 		ib_dereg_mr(cb->rdma_mr);
746 	if (cb->start_mr)
747 		ib_dereg_mr(cb->start_mr);
748 	if (cb->fastreg_mr)
749 		ib_dereg_mr(cb->fastreg_mr);
750 	if (cb->mw)
751 		ib_dealloc_mw(cb->mw);
752 
753 	dma_unmap_single(cb->pd->device->dma_device,
754 			 pci_unmap_addr(cb, recv_mapping),
755 			 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL);
756 	dma_unmap_single(cb->pd->device->dma_device,
757 			 pci_unmap_addr(cb, send_mapping),
758 			 sizeof(cb->send_buf), DMA_BIDIRECTIONAL);
759 	dma_unmap_single(cb->pd->device->dma_device,
760 			 pci_unmap_addr(cb, rdma_mapping),
761 			 cb->size, DMA_BIDIRECTIONAL);
762 	kfree(cb->rdma_buf);
763 	if (cb->start_buf) {
764 		dma_unmap_single(cb->pd->device->dma_device,
765 			 pci_unmap_addr(cb, start_mapping),
766 			 cb->size, DMA_BIDIRECTIONAL);
767 		kfree(cb->start_buf);
768 	}
769 }
770 
771 static int krping_create_qp(struct krping_cb *cb)
772 {
773 	struct ib_qp_init_attr init_attr;
774 	int ret;
775 
776 	memset(&init_attr, 0, sizeof(init_attr));
777 	init_attr.cap.max_send_wr = cb->txdepth;
778 	init_attr.cap.max_recv_wr = 2;
779 	init_attr.cap.max_recv_sge = 1;
780 	init_attr.cap.max_send_sge = 1;
781 	init_attr.qp_type = IB_QPT_RC;
782 	init_attr.send_cq = cb->cq;
783 	init_attr.recv_cq = cb->cq;
784 	init_attr.sq_sig_type = IB_SIGNAL_REQ_WR;
785 
786 	if (cb->server) {
787 		ret = rdma_create_qp(cb->child_cm_id, cb->pd, &init_attr);
788 		if (!ret)
789 			cb->qp = cb->child_cm_id->qp;
790 	} else {
791 		ret = rdma_create_qp(cb->cm_id, cb->pd, &init_attr);
792 		if (!ret)
793 			cb->qp = cb->cm_id->qp;
794 	}
795 
796 	return ret;
797 }
798 
799 static void krping_free_qp(struct krping_cb *cb)
800 {
801 	ib_destroy_qp(cb->qp);
802 	ib_destroy_cq(cb->cq);
803 	ib_dealloc_pd(cb->pd);
804 }
805 
806 static int krping_setup_qp(struct krping_cb *cb, struct rdma_cm_id *cm_id)
807 {
808 	int ret;
809 	cb->pd = ib_alloc_pd(cm_id->device);
810 	if (IS_ERR(cb->pd)) {
811 		PRINTF(cb, "ib_alloc_pd failed\n");
812 		return PTR_ERR(cb->pd);
813 	}
814 	DEBUG_LOG(cb, "created pd %p\n", cb->pd);
815 
816 	strlcpy(cb->stats.name, cb->pd->device->name, sizeof(cb->stats.name));
817 
818 	cb->cq = ib_create_cq(cm_id->device, krping_cq_event_handler, NULL,
819 			      cb, cb->txdepth * 2, 0);
820 	if (IS_ERR(cb->cq)) {
821 		PRINTF(cb, "ib_create_cq failed\n");
822 		ret = PTR_ERR(cb->cq);
823 		goto err1;
824 	}
825 	DEBUG_LOG(cb, "created cq %p\n", cb->cq);
826 
827 	if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest) {
828 		ret = ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
829 		if (ret) {
830 			PRINTF(cb, "ib_create_cq failed\n");
831 			goto err2;
832 		}
833 	}
834 
835 	ret = krping_create_qp(cb);
836 	if (ret) {
837 		PRINTF(cb, "krping_create_qp failed: %d\n", ret);
838 		goto err2;
839 	}
840 	DEBUG_LOG(cb, "created qp %p\n", cb->qp);
841 	return 0;
842 err2:
843 	ib_destroy_cq(cb->cq);
844 err1:
845 	ib_dealloc_pd(cb->pd);
846 	return ret;
847 }
848 
849 /*
850  * return the (possibly rebound) rkey for the rdma buffer.
851  * FASTREG mode: invalidate and rebind via fastreg wr.
852  * MW mode: rebind the MW.
853  * other modes: just return the mr rkey.
854  */
855 static u32 krping_rdma_rkey(struct krping_cb *cb, u64 buf, int post_inv)
856 {
857 	u32 rkey = 0xffffffff;
858 	u64 p;
859 	struct ib_send_wr *bad_wr;
860 	int i;
861 	int ret;
862 
863 	switch (cb->mem) {
864 	case FASTREG:
865 		cb->invalidate_wr.ex.invalidate_rkey = cb->fastreg_mr->rkey;
866 
867 		/*
868 		 * Update the fastreg key.
869 		 */
870 		ib_update_fast_reg_key(cb->fastreg_mr, ++cb->key);
871 		cb->fastreg_wr.wr.fast_reg.rkey = cb->fastreg_mr->rkey;
872 
873 		/*
874 		 * Update the fastreg WR with new buf info.
875 		 */
876 		if (buf == (u64)cb->start_dma_addr)
877 			cb->fastreg_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_READ;
878 		else
879 			cb->fastreg_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
880 		cb->fastreg_wr.wr.fast_reg.iova_start = buf;
881 		p = (u64)(buf & PAGE_MASK);
882 		for (i=0; i < cb->fastreg_wr.wr.fast_reg.page_list_len;
883 		     i++, p += PAGE_SIZE) {
884 			cb->page_list->page_list[i] = p;
885 			DEBUG_LOG(cb, "page_list[%d] 0x%jx\n", i, (uintmax_t)p);
886 		}
887 
888 		DEBUG_LOG(cb, "post_inv = %d, fastreg new rkey 0x%x shift %u len %u"
889 			" iova_start %jx page_list_len %u\n",
890 			post_inv,
891 			cb->fastreg_wr.wr.fast_reg.rkey,
892 			cb->fastreg_wr.wr.fast_reg.page_shift,
893 			(unsigned)cb->fastreg_wr.wr.fast_reg.length,
894 			(uintmax_t)cb->fastreg_wr.wr.fast_reg.iova_start,
895 			cb->fastreg_wr.wr.fast_reg.page_list_len);
896 
897 		if (post_inv)
898 			ret = ib_post_send(cb->qp, &cb->invalidate_wr, &bad_wr);
899 		else
900 			ret = ib_post_send(cb->qp, &cb->fastreg_wr, &bad_wr);
901 		if (ret) {
902 			PRINTF(cb, "post send error %d\n", ret);
903 			cb->state = ERROR;
904 		}
905 		rkey = cb->fastreg_mr->rkey;
906 		break;
907 	case MW:
908 		/*
909 		 * Update the MW with new buf info.
910 		 */
911 		if (buf == (u64)cb->start_dma_addr) {
912 #ifdef BIND_INFO
913 			cb->bind_attr.bind_info.mw_access_flags = IB_ACCESS_REMOTE_READ;
914 			cb->bind_attr.bind_info.mr = cb->start_mr;
915 #else
916 			cb->bind_attr.mw_access_flags = IB_ACCESS_REMOTE_READ;
917 			cb->bind_attr.mr = cb->start_mr;
918 #endif
919 		} else {
920 #ifdef BIND_INFO
921 			cb->bind_attr.bind_info.mw_access_flags = IB_ACCESS_REMOTE_WRITE;
922 			cb->bind_attr.bind_info.mr = cb->rdma_mr;
923 #else
924 			cb->bind_attr.mw_access_flags = IB_ACCESS_REMOTE_WRITE;
925 			cb->bind_attr.mr = cb->rdma_mr;
926 #endif
927 		}
928 #ifdef BIND_INFO
929 		cb->bind_attr.bind_info.addr = buf;
930 #else
931 		cb->bind_attr.addr = buf;
932 #endif
933 		DEBUG_LOG(cb, "binding mw rkey 0x%x to buf %jx mr rkey 0x%x\n",
934 #ifdef BIND_INFO
935 			cb->mw->rkey, (uintmax_t)buf, cb->bind_attr.bind_info.mr->rkey);
936 #else
937 			cb->mw->rkey, buf, cb->bind_attr.mr->rkey);
938 #endif
939 		ret = ib_bind_mw(cb->qp, cb->mw, &cb->bind_attr);
940 		if (ret) {
941 			PRINTF(cb, "bind mw error %d\n", ret);
942 			cb->state = ERROR;
943 		} else
944 			rkey = cb->mw->rkey;
945 		break;
946 	case MR:
947 		if (buf == (u64)cb->start_dma_addr)
948 			rkey = cb->start_mr->rkey;
949 		else
950 			rkey = cb->rdma_mr->rkey;
951 		break;
952 	case DMA:
953 		rkey = cb->dma_mr->rkey;
954 		break;
955 	default:
956 		PRINTF(cb, "%s:%d case ERROR\n", __func__, __LINE__);
957 		cb->state = ERROR;
958 		break;
959 	}
960 	return rkey;
961 }
962 
963 static void krping_format_send(struct krping_cb *cb, u64 buf)
964 {
965 	struct krping_rdma_info *info = &cb->send_buf;
966 	u32 rkey;
967 
968 	/*
969 	 * Client side will do fastreg or mw bind before
970 	 * advertising the rdma buffer.  Server side
971 	 * sends have no data.
972 	 */
973 	if (!cb->server || cb->wlat || cb->rlat || cb->bw || cb->frtest) {
974 		rkey = krping_rdma_rkey(cb, buf, !cb->server_invalidate);
975 		info->buf = htonll(buf);
976 		info->rkey = htonl(rkey);
977 		info->size = htonl(cb->size);
978 		DEBUG_LOG(cb, "RDMA addr %llx rkey %x len %d\n",
979 			  (unsigned long long)buf, rkey, cb->size);
980 	}
981 }
982 
983 static void krping_test_server(struct krping_cb *cb)
984 {
985 	struct ib_send_wr *bad_wr, inv;
986 	int ret;
987 
988 	while (1) {
989 		/* Wait for client's Start STAG/TO/Len */
990 		wait_event_interruptible(cb->sem, cb->state >= RDMA_READ_ADV);
991 		if (cb->state != RDMA_READ_ADV) {
992 			PRINTF(cb, "wait for RDMA_READ_ADV state %d\n",
993 				cb->state);
994 			break;
995 		}
996 
997 		DEBUG_LOG(cb, "server received sink adv\n");
998 
999 		cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey;
1000 		cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr;
1001 		cb->rdma_sq_wr.sg_list->length = cb->remote_len;
1002 		cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 1);
1003 
1004 		/* Issue RDMA Read. */
1005 		if (cb->read_inv)
1006 			cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ_WITH_INV;
1007 		else {
1008 
1009 			cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ;
1010 			if (cb->mem == FASTREG) {
1011 				/*
1012 				 * Immediately follow the read with a
1013 				 * fenced LOCAL_INV.
1014 				 */
1015 				cb->rdma_sq_wr.next = &inv;
1016 				memset(&inv, 0, sizeof inv);
1017 				inv.opcode = IB_WR_LOCAL_INV;
1018 				inv.ex.invalidate_rkey = cb->fastreg_mr->rkey;
1019 				inv.send_flags = IB_SEND_FENCE;
1020 			}
1021 		}
1022 
1023 		ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr);
1024 		if (ret) {
1025 			PRINTF(cb, "post send error %d\n", ret);
1026 			break;
1027 		}
1028 		cb->rdma_sq_wr.next = NULL;
1029 
1030 		DEBUG_LOG(cb, "server posted rdma read req \n");
1031 
1032 		/* Wait for read completion */
1033 		wait_event_interruptible(cb->sem,
1034 					 cb->state >= RDMA_READ_COMPLETE);
1035 		if (cb->state != RDMA_READ_COMPLETE) {
1036 			PRINTF(cb,
1037 			       "wait for RDMA_READ_COMPLETE state %d\n",
1038 			       cb->state);
1039 			break;
1040 		}
1041 		DEBUG_LOG(cb, "server received read complete\n");
1042 
1043 		/* Display data in recv buf */
1044 		if (cb->verbose) {
1045 			if (strlen(cb->rdma_buf) > 128) {
1046 				char msgbuf[128];
1047 
1048 				strlcpy(msgbuf, cb->rdma_buf, sizeof(msgbuf));
1049 				PRINTF(cb, "server ping data stripped: %s\n",
1050 				       msgbuf);
1051 			} else
1052 				PRINTF(cb, "server ping data: %s\n",
1053 				       cb->rdma_buf);
1054 		}
1055 
1056 		/* Tell client to continue */
1057 		if (cb->server && cb->server_invalidate) {
1058 			cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey;
1059 			cb->sq_wr.opcode = IB_WR_SEND_WITH_INV;
1060 			DEBUG_LOG(cb, "send-w-inv rkey 0x%x\n", cb->remote_rkey);
1061 		}
1062 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1063 		if (ret) {
1064 			PRINTF(cb, "post send error %d\n", ret);
1065 			break;
1066 		}
1067 		DEBUG_LOG(cb, "server posted go ahead\n");
1068 
1069 		/* Wait for client's RDMA STAG/TO/Len */
1070 		wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV);
1071 		if (cb->state != RDMA_WRITE_ADV) {
1072 			PRINTF(cb,
1073 			       "wait for RDMA_WRITE_ADV state %d\n",
1074 			       cb->state);
1075 			break;
1076 		}
1077 		DEBUG_LOG(cb, "server received sink adv\n");
1078 
1079 		/* RDMA Write echo data */
1080 		cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE;
1081 		cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey;
1082 		cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr;
1083 		cb->rdma_sq_wr.sg_list->length = strlen(cb->rdma_buf) + 1;
1084 		if (cb->local_dma_lkey)
1085 			cb->rdma_sgl.lkey = cb->qp->device->local_dma_lkey;
1086 		else
1087 			cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 0);
1088 
1089 		DEBUG_LOG(cb, "rdma write from lkey %x laddr %llx len %d\n",
1090 			  cb->rdma_sq_wr.sg_list->lkey,
1091 			  (unsigned long long)cb->rdma_sq_wr.sg_list->addr,
1092 			  cb->rdma_sq_wr.sg_list->length);
1093 
1094 		ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr);
1095 		if (ret) {
1096 			PRINTF(cb, "post send error %d\n", ret);
1097 			break;
1098 		}
1099 
1100 		/* Wait for completion */
1101 		ret = wait_event_interruptible(cb->sem, cb->state >=
1102 							 RDMA_WRITE_COMPLETE);
1103 		if (cb->state != RDMA_WRITE_COMPLETE) {
1104 			PRINTF(cb,
1105 			       "wait for RDMA_WRITE_COMPLETE state %d\n",
1106 			       cb->state);
1107 			break;
1108 		}
1109 		DEBUG_LOG(cb, "server rdma write complete \n");
1110 
1111 		cb->state = CONNECTED;
1112 
1113 		/* Tell client to begin again */
1114 		if (cb->server && cb->server_invalidate) {
1115 			cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey;
1116 			cb->sq_wr.opcode = IB_WR_SEND_WITH_INV;
1117 			DEBUG_LOG(cb, "send-w-inv rkey 0x%x\n", cb->remote_rkey);
1118 		}
1119 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1120 		if (ret) {
1121 			PRINTF(cb, "post send error %d\n", ret);
1122 			break;
1123 		}
1124 		DEBUG_LOG(cb, "server posted go ahead\n");
1125 	}
1126 }
1127 
1128 static void rlat_test(struct krping_cb *cb)
1129 {
1130 	int scnt;
1131 	int iters = cb->count;
1132 	struct timeval start_tv, stop_tv;
1133 	int ret;
1134 	struct ib_wc wc;
1135 	struct ib_send_wr *bad_wr;
1136 	int ne;
1137 
1138 	scnt = 0;
1139 	cb->rdma_sq_wr.opcode = IB_WR_RDMA_READ;
1140 	cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey;
1141 	cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr;
1142 	cb->rdma_sq_wr.sg_list->length = cb->size;
1143 
1144 	microtime(&start_tv);
1145 	if (!cb->poll) {
1146 		cb->state = RDMA_READ_ADV;
1147 		ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
1148 	}
1149 	while (scnt < iters) {
1150 
1151 		cb->state = RDMA_READ_ADV;
1152 		ret = ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr);
1153 		if (ret) {
1154 			PRINTF(cb,
1155 				"Couldn't post send: ret=%d scnt %d\n",
1156 				ret, scnt);
1157 			return;
1158 		}
1159 
1160 		do {
1161 			if (!cb->poll) {
1162 				wait_event_interruptible(cb->sem,
1163 					cb->state != RDMA_READ_ADV);
1164 				if (cb->state == RDMA_READ_COMPLETE) {
1165 					ne = 1;
1166 					ib_req_notify_cq(cb->cq,
1167 						IB_CQ_NEXT_COMP);
1168 				} else {
1169 					ne = -1;
1170 				}
1171 			} else
1172 				ne = ib_poll_cq(cb->cq, 1, &wc);
1173 			if (cb->state == ERROR) {
1174 				PRINTF(cb,
1175 					"state == ERROR...bailing scnt %d\n",
1176 					scnt);
1177 				return;
1178 			}
1179 		} while (ne == 0);
1180 
1181 		if (ne < 0) {
1182 			PRINTF(cb, "poll CQ failed %d\n", ne);
1183 			return;
1184 		}
1185 		if (cb->poll && wc.status != IB_WC_SUCCESS) {
1186 			PRINTF(cb, "Completion wth error at %s:\n",
1187 				cb->server ? "server" : "client");
1188 			PRINTF(cb, "Failed status %d: wr_id %d\n",
1189 				wc.status, (int) wc.wr_id);
1190 			return;
1191 		}
1192 		++scnt;
1193 	}
1194 	microtime(&stop_tv);
1195 
1196         if (stop_tv.tv_usec < start_tv.tv_usec) {
1197                 stop_tv.tv_usec += 1000000;
1198                 stop_tv.tv_sec  -= 1;
1199         }
1200 
1201 	PRINTF(cb, "delta sec %lu delta usec %lu iter %d size %d\n",
1202 		(unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1203 		(unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1204 		scnt, cb->size);
1205 }
1206 
1207 static void wlat_test(struct krping_cb *cb)
1208 {
1209 	int ccnt, scnt, rcnt;
1210 	int iters=cb->count;
1211 	volatile char *poll_buf = (char *) cb->start_buf;
1212 	char *buf = (char *)cb->rdma_buf;
1213 	struct timeval start_tv, stop_tv;
1214 	cycles_t *post_cycles_start, *post_cycles_stop;
1215 	cycles_t *poll_cycles_start, *poll_cycles_stop;
1216 	cycles_t *last_poll_cycles_start;
1217 	cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0;
1218 	int i;
1219 	int cycle_iters = 1000;
1220 
1221 	ccnt = 0;
1222 	scnt = 0;
1223 	rcnt = 0;
1224 
1225 	post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1226 	if (!post_cycles_start) {
1227 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1228 		return;
1229 	}
1230 	post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1231 	if (!post_cycles_stop) {
1232 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1233 		return;
1234 	}
1235 	poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1236 	if (!poll_cycles_start) {
1237 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1238 		return;
1239 	}
1240 	poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1241 	if (!poll_cycles_stop) {
1242 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1243 		return;
1244 	}
1245 	last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t),
1246 		GFP_KERNEL);
1247 	if (!last_poll_cycles_start) {
1248 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1249 		return;
1250 	}
1251 	cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE;
1252 	cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey;
1253 	cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr;
1254 	cb->rdma_sq_wr.sg_list->length = cb->size;
1255 
1256 	if (cycle_iters > iters)
1257 		cycle_iters = iters;
1258 	microtime(&start_tv);
1259 	while (scnt < iters || ccnt < iters || rcnt < iters) {
1260 
1261 		/* Wait till buffer changes. */
1262 		if (rcnt < iters && !(scnt < 1 && !cb->server)) {
1263 			++rcnt;
1264 			while (*poll_buf != (char)rcnt) {
1265 				if (cb->state == ERROR) {
1266 					PRINTF(cb,
1267 						"state = ERROR, bailing\n");
1268 					return;
1269 				}
1270 			}
1271 		}
1272 
1273 		if (scnt < iters) {
1274 			struct ib_send_wr *bad_wr;
1275 
1276 			*buf = (char)scnt+1;
1277 			if (scnt < cycle_iters)
1278 				post_cycles_start[scnt] = get_cycles();
1279 			if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) {
1280 				PRINTF(cb,
1281 					"Couldn't post send: scnt=%d\n",
1282 					scnt);
1283 				return;
1284 			}
1285 			if (scnt < cycle_iters)
1286 				post_cycles_stop[scnt] = get_cycles();
1287 			scnt++;
1288 		}
1289 
1290 		if (ccnt < iters) {
1291 			struct ib_wc wc;
1292 			int ne;
1293 
1294 			if (ccnt < cycle_iters)
1295 				poll_cycles_start[ccnt] = get_cycles();
1296 			do {
1297 				if (ccnt < cycle_iters)
1298 					last_poll_cycles_start[ccnt] =
1299 						get_cycles();
1300 				ne = ib_poll_cq(cb->cq, 1, &wc);
1301 			} while (ne == 0);
1302 			if (ccnt < cycle_iters)
1303 				poll_cycles_stop[ccnt] = get_cycles();
1304 			++ccnt;
1305 
1306 			if (ne < 0) {
1307 				PRINTF(cb, "poll CQ failed %d\n", ne);
1308 				return;
1309 			}
1310 			if (wc.status != IB_WC_SUCCESS) {
1311 				PRINTF(cb,
1312 					"Completion wth error at %s:\n",
1313 					cb->server ? "server" : "client");
1314 				PRINTF(cb,
1315 					"Failed status %d: wr_id %d\n",
1316 					wc.status, (int) wc.wr_id);
1317 				PRINTF(cb,
1318 					"scnt=%d, rcnt=%d, ccnt=%d\n",
1319 					scnt, rcnt, ccnt);
1320 				return;
1321 			}
1322 		}
1323 	}
1324 	microtime(&stop_tv);
1325 
1326         if (stop_tv.tv_usec < start_tv.tv_usec) {
1327                 stop_tv.tv_usec += 1000000;
1328                 stop_tv.tv_sec  -= 1;
1329         }
1330 
1331 	for (i=0; i < cycle_iters; i++) {
1332 		sum_post += post_cycles_stop[i] - post_cycles_start[i];
1333 		sum_poll += poll_cycles_stop[i] - poll_cycles_start[i];
1334 		sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i];
1335 	}
1336 	PRINTF(cb,
1337 		"delta sec %lu delta usec %lu iter %d size %d cycle_iters %d"
1338 		" sum_post %llu sum_poll %llu sum_last_poll %llu\n",
1339 		(unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1340 		(unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1341 		scnt, cb->size, cycle_iters,
1342 		(unsigned long long)sum_post, (unsigned long long)sum_poll,
1343 		(unsigned long long)sum_last_poll);
1344 	kfree(post_cycles_start);
1345 	kfree(post_cycles_stop);
1346 	kfree(poll_cycles_start);
1347 	kfree(poll_cycles_stop);
1348 	kfree(last_poll_cycles_start);
1349 }
1350 
1351 static void bw_test(struct krping_cb *cb)
1352 {
1353 	int ccnt, scnt, rcnt;
1354 	int iters=cb->count;
1355 	struct timeval start_tv, stop_tv;
1356 	cycles_t *post_cycles_start, *post_cycles_stop;
1357 	cycles_t *poll_cycles_start, *poll_cycles_stop;
1358 	cycles_t *last_poll_cycles_start;
1359 	cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0;
1360 	int i;
1361 	int cycle_iters = 1000;
1362 
1363 	ccnt = 0;
1364 	scnt = 0;
1365 	rcnt = 0;
1366 
1367 	post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1368 	if (!post_cycles_start) {
1369 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1370 		return;
1371 	}
1372 	post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1373 	if (!post_cycles_stop) {
1374 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1375 		return;
1376 	}
1377 	poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1378 	if (!poll_cycles_start) {
1379 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1380 		return;
1381 	}
1382 	poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1383 	if (!poll_cycles_stop) {
1384 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1385 		return;
1386 	}
1387 	last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t),
1388 		GFP_KERNEL);
1389 	if (!last_poll_cycles_start) {
1390 		PRINTF(cb, "%s kmalloc failed\n", __FUNCTION__);
1391 		return;
1392 	}
1393 	cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE;
1394 	cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey;
1395 	cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr;
1396 	cb->rdma_sq_wr.sg_list->length = cb->size;
1397 
1398 	if (cycle_iters > iters)
1399 		cycle_iters = iters;
1400 	microtime(&start_tv);
1401 	while (scnt < iters || ccnt < iters) {
1402 
1403 		while (scnt < iters && scnt - ccnt < cb->txdepth) {
1404 			struct ib_send_wr *bad_wr;
1405 
1406 			if (scnt < cycle_iters)
1407 				post_cycles_start[scnt] = get_cycles();
1408 			if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) {
1409 				PRINTF(cb,
1410 					"Couldn't post send: scnt=%d\n",
1411 					scnt);
1412 				return;
1413 			}
1414 			if (scnt < cycle_iters)
1415 				post_cycles_stop[scnt] = get_cycles();
1416 			++scnt;
1417 		}
1418 
1419 		if (ccnt < iters) {
1420 			int ne;
1421 			struct ib_wc wc;
1422 
1423 			if (ccnt < cycle_iters)
1424 				poll_cycles_start[ccnt] = get_cycles();
1425 			do {
1426 				if (ccnt < cycle_iters)
1427 					last_poll_cycles_start[ccnt] =
1428 						get_cycles();
1429 				ne = ib_poll_cq(cb->cq, 1, &wc);
1430 			} while (ne == 0);
1431 			if (ccnt < cycle_iters)
1432 				poll_cycles_stop[ccnt] = get_cycles();
1433 			ccnt += 1;
1434 
1435 			if (ne < 0) {
1436 				PRINTF(cb, "poll CQ failed %d\n", ne);
1437 				return;
1438 			}
1439 			if (wc.status != IB_WC_SUCCESS) {
1440 				PRINTF(cb,
1441 					"Completion wth error at %s:\n",
1442 					cb->server ? "server" : "client");
1443 				PRINTF(cb,
1444 					"Failed status %d: wr_id %d\n",
1445 					wc.status, (int) wc.wr_id);
1446 				return;
1447 			}
1448 		}
1449 	}
1450 	microtime(&stop_tv);
1451 
1452         if (stop_tv.tv_usec < start_tv.tv_usec) {
1453                 stop_tv.tv_usec += 1000000;
1454                 stop_tv.tv_sec  -= 1;
1455         }
1456 
1457 	for (i=0; i < cycle_iters; i++) {
1458 		sum_post += post_cycles_stop[i] - post_cycles_start[i];
1459 		sum_poll += poll_cycles_stop[i] - poll_cycles_start[i];
1460 		sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i];
1461 	}
1462 	PRINTF(cb,
1463 		"delta sec %lu delta usec %lu iter %d size %d cycle_iters %d"
1464 		" sum_post %llu sum_poll %llu sum_last_poll %llu\n",
1465 		(unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1466 		(unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1467 		scnt, cb->size, cycle_iters,
1468 		(unsigned long long)sum_post, (unsigned long long)sum_poll,
1469 		(unsigned long long)sum_last_poll);
1470 	kfree(post_cycles_start);
1471 	kfree(post_cycles_stop);
1472 	kfree(poll_cycles_start);
1473 	kfree(poll_cycles_stop);
1474 	kfree(last_poll_cycles_start);
1475 }
1476 
1477 static void krping_rlat_test_server(struct krping_cb *cb)
1478 {
1479 	struct ib_send_wr *bad_wr;
1480 	struct ib_wc wc;
1481 	int ret;
1482 
1483 	/* Spin waiting for client's Start STAG/TO/Len */
1484 	while (cb->state < RDMA_READ_ADV) {
1485 		krping_cq_event_handler(cb->cq, cb);
1486 	}
1487 
1488 	/* Send STAG/TO/Len to client */
1489 	krping_format_send(cb, cb->start_dma_addr);
1490 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1491 	if (ret) {
1492 		PRINTF(cb, "post send error %d\n", ret);
1493 		return;
1494 	}
1495 
1496 	/* Spin waiting for send completion */
1497 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1498 	if (ret < 0) {
1499 		PRINTF(cb, "poll error %d\n", ret);
1500 		return;
1501 	}
1502 	if (wc.status) {
1503 		PRINTF(cb, "send completiong error %d\n", wc.status);
1504 		return;
1505 	}
1506 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1507 }
1508 
1509 static void krping_wlat_test_server(struct krping_cb *cb)
1510 {
1511 	struct ib_send_wr *bad_wr;
1512 	struct ib_wc wc;
1513 	int ret;
1514 
1515 	/* Spin waiting for client's Start STAG/TO/Len */
1516 	while (cb->state < RDMA_READ_ADV) {
1517 		krping_cq_event_handler(cb->cq, cb);
1518 	}
1519 
1520 	/* Send STAG/TO/Len to client */
1521 	krping_format_send(cb, cb->start_dma_addr);
1522 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1523 	if (ret) {
1524 		PRINTF(cb, "post send error %d\n", ret);
1525 		return;
1526 	}
1527 
1528 	/* Spin waiting for send completion */
1529 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1530 	if (ret < 0) {
1531 		PRINTF(cb, "poll error %d\n", ret);
1532 		return;
1533 	}
1534 	if (wc.status) {
1535 		PRINTF(cb, "send completiong error %d\n", wc.status);
1536 		return;
1537 	}
1538 
1539 	wlat_test(cb);
1540 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1541 }
1542 
1543 static void krping_bw_test_server(struct krping_cb *cb)
1544 {
1545 	struct ib_send_wr *bad_wr;
1546 	struct ib_wc wc;
1547 	int ret;
1548 
1549 	/* Spin waiting for client's Start STAG/TO/Len */
1550 	while (cb->state < RDMA_READ_ADV) {
1551 		krping_cq_event_handler(cb->cq, cb);
1552 	}
1553 
1554 	/* Send STAG/TO/Len to client */
1555 	krping_format_send(cb, cb->start_dma_addr);
1556 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1557 	if (ret) {
1558 		PRINTF(cb, "post send error %d\n", ret);
1559 		return;
1560 	}
1561 
1562 	/* Spin waiting for send completion */
1563 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1564 	if (ret < 0) {
1565 		PRINTF(cb, "poll error %d\n", ret);
1566 		return;
1567 	}
1568 	if (wc.status) {
1569 		PRINTF(cb, "send completiong error %d\n", wc.status);
1570 		return;
1571 	}
1572 
1573 	if (cb->duplex)
1574 		bw_test(cb);
1575 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1576 }
1577 
1578 static int fastreg_supported(struct krping_cb *cb, int server)
1579 {
1580 	struct ib_device *dev = server?cb->child_cm_id->device:
1581 					cb->cm_id->device;
1582 	struct ib_device_attr attr;
1583 	int ret;
1584 
1585 	ret = ib_query_device(dev, &attr);
1586 	if (ret) {
1587 		PRINTF(cb, "ib_query_device failed ret %d\n", ret);
1588 		return 0;
1589 	}
1590 	if (!(attr.device_cap_flags & IB_DEVICE_MEM_MGT_EXTENSIONS)) {
1591 		PRINTF(cb, "Fastreg not supported - device_cap_flags 0x%llx\n",
1592 		    (unsigned long long)attr.device_cap_flags);
1593 		return 0;
1594 	}
1595 	DEBUG_LOG(cb, "Fastreg supported - device_cap_flags 0x%jx\n",
1596 		(uintmax_t)attr.device_cap_flags);
1597 	return 1;
1598 }
1599 
1600 static int krping_bind_server(struct krping_cb *cb)
1601 {
1602 	struct sockaddr_in sin;
1603 	int ret;
1604 
1605 	memset(&sin, 0, sizeof(sin));
1606 	sin.sin_len = sizeof sin;
1607 	sin.sin_family = AF_INET;
1608 	sin.sin_addr.s_addr = cb->addr.s_addr;
1609 	sin.sin_port = cb->port;
1610 
1611 	ret = rdma_bind_addr(cb->cm_id, (struct sockaddr *) &sin);
1612 	if (ret) {
1613 		PRINTF(cb, "rdma_bind_addr error %d\n", ret);
1614 		return ret;
1615 	}
1616 	DEBUG_LOG(cb, "rdma_bind_addr successful\n");
1617 
1618 	DEBUG_LOG(cb, "rdma_listen\n");
1619 	ret = rdma_listen(cb->cm_id, 3);
1620 	if (ret) {
1621 		PRINTF(cb, "rdma_listen failed: %d\n", ret);
1622 		return ret;
1623 	}
1624 
1625 	wait_event_interruptible(cb->sem, cb->state >= CONNECT_REQUEST);
1626 	if (cb->state != CONNECT_REQUEST) {
1627 		PRINTF(cb, "wait for CONNECT_REQUEST state %d\n",
1628 			cb->state);
1629 		return -1;
1630 	}
1631 
1632 	if (cb->mem == FASTREG && !fastreg_supported(cb, 1))
1633 		return -EINVAL;
1634 
1635 	return 0;
1636 }
1637 
1638 /*
1639  * sq-depth worth of fastreg + 0B read-inv pairs, reposting them as the reads
1640  * complete.
1641  * NOTE: every 9 seconds we sleep for 1 second to keep the kernel happy.
1642  */
1643 static void krping_fr_test5(struct krping_cb *cb)
1644 {
1645 	struct ib_fast_reg_page_list **pl;
1646 	struct ib_send_wr *fr, *read, *bad;
1647 	struct ib_wc wc;
1648 	struct ib_sge *sgl;
1649 	u8 key = 0;
1650 	struct ib_mr **mr;
1651 	u8 **buf;
1652 	dma_addr_t *dma_addr;
1653 	int i;
1654 	int ret;
1655 	int plen = (((cb->size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
1656 	time_t start;
1657 	int count = 0;
1658 	int scnt;
1659 	int depth = cb->txdepth >> 1;
1660 
1661 	if (!depth) {
1662 		PRINTF(cb, "txdepth must be > 1 for this test!\n");
1663 		return;
1664 	}
1665 
1666 	pl = kzalloc(sizeof *pl * depth, GFP_KERNEL);
1667 	DEBUG_LOG(cb, "%s pl %p size %zu\n", __func__, pl, sizeof *pl * depth);
1668 	mr = kzalloc(sizeof *mr * depth, GFP_KERNEL);
1669 	DEBUG_LOG(cb, "%s mr %p size %zu\n", __func__, mr, sizeof *mr * depth);
1670 	fr = kzalloc(sizeof *fr * depth, GFP_KERNEL);
1671 	DEBUG_LOG(cb, "%s fr %p size %zu\n", __func__, fr, sizeof *fr * depth);
1672 	sgl = kzalloc(sizeof *sgl * depth, GFP_KERNEL);
1673 	DEBUG_LOG(cb, "%s sgl %p size %zu\n", __func__, sgl, sizeof *sgl * depth);
1674 	read = kzalloc(sizeof *read * depth, GFP_KERNEL);
1675 	DEBUG_LOG(cb, "%s read %p size %zu\n", __func__, read, sizeof *read * depth);
1676 	buf = kzalloc(sizeof *buf * depth, GFP_KERNEL);
1677 	DEBUG_LOG(cb, "%s buf %p size %zu\n", __func__, buf, sizeof *buf * depth);
1678 	dma_addr = kzalloc(sizeof *dma_addr * depth, GFP_KERNEL);
1679 	DEBUG_LOG(cb, "%s dma_addr %p size %zu\n", __func__, dma_addr, sizeof *dma_addr * depth);
1680 	if (!pl || !mr || !fr || !read || !sgl || !buf || !dma_addr) {
1681 		PRINTF(cb, "kzalloc failed\n");
1682 		goto err1;
1683 	}
1684 
1685 	for (scnt = 0; scnt < depth; scnt++) {
1686 		pl[scnt] = ib_alloc_fast_reg_page_list(cb->qp->device, plen);
1687 		if (IS_ERR(pl[scnt])) {
1688 			PRINTF(cb, "alloc_fr_page_list failed %ld\n",
1689 			       PTR_ERR(pl[scnt]));
1690 			goto err2;
1691 		}
1692 		DEBUG_LOG(cb, "%s pl[%u] %p\n", __func__, scnt, pl[scnt]);
1693 
1694 		mr[scnt] = ib_alloc_fast_reg_mr(cb->pd, plen);
1695 		if (IS_ERR(mr[scnt])) {
1696 			PRINTF(cb, "alloc_fr failed %ld\n",
1697 			       PTR_ERR(mr[scnt]));
1698 			goto err2;
1699 		}
1700 		DEBUG_LOG(cb, "%s mr[%u] %p\n", __func__, scnt, mr[scnt]);
1701 		ib_update_fast_reg_key(mr[scnt], ++key);
1702 
1703 		buf[scnt] = kmalloc(cb->size, GFP_KERNEL);
1704 		if (!buf[scnt]) {
1705 			PRINTF(cb, "kmalloc failed\n");
1706 			ret = -ENOMEM;
1707 			goto err2;
1708 		}
1709 		DEBUG_LOG(cb, "%s buf[%u] %p\n", __func__, scnt, buf[scnt]);
1710 		dma_addr[scnt] = ib_dma_map_single(cb->pd->device,
1711 						   buf[scnt], cb->size,
1712 						   DMA_BIDIRECTIONAL);
1713 		if (dma_mapping_error(cb->pd->device->dma_device,
1714 		    dma_addr[scnt])) {
1715 			PRINTF(cb, "dma_map failed\n");
1716 			ret = -ENOMEM;
1717 			goto err2;
1718 		}
1719 		DEBUG_LOG(cb, "%s dma_addr[%u] %p\n", __func__, scnt, (void *)dma_addr[scnt]);
1720 		for (i=0; i<plen; i++) {
1721 			pl[scnt]->page_list[i] = ((unsigned long)dma_addr[scnt] & PAGE_MASK) + (i * PAGE_SIZE);
1722 			DEBUG_LOG(cb, "%s pl[%u]->page_list[%u] 0x%jx\n",
1723 				  __func__, scnt, i,  (uintmax_t)pl[scnt]->page_list[i]);
1724 		}
1725 
1726 		sgl[scnt].lkey = mr[scnt]->rkey;
1727 		sgl[scnt].length = cb->size;
1728 		sgl[scnt].addr = (u64)buf[scnt];
1729 		DEBUG_LOG(cb, "%s sgl[%u].lkey 0x%x length %u addr 0x%jx\n",
1730 			  __func__, scnt,  sgl[scnt].lkey, sgl[scnt].length,
1731 			  (uintmax_t)sgl[scnt].addr);
1732 
1733 		fr[scnt].opcode = IB_WR_FAST_REG_MR;
1734 		fr[scnt].wr_id = scnt;
1735 		fr[scnt].send_flags = 0;
1736 		fr[scnt].wr.fast_reg.page_shift = PAGE_SHIFT;
1737 		fr[scnt].wr.fast_reg.length = cb->size;
1738 		fr[scnt].wr.fast_reg.page_list = pl[scnt];
1739 		fr[scnt].wr.fast_reg.page_list_len = plen;
1740 		fr[scnt].wr.fast_reg.iova_start = (u64)buf[scnt];
1741 		fr[scnt].wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
1742 		fr[scnt].wr.fast_reg.rkey = mr[scnt]->rkey;
1743 		fr[scnt].next = &read[scnt];
1744 		read[scnt].opcode = IB_WR_RDMA_READ_WITH_INV;
1745 		read[scnt].wr_id = scnt;
1746 		read[scnt].send_flags = IB_SEND_SIGNALED;
1747 		read[scnt].wr.rdma.rkey = cb->remote_rkey;
1748 		read[scnt].wr.rdma.remote_addr = cb->remote_addr;
1749 		read[scnt].num_sge = 1;
1750 		read[scnt].sg_list = &sgl[scnt];
1751 		ret = ib_post_send(cb->qp, &fr[scnt], &bad);
1752 		if (ret) {
1753 			PRINTF(cb, "ib_post_send failed %d\n", ret);
1754 			goto err2;
1755 		}
1756 	}
1757 
1758 	start = time_uptime;
1759 	DEBUG_LOG(cb, "%s starting IO.\n", __func__);
1760 	while (!cb->count || cb->server || count < cb->count) {
1761 		if ((time_uptime - start) >= 9) {
1762 			DEBUG_LOG(cb, "%s pausing 1 tick! count %u\n", __func__,
1763 				  count);
1764 			wait_event_interruptible_timeout(cb->sem,
1765 							 cb->state == ERROR,
1766 							 1);
1767 			if (cb->state == ERROR)
1768 				break;
1769 			start = time_uptime;
1770 		}
1771 		do {
1772 			ret = ib_poll_cq(cb->cq, 1, &wc);
1773 			if (ret < 0) {
1774 				PRINTF(cb, "ib_poll_cq failed %d\n",
1775 				       ret);
1776 				goto err2;
1777 			}
1778 			if (ret == 1) {
1779 				if (wc.status) {
1780 					PRINTF(cb,
1781 					       "completion error %u wr_id %ju "
1782 					       "opcode %d\n", wc.status,
1783 					       (uintmax_t)wc.wr_id, wc.opcode);
1784 					goto err2;
1785 				}
1786 				count++;
1787 				if (count == cb->count)
1788 					break;
1789 				ib_update_fast_reg_key(mr[wc.wr_id], ++key);
1790 				fr[wc.wr_id].wr.fast_reg.rkey =
1791 					mr[wc.wr_id]->rkey;
1792 				sgl[wc.wr_id].lkey = mr[wc.wr_id]->rkey;
1793 				ret = ib_post_send(cb->qp, &fr[wc.wr_id], &bad);
1794 				if (ret) {
1795 					PRINTF(cb,
1796 					       "ib_post_send failed %d\n", ret);
1797 					goto err2;
1798 				}
1799 			} else if (krping_sigpending()) {
1800 				PRINTF(cb, "signal!\n");
1801 				goto err2;
1802 			}
1803 		} while (ret == 1);
1804 	}
1805 	DEBUG_LOG(cb, "%s done!\n", __func__);
1806 err2:
1807 	DEBUG_LOG(cb, "sleeping 1 second\n");
1808 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
1809 	DEBUG_LOG(cb, "draining the cq...\n");
1810 	do {
1811 		ret = ib_poll_cq(cb->cq, 1, &wc);
1812 		if (ret < 0) {
1813 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
1814 			break;
1815 		}
1816 		if (ret == 1) {
1817 			if (wc.status) {
1818 				PRINTF(cb, "completion error %u "
1819 				       "opcode %u\n", wc.status, wc.opcode);
1820 			}
1821 		}
1822 	} while (ret == 1);
1823 
1824 	DEBUG_LOG(cb, "destroying fr mrs!\n");
1825 	for (scnt = 0; scnt < depth; scnt++) {
1826 		if (mr[scnt]) {
1827 			ib_dereg_mr(mr[scnt]);
1828 			DEBUG_LOG(cb, "%s dereg mr %p\n", __func__, mr[scnt]);
1829 		}
1830 	}
1831 	DEBUG_LOG(cb, "unmapping/freeing bufs!\n");
1832 	for (scnt = 0; scnt < depth; scnt++) {
1833 		if (buf[scnt]) {
1834 			dma_unmap_single(cb->pd->device->dma_device,
1835 					 dma_addr[scnt], cb->size,
1836 					 DMA_BIDIRECTIONAL);
1837 			kfree(buf[scnt]);
1838 			DEBUG_LOG(cb, "%s unmap/free buf %p dma_addr %p\n", __func__, buf[scnt], (void *)dma_addr[scnt]);
1839 		}
1840 	}
1841 	DEBUG_LOG(cb, "destroying fr page lists!\n");
1842 	for (scnt = 0; scnt < depth; scnt++) {
1843 		if (pl[scnt]) {
1844 			DEBUG_LOG(cb, "%s free pl %p\n", __func__, pl[scnt]);
1845 			ib_free_fast_reg_page_list(pl[scnt]);
1846 		}
1847 	}
1848 err1:
1849 	if (pl)
1850 		kfree(pl);
1851 	if (mr)
1852 		kfree(mr);
1853 	if (fr)
1854 		kfree(fr);
1855 	if (read)
1856 		kfree(read);
1857 	if (sgl)
1858 		kfree(sgl);
1859 	if (buf)
1860 		kfree(buf);
1861 	if (dma_addr)
1862 		kfree(dma_addr);
1863 }
1864 static void krping_fr_test_server(struct krping_cb *cb)
1865 {
1866 	DEBUG_LOG(cb, "%s waiting for disconnect...\n", __func__);
1867 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1868 }
1869 
1870 static void krping_fr_test5_server(struct krping_cb *cb)
1871 {
1872 	struct ib_send_wr *bad_wr;
1873 	struct ib_wc wc;
1874 	int ret;
1875 
1876 	/* Spin waiting for client's Start STAG/TO/Len */
1877 	while (cb->state < RDMA_READ_ADV) {
1878 		krping_cq_event_handler(cb->cq, cb);
1879 	}
1880 	DEBUG_LOG(cb, "%s client STAG %x TO 0x%jx\n", __func__,
1881 		  cb->remote_rkey, (uintmax_t)cb->remote_addr);
1882 
1883 	/* Send STAG/TO/Len to client */
1884 	krping_format_send(cb, cb->start_dma_addr);
1885 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1886 	if (ret) {
1887 		PRINTF(cb, "post send error %d\n", ret);
1888 		return;
1889 	}
1890 
1891 	/* Spin waiting for send completion */
1892 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1893 	if (ret < 0) {
1894 		PRINTF(cb, "poll error %d\n", ret);
1895 		return;
1896 	}
1897 	if (wc.status) {
1898 		PRINTF(cb, "send completiong error %d\n", wc.status);
1899 		return;
1900 	}
1901 
1902 	if (cb->duplex)
1903 		krping_fr_test5(cb);
1904 	DEBUG_LOG(cb, "%s waiting for disconnect...\n", __func__);
1905 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1906 }
1907 
1908 static void krping_fr_test5_client(struct krping_cb *cb)
1909 {
1910 	struct ib_send_wr *bad;
1911 	struct ib_wc wc;
1912 	int ret;
1913 
1914 	cb->state = RDMA_READ_ADV;
1915 
1916 	/* Send STAG/TO/Len to server */
1917 	krping_format_send(cb, cb->start_dma_addr);
1918 	if (cb->state == ERROR) {
1919 		PRINTF(cb, "krping_format_send failed\n");
1920 		return;
1921 	}
1922 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad);
1923 	if (ret) {
1924 		PRINTF(cb, "post send error %d\n", ret);
1925 		return;
1926 	}
1927 
1928 	/* Spin waiting for send completion */
1929 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1930 	if (ret < 0) {
1931 		PRINTF(cb, "poll error %d\n", ret);
1932 		return;
1933 	}
1934 	if (wc.status) {
1935 		PRINTF(cb, "send completion error %d\n", wc.status);
1936 		return;
1937 	}
1938 
1939 	/* Spin waiting for server's Start STAG/TO/Len */
1940 	while (cb->state < RDMA_WRITE_ADV) {
1941 		krping_cq_event_handler(cb->cq, cb);
1942 	}
1943 	DEBUG_LOG(cb, "%s server STAG %x TO 0x%jx\n", __func__, cb->remote_rkey,
1944 	    (uintmax_t)cb->remote_addr);
1945 
1946 	return krping_fr_test5(cb);
1947 }
1948 
1949 /*
1950  * sq-depth worth of write + fastreg + inv, reposting them as the invs
1951  * complete.
1952  * NOTE: every 9 seconds we sleep for 1 second to keep the kernel happy.
1953  * If a count is given, then the last IO will have a bogus lkey in the
1954  * write work request.  This reproduces a fw bug where the connection
1955  * will get stuck if a fastreg is processed while the ulptx is failing
1956  * the bad write.
1957  */
1958 static void krping_fr_test6(struct krping_cb *cb)
1959 {
1960 	struct ib_fast_reg_page_list **pl;
1961 	struct ib_send_wr *fr, *write, *inv, *bad;
1962 	struct ib_wc wc;
1963 	struct ib_sge *sgl;
1964 	u8 key = 0;
1965 	struct ib_mr **mr;
1966 	u8 **buf;
1967 	dma_addr_t *dma_addr;
1968 	int i;
1969 	int ret;
1970 	int plen = (((cb->size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
1971 	unsigned long start;
1972 	int count = 0;
1973 	int scnt;
1974 	int depth = cb->txdepth  / 3;
1975 
1976 	if (!depth) {
1977 		PRINTF(cb, "txdepth must be > 3 for this test!\n");
1978 		return;
1979 	}
1980 
1981 	pl = kzalloc(sizeof *pl * depth, GFP_KERNEL);
1982 	DEBUG_LOG(cb, "%s pl %p size %zu\n", __func__, pl, sizeof *pl * depth);
1983 
1984 	mr = kzalloc(sizeof *mr * depth, GFP_KERNEL);
1985 	DEBUG_LOG(cb, "%s mr %p size %zu\n", __func__, mr, sizeof *mr * depth);
1986 
1987 	fr = kzalloc(sizeof *fr * depth, GFP_KERNEL);
1988 	DEBUG_LOG(cb, "%s fr %p size %zu\n", __func__, fr, sizeof *fr * depth);
1989 
1990 	sgl = kzalloc(sizeof *sgl * depth, GFP_KERNEL);
1991 	DEBUG_LOG(cb, "%s sgl %p size %zu\n", __func__, sgl, sizeof *sgl * depth);
1992 
1993 	write = kzalloc(sizeof *write * depth, GFP_KERNEL);
1994 	DEBUG_LOG(cb, "%s read %p size %zu\n", __func__, write, sizeof *write * depth);
1995 
1996 	inv = kzalloc(sizeof *inv * depth, GFP_KERNEL);
1997 	DEBUG_LOG(cb, "%s inv %p size %zu\n", __func__, inv, sizeof *inv * depth);
1998 
1999 	buf = kzalloc(sizeof *buf * depth, GFP_KERNEL);
2000 	DEBUG_LOG(cb, "%s buf %p size %zu\n", __func__, buf, sizeof *buf * depth);
2001 
2002 	dma_addr = kzalloc(sizeof *dma_addr * depth, GFP_KERNEL);
2003 	DEBUG_LOG(cb, "%s dma_addr %p size %zu\n", __func__, dma_addr, sizeof *dma_addr * depth);
2004 
2005 	if (!pl || !mr || !fr || !write || !sgl || !buf || !dma_addr) {
2006 		PRINTF(cb, "kzalloc failed\n");
2007 		goto err1;
2008 	}
2009 
2010 	for (scnt = 0; scnt < depth; scnt++) {
2011 		pl[scnt] = ib_alloc_fast_reg_page_list(cb->qp->device, plen);
2012 		if (IS_ERR(pl[scnt])) {
2013 			PRINTF(cb, "alloc_fr_page_list failed %ld\n",
2014 			       PTR_ERR(pl[scnt]));
2015 			goto err2;
2016 		}
2017 		DEBUG_LOG(cb, "%s pl[%u] %p\n", __func__, scnt, pl[scnt]);
2018 
2019 		mr[scnt] = ib_alloc_fast_reg_mr(cb->pd, plen);
2020 		if (IS_ERR(mr[scnt])) {
2021 			PRINTF(cb, "alloc_fr failed %ld\n",
2022 			       PTR_ERR(mr[scnt]));
2023 			goto err2;
2024 		}
2025 		DEBUG_LOG(cb, "%s mr[%u] %p\n", __func__, scnt, mr[scnt]);
2026 		ib_update_fast_reg_key(mr[scnt], ++key);
2027 
2028 		buf[scnt] = kmalloc(cb->size, GFP_KERNEL);
2029 		if (!buf[scnt]) {
2030 			PRINTF(cb, "kmalloc failed\n");
2031 			ret = -ENOMEM;
2032 			goto err2;
2033 		}
2034 		DEBUG_LOG(cb, "%s buf[%u] %p\n", __func__, scnt, buf[scnt]);
2035 		dma_addr[scnt] = ib_dma_map_single(cb->pd->device,
2036 						   buf[scnt], cb->size,
2037 						   DMA_BIDIRECTIONAL);
2038 		if (dma_mapping_error(cb->pd->device->dma_device,
2039 		    dma_addr[scnt])) {
2040 			PRINTF(cb, "dma_map failed\n");
2041 			ret = -ENOMEM;
2042 			goto err2;
2043 		}
2044 		DEBUG_LOG(cb, "%s dma_addr[%u] %p\n", __func__, scnt, (void *)dma_addr[scnt]);
2045 		for (i=0; i<plen; i++) {
2046 			pl[scnt]->page_list[i] = ((unsigned long)dma_addr[scnt] & PAGE_MASK) + (i * PAGE_SIZE);
2047 			DEBUG_LOG(cb, "%s pl[%u]->page_list[%u] 0x%jx\n",
2048 				  __func__, scnt, i,  (uintmax_t)pl[scnt]->page_list[i]);
2049 		}
2050 
2051 		write[scnt].opcode = IB_WR_RDMA_WRITE;
2052 		write[scnt].wr_id = scnt;
2053 		write[scnt].wr.rdma.rkey = cb->remote_rkey;
2054 		write[scnt].wr.rdma.remote_addr = cb->remote_addr;
2055 		write[scnt].num_sge = 1;
2056 		write[scnt].sg_list = &cb->rdma_sgl;
2057 		write[scnt].sg_list->length = cb->size;
2058 		write[scnt].next = &fr[scnt];
2059 
2060 		fr[scnt].opcode = IB_WR_FAST_REG_MR;
2061 		fr[scnt].wr_id = scnt;
2062 		fr[scnt].wr.fast_reg.page_shift = PAGE_SHIFT;
2063 		fr[scnt].wr.fast_reg.length = cb->size;
2064 		fr[scnt].wr.fast_reg.page_list = pl[scnt];
2065 		fr[scnt].wr.fast_reg.page_list_len = plen;
2066 		fr[scnt].wr.fast_reg.iova_start = (u64)buf[scnt];
2067 		fr[scnt].wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
2068 		fr[scnt].wr.fast_reg.rkey = mr[scnt]->rkey;
2069 		fr[scnt].next = &inv[scnt];
2070 
2071 		inv[scnt].opcode = IB_WR_LOCAL_INV;
2072 		inv[scnt].send_flags = IB_SEND_SIGNALED;
2073 		inv[scnt].ex.invalidate_rkey = mr[scnt]->rkey;
2074 
2075 		ret = ib_post_send(cb->qp, &write[scnt], &bad);
2076 		if (ret) {
2077 			PRINTF(cb, "ib_post_send failed %d\n", ret);
2078 			goto err2;
2079 		}
2080 	}
2081 
2082 	start = time_uptime;
2083 	DEBUG_LOG(cb, "%s starting IO.\n", __func__);
2084 	while (!cb->count || cb->server || count < cb->count) {
2085 		if ((time_uptime - start) >= 9) {
2086 			DEBUG_LOG(cb, "%s pausing 1 tick! count %u\n", __func__,
2087 				  count);
2088 			wait_event_interruptible_timeout(cb->sem,
2089 							 cb->state == ERROR,
2090 							 1);
2091 			if (cb->state == ERROR)
2092 				break;
2093 			start = time_uptime;
2094 		}
2095 		do {
2096 			ret = ib_poll_cq(cb->cq, 1, &wc);
2097 			if (ret < 0) {
2098 				PRINTF(cb, "ib_poll_cq failed %d\n",
2099 				       ret);
2100 				goto err2;
2101 			}
2102 			if (ret == 1) {
2103 				if (wc.status) {
2104 					PRINTF(cb,
2105 					       "completion error %u wr_id %ju "
2106 					       "opcode %d\n", wc.status,
2107 					       (uintmax_t)wc.wr_id, wc.opcode);
2108 					goto err2;
2109 				}
2110 				count++;
2111 				if (count == (cb->count -1))
2112 					cb->rdma_sgl.lkey = 0x00dead;
2113 				if (count == cb->count)
2114 					break;
2115 				ib_update_fast_reg_key(mr[wc.wr_id], ++key);
2116 				fr[wc.wr_id].wr.fast_reg.rkey =
2117 					mr[wc.wr_id]->rkey;
2118 				inv[wc.wr_id].ex.invalidate_rkey =
2119 					mr[wc.wr_id]->rkey;
2120 				ret = ib_post_send(cb->qp, &write[wc.wr_id], &bad);
2121 				if (ret) {
2122 					PRINTF(cb,
2123 					       "ib_post_send failed %d\n", ret);
2124 					goto err2;
2125 				}
2126 			} else if (krping_sigpending()){
2127 				PRINTF(cb, "signal!\n");
2128 				goto err2;
2129 			}
2130 		} while (ret == 1);
2131 	}
2132 	DEBUG_LOG(cb, "%s done!\n", __func__);
2133 err2:
2134 	DEBUG_LOG(cb, "sleeping 1 second\n");
2135 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2136 	DEBUG_LOG(cb, "draining the cq...\n");
2137 	do {
2138 		ret = ib_poll_cq(cb->cq, 1, &wc);
2139 		if (ret < 0) {
2140 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2141 			break;
2142 		}
2143 		if (ret == 1) {
2144 			if (wc.status) {
2145 				PRINTF(cb, "completion error %u "
2146 				       "opcode %u\n", wc.status, wc.opcode);
2147 			}
2148 		}
2149 	} while (ret == 1);
2150 
2151 	DEBUG_LOG(cb, "destroying fr mrs!\n");
2152 	for (scnt = 0; scnt < depth; scnt++) {
2153 		if (mr[scnt]) {
2154 			ib_dereg_mr(mr[scnt]);
2155 			DEBUG_LOG(cb, "%s dereg mr %p\n", __func__, mr[scnt]);
2156 		}
2157 	}
2158 	DEBUG_LOG(cb, "unmapping/freeing bufs!\n");
2159 	for (scnt = 0; scnt < depth; scnt++) {
2160 		if (buf[scnt]) {
2161 			dma_unmap_single(cb->pd->device->dma_device,
2162 					 dma_addr[scnt], cb->size,
2163 					 DMA_BIDIRECTIONAL);
2164 			kfree(buf[scnt]);
2165 			DEBUG_LOG(cb, "%s unmap/free buf %p dma_addr %p\n", __func__, buf[scnt], (void *)dma_addr[scnt]);
2166 		}
2167 	}
2168 	DEBUG_LOG(cb, "destroying fr page lists!\n");
2169 	for (scnt = 0; scnt < depth; scnt++) {
2170 		if (pl[scnt]) {
2171 			DEBUG_LOG(cb, "%s free pl %p\n", __func__, pl[scnt]);
2172 			ib_free_fast_reg_page_list(pl[scnt]);
2173 		}
2174 	}
2175 err1:
2176 	if (pl)
2177 		kfree(pl);
2178 	if (mr)
2179 		kfree(mr);
2180 	if (fr)
2181 		kfree(fr);
2182 	if (write)
2183 		kfree(write);
2184 	if (inv)
2185 		kfree(inv);
2186 	if (sgl)
2187 		kfree(sgl);
2188 	if (buf)
2189 		kfree(buf);
2190 	if (dma_addr)
2191 		kfree(dma_addr);
2192 }
2193 
2194 static void krping_fr_test6_server(struct krping_cb *cb)
2195 {
2196 	struct ib_send_wr *bad_wr;
2197 	struct ib_wc wc;
2198 	int ret;
2199 
2200 	/* Spin waiting for client's Start STAG/TO/Len */
2201 	while (cb->state < RDMA_READ_ADV) {
2202 		krping_cq_event_handler(cb->cq, cb);
2203 	}
2204 	DEBUG_LOG(cb, "%s client STAG %x TO 0x%jx\n", __func__,
2205 		  cb->remote_rkey, (uintmax_t)cb->remote_addr);
2206 
2207 	/* Send STAG/TO/Len to client */
2208 	krping_format_send(cb, cb->start_dma_addr);
2209 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
2210 	if (ret) {
2211 		PRINTF(cb, "post send error %d\n", ret);
2212 		return;
2213 	}
2214 
2215 	/* Spin waiting for send completion */
2216 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
2217 	if (ret < 0) {
2218 		PRINTF(cb, "poll error %d\n", ret);
2219 		return;
2220 	}
2221 	if (wc.status) {
2222 		PRINTF(cb, "send completiong error %d\n", wc.status);
2223 		return;
2224 	}
2225 
2226 	if (cb->duplex)
2227 		krping_fr_test6(cb);
2228 	DEBUG_LOG(cb, "%s waiting for disconnect...\n", __func__);
2229 	wait_event_interruptible(cb->sem, cb->state == ERROR);
2230 }
2231 
2232 static void krping_fr_test6_client(struct krping_cb *cb)
2233 {
2234 	struct ib_send_wr *bad;
2235 	struct ib_wc wc;
2236 	int ret;
2237 
2238 	cb->state = RDMA_READ_ADV;
2239 
2240 	/* Send STAG/TO/Len to server */
2241 	krping_format_send(cb, cb->start_dma_addr);
2242 	if (cb->state == ERROR) {
2243 		PRINTF(cb, "krping_format_send failed\n");
2244 		return;
2245 	}
2246 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad);
2247 	if (ret) {
2248 		PRINTF(cb, "post send error %d\n", ret);
2249 		return;
2250 	}
2251 
2252 	/* Spin waiting for send completion */
2253 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
2254 	if (ret < 0) {
2255 		PRINTF(cb, "poll error %d\n", ret);
2256 		return;
2257 	}
2258 	if (wc.status) {
2259 		PRINTF(cb, "send completion error %d\n", wc.status);
2260 		return;
2261 	}
2262 
2263 	/* Spin waiting for server's Start STAG/TO/Len */
2264 	while (cb->state < RDMA_WRITE_ADV) {
2265 		krping_cq_event_handler(cb->cq, cb);
2266 	}
2267 	DEBUG_LOG(cb, "%s server STAG %x TO 0x%jx\n", __func__, cb->remote_rkey,
2268 	    (uintmax_t)cb->remote_addr);
2269 
2270 	return krping_fr_test6(cb);
2271 }
2272 
2273 static void krping_run_server(struct krping_cb *cb)
2274 {
2275 	struct ib_recv_wr *bad_wr;
2276 	int ret;
2277 
2278 	ret = krping_bind_server(cb);
2279 	if (ret)
2280 		return;
2281 
2282 	ret = krping_setup_qp(cb, cb->child_cm_id);
2283 	if (ret) {
2284 		PRINTF(cb, "setup_qp failed: %d\n", ret);
2285 		goto err0;
2286 	}
2287 
2288 	ret = krping_setup_buffers(cb);
2289 	if (ret) {
2290 		PRINTF(cb, "krping_setup_buffers failed: %d\n", ret);
2291 		goto err1;
2292 	}
2293 
2294 	ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
2295 	if (ret) {
2296 		PRINTF(cb, "ib_post_recv failed: %d\n", ret);
2297 		goto err2;
2298 	}
2299 
2300 	ret = krping_accept(cb);
2301 	if (ret) {
2302 		PRINTF(cb, "connect error %d\n", ret);
2303 		goto err2;
2304 	}
2305 
2306 	if (cb->wlat)
2307 		krping_wlat_test_server(cb);
2308 	else if (cb->rlat)
2309 		krping_rlat_test_server(cb);
2310 	else if (cb->bw)
2311 		krping_bw_test_server(cb);
2312 	else if (cb->frtest) {
2313 		switch (cb->testnum) {
2314 		case 1:
2315 		case 2:
2316 		case 3:
2317 		case 4:
2318 			krping_fr_test_server(cb);
2319 			break;
2320 		case 5:
2321 			krping_fr_test5_server(cb);
2322 			break;
2323 		case 6:
2324 			krping_fr_test6_server(cb);
2325 			break;
2326 		default:
2327 			PRINTF(cb, "unknown fr test %d\n", cb->testnum);
2328 			goto err2;
2329 			break;
2330 		}
2331 	} else
2332 		krping_test_server(cb);
2333 	rdma_disconnect(cb->child_cm_id);
2334 err2:
2335 	krping_free_buffers(cb);
2336 err1:
2337 	krping_free_qp(cb);
2338 err0:
2339 	rdma_destroy_id(cb->child_cm_id);
2340 }
2341 
2342 static void krping_test_client(struct krping_cb *cb)
2343 {
2344 	int ping, start, cc, i, ret;
2345 	struct ib_send_wr *bad_wr;
2346 	unsigned char c;
2347 
2348 	start = 65;
2349 	for (ping = 0; !cb->count || ping < cb->count; ping++) {
2350 		cb->state = RDMA_READ_ADV;
2351 
2352 		/* Put some ascii text in the buffer. */
2353 		cc = sprintf(cb->start_buf, "rdma-ping-%d: ", ping);
2354 		for (i = cc, c = start; i < cb->size; i++) {
2355 			cb->start_buf[i] = c;
2356 			c++;
2357 			if (c > 122)
2358 				c = 65;
2359 		}
2360 		start++;
2361 		if (start > 122)
2362 			start = 65;
2363 		cb->start_buf[cb->size - 1] = 0;
2364 
2365 		krping_format_send(cb, cb->start_dma_addr);
2366 		if (cb->state == ERROR) {
2367 			PRINTF(cb, "krping_format_send failed\n");
2368 			break;
2369 		}
2370 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
2371 		if (ret) {
2372 			PRINTF(cb, "post send error %d\n", ret);
2373 			break;
2374 		}
2375 
2376 		/* Wait for server to ACK */
2377 		wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV);
2378 		if (cb->state != RDMA_WRITE_ADV) {
2379 			PRINTF(cb,
2380 			       "wait for RDMA_WRITE_ADV state %d\n",
2381 			       cb->state);
2382 			break;
2383 		}
2384 
2385 		krping_format_send(cb, cb->rdma_dma_addr);
2386 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
2387 		if (ret) {
2388 			PRINTF(cb, "post send error %d\n", ret);
2389 			break;
2390 		}
2391 
2392 		/* Wait for the server to say the RDMA Write is complete. */
2393 		wait_event_interruptible(cb->sem,
2394 					 cb->state >= RDMA_WRITE_COMPLETE);
2395 		if (cb->state != RDMA_WRITE_COMPLETE) {
2396 			PRINTF(cb,
2397 			       "wait for RDMA_WRITE_COMPLETE state %d\n",
2398 			       cb->state);
2399 			break;
2400 		}
2401 
2402 		if (cb->validate)
2403 			if (memcmp(cb->start_buf, cb->rdma_buf, cb->size)) {
2404 				PRINTF(cb, "data mismatch!\n");
2405 				break;
2406 			}
2407 
2408 		if (cb->verbose) {
2409 			if (strlen(cb->rdma_buf) > 128) {
2410 				char msgbuf[128];
2411 
2412 				strlcpy(msgbuf, cb->rdma_buf, sizeof(msgbuf));
2413 				PRINTF(cb, "ping data stripped: %s\n",
2414 				       msgbuf);
2415 			} else
2416 				PRINTF(cb, "ping data: %s\n", cb->rdma_buf);
2417 		}
2418 #ifdef SLOW_KRPING
2419 		wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2420 #endif
2421 	}
2422 }
2423 
2424 static void krping_rlat_test_client(struct krping_cb *cb)
2425 {
2426 	struct ib_send_wr *bad_wr;
2427 	struct ib_wc wc;
2428 	int ret;
2429 
2430 	cb->state = RDMA_READ_ADV;
2431 
2432 	/* Send STAG/TO/Len to client */
2433 	krping_format_send(cb, cb->start_dma_addr);
2434 	if (cb->state == ERROR) {
2435 		PRINTF(cb, "krping_format_send failed\n");
2436 		return;
2437 	}
2438 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
2439 	if (ret) {
2440 		PRINTF(cb, "post send error %d\n", ret);
2441 		return;
2442 	}
2443 
2444 	/* Spin waiting for send completion */
2445 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
2446 	if (ret < 0) {
2447 		PRINTF(cb, "poll error %d\n", ret);
2448 		return;
2449 	}
2450 	if (wc.status) {
2451 		PRINTF(cb, "send completion error %d\n", wc.status);
2452 		return;
2453 	}
2454 
2455 	/* Spin waiting for server's Start STAG/TO/Len */
2456 	while (cb->state < RDMA_WRITE_ADV) {
2457 		krping_cq_event_handler(cb->cq, cb);
2458 	}
2459 
2460 #if 0
2461 {
2462 	int i;
2463 	struct timeval start, stop;
2464 	time_t sec;
2465 	suseconds_t usec;
2466 	unsigned long long elapsed;
2467 	struct ib_wc wc;
2468 	struct ib_send_wr *bad_wr;
2469 	int ne;
2470 
2471 	cb->rdma_sq_wr.opcode = IB_WR_RDMA_WRITE;
2472 	cb->rdma_sq_wr.wr.rdma.rkey = cb->remote_rkey;
2473 	cb->rdma_sq_wr.wr.rdma.remote_addr = cb->remote_addr;
2474 	cb->rdma_sq_wr.sg_list->length = 0;
2475 	cb->rdma_sq_wr.num_sge = 0;
2476 
2477 	microtime(&start);
2478 	for (i=0; i < 100000; i++) {
2479 		if (ib_post_send(cb->qp, &cb->rdma_sq_wr, &bad_wr)) {
2480 			PRINTF(cb, "Couldn't post send\n");
2481 			return;
2482 		}
2483 		do {
2484 			ne = ib_poll_cq(cb->cq, 1, &wc);
2485 		} while (ne == 0);
2486 		if (ne < 0) {
2487 			PRINTF(cb, "poll CQ failed %d\n", ne);
2488 			return;
2489 		}
2490 		if (wc.status != IB_WC_SUCCESS) {
2491 			PRINTF(cb, "Completion wth error at %s:\n",
2492 				cb->server ? "server" : "client");
2493 			PRINTF(cb, "Failed status %d: wr_id %d\n",
2494 				wc.status, (int) wc.wr_id);
2495 			return;
2496 		}
2497 	}
2498 	microtime(&stop);
2499 
2500 	if (stop.tv_usec < start.tv_usec) {
2501 		stop.tv_usec += 1000000;
2502 		stop.tv_sec  -= 1;
2503 	}
2504 	sec     = stop.tv_sec - start.tv_sec;
2505 	usec    = stop.tv_usec - start.tv_usec;
2506 	elapsed = sec * 1000000 + usec;
2507 	PRINTF(cb, "0B-write-lat iters 100000 usec %llu\n", elapsed);
2508 }
2509 #endif
2510 
2511 	rlat_test(cb);
2512 }
2513 
2514 static void krping_wlat_test_client(struct krping_cb *cb)
2515 {
2516 	struct ib_send_wr *bad_wr;
2517 	struct ib_wc wc;
2518 	int ret;
2519 
2520 	cb->state = RDMA_READ_ADV;
2521 
2522 	/* Send STAG/TO/Len to client */
2523 	krping_format_send(cb, cb->start_dma_addr);
2524 	if (cb->state == ERROR) {
2525 		PRINTF(cb, "krping_format_send failed\n");
2526 		return;
2527 	}
2528 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
2529 	if (ret) {
2530 		PRINTF(cb, "post send error %d\n", ret);
2531 		return;
2532 	}
2533 
2534 	/* Spin waiting for send completion */
2535 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
2536 	if (ret < 0) {
2537 		PRINTF(cb, "poll error %d\n", ret);
2538 		return;
2539 	}
2540 	if (wc.status) {
2541 		PRINTF(cb, "send completion error %d\n", wc.status);
2542 		return;
2543 	}
2544 
2545 	/* Spin waiting for server's Start STAG/TO/Len */
2546 	while (cb->state < RDMA_WRITE_ADV) {
2547 		krping_cq_event_handler(cb->cq, cb);
2548 	}
2549 
2550 	wlat_test(cb);
2551 }
2552 
2553 static void krping_bw_test_client(struct krping_cb *cb)
2554 {
2555 	struct ib_send_wr *bad_wr;
2556 	struct ib_wc wc;
2557 	int ret;
2558 
2559 	cb->state = RDMA_READ_ADV;
2560 
2561 	/* Send STAG/TO/Len to client */
2562 	krping_format_send(cb, cb->start_dma_addr);
2563 	if (cb->state == ERROR) {
2564 		PRINTF(cb, "krping_format_send failed\n");
2565 		return;
2566 	}
2567 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
2568 	if (ret) {
2569 		PRINTF(cb, "post send error %d\n", ret);
2570 		return;
2571 	}
2572 
2573 	/* Spin waiting for send completion */
2574 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
2575 	if (ret < 0) {
2576 		PRINTF(cb, "poll error %d\n", ret);
2577 		return;
2578 	}
2579 	if (wc.status) {
2580 		PRINTF(cb, "send completion error %d\n", wc.status);
2581 		return;
2582 	}
2583 
2584 	/* Spin waiting for server's Start STAG/TO/Len */
2585 	while (cb->state < RDMA_WRITE_ADV) {
2586 		krping_cq_event_handler(cb->cq, cb);
2587 	}
2588 
2589 	bw_test(cb);
2590 }
2591 
2592 
2593 /*
2594  * fastreg 2 valid different mrs and verify the completions.
2595  */
2596 static void krping_fr_test1(struct krping_cb *cb)
2597 {
2598 	struct ib_fast_reg_page_list *pl;
2599 	struct ib_send_wr fr, *bad;
2600 	struct ib_wc wc;
2601 	struct ib_mr *mr1, *mr2;
2602 	int i;
2603 	int ret;
2604 	int size = cb->size;
2605 	int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
2606 	int count = 0;
2607 
2608 	pl = ib_alloc_fast_reg_page_list(cb->qp->device, plen);
2609 	if (IS_ERR(pl)) {
2610 		PRINTF(cb, "ib_alloc_fast_reg_page_list failed %ld\n", PTR_ERR(pl));
2611 		return;
2612 	}
2613 
2614 	mr1 = ib_alloc_fast_reg_mr(cb->pd, plen);
2615 	if (IS_ERR(mr1)) {
2616 		PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl));
2617 		goto err1;
2618 	}
2619 	mr2 = ib_alloc_fast_reg_mr(cb->pd, plen);
2620 	if (IS_ERR(mr2)) {
2621 		PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl));
2622 		goto err2;
2623 	}
2624 
2625 
2626 	for (i=0; i<plen; i++)
2627 		pl->page_list[i] = i * PAGE_SIZE;
2628 
2629 	memset(&fr, 0, sizeof fr);
2630 	fr.opcode = IB_WR_FAST_REG_MR;
2631 	fr.wr_id = 1;
2632 	fr.wr.fast_reg.page_shift = PAGE_SHIFT;
2633 	fr.wr.fast_reg.length = size;
2634 	fr.wr.fast_reg.page_list = pl;
2635 	fr.wr.fast_reg.page_list_len = plen;
2636 	fr.wr.fast_reg.iova_start = 0;
2637 	fr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
2638 	fr.send_flags = IB_SEND_SIGNALED;
2639 	fr.wr.fast_reg.rkey = mr1->rkey;
2640 	DEBUG_LOG(cb, "%s fr1: stag 0x%x plen %u size %u depth %u\n", __func__, fr.wr.fast_reg.rkey, plen, cb->size, cb->txdepth);
2641 	ret = ib_post_send(cb->qp, &fr, &bad);
2642 	if (ret) {
2643 		PRINTF(cb, "ib_post_send failed %d\n", ret);
2644 		goto err3;
2645 	}
2646 	fr.wr.fast_reg.rkey = mr2->rkey;
2647 	DEBUG_LOG(cb, "%s fr2: stag 0x%x plen %u size %u depth %u\n", __func__, fr.wr.fast_reg.rkey, plen, cb->size, cb->txdepth);
2648 	ret = ib_post_send(cb->qp, &fr, &bad);
2649 	if (ret) {
2650 		PRINTF(cb, "ib_post_send failed %d\n", ret);
2651 		goto err3;
2652 	}
2653 
2654 	DEBUG_LOG(cb, "sleeping 1 second\n");
2655 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2656 	do {
2657 		ret = ib_poll_cq(cb->cq, 1, &wc);
2658 		if (ret < 0) {
2659 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2660 			goto err3;
2661 		}
2662 		if (ret == 1) {
2663 			DEBUG_LOG(cb, "completion status %u wr %s\n",
2664 				  wc.status, wc.wr_id == 1 ? "fr" : "inv");
2665 			count++;
2666 		} else if (krping_sigpending()) {
2667 			PRINTF(cb, "signal!\n");
2668 			goto err3;
2669 		}
2670 
2671 		wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2672 	} while (count != 2);
2673 err3:
2674 	DEBUG_LOG(cb, "sleeping 1 second\n");
2675 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2676 	DEBUG_LOG(cb, "draining the cq...\n");
2677 	do {
2678 		ret = ib_poll_cq(cb->cq, 1, &wc);
2679 		if (ret < 0) {
2680 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2681 			break;
2682 		}
2683 		if (ret == 1) {
2684 			PRINTF(cb, "completion %u opcode %u\n", wc.status, wc.opcode);
2685 		}
2686 	} while (ret == 1);
2687 	DEBUG_LOG(cb, "destroying fr mr2!\n");
2688 
2689 	ib_dereg_mr(mr2);
2690 err2:
2691 	DEBUG_LOG(cb, "destroying fr mr1!\n");
2692 	ib_dereg_mr(mr1);
2693 err1:
2694 	DEBUG_LOG(cb, "destroying fr page list!\n");
2695 	ib_free_fast_reg_page_list(pl);
2696 	DEBUG_LOG(cb, "%s done!\n", __func__);
2697 }
2698 
2699 /*
2700  * fastreg the same mr twice, 2nd one should produce error cqe.
2701  */
2702 static void krping_fr_test2(struct krping_cb *cb)
2703 {
2704 	struct ib_fast_reg_page_list *pl;
2705 	struct ib_send_wr fr, *bad;
2706 	struct ib_wc wc;
2707 	struct ib_mr *mr1;
2708 	int i;
2709 	int ret;
2710 	int size = cb->size;
2711 	int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
2712 	int count = 0;
2713 
2714 	pl = ib_alloc_fast_reg_page_list(cb->qp->device, plen);
2715 	if (IS_ERR(pl)) {
2716 		PRINTF(cb, "ib_alloc_fast_reg_page_list failed %ld\n", PTR_ERR(pl));
2717 		return;
2718 	}
2719 
2720 	mr1 = ib_alloc_fast_reg_mr(cb->pd, plen);
2721 	if (IS_ERR(mr1)) {
2722 		PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl));
2723 		goto err1;
2724 	}
2725 
2726 	for (i=0; i<plen; i++)
2727 		pl->page_list[i] = i * PAGE_SIZE;
2728 
2729 	memset(&fr, 0, sizeof fr);
2730 	fr.opcode = IB_WR_FAST_REG_MR;
2731 	fr.wr_id = 1;
2732 	fr.wr.fast_reg.page_shift = PAGE_SHIFT;
2733 	fr.wr.fast_reg.length = size;
2734 	fr.wr.fast_reg.page_list = pl;
2735 	fr.wr.fast_reg.page_list_len = plen;
2736 	fr.wr.fast_reg.iova_start = 0;
2737 	fr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
2738 	fr.send_flags = IB_SEND_SIGNALED;
2739 	fr.wr.fast_reg.rkey = mr1->rkey;
2740 	DEBUG_LOG(cb, "%s fr1: stag 0x%x plen %u size %u depth %u\n", __func__, fr.wr.fast_reg.rkey, plen, cb->size, cb->txdepth);
2741 	ret = ib_post_send(cb->qp, &fr, &bad);
2742 	if (ret) {
2743 		PRINTF(cb, "ib_post_send failed %d\n", ret);
2744 		goto err3;
2745 	}
2746 	DEBUG_LOG(cb, "%s fr2: stag 0x%x plen %u size %u depth %u\n", __func__, fr.wr.fast_reg.rkey, plen, cb->size, cb->txdepth);
2747 	ret = ib_post_send(cb->qp, &fr, &bad);
2748 	if (ret) {
2749 		PRINTF(cb, "ib_post_send failed %d\n", ret);
2750 		goto err3;
2751 	}
2752 
2753 	DEBUG_LOG(cb, "sleeping 1 second\n");
2754 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2755 	do {
2756 		ret = ib_poll_cq(cb->cq, 1, &wc);
2757 		if (ret < 0) {
2758 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2759 			goto err3;
2760 		}
2761 		if (ret == 1) {
2762 			DEBUG_LOG(cb, "completion status %u wr %s\n",
2763 				  wc.status, wc.wr_id == 1 ? "fr" : "inv");
2764 			count++;
2765 		} else if (krping_sigpending()) {
2766 			PRINTF(cb, "signal!\n");
2767 			goto err3;
2768 		}
2769 		wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2770 	} while (count != 2);
2771 err3:
2772 	DEBUG_LOG(cb, "sleeping 1 second\n");
2773 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2774 	DEBUG_LOG(cb, "draining the cq...\n");
2775 	do {
2776 		ret = ib_poll_cq(cb->cq, 1, &wc);
2777 		if (ret < 0) {
2778 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2779 			break;
2780 		}
2781 		if (ret == 1) {
2782 			PRINTF(cb, "completion %u opcode %u\n", wc.status, wc.opcode);
2783 		}
2784 	} while (ret == 1);
2785 	DEBUG_LOG(cb, "destroying fr mr1!\n");
2786 	ib_dereg_mr(mr1);
2787 err1:
2788 	DEBUG_LOG(cb, "destroying fr page list!\n");
2789 	ib_free_fast_reg_page_list(pl);
2790 	DEBUG_LOG(cb, "%s done!\n", __func__);
2791 }
2792 
2793 /*
2794  * fastreg pipelined in a loop as fast as we can until the user interrupts.
2795  * NOTE: every 9 seconds we sleep for 1 second to keep the kernel happy.
2796  */
2797 static void krping_fr_test3(struct krping_cb *cb)
2798 {
2799 	struct ib_fast_reg_page_list *pl;
2800 	struct ib_send_wr fr, inv, *bad;
2801 	struct ib_wc wc;
2802 	u8 key = 0;
2803 	struct ib_mr *mr;
2804 	int i;
2805 	int ret;
2806 	int size = cb->size;
2807 	int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
2808 	unsigned long start;
2809 	int count = 0;
2810 	int scnt = 0;
2811 
2812 
2813 	pl = ib_alloc_fast_reg_page_list(cb->qp->device, plen);
2814 	if (IS_ERR(pl)) {
2815 		PRINTF(cb, "ib_alloc_fast_reg_page_list failed %ld\n", PTR_ERR(pl));
2816 		return;
2817 	}
2818 
2819 	mr = ib_alloc_fast_reg_mr(cb->pd, plen);
2820 	if (IS_ERR(mr)) {
2821 		PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl));
2822 		goto err1;
2823 	}
2824 
2825 	for (i=0; i<plen; i++)
2826 		pl->page_list[i] = i * PAGE_SIZE;
2827 
2828 	memset(&fr, 0, sizeof fr);
2829 	fr.opcode = IB_WR_FAST_REG_MR;
2830 	fr.wr.fast_reg.page_shift = PAGE_SHIFT;
2831 	fr.wr.fast_reg.length = size;
2832 	fr.wr.fast_reg.page_list = pl;
2833 	fr.wr.fast_reg.page_list_len = plen;
2834 	fr.wr.fast_reg.iova_start = 0;
2835 	fr.send_flags = IB_SEND_SIGNALED;
2836 	fr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
2837 	fr.next = &inv;
2838 	memset(&inv, 0, sizeof inv);
2839 	inv.opcode = IB_WR_LOCAL_INV;
2840 	inv.send_flags = IB_SEND_SIGNALED;
2841 
2842 	DEBUG_LOG(cb, "fr_test: stag index 0x%x plen %u size %u depth %u\n", mr->rkey >> 8, plen, cb->size, cb->txdepth);
2843 	start = time_uptime;
2844 	while (1) {
2845 		if ((time_uptime - start) >= 9) {
2846 			DEBUG_LOG(cb, "fr_test: pausing 1 second! count %u latest size %u plen %u\n", count, size, plen);
2847 			wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2848 			if (cb->state == ERROR)
2849 				break;
2850 			start = time_uptime;
2851 		}
2852 		while (scnt < (cb->txdepth>>1)) {
2853 			ib_update_fast_reg_key(mr, ++key);
2854 			fr.wr.fast_reg.rkey = mr->rkey;
2855 			inv.ex.invalidate_rkey = mr->rkey;
2856 			size = arc4random() % cb->size;
2857 			if (size == 0)
2858 				size = cb->size;
2859 			plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
2860 			fr.wr.fast_reg.length = size;
2861 			fr.wr.fast_reg.page_list_len = plen;
2862 			ret = ib_post_send(cb->qp, &fr, &bad);
2863 			if (ret) {
2864 				PRINTF(cb, "ib_post_send failed %d\n", ret);
2865 				goto err2;
2866 			}
2867 			scnt+=2;
2868 		}
2869 
2870 		do {
2871 			ret = ib_poll_cq(cb->cq, 1, &wc);
2872 			if (ret < 0) {
2873 				PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2874 				goto err2;
2875 			}
2876 			if (ret == 1) {
2877 				if (wc.status) {
2878 					PRINTF(cb, "completion error %u\n", wc.status);
2879 					goto err2;
2880 				}
2881 				count++;
2882 				scnt--;
2883 			}
2884 			else if (krping_sigpending()) {
2885 				PRINTF(cb, "signal!\n");
2886 				goto err2;
2887 			}
2888 		} while (ret == 1);
2889 	}
2890 err2:
2891 	DEBUG_LOG(cb, "sleeping 1 second\n");
2892 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2893 	DEBUG_LOG(cb, "draining the cq...\n");
2894 	do {
2895 		ret = ib_poll_cq(cb->cq, 1, &wc);
2896 		if (ret < 0) {
2897 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2898 			break;
2899 		}
2900 		if (ret == 1) {
2901 			if (wc.status) {
2902 				PRINTF(cb, "completion error %u opcode %u\n", wc.status, wc.opcode);
2903 			}
2904 		}
2905 	} while (ret == 1);
2906 	DEBUG_LOG(cb, "fr_test: done!\n");
2907 	ib_dereg_mr(mr);
2908 err1:
2909 	DEBUG_LOG(cb, "destroying fr page list!\n");
2910 	ib_free_fast_reg_page_list(pl);
2911 	DEBUG_LOG(cb, "%s done!\n", __func__);
2912 }
2913 
2914 /*
2915  * fastreg 1 and invalidate 1 mr and verify completion.
2916  */
2917 static void krping_fr_test4(struct krping_cb *cb)
2918 {
2919 	struct ib_fast_reg_page_list *pl;
2920 	struct ib_send_wr fr, inv, *bad;
2921 	struct ib_wc wc;
2922 	struct ib_mr *mr1;
2923 	int i;
2924 	int ret;
2925 	int size = cb->size;
2926 	int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
2927 	int count = 0;
2928 
2929 	pl = ib_alloc_fast_reg_page_list(cb->qp->device, plen);
2930 	if (IS_ERR(pl)) {
2931 		PRINTF(cb, "ib_alloc_fast_reg_page_list failed %ld\n", PTR_ERR(pl));
2932 		return;
2933 	}
2934 
2935 	mr1 = ib_alloc_fast_reg_mr(cb->pd, plen);
2936 	if (IS_ERR(mr1)) {
2937 		PRINTF(cb, "ib_alloc_fast_reg_mr failed %ld\n", PTR_ERR(pl));
2938 		goto err1;
2939 	}
2940 
2941 	for (i=0; i<plen; i++)
2942 		pl->page_list[i] = i * PAGE_SIZE;
2943 
2944 	memset(&fr, 0, sizeof fr);
2945 	fr.opcode = IB_WR_FAST_REG_MR;
2946 	fr.wr_id = 1;
2947 	fr.wr.fast_reg.page_shift = PAGE_SHIFT;
2948 	fr.wr.fast_reg.length = size;
2949 	fr.wr.fast_reg.page_list = pl;
2950 	fr.wr.fast_reg.page_list_len = plen;
2951 	fr.wr.fast_reg.iova_start = 0;
2952 	fr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
2953 	fr.send_flags = IB_SEND_SIGNALED;
2954 	fr.wr.fast_reg.rkey = mr1->rkey;
2955 	fr.next = &inv;
2956 	memset(&inv, 0, sizeof inv);
2957 	inv.opcode = IB_WR_LOCAL_INV;
2958 	inv.ex.invalidate_rkey = mr1->rkey;
2959 
2960 	DEBUG_LOG(cb, "%s fr1: stag 0x%x plen %u size %u depth %u\n", __func__, fr.wr.fast_reg.rkey, plen, cb->size, cb->txdepth);
2961 	ret = ib_post_send(cb->qp, &fr, &bad);
2962 	if (ret) {
2963 		PRINTF(cb, "ib_post_send failed %d\n", ret);
2964 		goto err3;
2965 	}
2966 	DEBUG_LOG(cb, "sleeping 1 second\n");
2967 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2968 	do {
2969 		ret = ib_poll_cq(cb->cq, 1, &wc);
2970 		if (ret < 0) {
2971 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2972 			goto err3;
2973 		}
2974 		if (ret == 1) {
2975 			DEBUG_LOG(cb, "completion status %u wr %s\n",
2976 				  wc.status, wc.wr_id == 1 ? "fr" : "inv");
2977 			count++;
2978 		} else if (krping_sigpending()) {
2979 			PRINTF(cb, "signal!\n");
2980 			goto err3;
2981 		}
2982 		wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2983 	} while (count != 1);
2984 err3:
2985 	DEBUG_LOG(cb, "sleeping 1 second\n");
2986 	wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
2987 	DEBUG_LOG(cb, "draining the cq...\n");
2988 	do {
2989 		ret = ib_poll_cq(cb->cq, 1, &wc);
2990 		if (ret < 0) {
2991 			PRINTF(cb, "ib_poll_cq failed %d\n", ret);
2992 			break;
2993 		}
2994 		if (ret == 1) {
2995 			PRINTF(cb, "completion %u opcode %u\n", wc.status, wc.opcode);
2996 		}
2997 	} while (ret == 1);
2998 	DEBUG_LOG(cb, "destroying fr mr1!\n");
2999 	ib_dereg_mr(mr1);
3000 err1:
3001 	DEBUG_LOG(cb, "destroying fr page list!\n");
3002 	ib_free_fast_reg_page_list(pl);
3003 	DEBUG_LOG(cb, "%s done!\n", __func__);
3004 }
3005 
3006 static void krping_fr_test(struct krping_cb *cb)
3007 {
3008 	switch (cb->testnum) {
3009 	case 1:
3010 		krping_fr_test1(cb);
3011 		break;
3012 	case 2:
3013 		krping_fr_test2(cb);
3014 		break;
3015 	case 3:
3016 		krping_fr_test3(cb);
3017 		break;
3018 	case 4:
3019 		krping_fr_test4(cb);
3020 		break;
3021 	case 5:
3022 		krping_fr_test5_client(cb);
3023 		break;
3024 	case 6:
3025 		krping_fr_test6_client(cb);
3026 		break;
3027 	default:
3028 		PRINTF(cb, "Unkown frtest num %u\n", cb->testnum);
3029 		break;
3030 	}
3031 }
3032 
3033 static int krping_connect_client(struct krping_cb *cb)
3034 {
3035 	struct rdma_conn_param conn_param;
3036 	int ret;
3037 
3038 	memset(&conn_param, 0, sizeof conn_param);
3039 	conn_param.responder_resources = 1;
3040 	conn_param.initiator_depth = 1;
3041 	conn_param.retry_count = 10;
3042 
3043 	ret = rdma_connect(cb->cm_id, &conn_param);
3044 	if (ret) {
3045 		PRINTF(cb, "rdma_connect error %d\n", ret);
3046 		return ret;
3047 	}
3048 
3049 	wait_event_interruptible(cb->sem, cb->state >= CONNECTED);
3050 	if (cb->state == ERROR) {
3051 		PRINTF(cb, "wait for CONNECTED state %d\n", cb->state);
3052 		return -1;
3053 	}
3054 
3055 	DEBUG_LOG(cb, "rdma_connect successful\n");
3056 	return 0;
3057 }
3058 
3059 static int krping_bind_client(struct krping_cb *cb)
3060 {
3061 	struct sockaddr_in sin;
3062 	int ret;
3063 
3064 	memset(&sin, 0, sizeof(sin));
3065 	sin.sin_len = sizeof sin;
3066 	sin.sin_family = AF_INET;
3067 	sin.sin_addr.s_addr = cb->addr.s_addr;
3068 	sin.sin_port = cb->port;
3069 
3070 	ret = rdma_resolve_addr(cb->cm_id, NULL, (struct sockaddr *) &sin,
3071 				2000);
3072 	if (ret) {
3073 		PRINTF(cb, "rdma_resolve_addr error %d\n", ret);
3074 		return ret;
3075 	}
3076 
3077 	wait_event_interruptible(cb->sem, cb->state >= ROUTE_RESOLVED);
3078 	if (cb->state != ROUTE_RESOLVED) {
3079 		PRINTF(cb,
3080 		       "addr/route resolution did not resolve: state %d\n",
3081 		       cb->state);
3082 		return -EINTR;
3083 	}
3084 
3085 	if (cb->mem == FASTREG && !fastreg_supported(cb, 0))
3086 		return -EINVAL;
3087 
3088 	DEBUG_LOG(cb, "rdma_resolve_addr - rdma_resolve_route successful\n");
3089 	return 0;
3090 }
3091 
3092 static void krping_run_client(struct krping_cb *cb)
3093 {
3094 	struct ib_recv_wr *bad_wr;
3095 	int ret;
3096 
3097 	ret = krping_bind_client(cb);
3098 	if (ret)
3099 		return;
3100 
3101 	ret = krping_setup_qp(cb, cb->cm_id);
3102 	if (ret) {
3103 		PRINTF(cb, "setup_qp failed: %d\n", ret);
3104 		return;
3105 	}
3106 
3107 	ret = krping_setup_buffers(cb);
3108 	if (ret) {
3109 		PRINTF(cb, "krping_setup_buffers failed: %d\n", ret);
3110 		goto err1;
3111 	}
3112 
3113 	ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
3114 	if (ret) {
3115 		PRINTF(cb, "ib_post_recv failed: %d\n", ret);
3116 		goto err2;
3117 	}
3118 
3119 	ret = krping_connect_client(cb);
3120 	if (ret) {
3121 		PRINTF(cb, "connect error %d\n", ret);
3122 		goto err2;
3123 	}
3124 
3125 	if (cb->wlat)
3126 		krping_wlat_test_client(cb);
3127 	else if (cb->rlat)
3128 		krping_rlat_test_client(cb);
3129 	else if (cb->bw)
3130 		krping_bw_test_client(cb);
3131 	else if (cb->frtest)
3132 		krping_fr_test(cb);
3133 	else
3134 		krping_test_client(cb);
3135 	rdma_disconnect(cb->cm_id);
3136 err2:
3137 	krping_free_buffers(cb);
3138 err1:
3139 	krping_free_qp(cb);
3140 }
3141 
3142 int krping_doit(char *cmd, void *cookie)
3143 {
3144 	struct krping_cb *cb;
3145 	int op;
3146 	int ret = 0;
3147 	char *optarg;
3148 	unsigned long optint;
3149 
3150 	cb = kzalloc(sizeof(*cb), GFP_KERNEL);
3151 	if (!cb)
3152 		return -ENOMEM;
3153 
3154 	mutex_lock(&krping_mutex);
3155 	list_add_tail(&cb->list, &krping_cbs);
3156 	mutex_unlock(&krping_mutex);
3157 
3158 	cb->cookie = cookie;
3159 	cb->server = -1;
3160 	cb->state = IDLE;
3161 	cb->size = 64;
3162 	cb->txdepth = RPING_SQ_DEPTH;
3163 	cb->mem = DMA;
3164 	init_waitqueue_head(&cb->sem);
3165 
3166 	while ((op = krping_getopt("krping", &cmd, krping_opts, NULL, &optarg,
3167 			      &optint)) != 0) {
3168 		switch (op) {
3169 		case 'a':
3170 			cb->addr_str = optarg;
3171 			DEBUG_LOG(cb, "ipaddr (%s)\n", optarg);
3172 			if (!inet_aton(optarg, &cb->addr)) {
3173 				PRINTF(cb, "bad addr string %s\n",
3174 				    optarg);
3175 				ret = EINVAL;
3176 			}
3177 			break;
3178 		case 'p':
3179 			cb->port = htons(optint);
3180 			DEBUG_LOG(cb, "port %d\n", (int)optint);
3181 			break;
3182 		case 'P':
3183 			cb->poll = 1;
3184 			DEBUG_LOG(cb, "server\n");
3185 			break;
3186 		case 's':
3187 			cb->server = 1;
3188 			DEBUG_LOG(cb, "server\n");
3189 			break;
3190 		case 'c':
3191 			cb->server = 0;
3192 			DEBUG_LOG(cb, "client\n");
3193 			break;
3194 		case 'S':
3195 			cb->size = optint;
3196 			if ((cb->size < 1) ||
3197 			    (cb->size > RPING_BUFSIZE)) {
3198 				PRINTF(cb, "Invalid size %d "
3199 				       "(valid range is 1 to %d)\n",
3200 				       cb->size, RPING_BUFSIZE);
3201 				ret = EINVAL;
3202 			} else
3203 				DEBUG_LOG(cb, "size %d\n", (int)optint);
3204 			break;
3205 		case 'C':
3206 			cb->count = optint;
3207 			if (cb->count < 0) {
3208 				PRINTF(cb, "Invalid count %d\n",
3209 					cb->count);
3210 				ret = EINVAL;
3211 			} else
3212 				DEBUG_LOG(cb, "count %d\n", (int) cb->count);
3213 			break;
3214 		case 'v':
3215 			cb->verbose++;
3216 			DEBUG_LOG(cb, "verbose\n");
3217 			break;
3218 		case 'V':
3219 			cb->validate++;
3220 			DEBUG_LOG(cb, "validate data\n");
3221 			break;
3222 		case 'l':
3223 			cb->wlat++;
3224 			break;
3225 		case 'L':
3226 			cb->rlat++;
3227 			break;
3228 		case 'B':
3229 			cb->bw++;
3230 			break;
3231 		case 'd':
3232 			cb->duplex++;
3233 			break;
3234 		case 'm':
3235 			if (!strncmp(optarg, "dma", 3))
3236 				cb->mem = DMA;
3237 			else if (!strncmp(optarg, "fastreg", 7))
3238 				cb->mem = FASTREG;
3239 			else if (!strncmp(optarg, "mw", 2))
3240 				cb->mem = MW;
3241 			else if (!strncmp(optarg, "mr", 2))
3242 				cb->mem = MR;
3243 			else {
3244 				PRINTF(cb, "unknown mem mode %s.  "
3245 					"Must be dma, fastreg, mw, or mr\n",
3246 					optarg);
3247 				ret = -EINVAL;
3248 				break;
3249 			}
3250 			break;
3251 		case 'I':
3252 			cb->server_invalidate = 1;
3253 			break;
3254 		case 'T':
3255 			cb->txdepth = optint;
3256 			DEBUG_LOG(cb, "txdepth %d\n", (int) cb->txdepth);
3257 			break;
3258 		case 'Z':
3259 			cb->local_dma_lkey = 1;
3260 			DEBUG_LOG(cb, "using local dma lkey\n");
3261 			break;
3262 		case 'R':
3263 			cb->read_inv = 1;
3264 			DEBUG_LOG(cb, "using read-with-inv\n");
3265 			break;
3266 		case 'f':
3267 			cb->frtest = 1;
3268 			cb->testnum = optint;
3269 			DEBUG_LOG(cb, "fast-reg test!\n");
3270 			break;
3271 		default:
3272 			PRINTF(cb, "unknown opt %s\n", optarg);
3273 			ret = -EINVAL;
3274 			break;
3275 		}
3276 	}
3277 	if (ret)
3278 		goto out;
3279 
3280 	if (cb->server == -1) {
3281 		PRINTF(cb, "must be either client or server\n");
3282 		ret = -EINVAL;
3283 		goto out;
3284 	}
3285 
3286 	if ((cb->frtest + cb->bw + cb->rlat + cb->wlat) > 1) {
3287 		PRINTF(cb, "Pick only one test: fr, bw, rlat, wlat\n");
3288 		ret = -EINVAL;
3289 		goto out;
3290 	}
3291 	if (cb->server_invalidate && cb->mem != FASTREG) {
3292 		PRINTF(cb, "server_invalidate only valid with fastreg mem_mode\n");
3293 		ret = -EINVAL;
3294 		goto out;
3295 	}
3296 
3297 	if (cb->read_inv && cb->mem != FASTREG) {
3298 		PRINTF(cb, "read_inv only valid with fastreg mem_mode\n");
3299 		ret = -EINVAL;
3300 		goto out;
3301 	}
3302 
3303 	if (cb->mem != MR && (cb->wlat || cb->rlat || cb->bw || cb->frtest)) {
3304 		PRINTF(cb, "wlat, rlat, and bw tests only support mem_mode MR\n");
3305 		ret = -EINVAL;
3306 		goto out;
3307 	}
3308 
3309 	cb->cm_id = rdma_create_id(krping_cma_event_handler, cb, RDMA_PS_TCP, IB_QPT_RC);
3310 	if (IS_ERR(cb->cm_id)) {
3311 		ret = PTR_ERR(cb->cm_id);
3312 		PRINTF(cb, "rdma_create_id error %d\n", ret);
3313 		goto out;
3314 	}
3315 	DEBUG_LOG(cb, "created cm_id %p\n", cb->cm_id);
3316 
3317 	if (cb->server)
3318 		krping_run_server(cb);
3319 	else
3320 		krping_run_client(cb);
3321 
3322 	DEBUG_LOG(cb, "destroy cm_id %p\n", cb->cm_id);
3323 	rdma_destroy_id(cb->cm_id);
3324 out:
3325 	mutex_lock(&krping_mutex);
3326 	list_del(&cb->list);
3327 	mutex_unlock(&krping_mutex);
3328 	kfree(cb);
3329 	return ret;
3330 }
3331 
3332 void
3333 krping_walk_cb_list(void (*f)(struct krping_stats *, void *), void *arg)
3334 {
3335 	struct krping_cb *cb;
3336 
3337 	mutex_lock(&krping_mutex);
3338 	list_for_each_entry(cb, &krping_cbs, list)
3339 	    (*f)(cb->pd ? &cb->stats : NULL, arg);
3340 	mutex_unlock(&krping_mutex);
3341 }
3342 
3343 void krping_init(void)
3344 {
3345 
3346 	mutex_init(&krping_mutex);
3347 }
3348