xref: /freebsd/sys/contrib/rdma/krping/krping.c (revision cfaab41c950f791f7a528814c8bef91116fd7af1)
1 /*
2  * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
3  * Copyright (c) 2006-2009 Open Grid Computing, Inc. All rights reserved.
4  *
5  * This software is available to you under a choice of one of two
6  * licenses.  You may choose to be licensed under the terms of the GNU
7  * General Public License (GPL) Version 2, available from the file
8  * COPYING in the main directory of this source tree, or the
9  * OpenIB.org BSD license below:
10  *
11  *     Redistribution and use in source and binary forms, with or
12  *     without modification, are permitted provided that the following
13  *     conditions are met:
14  *
15  *      - Redistributions of source code must retain the above
16  *        copyright notice, this list of conditions and the following
17  *        disclaimer.
18  *
19  *      - Redistributions in binary form must reproduce the above
20  *        copyright notice, this list of conditions and the following
21  *        disclaimer in the documentation and/or other materials
22  *        provided with the distribution.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31  * SOFTWARE.
32  */
33 
34 #include <sys/cdefs.h>
35 __FBSDID("$FreeBSD$");
36 
37 #include <linux/module.h>
38 #include <linux/moduleparam.h>
39 #include <linux/slab.h>
40 #include <linux/err.h>
41 #include <linux/string.h>
42 #include <linux/list.h>
43 #include <linux/in.h>
44 #include <linux/device.h>
45 #include <linux/pci.h>
46 #include <linux/sched.h>
47 #include <linux/wait.h>
48 
49 #include <asm/atomic.h>
50 
51 #include <rdma/ib_verbs.h>
52 #include <rdma/rdma_cm.h>
53 
54 #include "krping.h"
55 #include "getopt.h"
56 
57 #define PFX "krping: "
58 
59 extern int krping_debug;
60 #define DEBUG_LOG(...) do { if (krping_debug) log(LOG_INFO, __VA_ARGS__); } while (0)
61 #define BIND_INFO 1
62 
63 MODULE_AUTHOR("Steve Wise");
64 MODULE_DESCRIPTION("RDMA ping server");
65 MODULE_LICENSE("Dual BSD/GPL");
66 MODULE_VERSION(krping, 1);
67 MODULE_DEPEND(krping, linuxkpi, 1, 1, 1);
68 
69 static __inline uint64_t
get_cycles(void)70 get_cycles(void)
71 {
72 	return (get_cyclecount());
73 }
74 
75 typedef uint64_t cycles_t;
76 
77 enum mem_type {
78 	DMA = 1,
79 	REG = 2,
80 };
81 
82 static const struct krping_option krping_opts[] = {
83 	{"count", OPT_INT, 'C'},
84 	{"size", OPT_INT, 'S'},
85 	{"addr", OPT_STRING, 'a'},
86 	{"addr6", OPT_STRING, 'A'},
87 	{"port", OPT_INT, 'p'},
88 	{"verbose", OPT_NOPARAM, 'v'},
89 	{"validate", OPT_NOPARAM, 'V'},
90 	{"server", OPT_NOPARAM, 's'},
91 	{"client", OPT_NOPARAM, 'c'},
92 	{"server_inv", OPT_NOPARAM, 'I'},
93  	{"wlat", OPT_NOPARAM, 'l'},
94  	{"rlat", OPT_NOPARAM, 'L'},
95  	{"bw", OPT_NOPARAM, 'B'},
96  	{"duplex", OPT_NOPARAM, 'd'},
97 	{"tos", OPT_INT, 't'},
98  	{"txdepth", OPT_INT, 'T'},
99  	{"poll", OPT_NOPARAM, 'P'},
100  	{"local_dma_lkey", OPT_NOPARAM, 'Z'},
101  	{"read_inv", OPT_NOPARAM, 'R'},
102  	{"fr", OPT_NOPARAM, 'f'},
103 	{NULL, 0, 0}
104 };
105 
106 #define htonll(x) cpu_to_be64((x))
107 #define ntohll(x) cpu_to_be64((x))
108 
109 static DEFINE_MUTEX(krping_mutex);
110 
111 /*
112  * List of running krping threads.
113  */
114 static LIST_HEAD(krping_cbs);
115 
116 /*
117  * Invoke like this, one on each side, using the server's address on
118  * the RDMA device (iw%d):
119  *
120  * /bin/echo server,port=9999,addr=192.168.69.142,validate > /proc/krping
121  * /bin/echo client,port=9999,addr=192.168.69.142,validate > /proc/krping
122  * /bin/echo client,port=9999,addr6=2001:db8:0:f101::1,validate > /proc/krping
123  *
124  * krping "ping/pong" loop:
125  * 	client sends source rkey/addr/len
126  *	server receives source rkey/add/len
127  *	server rdma reads "ping" data from source
128  * 	server sends "go ahead" on rdma read completion
129  *	client sends sink rkey/addr/len
130  * 	server receives sink rkey/addr/len
131  * 	server rdma writes "pong" data to sink
132  * 	server sends "go ahead" on rdma write completion
133  * 	<repeat loop>
134  */
135 
136 /*
137  * These states are used to signal events between the completion handler
138  * and the main client or server thread.
139  *
140  * Once CONNECTED, they cycle through RDMA_READ_ADV, RDMA_WRITE_ADV,
141  * and RDMA_WRITE_COMPLETE for each ping.
142  */
143 enum test_state {
144 	IDLE = 1,
145 	CONNECT_REQUEST,
146 	ADDR_RESOLVED,
147 	ROUTE_RESOLVED,
148 	CONNECTED,
149 	RDMA_READ_ADV,
150 	RDMA_READ_COMPLETE,
151 	RDMA_WRITE_ADV,
152 	RDMA_WRITE_COMPLETE,
153 	ERROR
154 };
155 
156 struct krping_rdma_info {
157 	uint64_t buf;
158 	uint32_t rkey;
159 	uint32_t size;
160 };
161 
162 /*
163  * Default max buffer size for IO...
164  */
165 #define RPING_BUFSIZE 128*1024
166 #define RPING_SQ_DEPTH 64
167 
168 /*
169  * Control block struct.
170  */
171 struct krping_cb {
172 	int server;			/* 0 iff client */
173 	struct ib_cq *cq;
174 	struct ib_pd *pd;
175 	struct ib_qp *qp;
176 
177 	struct ib_mr *dma_mr;
178 
179 	struct ib_fast_reg_page_list *page_list;
180 	int page_list_len;
181 	struct ib_reg_wr reg_mr_wr;
182 	struct ib_send_wr invalidate_wr;
183 	struct ib_mr *reg_mr;
184 	int server_invalidate;
185 	int read_inv;
186 	u8 key;
187 
188 	struct ib_recv_wr rq_wr;	/* recv work request record */
189 	struct ib_sge recv_sgl;		/* recv single SGE */
190 	struct krping_rdma_info recv_buf __aligned(16);	/* malloc'd buffer */
191 	u64 recv_dma_addr;
192 	DECLARE_PCI_UNMAP_ADDR(recv_mapping)
193 
194 	struct ib_send_wr sq_wr;	/* send work requrest record */
195 	struct ib_sge send_sgl;
196 	struct krping_rdma_info send_buf __aligned(16); /* single send buf */
197 	u64 send_dma_addr;
198 	DECLARE_PCI_UNMAP_ADDR(send_mapping)
199 
200 	struct ib_rdma_wr rdma_sq_wr;	/* rdma work request record */
201 	struct ib_sge rdma_sgl;		/* rdma single SGE */
202 	char *rdma_buf;			/* used as rdma sink */
203 	u64  rdma_dma_addr;
204 	DECLARE_PCI_UNMAP_ADDR(rdma_mapping)
205 	struct ib_mr *rdma_mr;
206 
207 	uint32_t remote_rkey;		/* remote guys RKEY */
208 	uint64_t remote_addr;		/* remote guys TO */
209 	uint32_t remote_len;		/* remote guys LEN */
210 
211 	char *start_buf;		/* rdma read src */
212 	u64  start_dma_addr;
213 	DECLARE_PCI_UNMAP_ADDR(start_mapping)
214 	struct ib_mr *start_mr;
215 
216 	enum test_state state;		/* used for cond/signalling */
217 	wait_queue_head_t sem;
218 	struct krping_stats stats;
219 
220 	uint16_t port;			/* dst port in NBO */
221 	u8 addr[16] __aligned(8);	/* dst addr in NBO */
222 	char *addr_str;			/* dst addr string */
223 	uint8_t addr_type;		/* ADDR_FAMILY - IPv4/V6 */
224 	int verbose;			/* verbose logging */
225 	int count;			/* ping count */
226 	int size;			/* ping data size */
227 	int validate;			/* validate ping data */
228 	int wlat;			/* run wlat test */
229 	int rlat;			/* run rlat test */
230 	int bw;				/* run bw test */
231 	int duplex;			/* run bw full duplex test */
232 	int poll;			/* poll or block for rlat test */
233 	int txdepth;			/* SQ depth */
234 	int local_dma_lkey;		/* use 0 for lkey */
235 	int frtest;			/* reg test */
236 	int tos;			/* type of service */
237 
238 	/* CM stuff */
239 	struct rdma_cm_id *cm_id;	/* connection on client side,*/
240 					/* listener on server side. */
241 	struct rdma_cm_id *child_cm_id;	/* connection on server side */
242 	struct list_head list;
243 };
244 
krping_cma_event_handler(struct rdma_cm_id * cma_id,struct rdma_cm_event * event)245 static int krping_cma_event_handler(struct rdma_cm_id *cma_id,
246 				   struct rdma_cm_event *event)
247 {
248 	int ret;
249 	struct krping_cb *cb = cma_id->context;
250 
251 	DEBUG_LOG("cma_event type %d cma_id %p (%s)\n", event->event, cma_id,
252 		  (cma_id == cb->cm_id) ? "parent" : "child");
253 
254 	switch (event->event) {
255 	case RDMA_CM_EVENT_ADDR_RESOLVED:
256 		cb->state = ADDR_RESOLVED;
257 		ret = rdma_resolve_route(cma_id, 2000);
258 		if (ret) {
259 			printk(KERN_ERR PFX "rdma_resolve_route error %d\n",
260 			       ret);
261 			wake_up_interruptible(&cb->sem);
262 		}
263 		break;
264 
265 	case RDMA_CM_EVENT_ROUTE_RESOLVED:
266 		cb->state = ROUTE_RESOLVED;
267 		wake_up_interruptible(&cb->sem);
268 		break;
269 
270 	case RDMA_CM_EVENT_CONNECT_REQUEST:
271 		cb->state = CONNECT_REQUEST;
272 		cb->child_cm_id = cma_id;
273 		DEBUG_LOG("child cma %p\n", cb->child_cm_id);
274 		wake_up_interruptible(&cb->sem);
275 		break;
276 
277 	case RDMA_CM_EVENT_ESTABLISHED:
278 		DEBUG_LOG("ESTABLISHED\n");
279 		if (!cb->server) {
280 			cb->state = CONNECTED;
281 		}
282 		wake_up_interruptible(&cb->sem);
283 		break;
284 
285 	case RDMA_CM_EVENT_ADDR_ERROR:
286 	case RDMA_CM_EVENT_ROUTE_ERROR:
287 	case RDMA_CM_EVENT_CONNECT_ERROR:
288 	case RDMA_CM_EVENT_UNREACHABLE:
289 	case RDMA_CM_EVENT_REJECTED:
290 		printk(KERN_ERR PFX "cma event %d, error %d\n", event->event,
291 		       event->status);
292 		cb->state = ERROR;
293 		wake_up_interruptible(&cb->sem);
294 		break;
295 
296 	case RDMA_CM_EVENT_DISCONNECTED:
297 		printk(KERN_ERR PFX "DISCONNECT EVENT...\n");
298 		cb->state = ERROR;
299 		wake_up_interruptible(&cb->sem);
300 		break;
301 
302 	case RDMA_CM_EVENT_DEVICE_REMOVAL:
303 		printk(KERN_ERR PFX "cma detected device removal!!!!\n");
304 		cb->state = ERROR;
305 		wake_up_interruptible(&cb->sem);
306 		break;
307 
308 	default:
309 		printk(KERN_ERR PFX "oof bad type!\n");
310 		wake_up_interruptible(&cb->sem);
311 		break;
312 	}
313 	return 0;
314 }
315 
server_recv(struct krping_cb * cb,struct ib_wc * wc)316 static int server_recv(struct krping_cb *cb, struct ib_wc *wc)
317 {
318 	if (wc->byte_len != sizeof(cb->recv_buf)) {
319 		printk(KERN_ERR PFX "Received bogus data, size %d\n",
320 		       wc->byte_len);
321 		return -1;
322 	}
323 
324 	cb->remote_rkey = ntohl(cb->recv_buf.rkey);
325 	cb->remote_addr = ntohll(cb->recv_buf.buf);
326 	cb->remote_len  = ntohl(cb->recv_buf.size);
327 	DEBUG_LOG("Received rkey %x addr %llx len %d from peer\n",
328 		  cb->remote_rkey, (unsigned long long)cb->remote_addr,
329 		  cb->remote_len);
330 
331 	if (cb->state <= CONNECTED || cb->state == RDMA_WRITE_COMPLETE)
332 		cb->state = RDMA_READ_ADV;
333 	else
334 		cb->state = RDMA_WRITE_ADV;
335 
336 	return 0;
337 }
338 
client_recv(struct krping_cb * cb,struct ib_wc * wc)339 static int client_recv(struct krping_cb *cb, struct ib_wc *wc)
340 {
341 	if (wc->byte_len != sizeof(cb->recv_buf)) {
342 		printk(KERN_ERR PFX "Received bogus data, size %d\n",
343 		       wc->byte_len);
344 		return -1;
345 	}
346 
347 	if (cb->state == RDMA_READ_ADV)
348 		cb->state = RDMA_WRITE_ADV;
349 	else
350 		cb->state = RDMA_WRITE_COMPLETE;
351 
352 	return 0;
353 }
354 
krping_cq_event_handler(struct ib_cq * cq,void * ctx)355 static void krping_cq_event_handler(struct ib_cq *cq, void *ctx)
356 {
357 	struct krping_cb *cb = ctx;
358 	struct ib_wc wc;
359 	const struct ib_recv_wr *bad_wr;
360 	int ret;
361 
362 	BUG_ON(cb->cq != cq);
363 	if (cb->frtest) {
364 		printk(KERN_ERR PFX "cq completion event in frtest!\n");
365 		return;
366 	}
367 	if (!cb->wlat && !cb->rlat && !cb->bw)
368 		ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
369 	while ((ret = ib_poll_cq(cb->cq, 1, &wc)) == 1) {
370 		if (wc.status) {
371 			if (wc.status == IB_WC_WR_FLUSH_ERR) {
372 				DEBUG_LOG("cq flushed\n");
373 				continue;
374 			} else {
375 				printk(KERN_ERR PFX "cq completion failed with "
376 				       "wr_id %jx status %d opcode %d vender_err %x\n",
377 					(uintmax_t)wc.wr_id, wc.status, wc.opcode, wc.vendor_err);
378 				goto error;
379 			}
380 		}
381 		if (cb->state == ERROR) {
382 			printk(KERN_ERR PFX "cq completion in ERROR state\n");
383 			return;
384 		}
385 		switch (wc.opcode) {
386 		case IB_WC_SEND:
387 			DEBUG_LOG("send completion\n");
388 			cb->stats.send_bytes += cb->send_sgl.length;
389 			cb->stats.send_msgs++;
390 			break;
391 
392 		case IB_WC_RDMA_WRITE:
393 			DEBUG_LOG("rdma write completion\n");
394 			cb->stats.write_bytes += cb->rdma_sq_wr.wr.sg_list->length;
395 			cb->stats.write_msgs++;
396 			cb->state = RDMA_WRITE_COMPLETE;
397 			wake_up_interruptible(&cb->sem);
398 			break;
399 
400 		case IB_WC_RDMA_READ:
401 			DEBUG_LOG("rdma read completion\n");
402 			cb->stats.read_bytes += cb->rdma_sq_wr.wr.sg_list->length;
403 			cb->stats.read_msgs++;
404 			cb->state = RDMA_READ_COMPLETE;
405 			wake_up_interruptible(&cb->sem);
406 			break;
407 
408 		case IB_WC_RECV:
409 			DEBUG_LOG("recv completion\n");
410 			cb->stats.recv_bytes += sizeof(cb->recv_buf);
411 			cb->stats.recv_msgs++;
412 			if (cb->wlat || cb->rlat || cb->bw)
413 				ret = server_recv(cb, &wc);
414 			else
415 				ret = cb->server ? server_recv(cb, &wc) :
416 						   client_recv(cb, &wc);
417 			if (ret) {
418 				printk(KERN_ERR PFX "recv wc error: %d\n", ret);
419 				goto error;
420 			}
421 
422 			ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
423 			if (ret) {
424 				printk(KERN_ERR PFX "post recv error: %d\n",
425 				       ret);
426 				goto error;
427 			}
428 			wake_up_interruptible(&cb->sem);
429 			break;
430 
431 		default:
432 			printk(KERN_ERR PFX
433 			       "%s:%d Unexpected opcode %d, Shutting down\n",
434 			       __func__, __LINE__, wc.opcode);
435 			goto error;
436 		}
437 	}
438 	if (ret) {
439 		printk(KERN_ERR PFX "poll error %d\n", ret);
440 		goto error;
441 	}
442 	return;
443 error:
444 	cb->state = ERROR;
445 	wake_up_interruptible(&cb->sem);
446 }
447 
krping_accept(struct krping_cb * cb)448 static int krping_accept(struct krping_cb *cb)
449 {
450 	struct rdma_conn_param conn_param;
451 	int ret;
452 
453 	DEBUG_LOG("accepting client connection request\n");
454 
455 	memset(&conn_param, 0, sizeof conn_param);
456 	conn_param.responder_resources = 1;
457 	conn_param.initiator_depth = 1;
458 
459 	ret = rdma_accept(cb->child_cm_id, &conn_param);
460 	if (ret) {
461 		printk(KERN_ERR PFX "rdma_accept error: %d\n", ret);
462 		return ret;
463 	}
464 
465 	if (!cb->wlat && !cb->rlat && !cb->bw) {
466 		wait_event_interruptible(cb->sem, cb->state >= CONNECTED);
467 		if (cb->state == ERROR) {
468 			printk(KERN_ERR PFX "wait for CONNECTED state %d\n",
469 				cb->state);
470 			return -1;
471 		}
472 	}
473 	return 0;
474 }
475 
krping_setup_wr(struct krping_cb * cb)476 static void krping_setup_wr(struct krping_cb *cb)
477 {
478 	cb->recv_sgl.addr = cb->recv_dma_addr;
479 	cb->recv_sgl.length = sizeof cb->recv_buf;
480 	cb->recv_sgl.lkey = cb->pd->local_dma_lkey;
481 	cb->rq_wr.sg_list = &cb->recv_sgl;
482 	cb->rq_wr.num_sge = 1;
483 
484 	cb->send_sgl.addr = cb->send_dma_addr;
485 	cb->send_sgl.length = sizeof cb->send_buf;
486 	cb->send_sgl.lkey = cb->pd->local_dma_lkey;
487 
488 	cb->sq_wr.opcode = IB_WR_SEND;
489 	cb->sq_wr.send_flags = IB_SEND_SIGNALED;
490 	cb->sq_wr.sg_list = &cb->send_sgl;
491 	cb->sq_wr.num_sge = 1;
492 
493 	if (cb->server || cb->wlat || cb->rlat || cb->bw) {
494 		cb->rdma_sgl.addr = cb->rdma_dma_addr;
495 		cb->rdma_sq_wr.wr.send_flags = IB_SEND_SIGNALED;
496 		cb->rdma_sq_wr.wr.sg_list = &cb->rdma_sgl;
497 		cb->rdma_sq_wr.wr.num_sge = 1;
498 	}
499 
500 	/*
501 	 * A chain of 2 WRs, INVALDATE_MR + REG_MR.
502 	 * both unsignaled.  The client uses them to reregister
503 	 * the rdma buffers with a new key each iteration.
504 	 */
505 	cb->reg_mr_wr.wr.opcode = IB_WR_REG_MR;
506 	cb->reg_mr_wr.mr = cb->reg_mr;
507 
508 	cb->invalidate_wr.next = &cb->reg_mr_wr.wr;
509 	cb->invalidate_wr.opcode = IB_WR_LOCAL_INV;
510 }
511 
krping_setup_buffers(struct krping_cb * cb)512 static int krping_setup_buffers(struct krping_cb *cb)
513 {
514 	int ret;
515 
516 	DEBUG_LOG(PFX "krping_setup_buffers called on cb %p\n", cb);
517 
518 	cb->recv_dma_addr = ib_dma_map_single(cb->pd->device,
519 				   &cb->recv_buf,
520 				   sizeof(cb->recv_buf), DMA_BIDIRECTIONAL);
521 	pci_unmap_addr_set(cb, recv_mapping, cb->recv_dma_addr);
522 	cb->send_dma_addr = ib_dma_map_single(cb->pd->device,
523 					   &cb->send_buf, sizeof(cb->send_buf),
524 					   DMA_BIDIRECTIONAL);
525 	pci_unmap_addr_set(cb, send_mapping, cb->send_dma_addr);
526 
527 	cb->rdma_buf = ib_dma_alloc_coherent(cb->pd->device, cb->size,
528 					     &cb->rdma_dma_addr,
529 					     GFP_KERNEL);
530 	if (!cb->rdma_buf) {
531 		DEBUG_LOG(PFX "rdma_buf allocation failed\n");
532 		ret = -ENOMEM;
533 		goto bail;
534 	}
535 	pci_unmap_addr_set(cb, rdma_mapping, cb->rdma_dma_addr);
536 	cb->page_list_len = (((cb->size - 1) & PAGE_MASK) + PAGE_SIZE)
537 				>> PAGE_SHIFT;
538 	cb->reg_mr = ib_alloc_mr(cb->pd,  IB_MR_TYPE_MEM_REG,
539 				 cb->page_list_len);
540 	if (IS_ERR(cb->reg_mr)) {
541 		ret = PTR_ERR(cb->reg_mr);
542 		DEBUG_LOG(PFX "recv_buf reg_mr failed %d\n", ret);
543 		goto bail;
544 	}
545 	DEBUG_LOG(PFX "reg rkey 0x%x page_list_len %u\n",
546 		cb->reg_mr->rkey, cb->page_list_len);
547 
548 	if (!cb->server || cb->wlat || cb->rlat || cb->bw) {
549 
550 		cb->start_buf = ib_dma_alloc_coherent(cb->pd->device, cb->size,
551 						      &cb->start_dma_addr,
552 						      GFP_KERNEL);
553 		if (!cb->start_buf) {
554 			DEBUG_LOG(PFX "start_buf malloc failed\n");
555 			ret = -ENOMEM;
556 			goto bail;
557 		}
558 		pci_unmap_addr_set(cb, start_mapping, cb->start_dma_addr);
559 	}
560 
561 	krping_setup_wr(cb);
562 	DEBUG_LOG(PFX "allocated & registered buffers...\n");
563 	return 0;
564 bail:
565 	if (cb->reg_mr && !IS_ERR(cb->reg_mr))
566 		ib_dereg_mr(cb->reg_mr);
567 	if (cb->rdma_mr && !IS_ERR(cb->rdma_mr))
568 		ib_dereg_mr(cb->rdma_mr);
569 	if (cb->dma_mr && !IS_ERR(cb->dma_mr))
570 		ib_dereg_mr(cb->dma_mr);
571 	if (cb->rdma_buf) {
572 		ib_dma_free_coherent(cb->pd->device, cb->size, cb->rdma_buf,
573 				     cb->rdma_dma_addr);
574 	}
575 	if (cb->start_buf) {
576 		ib_dma_free_coherent(cb->pd->device, cb->size, cb->start_buf,
577 				     cb->start_dma_addr);
578 	}
579 	return ret;
580 }
581 
krping_free_buffers(struct krping_cb * cb)582 static void krping_free_buffers(struct krping_cb *cb)
583 {
584 	DEBUG_LOG("krping_free_buffers called on cb %p\n", cb);
585 
586 	if (cb->dma_mr)
587 		ib_dereg_mr(cb->dma_mr);
588 	if (cb->rdma_mr)
589 		ib_dereg_mr(cb->rdma_mr);
590 	if (cb->start_mr)
591 		ib_dereg_mr(cb->start_mr);
592 	if (cb->reg_mr)
593 		ib_dereg_mr(cb->reg_mr);
594 
595 	dma_unmap_single(cb->pd->device->dma_device,
596 			 pci_unmap_addr(cb, recv_mapping),
597 			 sizeof(cb->recv_buf), DMA_BIDIRECTIONAL);
598 	dma_unmap_single(cb->pd->device->dma_device,
599 			 pci_unmap_addr(cb, send_mapping),
600 			 sizeof(cb->send_buf), DMA_BIDIRECTIONAL);
601 
602 	ib_dma_free_coherent(cb->pd->device, cb->size, cb->rdma_buf,
603 			     cb->rdma_dma_addr);
604 
605 	if (cb->start_buf) {
606 		ib_dma_free_coherent(cb->pd->device, cb->size, cb->start_buf,
607 				     cb->start_dma_addr);
608 	}
609 }
610 
krping_create_qp(struct krping_cb * cb)611 static int krping_create_qp(struct krping_cb *cb)
612 {
613 	struct ib_qp_init_attr init_attr;
614 	int ret;
615 
616 	memset(&init_attr, 0, sizeof(init_attr));
617 	init_attr.cap.max_send_wr = cb->txdepth;
618 	init_attr.cap.max_recv_wr = 2;
619 
620 	/* For flush_qp() */
621 	init_attr.cap.max_send_wr++;
622 	init_attr.cap.max_recv_wr++;
623 
624 	init_attr.cap.max_recv_sge = 1;
625 	init_attr.cap.max_send_sge = 1;
626 	init_attr.qp_type = IB_QPT_RC;
627 	init_attr.send_cq = cb->cq;
628 	init_attr.recv_cq = cb->cq;
629 	init_attr.sq_sig_type = IB_SIGNAL_REQ_WR;
630 
631 	if (cb->server) {
632 		ret = rdma_create_qp(cb->child_cm_id, cb->pd, &init_attr);
633 		if (!ret)
634 			cb->qp = cb->child_cm_id->qp;
635 	} else {
636 		ret = rdma_create_qp(cb->cm_id, cb->pd, &init_attr);
637 		if (!ret)
638 			cb->qp = cb->cm_id->qp;
639 	}
640 
641 	return ret;
642 }
643 
krping_free_qp(struct krping_cb * cb)644 static void krping_free_qp(struct krping_cb *cb)
645 {
646 	ib_destroy_qp(cb->qp);
647 	ib_destroy_cq(cb->cq);
648 	ib_dealloc_pd(cb->pd);
649 }
650 
krping_setup_qp(struct krping_cb * cb,struct rdma_cm_id * cm_id)651 static int krping_setup_qp(struct krping_cb *cb, struct rdma_cm_id *cm_id)
652 {
653 	int ret;
654 	struct ib_cq_init_attr attr = {0};
655 
656 	cb->pd = ib_alloc_pd(cm_id->device, 0);
657 	if (IS_ERR(cb->pd)) {
658 		printk(KERN_ERR PFX "ib_alloc_pd failed\n");
659 		return PTR_ERR(cb->pd);
660 	}
661 	DEBUG_LOG("created pd %p\n", cb->pd);
662 
663 	strlcpy(cb->stats.name, cb->pd->device->name, sizeof(cb->stats.name));
664 
665 	attr.cqe = cb->txdepth * 2;
666 	attr.comp_vector = 0;
667 	cb->cq = ib_create_cq(cm_id->device, krping_cq_event_handler, NULL,
668 			      cb, &attr);
669 	if (IS_ERR(cb->cq)) {
670 		printk(KERN_ERR PFX "ib_create_cq failed\n");
671 		ret = PTR_ERR(cb->cq);
672 		goto err1;
673 	}
674 	DEBUG_LOG("created cq %p\n", cb->cq);
675 
676 	if (!cb->wlat && !cb->rlat && !cb->bw && !cb->frtest) {
677 		ret = ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
678 		if (ret) {
679 			printk(KERN_ERR PFX "ib_create_cq failed\n");
680 			goto err2;
681 		}
682 	}
683 
684 	ret = krping_create_qp(cb);
685 	if (ret) {
686 		printk(KERN_ERR PFX "krping_create_qp failed: %d\n", ret);
687 		goto err2;
688 	}
689 	DEBUG_LOG("created qp %p\n", cb->qp);
690 	return 0;
691 err2:
692 	ib_destroy_cq(cb->cq);
693 err1:
694 	ib_dealloc_pd(cb->pd);
695 	return ret;
696 }
697 
698 /*
699  * return the (possibly rebound) rkey for the rdma buffer.
700  * REG mode: invalidate and rebind via reg wr.
701  * other modes: just return the mr rkey.
702  */
krping_rdma_rkey(struct krping_cb * cb,u64 buf,int post_inv)703 static u32 krping_rdma_rkey(struct krping_cb *cb, u64 buf, int post_inv)
704 {
705 	u32 rkey;
706 	const struct ib_send_wr *bad_wr;
707 	int ret;
708 	struct scatterlist sg = {0};
709 
710 	cb->invalidate_wr.ex.invalidate_rkey = cb->reg_mr->rkey;
711 
712 	/*
713 	 * Update the reg key.
714 	 */
715 	ib_update_fast_reg_key(cb->reg_mr, ++cb->key);
716 	cb->reg_mr_wr.key = cb->reg_mr->rkey;
717 
718 	/*
719 	 * Update the reg WR with new buf info.
720 	 */
721 	if (buf == (u64)cb->start_dma_addr)
722 		cb->reg_mr_wr.access = IB_ACCESS_REMOTE_READ;
723 	else
724 		cb->reg_mr_wr.access = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
725 	sg_dma_address(&sg) = buf;
726 	sg_dma_len(&sg) = cb->size;
727 
728 	ret = ib_map_mr_sg(cb->reg_mr, &sg, 1, NULL, PAGE_SIZE);
729 	BUG_ON(ret <= 0 || ret > cb->page_list_len);
730 
731 	DEBUG_LOG(PFX "post_inv = %d, reg_mr new rkey 0x%x pgsz %u len %u"
732 		" iova_start %llx\n",
733 		post_inv,
734 		cb->reg_mr_wr.key,
735 		cb->reg_mr->page_size,
736 		(unsigned)cb->reg_mr->length,
737 	        (unsigned long long)cb->reg_mr->iova);
738 
739 	if (post_inv)
740 		ret = ib_post_send(cb->qp, &cb->invalidate_wr, &bad_wr);
741 	else
742 		ret = ib_post_send(cb->qp, &cb->reg_mr_wr.wr, &bad_wr);
743 	if (ret) {
744 		printk(KERN_ERR PFX "post send error %d\n", ret);
745 		cb->state = ERROR;
746 	}
747 	rkey = cb->reg_mr->rkey;
748 	return rkey;
749 }
750 
krping_format_send(struct krping_cb * cb,u64 buf)751 static void krping_format_send(struct krping_cb *cb, u64 buf)
752 {
753 	struct krping_rdma_info *info = &cb->send_buf;
754 	u32 rkey;
755 
756 	/*
757 	 * Client side will do reg or mw bind before
758 	 * advertising the rdma buffer.  Server side
759 	 * sends have no data.
760 	 */
761 	if (!cb->server || cb->wlat || cb->rlat || cb->bw) {
762 		rkey = krping_rdma_rkey(cb, buf, !cb->server_invalidate);
763 		info->buf = htonll(buf);
764 		info->rkey = htonl(rkey);
765 		info->size = htonl(cb->size);
766 		DEBUG_LOG("RDMA addr %llx rkey %x len %d\n",
767 			  (unsigned long long)buf, rkey, cb->size);
768 	}
769 }
770 
krping_test_server(struct krping_cb * cb)771 static void krping_test_server(struct krping_cb *cb)
772 {
773 	const struct ib_send_wr *bad_wr;
774 	struct ib_send_wr inv;
775 	int ret;
776 
777 	while (1) {
778 		/* Wait for client's Start STAG/TO/Len */
779 		wait_event_interruptible(cb->sem, cb->state >= RDMA_READ_ADV);
780 		if (cb->state != RDMA_READ_ADV) {
781 			printk(KERN_ERR PFX "wait for RDMA_READ_ADV state %d\n",
782 				cb->state);
783 			break;
784 		}
785 
786 		DEBUG_LOG("server received sink adv\n");
787 
788 		cb->rdma_sq_wr.rkey = cb->remote_rkey;
789 		cb->rdma_sq_wr.remote_addr = cb->remote_addr;
790 		cb->rdma_sq_wr.wr.sg_list->length = cb->remote_len;
791 		cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, !cb->read_inv);
792 		cb->rdma_sq_wr.wr.next = NULL;
793 
794 		/* Issue RDMA Read. */
795 		if (cb->read_inv)
796 			cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_READ_WITH_INV;
797 		else {
798 
799 			cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_READ;
800 			/*
801 			 * Immediately follow the read with a
802 			 * fenced LOCAL_INV.
803 			 */
804 			cb->rdma_sq_wr.wr.next = &inv;
805 			memset(&inv, 0, sizeof inv);
806 			inv.opcode = IB_WR_LOCAL_INV;
807 			inv.ex.invalidate_rkey = cb->reg_mr->rkey;
808 			inv.send_flags = IB_SEND_FENCE;
809 		}
810 
811 		ret = ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr);
812 		if (ret) {
813 			printk(KERN_ERR PFX "post send error %d\n", ret);
814 			break;
815 		}
816 		cb->rdma_sq_wr.wr.next = NULL;
817 
818 		DEBUG_LOG("server posted rdma read req \n");
819 
820 		/* Wait for read completion */
821 		wait_event_interruptible(cb->sem,
822 					 cb->state >= RDMA_READ_COMPLETE);
823 		if (cb->state != RDMA_READ_COMPLETE) {
824 			printk(KERN_ERR PFX
825 			       "wait for RDMA_READ_COMPLETE state %d\n",
826 			       cb->state);
827 			break;
828 		}
829 		DEBUG_LOG("server received read complete\n");
830 
831 		/* Display data in recv buf */
832 		if (cb->verbose)
833 			printk(KERN_INFO PFX "server ping data: %s\n",
834 				cb->rdma_buf);
835 
836 		/* Tell client to continue */
837 		if (cb->server && cb->server_invalidate) {
838 			cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey;
839 			cb->sq_wr.opcode = IB_WR_SEND_WITH_INV;
840 			DEBUG_LOG("send-w-inv rkey 0x%x\n", cb->remote_rkey);
841 		}
842 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
843 		if (ret) {
844 			printk(KERN_ERR PFX "post send error %d\n", ret);
845 			break;
846 		}
847 		DEBUG_LOG("server posted go ahead\n");
848 
849 		/* Wait for client's RDMA STAG/TO/Len */
850 		wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV);
851 		if (cb->state != RDMA_WRITE_ADV) {
852 			printk(KERN_ERR PFX
853 			       "wait for RDMA_WRITE_ADV state %d\n",
854 			       cb->state);
855 			break;
856 		}
857 		DEBUG_LOG("server received sink adv\n");
858 
859 		/* RDMA Write echo data */
860 		cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
861 		cb->rdma_sq_wr.rkey = cb->remote_rkey;
862 		cb->rdma_sq_wr.remote_addr = cb->remote_addr;
863 		cb->rdma_sq_wr.wr.sg_list->length = strlen(cb->rdma_buf) + 1;
864 		if (cb->local_dma_lkey)
865 			cb->rdma_sgl.lkey = cb->pd->local_dma_lkey;
866 		else
867 			cb->rdma_sgl.lkey = krping_rdma_rkey(cb, cb->rdma_dma_addr, 0);
868 
869 		DEBUG_LOG("rdma write from lkey %x laddr %llx len %d\n",
870 			  cb->rdma_sq_wr.wr.sg_list->lkey,
871 			  (unsigned long long)cb->rdma_sq_wr.wr.sg_list->addr,
872 			  cb->rdma_sq_wr.wr.sg_list->length);
873 
874 		ret = ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr);
875 		if (ret) {
876 			printk(KERN_ERR PFX "post send error %d\n", ret);
877 			break;
878 		}
879 
880 		/* Wait for completion */
881 		ret = wait_event_interruptible(cb->sem, cb->state >=
882 							 RDMA_WRITE_COMPLETE);
883 		if (cb->state != RDMA_WRITE_COMPLETE) {
884 			printk(KERN_ERR PFX
885 			       "wait for RDMA_WRITE_COMPLETE state %d\n",
886 			       cb->state);
887 			break;
888 		}
889 		DEBUG_LOG("server rdma write complete \n");
890 
891 		cb->state = CONNECTED;
892 
893 		/* Tell client to begin again */
894 		if (cb->server && cb->server_invalidate) {
895 			cb->sq_wr.ex.invalidate_rkey = cb->remote_rkey;
896 			cb->sq_wr.opcode = IB_WR_SEND_WITH_INV;
897 			DEBUG_LOG("send-w-inv rkey 0x%x\n", cb->remote_rkey);
898 		}
899 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
900 		if (ret) {
901 			printk(KERN_ERR PFX "post send error %d\n", ret);
902 			break;
903 		}
904 		DEBUG_LOG("server posted go ahead\n");
905 	}
906 }
907 
rlat_test(struct krping_cb * cb)908 static void rlat_test(struct krping_cb *cb)
909 {
910 	int scnt;
911 	int iters = cb->count;
912 	struct timeval start_tv, stop_tv;
913 	int ret;
914 	struct ib_wc wc;
915 	const struct ib_send_wr *bad_wr;
916 	int ne;
917 
918 	scnt = 0;
919 	cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_READ;
920 	cb->rdma_sq_wr.rkey = cb->remote_rkey;
921 	cb->rdma_sq_wr.remote_addr = cb->remote_addr;
922 	cb->rdma_sq_wr.wr.sg_list->length = cb->size;
923 
924 	microtime(&start_tv);
925 	if (!cb->poll) {
926 		cb->state = RDMA_READ_ADV;
927 		ib_req_notify_cq(cb->cq, IB_CQ_NEXT_COMP);
928 	}
929 	while (scnt < iters) {
930 
931 		cb->state = RDMA_READ_ADV;
932 		ret = ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr);
933 		if (ret) {
934 			printk(KERN_ERR PFX
935 				"Couldn't post send: ret=%d scnt %d\n",
936 				ret, scnt);
937 			return;
938 		}
939 
940 		do {
941 			if (!cb->poll) {
942 				wait_event_interruptible(cb->sem,
943 					cb->state != RDMA_READ_ADV);
944 				if (cb->state == RDMA_READ_COMPLETE) {
945 					ne = 1;
946 					ib_req_notify_cq(cb->cq,
947 						IB_CQ_NEXT_COMP);
948 				} else {
949 					ne = -1;
950 				}
951 			} else
952 				ne = ib_poll_cq(cb->cq, 1, &wc);
953 			if (cb->state == ERROR) {
954 				printk(KERN_ERR PFX
955 					"state == ERROR...bailing scnt %d\n",
956 					scnt);
957 				return;
958 			}
959 		} while (ne == 0);
960 
961 		if (ne < 0) {
962 			printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
963 			return;
964 		}
965 		if (cb->poll && wc.status != IB_WC_SUCCESS) {
966 			printk(KERN_ERR PFX "Completion wth error at %s:\n",
967 				cb->server ? "server" : "client");
968 			printk(KERN_ERR PFX "Failed status %d: wr_id %d\n",
969 				wc.status, (int) wc.wr_id);
970 			return;
971 		}
972 		++scnt;
973 	}
974 	microtime(&stop_tv);
975 
976         if (stop_tv.tv_usec < start_tv.tv_usec) {
977                 stop_tv.tv_usec += 1000000;
978                 stop_tv.tv_sec  -= 1;
979         }
980 
981 	printk(KERN_ERR PFX "delta sec %lu delta usec %lu iter %d size %d\n",
982 		(unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
983 		(unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
984 		scnt, cb->size);
985 }
986 
wlat_test(struct krping_cb * cb)987 static void wlat_test(struct krping_cb *cb)
988 {
989 	int ccnt, scnt, rcnt;
990 	int iters=cb->count;
991 	volatile char *poll_buf = (char *) cb->start_buf;
992 	char *buf = (char *)cb->rdma_buf;
993 	struct timeval start_tv, stop_tv;
994 	cycles_t *post_cycles_start = NULL;
995 	cycles_t *post_cycles_stop = NULL;
996 	cycles_t *poll_cycles_start = NULL;
997 	cycles_t *poll_cycles_stop = NULL;
998 	cycles_t *last_poll_cycles_start = NULL;
999 	cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0;
1000 	int i;
1001 	int cycle_iters = 1000;
1002 
1003 	ccnt = 0;
1004 	scnt = 0;
1005 	rcnt = 0;
1006 
1007 	post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1008 	if (!post_cycles_start) {
1009 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1010 		goto done;
1011 	}
1012 	post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1013 	if (!post_cycles_stop) {
1014 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1015 		goto done;
1016 	}
1017 	poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1018 	if (!poll_cycles_start) {
1019 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1020 		goto done;
1021 	}
1022 	poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1023 	if (!poll_cycles_stop) {
1024 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1025 		goto done;
1026 	}
1027 	last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t),
1028 		GFP_KERNEL);
1029 	if (!last_poll_cycles_start) {
1030 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1031 		goto done;
1032 	}
1033 	cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
1034 	cb->rdma_sq_wr.rkey = cb->remote_rkey;
1035 	cb->rdma_sq_wr.remote_addr = cb->remote_addr;
1036 	cb->rdma_sq_wr.wr.sg_list->length = cb->size;
1037 
1038 	if (cycle_iters > iters)
1039 		cycle_iters = iters;
1040 	microtime(&start_tv);
1041 	while (scnt < iters || ccnt < iters || rcnt < iters) {
1042 
1043 		/* Wait till buffer changes. */
1044 		if (rcnt < iters && !(scnt < 1 && !cb->server)) {
1045 			++rcnt;
1046 			while (*poll_buf != (char)rcnt) {
1047 				if (cb->state == ERROR) {
1048 					printk(KERN_ERR PFX
1049 						"state = ERROR, bailing\n");
1050 					goto done;
1051 				}
1052 			}
1053 		}
1054 
1055 		if (scnt < iters) {
1056 			const struct ib_send_wr *bad_wr;
1057 
1058 			*buf = (char)scnt+1;
1059 			if (scnt < cycle_iters)
1060 				post_cycles_start[scnt] = get_cycles();
1061 			if (ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr)) {
1062 				printk(KERN_ERR PFX
1063 					"Couldn't post send: scnt=%d\n",
1064 					scnt);
1065 				goto done;
1066 			}
1067 			if (scnt < cycle_iters)
1068 				post_cycles_stop[scnt] = get_cycles();
1069 			scnt++;
1070 		}
1071 
1072 		if (ccnt < iters) {
1073 			struct ib_wc wc;
1074 			int ne;
1075 
1076 			if (ccnt < cycle_iters)
1077 				poll_cycles_start[ccnt] = get_cycles();
1078 			do {
1079 				if (ccnt < cycle_iters)
1080 					last_poll_cycles_start[ccnt] =
1081 						get_cycles();
1082 				ne = ib_poll_cq(cb->cq, 1, &wc);
1083 			} while (ne == 0);
1084 			if (ccnt < cycle_iters)
1085 				poll_cycles_stop[ccnt] = get_cycles();
1086 			++ccnt;
1087 
1088 			if (ne < 0) {
1089 				printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
1090 				goto done;
1091 			}
1092 			if (wc.status != IB_WC_SUCCESS) {
1093 				printk(KERN_ERR PFX
1094 					"Completion wth error at %s:\n",
1095 					cb->server ? "server" : "client");
1096 				printk(KERN_ERR PFX
1097 					"Failed status %d: wr_id %d\n",
1098 					wc.status, (int) wc.wr_id);
1099 				printk(KERN_ERR PFX
1100 					"scnt=%d, rcnt=%d, ccnt=%d\n",
1101 					scnt, rcnt, ccnt);
1102 				goto done;
1103 			}
1104 		}
1105 	}
1106 	microtime(&stop_tv);
1107 
1108         if (stop_tv.tv_usec < start_tv.tv_usec) {
1109                 stop_tv.tv_usec += 1000000;
1110                 stop_tv.tv_sec  -= 1;
1111         }
1112 
1113 	for (i=0; i < cycle_iters; i++) {
1114 		sum_post += post_cycles_stop[i] - post_cycles_start[i];
1115 		sum_poll += poll_cycles_stop[i] - poll_cycles_start[i];
1116 		sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i];
1117 	}
1118 	printk(KERN_ERR PFX
1119 		"delta sec %lu delta usec %lu iter %d size %d cycle_iters %d"
1120 		" sum_post %llu sum_poll %llu sum_last_poll %llu\n",
1121 		(unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1122 		(unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1123 		scnt, cb->size, cycle_iters,
1124 		(unsigned long long)sum_post, (unsigned long long)sum_poll,
1125 		(unsigned long long)sum_last_poll);
1126 done:
1127 	kfree(post_cycles_start);
1128 	kfree(post_cycles_stop);
1129 	kfree(poll_cycles_start);
1130 	kfree(poll_cycles_stop);
1131 	kfree(last_poll_cycles_start);
1132 }
1133 
bw_test(struct krping_cb * cb)1134 static void bw_test(struct krping_cb *cb)
1135 {
1136 	int ccnt, scnt;
1137 	int iters=cb->count;
1138 	struct timeval start_tv, stop_tv;
1139 	cycles_t *post_cycles_start = NULL;
1140 	cycles_t *post_cycles_stop = NULL;
1141 	cycles_t *poll_cycles_start = NULL;
1142 	cycles_t *poll_cycles_stop = NULL;
1143 	cycles_t *last_poll_cycles_start = NULL;
1144 	cycles_t sum_poll = 0, sum_post = 0, sum_last_poll = 0;
1145 	int i;
1146 	int cycle_iters = 1000;
1147 
1148 	ccnt = 0;
1149 	scnt = 0;
1150 
1151 	post_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1152 	if (!post_cycles_start) {
1153 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1154 		goto done;
1155 	}
1156 	post_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1157 	if (!post_cycles_stop) {
1158 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1159 		goto done;
1160 	}
1161 	poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1162 	if (!poll_cycles_start) {
1163 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1164 		goto done;
1165 	}
1166 	poll_cycles_stop = kmalloc(cycle_iters * sizeof(cycles_t), GFP_KERNEL);
1167 	if (!poll_cycles_stop) {
1168 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1169 		goto done;
1170 	}
1171 	last_poll_cycles_start = kmalloc(cycle_iters * sizeof(cycles_t),
1172 		GFP_KERNEL);
1173 	if (!last_poll_cycles_start) {
1174 		printk(KERN_ERR PFX "%s kmalloc failed\n", __FUNCTION__);
1175 		goto done;
1176 	}
1177 	cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
1178 	cb->rdma_sq_wr.rkey = cb->remote_rkey;
1179 	cb->rdma_sq_wr.remote_addr = cb->remote_addr;
1180 	cb->rdma_sq_wr.wr.sg_list->length = cb->size;
1181 
1182 	if (cycle_iters > iters)
1183 		cycle_iters = iters;
1184 	microtime(&start_tv);
1185 	while (scnt < iters || ccnt < iters) {
1186 
1187 		while (scnt < iters && scnt - ccnt < cb->txdepth) {
1188 			const struct ib_send_wr *bad_wr;
1189 
1190 			if (scnt < cycle_iters)
1191 				post_cycles_start[scnt] = get_cycles();
1192 			if (ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr)) {
1193 				printk(KERN_ERR PFX
1194 					"Couldn't post send: scnt=%d\n",
1195 					scnt);
1196 				goto done;
1197 			}
1198 			if (scnt < cycle_iters)
1199 				post_cycles_stop[scnt] = get_cycles();
1200 			++scnt;
1201 		}
1202 
1203 		if (ccnt < iters) {
1204 			int ne;
1205 			struct ib_wc wc;
1206 
1207 			if (ccnt < cycle_iters)
1208 				poll_cycles_start[ccnt] = get_cycles();
1209 			do {
1210 				if (ccnt < cycle_iters)
1211 					last_poll_cycles_start[ccnt] =
1212 						get_cycles();
1213 				ne = ib_poll_cq(cb->cq, 1, &wc);
1214 			} while (ne == 0);
1215 			if (ccnt < cycle_iters)
1216 				poll_cycles_stop[ccnt] = get_cycles();
1217 			ccnt += 1;
1218 
1219 			if (ne < 0) {
1220 				printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
1221 				goto done;
1222 			}
1223 			if (wc.status != IB_WC_SUCCESS) {
1224 				printk(KERN_ERR PFX
1225 					"Completion wth error at %s:\n",
1226 					cb->server ? "server" : "client");
1227 				printk(KERN_ERR PFX
1228 					"Failed status %d: wr_id %d\n",
1229 					wc.status, (int) wc.wr_id);
1230 				goto done;
1231 			}
1232 		}
1233 	}
1234 	microtime(&stop_tv);
1235 
1236         if (stop_tv.tv_usec < start_tv.tv_usec) {
1237                 stop_tv.tv_usec += 1000000;
1238                 stop_tv.tv_sec  -= 1;
1239         }
1240 
1241 	for (i=0; i < cycle_iters; i++) {
1242 		sum_post += post_cycles_stop[i] - post_cycles_start[i];
1243 		sum_poll += poll_cycles_stop[i] - poll_cycles_start[i];
1244 		sum_last_poll += poll_cycles_stop[i]-last_poll_cycles_start[i];
1245 	}
1246 	printk(KERN_ERR PFX
1247 		"delta sec %lu delta usec %lu iter %d size %d cycle_iters %d"
1248 		" sum_post %llu sum_poll %llu sum_last_poll %llu\n",
1249 		(unsigned long)(stop_tv.tv_sec - start_tv.tv_sec),
1250 		(unsigned long)(stop_tv.tv_usec - start_tv.tv_usec),
1251 		scnt, cb->size, cycle_iters,
1252 		(unsigned long long)sum_post, (unsigned long long)sum_poll,
1253 		(unsigned long long)sum_last_poll);
1254 done:
1255 	kfree(post_cycles_start);
1256 	kfree(post_cycles_stop);
1257 	kfree(poll_cycles_start);
1258 	kfree(poll_cycles_stop);
1259 	kfree(last_poll_cycles_start);
1260 }
1261 
krping_rlat_test_server(struct krping_cb * cb)1262 static void krping_rlat_test_server(struct krping_cb *cb)
1263 {
1264 	const struct ib_send_wr *bad_wr;
1265 	struct ib_wc wc;
1266 	int ret;
1267 
1268 	/* Spin waiting for client's Start STAG/TO/Len */
1269 	while (cb->state < RDMA_READ_ADV) {
1270 		krping_cq_event_handler(cb->cq, cb);
1271 	}
1272 
1273 	/* Send STAG/TO/Len to client */
1274 	krping_format_send(cb, cb->start_dma_addr);
1275 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1276 	if (ret) {
1277 		printk(KERN_ERR PFX "post send error %d\n", ret);
1278 		return;
1279 	}
1280 
1281 	/* Spin waiting for send completion */
1282 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1283 	if (ret < 0) {
1284 		printk(KERN_ERR PFX "poll error %d\n", ret);
1285 		return;
1286 	}
1287 	if (wc.status) {
1288 		printk(KERN_ERR PFX "send completiong error %d\n", wc.status);
1289 		return;
1290 	}
1291 
1292 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1293 }
1294 
krping_wlat_test_server(struct krping_cb * cb)1295 static void krping_wlat_test_server(struct krping_cb *cb)
1296 {
1297 	const struct ib_send_wr *bad_wr;
1298 	struct ib_wc wc;
1299 	int ret;
1300 
1301 	/* Spin waiting for client's Start STAG/TO/Len */
1302 	while (cb->state < RDMA_READ_ADV) {
1303 		krping_cq_event_handler(cb->cq, cb);
1304 	}
1305 
1306 	/* Send STAG/TO/Len to client */
1307 	krping_format_send(cb, cb->start_dma_addr);
1308 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1309 	if (ret) {
1310 		printk(KERN_ERR PFX "post send error %d\n", ret);
1311 		return;
1312 	}
1313 
1314 	/* Spin waiting for send completion */
1315 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1316 	if (ret < 0) {
1317 		printk(KERN_ERR PFX "poll error %d\n", ret);
1318 		return;
1319 	}
1320 	if (wc.status) {
1321 		printk(KERN_ERR PFX "send completiong error %d\n", wc.status);
1322 		return;
1323 	}
1324 
1325 	wlat_test(cb);
1326 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1327 }
1328 
krping_bw_test_server(struct krping_cb * cb)1329 static void krping_bw_test_server(struct krping_cb *cb)
1330 {
1331 	const struct ib_send_wr *bad_wr;
1332 	struct ib_wc wc;
1333 	int ret;
1334 
1335 	/* Spin waiting for client's Start STAG/TO/Len */
1336 	while (cb->state < RDMA_READ_ADV) {
1337 		krping_cq_event_handler(cb->cq, cb);
1338 	}
1339 
1340 	/* Send STAG/TO/Len to client */
1341 	krping_format_send(cb, cb->start_dma_addr);
1342 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1343 	if (ret) {
1344 		printk(KERN_ERR PFX "post send error %d\n", ret);
1345 		return;
1346 	}
1347 
1348 	/* Spin waiting for send completion */
1349 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1350 	if (ret < 0) {
1351 		printk(KERN_ERR PFX "poll error %d\n", ret);
1352 		return;
1353 	}
1354 	if (wc.status) {
1355 		printk(KERN_ERR PFX "send completiong error %d\n", wc.status);
1356 		return;
1357 	}
1358 
1359 	if (cb->duplex)
1360 		bw_test(cb);
1361 	wait_event_interruptible(cb->sem, cb->state == ERROR);
1362 }
1363 
reg_supported(struct ib_device * dev)1364 static int reg_supported(struct ib_device *dev)
1365 {
1366 	u64 needed_flags = IB_DEVICE_MEM_MGT_EXTENSIONS;
1367 
1368 	if ((dev->attrs.device_cap_flags & needed_flags) != needed_flags) {
1369 		printk(KERN_ERR PFX
1370 			"Fastreg not supported - device_cap_flags 0x%llx\n",
1371 			(unsigned long long)dev->attrs.device_cap_flags);
1372 		return 0;
1373 	}
1374 	DEBUG_LOG("Fastreg supported - device_cap_flags 0x%llx\n",
1375 		(unsigned long long)dev->attrs.device_cap_flags);
1376 	return 1;
1377 }
1378 
fill_sockaddr(struct sockaddr_storage * sin,struct krping_cb * cb)1379 static void fill_sockaddr(struct sockaddr_storage *sin, struct krping_cb *cb)
1380 {
1381 	memset(sin, 0, sizeof(*sin));
1382 
1383 	if (cb->addr_type == AF_INET) {
1384 		struct sockaddr_in *sin4 = (struct sockaddr_in *)sin;
1385 		sin4->sin_len = sizeof(*sin4);
1386 		sin4->sin_family = AF_INET;
1387 		memcpy((void *)&sin4->sin_addr.s_addr, cb->addr, 4);
1388 		sin4->sin_port = cb->port;
1389 	} else if (cb->addr_type == AF_INET6) {
1390 		struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)sin;
1391 		sin6->sin6_len = sizeof(*sin6);
1392 		sin6->sin6_family = AF_INET6;
1393 		memcpy((void *)&sin6->sin6_addr, cb->addr, 16);
1394 		sin6->sin6_port = cb->port;
1395 	}
1396 }
1397 
krping_bind_server(struct krping_cb * cb)1398 static int krping_bind_server(struct krping_cb *cb)
1399 {
1400 	struct sockaddr_storage sin;
1401 	int ret;
1402 
1403 
1404 	fill_sockaddr(&sin, cb);
1405 
1406 	ret = rdma_bind_addr(cb->cm_id, (struct sockaddr *)&sin);
1407 	if (ret) {
1408 		printk(KERN_ERR PFX "rdma_bind_addr error %d\n", ret);
1409 		return ret;
1410 	}
1411 	DEBUG_LOG("rdma_bind_addr successful\n");
1412 
1413 	DEBUG_LOG("rdma_listen\n");
1414 	ret = rdma_listen(cb->cm_id, 3);
1415 	if (ret) {
1416 		printk(KERN_ERR PFX "rdma_listen failed: %d\n", ret);
1417 		return ret;
1418 	}
1419 
1420 	wait_event_interruptible(cb->sem, cb->state >= CONNECT_REQUEST);
1421 	if (cb->state != CONNECT_REQUEST) {
1422 		printk(KERN_ERR PFX "wait for CONNECT_REQUEST state %d\n",
1423 			cb->state);
1424 		return -1;
1425 	}
1426 
1427 	if (!reg_supported(cb->child_cm_id->device))
1428 		return -EINVAL;
1429 
1430 	return 0;
1431 }
1432 
krping_run_server(struct krping_cb * cb)1433 static void krping_run_server(struct krping_cb *cb)
1434 {
1435 	const struct ib_recv_wr *bad_wr;
1436 	int ret;
1437 
1438 	ret = krping_bind_server(cb);
1439 	if (ret)
1440 		return;
1441 
1442 	ret = krping_setup_qp(cb, cb->child_cm_id);
1443 	if (ret) {
1444 		printk(KERN_ERR PFX "setup_qp failed: %d\n", ret);
1445 		goto err0;
1446 	}
1447 
1448 	ret = krping_setup_buffers(cb);
1449 	if (ret) {
1450 		printk(KERN_ERR PFX "krping_setup_buffers failed: %d\n", ret);
1451 		goto err1;
1452 	}
1453 
1454 	ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
1455 	if (ret) {
1456 		printk(KERN_ERR PFX "ib_post_recv failed: %d\n", ret);
1457 		goto err2;
1458 	}
1459 
1460 	ret = krping_accept(cb);
1461 	if (ret) {
1462 		printk(KERN_ERR PFX "connect error %d\n", ret);
1463 		goto err2;
1464 	}
1465 
1466 	if (cb->wlat)
1467 		krping_wlat_test_server(cb);
1468 	else if (cb->rlat)
1469 		krping_rlat_test_server(cb);
1470 	else if (cb->bw)
1471 		krping_bw_test_server(cb);
1472 	else
1473 		krping_test_server(cb);
1474 	rdma_disconnect(cb->child_cm_id);
1475 err2:
1476 	krping_free_buffers(cb);
1477 err1:
1478 	krping_free_qp(cb);
1479 err0:
1480 	rdma_destroy_id(cb->child_cm_id);
1481 }
1482 
krping_test_client(struct krping_cb * cb)1483 static void krping_test_client(struct krping_cb *cb)
1484 {
1485 	int ping, start, cc, i, ret;
1486 	const struct ib_send_wr *bad_wr;
1487 	unsigned char c;
1488 
1489 	start = 65;
1490 	for (ping = 0; !cb->count || ping < cb->count; ping++) {
1491 		cb->state = RDMA_READ_ADV;
1492 
1493 		/* Put some ascii text in the buffer. */
1494 		cc = sprintf(cb->start_buf, "rdma-ping-%d: ", ping);
1495 		for (i = cc, c = start; i < cb->size; i++) {
1496 			cb->start_buf[i] = c;
1497 			c++;
1498 			if (c > 122)
1499 				c = 65;
1500 		}
1501 		start++;
1502 		if (start > 122)
1503 			start = 65;
1504 		cb->start_buf[cb->size - 1] = 0;
1505 
1506 		krping_format_send(cb, cb->start_dma_addr);
1507 		if (cb->state == ERROR) {
1508 			printk(KERN_ERR PFX "krping_format_send failed\n");
1509 			break;
1510 		}
1511 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1512 		if (ret) {
1513 			printk(KERN_ERR PFX "post send error %d\n", ret);
1514 			break;
1515 		}
1516 
1517 		/* Wait for server to ACK */
1518 		wait_event_interruptible(cb->sem, cb->state >= RDMA_WRITE_ADV);
1519 		if (cb->state != RDMA_WRITE_ADV) {
1520 			printk(KERN_ERR PFX
1521 			       "wait for RDMA_WRITE_ADV state %d\n",
1522 			       cb->state);
1523 			break;
1524 		}
1525 
1526 		krping_format_send(cb, cb->rdma_dma_addr);
1527 		ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1528 		if (ret) {
1529 			printk(KERN_ERR PFX "post send error %d\n", ret);
1530 			break;
1531 		}
1532 
1533 		/* Wait for the server to say the RDMA Write is complete. */
1534 		wait_event_interruptible(cb->sem,
1535 					 cb->state >= RDMA_WRITE_COMPLETE);
1536 		if (cb->state != RDMA_WRITE_COMPLETE) {
1537 			printk(KERN_ERR PFX
1538 			       "wait for RDMA_WRITE_COMPLETE state %d\n",
1539 			       cb->state);
1540 			break;
1541 		}
1542 
1543 		if (cb->validate)
1544 			if (memcmp(cb->start_buf, cb->rdma_buf, cb->size)) {
1545 				printk(KERN_ERR PFX "data mismatch!\n");
1546 				break;
1547 			}
1548 
1549 		if (cb->verbose)
1550 			printk(KERN_INFO PFX "ping data: %s\n", cb->rdma_buf);
1551 #ifdef SLOW_KRPING
1552 		wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
1553 #endif
1554 	}
1555 }
1556 
krping_rlat_test_client(struct krping_cb * cb)1557 static void krping_rlat_test_client(struct krping_cb *cb)
1558 {
1559 	const struct ib_send_wr *bad_wr;
1560 	struct ib_wc wc;
1561 	int ret;
1562 
1563 	cb->state = RDMA_READ_ADV;
1564 
1565 	/* Send STAG/TO/Len to client */
1566 	krping_format_send(cb, cb->start_dma_addr);
1567 	if (cb->state == ERROR) {
1568 		printk(KERN_ERR PFX "krping_format_send failed\n");
1569 		return;
1570 	}
1571 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1572 	if (ret) {
1573 		printk(KERN_ERR PFX "post send error %d\n", ret);
1574 		return;
1575 	}
1576 
1577 	/* Spin waiting for send completion */
1578 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1579 	if (ret < 0) {
1580 		printk(KERN_ERR PFX "poll error %d\n", ret);
1581 		return;
1582 	}
1583 	if (wc.status) {
1584 		printk(KERN_ERR PFX "send completion error %d\n", wc.status);
1585 		return;
1586 	}
1587 
1588 	/* Spin waiting for server's Start STAG/TO/Len */
1589 	while (cb->state < RDMA_WRITE_ADV) {
1590 		krping_cq_event_handler(cb->cq, cb);
1591 	}
1592 
1593 #if 0
1594 {
1595 	int i;
1596 	struct timeval start, stop;
1597 	time_t sec;
1598 	suseconds_t usec;
1599 	unsigned long long elapsed;
1600 	struct ib_wc wc;
1601 	const struct ib_send_wr *bad_wr;
1602 	int ne;
1603 
1604 	cb->rdma_sq_wr.wr.opcode = IB_WR_RDMA_WRITE;
1605 	cb->rdma_sq_wr.rkey = cb->remote_rkey;
1606 	cb->rdma_sq_wr.remote_addr = cb->remote_addr;
1607 	cb->rdma_sq_wr.wr.sg_list->length = 0;
1608 	cb->rdma_sq_wr.wr.num_sge = 0;
1609 
1610 	microtime(&start);
1611 	for (i=0; i < 100000; i++) {
1612 		if (ib_post_send(cb->qp, &cb->rdma_sq_wr.wr, &bad_wr)) {
1613 			printk(KERN_ERR PFX  "Couldn't post send\n");
1614 			return;
1615 		}
1616 		do {
1617 			ne = ib_poll_cq(cb->cq, 1, &wc);
1618 		} while (ne == 0);
1619 		if (ne < 0) {
1620 			printk(KERN_ERR PFX "poll CQ failed %d\n", ne);
1621 			return;
1622 		}
1623 		if (wc.status != IB_WC_SUCCESS) {
1624 			printk(KERN_ERR PFX "Completion wth error at %s:\n",
1625 				cb->server ? "server" : "client");
1626 			printk(KERN_ERR PFX "Failed status %d: wr_id %d\n",
1627 				wc.status, (int) wc.wr_id);
1628 			return;
1629 		}
1630 	}
1631 	microtime(&stop);
1632 
1633 	if (stop.tv_usec < start.tv_usec) {
1634 		stop.tv_usec += 1000000;
1635 		stop.tv_sec  -= 1;
1636 	}
1637 	sec     = stop.tv_sec - start.tv_sec;
1638 	usec    = stop.tv_usec - start.tv_usec;
1639 	elapsed = sec * 1000000 + usec;
1640 	printk(KERN_ERR PFX "0B-write-lat iters 100000 usec %llu\n", elapsed);
1641 }
1642 #endif
1643 
1644 	rlat_test(cb);
1645 }
1646 
krping_wlat_test_client(struct krping_cb * cb)1647 static void krping_wlat_test_client(struct krping_cb *cb)
1648 {
1649 	const struct ib_send_wr *bad_wr;
1650 	struct ib_wc wc;
1651 	int ret;
1652 
1653 	cb->state = RDMA_READ_ADV;
1654 
1655 	/* Send STAG/TO/Len to client */
1656 	krping_format_send(cb, cb->start_dma_addr);
1657 	if (cb->state == ERROR) {
1658 		printk(KERN_ERR PFX "krping_format_send failed\n");
1659 		return;
1660 	}
1661 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1662 	if (ret) {
1663 		printk(KERN_ERR PFX "post send error %d\n", ret);
1664 		return;
1665 	}
1666 
1667 	/* Spin waiting for send completion */
1668 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1669 	if (ret < 0) {
1670 		printk(KERN_ERR PFX "poll error %d\n", ret);
1671 		return;
1672 	}
1673 	if (wc.status) {
1674 		printk(KERN_ERR PFX "send completion error %d\n", wc.status);
1675 		return;
1676 	}
1677 
1678 	/* Spin waiting for server's Start STAG/TO/Len */
1679 	while (cb->state < RDMA_WRITE_ADV) {
1680 		krping_cq_event_handler(cb->cq, cb);
1681 	}
1682 
1683 	wlat_test(cb);
1684 }
1685 
krping_bw_test_client(struct krping_cb * cb)1686 static void krping_bw_test_client(struct krping_cb *cb)
1687 {
1688 	const struct ib_send_wr *bad_wr;
1689 	struct ib_wc wc;
1690 	int ret;
1691 
1692 	cb->state = RDMA_READ_ADV;
1693 
1694 	/* Send STAG/TO/Len to client */
1695 	krping_format_send(cb, cb->start_dma_addr);
1696 	if (cb->state == ERROR) {
1697 		printk(KERN_ERR PFX "krping_format_send failed\n");
1698 		return;
1699 	}
1700 	ret = ib_post_send(cb->qp, &cb->sq_wr, &bad_wr);
1701 	if (ret) {
1702 		printk(KERN_ERR PFX "post send error %d\n", ret);
1703 		return;
1704 	}
1705 
1706 	/* Spin waiting for send completion */
1707 	while ((ret = ib_poll_cq(cb->cq, 1, &wc) == 0));
1708 	if (ret < 0) {
1709 		printk(KERN_ERR PFX "poll error %d\n", ret);
1710 		return;
1711 	}
1712 	if (wc.status) {
1713 		printk(KERN_ERR PFX "send completion error %d\n", wc.status);
1714 		return;
1715 	}
1716 
1717 	/* Spin waiting for server's Start STAG/TO/Len */
1718 	while (cb->state < RDMA_WRITE_ADV) {
1719 		krping_cq_event_handler(cb->cq, cb);
1720 	}
1721 
1722 	bw_test(cb);
1723 }
1724 
1725 /*
1726  * Manual qp flush test
1727  */
flush_qp(struct krping_cb * cb)1728 static void flush_qp(struct krping_cb *cb)
1729 {
1730 	struct ib_send_wr wr = { 0 };
1731 	const struct ib_send_wr *bad;
1732 	struct ib_recv_wr recv_wr = { 0 };
1733 	const struct ib_recv_wr *recv_bad;
1734 	struct ib_wc wc;
1735 	int ret;
1736 	int flushed = 0;
1737 	int ccnt = 0;
1738 
1739 	rdma_disconnect(cb->cm_id);
1740 	DEBUG_LOG("disconnected!\n");
1741 
1742 	wr.opcode = IB_WR_SEND;
1743 	wr.wr_id = 0xdeadbeefcafebabe;
1744 	ret = ib_post_send(cb->qp, &wr, &bad);
1745 	if (ret) {
1746 		printk(KERN_ERR PFX "%s post_send failed ret %d\n", __func__, ret);
1747 		return;
1748 	}
1749 
1750 	recv_wr.wr_id = 0xcafebabedeadbeef;
1751 	ret = ib_post_recv(cb->qp, &recv_wr, &recv_bad);
1752 	if (ret) {
1753 		printk(KERN_ERR PFX "%s post_recv failed ret %d\n", __func__, ret);
1754 		return;
1755 	}
1756 
1757 	/* poll until the flush WRs complete */
1758 	do {
1759 		ret = ib_poll_cq(cb->cq, 1, &wc);
1760 		if (ret < 0) {
1761 			printk(KERN_ERR PFX "ib_poll_cq failed %d\n", ret);
1762 			return;
1763 		}
1764 		if (ret == 0)
1765 			continue;
1766 		ccnt++;
1767 		if (wc.wr_id == 0xdeadbeefcafebabe ||
1768 		    wc.wr_id == 0xcafebabedeadbeef)
1769 			flushed++;
1770 	} while (flushed != 2);
1771 	DEBUG_LOG("qp_flushed! ccnt %u\n", ccnt);
1772 }
1773 
krping_fr_test(struct krping_cb * cb)1774 static void krping_fr_test(struct krping_cb *cb)
1775 {
1776 	struct ib_send_wr inv;
1777 	const struct ib_send_wr *bad;
1778 	struct ib_reg_wr fr;
1779 	struct ib_wc wc;
1780 	u8 key = 0;
1781 	struct ib_mr *mr;
1782 	int ret;
1783 	int size = cb->size;
1784 	int plen = (((size - 1) & PAGE_MASK) + PAGE_SIZE) >> PAGE_SHIFT;
1785 	unsigned long start;
1786 	int count = 0;
1787 	int scnt = 0;
1788 	struct scatterlist sg = {0};
1789 
1790 	mr = ib_alloc_mr(cb->pd, IB_MR_TYPE_MEM_REG, plen);
1791 	if (IS_ERR(mr)) {
1792 		printk(KERN_ERR PFX "ib_alloc_mr failed %ld\n", PTR_ERR(mr));
1793 		return;
1794 	}
1795 
1796 	sg_dma_address(&sg) = (dma_addr_t)0xcafebabe0000ULL;
1797 	sg_dma_len(&sg) = size;
1798 	ret = ib_map_mr_sg(mr, &sg, 1, NULL, PAGE_SIZE);
1799 	if (ret <= 0) {
1800 		printk(KERN_ERR PFX "ib_map_mr_sge err %d\n", ret);
1801 		goto err2;
1802 	}
1803 
1804 	memset(&fr, 0, sizeof fr);
1805 	fr.wr.opcode = IB_WR_REG_MR;
1806 	fr.access = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE;
1807 	fr.mr = mr;
1808 	fr.wr.next = &inv;
1809 
1810 	memset(&inv, 0, sizeof inv);
1811 	inv.opcode = IB_WR_LOCAL_INV;
1812 	inv.send_flags = IB_SEND_SIGNALED;
1813 
1814 	DEBUG_LOG("fr_test: stag index 0x%x plen %u size %u depth %u\n", mr->rkey >> 8, plen, cb->size, cb->txdepth);
1815 	start = time_uptime;
1816 	while (!cb->count || count <= cb->count) {
1817 		if (SIGPENDING(curthread)) {
1818 			printk(KERN_ERR PFX "signal!\n");
1819 			break;
1820 		}
1821 		if ((time_uptime - start) >= 9) {
1822 			DEBUG_LOG("fr_test: pausing 1 second! count %u latest size %u plen %u\n", count, size, plen);
1823 			wait_event_interruptible_timeout(cb->sem, cb->state == ERROR, HZ);
1824 			if (cb->state == ERROR)
1825 				break;
1826 			start = time_uptime;
1827 		}
1828 		while (scnt < (cb->txdepth>>1)) {
1829 			ib_update_fast_reg_key(mr, ++key);
1830 			fr.key = mr->rkey;
1831 			inv.ex.invalidate_rkey = mr->rkey;
1832 
1833 			size = arc4random() % cb->size;
1834 			if (size == 0)
1835 				size = cb->size;
1836 			sg_dma_len(&sg) = size;
1837 			ret = ib_map_mr_sg(mr, &sg, 1, NULL, PAGE_SIZE);
1838 			if (ret <= 0) {
1839 				printk(KERN_ERR PFX "ib_map_mr_sge err %d\n", ret);
1840 				goto err2;
1841 			}
1842 			ret = ib_post_send(cb->qp, &fr.wr, &bad);
1843 			if (ret) {
1844 				printk(KERN_ERR PFX "ib_post_send failed %d\n", ret);
1845 				goto err2;
1846 			}
1847 			scnt++;
1848 		}
1849 
1850 		ret = ib_poll_cq(cb->cq, 1, &wc);
1851 		if (ret < 0) {
1852 			printk(KERN_ERR PFX "ib_poll_cq failed %d\n", ret);
1853 			goto err2;
1854 		}
1855 		if (ret == 1) {
1856 			if (wc.status) {
1857 				printk(KERN_ERR PFX "completion error %u\n", wc.status);
1858 				goto err2;
1859 			}
1860 			count++;
1861 			scnt--;
1862 		}
1863 	}
1864 err2:
1865 	flush_qp(cb);
1866 	DEBUG_LOG("fr_test: done!\n");
1867 	ib_dereg_mr(mr);
1868 }
1869 
krping_connect_client(struct krping_cb * cb)1870 static int krping_connect_client(struct krping_cb *cb)
1871 {
1872 	struct rdma_conn_param conn_param;
1873 	int ret;
1874 
1875 	memset(&conn_param, 0, sizeof conn_param);
1876 	conn_param.responder_resources = 1;
1877 	conn_param.initiator_depth = 1;
1878 	conn_param.retry_count = 10;
1879 
1880 	ret = rdma_connect(cb->cm_id, &conn_param);
1881 	if (ret) {
1882 		printk(KERN_ERR PFX "rdma_connect error %d\n", ret);
1883 		return ret;
1884 	}
1885 
1886 	wait_event_interruptible(cb->sem, cb->state >= CONNECTED);
1887 	if (cb->state == ERROR) {
1888 		printk(KERN_ERR PFX "wait for CONNECTED state %d\n", cb->state);
1889 		return -1;
1890 	}
1891 
1892 	DEBUG_LOG("rdma_connect successful\n");
1893 	return 0;
1894 }
1895 
krping_bind_client(struct krping_cb * cb)1896 static int krping_bind_client(struct krping_cb *cb)
1897 {
1898 	struct sockaddr_storage sin;
1899 	int ret;
1900 
1901 	fill_sockaddr(&sin, cb);
1902 
1903 	ret = rdma_resolve_addr(cb->cm_id, NULL, (struct sockaddr *)&sin, 2000);
1904 	if (ret) {
1905 		printk(KERN_ERR PFX "rdma_resolve_addr error %d\n", ret);
1906 		return ret;
1907 	}
1908 
1909 	wait_event_interruptible(cb->sem, cb->state >= ROUTE_RESOLVED);
1910 	if (cb->state != ROUTE_RESOLVED) {
1911 		printk(KERN_ERR PFX
1912 		       "addr/route resolution did not resolve: state %d\n",
1913 		       cb->state);
1914 		return -EINTR;
1915 	}
1916 
1917 	if (!reg_supported(cb->cm_id->device))
1918 		return -EINVAL;
1919 
1920 	DEBUG_LOG("rdma_resolve_addr - rdma_resolve_route successful\n");
1921 	return 0;
1922 }
1923 
krping_run_client(struct krping_cb * cb)1924 static void krping_run_client(struct krping_cb *cb)
1925 {
1926 	const struct ib_recv_wr *bad_wr;
1927 	int ret;
1928 
1929 	/* set type of service, if any */
1930 	if (cb->tos != 0)
1931 		rdma_set_service_type(cb->cm_id, cb->tos);
1932 
1933 	ret = krping_bind_client(cb);
1934 	if (ret)
1935 		return;
1936 
1937 	ret = krping_setup_qp(cb, cb->cm_id);
1938 	if (ret) {
1939 		printk(KERN_ERR PFX "setup_qp failed: %d\n", ret);
1940 		return;
1941 	}
1942 
1943 	ret = krping_setup_buffers(cb);
1944 	if (ret) {
1945 		printk(KERN_ERR PFX "krping_setup_buffers failed: %d\n", ret);
1946 		goto err1;
1947 	}
1948 
1949 	ret = ib_post_recv(cb->qp, &cb->rq_wr, &bad_wr);
1950 	if (ret) {
1951 		printk(KERN_ERR PFX "ib_post_recv failed: %d\n", ret);
1952 		goto err2;
1953 	}
1954 
1955 	ret = krping_connect_client(cb);
1956 	if (ret) {
1957 		printk(KERN_ERR PFX "connect error %d\n", ret);
1958 		goto err2;
1959 	}
1960 
1961 	if (cb->wlat)
1962 		krping_wlat_test_client(cb);
1963 	else if (cb->rlat)
1964 		krping_rlat_test_client(cb);
1965 	else if (cb->bw)
1966 		krping_bw_test_client(cb);
1967 	else if (cb->frtest)
1968 		krping_fr_test(cb);
1969 	else
1970 		krping_test_client(cb);
1971 	rdma_disconnect(cb->cm_id);
1972 err2:
1973 	krping_free_buffers(cb);
1974 err1:
1975 	krping_free_qp(cb);
1976 }
1977 
1978 static uint16_t
krping_get_ipv6_scope_id(char * name)1979 krping_get_ipv6_scope_id(char *name)
1980 {
1981 	struct ifnet *ifp;
1982 	uint16_t retval;
1983 
1984 	if (name == NULL)
1985 		return (0);
1986 	CURVNET_SET_QUIET(TD_TO_VNET(curthread));
1987 	ifp = ifunit_ref(name);
1988 	CURVNET_RESTORE();
1989 	if (ifp == NULL)
1990 		return (0);
1991 	retval = if_getindex(ifp);
1992 	if_rele(ifp);
1993 	return (retval);
1994 }
1995 
krping_doit(char * cmd)1996 int krping_doit(char *cmd)
1997 {
1998 	struct krping_cb *cb;
1999 	int op;
2000 	int ret = 0;
2001 	char *optarg;
2002 	char *scope;
2003 	unsigned long optint;
2004 
2005 	cb = kzalloc(sizeof(*cb), GFP_KERNEL);
2006 	if (!cb)
2007 		return -ENOMEM;
2008 
2009 	mutex_lock(&krping_mutex);
2010 	list_add_tail(&cb->list, &krping_cbs);
2011 	mutex_unlock(&krping_mutex);
2012 
2013 	cb->server = -1;
2014 	cb->state = IDLE;
2015 	cb->size = 64;
2016 	cb->txdepth = RPING_SQ_DEPTH;
2017 	init_waitqueue_head(&cb->sem);
2018 
2019 	while ((op = krping_getopt("krping", &cmd, krping_opts, NULL, &optarg,
2020 			      &optint)) != 0) {
2021 		switch (op) {
2022 		case 'a':
2023 			cb->addr_str = optarg;
2024 			cb->addr_type = AF_INET;
2025 			DEBUG_LOG("ipaddr (%s)\n", optarg);
2026 			if (inet_pton(AF_INET, optarg, cb->addr) != 1) {
2027 				printk(KERN_ERR PFX "bad addr string %s\n",
2028 				    optarg);
2029 				ret = EINVAL;
2030 			}
2031 			break;
2032 		case 'A':
2033 			cb->addr_str = optarg;
2034 			cb->addr_type = AF_INET6;
2035 			DEBUG_LOG("ipv6addr (%s)\n", optarg);
2036 			scope = strstr(optarg, "%");
2037 			/* extract scope ID, if any */
2038 			if (scope != NULL)
2039 				*scope++ = 0;
2040 			/* extract IPv6 network address */
2041 			if (inet_pton(AF_INET6, optarg, cb->addr) != 1) {
2042 				printk(KERN_ERR PFX "bad addr string %s\n",
2043 				    optarg);
2044 				ret = EINVAL;
2045 			} else if (IN6_IS_SCOPE_LINKLOCAL((struct in6_addr *)cb->addr) ||
2046 			    IN6_IS_ADDR_MC_INTFACELOCAL((struct in6_addr *)cb->addr)) {
2047 				uint16_t scope_id = krping_get_ipv6_scope_id(scope);
2048 				DEBUG_LOG("ipv6 scope ID = %d\n", scope_id);
2049 				cb->addr[2] = scope_id >> 8;
2050 				cb->addr[3] = scope_id & 0xFF;
2051 			}
2052 			break;
2053 		case 'p':
2054 			cb->port = htons(optint);
2055 			DEBUG_LOG("port %d\n", (int)optint);
2056 			break;
2057 		case 'P':
2058 			cb->poll = 1;
2059 			DEBUG_LOG("server\n");
2060 			break;
2061 		case 's':
2062 			cb->server = 1;
2063 			DEBUG_LOG("server\n");
2064 			break;
2065 		case 'c':
2066 			cb->server = 0;
2067 			DEBUG_LOG("client\n");
2068 			break;
2069 		case 'S':
2070 			cb->size = optint;
2071 			if ((cb->size < 1) ||
2072 			    (cb->size > RPING_BUFSIZE)) {
2073 				printk(KERN_ERR PFX "Invalid size %d "
2074 				       "(valid range is 1 to %d)\n",
2075 				       cb->size, RPING_BUFSIZE);
2076 				ret = EINVAL;
2077 			} else
2078 				DEBUG_LOG("size %d\n", (int)optint);
2079 			break;
2080 		case 'C':
2081 			cb->count = optint;
2082 			if (cb->count < 0) {
2083 				printk(KERN_ERR PFX "Invalid count %d\n",
2084 					cb->count);
2085 				ret = EINVAL;
2086 			} else
2087 				DEBUG_LOG("count %d\n", (int) cb->count);
2088 			break;
2089 		case 'v':
2090 			cb->verbose++;
2091 			DEBUG_LOG("verbose\n");
2092 			break;
2093 		case 'V':
2094 			cb->validate++;
2095 			DEBUG_LOG("validate data\n");
2096 			break;
2097 		case 'l':
2098 			cb->wlat++;
2099 			break;
2100 		case 'L':
2101 			cb->rlat++;
2102 			break;
2103 		case 'B':
2104 			cb->bw++;
2105 			break;
2106 		case 'd':
2107 			cb->duplex++;
2108 			break;
2109 		case 'I':
2110 			cb->server_invalidate = 1;
2111 			break;
2112 		case 't':
2113 			cb->tos = optint;
2114 			DEBUG_LOG("type of service, tos=%d\n", (int) cb->tos);
2115 			break;
2116 		case 'T':
2117 			cb->txdepth = optint;
2118 			DEBUG_LOG("txdepth %d\n", (int) cb->txdepth);
2119 			break;
2120 		case 'Z':
2121 			cb->local_dma_lkey = 1;
2122 			DEBUG_LOG("using local dma lkey\n");
2123 			break;
2124 		case 'R':
2125 			cb->read_inv = 1;
2126 			DEBUG_LOG("using read-with-inv\n");
2127 			break;
2128 		case 'f':
2129 			cb->frtest = 1;
2130 			DEBUG_LOG("fast-reg test!\n");
2131 			break;
2132 		default:
2133 			printk(KERN_ERR PFX "unknown opt %s\n", optarg);
2134 			ret = -EINVAL;
2135 			break;
2136 		}
2137 	}
2138 	if (ret)
2139 		goto out;
2140 
2141 	if (cb->server == -1) {
2142 		printk(KERN_ERR PFX "must be either client or server\n");
2143 		ret = -EINVAL;
2144 		goto out;
2145 	}
2146 
2147 	if (cb->server && cb->frtest) {
2148 		printk(KERN_ERR PFX "must be client to run frtest\n");
2149 		ret = -EINVAL;
2150 		goto out;
2151 	}
2152 
2153 	if ((cb->frtest + cb->bw + cb->rlat + cb->wlat) > 1) {
2154 		printk(KERN_ERR PFX "Pick only one test: fr, bw, rlat, wlat\n");
2155 		ret = -EINVAL;
2156 		goto out;
2157 	}
2158 
2159 	if (cb->wlat || cb->rlat || cb->bw) {
2160 		printk(KERN_ERR PFX "wlat, rlat, and bw tests only support mem_mode MR - which is no longer supported\n");
2161 		ret = -EINVAL;
2162 		goto out;
2163 	}
2164 
2165 	cb->cm_id = rdma_create_id(TD_TO_VNET(curthread), krping_cma_event_handler, cb, RDMA_PS_TCP, IB_QPT_RC);
2166 	if (IS_ERR(cb->cm_id)) {
2167 		ret = PTR_ERR(cb->cm_id);
2168 		printk(KERN_ERR PFX "rdma_create_id error %d\n", ret);
2169 		goto out;
2170 	}
2171 	DEBUG_LOG("created cm_id %p\n", cb->cm_id);
2172 
2173 	if (cb->server)
2174 		krping_run_server(cb);
2175 	else
2176 		krping_run_client(cb);
2177 
2178 	DEBUG_LOG("destroy cm_id %p\n", cb->cm_id);
2179 	rdma_destroy_id(cb->cm_id);
2180 out:
2181 	mutex_lock(&krping_mutex);
2182 	list_del(&cb->list);
2183 	mutex_unlock(&krping_mutex);
2184 	kfree(cb);
2185 	return ret;
2186 }
2187 
2188 void
krping_walk_cb_list(void (* f)(struct krping_stats *,void *),void * arg)2189 krping_walk_cb_list(void (*f)(struct krping_stats *, void *), void *arg)
2190 {
2191 	struct krping_cb *cb;
2192 
2193 	mutex_lock(&krping_mutex);
2194 	list_for_each_entry(cb, &krping_cbs, list)
2195 	    (*f)(cb->pd ? &cb->stats : NULL, arg);
2196 	mutex_unlock(&krping_mutex);
2197 }
2198 
2199 void
krping_cancel_all(void)2200 krping_cancel_all(void)
2201 {
2202 	struct krping_cb *cb;
2203 
2204 	mutex_lock(&krping_mutex);
2205 	list_for_each_entry(cb, &krping_cbs, list) {
2206 		cb->state = ERROR;
2207 		wake_up_interruptible(&cb->sem);
2208 	}
2209 	mutex_unlock(&krping_mutex);
2210 }
2211 
2212