xref: /linux/drivers/infiniband/sw/siw/siw_verbs.c (revision dd91b5e1d6448794c07378d1be12e3261c8769e7)
1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
2 
3 /* Authors: Bernard Metzler <bmt@zurich.ibm.com> */
4 /* Copyright (c) 2008-2019, IBM Corporation */
5 
6 #include <linux/errno.h>
7 #include <linux/types.h>
8 #include <linux/uaccess.h>
9 #include <linux/vmalloc.h>
10 #include <linux/xarray.h>
11 #include <net/addrconf.h>
12 
13 #include <rdma/iw_cm.h>
14 #include <rdma/ib_verbs.h>
15 #include <rdma/ib_user_verbs.h>
16 #include <rdma/uverbs_ioctl.h>
17 
18 #include "siw.h"
19 #include "siw_verbs.h"
20 #include "siw_mem.h"
21 
22 static int siw_qp_state_to_ib_qp_state[SIW_QP_STATE_COUNT] = {
23 	[SIW_QP_STATE_IDLE] = IB_QPS_INIT,
24 	[SIW_QP_STATE_RTR] = IB_QPS_RTR,
25 	[SIW_QP_STATE_RTS] = IB_QPS_RTS,
26 	[SIW_QP_STATE_CLOSING] = IB_QPS_SQD,
27 	[SIW_QP_STATE_TERMINATE] = IB_QPS_SQE,
28 	[SIW_QP_STATE_ERROR] = IB_QPS_ERR
29 };
30 
31 static int ib_qp_state_to_siw_qp_state[IB_QPS_ERR + 1] = {
32 	[IB_QPS_RESET] = SIW_QP_STATE_IDLE,
33 	[IB_QPS_INIT] = SIW_QP_STATE_IDLE,
34 	[IB_QPS_RTR] = SIW_QP_STATE_RTR,
35 	[IB_QPS_RTS] = SIW_QP_STATE_RTS,
36 	[IB_QPS_SQD] = SIW_QP_STATE_CLOSING,
37 	[IB_QPS_SQE] = SIW_QP_STATE_TERMINATE,
38 	[IB_QPS_ERR] = SIW_QP_STATE_ERROR
39 };
40 
41 static char ib_qp_state_to_string[IB_QPS_ERR + 1][sizeof("RESET")] = {
42 	[IB_QPS_RESET] = "RESET", [IB_QPS_INIT] = "INIT", [IB_QPS_RTR] = "RTR",
43 	[IB_QPS_RTS] = "RTS",     [IB_QPS_SQD] = "SQD",   [IB_QPS_SQE] = "SQE",
44 	[IB_QPS_ERR] = "ERR"
45 };
46 
siw_mmap_free(struct rdma_user_mmap_entry * rdma_entry)47 void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry)
48 {
49 	struct siw_user_mmap_entry *entry = to_siw_mmap_entry(rdma_entry);
50 
51 	kfree(entry);
52 }
53 
siw_mmap(struct ib_ucontext * ctx,struct vm_area_struct * vma)54 int siw_mmap(struct ib_ucontext *ctx, struct vm_area_struct *vma)
55 {
56 	struct siw_ucontext *uctx = to_siw_ctx(ctx);
57 	size_t size = vma->vm_end - vma->vm_start;
58 	struct rdma_user_mmap_entry *rdma_entry;
59 	struct siw_user_mmap_entry *entry;
60 	int rv = -EINVAL;
61 
62 	/*
63 	 * Must be page aligned
64 	 */
65 	if (vma->vm_start & (PAGE_SIZE - 1)) {
66 		pr_warn("siw: mmap not page aligned\n");
67 		return -EINVAL;
68 	}
69 	rdma_entry = rdma_user_mmap_entry_get(&uctx->base_ucontext, vma);
70 	if (!rdma_entry) {
71 		siw_dbg(&uctx->sdev->base_dev, "mmap lookup failed: %lu, %#zx\n",
72 			vma->vm_pgoff, size);
73 		return -EINVAL;
74 	}
75 	entry = to_siw_mmap_entry(rdma_entry);
76 
77 	rv = remap_vmalloc_range(vma, entry->address, 0);
78 	if (rv)
79 		pr_warn("remap_vmalloc_range failed: %lu, %zu\n", vma->vm_pgoff,
80 			size);
81 	rdma_user_mmap_entry_put(rdma_entry);
82 
83 	return rv;
84 }
85 
siw_alloc_ucontext(struct ib_ucontext * base_ctx,struct ib_udata * udata)86 int siw_alloc_ucontext(struct ib_ucontext *base_ctx, struct ib_udata *udata)
87 {
88 	struct siw_device *sdev = to_siw_dev(base_ctx->device);
89 	struct siw_ucontext *ctx = to_siw_ctx(base_ctx);
90 	struct siw_uresp_alloc_ctx uresp = {};
91 	int rv;
92 
93 	if (atomic_inc_return(&sdev->num_ctx) > SIW_MAX_CONTEXT) {
94 		rv = -ENOMEM;
95 		goto err_out;
96 	}
97 	ctx->sdev = sdev;
98 
99 	uresp.dev_id = sdev->vendor_part_id;
100 
101 	if (udata->outlen < sizeof(uresp)) {
102 		rv = -EINVAL;
103 		goto err_out;
104 	}
105 	rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp));
106 	if (rv)
107 		goto err_out;
108 
109 	siw_dbg(base_ctx->device, "success. now %d context(s)\n",
110 		atomic_read(&sdev->num_ctx));
111 
112 	return 0;
113 
114 err_out:
115 	atomic_dec(&sdev->num_ctx);
116 	siw_dbg(base_ctx->device, "failure %d. now %d context(s)\n", rv,
117 		atomic_read(&sdev->num_ctx));
118 
119 	return rv;
120 }
121 
siw_dealloc_ucontext(struct ib_ucontext * base_ctx)122 void siw_dealloc_ucontext(struct ib_ucontext *base_ctx)
123 {
124 	struct siw_ucontext *uctx = to_siw_ctx(base_ctx);
125 
126 	atomic_dec(&uctx->sdev->num_ctx);
127 }
128 
siw_query_device(struct ib_device * base_dev,struct ib_device_attr * attr,struct ib_udata * udata)129 int siw_query_device(struct ib_device *base_dev, struct ib_device_attr *attr,
130 		     struct ib_udata *udata)
131 {
132 	struct siw_device *sdev = to_siw_dev(base_dev);
133 
134 	if (udata->inlen || udata->outlen)
135 		return -EINVAL;
136 
137 	memset(attr, 0, sizeof(*attr));
138 
139 	/* Revisit atomic caps if RFC 7306 gets supported */
140 	attr->atomic_cap = 0;
141 	attr->device_cap_flags = IB_DEVICE_MEM_MGT_EXTENSIONS;
142 	attr->kernel_cap_flags = IBK_ALLOW_USER_UNREG;
143 	attr->max_cq = sdev->attrs.max_cq;
144 	attr->max_cqe = sdev->attrs.max_cqe;
145 	attr->max_fast_reg_page_list_len = SIW_MAX_SGE_PBL;
146 	attr->max_mr = sdev->attrs.max_mr;
147 	attr->max_mw = sdev->attrs.max_mw;
148 	attr->max_mr_size = ~0ull;
149 	attr->max_pd = sdev->attrs.max_pd;
150 	attr->max_qp = sdev->attrs.max_qp;
151 	attr->max_qp_init_rd_atom = sdev->attrs.max_ird;
152 	attr->max_qp_rd_atom = sdev->attrs.max_ord;
153 	attr->max_qp_wr = sdev->attrs.max_qp_wr;
154 	attr->max_recv_sge = sdev->attrs.max_sge;
155 	attr->max_res_rd_atom = sdev->attrs.max_qp * sdev->attrs.max_ird;
156 	attr->max_send_sge = sdev->attrs.max_sge;
157 	attr->max_sge_rd = sdev->attrs.max_sge_rd;
158 	attr->max_srq = sdev->attrs.max_srq;
159 	attr->max_srq_sge = sdev->attrs.max_srq_sge;
160 	attr->max_srq_wr = sdev->attrs.max_srq_wr;
161 	attr->page_size_cap = PAGE_SIZE;
162 	attr->vendor_id = SIW_VENDOR_ID;
163 	attr->vendor_part_id = sdev->vendor_part_id;
164 
165 	addrconf_addr_eui48((u8 *)&attr->sys_image_guid,
166 			    sdev->raw_gid);
167 
168 	return 0;
169 }
170 
siw_query_port(struct ib_device * base_dev,u32 port,struct ib_port_attr * attr)171 int siw_query_port(struct ib_device *base_dev, u32 port,
172 		   struct ib_port_attr *attr)
173 {
174 	struct net_device *ndev;
175 	int rv;
176 
177 	memset(attr, 0, sizeof(*attr));
178 
179 	rv = ib_get_eth_speed(base_dev, port, &attr->active_speed,
180 			 &attr->active_width);
181 	if (rv)
182 		return rv;
183 
184 	ndev = ib_device_get_netdev(base_dev, SIW_PORT);
185 	if (!ndev)
186 		return -ENODEV;
187 
188 	attr->gid_tbl_len = 1;
189 	attr->max_msg_sz = -1;
190 	attr->max_mtu = ib_mtu_int_to_enum(ndev->max_mtu);
191 	attr->active_mtu = ib_mtu_int_to_enum(READ_ONCE(ndev->mtu));
192 	attr->state = ib_get_curr_port_state(ndev);
193 	attr->phys_state = attr->state == IB_PORT_ACTIVE ?
194 		IB_PORT_PHYS_STATE_LINK_UP : IB_PORT_PHYS_STATE_DISABLED;
195 	attr->port_cap_flags = IB_PORT_CM_SUP | IB_PORT_DEVICE_MGMT_SUP;
196 	/*
197 	 * All zero
198 	 *
199 	 * attr->lid = 0;
200 	 * attr->bad_pkey_cntr = 0;
201 	 * attr->qkey_viol_cntr = 0;
202 	 * attr->sm_lid = 0;
203 	 * attr->lmc = 0;
204 	 * attr->max_vl_num = 0;
205 	 * attr->sm_sl = 0;
206 	 * attr->subnet_timeout = 0;
207 	 * attr->init_type_repy = 0;
208 	 */
209 	dev_put(ndev);
210 	return rv;
211 }
212 
siw_get_port_immutable(struct ib_device * base_dev,u32 port,struct ib_port_immutable * port_immutable)213 int siw_get_port_immutable(struct ib_device *base_dev, u32 port,
214 			   struct ib_port_immutable *port_immutable)
215 {
216 	struct ib_port_attr attr;
217 	int rv = siw_query_port(base_dev, port, &attr);
218 
219 	if (rv)
220 		return rv;
221 
222 	port_immutable->gid_tbl_len = attr.gid_tbl_len;
223 	port_immutable->core_cap_flags = RDMA_CORE_PORT_IWARP;
224 
225 	return 0;
226 }
227 
siw_query_gid(struct ib_device * base_dev,u32 port,int idx,union ib_gid * gid)228 int siw_query_gid(struct ib_device *base_dev, u32 port, int idx,
229 		  union ib_gid *gid)
230 {
231 	struct siw_device *sdev = to_siw_dev(base_dev);
232 
233 	/* subnet_prefix == interface_id == 0; */
234 	memset(gid, 0, sizeof(*gid));
235 	memcpy(gid->raw, sdev->raw_gid, ETH_ALEN);
236 
237 	return 0;
238 }
239 
siw_alloc_pd(struct ib_pd * pd,struct ib_udata * udata)240 int siw_alloc_pd(struct ib_pd *pd, struct ib_udata *udata)
241 {
242 	struct siw_device *sdev = to_siw_dev(pd->device);
243 
244 	if (atomic_inc_return(&sdev->num_pd) > SIW_MAX_PD) {
245 		atomic_dec(&sdev->num_pd);
246 		return -ENOMEM;
247 	}
248 	siw_dbg_pd(pd, "now %d PD's(s)\n", atomic_read(&sdev->num_pd));
249 
250 	return 0;
251 }
252 
siw_dealloc_pd(struct ib_pd * pd,struct ib_udata * udata)253 int siw_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
254 {
255 	struct siw_device *sdev = to_siw_dev(pd->device);
256 
257 	siw_dbg_pd(pd, "free PD\n");
258 	atomic_dec(&sdev->num_pd);
259 	return 0;
260 }
261 
siw_qp_get_ref(struct ib_qp * base_qp)262 void siw_qp_get_ref(struct ib_qp *base_qp)
263 {
264 	siw_qp_get(to_siw_qp(base_qp));
265 }
266 
siw_qp_put_ref(struct ib_qp * base_qp)267 void siw_qp_put_ref(struct ib_qp *base_qp)
268 {
269 	siw_qp_put(to_siw_qp(base_qp));
270 }
271 
272 static struct rdma_user_mmap_entry *
siw_mmap_entry_insert(struct siw_ucontext * uctx,void * address,size_t length,u64 * offset)273 siw_mmap_entry_insert(struct siw_ucontext *uctx,
274 		      void *address, size_t length,
275 		      u64 *offset)
276 {
277 	struct siw_user_mmap_entry *entry = kzalloc(sizeof(*entry), GFP_KERNEL);
278 	int rv;
279 
280 	*offset = SIW_INVAL_UOBJ_KEY;
281 	if (!entry)
282 		return NULL;
283 
284 	entry->address = address;
285 
286 	rv = rdma_user_mmap_entry_insert(&uctx->base_ucontext,
287 					 &entry->rdma_entry,
288 					 length);
289 	if (rv) {
290 		kfree(entry);
291 		return NULL;
292 	}
293 
294 	*offset = rdma_user_mmap_get_offset(&entry->rdma_entry);
295 
296 	return &entry->rdma_entry;
297 }
298 
299 /*
300  * siw_create_qp()
301  *
302  * Create QP of requested size on given device.
303  *
304  * @qp:		Queue pait
305  * @attrs:	Initial QP attributes.
306  * @udata:	used to provide QP ID, SQ and RQ size back to user.
307  */
308 
siw_create_qp(struct ib_qp * ibqp,struct ib_qp_init_attr * attrs,struct ib_udata * udata)309 int siw_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *attrs,
310 		  struct ib_udata *udata)
311 {
312 	struct ib_pd *pd = ibqp->pd;
313 	struct siw_qp *qp = to_siw_qp(ibqp);
314 	struct ib_device *base_dev = pd->device;
315 	struct siw_device *sdev = to_siw_dev(base_dev);
316 	struct siw_ucontext *uctx =
317 		rdma_udata_to_drv_context(udata, struct siw_ucontext,
318 					  base_ucontext);
319 	unsigned long flags;
320 	int num_sqe, num_rqe, rv = 0;
321 	size_t length;
322 
323 	siw_dbg(base_dev, "create new QP\n");
324 
325 	if (attrs->create_flags)
326 		return -EOPNOTSUPP;
327 
328 	if (atomic_inc_return(&sdev->num_qp) > SIW_MAX_QP) {
329 		siw_dbg(base_dev, "too many QP's\n");
330 		rv = -ENOMEM;
331 		goto err_atomic;
332 	}
333 	if (attrs->qp_type != IB_QPT_RC) {
334 		siw_dbg(base_dev, "only RC QP's supported\n");
335 		rv = -EOPNOTSUPP;
336 		goto err_atomic;
337 	}
338 	if ((attrs->cap.max_send_wr > SIW_MAX_QP_WR) ||
339 	    (attrs->cap.max_recv_wr > SIW_MAX_QP_WR) ||
340 	    (attrs->cap.max_send_sge > SIW_MAX_SGE) ||
341 	    (attrs->cap.max_recv_sge > SIW_MAX_SGE)) {
342 		siw_dbg(base_dev, "QP size error\n");
343 		rv = -EINVAL;
344 		goto err_atomic;
345 	}
346 	if (attrs->cap.max_inline_data > SIW_MAX_INLINE) {
347 		siw_dbg(base_dev, "max inline send: %d > %d\n",
348 			attrs->cap.max_inline_data, (int)SIW_MAX_INLINE);
349 		rv = -EINVAL;
350 		goto err_atomic;
351 	}
352 	/*
353 	 * NOTE: we don't allow for a QP unable to hold any SQ WQE
354 	 */
355 	if (attrs->cap.max_send_wr == 0) {
356 		siw_dbg(base_dev, "QP must have send queue\n");
357 		rv = -EINVAL;
358 		goto err_atomic;
359 	}
360 
361 	if (!attrs->send_cq || (!attrs->recv_cq && !attrs->srq)) {
362 		siw_dbg(base_dev, "send CQ or receive CQ invalid\n");
363 		rv = -EINVAL;
364 		goto err_atomic;
365 	}
366 
367 	init_rwsem(&qp->state_lock);
368 	spin_lock_init(&qp->sq_lock);
369 	spin_lock_init(&qp->rq_lock);
370 	spin_lock_init(&qp->orq_lock);
371 
372 	rv = siw_qp_add(sdev, qp);
373 	if (rv)
374 		goto err_atomic;
375 
376 
377 	/* All queue indices are derived from modulo operations
378 	 * on a free running 'get' (consumer) and 'put' (producer)
379 	 * unsigned counter. Having queue sizes at power of two
380 	 * avoids handling counter wrap around.
381 	 */
382 	num_sqe = roundup_pow_of_two(attrs->cap.max_send_wr);
383 	num_rqe = attrs->cap.max_recv_wr;
384 	if (num_rqe)
385 		num_rqe = roundup_pow_of_two(num_rqe);
386 
387 	if (udata)
388 		qp->sendq = vmalloc_user(num_sqe * sizeof(struct siw_sqe));
389 	else
390 		qp->sendq = vcalloc(num_sqe, sizeof(struct siw_sqe));
391 
392 	if (qp->sendq == NULL) {
393 		rv = -ENOMEM;
394 		goto err_out_xa;
395 	}
396 	if (attrs->sq_sig_type != IB_SIGNAL_REQ_WR) {
397 		if (attrs->sq_sig_type == IB_SIGNAL_ALL_WR)
398 			qp->attrs.flags |= SIW_SIGNAL_ALL_WR;
399 		else {
400 			rv = -EINVAL;
401 			goto err_out_xa;
402 		}
403 	}
404 	qp->pd = pd;
405 	qp->scq = to_siw_cq(attrs->send_cq);
406 	qp->rcq = to_siw_cq(attrs->recv_cq);
407 
408 	if (attrs->srq) {
409 		/*
410 		 * SRQ support.
411 		 * Verbs 6.3.7: ignore RQ size, if SRQ present
412 		 * Verbs 6.3.5: do not check PD of SRQ against PD of QP
413 		 */
414 		qp->srq = to_siw_srq(attrs->srq);
415 		qp->attrs.rq_size = 0;
416 		siw_dbg(base_dev, "QP [%u]: SRQ attached\n",
417 			qp->base_qp.qp_num);
418 	} else if (num_rqe) {
419 		if (udata)
420 			qp->recvq =
421 				vmalloc_user(num_rqe * sizeof(struct siw_rqe));
422 		else
423 			qp->recvq = vcalloc(num_rqe, sizeof(struct siw_rqe));
424 
425 		if (qp->recvq == NULL) {
426 			rv = -ENOMEM;
427 			goto err_out_xa;
428 		}
429 		qp->attrs.rq_size = num_rqe;
430 	}
431 	qp->attrs.sq_size = num_sqe;
432 	qp->attrs.sq_max_sges = attrs->cap.max_send_sge;
433 	qp->attrs.rq_max_sges = attrs->cap.max_recv_sge;
434 
435 	/* Make those two tunables fixed for now. */
436 	qp->tx_ctx.gso_seg_limit = 1;
437 	qp->tx_ctx.zcopy_tx = zcopy_tx;
438 
439 	qp->attrs.state = SIW_QP_STATE_IDLE;
440 
441 	if (udata) {
442 		struct siw_uresp_create_qp uresp = {};
443 
444 		uresp.num_sqe = num_sqe;
445 		uresp.num_rqe = num_rqe;
446 		uresp.qp_id = qp_id(qp);
447 
448 		if (qp->sendq) {
449 			length = num_sqe * sizeof(struct siw_sqe);
450 			qp->sq_entry =
451 				siw_mmap_entry_insert(uctx, qp->sendq,
452 						      length, &uresp.sq_key);
453 			if (!qp->sq_entry) {
454 				rv = -ENOMEM;
455 				goto err_out_xa;
456 			}
457 		}
458 
459 		if (qp->recvq) {
460 			length = num_rqe * sizeof(struct siw_rqe);
461 			qp->rq_entry =
462 				siw_mmap_entry_insert(uctx, qp->recvq,
463 						      length, &uresp.rq_key);
464 			if (!qp->rq_entry) {
465 				uresp.sq_key = SIW_INVAL_UOBJ_KEY;
466 				rv = -ENOMEM;
467 				goto err_out_xa;
468 			}
469 		}
470 
471 		if (udata->outlen < sizeof(uresp)) {
472 			rv = -EINVAL;
473 			goto err_out_xa;
474 		}
475 		rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp));
476 		if (rv)
477 			goto err_out_xa;
478 	}
479 	qp->tx_cpu = siw_get_tx_cpu(sdev);
480 	if (qp->tx_cpu < 0) {
481 		rv = -EINVAL;
482 		goto err_out_xa;
483 	}
484 	INIT_LIST_HEAD(&qp->devq);
485 	spin_lock_irqsave(&sdev->lock, flags);
486 	list_add_tail(&qp->devq, &sdev->qp_list);
487 	spin_unlock_irqrestore(&sdev->lock, flags);
488 
489 	init_completion(&qp->qp_free);
490 
491 	return 0;
492 
493 err_out_xa:
494 	xa_erase(&sdev->qp_xa, qp_id(qp));
495 	if (uctx) {
496 		rdma_user_mmap_entry_remove(qp->sq_entry);
497 		rdma_user_mmap_entry_remove(qp->rq_entry);
498 	}
499 	vfree(qp->sendq);
500 	vfree(qp->recvq);
501 
502 err_atomic:
503 	atomic_dec(&sdev->num_qp);
504 	return rv;
505 }
506 
507 /*
508  * Minimum siw_query_qp() verb interface.
509  *
510  * @qp_attr_mask is not used but all available information is provided
511  */
siw_query_qp(struct ib_qp * base_qp,struct ib_qp_attr * qp_attr,int qp_attr_mask,struct ib_qp_init_attr * qp_init_attr)512 int siw_query_qp(struct ib_qp *base_qp, struct ib_qp_attr *qp_attr,
513 		 int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr)
514 {
515 	struct siw_qp *qp;
516 	struct net_device *ndev;
517 
518 	if (base_qp && qp_attr && qp_init_attr)
519 		qp = to_siw_qp(base_qp);
520 	else
521 		return -EINVAL;
522 
523 	ndev = ib_device_get_netdev(base_qp->device, SIW_PORT);
524 	if (!ndev)
525 		return -ENODEV;
526 
527 	qp_attr->qp_state = siw_qp_state_to_ib_qp_state[qp->attrs.state];
528 	qp_attr->cap.max_inline_data = SIW_MAX_INLINE;
529 	qp_attr->cap.max_send_wr = qp->attrs.sq_size;
530 	qp_attr->cap.max_send_sge = qp->attrs.sq_max_sges;
531 	qp_attr->cap.max_recv_wr = qp->attrs.rq_size;
532 	qp_attr->cap.max_recv_sge = qp->attrs.rq_max_sges;
533 	qp_attr->path_mtu = ib_mtu_int_to_enum(READ_ONCE(ndev->mtu));
534 	qp_attr->max_rd_atomic = qp->attrs.irq_size;
535 	qp_attr->max_dest_rd_atomic = qp->attrs.orq_size;
536 
537 	qp_attr->qp_access_flags = IB_ACCESS_LOCAL_WRITE |
538 				   IB_ACCESS_REMOTE_WRITE |
539 				   IB_ACCESS_REMOTE_READ;
540 
541 	qp_init_attr->qp_type = base_qp->qp_type;
542 	qp_init_attr->send_cq = base_qp->send_cq;
543 	qp_init_attr->recv_cq = base_qp->recv_cq;
544 	qp_init_attr->srq = base_qp->srq;
545 
546 	qp_init_attr->cap = qp_attr->cap;
547 
548 	dev_put(ndev);
549 	return 0;
550 }
551 
siw_verbs_modify_qp(struct ib_qp * base_qp,struct ib_qp_attr * attr,int attr_mask,struct ib_udata * udata)552 int siw_verbs_modify_qp(struct ib_qp *base_qp, struct ib_qp_attr *attr,
553 			int attr_mask, struct ib_udata *udata)
554 {
555 	struct siw_qp_attrs new_attrs;
556 	enum siw_qp_attr_mask siw_attr_mask = 0;
557 	struct siw_qp *qp = to_siw_qp(base_qp);
558 	int rv = 0;
559 
560 	if (!attr_mask)
561 		return 0;
562 
563 	if (attr_mask & ~IB_QP_ATTR_STANDARD_BITS)
564 		return -EOPNOTSUPP;
565 
566 	memset(&new_attrs, 0, sizeof(new_attrs));
567 
568 	if (attr_mask & IB_QP_ACCESS_FLAGS) {
569 		siw_attr_mask = SIW_QP_ATTR_ACCESS_FLAGS;
570 
571 		if (attr->qp_access_flags & IB_ACCESS_REMOTE_READ)
572 			new_attrs.flags |= SIW_RDMA_READ_ENABLED;
573 		if (attr->qp_access_flags & IB_ACCESS_REMOTE_WRITE)
574 			new_attrs.flags |= SIW_RDMA_WRITE_ENABLED;
575 		if (attr->qp_access_flags & IB_ACCESS_MW_BIND)
576 			new_attrs.flags |= SIW_RDMA_BIND_ENABLED;
577 	}
578 	if (attr_mask & IB_QP_STATE) {
579 		siw_dbg_qp(qp, "desired IB QP state: %s\n",
580 			   ib_qp_state_to_string[attr->qp_state]);
581 
582 		new_attrs.state = ib_qp_state_to_siw_qp_state[attr->qp_state];
583 
584 		if (new_attrs.state > SIW_QP_STATE_RTS)
585 			qp->tx_ctx.tx_suspend = 1;
586 
587 		siw_attr_mask |= SIW_QP_ATTR_STATE;
588 	}
589 	if (!siw_attr_mask)
590 		goto out;
591 
592 	down_write(&qp->state_lock);
593 
594 	rv = siw_qp_modify(qp, &new_attrs, siw_attr_mask);
595 
596 	up_write(&qp->state_lock);
597 out:
598 	return rv;
599 }
600 
siw_destroy_qp(struct ib_qp * base_qp,struct ib_udata * udata)601 int siw_destroy_qp(struct ib_qp *base_qp, struct ib_udata *udata)
602 {
603 	struct siw_qp *qp = to_siw_qp(base_qp);
604 	struct siw_ucontext *uctx =
605 		rdma_udata_to_drv_context(udata, struct siw_ucontext,
606 					  base_ucontext);
607 	struct siw_qp_attrs qp_attrs;
608 
609 	siw_dbg_qp(qp, "state %d\n", qp->attrs.state);
610 
611 	/*
612 	 * Mark QP as in process of destruction to prevent from
613 	 * any async callbacks to RDMA core
614 	 */
615 	qp->attrs.flags |= SIW_QP_IN_DESTROY;
616 	qp->rx_stream.rx_suspend = 1;
617 
618 	if (uctx) {
619 		rdma_user_mmap_entry_remove(qp->sq_entry);
620 		rdma_user_mmap_entry_remove(qp->rq_entry);
621 	}
622 
623 	down_write(&qp->state_lock);
624 
625 	qp_attrs.state = SIW_QP_STATE_ERROR;
626 	siw_qp_modify(qp, &qp_attrs, SIW_QP_ATTR_STATE);
627 
628 	if (qp->cep) {
629 		siw_cep_put(qp->cep);
630 		qp->cep = NULL;
631 	}
632 	up_write(&qp->state_lock);
633 
634 	qp->scq = qp->rcq = NULL;
635 
636 	siw_qp_put(qp);
637 	wait_for_completion(&qp->qp_free);
638 
639 	return 0;
640 }
641 
642 /*
643  * siw_copy_inline_sgl()
644  *
645  * Prepare sgl of inlined data for sending. For userland callers
646  * function checks if given buffer addresses and len's are within
647  * process context bounds.
648  * Data from all provided sge's are copied together into the wqe,
649  * referenced by a single sge.
650  */
siw_copy_inline_sgl(const struct ib_send_wr * core_wr,struct siw_sqe * sqe)651 static int siw_copy_inline_sgl(const struct ib_send_wr *core_wr,
652 			       struct siw_sqe *sqe)
653 {
654 	struct ib_sge *core_sge = core_wr->sg_list;
655 	void *kbuf = &sqe->sge[1];
656 	int num_sge = core_wr->num_sge, bytes = 0;
657 
658 	sqe->sge[0].laddr = (uintptr_t)kbuf;
659 	sqe->sge[0].lkey = 0;
660 
661 	while (num_sge--) {
662 		if (!core_sge->length) {
663 			core_sge++;
664 			continue;
665 		}
666 		bytes += core_sge->length;
667 		if (bytes > SIW_MAX_INLINE) {
668 			bytes = -EINVAL;
669 			break;
670 		}
671 		memcpy(kbuf, ib_virt_dma_to_ptr(core_sge->addr),
672 		       core_sge->length);
673 
674 		kbuf += core_sge->length;
675 		core_sge++;
676 	}
677 	sqe->sge[0].length = max(bytes, 0);
678 	sqe->num_sge = bytes > 0 ? 1 : 0;
679 
680 	return bytes;
681 }
682 
683 /* Complete SQ WR's without processing */
siw_sq_flush_wr(struct siw_qp * qp,const struct ib_send_wr * wr,const struct ib_send_wr ** bad_wr)684 static int siw_sq_flush_wr(struct siw_qp *qp, const struct ib_send_wr *wr,
685 			   const struct ib_send_wr **bad_wr)
686 {
687 	int rv = 0;
688 
689 	while (wr) {
690 		struct siw_sqe sqe = {};
691 
692 		switch (wr->opcode) {
693 		case IB_WR_RDMA_WRITE:
694 			sqe.opcode = SIW_OP_WRITE;
695 			break;
696 		case IB_WR_RDMA_READ:
697 			sqe.opcode = SIW_OP_READ;
698 			break;
699 		case IB_WR_RDMA_READ_WITH_INV:
700 			sqe.opcode = SIW_OP_READ_LOCAL_INV;
701 			break;
702 		case IB_WR_SEND:
703 			sqe.opcode = SIW_OP_SEND;
704 			break;
705 		case IB_WR_SEND_WITH_IMM:
706 			sqe.opcode = SIW_OP_SEND_WITH_IMM;
707 			break;
708 		case IB_WR_SEND_WITH_INV:
709 			sqe.opcode = SIW_OP_SEND_REMOTE_INV;
710 			break;
711 		case IB_WR_LOCAL_INV:
712 			sqe.opcode = SIW_OP_INVAL_STAG;
713 			break;
714 		case IB_WR_REG_MR:
715 			sqe.opcode = SIW_OP_REG_MR;
716 			break;
717 		default:
718 			rv = -EINVAL;
719 			break;
720 		}
721 		if (!rv) {
722 			sqe.id = wr->wr_id;
723 			rv = siw_sqe_complete(qp, &sqe, 0,
724 					      SIW_WC_WR_FLUSH_ERR);
725 		}
726 		if (rv) {
727 			if (bad_wr)
728 				*bad_wr = wr;
729 			break;
730 		}
731 		wr = wr->next;
732 	}
733 	return rv;
734 }
735 
736 /* Complete RQ WR's without processing */
siw_rq_flush_wr(struct siw_qp * qp,const struct ib_recv_wr * wr,const struct ib_recv_wr ** bad_wr)737 static int siw_rq_flush_wr(struct siw_qp *qp, const struct ib_recv_wr *wr,
738 			   const struct ib_recv_wr **bad_wr)
739 {
740 	struct siw_rqe rqe = {};
741 	int rv = 0;
742 
743 	while (wr) {
744 		rqe.id = wr->wr_id;
745 		rv = siw_rqe_complete(qp, &rqe, 0, 0, SIW_WC_WR_FLUSH_ERR);
746 		if (rv) {
747 			if (bad_wr)
748 				*bad_wr = wr;
749 			break;
750 		}
751 		wr = wr->next;
752 	}
753 	return rv;
754 }
755 
756 /*
757  * siw_post_send()
758  *
759  * Post a list of S-WR's to a SQ.
760  *
761  * @base_qp:	Base QP contained in siw QP
762  * @wr:		Null terminated list of user WR's
763  * @bad_wr:	Points to failing WR in case of synchronous failure.
764  */
siw_post_send(struct ib_qp * base_qp,const struct ib_send_wr * wr,const struct ib_send_wr ** bad_wr)765 int siw_post_send(struct ib_qp *base_qp, const struct ib_send_wr *wr,
766 		  const struct ib_send_wr **bad_wr)
767 {
768 	struct siw_qp *qp = to_siw_qp(base_qp);
769 	struct siw_wqe *wqe = tx_wqe(qp);
770 
771 	unsigned long flags;
772 	int rv = 0;
773 
774 	if (wr && !rdma_is_kernel_res(&qp->base_qp.res)) {
775 		siw_dbg_qp(qp, "wr must be empty for user mapped sq\n");
776 		*bad_wr = wr;
777 		return -EINVAL;
778 	}
779 
780 	/*
781 	 * Try to acquire QP state lock. Must be non-blocking
782 	 * to accommodate kernel clients needs.
783 	 */
784 	if (!down_read_trylock(&qp->state_lock)) {
785 		if (qp->attrs.state == SIW_QP_STATE_ERROR) {
786 			/*
787 			 * ERROR state is final, so we can be sure
788 			 * this state will not change as long as the QP
789 			 * exists.
790 			 *
791 			 * This handles an ib_drain_sq() call with
792 			 * a concurrent request to set the QP state
793 			 * to ERROR.
794 			 */
795 			rv = siw_sq_flush_wr(qp, wr, bad_wr);
796 		} else {
797 			siw_dbg_qp(qp, "QP locked, state %d\n",
798 				   qp->attrs.state);
799 			*bad_wr = wr;
800 			rv = -ENOTCONN;
801 		}
802 		return rv;
803 	}
804 	if (unlikely(qp->attrs.state != SIW_QP_STATE_RTS)) {
805 		if (qp->attrs.state == SIW_QP_STATE_ERROR) {
806 			/*
807 			 * Immediately flush this WR to CQ, if QP
808 			 * is in ERROR state. SQ is guaranteed to
809 			 * be empty, so WR complets in-order.
810 			 *
811 			 * Typically triggered by ib_drain_sq().
812 			 */
813 			rv = siw_sq_flush_wr(qp, wr, bad_wr);
814 		} else {
815 			siw_dbg_qp(qp, "QP out of state %d\n",
816 				   qp->attrs.state);
817 			*bad_wr = wr;
818 			rv = -ENOTCONN;
819 		}
820 		up_read(&qp->state_lock);
821 		return rv;
822 	}
823 	spin_lock_irqsave(&qp->sq_lock, flags);
824 
825 	while (wr) {
826 		u32 idx = qp->sq_put % qp->attrs.sq_size;
827 		struct siw_sqe *sqe = &qp->sendq[idx];
828 
829 		if (sqe->flags) {
830 			siw_dbg_qp(qp, "sq full\n");
831 			rv = -ENOMEM;
832 			break;
833 		}
834 		if (wr->num_sge > qp->attrs.sq_max_sges) {
835 			siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge);
836 			rv = -EINVAL;
837 			break;
838 		}
839 		sqe->id = wr->wr_id;
840 
841 		if ((wr->send_flags & IB_SEND_SIGNALED) ||
842 		    (qp->attrs.flags & SIW_SIGNAL_ALL_WR))
843 			sqe->flags |= SIW_WQE_SIGNALLED;
844 
845 		if (wr->send_flags & IB_SEND_FENCE)
846 			sqe->flags |= SIW_WQE_READ_FENCE;
847 
848 		switch (wr->opcode) {
849 		case IB_WR_SEND:
850 		case IB_WR_SEND_WITH_INV:
851 			if (wr->send_flags & IB_SEND_SOLICITED)
852 				sqe->flags |= SIW_WQE_SOLICITED;
853 
854 			if (!(wr->send_flags & IB_SEND_INLINE)) {
855 				siw_copy_sgl(wr->sg_list, sqe->sge,
856 					     wr->num_sge);
857 				sqe->num_sge = wr->num_sge;
858 			} else {
859 				rv = siw_copy_inline_sgl(wr, sqe);
860 				if (rv <= 0) {
861 					rv = -EINVAL;
862 					break;
863 				}
864 				sqe->flags |= SIW_WQE_INLINE;
865 				sqe->num_sge = 1;
866 			}
867 			if (wr->opcode == IB_WR_SEND)
868 				sqe->opcode = SIW_OP_SEND;
869 			else {
870 				sqe->opcode = SIW_OP_SEND_REMOTE_INV;
871 				sqe->rkey = wr->ex.invalidate_rkey;
872 			}
873 			break;
874 
875 		case IB_WR_RDMA_READ_WITH_INV:
876 		case IB_WR_RDMA_READ:
877 			/*
878 			 * iWarp restricts RREAD sink to SGL containing
879 			 * 1 SGE only. we could relax to SGL with multiple
880 			 * elements referring the SAME ltag or even sending
881 			 * a private per-rreq tag referring to a checked
882 			 * local sgl with MULTIPLE ltag's.
883 			 */
884 			if (unlikely(wr->num_sge != 1)) {
885 				rv = -EINVAL;
886 				break;
887 			}
888 			siw_copy_sgl(wr->sg_list, &sqe->sge[0], 1);
889 			/*
890 			 * NOTE: zero length RREAD is allowed!
891 			 */
892 			sqe->raddr = rdma_wr(wr)->remote_addr;
893 			sqe->rkey = rdma_wr(wr)->rkey;
894 			sqe->num_sge = 1;
895 
896 			if (wr->opcode == IB_WR_RDMA_READ)
897 				sqe->opcode = SIW_OP_READ;
898 			else
899 				sqe->opcode = SIW_OP_READ_LOCAL_INV;
900 			break;
901 
902 		case IB_WR_RDMA_WRITE:
903 			if (!(wr->send_flags & IB_SEND_INLINE)) {
904 				siw_copy_sgl(wr->sg_list, &sqe->sge[0],
905 					     wr->num_sge);
906 				sqe->num_sge = wr->num_sge;
907 			} else {
908 				rv = siw_copy_inline_sgl(wr, sqe);
909 				if (unlikely(rv < 0)) {
910 					rv = -EINVAL;
911 					break;
912 				}
913 				sqe->flags |= SIW_WQE_INLINE;
914 				sqe->num_sge = 1;
915 			}
916 			sqe->raddr = rdma_wr(wr)->remote_addr;
917 			sqe->rkey = rdma_wr(wr)->rkey;
918 			sqe->opcode = SIW_OP_WRITE;
919 			break;
920 
921 		case IB_WR_REG_MR:
922 			sqe->base_mr = (uintptr_t)reg_wr(wr)->mr;
923 			sqe->rkey = reg_wr(wr)->key;
924 			sqe->access = reg_wr(wr)->access & IWARP_ACCESS_MASK;
925 			sqe->opcode = SIW_OP_REG_MR;
926 			break;
927 
928 		case IB_WR_LOCAL_INV:
929 			sqe->rkey = wr->ex.invalidate_rkey;
930 			sqe->opcode = SIW_OP_INVAL_STAG;
931 			break;
932 
933 		default:
934 			siw_dbg_qp(qp, "ib wr type %d unsupported\n",
935 				   wr->opcode);
936 			rv = -EINVAL;
937 			break;
938 		}
939 		siw_dbg_qp(qp, "opcode %d, flags 0x%x, wr_id 0x%p\n",
940 			   sqe->opcode, sqe->flags,
941 			   (void *)(uintptr_t)sqe->id);
942 
943 		if (unlikely(rv < 0))
944 			break;
945 
946 		/* make SQE only valid after completely written */
947 		smp_wmb();
948 		sqe->flags |= SIW_WQE_VALID;
949 
950 		qp->sq_put++;
951 		wr = wr->next;
952 	}
953 
954 	/*
955 	 * Send directly if SQ processing is not in progress.
956 	 * Eventual immediate errors (rv < 0) do not affect the involved
957 	 * RI resources (Verbs, 8.3.1) and thus do not prevent from SQ
958 	 * processing, if new work is already pending. But rv must be passed
959 	 * to caller.
960 	 */
961 	if (wqe->wr_status != SIW_WR_IDLE) {
962 		spin_unlock_irqrestore(&qp->sq_lock, flags);
963 		goto skip_direct_sending;
964 	}
965 	rv = siw_activate_tx(qp);
966 	spin_unlock_irqrestore(&qp->sq_lock, flags);
967 
968 	if (rv <= 0)
969 		goto skip_direct_sending;
970 
971 	if (rdma_is_kernel_res(&qp->base_qp.res)) {
972 		rv = siw_sq_start(qp);
973 	} else {
974 		qp->tx_ctx.in_syscall = 1;
975 
976 		if (siw_qp_sq_process(qp) != 0 && !(qp->tx_ctx.tx_suspend))
977 			siw_qp_cm_drop(qp, 0);
978 
979 		qp->tx_ctx.in_syscall = 0;
980 	}
981 skip_direct_sending:
982 
983 	up_read(&qp->state_lock);
984 
985 	if (rv >= 0)
986 		return 0;
987 	/*
988 	 * Immediate error
989 	 */
990 	siw_dbg_qp(qp, "error %d\n", rv);
991 
992 	*bad_wr = wr;
993 	return rv;
994 }
995 
996 /*
997  * siw_post_receive()
998  *
999  * Post a list of R-WR's to a RQ.
1000  *
1001  * @base_qp:	Base QP contained in siw QP
1002  * @wr:		Null terminated list of user WR's
1003  * @bad_wr:	Points to failing WR in case of synchronous failure.
1004  */
siw_post_receive(struct ib_qp * base_qp,const struct ib_recv_wr * wr,const struct ib_recv_wr ** bad_wr)1005 int siw_post_receive(struct ib_qp *base_qp, const struct ib_recv_wr *wr,
1006 		     const struct ib_recv_wr **bad_wr)
1007 {
1008 	struct siw_qp *qp = to_siw_qp(base_qp);
1009 	unsigned long flags;
1010 	int rv = 0;
1011 
1012 	if (qp->srq || qp->attrs.rq_size == 0) {
1013 		*bad_wr = wr;
1014 		return -EINVAL;
1015 	}
1016 	if (!rdma_is_kernel_res(&qp->base_qp.res)) {
1017 		siw_dbg_qp(qp, "no kernel post_recv for user mapped rq\n");
1018 		*bad_wr = wr;
1019 		return -EINVAL;
1020 	}
1021 
1022 	/*
1023 	 * Try to acquire QP state lock. Must be non-blocking
1024 	 * to accommodate kernel clients needs.
1025 	 */
1026 	if (!down_read_trylock(&qp->state_lock)) {
1027 		if (qp->attrs.state == SIW_QP_STATE_ERROR) {
1028 			/*
1029 			 * ERROR state is final, so we can be sure
1030 			 * this state will not change as long as the QP
1031 			 * exists.
1032 			 *
1033 			 * This handles an ib_drain_rq() call with
1034 			 * a concurrent request to set the QP state
1035 			 * to ERROR.
1036 			 */
1037 			rv = siw_rq_flush_wr(qp, wr, bad_wr);
1038 		} else {
1039 			siw_dbg_qp(qp, "QP locked, state %d\n",
1040 				   qp->attrs.state);
1041 			*bad_wr = wr;
1042 			rv = -ENOTCONN;
1043 		}
1044 		return rv;
1045 	}
1046 	if (qp->attrs.state > SIW_QP_STATE_RTS) {
1047 		if (qp->attrs.state == SIW_QP_STATE_ERROR) {
1048 			/*
1049 			 * Immediately flush this WR to CQ, if QP
1050 			 * is in ERROR state. RQ is guaranteed to
1051 			 * be empty, so WR complets in-order.
1052 			 *
1053 			 * Typically triggered by ib_drain_rq().
1054 			 */
1055 			rv = siw_rq_flush_wr(qp, wr, bad_wr);
1056 		} else {
1057 			siw_dbg_qp(qp, "QP out of state %d\n",
1058 				   qp->attrs.state);
1059 			*bad_wr = wr;
1060 			rv = -ENOTCONN;
1061 		}
1062 		up_read(&qp->state_lock);
1063 		return rv;
1064 	}
1065 	/*
1066 	 * Serialize potentially multiple producers.
1067 	 * Not needed for single threaded consumer side.
1068 	 */
1069 	spin_lock_irqsave(&qp->rq_lock, flags);
1070 
1071 	while (wr) {
1072 		u32 idx = qp->rq_put % qp->attrs.rq_size;
1073 		struct siw_rqe *rqe = &qp->recvq[idx];
1074 
1075 		if (rqe->flags) {
1076 			siw_dbg_qp(qp, "RQ full\n");
1077 			rv = -ENOMEM;
1078 			break;
1079 		}
1080 		if (wr->num_sge > qp->attrs.rq_max_sges) {
1081 			siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge);
1082 			rv = -EINVAL;
1083 			break;
1084 		}
1085 		rqe->id = wr->wr_id;
1086 		rqe->num_sge = wr->num_sge;
1087 		siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
1088 
1089 		/* make sure RQE is completely written before valid */
1090 		smp_wmb();
1091 
1092 		rqe->flags = SIW_WQE_VALID;
1093 
1094 		qp->rq_put++;
1095 		wr = wr->next;
1096 	}
1097 	spin_unlock_irqrestore(&qp->rq_lock, flags);
1098 
1099 	up_read(&qp->state_lock);
1100 
1101 	if (rv < 0) {
1102 		siw_dbg_qp(qp, "error %d\n", rv);
1103 		*bad_wr = wr;
1104 	}
1105 	return rv;
1106 }
1107 
siw_destroy_cq(struct ib_cq * base_cq,struct ib_udata * udata)1108 int siw_destroy_cq(struct ib_cq *base_cq, struct ib_udata *udata)
1109 {
1110 	struct siw_cq *cq = to_siw_cq(base_cq);
1111 	struct siw_device *sdev = to_siw_dev(base_cq->device);
1112 	struct siw_ucontext *ctx =
1113 		rdma_udata_to_drv_context(udata, struct siw_ucontext,
1114 					  base_ucontext);
1115 
1116 	siw_dbg_cq(cq, "free CQ resources\n");
1117 
1118 	siw_cq_flush(cq);
1119 
1120 	if (ctx)
1121 		rdma_user_mmap_entry_remove(cq->cq_entry);
1122 
1123 	atomic_dec(&sdev->num_cq);
1124 
1125 	vfree(cq->queue);
1126 	return 0;
1127 }
1128 
1129 /*
1130  * siw_create_cq()
1131  *
1132  * Populate CQ of requested size
1133  *
1134  * @base_cq: CQ as allocated by RDMA midlayer
1135  * @attr: Initial CQ attributes
1136  * @attrs: uverbs bundle
1137  */
1138 
siw_create_cq(struct ib_cq * base_cq,const struct ib_cq_init_attr * attr,struct uverbs_attr_bundle * attrs)1139 int siw_create_cq(struct ib_cq *base_cq, const struct ib_cq_init_attr *attr,
1140 		  struct uverbs_attr_bundle *attrs)
1141 {
1142 	struct ib_udata *udata = &attrs->driver_udata;
1143 	struct siw_device *sdev = to_siw_dev(base_cq->device);
1144 	struct siw_cq *cq = to_siw_cq(base_cq);
1145 	int rv, size = attr->cqe;
1146 
1147 	if (attr->flags)
1148 		return -EOPNOTSUPP;
1149 
1150 	if (atomic_inc_return(&sdev->num_cq) > SIW_MAX_CQ) {
1151 		siw_dbg(base_cq->device, "too many CQ's\n");
1152 		rv = -ENOMEM;
1153 		goto err_out;
1154 	}
1155 	if (size < 1 || size > sdev->attrs.max_cqe) {
1156 		siw_dbg(base_cq->device, "CQ size error: %d\n", size);
1157 		rv = -EINVAL;
1158 		goto err_out;
1159 	}
1160 	size = roundup_pow_of_two(size);
1161 	cq->base_cq.cqe = size;
1162 	cq->num_cqe = size;
1163 
1164 	if (udata)
1165 		cq->queue = vmalloc_user(size * sizeof(struct siw_cqe) +
1166 					 sizeof(struct siw_cq_ctrl));
1167 	else
1168 		cq->queue = vzalloc(size * sizeof(struct siw_cqe) +
1169 				    sizeof(struct siw_cq_ctrl));
1170 
1171 	if (cq->queue == NULL) {
1172 		rv = -ENOMEM;
1173 		goto err_out;
1174 	}
1175 	get_random_bytes(&cq->id, 4);
1176 	siw_dbg(base_cq->device, "new CQ [%u]\n", cq->id);
1177 
1178 	spin_lock_init(&cq->lock);
1179 
1180 	cq->notify = (struct siw_cq_ctrl *)&cq->queue[size];
1181 
1182 	if (udata) {
1183 		struct siw_uresp_create_cq uresp = {};
1184 		struct siw_ucontext *ctx =
1185 			rdma_udata_to_drv_context(udata, struct siw_ucontext,
1186 						  base_ucontext);
1187 		size_t length = size * sizeof(struct siw_cqe) +
1188 			sizeof(struct siw_cq_ctrl);
1189 
1190 		cq->cq_entry =
1191 			siw_mmap_entry_insert(ctx, cq->queue,
1192 					      length, &uresp.cq_key);
1193 		if (!cq->cq_entry) {
1194 			rv = -ENOMEM;
1195 			goto err_out;
1196 		}
1197 
1198 		uresp.cq_id = cq->id;
1199 		uresp.num_cqe = size;
1200 
1201 		if (udata->outlen < sizeof(uresp)) {
1202 			rv = -EINVAL;
1203 			goto err_out;
1204 		}
1205 		rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp));
1206 		if (rv)
1207 			goto err_out;
1208 	}
1209 	return 0;
1210 
1211 err_out:
1212 	siw_dbg(base_cq->device, "CQ creation failed: %d", rv);
1213 
1214 	if (cq->queue) {
1215 		struct siw_ucontext *ctx =
1216 			rdma_udata_to_drv_context(udata, struct siw_ucontext,
1217 						  base_ucontext);
1218 		if (ctx)
1219 			rdma_user_mmap_entry_remove(cq->cq_entry);
1220 		vfree(cq->queue);
1221 	}
1222 	atomic_dec(&sdev->num_cq);
1223 
1224 	return rv;
1225 }
1226 
1227 /*
1228  * siw_poll_cq()
1229  *
1230  * Reap CQ entries if available and copy work completion status into
1231  * array of WC's provided by caller. Returns number of reaped CQE's.
1232  *
1233  * @base_cq:	Base CQ contained in siw CQ.
1234  * @num_cqe:	Maximum number of CQE's to reap.
1235  * @wc:		Array of work completions to be filled by siw.
1236  */
siw_poll_cq(struct ib_cq * base_cq,int num_cqe,struct ib_wc * wc)1237 int siw_poll_cq(struct ib_cq *base_cq, int num_cqe, struct ib_wc *wc)
1238 {
1239 	struct siw_cq *cq = to_siw_cq(base_cq);
1240 	int i;
1241 
1242 	for (i = 0; i < num_cqe; i++) {
1243 		if (!siw_reap_cqe(cq, wc))
1244 			break;
1245 		wc++;
1246 	}
1247 	return i;
1248 }
1249 
1250 /*
1251  * siw_req_notify_cq()
1252  *
1253  * Request notification for new CQE's added to that CQ.
1254  * Defined flags:
1255  * o SIW_CQ_NOTIFY_SOLICITED lets siw trigger a notification
1256  *   event if a WQE with notification flag set enters the CQ
1257  * o SIW_CQ_NOTIFY_NEXT_COMP lets siw trigger a notification
1258  *   event if a WQE enters the CQ.
1259  * o IB_CQ_REPORT_MISSED_EVENTS: return value will provide the
1260  *   number of not reaped CQE's regardless of its notification
1261  *   type and current or new CQ notification settings.
1262  *
1263  * @base_cq:	Base CQ contained in siw CQ.
1264  * @flags:	Requested notification flags.
1265  */
siw_req_notify_cq(struct ib_cq * base_cq,enum ib_cq_notify_flags flags)1266 int siw_req_notify_cq(struct ib_cq *base_cq, enum ib_cq_notify_flags flags)
1267 {
1268 	struct siw_cq *cq = to_siw_cq(base_cq);
1269 
1270 	siw_dbg_cq(cq, "flags: 0x%02x\n", flags);
1271 
1272 	if ((flags & IB_CQ_SOLICITED_MASK) == IB_CQ_SOLICITED)
1273 		/*
1274 		 * Enable CQ event for next solicited completion.
1275 		 * and make it visible to all associated producers.
1276 		 */
1277 		smp_store_mb(cq->notify->flags, SIW_NOTIFY_SOLICITED);
1278 	else
1279 		/*
1280 		 * Enable CQ event for any signalled completion.
1281 		 * and make it visible to all associated producers.
1282 		 */
1283 		smp_store_mb(cq->notify->flags, SIW_NOTIFY_ALL);
1284 
1285 	if (flags & IB_CQ_REPORT_MISSED_EVENTS)
1286 		return cq->cq_put - cq->cq_get;
1287 
1288 	return 0;
1289 }
1290 
1291 /*
1292  * siw_dereg_mr()
1293  *
1294  * Release Memory Region.
1295  *
1296  * @base_mr: Base MR contained in siw MR.
1297  * @udata: points to user context, unused.
1298  */
siw_dereg_mr(struct ib_mr * base_mr,struct ib_udata * udata)1299 int siw_dereg_mr(struct ib_mr *base_mr, struct ib_udata *udata)
1300 {
1301 	struct siw_mr *mr = to_siw_mr(base_mr);
1302 	struct siw_device *sdev = to_siw_dev(base_mr->device);
1303 
1304 	siw_dbg_mem(mr->mem, "deregister MR\n");
1305 
1306 	atomic_dec(&sdev->num_mr);
1307 
1308 	siw_mr_drop_mem(mr);
1309 	kfree_rcu(mr, rcu);
1310 
1311 	return 0;
1312 }
1313 
1314 /*
1315  * siw_reg_user_mr()
1316  *
1317  * Register Memory Region.
1318  *
1319  * @pd:		Protection Domain
1320  * @start:	starting address of MR (virtual address)
1321  * @len:	len of MR
1322  * @rnic_va:	not used by siw
1323  * @rights:	MR access rights
1324  * @udata:	user buffer to communicate STag and Key.
1325  */
siw_reg_user_mr(struct ib_pd * pd,u64 start,u64 len,u64 rnic_va,int rights,struct ib_udata * udata)1326 struct ib_mr *siw_reg_user_mr(struct ib_pd *pd, u64 start, u64 len,
1327 			      u64 rnic_va, int rights, struct ib_udata *udata)
1328 {
1329 	struct siw_mr *mr = NULL;
1330 	struct siw_umem *umem = NULL;
1331 	struct siw_ureq_reg_mr ureq;
1332 	struct siw_device *sdev = to_siw_dev(pd->device);
1333 	int rv;
1334 
1335 	siw_dbg_pd(pd, "start: 0x%p, va: 0x%p, len: %llu\n",
1336 		   (void *)(uintptr_t)start, (void *)(uintptr_t)rnic_va,
1337 		   (unsigned long long)len);
1338 
1339 	if (atomic_inc_return(&sdev->num_mr) > SIW_MAX_MR) {
1340 		siw_dbg_pd(pd, "too many mr's\n");
1341 		rv = -ENOMEM;
1342 		goto err_out;
1343 	}
1344 	if (!len) {
1345 		rv = -EINVAL;
1346 		goto err_out;
1347 	}
1348 	umem = siw_umem_get(pd->device, start, len, rights);
1349 	if (IS_ERR(umem)) {
1350 		rv = PTR_ERR(umem);
1351 		siw_dbg_pd(pd, "getting user memory failed: %d\n", rv);
1352 		umem = NULL;
1353 		goto err_out;
1354 	}
1355 	mr = kzalloc(sizeof(*mr), GFP_KERNEL);
1356 	if (!mr) {
1357 		rv = -ENOMEM;
1358 		goto err_out;
1359 	}
1360 	rv = siw_mr_add_mem(mr, pd, umem, start, len, rights);
1361 	if (rv)
1362 		goto err_out;
1363 
1364 	if (udata) {
1365 		struct siw_uresp_reg_mr uresp = {};
1366 		struct siw_mem *mem = mr->mem;
1367 
1368 		if (udata->inlen < sizeof(ureq)) {
1369 			rv = -EINVAL;
1370 			goto err_out;
1371 		}
1372 		rv = ib_copy_from_udata(&ureq, udata, sizeof(ureq));
1373 		if (rv)
1374 			goto err_out;
1375 
1376 		mr->base_mr.lkey |= ureq.stag_key;
1377 		mr->base_mr.rkey |= ureq.stag_key;
1378 		mem->stag |= ureq.stag_key;
1379 		uresp.stag = mem->stag;
1380 
1381 		if (udata->outlen < sizeof(uresp)) {
1382 			rv = -EINVAL;
1383 			goto err_out;
1384 		}
1385 		rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp));
1386 		if (rv)
1387 			goto err_out;
1388 	}
1389 	mr->mem->stag_valid = 1;
1390 
1391 	return &mr->base_mr;
1392 
1393 err_out:
1394 	atomic_dec(&sdev->num_mr);
1395 	if (mr) {
1396 		if (mr->mem)
1397 			siw_mr_drop_mem(mr);
1398 		kfree_rcu(mr, rcu);
1399 	} else {
1400 		if (umem)
1401 			siw_umem_release(umem);
1402 	}
1403 	return ERR_PTR(rv);
1404 }
1405 
siw_alloc_mr(struct ib_pd * pd,enum ib_mr_type mr_type,u32 max_sge)1406 struct ib_mr *siw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
1407 			   u32 max_sge)
1408 {
1409 	struct siw_device *sdev = to_siw_dev(pd->device);
1410 	struct siw_mr *mr = NULL;
1411 	struct siw_pbl *pbl = NULL;
1412 	int rv;
1413 
1414 	if (atomic_inc_return(&sdev->num_mr) > SIW_MAX_MR) {
1415 		siw_dbg_pd(pd, "too many mr's\n");
1416 		rv = -ENOMEM;
1417 		goto err_out;
1418 	}
1419 	if (mr_type != IB_MR_TYPE_MEM_REG) {
1420 		siw_dbg_pd(pd, "mr type %d unsupported\n", mr_type);
1421 		rv = -EOPNOTSUPP;
1422 		goto err_out;
1423 	}
1424 	if (max_sge > SIW_MAX_SGE_PBL) {
1425 		siw_dbg_pd(pd, "too many sge's: %d\n", max_sge);
1426 		rv = -ENOMEM;
1427 		goto err_out;
1428 	}
1429 	pbl = siw_pbl_alloc(max_sge);
1430 	if (IS_ERR(pbl)) {
1431 		rv = PTR_ERR(pbl);
1432 		siw_dbg_pd(pd, "pbl allocation failed: %d\n", rv);
1433 		pbl = NULL;
1434 		goto err_out;
1435 	}
1436 	mr = kzalloc(sizeof(*mr), GFP_KERNEL);
1437 	if (!mr) {
1438 		rv = -ENOMEM;
1439 		goto err_out;
1440 	}
1441 	rv = siw_mr_add_mem(mr, pd, pbl, 0, max_sge * PAGE_SIZE, 0);
1442 	if (rv)
1443 		goto err_out;
1444 
1445 	mr->mem->is_pbl = 1;
1446 
1447 	siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag);
1448 
1449 	return &mr->base_mr;
1450 
1451 err_out:
1452 	atomic_dec(&sdev->num_mr);
1453 
1454 	if (!mr) {
1455 		kfree(pbl);
1456 	} else {
1457 		if (mr->mem)
1458 			siw_mr_drop_mem(mr);
1459 		kfree_rcu(mr, rcu);
1460 	}
1461 	siw_dbg_pd(pd, "failed: %d\n", rv);
1462 
1463 	return ERR_PTR(rv);
1464 }
1465 
1466 /* Just used to count number of pages being mapped */
siw_set_pbl_page(struct ib_mr * base_mr,u64 buf_addr)1467 static int siw_set_pbl_page(struct ib_mr *base_mr, u64 buf_addr)
1468 {
1469 	return 0;
1470 }
1471 
siw_map_mr_sg(struct ib_mr * base_mr,struct scatterlist * sl,int num_sle,unsigned int * sg_off)1472 int siw_map_mr_sg(struct ib_mr *base_mr, struct scatterlist *sl, int num_sle,
1473 		  unsigned int *sg_off)
1474 {
1475 	struct scatterlist *slp;
1476 	struct siw_mr *mr = to_siw_mr(base_mr);
1477 	struct siw_mem *mem = mr->mem;
1478 	struct siw_pbl *pbl = mem->pbl;
1479 	struct siw_pble *pble;
1480 	unsigned long pbl_size;
1481 	int i, rv;
1482 
1483 	if (!pbl) {
1484 		siw_dbg_mem(mem, "no PBL allocated\n");
1485 		return -EINVAL;
1486 	}
1487 	pble = pbl->pbe;
1488 
1489 	if (pbl->max_buf < num_sle) {
1490 		siw_dbg_mem(mem, "too many SGE's: %d > %d\n",
1491 			    num_sle, pbl->max_buf);
1492 		return -ENOMEM;
1493 	}
1494 	for_each_sg(sl, slp, num_sle, i) {
1495 		if (sg_dma_len(slp) == 0) {
1496 			siw_dbg_mem(mem, "empty SGE\n");
1497 			return -EINVAL;
1498 		}
1499 		if (i == 0) {
1500 			pble->addr = sg_dma_address(slp);
1501 			pble->size = sg_dma_len(slp);
1502 			pble->pbl_off = 0;
1503 			pbl_size = pble->size;
1504 			pbl->num_buf = 1;
1505 		} else {
1506 			/* Merge PBL entries if adjacent */
1507 			if (pble->addr + pble->size == sg_dma_address(slp)) {
1508 				pble->size += sg_dma_len(slp);
1509 			} else {
1510 				pble++;
1511 				pbl->num_buf++;
1512 				pble->addr = sg_dma_address(slp);
1513 				pble->size = sg_dma_len(slp);
1514 				pble->pbl_off = pbl_size;
1515 			}
1516 			pbl_size += sg_dma_len(slp);
1517 		}
1518 		siw_dbg_mem(mem,
1519 			"sge[%d], size %u, addr 0x%p, total %lu\n",
1520 			i, pble->size, ib_virt_dma_to_ptr(pble->addr),
1521 			pbl_size);
1522 	}
1523 	rv = ib_sg_to_pages(base_mr, sl, num_sle, sg_off, siw_set_pbl_page);
1524 	if (rv > 0) {
1525 		mem->len = base_mr->length;
1526 		mem->va = base_mr->iova;
1527 		siw_dbg_mem(mem,
1528 			"%llu bytes, start 0x%p, %u SLE to %u entries\n",
1529 			mem->len, (void *)(uintptr_t)mem->va, num_sle,
1530 			pbl->num_buf);
1531 	}
1532 	return rv;
1533 }
1534 
1535 /*
1536  * siw_get_dma_mr()
1537  *
1538  * Create a (empty) DMA memory region, where no umem is attached.
1539  */
siw_get_dma_mr(struct ib_pd * pd,int rights)1540 struct ib_mr *siw_get_dma_mr(struct ib_pd *pd, int rights)
1541 {
1542 	struct siw_device *sdev = to_siw_dev(pd->device);
1543 	struct siw_mr *mr = NULL;
1544 	int rv;
1545 
1546 	if (atomic_inc_return(&sdev->num_mr) > SIW_MAX_MR) {
1547 		siw_dbg_pd(pd, "too many mr's\n");
1548 		rv = -ENOMEM;
1549 		goto err_out;
1550 	}
1551 	mr = kzalloc(sizeof(*mr), GFP_KERNEL);
1552 	if (!mr) {
1553 		rv = -ENOMEM;
1554 		goto err_out;
1555 	}
1556 	rv = siw_mr_add_mem(mr, pd, NULL, 0, ULONG_MAX, rights);
1557 	if (rv)
1558 		goto err_out;
1559 
1560 	mr->mem->stag_valid = 1;
1561 
1562 	siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag);
1563 
1564 	return &mr->base_mr;
1565 
1566 err_out:
1567 	if (rv)
1568 		kfree(mr);
1569 
1570 	atomic_dec(&sdev->num_mr);
1571 
1572 	return ERR_PTR(rv);
1573 }
1574 
1575 /*
1576  * siw_create_srq()
1577  *
1578  * Create Shared Receive Queue of attributes @init_attrs
1579  * within protection domain given by @pd.
1580  *
1581  * @base_srq:	Base SRQ contained in siw SRQ.
1582  * @init_attrs:	SRQ init attributes.
1583  * @udata:	points to user context
1584  */
siw_create_srq(struct ib_srq * base_srq,struct ib_srq_init_attr * init_attrs,struct ib_udata * udata)1585 int siw_create_srq(struct ib_srq *base_srq,
1586 		   struct ib_srq_init_attr *init_attrs, struct ib_udata *udata)
1587 {
1588 	struct siw_srq *srq = to_siw_srq(base_srq);
1589 	struct ib_srq_attr *attrs = &init_attrs->attr;
1590 	struct siw_device *sdev = to_siw_dev(base_srq->device);
1591 	struct siw_ucontext *ctx =
1592 		rdma_udata_to_drv_context(udata, struct siw_ucontext,
1593 					  base_ucontext);
1594 	int rv;
1595 
1596 	if (init_attrs->srq_type != IB_SRQT_BASIC)
1597 		return -EOPNOTSUPP;
1598 
1599 	if (atomic_inc_return(&sdev->num_srq) > SIW_MAX_SRQ) {
1600 		siw_dbg_pd(base_srq->pd, "too many SRQ's\n");
1601 		rv = -ENOMEM;
1602 		goto err_out;
1603 	}
1604 	if (attrs->max_wr == 0 || attrs->max_wr > SIW_MAX_SRQ_WR ||
1605 	    attrs->max_sge > SIW_MAX_SGE || attrs->srq_limit > attrs->max_wr) {
1606 		rv = -EINVAL;
1607 		goto err_out;
1608 	}
1609 	srq->max_sge = attrs->max_sge;
1610 	srq->num_rqe = roundup_pow_of_two(attrs->max_wr);
1611 	srq->limit = attrs->srq_limit;
1612 	if (srq->limit)
1613 		srq->armed = true;
1614 
1615 	srq->is_kernel_res = !udata;
1616 
1617 	if (udata)
1618 		srq->recvq =
1619 			vmalloc_user(srq->num_rqe * sizeof(struct siw_rqe));
1620 	else
1621 		srq->recvq = vcalloc(srq->num_rqe, sizeof(struct siw_rqe));
1622 
1623 	if (srq->recvq == NULL) {
1624 		rv = -ENOMEM;
1625 		goto err_out;
1626 	}
1627 	if (udata) {
1628 		struct siw_uresp_create_srq uresp = {};
1629 		size_t length = srq->num_rqe * sizeof(struct siw_rqe);
1630 
1631 		srq->srq_entry =
1632 			siw_mmap_entry_insert(ctx, srq->recvq,
1633 					      length, &uresp.srq_key);
1634 		if (!srq->srq_entry) {
1635 			rv = -ENOMEM;
1636 			goto err_out;
1637 		}
1638 
1639 		uresp.num_rqe = srq->num_rqe;
1640 
1641 		if (udata->outlen < sizeof(uresp)) {
1642 			rv = -EINVAL;
1643 			goto err_out;
1644 		}
1645 		rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp));
1646 		if (rv)
1647 			goto err_out;
1648 	}
1649 	spin_lock_init(&srq->lock);
1650 
1651 	siw_dbg_pd(base_srq->pd, "[SRQ]: success\n");
1652 
1653 	return 0;
1654 
1655 err_out:
1656 	if (srq->recvq) {
1657 		if (ctx)
1658 			rdma_user_mmap_entry_remove(srq->srq_entry);
1659 		vfree(srq->recvq);
1660 	}
1661 	atomic_dec(&sdev->num_srq);
1662 
1663 	return rv;
1664 }
1665 
1666 /*
1667  * siw_modify_srq()
1668  *
1669  * Modify SRQ. The caller may resize SRQ and/or set/reset notification
1670  * limit and (re)arm IB_EVENT_SRQ_LIMIT_REACHED notification.
1671  *
1672  * NOTE: it is unclear if RDMA core allows for changing the MAX_SGE
1673  * parameter. siw_modify_srq() does not check the attrs->max_sge param.
1674  */
siw_modify_srq(struct ib_srq * base_srq,struct ib_srq_attr * attrs,enum ib_srq_attr_mask attr_mask,struct ib_udata * udata)1675 int siw_modify_srq(struct ib_srq *base_srq, struct ib_srq_attr *attrs,
1676 		   enum ib_srq_attr_mask attr_mask, struct ib_udata *udata)
1677 {
1678 	struct siw_srq *srq = to_siw_srq(base_srq);
1679 	unsigned long flags;
1680 	int rv = 0;
1681 
1682 	spin_lock_irqsave(&srq->lock, flags);
1683 
1684 	if (attr_mask & IB_SRQ_MAX_WR) {
1685 		/* resize request not yet supported */
1686 		rv = -EOPNOTSUPP;
1687 		goto out;
1688 	}
1689 	if (attr_mask & IB_SRQ_LIMIT) {
1690 		if (attrs->srq_limit) {
1691 			if (unlikely(attrs->srq_limit > srq->num_rqe)) {
1692 				rv = -EINVAL;
1693 				goto out;
1694 			}
1695 			srq->armed = true;
1696 		} else {
1697 			srq->armed = false;
1698 		}
1699 		srq->limit = attrs->srq_limit;
1700 	}
1701 out:
1702 	spin_unlock_irqrestore(&srq->lock, flags);
1703 
1704 	return rv;
1705 }
1706 
1707 /*
1708  * siw_query_srq()
1709  *
1710  * Query SRQ attributes.
1711  */
siw_query_srq(struct ib_srq * base_srq,struct ib_srq_attr * attrs)1712 int siw_query_srq(struct ib_srq *base_srq, struct ib_srq_attr *attrs)
1713 {
1714 	struct siw_srq *srq = to_siw_srq(base_srq);
1715 	unsigned long flags;
1716 
1717 	spin_lock_irqsave(&srq->lock, flags);
1718 
1719 	attrs->max_wr = srq->num_rqe;
1720 	attrs->max_sge = srq->max_sge;
1721 	attrs->srq_limit = srq->limit;
1722 
1723 	spin_unlock_irqrestore(&srq->lock, flags);
1724 
1725 	return 0;
1726 }
1727 
1728 /*
1729  * siw_destroy_srq()
1730  *
1731  * Destroy SRQ.
1732  * It is assumed that the SRQ is not referenced by any
1733  * QP anymore - the code trusts the RDMA core environment to keep track
1734  * of QP references.
1735  */
siw_destroy_srq(struct ib_srq * base_srq,struct ib_udata * udata)1736 int siw_destroy_srq(struct ib_srq *base_srq, struct ib_udata *udata)
1737 {
1738 	struct siw_srq *srq = to_siw_srq(base_srq);
1739 	struct siw_device *sdev = to_siw_dev(base_srq->device);
1740 	struct siw_ucontext *ctx =
1741 		rdma_udata_to_drv_context(udata, struct siw_ucontext,
1742 					  base_ucontext);
1743 
1744 	if (ctx)
1745 		rdma_user_mmap_entry_remove(srq->srq_entry);
1746 	vfree(srq->recvq);
1747 	atomic_dec(&sdev->num_srq);
1748 	return 0;
1749 }
1750 
1751 /*
1752  * siw_post_srq_recv()
1753  *
1754  * Post a list of receive queue elements to SRQ.
1755  * NOTE: The function does not check or lock a certain SRQ state
1756  *       during the post operation. The code simply trusts the
1757  *       RDMA core environment.
1758  *
1759  * @base_srq:	Base SRQ contained in siw SRQ
1760  * @wr:		List of R-WR's
1761  * @bad_wr:	Updated to failing WR if posting fails.
1762  */
siw_post_srq_recv(struct ib_srq * base_srq,const struct ib_recv_wr * wr,const struct ib_recv_wr ** bad_wr)1763 int siw_post_srq_recv(struct ib_srq *base_srq, const struct ib_recv_wr *wr,
1764 		      const struct ib_recv_wr **bad_wr)
1765 {
1766 	struct siw_srq *srq = to_siw_srq(base_srq);
1767 	unsigned long flags;
1768 	int rv = 0;
1769 
1770 	if (unlikely(!srq->is_kernel_res)) {
1771 		siw_dbg_pd(base_srq->pd,
1772 			   "[SRQ]: no kernel post_recv for mapped srq\n");
1773 		rv = -EINVAL;
1774 		goto out;
1775 	}
1776 	/*
1777 	 * Serialize potentially multiple producers.
1778 	 * Also needed to serialize potentially multiple
1779 	 * consumers.
1780 	 */
1781 	spin_lock_irqsave(&srq->lock, flags);
1782 
1783 	while (wr) {
1784 		u32 idx = srq->rq_put % srq->num_rqe;
1785 		struct siw_rqe *rqe = &srq->recvq[idx];
1786 
1787 		if (rqe->flags) {
1788 			siw_dbg_pd(base_srq->pd, "SRQ full\n");
1789 			rv = -ENOMEM;
1790 			break;
1791 		}
1792 		if (unlikely(wr->num_sge > srq->max_sge)) {
1793 			siw_dbg_pd(base_srq->pd,
1794 				   "[SRQ]: too many sge's: %d\n", wr->num_sge);
1795 			rv = -EINVAL;
1796 			break;
1797 		}
1798 		rqe->id = wr->wr_id;
1799 		rqe->num_sge = wr->num_sge;
1800 		siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
1801 
1802 		/* Make sure S-RQE is completely written before valid */
1803 		smp_wmb();
1804 
1805 		rqe->flags = SIW_WQE_VALID;
1806 
1807 		srq->rq_put++;
1808 		wr = wr->next;
1809 	}
1810 	spin_unlock_irqrestore(&srq->lock, flags);
1811 out:
1812 	if (unlikely(rv < 0)) {
1813 		siw_dbg_pd(base_srq->pd, "[SRQ]: error %d\n", rv);
1814 		*bad_wr = wr;
1815 	}
1816 	return rv;
1817 }
1818 
siw_qp_event(struct siw_qp * qp,enum ib_event_type etype)1819 void siw_qp_event(struct siw_qp *qp, enum ib_event_type etype)
1820 {
1821 	struct ib_event event;
1822 	struct ib_qp *base_qp = &qp->base_qp;
1823 
1824 	/*
1825 	 * Do not report asynchronous errors on QP which gets
1826 	 * destroyed via verbs interface (siw_destroy_qp())
1827 	 */
1828 	if (qp->attrs.flags & SIW_QP_IN_DESTROY)
1829 		return;
1830 
1831 	event.event = etype;
1832 	event.device = base_qp->device;
1833 	event.element.qp = base_qp;
1834 
1835 	if (base_qp->event_handler) {
1836 		siw_dbg_qp(qp, "reporting event %d\n", etype);
1837 		base_qp->event_handler(&event, base_qp->qp_context);
1838 	}
1839 }
1840 
siw_cq_event(struct siw_cq * cq,enum ib_event_type etype)1841 void siw_cq_event(struct siw_cq *cq, enum ib_event_type etype)
1842 {
1843 	struct ib_event event;
1844 	struct ib_cq *base_cq = &cq->base_cq;
1845 
1846 	event.event = etype;
1847 	event.device = base_cq->device;
1848 	event.element.cq = base_cq;
1849 
1850 	if (base_cq->event_handler) {
1851 		siw_dbg_cq(cq, "reporting CQ event %d\n", etype);
1852 		base_cq->event_handler(&event, base_cq->cq_context);
1853 	}
1854 }
1855 
siw_srq_event(struct siw_srq * srq,enum ib_event_type etype)1856 void siw_srq_event(struct siw_srq *srq, enum ib_event_type etype)
1857 {
1858 	struct ib_event event;
1859 	struct ib_srq *base_srq = &srq->base_srq;
1860 
1861 	event.event = etype;
1862 	event.device = base_srq->device;
1863 	event.element.srq = base_srq;
1864 
1865 	if (base_srq->event_handler) {
1866 		siw_dbg_pd(srq->base_srq.pd,
1867 			   "reporting SRQ event %d\n", etype);
1868 		base_srq->event_handler(&event, base_srq->srq_context);
1869 	}
1870 }
1871 
siw_port_event(struct siw_device * sdev,u32 port,enum ib_event_type etype)1872 void siw_port_event(struct siw_device *sdev, u32 port, enum ib_event_type etype)
1873 {
1874 	struct ib_event event;
1875 
1876 	event.event = etype;
1877 	event.device = &sdev->base_dev;
1878 	event.element.port_num = port;
1879 
1880 	siw_dbg(&sdev->base_dev, "reporting port event %d\n", etype);
1881 
1882 	ib_dispatch_event(&event);
1883 }
1884