xref: /linux/drivers/infiniband/hw/irdma/uk.c (revision c532de5a67a70f8533d495f8f2aaa9a0491c3ad0)
1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
2 /* Copyright (c) 2015 - 2021 Intel Corporation */
3 #include "osdep.h"
4 #include "defs.h"
5 #include "user.h"
6 #include "irdma.h"
7 
8 /**
9  * irdma_set_fragment - set fragment in wqe
10  * @wqe: wqe for setting fragment
11  * @offset: offset value
12  * @sge: sge length and stag
13  * @valid: The wqe valid
14  */
15 static void irdma_set_fragment(__le64 *wqe, u32 offset, struct ib_sge *sge,
16 			       u8 valid)
17 {
18 	if (sge) {
19 		set_64bit_val(wqe, offset,
20 			      FIELD_PREP(IRDMAQPSQ_FRAG_TO, sge->addr));
21 		set_64bit_val(wqe, offset + 8,
22 			      FIELD_PREP(IRDMAQPSQ_VALID, valid) |
23 			      FIELD_PREP(IRDMAQPSQ_FRAG_LEN, sge->length) |
24 			      FIELD_PREP(IRDMAQPSQ_FRAG_STAG, sge->lkey));
25 	} else {
26 		set_64bit_val(wqe, offset, 0);
27 		set_64bit_val(wqe, offset + 8,
28 			      FIELD_PREP(IRDMAQPSQ_VALID, valid));
29 	}
30 }
31 
32 /**
33  * irdma_set_fragment_gen_1 - set fragment in wqe
34  * @wqe: wqe for setting fragment
35  * @offset: offset value
36  * @sge: sge length and stag
37  * @valid: wqe valid flag
38  */
39 static void irdma_set_fragment_gen_1(__le64 *wqe, u32 offset,
40 				     struct ib_sge *sge, u8 valid)
41 {
42 	if (sge) {
43 		set_64bit_val(wqe, offset,
44 			      FIELD_PREP(IRDMAQPSQ_FRAG_TO, sge->addr));
45 		set_64bit_val(wqe, offset + 8,
46 			      FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_LEN, sge->length) |
47 			      FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_STAG, sge->lkey));
48 	} else {
49 		set_64bit_val(wqe, offset, 0);
50 		set_64bit_val(wqe, offset + 8, 0);
51 	}
52 }
53 
54 /**
55  * irdma_nop_1 - insert a NOP wqe
56  * @qp: hw qp ptr
57  */
58 static int irdma_nop_1(struct irdma_qp_uk *qp)
59 {
60 	u64 hdr;
61 	__le64 *wqe;
62 	u32 wqe_idx;
63 	bool signaled = false;
64 
65 	if (!qp->sq_ring.head)
66 		return -EINVAL;
67 
68 	wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring);
69 	wqe = qp->sq_base[wqe_idx].elem;
70 
71 	qp->sq_wrtrk_array[wqe_idx].quanta = IRDMA_QP_WQE_MIN_QUANTA;
72 
73 	set_64bit_val(wqe, 0, 0);
74 	set_64bit_val(wqe, 8, 0);
75 	set_64bit_val(wqe, 16, 0);
76 
77 	hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_NOP) |
78 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, signaled) |
79 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
80 
81 	/* make sure WQE is written before valid bit is set */
82 	dma_wmb();
83 
84 	set_64bit_val(wqe, 24, hdr);
85 
86 	return 0;
87 }
88 
89 /**
90  * irdma_clr_wqes - clear next 128 sq entries
91  * @qp: hw qp ptr
92  * @qp_wqe_idx: wqe_idx
93  */
94 void irdma_clr_wqes(struct irdma_qp_uk *qp, u32 qp_wqe_idx)
95 {
96 	struct irdma_qp_quanta *sq;
97 	u32 wqe_idx;
98 
99 	if (!(qp_wqe_idx & 0x7F)) {
100 		wqe_idx = (qp_wqe_idx + 128) % qp->sq_ring.size;
101 		sq = qp->sq_base + wqe_idx;
102 		if (wqe_idx)
103 			memset(sq, qp->swqe_polarity ? 0 : 0xFF,
104 			       128 * sizeof(*sq));
105 		else
106 			memset(sq, qp->swqe_polarity ? 0xFF : 0,
107 			       128 * sizeof(*sq));
108 	}
109 }
110 
111 /**
112  * irdma_uk_qp_post_wr - ring doorbell
113  * @qp: hw qp ptr
114  */
115 void irdma_uk_qp_post_wr(struct irdma_qp_uk *qp)
116 {
117 	u64 temp;
118 	u32 hw_sq_tail;
119 	u32 sw_sq_head;
120 
121 	/* valid bit is written and loads completed before reading shadow */
122 	mb();
123 
124 	/* read the doorbell shadow area */
125 	get_64bit_val(qp->shadow_area, 0, &temp);
126 
127 	hw_sq_tail = (u32)FIELD_GET(IRDMA_QP_DBSA_HW_SQ_TAIL, temp);
128 	sw_sq_head = IRDMA_RING_CURRENT_HEAD(qp->sq_ring);
129 	if (sw_sq_head != qp->initial_ring.head) {
130 		if (sw_sq_head != hw_sq_tail) {
131 			if (sw_sq_head > qp->initial_ring.head) {
132 				if (hw_sq_tail >= qp->initial_ring.head &&
133 				    hw_sq_tail < sw_sq_head)
134 					writel(qp->qp_id, qp->wqe_alloc_db);
135 			} else {
136 				if (hw_sq_tail >= qp->initial_ring.head ||
137 				    hw_sq_tail < sw_sq_head)
138 					writel(qp->qp_id, qp->wqe_alloc_db);
139 			}
140 		}
141 	}
142 
143 	qp->initial_ring.head = qp->sq_ring.head;
144 }
145 
146 /**
147  * irdma_qp_get_next_send_wqe - pad with NOP if needed, return where next WR should go
148  * @qp: hw qp ptr
149  * @wqe_idx: return wqe index
150  * @quanta: size of WR in quanta
151  * @total_size: size of WR in bytes
152  * @info: info on WR
153  */
154 __le64 *irdma_qp_get_next_send_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx,
155 				   u16 quanta, u32 total_size,
156 				   struct irdma_post_sq_info *info)
157 {
158 	__le64 *wqe;
159 	__le64 *wqe_0 = NULL;
160 	u16 avail_quanta;
161 	u16 i;
162 
163 	avail_quanta = qp->uk_attrs->max_hw_sq_chunk -
164 		       (IRDMA_RING_CURRENT_HEAD(qp->sq_ring) %
165 		       qp->uk_attrs->max_hw_sq_chunk);
166 	if (quanta <= avail_quanta) {
167 		/* WR fits in current chunk */
168 		if (quanta > IRDMA_SQ_RING_FREE_QUANTA(qp->sq_ring))
169 			return NULL;
170 	} else {
171 		/* Need to pad with NOP */
172 		if (quanta + avail_quanta >
173 			IRDMA_SQ_RING_FREE_QUANTA(qp->sq_ring))
174 			return NULL;
175 
176 		for (i = 0; i < avail_quanta; i++) {
177 			irdma_nop_1(qp);
178 			IRDMA_RING_MOVE_HEAD_NOCHECK(qp->sq_ring);
179 		}
180 	}
181 
182 	*wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring);
183 	if (!*wqe_idx)
184 		qp->swqe_polarity = !qp->swqe_polarity;
185 
186 	IRDMA_RING_MOVE_HEAD_BY_COUNT_NOCHECK(qp->sq_ring, quanta);
187 
188 	wqe = qp->sq_base[*wqe_idx].elem;
189 	if (qp->uk_attrs->hw_rev == IRDMA_GEN_1 && quanta == 1 &&
190 	    (IRDMA_RING_CURRENT_HEAD(qp->sq_ring) & 1)) {
191 		wqe_0 = qp->sq_base[IRDMA_RING_CURRENT_HEAD(qp->sq_ring)].elem;
192 		wqe_0[3] = cpu_to_le64(FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity ? 0 : 1));
193 	}
194 	qp->sq_wrtrk_array[*wqe_idx].wrid = info->wr_id;
195 	qp->sq_wrtrk_array[*wqe_idx].wr_len = total_size;
196 	qp->sq_wrtrk_array[*wqe_idx].quanta = quanta;
197 
198 	return wqe;
199 }
200 
201 /**
202  * irdma_qp_get_next_recv_wqe - get next qp's rcv wqe
203  * @qp: hw qp ptr
204  * @wqe_idx: return wqe index
205  */
206 __le64 *irdma_qp_get_next_recv_wqe(struct irdma_qp_uk *qp, u32 *wqe_idx)
207 {
208 	__le64 *wqe;
209 	int ret_code;
210 
211 	if (IRDMA_RING_FULL_ERR(qp->rq_ring))
212 		return NULL;
213 
214 	IRDMA_ATOMIC_RING_MOVE_HEAD(qp->rq_ring, *wqe_idx, ret_code);
215 	if (ret_code)
216 		return NULL;
217 
218 	if (!*wqe_idx)
219 		qp->rwqe_polarity = !qp->rwqe_polarity;
220 	/* rq_wqe_size_multiplier is no of 32 byte quanta in one rq wqe */
221 	wqe = qp->rq_base[*wqe_idx * qp->rq_wqe_size_multiplier].elem;
222 
223 	return wqe;
224 }
225 
226 /**
227  * irdma_uk_rdma_write - rdma write operation
228  * @qp: hw qp ptr
229  * @info: post sq information
230  * @post_sq: flag to post sq
231  */
232 int irdma_uk_rdma_write(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info,
233 			bool post_sq)
234 {
235 	u64 hdr;
236 	__le64 *wqe;
237 	struct irdma_rdma_write *op_info;
238 	u32 i, wqe_idx;
239 	u32 total_size = 0, byte_off;
240 	int ret_code;
241 	u32 frag_cnt, addl_frag_cnt;
242 	bool read_fence = false;
243 	u16 quanta;
244 
245 	op_info = &info->op.rdma_write;
246 	if (op_info->num_lo_sges > qp->max_sq_frag_cnt)
247 		return -EINVAL;
248 
249 	for (i = 0; i < op_info->num_lo_sges; i++)
250 		total_size += op_info->lo_sg_list[i].length;
251 
252 	read_fence |= info->read_fence;
253 
254 	if (info->imm_data_valid)
255 		frag_cnt = op_info->num_lo_sges + 1;
256 	else
257 		frag_cnt = op_info->num_lo_sges;
258 	addl_frag_cnt = frag_cnt > 1 ? (frag_cnt - 1) : 0;
259 	ret_code = irdma_fragcnt_to_quanta_sq(frag_cnt, &quanta);
260 	if (ret_code)
261 		return ret_code;
262 
263 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, quanta, total_size,
264 					 info);
265 	if (!wqe)
266 		return -ENOMEM;
267 
268 	irdma_clr_wqes(qp, wqe_idx);
269 
270 	set_64bit_val(wqe, 16,
271 		      FIELD_PREP(IRDMAQPSQ_FRAG_TO, op_info->rem_addr.addr));
272 
273 	if (info->imm_data_valid) {
274 		set_64bit_val(wqe, 0,
275 			      FIELD_PREP(IRDMAQPSQ_IMMDATA, info->imm_data));
276 		i = 0;
277 	} else {
278 		qp->wqe_ops.iw_set_fragment(wqe, 0,
279 					    op_info->lo_sg_list,
280 					    qp->swqe_polarity);
281 		i = 1;
282 	}
283 
284 	for (byte_off = 32; i < op_info->num_lo_sges; i++) {
285 		qp->wqe_ops.iw_set_fragment(wqe, byte_off,
286 					    &op_info->lo_sg_list[i],
287 					    qp->swqe_polarity);
288 		byte_off += 16;
289 	}
290 
291 	/* if not an odd number set valid bit in next fragment */
292 	if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) &&
293 	    frag_cnt) {
294 		qp->wqe_ops.iw_set_fragment(wqe, byte_off, NULL,
295 					    qp->swqe_polarity);
296 		if (qp->uk_attrs->hw_rev == IRDMA_GEN_2)
297 			++addl_frag_cnt;
298 	}
299 
300 	hdr = FIELD_PREP(IRDMAQPSQ_REMSTAG, op_info->rem_addr.lkey) |
301 	      FIELD_PREP(IRDMAQPSQ_OPCODE, info->op_type) |
302 	      FIELD_PREP(IRDMAQPSQ_IMMDATAFLAG, info->imm_data_valid) |
303 	      FIELD_PREP(IRDMAQPSQ_REPORTRTT, info->report_rtt) |
304 	      FIELD_PREP(IRDMAQPSQ_ADDFRAGCNT, addl_frag_cnt) |
305 	      FIELD_PREP(IRDMAQPSQ_READFENCE, read_fence) |
306 	      FIELD_PREP(IRDMAQPSQ_LOCALFENCE, info->local_fence) |
307 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
308 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
309 
310 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
311 
312 	set_64bit_val(wqe, 24, hdr);
313 
314 	if (post_sq)
315 		irdma_uk_qp_post_wr(qp);
316 
317 	return 0;
318 }
319 
320 /**
321  * irdma_uk_rdma_read - rdma read command
322  * @qp: hw qp ptr
323  * @info: post sq information
324  * @inv_stag: flag for inv_stag
325  * @post_sq: flag to post sq
326  */
327 int irdma_uk_rdma_read(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info,
328 		       bool inv_stag, bool post_sq)
329 {
330 	struct irdma_rdma_read *op_info;
331 	int ret_code;
332 	u32 i, byte_off, total_size = 0;
333 	bool local_fence = false;
334 	u32 addl_frag_cnt;
335 	__le64 *wqe;
336 	u32 wqe_idx;
337 	u16 quanta;
338 	u64 hdr;
339 
340 	op_info = &info->op.rdma_read;
341 	if (qp->max_sq_frag_cnt < op_info->num_lo_sges)
342 		return -EINVAL;
343 
344 	for (i = 0; i < op_info->num_lo_sges; i++)
345 		total_size += op_info->lo_sg_list[i].length;
346 
347 	ret_code = irdma_fragcnt_to_quanta_sq(op_info->num_lo_sges, &quanta);
348 	if (ret_code)
349 		return ret_code;
350 
351 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, quanta, total_size,
352 					 info);
353 	if (!wqe)
354 		return -ENOMEM;
355 
356 	irdma_clr_wqes(qp, wqe_idx);
357 
358 	addl_frag_cnt = op_info->num_lo_sges > 1 ?
359 			(op_info->num_lo_sges - 1) : 0;
360 	local_fence |= info->local_fence;
361 
362 	qp->wqe_ops.iw_set_fragment(wqe, 0, op_info->lo_sg_list,
363 				    qp->swqe_polarity);
364 	for (i = 1, byte_off = 32; i < op_info->num_lo_sges; ++i) {
365 		qp->wqe_ops.iw_set_fragment(wqe, byte_off,
366 					    &op_info->lo_sg_list[i],
367 					    qp->swqe_polarity);
368 		byte_off += 16;
369 	}
370 
371 	/* if not an odd number set valid bit in next fragment */
372 	if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 &&
373 	    !(op_info->num_lo_sges & 0x01) && op_info->num_lo_sges) {
374 		qp->wqe_ops.iw_set_fragment(wqe, byte_off, NULL,
375 					    qp->swqe_polarity);
376 		if (qp->uk_attrs->hw_rev == IRDMA_GEN_2)
377 			++addl_frag_cnt;
378 	}
379 	set_64bit_val(wqe, 16,
380 		      FIELD_PREP(IRDMAQPSQ_FRAG_TO, op_info->rem_addr.addr));
381 	hdr = FIELD_PREP(IRDMAQPSQ_REMSTAG, op_info->rem_addr.lkey) |
382 	      FIELD_PREP(IRDMAQPSQ_REPORTRTT, (info->report_rtt ? 1 : 0)) |
383 	      FIELD_PREP(IRDMAQPSQ_ADDFRAGCNT, addl_frag_cnt) |
384 	      FIELD_PREP(IRDMAQPSQ_OPCODE,
385 			 (inv_stag ? IRDMAQP_OP_RDMA_READ_LOC_INV : IRDMAQP_OP_RDMA_READ)) |
386 	      FIELD_PREP(IRDMAQPSQ_READFENCE, info->read_fence) |
387 	      FIELD_PREP(IRDMAQPSQ_LOCALFENCE, local_fence) |
388 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
389 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
390 
391 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
392 
393 	set_64bit_val(wqe, 24, hdr);
394 
395 	if (post_sq)
396 		irdma_uk_qp_post_wr(qp);
397 
398 	return 0;
399 }
400 
401 /**
402  * irdma_uk_send - rdma send command
403  * @qp: hw qp ptr
404  * @info: post sq information
405  * @post_sq: flag to post sq
406  */
407 int irdma_uk_send(struct irdma_qp_uk *qp, struct irdma_post_sq_info *info,
408 		  bool post_sq)
409 {
410 	__le64 *wqe;
411 	struct irdma_post_send *op_info;
412 	u64 hdr;
413 	u32 i, wqe_idx, total_size = 0, byte_off;
414 	int ret_code;
415 	u32 frag_cnt, addl_frag_cnt;
416 	bool read_fence = false;
417 	u16 quanta;
418 
419 	op_info = &info->op.send;
420 	if (qp->max_sq_frag_cnt < op_info->num_sges)
421 		return -EINVAL;
422 
423 	for (i = 0; i < op_info->num_sges; i++)
424 		total_size += op_info->sg_list[i].length;
425 
426 	if (info->imm_data_valid)
427 		frag_cnt = op_info->num_sges + 1;
428 	else
429 		frag_cnt = op_info->num_sges;
430 	ret_code = irdma_fragcnt_to_quanta_sq(frag_cnt, &quanta);
431 	if (ret_code)
432 		return ret_code;
433 
434 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, quanta, total_size,
435 					 info);
436 	if (!wqe)
437 		return -ENOMEM;
438 
439 	irdma_clr_wqes(qp, wqe_idx);
440 
441 	read_fence |= info->read_fence;
442 	addl_frag_cnt = frag_cnt > 1 ? (frag_cnt - 1) : 0;
443 	if (info->imm_data_valid) {
444 		set_64bit_val(wqe, 0,
445 			      FIELD_PREP(IRDMAQPSQ_IMMDATA, info->imm_data));
446 		i = 0;
447 	} else {
448 		qp->wqe_ops.iw_set_fragment(wqe, 0,
449 					    frag_cnt ? op_info->sg_list : NULL,
450 					    qp->swqe_polarity);
451 		i = 1;
452 	}
453 
454 	for (byte_off = 32; i < op_info->num_sges; i++) {
455 		qp->wqe_ops.iw_set_fragment(wqe, byte_off, &op_info->sg_list[i],
456 					    qp->swqe_polarity);
457 		byte_off += 16;
458 	}
459 
460 	/* if not an odd number set valid bit in next fragment */
461 	if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) &&
462 	    frag_cnt) {
463 		qp->wqe_ops.iw_set_fragment(wqe, byte_off, NULL,
464 					    qp->swqe_polarity);
465 		if (qp->uk_attrs->hw_rev == IRDMA_GEN_2)
466 			++addl_frag_cnt;
467 	}
468 
469 	set_64bit_val(wqe, 16,
470 		      FIELD_PREP(IRDMAQPSQ_DESTQKEY, op_info->qkey) |
471 		      FIELD_PREP(IRDMAQPSQ_DESTQPN, op_info->dest_qp));
472 	hdr = FIELD_PREP(IRDMAQPSQ_REMSTAG, info->stag_to_inv) |
473 	      FIELD_PREP(IRDMAQPSQ_AHID, op_info->ah_id) |
474 	      FIELD_PREP(IRDMAQPSQ_IMMDATAFLAG,
475 			 (info->imm_data_valid ? 1 : 0)) |
476 	      FIELD_PREP(IRDMAQPSQ_REPORTRTT, (info->report_rtt ? 1 : 0)) |
477 	      FIELD_PREP(IRDMAQPSQ_OPCODE, info->op_type) |
478 	      FIELD_PREP(IRDMAQPSQ_ADDFRAGCNT, addl_frag_cnt) |
479 	      FIELD_PREP(IRDMAQPSQ_READFENCE, read_fence) |
480 	      FIELD_PREP(IRDMAQPSQ_LOCALFENCE, info->local_fence) |
481 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
482 	      FIELD_PREP(IRDMAQPSQ_UDPHEADER, info->udp_hdr) |
483 	      FIELD_PREP(IRDMAQPSQ_L4LEN, info->l4len) |
484 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
485 
486 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
487 
488 	set_64bit_val(wqe, 24, hdr);
489 
490 	if (post_sq)
491 		irdma_uk_qp_post_wr(qp);
492 
493 	return 0;
494 }
495 
496 /**
497  * irdma_set_mw_bind_wqe_gen_1 - set mw bind wqe
498  * @wqe: wqe for setting fragment
499  * @op_info: info for setting bind wqe values
500  */
501 static void irdma_set_mw_bind_wqe_gen_1(__le64 *wqe,
502 					struct irdma_bind_window *op_info)
503 {
504 	set_64bit_val(wqe, 0, (uintptr_t)op_info->va);
505 	set_64bit_val(wqe, 8,
506 		      FIELD_PREP(IRDMAQPSQ_PARENTMRSTAG, op_info->mw_stag) |
507 		      FIELD_PREP(IRDMAQPSQ_MWSTAG, op_info->mr_stag));
508 	set_64bit_val(wqe, 16, op_info->bind_len);
509 }
510 
511 /**
512  * irdma_copy_inline_data_gen_1 - Copy inline data to wqe
513  * @wqe: pointer to wqe
514  * @sge_list: table of pointers to inline data
515  * @num_sges: Total inline data length
516  * @polarity: compatibility parameter
517  */
518 static void irdma_copy_inline_data_gen_1(u8 *wqe, struct ib_sge *sge_list,
519 					 u32 num_sges, u8 polarity)
520 {
521 	u32 quanta_bytes_remaining = 16;
522 	int i;
523 
524 	for (i = 0; i < num_sges; i++) {
525 		u8 *cur_sge = (u8 *)(uintptr_t)sge_list[i].addr;
526 		u32 sge_len = sge_list[i].length;
527 
528 		while (sge_len) {
529 			u32 bytes_copied;
530 
531 			bytes_copied = min(sge_len, quanta_bytes_remaining);
532 			memcpy(wqe, cur_sge, bytes_copied);
533 			wqe += bytes_copied;
534 			cur_sge += bytes_copied;
535 			quanta_bytes_remaining -= bytes_copied;
536 			sge_len -= bytes_copied;
537 
538 			if (!quanta_bytes_remaining) {
539 				/* Remaining inline bytes reside after hdr */
540 				wqe += 16;
541 				quanta_bytes_remaining = 32;
542 			}
543 		}
544 	}
545 }
546 
547 /**
548  * irdma_inline_data_size_to_quanta_gen_1 - based on inline data, quanta
549  * @data_size: data size for inline
550  *
551  * Gets the quanta based on inline and immediate data.
552  */
553 static inline u16 irdma_inline_data_size_to_quanta_gen_1(u32 data_size)
554 {
555 	return data_size <= 16 ? IRDMA_QP_WQE_MIN_QUANTA : 2;
556 }
557 
558 /**
559  * irdma_set_mw_bind_wqe - set mw bind in wqe
560  * @wqe: wqe for setting mw bind
561  * @op_info: info for setting wqe values
562  */
563 static void irdma_set_mw_bind_wqe(__le64 *wqe,
564 				  struct irdma_bind_window *op_info)
565 {
566 	set_64bit_val(wqe, 0, (uintptr_t)op_info->va);
567 	set_64bit_val(wqe, 8,
568 		      FIELD_PREP(IRDMAQPSQ_PARENTMRSTAG, op_info->mr_stag) |
569 		      FIELD_PREP(IRDMAQPSQ_MWSTAG, op_info->mw_stag));
570 	set_64bit_val(wqe, 16, op_info->bind_len);
571 }
572 
573 /**
574  * irdma_copy_inline_data - Copy inline data to wqe
575  * @wqe: pointer to wqe
576  * @sge_list: table of pointers to inline data
577  * @num_sges: number of SGE's
578  * @polarity: polarity of wqe valid bit
579  */
580 static void irdma_copy_inline_data(u8 *wqe, struct ib_sge *sge_list,
581 				   u32 num_sges, u8 polarity)
582 {
583 	u8 inline_valid = polarity << IRDMA_INLINE_VALID_S;
584 	u32 quanta_bytes_remaining = 8;
585 	bool first_quanta = true;
586 	int i;
587 
588 	wqe += 8;
589 
590 	for (i = 0; i < num_sges; i++) {
591 		u8 *cur_sge = (u8 *)(uintptr_t)sge_list[i].addr;
592 		u32 sge_len = sge_list[i].length;
593 
594 		while (sge_len) {
595 			u32 bytes_copied;
596 
597 			bytes_copied = min(sge_len, quanta_bytes_remaining);
598 			memcpy(wqe, cur_sge, bytes_copied);
599 			wqe += bytes_copied;
600 			cur_sge += bytes_copied;
601 			quanta_bytes_remaining -= bytes_copied;
602 			sge_len -= bytes_copied;
603 
604 			if (!quanta_bytes_remaining) {
605 				quanta_bytes_remaining = 31;
606 
607 				/* Remaining inline bytes reside after hdr */
608 				if (first_quanta) {
609 					first_quanta = false;
610 					wqe += 16;
611 				} else {
612 					*wqe = inline_valid;
613 					wqe++;
614 				}
615 			}
616 		}
617 	}
618 	if (!first_quanta && quanta_bytes_remaining < 31)
619 		*(wqe + quanta_bytes_remaining) = inline_valid;
620 }
621 
622 /**
623  * irdma_inline_data_size_to_quanta - based on inline data, quanta
624  * @data_size: data size for inline
625  *
626  * Gets the quanta based on inline and immediate data.
627  */
628 static u16 irdma_inline_data_size_to_quanta(u32 data_size)
629 {
630 	if (data_size <= 8)
631 		return IRDMA_QP_WQE_MIN_QUANTA;
632 	else if (data_size <= 39)
633 		return 2;
634 	else if (data_size <= 70)
635 		return 3;
636 	else if (data_size <= 101)
637 		return 4;
638 	else if (data_size <= 132)
639 		return 5;
640 	else if (data_size <= 163)
641 		return 6;
642 	else if (data_size <= 194)
643 		return 7;
644 	else
645 		return 8;
646 }
647 
648 /**
649  * irdma_uk_inline_rdma_write - inline rdma write operation
650  * @qp: hw qp ptr
651  * @info: post sq information
652  * @post_sq: flag to post sq
653  */
654 int irdma_uk_inline_rdma_write(struct irdma_qp_uk *qp,
655 			       struct irdma_post_sq_info *info, bool post_sq)
656 {
657 	__le64 *wqe;
658 	struct irdma_rdma_write *op_info;
659 	u64 hdr = 0;
660 	u32 wqe_idx;
661 	bool read_fence = false;
662 	u32 i, total_size = 0;
663 	u16 quanta;
664 
665 	op_info = &info->op.rdma_write;
666 
667 	if (unlikely(qp->max_sq_frag_cnt < op_info->num_lo_sges))
668 		return -EINVAL;
669 
670 	for (i = 0; i < op_info->num_lo_sges; i++)
671 		total_size += op_info->lo_sg_list[i].length;
672 
673 	if (unlikely(total_size > qp->max_inline_data))
674 		return -EINVAL;
675 
676 	quanta = qp->wqe_ops.iw_inline_data_size_to_quanta(total_size);
677 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, quanta, total_size,
678 					 info);
679 	if (!wqe)
680 		return -ENOMEM;
681 
682 	irdma_clr_wqes(qp, wqe_idx);
683 
684 	read_fence |= info->read_fence;
685 	set_64bit_val(wqe, 16,
686 		      FIELD_PREP(IRDMAQPSQ_FRAG_TO, op_info->rem_addr.addr));
687 
688 	hdr = FIELD_PREP(IRDMAQPSQ_REMSTAG, op_info->rem_addr.lkey) |
689 	      FIELD_PREP(IRDMAQPSQ_OPCODE, info->op_type) |
690 	      FIELD_PREP(IRDMAQPSQ_INLINEDATALEN, total_size) |
691 	      FIELD_PREP(IRDMAQPSQ_REPORTRTT, info->report_rtt ? 1 : 0) |
692 	      FIELD_PREP(IRDMAQPSQ_INLINEDATAFLAG, 1) |
693 	      FIELD_PREP(IRDMAQPSQ_IMMDATAFLAG, info->imm_data_valid ? 1 : 0) |
694 	      FIELD_PREP(IRDMAQPSQ_READFENCE, read_fence) |
695 	      FIELD_PREP(IRDMAQPSQ_LOCALFENCE, info->local_fence) |
696 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
697 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
698 
699 	if (info->imm_data_valid)
700 		set_64bit_val(wqe, 0,
701 			      FIELD_PREP(IRDMAQPSQ_IMMDATA, info->imm_data));
702 
703 	qp->wqe_ops.iw_copy_inline_data((u8 *)wqe, op_info->lo_sg_list,
704 					op_info->num_lo_sges,
705 					qp->swqe_polarity);
706 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
707 
708 	set_64bit_val(wqe, 24, hdr);
709 
710 	if (post_sq)
711 		irdma_uk_qp_post_wr(qp);
712 
713 	return 0;
714 }
715 
716 /**
717  * irdma_uk_inline_send - inline send operation
718  * @qp: hw qp ptr
719  * @info: post sq information
720  * @post_sq: flag to post sq
721  */
722 int irdma_uk_inline_send(struct irdma_qp_uk *qp,
723 			 struct irdma_post_sq_info *info, bool post_sq)
724 {
725 	__le64 *wqe;
726 	struct irdma_post_send *op_info;
727 	u64 hdr;
728 	u32 wqe_idx;
729 	bool read_fence = false;
730 	u32 i, total_size = 0;
731 	u16 quanta;
732 
733 	op_info = &info->op.send;
734 
735 	if (unlikely(qp->max_sq_frag_cnt < op_info->num_sges))
736 		return -EINVAL;
737 
738 	for (i = 0; i < op_info->num_sges; i++)
739 		total_size += op_info->sg_list[i].length;
740 
741 	if (unlikely(total_size > qp->max_inline_data))
742 		return -EINVAL;
743 
744 	quanta = qp->wqe_ops.iw_inline_data_size_to_quanta(total_size);
745 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, quanta, total_size,
746 					 info);
747 	if (!wqe)
748 		return -ENOMEM;
749 
750 	irdma_clr_wqes(qp, wqe_idx);
751 
752 	set_64bit_val(wqe, 16,
753 		      FIELD_PREP(IRDMAQPSQ_DESTQKEY, op_info->qkey) |
754 		      FIELD_PREP(IRDMAQPSQ_DESTQPN, op_info->dest_qp));
755 
756 	read_fence |= info->read_fence;
757 	hdr = FIELD_PREP(IRDMAQPSQ_REMSTAG, info->stag_to_inv) |
758 	      FIELD_PREP(IRDMAQPSQ_AHID, op_info->ah_id) |
759 	      FIELD_PREP(IRDMAQPSQ_OPCODE, info->op_type) |
760 	      FIELD_PREP(IRDMAQPSQ_INLINEDATALEN, total_size) |
761 	      FIELD_PREP(IRDMAQPSQ_IMMDATAFLAG,
762 			 (info->imm_data_valid ? 1 : 0)) |
763 	      FIELD_PREP(IRDMAQPSQ_REPORTRTT, (info->report_rtt ? 1 : 0)) |
764 	      FIELD_PREP(IRDMAQPSQ_INLINEDATAFLAG, 1) |
765 	      FIELD_PREP(IRDMAQPSQ_READFENCE, read_fence) |
766 	      FIELD_PREP(IRDMAQPSQ_LOCALFENCE, info->local_fence) |
767 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
768 	      FIELD_PREP(IRDMAQPSQ_UDPHEADER, info->udp_hdr) |
769 	      FIELD_PREP(IRDMAQPSQ_L4LEN, info->l4len) |
770 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
771 
772 	if (info->imm_data_valid)
773 		set_64bit_val(wqe, 0,
774 			      FIELD_PREP(IRDMAQPSQ_IMMDATA, info->imm_data));
775 	qp->wqe_ops.iw_copy_inline_data((u8 *)wqe, op_info->sg_list,
776 					op_info->num_sges, qp->swqe_polarity);
777 
778 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
779 
780 	set_64bit_val(wqe, 24, hdr);
781 
782 	if (post_sq)
783 		irdma_uk_qp_post_wr(qp);
784 
785 	return 0;
786 }
787 
788 /**
789  * irdma_uk_stag_local_invalidate - stag invalidate operation
790  * @qp: hw qp ptr
791  * @info: post sq information
792  * @post_sq: flag to post sq
793  */
794 int irdma_uk_stag_local_invalidate(struct irdma_qp_uk *qp,
795 				   struct irdma_post_sq_info *info,
796 				   bool post_sq)
797 {
798 	__le64 *wqe;
799 	struct irdma_inv_local_stag *op_info;
800 	u64 hdr;
801 	u32 wqe_idx;
802 	bool local_fence = false;
803 	struct ib_sge sge = {};
804 
805 	op_info = &info->op.inv_local_stag;
806 	local_fence = info->local_fence;
807 
808 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, IRDMA_QP_WQE_MIN_QUANTA,
809 					 0, info);
810 	if (!wqe)
811 		return -ENOMEM;
812 
813 	irdma_clr_wqes(qp, wqe_idx);
814 
815 	sge.lkey = op_info->target_stag;
816 	qp->wqe_ops.iw_set_fragment(wqe, 0, &sge, 0);
817 
818 	set_64bit_val(wqe, 16, 0);
819 
820 	hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMA_OP_TYPE_INV_STAG) |
821 	      FIELD_PREP(IRDMAQPSQ_READFENCE, info->read_fence) |
822 	      FIELD_PREP(IRDMAQPSQ_LOCALFENCE, local_fence) |
823 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, info->signaled) |
824 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
825 
826 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
827 
828 	set_64bit_val(wqe, 24, hdr);
829 
830 	if (post_sq)
831 		irdma_uk_qp_post_wr(qp);
832 
833 	return 0;
834 }
835 
836 /**
837  * irdma_uk_post_receive - post receive wqe
838  * @qp: hw qp ptr
839  * @info: post rq information
840  */
841 int irdma_uk_post_receive(struct irdma_qp_uk *qp,
842 			  struct irdma_post_rq_info *info)
843 {
844 	u32 wqe_idx, i, byte_off;
845 	u32 addl_frag_cnt;
846 	__le64 *wqe;
847 	u64 hdr;
848 
849 	if (qp->max_rq_frag_cnt < info->num_sges)
850 		return -EINVAL;
851 
852 	wqe = irdma_qp_get_next_recv_wqe(qp, &wqe_idx);
853 	if (!wqe)
854 		return -ENOMEM;
855 
856 	qp->rq_wrid_array[wqe_idx] = info->wr_id;
857 	addl_frag_cnt = info->num_sges > 1 ? (info->num_sges - 1) : 0;
858 	qp->wqe_ops.iw_set_fragment(wqe, 0, info->sg_list,
859 				    qp->rwqe_polarity);
860 
861 	for (i = 1, byte_off = 32; i < info->num_sges; i++) {
862 		qp->wqe_ops.iw_set_fragment(wqe, byte_off, &info->sg_list[i],
863 					    qp->rwqe_polarity);
864 		byte_off += 16;
865 	}
866 
867 	/* if not an odd number set valid bit in next fragment */
868 	if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(info->num_sges & 0x01) &&
869 	    info->num_sges) {
870 		qp->wqe_ops.iw_set_fragment(wqe, byte_off, NULL,
871 					    qp->rwqe_polarity);
872 		if (qp->uk_attrs->hw_rev == IRDMA_GEN_2)
873 			++addl_frag_cnt;
874 	}
875 
876 	set_64bit_val(wqe, 16, 0);
877 	hdr = FIELD_PREP(IRDMAQPSQ_ADDFRAGCNT, addl_frag_cnt) |
878 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->rwqe_polarity);
879 
880 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
881 
882 	set_64bit_val(wqe, 24, hdr);
883 
884 	return 0;
885 }
886 
887 /**
888  * irdma_uk_cq_resize - reset the cq buffer info
889  * @cq: cq to resize
890  * @cq_base: new cq buffer addr
891  * @cq_size: number of cqes
892  */
893 void irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size)
894 {
895 	cq->cq_base = cq_base;
896 	cq->cq_size = cq_size;
897 	IRDMA_RING_INIT(cq->cq_ring, cq->cq_size);
898 	cq->polarity = 1;
899 }
900 
901 /**
902  * irdma_uk_cq_set_resized_cnt - record the count of the resized buffers
903  * @cq: cq to resize
904  * @cq_cnt: the count of the resized cq buffers
905  */
906 void irdma_uk_cq_set_resized_cnt(struct irdma_cq_uk *cq, u16 cq_cnt)
907 {
908 	u64 temp_val;
909 	u16 sw_cq_sel;
910 	u8 arm_next_se;
911 	u8 arm_next;
912 	u8 arm_seq_num;
913 
914 	get_64bit_val(cq->shadow_area, 32, &temp_val);
915 
916 	sw_cq_sel = (u16)FIELD_GET(IRDMA_CQ_DBSA_SW_CQ_SELECT, temp_val);
917 	sw_cq_sel += cq_cnt;
918 
919 	arm_seq_num = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_SEQ_NUM, temp_val);
920 	arm_next_se = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_NEXT_SE, temp_val);
921 	arm_next = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_NEXT, temp_val);
922 
923 	temp_val = FIELD_PREP(IRDMA_CQ_DBSA_ARM_SEQ_NUM, arm_seq_num) |
924 		   FIELD_PREP(IRDMA_CQ_DBSA_SW_CQ_SELECT, sw_cq_sel) |
925 		   FIELD_PREP(IRDMA_CQ_DBSA_ARM_NEXT_SE, arm_next_se) |
926 		   FIELD_PREP(IRDMA_CQ_DBSA_ARM_NEXT, arm_next);
927 
928 	set_64bit_val(cq->shadow_area, 32, temp_val);
929 }
930 
931 /**
932  * irdma_uk_cq_request_notification - cq notification request (door bell)
933  * @cq: hw cq
934  * @cq_notify: notification type
935  */
936 void irdma_uk_cq_request_notification(struct irdma_cq_uk *cq,
937 				      enum irdma_cmpl_notify cq_notify)
938 {
939 	u64 temp_val;
940 	u16 sw_cq_sel;
941 	u8 arm_next_se = 0;
942 	u8 arm_next = 0;
943 	u8 arm_seq_num;
944 
945 	get_64bit_val(cq->shadow_area, 32, &temp_val);
946 	arm_seq_num = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_SEQ_NUM, temp_val);
947 	arm_seq_num++;
948 	sw_cq_sel = (u16)FIELD_GET(IRDMA_CQ_DBSA_SW_CQ_SELECT, temp_val);
949 	arm_next_se = (u8)FIELD_GET(IRDMA_CQ_DBSA_ARM_NEXT_SE, temp_val);
950 	arm_next_se |= 1;
951 	if (cq_notify == IRDMA_CQ_COMPL_EVENT)
952 		arm_next = 1;
953 	temp_val = FIELD_PREP(IRDMA_CQ_DBSA_ARM_SEQ_NUM, arm_seq_num) |
954 		   FIELD_PREP(IRDMA_CQ_DBSA_SW_CQ_SELECT, sw_cq_sel) |
955 		   FIELD_PREP(IRDMA_CQ_DBSA_ARM_NEXT_SE, arm_next_se) |
956 		   FIELD_PREP(IRDMA_CQ_DBSA_ARM_NEXT, arm_next);
957 
958 	set_64bit_val(cq->shadow_area, 32, temp_val);
959 
960 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
961 
962 	writel(cq->cq_id, cq->cqe_alloc_db);
963 }
964 
965 /**
966  * irdma_uk_cq_poll_cmpl - get cq completion info
967  * @cq: hw cq
968  * @info: cq poll information returned
969  */
970 int irdma_uk_cq_poll_cmpl(struct irdma_cq_uk *cq,
971 			  struct irdma_cq_poll_info *info)
972 {
973 	u64 comp_ctx, qword0, qword2, qword3;
974 	__le64 *cqe;
975 	struct irdma_qp_uk *qp;
976 	struct irdma_ring *pring = NULL;
977 	u32 wqe_idx;
978 	int ret_code;
979 	bool move_cq_head = true;
980 	u8 polarity;
981 	bool ext_valid;
982 	__le64 *ext_cqe;
983 
984 	if (cq->avoid_mem_cflct)
985 		cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(cq);
986 	else
987 		cqe = IRDMA_GET_CURRENT_CQ_ELEM(cq);
988 
989 	get_64bit_val(cqe, 24, &qword3);
990 	polarity = (u8)FIELD_GET(IRDMA_CQ_VALID, qword3);
991 	if (polarity != cq->polarity)
992 		return -ENOENT;
993 
994 	/* Ensure CQE contents are read after valid bit is checked */
995 	dma_rmb();
996 
997 	ext_valid = (bool)FIELD_GET(IRDMA_CQ_EXTCQE, qword3);
998 	if (ext_valid) {
999 		u64 qword6, qword7;
1000 		u32 peek_head;
1001 
1002 		if (cq->avoid_mem_cflct) {
1003 			ext_cqe = (__le64 *)((u8 *)cqe + 32);
1004 			get_64bit_val(ext_cqe, 24, &qword7);
1005 			polarity = (u8)FIELD_GET(IRDMA_CQ_VALID, qword7);
1006 		} else {
1007 			peek_head = (cq->cq_ring.head + 1) % cq->cq_ring.size;
1008 			ext_cqe = cq->cq_base[peek_head].buf;
1009 			get_64bit_val(ext_cqe, 24, &qword7);
1010 			polarity = (u8)FIELD_GET(IRDMA_CQ_VALID, qword7);
1011 			if (!peek_head)
1012 				polarity ^= 1;
1013 		}
1014 		if (polarity != cq->polarity)
1015 			return -ENOENT;
1016 
1017 		/* Ensure ext CQE contents are read after ext valid bit is checked */
1018 		dma_rmb();
1019 
1020 		info->imm_valid = (bool)FIELD_GET(IRDMA_CQ_IMMVALID, qword7);
1021 		if (info->imm_valid) {
1022 			u64 qword4;
1023 
1024 			get_64bit_val(ext_cqe, 0, &qword4);
1025 			info->imm_data = (u32)FIELD_GET(IRDMA_CQ_IMMDATALOW32, qword4);
1026 		}
1027 		info->ud_smac_valid = (bool)FIELD_GET(IRDMA_CQ_UDSMACVALID, qword7);
1028 		info->ud_vlan_valid = (bool)FIELD_GET(IRDMA_CQ_UDVLANVALID, qword7);
1029 		if (info->ud_smac_valid || info->ud_vlan_valid) {
1030 			get_64bit_val(ext_cqe, 16, &qword6);
1031 			if (info->ud_vlan_valid)
1032 				info->ud_vlan = (u16)FIELD_GET(IRDMA_CQ_UDVLAN, qword6);
1033 			if (info->ud_smac_valid) {
1034 				info->ud_smac[5] = qword6 & 0xFF;
1035 				info->ud_smac[4] = (qword6 >> 8) & 0xFF;
1036 				info->ud_smac[3] = (qword6 >> 16) & 0xFF;
1037 				info->ud_smac[2] = (qword6 >> 24) & 0xFF;
1038 				info->ud_smac[1] = (qword6 >> 32) & 0xFF;
1039 				info->ud_smac[0] = (qword6 >> 40) & 0xFF;
1040 			}
1041 		}
1042 	} else {
1043 		info->imm_valid = false;
1044 		info->ud_smac_valid = false;
1045 		info->ud_vlan_valid = false;
1046 	}
1047 
1048 	info->q_type = (u8)FIELD_GET(IRDMA_CQ_SQ, qword3);
1049 	info->error = (bool)FIELD_GET(IRDMA_CQ_ERROR, qword3);
1050 	info->ipv4 = (bool)FIELD_GET(IRDMACQ_IPV4, qword3);
1051 	if (info->error) {
1052 		info->major_err = FIELD_GET(IRDMA_CQ_MAJERR, qword3);
1053 		info->minor_err = FIELD_GET(IRDMA_CQ_MINERR, qword3);
1054 		if (info->major_err == IRDMA_FLUSH_MAJOR_ERR) {
1055 			info->comp_status = IRDMA_COMPL_STATUS_FLUSHED;
1056 			/* Set the min error to standard flush error code for remaining cqes */
1057 			if (info->minor_err != FLUSH_GENERAL_ERR) {
1058 				qword3 &= ~IRDMA_CQ_MINERR;
1059 				qword3 |= FIELD_PREP(IRDMA_CQ_MINERR, FLUSH_GENERAL_ERR);
1060 				set_64bit_val(cqe, 24, qword3);
1061 			}
1062 		} else {
1063 			info->comp_status = IRDMA_COMPL_STATUS_UNKNOWN;
1064 		}
1065 	} else {
1066 		info->comp_status = IRDMA_COMPL_STATUS_SUCCESS;
1067 	}
1068 
1069 	get_64bit_val(cqe, 0, &qword0);
1070 	get_64bit_val(cqe, 16, &qword2);
1071 
1072 	info->tcp_seq_num_rtt = (u32)FIELD_GET(IRDMACQ_TCPSEQNUMRTT, qword0);
1073 	info->qp_id = (u32)FIELD_GET(IRDMACQ_QPID, qword2);
1074 	info->ud_src_qpn = (u32)FIELD_GET(IRDMACQ_UDSRCQPN, qword2);
1075 
1076 	get_64bit_val(cqe, 8, &comp_ctx);
1077 
1078 	info->solicited_event = (bool)FIELD_GET(IRDMACQ_SOEVENT, qword3);
1079 	qp = (struct irdma_qp_uk *)(unsigned long)comp_ctx;
1080 	if (!qp || qp->destroy_pending) {
1081 		ret_code = -EFAULT;
1082 		goto exit;
1083 	}
1084 	wqe_idx = (u32)FIELD_GET(IRDMA_CQ_WQEIDX, qword3);
1085 	info->qp_handle = (irdma_qp_handle)(unsigned long)qp;
1086 	info->op_type = (u8)FIELD_GET(IRDMACQ_OP, qword3);
1087 
1088 	if (info->q_type == IRDMA_CQE_QTYPE_RQ) {
1089 		u32 array_idx;
1090 
1091 		array_idx = wqe_idx / qp->rq_wqe_size_multiplier;
1092 
1093 		if (info->comp_status == IRDMA_COMPL_STATUS_FLUSHED ||
1094 		    info->comp_status == IRDMA_COMPL_STATUS_UNKNOWN) {
1095 			if (!IRDMA_RING_MORE_WORK(qp->rq_ring)) {
1096 				ret_code = -ENOENT;
1097 				goto exit;
1098 			}
1099 
1100 			info->wr_id = qp->rq_wrid_array[qp->rq_ring.tail];
1101 			array_idx = qp->rq_ring.tail;
1102 		} else {
1103 			info->wr_id = qp->rq_wrid_array[array_idx];
1104 		}
1105 
1106 		info->bytes_xfered = (u32)FIELD_GET(IRDMACQ_PAYLDLEN, qword0);
1107 
1108 		if (qword3 & IRDMACQ_STAG) {
1109 			info->stag_invalid_set = true;
1110 			info->inv_stag = (u32)FIELD_GET(IRDMACQ_INVSTAG, qword2);
1111 		} else {
1112 			info->stag_invalid_set = false;
1113 		}
1114 		IRDMA_RING_SET_TAIL(qp->rq_ring, array_idx + 1);
1115 		if (info->comp_status == IRDMA_COMPL_STATUS_FLUSHED) {
1116 			qp->rq_flush_seen = true;
1117 			if (!IRDMA_RING_MORE_WORK(qp->rq_ring))
1118 				qp->rq_flush_complete = true;
1119 			else
1120 				move_cq_head = false;
1121 		}
1122 		pring = &qp->rq_ring;
1123 	} else { /* q_type is IRDMA_CQE_QTYPE_SQ */
1124 		if (qp->first_sq_wq) {
1125 			if (wqe_idx + 1 >= qp->conn_wqes)
1126 				qp->first_sq_wq = false;
1127 
1128 			if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) {
1129 				IRDMA_RING_MOVE_HEAD_NOCHECK(cq->cq_ring);
1130 				IRDMA_RING_MOVE_TAIL(cq->cq_ring);
1131 				set_64bit_val(cq->shadow_area, 0,
1132 					      IRDMA_RING_CURRENT_HEAD(cq->cq_ring));
1133 				memset(info, 0,
1134 				       sizeof(struct irdma_cq_poll_info));
1135 				return irdma_uk_cq_poll_cmpl(cq, info);
1136 			}
1137 		}
1138 		if (info->comp_status != IRDMA_COMPL_STATUS_FLUSHED) {
1139 			info->wr_id = qp->sq_wrtrk_array[wqe_idx].wrid;
1140 			if (!info->comp_status)
1141 				info->bytes_xfered = qp->sq_wrtrk_array[wqe_idx].wr_len;
1142 			info->op_type = (u8)FIELD_GET(IRDMACQ_OP, qword3);
1143 			IRDMA_RING_SET_TAIL(qp->sq_ring,
1144 					    wqe_idx + qp->sq_wrtrk_array[wqe_idx].quanta);
1145 		} else {
1146 			if (!IRDMA_RING_MORE_WORK(qp->sq_ring)) {
1147 				ret_code = -ENOENT;
1148 				goto exit;
1149 			}
1150 
1151 			do {
1152 				__le64 *sw_wqe;
1153 				u64 wqe_qword;
1154 				u32 tail;
1155 
1156 				tail = qp->sq_ring.tail;
1157 				sw_wqe = qp->sq_base[tail].elem;
1158 				get_64bit_val(sw_wqe, 24,
1159 					      &wqe_qword);
1160 				info->op_type = (u8)FIELD_GET(IRDMAQPSQ_OPCODE,
1161 							      wqe_qword);
1162 				IRDMA_RING_SET_TAIL(qp->sq_ring,
1163 						    tail + qp->sq_wrtrk_array[tail].quanta);
1164 				if (info->op_type != IRDMAQP_OP_NOP) {
1165 					info->wr_id = qp->sq_wrtrk_array[tail].wrid;
1166 					info->bytes_xfered = qp->sq_wrtrk_array[tail].wr_len;
1167 					break;
1168 				}
1169 			} while (1);
1170 			if (info->op_type == IRDMA_OP_TYPE_BIND_MW &&
1171 			    info->minor_err == FLUSH_PROT_ERR)
1172 				info->minor_err = FLUSH_MW_BIND_ERR;
1173 			qp->sq_flush_seen = true;
1174 			if (!IRDMA_RING_MORE_WORK(qp->sq_ring))
1175 				qp->sq_flush_complete = true;
1176 		}
1177 		pring = &qp->sq_ring;
1178 	}
1179 
1180 	ret_code = 0;
1181 
1182 exit:
1183 	if (!ret_code && info->comp_status == IRDMA_COMPL_STATUS_FLUSHED)
1184 		if (pring && IRDMA_RING_MORE_WORK(*pring))
1185 			move_cq_head = false;
1186 
1187 	if (move_cq_head) {
1188 		IRDMA_RING_MOVE_HEAD_NOCHECK(cq->cq_ring);
1189 		if (!IRDMA_RING_CURRENT_HEAD(cq->cq_ring))
1190 			cq->polarity ^= 1;
1191 
1192 		if (ext_valid && !cq->avoid_mem_cflct) {
1193 			IRDMA_RING_MOVE_HEAD_NOCHECK(cq->cq_ring);
1194 			if (!IRDMA_RING_CURRENT_HEAD(cq->cq_ring))
1195 				cq->polarity ^= 1;
1196 		}
1197 
1198 		IRDMA_RING_MOVE_TAIL(cq->cq_ring);
1199 		if (!cq->avoid_mem_cflct && ext_valid)
1200 			IRDMA_RING_MOVE_TAIL(cq->cq_ring);
1201 		set_64bit_val(cq->shadow_area, 0,
1202 			      IRDMA_RING_CURRENT_HEAD(cq->cq_ring));
1203 	} else {
1204 		qword3 &= ~IRDMA_CQ_WQEIDX;
1205 		qword3 |= FIELD_PREP(IRDMA_CQ_WQEIDX, pring->tail);
1206 		set_64bit_val(cqe, 24, qword3);
1207 	}
1208 
1209 	return ret_code;
1210 }
1211 
1212 /**
1213  * irdma_qp_round_up - return round up qp wq depth
1214  * @wqdepth: wq depth in quanta to round up
1215  */
1216 static int irdma_qp_round_up(u32 wqdepth)
1217 {
1218 	int scount = 1;
1219 
1220 	for (wqdepth--; scount <= 16; scount *= 2)
1221 		wqdepth |= wqdepth >> scount;
1222 
1223 	return ++wqdepth;
1224 }
1225 
1226 /**
1227  * irdma_get_wqe_shift - get shift count for maximum wqe size
1228  * @uk_attrs: qp HW attributes
1229  * @sge: Maximum Scatter Gather Elements wqe
1230  * @inline_data: Maximum inline data size
1231  * @shift: Returns the shift needed based on sge
1232  *
1233  * Shift can be used to left shift the wqe size based on number of SGEs and inlind data size.
1234  * For 1 SGE or inline data <= 8, shift = 0 (wqe size of 32
1235  * bytes). For 2 or 3 SGEs or inline data <= 39, shift = 1 (wqe
1236  * size of 64 bytes).
1237  * For 4-7 SGE's and inline <= 101 Shift of 2 otherwise (wqe
1238  * size of 256 bytes).
1239  */
1240 void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge,
1241 			 u32 inline_data, u8 *shift)
1242 {
1243 	*shift = 0;
1244 	if (uk_attrs->hw_rev >= IRDMA_GEN_2) {
1245 		if (sge > 1 || inline_data > 8) {
1246 			if (sge < 4 && inline_data <= 39)
1247 				*shift = 1;
1248 			else if (sge < 8 && inline_data <= 101)
1249 				*shift = 2;
1250 			else
1251 				*shift = 3;
1252 		}
1253 	} else if (sge > 1 || inline_data > 16) {
1254 		*shift = (sge < 4 && inline_data <= 48) ? 1 : 2;
1255 	}
1256 }
1257 
1258 /*
1259  * irdma_get_sqdepth - get SQ depth (quanta)
1260  * @uk_attrs: qp HW attributes
1261  * @sq_size: SQ size
1262  * @shift: shift which determines size of WQE
1263  * @sqdepth: depth of SQ
1264  *
1265  */
1266 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift,
1267 		      u32 *sqdepth)
1268 {
1269 	u32 min_size = (u32)uk_attrs->min_hw_wq_size << shift;
1270 
1271 	*sqdepth = irdma_qp_round_up((sq_size << shift) + IRDMA_SQ_RSVD);
1272 
1273 	if (*sqdepth < min_size)
1274 		*sqdepth = min_size;
1275 	else if (*sqdepth > uk_attrs->max_hw_wq_quanta)
1276 		return -EINVAL;
1277 
1278 	return 0;
1279 }
1280 
1281 /*
1282  * irdma_get_rqdepth - get RQ depth (quanta)
1283  * @uk_attrs: qp HW attributes
1284  * @rq_size: RQ size
1285  * @shift: shift which determines size of WQE
1286  * @rqdepth: depth of RQ
1287  */
1288 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift,
1289 		      u32 *rqdepth)
1290 {
1291 	u32 min_size = (u32)uk_attrs->min_hw_wq_size << shift;
1292 
1293 	*rqdepth = irdma_qp_round_up((rq_size << shift) + IRDMA_RQ_RSVD);
1294 
1295 	if (*rqdepth < min_size)
1296 		*rqdepth = min_size;
1297 	else if (*rqdepth > uk_attrs->max_hw_rq_quanta)
1298 		return -EINVAL;
1299 
1300 	return 0;
1301 }
1302 
1303 static const struct irdma_wqe_uk_ops iw_wqe_uk_ops = {
1304 	.iw_copy_inline_data = irdma_copy_inline_data,
1305 	.iw_inline_data_size_to_quanta = irdma_inline_data_size_to_quanta,
1306 	.iw_set_fragment = irdma_set_fragment,
1307 	.iw_set_mw_bind_wqe = irdma_set_mw_bind_wqe,
1308 };
1309 
1310 static const struct irdma_wqe_uk_ops iw_wqe_uk_ops_gen_1 = {
1311 	.iw_copy_inline_data = irdma_copy_inline_data_gen_1,
1312 	.iw_inline_data_size_to_quanta = irdma_inline_data_size_to_quanta_gen_1,
1313 	.iw_set_fragment = irdma_set_fragment_gen_1,
1314 	.iw_set_mw_bind_wqe = irdma_set_mw_bind_wqe_gen_1,
1315 };
1316 
1317 /**
1318  * irdma_setup_connection_wqes - setup WQEs necessary to complete
1319  * connection.
1320  * @qp: hw qp (user and kernel)
1321  * @info: qp initialization info
1322  */
1323 static void irdma_setup_connection_wqes(struct irdma_qp_uk *qp,
1324 					struct irdma_qp_uk_init_info *info)
1325 {
1326 	u16 move_cnt = 1;
1327 
1328 	if (!info->legacy_mode &&
1329 	    (qp->uk_attrs->feature_flags & IRDMA_FEATURE_RTS_AE))
1330 		move_cnt = 3;
1331 
1332 	qp->conn_wqes = move_cnt;
1333 	IRDMA_RING_MOVE_HEAD_BY_COUNT_NOCHECK(qp->sq_ring, move_cnt);
1334 	IRDMA_RING_MOVE_TAIL_BY_COUNT(qp->sq_ring, move_cnt);
1335 	IRDMA_RING_MOVE_HEAD_BY_COUNT_NOCHECK(qp->initial_ring, move_cnt);
1336 }
1337 
1338 /**
1339  * irdma_uk_calc_shift_wq - calculate WQE shift for both SQ and RQ
1340  * @ukinfo: qp initialization info
1341  * @sq_shift: Returns shift of SQ
1342  * @rq_shift: Returns shift of RQ
1343  */
1344 void irdma_uk_calc_shift_wq(struct irdma_qp_uk_init_info *ukinfo, u8 *sq_shift,
1345 			    u8 *rq_shift)
1346 {
1347 	bool imm_support = ukinfo->uk_attrs->hw_rev >= IRDMA_GEN_2;
1348 
1349 	irdma_get_wqe_shift(ukinfo->uk_attrs,
1350 			    imm_support ? ukinfo->max_sq_frag_cnt + 1 :
1351 					  ukinfo->max_sq_frag_cnt,
1352 			    ukinfo->max_inline_data, sq_shift);
1353 
1354 	irdma_get_wqe_shift(ukinfo->uk_attrs, ukinfo->max_rq_frag_cnt, 0,
1355 			    rq_shift);
1356 
1357 	if (ukinfo->uk_attrs->hw_rev == IRDMA_GEN_1) {
1358 		if (ukinfo->abi_ver > 4)
1359 			*rq_shift = IRDMA_MAX_RQ_WQE_SHIFT_GEN1;
1360 	}
1361 }
1362 
1363 /**
1364  * irdma_uk_calc_depth_shift_sq - calculate depth and shift for SQ size.
1365  * @ukinfo: qp initialization info
1366  * @sq_depth: Returns depth of SQ
1367  * @sq_shift: Returns shift of SQ
1368  */
1369 int irdma_uk_calc_depth_shift_sq(struct irdma_qp_uk_init_info *ukinfo,
1370 				 u32 *sq_depth, u8 *sq_shift)
1371 {
1372 	bool imm_support = ukinfo->uk_attrs->hw_rev >= IRDMA_GEN_2;
1373 	int status;
1374 
1375 	irdma_get_wqe_shift(ukinfo->uk_attrs,
1376 			    imm_support ? ukinfo->max_sq_frag_cnt + 1 :
1377 			    ukinfo->max_sq_frag_cnt,
1378 			    ukinfo->max_inline_data, sq_shift);
1379 	status = irdma_get_sqdepth(ukinfo->uk_attrs, ukinfo->sq_size,
1380 				   *sq_shift, sq_depth);
1381 
1382 	return status;
1383 }
1384 
1385 /**
1386  * irdma_uk_calc_depth_shift_rq - calculate depth and shift for RQ size.
1387  * @ukinfo: qp initialization info
1388  * @rq_depth: Returns depth of RQ
1389  * @rq_shift: Returns shift of RQ
1390  */
1391 int irdma_uk_calc_depth_shift_rq(struct irdma_qp_uk_init_info *ukinfo,
1392 				 u32 *rq_depth, u8 *rq_shift)
1393 {
1394 	int status;
1395 
1396 	irdma_get_wqe_shift(ukinfo->uk_attrs, ukinfo->max_rq_frag_cnt, 0,
1397 			    rq_shift);
1398 
1399 	if (ukinfo->uk_attrs->hw_rev == IRDMA_GEN_1) {
1400 		if (ukinfo->abi_ver > 4)
1401 			*rq_shift = IRDMA_MAX_RQ_WQE_SHIFT_GEN1;
1402 	}
1403 
1404 	status = irdma_get_rqdepth(ukinfo->uk_attrs, ukinfo->rq_size,
1405 				   *rq_shift, rq_depth);
1406 
1407 	return status;
1408 }
1409 
1410 /**
1411  * irdma_uk_qp_init - initialize shared qp
1412  * @qp: hw qp (user and kernel)
1413  * @info: qp initialization info
1414  *
1415  * initializes the vars used in both user and kernel mode.
1416  * size of the wqe depends on numbers of max. fragements
1417  * allowed. Then size of wqe * the number of wqes should be the
1418  * amount of memory allocated for sq and rq.
1419  */
1420 int irdma_uk_qp_init(struct irdma_qp_uk *qp, struct irdma_qp_uk_init_info *info)
1421 {
1422 	int ret_code = 0;
1423 	u32 sq_ring_size;
1424 
1425 	qp->uk_attrs = info->uk_attrs;
1426 	if (info->max_sq_frag_cnt > qp->uk_attrs->max_hw_wq_frags ||
1427 	    info->max_rq_frag_cnt > qp->uk_attrs->max_hw_wq_frags)
1428 		return -EINVAL;
1429 
1430 	qp->qp_caps = info->qp_caps;
1431 	qp->sq_base = info->sq;
1432 	qp->rq_base = info->rq;
1433 	qp->qp_type = info->type ? info->type : IRDMA_QP_TYPE_IWARP;
1434 	qp->shadow_area = info->shadow_area;
1435 	qp->sq_wrtrk_array = info->sq_wrtrk_array;
1436 
1437 	qp->rq_wrid_array = info->rq_wrid_array;
1438 	qp->wqe_alloc_db = info->wqe_alloc_db;
1439 	qp->qp_id = info->qp_id;
1440 	qp->sq_size = info->sq_size;
1441 	qp->max_sq_frag_cnt = info->max_sq_frag_cnt;
1442 	sq_ring_size = qp->sq_size << info->sq_shift;
1443 	IRDMA_RING_INIT(qp->sq_ring, sq_ring_size);
1444 	IRDMA_RING_INIT(qp->initial_ring, sq_ring_size);
1445 	if (info->first_sq_wq) {
1446 		irdma_setup_connection_wqes(qp, info);
1447 		qp->swqe_polarity = 1;
1448 		qp->first_sq_wq = true;
1449 	} else {
1450 		qp->swqe_polarity = 0;
1451 	}
1452 	qp->swqe_polarity_deferred = 1;
1453 	qp->rwqe_polarity = 0;
1454 	qp->rq_size = info->rq_size;
1455 	qp->max_rq_frag_cnt = info->max_rq_frag_cnt;
1456 	qp->max_inline_data = info->max_inline_data;
1457 	qp->rq_wqe_size = info->rq_shift;
1458 	IRDMA_RING_INIT(qp->rq_ring, qp->rq_size);
1459 	qp->rq_wqe_size_multiplier = 1 << info->rq_shift;
1460 	if (qp->uk_attrs->hw_rev == IRDMA_GEN_1)
1461 		qp->wqe_ops = iw_wqe_uk_ops_gen_1;
1462 	else
1463 		qp->wqe_ops = iw_wqe_uk_ops;
1464 	return ret_code;
1465 }
1466 
1467 /**
1468  * irdma_uk_cq_init - initialize shared cq (user and kernel)
1469  * @cq: hw cq
1470  * @info: hw cq initialization info
1471  */
1472 void irdma_uk_cq_init(struct irdma_cq_uk *cq,
1473 		      struct irdma_cq_uk_init_info *info)
1474 {
1475 	cq->cq_base = info->cq_base;
1476 	cq->cq_id = info->cq_id;
1477 	cq->cq_size = info->cq_size;
1478 	cq->cqe_alloc_db = info->cqe_alloc_db;
1479 	cq->cq_ack_db = info->cq_ack_db;
1480 	cq->shadow_area = info->shadow_area;
1481 	cq->avoid_mem_cflct = info->avoid_mem_cflct;
1482 	IRDMA_RING_INIT(cq->cq_ring, cq->cq_size);
1483 	cq->polarity = 1;
1484 }
1485 
1486 /**
1487  * irdma_uk_clean_cq - clean cq entries
1488  * @q: completion context
1489  * @cq: cq to clean
1490  */
1491 void irdma_uk_clean_cq(void *q, struct irdma_cq_uk *cq)
1492 {
1493 	__le64 *cqe;
1494 	u64 qword3, comp_ctx;
1495 	u32 cq_head;
1496 	u8 polarity, temp;
1497 
1498 	cq_head = cq->cq_ring.head;
1499 	temp = cq->polarity;
1500 	do {
1501 		if (cq->avoid_mem_cflct)
1502 			cqe = ((struct irdma_extended_cqe *)(cq->cq_base))[cq_head].buf;
1503 		else
1504 			cqe = cq->cq_base[cq_head].buf;
1505 		get_64bit_val(cqe, 24, &qword3);
1506 		polarity = (u8)FIELD_GET(IRDMA_CQ_VALID, qword3);
1507 
1508 		if (polarity != temp)
1509 			break;
1510 
1511 		/* Ensure CQE contents are read after valid bit is checked */
1512 		dma_rmb();
1513 
1514 		get_64bit_val(cqe, 8, &comp_ctx);
1515 		if ((void *)(unsigned long)comp_ctx == q)
1516 			set_64bit_val(cqe, 8, 0);
1517 
1518 		cq_head = (cq_head + 1) % cq->cq_ring.size;
1519 		if (!cq_head)
1520 			temp ^= 1;
1521 	} while (true);
1522 }
1523 
1524 /**
1525  * irdma_nop - post a nop
1526  * @qp: hw qp ptr
1527  * @wr_id: work request id
1528  * @signaled: signaled for completion
1529  * @post_sq: ring doorbell
1530  */
1531 int irdma_nop(struct irdma_qp_uk *qp, u64 wr_id, bool signaled, bool post_sq)
1532 {
1533 	__le64 *wqe;
1534 	u64 hdr;
1535 	u32 wqe_idx;
1536 	struct irdma_post_sq_info info = {};
1537 
1538 	info.wr_id = wr_id;
1539 	wqe = irdma_qp_get_next_send_wqe(qp, &wqe_idx, IRDMA_QP_WQE_MIN_QUANTA,
1540 					 0, &info);
1541 	if (!wqe)
1542 		return -ENOMEM;
1543 
1544 	irdma_clr_wqes(qp, wqe_idx);
1545 
1546 	set_64bit_val(wqe, 0, 0);
1547 	set_64bit_val(wqe, 8, 0);
1548 	set_64bit_val(wqe, 16, 0);
1549 
1550 	hdr = FIELD_PREP(IRDMAQPSQ_OPCODE, IRDMAQP_OP_NOP) |
1551 	      FIELD_PREP(IRDMAQPSQ_SIGCOMPL, signaled) |
1552 	      FIELD_PREP(IRDMAQPSQ_VALID, qp->swqe_polarity);
1553 
1554 	dma_wmb(); /* make sure WQE is populated before valid bit is set */
1555 
1556 	set_64bit_val(wqe, 24, hdr);
1557 	if (post_sq)
1558 		irdma_uk_qp_post_wr(qp);
1559 
1560 	return 0;
1561 }
1562 
1563 /**
1564  * irdma_fragcnt_to_quanta_sq - calculate quanta based on fragment count for SQ
1565  * @frag_cnt: number of fragments
1566  * @quanta: quanta for frag_cnt
1567  */
1568 int irdma_fragcnt_to_quanta_sq(u32 frag_cnt, u16 *quanta)
1569 {
1570 	switch (frag_cnt) {
1571 	case 0:
1572 	case 1:
1573 		*quanta = IRDMA_QP_WQE_MIN_QUANTA;
1574 		break;
1575 	case 2:
1576 	case 3:
1577 		*quanta = 2;
1578 		break;
1579 	case 4:
1580 	case 5:
1581 		*quanta = 3;
1582 		break;
1583 	case 6:
1584 	case 7:
1585 		*quanta = 4;
1586 		break;
1587 	case 8:
1588 	case 9:
1589 		*quanta = 5;
1590 		break;
1591 	case 10:
1592 	case 11:
1593 		*quanta = 6;
1594 		break;
1595 	case 12:
1596 	case 13:
1597 		*quanta = 7;
1598 		break;
1599 	case 14:
1600 	case 15: /* when immediate data is present */
1601 		*quanta = 8;
1602 		break;
1603 	default:
1604 		return -EINVAL;
1605 	}
1606 
1607 	return 0;
1608 }
1609 
1610 /**
1611  * irdma_fragcnt_to_wqesize_rq - calculate wqe size based on fragment count for RQ
1612  * @frag_cnt: number of fragments
1613  * @wqe_size: size in bytes given frag_cnt
1614  */
1615 int irdma_fragcnt_to_wqesize_rq(u32 frag_cnt, u16 *wqe_size)
1616 {
1617 	switch (frag_cnt) {
1618 	case 0:
1619 	case 1:
1620 		*wqe_size = 32;
1621 		break;
1622 	case 2:
1623 	case 3:
1624 		*wqe_size = 64;
1625 		break;
1626 	case 4:
1627 	case 5:
1628 	case 6:
1629 	case 7:
1630 		*wqe_size = 128;
1631 		break;
1632 	case 8:
1633 	case 9:
1634 	case 10:
1635 	case 11:
1636 	case 12:
1637 	case 13:
1638 	case 14:
1639 		*wqe_size = 256;
1640 		break;
1641 	default:
1642 		return -EINVAL;
1643 	}
1644 
1645 	return 0;
1646 }
1647