xref: /illumos-gate/usr/src/uts/common/io/mlxcx/mlxcx_ring.c (revision a92282e44f968185a6bba094d1e5fece2da819cf)
1 /*
2  * This file and its contents are supplied under the terms of the
3  * Common Development and Distribution License ("CDDL"), version 1.0.
4  * You may only use this file in accordance with the terms of version
5  * 1.0 of the CDDL.
6  *
7  * A full copy of the text of the CDDL should have accompanied this
8  * source.  A copy of the CDDL is also available via the Internet at
9  * http://www.illumos.org/license/CDDL.
10  */
11 
12 /*
13  * Copyright 2020, The University of Queensland
14  * Copyright (c) 2018, Joyent, Inc.
15  * Copyright 2020 RackTop Systems, Inc.
16  */
17 
18 /*
19  * Mellanox Connect-X 4/5/6 driver.
20  */
21 
22 #include <sys/modctl.h>
23 #include <sys/conf.h>
24 #include <sys/devops.h>
25 #include <sys/sysmacros.h>
26 #include <sys/atomic.h>
27 #include <sys/cpuvar.h>
28 #include <sys/sdt.h>
29 
30 #include <sys/pattr.h>
31 #include <sys/dlpi.h>
32 
33 #include <sys/mac_provider.h>
34 
35 #include <sys/random.h>
36 
37 #include <mlxcx.h>
38 
39 boolean_t
40 mlxcx_wq_alloc_dma(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq)
41 {
42 	ddi_device_acc_attr_t acc;
43 	ddi_dma_attr_t attr;
44 	boolean_t ret;
45 	size_t sz;
46 
47 	VERIFY0(mlwq->mlwq_state & MLXCX_WQ_ALLOC);
48 
49 	/* Receive and send queue entries might be different sizes. */
50 	switch (mlwq->mlwq_type) {
51 	case MLXCX_WQ_TYPE_SENDQ:
52 		mlwq->mlwq_entshift = mlxp->mlx_props.mldp_sq_size_shift;
53 		mlwq->mlwq_nents = (1 << mlwq->mlwq_entshift);
54 		sz = mlwq->mlwq_nents * sizeof (mlxcx_sendq_ent_t);
55 		break;
56 	case MLXCX_WQ_TYPE_RECVQ:
57 		mlwq->mlwq_entshift = mlxp->mlx_props.mldp_rq_size_shift;
58 		mlwq->mlwq_nents = (1 << mlwq->mlwq_entshift);
59 		sz = mlwq->mlwq_nents * sizeof (mlxcx_recvq_ent_t);
60 		break;
61 	default:
62 		VERIFY(0);
63 		return (B_FALSE);
64 	}
65 	ASSERT3U(sz & (MLXCX_HW_PAGE_SIZE - 1), ==, 0);
66 
67 	mlxcx_dma_acc_attr(mlxp, &acc);
68 	mlxcx_dma_queue_attr(mlxp, &attr);
69 
70 	ret = mlxcx_dma_alloc(mlxp, &mlwq->mlwq_dma, &attr, &acc,
71 	    B_TRUE, sz, B_TRUE);
72 	if (!ret) {
73 		mlxcx_warn(mlxp, "failed to allocate WQ memory");
74 		return (B_FALSE);
75 	}
76 
77 	/*
78 	 * Just set the first pointer in the union. Yes, this is a strict
79 	 * aliasing violation. No, I don't care.
80 	 */
81 	mlwq->mlwq_send_ent = (mlxcx_sendq_ent_t *)mlwq->mlwq_dma.mxdb_va;
82 
83 	mlxcx_dma_acc_attr(mlxp, &acc);
84 	mlxcx_dma_qdbell_attr(mlxp, &attr);
85 	sz = sizeof (mlxcx_workq_doorbell_t);
86 	ret = mlxcx_dma_alloc(mlxp, &mlwq->mlwq_doorbell_dma, &attr, &acc,
87 	    B_TRUE, sz, B_TRUE);
88 	if (!ret) {
89 		mlxcx_warn(mlxp, "failed to allocate WQ doorbell memory");
90 		mlxcx_dma_free(&mlwq->mlwq_dma);
91 		mlwq->mlwq_send_ent = NULL;
92 		return (B_FALSE);
93 	}
94 
95 	mlwq->mlwq_doorbell =
96 	    (mlxcx_workq_doorbell_t *)mlwq->mlwq_doorbell_dma.mxdb_va;
97 
98 	mlwq->mlwq_state |= MLXCX_WQ_ALLOC;
99 
100 	return (B_TRUE);
101 }
102 
103 void
104 mlxcx_wq_rele_dma(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq)
105 {
106 	VERIFY(mlwq->mlwq_state & MLXCX_WQ_ALLOC);
107 	if (mlwq->mlwq_state & MLXCX_WQ_CREATED)
108 		VERIFY(mlwq->mlwq_state & MLXCX_WQ_DESTROYED);
109 
110 	mlxcx_dma_free(&mlwq->mlwq_dma);
111 	mlwq->mlwq_send_ent = NULL;
112 	mlxcx_dma_free(&mlwq->mlwq_doorbell_dma);
113 	mlwq->mlwq_doorbell = NULL;
114 
115 	mlwq->mlwq_state &= ~MLXCX_CQ_ALLOC;
116 }
117 
118 static boolean_t
119 mlxcx_cq_alloc_dma(mlxcx_t *mlxp, mlxcx_completion_queue_t *mlcq,
120     uint_t ent_shift)
121 {
122 	ddi_device_acc_attr_t acc;
123 	ddi_dma_attr_t attr;
124 	boolean_t ret;
125 	size_t sz, i;
126 
127 	VERIFY0(mlcq->mlcq_state & MLXCX_EQ_ALLOC);
128 
129 	mlcq->mlcq_entshift = ent_shift;
130 	mlcq->mlcq_nents = (1 << mlcq->mlcq_entshift);
131 	sz = mlcq->mlcq_nents * sizeof (mlxcx_completionq_ent_t);
132 	ASSERT3U(sz & (MLXCX_HW_PAGE_SIZE - 1), ==, 0);
133 
134 	mlxcx_dma_acc_attr(mlxp, &acc);
135 	mlxcx_dma_queue_attr(mlxp, &attr);
136 
137 	ret = mlxcx_dma_alloc(mlxp, &mlcq->mlcq_dma, &attr, &acc,
138 	    B_TRUE, sz, B_TRUE);
139 	if (!ret) {
140 		mlxcx_warn(mlxp, "failed to allocate CQ memory");
141 		return (B_FALSE);
142 	}
143 
144 	mlcq->mlcq_ent = (mlxcx_completionq_ent_t *)mlcq->mlcq_dma.mxdb_va;
145 
146 	for (i = 0; i < mlcq->mlcq_nents; ++i) {
147 		mlcq->mlcq_ent[i].mlcqe_opcode = MLXCX_CQE_OP_INVALID;
148 		mlcq->mlcq_ent[i].mlcqe_owner = MLXCX_CQE_OWNER_INIT;
149 	}
150 
151 	mlxcx_dma_acc_attr(mlxp, &acc);
152 	mlxcx_dma_qdbell_attr(mlxp, &attr);
153 	sz = sizeof (mlxcx_completionq_doorbell_t);
154 	ret = mlxcx_dma_alloc(mlxp, &mlcq->mlcq_doorbell_dma, &attr, &acc,
155 	    B_TRUE, sz, B_TRUE);
156 	if (!ret) {
157 		mlxcx_warn(mlxp, "failed to allocate CQ doorbell memory");
158 		mlxcx_dma_free(&mlcq->mlcq_dma);
159 		mlcq->mlcq_ent = NULL;
160 		return (B_FALSE);
161 	}
162 
163 	mlcq->mlcq_doorbell =
164 	    (mlxcx_completionq_doorbell_t *)mlcq->mlcq_doorbell_dma.mxdb_va;
165 
166 	atomic_or_uint(&mlcq->mlcq_state, MLXCX_CQ_ALLOC);
167 
168 	return (B_TRUE);
169 }
170 
171 static void
172 mlxcx_cq_rele_dma(mlxcx_t *mlxp, mlxcx_completion_queue_t *mlcq)
173 {
174 	VERIFY(mlcq->mlcq_state & MLXCX_CQ_ALLOC);
175 	if (mlcq->mlcq_state & MLXCX_CQ_CREATED)
176 		VERIFY(mlcq->mlcq_state & MLXCX_CQ_DESTROYED);
177 
178 	mlxcx_dma_free(&mlcq->mlcq_dma);
179 	mlcq->mlcq_ent = NULL;
180 	mlxcx_dma_free(&mlcq->mlcq_doorbell_dma);
181 	mlcq->mlcq_doorbell = NULL;
182 
183 	atomic_and_uint(&mlcq->mlcq_state, ~MLXCX_CQ_ALLOC);
184 }
185 
186 void
187 mlxcx_wq_teardown(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq)
188 {
189 	mlxcx_completion_queue_t *mlcq;
190 
191 	/*
192 	 * If something is holding the lock on a long operation like a
193 	 * refill, setting this flag asks them to exit early if possible.
194 	 */
195 	atomic_or_uint(&mlwq->mlwq_state, MLXCX_WQ_TEARDOWN);
196 
197 	mutex_enter(&mlwq->mlwq_mtx);
198 
199 	list_remove(&mlxp->mlx_wqs, mlwq);
200 
201 	if ((mlwq->mlwq_state & MLXCX_WQ_CREATED) &&
202 	    !(mlwq->mlwq_state & MLXCX_WQ_DESTROYED)) {
203 		if (mlwq->mlwq_type == MLXCX_WQ_TYPE_RECVQ &&
204 		    mlwq->mlwq_state & MLXCX_WQ_STARTED &&
205 		    !mlxcx_cmd_stop_rq(mlxp, mlwq)) {
206 			mlxcx_warn(mlxp, "failed to stop "
207 			    "recv queue num %x", mlwq->mlwq_num);
208 		}
209 		if (mlwq->mlwq_type == MLXCX_WQ_TYPE_SENDQ &&
210 		    mlwq->mlwq_state & MLXCX_WQ_STARTED &&
211 		    !mlxcx_cmd_stop_sq(mlxp, mlwq)) {
212 			mlxcx_warn(mlxp, "failed to stop "
213 			    "send queue num %x", mlwq->mlwq_num);
214 		}
215 		if (mlwq->mlwq_type == MLXCX_WQ_TYPE_RECVQ &&
216 		    !mlxcx_cmd_destroy_rq(mlxp, mlwq)) {
217 			mlxcx_warn(mlxp, "failed to destroy "
218 			    "recv queue num %x", mlwq->mlwq_num);
219 		}
220 		if (mlwq->mlwq_type == MLXCX_WQ_TYPE_SENDQ &&
221 		    !mlxcx_cmd_destroy_sq(mlxp, mlwq)) {
222 			mlxcx_warn(mlxp, "failed to destroy "
223 			    "send queue num %x", mlwq->mlwq_num);
224 		}
225 	}
226 	if (mlwq->mlwq_state & MLXCX_WQ_ALLOC) {
227 		mlxcx_wq_rele_dma(mlxp, mlwq);
228 	}
229 	mlcq = mlwq->mlwq_cq;
230 
231 	/* These will be released by mlxcx_teardown_bufs() */
232 	mlwq->mlwq_bufs = NULL;
233 	mlwq->mlwq_foreign_bufs = NULL;
234 
235 	mutex_exit(&mlwq->mlwq_mtx);
236 
237 	mutex_enter(&mlcq->mlcq_mtx);
238 	mutex_enter(&mlwq->mlwq_mtx);
239 	ASSERT3P(mlcq->mlcq_wq, ==, mlwq);
240 	mlcq->mlcq_wq = NULL;
241 	mutex_exit(&mlwq->mlwq_mtx);
242 	mutex_exit(&mlcq->mlcq_mtx);
243 
244 	mutex_destroy(&mlwq->mlwq_mtx);
245 }
246 
247 void
248 mlxcx_cq_teardown(mlxcx_t *mlxp, mlxcx_completion_queue_t *mlcq)
249 {
250 	mlxcx_event_queue_t *mleq;
251 	mlxcx_buffer_t *b;
252 
253 	/*
254 	 * If something is holding the lock on a long operation like polling
255 	 * which we're going to abort anyway, this flag asks them to exit
256 	 * early if possible.
257 	 */
258 	atomic_or_uint(&mlcq->mlcq_state, MLXCX_CQ_TEARDOWN);
259 
260 	mutex_enter(&mlcq->mlcq_mtx);
261 
262 	list_remove(&mlxp->mlx_cqs, mlcq);
263 
264 	if ((mlcq->mlcq_state & MLXCX_CQ_CREATED) &&
265 	    !(mlcq->mlcq_state & MLXCX_CQ_DESTROYED)) {
266 		if (!mlxcx_cmd_destroy_cq(mlxp, mlcq)) {
267 			mlxcx_warn(mlxp, "failed to destroy "
268 			    "completion queue num %u",
269 			    mlcq->mlcq_num);
270 		}
271 	}
272 	if (mlcq->mlcq_state & MLXCX_CQ_ALLOC) {
273 		mlxcx_cq_rele_dma(mlxp, mlcq);
274 	}
275 	/*
276 	 * If we're on an EQ AVL tree, then we need to grab
277 	 * the EQ's mutex to take it off. The ISR always takes
278 	 * EQ mutex before CQ mutex, so we have to let go of
279 	 * the CQ mutex then come back again.
280 	 *
281 	 * The ISR will bail out if tries to touch this CQ now since
282 	 * we added the CQ_DESTROYED flag above.
283 	 */
284 	if (mlcq->mlcq_state & MLXCX_CQ_EQAVL) {
285 		mleq = mlcq->mlcq_eq;
286 	} else {
287 		mleq = NULL;
288 	}
289 
290 	/* Return any outstanding buffers to the free pool. */
291 	while ((b = list_remove_head(&mlcq->mlcq_buffers)) != NULL) {
292 		mlxcx_buf_return_chain(mlxp, b, B_FALSE);
293 	}
294 	mutex_enter(&mlcq->mlcq_bufbmtx);
295 	while ((b = list_remove_head(&mlcq->mlcq_buffers_b)) != NULL) {
296 		mlxcx_buf_return_chain(mlxp, b, B_FALSE);
297 	}
298 	mutex_exit(&mlcq->mlcq_bufbmtx);
299 
300 	/*
301 	 * Since the interrupt handlers take the EQ lock before the CQ one,
302 	 * we must do the same here. That means letting go of the lock
303 	 * for a brief window here (we'll double-check the state when we
304 	 * get back in).
305 	 */
306 	mutex_exit(&mlcq->mlcq_mtx);
307 
308 	if (mleq != NULL) {
309 		mutex_enter(&mleq->mleq_mtx);
310 		mutex_enter(&mlcq->mlcq_mtx);
311 		/*
312 		 * Double-check the state, we let go of the
313 		 * mutex briefly.
314 		 */
315 		if (mlcq->mlcq_state & MLXCX_CQ_EQAVL) {
316 			avl_remove(&mleq->mleq_cqs, mlcq);
317 			atomic_and_uint(&mlcq->mlcq_state, ~MLXCX_CQ_EQAVL);
318 		}
319 		mutex_exit(&mlcq->mlcq_mtx);
320 		mutex_exit(&mleq->mleq_mtx);
321 	}
322 
323 	mutex_enter(&mlcq->mlcq_mtx);
324 	ASSERT0(mlcq->mlcq_state & ~(MLXCX_CQ_CREATED | MLXCX_CQ_DESTROYED |
325 	    MLXCX_CQ_TEARDOWN | MLXCX_CQ_ARMED));
326 	mutex_exit(&mlcq->mlcq_mtx);
327 
328 	mutex_destroy(&mlcq->mlcq_mtx);
329 	mutex_destroy(&mlcq->mlcq_arm_mtx);
330 	mutex_destroy(&mlcq->mlcq_bufbmtx);
331 	list_destroy(&mlcq->mlcq_buffers);
332 	list_destroy(&mlcq->mlcq_buffers_b);
333 	kmem_free(mlcq, sizeof (mlxcx_completion_queue_t));
334 }
335 
336 static boolean_t
337 mlxcx_cq_setup(mlxcx_t *mlxp, mlxcx_event_queue_t *eq,
338     mlxcx_completion_queue_t **cqp, uint_t ent_shift)
339 {
340 	mlxcx_completion_queue_t *cq;
341 
342 	cq = kmem_zalloc(sizeof (mlxcx_completion_queue_t), KM_SLEEP);
343 	mutex_init(&cq->mlcq_mtx, NULL, MUTEX_DRIVER,
344 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
345 	mutex_init(&cq->mlcq_arm_mtx, NULL, MUTEX_DRIVER,
346 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
347 	mutex_init(&cq->mlcq_bufbmtx, NULL, MUTEX_DRIVER,
348 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
349 	list_create(&cq->mlcq_buffers, sizeof (mlxcx_buffer_t),
350 	    offsetof(mlxcx_buffer_t, mlb_cq_entry));
351 	list_create(&cq->mlcq_buffers_b, sizeof (mlxcx_buffer_t),
352 	    offsetof(mlxcx_buffer_t, mlb_cq_entry));
353 
354 	cq->mlcq_mlx = mlxp;
355 	list_insert_tail(&mlxp->mlx_cqs, cq);
356 
357 	mutex_enter(&cq->mlcq_mtx);
358 
359 	if (!mlxcx_cq_alloc_dma(mlxp, cq, ent_shift)) {
360 		mutex_exit(&cq->mlcq_mtx);
361 		return (B_FALSE);
362 	}
363 
364 	cq->mlcq_bufhwm = cq->mlcq_nents - MLXCX_CQ_HWM_GAP;
365 	cq->mlcq_buflwm = cq->mlcq_nents - MLXCX_CQ_LWM_GAP;
366 
367 	cq->mlcq_uar = &mlxp->mlx_uar;
368 	cq->mlcq_eq = eq;
369 
370 	cq->mlcq_cqemod_period_usec = mlxp->mlx_props.mldp_cqemod_period_usec;
371 	cq->mlcq_cqemod_count = mlxp->mlx_props.mldp_cqemod_count;
372 
373 	if (!mlxcx_cmd_create_cq(mlxp, cq)) {
374 		mutex_exit(&cq->mlcq_mtx);
375 		return (B_FALSE);
376 	}
377 
378 	mutex_exit(&cq->mlcq_mtx);
379 
380 	mutex_enter(&eq->mleq_mtx);
381 	mutex_enter(&cq->mlcq_arm_mtx);
382 	mutex_enter(&cq->mlcq_mtx);
383 	ASSERT0(cq->mlcq_state & MLXCX_CQ_EQAVL);
384 	avl_add(&eq->mleq_cqs, cq);
385 	atomic_or_uint(&cq->mlcq_state, MLXCX_CQ_EQAVL);
386 	mlxcx_arm_cq(mlxp, cq);
387 	mutex_exit(&cq->mlcq_mtx);
388 	mutex_exit(&cq->mlcq_arm_mtx);
389 	mutex_exit(&eq->mleq_mtx);
390 
391 	*cqp = cq;
392 	return (B_TRUE);
393 }
394 
395 static boolean_t
396 mlxcx_rq_setup(mlxcx_t *mlxp, mlxcx_completion_queue_t *cq,
397     mlxcx_work_queue_t *wq)
398 {
399 	mutex_init(&wq->mlwq_mtx, NULL, MUTEX_DRIVER,
400 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
401 
402 	list_insert_tail(&mlxp->mlx_wqs, wq);
403 
404 	mutex_enter(&wq->mlwq_mtx);
405 
406 	wq->mlwq_mlx = mlxp;
407 	wq->mlwq_type = MLXCX_WQ_TYPE_RECVQ;
408 	wq->mlwq_cq = cq;
409 	wq->mlwq_pd = &mlxp->mlx_pd;
410 	wq->mlwq_uar = &mlxp->mlx_uar;
411 
412 	wq->mlwq_bufs = mlxcx_mlbs_create(mlxp);
413 
414 	if (!mlxcx_wq_alloc_dma(mlxp, wq)) {
415 		mutex_exit(&wq->mlwq_mtx);
416 		return (B_FALSE);
417 	}
418 
419 	if (!mlxcx_cmd_create_rq(mlxp, wq)) {
420 		mutex_exit(&wq->mlwq_mtx);
421 		return (B_FALSE);
422 	}
423 
424 	wq->mlwq_bufhwm = wq->mlwq_nents - MLXCX_WQ_HWM_GAP;
425 	wq->mlwq_buflwm = wq->mlwq_nents - MLXCX_WQ_LWM_GAP;
426 
427 	mutex_exit(&wq->mlwq_mtx);
428 
429 	mutex_enter(&cq->mlcq_mtx);
430 	mutex_enter(&wq->mlwq_mtx);
431 	ASSERT3P(cq->mlcq_wq, ==, NULL);
432 	cq->mlcq_wq = wq;
433 	mutex_exit(&wq->mlwq_mtx);
434 	mutex_exit(&cq->mlcq_mtx);
435 
436 	return (B_TRUE);
437 }
438 
439 static boolean_t
440 mlxcx_sq_setup(mlxcx_t *mlxp, mlxcx_port_t *port, mlxcx_completion_queue_t *cq,
441     mlxcx_tis_t *tis, mlxcx_work_queue_t *wq)
442 {
443 	mutex_init(&wq->mlwq_mtx, NULL, MUTEX_DRIVER,
444 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
445 
446 	list_insert_tail(&mlxp->mlx_wqs, wq);
447 
448 	mutex_enter(&wq->mlwq_mtx);
449 
450 	wq->mlwq_mlx = mlxp;
451 	wq->mlwq_type = MLXCX_WQ_TYPE_SENDQ;
452 	wq->mlwq_cq = cq;
453 	wq->mlwq_pd = &mlxp->mlx_pd;
454 	wq->mlwq_uar = &mlxp->mlx_uar;
455 	wq->mlwq_tis = tis;
456 
457 	wq->mlwq_bufs = mlxcx_mlbs_create(mlxp);
458 	wq->mlwq_foreign_bufs = mlxcx_mlbs_create(mlxp);
459 
460 	VERIFY3U(port->mlp_wqe_min_inline, <=, MLXCX_ETH_INLINE_L2);
461 	wq->mlwq_inline_mode = MLXCX_ETH_INLINE_L2;
462 
463 	if (!mlxcx_wq_alloc_dma(mlxp, wq)) {
464 		mutex_exit(&wq->mlwq_mtx);
465 		return (B_FALSE);
466 	}
467 
468 	if (!mlxcx_cmd_create_sq(mlxp, wq)) {
469 		mutex_exit(&wq->mlwq_mtx);
470 		return (B_FALSE);
471 	}
472 
473 	wq->mlwq_bufhwm = wq->mlwq_nents - MLXCX_WQ_HWM_GAP;
474 	wq->mlwq_buflwm = wq->mlwq_nents - MLXCX_WQ_LWM_GAP;
475 
476 	mutex_exit(&wq->mlwq_mtx);
477 
478 	mutex_enter(&cq->mlcq_mtx);
479 	mutex_enter(&wq->mlwq_mtx);
480 	ASSERT3P(cq->mlcq_wq, ==, NULL);
481 	cq->mlcq_wq = wq;
482 	mutex_exit(&wq->mlwq_mtx);
483 	mutex_exit(&cq->mlcq_mtx);
484 
485 	return (B_TRUE);
486 }
487 
488 /*
489  * Before we tear down the queues associated with the rx group,
490  * flag each cq as being torn down and wake up any tasks.
491  */
492 static void
493 mlxcx_quiesce_rx_cqs(mlxcx_t *mlxp, mlxcx_ring_group_t *g)
494 {
495 	mlxcx_work_queue_t *wq;
496 	mlxcx_completion_queue_t *cq;
497 	mlxcx_buf_shard_t *s;
498 	uint_t i;
499 
500 	mutex_enter(&g->mlg_mtx);
501 
502 	for (i = 0; i < g->mlg_nwqs; ++i) {
503 		wq = &g->mlg_wqs[i];
504 		cq = wq->mlwq_cq;
505 		if (cq != NULL) {
506 			s = wq->mlwq_bufs;
507 			mutex_enter(&s->mlbs_mtx);
508 			atomic_or_uint(&cq->mlcq_state, MLXCX_CQ_TEARDOWN);
509 			cv_broadcast(&s->mlbs_free_nonempty);
510 			mutex_exit(&s->mlbs_mtx);
511 		}
512 	}
513 
514 	mutex_exit(&g->mlg_mtx);
515 }
516 
517 void
518 mlxcx_teardown_rx_group(mlxcx_t *mlxp, mlxcx_ring_group_t *g)
519 {
520 	mlxcx_work_queue_t *wq;
521 	mlxcx_completion_queue_t *cq;
522 	mlxcx_flow_entry_t *fe;
523 	mlxcx_flow_group_t *fg;
524 	mlxcx_flow_table_t *ft;
525 	uint_t i;
526 
527 	mutex_enter(&g->mlg_port->mlp_mtx);
528 	mutex_enter(&g->mlg_mtx);
529 
530 	if (g->mlg_state & MLXCX_GROUP_FLOWS) {
531 		mlxcx_remove_all_umcast_entries(mlxp, g->mlg_port, g);
532 
533 		if (g->mlg_rx_vlan_ft != NULL)
534 			mlxcx_remove_all_vlan_entries(mlxp, g);
535 
536 		if (g == &mlxp->mlx_rx_groups[0]) {
537 			ft = g->mlg_port->mlp_rx_flow;
538 			mutex_enter(&ft->mlft_mtx);
539 
540 			fg = g->mlg_port->mlp_bcast;
541 			fe = list_head(&fg->mlfg_entries);
542 			if (fe->mlfe_state & MLXCX_FLOW_ENTRY_CREATED) {
543 				(void) mlxcx_cmd_delete_flow_table_entry(
544 				    mlxp, fe);
545 			}
546 
547 			fg = g->mlg_port->mlp_promisc;
548 			fe = list_head(&fg->mlfg_entries);
549 			if (fe->mlfe_state & MLXCX_FLOW_ENTRY_CREATED) {
550 				(void) mlxcx_cmd_delete_flow_table_entry(
551 				    mlxp, fe);
552 			}
553 
554 			mutex_exit(&ft->mlft_mtx);
555 		}
556 
557 		if (g->mlg_rx_vlan_ft != NULL) {
558 			mutex_enter(&g->mlg_rx_vlan_ft->mlft_mtx);
559 			ASSERT(list_is_empty(&g->mlg_rx_vlans));
560 			fg = g->mlg_rx_vlan_def_fg;
561 			if (fg != NULL) {
562 				fe = list_head(&fg->mlfg_entries);
563 				if (fe->mlfe_state & MLXCX_FLOW_ENTRY_CREATED) {
564 					(void)
565 					    mlxcx_cmd_delete_flow_table_entry(
566 					    mlxp, fe);
567 				}
568 			}
569 			fg = g->mlg_rx_vlan_promisc_fg;
570 			if (fg != NULL) {
571 				fe = list_head(&fg->mlfg_entries);
572 				if (fe->mlfe_state & MLXCX_FLOW_ENTRY_CREATED) {
573 					(void)
574 					    mlxcx_cmd_delete_flow_table_entry(
575 					    mlxp, fe);
576 				}
577 			}
578 			mlxcx_teardown_flow_table(mlxp, g->mlg_rx_vlan_ft);
579 			list_destroy(&g->mlg_rx_vlans);
580 
581 			g->mlg_rx_vlan_ft = NULL;
582 		}
583 
584 		mutex_enter(&g->mlg_rx_hash_ft->mlft_mtx);
585 		mlxcx_teardown_flow_table(mlxp, g->mlg_rx_hash_ft);
586 		g->mlg_rx_hash_ft = NULL;
587 
588 		avl_destroy(&g->mlg_rx_macs);
589 		g->mlg_state &= ~MLXCX_GROUP_FLOWS;
590 	}
591 
592 	if (g->mlg_state & MLXCX_GROUP_RUNNING) {
593 		for (i = 0; i < g->mlg_nwqs; ++i) {
594 			wq = &g->mlg_wqs[i];
595 			mutex_enter(&wq->mlwq_mtx);
596 			if (wq->mlwq_state & MLXCX_WQ_STARTED &&
597 			    !mlxcx_cmd_stop_rq(mlxp, wq)) {
598 				mlxcx_warn(mlxp, "failed to stop rq %x",
599 				    wq->mlwq_num);
600 			}
601 			mutex_exit(&wq->mlwq_mtx);
602 		}
603 		taskq_destroy(g->mlg_refill_tq);
604 		g->mlg_state &= ~MLXCX_GROUP_RUNNING;
605 	}
606 
607 	if (g->mlg_state & MLXCX_GROUP_TIRTIS) {
608 		for (i = 0; i < MLXCX_TIRS_PER_GROUP; ++i) {
609 			mlxcx_tir_t *tir = &g->mlg_tir[i];
610 			if (tir->mltir_state & MLXCX_TIR_CREATED &&
611 			    !(tir->mltir_state & MLXCX_TIR_DESTROYED)) {
612 				if (!mlxcx_cmd_destroy_tir(mlxp, tir)) {
613 					mlxcx_warn(mlxp,
614 					    "failed to destroy tir %u "
615 					    "for rx ring", tir->mltir_num);
616 				}
617 			}
618 		}
619 		g->mlg_state &= ~MLXCX_GROUP_TIRTIS;
620 	}
621 
622 	if (g->mlg_state & MLXCX_GROUP_RQT) {
623 		if (g->mlg_rqt->mlrqt_state & MLXCX_RQT_CREATED &&
624 		    !(g->mlg_rqt->mlrqt_state & MLXCX_RQT_DESTROYED)) {
625 			if (!mlxcx_cmd_destroy_rqt(mlxp, g->mlg_rqt)) {
626 				mlxcx_warn(mlxp, "failed to destroy rqt %u "
627 				    "for rx ring", g->mlg_rqt->mlrqt_num);
628 			}
629 			kmem_free(g->mlg_rqt->mlrqt_rq,
630 			    g->mlg_rqt->mlrqt_rq_size);
631 			g->mlg_rqt->mlrqt_rq = NULL;
632 			kmem_free(g->mlg_rqt, sizeof (mlxcx_rqtable_t));
633 			g->mlg_rqt = NULL;
634 		}
635 		g->mlg_state &= ~MLXCX_GROUP_RQT;
636 	}
637 
638 	for (i = 0; i < g->mlg_nwqs; ++i) {
639 		wq = &g->mlg_wqs[i];
640 		cq = wq->mlwq_cq;
641 		mlxcx_wq_teardown(mlxp, wq);
642 		if (cq != NULL)
643 			mlxcx_cq_teardown(mlxp, cq);
644 	}
645 	kmem_free(g->mlg_wqs, g->mlg_wqs_size);
646 	g->mlg_wqs = NULL;
647 	g->mlg_state &= ~MLXCX_GROUP_WQS;
648 
649 	mutex_exit(&g->mlg_mtx);
650 	mutex_exit(&g->mlg_port->mlp_mtx);
651 
652 	mutex_destroy(&g->mlg_mtx);
653 
654 	g->mlg_state &= ~MLXCX_GROUP_INIT;
655 	ASSERT3S(g->mlg_state, ==, 0);
656 }
657 
658 void
659 mlxcx_teardown_tx_group(mlxcx_t *mlxp, mlxcx_ring_group_t *g)
660 {
661 	mlxcx_work_queue_t *wq;
662 	mlxcx_completion_queue_t *cq;
663 	uint_t i;
664 
665 	mutex_enter(&g->mlg_mtx);
666 
667 	if (g->mlg_state & MLXCX_GROUP_WQS) {
668 		for (i = 0; i < g->mlg_nwqs; ++i) {
669 			wq = &g->mlg_wqs[i];
670 			mutex_enter(&wq->mlwq_mtx);
671 			cq = wq->mlwq_cq;
672 			if (wq->mlwq_state & MLXCX_WQ_STARTED &&
673 			    !mlxcx_cmd_stop_sq(mlxp, wq)) {
674 				mlxcx_warn(mlxp, "failed to stop sq %x",
675 				    wq->mlwq_num);
676 			}
677 			mutex_exit(&wq->mlwq_mtx);
678 			mlxcx_wq_teardown(mlxp, wq);
679 			if (cq != NULL)
680 				mlxcx_cq_teardown(mlxp, cq);
681 		}
682 		g->mlg_state &= ~MLXCX_GROUP_RUNNING;
683 		kmem_free(g->mlg_wqs, g->mlg_wqs_size);
684 		g->mlg_wqs = NULL;
685 		g->mlg_state &= ~MLXCX_GROUP_WQS;
686 	}
687 
688 	if ((g->mlg_state & MLXCX_GROUP_TIRTIS) &&
689 	    g->mlg_tis.mltis_state & MLXCX_TIS_CREATED &&
690 	    !(g->mlg_tis.mltis_state & MLXCX_TIS_DESTROYED)) {
691 		if (!mlxcx_cmd_destroy_tis(mlxp, &g->mlg_tis)) {
692 			mlxcx_warn(mlxp, "failed to destroy tis %u for tx ring",
693 			    g->mlg_tis.mltis_num);
694 		}
695 	}
696 	g->mlg_state &= ~MLXCX_GROUP_TIRTIS;
697 
698 	mutex_exit(&g->mlg_mtx);
699 	mutex_destroy(&g->mlg_mtx);
700 	g->mlg_state &= ~MLXCX_GROUP_INIT;
701 	ASSERT3S(g->mlg_state, ==, 0);
702 }
703 
704 void
705 mlxcx_teardown_groups(mlxcx_t *mlxp)
706 {
707 	mlxcx_ring_group_t *g;
708 	uint_t i;
709 
710 	for (i = 0; i < mlxp->mlx_rx_ngroups; ++i) {
711 		g = &mlxp->mlx_rx_groups[i];
712 		if (!(g->mlg_state & MLXCX_GROUP_INIT))
713 			continue;
714 		ASSERT3S(g->mlg_type, ==, MLXCX_GROUP_RX);
715 		mlxcx_quiesce_rx_cqs(mlxp, g);
716 	}
717 
718 	for (i = 0; i < mlxp->mlx_rx_ngroups; ++i) {
719 		g = &mlxp->mlx_rx_groups[i];
720 		if (!(g->mlg_state & MLXCX_GROUP_INIT))
721 			continue;
722 		mlxcx_teardown_rx_group(mlxp, g);
723 	}
724 
725 	kmem_free(mlxp->mlx_rx_groups, mlxp->mlx_rx_groups_size);
726 	mlxp->mlx_rx_groups = NULL;
727 
728 	for (i = 0; i < mlxp->mlx_tx_ngroups; ++i) {
729 		g = &mlxp->mlx_tx_groups[i];
730 		if (!(g->mlg_state & MLXCX_GROUP_INIT))
731 			continue;
732 		ASSERT3S(g->mlg_type, ==, MLXCX_GROUP_TX);
733 		mlxcx_teardown_tx_group(mlxp, g);
734 	}
735 
736 	kmem_free(mlxp->mlx_tx_groups, mlxp->mlx_tx_groups_size);
737 	mlxp->mlx_tx_groups = NULL;
738 }
739 
740 boolean_t
741 mlxcx_rx_group_setup(mlxcx_t *mlxp, mlxcx_ring_group_t *g)
742 {
743 	mlxcx_event_queue_t *eq;
744 	mlxcx_completion_queue_t *cq;
745 	mlxcx_work_queue_t *rq;
746 	mlxcx_flow_table_t *ft;
747 	mlxcx_flow_group_t *fg;
748 	mlxcx_flow_entry_t *fe;
749 	uint_t ent_shift;
750 	uint_t i, j;
751 
752 	ASSERT3S(g->mlg_state, ==, 0);
753 
754 	mutex_init(&g->mlg_mtx, NULL, MUTEX_DRIVER,
755 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
756 	mutex_enter(&g->mlg_mtx);
757 	g->mlg_mlx = mlxp;
758 	g->mlg_type = MLXCX_GROUP_RX;
759 	g->mlg_port = &mlxp->mlx_ports[0];
760 	g->mlg_state |= MLXCX_GROUP_INIT;
761 
762 	g->mlg_nwqs = mlxp->mlx_props.mldp_rx_nrings_per_small_group;
763 	i = g - &mlxp->mlx_rx_groups[0];
764 	if (i < mlxp->mlx_props.mldp_rx_ngroups_large)
765 		g->mlg_nwqs = mlxp->mlx_props.mldp_rx_nrings_per_large_group;
766 
767 	g->mlg_wqs_size = g->mlg_nwqs * sizeof (mlxcx_work_queue_t);
768 	g->mlg_wqs = kmem_zalloc(g->mlg_wqs_size, KM_SLEEP);
769 	g->mlg_state |= MLXCX_GROUP_WQS;
770 
771 	g->mlg_rqt = kmem_zalloc(sizeof (mlxcx_rqtable_t), KM_SLEEP);
772 	g->mlg_rqt->mlrqt_max = 2;
773 	while (g->mlg_rqt->mlrqt_max < g->mlg_nwqs)
774 		g->mlg_rqt->mlrqt_max <<= 1;
775 	g->mlg_rqt->mlrqt_rq_size = g->mlg_rqt->mlrqt_max *
776 	    sizeof (mlxcx_work_queue_t *);
777 	g->mlg_rqt->mlrqt_rq = kmem_zalloc(g->mlg_rqt->mlrqt_rq_size, KM_SLEEP);
778 	g->mlg_state |= MLXCX_GROUP_RQT;
779 
780 	for (i = 0; i < g->mlg_nwqs; ++i) {
781 		eq = NULL;
782 		while (eq == NULL) {
783 			eq = &mlxp->mlx_eqs[mlxp->mlx_next_eq++];
784 			if (mlxp->mlx_next_eq >= mlxp->mlx_intr_count)
785 				mlxp->mlx_next_eq = mlxp->mlx_intr_cq0;
786 			if (eq->mleq_type != MLXCX_EQ_TYPE_ANY &&
787 			    eq->mleq_type != MLXCX_EQ_TYPE_RX) {
788 				/* Try the next one */
789 				eq = NULL;
790 			}
791 		}
792 
793 		/*
794 		 * A single completion is indicated for each rq entry as
795 		 * it is used. So, the number of cq entries never needs
796 		 * to be larger than the rq.
797 		 */
798 		ent_shift = MIN(mlxp->mlx_props.mldp_cq_size_shift,
799 		    mlxp->mlx_props.mldp_rq_size_shift);
800 		if (!mlxcx_cq_setup(mlxp, eq, &cq, ent_shift)) {
801 			g->mlg_nwqs = i;
802 			break;
803 		}
804 
805 		cq->mlcq_stats = &g->mlg_port->mlp_stats;
806 
807 		rq = &g->mlg_wqs[i];
808 		if (!mlxcx_rq_setup(mlxp, cq, rq)) {
809 			g->mlg_nwqs = i;
810 			break;
811 		}
812 		g->mlg_rqt->mlrqt_rq[g->mlg_rqt->mlrqt_used++] = rq;
813 		g->mlg_rqt->mlrqt_state |= MLXCX_RQT_DIRTY;
814 		rq->mlwq_group = g;
815 	}
816 	if (g->mlg_nwqs == 0) {
817 		mutex_exit(&g->mlg_mtx);
818 		return (B_FALSE);
819 	}
820 
821 	if (!mlxcx_cmd_create_rqt(mlxp, g->mlg_rqt)) {
822 		mutex_exit(&g->mlg_mtx);
823 		return (B_FALSE);
824 	}
825 
826 	for (i = 0; i < MLXCX_TIRS_PER_GROUP; ++i) {
827 		mlxcx_tir_t *tir = &g->mlg_tir[i];
828 		tir->mltir_tdom = &mlxp->mlx_tdom;
829 		switch (i) {
830 		case MLXCX_TIR_ROLE_OTHER:
831 			tir->mltir_type = MLXCX_TIR_DIRECT;
832 			tir->mltir_rq = &g->mlg_wqs[0];
833 			break;
834 		case MLXCX_TIR_ROLE_IPv4:
835 		case MLXCX_TIR_ROLE_IPv6:
836 		case MLXCX_TIR_ROLE_TCPv4:
837 		case MLXCX_TIR_ROLE_TCPv6:
838 		case MLXCX_TIR_ROLE_UDPv4:
839 		case MLXCX_TIR_ROLE_UDPv6:
840 			tir->mltir_type = MLXCX_TIR_INDIRECT;
841 			tir->mltir_rqtable = g->mlg_rqt;
842 			tir->mltir_hash_fn = MLXCX_TIR_HASH_TOEPLITZ;
843 			(void) random_get_pseudo_bytes(tir->mltir_toeplitz_key,
844 			    sizeof (tir->mltir_toeplitz_key));
845 			break;
846 		}
847 		switch (i) {
848 		case MLXCX_TIR_ROLE_OTHER:
849 			break;
850 		case MLXCX_TIR_ROLE_IPv4:
851 		case MLXCX_TIR_ROLE_TCPv4:
852 		case MLXCX_TIR_ROLE_UDPv4:
853 			tir->mltir_l3_type = MLXCX_RX_HASH_L3_IPv4;
854 			tir->mltir_hash_fields =
855 			    MLXCX_RX_HASH_SRC_IP | MLXCX_RX_HASH_DST_IP;
856 			break;
857 		case MLXCX_TIR_ROLE_IPv6:
858 		case MLXCX_TIR_ROLE_TCPv6:
859 		case MLXCX_TIR_ROLE_UDPv6:
860 			tir->mltir_l3_type = MLXCX_RX_HASH_L3_IPv6;
861 			tir->mltir_hash_fields =
862 			    MLXCX_RX_HASH_SRC_IP | MLXCX_RX_HASH_DST_IP;
863 			break;
864 		}
865 		switch (i) {
866 		case MLXCX_TIR_ROLE_OTHER:
867 		case MLXCX_TIR_ROLE_IPv4:
868 		case MLXCX_TIR_ROLE_IPv6:
869 			break;
870 		case MLXCX_TIR_ROLE_TCPv4:
871 		case MLXCX_TIR_ROLE_TCPv6:
872 			tir->mltir_l4_type = MLXCX_RX_HASH_L4_TCP;
873 			tir->mltir_hash_fields |=
874 			    MLXCX_RX_HASH_L4_SPORT | MLXCX_RX_HASH_L4_DPORT;
875 			break;
876 		case MLXCX_TIR_ROLE_UDPv4:
877 		case MLXCX_TIR_ROLE_UDPv6:
878 			tir->mltir_l4_type = MLXCX_RX_HASH_L4_UDP;
879 			tir->mltir_hash_fields |=
880 			    MLXCX_RX_HASH_L4_SPORT | MLXCX_RX_HASH_L4_DPORT;
881 			break;
882 		}
883 
884 		if (!mlxcx_cmd_create_tir(mlxp, tir)) {
885 			mutex_exit(&g->mlg_mtx);
886 			return (B_FALSE);
887 		}
888 
889 		g->mlg_state |= MLXCX_GROUP_TIRTIS;
890 	}
891 
892 	/*
893 	 * Flow table: our RX hashing breakout table for RSS
894 	 */
895 
896 	g->mlg_rx_hash_ft = (ft = kmem_zalloc(sizeof (mlxcx_flow_table_t),
897 	    KM_SLEEP));
898 	mutex_init(&ft->mlft_mtx, NULL, MUTEX_DRIVER,
899 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
900 	avl_create(&g->mlg_rx_macs, mlxcx_grmac_compare,
901 	    sizeof (mlxcx_group_mac_t),
902 	    offsetof(mlxcx_group_mac_t, mlgm_group_entry));
903 	g->mlg_state |= MLXCX_GROUP_FLOWS;
904 
905 	mutex_enter(&ft->mlft_mtx);
906 
907 	ft->mlft_type = MLXCX_FLOW_TABLE_NIC_RX;
908 	ft->mlft_level = 2;
909 	ft->mlft_port = g->mlg_port;
910 	ft->mlft_entshift = MLXCX_RX_HASH_FT_SIZE_SHIFT;
911 	ft->mlft_nents = (1 << ft->mlft_entshift);
912 	ASSERT3U(ft->mlft_nents, >=, MLXCX_TIRS_PER_GROUP);
913 	ft->mlft_entsize = ft->mlft_nents * sizeof (mlxcx_flow_entry_t);
914 	ft->mlft_ent = kmem_zalloc(ft->mlft_entsize, KM_SLEEP);
915 	list_create(&ft->mlft_groups, sizeof (mlxcx_flow_group_t),
916 	    offsetof(mlxcx_flow_group_t, mlfg_entry));
917 
918 	for (j = 0; j < ft->mlft_nents; ++j) {
919 		ft->mlft_ent[j].mlfe_table = ft;
920 		ft->mlft_ent[j].mlfe_index = j;
921 	}
922 
923 	if (!mlxcx_cmd_create_flow_table(mlxp, ft)) {
924 		mutex_exit(&ft->mlft_mtx);
925 		mutex_exit(&g->mlg_mtx);
926 		return (B_FALSE);
927 	}
928 
929 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
930 	list_insert_tail(&ft->mlft_groups, fg);
931 	fg->mlfg_table = ft;
932 	fg->mlfg_size = 1;
933 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_IP_VER | MLXCX_FLOW_MATCH_IP_PROTO;
934 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
935 		mutex_exit(&ft->mlft_mtx);
936 		mutex_exit(&g->mlg_mtx);
937 		return (B_FALSE);
938 	}
939 	fe = list_head(&fg->mlfg_entries);
940 	fe->mlfe_ip_version = 6;
941 	fe->mlfe_ip_proto = IPPROTO_UDP;
942 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
943 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
944 	    &g->mlg_tir[MLXCX_TIR_ROLE_UDPv6];
945 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
946 		mutex_exit(&ft->mlft_mtx);
947 		mutex_exit(&g->mlg_mtx);
948 		return (B_FALSE);
949 	}
950 
951 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
952 	list_insert_tail(&ft->mlft_groups, fg);
953 	fg->mlfg_table = ft;
954 	fg->mlfg_size = 1;
955 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_IP_VER | MLXCX_FLOW_MATCH_IP_PROTO;
956 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
957 		mutex_exit(&ft->mlft_mtx);
958 		mutex_exit(&g->mlg_mtx);
959 		return (B_FALSE);
960 	}
961 	fe = list_head(&fg->mlfg_entries);
962 	fe->mlfe_ip_version = 4;
963 	fe->mlfe_ip_proto = IPPROTO_UDP;
964 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
965 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
966 	    &g->mlg_tir[MLXCX_TIR_ROLE_UDPv4];
967 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
968 		mutex_exit(&ft->mlft_mtx);
969 		mutex_exit(&g->mlg_mtx);
970 		return (B_FALSE);
971 	}
972 
973 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
974 	list_insert_tail(&ft->mlft_groups, fg);
975 	fg->mlfg_table = ft;
976 	fg->mlfg_size = 1;
977 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_IP_VER | MLXCX_FLOW_MATCH_IP_PROTO;
978 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
979 		mutex_exit(&ft->mlft_mtx);
980 		mutex_exit(&g->mlg_mtx);
981 		return (B_FALSE);
982 	}
983 	fe = list_head(&fg->mlfg_entries);
984 	fe->mlfe_ip_version = 6;
985 	fe->mlfe_ip_proto = IPPROTO_TCP;
986 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
987 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
988 	    &g->mlg_tir[MLXCX_TIR_ROLE_TCPv6];
989 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
990 		mutex_exit(&ft->mlft_mtx);
991 		mutex_exit(&g->mlg_mtx);
992 		return (B_FALSE);
993 	}
994 
995 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
996 	list_insert_tail(&ft->mlft_groups, fg);
997 	fg->mlfg_table = ft;
998 	fg->mlfg_size = 1;
999 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_IP_VER | MLXCX_FLOW_MATCH_IP_PROTO;
1000 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1001 		mutex_exit(&ft->mlft_mtx);
1002 		mutex_exit(&g->mlg_mtx);
1003 		return (B_FALSE);
1004 	}
1005 	fe = list_head(&fg->mlfg_entries);
1006 	fe->mlfe_ip_version = 4;
1007 	fe->mlfe_ip_proto = IPPROTO_TCP;
1008 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
1009 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
1010 	    &g->mlg_tir[MLXCX_TIR_ROLE_TCPv4];
1011 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
1012 		mutex_exit(&ft->mlft_mtx);
1013 		mutex_exit(&g->mlg_mtx);
1014 		return (B_FALSE);
1015 	}
1016 
1017 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
1018 	list_insert_tail(&ft->mlft_groups, fg);
1019 	fg->mlfg_table = ft;
1020 	fg->mlfg_size = 1;
1021 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_IP_VER;
1022 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1023 		mutex_exit(&ft->mlft_mtx);
1024 		mutex_exit(&g->mlg_mtx);
1025 		return (B_FALSE);
1026 	}
1027 	fe = list_head(&fg->mlfg_entries);
1028 	fe->mlfe_ip_version = 6;
1029 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
1030 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
1031 	    &g->mlg_tir[MLXCX_TIR_ROLE_IPv6];
1032 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
1033 		mutex_exit(&ft->mlft_mtx);
1034 		mutex_exit(&g->mlg_mtx);
1035 		return (B_FALSE);
1036 	}
1037 
1038 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
1039 	list_insert_tail(&ft->mlft_groups, fg);
1040 	fg->mlfg_table = ft;
1041 	fg->mlfg_size = 1;
1042 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_IP_VER;
1043 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1044 		mutex_exit(&ft->mlft_mtx);
1045 		mutex_exit(&g->mlg_mtx);
1046 		return (B_FALSE);
1047 	}
1048 	fe = list_head(&fg->mlfg_entries);
1049 	fe->mlfe_ip_version = 4;
1050 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
1051 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
1052 	    &g->mlg_tir[MLXCX_TIR_ROLE_IPv4];
1053 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
1054 		mutex_exit(&ft->mlft_mtx);
1055 		mutex_exit(&g->mlg_mtx);
1056 		return (B_FALSE);
1057 	}
1058 
1059 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
1060 	list_insert_tail(&ft->mlft_groups, fg);
1061 	fg->mlfg_table = ft;
1062 	fg->mlfg_size = 1;
1063 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1064 		mutex_exit(&ft->mlft_mtx);
1065 		mutex_exit(&g->mlg_mtx);
1066 		return (B_FALSE);
1067 	}
1068 	fe = list_head(&fg->mlfg_entries);
1069 	fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
1070 	fe->mlfe_dest[fe->mlfe_ndest++].mlfed_tir =
1071 	    &g->mlg_tir[MLXCX_TIR_ROLE_OTHER];
1072 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
1073 		mutex_exit(&ft->mlft_mtx);
1074 		mutex_exit(&g->mlg_mtx);
1075 		return (B_FALSE);
1076 	}
1077 
1078 	mutex_exit(&ft->mlft_mtx);
1079 
1080 	/*
1081 	 * Flow table: the VLAN breakout table for doing VLAN filtering after
1082 	 * we've matched a MAC address.
1083 	 */
1084 
1085 	g->mlg_rx_vlan_ft = (ft = kmem_zalloc(sizeof (mlxcx_flow_table_t),
1086 	    KM_SLEEP));
1087 	mutex_init(&ft->mlft_mtx, NULL, MUTEX_DRIVER,
1088 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
1089 	list_create(&g->mlg_rx_vlans, sizeof (mlxcx_group_vlan_t),
1090 	    offsetof(mlxcx_group_vlan_t, mlgv_entry));
1091 
1092 	mutex_enter(&ft->mlft_mtx);
1093 
1094 	ft->mlft_type = MLXCX_FLOW_TABLE_NIC_RX;
1095 	ft->mlft_level = 1;
1096 	ft->mlft_port = g->mlg_port;
1097 	ft->mlft_entshift = mlxp->mlx_props.mldp_ftbl_vlan_size_shift;
1098 	ft->mlft_nents = (1 << ft->mlft_entshift);
1099 	ft->mlft_entsize = ft->mlft_nents * sizeof (mlxcx_flow_entry_t);
1100 	ft->mlft_ent = kmem_zalloc(ft->mlft_entsize, KM_SLEEP);
1101 	list_create(&ft->mlft_groups, sizeof (mlxcx_flow_group_t),
1102 	    offsetof(mlxcx_flow_group_t, mlfg_entry));
1103 
1104 	for (j = 0; j < ft->mlft_nents; ++j) {
1105 		fe = &ft->mlft_ent[j];
1106 		fe->mlfe_table = ft;
1107 		fe->mlfe_index = j;
1108 		fe->mlfe_action = MLXCX_FLOW_ACTION_FORWARD;
1109 		fe->mlfe_dest[fe->mlfe_ndest++].mlfed_flow = g->mlg_rx_hash_ft;
1110 	}
1111 
1112 	if (!mlxcx_cmd_create_flow_table(mlxp, ft)) {
1113 		mutex_exit(&ft->mlft_mtx);
1114 		mutex_exit(&g->mlg_mtx);
1115 		return (B_FALSE);
1116 	}
1117 
1118 	/* First group is all actual matched VLANs */
1119 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
1120 	g->mlg_rx_vlan_fg = fg;
1121 	list_insert_tail(&ft->mlft_groups, fg);
1122 	fg->mlfg_table = ft;
1123 	fg->mlfg_size = ft->mlft_nents - 2;
1124 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_VLAN;
1125 	fg->mlfg_mask |= MLXCX_FLOW_MATCH_VID;
1126 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1127 		mutex_exit(&ft->mlft_mtx);
1128 		mutex_exit(&g->mlg_mtx);
1129 		return (B_FALSE);
1130 	}
1131 
1132 	/*
1133 	 * Then the "default" entry which we enable when we have no VLAN IDs
1134 	 * added to the group (we start with this enabled).
1135 	 */
1136 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
1137 	g->mlg_rx_vlan_def_fg = fg;
1138 	list_insert_tail(&ft->mlft_groups, fg);
1139 	fg->mlfg_table = ft;
1140 	fg->mlfg_size = 1;
1141 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1142 		mutex_exit(&ft->mlft_mtx);
1143 		mutex_exit(&g->mlg_mtx);
1144 		return (B_FALSE);
1145 	}
1146 	fe = list_head(&fg->mlfg_entries);
1147 	if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
1148 		mutex_exit(&ft->mlft_mtx);
1149 		mutex_exit(&g->mlg_mtx);
1150 		return (B_FALSE);
1151 	}
1152 
1153 	/*
1154 	 * Finally, the promisc entry which points at the *hash ft* from the
1155 	 * default group. We only enable this when we have promisc on.
1156 	 */
1157 	fg = kmem_zalloc(sizeof (mlxcx_flow_group_t), KM_SLEEP);
1158 	g->mlg_rx_vlan_promisc_fg = fg;
1159 	list_insert_tail(&ft->mlft_groups, fg);
1160 	fg->mlfg_table = ft;
1161 	fg->mlfg_size = 1;
1162 	if (!mlxcx_setup_flow_group(mlxp, ft, fg)) {
1163 		mutex_exit(&ft->mlft_mtx);
1164 		mutex_exit(&g->mlg_mtx);
1165 		return (B_FALSE);
1166 	}
1167 	fe = list_head(&fg->mlfg_entries);
1168 	fe->mlfe_ndest = 1;
1169 	fe->mlfe_dest[0].mlfed_flow = mlxp->mlx_rx_groups[0].mlg_rx_hash_ft;
1170 
1171 	mutex_exit(&ft->mlft_mtx);
1172 
1173 	mutex_exit(&g->mlg_mtx);
1174 
1175 	return (B_TRUE);
1176 }
1177 
1178 boolean_t
1179 mlxcx_rx_ring_start(mlxcx_t *mlxp, mlxcx_ring_group_t *g,
1180     mlxcx_work_queue_t *rq)
1181 {
1182 	uint_t j;
1183 	mlxcx_buffer_t *b;
1184 	mlxcx_completion_queue_t *cq;
1185 
1186 	mutex_enter(&g->mlg_mtx);
1187 	/*
1188 	 * Sadly, even though MAC has the mgi_start callback, it is not always
1189 	 * called -- in particular when we are being managed under an aggr, the
1190 	 * mgi_start callback will only ever be called on the default group.
1191 	 *
1192 	 * So instead of asserting about the group state here, we have to
1193 	 * check it and call group start if needed.
1194 	 */
1195 	if (!(g->mlg_state & MLXCX_GROUP_RUNNING)) {
1196 		mutex_exit(&g->mlg_mtx);
1197 		if (!mlxcx_rx_group_start(mlxp, g))
1198 			return (B_FALSE);
1199 		mutex_enter(&g->mlg_mtx);
1200 	}
1201 	ASSERT(g->mlg_state & MLXCX_GROUP_RUNNING);
1202 
1203 	cq = rq->mlwq_cq;
1204 	ASSERT(cq != NULL);
1205 
1206 	mutex_enter(&cq->mlcq_mtx);
1207 	mutex_enter(&rq->mlwq_mtx);
1208 
1209 	if (rq->mlwq_state & MLXCX_WQ_STARTED) {
1210 		mutex_exit(&rq->mlwq_mtx);
1211 		mutex_exit(&cq->mlcq_mtx);
1212 		mutex_exit(&g->mlg_mtx);
1213 		return (B_TRUE);
1214 	}
1215 
1216 	if (!mlxcx_cmd_start_rq(mlxp, rq)) {
1217 		mutex_exit(&rq->mlwq_mtx);
1218 		mutex_exit(&cq->mlcq_mtx);
1219 		mutex_exit(&g->mlg_mtx);
1220 		return (B_FALSE);
1221 	}
1222 	ASSERT(rq->mlwq_state & MLXCX_WQ_STARTED);
1223 
1224 	ASSERT0(rq->mlwq_state & MLXCX_WQ_BUFFERS);
1225 	rq->mlwq_state |= MLXCX_WQ_BUFFERS;
1226 
1227 	mlxcx_shard_ready(rq->mlwq_bufs);
1228 
1229 	for (j = 0; j < rq->mlwq_nents; ++j) {
1230 		if (!mlxcx_buf_create(mlxp, rq->mlwq_bufs, &b))
1231 			break;
1232 		mlxcx_buf_return(mlxp, b);
1233 	}
1234 	for (j = 0; j < rq->mlwq_nents / 2; ++j) {
1235 		if (!mlxcx_buf_create(mlxp, rq->mlwq_bufs, &b))
1236 			break;
1237 		mlxcx_buf_return(mlxp, b);
1238 	}
1239 
1240 	mlxcx_rq_refill(mlxp, rq);
1241 
1242 	mutex_exit(&rq->mlwq_mtx);
1243 	mutex_exit(&cq->mlcq_mtx);
1244 	mutex_exit(&g->mlg_mtx);
1245 
1246 	return (B_TRUE);
1247 }
1248 
1249 boolean_t
1250 mlxcx_rx_group_start(mlxcx_t *mlxp, mlxcx_ring_group_t *g)
1251 {
1252 	mlxcx_flow_table_t *ft;
1253 	mlxcx_flow_group_t *fg;
1254 	mlxcx_flow_entry_t *fe;
1255 	char tq_name[TASKQ_NAMELEN];
1256 
1257 	mutex_enter(&g->mlg_mtx);
1258 
1259 	if (g->mlg_state & MLXCX_GROUP_RUNNING) {
1260 		mutex_exit(&g->mlg_mtx);
1261 		return (B_TRUE);
1262 	}
1263 
1264 	ASSERT0(g->mlg_state & MLXCX_GROUP_RUNNING);
1265 
1266 	g->mlg_state |= MLXCX_GROUP_RUNNING;
1267 
1268 	(void) snprintf(tq_name, sizeof (tq_name), "%s_refill_%d_%ld",
1269 	    ddi_driver_name(mlxp->mlx_dip), mlxp->mlx_inst,
1270 	    g - &mlxp->mlx_rx_groups[0]);
1271 
1272 	/*
1273 	 * Create one refill taskq per group with one thread per work queue.
1274 	 * The refill task may block waiting for resources, so by effectively
1275 	 * having one thread per work queue we avoid work queues blocking each
1276 	 * other.
1277 	 */
1278 	if ((g->mlg_refill_tq = taskq_create(tq_name, g->mlg_nwqs, minclsyspri,
1279 	    g->mlg_nwqs, INT_MAX, TASKQ_PREPOPULATE)) == NULL) {
1280 		mlxcx_warn(mlxp, "failed to create rq refill task queue");
1281 		mutex_exit(&g->mlg_mtx);
1282 		return (B_FALSE);
1283 	}
1284 
1285 	if (g == &mlxp->mlx_rx_groups[0]) {
1286 		ft = g->mlg_port->mlp_rx_flow;
1287 		mutex_enter(&ft->mlft_mtx);
1288 
1289 		/*
1290 		 * Broadcast and promisc entries go directly to group 0's
1291 		 * RSS hash fanout flow table. They bypass VLAN filtering.
1292 		 */
1293 		fg = g->mlg_port->mlp_bcast;
1294 		fe = list_head(&fg->mlfg_entries);
1295 		fe->mlfe_dest[fe->mlfe_ndest++].mlfed_flow = g->mlg_rx_hash_ft;
1296 		if (!mlxcx_cmd_set_flow_table_entry(mlxp, fe)) {
1297 			mutex_exit(&ft->mlft_mtx);
1298 			g->mlg_state &= ~MLXCX_GROUP_RUNNING;
1299 			taskq_destroy(g->mlg_refill_tq);
1300 			mutex_exit(&g->mlg_mtx);
1301 			return (B_FALSE);
1302 		}
1303 
1304 		fg = g->mlg_port->mlp_promisc;
1305 		fe = list_head(&fg->mlfg_entries);
1306 		fe->mlfe_dest[fe->mlfe_ndest++].mlfed_flow = g->mlg_rx_hash_ft;
1307 		/*
1308 		 * Don't actually set the promisc entry until promisc is
1309 		 * enabled.
1310 		 */
1311 
1312 		mutex_exit(&ft->mlft_mtx);
1313 	}
1314 
1315 	mutex_exit(&g->mlg_mtx);
1316 
1317 	return (B_TRUE);
1318 }
1319 
1320 boolean_t
1321 mlxcx_tx_group_setup(mlxcx_t *mlxp, mlxcx_ring_group_t *g)
1322 {
1323 	mlxcx_event_queue_t *eq;
1324 	mlxcx_completion_queue_t *cq;
1325 	mlxcx_work_queue_t *sq;
1326 	uint_t i;
1327 
1328 	ASSERT3S(g->mlg_state, ==, 0);
1329 
1330 	mutex_init(&g->mlg_mtx, NULL, MUTEX_DRIVER,
1331 	    DDI_INTR_PRI(mlxp->mlx_intr_pri));
1332 	g->mlg_state |= MLXCX_GROUP_INIT;
1333 	mutex_enter(&g->mlg_mtx);
1334 
1335 	g->mlg_mlx = mlxp;
1336 	g->mlg_type = MLXCX_GROUP_TX;
1337 	g->mlg_port = &mlxp->mlx_ports[0];
1338 
1339 	g->mlg_nwqs = mlxp->mlx_props.mldp_tx_nrings_per_group;
1340 	g->mlg_wqs_size = g->mlg_nwqs * sizeof (mlxcx_work_queue_t);
1341 	g->mlg_wqs = kmem_zalloc(g->mlg_wqs_size, KM_SLEEP);
1342 	g->mlg_state |= MLXCX_GROUP_WQS;
1343 
1344 	g->mlg_tis.mltis_tdom = &mlxp->mlx_tdom;
1345 
1346 	if (!mlxcx_cmd_create_tis(mlxp, &g->mlg_tis)) {
1347 		mutex_exit(&g->mlg_mtx);
1348 		return (B_FALSE);
1349 	}
1350 
1351 	g->mlg_state |= MLXCX_GROUP_TIRTIS;
1352 
1353 	for (i = 0; i < g->mlg_nwqs; ++i) {
1354 		eq = NULL;
1355 		while (eq == NULL) {
1356 			eq = &mlxp->mlx_eqs[mlxp->mlx_next_eq++];
1357 			if (mlxp->mlx_next_eq >= mlxp->mlx_intr_count)
1358 				mlxp->mlx_next_eq = mlxp->mlx_intr_cq0;
1359 			if (eq->mleq_type != MLXCX_EQ_TYPE_ANY &&
1360 			    eq->mleq_type != MLXCX_EQ_TYPE_TX) {
1361 				/* Try the next one */
1362 				eq = NULL;
1363 			}
1364 		}
1365 
1366 		if (!mlxcx_cq_setup(mlxp, eq, &cq,
1367 		    mlxp->mlx_props.mldp_cq_size_shift))
1368 			return (B_FALSE);
1369 
1370 		cq->mlcq_stats = &g->mlg_port->mlp_stats;
1371 
1372 		sq = &g->mlg_wqs[i];
1373 		if (!mlxcx_sq_setup(mlxp, g->mlg_port, cq, &g->mlg_tis, sq)) {
1374 			mutex_exit(&g->mlg_mtx);
1375 			return (B_FALSE);
1376 		}
1377 		sq->mlwq_group = g;
1378 	}
1379 
1380 	mutex_exit(&g->mlg_mtx);
1381 
1382 	return (B_TRUE);
1383 }
1384 
1385 boolean_t
1386 mlxcx_tx_ring_start(mlxcx_t *mlxp, mlxcx_ring_group_t *g,
1387     mlxcx_work_queue_t *sq)
1388 {
1389 	uint_t i;
1390 	mlxcx_buffer_t *b;
1391 	mlxcx_completion_queue_t *cq;
1392 
1393 	mutex_enter(&g->mlg_mtx);
1394 
1395 	cq = sq->mlwq_cq;
1396 	ASSERT(cq != NULL);
1397 
1398 	mutex_enter(&cq->mlcq_mtx);
1399 	mutex_enter(&sq->mlwq_mtx);
1400 	if (sq->mlwq_state & MLXCX_WQ_STARTED) {
1401 		mutex_exit(&sq->mlwq_mtx);
1402 		mutex_exit(&cq->mlcq_mtx);
1403 		mutex_exit(&g->mlg_mtx);
1404 		return (B_TRUE);
1405 	}
1406 
1407 	ASSERT0(sq->mlwq_state & MLXCX_WQ_BUFFERS);
1408 	for (i = 0; i < sq->mlwq_nents; ++i) {
1409 		if (!mlxcx_buf_create_foreign(mlxp, sq->mlwq_foreign_bufs, &b))
1410 			break;
1411 		mlxcx_buf_return(mlxp, b);
1412 	}
1413 	for (i = 0; i < sq->mlwq_nents / 2; ++i) {
1414 		if (!mlxcx_buf_create_foreign(mlxp, sq->mlwq_foreign_bufs, &b))
1415 			break;
1416 		mlxcx_buf_return(mlxp, b);
1417 	}
1418 	for (i = 0; i < sq->mlwq_nents; ++i) {
1419 		if (!mlxcx_buf_create(mlxp, sq->mlwq_bufs, &b))
1420 			break;
1421 		mlxcx_buf_return(mlxp, b);
1422 	}
1423 	sq->mlwq_state |= MLXCX_WQ_BUFFERS;
1424 
1425 	mlxcx_shard_ready(sq->mlwq_bufs);
1426 	mlxcx_shard_ready(sq->mlwq_foreign_bufs);
1427 
1428 	if (!mlxcx_cmd_start_sq(mlxp, sq)) {
1429 		mutex_exit(&sq->mlwq_mtx);
1430 		mutex_exit(&cq->mlcq_mtx);
1431 		mutex_exit(&g->mlg_mtx);
1432 		return (B_FALSE);
1433 	}
1434 	g->mlg_state |= MLXCX_GROUP_RUNNING;
1435 
1436 	(void) mlxcx_sq_add_nop(mlxp, sq);
1437 
1438 	mutex_exit(&sq->mlwq_mtx);
1439 	mutex_exit(&cq->mlcq_mtx);
1440 	mutex_exit(&g->mlg_mtx);
1441 
1442 	return (B_TRUE);
1443 }
1444 
1445 static boolean_t
1446 mlxcx_sq_ring_dbell(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq, uint_t first)
1447 {
1448 	uint_t idx;
1449 	mlxcx_bf_t *bf;
1450 	ddi_fm_error_t err;
1451 	uint_t try = 0;
1452 
1453 	ASSERT3U(mlwq->mlwq_type, ==, MLXCX_WQ_TYPE_SENDQ);
1454 	ASSERT(mutex_owned(&mlwq->mlwq_mtx));
1455 
1456 	mlwq->mlwq_doorbell->mlwqd_send_counter = to_be16(mlwq->mlwq_pc);
1457 
1458 	ASSERT(mlwq->mlwq_cq != NULL);
1459 	ASSERT(mlwq->mlwq_cq->mlcq_eq != NULL);
1460 	idx = mlwq->mlwq_cq->mlcq_eq->mleq_intr_index & MLXCX_BF_PER_UAR_MASK;
1461 	bf = &mlwq->mlwq_uar->mlu_bf[idx];
1462 
1463 retry:
1464 	MLXCX_DMA_SYNC(mlwq->mlwq_doorbell_dma, DDI_DMA_SYNC_FORDEV);
1465 	ddi_fm_dma_err_get(mlwq->mlwq_doorbell_dma.mxdb_dma_handle, &err,
1466 	    DDI_FME_VERSION);
1467 	if (err.fme_status != DDI_FM_OK) {
1468 		if (try++ < mlxcx_doorbell_tries) {
1469 			ddi_fm_dma_err_clear(
1470 			    mlwq->mlwq_doorbell_dma.mxdb_dma_handle,
1471 			    DDI_FME_VERSION);
1472 			goto retry;
1473 		} else {
1474 			goto err;
1475 		}
1476 	}
1477 
1478 	mlxcx_put64(mlxp, bf->mbf_even, from_be64(
1479 	    mlwq->mlwq_bf_ent[first].mlsqbf_qwords[0]));
1480 	ddi_fm_acc_err_get(mlxp->mlx_regs_handle, &err,
1481 	    DDI_FME_VERSION);
1482 	if (err.fme_status == DDI_FM_OK)
1483 		return (B_TRUE);
1484 	if (try++ < mlxcx_doorbell_tries) {
1485 		ddi_fm_acc_err_clear(mlxp->mlx_regs_handle, DDI_FME_VERSION);
1486 		goto retry;
1487 	}
1488 
1489 err:
1490 	ddi_fm_service_impact(mlxp->mlx_dip, DDI_SERVICE_LOST);
1491 	return (B_FALSE);
1492 }
1493 
1494 boolean_t
1495 mlxcx_sq_add_nop(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq)
1496 {
1497 	uint_t index, start_pc;
1498 	mlxcx_sendq_ent_t *ent0;
1499 	ddi_fm_error_t err;
1500 
1501 	ASSERT(mutex_owned(&mlwq->mlwq_mtx));
1502 
1503 	index = mlwq->mlwq_pc & (mlwq->mlwq_nents - 1);
1504 	ent0 = &mlwq->mlwq_send_ent[index];
1505 	start_pc = mlwq->mlwq_pc;
1506 	++mlwq->mlwq_pc;
1507 	/*
1508 	 * This counter is manipulated in the interrupt handler, which
1509 	 * does not hold the mlwq_mtx, hence the atomic.
1510 	 */
1511 	atomic_inc_64(&mlwq->mlwq_wqebb_used);
1512 
1513 	bzero(ent0, sizeof (mlxcx_sendq_ent_t));
1514 	ent0->mlsqe_control.mlcs_opcode = MLXCX_WQE_OP_NOP;
1515 	ent0->mlsqe_control.mlcs_qp_or_sq = to_be24(mlwq->mlwq_num);
1516 	ent0->mlsqe_control.mlcs_wqe_index = to_be16(start_pc);
1517 
1518 	set_bits8(&ent0->mlsqe_control.mlcs_flags,
1519 	    MLXCX_SQE_FENCE_MODE, MLXCX_SQE_FENCE_NONE);
1520 	set_bits8(&ent0->mlsqe_control.mlcs_flags,
1521 	    MLXCX_SQE_COMPLETION_MODE, MLXCX_SQE_CQE_ALWAYS);
1522 
1523 	ent0->mlsqe_control.mlcs_ds = 1;
1524 
1525 	VERIFY0(ddi_dma_sync(mlwq->mlwq_dma.mxdb_dma_handle,
1526 	    (uintptr_t)ent0 - (uintptr_t)mlwq->mlwq_send_ent,
1527 	    sizeof (mlxcx_sendq_ent_t), DDI_DMA_SYNC_FORDEV));
1528 	ddi_fm_dma_err_get(mlwq->mlwq_dma.mxdb_dma_handle, &err,
1529 	    DDI_FME_VERSION);
1530 	if (err.fme_status != DDI_FM_OK) {
1531 		return (B_FALSE);
1532 	}
1533 	if (!mlxcx_sq_ring_dbell(mlxp, mlwq, index)) {
1534 		return (B_FALSE);
1535 	}
1536 	return (B_TRUE);
1537 }
1538 
1539 boolean_t
1540 mlxcx_sq_add_buffer(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq,
1541     uint8_t *inlinehdrs, size_t inlinelen, uint32_t chkflags,
1542     mlxcx_buffer_t *b0)
1543 {
1544 	uint_t index, first, ents;
1545 	mlxcx_completion_queue_t *cq;
1546 	mlxcx_sendq_ent_t *ent0;
1547 	mlxcx_sendq_extra_ent_t *ent;
1548 	mlxcx_wqe_data_seg_t *seg;
1549 	uint_t ptri, nptr;
1550 	const ddi_dma_cookie_t *c;
1551 	size_t rem;
1552 	uint64_t wqebb_used;
1553 	mlxcx_buffer_t *b;
1554 	ddi_fm_error_t err;
1555 	boolean_t rv;
1556 
1557 	ASSERT(mutex_owned(&mlwq->mlwq_mtx));
1558 	ASSERT3P(b0->mlb_tx_head, ==, b0);
1559 	ASSERT3U(b0->mlb_state, ==, MLXCX_BUFFER_ON_WQ);
1560 	cq = mlwq->mlwq_cq;
1561 
1562 	index = mlwq->mlwq_pc & (mlwq->mlwq_nents - 1);
1563 	ent0 = &mlwq->mlwq_send_ent[index];
1564 	b0->mlb_wqe_index = mlwq->mlwq_pc;
1565 	ents = 1;
1566 
1567 	first = index;
1568 
1569 	bzero(ent0, sizeof (mlxcx_sendq_ent_t));
1570 	ent0->mlsqe_control.mlcs_opcode = MLXCX_WQE_OP_SEND;
1571 	ent0->mlsqe_control.mlcs_qp_or_sq = to_be24(mlwq->mlwq_num);
1572 	ent0->mlsqe_control.mlcs_wqe_index = to_be16(b0->mlb_wqe_index);
1573 
1574 	set_bits8(&ent0->mlsqe_control.mlcs_flags,
1575 	    MLXCX_SQE_FENCE_MODE, MLXCX_SQE_FENCE_WAIT_OTHERS);
1576 	set_bits8(&ent0->mlsqe_control.mlcs_flags,
1577 	    MLXCX_SQE_COMPLETION_MODE, MLXCX_SQE_CQE_ALWAYS);
1578 
1579 	VERIFY3U(inlinelen, <=, sizeof (ent0->mlsqe_eth.mles_inline_headers));
1580 	set_bits16(&ent0->mlsqe_eth.mles_szflags,
1581 	    MLXCX_SQE_ETH_INLINE_HDR_SZ, inlinelen);
1582 	if (inlinelen > 0) {
1583 		bcopy(inlinehdrs, ent0->mlsqe_eth.mles_inline_headers,
1584 		    inlinelen);
1585 	}
1586 
1587 	ent0->mlsqe_control.mlcs_ds = offsetof(mlxcx_sendq_ent_t, mlsqe_data) /
1588 	    MLXCX_WQE_OCTOWORD;
1589 
1590 	if (chkflags & HCK_IPV4_HDRCKSUM) {
1591 		ASSERT(mlxp->mlx_caps->mlc_checksum);
1592 		set_bit8(&ent0->mlsqe_eth.mles_csflags,
1593 		    MLXCX_SQE_ETH_CSFLAG_L3_CHECKSUM);
1594 	}
1595 	if (chkflags & HCK_FULLCKSUM) {
1596 		ASSERT(mlxp->mlx_caps->mlc_checksum);
1597 		set_bit8(&ent0->mlsqe_eth.mles_csflags,
1598 		    MLXCX_SQE_ETH_CSFLAG_L4_CHECKSUM);
1599 	}
1600 
1601 	/*
1602 	 * mlwq_wqebb_used is only incremented whilst holding
1603 	 * the mlwq_mtx mutex, but it is decremented (atomically) in
1604 	 * the interrupt context *not* under mlwq_mtx mutex.
1605 	 * So, now take a snapshot of the number of used wqes which will
1606 	 * be a conistent maximum we can use whilst iterating through
1607 	 * the buffers and DMA cookies.
1608 	 */
1609 	wqebb_used = mlwq->mlwq_wqebb_used;
1610 
1611 	b = b0;
1612 	ptri = 0;
1613 	nptr = sizeof (ent0->mlsqe_data) / sizeof (mlxcx_wqe_data_seg_t);
1614 	seg = ent0->mlsqe_data;
1615 	while (b != NULL) {
1616 		rem = b->mlb_used;
1617 
1618 		c = NULL;
1619 		while (rem > 0 &&
1620 		    (c = mlxcx_dma_cookie_iter(&b->mlb_dma, c)) != NULL) {
1621 			if (ptri >= nptr) {
1622 				if ((ents + wqebb_used) >= mlwq->mlwq_nents)
1623 					return (B_FALSE);
1624 
1625 				index = (mlwq->mlwq_pc + ents) &
1626 				    (mlwq->mlwq_nents - 1);
1627 				ent = &mlwq->mlwq_send_extra_ent[index];
1628 				++ents;
1629 
1630 				seg = ent->mlsqe_data;
1631 				ptri = 0;
1632 				nptr = sizeof (ent->mlsqe_data) /
1633 				    sizeof (mlxcx_wqe_data_seg_t);
1634 			}
1635 
1636 			seg->mlds_lkey = to_be32(mlxp->mlx_rsvd_lkey);
1637 			if (c->dmac_size > rem) {
1638 				seg->mlds_byte_count = to_be32(rem);
1639 				rem = 0;
1640 			} else {
1641 				seg->mlds_byte_count = to_be32(c->dmac_size);
1642 				rem -= c->dmac_size;
1643 			}
1644 			seg->mlds_address = to_be64(c->dmac_laddress);
1645 			++seg;
1646 			++ptri;
1647 			++ent0->mlsqe_control.mlcs_ds;
1648 
1649 			ASSERT3U(ent0->mlsqe_control.mlcs_ds, <=,
1650 			    MLXCX_SQE_MAX_DS);
1651 		}
1652 
1653 		if (b == b0) {
1654 			b = list_head(&b0->mlb_tx_chain);
1655 		} else {
1656 			b = list_next(&b0->mlb_tx_chain, b);
1657 		}
1658 	}
1659 
1660 	b0->mlb_wqebbs = ents;
1661 	mlwq->mlwq_pc += ents;
1662 	atomic_add_64(&mlwq->mlwq_wqebb_used, ents);
1663 
1664 	for (; ptri < nptr; ++ptri, ++seg) {
1665 		seg->mlds_lkey = to_be32(MLXCX_NULL_LKEY);
1666 		seg->mlds_byte_count = to_be32(0);
1667 		seg->mlds_address = to_be64(0);
1668 	}
1669 
1670 	/*
1671 	 * Make sure the workqueue entry is flushed out before updating
1672 	 * the doorbell.
1673 	 * If the ring has wrapped, we need to flush the front and back.
1674 	 */
1675 	if ((first + ents) > mlwq->mlwq_nents) {
1676 		uint_t sync_cnt = mlwq->mlwq_nents - first;
1677 
1678 		VERIFY0(ddi_dma_sync(mlwq->mlwq_dma.mxdb_dma_handle,
1679 		    (uintptr_t)ent0 - (uintptr_t)mlwq->mlwq_send_ent,
1680 		    sync_cnt * sizeof (mlxcx_sendq_ent_t),
1681 		    DDI_DMA_SYNC_FORDEV));
1682 
1683 		ent0 = &mlwq->mlwq_send_ent[0];
1684 		ents -= sync_cnt;
1685 	}
1686 
1687 	VERIFY0(ddi_dma_sync(mlwq->mlwq_dma.mxdb_dma_handle,
1688 	    (uintptr_t)ent0 - (uintptr_t)mlwq->mlwq_send_ent,
1689 	    ents * sizeof (mlxcx_sendq_ent_t), DDI_DMA_SYNC_FORDEV));
1690 	ddi_fm_dma_err_get(mlwq->mlwq_dma.mxdb_dma_handle, &err,
1691 	    DDI_FME_VERSION);
1692 	if (err.fme_status != DDI_FM_OK) {
1693 		return (B_FALSE);
1694 	}
1695 
1696 	/*
1697 	 * Hold the bufmtx whilst ringing the doorbell, to prevent
1698 	 * the buffer from being moved to another list, so we can
1699 	 * safely remove it should the ring fail.
1700 	 */
1701 	mutex_enter(&cq->mlcq_bufbmtx);
1702 
1703 	list_insert_tail(&cq->mlcq_buffers_b, b0);
1704 	if ((rv = mlxcx_sq_ring_dbell(mlxp, mlwq, first))) {
1705 		atomic_inc_64(&cq->mlcq_bufcnt);
1706 	} else {
1707 		list_remove(&cq->mlcq_buffers_b, b0);
1708 	}
1709 
1710 	mutex_exit(&cq->mlcq_bufbmtx);
1711 
1712 	return (rv);
1713 }
1714 
1715 boolean_t
1716 mlxcx_rq_add_buffer(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq,
1717     mlxcx_buffer_t *buf)
1718 {
1719 	return (mlxcx_rq_add_buffers(mlxp, mlwq, &buf, 1));
1720 }
1721 
1722 boolean_t
1723 mlxcx_rq_add_buffers(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq,
1724     mlxcx_buffer_t **bufs, size_t nbufs)
1725 {
1726 	uint_t index;
1727 	mlxcx_recvq_ent_t *ent;
1728 	mlxcx_completion_queue_t *cq;
1729 	mlxcx_wqe_data_seg_t *seg;
1730 	uint_t bi, ptri;
1731 	const ddi_dma_cookie_t *c;
1732 	mlxcx_buffer_t *buf;
1733 	ddi_fm_error_t err;
1734 
1735 	ASSERT(mutex_owned(&mlwq->mlwq_mtx));
1736 	cq = mlwq->mlwq_cq;
1737 	ASSERT(mutex_owned(&cq->mlcq_mtx));
1738 
1739 	for (bi = 0; bi < nbufs; ++bi) {
1740 		buf = bufs[bi];
1741 		bufs[bi] = NULL;
1742 		ASSERT3U(buf->mlb_state, ==, MLXCX_BUFFER_ON_WQ);
1743 
1744 		index = mlwq->mlwq_pc & (mlwq->mlwq_nents - 1);
1745 		ent = &mlwq->mlwq_recv_ent[index];
1746 		buf->mlb_wqe_index = mlwq->mlwq_pc;
1747 		buf->mlb_wqebbs = 1;
1748 
1749 		++mlwq->mlwq_pc;
1750 		atomic_inc_64(&mlwq->mlwq_wqebb_used);
1751 
1752 		mutex_enter(&cq->mlcq_bufbmtx);
1753 		list_insert_tail(&cq->mlcq_buffers, buf);
1754 		atomic_inc_64(&cq->mlcq_bufcnt);
1755 		mutex_exit(&cq->mlcq_bufbmtx);
1756 
1757 		ASSERT3U(buf->mlb_dma.mxdb_ncookies, <=, MLXCX_RECVQ_MAX_PTRS);
1758 		ptri = 0;
1759 		c = NULL;
1760 		while ((c = mlxcx_dma_cookie_iter(&buf->mlb_dma, c)) != NULL) {
1761 			seg = &ent->mlrqe_data[ptri++];
1762 			seg->mlds_lkey = to_be32(mlxp->mlx_rsvd_lkey);
1763 			seg->mlds_byte_count = to_be32(c->dmac_size);
1764 			seg->mlds_address = to_be64(c->dmac_laddress);
1765 		}
1766 		/*
1767 		 * Fill any unused scatter pointers with the special null
1768 		 * value.
1769 		 */
1770 		for (; ptri < MLXCX_RECVQ_MAX_PTRS; ++ptri) {
1771 			seg = &ent->mlrqe_data[ptri];
1772 			seg->mlds_lkey = to_be32(MLXCX_NULL_LKEY);
1773 			seg->mlds_byte_count = to_be32(0);
1774 			seg->mlds_address = to_be64(0);
1775 		}
1776 
1777 		/*
1778 		 * Make sure the workqueue entry is flushed out before updating
1779 		 * the doorbell.
1780 		 */
1781 		VERIFY0(ddi_dma_sync(mlwq->mlwq_dma.mxdb_dma_handle,
1782 		    (uintptr_t)ent - (uintptr_t)mlwq->mlwq_recv_ent,
1783 		    sizeof (mlxcx_recvq_ent_t), DDI_DMA_SYNC_FORDEV));
1784 		ddi_fm_dma_err_get(mlwq->mlwq_dma.mxdb_dma_handle, &err,
1785 		    DDI_FME_VERSION);
1786 		if (err.fme_status != DDI_FM_OK) {
1787 			return (B_FALSE);
1788 		}
1789 	}
1790 
1791 	mlwq->mlwq_doorbell->mlwqd_recv_counter = to_be16(mlwq->mlwq_pc);
1792 	/*
1793 	 * Flush the CQ doorbell as well so that HW knows how many
1794 	 * completions we've consumed.
1795 	 */
1796 	MLXCX_DMA_SYNC(cq->mlcq_doorbell_dma, DDI_DMA_SYNC_FORDEV);
1797 	ddi_fm_dma_err_get(cq->mlcq_doorbell_dma.mxdb_dma_handle, &err,
1798 	    DDI_FME_VERSION);
1799 	if (err.fme_status != DDI_FM_OK) {
1800 		return (B_FALSE);
1801 	}
1802 	MLXCX_DMA_SYNC(mlwq->mlwq_doorbell_dma, DDI_DMA_SYNC_FORDEV);
1803 	ddi_fm_dma_err_get(mlwq->mlwq_doorbell_dma.mxdb_dma_handle, &err,
1804 	    DDI_FME_VERSION);
1805 	if (err.fme_status != DDI_FM_OK) {
1806 		return (B_FALSE);
1807 	}
1808 	return (B_TRUE);
1809 }
1810 
1811 static void
1812 mlxcx_rq_refill_task(void *arg)
1813 {
1814 	mlxcx_work_queue_t *wq = arg;
1815 	mlxcx_completion_queue_t *cq = wq->mlwq_cq;
1816 	mlxcx_t *mlxp = wq->mlwq_mlx;
1817 	mlxcx_buf_shard_t *s = wq->mlwq_bufs;
1818 	boolean_t refill, draining;
1819 
1820 	do {
1821 		/*
1822 		 * Wait here until one of 3 conditions:
1823 		 * 1. The shard is draining, or
1824 		 * 2. There are buffers on the free list, or
1825 		 * 3. The WQ is being shut down.
1826 		 */
1827 		mutex_enter(&s->mlbs_mtx);
1828 		while (s->mlbs_state != MLXCX_SHARD_DRAINING &&
1829 		    list_is_empty(&s->mlbs_free) &&
1830 		    (cq->mlcq_state & MLXCX_CQ_TEARDOWN) == 0) {
1831 			cv_wait(&s->mlbs_free_nonempty, &s->mlbs_mtx);
1832 		}
1833 
1834 		draining = (s->mlbs_state == MLXCX_SHARD_DRAINING);
1835 		mutex_exit(&s->mlbs_mtx);
1836 
1837 		mutex_enter(&cq->mlcq_mtx);
1838 		mutex_enter(&wq->mlwq_mtx);
1839 
1840 		if (draining || (cq->mlcq_state & MLXCX_CQ_TEARDOWN) != 0) {
1841 			refill = B_FALSE;
1842 			wq->mlwq_state &= ~MLXCX_WQ_REFILLING;
1843 		} else {
1844 			mlxcx_rq_refill(mlxp, wq);
1845 
1846 			if (cq->mlcq_bufcnt < MLXCX_RQ_REFILL_STEP) {
1847 				refill = B_TRUE;
1848 			} else {
1849 				refill = B_FALSE;
1850 				wq->mlwq_state &= ~MLXCX_WQ_REFILLING;
1851 			}
1852 		}
1853 
1854 		mutex_exit(&wq->mlwq_mtx);
1855 		mutex_exit(&cq->mlcq_mtx);
1856 	} while (refill);
1857 }
1858 
1859 void
1860 mlxcx_rq_refill(mlxcx_t *mlxp, mlxcx_work_queue_t *mlwq)
1861 {
1862 	size_t target, current, want, done, n;
1863 	mlxcx_completion_queue_t *cq;
1864 	mlxcx_ring_group_t *g;
1865 	mlxcx_buffer_t *b[MLXCX_RQ_REFILL_STEP];
1866 	uint_t i;
1867 
1868 	ASSERT(mutex_owned(&mlwq->mlwq_mtx));
1869 	cq = mlwq->mlwq_cq;
1870 	ASSERT(mutex_owned(&cq->mlcq_mtx));
1871 
1872 	ASSERT(mlwq->mlwq_state & MLXCX_WQ_BUFFERS);
1873 
1874 	target = mlwq->mlwq_nents - MLXCX_RQ_REFILL_STEP;
1875 	cq = mlwq->mlwq_cq;
1876 
1877 	if ((mlwq->mlwq_state & MLXCX_WQ_STARTED) == 0)
1878 		return;
1879 
1880 	if ((cq->mlcq_state & MLXCX_CQ_TEARDOWN) != 0)
1881 		return;
1882 
1883 	current = cq->mlcq_bufcnt;
1884 
1885 	if (current >= target - MLXCX_RQ_REFILL_STEP)
1886 		return;
1887 
1888 	want = target - current;
1889 	done = 0;
1890 
1891 	while (!(mlwq->mlwq_state & MLXCX_WQ_TEARDOWN) && done < want) {
1892 		n = mlxcx_buf_take_n(mlxp, mlwq, b, MLXCX_RQ_REFILL_STEP);
1893 		if (n == 0) {
1894 			/*
1895 			 * We didn't get any buffers from the free queue.
1896 			 * It might not be an issue, schedule a taskq
1897 			 * to wait for free buffers if the completion
1898 			 * queue is low.
1899 			 */
1900 			if (current < MLXCX_RQ_REFILL_STEP &&
1901 			    (mlwq->mlwq_state & MLXCX_WQ_REFILLING) == 0) {
1902 				mlwq->mlwq_state |= MLXCX_WQ_REFILLING;
1903 				g = mlwq->mlwq_group;
1904 				taskq_dispatch_ent(g->mlg_refill_tq,
1905 				    mlxcx_rq_refill_task, mlwq, TQ_NOSLEEP,
1906 				    &mlwq->mlwq_tqe);
1907 			}
1908 
1909 			return;
1910 		}
1911 
1912 		if ((mlwq->mlwq_state & MLXCX_WQ_TEARDOWN) != 0) {
1913 			for (i = 0; i < n; ++i)
1914 				mlxcx_buf_return(mlxp, b[i]);
1915 			return;
1916 		}
1917 		if (!mlxcx_rq_add_buffers(mlxp, mlwq, b, n)) {
1918 			/*
1919 			 * mlxcx_rq_add_buffers NULLs out the buffers as it
1920 			 * enqueues them, so any that are non-NULL we have to
1921 			 * free now. The others now belong to the WQ, even if
1922 			 * we failed.
1923 			 */
1924 			for (i = 0; i < n; ++i) {
1925 				if (b[i] != NULL) {
1926 					mlxcx_buf_return(mlxp, b[i]);
1927 				}
1928 			}
1929 			return;
1930 		}
1931 		done += n;
1932 	}
1933 }
1934 
1935 static const char *
1936 mlxcx_cq_err_syndrome_string(mlxcx_cq_error_syndrome_t sy)
1937 {
1938 	switch (sy) {
1939 	case MLXCX_CQ_ERR_LOCAL_LENGTH:
1940 		return ("LOCAL_LENGTH");
1941 	case MLXCX_CQ_ERR_LOCAL_QP_OP:
1942 		return ("LOCAL_QP_OP");
1943 	case MLXCX_CQ_ERR_LOCAL_PROTECTION:
1944 		return ("LOCAL_PROTECTION");
1945 	case MLXCX_CQ_ERR_WR_FLUSHED:
1946 		return ("WR_FLUSHED");
1947 	case MLXCX_CQ_ERR_MEM_WINDOW_BIND:
1948 		return ("MEM_WINDOW_BIND");
1949 	case MLXCX_CQ_ERR_BAD_RESPONSE:
1950 		return ("BAD_RESPONSE");
1951 	case MLXCX_CQ_ERR_LOCAL_ACCESS:
1952 		return ("LOCAL_ACCESS");
1953 	case MLXCX_CQ_ERR_XPORT_RETRY_CTR:
1954 		return ("XPORT_RETRY_CTR");
1955 	case MLXCX_CQ_ERR_RNR_RETRY_CTR:
1956 		return ("RNR_RETRY_CTR");
1957 	case MLXCX_CQ_ERR_ABORTED:
1958 		return ("ABORTED");
1959 	default:
1960 		return ("UNKNOWN");
1961 	}
1962 }
1963 
1964 static void
1965 mlxcx_fm_cqe_ereport(mlxcx_t *mlxp, mlxcx_completion_queue_t *mlcq,
1966     mlxcx_completionq_error_ent_t *ent)
1967 {
1968 	uint64_t ena;
1969 	char buf[FM_MAX_CLASS];
1970 	const char *name = mlxcx_cq_err_syndrome_string(ent->mlcqee_syndrome);
1971 
1972 	if (!DDI_FM_EREPORT_CAP(mlxp->mlx_fm_caps))
1973 		return;
1974 
1975 	(void) snprintf(buf, FM_MAX_CLASS, "%s.%s",
1976 	    MLXCX_FM_SERVICE_MLXCX, "cqe.err");
1977 	ena = fm_ena_generate(0, FM_ENA_FMT1);
1978 
1979 	ddi_fm_ereport_post(mlxp->mlx_dip, buf, ena, DDI_NOSLEEP,
1980 	    FM_VERSION, DATA_TYPE_UINT8, FM_EREPORT_VERS0,
1981 	    "syndrome", DATA_TYPE_STRING, name,
1982 	    "syndrome_num", DATA_TYPE_UINT8, ent->mlcqee_syndrome,
1983 	    "vendor_syndrome", DATA_TYPE_UINT8,
1984 	    ent->mlcqee_vendor_error_syndrome,
1985 	    "wqe_counter", DATA_TYPE_UINT16, from_be16(ent->mlcqee_wqe_counter),
1986 	    "wq_type", DATA_TYPE_STRING,
1987 	    (mlcq->mlcq_wq->mlwq_type == MLXCX_WQ_TYPE_SENDQ) ? "send": "recv",
1988 	    "cq_num", DATA_TYPE_UINT32, mlcq->mlcq_num,
1989 	    "wq_num", DATA_TYPE_UINT32, mlcq->mlcq_wq->mlwq_num,
1990 	    NULL);
1991 	ddi_fm_service_impact(mlxp->mlx_dip, DDI_SERVICE_DEGRADED);
1992 }
1993 
1994 void
1995 mlxcx_tx_completion(mlxcx_t *mlxp, mlxcx_completion_queue_t *mlcq,
1996     mlxcx_completionq_ent_t *ent, mlxcx_buffer_t *buf)
1997 {
1998 	ASSERT(mutex_owned(&mlcq->mlcq_mtx));
1999 	if (ent->mlcqe_opcode == MLXCX_CQE_OP_REQ_ERR) {
2000 		mlxcx_completionq_error_ent_t *eent =
2001 		    (mlxcx_completionq_error_ent_t *)ent;
2002 		mlxcx_fm_cqe_ereport(mlxp, mlcq, eent);
2003 		mlxcx_buf_return_chain(mlxp, buf, B_FALSE);
2004 		mutex_enter(&mlcq->mlcq_wq->mlwq_mtx);
2005 		mlxcx_check_sq(mlxp, mlcq->mlcq_wq);
2006 		mutex_exit(&mlcq->mlcq_wq->mlwq_mtx);
2007 		return;
2008 	}
2009 
2010 	if (ent->mlcqe_opcode != MLXCX_CQE_OP_REQ) {
2011 		mlxcx_warn(mlxp, "!got weird cq opcode: %x", ent->mlcqe_opcode);
2012 		mlxcx_buf_return_chain(mlxp, buf, B_FALSE);
2013 		return;
2014 	}
2015 
2016 	if (ent->mlcqe_send_wqe_opcode != MLXCX_WQE_OP_SEND) {
2017 		mlxcx_warn(mlxp, "!got weird cq wqe opcode: %x",
2018 		    ent->mlcqe_send_wqe_opcode);
2019 		mlxcx_buf_return_chain(mlxp, buf, B_FALSE);
2020 		return;
2021 	}
2022 
2023 	if (ent->mlcqe_format != MLXCX_CQE_FORMAT_BASIC) {
2024 		mlxcx_warn(mlxp, "!got weird cq format: %x", ent->mlcqe_format);
2025 		mlxcx_buf_return_chain(mlxp, buf, B_FALSE);
2026 		return;
2027 	}
2028 
2029 	mlxcx_buf_return_chain(mlxp, buf, B_FALSE);
2030 }
2031 
2032 mblk_t *
2033 mlxcx_rx_completion(mlxcx_t *mlxp, mlxcx_completion_queue_t *mlcq,
2034     mlxcx_completionq_ent_t *ent, mlxcx_buffer_t *buf)
2035 {
2036 	uint32_t chkflags = 0;
2037 	uint_t wqe_index;
2038 	ddi_fm_error_t err;
2039 
2040 	ASSERT(mutex_owned(&mlcq->mlcq_mtx));
2041 
2042 	if (ent->mlcqe_opcode == MLXCX_CQE_OP_RESP_ERR) {
2043 		mlxcx_completionq_error_ent_t *eent =
2044 		    (mlxcx_completionq_error_ent_t *)ent;
2045 		mlxcx_fm_cqe_ereport(mlxp, mlcq, eent);
2046 		mlxcx_buf_return(mlxp, buf);
2047 		mutex_enter(&mlcq->mlcq_wq->mlwq_mtx);
2048 		mlxcx_check_rq(mlxp, mlcq->mlcq_wq);
2049 		mutex_exit(&mlcq->mlcq_wq->mlwq_mtx);
2050 		return (NULL);
2051 	}
2052 
2053 	if (ent->mlcqe_opcode != MLXCX_CQE_OP_RESP) {
2054 		mlxcx_warn(mlxp, "!got weird cq opcode: %x", ent->mlcqe_opcode);
2055 		mlxcx_buf_return(mlxp, buf);
2056 		return (NULL);
2057 	}
2058 
2059 	if (ent->mlcqe_format != MLXCX_CQE_FORMAT_BASIC) {
2060 		mlxcx_warn(mlxp, "!got weird cq format: %x", ent->mlcqe_format);
2061 		mlxcx_buf_return(mlxp, buf);
2062 		return (NULL);
2063 	}
2064 
2065 	if (ent->mlcqe_rx_drop_counter > 0) {
2066 		atomic_add_64(&mlcq->mlcq_stats->mlps_rx_drops,
2067 		    ent->mlcqe_rx_drop_counter);
2068 	}
2069 
2070 	MLXCX_DMA_SYNC(buf->mlb_dma, DDI_DMA_SYNC_FORCPU);
2071 	ddi_fm_dma_err_get(buf->mlb_dma.mxdb_dma_handle, &err,
2072 	    DDI_FME_VERSION);
2073 	if (err.fme_status != DDI_FM_OK) {
2074 		ddi_fm_dma_err_clear(buf->mlb_dma.mxdb_dma_handle,
2075 		    DDI_FME_VERSION);
2076 		mlxcx_buf_return(mlxp, buf);
2077 		return (NULL);
2078 	}
2079 
2080 	/*
2081 	 * mlxcx_buf_loan() will set mlb_wqe_index to zero.
2082 	 * Remember it for later.
2083 	 */
2084 	wqe_index = buf->mlb_wqe_index;
2085 
2086 	if (!mlxcx_buf_loan(mlxp, buf)) {
2087 		mlxcx_buf_return(mlxp, buf);
2088 		return (NULL);
2089 	}
2090 
2091 	buf->mlb_mp->b_next = NULL;
2092 	buf->mlb_mp->b_cont = NULL;
2093 	buf->mlb_mp->b_wptr = buf->mlb_mp->b_rptr +
2094 	    from_be32(ent->mlcqe_byte_cnt);
2095 
2096 	if (get_bit8(ent->mlcqe_csflags, MLXCX_CQE_CSFLAGS_L4_OK)) {
2097 		chkflags |= HCK_FULLCKSUM_OK;
2098 	}
2099 	if (get_bit8(ent->mlcqe_csflags, MLXCX_CQE_CSFLAGS_L3_OK)) {
2100 		chkflags |= HCK_IPV4_HDRCKSUM_OK;
2101 	}
2102 	if (chkflags != 0) {
2103 		mac_hcksum_set(buf->mlb_mp, 0, 0, 0,
2104 		    from_be16(ent->mlcqe_checksum), chkflags);
2105 	}
2106 
2107 	/*
2108 	 * Don't check if a refill is needed on every single completion,
2109 	 * since checking involves taking the RQ lock.
2110 	 */
2111 	if ((wqe_index & 0x7) == 0) {
2112 		mlxcx_work_queue_t *wq = mlcq->mlcq_wq;
2113 		ASSERT(wq != NULL);
2114 		mutex_enter(&wq->mlwq_mtx);
2115 		if (!(wq->mlwq_state & MLXCX_WQ_TEARDOWN))
2116 			mlxcx_rq_refill(mlxp, wq);
2117 		mutex_exit(&wq->mlwq_mtx);
2118 	}
2119 
2120 	return (buf->mlb_mp);
2121 }
2122 
2123 static void
2124 mlxcx_buf_mp_return(caddr_t arg)
2125 {
2126 	mlxcx_buffer_t *b = (mlxcx_buffer_t *)arg;
2127 	mlxcx_t *mlxp = b->mlb_mlx;
2128 
2129 	/* The mblk has been used now, so NULL it out. */
2130 	b->mlb_mp = NULL;
2131 
2132 	if (b->mlb_state == MLXCX_BUFFER_ON_LOAN)
2133 		mlxcx_buf_return(mlxp, b);
2134 }
2135 
2136 boolean_t
2137 mlxcx_buf_create(mlxcx_t *mlxp, mlxcx_buf_shard_t *shard, mlxcx_buffer_t **bp)
2138 {
2139 	mlxcx_buffer_t *b;
2140 	ddi_device_acc_attr_t acc;
2141 	ddi_dma_attr_t attr;
2142 	boolean_t ret;
2143 
2144 	b = kmem_cache_alloc(mlxp->mlx_bufs_cache, KM_SLEEP);
2145 	b->mlb_shard = shard;
2146 	b->mlb_foreign = B_FALSE;
2147 
2148 	mlxcx_dma_acc_attr(mlxp, &acc);
2149 	mlxcx_dma_buf_attr(mlxp, &attr);
2150 
2151 	ret = mlxcx_dma_alloc_offset(mlxp, &b->mlb_dma, &attr, &acc,
2152 	    B_FALSE, mlxp->mlx_ports[0].mlp_mtu, 2, B_TRUE);
2153 	if (!ret) {
2154 		kmem_cache_free(mlxp->mlx_bufs_cache, b);
2155 		return (B_FALSE);
2156 	}
2157 
2158 	b->mlb_frtn.free_func = mlxcx_buf_mp_return;
2159 	b->mlb_frtn.free_arg = (caddr_t)b;
2160 	b->mlb_mp = desballoc((unsigned char *)b->mlb_dma.mxdb_va,
2161 	    b->mlb_dma.mxdb_len, 0, &b->mlb_frtn);
2162 
2163 	*bp = b;
2164 
2165 	return (B_TRUE);
2166 }
2167 
2168 boolean_t
2169 mlxcx_buf_create_foreign(mlxcx_t *mlxp, mlxcx_buf_shard_t *shard,
2170     mlxcx_buffer_t **bp)
2171 {
2172 	mlxcx_buffer_t *b;
2173 	ddi_dma_attr_t attr;
2174 	boolean_t ret;
2175 
2176 	b = kmem_cache_alloc(mlxp->mlx_bufs_cache, KM_SLEEP);
2177 	b->mlb_shard = shard;
2178 	b->mlb_foreign = B_TRUE;
2179 
2180 	mlxcx_dma_buf_attr(mlxp, &attr);
2181 
2182 	ret = mlxcx_dma_init(mlxp, &b->mlb_dma, &attr, B_TRUE);
2183 	if (!ret) {
2184 		kmem_cache_free(mlxp->mlx_bufs_cache, b);
2185 		return (B_FALSE);
2186 	}
2187 
2188 	*bp = b;
2189 
2190 	return (B_TRUE);
2191 }
2192 
2193 static mlxcx_buffer_t *
2194 mlxcx_buf_take_foreign(mlxcx_t *mlxp, mlxcx_work_queue_t *wq)
2195 {
2196 	mlxcx_buffer_t *b;
2197 	mlxcx_buf_shard_t *s = wq->mlwq_foreign_bufs;
2198 
2199 	mutex_enter(&s->mlbs_mtx);
2200 	if (s->mlbs_state != MLXCX_SHARD_READY) {
2201 		mutex_exit(&s->mlbs_mtx);
2202 		return (NULL);
2203 	}
2204 
2205 	if ((b = list_remove_head(&s->mlbs_free)) != NULL) {
2206 		ASSERT3U(b->mlb_state, ==, MLXCX_BUFFER_FREE);
2207 		ASSERT(b->mlb_foreign);
2208 		b->mlb_state = MLXCX_BUFFER_ON_WQ;
2209 		list_insert_tail(&s->mlbs_busy, b);
2210 	}
2211 	mutex_exit(&s->mlbs_mtx);
2212 
2213 	return (b);
2214 }
2215 
2216 static mlxcx_buffer_t *
2217 mlxcx_copy_data(mlxcx_t *mlxp, mlxcx_work_queue_t *wq, uint8_t *rptr, size_t sz)
2218 {
2219 	ddi_fm_error_t err;
2220 	mlxcx_buffer_t *b;
2221 	uint_t attempts = 0;
2222 
2223 copyb:
2224 	if ((b = mlxcx_buf_take(mlxp, wq)) == NULL)
2225 		return (NULL);
2226 
2227 	ASSERT3U(b->mlb_dma.mxdb_len, >=, sz);
2228 	bcopy(rptr, b->mlb_dma.mxdb_va, sz);
2229 
2230 	MLXCX_DMA_SYNC(b->mlb_dma, DDI_DMA_SYNC_FORDEV);
2231 
2232 	ddi_fm_dma_err_get(b->mlb_dma.mxdb_dma_handle, &err,
2233 	    DDI_FME_VERSION);
2234 	if (err.fme_status != DDI_FM_OK) {
2235 		ddi_fm_dma_err_clear(b->mlb_dma.mxdb_dma_handle,
2236 		    DDI_FME_VERSION);
2237 		mlxcx_buf_return(mlxp, b);
2238 		if (++attempts > MLXCX_BUF_BIND_MAX_ATTEMTPS) {
2239 			return (NULL);
2240 		}
2241 		goto copyb;
2242 	}
2243 
2244 	return (b);
2245 }
2246 
2247 static mlxcx_buffer_t *
2248 mlxcx_bind_or_copy_mblk(mlxcx_t *mlxp, mlxcx_work_queue_t *wq,
2249     mblk_t *mp, size_t off)
2250 {
2251 	mlxcx_buffer_t *b;
2252 	uint8_t *rptr;
2253 	size_t sz;
2254 	boolean_t ret;
2255 
2256 	rptr = mp->b_rptr;
2257 	sz = MBLKL(mp);
2258 
2259 #ifdef DEBUG
2260 	if (off > 0) {
2261 		ASSERT3U(off, <, sz);
2262 	}
2263 #endif
2264 
2265 	rptr += off;
2266 	sz -= off;
2267 
2268 	if (sz < mlxp->mlx_props.mldp_tx_bind_threshold) {
2269 		b = mlxcx_copy_data(mlxp, wq, rptr, sz);
2270 	} else {
2271 		b = mlxcx_buf_take_foreign(mlxp, wq);
2272 		if (b == NULL)
2273 			return (NULL);
2274 
2275 		ret = mlxcx_dma_bind_mblk(mlxp, &b->mlb_dma, mp, off,
2276 		    B_FALSE);
2277 
2278 		if (!ret) {
2279 			mlxcx_buf_return(mlxp, b);
2280 
2281 			b = mlxcx_copy_data(mlxp, wq, rptr, sz);
2282 		}
2283 	}
2284 
2285 	return (b);
2286 }
2287 
2288 uint_t
2289 mlxcx_buf_bind_or_copy(mlxcx_t *mlxp, mlxcx_work_queue_t *wq,
2290     mblk_t *mpb, size_t off, mlxcx_buffer_t **bp)
2291 {
2292 	mlxcx_buffer_t *b, *b0 = NULL;
2293 	boolean_t first = B_TRUE;
2294 	mblk_t *mp;
2295 	size_t offset = off;
2296 	size_t ncookies = 0;
2297 	uint_t count = 0;
2298 
2299 	for (mp = mpb; mp != NULL && ncookies <= MLXCX_SQE_MAX_PTRS;
2300 	    mp = mp->b_cont) {
2301 		b = mlxcx_bind_or_copy_mblk(mlxp, wq, mp, offset);
2302 		if (b == NULL)
2303 			goto failed;
2304 
2305 		ncookies += b->mlb_dma.mxdb_ncookies;
2306 
2307 		if (first)
2308 			b0 = b;
2309 
2310 		if (!first)
2311 			b->mlb_state = MLXCX_BUFFER_ON_CHAIN;
2312 
2313 		b->mlb_tx_mp = mp;
2314 		b->mlb_tx_head = b0;
2315 		b->mlb_used = MBLKL(mp) - offset;
2316 
2317 		if (!first)
2318 			list_insert_tail(&b0->mlb_tx_chain, b);
2319 		first = B_FALSE;
2320 		offset = 0;
2321 
2322 		count++;
2323 	}
2324 
2325 	/*
2326 	 * The chain of mblks has resulted in too many cookies for
2327 	 * a single message. This is unusual, so take the hit to tidy
2328 	 * up, do a pullup to a single mblk and allocate the requisite
2329 	 * buf.
2330 	 */
2331 	if (ncookies > MLXCX_SQE_MAX_PTRS) {
2332 		DTRACE_PROBE4(pullup, mlxcx_t *, mlxp, mlxcx_work_queue_t *, wq,
2333 		    mblk_t *, mpb, size_t, ncookies);
2334 
2335 		if (b0 != NULL)
2336 			mlxcx_buf_return_chain(mlxp, b0, B_TRUE);
2337 
2338 		if ((mp = msgpullup(mpb, -1)) == NULL)
2339 			return (0);
2340 
2341 		b0 = mlxcx_bind_or_copy_mblk(mlxp, wq, mp, off);
2342 		if (b0 == NULL) {
2343 			freemsg(mp);
2344 			return (0);
2345 		}
2346 		freemsg(mpb);
2347 
2348 		b0->mlb_tx_mp = mp;
2349 		b0->mlb_tx_head = b0;
2350 		b0->mlb_used = MBLKL(mp) - off;
2351 
2352 		count = 1;
2353 	}
2354 
2355 	*bp = b0;
2356 
2357 	return (count);
2358 
2359 failed:
2360 	if (b0 != NULL)
2361 		mlxcx_buf_return_chain(mlxp, b0, B_TRUE);
2362 
2363 	return (0);
2364 }
2365 
2366 mlxcx_buffer_t *
2367 mlxcx_buf_take(mlxcx_t *mlxp, mlxcx_work_queue_t *wq)
2368 {
2369 	mlxcx_buffer_t *b;
2370 	mlxcx_buf_shard_t *s = wq->mlwq_bufs;
2371 
2372 	mutex_enter(&s->mlbs_mtx);
2373 	if (s->mlbs_state != MLXCX_SHARD_READY) {
2374 		mutex_exit(&s->mlbs_mtx);
2375 		return (NULL);
2376 	}
2377 
2378 	if ((b = list_remove_head(&s->mlbs_free)) != NULL) {
2379 		ASSERT3U(b->mlb_state, ==, MLXCX_BUFFER_FREE);
2380 		b->mlb_state = MLXCX_BUFFER_ON_WQ;
2381 		list_insert_tail(&s->mlbs_busy, b);
2382 	}
2383 	mutex_exit(&s->mlbs_mtx);
2384 
2385 	return (b);
2386 }
2387 
2388 size_t
2389 mlxcx_buf_take_n(mlxcx_t *mlxp, mlxcx_work_queue_t *wq, mlxcx_buffer_t **bp,
2390     size_t nbufs)
2391 {
2392 	mlxcx_buffer_t *b;
2393 	size_t done = 0;
2394 	mlxcx_buf_shard_t *s;
2395 
2396 	s = wq->mlwq_bufs;
2397 
2398 	mutex_enter(&s->mlbs_mtx);
2399 	if (s->mlbs_state != MLXCX_SHARD_READY) {
2400 		mutex_exit(&s->mlbs_mtx);
2401 		return (0);
2402 	}
2403 
2404 	while (done < nbufs && (b = list_remove_head(&s->mlbs_free)) != NULL) {
2405 		ASSERT3U(b->mlb_state, ==, MLXCX_BUFFER_FREE);
2406 		b->mlb_state = MLXCX_BUFFER_ON_WQ;
2407 		list_insert_tail(&s->mlbs_busy, b);
2408 		bp[done++] = b;
2409 	}
2410 	mutex_exit(&s->mlbs_mtx);
2411 	return (done);
2412 }
2413 
2414 boolean_t
2415 mlxcx_buf_loan(mlxcx_t *mlxp, mlxcx_buffer_t *b)
2416 {
2417 	mlxcx_buf_shard_t *s = b->mlb_shard;
2418 
2419 	VERIFY3U(b->mlb_state, ==, MLXCX_BUFFER_ON_WQ);
2420 	ASSERT3P(b->mlb_mlx, ==, mlxp);
2421 
2422 	if (b->mlb_mp == NULL) {
2423 		b->mlb_mp = desballoc((unsigned char *)b->mlb_dma.mxdb_va,
2424 		    b->mlb_dma.mxdb_len, 0, &b->mlb_frtn);
2425 		if (b->mlb_mp == NULL)
2426 			return (B_FALSE);
2427 	}
2428 
2429 	b->mlb_state = MLXCX_BUFFER_ON_LOAN;
2430 	b->mlb_wqe_index = 0;
2431 
2432 	mutex_enter(&s->mlbs_mtx);
2433 	list_remove(&s->mlbs_busy, b);
2434 	list_insert_tail(&s->mlbs_loaned, b);
2435 	mutex_exit(&s->mlbs_mtx);
2436 
2437 	return (B_TRUE);
2438 }
2439 
2440 void
2441 mlxcx_buf_return_chain(mlxcx_t *mlxp, mlxcx_buffer_t *b0, boolean_t keepmp)
2442 {
2443 	mlxcx_buffer_t *b;
2444 
2445 	if (b0->mlb_tx_head != b0) {
2446 		mlxcx_buf_return(mlxp, b0);
2447 		return;
2448 	}
2449 
2450 	while ((b = list_head(&b0->mlb_tx_chain)) != NULL) {
2451 		mlxcx_buf_return(mlxp, b);
2452 	}
2453 	if (keepmp) {
2454 		b0->mlb_tx_mp = NULL;
2455 		b0->mlb_tx_head = NULL;
2456 	}
2457 	mlxcx_buf_return(mlxp, b0);
2458 }
2459 
2460 void
2461 mlxcx_buf_return(mlxcx_t *mlxp, mlxcx_buffer_t *b)
2462 {
2463 	mlxcx_buffer_state_t oldstate = b->mlb_state;
2464 	mlxcx_buffer_t *txhead = b->mlb_tx_head;
2465 	mlxcx_buf_shard_t *s = b->mlb_shard;
2466 	mblk_t *mp = b->mlb_tx_mp;
2467 
2468 	VERIFY3U(oldstate, !=, MLXCX_BUFFER_FREE);
2469 	ASSERT3P(b->mlb_mlx, ==, mlxp);
2470 
2471 	/*
2472 	 * The mlbs_mtx held below is a heavily contended lock, so it is
2473 	 * imperative we do as much of the buffer clean up outside the lock
2474 	 * as is possible.
2475 	 */
2476 	b->mlb_state = MLXCX_BUFFER_FREE;
2477 	b->mlb_wqe_index = 0;
2478 	b->mlb_tx_head = NULL;
2479 	b->mlb_tx_mp = NULL;
2480 	b->mlb_used = 0;
2481 	b->mlb_wqebbs = 0;
2482 	ASSERT(list_is_empty(&b->mlb_tx_chain));
2483 
2484 	if (b->mlb_foreign) {
2485 		if (b->mlb_dma.mxdb_flags & MLXCX_DMABUF_BOUND) {
2486 			mlxcx_dma_unbind(mlxp, &b->mlb_dma);
2487 		}
2488 	}
2489 
2490 	mutex_enter(&s->mlbs_mtx);
2491 	switch (oldstate) {
2492 	case MLXCX_BUFFER_INIT:
2493 		break;
2494 	case MLXCX_BUFFER_ON_WQ:
2495 		list_remove(&s->mlbs_busy, b);
2496 		break;
2497 	case MLXCX_BUFFER_ON_LOAN:
2498 		ASSERT(!b->mlb_foreign);
2499 		list_remove(&s->mlbs_loaned, b);
2500 		if (s->mlbs_state == MLXCX_SHARD_DRAINING) {
2501 			/*
2502 			 * When we're draining, Eg during mac_stop(),
2503 			 * we destroy the buffer immediately rather than
2504 			 * recycling it. Otherwise we risk leaving it
2505 			 * on the free list and leaking it.
2506 			 */
2507 			list_insert_tail(&s->mlbs_free, b);
2508 			mlxcx_buf_destroy(mlxp, b);
2509 			/*
2510 			 * Teardown might be waiting for loaned list to empty.
2511 			 */
2512 			cv_broadcast(&s->mlbs_free_nonempty);
2513 			mutex_exit(&s->mlbs_mtx);
2514 			return;
2515 		}
2516 		break;
2517 	case MLXCX_BUFFER_FREE:
2518 		VERIFY(0);
2519 		break;
2520 	case MLXCX_BUFFER_ON_CHAIN:
2521 		ASSERT(txhead != NULL);
2522 		list_remove(&txhead->mlb_tx_chain, b);
2523 		list_remove(&s->mlbs_busy, b);
2524 		break;
2525 	}
2526 
2527 	list_insert_tail(&s->mlbs_free, b);
2528 	cv_broadcast(&s->mlbs_free_nonempty);
2529 
2530 	mutex_exit(&s->mlbs_mtx);
2531 
2532 	/*
2533 	 * For TX chain heads, free the mblk_t after we let go of the lock.
2534 	 * This might be a borrowed buf that we in turn loaned to MAC, in which
2535 	 * case calling freemsg() on it will re-enter this very function -- so
2536 	 * we better not be holding the lock!
2537 	 */
2538 	if (txhead == b)
2539 		freemsg(mp);
2540 }
2541 
2542 void
2543 mlxcx_buf_destroy(mlxcx_t *mlxp, mlxcx_buffer_t *b)
2544 {
2545 	mlxcx_buf_shard_t *s = b->mlb_shard;
2546 
2547 	VERIFY(b->mlb_state == MLXCX_BUFFER_FREE ||
2548 	    b->mlb_state == MLXCX_BUFFER_INIT);
2549 	ASSERT(mutex_owned(&s->mlbs_mtx));
2550 
2551 	if (b->mlb_state == MLXCX_BUFFER_FREE)
2552 		list_remove(&s->mlbs_free, b);
2553 
2554 	/*
2555 	 * This is going back to the kmem cache, so it needs to be set up in
2556 	 * the same way we expect a new buffer to come out (state INIT, other
2557 	 * fields NULL'd)
2558 	 */
2559 	b->mlb_state = MLXCX_BUFFER_INIT;
2560 	b->mlb_shard = NULL;
2561 	if (b->mlb_mp != NULL) {
2562 		freeb(b->mlb_mp);
2563 		ASSERT(b->mlb_mp == NULL);
2564 	}
2565 	mlxcx_dma_free(&b->mlb_dma);
2566 	ASSERT(list_is_empty(&b->mlb_tx_chain));
2567 
2568 	kmem_cache_free(mlxp->mlx_bufs_cache, b);
2569 }
2570 
2571 void
2572 mlxcx_shard_ready(mlxcx_buf_shard_t *s)
2573 {
2574 	mutex_enter(&s->mlbs_mtx);
2575 	s->mlbs_state = MLXCX_SHARD_READY;
2576 	mutex_exit(&s->mlbs_mtx);
2577 }
2578 
2579 void
2580 mlxcx_shard_draining(mlxcx_buf_shard_t *s)
2581 {
2582 	mutex_enter(&s->mlbs_mtx);
2583 	s->mlbs_state = MLXCX_SHARD_DRAINING;
2584 	cv_broadcast(&s->mlbs_free_nonempty);
2585 	mutex_exit(&s->mlbs_mtx);
2586 }
2587