xref: /illumos-gate/usr/src/uts/common/io/nxge/nxge_hio.c (revision 0dc2366f7b9f9f36e10909b1e95edbf2a261c2ac)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 /*
28  * nxge_hio.c
29  *
30  * This file manages the virtualization resources for Neptune
31  * devices.  That is, it implements a hybrid I/O (HIO) approach in the
32  * Solaris kernel, whereby a guest domain on an LDOMs server may
33  * request & use hardware resources from the service domain.
34  *
35  */
36 
37 #include <sys/mac_provider.h>
38 #include <sys/nxge/nxge_impl.h>
39 #include <sys/nxge/nxge_fzc.h>
40 #include <sys/nxge/nxge_rxdma.h>
41 #include <sys/nxge/nxge_txdma.h>
42 #include <sys/nxge/nxge_hio.h>
43 
44 /*
45  * External prototypes
46  */
47 extern npi_status_t npi_rxdma_dump_rdc_table(npi_handle_t, uint8_t);
48 
49 /* The following function may be found in nxge_main.c */
50 extern int nxge_m_mmac_remove(void *arg, int slot);
51 extern int nxge_m_mmac_add_g(void *arg, const uint8_t *maddr, int rdctbl,
52 	boolean_t usetbl);
53 extern int nxge_rx_ring_start(mac_ring_driver_t rdriver, uint64_t mr_gen_num);
54 
55 /* The following function may be found in nxge_[t|r]xdma.c */
56 extern npi_status_t nxge_txdma_channel_disable(nxge_t *, int);
57 extern nxge_status_t nxge_disable_rxdma_channel(nxge_t *, uint16_t);
58 
59 /*
60  * Local prototypes
61  */
62 static void nxge_grp_dc_append(nxge_t *, nxge_grp_t *, nxge_hio_dc_t *);
63 static nxge_hio_dc_t *nxge_grp_dc_unlink(nxge_t *, nxge_grp_t *, int);
64 static void nxge_grp_dc_map(nxge_grp_t *group);
65 
66 /*
67  * These functions are used by both service & guest domains to
68  * decide whether they're running in an LDOMs/XEN environment
69  * or not.  If so, then the Hybrid I/O (HIO) module is initialized.
70  */
71 
72 /*
73  * nxge_get_environs
74  *
75  *	Figure out if we are in a guest domain or not.
76  *
77  * Arguments:
78  * 	nxge
79  *
80  * Notes:
81  *
82  * Context:
83  *	Any domain
84  */
85 void
86 nxge_get_environs(
87 	nxge_t *nxge)
88 {
89 	char *string;
90 
91 	/*
92 	 * In the beginning, assume that we are running sans LDOMs/XEN.
93 	 */
94 	nxge->environs = SOLARIS_DOMAIN;
95 
96 	/*
97 	 * Are we a hybrid I/O (HIO) guest domain driver?
98 	 */
99 	if ((ddi_prop_lookup_string(DDI_DEV_T_ANY, nxge->dip,
100 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
101 	    "niutype", &string)) == DDI_PROP_SUCCESS) {
102 		if (strcmp(string, "n2niu") == 0) {
103 			nxge->environs = SOLARIS_GUEST_DOMAIN;
104 			/* So we can allocate properly-aligned memory. */
105 			nxge->niu_type = N2_NIU;
106 			NXGE_DEBUG_MSG((nxge, HIO_CTL,
107 			    "Hybrid IO-capable guest domain"));
108 		}
109 		ddi_prop_free(string);
110 	}
111 }
112 
113 #if !defined(sun4v)
114 
115 /*
116  * nxge_hio_init
117  *
118  *	Initialize the HIO module of the NXGE driver.
119  *
120  * Arguments:
121  * 	nxge
122  *
123  * Notes:
124  *	This is the non-hybrid I/O version of this function.
125  *
126  * Context:
127  *	Any domain
128  */
129 int
130 nxge_hio_init(nxge_t *nxge)
131 {
132 	nxge_hio_data_t *nhd;
133 	int i;
134 
135 	nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
136 	if (nhd == NULL) {
137 		nhd = KMEM_ZALLOC(sizeof (*nhd), KM_SLEEP);
138 		MUTEX_INIT(&nhd->lock, NULL, MUTEX_DRIVER, NULL);
139 		nhd->type = NXGE_HIO_TYPE_SERVICE;
140 		nxge->nxge_hw_p->hio = (uintptr_t)nhd;
141 	}
142 
143 	/*
144 	 * Initialize share and ring group structures.
145 	 */
146 	for (i = 0; i < NXGE_MAX_TDCS; i++)
147 		nxge->tdc_is_shared[i] = B_FALSE;
148 
149 	for (i = 0; i < NXGE_MAX_TDC_GROUPS; i++) {
150 		nxge->tx_hio_groups[i].ghandle = NULL;
151 		nxge->tx_hio_groups[i].nxgep = nxge;
152 		nxge->tx_hio_groups[i].type = MAC_RING_TYPE_TX;
153 		nxge->tx_hio_groups[i].gindex = 0;
154 		nxge->tx_hio_groups[i].sindex = 0;
155 	}
156 
157 	for (i = 0; i < NXGE_MAX_RDC_GROUPS; i++) {
158 		nxge->rx_hio_groups[i].ghandle = NULL;
159 		nxge->rx_hio_groups[i].nxgep = nxge;
160 		nxge->rx_hio_groups[i].type = MAC_RING_TYPE_RX;
161 		nxge->rx_hio_groups[i].gindex = 0;
162 		nxge->rx_hio_groups[i].sindex = 0;
163 		nxge->rx_hio_groups[i].started = B_FALSE;
164 		nxge->rx_hio_groups[i].port_default_grp = B_FALSE;
165 		nxge->rx_hio_groups[i].rdctbl = -1;
166 		nxge->rx_hio_groups[i].n_mac_addrs = 0;
167 	}
168 
169 	nhd->hio.ldoms = B_FALSE;
170 
171 	return (NXGE_OK);
172 }
173 
174 #endif
175 
176 void
177 nxge_hio_uninit(nxge_t *nxge)
178 {
179 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
180 
181 	ASSERT(nxge->nxge_hw_p->ndevs == 0);
182 
183 	if (nhd != NULL) {
184 		MUTEX_DESTROY(&nhd->lock);
185 		KMEM_FREE(nhd, sizeof (*nhd));
186 		nxge->nxge_hw_p->hio = 0;
187 	}
188 }
189 
190 /*
191  * nxge_dci_map
192  *
193  *	Map a DMA channel index to a channel number.
194  *
195  * Arguments:
196  * 	instance	The instance number of the driver.
197  * 	type		The type of channel this is: Tx or Rx.
198  * 	index		The index to convert to a channel number
199  *
200  * Notes:
201  *	This function is called by nxge_ndd.c:nxge_param_set_port_rdc()
202  *
203  * Context:
204  *	Any domain
205  */
206 int
207 nxge_dci_map(
208 	nxge_t *nxge,
209 	vpc_type_t type,
210 	int index)
211 {
212 	nxge_grp_set_t *set;
213 	int dc;
214 
215 	switch (type) {
216 	case VP_BOUND_TX:
217 		set = &nxge->tx_set;
218 		break;
219 	case VP_BOUND_RX:
220 		set = &nxge->rx_set;
221 		break;
222 	}
223 
224 	for (dc = 0; dc < NXGE_MAX_TDCS; dc++) {
225 		if ((1 << dc) & set->owned.map) {
226 			if (index == 0)
227 				return (dc);
228 			else
229 				index--;
230 		}
231 	}
232 
233 	return (-1);
234 }
235 
236 /*
237  * ---------------------------------------------------------------------
238  * These are the general-purpose DMA channel group functions.  That is,
239  * these functions are used to manage groups of TDCs or RDCs in an HIO
240  * environment.
241  *
242  * But is also expected that in the future they will be able to manage
243  * Crossbow groups.
244  * ---------------------------------------------------------------------
245  */
246 
247 /*
248  * nxge_grp_cleanup(p_nxge_t nxge)
249  *
250  *	Remove all outstanding groups.
251  *
252  * Arguments:
253  *	nxge
254  */
255 void
256 nxge_grp_cleanup(p_nxge_t nxge)
257 {
258 	nxge_grp_set_t *set;
259 	int i;
260 
261 	MUTEX_ENTER(&nxge->group_lock);
262 
263 	/*
264 	 * Find RX groups that need to be cleaned up.
265 	 */
266 	set = &nxge->rx_set;
267 	for (i = 0; i < NXGE_LOGICAL_GROUP_MAX; i++) {
268 		if (set->group[i] != NULL) {
269 			KMEM_FREE(set->group[i], sizeof (nxge_grp_t));
270 			set->group[i] = NULL;
271 		}
272 	}
273 
274 	/*
275 	 * Find TX groups that need to be cleaned up.
276 	 */
277 	set = &nxge->tx_set;
278 	for (i = 0; i < NXGE_LOGICAL_GROUP_MAX; i++) {
279 		if (set->group[i] != NULL) {
280 			KMEM_FREE(set->group[i], sizeof (nxge_grp_t));
281 			set->group[i] = NULL;
282 		}
283 	}
284 	MUTEX_EXIT(&nxge->group_lock);
285 }
286 
287 
288 /*
289  * nxge_grp_add
290  *
291  *	Add a group to an instance of NXGE.
292  *
293  * Arguments:
294  * 	nxge
295  * 	type	Tx or Rx
296  *
297  * Notes:
298  *
299  * Context:
300  *	Any domain
301  */
302 nxge_grp_t *
303 nxge_grp_add(
304 	nxge_t *nxge,
305 	nxge_grp_type_t type)
306 {
307 	nxge_grp_set_t *set;
308 	nxge_grp_t *group;
309 	int i;
310 
311 	group = KMEM_ZALLOC(sizeof (*group), KM_SLEEP);
312 	group->nxge = nxge;
313 
314 	MUTEX_ENTER(&nxge->group_lock);
315 	switch (type) {
316 	case NXGE_TRANSMIT_GROUP:
317 	case EXT_TRANSMIT_GROUP:
318 		set = &nxge->tx_set;
319 		break;
320 	default:
321 		set = &nxge->rx_set;
322 		break;
323 	}
324 
325 	group->type = type;
326 	group->active = B_TRUE;
327 	group->sequence = set->sequence++;
328 
329 	/* Find an empty slot for this logical group. */
330 	for (i = 0; i < NXGE_LOGICAL_GROUP_MAX; i++) {
331 		if (set->group[i] == 0) {
332 			group->index = i;
333 			set->group[i] = group;
334 			NXGE_DC_SET(set->lg.map, i);
335 			set->lg.count++;
336 			break;
337 		}
338 	}
339 	MUTEX_EXIT(&nxge->group_lock);
340 
341 	NXGE_DEBUG_MSG((nxge, HIO_CTL,
342 	    "nxge_grp_add: %cgroup = %d.%d",
343 	    type == NXGE_TRANSMIT_GROUP ? 't' : 'r',
344 	    nxge->mac.portnum, group->sequence));
345 
346 	return (group);
347 }
348 
349 void
350 nxge_grp_remove(
351 	nxge_t *nxge,
352 	nxge_grp_t *group)	/* The group to remove. */
353 {
354 	nxge_grp_set_t *set;
355 	vpc_type_t type;
356 
357 	if (group == NULL)
358 		return;
359 
360 	MUTEX_ENTER(&nxge->group_lock);
361 	switch (group->type) {
362 	case NXGE_TRANSMIT_GROUP:
363 	case EXT_TRANSMIT_GROUP:
364 		set = &nxge->tx_set;
365 		break;
366 	default:
367 		set = &nxge->rx_set;
368 		break;
369 	}
370 
371 	if (set->group[group->index] != group) {
372 		MUTEX_EXIT(&nxge->group_lock);
373 		return;
374 	}
375 
376 	set->group[group->index] = 0;
377 	NXGE_DC_RESET(set->lg.map, group->index);
378 	set->lg.count--;
379 
380 	/* While inside the mutex, deactivate <group>. */
381 	group->active = B_FALSE;
382 
383 	MUTEX_EXIT(&nxge->group_lock);
384 
385 	NXGE_DEBUG_MSG((nxge, HIO_CTL,
386 	    "nxge_grp_remove(%c.%d.%d) called",
387 	    group->type == NXGE_TRANSMIT_GROUP ? 't' : 'r',
388 	    nxge->mac.portnum, group->sequence));
389 
390 	/* Now, remove any DCs which are still active. */
391 	switch (group->type) {
392 	default:
393 		type = VP_BOUND_TX;
394 		break;
395 	case NXGE_RECEIVE_GROUP:
396 	case EXT_RECEIVE_GROUP:
397 		type = VP_BOUND_RX;
398 	}
399 
400 	while (group->dc) {
401 		nxge_grp_dc_remove(nxge, type, group->dc->channel);
402 	}
403 
404 	KMEM_FREE(group, sizeof (*group));
405 }
406 
407 /*
408  * nxge_grp_dc_add
409  *
410  *	Add a DMA channel to a VR/Group.
411  *
412  * Arguments:
413  * 	nxge
414  * 	channel	The channel to add.
415  * Notes:
416  *
417  * Context:
418  *	Any domain
419  */
420 /* ARGSUSED */
421 int
422 nxge_grp_dc_add(
423 	nxge_t *nxge,
424 	nxge_grp_t *group,	/* The group to add <channel> to. */
425 	vpc_type_t type,	/* Rx or Tx */
426 	int channel)		/* A physical/logical channel number */
427 {
428 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
429 	nxge_hio_dc_t *dc;
430 	nxge_grp_set_t *set;
431 	nxge_status_t status = NXGE_OK;
432 	int error = 0;
433 
434 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_grp_dc_add"));
435 
436 	if (group == 0)
437 		return (0);
438 
439 	switch (type) {
440 	case VP_BOUND_TX:
441 		set = &nxge->tx_set;
442 		if (channel > NXGE_MAX_TDCS) {
443 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
444 			    "nxge_grp_dc_add: TDC = %d", channel));
445 			return (NXGE_ERROR);
446 		}
447 		break;
448 	case VP_BOUND_RX:
449 		set = &nxge->rx_set;
450 		if (channel > NXGE_MAX_RDCS) {
451 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
452 			    "nxge_grp_dc_add: RDC = %d", channel));
453 			return (NXGE_ERROR);
454 		}
455 		break;
456 
457 	default:
458 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
459 		    "nxge_grp_dc_add: unknown type channel(%d)", channel));
460 	}
461 
462 	NXGE_DEBUG_MSG((nxge, HIO_CTL,
463 	    "nxge_grp_dc_add: %cgroup = %d.%d.%d, channel = %d",
464 	    type == VP_BOUND_TX ? 't' : 'r',
465 	    nxge->mac.portnum, group->sequence, group->count, channel));
466 
467 	MUTEX_ENTER(&nxge->group_lock);
468 	if (group->active != B_TRUE) {
469 		/* We may be in the process of removing this group. */
470 		MUTEX_EXIT(&nxge->group_lock);
471 		return (NXGE_ERROR);
472 	}
473 	MUTEX_EXIT(&nxge->group_lock);
474 
475 	if (!(dc = nxge_grp_dc_find(nxge, type, channel))) {
476 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
477 		    "nxge_grp_dc_add(%d): DC FIND failed", channel));
478 		return (NXGE_ERROR);
479 	}
480 
481 	MUTEX_ENTER(&nhd->lock);
482 
483 	if (dc->group) {
484 		MUTEX_EXIT(&nhd->lock);
485 		/* This channel is already in use! */
486 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
487 		    "nxge_grp_dc_add(%d): channel already in group", channel));
488 		return (NXGE_ERROR);
489 	}
490 
491 	dc->next = 0;
492 	dc->page = channel;
493 	dc->channel = (nxge_channel_t)channel;
494 
495 	dc->type = type;
496 	if (type == VP_BOUND_RX) {
497 		dc->init = nxge_init_rxdma_channel;
498 		dc->uninit = nxge_uninit_rxdma_channel;
499 	} else {
500 		dc->init = nxge_init_txdma_channel;
501 		dc->uninit = nxge_uninit_txdma_channel;
502 	}
503 
504 	dc->group = group;
505 
506 	if (isLDOMguest(nxge)) {
507 		error = nxge_hio_ldsv_add(nxge, dc);
508 		if (error != 0) {
509 			MUTEX_EXIT(&nhd->lock);
510 			return (NXGE_ERROR);
511 		}
512 	}
513 
514 	NXGE_DC_SET(set->owned.map, channel);
515 	set->owned.count++;
516 
517 	MUTEX_EXIT(&nhd->lock);
518 
519 	if ((status = (*dc->init)(nxge, channel)) != NXGE_OK) {
520 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
521 		    "nxge_grp_dc_add(%d): channel init failed", channel));
522 		MUTEX_ENTER(&nhd->lock);
523 		(void) memset(dc, 0, sizeof (*dc));
524 		NXGE_DC_RESET(set->owned.map, channel);
525 		set->owned.count--;
526 		MUTEX_EXIT(&nhd->lock);
527 		return (NXGE_ERROR);
528 	}
529 
530 	nxge_grp_dc_append(nxge, group, dc);
531 
532 	if (type == VP_BOUND_TX) {
533 		MUTEX_ENTER(&nhd->lock);
534 		nxge->tdc_is_shared[channel] = B_FALSE;
535 		MUTEX_EXIT(&nhd->lock);
536 	}
537 
538 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_grp_dc_add"));
539 
540 	return ((int)status);
541 }
542 
543 void
544 nxge_grp_dc_remove(
545 	nxge_t *nxge,
546 	vpc_type_t type,
547 	int channel)
548 {
549 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
550 	nxge_hio_dc_t *dc;
551 	nxge_grp_set_t *set;
552 	nxge_grp_t *group;
553 
554 	dc_uninit_t uninit;
555 
556 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_grp_dc_remove"));
557 
558 	if ((dc = nxge_grp_dc_find(nxge, type, channel)) == 0)
559 		goto nxge_grp_dc_remove_exit;
560 
561 	if ((dc->group == NULL) && (dc->next == 0) &&
562 	    (dc->channel == 0) && (dc->page == 0) && (dc->type == 0)) {
563 		goto nxge_grp_dc_remove_exit;
564 	}
565 
566 	group = (nxge_grp_t *)dc->group;
567 
568 	if (isLDOMguest(nxge)) {
569 		(void) nxge_hio_intr_remove(nxge, type, channel);
570 	}
571 
572 	NXGE_DEBUG_MSG((nxge, HIO_CTL,
573 	    "DC remove: group = %d.%d.%d, %cdc %d",
574 	    nxge->mac.portnum, group->sequence, group->count,
575 	    type == VP_BOUND_TX ? 't' : 'r', dc->channel));
576 
577 	MUTEX_ENTER(&nhd->lock);
578 
579 	set = dc->type == VP_BOUND_TX ? &nxge->tx_set : &nxge->rx_set;
580 
581 	/* Remove the DC from its group. */
582 	if (nxge_grp_dc_unlink(nxge, group, channel) != dc) {
583 		MUTEX_EXIT(&nhd->lock);
584 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
585 		    "nxge_grp_dc_remove(%d) failed", channel));
586 		goto nxge_grp_dc_remove_exit;
587 	}
588 
589 	uninit = dc->uninit;
590 	channel = dc->channel;
591 
592 	NXGE_DC_RESET(set->owned.map, channel);
593 	set->owned.count--;
594 
595 	(void) memset(dc, 0, sizeof (*dc));
596 
597 	MUTEX_EXIT(&nhd->lock);
598 
599 	(*uninit)(nxge, channel);
600 
601 nxge_grp_dc_remove_exit:
602 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_grp_dc_remove"));
603 }
604 
605 nxge_hio_dc_t *
606 nxge_grp_dc_find(
607 	nxge_t *nxge,
608 	vpc_type_t type,	/* Rx or Tx */
609 	int channel)
610 {
611 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
612 	nxge_hio_dc_t *current;
613 
614 	current = (type == VP_BOUND_TX) ? &nhd->tdc[0] : &nhd->rdc[0];
615 
616 	if (!isLDOMguest(nxge)) {
617 		return (&current[channel]);
618 	} else {
619 		/* We're in a guest domain. */
620 		int i, limit = (type == VP_BOUND_TX) ?
621 		    NXGE_MAX_TDCS : NXGE_MAX_RDCS;
622 
623 		MUTEX_ENTER(&nhd->lock);
624 		for (i = 0; i < limit; i++, current++) {
625 			if (current->channel == channel) {
626 				if (current->vr && current->vr->nxge ==
627 				    (uintptr_t)nxge) {
628 					MUTEX_EXIT(&nhd->lock);
629 					return (current);
630 				}
631 			}
632 		}
633 		MUTEX_EXIT(&nhd->lock);
634 	}
635 
636 	return (0);
637 }
638 
639 /*
640  * nxge_grp_dc_append
641  *
642  *	Append a DMA channel to a group.
643  *
644  * Arguments:
645  * 	nxge
646  * 	group	The group to append to
647  * 	dc	The DMA channel to append
648  *
649  * Notes:
650  *
651  * Context:
652  *	Any domain
653  */
654 static
655 void
656 nxge_grp_dc_append(
657 	nxge_t *nxge,
658 	nxge_grp_t *group,
659 	nxge_hio_dc_t *dc)
660 {
661 	MUTEX_ENTER(&nxge->group_lock);
662 
663 	if (group->dc == 0) {
664 		group->dc = dc;
665 	} else {
666 		nxge_hio_dc_t *current = group->dc;
667 		do {
668 			if (current->next == 0) {
669 				current->next = dc;
670 				break;
671 			}
672 			current = current->next;
673 		} while (current);
674 	}
675 
676 	NXGE_DC_SET(group->map, dc->channel);
677 
678 	nxge_grp_dc_map(group);
679 	group->count++;
680 
681 	MUTEX_EXIT(&nxge->group_lock);
682 }
683 
684 /*
685  * nxge_grp_dc_unlink
686  *
687  *	Unlink a DMA channel fromits linked list (group).
688  *
689  * Arguments:
690  * 	nxge
691  * 	group	The group (linked list) to unlink from
692  * 	dc	The DMA channel to append
693  *
694  * Notes:
695  *
696  * Context:
697  *	Any domain
698  */
699 nxge_hio_dc_t *
700 nxge_grp_dc_unlink(
701 	nxge_t *nxge,
702 	nxge_grp_t *group,
703 	int channel)
704 {
705 	nxge_hio_dc_t *current, *previous;
706 
707 	MUTEX_ENTER(&nxge->group_lock);
708 
709 	if (group == NULL) {
710 		MUTEX_EXIT(&nxge->group_lock);
711 		return (0);
712 	}
713 
714 	if ((current = group->dc) == 0) {
715 		MUTEX_EXIT(&nxge->group_lock);
716 		return (0);
717 	}
718 
719 	previous = 0;
720 	do {
721 		if (current->channel == channel) {
722 			if (previous)
723 				previous->next = current->next;
724 			else
725 				group->dc = current->next;
726 			break;
727 		}
728 		previous = current;
729 		current = current->next;
730 	} while (current);
731 
732 	if (current == 0) {
733 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
734 		    "DC unlink: DC %d not found", channel));
735 	} else {
736 		current->next = 0;
737 		current->group = 0;
738 
739 		NXGE_DC_RESET(group->map, channel);
740 		group->count--;
741 	}
742 
743 	nxge_grp_dc_map(group);
744 
745 	MUTEX_EXIT(&nxge->group_lock);
746 
747 	return (current);
748 }
749 
750 /*
751  * nxge_grp_dc_map
752  *
753  *	Map a linked list to an array of channel numbers.
754  *
755  * Arguments:
756  * 	nxge
757  * 	group	The group to remap.
758  *
759  * Notes:
760  *	It is expected that the caller will hold the correct mutex.
761  *
762  * Context:
763  *	Service domain
764  */
765 void
766 nxge_grp_dc_map(
767 	nxge_grp_t *group)
768 {
769 	nxge_channel_t *legend;
770 	nxge_hio_dc_t *dc;
771 
772 	(void) memset(group->legend, 0, sizeof (group->legend));
773 
774 	legend = group->legend;
775 	dc = group->dc;
776 	while (dc) {
777 		*legend = dc->channel;
778 		legend++;
779 		dc = dc->next;
780 	}
781 }
782 
783 /*
784  * ---------------------------------------------------------------------
785  * These are HIO debugging functions.
786  * ---------------------------------------------------------------------
787  */
788 
789 /*
790  * nxge_delay
791  *
792  *	Delay <seconds> number of seconds.
793  *
794  * Arguments:
795  * 	nxge
796  * 	group	The group to append to
797  * 	dc	The DMA channel to append
798  *
799  * Notes:
800  *	This is a developer-only function.
801  *
802  * Context:
803  *	Any domain
804  */
805 void
806 nxge_delay(
807 	int seconds)
808 {
809 	delay(drv_usectohz(seconds * 1000000));
810 }
811 
812 static dmc_reg_name_t rx_names[] = {
813 	{ "RXDMA_CFIG1",	0 },
814 	{ "RXDMA_CFIG2",	8 },
815 	{ "RBR_CFIG_A",		0x10 },
816 	{ "RBR_CFIG_B",		0x18 },
817 	{ "RBR_KICK",		0x20 },
818 	{ "RBR_STAT",		0x28 },
819 	{ "RBR_HDH",		0x30 },
820 	{ "RBR_HDL",		0x38 },
821 	{ "RCRCFIG_A",		0x40 },
822 	{ "RCRCFIG_B",		0x48 },
823 	{ "RCRSTAT_A",		0x50 },
824 	{ "RCRSTAT_B",		0x58 },
825 	{ "RCRSTAT_C",		0x60 },
826 	{ "RX_DMA_ENT_MSK",	0x68 },
827 	{ "RX_DMA_CTL_STAT",	0x70 },
828 	{ "RCR_FLSH",		0x78 },
829 	{ "RXMISC",		0x90 },
830 	{ "RX_DMA_CTL_STAT_DBG", 0x98 },
831 	{ 0, -1 }
832 };
833 
834 static dmc_reg_name_t tx_names[] = {
835 	{ "Tx_RNG_CFIG",	0 },
836 	{ "Tx_RNG_HDL",		0x10 },
837 	{ "Tx_RNG_KICK",	0x18 },
838 	{ "Tx_ENT_MASK",	0x20 },
839 	{ "Tx_CS",		0x28 },
840 	{ "TxDMA_MBH",		0x30 },
841 	{ "TxDMA_MBL",		0x38 },
842 	{ "TxDMA_PRE_ST",	0x40 },
843 	{ "Tx_RNG_ERR_LOGH",	0x48 },
844 	{ "Tx_RNG_ERR_LOGL",	0x50 },
845 	{ "TDMC_INTR_DBG",	0x60 },
846 	{ "Tx_CS_DBG",		0x68 },
847 	{ 0, -1 }
848 };
849 
850 /*
851  * nxge_xx2str
852  *
853  *	Translate a register address into a string.
854  *
855  * Arguments:
856  * 	offset	The address of the register to translate.
857  *
858  * Notes:
859  *	These are developer-only function.
860  *
861  * Context:
862  *	Any domain
863  */
864 const char *
865 nxge_rx2str(
866 	int offset)
867 {
868 	dmc_reg_name_t *reg = &rx_names[0];
869 
870 	offset &= DMA_CSR_MASK;
871 
872 	while (reg->name) {
873 		if (offset == reg->offset)
874 			return (reg->name);
875 		reg++;
876 	}
877 
878 	return (0);
879 }
880 
881 const char *
882 nxge_tx2str(
883 	int offset)
884 {
885 	dmc_reg_name_t *reg = &tx_names[0];
886 
887 	offset &= DMA_CSR_MASK;
888 
889 	while (reg->name) {
890 		if (offset == reg->offset)
891 			return (reg->name);
892 		reg++;
893 	}
894 
895 	return (0);
896 }
897 
898 /*
899  * nxge_ddi_perror
900  *
901  *	Map a DDI error number to a string.
902  *
903  * Arguments:
904  * 	ddi_error	The DDI error number to map.
905  *
906  * Notes:
907  *
908  * Context:
909  *	Any domain
910  */
911 const char *
912 nxge_ddi_perror(
913 	int ddi_error)
914 {
915 	switch (ddi_error) {
916 	case DDI_SUCCESS:
917 		return ("DDI_SUCCESS");
918 	case DDI_FAILURE:
919 		return ("DDI_FAILURE");
920 	case DDI_NOT_WELL_FORMED:
921 		return ("DDI_NOT_WELL_FORMED");
922 	case DDI_EAGAIN:
923 		return ("DDI_EAGAIN");
924 	case DDI_EINVAL:
925 		return ("DDI_EINVAL");
926 	case DDI_ENOTSUP:
927 		return ("DDI_ENOTSUP");
928 	case DDI_EPENDING:
929 		return ("DDI_EPENDING");
930 	case DDI_ENOMEM:
931 		return ("DDI_ENOMEM");
932 	case DDI_EBUSY:
933 		return ("DDI_EBUSY");
934 	case DDI_ETRANSPORT:
935 		return ("DDI_ETRANSPORT");
936 	case DDI_ECONTEXT:
937 		return ("DDI_ECONTEXT");
938 	default:
939 		return ("Unknown error");
940 	}
941 }
942 
943 /*
944  * ---------------------------------------------------------------------
945  * These are Sun4v HIO function definitions
946  * ---------------------------------------------------------------------
947  */
948 
949 #if defined(sun4v)
950 
951 /*
952  * Local prototypes
953  */
954 static nxge_hio_vr_t *nxge_hio_vr_share(nxge_t *);
955 static void nxge_hio_unshare(nxge_hio_vr_t *);
956 
957 static int nxge_hio_addres(nxge_hio_vr_t *, mac_ring_type_t, uint64_t *);
958 static void nxge_hio_remres(nxge_hio_vr_t *, mac_ring_type_t, res_map_t);
959 
960 static void nxge_hio_tdc_unshare(nxge_t *nxge, int dev_grpid, int channel);
961 static void nxge_hio_rdc_unshare(nxge_t *nxge, int dev_grpid, int channel);
962 static int nxge_hio_dc_share(nxge_t *, nxge_hio_vr_t *, mac_ring_type_t, int);
963 static void nxge_hio_dc_unshare(nxge_t *, nxge_hio_vr_t *,
964     mac_ring_type_t, int);
965 
966 /*
967  * nxge_hio_init
968  *
969  *	Initialize the HIO module of the NXGE driver.
970  *
971  * Arguments:
972  * 	nxge
973  *
974  * Notes:
975  *
976  * Context:
977  *	Any domain
978  */
979 int
980 nxge_hio_init(nxge_t *nxge)
981 {
982 	nxge_hio_data_t *nhd;
983 	int i, region;
984 
985 	nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
986 	if (nhd == 0) {
987 		nhd = KMEM_ZALLOC(sizeof (*nhd), KM_SLEEP);
988 		MUTEX_INIT(&nhd->lock, NULL, MUTEX_DRIVER, NULL);
989 		if (isLDOMguest(nxge))
990 			nhd->type = NXGE_HIO_TYPE_GUEST;
991 		else
992 			nhd->type = NXGE_HIO_TYPE_SERVICE;
993 		nxge->nxge_hw_p->hio = (uintptr_t)nhd;
994 	}
995 
996 	if ((nxge->environs == SOLARIS_DOMAIN) &&
997 	    (nxge->niu_type == N2_NIU)) {
998 		if (nxge->niu_hsvc_available == B_TRUE) {
999 			hsvc_info_t *niu_hsvc = &nxge->niu_hsvc;
1000 			/*
1001 			 * Versions supported now are:
1002 			 *  - major number >= 1 (NIU_MAJOR_VER).
1003 			 */
1004 			if ((niu_hsvc->hsvc_major >= NIU_MAJOR_VER) ||
1005 			    (niu_hsvc->hsvc_major == 1 &&
1006 			    niu_hsvc->hsvc_minor == 1)) {
1007 				nxge->environs = SOLARIS_SERVICE_DOMAIN;
1008 				NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
1009 				    "nxge_hio_init: hypervisor services "
1010 				    "version %d.%d",
1011 				    niu_hsvc->hsvc_major,
1012 				    niu_hsvc->hsvc_minor));
1013 			}
1014 		}
1015 	}
1016 
1017 	/*
1018 	 * Initialize share and ring group structures.
1019 	 */
1020 	for (i = 0; i < NXGE_MAX_TDC_GROUPS; i++) {
1021 		nxge->tx_hio_groups[i].ghandle = NULL;
1022 		nxge->tx_hio_groups[i].nxgep = nxge;
1023 		nxge->tx_hio_groups[i].type = MAC_RING_TYPE_TX;
1024 		nxge->tx_hio_groups[i].gindex = 0;
1025 		nxge->tx_hio_groups[i].sindex = 0;
1026 	}
1027 
1028 	for (i = 0; i < NXGE_MAX_RDC_GROUPS; i++) {
1029 		nxge->rx_hio_groups[i].ghandle = NULL;
1030 		nxge->rx_hio_groups[i].nxgep = nxge;
1031 		nxge->rx_hio_groups[i].type = MAC_RING_TYPE_RX;
1032 		nxge->rx_hio_groups[i].gindex = 0;
1033 		nxge->rx_hio_groups[i].sindex = 0;
1034 		nxge->rx_hio_groups[i].started = B_FALSE;
1035 		nxge->rx_hio_groups[i].port_default_grp = B_FALSE;
1036 		nxge->rx_hio_groups[i].rdctbl = -1;
1037 		nxge->rx_hio_groups[i].n_mac_addrs = 0;
1038 	}
1039 
1040 	if (!isLDOMs(nxge)) {
1041 		nhd->hio.ldoms = B_FALSE;
1042 		return (NXGE_OK);
1043 	}
1044 
1045 	nhd->hio.ldoms = B_TRUE;
1046 
1047 	/*
1048 	 * Fill in what we can.
1049 	 */
1050 	for (region = 0; region < NXGE_VR_SR_MAX; region++) {
1051 		nhd->vr[region].region = region;
1052 	}
1053 	nhd->vrs = NXGE_VR_SR_MAX - 2;
1054 
1055 	/*
1056 	 * Initialize the share stuctures.
1057 	 */
1058 	for (i = 0; i < NXGE_MAX_TDCS; i++)
1059 		nxge->tdc_is_shared[i] = B_FALSE;
1060 
1061 	for (i = 0; i < NXGE_VR_SR_MAX; i++) {
1062 		nxge->shares[i].nxgep = nxge;
1063 		nxge->shares[i].index = 0;
1064 		nxge->shares[i].vrp = NULL;
1065 		nxge->shares[i].tmap = 0;
1066 		nxge->shares[i].rmap = 0;
1067 		nxge->shares[i].rxgroup = 0;
1068 		nxge->shares[i].active = B_FALSE;
1069 	}
1070 
1071 	/* Fill in the HV HIO function pointers. */
1072 	nxge_hio_hv_init(nxge);
1073 
1074 	if (isLDOMservice(nxge)) {
1075 		NXGE_DEBUG_MSG((nxge, HIO_CTL,
1076 		    "Hybrid IO-capable service domain"));
1077 		return (NXGE_OK);
1078 	}
1079 
1080 	return (0);
1081 }
1082 #endif /* defined(sun4v) */
1083 
1084 static int
1085 nxge_hio_group_mac_add(nxge_t *nxge, nxge_ring_group_t *g,
1086     const uint8_t *macaddr)
1087 {
1088 	int rv;
1089 	nxge_rdc_grp_t *group;
1090 
1091 	mutex_enter(nxge->genlock);
1092 
1093 	/*
1094 	 * Initialize the NXGE RDC table data structure.
1095 	 */
1096 	group = &nxge->pt_config.rdc_grps[g->rdctbl];
1097 	if (!group->flag) {
1098 		group->port = NXGE_GET_PORT_NUM(nxge->function_num);
1099 		group->config_method = RDC_TABLE_ENTRY_METHOD_REP;
1100 		group->flag = B_TRUE;	/* This group has been configured. */
1101 	}
1102 
1103 	mutex_exit(nxge->genlock);
1104 
1105 	/*
1106 	 * Add the MAC address.
1107 	 */
1108 	if ((rv = nxge_m_mmac_add_g((void *)nxge, macaddr,
1109 	    g->rdctbl, B_TRUE)) != 0) {
1110 		return (rv);
1111 	}
1112 
1113 	mutex_enter(nxge->genlock);
1114 	g->n_mac_addrs++;
1115 	mutex_exit(nxge->genlock);
1116 	return (0);
1117 }
1118 
1119 static int
1120 nxge_hio_set_unicst(void *arg, const uint8_t *macaddr)
1121 {
1122 	p_nxge_t		nxgep = (p_nxge_t)arg;
1123 	struct ether_addr	addrp;
1124 
1125 	bcopy(macaddr, (uint8_t *)&addrp, ETHERADDRL);
1126 	if (nxge_set_mac_addr(nxgep, &addrp)) {
1127 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1128 		    "<== nxge_m_unicst: set unitcast failed"));
1129 		return (EINVAL);
1130 	}
1131 
1132 	nxgep->primary = B_TRUE;
1133 
1134 	return (0);
1135 }
1136 
1137 /*ARGSUSED*/
1138 static int
1139 nxge_hio_clear_unicst(p_nxge_t nxgep, const uint8_t *mac_addr)
1140 {
1141 	nxgep->primary = B_FALSE;
1142 	return (0);
1143 }
1144 
1145 static int
1146 nxge_hio_add_mac(void *arg, const uint8_t *mac_addr)
1147 {
1148 	nxge_ring_group_t	*group = (nxge_ring_group_t *)arg;
1149 	p_nxge_t		nxge = group->nxgep;
1150 	int			rv;
1151 	nxge_hio_vr_t		*vr;	/* The Virtualization Region */
1152 
1153 	ASSERT(group->type == MAC_RING_TYPE_RX);
1154 	ASSERT(group->nxgep != NULL);
1155 
1156 	if (isLDOMguest(group->nxgep))
1157 		return (0);
1158 
1159 	mutex_enter(nxge->genlock);
1160 
1161 	if (!nxge->primary && group->port_default_grp) {
1162 		rv = nxge_hio_set_unicst((void *)nxge, mac_addr);
1163 		mutex_exit(nxge->genlock);
1164 		return (rv);
1165 	}
1166 
1167 	/*
1168 	 * If the group is associated with a VR, then only one
1169 	 * address may be assigned to the group.
1170 	 */
1171 	vr = (nxge_hio_vr_t *)nxge->shares[group->sindex].vrp;
1172 	if ((vr != NULL) && (group->n_mac_addrs)) {
1173 		mutex_exit(nxge->genlock);
1174 		return (ENOSPC);
1175 	}
1176 
1177 	mutex_exit(nxge->genlock);
1178 
1179 	/*
1180 	 * Program the mac address for the group.
1181 	 */
1182 	if ((rv = nxge_hio_group_mac_add(nxge, group, mac_addr)) != 0) {
1183 		return (rv);
1184 	}
1185 
1186 	return (0);
1187 }
1188 
1189 static int
1190 find_mac_slot(nxge_mmac_t *mmac_info, const uint8_t *mac_addr)
1191 {
1192 	int i;
1193 	for (i = 0; i <= mmac_info->num_mmac; i++) {
1194 		if (memcmp(mmac_info->mac_pool[i].addr, mac_addr,
1195 		    ETHERADDRL) == 0) {
1196 			return (i);
1197 		}
1198 	}
1199 	return (-1);
1200 }
1201 
1202 /* ARGSUSED */
1203 static int
1204 nxge_hio_rem_mac(void *arg, const uint8_t *mac_addr)
1205 {
1206 	nxge_ring_group_t *group = (nxge_ring_group_t *)arg;
1207 	struct ether_addr addrp;
1208 	p_nxge_t nxge = group->nxgep;
1209 	nxge_mmac_t *mmac_info;
1210 	int rv, slot;
1211 
1212 	ASSERT(group->type == MAC_RING_TYPE_RX);
1213 	ASSERT(group->nxgep != NULL);
1214 
1215 	if (isLDOMguest(group->nxgep))
1216 		return (0);
1217 
1218 	mutex_enter(nxge->genlock);
1219 
1220 	mmac_info = &nxge->nxge_mmac_info;
1221 	slot = find_mac_slot(mmac_info, mac_addr);
1222 	if (slot < 0) {
1223 		if (group->port_default_grp && nxge->primary) {
1224 			bcopy(mac_addr, (uint8_t *)&addrp, ETHERADDRL);
1225 			if (ether_cmp(&addrp, &nxge->ouraddr) == 0) {
1226 				rv = nxge_hio_clear_unicst(nxge, mac_addr);
1227 				mutex_exit(nxge->genlock);
1228 				return (rv);
1229 			} else {
1230 				mutex_exit(nxge->genlock);
1231 				return (EINVAL);
1232 			}
1233 		} else {
1234 			mutex_exit(nxge->genlock);
1235 			return (EINVAL);
1236 		}
1237 	}
1238 
1239 	mutex_exit(nxge->genlock);
1240 
1241 	/*
1242 	 * Remove the mac address for the group
1243 	 */
1244 	if ((rv = nxge_m_mmac_remove(nxge, slot)) != 0) {
1245 		return (rv);
1246 	}
1247 
1248 	mutex_enter(nxge->genlock);
1249 	group->n_mac_addrs--;
1250 	mutex_exit(nxge->genlock);
1251 
1252 	return (0);
1253 }
1254 
1255 static int
1256 nxge_hio_group_start(mac_group_driver_t gdriver)
1257 {
1258 	nxge_ring_group_t	*group = (nxge_ring_group_t *)gdriver;
1259 	nxge_rdc_grp_t		*rdc_grp_p;
1260 	int			rdctbl;
1261 	int			dev_gindex;
1262 
1263 	ASSERT(group->type == MAC_RING_TYPE_RX);
1264 	ASSERT(group->nxgep != NULL);
1265 
1266 	ASSERT(group->nxgep->nxge_mac_state == NXGE_MAC_STARTED);
1267 	if (group->nxgep->nxge_mac_state != NXGE_MAC_STARTED)
1268 		return (ENXIO);
1269 
1270 	mutex_enter(group->nxgep->genlock);
1271 	if (isLDOMguest(group->nxgep))
1272 		goto nxge_hio_group_start_exit;
1273 
1274 	dev_gindex = group->nxgep->pt_config.hw_config.def_mac_rxdma_grpid +
1275 	    group->gindex;
1276 	rdc_grp_p = &group->nxgep->pt_config.rdc_grps[dev_gindex];
1277 
1278 	/*
1279 	 * Get an rdc table for this group.
1280 	 * Group ID is given by the caller, and that's the group it needs
1281 	 * to bind to.  The default group is already bound when the driver
1282 	 * was attached.
1283 	 *
1284 	 * For Group 0, it's RDC table was allocated at attach time
1285 	 * no need to allocate a new table.
1286 	 */
1287 	if (group->gindex != 0) {
1288 		rdctbl = nxge_fzc_rdc_tbl_bind(group->nxgep,
1289 		    dev_gindex, B_TRUE);
1290 		if (rdctbl < 0) {
1291 			mutex_exit(group->nxgep->genlock);
1292 			return (rdctbl);
1293 		}
1294 	} else {
1295 		rdctbl = group->nxgep->pt_config.hw_config.def_mac_rxdma_grpid;
1296 	}
1297 
1298 	group->rdctbl = rdctbl;
1299 
1300 	(void) nxge_init_fzc_rdc_tbl(group->nxgep, rdc_grp_p, rdctbl);
1301 
1302 nxge_hio_group_start_exit:
1303 	group->started = B_TRUE;
1304 	mutex_exit(group->nxgep->genlock);
1305 	return (0);
1306 }
1307 
1308 static void
1309 nxge_hio_group_stop(mac_group_driver_t gdriver)
1310 {
1311 	nxge_ring_group_t *group = (nxge_ring_group_t *)gdriver;
1312 
1313 	ASSERT(group->type == MAC_RING_TYPE_RX);
1314 
1315 	mutex_enter(group->nxgep->genlock);
1316 	group->started = B_FALSE;
1317 
1318 	if (isLDOMguest(group->nxgep))
1319 		goto nxge_hio_group_stop_exit;
1320 
1321 	/*
1322 	 * Unbind the RDC table previously bound for this group.
1323 	 *
1324 	 * Since RDC table for group 0 was allocated at attach
1325 	 * time, no need to unbind the table here.
1326 	 */
1327 	if (group->gindex != 0)
1328 		(void) nxge_fzc_rdc_tbl_unbind(group->nxgep, group->rdctbl);
1329 
1330 nxge_hio_group_stop_exit:
1331 	mutex_exit(group->nxgep->genlock);
1332 }
1333 
1334 /* ARGSUSED */
1335 void
1336 nxge_hio_group_get(void *arg, mac_ring_type_t type, int groupid,
1337 	mac_group_info_t *infop, mac_group_handle_t ghdl)
1338 {
1339 	p_nxge_t		nxgep = (p_nxge_t)arg;
1340 	nxge_ring_group_t	*group;
1341 	int			dev_gindex;
1342 
1343 	switch (type) {
1344 	case MAC_RING_TYPE_RX:
1345 		group = &nxgep->rx_hio_groups[groupid];
1346 		group->nxgep = nxgep;
1347 		group->ghandle = ghdl;
1348 		group->gindex = groupid;
1349 		group->sindex = 0;	/* not yet bound to a share */
1350 
1351 		if (!isLDOMguest(nxgep)) {
1352 			dev_gindex =
1353 			    nxgep->pt_config.hw_config.def_mac_rxdma_grpid +
1354 			    groupid;
1355 
1356 			if (nxgep->pt_config.hw_config.def_mac_rxdma_grpid ==
1357 			    dev_gindex)
1358 				group->port_default_grp = B_TRUE;
1359 
1360 			infop->mgi_count =
1361 			    nxgep->pt_config.rdc_grps[dev_gindex].max_rdcs;
1362 		} else {
1363 			infop->mgi_count = NXGE_HIO_SHARE_MAX_CHANNELS;
1364 		}
1365 
1366 		infop->mgi_driver = (mac_group_driver_t)group;
1367 		infop->mgi_start = nxge_hio_group_start;
1368 		infop->mgi_stop = nxge_hio_group_stop;
1369 		infop->mgi_addmac = nxge_hio_add_mac;
1370 		infop->mgi_remmac = nxge_hio_rem_mac;
1371 		break;
1372 
1373 	case MAC_RING_TYPE_TX:
1374 		/*
1375 		 * 'groupid' for TX should be incremented by one since
1376 		 * the default group (groupid 0) is not known by the MAC layer
1377 		 */
1378 		group = &nxgep->tx_hio_groups[groupid + 1];
1379 		group->nxgep = nxgep;
1380 		group->ghandle = ghdl;
1381 		group->gindex = groupid + 1;
1382 		group->sindex = 0;	/* not yet bound to a share */
1383 
1384 		infop->mgi_driver = (mac_group_driver_t)group;
1385 		infop->mgi_start = NULL;
1386 		infop->mgi_stop = NULL;
1387 		infop->mgi_addmac = NULL;	/* not needed */
1388 		infop->mgi_remmac = NULL;	/* not needed */
1389 		/* no rings associated with group initially */
1390 		infop->mgi_count = 0;
1391 		break;
1392 	}
1393 }
1394 
1395 #if defined(sun4v)
1396 
1397 int
1398 nxge_hio_share_assign(
1399 	nxge_t *nxge,
1400 	uint64_t cookie,
1401 	res_map_t *tmap,
1402 	res_map_t *rmap,
1403 	nxge_hio_vr_t *vr)
1404 {
1405 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
1406 	uint64_t slot, hv_rv;
1407 	nxge_hio_dc_t *dc;
1408 	nxhv_vr_fp_t *fp;
1409 	int i;
1410 	uint64_t major;
1411 
1412 	/*
1413 	 * Ask the Hypervisor to set up the VR for us
1414 	 */
1415 	fp = &nhd->hio.vr;
1416 	major = nxge->niu_hsvc.hsvc_major;
1417 	switch (major) {
1418 	case NIU_MAJOR_VER: /* 1 */
1419 		if ((hv_rv = (*fp->assign)(vr->region, cookie, &vr->cookie))) {
1420 			NXGE_ERROR_MSG((nxge, HIO_CTL,
1421 			    "nxge_hio_share_assign: major %d "
1422 			    "vr->assign() returned %d", major, hv_rv));
1423 			nxge_hio_unshare(vr);
1424 			return (-EIO);
1425 		}
1426 
1427 		break;
1428 
1429 	case NIU_MAJOR_VER_2: /* 2 */
1430 	default:
1431 		if ((hv_rv = (*fp->cfgh_assign)
1432 		    (nxge->niu_cfg_hdl, vr->region, cookie, &vr->cookie))) {
1433 			NXGE_ERROR_MSG((nxge, HIO_CTL,
1434 			    "nxge_hio_share_assign: major %d "
1435 			    "vr->assign() returned %d", major, hv_rv));
1436 			nxge_hio_unshare(vr);
1437 			return (-EIO);
1438 		}
1439 
1440 		break;
1441 	}
1442 
1443 	NXGE_DEBUG_MSG((nxge, HIO_CTL,
1444 	    "nxge_hio_share_assign: major %d "
1445 	    "vr->assign() success", major));
1446 
1447 	/*
1448 	 * For each shared TDC, ask the HV to find us an empty slot.
1449 	 */
1450 	dc = vr->tx_group.dc;
1451 	for (i = 0; i < NXGE_MAX_TDCS; i++) {
1452 		nxhv_dc_fp_t *tx = &nhd->hio.tx;
1453 		while (dc) {
1454 			hv_rv = (*tx->assign)
1455 			    (vr->cookie, dc->channel, &slot);
1456 			if (hv_rv != 0) {
1457 				NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
1458 				    "nxge_hio_share_assign: "
1459 				    "tx->assign(%x, %d) failed: %ld",
1460 				    vr->cookie, dc->channel, hv_rv));
1461 				return (-EIO);
1462 			}
1463 
1464 			dc->cookie = vr->cookie;
1465 			dc->page = (vp_channel_t)slot;
1466 
1467 			/* Inform the caller about the slot chosen. */
1468 			(*tmap) |= 1 << slot;
1469 
1470 			dc = dc->next;
1471 		}
1472 	}
1473 
1474 	/*
1475 	 * For each shared RDC, ask the HV to find us an empty slot.
1476 	 */
1477 	dc = vr->rx_group.dc;
1478 	for (i = 0; i < NXGE_MAX_RDCS; i++) {
1479 		nxhv_dc_fp_t *rx = &nhd->hio.rx;
1480 		while (dc) {
1481 			hv_rv = (*rx->assign)
1482 			    (vr->cookie, dc->channel, &slot);
1483 			if (hv_rv != 0) {
1484 				NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
1485 				    "nxge_hio_share_assign: "
1486 				    "rx->assign(%x, %d) failed: %ld",
1487 				    vr->cookie, dc->channel, hv_rv));
1488 				return (-EIO);
1489 			}
1490 
1491 			dc->cookie = vr->cookie;
1492 			dc->page = (vp_channel_t)slot;
1493 
1494 			/* Inform the caller about the slot chosen. */
1495 			(*rmap) |= 1 << slot;
1496 
1497 			dc = dc->next;
1498 		}
1499 	}
1500 
1501 	return (0);
1502 }
1503 
1504 void
1505 nxge_hio_share_unassign(
1506 	nxge_hio_vr_t *vr)
1507 {
1508 	nxge_t *nxge = (nxge_t *)vr->nxge;
1509 	nxge_hio_data_t *nhd;
1510 	nxge_hio_dc_t *dc;
1511 	nxhv_vr_fp_t *fp;
1512 	uint64_t hv_rv;
1513 
1514 	nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
1515 
1516 	dc = vr->tx_group.dc;
1517 	while (dc) {
1518 		nxhv_dc_fp_t *tx = &nhd->hio.tx;
1519 		hv_rv = (*tx->unassign)(vr->cookie, dc->page);
1520 		if (hv_rv != 0) {
1521 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
1522 			    "nxge_hio_share_unassign: "
1523 			    "tx->unassign(%x, %d) failed: %ld",
1524 			    vr->cookie, dc->page, hv_rv));
1525 		}
1526 		dc = dc->next;
1527 	}
1528 
1529 	dc = vr->rx_group.dc;
1530 	while (dc) {
1531 		nxhv_dc_fp_t *rx = &nhd->hio.rx;
1532 		hv_rv = (*rx->unassign)(vr->cookie, dc->page);
1533 		if (hv_rv != 0) {
1534 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
1535 			    "nxge_hio_share_unassign: "
1536 			    "rx->unassign(%x, %d) failed: %ld",
1537 			    vr->cookie, dc->page, hv_rv));
1538 		}
1539 		dc = dc->next;
1540 	}
1541 
1542 	fp = &nhd->hio.vr;
1543 	if (fp->unassign) {
1544 		hv_rv = (*fp->unassign)(vr->cookie);
1545 		if (hv_rv != 0) {
1546 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
1547 			    "nxge_hio_share_unassign: "
1548 			    "vr->assign(%x) failed: %ld",
1549 			    vr->cookie, hv_rv));
1550 		}
1551 	}
1552 }
1553 
1554 int
1555 nxge_hio_share_alloc(void *arg, mac_share_handle_t *shandle)
1556 {
1557 	p_nxge_t		nxge = (p_nxge_t)arg;
1558 	nxge_share_handle_t	*shp;
1559 	nxge_hio_vr_t		*vr;	/* The Virtualization Region */
1560 	nxge_hio_data_t		*nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
1561 
1562 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_share"));
1563 
1564 	if (nhd->hio.vr.assign == 0 || nhd->hio.tx.assign == 0 ||
1565 	    nhd->hio.rx.assign == 0) {
1566 		NXGE_ERROR_MSG((nxge, HIO_CTL, "HV assign function(s) NULL"));
1567 		return (EIO);
1568 	}
1569 
1570 	/*
1571 	 * Get a VR.
1572 	 */
1573 	if ((vr = nxge_hio_vr_share(nxge)) == 0)
1574 		return (EAGAIN);
1575 
1576 	shp = &nxge->shares[vr->region];
1577 	shp->nxgep = nxge;
1578 	shp->index = vr->region;
1579 	shp->vrp = (void *)vr;
1580 	shp->tmap = shp->rmap = 0;	/* to be assigned by ms_sbind */
1581 	shp->rxgroup = 0;		/* to be assigned by ms_sadd */
1582 	shp->active = B_FALSE;		/* not bound yet */
1583 
1584 	*shandle = (mac_share_handle_t)shp;
1585 
1586 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_share"));
1587 	return (0);
1588 }
1589 
1590 
1591 void
1592 nxge_hio_share_free(mac_share_handle_t shandle)
1593 {
1594 	nxge_share_handle_t	*shp = (nxge_share_handle_t *)shandle;
1595 	nxge_hio_vr_t		*vr;
1596 
1597 	/*
1598 	 * Clear internal handle state.
1599 	 */
1600 	vr = shp->vrp;
1601 	shp->vrp = (void *)NULL;
1602 	shp->index = 0;
1603 	shp->tmap = 0;
1604 	shp->rmap = 0;
1605 	shp->rxgroup = 0;
1606 	shp->active = B_FALSE;
1607 
1608 	/*
1609 	 * Free VR resource.
1610 	 */
1611 	nxge_hio_unshare(vr);
1612 }
1613 
1614 
1615 void
1616 nxge_hio_share_query(mac_share_handle_t shandle, mac_ring_type_t type,
1617     mac_ring_handle_t *rings, uint_t *n_rings)
1618 {
1619 	nxge_t			*nxge;
1620 	nxge_share_handle_t	*shp = (nxge_share_handle_t *)shandle;
1621 	nxge_ring_handle_t	*rh;
1622 	uint32_t		offset;
1623 
1624 	nxge = shp->nxgep;
1625 
1626 	switch (type) {
1627 	case MAC_RING_TYPE_RX:
1628 		rh = nxge->rx_ring_handles;
1629 		offset = nxge->pt_config.hw_config.start_rdc;
1630 		break;
1631 
1632 	case MAC_RING_TYPE_TX:
1633 		rh = nxge->tx_ring_handles;
1634 		offset = nxge->pt_config.hw_config.tdc.start;
1635 		break;
1636 	}
1637 
1638 	/*
1639 	 * In version 1.0, we may only give a VR 2 RDCs/TDCs.  Not only that,
1640 	 * but the HV has statically assigned the channels like so:
1641 	 * VR0: RDC0 & RDC1
1642 	 * VR1: RDC2 & RDC3, etc.
1643 	 * The TDCs are assigned in exactly the same way.
1644 	 */
1645 	if (rings != NULL) {
1646 		rings[0] = rh[(shp->index * 2) - offset].ring_handle;
1647 		rings[1] = rh[(shp->index * 2 + 1) - offset].ring_handle;
1648 	}
1649 	if (n_rings != NULL) {
1650 		*n_rings = 2;
1651 	}
1652 }
1653 
1654 int
1655 nxge_hio_share_add_group(mac_share_handle_t shandle,
1656     mac_group_driver_t ghandle)
1657 {
1658 	nxge_t			*nxge;
1659 	nxge_share_handle_t	*shp = (nxge_share_handle_t *)shandle;
1660 	nxge_ring_group_t	*rg = (nxge_ring_group_t *)ghandle;
1661 	nxge_hio_vr_t		*vr;	/* The Virtualization Region */
1662 	nxge_grp_t		*group;
1663 	int			i;
1664 
1665 	if (rg->sindex != 0) {
1666 		/* the group is already bound to a share */
1667 		return (EALREADY);
1668 	}
1669 
1670 	/*
1671 	 * If we are adding a group 0 to a share, this
1672 	 * is not correct.
1673 	 */
1674 	ASSERT(rg->gindex != 0);
1675 
1676 	nxge = rg->nxgep;
1677 	vr = shp->vrp;
1678 
1679 	switch (rg->type) {
1680 	case MAC_RING_TYPE_RX:
1681 		/*
1682 		 * Make sure that the group has the right rings associated
1683 		 * for the share. In version 1.0, we may only give a VR
1684 		 * 2 RDCs.  Not only that, but the HV has statically
1685 		 * assigned the channels like so:
1686 		 * VR0: RDC0 & RDC1
1687 		 * VR1: RDC2 & RDC3, etc.
1688 		 */
1689 		group = nxge->rx_set.group[rg->gindex];
1690 
1691 		if (group->count > 2) {
1692 			/* a share can have at most 2 rings */
1693 			return (EINVAL);
1694 		}
1695 
1696 		for (i = 0; i < NXGE_MAX_RDCS; i++) {
1697 			if (group->map & (1 << i)) {
1698 				if ((i != shp->index * 2) &&
1699 				    (i != (shp->index * 2 + 1))) {
1700 					/*
1701 					 * A group with invalid rings was
1702 					 * attempted to bind to this share
1703 					 */
1704 					return (EINVAL);
1705 				}
1706 			}
1707 		}
1708 
1709 		rg->sindex = vr->region;
1710 		vr->rdc_tbl = rg->rdctbl;
1711 		shp->rxgroup = vr->rdc_tbl;
1712 		break;
1713 
1714 	case MAC_RING_TYPE_TX:
1715 		/*
1716 		 * Make sure that the group has the right rings associated
1717 		 * for the share. In version 1.0, we may only give a VR
1718 		 * 2 TDCs.  Not only that, but the HV has statically
1719 		 * assigned the channels like so:
1720 		 * VR0: TDC0 & TDC1
1721 		 * VR1: TDC2 & TDC3, etc.
1722 		 */
1723 		group = nxge->tx_set.group[rg->gindex];
1724 
1725 		if (group->count > 2) {
1726 			/* a share can have at most 2 rings */
1727 			return (EINVAL);
1728 		}
1729 
1730 		for (i = 0; i < NXGE_MAX_TDCS; i++) {
1731 			if (group->map & (1 << i)) {
1732 				if ((i != shp->index * 2) &&
1733 				    (i != (shp->index * 2 + 1))) {
1734 					/*
1735 					 * A group with invalid rings was
1736 					 * attempted to bind to this share
1737 					 */
1738 					return (EINVAL);
1739 				}
1740 			}
1741 		}
1742 
1743 		vr->tdc_tbl = nxge->pt_config.hw_config.def_mac_txdma_grpid +
1744 		    rg->gindex;
1745 		rg->sindex = vr->region;
1746 		break;
1747 	}
1748 	return (0);
1749 }
1750 
1751 int
1752 nxge_hio_share_rem_group(mac_share_handle_t shandle,
1753     mac_group_driver_t ghandle)
1754 {
1755 	nxge_share_handle_t	*shp = (nxge_share_handle_t *)shandle;
1756 	nxge_ring_group_t	*group = (nxge_ring_group_t *)ghandle;
1757 	nxge_hio_vr_t		*vr;	/* The Virtualization Region */
1758 	int			rv = 0;
1759 
1760 	vr = shp->vrp;
1761 
1762 	switch (group->type) {
1763 	case MAC_RING_TYPE_RX:
1764 		group->sindex = 0;
1765 		vr->rdc_tbl = 0;
1766 		shp->rxgroup = 0;
1767 		break;
1768 
1769 	case MAC_RING_TYPE_TX:
1770 		group->sindex = 0;
1771 		vr->tdc_tbl = 0;
1772 		break;
1773 	}
1774 
1775 	return (rv);
1776 }
1777 
1778 int
1779 nxge_hio_share_bind(mac_share_handle_t shandle, uint64_t cookie,
1780     uint64_t *rcookie)
1781 {
1782 	nxge_t			*nxge;
1783 	nxge_share_handle_t	*shp = (nxge_share_handle_t *)shandle;
1784 	nxge_hio_vr_t		*vr;
1785 	uint64_t		rmap, tmap, hv_rmap, hv_tmap;
1786 	int			rv;
1787 
1788 	ASSERT(shp != NULL);
1789 	ASSERT(shp->nxgep != NULL);
1790 	ASSERT(shp->vrp != NULL);
1791 
1792 	nxge = shp->nxgep;
1793 	vr = (nxge_hio_vr_t *)shp->vrp;
1794 
1795 	/*
1796 	 * Add resources to the share.
1797 	 * For each DMA channel associated with the VR, bind its resources
1798 	 * to the VR.
1799 	 */
1800 	tmap = 0;
1801 	rv = nxge_hio_addres(vr, MAC_RING_TYPE_TX, &tmap);
1802 	if (rv != 0) {
1803 		return (rv);
1804 	}
1805 
1806 	rmap = 0;
1807 	rv = nxge_hio_addres(vr, MAC_RING_TYPE_RX, &rmap);
1808 	if (rv != 0) {
1809 		nxge_hio_remres(vr, MAC_RING_TYPE_TX, tmap);
1810 		return (rv);
1811 	}
1812 
1813 	/*
1814 	 * Ask the Hypervisor to set up the VR and allocate slots for
1815 	 * each rings associated with the VR.
1816 	 */
1817 	hv_tmap = hv_rmap = 0;
1818 	if ((rv = nxge_hio_share_assign(nxge, cookie,
1819 	    &hv_tmap, &hv_rmap, vr))) {
1820 		nxge_hio_remres(vr, MAC_RING_TYPE_TX, tmap);
1821 		nxge_hio_remres(vr, MAC_RING_TYPE_RX, rmap);
1822 		return (rv);
1823 	}
1824 
1825 	shp->active = B_TRUE;
1826 	shp->tmap = hv_tmap;
1827 	shp->rmap = hv_rmap;
1828 
1829 	/* high 32 bits are cfg_hdl and low 32 bits are HV cookie */
1830 	*rcookie = (((uint64_t)nxge->niu_cfg_hdl) << 32) | vr->cookie;
1831 
1832 	return (0);
1833 }
1834 
1835 void
1836 nxge_hio_share_unbind(mac_share_handle_t shandle)
1837 {
1838 	nxge_share_handle_t *shp = (nxge_share_handle_t *)shandle;
1839 
1840 	/*
1841 	 * First, unassign the VR (take it back),
1842 	 * so we can enable interrupts again.
1843 	 */
1844 	nxge_hio_share_unassign(shp->vrp);
1845 
1846 	/*
1847 	 * Free Ring Resources for TX and RX
1848 	 */
1849 	nxge_hio_remres(shp->vrp, MAC_RING_TYPE_TX, shp->tmap);
1850 	nxge_hio_remres(shp->vrp, MAC_RING_TYPE_RX, shp->rmap);
1851 }
1852 
1853 
1854 /*
1855  * nxge_hio_vr_share
1856  *
1857  *	Find an unused Virtualization Region (VR).
1858  *
1859  * Arguments:
1860  * 	nxge
1861  *
1862  * Notes:
1863  *
1864  * Context:
1865  *	Service domain
1866  */
1867 nxge_hio_vr_t *
1868 nxge_hio_vr_share(
1869 	nxge_t *nxge)
1870 {
1871 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
1872 	nxge_hio_vr_t *vr;
1873 
1874 	int first, limit, region;
1875 
1876 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_vr_share"));
1877 
1878 	MUTEX_ENTER(&nhd->lock);
1879 
1880 	if (nhd->vrs == 0) {
1881 		MUTEX_EXIT(&nhd->lock);
1882 		return (0);
1883 	}
1884 
1885 	/* Find an empty virtual region (VR). */
1886 	if (nxge->function_num == 0) {
1887 		// FUNC0_VIR0 'belongs' to NIU port 0.
1888 		first = FUNC0_VIR1;
1889 		limit = FUNC2_VIR0;
1890 	} else if (nxge->function_num == 1) {
1891 		// FUNC2_VIR0 'belongs' to NIU port 1.
1892 		first = FUNC2_VIR1;
1893 		limit = FUNC_VIR_MAX;
1894 	} else {
1895 		cmn_err(CE_WARN,
1896 		    "Shares not supported on function(%d) at this time.\n",
1897 		    nxge->function_num);
1898 	}
1899 
1900 	for (region = first; region < limit; region++) {
1901 		if (nhd->vr[region].nxge == 0)
1902 			break;
1903 	}
1904 
1905 	if (region == limit) {
1906 		MUTEX_EXIT(&nhd->lock);
1907 		return (0);
1908 	}
1909 
1910 	vr = &nhd->vr[region];
1911 	vr->nxge = (uintptr_t)nxge;
1912 	vr->region = (uintptr_t)region;
1913 
1914 	nhd->vrs--;
1915 
1916 	MUTEX_EXIT(&nhd->lock);
1917 
1918 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_vr_share"));
1919 
1920 	return (vr);
1921 }
1922 
1923 void
1924 nxge_hio_unshare(
1925 	nxge_hio_vr_t *vr)
1926 {
1927 	nxge_t *nxge = (nxge_t *)vr->nxge;
1928 	nxge_hio_data_t *nhd;
1929 
1930 	vr_region_t region;
1931 
1932 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_unshare"));
1933 
1934 	if (!nxge) {
1935 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_unshare: "
1936 		    "vr->nxge is NULL"));
1937 		return;
1938 	}
1939 
1940 	/*
1941 	 * This function is no longer called, but I will keep it
1942 	 * here in case we want to revisit this topic in the future.
1943 	 *
1944 	 * nxge_hio_hostinfo_uninit(nxge, vr);
1945 	 */
1946 
1947 	/*
1948 	 * XXX: This is done by ms_sremove?
1949 	 * (void) nxge_fzc_rdc_tbl_unbind(nxge, vr->rdc_tbl);
1950 	 */
1951 
1952 	nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
1953 
1954 	MUTEX_ENTER(&nhd->lock);
1955 
1956 	region = vr->region;
1957 	(void) memset(vr, 0, sizeof (*vr));
1958 	vr->region = region;
1959 
1960 	nhd->vrs++;
1961 
1962 	MUTEX_EXIT(&nhd->lock);
1963 
1964 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_unshare"));
1965 }
1966 
1967 int
1968 nxge_hio_addres(nxge_hio_vr_t *vr, mac_ring_type_t type, uint64_t *map)
1969 {
1970 	nxge_t		*nxge;
1971 	nxge_grp_t	*group;
1972 	int		groupid;
1973 	int		i, rv = 0;
1974 	int		max_dcs;
1975 
1976 	ASSERT(vr != NULL);
1977 	ASSERT(vr->nxge != NULL);
1978 	nxge = (nxge_t *)vr->nxge;
1979 
1980 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_addres"));
1981 
1982 	/*
1983 	 * For each ring associated with the group, add the resources
1984 	 * to the group and bind.
1985 	 */
1986 	max_dcs = (type == MAC_RING_TYPE_TX) ? NXGE_MAX_TDCS : NXGE_MAX_RDCS;
1987 	if (type == MAC_RING_TYPE_TX) {
1988 		/* set->group is an array of group indexed by a port group id */
1989 		groupid = vr->tdc_tbl -
1990 		    nxge->pt_config.hw_config.def_mac_txdma_grpid;
1991 		group = nxge->tx_set.group[groupid];
1992 	} else {
1993 		/* set->group is an array of group indexed by a port group id */
1994 		groupid = vr->rdc_tbl -
1995 		    nxge->pt_config.hw_config.def_mac_rxdma_grpid;
1996 		group = nxge->rx_set.group[groupid];
1997 	}
1998 
1999 	ASSERT(group != NULL);
2000 
2001 	if (group->map == 0) {
2002 		NXGE_DEBUG_MSG((nxge, HIO_CTL, "There is no rings associated "
2003 		    "with this VR"));
2004 		return (EINVAL);
2005 	}
2006 
2007 	for (i = 0; i < max_dcs; i++) {
2008 		if (group->map & (1 << i)) {
2009 			if ((rv = nxge_hio_dc_share(nxge, vr, type, i)) < 0) {
2010 				if (*map == 0) /* Couldn't get even one DC. */
2011 					return (-rv);
2012 				else
2013 					break;
2014 			}
2015 			*map |= (1 << i);
2016 		}
2017 	}
2018 
2019 	if ((*map == 0) || (rv != 0)) {
2020 		NXGE_DEBUG_MSG((nxge, HIO_CTL,
2021 		    "<== nxge_hio_addres: rv(%x)", rv));
2022 		return (EIO);
2023 	}
2024 
2025 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_addres"));
2026 	return (0);
2027 }
2028 
2029 /* ARGSUSED */
2030 void
2031 nxge_hio_remres(
2032 	nxge_hio_vr_t *vr,
2033 	mac_ring_type_t type,
2034 	res_map_t res_map)
2035 {
2036 	nxge_t *nxge = (nxge_t *)vr->nxge;
2037 	nxge_grp_t *group;
2038 
2039 	if (!nxge) {
2040 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_remres: "
2041 		    "vr->nxge is NULL"));
2042 		return;
2043 	}
2044 
2045 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_remres(%lx)", res_map));
2046 
2047 	/*
2048 	 * For each ring bound to the group, remove the DMA resources
2049 	 * from the group and unbind.
2050 	 */
2051 	group = (type == MAC_RING_TYPE_TX ? &vr->tx_group : &vr->rx_group);
2052 	while (group->dc) {
2053 		nxge_hio_dc_t *dc = group->dc;
2054 		NXGE_DC_RESET(res_map, dc->page);
2055 		nxge_hio_dc_unshare(nxge, vr, type, dc->channel);
2056 	}
2057 
2058 	if (res_map) {
2059 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_remres: "
2060 		    "res_map %lx", res_map));
2061 	}
2062 
2063 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_remres"));
2064 }
2065 
2066 /*
2067  * nxge_hio_tdc_share
2068  *
2069  *	Share an unused TDC channel.
2070  *
2071  * Arguments:
2072  * 	nxge
2073  *
2074  * Notes:
2075  *
2076  * A.7.3 Reconfigure Tx DMA channel
2077  *	Disable TxDMA			A.9.6.10
2078  *     [Rebind TxDMA channel to Port	A.9.6.7]
2079  *
2080  * We don't have to Rebind the TDC to the port - it always already bound.
2081  *
2082  *	Soft Reset TxDMA		A.9.6.2
2083  *
2084  * This procedure will be executed by nxge_init_txdma_channel() in the
2085  * guest domain:
2086  *
2087  *	Re-initialize TxDMA		A.9.6.8
2088  *	Reconfigure TxDMA
2089  *	Enable TxDMA			A.9.6.9
2090  *
2091  * Context:
2092  *	Service domain
2093  */
2094 int
2095 nxge_hio_tdc_share(
2096 	nxge_t *nxge,
2097 	int channel)
2098 {
2099 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
2100 	nxge_grp_set_t *set = &nxge->tx_set;
2101 	tx_ring_t *ring;
2102 	int count;
2103 
2104 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_tdc_share"));
2105 
2106 	/*
2107 	 * Wait until this channel is idle.
2108 	 */
2109 	ring = nxge->tx_rings->rings[channel];
2110 	ASSERT(ring != NULL);
2111 
2112 	(void) atomic_swap_32(&ring->tx_ring_offline, NXGE_TX_RING_OFFLINING);
2113 	if (ring->tx_ring_busy) {
2114 		/*
2115 		 * Wait for 30 seconds.
2116 		 */
2117 		for (count = 30 * 1000; count; count--) {
2118 			if (ring->tx_ring_offline & NXGE_TX_RING_OFFLINED) {
2119 				break;
2120 			}
2121 
2122 			drv_usecwait(1000);
2123 		}
2124 
2125 		if (count == 0) {
2126 			(void) atomic_swap_32(&ring->tx_ring_offline,
2127 			    NXGE_TX_RING_ONLINE);
2128 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2129 			    "nxge_hio_tdc_share: "
2130 			    "Tx ring %d was always BUSY", channel));
2131 			return (-EIO);
2132 		}
2133 	} else {
2134 		(void) atomic_swap_32(&ring->tx_ring_offline,
2135 		    NXGE_TX_RING_OFFLINED);
2136 	}
2137 
2138 	MUTEX_ENTER(&nhd->lock);
2139 	nxge->tdc_is_shared[channel] = B_TRUE;
2140 	MUTEX_EXIT(&nhd->lock);
2141 
2142 	if (nxge_intr_remove(nxge, VP_BOUND_TX, channel) != NXGE_OK) {
2143 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_tdc_share: "
2144 		    "Failed to remove interrupt for TxDMA channel %d",
2145 		    channel));
2146 		return (-EINVAL);
2147 	}
2148 
2149 	/* Disable TxDMA A.9.6.10 */
2150 	(void) nxge_txdma_channel_disable(nxge, channel);
2151 
2152 	/* The SD is sharing this channel. */
2153 	NXGE_DC_SET(set->shared.map, channel);
2154 	set->shared.count++;
2155 
2156 	/* Soft Reset TxDMA A.9.6.2 */
2157 	nxge_grp_dc_remove(nxge, VP_BOUND_TX, channel);
2158 
2159 	/*
2160 	 * Initialize the DC-specific FZC control registers.
2161 	 * -----------------------------------------------------
2162 	 */
2163 	if (nxge_init_fzc_tdc(nxge, channel) != NXGE_OK) {
2164 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2165 		    "nxge_hio_tdc_share: FZC TDC failed: %d", channel));
2166 		return (-EIO);
2167 	}
2168 
2169 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_tdc_share"));
2170 
2171 	return (0);
2172 }
2173 
2174 /*
2175  * nxge_hio_rdc_share
2176  *
2177  *	Share an unused RDC channel.
2178  *
2179  * Arguments:
2180  * 	nxge
2181  *
2182  * Notes:
2183  *
2184  * This is the latest version of the procedure to
2185  * Reconfigure an Rx DMA channel:
2186  *
2187  * A.6.3 Reconfigure Rx DMA channel
2188  *	Stop RxMAC		A.9.2.6
2189  *	Drain IPP Port		A.9.3.6
2190  *	Stop and reset RxDMA	A.9.5.3
2191  *
2192  * This procedure will be executed by nxge_init_rxdma_channel() in the
2193  * guest domain:
2194  *
2195  *	Initialize RxDMA	A.9.5.4
2196  *	Reconfigure RxDMA
2197  *	Enable RxDMA		A.9.5.5
2198  *
2199  * We will do this here, since the RDC is a canalis non grata:
2200  *	Enable RxMAC		A.9.2.10
2201  *
2202  * Context:
2203  *	Service domain
2204  */
2205 int
2206 nxge_hio_rdc_share(
2207 	nxge_t *nxge,
2208 	nxge_hio_vr_t *vr,
2209 	int channel)
2210 {
2211 	nxge_grp_set_t *set = &nxge->rx_set;
2212 	nxge_rdc_grp_t *rdc_grp;
2213 
2214 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_rdc_share"));
2215 
2216 	/* Disable interrupts. */
2217 	if (nxge_intr_remove(nxge, VP_BOUND_RX, channel) != NXGE_OK) {
2218 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_rdc_share: "
2219 		    "Failed to remove interrupt for RxDMA channel %d",
2220 		    channel));
2221 		return (NXGE_ERROR);
2222 	}
2223 
2224 	/* Stop RxMAC = A.9.2.6 */
2225 	if (nxge_rx_mac_disable(nxge) != NXGE_OK) {
2226 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_rdc_share: "
2227 		    "Failed to disable RxMAC"));
2228 	}
2229 
2230 	/* Drain IPP Port = A.9.3.6 */
2231 	(void) nxge_ipp_drain(nxge);
2232 
2233 	/* Stop and reset RxDMA = A.9.5.3 */
2234 	// De-assert EN: RXDMA_CFIG1[31] = 0 (DMC+00000 )
2235 	if (nxge_disable_rxdma_channel(nxge, channel) != NXGE_OK) {
2236 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_rdc_share: "
2237 		    "Failed to disable RxDMA channel %d", channel));
2238 	}
2239 
2240 	/* The SD is sharing this channel. */
2241 	NXGE_DC_SET(set->shared.map, channel);
2242 	set->shared.count++;
2243 
2244 	// Assert RST: RXDMA_CFIG1[30] = 1
2245 	nxge_grp_dc_remove(nxge, VP_BOUND_RX, channel);
2246 
2247 	/*
2248 	 * The guest domain will reconfigure the RDC later.
2249 	 *
2250 	 * But in the meantime, we must re-enable the Rx MAC so
2251 	 * that we can start receiving packets again on the
2252 	 * remaining RDCs:
2253 	 *
2254 	 * Enable RxMAC = A.9.2.10
2255 	 */
2256 	if (nxge_rx_mac_enable(nxge) != NXGE_OK) {
2257 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2258 		    "nxge_hio_rdc_share: Rx MAC still disabled"));
2259 	}
2260 
2261 	/*
2262 	 * Initialize the DC-specific FZC control registers.
2263 	 * -----------------------------------------------------
2264 	 */
2265 	if (nxge_init_fzc_rdc(nxge, channel) != NXGE_OK) {
2266 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2267 		    "nxge_hio_rdc_share: RZC RDC failed: %ld", channel));
2268 		return (-EIO);
2269 	}
2270 
2271 	/*
2272 	 * Update the RDC group.
2273 	 */
2274 	rdc_grp = &nxge->pt_config.rdc_grps[vr->rdc_tbl];
2275 	NXGE_DC_SET(rdc_grp->map, channel);
2276 
2277 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_rdc_share"));
2278 
2279 	return (0);
2280 }
2281 
2282 /*
2283  * nxge_hio_dc_share
2284  *
2285  *	Share a DMA channel with a guest domain.
2286  *
2287  * Arguments:
2288  * 	nxge
2289  * 	vr	The VR that <channel> will belong to.
2290  * 	type	Tx or Rx.
2291  * 	channel	Channel to share
2292  *
2293  * Notes:
2294  *
2295  * Context:
2296  *	Service domain
2297  */
2298 int
2299 nxge_hio_dc_share(
2300 	nxge_t *nxge,
2301 	nxge_hio_vr_t *vr,
2302 	mac_ring_type_t type,
2303 	int channel)
2304 {
2305 	nxge_hio_data_t *nhd = (nxge_hio_data_t *)nxge->nxge_hw_p->hio;
2306 	nxge_hio_dc_t *dc;
2307 	nxge_grp_t *group;
2308 	int slot;
2309 
2310 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_dc_share(%cdc %d",
2311 	    type == MAC_RING_TYPE_TX ? 't' : 'r', channel));
2312 
2313 
2314 	/* -------------------------------------------------- */
2315 	slot = (type == MAC_RING_TYPE_TX) ?
2316 	    nxge_hio_tdc_share(nxge, channel) :
2317 	    nxge_hio_rdc_share(nxge, vr, channel);
2318 
2319 	if (slot < 0) {
2320 		if (type == MAC_RING_TYPE_RX) {
2321 			nxge_hio_rdc_unshare(nxge, vr->rdc_tbl, channel);
2322 		} else {
2323 			nxge_hio_tdc_unshare(nxge, vr->tdc_tbl, channel);
2324 		}
2325 		return (slot);
2326 	}
2327 
2328 	MUTEX_ENTER(&nhd->lock);
2329 
2330 	/*
2331 	 * Tag this channel.
2332 	 * --------------------------------------------------
2333 	 */
2334 	dc = type == MAC_RING_TYPE_TX ? &nhd->tdc[channel] : &nhd->rdc[channel];
2335 
2336 	dc->vr = vr;
2337 	dc->channel = (nxge_channel_t)channel;
2338 
2339 	MUTEX_EXIT(&nhd->lock);
2340 
2341 	/*
2342 	 * vr->[t|r]x_group is used by the service domain to
2343 	 * keep track of its shared DMA channels.
2344 	 */
2345 	MUTEX_ENTER(&nxge->group_lock);
2346 	group = (type == MAC_RING_TYPE_TX ? &vr->tx_group : &vr->rx_group);
2347 
2348 	dc->group = group;
2349 	/* Initialize <group>, if necessary */
2350 	if (group->count == 0) {
2351 		group->nxge = nxge;
2352 		group->type = (type == MAC_RING_TYPE_TX) ?
2353 		    VP_BOUND_TX : VP_BOUND_RX;
2354 		group->sequence	= nhd->sequence++;
2355 		group->active = B_TRUE;
2356 	}
2357 
2358 	MUTEX_EXIT(&nxge->group_lock);
2359 
2360 	NXGE_ERROR_MSG((nxge, HIO_CTL,
2361 	    "DC share: %cDC %d was assigned to slot %d",
2362 	    type == MAC_RING_TYPE_TX ? 'T' : 'R', channel, slot));
2363 
2364 	nxge_grp_dc_append(nxge, group, dc);
2365 
2366 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_dc_share"));
2367 
2368 	return (0);
2369 }
2370 
2371 /*
2372  * nxge_hio_tdc_unshare
2373  *
2374  *	Unshare a TDC.
2375  *
2376  * Arguments:
2377  * 	nxge
2378  * 	channel	The channel to unshare (add again).
2379  *
2380  * Notes:
2381  *
2382  * Context:
2383  *	Service domain
2384  */
2385 void
2386 nxge_hio_tdc_unshare(
2387 	nxge_t *nxge,
2388 	int dev_grpid,
2389 	int channel)
2390 {
2391 	nxge_grp_set_t *set = &nxge->tx_set;
2392 	nxge_grp_t *group;
2393 	int grpid;
2394 
2395 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_tdc_unshare"));
2396 
2397 	NXGE_DC_RESET(set->shared.map, channel);
2398 	set->shared.count--;
2399 
2400 	grpid = dev_grpid - nxge->pt_config.hw_config.def_mac_txdma_grpid;
2401 	group = set->group[grpid];
2402 
2403 	if ((nxge_grp_dc_add(nxge, group, VP_BOUND_TX, channel))) {
2404 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_tdc_unshare: "
2405 		    "Failed to initialize TxDMA channel %d", channel));
2406 		return;
2407 	}
2408 
2409 	/* Re-add this interrupt. */
2410 	if (nxge_intr_add(nxge, VP_BOUND_TX, channel) != NXGE_OK) {
2411 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_tdc_unshare: "
2412 		    "Failed to add interrupt for TxDMA channel %d", channel));
2413 	}
2414 
2415 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_tdc_unshare"));
2416 }
2417 
2418 /*
2419  * nxge_hio_rdc_unshare
2420  *
2421  *	Unshare an RDC: add it to the SD's RDC groups (tables).
2422  *
2423  * Arguments:
2424  * 	nxge
2425  * 	channel	The channel to unshare (add again).
2426  *
2427  * Notes:
2428  *
2429  * Context:
2430  *	Service domain
2431  */
2432 void
2433 nxge_hio_rdc_unshare(
2434 	nxge_t *nxge,
2435 	int dev_grpid,
2436 	int channel)
2437 {
2438 	nxge_grp_set_t		*set = &nxge->rx_set;
2439 	nxge_grp_t		*group;
2440 	int			grpid;
2441 	int			i;
2442 
2443 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_rdc_unshare"));
2444 
2445 	/* Stop RxMAC = A.9.2.6 */
2446 	if (nxge_rx_mac_disable(nxge) != NXGE_OK) {
2447 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_rdc_unshare: "
2448 		    "Failed to disable RxMAC"));
2449 	}
2450 
2451 	/* Drain IPP Port = A.9.3.6 */
2452 	(void) nxge_ipp_drain(nxge);
2453 
2454 	/* Stop and reset RxDMA = A.9.5.3 */
2455 	// De-assert EN: RXDMA_CFIG1[31] = 0 (DMC+00000 )
2456 	if (nxge_disable_rxdma_channel(nxge, channel) != NXGE_OK) {
2457 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_rdc_unshare: "
2458 		    "Failed to disable RxDMA channel %d", channel));
2459 	}
2460 
2461 	NXGE_DC_RESET(set->shared.map, channel);
2462 	set->shared.count--;
2463 
2464 	grpid = dev_grpid - nxge->pt_config.hw_config.def_mac_rxdma_grpid;
2465 	group = set->group[grpid];
2466 
2467 	/*
2468 	 * Assert RST: RXDMA_CFIG1[30] = 1
2469 	 *
2470 	 * Initialize RxDMA	A.9.5.4
2471 	 * Reconfigure RxDMA
2472 	 * Enable RxDMA		A.9.5.5
2473 	 */
2474 	if ((nxge_grp_dc_add(nxge, group, VP_BOUND_RX, channel))) {
2475 		/* Be sure to re-enable the RX MAC. */
2476 		if (nxge_rx_mac_enable(nxge) != NXGE_OK) {
2477 			NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2478 			    "nxge_hio_rdc_share: Rx MAC still disabled"));
2479 		}
2480 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL, "nxge_hio_rdc_unshare: "
2481 		    "Failed to initialize RxDMA channel %d", channel));
2482 		return;
2483 	}
2484 
2485 	/*
2486 	 * Enable RxMAC = A.9.2.10
2487 	 */
2488 	if (nxge_rx_mac_enable(nxge) != NXGE_OK) {
2489 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2490 		    "nxge_hio_rdc_share: Rx MAC still disabled"));
2491 		return;
2492 	}
2493 
2494 	/* Re-add this interrupt. */
2495 	if (nxge_intr_add(nxge, VP_BOUND_RX, channel) != NXGE_OK) {
2496 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2497 		    "nxge_hio_rdc_unshare: Failed to add interrupt for "
2498 		    "RxDMA CHANNEL %d", channel));
2499 	}
2500 
2501 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_rdc_unshare"));
2502 
2503 	for (i = 0; i < NXGE_MAX_RDCS; i++) {
2504 		if (nxge->rx_ring_handles[i].channel == channel) {
2505 			nxge_rx_ring_start(
2506 			    (mac_ring_driver_t)&nxge->rx_ring_handles[i],
2507 			    nxge->rx_ring_handles[i].ring_gen_num);
2508 		}
2509 	}
2510 }
2511 
2512 /*
2513  * nxge_hio_dc_unshare
2514  *
2515  *	Unshare (reuse) a DMA channel.
2516  *
2517  * Arguments:
2518  * 	nxge
2519  * 	vr	The VR that <channel> belongs to.
2520  * 	type	Tx or Rx.
2521  * 	channel	The DMA channel to reuse.
2522  *
2523  * Notes:
2524  *
2525  * Context:
2526  *	Service domain
2527  */
2528 void
2529 nxge_hio_dc_unshare(
2530 	nxge_t *nxge,
2531 	nxge_hio_vr_t *vr,
2532 	mac_ring_type_t type,
2533 	int channel)
2534 {
2535 	nxge_grp_t *group;
2536 	nxge_hio_dc_t *dc;
2537 
2538 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "==> nxge_hio_dc_unshare(%cdc %d)",
2539 	    type == MAC_RING_TYPE_TX ? 't' : 'r', channel));
2540 
2541 	/* Unlink the channel from its group. */
2542 	/* -------------------------------------------------- */
2543 	group = (type == MAC_RING_TYPE_TX) ? &vr->tx_group : &vr->rx_group;
2544 	NXGE_DC_RESET(group->map, channel);
2545 	if ((dc = nxge_grp_dc_unlink(nxge, group, channel)) == 0) {
2546 		NXGE_ERROR_MSG((nxge, NXGE_ERR_CTL,
2547 		    "nxge_hio_dc_unshare(%d) failed", channel));
2548 		return;
2549 	}
2550 
2551 	dc->vr = 0;
2552 	dc->cookie = 0;
2553 
2554 	if (type == MAC_RING_TYPE_RX) {
2555 		nxge_hio_rdc_unshare(nxge, vr->rdc_tbl, channel);
2556 	} else {
2557 		nxge_hio_tdc_unshare(nxge, vr->tdc_tbl, channel);
2558 	}
2559 
2560 	NXGE_DEBUG_MSG((nxge, HIO_CTL, "<== nxge_hio_dc_unshare"));
2561 }
2562 
2563 
2564 /*
2565  * nxge_hio_rxdma_bind_intr():
2566  *
2567  *	For the guest domain driver, need to bind the interrupt group
2568  *	and state to the rx_rcr_ring_t.
2569  */
2570 
2571 int
2572 nxge_hio_rxdma_bind_intr(nxge_t *nxge, rx_rcr_ring_t *ring, int channel)
2573 {
2574 	nxge_hio_dc_t	*dc;
2575 	nxge_ldgv_t	*control;
2576 	nxge_ldg_t	*group;
2577 	nxge_ldv_t	*device;
2578 
2579 	/*
2580 	 * Find the DMA channel.
2581 	 */
2582 	if (!(dc = nxge_grp_dc_find(nxge, VP_BOUND_RX, channel))) {
2583 		return (NXGE_ERROR);
2584 	}
2585 
2586 	/*
2587 	 * Get the control structure.
2588 	 */
2589 	control = nxge->ldgvp;
2590 	if (control == NULL) {
2591 		return (NXGE_ERROR);
2592 	}
2593 
2594 	group = &control->ldgp[dc->ldg.vector];
2595 	device = &control->ldvp[dc->ldg.ldsv];
2596 
2597 	MUTEX_ENTER(&ring->lock);
2598 	ring->ldgp = group;
2599 	ring->ldvp = device;
2600 	MUTEX_EXIT(&ring->lock);
2601 
2602 	return (NXGE_OK);
2603 }
2604 #endif	/* if defined(sun4v) */
2605