xref: /titanic_50/usr/src/uts/common/io/nxge/nxge_main.c (revision 567cc2e6a4ceb0e421e1cf9bead0f43c55603d27)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #pragma ident	"%Z%%M%	%I%	%E% SMI"
27 
28 /*
29  * SunOs MT STREAMS NIU/Neptune 10Gb Ethernet Device Driver.
30  */
31 #include	<sys/nxge/nxge_impl.h>
32 #include	<sys/pcie.h>
33 
34 uint32_t 	nxge_use_partition = 0;		/* debug partition flag */
35 uint32_t 	nxge_dma_obp_props_only = 1;	/* use obp published props */
36 uint32_t 	nxge_use_rdc_intr = 1;		/* debug to assign rdc intr */
37 /*
38  * PSARC/2007/453 MSI-X interrupt limit override
39  * (This PSARC case is limited to MSI-X vectors
40  *  and SPARC platforms only).
41  */
42 #if defined(_BIG_ENDIAN)
43 uint32_t	nxge_msi_enable = 2;
44 #else
45 uint32_t	nxge_msi_enable = 1;
46 #endif
47 
48 /*
49  * Globals: tunable parameters (/etc/system or adb)
50  *
51  */
52 uint32_t 	nxge_rbr_size = NXGE_RBR_RBB_DEFAULT;
53 uint32_t 	nxge_rbr_spare_size = 0;
54 uint32_t 	nxge_rcr_size = NXGE_RCR_DEFAULT;
55 uint32_t 	nxge_tx_ring_size = NXGE_TX_RING_DEFAULT;
56 boolean_t 	nxge_no_msg = B_TRUE;		/* control message display */
57 uint32_t 	nxge_no_link_notify = 0;	/* control DL_NOTIFY */
58 uint32_t 	nxge_bcopy_thresh = TX_BCOPY_MAX;
59 uint32_t 	nxge_dvma_thresh = TX_FASTDVMA_MIN;
60 uint32_t 	nxge_dma_stream_thresh = TX_STREAM_MIN;
61 uint32_t	nxge_jumbo_mtu	= TX_JUMBO_MTU;
62 boolean_t	nxge_jumbo_enable = B_FALSE;
63 uint16_t	nxge_rcr_timeout = NXGE_RDC_RCR_TIMEOUT;
64 uint16_t	nxge_rcr_threshold = NXGE_RDC_RCR_THRESHOLD;
65 nxge_tx_mode_t	nxge_tx_scheme = NXGE_USE_SERIAL;
66 
67 /* MAX LSO size */
68 #define		NXGE_LSO_MAXLEN	65535
69 uint32_t	nxge_lso_max = NXGE_LSO_MAXLEN;
70 
71 /*
72  * Debugging flags:
73  *		nxge_no_tx_lb : transmit load balancing
74  *		nxge_tx_lb_policy: 0 - TCP port (default)
75  *				   3 - DEST MAC
76  */
77 uint32_t 	nxge_no_tx_lb = 0;
78 uint32_t 	nxge_tx_lb_policy = NXGE_TX_LB_TCPUDP;
79 
80 /*
81  * Add tunable to reduce the amount of time spent in the
82  * ISR doing Rx Processing.
83  */
84 uint32_t nxge_max_rx_pkts = 1024;
85 
86 /*
87  * Tunables to manage the receive buffer blocks.
88  *
89  * nxge_rx_threshold_hi: copy all buffers.
90  * nxge_rx_bcopy_size_type: receive buffer block size type.
91  * nxge_rx_threshold_lo: copy only up to tunable block size type.
92  */
93 nxge_rxbuf_threshold_t nxge_rx_threshold_hi = NXGE_RX_COPY_6;
94 nxge_rxbuf_type_t nxge_rx_buf_size_type = RCR_PKTBUFSZ_0;
95 nxge_rxbuf_threshold_t nxge_rx_threshold_lo = NXGE_RX_COPY_3;
96 
97 rtrace_t npi_rtracebuf;
98 
99 #if	defined(sun4v)
100 /*
101  * Hypervisor N2/NIU services information.
102  */
103 static hsvc_info_t niu_hsvc = {
104 	HSVC_REV_1, NULL, HSVC_GROUP_NIU, NIU_MAJOR_VER,
105 	NIU_MINOR_VER, "nxge"
106 };
107 #endif
108 
109 /*
110  * Function Prototypes
111  */
112 static int nxge_attach(dev_info_t *, ddi_attach_cmd_t);
113 static int nxge_detach(dev_info_t *, ddi_detach_cmd_t);
114 static void nxge_unattach(p_nxge_t);
115 
116 #if NXGE_PROPERTY
117 static void nxge_remove_hard_properties(p_nxge_t);
118 #endif
119 
120 static nxge_status_t nxge_setup_system_dma_pages(p_nxge_t);
121 
122 static nxge_status_t nxge_setup_mutexes(p_nxge_t);
123 static void nxge_destroy_mutexes(p_nxge_t);
124 
125 static nxge_status_t nxge_map_regs(p_nxge_t nxgep);
126 static void nxge_unmap_regs(p_nxge_t nxgep);
127 #ifdef	NXGE_DEBUG
128 static void nxge_test_map_regs(p_nxge_t nxgep);
129 #endif
130 
131 static nxge_status_t nxge_add_intrs(p_nxge_t nxgep);
132 static nxge_status_t nxge_add_soft_intrs(p_nxge_t nxgep);
133 static void nxge_remove_intrs(p_nxge_t nxgep);
134 static void nxge_remove_soft_intrs(p_nxge_t nxgep);
135 
136 static nxge_status_t nxge_add_intrs_adv(p_nxge_t nxgep);
137 static nxge_status_t nxge_add_intrs_adv_type(p_nxge_t, uint32_t);
138 static nxge_status_t nxge_add_intrs_adv_type_fix(p_nxge_t, uint32_t);
139 static void nxge_intrs_enable(p_nxge_t nxgep);
140 static void nxge_intrs_disable(p_nxge_t nxgep);
141 
142 static void nxge_suspend(p_nxge_t);
143 static nxge_status_t nxge_resume(p_nxge_t);
144 
145 static nxge_status_t nxge_setup_dev(p_nxge_t);
146 static void nxge_destroy_dev(p_nxge_t);
147 
148 static nxge_status_t nxge_alloc_mem_pool(p_nxge_t);
149 static void nxge_free_mem_pool(p_nxge_t);
150 
151 static nxge_status_t nxge_alloc_rx_mem_pool(p_nxge_t);
152 static void nxge_free_rx_mem_pool(p_nxge_t);
153 
154 static nxge_status_t nxge_alloc_tx_mem_pool(p_nxge_t);
155 static void nxge_free_tx_mem_pool(p_nxge_t);
156 
157 static nxge_status_t nxge_dma_mem_alloc(p_nxge_t, dma_method_t,
158 	struct ddi_dma_attr *,
159 	size_t, ddi_device_acc_attr_t *, uint_t,
160 	p_nxge_dma_common_t);
161 
162 static void nxge_dma_mem_free(p_nxge_dma_common_t);
163 
164 static nxge_status_t nxge_alloc_rx_buf_dma(p_nxge_t, uint16_t,
165 	p_nxge_dma_common_t *, size_t, size_t, uint32_t *);
166 static void nxge_free_rx_buf_dma(p_nxge_t, p_nxge_dma_common_t, uint32_t);
167 
168 static nxge_status_t nxge_alloc_rx_cntl_dma(p_nxge_t, uint16_t,
169 	p_nxge_dma_common_t *, size_t);
170 static void nxge_free_rx_cntl_dma(p_nxge_t, p_nxge_dma_common_t);
171 
172 static nxge_status_t nxge_alloc_tx_buf_dma(p_nxge_t, uint16_t,
173 	p_nxge_dma_common_t *, size_t, size_t, uint32_t *);
174 static void nxge_free_tx_buf_dma(p_nxge_t, p_nxge_dma_common_t, uint32_t);
175 
176 static nxge_status_t nxge_alloc_tx_cntl_dma(p_nxge_t, uint16_t,
177 	p_nxge_dma_common_t *,
178 	size_t);
179 static void nxge_free_tx_cntl_dma(p_nxge_t, p_nxge_dma_common_t);
180 
181 static int nxge_init_common_dev(p_nxge_t);
182 static void nxge_uninit_common_dev(p_nxge_t);
183 
184 /*
185  * The next declarations are for the GLDv3 interface.
186  */
187 static int nxge_m_start(void *);
188 static void nxge_m_stop(void *);
189 static int nxge_m_unicst(void *, const uint8_t *);
190 static int nxge_m_multicst(void *, boolean_t, const uint8_t *);
191 static int nxge_m_promisc(void *, boolean_t);
192 static void nxge_m_ioctl(void *, queue_t *, mblk_t *);
193 static void nxge_m_resources(void *);
194 mblk_t *nxge_m_tx(void *arg, mblk_t *);
195 static nxge_status_t nxge_mac_register(p_nxge_t);
196 static int nxge_altmac_set(p_nxge_t nxgep, uint8_t *mac_addr,
197 	mac_addr_slot_t slot);
198 static void nxge_mmac_kstat_update(p_nxge_t nxgep, mac_addr_slot_t slot,
199 	boolean_t factory);
200 static int nxge_m_mmac_add(void *arg, mac_multi_addr_t *maddr);
201 static int nxge_m_mmac_reserve(void *arg, mac_multi_addr_t *maddr);
202 static int nxge_m_mmac_remove(void *arg, mac_addr_slot_t slot);
203 static int nxge_m_mmac_modify(void *arg, mac_multi_addr_t *maddr);
204 static int nxge_m_mmac_get(void *arg, mac_multi_addr_t *maddr);
205 
206 #define	NXGE_NEPTUNE_MAGIC	0x4E584745UL
207 #define	MAX_DUMP_SZ 256
208 
209 #define	NXGE_M_CALLBACK_FLAGS	(MC_RESOURCES | MC_IOCTL | MC_GETCAPAB)
210 
211 static	boolean_t	nxge_m_getcapab(void *, mac_capab_t, void *);
212 static mac_callbacks_t nxge_m_callbacks = {
213 	NXGE_M_CALLBACK_FLAGS,
214 	nxge_m_stat,
215 	nxge_m_start,
216 	nxge_m_stop,
217 	nxge_m_promisc,
218 	nxge_m_multicst,
219 	nxge_m_unicst,
220 	nxge_m_tx,
221 	nxge_m_resources,
222 	nxge_m_ioctl,
223 	nxge_m_getcapab
224 };
225 
226 void
227 nxge_err_inject(p_nxge_t, queue_t *, mblk_t *);
228 
229 /* PSARC/2007/453 MSI-X interrupt limit override. */
230 #define	NXGE_MSIX_REQUEST_10G	8
231 #define	NXGE_MSIX_REQUEST_1G	2
232 static int nxge_create_msi_property(p_nxge_t);
233 
234 /*
235  * These global variables control the message
236  * output.
237  */
238 out_dbgmsg_t nxge_dbgmsg_out = DBG_CONSOLE | STR_LOG;
239 uint64_t nxge_debug_level = 0;
240 
241 /*
242  * This list contains the instance structures for the Neptune
243  * devices present in the system. The lock exists to guarantee
244  * mutually exclusive access to the list.
245  */
246 void 			*nxge_list = NULL;
247 
248 void			*nxge_hw_list = NULL;
249 nxge_os_mutex_t 	nxge_common_lock;
250 
251 extern uint64_t 	npi_debug_level;
252 
253 extern nxge_status_t	nxge_ldgv_init(p_nxge_t, int *, int *);
254 extern nxge_status_t	nxge_ldgv_init_n2(p_nxge_t, int *, int *);
255 extern nxge_status_t	nxge_ldgv_uninit(p_nxge_t);
256 extern nxge_status_t	nxge_intr_ldgv_init(p_nxge_t);
257 extern void		nxge_fm_init(p_nxge_t,
258 					ddi_device_acc_attr_t *,
259 					ddi_device_acc_attr_t *,
260 					ddi_dma_attr_t *);
261 extern void		nxge_fm_fini(p_nxge_t);
262 extern npi_status_t	npi_mac_altaddr_disable(npi_handle_t, uint8_t, uint8_t);
263 
264 /*
265  * Count used to maintain the number of buffers being used
266  * by Neptune instances and loaned up to the upper layers.
267  */
268 uint32_t nxge_mblks_pending = 0;
269 
270 /*
271  * Device register access attributes for PIO.
272  */
273 static ddi_device_acc_attr_t nxge_dev_reg_acc_attr = {
274 	DDI_DEVICE_ATTR_V0,
275 	DDI_STRUCTURE_LE_ACC,
276 	DDI_STRICTORDER_ACC,
277 };
278 
279 /*
280  * Device descriptor access attributes for DMA.
281  */
282 static ddi_device_acc_attr_t nxge_dev_desc_dma_acc_attr = {
283 	DDI_DEVICE_ATTR_V0,
284 	DDI_STRUCTURE_LE_ACC,
285 	DDI_STRICTORDER_ACC
286 };
287 
288 /*
289  * Device buffer access attributes for DMA.
290  */
291 static ddi_device_acc_attr_t nxge_dev_buf_dma_acc_attr = {
292 	DDI_DEVICE_ATTR_V0,
293 	DDI_STRUCTURE_BE_ACC,
294 	DDI_STRICTORDER_ACC
295 };
296 
297 ddi_dma_attr_t nxge_desc_dma_attr = {
298 	DMA_ATTR_V0,		/* version number. */
299 	0,			/* low address */
300 	0xffffffffffffffff,	/* high address */
301 	0xffffffffffffffff,	/* address counter max */
302 #ifndef NIU_PA_WORKAROUND
303 	0x100000,		/* alignment */
304 #else
305 	0x2000,
306 #endif
307 	0xfc00fc,		/* dlim_burstsizes */
308 	0x1,			/* minimum transfer size */
309 	0xffffffffffffffff,	/* maximum transfer size */
310 	0xffffffffffffffff,	/* maximum segment size */
311 	1,			/* scatter/gather list length */
312 	(unsigned int) 1,	/* granularity */
313 	0			/* attribute flags */
314 };
315 
316 ddi_dma_attr_t nxge_tx_dma_attr = {
317 	DMA_ATTR_V0,		/* version number. */
318 	0,			/* low address */
319 	0xffffffffffffffff,	/* high address */
320 	0xffffffffffffffff,	/* address counter max */
321 #if defined(_BIG_ENDIAN)
322 	0x2000,			/* alignment */
323 #else
324 	0x1000,			/* alignment */
325 #endif
326 	0xfc00fc,		/* dlim_burstsizes */
327 	0x1,			/* minimum transfer size */
328 	0xffffffffffffffff,	/* maximum transfer size */
329 	0xffffffffffffffff,	/* maximum segment size */
330 	5,			/* scatter/gather list length */
331 	(unsigned int) 1,	/* granularity */
332 	0			/* attribute flags */
333 };
334 
335 ddi_dma_attr_t nxge_rx_dma_attr = {
336 	DMA_ATTR_V0,		/* version number. */
337 	0,			/* low address */
338 	0xffffffffffffffff,	/* high address */
339 	0xffffffffffffffff,	/* address counter max */
340 	0x2000,			/* alignment */
341 	0xfc00fc,		/* dlim_burstsizes */
342 	0x1,			/* minimum transfer size */
343 	0xffffffffffffffff,	/* maximum transfer size */
344 	0xffffffffffffffff,	/* maximum segment size */
345 	1,			/* scatter/gather list length */
346 	(unsigned int) 1,	/* granularity */
347 	DDI_DMA_RELAXED_ORDERING /* attribute flags */
348 };
349 
350 ddi_dma_lim_t nxge_dma_limits = {
351 	(uint_t)0,		/* dlim_addr_lo */
352 	(uint_t)0xffffffff,	/* dlim_addr_hi */
353 	(uint_t)0xffffffff,	/* dlim_cntr_max */
354 	(uint_t)0xfc00fc,	/* dlim_burstsizes for 32 and 64 bit xfers */
355 	0x1,			/* dlim_minxfer */
356 	1024			/* dlim_speed */
357 };
358 
359 dma_method_t nxge_force_dma = DVMA;
360 
361 /*
362  * dma chunk sizes.
363  *
364  * Try to allocate the largest possible size
365  * so that fewer number of dma chunks would be managed
366  */
367 #ifdef NIU_PA_WORKAROUND
368 size_t alloc_sizes [] = {0x2000};
369 #else
370 size_t alloc_sizes [] = {0x1000, 0x2000, 0x4000, 0x8000,
371 		0x10000, 0x20000, 0x40000, 0x80000,
372 		0x100000, 0x200000, 0x400000, 0x800000,
373 		0x1000000, 0x2000000, 0x4000000};
374 #endif
375 
376 /*
377  * Translate "dev_t" to a pointer to the associated "dev_info_t".
378  */
379 
380 static int
381 nxge_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
382 {
383 	p_nxge_t	nxgep = NULL;
384 	int		instance;
385 	int		status = DDI_SUCCESS;
386 	uint8_t		portn;
387 	nxge_mmac_t	*mmac_info;
388 
389 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_attach"));
390 
391 	/*
392 	 * Get the device instance since we'll need to setup
393 	 * or retrieve a soft state for this instance.
394 	 */
395 	instance = ddi_get_instance(dip);
396 
397 	switch (cmd) {
398 	case DDI_ATTACH:
399 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_ATTACH"));
400 		break;
401 
402 	case DDI_RESUME:
403 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_RESUME"));
404 		nxgep = (p_nxge_t)ddi_get_soft_state(nxge_list, instance);
405 		if (nxgep == NULL) {
406 			status = DDI_FAILURE;
407 			break;
408 		}
409 		if (nxgep->dip != dip) {
410 			status = DDI_FAILURE;
411 			break;
412 		}
413 		if (nxgep->suspended == DDI_PM_SUSPEND) {
414 			status = ddi_dev_is_needed(nxgep->dip, 0, 1);
415 		} else {
416 			status = nxge_resume(nxgep);
417 		}
418 		goto nxge_attach_exit;
419 
420 	case DDI_PM_RESUME:
421 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_PM_RESUME"));
422 		nxgep = (p_nxge_t)ddi_get_soft_state(nxge_list, instance);
423 		if (nxgep == NULL) {
424 			status = DDI_FAILURE;
425 			break;
426 		}
427 		if (nxgep->dip != dip) {
428 			status = DDI_FAILURE;
429 			break;
430 		}
431 		status = nxge_resume(nxgep);
432 		goto nxge_attach_exit;
433 
434 	default:
435 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing unknown"));
436 		status = DDI_FAILURE;
437 		goto nxge_attach_exit;
438 	}
439 
440 
441 	if (ddi_soft_state_zalloc(nxge_list, instance) == DDI_FAILURE) {
442 		status = DDI_FAILURE;
443 		goto nxge_attach_exit;
444 	}
445 
446 	nxgep = ddi_get_soft_state(nxge_list, instance);
447 	if (nxgep == NULL) {
448 		status = NXGE_ERROR;
449 		goto nxge_attach_fail2;
450 	}
451 
452 	nxgep->nxge_magic = NXGE_MAGIC;
453 
454 	nxgep->drv_state = 0;
455 	nxgep->dip = dip;
456 	nxgep->instance = instance;
457 	nxgep->p_dip = ddi_get_parent(dip);
458 	nxgep->nxge_debug_level = nxge_debug_level;
459 	npi_debug_level = nxge_debug_level;
460 
461 	nxge_fm_init(nxgep, &nxge_dev_reg_acc_attr, &nxge_dev_desc_dma_acc_attr,
462 				&nxge_rx_dma_attr);
463 
464 	status = nxge_map_regs(nxgep);
465 	if (status != NXGE_OK) {
466 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "nxge_map_regs failed"));
467 		goto nxge_attach_fail3;
468 	}
469 
470 	status = nxge_init_common_dev(nxgep);
471 	if (status != NXGE_OK) {
472 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
473 			"nxge_init_common_dev failed"));
474 		goto nxge_attach_fail4;
475 	}
476 
477 	if (nxgep->niu_type == NEPTUNE_2_10GF) {
478 		if (nxgep->function_num > 1) {
479 			NXGE_DEBUG_MSG((nxgep, DDI_CTL, "Unsupported"
480 			    " function %d. Only functions 0 and 1 are "
481 			    "supported for this card.", nxgep->function_num));
482 			status = NXGE_ERROR;
483 			goto nxge_attach_fail4;
484 		}
485 	}
486 
487 	portn = NXGE_GET_PORT_NUM(nxgep->function_num);
488 	nxgep->mac.portnum = portn;
489 	if ((portn == 0) || (portn == 1))
490 		nxgep->mac.porttype = PORT_TYPE_XMAC;
491 	else
492 		nxgep->mac.porttype = PORT_TYPE_BMAC;
493 	/*
494 	 * Neptune has 4 ports, the first 2 ports use XMAC (10G MAC)
495 	 * internally, the rest 2 ports use BMAC (1G "Big" MAC).
496 	 * The two types of MACs have different characterizations.
497 	 */
498 	mmac_info = &nxgep->nxge_mmac_info;
499 	if (nxgep->function_num < 2) {
500 		mmac_info->num_mmac = XMAC_MAX_ALT_ADDR_ENTRY;
501 		mmac_info->naddrfree = XMAC_MAX_ALT_ADDR_ENTRY;
502 	} else {
503 		mmac_info->num_mmac = BMAC_MAX_ALT_ADDR_ENTRY;
504 		mmac_info->naddrfree = BMAC_MAX_ALT_ADDR_ENTRY;
505 	}
506 	/*
507 	 * Setup the Ndd parameters for the this instance.
508 	 */
509 	nxge_init_param(nxgep);
510 
511 	/*
512 	 * Setup Register Tracing Buffer.
513 	 */
514 	npi_rtrace_buf_init((rtrace_t *)&npi_rtracebuf);
515 
516 	/* init stats ptr */
517 	nxge_init_statsp(nxgep);
518 
519 	/*
520 	 * read the vpd info from the eeprom into local data
521 	 * structure and check for the VPD info validity
522 	 */
523 	nxge_vpd_info_get(nxgep);
524 
525 	status = nxge_xcvr_find(nxgep);
526 
527 	if (status != NXGE_OK) {
528 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "nxge_attach: "
529 				    " Couldn't determine card type"
530 				    " .... exit "));
531 		goto nxge_attach_fail5;
532 	}
533 
534 	status = nxge_get_config_properties(nxgep);
535 
536 	if (status != NXGE_OK) {
537 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "get_hw create failed"));
538 		goto nxge_attach_fail;
539 	}
540 
541 	/*
542 	 * Setup the Kstats for the driver.
543 	 */
544 	nxge_setup_kstats(nxgep);
545 
546 	nxge_setup_param(nxgep);
547 
548 	status = nxge_setup_system_dma_pages(nxgep);
549 	if (status != NXGE_OK) {
550 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "set dma page failed"));
551 		goto nxge_attach_fail;
552 	}
553 
554 #if	defined(sun4v)
555 	if (nxgep->niu_type == N2_NIU) {
556 		nxgep->niu_hsvc_available = B_FALSE;
557 		bcopy(&niu_hsvc, &nxgep->niu_hsvc, sizeof (hsvc_info_t));
558 		if ((status =
559 			hsvc_register(&nxgep->niu_hsvc,
560 					&nxgep->niu_min_ver)) != 0) {
561 				NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
562 					"nxge_attach: "
563 					"%s: cannot negotiate "
564 					"hypervisor services "
565 					"revision %d "
566 					"group: 0x%lx "
567 					"major: 0x%lx minor: 0x%lx "
568 					"errno: %d",
569 					niu_hsvc.hsvc_modname,
570 					niu_hsvc.hsvc_rev,
571 					niu_hsvc.hsvc_group,
572 					niu_hsvc.hsvc_major,
573 					niu_hsvc.hsvc_minor,
574 					status));
575 				status = DDI_FAILURE;
576 				goto nxge_attach_fail;
577 		}
578 
579 		nxgep->niu_hsvc_available = B_TRUE;
580 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
581 			"NIU Hypervisor service enabled"));
582 	}
583 #endif
584 
585 	nxge_hw_id_init(nxgep);
586 	nxge_hw_init_niu_common(nxgep);
587 
588 	status = nxge_setup_mutexes(nxgep);
589 	if (status != NXGE_OK) {
590 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "set mutex failed"));
591 		goto nxge_attach_fail;
592 	}
593 
594 	status = nxge_setup_dev(nxgep);
595 	if (status != DDI_SUCCESS) {
596 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "set dev failed"));
597 		goto nxge_attach_fail;
598 	}
599 
600 	status = nxge_add_intrs(nxgep);
601 	if (status != DDI_SUCCESS) {
602 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "add_intr failed"));
603 		goto nxge_attach_fail;
604 	}
605 	status = nxge_add_soft_intrs(nxgep);
606 	if (status != DDI_SUCCESS) {
607 		NXGE_DEBUG_MSG((nxgep, NXGE_ERR_CTL, "add_soft_intr failed"));
608 		goto nxge_attach_fail;
609 	}
610 
611 	/*
612 	 * Enable interrupts.
613 	 */
614 	nxge_intrs_enable(nxgep);
615 
616 	if ((status = nxge_mac_register(nxgep)) != NXGE_OK) {
617 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
618 			"unable to register to mac layer (%d)", status));
619 		goto nxge_attach_fail;
620 	}
621 
622 	mac_link_update(nxgep->mach, LINK_STATE_UNKNOWN);
623 
624 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "registered to mac (instance %d)",
625 		instance));
626 
627 	(void) nxge_link_monitor(nxgep, LINK_MONITOR_START);
628 
629 	goto nxge_attach_exit;
630 
631 nxge_attach_fail:
632 	nxge_unattach(nxgep);
633 	goto nxge_attach_fail1;
634 
635 nxge_attach_fail5:
636 	/*
637 	 * Tear down the ndd parameters setup.
638 	 */
639 	nxge_destroy_param(nxgep);
640 
641 	/*
642 	 * Tear down the kstat setup.
643 	 */
644 	nxge_destroy_kstats(nxgep);
645 
646 nxge_attach_fail4:
647 	if (nxgep->nxge_hw_p) {
648 		nxge_uninit_common_dev(nxgep);
649 		nxgep->nxge_hw_p = NULL;
650 	}
651 
652 nxge_attach_fail3:
653 	/*
654 	 * Unmap the register setup.
655 	 */
656 	nxge_unmap_regs(nxgep);
657 
658 	nxge_fm_fini(nxgep);
659 
660 nxge_attach_fail2:
661 	ddi_soft_state_free(nxge_list, nxgep->instance);
662 
663 nxge_attach_fail1:
664 	if (status != NXGE_OK)
665 		status = (NXGE_ERROR | NXGE_DDI_FAILED);
666 	nxgep = NULL;
667 
668 nxge_attach_exit:
669 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_attach status = 0x%08x",
670 		status));
671 
672 	return (status);
673 }
674 
675 static int
676 nxge_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
677 {
678 	int 		status = DDI_SUCCESS;
679 	int 		instance;
680 	p_nxge_t 	nxgep = NULL;
681 
682 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_detach"));
683 	instance = ddi_get_instance(dip);
684 	nxgep = ddi_get_soft_state(nxge_list, instance);
685 	if (nxgep == NULL) {
686 		status = DDI_FAILURE;
687 		goto nxge_detach_exit;
688 	}
689 
690 	switch (cmd) {
691 	case DDI_DETACH:
692 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_DETACH"));
693 		break;
694 
695 	case DDI_PM_SUSPEND:
696 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_PM_SUSPEND"));
697 		nxgep->suspended = DDI_PM_SUSPEND;
698 		nxge_suspend(nxgep);
699 		break;
700 
701 	case DDI_SUSPEND:
702 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "doing DDI_SUSPEND"));
703 		if (nxgep->suspended != DDI_PM_SUSPEND) {
704 			nxgep->suspended = DDI_SUSPEND;
705 			nxge_suspend(nxgep);
706 		}
707 		break;
708 
709 	default:
710 		status = DDI_FAILURE;
711 	}
712 
713 	if (cmd != DDI_DETACH)
714 		goto nxge_detach_exit;
715 
716 	/*
717 	 * Stop the xcvr polling.
718 	 */
719 	nxgep->suspended = cmd;
720 
721 	(void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP);
722 
723 	if (nxgep->mach && (status = mac_unregister(nxgep->mach)) != 0) {
724 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
725 			"<== nxge_detach status = 0x%08X", status));
726 		return (DDI_FAILURE);
727 	}
728 
729 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
730 		"<== nxge_detach (mac_unregister) status = 0x%08X", status));
731 
732 	nxge_unattach(nxgep);
733 	nxgep = NULL;
734 
735 nxge_detach_exit:
736 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_detach status = 0x%08X",
737 		status));
738 
739 	return (status);
740 }
741 
742 static void
743 nxge_unattach(p_nxge_t nxgep)
744 {
745 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_unattach"));
746 
747 	if (nxgep == NULL || nxgep->dev_regs == NULL) {
748 		return;
749 	}
750 
751 	nxgep->nxge_magic = 0;
752 
753 	if (nxgep->nxge_timerid) {
754 		nxge_stop_timer(nxgep, nxgep->nxge_timerid);
755 		nxgep->nxge_timerid = 0;
756 	}
757 
758 	if (nxgep->nxge_hw_p) {
759 		nxge_uninit_common_dev(nxgep);
760 		nxgep->nxge_hw_p = NULL;
761 	}
762 
763 #if	defined(sun4v)
764 	if (nxgep->niu_type == N2_NIU && nxgep->niu_hsvc_available == B_TRUE) {
765 		(void) hsvc_unregister(&nxgep->niu_hsvc);
766 		nxgep->niu_hsvc_available = B_FALSE;
767 	}
768 #endif
769 	/*
770 	 * Stop any further interrupts.
771 	 */
772 	nxge_remove_intrs(nxgep);
773 
774 	/* remove soft interrups */
775 	nxge_remove_soft_intrs(nxgep);
776 
777 	/*
778 	 * Stop the device and free resources.
779 	 */
780 	nxge_destroy_dev(nxgep);
781 
782 	/*
783 	 * Tear down the ndd parameters setup.
784 	 */
785 	nxge_destroy_param(nxgep);
786 
787 	/*
788 	 * Tear down the kstat setup.
789 	 */
790 	nxge_destroy_kstats(nxgep);
791 
792 	/*
793 	 * Destroy all mutexes.
794 	 */
795 	nxge_destroy_mutexes(nxgep);
796 
797 	/*
798 	 * Remove the list of ndd parameters which
799 	 * were setup during attach.
800 	 */
801 	if (nxgep->dip) {
802 		NXGE_DEBUG_MSG((nxgep, OBP_CTL,
803 				    " nxge_unattach: remove all properties"));
804 
805 		(void) ddi_prop_remove_all(nxgep->dip);
806 	}
807 
808 #if NXGE_PROPERTY
809 	nxge_remove_hard_properties(nxgep);
810 #endif
811 
812 	/*
813 	 * Unmap the register setup.
814 	 */
815 	nxge_unmap_regs(nxgep);
816 
817 	nxge_fm_fini(nxgep);
818 
819 	ddi_soft_state_free(nxge_list, nxgep->instance);
820 
821 	NXGE_DEBUG_MSG((NULL, DDI_CTL, "<== nxge_unattach"));
822 }
823 
824 static char n2_siu_name[] = "niu";
825 
826 static nxge_status_t
827 nxge_map_regs(p_nxge_t nxgep)
828 {
829 	int		ddi_status = DDI_SUCCESS;
830 	p_dev_regs_t 	dev_regs;
831 	char		buf[MAXPATHLEN + 1];
832 	char 		*devname;
833 #ifdef	NXGE_DEBUG
834 	char 		*sysname;
835 #endif
836 	off_t		regsize;
837 	nxge_status_t	status = NXGE_OK;
838 #if !defined(_BIG_ENDIAN)
839 	off_t pci_offset;
840 	uint16_t pcie_devctl;
841 #endif
842 
843 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_map_regs"));
844 	nxgep->dev_regs = NULL;
845 	dev_regs = KMEM_ZALLOC(sizeof (dev_regs_t), KM_SLEEP);
846 	dev_regs->nxge_regh = NULL;
847 	dev_regs->nxge_pciregh = NULL;
848 	dev_regs->nxge_msix_regh = NULL;
849 	dev_regs->nxge_vir_regh = NULL;
850 	dev_regs->nxge_vir2_regh = NULL;
851 	nxgep->niu_type = NIU_TYPE_NONE;
852 
853 	devname = ddi_pathname(nxgep->dip, buf);
854 	ASSERT(strlen(devname) > 0);
855 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
856 		"nxge_map_regs: pathname devname %s", devname));
857 
858 	if (strstr(devname, n2_siu_name)) {
859 		/* N2/NIU */
860 		nxgep->niu_type = N2_NIU;
861 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
862 			"nxge_map_regs: N2/NIU devname %s", devname));
863 		/* get function number */
864 		nxgep->function_num =
865 			(devname[strlen(devname) -1] == '1' ? 1 : 0);
866 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
867 			"nxge_map_regs: N2/NIU function number %d",
868 			nxgep->function_num));
869 	} else {
870 		int		*prop_val;
871 		uint_t 		prop_len;
872 		uint8_t 	func_num;
873 
874 		if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, nxgep->dip,
875 				0, "reg",
876 				&prop_val, &prop_len) != DDI_PROP_SUCCESS) {
877 			NXGE_DEBUG_MSG((nxgep, VPD_CTL,
878 				"Reg property not found"));
879 			ddi_status = DDI_FAILURE;
880 			goto nxge_map_regs_fail0;
881 
882 		} else {
883 			func_num = (prop_val[0] >> 8) & 0x7;
884 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
885 				"Reg property found: fun # %d",
886 				func_num));
887 			nxgep->function_num = func_num;
888 			ddi_prop_free(prop_val);
889 		}
890 	}
891 
892 	switch (nxgep->niu_type) {
893 	default:
894 		(void) ddi_dev_regsize(nxgep->dip, 0, &regsize);
895 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
896 			"nxge_map_regs: pci config size 0x%x", regsize));
897 
898 		ddi_status = ddi_regs_map_setup(nxgep->dip, 0,
899 			(caddr_t *)&(dev_regs->nxge_pciregp), 0, 0,
900 			&nxge_dev_reg_acc_attr, &dev_regs->nxge_pciregh);
901 		if (ddi_status != DDI_SUCCESS) {
902 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
903 				"ddi_map_regs, nxge bus config regs failed"));
904 			goto nxge_map_regs_fail0;
905 		}
906 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
907 			"nxge_map_reg: PCI config addr 0x%0llx "
908 			" handle 0x%0llx", dev_regs->nxge_pciregp,
909 			dev_regs->nxge_pciregh));
910 			/*
911 			 * IMP IMP
912 			 * workaround  for bit swapping bug in HW
913 			 * which ends up in no-snoop = yes
914 			 * resulting, in DMA not synched properly
915 			 */
916 #if !defined(_BIG_ENDIAN)
917 		/* workarounds for x86 systems */
918 		pci_offset = 0x80 + PCIE_DEVCTL;
919 		pcie_devctl = 0x0;
920 		pcie_devctl &= PCIE_DEVCTL_ENABLE_NO_SNOOP;
921 		pcie_devctl |= PCIE_DEVCTL_RO_EN;
922 		pci_config_put16(dev_regs->nxge_pciregh, pci_offset,
923 				    pcie_devctl);
924 #endif
925 
926 		(void) ddi_dev_regsize(nxgep->dip, 1, &regsize);
927 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
928 			"nxge_map_regs: pio size 0x%x", regsize));
929 		/* set up the device mapped register */
930 		ddi_status = ddi_regs_map_setup(nxgep->dip, 1,
931 			(caddr_t *)&(dev_regs->nxge_regp), 0, 0,
932 			&nxge_dev_reg_acc_attr, &dev_regs->nxge_regh);
933 		if (ddi_status != DDI_SUCCESS) {
934 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
935 				"ddi_map_regs for Neptune global reg failed"));
936 			goto nxge_map_regs_fail1;
937 		}
938 
939 		/* set up the msi/msi-x mapped register */
940 		(void) ddi_dev_regsize(nxgep->dip, 2, &regsize);
941 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
942 			"nxge_map_regs: msix size 0x%x", regsize));
943 		ddi_status = ddi_regs_map_setup(nxgep->dip, 2,
944 			(caddr_t *)&(dev_regs->nxge_msix_regp), 0, 0,
945 			&nxge_dev_reg_acc_attr, &dev_regs->nxge_msix_regh);
946 		if (ddi_status != DDI_SUCCESS) {
947 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
948 				"ddi_map_regs for msi reg failed"));
949 			goto nxge_map_regs_fail2;
950 		}
951 
952 		/* set up the vio region mapped register */
953 		(void) ddi_dev_regsize(nxgep->dip, 3, &regsize);
954 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
955 			"nxge_map_regs: vio size 0x%x", regsize));
956 		ddi_status = ddi_regs_map_setup(nxgep->dip, 3,
957 			(caddr_t *)&(dev_regs->nxge_vir_regp), 0, 0,
958 			&nxge_dev_reg_acc_attr, &dev_regs->nxge_vir_regh);
959 
960 		if (ddi_status != DDI_SUCCESS) {
961 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
962 				"ddi_map_regs for nxge vio reg failed"));
963 			goto nxge_map_regs_fail3;
964 		}
965 		nxgep->dev_regs = dev_regs;
966 
967 		NPI_PCI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_pciregh);
968 		NPI_PCI_ADD_HANDLE_SET(nxgep,
969 			(npi_reg_ptr_t)dev_regs->nxge_pciregp);
970 		NPI_MSI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_msix_regh);
971 		NPI_MSI_ADD_HANDLE_SET(nxgep,
972 			(npi_reg_ptr_t)dev_regs->nxge_msix_regp);
973 
974 		NPI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh);
975 		NPI_ADD_HANDLE_SET(nxgep, (npi_reg_ptr_t)dev_regs->nxge_regp);
976 
977 		NPI_REG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh);
978 		NPI_REG_ADD_HANDLE_SET(nxgep,
979 			(npi_reg_ptr_t)dev_regs->nxge_regp);
980 
981 		NPI_VREG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_vir_regh);
982 		NPI_VREG_ADD_HANDLE_SET(nxgep,
983 			(npi_reg_ptr_t)dev_regs->nxge_vir_regp);
984 
985 		break;
986 
987 	case N2_NIU:
988 		NXGE_DEBUG_MSG((nxgep, DDI_CTL, "ddi_map_regs, NIU"));
989 		/*
990 		 * Set up the device mapped register (FWARC 2006/556)
991 		 * (changed back to 1: reg starts at 1!)
992 		 */
993 		(void) ddi_dev_regsize(nxgep->dip, 1, &regsize);
994 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
995 			"nxge_map_regs: dev size 0x%x", regsize));
996 		ddi_status = ddi_regs_map_setup(nxgep->dip, 1,
997 				(caddr_t *)&(dev_regs->nxge_regp), 0, 0,
998 				&nxge_dev_reg_acc_attr, &dev_regs->nxge_regh);
999 
1000 		if (ddi_status != DDI_SUCCESS) {
1001 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1002 				"ddi_map_regs for N2/NIU, global reg failed "));
1003 			goto nxge_map_regs_fail1;
1004 		}
1005 
1006 		/* set up the vio region mapped register */
1007 		(void) ddi_dev_regsize(nxgep->dip, 2, &regsize);
1008 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1009 			"nxge_map_regs: vio (1) size 0x%x", regsize));
1010 		ddi_status = ddi_regs_map_setup(nxgep->dip, 2,
1011 			(caddr_t *)&(dev_regs->nxge_vir_regp), 0, 0,
1012 			&nxge_dev_reg_acc_attr, &dev_regs->nxge_vir_regh);
1013 
1014 		if (ddi_status != DDI_SUCCESS) {
1015 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1016 				"ddi_map_regs for nxge vio reg failed"));
1017 			goto nxge_map_regs_fail2;
1018 		}
1019 		/* set up the vio region mapped register */
1020 		(void) ddi_dev_regsize(nxgep->dip, 3, &regsize);
1021 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1022 			"nxge_map_regs: vio (3) size 0x%x", regsize));
1023 		ddi_status = ddi_regs_map_setup(nxgep->dip, 3,
1024 			(caddr_t *)&(dev_regs->nxge_vir2_regp), 0, 0,
1025 			&nxge_dev_reg_acc_attr, &dev_regs->nxge_vir2_regh);
1026 
1027 		if (ddi_status != DDI_SUCCESS) {
1028 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1029 				"ddi_map_regs for nxge vio2 reg failed"));
1030 			goto nxge_map_regs_fail3;
1031 		}
1032 		nxgep->dev_regs = dev_regs;
1033 
1034 		NPI_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh);
1035 		NPI_ADD_HANDLE_SET(nxgep, (npi_reg_ptr_t)dev_regs->nxge_regp);
1036 
1037 		NPI_REG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_regh);
1038 		NPI_REG_ADD_HANDLE_SET(nxgep,
1039 			(npi_reg_ptr_t)dev_regs->nxge_regp);
1040 
1041 		NPI_VREG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_vir_regh);
1042 		NPI_VREG_ADD_HANDLE_SET(nxgep,
1043 			(npi_reg_ptr_t)dev_regs->nxge_vir_regp);
1044 
1045 		NPI_V2REG_ACC_HANDLE_SET(nxgep, dev_regs->nxge_vir2_regh);
1046 		NPI_V2REG_ADD_HANDLE_SET(nxgep,
1047 			(npi_reg_ptr_t)dev_regs->nxge_vir2_regp);
1048 
1049 		break;
1050 	}
1051 
1052 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "nxge_map_reg: hardware addr 0x%0llx "
1053 		" handle 0x%0llx", dev_regs->nxge_regp, dev_regs->nxge_regh));
1054 
1055 	goto nxge_map_regs_exit;
1056 nxge_map_regs_fail3:
1057 	if (dev_regs->nxge_msix_regh) {
1058 		ddi_regs_map_free(&dev_regs->nxge_msix_regh);
1059 	}
1060 	if (dev_regs->nxge_vir_regh) {
1061 		ddi_regs_map_free(&dev_regs->nxge_regh);
1062 	}
1063 nxge_map_regs_fail2:
1064 	if (dev_regs->nxge_regh) {
1065 		ddi_regs_map_free(&dev_regs->nxge_regh);
1066 	}
1067 nxge_map_regs_fail1:
1068 	if (dev_regs->nxge_pciregh) {
1069 		ddi_regs_map_free(&dev_regs->nxge_pciregh);
1070 	}
1071 nxge_map_regs_fail0:
1072 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "Freeing register set memory"));
1073 	kmem_free(dev_regs, sizeof (dev_regs_t));
1074 
1075 nxge_map_regs_exit:
1076 	if (ddi_status != DDI_SUCCESS)
1077 		status |= (NXGE_ERROR | NXGE_DDI_FAILED);
1078 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_map_regs"));
1079 	return (status);
1080 }
1081 
1082 static void
1083 nxge_unmap_regs(p_nxge_t nxgep)
1084 {
1085 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_unmap_regs"));
1086 	if (nxgep->dev_regs) {
1087 		if (nxgep->dev_regs->nxge_pciregh) {
1088 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1089 				"==> nxge_unmap_regs: bus"));
1090 			ddi_regs_map_free(&nxgep->dev_regs->nxge_pciregh);
1091 			nxgep->dev_regs->nxge_pciregh = NULL;
1092 		}
1093 		if (nxgep->dev_regs->nxge_regh) {
1094 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1095 				"==> nxge_unmap_regs: device registers"));
1096 			ddi_regs_map_free(&nxgep->dev_regs->nxge_regh);
1097 			nxgep->dev_regs->nxge_regh = NULL;
1098 		}
1099 		if (nxgep->dev_regs->nxge_msix_regh) {
1100 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1101 				"==> nxge_unmap_regs: device interrupts"));
1102 			ddi_regs_map_free(&nxgep->dev_regs->nxge_msix_regh);
1103 			nxgep->dev_regs->nxge_msix_regh = NULL;
1104 		}
1105 		if (nxgep->dev_regs->nxge_vir_regh) {
1106 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1107 				"==> nxge_unmap_regs: vio region"));
1108 			ddi_regs_map_free(&nxgep->dev_regs->nxge_vir_regh);
1109 			nxgep->dev_regs->nxge_vir_regh = NULL;
1110 		}
1111 		if (nxgep->dev_regs->nxge_vir2_regh) {
1112 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1113 				"==> nxge_unmap_regs: vio2 region"));
1114 			ddi_regs_map_free(&nxgep->dev_regs->nxge_vir2_regh);
1115 			nxgep->dev_regs->nxge_vir2_regh = NULL;
1116 		}
1117 
1118 		kmem_free(nxgep->dev_regs, sizeof (dev_regs_t));
1119 		nxgep->dev_regs = NULL;
1120 	}
1121 
1122 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_unmap_regs"));
1123 }
1124 
1125 static nxge_status_t
1126 nxge_setup_mutexes(p_nxge_t nxgep)
1127 {
1128 	int ddi_status = DDI_SUCCESS;
1129 	nxge_status_t status = NXGE_OK;
1130 	nxge_classify_t *classify_ptr;
1131 	int partition;
1132 
1133 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_setup_mutexes"));
1134 
1135 	/*
1136 	 * Get the interrupt cookie so the mutexes can be
1137 	 * Initialized.
1138 	 */
1139 	ddi_status = ddi_get_iblock_cookie(nxgep->dip, 0,
1140 					&nxgep->interrupt_cookie);
1141 	if (ddi_status != DDI_SUCCESS) {
1142 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1143 			"<== nxge_setup_mutexes: failed 0x%x", ddi_status));
1144 		goto nxge_setup_mutexes_exit;
1145 	}
1146 
1147 	cv_init(&nxgep->poll_cv, NULL, CV_DRIVER, NULL);
1148 	MUTEX_INIT(&nxgep->poll_lock, NULL,
1149 	    MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1150 
1151 	/*
1152 	 * Initialize mutexes for this device.
1153 	 */
1154 	MUTEX_INIT(nxgep->genlock, NULL,
1155 		MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1156 	MUTEX_INIT(&nxgep->ouraddr_lock, NULL,
1157 		MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1158 	MUTEX_INIT(&nxgep->mif_lock, NULL,
1159 		MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1160 	RW_INIT(&nxgep->filter_lock, NULL,
1161 		RW_DRIVER, (void *)nxgep->interrupt_cookie);
1162 
1163 	classify_ptr = &nxgep->classifier;
1164 		/*
1165 		 * FFLP Mutexes are never used in interrupt context
1166 		 * as fflp operation can take very long time to
1167 		 * complete and hence not suitable to invoke from interrupt
1168 		 * handlers.
1169 		 */
1170 	MUTEX_INIT(&classify_ptr->tcam_lock, NULL,
1171 	    NXGE_MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1172 	if (NXGE_IS_VALID_NEPTUNE_TYPE(nxgep)) {
1173 		MUTEX_INIT(&classify_ptr->fcram_lock, NULL,
1174 		    NXGE_MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1175 		for (partition = 0; partition < MAX_PARTITION; partition++) {
1176 			MUTEX_INIT(&classify_ptr->hash_lock[partition], NULL,
1177 			    NXGE_MUTEX_DRIVER, (void *)nxgep->interrupt_cookie);
1178 		}
1179 	}
1180 
1181 nxge_setup_mutexes_exit:
1182 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1183 	    "<== nxge_setup_mutexes status = %x", status));
1184 
1185 	if (ddi_status != DDI_SUCCESS)
1186 		status |= (NXGE_ERROR | NXGE_DDI_FAILED);
1187 
1188 	return (status);
1189 }
1190 
1191 static void
1192 nxge_destroy_mutexes(p_nxge_t nxgep)
1193 {
1194 	int partition;
1195 	nxge_classify_t *classify_ptr;
1196 
1197 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_destroy_mutexes"));
1198 	RW_DESTROY(&nxgep->filter_lock);
1199 	MUTEX_DESTROY(&nxgep->mif_lock);
1200 	MUTEX_DESTROY(&nxgep->ouraddr_lock);
1201 	MUTEX_DESTROY(nxgep->genlock);
1202 
1203 	classify_ptr = &nxgep->classifier;
1204 	MUTEX_DESTROY(&classify_ptr->tcam_lock);
1205 
1206 	/* Destroy all polling resources. */
1207 	MUTEX_DESTROY(&nxgep->poll_lock);
1208 	cv_destroy(&nxgep->poll_cv);
1209 
1210 	/* free data structures, based on HW type */
1211 	if (NXGE_IS_VALID_NEPTUNE_TYPE(nxgep)) {
1212 		MUTEX_DESTROY(&classify_ptr->fcram_lock);
1213 		for (partition = 0; partition < MAX_PARTITION; partition++) {
1214 			MUTEX_DESTROY(&classify_ptr->hash_lock[partition]);
1215 		}
1216 	}
1217 
1218 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_destroy_mutexes"));
1219 }
1220 
1221 nxge_status_t
1222 nxge_init(p_nxge_t nxgep)
1223 {
1224 	nxge_status_t	status = NXGE_OK;
1225 
1226 	NXGE_DEBUG_MSG((nxgep, STR_CTL, "==> nxge_init"));
1227 
1228 	if (nxgep->drv_state & STATE_HW_INITIALIZED) {
1229 		return (status);
1230 	}
1231 
1232 	/*
1233 	 * Allocate system memory for the receive/transmit buffer blocks
1234 	 * and receive/transmit descriptor rings.
1235 	 */
1236 	status = nxge_alloc_mem_pool(nxgep);
1237 	if (status != NXGE_OK) {
1238 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "alloc mem failed\n"));
1239 		goto nxge_init_fail1;
1240 	}
1241 
1242 	/*
1243 	 * Initialize and enable TXC registers
1244 	 * (Globally enable TX controller,
1245 	 *  enable a port, configure dma channel bitmap,
1246 	 *  configure the max burst size).
1247 	 */
1248 	status = nxge_txc_init(nxgep);
1249 	if (status != NXGE_OK) {
1250 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init txc failed\n"));
1251 		goto nxge_init_fail2;
1252 	}
1253 
1254 	/*
1255 	 * Initialize and enable TXDMA channels.
1256 	 */
1257 	status = nxge_init_txdma_channels(nxgep);
1258 	if (status != NXGE_OK) {
1259 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init txdma failed\n"));
1260 		goto nxge_init_fail3;
1261 	}
1262 
1263 	/*
1264 	 * Initialize and enable RXDMA channels.
1265 	 */
1266 	status = nxge_init_rxdma_channels(nxgep);
1267 	if (status != NXGE_OK) {
1268 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init rxdma failed\n"));
1269 		goto nxge_init_fail4;
1270 	}
1271 
1272 	/*
1273 	 * Initialize TCAM and FCRAM (Neptune).
1274 	 */
1275 	status = nxge_classify_init(nxgep);
1276 	if (status != NXGE_OK) {
1277 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init classify failed\n"));
1278 		goto nxge_init_fail5;
1279 	}
1280 
1281 	/*
1282 	 * Initialize ZCP
1283 	 */
1284 	status = nxge_zcp_init(nxgep);
1285 	if (status != NXGE_OK) {
1286 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init ZCP failed\n"));
1287 		goto nxge_init_fail5;
1288 	}
1289 
1290 	/*
1291 	 * Initialize IPP.
1292 	 */
1293 	status = nxge_ipp_init(nxgep);
1294 	if (status != NXGE_OK) {
1295 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init IPP failed\n"));
1296 		goto nxge_init_fail5;
1297 	}
1298 
1299 	/*
1300 	 * Initialize the MAC block.
1301 	 */
1302 	status = nxge_mac_init(nxgep);
1303 	if (status != NXGE_OK) {
1304 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "init MAC failed\n"));
1305 		goto nxge_init_fail5;
1306 	}
1307 
1308 	nxge_intrs_enable(nxgep);
1309 
1310 	/*
1311 	 * Enable hardware interrupts.
1312 	 */
1313 	nxge_intr_hw_enable(nxgep);
1314 	nxgep->drv_state |= STATE_HW_INITIALIZED;
1315 
1316 	goto nxge_init_exit;
1317 
1318 nxge_init_fail5:
1319 	nxge_uninit_rxdma_channels(nxgep);
1320 nxge_init_fail4:
1321 	nxge_uninit_txdma_channels(nxgep);
1322 nxge_init_fail3:
1323 	(void) nxge_txc_uninit(nxgep);
1324 nxge_init_fail2:
1325 	nxge_free_mem_pool(nxgep);
1326 nxge_init_fail1:
1327 	NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1328 		"<== nxge_init status (failed) = 0x%08x", status));
1329 	return (status);
1330 
1331 nxge_init_exit:
1332 
1333 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_init status = 0x%08x",
1334 		status));
1335 	return (status);
1336 }
1337 
1338 
1339 timeout_id_t
1340 nxge_start_timer(p_nxge_t nxgep, fptrv_t func, int msec)
1341 {
1342 	if ((nxgep->suspended == 0) ||
1343 			(nxgep->suspended == DDI_RESUME)) {
1344 		return (timeout(func, (caddr_t)nxgep,
1345 			drv_usectohz(1000 * msec)));
1346 	}
1347 	return (NULL);
1348 }
1349 
1350 /*ARGSUSED*/
1351 void
1352 nxge_stop_timer(p_nxge_t nxgep, timeout_id_t timerid)
1353 {
1354 	if (timerid) {
1355 		(void) untimeout(timerid);
1356 	}
1357 }
1358 
1359 void
1360 nxge_uninit(p_nxge_t nxgep)
1361 {
1362 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_uninit"));
1363 
1364 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
1365 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1366 			"==> nxge_uninit: not initialized"));
1367 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1368 			"<== nxge_uninit"));
1369 		return;
1370 	}
1371 
1372 	/* stop timer */
1373 	if (nxgep->nxge_timerid) {
1374 		nxge_stop_timer(nxgep, nxgep->nxge_timerid);
1375 		nxgep->nxge_timerid = 0;
1376 	}
1377 
1378 	(void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP);
1379 	(void) nxge_intr_hw_disable(nxgep);
1380 
1381 	/*
1382 	 * Reset the receive MAC side.
1383 	 */
1384 	(void) nxge_rx_mac_disable(nxgep);
1385 
1386 	/* Disable and soft reset the IPP */
1387 	(void) nxge_ipp_disable(nxgep);
1388 
1389 	/* Free classification resources */
1390 	(void) nxge_classify_uninit(nxgep);
1391 
1392 	/*
1393 	 * Reset the transmit/receive DMA side.
1394 	 */
1395 	(void) nxge_txdma_hw_mode(nxgep, NXGE_DMA_STOP);
1396 	(void) nxge_rxdma_hw_mode(nxgep, NXGE_DMA_STOP);
1397 
1398 	nxge_uninit_txdma_channels(nxgep);
1399 	nxge_uninit_rxdma_channels(nxgep);
1400 
1401 	/*
1402 	 * Reset the transmit MAC side.
1403 	 */
1404 	(void) nxge_tx_mac_disable(nxgep);
1405 
1406 	nxge_free_mem_pool(nxgep);
1407 
1408 	(void) nxge_link_monitor(nxgep, LINK_MONITOR_START);
1409 
1410 	nxgep->drv_state &= ~STATE_HW_INITIALIZED;
1411 
1412 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_uninit: "
1413 		"nxge_mblks_pending %d", nxge_mblks_pending));
1414 }
1415 
1416 void
1417 nxge_get64(p_nxge_t nxgep, p_mblk_t mp)
1418 {
1419 #if defined(__i386)
1420 	size_t		reg;
1421 #else
1422 	uint64_t	reg;
1423 #endif
1424 	uint64_t	regdata;
1425 	int		i, retry;
1426 
1427 	bcopy((char *)mp->b_rptr, (char *)&reg, sizeof (uint64_t));
1428 	regdata = 0;
1429 	retry = 1;
1430 
1431 	for (i = 0; i < retry; i++) {
1432 		NXGE_REG_RD64(nxgep->npi_handle, reg, &regdata);
1433 	}
1434 	bcopy((char *)&regdata, (char *)mp->b_rptr, sizeof (uint64_t));
1435 }
1436 
1437 void
1438 nxge_put64(p_nxge_t nxgep, p_mblk_t mp)
1439 {
1440 #if defined(__i386)
1441 	size_t		reg;
1442 #else
1443 	uint64_t	reg;
1444 #endif
1445 	uint64_t	buf[2];
1446 
1447 	bcopy((char *)mp->b_rptr, (char *)&buf[0], 2 * sizeof (uint64_t));
1448 #if defined(__i386)
1449 	reg = (size_t)buf[0];
1450 #else
1451 	reg = buf[0];
1452 #endif
1453 
1454 	NXGE_NPI_PIO_WRITE64(nxgep->npi_handle, reg, buf[1]);
1455 }
1456 
1457 
1458 nxge_os_mutex_t nxgedebuglock;
1459 int nxge_debug_init = 0;
1460 
1461 /*ARGSUSED*/
1462 /*VARARGS*/
1463 void
1464 nxge_debug_msg(p_nxge_t nxgep, uint64_t level, char *fmt, ...)
1465 {
1466 	char msg_buffer[1048];
1467 	char prefix_buffer[32];
1468 	int instance;
1469 	uint64_t debug_level;
1470 	int cmn_level = CE_CONT;
1471 	va_list ap;
1472 
1473 	debug_level = (nxgep == NULL) ? nxge_debug_level :
1474 		nxgep->nxge_debug_level;
1475 
1476 	if ((level & debug_level) ||
1477 		(level == NXGE_NOTE) ||
1478 		(level == NXGE_ERR_CTL)) {
1479 		/* do the msg processing */
1480 		if (nxge_debug_init == 0) {
1481 			MUTEX_INIT(&nxgedebuglock, NULL, MUTEX_DRIVER, NULL);
1482 			nxge_debug_init = 1;
1483 		}
1484 
1485 		MUTEX_ENTER(&nxgedebuglock);
1486 
1487 		if ((level & NXGE_NOTE)) {
1488 			cmn_level = CE_NOTE;
1489 		}
1490 
1491 		if (level & NXGE_ERR_CTL) {
1492 			cmn_level = CE_WARN;
1493 		}
1494 
1495 		va_start(ap, fmt);
1496 		(void) vsprintf(msg_buffer, fmt, ap);
1497 		va_end(ap);
1498 		if (nxgep == NULL) {
1499 			instance = -1;
1500 			(void) sprintf(prefix_buffer, "%s :", "nxge");
1501 		} else {
1502 			instance = nxgep->instance;
1503 			(void) sprintf(prefix_buffer,
1504 						    "%s%d :", "nxge", instance);
1505 		}
1506 
1507 		MUTEX_EXIT(&nxgedebuglock);
1508 		cmn_err(cmn_level, "!%s %s\n",
1509 				prefix_buffer, msg_buffer);
1510 
1511 	}
1512 }
1513 
1514 char *
1515 nxge_dump_packet(char *addr, int size)
1516 {
1517 	uchar_t *ap = (uchar_t *)addr;
1518 	int i;
1519 	static char etherbuf[1024];
1520 	char *cp = etherbuf;
1521 	char digits[] = "0123456789abcdef";
1522 
1523 	if (!size)
1524 		size = 60;
1525 
1526 	if (size > MAX_DUMP_SZ) {
1527 		/* Dump the leading bytes */
1528 		for (i = 0; i < MAX_DUMP_SZ/2; i++) {
1529 			if (*ap > 0x0f)
1530 				*cp++ = digits[*ap >> 4];
1531 			*cp++ = digits[*ap++ & 0xf];
1532 			*cp++ = ':';
1533 		}
1534 		for (i = 0; i < 20; i++)
1535 			*cp++ = '.';
1536 		/* Dump the last MAX_DUMP_SZ/2 bytes */
1537 		ap = (uchar_t *)(addr + (size - MAX_DUMP_SZ/2));
1538 		for (i = 0; i < MAX_DUMP_SZ/2; i++) {
1539 			if (*ap > 0x0f)
1540 				*cp++ = digits[*ap >> 4];
1541 			*cp++ = digits[*ap++ & 0xf];
1542 			*cp++ = ':';
1543 		}
1544 	} else {
1545 		for (i = 0; i < size; i++) {
1546 			if (*ap > 0x0f)
1547 				*cp++ = digits[*ap >> 4];
1548 			*cp++ = digits[*ap++ & 0xf];
1549 			*cp++ = ':';
1550 		}
1551 	}
1552 	*--cp = 0;
1553 	return (etherbuf);
1554 }
1555 
1556 #ifdef	NXGE_DEBUG
1557 static void
1558 nxge_test_map_regs(p_nxge_t nxgep)
1559 {
1560 	ddi_acc_handle_t cfg_handle;
1561 	p_pci_cfg_t	cfg_ptr;
1562 	ddi_acc_handle_t dev_handle;
1563 	char		*dev_ptr;
1564 	ddi_acc_handle_t pci_config_handle;
1565 	uint32_t	regval;
1566 	int		i;
1567 
1568 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_test_map_regs"));
1569 
1570 	dev_handle = nxgep->dev_regs->nxge_regh;
1571 	dev_ptr = (char *)nxgep->dev_regs->nxge_regp;
1572 
1573 	if (NXGE_IS_VALID_NEPTUNE_TYPE(nxgep)) {
1574 		cfg_handle = nxgep->dev_regs->nxge_pciregh;
1575 		cfg_ptr = (void *)nxgep->dev_regs->nxge_pciregp;
1576 
1577 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1578 		    "Neptune PCI regp cfg_ptr 0x%llx", (char *)cfg_ptr));
1579 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1580 		    "Neptune PCI cfg_ptr vendor id ptr 0x%llx",
1581 		    &cfg_ptr->vendorid));
1582 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1583 		    "\tvendorid 0x%x devid 0x%x",
1584 		    NXGE_PIO_READ16(cfg_handle, &cfg_ptr->vendorid, 0),
1585 		    NXGE_PIO_READ16(cfg_handle, &cfg_ptr->devid,    0)));
1586 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1587 		    "PCI BAR: base 0x%x base14 0x%x base 18 0x%x "
1588 		    "bar1c 0x%x",
1589 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base,   0),
1590 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base14, 0),
1591 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base18, 0),
1592 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base1c, 0)));
1593 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1594 		    "\nNeptune PCI BAR: base20 0x%x base24 0x%x "
1595 		    "base 28 0x%x bar2c 0x%x\n",
1596 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base20, 0),
1597 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base24, 0),
1598 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base28, 0),
1599 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base2c, 0)));
1600 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1601 		    "\nNeptune PCI BAR: base30 0x%x\n",
1602 		    NXGE_PIO_READ32(cfg_handle, &cfg_ptr->base30, 0)));
1603 
1604 		cfg_handle = nxgep->dev_regs->nxge_pciregh;
1605 		cfg_ptr = (void *)nxgep->dev_regs->nxge_pciregp;
1606 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1607 		    "first  0x%llx second 0x%llx third 0x%llx "
1608 		    "last 0x%llx ",
1609 		    NXGE_PIO_READ64(dev_handle,
1610 		    (uint64_t *)(dev_ptr + 0),  0),
1611 		    NXGE_PIO_READ64(dev_handle,
1612 		    (uint64_t *)(dev_ptr + 8),  0),
1613 		    NXGE_PIO_READ64(dev_handle,
1614 		    (uint64_t *)(dev_ptr + 16), 0),
1615 		    NXGE_PIO_READ64(cfg_handle,
1616 		    (uint64_t *)(dev_ptr + 24), 0)));
1617 	}
1618 }
1619 
1620 #endif
1621 
1622 static void
1623 nxge_suspend(p_nxge_t nxgep)
1624 {
1625 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_suspend"));
1626 
1627 	nxge_intrs_disable(nxgep);
1628 	nxge_destroy_dev(nxgep);
1629 
1630 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_suspend"));
1631 }
1632 
1633 static nxge_status_t
1634 nxge_resume(p_nxge_t nxgep)
1635 {
1636 	nxge_status_t status = NXGE_OK;
1637 
1638 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_resume"));
1639 
1640 	nxgep->suspended = DDI_RESUME;
1641 	(void) nxge_link_monitor(nxgep, LINK_MONITOR_START);
1642 	(void) nxge_rxdma_hw_mode(nxgep, NXGE_DMA_START);
1643 	(void) nxge_txdma_hw_mode(nxgep, NXGE_DMA_START);
1644 	(void) nxge_rx_mac_enable(nxgep);
1645 	(void) nxge_tx_mac_enable(nxgep);
1646 	nxge_intrs_enable(nxgep);
1647 	nxgep->suspended = 0;
1648 
1649 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1650 			"<== nxge_resume status = 0x%x", status));
1651 	return (status);
1652 }
1653 
1654 static nxge_status_t
1655 nxge_setup_dev(p_nxge_t nxgep)
1656 {
1657 	nxge_status_t	status = NXGE_OK;
1658 
1659 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_setup_dev port %d",
1660 	    nxgep->mac.portnum));
1661 
1662 	status = nxge_link_init(nxgep);
1663 
1664 	if (fm_check_acc_handle(nxgep->dev_regs->nxge_regh) != DDI_FM_OK) {
1665 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1666 			"port%d Bad register acc handle", nxgep->mac.portnum));
1667 		status = NXGE_ERROR;
1668 	}
1669 
1670 	if (status != NXGE_OK) {
1671 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1672 			    " nxge_setup_dev status "
1673 			    "(xcvr init 0x%08x)", status));
1674 		goto nxge_setup_dev_exit;
1675 	}
1676 
1677 nxge_setup_dev_exit:
1678 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1679 		"<== nxge_setup_dev port %d status = 0x%08x",
1680 		nxgep->mac.portnum, status));
1681 
1682 	return (status);
1683 }
1684 
1685 static void
1686 nxge_destroy_dev(p_nxge_t nxgep)
1687 {
1688 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_destroy_dev"));
1689 
1690 	(void) nxge_link_monitor(nxgep, LINK_MONITOR_STOP);
1691 
1692 	(void) nxge_hw_stop(nxgep);
1693 
1694 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_destroy_dev"));
1695 }
1696 
1697 static nxge_status_t
1698 nxge_setup_system_dma_pages(p_nxge_t nxgep)
1699 {
1700 	int 			ddi_status = DDI_SUCCESS;
1701 	uint_t 			count;
1702 	ddi_dma_cookie_t 	cookie;
1703 	uint_t 			iommu_pagesize;
1704 	nxge_status_t		status = NXGE_OK;
1705 
1706 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_setup_system_dma_pages"));
1707 	nxgep->sys_page_sz = ddi_ptob(nxgep->dip, (ulong_t)1);
1708 	if (nxgep->niu_type != N2_NIU) {
1709 		iommu_pagesize = dvma_pagesize(nxgep->dip);
1710 		NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1711 			" nxge_setup_system_dma_pages: page %d (ddi_ptob %d) "
1712 			" default_block_size %d iommu_pagesize %d",
1713 			nxgep->sys_page_sz,
1714 			ddi_ptob(nxgep->dip, (ulong_t)1),
1715 			nxgep->rx_default_block_size,
1716 			iommu_pagesize));
1717 
1718 		if (iommu_pagesize != 0) {
1719 			if (nxgep->sys_page_sz == iommu_pagesize) {
1720 				if (iommu_pagesize > 0x4000)
1721 					nxgep->sys_page_sz = 0x4000;
1722 			} else {
1723 				if (nxgep->sys_page_sz > iommu_pagesize)
1724 					nxgep->sys_page_sz = iommu_pagesize;
1725 			}
1726 		}
1727 	}
1728 	nxgep->sys_page_mask = ~(nxgep->sys_page_sz - 1);
1729 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1730 		"==> nxge_setup_system_dma_pages: page %d (ddi_ptob %d) "
1731 		"default_block_size %d page mask %d",
1732 		nxgep->sys_page_sz,
1733 		ddi_ptob(nxgep->dip, (ulong_t)1),
1734 		nxgep->rx_default_block_size,
1735 		nxgep->sys_page_mask));
1736 
1737 
1738 	switch (nxgep->sys_page_sz) {
1739 	default:
1740 		nxgep->sys_page_sz = 0x1000;
1741 		nxgep->sys_page_mask = ~(nxgep->sys_page_sz - 1);
1742 		nxgep->rx_default_block_size = 0x1000;
1743 		nxgep->rx_bksize_code = RBR_BKSIZE_4K;
1744 		break;
1745 	case 0x1000:
1746 		nxgep->rx_default_block_size = 0x1000;
1747 		nxgep->rx_bksize_code = RBR_BKSIZE_4K;
1748 		break;
1749 	case 0x2000:
1750 		nxgep->rx_default_block_size = 0x2000;
1751 		nxgep->rx_bksize_code = RBR_BKSIZE_8K;
1752 		break;
1753 	case 0x4000:
1754 		nxgep->rx_default_block_size = 0x4000;
1755 		nxgep->rx_bksize_code = RBR_BKSIZE_16K;
1756 		break;
1757 	case 0x8000:
1758 		nxgep->rx_default_block_size = 0x8000;
1759 		nxgep->rx_bksize_code = RBR_BKSIZE_32K;
1760 		break;
1761 	}
1762 
1763 #ifndef USE_RX_BIG_BUF
1764 	nxge_rx_dma_attr.dma_attr_align = nxgep->sys_page_sz;
1765 #else
1766 		nxgep->rx_default_block_size = 0x2000;
1767 		nxgep->rx_bksize_code = RBR_BKSIZE_8K;
1768 #endif
1769 	/*
1770 	 * Get the system DMA burst size.
1771 	 */
1772 	ddi_status = ddi_dma_alloc_handle(nxgep->dip, &nxge_tx_dma_attr,
1773 			DDI_DMA_DONTWAIT, 0,
1774 			&nxgep->dmasparehandle);
1775 	if (ddi_status != DDI_SUCCESS) {
1776 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1777 			"ddi_dma_alloc_handle: failed "
1778 			" status 0x%x", ddi_status));
1779 		goto nxge_get_soft_properties_exit;
1780 	}
1781 
1782 	ddi_status = ddi_dma_addr_bind_handle(nxgep->dmasparehandle, NULL,
1783 				(caddr_t)nxgep->dmasparehandle,
1784 				sizeof (nxgep->dmasparehandle),
1785 				DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
1786 				DDI_DMA_DONTWAIT, 0,
1787 				&cookie, &count);
1788 	if (ddi_status != DDI_DMA_MAPPED) {
1789 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1790 			"Binding spare handle to find system"
1791 			" burstsize failed."));
1792 		ddi_status = DDI_FAILURE;
1793 		goto nxge_get_soft_properties_fail1;
1794 	}
1795 
1796 	nxgep->sys_burst_sz = ddi_dma_burstsizes(nxgep->dmasparehandle);
1797 	(void) ddi_dma_unbind_handle(nxgep->dmasparehandle);
1798 
1799 nxge_get_soft_properties_fail1:
1800 	ddi_dma_free_handle(&nxgep->dmasparehandle);
1801 
1802 nxge_get_soft_properties_exit:
1803 
1804 	if (ddi_status != DDI_SUCCESS)
1805 		status |= (NXGE_ERROR | NXGE_DDI_FAILED);
1806 
1807 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
1808 		"<== nxge_setup_system_dma_pages status = 0x%08x", status));
1809 	return (status);
1810 }
1811 
1812 static nxge_status_t
1813 nxge_alloc_mem_pool(p_nxge_t nxgep)
1814 {
1815 	nxge_status_t	status = NXGE_OK;
1816 
1817 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_alloc_mem_pool"));
1818 
1819 	status = nxge_alloc_rx_mem_pool(nxgep);
1820 	if (status != NXGE_OK) {
1821 		return (NXGE_ERROR);
1822 	}
1823 
1824 	status = nxge_alloc_tx_mem_pool(nxgep);
1825 	if (status != NXGE_OK) {
1826 		nxge_free_rx_mem_pool(nxgep);
1827 		return (NXGE_ERROR);
1828 	}
1829 
1830 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_alloc_mem_pool"));
1831 	return (NXGE_OK);
1832 }
1833 
1834 static void
1835 nxge_free_mem_pool(p_nxge_t nxgep)
1836 {
1837 	NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_free_mem_pool"));
1838 
1839 	nxge_free_rx_mem_pool(nxgep);
1840 	nxge_free_tx_mem_pool(nxgep);
1841 
1842 	NXGE_DEBUG_MSG((nxgep, MEM_CTL, "<== nxge_free_mem_pool"));
1843 }
1844 
1845 static nxge_status_t
1846 nxge_alloc_rx_mem_pool(p_nxge_t nxgep)
1847 {
1848 	int			i, j;
1849 	uint32_t		ndmas, st_rdc;
1850 	p_nxge_dma_pt_cfg_t	p_all_cfgp;
1851 	p_nxge_hw_pt_cfg_t	p_cfgp;
1852 	p_nxge_dma_pool_t	dma_poolp;
1853 	p_nxge_dma_common_t	*dma_buf_p;
1854 	p_nxge_dma_pool_t	dma_cntl_poolp;
1855 	p_nxge_dma_common_t	*dma_cntl_p;
1856 	size_t			rx_buf_alloc_size;
1857 	size_t			rx_cntl_alloc_size;
1858 	uint32_t 		*num_chunks; /* per dma */
1859 	nxge_status_t		status = NXGE_OK;
1860 
1861 	uint32_t		nxge_port_rbr_size;
1862 	uint32_t		nxge_port_rbr_spare_size;
1863 	uint32_t		nxge_port_rcr_size;
1864 
1865 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rx_mem_pool"));
1866 
1867 	p_all_cfgp = (p_nxge_dma_pt_cfg_t)&nxgep->pt_config;
1868 	p_cfgp = (p_nxge_hw_pt_cfg_t)&p_all_cfgp->hw_config;
1869 	st_rdc = p_cfgp->start_rdc;
1870 	ndmas = p_cfgp->max_rdcs;
1871 
1872 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
1873 		" nxge_alloc_rx_mem_pool st_rdc %d ndmas %d", st_rdc, ndmas));
1874 
1875 	/*
1876 	 * Allocate memory for each receive DMA channel.
1877 	 */
1878 	dma_poolp = (p_nxge_dma_pool_t)KMEM_ZALLOC(sizeof (nxge_dma_pool_t),
1879 			KM_SLEEP);
1880 	dma_buf_p = (p_nxge_dma_common_t *)KMEM_ZALLOC(
1881 			sizeof (p_nxge_dma_common_t) * ndmas, KM_SLEEP);
1882 
1883 	dma_cntl_poolp = (p_nxge_dma_pool_t)
1884 				KMEM_ZALLOC(sizeof (nxge_dma_pool_t), KM_SLEEP);
1885 	dma_cntl_p = (p_nxge_dma_common_t *)KMEM_ZALLOC(
1886 			sizeof (p_nxge_dma_common_t) * ndmas, KM_SLEEP);
1887 
1888 	num_chunks = (uint32_t *)KMEM_ZALLOC(
1889 			sizeof (uint32_t) * ndmas, KM_SLEEP);
1890 
1891 	/*
1892 	 * Assume that each DMA channel will be configured with default
1893 	 * block size.
1894 	 * rbr block counts are mod of batch count (16).
1895 	 */
1896 	nxge_port_rbr_size = p_all_cfgp->rbr_size;
1897 	nxge_port_rcr_size = p_all_cfgp->rcr_size;
1898 
1899 	if (!nxge_port_rbr_size) {
1900 		nxge_port_rbr_size = NXGE_RBR_RBB_DEFAULT;
1901 	}
1902 	if (nxge_port_rbr_size % NXGE_RXDMA_POST_BATCH) {
1903 		nxge_port_rbr_size = (NXGE_RXDMA_POST_BATCH *
1904 			(nxge_port_rbr_size / NXGE_RXDMA_POST_BATCH + 1));
1905 	}
1906 
1907 	p_all_cfgp->rbr_size = nxge_port_rbr_size;
1908 	nxge_port_rbr_spare_size = nxge_rbr_spare_size;
1909 
1910 	if (nxge_port_rbr_spare_size % NXGE_RXDMA_POST_BATCH) {
1911 		nxge_port_rbr_spare_size = (NXGE_RXDMA_POST_BATCH *
1912 			(nxge_port_rbr_spare_size / NXGE_RXDMA_POST_BATCH + 1));
1913 	}
1914 	if (nxge_port_rbr_size > RBR_DEFAULT_MAX_BLKS) {
1915 		NXGE_DEBUG_MSG((nxgep, MEM_CTL,
1916 		    "nxge_alloc_rx_mem_pool: RBR size too high %d, "
1917 		    "set to default %d",
1918 		    nxge_port_rbr_size, RBR_DEFAULT_MAX_BLKS));
1919 		nxge_port_rbr_size = RBR_DEFAULT_MAX_BLKS;
1920 	}
1921 	if (nxge_port_rcr_size > RCR_DEFAULT_MAX) {
1922 		NXGE_DEBUG_MSG((nxgep, MEM_CTL,
1923 		    "nxge_alloc_rx_mem_pool: RCR too high %d, "
1924 		    "set to default %d",
1925 		    nxge_port_rcr_size, RCR_DEFAULT_MAX));
1926 		nxge_port_rcr_size = RCR_DEFAULT_MAX;
1927 	}
1928 
1929 	/*
1930 	 * N2/NIU has limitation on the descriptor sizes (contiguous
1931 	 * memory allocation on data buffers to 4M (contig_mem_alloc)
1932 	 * and little endian for control buffers (must use the ddi/dki mem alloc
1933 	 * function).
1934 	 */
1935 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
1936 	if (nxgep->niu_type == N2_NIU) {
1937 		nxge_port_rbr_spare_size = 0;
1938 		if ((nxge_port_rbr_size > NXGE_NIU_CONTIG_RBR_MAX) ||
1939 				(!ISP2(nxge_port_rbr_size))) {
1940 			nxge_port_rbr_size = NXGE_NIU_CONTIG_RBR_MAX;
1941 		}
1942 		if ((nxge_port_rcr_size > NXGE_NIU_CONTIG_RCR_MAX) ||
1943 				(!ISP2(nxge_port_rcr_size))) {
1944 			nxge_port_rcr_size = NXGE_NIU_CONTIG_RCR_MAX;
1945 		}
1946 	}
1947 #endif
1948 
1949 	rx_buf_alloc_size = (nxgep->rx_default_block_size *
1950 		(nxge_port_rbr_size + nxge_port_rbr_spare_size));
1951 
1952 	/*
1953 	 * Addresses of receive block ring, receive completion ring and the
1954 	 * mailbox must be all cache-aligned (64 bytes).
1955 	 */
1956 	rx_cntl_alloc_size = nxge_port_rbr_size + nxge_port_rbr_spare_size;
1957 	rx_cntl_alloc_size *= (sizeof (rx_desc_t));
1958 	rx_cntl_alloc_size += (sizeof (rcr_entry_t) * nxge_port_rcr_size);
1959 	rx_cntl_alloc_size += sizeof (rxdma_mailbox_t);
1960 
1961 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_alloc_rx_mem_pool: "
1962 		"nxge_port_rbr_size = %d nxge_port_rbr_spare_size = %d "
1963 		"nxge_port_rcr_size = %d "
1964 		"rx_cntl_alloc_size = %d",
1965 		nxge_port_rbr_size, nxge_port_rbr_spare_size,
1966 		nxge_port_rcr_size,
1967 		rx_cntl_alloc_size));
1968 
1969 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
1970 	if (nxgep->niu_type == N2_NIU) {
1971 		if (!ISP2(rx_buf_alloc_size)) {
1972 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1973 				"==> nxge_alloc_rx_mem_pool: "
1974 				" must be power of 2"));
1975 			status |= (NXGE_ERROR | NXGE_DDI_FAILED);
1976 			goto nxge_alloc_rx_mem_pool_exit;
1977 		}
1978 
1979 		if (rx_buf_alloc_size > (1 << 22)) {
1980 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
1981 				"==> nxge_alloc_rx_mem_pool: "
1982 				" limit size to 4M"));
1983 			status |= (NXGE_ERROR | NXGE_DDI_FAILED);
1984 			goto nxge_alloc_rx_mem_pool_exit;
1985 		}
1986 
1987 		if (rx_cntl_alloc_size < 0x2000) {
1988 			rx_cntl_alloc_size = 0x2000;
1989 		}
1990 	}
1991 #endif
1992 	nxgep->nxge_port_rbr_size = nxge_port_rbr_size;
1993 	nxgep->nxge_port_rcr_size = nxge_port_rcr_size;
1994 
1995 	/*
1996 	 * Allocate memory for receive buffers and descriptor rings.
1997 	 * Replace allocation functions with interface functions provided
1998 	 * by the partition manager when it is available.
1999 	 */
2000 	/*
2001 	 * Allocate memory for the receive buffer blocks.
2002 	 */
2003 	for (i = 0; i < ndmas; i++) {
2004 		NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2005 			" nxge_alloc_rx_mem_pool to alloc mem: "
2006 			" dma %d dma_buf_p %llx &dma_buf_p %llx",
2007 			i, dma_buf_p[i], &dma_buf_p[i]));
2008 		num_chunks[i] = 0;
2009 		status = nxge_alloc_rx_buf_dma(nxgep, st_rdc, &dma_buf_p[i],
2010 				rx_buf_alloc_size,
2011 				nxgep->rx_default_block_size, &num_chunks[i]);
2012 		if (status != NXGE_OK) {
2013 			break;
2014 		}
2015 		st_rdc++;
2016 		NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2017 			" nxge_alloc_rx_mem_pool DONE  alloc mem: "
2018 			"dma %d dma_buf_p %llx &dma_buf_p %llx", i,
2019 			dma_buf_p[i], &dma_buf_p[i]));
2020 	}
2021 	if (i < ndmas) {
2022 		goto nxge_alloc_rx_mem_fail1;
2023 	}
2024 	/*
2025 	 * Allocate memory for descriptor rings and mailbox.
2026 	 */
2027 	st_rdc = p_cfgp->start_rdc;
2028 	for (j = 0; j < ndmas; j++) {
2029 		status = nxge_alloc_rx_cntl_dma(nxgep, st_rdc, &dma_cntl_p[j],
2030 					rx_cntl_alloc_size);
2031 		if (status != NXGE_OK) {
2032 			break;
2033 		}
2034 		st_rdc++;
2035 	}
2036 	if (j < ndmas) {
2037 		goto nxge_alloc_rx_mem_fail2;
2038 	}
2039 
2040 	dma_poolp->ndmas = ndmas;
2041 	dma_poolp->num_chunks = num_chunks;
2042 	dma_poolp->buf_allocated = B_TRUE;
2043 	nxgep->rx_buf_pool_p = dma_poolp;
2044 	dma_poolp->dma_buf_pool_p = dma_buf_p;
2045 
2046 	dma_cntl_poolp->ndmas = ndmas;
2047 	dma_cntl_poolp->buf_allocated = B_TRUE;
2048 	nxgep->rx_cntl_pool_p = dma_cntl_poolp;
2049 	dma_cntl_poolp->dma_buf_pool_p = dma_cntl_p;
2050 
2051 	goto nxge_alloc_rx_mem_pool_exit;
2052 
2053 nxge_alloc_rx_mem_fail2:
2054 	/* Free control buffers */
2055 	j--;
2056 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2057 		"==> nxge_alloc_rx_mem_pool: freeing control bufs (%d)", j));
2058 	for (; j >= 0; j--) {
2059 		nxge_free_rx_cntl_dma(nxgep,
2060 			(p_nxge_dma_common_t)dma_cntl_p[j]);
2061 		NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2062 			"==> nxge_alloc_rx_mem_pool: control bufs freed (%d)",
2063 			j));
2064 	}
2065 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2066 		"==> nxge_alloc_rx_mem_pool: control bufs freed (%d)", j));
2067 
2068 nxge_alloc_rx_mem_fail1:
2069 	/* Free data buffers */
2070 	i--;
2071 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2072 		"==> nxge_alloc_rx_mem_pool: freeing data bufs (%d)", i));
2073 	for (; i >= 0; i--) {
2074 		nxge_free_rx_buf_dma(nxgep, (p_nxge_dma_common_t)dma_buf_p[i],
2075 			num_chunks[i]);
2076 	}
2077 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2078 		"==> nxge_alloc_rx_mem_pool: data bufs freed (%d)", i));
2079 
2080 	KMEM_FREE(num_chunks, sizeof (uint32_t) * ndmas);
2081 	KMEM_FREE(dma_poolp, sizeof (nxge_dma_pool_t));
2082 	KMEM_FREE(dma_buf_p, ndmas * sizeof (p_nxge_dma_common_t));
2083 	KMEM_FREE(dma_cntl_poolp, sizeof (nxge_dma_pool_t));
2084 	KMEM_FREE(dma_cntl_p, ndmas * sizeof (p_nxge_dma_common_t));
2085 
2086 nxge_alloc_rx_mem_pool_exit:
2087 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2088 		"<== nxge_alloc_rx_mem_pool:status 0x%08x", status));
2089 
2090 	return (status);
2091 }
2092 
2093 static void
2094 nxge_free_rx_mem_pool(p_nxge_t nxgep)
2095 {
2096 	uint32_t		i, ndmas;
2097 	p_nxge_dma_pool_t	dma_poolp;
2098 	p_nxge_dma_common_t	*dma_buf_p;
2099 	p_nxge_dma_pool_t	dma_cntl_poolp;
2100 	p_nxge_dma_common_t	*dma_cntl_p;
2101 	uint32_t 		*num_chunks;
2102 
2103 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_free_rx_mem_pool"));
2104 
2105 	dma_poolp = nxgep->rx_buf_pool_p;
2106 	if (dma_poolp == NULL || (!dma_poolp->buf_allocated)) {
2107 		NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2108 			"<== nxge_free_rx_mem_pool "
2109 			"(null rx buf pool or buf not allocated"));
2110 		return;
2111 	}
2112 
2113 	dma_cntl_poolp = nxgep->rx_cntl_pool_p;
2114 	if (dma_cntl_poolp == NULL || (!dma_cntl_poolp->buf_allocated)) {
2115 		NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2116 			"<== nxge_free_rx_mem_pool "
2117 			"(null rx cntl buf pool or cntl buf not allocated"));
2118 		return;
2119 	}
2120 
2121 	dma_buf_p = dma_poolp->dma_buf_pool_p;
2122 	num_chunks = dma_poolp->num_chunks;
2123 
2124 	dma_cntl_p = dma_cntl_poolp->dma_buf_pool_p;
2125 	ndmas = dma_cntl_poolp->ndmas;
2126 
2127 	for (i = 0; i < ndmas; i++) {
2128 		nxge_free_rx_buf_dma(nxgep, dma_buf_p[i], num_chunks[i]);
2129 	}
2130 
2131 	for (i = 0; i < ndmas; i++) {
2132 		nxge_free_rx_cntl_dma(nxgep, dma_cntl_p[i]);
2133 	}
2134 
2135 	for (i = 0; i < ndmas; i++) {
2136 		KMEM_FREE(dma_buf_p[i],
2137 			sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK);
2138 		KMEM_FREE(dma_cntl_p[i], sizeof (nxge_dma_common_t));
2139 	}
2140 
2141 	KMEM_FREE(num_chunks, sizeof (uint32_t) * ndmas);
2142 	KMEM_FREE(dma_cntl_p, ndmas * sizeof (p_nxge_dma_common_t));
2143 	KMEM_FREE(dma_cntl_poolp, sizeof (nxge_dma_pool_t));
2144 	KMEM_FREE(dma_buf_p, ndmas * sizeof (p_nxge_dma_common_t));
2145 	KMEM_FREE(dma_poolp, sizeof (nxge_dma_pool_t));
2146 
2147 	nxgep->rx_buf_pool_p = NULL;
2148 	nxgep->rx_cntl_pool_p = NULL;
2149 
2150 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "<== nxge_free_rx_mem_pool"));
2151 }
2152 
2153 
2154 static nxge_status_t
2155 nxge_alloc_rx_buf_dma(p_nxge_t nxgep, uint16_t dma_channel,
2156 	p_nxge_dma_common_t *dmap,
2157 	size_t alloc_size, size_t block_size, uint32_t *num_chunks)
2158 {
2159 	p_nxge_dma_common_t 	rx_dmap;
2160 	nxge_status_t		status = NXGE_OK;
2161 	size_t			total_alloc_size;
2162 	size_t			allocated = 0;
2163 	int			i, size_index, array_size;
2164 
2165 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rx_buf_dma"));
2166 
2167 	rx_dmap = (p_nxge_dma_common_t)
2168 			KMEM_ZALLOC(sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK,
2169 			KM_SLEEP);
2170 
2171 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2172 		" alloc_rx_buf_dma rdc %d asize %x bsize %x bbuf %llx ",
2173 		dma_channel, alloc_size, block_size, dmap));
2174 
2175 	total_alloc_size = alloc_size;
2176 
2177 #if defined(RX_USE_RECLAIM_POST)
2178 	total_alloc_size = alloc_size + alloc_size/4;
2179 #endif
2180 
2181 	i = 0;
2182 	size_index = 0;
2183 	array_size =  sizeof (alloc_sizes)/sizeof (size_t);
2184 	while ((alloc_sizes[size_index] < alloc_size) &&
2185 			(size_index < array_size))
2186 			size_index++;
2187 	if (size_index >= array_size) {
2188 		size_index = array_size - 1;
2189 	}
2190 
2191 	while ((allocated < total_alloc_size) &&
2192 			(size_index >= 0) && (i < NXGE_DMA_BLOCK)) {
2193 		rx_dmap[i].dma_chunk_index = i;
2194 		rx_dmap[i].block_size = block_size;
2195 		rx_dmap[i].alength = alloc_sizes[size_index];
2196 		rx_dmap[i].orig_alength = rx_dmap[i].alength;
2197 		rx_dmap[i].nblocks = alloc_sizes[size_index] / block_size;
2198 		rx_dmap[i].dma_channel = dma_channel;
2199 		rx_dmap[i].contig_alloc_type = B_FALSE;
2200 
2201 		/*
2202 		 * N2/NIU: data buffers must be contiguous as the driver
2203 		 *	   needs to call Hypervisor api to set up
2204 		 *	   logical pages.
2205 		 */
2206 		if ((nxgep->niu_type == N2_NIU) && (NXGE_DMA_BLOCK == 1)) {
2207 			rx_dmap[i].contig_alloc_type = B_TRUE;
2208 		}
2209 
2210 		NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2211 			"alloc_rx_buf_dma rdc %d chunk %d bufp %llx size %x "
2212 			"i %d nblocks %d alength %d",
2213 			dma_channel, i, &rx_dmap[i], block_size,
2214 			i, rx_dmap[i].nblocks,
2215 			rx_dmap[i].alength));
2216 		status = nxge_dma_mem_alloc(nxgep, nxge_force_dma,
2217 			&nxge_rx_dma_attr,
2218 			rx_dmap[i].alength,
2219 			&nxge_dev_buf_dma_acc_attr,
2220 			DDI_DMA_READ | DDI_DMA_STREAMING,
2221 			(p_nxge_dma_common_t)(&rx_dmap[i]));
2222 		if (status != NXGE_OK) {
2223 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2224 				" nxge_alloc_rx_buf_dma: Alloc Failed "));
2225 			size_index--;
2226 		} else {
2227 			NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2228 				" alloc_rx_buf_dma allocated rdc %d "
2229 				"chunk %d size %x dvma %x bufp %llx ",
2230 				dma_channel, i, rx_dmap[i].alength,
2231 				rx_dmap[i].ioaddr_pp, &rx_dmap[i]));
2232 			i++;
2233 			allocated += alloc_sizes[size_index];
2234 		}
2235 	}
2236 
2237 
2238 	if (allocated < total_alloc_size) {
2239 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2240 		    "==> nxge_alloc_rx_buf_dma: not enough for channe %d "
2241 		    "allocated 0x%x requested 0x%x",
2242 		    dma_channel,
2243 		    allocated, total_alloc_size));
2244 		status = NXGE_ERROR;
2245 		goto nxge_alloc_rx_mem_fail1;
2246 	}
2247 
2248 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2249 	    "==> nxge_alloc_rx_buf_dma: Allocated for channe %d "
2250 	    "allocated 0x%x requested 0x%x",
2251 	    dma_channel,
2252 	    allocated, total_alloc_size));
2253 
2254 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2255 		" alloc_rx_buf_dma rdc %d allocated %d chunks",
2256 		dma_channel, i));
2257 	*num_chunks = i;
2258 	*dmap = rx_dmap;
2259 
2260 	goto nxge_alloc_rx_mem_exit;
2261 
2262 nxge_alloc_rx_mem_fail1:
2263 	KMEM_FREE(rx_dmap, sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK);
2264 
2265 nxge_alloc_rx_mem_exit:
2266 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2267 		"<== nxge_alloc_rx_buf_dma status 0x%08x", status));
2268 
2269 	return (status);
2270 }
2271 
2272 /*ARGSUSED*/
2273 static void
2274 nxge_free_rx_buf_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap,
2275     uint32_t num_chunks)
2276 {
2277 	int		i;
2278 
2279 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2280 		"==> nxge_free_rx_buf_dma: # of chunks %d", num_chunks));
2281 
2282 	for (i = 0; i < num_chunks; i++) {
2283 		NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2284 			"==> nxge_free_rx_buf_dma: chunk %d dmap 0x%llx",
2285 				i, dmap));
2286 		nxge_dma_mem_free(dmap++);
2287 	}
2288 
2289 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL, "==> nxge_free_rx_buf_dma"));
2290 }
2291 
2292 /*ARGSUSED*/
2293 static nxge_status_t
2294 nxge_alloc_rx_cntl_dma(p_nxge_t nxgep, uint16_t dma_channel,
2295     p_nxge_dma_common_t *dmap, size_t size)
2296 {
2297 	p_nxge_dma_common_t 	rx_dmap;
2298 	nxge_status_t		status = NXGE_OK;
2299 
2300 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_rx_cntl_dma"));
2301 
2302 	rx_dmap = (p_nxge_dma_common_t)
2303 			KMEM_ZALLOC(sizeof (nxge_dma_common_t), KM_SLEEP);
2304 
2305 	rx_dmap->contig_alloc_type = B_FALSE;
2306 
2307 	status = nxge_dma_mem_alloc(nxgep, nxge_force_dma,
2308 			&nxge_desc_dma_attr,
2309 			size,
2310 			&nxge_dev_desc_dma_acc_attr,
2311 			DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
2312 			rx_dmap);
2313 	if (status != NXGE_OK) {
2314 		goto nxge_alloc_rx_cntl_dma_fail1;
2315 	}
2316 
2317 	*dmap = rx_dmap;
2318 	goto nxge_alloc_rx_cntl_dma_exit;
2319 
2320 nxge_alloc_rx_cntl_dma_fail1:
2321 	KMEM_FREE(rx_dmap, sizeof (nxge_dma_common_t));
2322 
2323 nxge_alloc_rx_cntl_dma_exit:
2324 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2325 		"<== nxge_alloc_rx_cntl_dma status 0x%08x", status));
2326 
2327 	return (status);
2328 }
2329 
2330 /*ARGSUSED*/
2331 static void
2332 nxge_free_rx_cntl_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap)
2333 {
2334 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_free_rx_cntl_dma"));
2335 
2336 	nxge_dma_mem_free(dmap);
2337 
2338 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_free_rx_cntl_dma"));
2339 }
2340 
2341 static nxge_status_t
2342 nxge_alloc_tx_mem_pool(p_nxge_t nxgep)
2343 {
2344 	nxge_status_t		status = NXGE_OK;
2345 	int			i, j;
2346 	uint32_t		ndmas, st_tdc;
2347 	p_nxge_dma_pt_cfg_t	p_all_cfgp;
2348 	p_nxge_hw_pt_cfg_t	p_cfgp;
2349 	p_nxge_dma_pool_t	dma_poolp;
2350 	p_nxge_dma_common_t	*dma_buf_p;
2351 	p_nxge_dma_pool_t	dma_cntl_poolp;
2352 	p_nxge_dma_common_t	*dma_cntl_p;
2353 	size_t			tx_buf_alloc_size;
2354 	size_t			tx_cntl_alloc_size;
2355 	uint32_t		*num_chunks; /* per dma */
2356 	uint32_t		bcopy_thresh;
2357 
2358 	NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_alloc_tx_mem_pool"));
2359 
2360 	p_all_cfgp = (p_nxge_dma_pt_cfg_t)&nxgep->pt_config;
2361 	p_cfgp = (p_nxge_hw_pt_cfg_t)&p_all_cfgp->hw_config;
2362 	st_tdc = p_cfgp->start_tdc;
2363 	ndmas = p_cfgp->max_tdcs;
2364 
2365 	NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_alloc_tx_mem_pool: "
2366 		"p_cfgp 0x%016llx start_tdc %d ndmas %d nxgep->max_tdcs %d",
2367 		p_cfgp, p_cfgp->start_tdc, p_cfgp->max_tdcs, nxgep->max_tdcs));
2368 	/*
2369 	 * Allocate memory for each transmit DMA channel.
2370 	 */
2371 	dma_poolp = (p_nxge_dma_pool_t)KMEM_ZALLOC(sizeof (nxge_dma_pool_t),
2372 			KM_SLEEP);
2373 	dma_buf_p = (p_nxge_dma_common_t *)KMEM_ZALLOC(
2374 			sizeof (p_nxge_dma_common_t) * ndmas, KM_SLEEP);
2375 
2376 	dma_cntl_poolp = (p_nxge_dma_pool_t)
2377 			KMEM_ZALLOC(sizeof (nxge_dma_pool_t), KM_SLEEP);
2378 	dma_cntl_p = (p_nxge_dma_common_t *)KMEM_ZALLOC(
2379 			sizeof (p_nxge_dma_common_t) * ndmas, KM_SLEEP);
2380 
2381 	if (nxge_tx_ring_size > TDC_DEFAULT_MAX) {
2382 		NXGE_DEBUG_MSG((nxgep, MEM_CTL,
2383 		    "nxge_alloc_tx_mem_pool: TDC too high %d, "
2384 		    "set to default %d",
2385 		    nxge_tx_ring_size, TDC_DEFAULT_MAX));
2386 		nxge_tx_ring_size = TDC_DEFAULT_MAX;
2387 	}
2388 
2389 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
2390 	/*
2391 	 * N2/NIU has limitation on the descriptor sizes (contiguous
2392 	 * memory allocation on data buffers to 4M (contig_mem_alloc)
2393 	 * and little endian for control buffers (must use the ddi/dki mem alloc
2394 	 * function). The transmit ring is limited to 8K (includes the
2395 	 * mailbox).
2396 	 */
2397 	if (nxgep->niu_type == N2_NIU) {
2398 		if ((nxge_tx_ring_size > NXGE_NIU_CONTIG_TX_MAX) ||
2399 			(!ISP2(nxge_tx_ring_size))) {
2400 			nxge_tx_ring_size = NXGE_NIU_CONTIG_TX_MAX;
2401 		}
2402 	}
2403 #endif
2404 
2405 	nxgep->nxge_port_tx_ring_size = nxge_tx_ring_size;
2406 
2407 	/*
2408 	 * Assume that each DMA channel will be configured with default
2409 	 * transmit bufer size for copying transmit data.
2410 	 * (For packet payload over this limit, packets will not be
2411 	 *  copied.)
2412 	 */
2413 	if (nxgep->niu_type == N2_NIU) {
2414 		bcopy_thresh = TX_BCOPY_SIZE;
2415 	} else {
2416 		bcopy_thresh = nxge_bcopy_thresh;
2417 	}
2418 	tx_buf_alloc_size = (bcopy_thresh * nxge_tx_ring_size);
2419 
2420 	/*
2421 	 * Addresses of transmit descriptor ring and the
2422 	 * mailbox must be all cache-aligned (64 bytes).
2423 	 */
2424 	tx_cntl_alloc_size = nxge_tx_ring_size;
2425 	tx_cntl_alloc_size *= (sizeof (tx_desc_t));
2426 	tx_cntl_alloc_size += sizeof (txdma_mailbox_t);
2427 
2428 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
2429 	if (nxgep->niu_type == N2_NIU) {
2430 		if (!ISP2(tx_buf_alloc_size)) {
2431 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2432 				"==> nxge_alloc_tx_mem_pool: "
2433 				" must be power of 2"));
2434 			status |= (NXGE_ERROR | NXGE_DDI_FAILED);
2435 			goto nxge_alloc_tx_mem_pool_exit;
2436 		}
2437 
2438 		if (tx_buf_alloc_size > (1 << 22)) {
2439 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2440 				"==> nxge_alloc_tx_mem_pool: "
2441 				" limit size to 4M"));
2442 			status |= (NXGE_ERROR | NXGE_DDI_FAILED);
2443 			goto nxge_alloc_tx_mem_pool_exit;
2444 		}
2445 
2446 		if (tx_cntl_alloc_size < 0x2000) {
2447 			tx_cntl_alloc_size = 0x2000;
2448 		}
2449 	}
2450 #endif
2451 
2452 	num_chunks = (uint32_t *)KMEM_ZALLOC(
2453 			sizeof (uint32_t) * ndmas, KM_SLEEP);
2454 
2455 	/*
2456 	 * Allocate memory for transmit buffers and descriptor rings.
2457 	 * Replace allocation functions with interface functions provided
2458 	 * by the partition manager when it is available.
2459 	 *
2460 	 * Allocate memory for the transmit buffer pool.
2461 	 */
2462 	for (i = 0; i < ndmas; i++) {
2463 		num_chunks[i] = 0;
2464 		status = nxge_alloc_tx_buf_dma(nxgep, st_tdc, &dma_buf_p[i],
2465 					tx_buf_alloc_size,
2466 					bcopy_thresh, &num_chunks[i]);
2467 		if (status != NXGE_OK) {
2468 			break;
2469 		}
2470 		st_tdc++;
2471 	}
2472 	if (i < ndmas) {
2473 		goto nxge_alloc_tx_mem_pool_fail1;
2474 	}
2475 
2476 	st_tdc = p_cfgp->start_tdc;
2477 	/*
2478 	 * Allocate memory for descriptor rings and mailbox.
2479 	 */
2480 	for (j = 0; j < ndmas; j++) {
2481 		status = nxge_alloc_tx_cntl_dma(nxgep, st_tdc, &dma_cntl_p[j],
2482 					tx_cntl_alloc_size);
2483 		if (status != NXGE_OK) {
2484 			break;
2485 		}
2486 		st_tdc++;
2487 	}
2488 	if (j < ndmas) {
2489 		goto nxge_alloc_tx_mem_pool_fail2;
2490 	}
2491 
2492 	dma_poolp->ndmas = ndmas;
2493 	dma_poolp->num_chunks = num_chunks;
2494 	dma_poolp->buf_allocated = B_TRUE;
2495 	dma_poolp->dma_buf_pool_p = dma_buf_p;
2496 	nxgep->tx_buf_pool_p = dma_poolp;
2497 
2498 	dma_cntl_poolp->ndmas = ndmas;
2499 	dma_cntl_poolp->buf_allocated = B_TRUE;
2500 	dma_cntl_poolp->dma_buf_pool_p = dma_cntl_p;
2501 	nxgep->tx_cntl_pool_p = dma_cntl_poolp;
2502 
2503 	NXGE_DEBUG_MSG((nxgep, MEM_CTL,
2504 		"==> nxge_alloc_tx_mem_pool: start_tdc %d "
2505 		"ndmas %d poolp->ndmas %d",
2506 		st_tdc, ndmas, dma_poolp->ndmas));
2507 
2508 	goto nxge_alloc_tx_mem_pool_exit;
2509 
2510 nxge_alloc_tx_mem_pool_fail2:
2511 	/* Free control buffers */
2512 	j--;
2513 	for (; j >= 0; j--) {
2514 		nxge_free_tx_cntl_dma(nxgep,
2515 			(p_nxge_dma_common_t)dma_cntl_p[j]);
2516 	}
2517 
2518 nxge_alloc_tx_mem_pool_fail1:
2519 	/* Free data buffers */
2520 	i--;
2521 	for (; i >= 0; i--) {
2522 		nxge_free_tx_buf_dma(nxgep, (p_nxge_dma_common_t)dma_buf_p[i],
2523 			num_chunks[i]);
2524 	}
2525 
2526 	KMEM_FREE(dma_poolp, sizeof (nxge_dma_pool_t));
2527 	KMEM_FREE(dma_buf_p, ndmas * sizeof (p_nxge_dma_common_t));
2528 	KMEM_FREE(dma_cntl_poolp, sizeof (nxge_dma_pool_t));
2529 	KMEM_FREE(dma_cntl_p, ndmas * sizeof (p_nxge_dma_common_t));
2530 	KMEM_FREE(num_chunks, sizeof (uint32_t) * ndmas);
2531 
2532 nxge_alloc_tx_mem_pool_exit:
2533 	NXGE_DEBUG_MSG((nxgep, MEM_CTL,
2534 		"<== nxge_alloc_tx_mem_pool:status 0x%08x", status));
2535 
2536 	return (status);
2537 }
2538 
2539 static nxge_status_t
2540 nxge_alloc_tx_buf_dma(p_nxge_t nxgep, uint16_t dma_channel,
2541     p_nxge_dma_common_t *dmap, size_t alloc_size,
2542     size_t block_size, uint32_t *num_chunks)
2543 {
2544 	p_nxge_dma_common_t 	tx_dmap;
2545 	nxge_status_t		status = NXGE_OK;
2546 	size_t			total_alloc_size;
2547 	size_t			allocated = 0;
2548 	int			i, size_index, array_size;
2549 
2550 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_tx_buf_dma"));
2551 
2552 	tx_dmap = (p_nxge_dma_common_t)
2553 		KMEM_ZALLOC(sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK,
2554 			KM_SLEEP);
2555 
2556 	total_alloc_size = alloc_size;
2557 	i = 0;
2558 	size_index = 0;
2559 	array_size =  sizeof (alloc_sizes) /  sizeof (size_t);
2560 	while ((alloc_sizes[size_index] < alloc_size) &&
2561 		(size_index < array_size))
2562 		size_index++;
2563 	if (size_index >= array_size) {
2564 		size_index = array_size - 1;
2565 	}
2566 
2567 	while ((allocated < total_alloc_size) &&
2568 			(size_index >= 0) && (i < NXGE_DMA_BLOCK)) {
2569 
2570 		tx_dmap[i].dma_chunk_index = i;
2571 		tx_dmap[i].block_size = block_size;
2572 		tx_dmap[i].alength = alloc_sizes[size_index];
2573 		tx_dmap[i].orig_alength = tx_dmap[i].alength;
2574 		tx_dmap[i].nblocks = alloc_sizes[size_index] / block_size;
2575 		tx_dmap[i].dma_channel = dma_channel;
2576 		tx_dmap[i].contig_alloc_type = B_FALSE;
2577 
2578 		/*
2579 		 * N2/NIU: data buffers must be contiguous as the driver
2580 		 *	   needs to call Hypervisor api to set up
2581 		 *	   logical pages.
2582 		 */
2583 		if ((nxgep->niu_type == N2_NIU) && (NXGE_DMA_BLOCK == 1)) {
2584 			tx_dmap[i].contig_alloc_type = B_TRUE;
2585 		}
2586 
2587 		status = nxge_dma_mem_alloc(nxgep, nxge_force_dma,
2588 			&nxge_tx_dma_attr,
2589 			tx_dmap[i].alength,
2590 			&nxge_dev_buf_dma_acc_attr,
2591 			DDI_DMA_WRITE | DDI_DMA_STREAMING,
2592 			(p_nxge_dma_common_t)(&tx_dmap[i]));
2593 		if (status != NXGE_OK) {
2594 			size_index--;
2595 		} else {
2596 			i++;
2597 			allocated += alloc_sizes[size_index];
2598 		}
2599 	}
2600 
2601 	if (allocated < total_alloc_size) {
2602 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2603 		    "==> nxge_alloc_tx_buf_dma: not enough channel %d: "
2604 		    "allocated 0x%x requested 0x%x",
2605 		    dma_channel,
2606 		    allocated, total_alloc_size));
2607 		status = NXGE_ERROR;
2608 		goto nxge_alloc_tx_mem_fail1;
2609 	}
2610 
2611 	NXGE_DEBUG_MSG((nxgep, MEM2_CTL,
2612 	    "==> nxge_alloc_tx_buf_dma: Allocated for channel %d: "
2613 	    "allocated 0x%x requested 0x%x",
2614 	    dma_channel,
2615 	    allocated, total_alloc_size));
2616 
2617 	*num_chunks = i;
2618 	*dmap = tx_dmap;
2619 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2620 		"==> nxge_alloc_tx_buf_dma dmap 0x%016llx num chunks %d",
2621 		*dmap, i));
2622 	goto nxge_alloc_tx_mem_exit;
2623 
2624 nxge_alloc_tx_mem_fail1:
2625 	KMEM_FREE(tx_dmap, sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK);
2626 
2627 nxge_alloc_tx_mem_exit:
2628 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2629 		"<== nxge_alloc_tx_buf_dma status 0x%08x", status));
2630 
2631 	return (status);
2632 }
2633 
2634 /*ARGSUSED*/
2635 static void
2636 nxge_free_tx_buf_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap,
2637     uint32_t num_chunks)
2638 {
2639 	int		i;
2640 
2641 	NXGE_DEBUG_MSG((nxgep, MEM_CTL, "==> nxge_free_tx_buf_dma"));
2642 
2643 	for (i = 0; i < num_chunks; i++) {
2644 		nxge_dma_mem_free(dmap++);
2645 	}
2646 
2647 	NXGE_DEBUG_MSG((nxgep, MEM_CTL, "<== nxge_free_tx_buf_dma"));
2648 }
2649 
2650 /*ARGSUSED*/
2651 static nxge_status_t
2652 nxge_alloc_tx_cntl_dma(p_nxge_t nxgep, uint16_t dma_channel,
2653     p_nxge_dma_common_t *dmap, size_t size)
2654 {
2655 	p_nxge_dma_common_t 	tx_dmap;
2656 	nxge_status_t		status = NXGE_OK;
2657 
2658 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_alloc_tx_cntl_dma"));
2659 	tx_dmap = (p_nxge_dma_common_t)
2660 			KMEM_ZALLOC(sizeof (nxge_dma_common_t), KM_SLEEP);
2661 
2662 	tx_dmap->contig_alloc_type = B_FALSE;
2663 
2664 	status = nxge_dma_mem_alloc(nxgep, nxge_force_dma,
2665 			&nxge_desc_dma_attr,
2666 			size,
2667 			&nxge_dev_desc_dma_acc_attr,
2668 			DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
2669 			tx_dmap);
2670 	if (status != NXGE_OK) {
2671 		goto nxge_alloc_tx_cntl_dma_fail1;
2672 	}
2673 
2674 	*dmap = tx_dmap;
2675 	goto nxge_alloc_tx_cntl_dma_exit;
2676 
2677 nxge_alloc_tx_cntl_dma_fail1:
2678 	KMEM_FREE(tx_dmap, sizeof (nxge_dma_common_t));
2679 
2680 nxge_alloc_tx_cntl_dma_exit:
2681 	NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2682 		"<== nxge_alloc_tx_cntl_dma status 0x%08x", status));
2683 
2684 	return (status);
2685 }
2686 
2687 /*ARGSUSED*/
2688 static void
2689 nxge_free_tx_cntl_dma(p_nxge_t nxgep, p_nxge_dma_common_t dmap)
2690 {
2691 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "==> nxge_free_tx_cntl_dma"));
2692 
2693 	nxge_dma_mem_free(dmap);
2694 
2695 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_free_tx_cntl_dma"));
2696 }
2697 
2698 static void
2699 nxge_free_tx_mem_pool(p_nxge_t nxgep)
2700 {
2701 	uint32_t		i, ndmas;
2702 	p_nxge_dma_pool_t	dma_poolp;
2703 	p_nxge_dma_common_t	*dma_buf_p;
2704 	p_nxge_dma_pool_t	dma_cntl_poolp;
2705 	p_nxge_dma_common_t	*dma_cntl_p;
2706 	uint32_t 		*num_chunks;
2707 
2708 	NXGE_DEBUG_MSG((nxgep, MEM3_CTL, "==> nxge_free_tx_mem_pool"));
2709 
2710 	dma_poolp = nxgep->tx_buf_pool_p;
2711 	if (dma_poolp == NULL || (!dma_poolp->buf_allocated)) {
2712 		NXGE_DEBUG_MSG((nxgep, MEM3_CTL,
2713 			"<== nxge_free_tx_mem_pool "
2714 			"(null rx buf pool or buf not allocated"));
2715 		return;
2716 	}
2717 
2718 	dma_cntl_poolp = nxgep->tx_cntl_pool_p;
2719 	if (dma_cntl_poolp == NULL || (!dma_cntl_poolp->buf_allocated)) {
2720 		NXGE_DEBUG_MSG((nxgep, MEM3_CTL,
2721 			"<== nxge_free_tx_mem_pool "
2722 			"(null tx cntl buf pool or cntl buf not allocated"));
2723 		return;
2724 	}
2725 
2726 	dma_buf_p = dma_poolp->dma_buf_pool_p;
2727 	num_chunks = dma_poolp->num_chunks;
2728 
2729 	dma_cntl_p = dma_cntl_poolp->dma_buf_pool_p;
2730 	ndmas = dma_cntl_poolp->ndmas;
2731 
2732 	for (i = 0; i < ndmas; i++) {
2733 		nxge_free_tx_buf_dma(nxgep, dma_buf_p[i], num_chunks[i]);
2734 	}
2735 
2736 	for (i = 0; i < ndmas; i++) {
2737 		nxge_free_tx_cntl_dma(nxgep, dma_cntl_p[i]);
2738 	}
2739 
2740 	for (i = 0; i < ndmas; i++) {
2741 		KMEM_FREE(dma_buf_p[i],
2742 			sizeof (nxge_dma_common_t) * NXGE_DMA_BLOCK);
2743 		KMEM_FREE(dma_cntl_p[i], sizeof (nxge_dma_common_t));
2744 	}
2745 
2746 	KMEM_FREE(num_chunks, sizeof (uint32_t) * ndmas);
2747 	KMEM_FREE(dma_cntl_p, ndmas * sizeof (p_nxge_dma_common_t));
2748 	KMEM_FREE(dma_cntl_poolp, sizeof (nxge_dma_pool_t));
2749 	KMEM_FREE(dma_buf_p, ndmas * sizeof (p_nxge_dma_common_t));
2750 	KMEM_FREE(dma_poolp, sizeof (nxge_dma_pool_t));
2751 
2752 	nxgep->tx_buf_pool_p = NULL;
2753 	nxgep->tx_cntl_pool_p = NULL;
2754 
2755 	NXGE_DEBUG_MSG((nxgep, MEM3_CTL, "<== nxge_free_tx_mem_pool"));
2756 }
2757 
2758 /*ARGSUSED*/
2759 static nxge_status_t
2760 nxge_dma_mem_alloc(p_nxge_t nxgep, dma_method_t method,
2761 	struct ddi_dma_attr *dma_attrp,
2762 	size_t length, ddi_device_acc_attr_t *acc_attr_p, uint_t xfer_flags,
2763 	p_nxge_dma_common_t dma_p)
2764 {
2765 	caddr_t 		kaddrp;
2766 	int			ddi_status = DDI_SUCCESS;
2767 	boolean_t		contig_alloc_type;
2768 
2769 	contig_alloc_type = dma_p->contig_alloc_type;
2770 
2771 	if (contig_alloc_type && (nxgep->niu_type != N2_NIU)) {
2772 		/*
2773 		 * contig_alloc_type for contiguous memory only allowed
2774 		 * for N2/NIU.
2775 		 */
2776 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2777 			"nxge_dma_mem_alloc: alloc type not allows (%d)",
2778 			dma_p->contig_alloc_type));
2779 		return (NXGE_ERROR | NXGE_DDI_FAILED);
2780 	}
2781 
2782 	dma_p->dma_handle = NULL;
2783 	dma_p->acc_handle = NULL;
2784 	dma_p->kaddrp = dma_p->last_kaddrp = NULL;
2785 	dma_p->first_ioaddr_pp = dma_p->last_ioaddr_pp = NULL;
2786 	ddi_status = ddi_dma_alloc_handle(nxgep->dip, dma_attrp,
2787 		DDI_DMA_DONTWAIT, NULL, &dma_p->dma_handle);
2788 	if (ddi_status != DDI_SUCCESS) {
2789 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2790 			"nxge_dma_mem_alloc:ddi_dma_alloc_handle failed."));
2791 		return (NXGE_ERROR | NXGE_DDI_FAILED);
2792 	}
2793 
2794 	switch (contig_alloc_type) {
2795 	case B_FALSE:
2796 		ddi_status = ddi_dma_mem_alloc(dma_p->dma_handle, length,
2797 			acc_attr_p,
2798 			xfer_flags,
2799 			DDI_DMA_DONTWAIT, 0, &kaddrp, &dma_p->alength,
2800 			&dma_p->acc_handle);
2801 		if (ddi_status != DDI_SUCCESS) {
2802 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2803 				"nxge_dma_mem_alloc:ddi_dma_mem_alloc failed"));
2804 			ddi_dma_free_handle(&dma_p->dma_handle);
2805 			dma_p->dma_handle = NULL;
2806 			return (NXGE_ERROR | NXGE_DDI_FAILED);
2807 		}
2808 		if (dma_p->alength < length) {
2809 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2810 				"nxge_dma_mem_alloc:ddi_dma_mem_alloc "
2811 				"< length."));
2812 			ddi_dma_mem_free(&dma_p->acc_handle);
2813 			ddi_dma_free_handle(&dma_p->dma_handle);
2814 			dma_p->acc_handle = NULL;
2815 			dma_p->dma_handle = NULL;
2816 			return (NXGE_ERROR);
2817 		}
2818 
2819 		ddi_status = ddi_dma_addr_bind_handle(dma_p->dma_handle, NULL,
2820 			kaddrp, dma_p->alength, xfer_flags, DDI_DMA_DONTWAIT, 0,
2821 			&dma_p->dma_cookie, &dma_p->ncookies);
2822 		if (ddi_status != DDI_DMA_MAPPED) {
2823 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2824 				"nxge_dma_mem_alloc:di_dma_addr_bind failed "
2825 				"(staus 0x%x ncookies %d.)", ddi_status,
2826 				dma_p->ncookies));
2827 			if (dma_p->acc_handle) {
2828 				ddi_dma_mem_free(&dma_p->acc_handle);
2829 				dma_p->acc_handle = NULL;
2830 			}
2831 			ddi_dma_free_handle(&dma_p->dma_handle);
2832 			dma_p->dma_handle = NULL;
2833 			return (NXGE_ERROR | NXGE_DDI_FAILED);
2834 		}
2835 
2836 		if (dma_p->ncookies != 1) {
2837 			NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2838 				"nxge_dma_mem_alloc:ddi_dma_addr_bind "
2839 				"> 1 cookie"
2840 				"(staus 0x%x ncookies %d.)", ddi_status,
2841 				dma_p->ncookies));
2842 			if (dma_p->acc_handle) {
2843 				ddi_dma_mem_free(&dma_p->acc_handle);
2844 				dma_p->acc_handle = NULL;
2845 			}
2846 			(void) ddi_dma_unbind_handle(dma_p->dma_handle);
2847 			ddi_dma_free_handle(&dma_p->dma_handle);
2848 			dma_p->dma_handle = NULL;
2849 			return (NXGE_ERROR);
2850 		}
2851 		break;
2852 
2853 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
2854 	case B_TRUE:
2855 		kaddrp = (caddr_t)contig_mem_alloc(length);
2856 		if (kaddrp == NULL) {
2857 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2858 				"nxge_dma_mem_alloc:contig_mem_alloc failed."));
2859 			ddi_dma_free_handle(&dma_p->dma_handle);
2860 			return (NXGE_ERROR | NXGE_DDI_FAILED);
2861 		}
2862 
2863 		dma_p->alength = length;
2864 		ddi_status = ddi_dma_addr_bind_handle(dma_p->dma_handle, NULL,
2865 			kaddrp, dma_p->alength, xfer_flags, DDI_DMA_DONTWAIT, 0,
2866 			&dma_p->dma_cookie, &dma_p->ncookies);
2867 		if (ddi_status != DDI_DMA_MAPPED) {
2868 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2869 				"nxge_dma_mem_alloc:di_dma_addr_bind failed "
2870 				"(status 0x%x ncookies %d.)", ddi_status,
2871 				dma_p->ncookies));
2872 
2873 			NXGE_DEBUG_MSG((nxgep, DMA_CTL,
2874 				"==> nxge_dma_mem_alloc: (not mapped)"
2875 				"length %lu (0x%x) "
2876 				"free contig kaddrp $%p "
2877 				"va_to_pa $%p",
2878 				length, length,
2879 				kaddrp,
2880 				va_to_pa(kaddrp)));
2881 
2882 
2883 			contig_mem_free((void *)kaddrp, length);
2884 			ddi_dma_free_handle(&dma_p->dma_handle);
2885 
2886 			dma_p->dma_handle = NULL;
2887 			dma_p->acc_handle = NULL;
2888 			dma_p->alength = NULL;
2889 			dma_p->kaddrp = NULL;
2890 
2891 			return (NXGE_ERROR | NXGE_DDI_FAILED);
2892 		}
2893 
2894 		if (dma_p->ncookies != 1 ||
2895 			(dma_p->dma_cookie.dmac_laddress == NULL)) {
2896 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2897 				"nxge_dma_mem_alloc:di_dma_addr_bind > 1 "
2898 				"cookie or "
2899 				"dmac_laddress is NULL $%p size %d "
2900 				" (status 0x%x ncookies %d.)",
2901 				ddi_status,
2902 				dma_p->dma_cookie.dmac_laddress,
2903 				dma_p->dma_cookie.dmac_size,
2904 				dma_p->ncookies));
2905 
2906 			contig_mem_free((void *)kaddrp, length);
2907 			(void) ddi_dma_unbind_handle(dma_p->dma_handle);
2908 			ddi_dma_free_handle(&dma_p->dma_handle);
2909 
2910 			dma_p->alength = 0;
2911 			dma_p->dma_handle = NULL;
2912 			dma_p->acc_handle = NULL;
2913 			dma_p->kaddrp = NULL;
2914 
2915 			return (NXGE_ERROR | NXGE_DDI_FAILED);
2916 		}
2917 		break;
2918 
2919 #else
2920 	case B_TRUE:
2921 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
2922 			"nxge_dma_mem_alloc: invalid alloc type for !sun4v"));
2923 		return (NXGE_ERROR | NXGE_DDI_FAILED);
2924 #endif
2925 	}
2926 
2927 	dma_p->kaddrp = kaddrp;
2928 	dma_p->last_kaddrp = (unsigned char *)kaddrp +
2929 			dma_p->alength - RXBUF_64B_ALIGNED;
2930 #if defined(__i386)
2931 	dma_p->ioaddr_pp =
2932 		(unsigned char *)(uint32_t)dma_p->dma_cookie.dmac_laddress;
2933 #else
2934 	dma_p->ioaddr_pp = (unsigned char *)dma_p->dma_cookie.dmac_laddress;
2935 #endif
2936 	dma_p->last_ioaddr_pp =
2937 #if defined(__i386)
2938 		(unsigned char *)(uint32_t)dma_p->dma_cookie.dmac_laddress +
2939 #else
2940 		(unsigned char *)dma_p->dma_cookie.dmac_laddress +
2941 #endif
2942 				dma_p->alength - RXBUF_64B_ALIGNED;
2943 
2944 	NPI_DMA_ACC_HANDLE_SET(dma_p, dma_p->acc_handle);
2945 
2946 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
2947 	dma_p->orig_ioaddr_pp =
2948 		(unsigned char *)dma_p->dma_cookie.dmac_laddress;
2949 	dma_p->orig_alength = length;
2950 	dma_p->orig_kaddrp = kaddrp;
2951 	dma_p->orig_vatopa = (uint64_t)va_to_pa(kaddrp);
2952 #endif
2953 
2954 	NXGE_DEBUG_MSG((nxgep, DMA_CTL, "<== nxge_dma_mem_alloc: "
2955 		"dma buffer allocated: dma_p $%p "
2956 		"return dmac_ladress from cookie $%p cookie dmac_size %d "
2957 		"dma_p->ioaddr_p $%p "
2958 		"dma_p->orig_ioaddr_p $%p "
2959 		"orig_vatopa $%p "
2960 		"alength %d (0x%x) "
2961 		"kaddrp $%p "
2962 		"length %d (0x%x)",
2963 		dma_p,
2964 		dma_p->dma_cookie.dmac_laddress, dma_p->dma_cookie.dmac_size,
2965 		dma_p->ioaddr_pp,
2966 		dma_p->orig_ioaddr_pp,
2967 		dma_p->orig_vatopa,
2968 		dma_p->alength, dma_p->alength,
2969 		kaddrp,
2970 		length, length));
2971 
2972 	return (NXGE_OK);
2973 }
2974 
2975 static void
2976 nxge_dma_mem_free(p_nxge_dma_common_t dma_p)
2977 {
2978 	if (dma_p->dma_handle != NULL) {
2979 		if (dma_p->ncookies) {
2980 			(void) ddi_dma_unbind_handle(dma_p->dma_handle);
2981 			dma_p->ncookies = 0;
2982 		}
2983 		ddi_dma_free_handle(&dma_p->dma_handle);
2984 		dma_p->dma_handle = NULL;
2985 	}
2986 
2987 	if (dma_p->acc_handle != NULL) {
2988 		ddi_dma_mem_free(&dma_p->acc_handle);
2989 		dma_p->acc_handle = NULL;
2990 		NPI_DMA_ACC_HANDLE_SET(dma_p, NULL);
2991 	}
2992 
2993 #if	defined(sun4v) && defined(NIU_LP_WORKAROUND)
2994 	if (dma_p->contig_alloc_type &&
2995 			dma_p->orig_kaddrp && dma_p->orig_alength) {
2996 		NXGE_DEBUG_MSG((NULL, DMA_CTL, "nxge_dma_mem_free: "
2997 			"kaddrp $%p (orig_kaddrp $%p)"
2998 			"mem type %d ",
2999 			"orig_alength %d "
3000 			"alength 0x%x (%d)",
3001 			dma_p->kaddrp,
3002 			dma_p->orig_kaddrp,
3003 			dma_p->contig_alloc_type,
3004 			dma_p->orig_alength,
3005 			dma_p->alength, dma_p->alength));
3006 
3007 		contig_mem_free(dma_p->orig_kaddrp, dma_p->orig_alength);
3008 		dma_p->orig_alength = NULL;
3009 		dma_p->orig_kaddrp = NULL;
3010 		dma_p->contig_alloc_type = B_FALSE;
3011 	}
3012 #endif
3013 	dma_p->kaddrp = NULL;
3014 	dma_p->alength = NULL;
3015 }
3016 
3017 /*
3018  *	nxge_m_start() -- start transmitting and receiving.
3019  *
3020  *	This function is called by the MAC layer when the first
3021  *	stream is open to prepare the hardware ready for sending
3022  *	and transmitting packets.
3023  */
3024 static int
3025 nxge_m_start(void *arg)
3026 {
3027 	p_nxge_t 	nxgep = (p_nxge_t)arg;
3028 
3029 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_start"));
3030 
3031 	MUTEX_ENTER(nxgep->genlock);
3032 	if (nxge_init(nxgep) != NXGE_OK) {
3033 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
3034 			"<== nxge_m_start: initialization failed"));
3035 		MUTEX_EXIT(nxgep->genlock);
3036 		return (EIO);
3037 	}
3038 
3039 	if (nxgep->nxge_mac_state == NXGE_MAC_STARTED)
3040 		goto nxge_m_start_exit;
3041 	/*
3042 	 * Start timer to check the system error and tx hangs
3043 	 */
3044 	nxgep->nxge_timerid = nxge_start_timer(nxgep, nxge_check_hw_state,
3045 		NXGE_CHECK_TIMER);
3046 
3047 	nxgep->link_notify = B_TRUE;
3048 
3049 	nxgep->nxge_mac_state = NXGE_MAC_STARTED;
3050 
3051 nxge_m_start_exit:
3052 	MUTEX_EXIT(nxgep->genlock);
3053 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_start"));
3054 	return (0);
3055 }
3056 
3057 /*
3058  *	nxge_m_stop(): stop transmitting and receiving.
3059  */
3060 static void
3061 nxge_m_stop(void *arg)
3062 {
3063 	p_nxge_t 	nxgep = (p_nxge_t)arg;
3064 
3065 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_stop"));
3066 
3067 	if (nxgep->nxge_timerid) {
3068 		nxge_stop_timer(nxgep, nxgep->nxge_timerid);
3069 		nxgep->nxge_timerid = 0;
3070 	}
3071 
3072 	MUTEX_ENTER(nxgep->genlock);
3073 	nxge_uninit(nxgep);
3074 
3075 	nxgep->nxge_mac_state = NXGE_MAC_STOPPED;
3076 
3077 	MUTEX_EXIT(nxgep->genlock);
3078 
3079 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_stop"));
3080 }
3081 
3082 static int
3083 nxge_m_unicst(void *arg, const uint8_t *macaddr)
3084 {
3085 	p_nxge_t 	nxgep = (p_nxge_t)arg;
3086 	struct 		ether_addr addrp;
3087 
3088 	NXGE_DEBUG_MSG((nxgep, MAC_CTL, "==> nxge_m_unicst"));
3089 
3090 	bcopy(macaddr, (uint8_t *)&addrp, ETHERADDRL);
3091 	if (nxge_set_mac_addr(nxgep, &addrp)) {
3092 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
3093 			"<== nxge_m_unicst: set unitcast failed"));
3094 		return (EINVAL);
3095 	}
3096 
3097 	NXGE_DEBUG_MSG((nxgep, MAC_CTL, "<== nxge_m_unicst"));
3098 
3099 	return (0);
3100 }
3101 
3102 static int
3103 nxge_m_multicst(void *arg, boolean_t add, const uint8_t *mca)
3104 {
3105 	p_nxge_t 	nxgep = (p_nxge_t)arg;
3106 	struct 		ether_addr addrp;
3107 
3108 	NXGE_DEBUG_MSG((nxgep, MAC_CTL,
3109 		"==> nxge_m_multicst: add %d", add));
3110 
3111 	bcopy(mca, (uint8_t *)&addrp, ETHERADDRL);
3112 	if (add) {
3113 		if (nxge_add_mcast_addr(nxgep, &addrp)) {
3114 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
3115 				"<== nxge_m_multicst: add multicast failed"));
3116 			return (EINVAL);
3117 		}
3118 	} else {
3119 		if (nxge_del_mcast_addr(nxgep, &addrp)) {
3120 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
3121 				"<== nxge_m_multicst: del multicast failed"));
3122 			return (EINVAL);
3123 		}
3124 	}
3125 
3126 	NXGE_DEBUG_MSG((nxgep, MAC_CTL, "<== nxge_m_multicst"));
3127 
3128 	return (0);
3129 }
3130 
3131 static int
3132 nxge_m_promisc(void *arg, boolean_t on)
3133 {
3134 	p_nxge_t 	nxgep = (p_nxge_t)arg;
3135 
3136 	NXGE_DEBUG_MSG((nxgep, MAC_CTL,
3137 		"==> nxge_m_promisc: on %d", on));
3138 
3139 	if (nxge_set_promisc(nxgep, on)) {
3140 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
3141 			"<== nxge_m_promisc: set promisc failed"));
3142 		return (EINVAL);
3143 	}
3144 
3145 	NXGE_DEBUG_MSG((nxgep, MAC_CTL,
3146 		"<== nxge_m_promisc: on %d", on));
3147 
3148 	return (0);
3149 }
3150 
3151 static void
3152 nxge_m_ioctl(void *arg,  queue_t *wq, mblk_t *mp)
3153 {
3154 	p_nxge_t 	nxgep = (p_nxge_t)arg;
3155 	struct 		iocblk *iocp;
3156 	boolean_t 	need_privilege;
3157 	int 		err;
3158 	int 		cmd;
3159 
3160 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_ioctl"));
3161 
3162 	iocp = (struct iocblk *)mp->b_rptr;
3163 	iocp->ioc_error = 0;
3164 	need_privilege = B_TRUE;
3165 	cmd = iocp->ioc_cmd;
3166 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "==> nxge_m_ioctl: cmd 0x%08x", cmd));
3167 	switch (cmd) {
3168 	default:
3169 		miocnak(wq, mp, 0, EINVAL);
3170 		NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_ioctl: invalid"));
3171 		return;
3172 
3173 	case LB_GET_INFO_SIZE:
3174 	case LB_GET_INFO:
3175 	case LB_GET_MODE:
3176 		need_privilege = B_FALSE;
3177 		break;
3178 	case LB_SET_MODE:
3179 		break;
3180 
3181 	case ND_GET:
3182 		need_privilege = B_FALSE;
3183 		break;
3184 	case ND_SET:
3185 		break;
3186 
3187 	case NXGE_GET_MII:
3188 	case NXGE_PUT_MII:
3189 	case NXGE_GET64:
3190 	case NXGE_PUT64:
3191 	case NXGE_GET_TX_RING_SZ:
3192 	case NXGE_GET_TX_DESC:
3193 	case NXGE_TX_SIDE_RESET:
3194 	case NXGE_RX_SIDE_RESET:
3195 	case NXGE_GLOBAL_RESET:
3196 	case NXGE_RESET_MAC:
3197 	case NXGE_TX_REGS_DUMP:
3198 	case NXGE_RX_REGS_DUMP:
3199 	case NXGE_INT_REGS_DUMP:
3200 	case NXGE_VIR_INT_REGS_DUMP:
3201 	case NXGE_PUT_TCAM:
3202 	case NXGE_GET_TCAM:
3203 	case NXGE_RTRACE:
3204 	case NXGE_RDUMP:
3205 
3206 		need_privilege = B_FALSE;
3207 		break;
3208 	case NXGE_INJECT_ERR:
3209 		cmn_err(CE_NOTE, "!nxge_m_ioctl: Inject error\n");
3210 		nxge_err_inject(nxgep, wq, mp);
3211 		break;
3212 	}
3213 
3214 	if (need_privilege) {
3215 		err = secpolicy_net_config(iocp->ioc_cr, B_FALSE);
3216 		if (err != 0) {
3217 			miocnak(wq, mp, 0, err);
3218 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
3219 				"<== nxge_m_ioctl: no priv"));
3220 			return;
3221 		}
3222 	}
3223 
3224 	switch (cmd) {
3225 	case ND_GET:
3226 		NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "ND_GET command"));
3227 	case ND_SET:
3228 		NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "ND_SET command"));
3229 		nxge_param_ioctl(nxgep, wq, mp, iocp);
3230 		break;
3231 
3232 	case LB_GET_MODE:
3233 	case LB_SET_MODE:
3234 	case LB_GET_INFO_SIZE:
3235 	case LB_GET_INFO:
3236 		nxge_loopback_ioctl(nxgep, wq, mp, iocp);
3237 		break;
3238 
3239 	case NXGE_GET_MII:
3240 	case NXGE_PUT_MII:
3241 	case NXGE_PUT_TCAM:
3242 	case NXGE_GET_TCAM:
3243 	case NXGE_GET64:
3244 	case NXGE_PUT64:
3245 	case NXGE_GET_TX_RING_SZ:
3246 	case NXGE_GET_TX_DESC:
3247 	case NXGE_TX_SIDE_RESET:
3248 	case NXGE_RX_SIDE_RESET:
3249 	case NXGE_GLOBAL_RESET:
3250 	case NXGE_RESET_MAC:
3251 	case NXGE_TX_REGS_DUMP:
3252 	case NXGE_RX_REGS_DUMP:
3253 	case NXGE_INT_REGS_DUMP:
3254 	case NXGE_VIR_INT_REGS_DUMP:
3255 		NXGE_DEBUG_MSG((nxgep, NXGE_CTL,
3256 			"==> nxge_m_ioctl: cmd 0x%x", cmd));
3257 		nxge_hw_ioctl(nxgep, wq, mp, iocp);
3258 		break;
3259 	}
3260 
3261 	NXGE_DEBUG_MSG((nxgep, NXGE_CTL, "<== nxge_m_ioctl"));
3262 }
3263 
3264 extern void nxge_rx_hw_blank(void *arg, time_t ticks, uint_t count);
3265 
3266 static void
3267 nxge_m_resources(void *arg)
3268 {
3269 	p_nxge_t		nxgep = arg;
3270 	mac_rx_fifo_t 		mrf;
3271 	p_rx_rcr_rings_t	rcr_rings;
3272 	p_rx_rcr_ring_t		*rcr_p;
3273 	uint32_t		i, ndmas;
3274 	nxge_status_t		status;
3275 
3276 	NXGE_DEBUG_MSG((nxgep, RX_CTL, "==> nxge_m_resources"));
3277 
3278 	MUTEX_ENTER(nxgep->genlock);
3279 
3280 	/*
3281 	 * CR 6492541 Check to see if the drv_state has been initialized,
3282 	 * if not * call nxge_init().
3283 	 */
3284 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
3285 		status = nxge_init(nxgep);
3286 		if (status != NXGE_OK)
3287 			goto nxge_m_resources_exit;
3288 	}
3289 
3290 	mrf.mrf_type = MAC_RX_FIFO;
3291 	mrf.mrf_blank = nxge_rx_hw_blank;
3292 	mrf.mrf_arg = (void *)nxgep;
3293 
3294 	mrf.mrf_normal_blank_time = 128;
3295 	mrf.mrf_normal_pkt_count = 8;
3296 	rcr_rings = nxgep->rx_rcr_rings;
3297 	rcr_p = rcr_rings->rcr_rings;
3298 	ndmas = rcr_rings->ndmas;
3299 
3300 	/*
3301 	 * Export our receive resources to the MAC layer.
3302 	 */
3303 	for (i = 0; i < ndmas; i++) {
3304 		((p_rx_rcr_ring_t)rcr_p[i])->rcr_mac_handle =
3305 				mac_resource_add(nxgep->mach,
3306 				    (mac_resource_t *)&mrf);
3307 
3308 		NXGE_DEBUG_MSG((nxgep, NXGE_CTL,
3309 			"==> nxge_m_resources: vdma %d dma %d "
3310 			"rcrptr 0x%016llx mac_handle 0x%016llx",
3311 			i, ((p_rx_rcr_ring_t)rcr_p[i])->rdc,
3312 			rcr_p[i],
3313 			((p_rx_rcr_ring_t)rcr_p[i])->rcr_mac_handle));
3314 	}
3315 
3316 nxge_m_resources_exit:
3317 	MUTEX_EXIT(nxgep->genlock);
3318 	NXGE_DEBUG_MSG((nxgep, RX_CTL, "<== nxge_m_resources"));
3319 }
3320 
3321 static void
3322 nxge_mmac_kstat_update(p_nxge_t nxgep, mac_addr_slot_t slot, boolean_t factory)
3323 {
3324 	p_nxge_mmac_stats_t mmac_stats;
3325 	int i;
3326 	nxge_mmac_t *mmac_info;
3327 
3328 	mmac_info = &nxgep->nxge_mmac_info;
3329 
3330 	mmac_stats = &nxgep->statsp->mmac_stats;
3331 	mmac_stats->mmac_max_cnt = mmac_info->num_mmac;
3332 	mmac_stats->mmac_avail_cnt = mmac_info->naddrfree;
3333 
3334 	for (i = 0; i < ETHERADDRL; i++) {
3335 		if (factory) {
3336 			mmac_stats->mmac_avail_pool[slot-1].ether_addr_octet[i]
3337 			= mmac_info->factory_mac_pool[slot][(ETHERADDRL-1) - i];
3338 		} else {
3339 			mmac_stats->mmac_avail_pool[slot-1].ether_addr_octet[i]
3340 			= mmac_info->mac_pool[slot].addr[(ETHERADDRL - 1) - i];
3341 		}
3342 	}
3343 }
3344 
3345 /*
3346  * nxge_altmac_set() -- Set an alternate MAC address
3347  */
3348 static int
3349 nxge_altmac_set(p_nxge_t nxgep, uint8_t *maddr, mac_addr_slot_t slot)
3350 {
3351 	uint8_t addrn;
3352 	uint8_t portn;
3353 	npi_mac_addr_t altmac;
3354 	hostinfo_t mac_rdc;
3355 	p_nxge_class_pt_cfg_t clscfgp;
3356 
3357 	altmac.w2 = ((uint16_t)maddr[0] << 8) | ((uint16_t)maddr[1] & 0x0ff);
3358 	altmac.w1 = ((uint16_t)maddr[2] << 8) | ((uint16_t)maddr[3] & 0x0ff);
3359 	altmac.w0 = ((uint16_t)maddr[4] << 8) | ((uint16_t)maddr[5] & 0x0ff);
3360 
3361 	portn = nxgep->mac.portnum;
3362 	addrn = (uint8_t)slot - 1;
3363 
3364 	if (npi_mac_altaddr_entry(nxgep->npi_handle, OP_SET, portn,
3365 		addrn, &altmac) != NPI_SUCCESS)
3366 		return (EIO);
3367 
3368 	/*
3369 	 * Set the rdc table number for the host info entry
3370 	 * for this mac address slot.
3371 	 */
3372 	clscfgp = (p_nxge_class_pt_cfg_t)&nxgep->class_config;
3373 	mac_rdc.value = 0;
3374 	mac_rdc.bits.w0.rdc_tbl_num = clscfgp->mac_host_info[addrn].rdctbl;
3375 	mac_rdc.bits.w0.mac_pref = clscfgp->mac_host_info[addrn].mpr_npr;
3376 
3377 	if (npi_mac_hostinfo_entry(nxgep->npi_handle, OP_SET,
3378 	    nxgep->function_num, addrn, &mac_rdc) != NPI_SUCCESS) {
3379 		return (EIO);
3380 	}
3381 
3382 	/*
3383 	 * Enable comparison with the alternate MAC address.
3384 	 * While the first alternate addr is enabled by bit 1 of register
3385 	 * BMAC_ALTAD_CMPEN, it is enabled by bit 0 of register
3386 	 * XMAC_ADDR_CMPEN, so slot needs to be converted to addrn
3387 	 * accordingly before calling npi_mac_altaddr_entry.
3388 	 */
3389 	if (portn == XMAC_PORT_0 || portn == XMAC_PORT_1)
3390 		addrn = (uint8_t)slot - 1;
3391 	else
3392 		addrn = (uint8_t)slot;
3393 
3394 	if (npi_mac_altaddr_enable(nxgep->npi_handle, portn, addrn)
3395 		!= NPI_SUCCESS)
3396 		return (EIO);
3397 
3398 	return (0);
3399 }
3400 
3401 /*
3402  * nxeg_m_mmac_add() - find an unused address slot, set the address
3403  * value to the one specified, enable the port to start filtering on
3404  * the new MAC address.  Returns 0 on success.
3405  */
3406 static int
3407 nxge_m_mmac_add(void *arg, mac_multi_addr_t *maddr)
3408 {
3409 	p_nxge_t nxgep = arg;
3410 	mac_addr_slot_t slot;
3411 	nxge_mmac_t *mmac_info;
3412 	int err;
3413 	nxge_status_t status;
3414 
3415 	mutex_enter(nxgep->genlock);
3416 
3417 	/*
3418 	 * Make sure that nxge is initialized, if _start() has
3419 	 * not been called.
3420 	 */
3421 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
3422 		status = nxge_init(nxgep);
3423 		if (status != NXGE_OK) {
3424 			mutex_exit(nxgep->genlock);
3425 			return (ENXIO);
3426 		}
3427 	}
3428 
3429 	mmac_info = &nxgep->nxge_mmac_info;
3430 	if (mmac_info->naddrfree == 0) {
3431 		mutex_exit(nxgep->genlock);
3432 		return (ENOSPC);
3433 	}
3434 	if (!mac_unicst_verify(nxgep->mach, maddr->mma_addr,
3435 		maddr->mma_addrlen)) {
3436 		mutex_exit(nxgep->genlock);
3437 		return (EINVAL);
3438 	}
3439 	/*
3440 	 * 	Search for the first available slot. Because naddrfree
3441 	 * is not zero, we are guaranteed to find one.
3442 	 * 	Slot 0 is for unique (primary) MAC. The first alternate
3443 	 * MAC slot is slot 1.
3444 	 *	Each of the first two ports of Neptune has 16 alternate
3445 	 * MAC slots but only the first 7 (or 15) slots have assigned factory
3446 	 * MAC addresses. We first search among the slots without bundled
3447 	 * factory MACs. If we fail to find one in that range, then we
3448 	 * search the slots with bundled factory MACs.  A factory MAC
3449 	 * will be wasted while the slot is used with a user MAC address.
3450 	 * But the slot could be used by factory MAC again after calling
3451 	 * nxge_m_mmac_remove and nxge_m_mmac_reserve.
3452 	 */
3453 	if (mmac_info->num_factory_mmac < mmac_info->num_mmac) {
3454 		for (slot = mmac_info->num_factory_mmac + 1;
3455 			slot <= mmac_info->num_mmac; slot++) {
3456 			if (!(mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED))
3457 				break;
3458 		}
3459 		if (slot > mmac_info->num_mmac) {
3460 			for (slot = 1; slot <= mmac_info->num_factory_mmac;
3461 				slot++) {
3462 				if (!(mmac_info->mac_pool[slot].flags
3463 					& MMAC_SLOT_USED))
3464 					break;
3465 			}
3466 		}
3467 	} else {
3468 		for (slot = 1; slot <= mmac_info->num_mmac; slot++) {
3469 			if (!(mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED))
3470 				break;
3471 		}
3472 	}
3473 	ASSERT(slot <= mmac_info->num_mmac);
3474 	if ((err = nxge_altmac_set(nxgep, maddr->mma_addr, slot)) != 0) {
3475 		mutex_exit(nxgep->genlock);
3476 		return (err);
3477 	}
3478 	bcopy(maddr->mma_addr, mmac_info->mac_pool[slot].addr, ETHERADDRL);
3479 	mmac_info->mac_pool[slot].flags |= MMAC_SLOT_USED;
3480 	mmac_info->mac_pool[slot].flags &= ~MMAC_VENDOR_ADDR;
3481 	mmac_info->naddrfree--;
3482 	nxge_mmac_kstat_update(nxgep, slot, B_FALSE);
3483 
3484 	maddr->mma_slot = slot;
3485 
3486 	mutex_exit(nxgep->genlock);
3487 	return (0);
3488 }
3489 
3490 /*
3491  * This function reserves an unused slot and programs the slot and the HW
3492  * with a factory mac address.
3493  */
3494 static int
3495 nxge_m_mmac_reserve(void *arg, mac_multi_addr_t *maddr)
3496 {
3497 	p_nxge_t nxgep = arg;
3498 	mac_addr_slot_t slot;
3499 	nxge_mmac_t *mmac_info;
3500 	int err;
3501 	nxge_status_t status;
3502 
3503 	mutex_enter(nxgep->genlock);
3504 
3505 	/*
3506 	 * Make sure that nxge is initialized, if _start() has
3507 	 * not been called.
3508 	 */
3509 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
3510 		status = nxge_init(nxgep);
3511 		if (status != NXGE_OK) {
3512 			mutex_exit(nxgep->genlock);
3513 			return (ENXIO);
3514 		}
3515 	}
3516 
3517 	mmac_info = &nxgep->nxge_mmac_info;
3518 	if (mmac_info->naddrfree == 0) {
3519 		mutex_exit(nxgep->genlock);
3520 		return (ENOSPC);
3521 	}
3522 
3523 	slot = maddr->mma_slot;
3524 	if (slot == -1) {  /* -1: Take the first available slot */
3525 		for (slot = 1; slot <= mmac_info->num_factory_mmac; slot++) {
3526 			if (!(mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED))
3527 				break;
3528 		}
3529 		if (slot > mmac_info->num_factory_mmac) {
3530 			mutex_exit(nxgep->genlock);
3531 			return (ENOSPC);
3532 		}
3533 	}
3534 	if (slot < 1 || slot > mmac_info->num_factory_mmac) {
3535 		/*
3536 		 * Do not support factory MAC at a slot greater than
3537 		 * num_factory_mmac even when there are available factory
3538 		 * MAC addresses because the alternate MACs are bundled with
3539 		 * slot[1] through slot[num_factory_mmac]
3540 		 */
3541 		mutex_exit(nxgep->genlock);
3542 		return (EINVAL);
3543 	}
3544 	if (mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED) {
3545 		mutex_exit(nxgep->genlock);
3546 		return (EBUSY);
3547 	}
3548 	/* Verify the address to be reserved */
3549 	if (!mac_unicst_verify(nxgep->mach,
3550 		mmac_info->factory_mac_pool[slot], ETHERADDRL)) {
3551 		mutex_exit(nxgep->genlock);
3552 		return (EINVAL);
3553 	}
3554 	if (err = nxge_altmac_set(nxgep,
3555 		mmac_info->factory_mac_pool[slot], slot)) {
3556 		mutex_exit(nxgep->genlock);
3557 		return (err);
3558 	}
3559 	bcopy(mmac_info->factory_mac_pool[slot], maddr->mma_addr, ETHERADDRL);
3560 	mmac_info->mac_pool[slot].flags |= MMAC_SLOT_USED | MMAC_VENDOR_ADDR;
3561 	mmac_info->naddrfree--;
3562 
3563 	nxge_mmac_kstat_update(nxgep, slot, B_TRUE);
3564 	mutex_exit(nxgep->genlock);
3565 
3566 	/* Pass info back to the caller */
3567 	maddr->mma_slot = slot;
3568 	maddr->mma_addrlen = ETHERADDRL;
3569 	maddr->mma_flags = MMAC_SLOT_USED | MMAC_VENDOR_ADDR;
3570 
3571 	return (0);
3572 }
3573 
3574 /*
3575  * Remove the specified mac address and update the HW not to filter
3576  * the mac address anymore.
3577  */
3578 static int
3579 nxge_m_mmac_remove(void *arg, mac_addr_slot_t slot)
3580 {
3581 	p_nxge_t nxgep = arg;
3582 	nxge_mmac_t *mmac_info;
3583 	uint8_t addrn;
3584 	uint8_t portn;
3585 	int err = 0;
3586 	nxge_status_t status;
3587 
3588 	mutex_enter(nxgep->genlock);
3589 
3590 	/*
3591 	 * Make sure that nxge is initialized, if _start() has
3592 	 * not been called.
3593 	 */
3594 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
3595 		status = nxge_init(nxgep);
3596 		if (status != NXGE_OK) {
3597 			mutex_exit(nxgep->genlock);
3598 			return (ENXIO);
3599 		}
3600 	}
3601 
3602 	mmac_info = &nxgep->nxge_mmac_info;
3603 	if (slot < 1 || slot > mmac_info->num_mmac) {
3604 		mutex_exit(nxgep->genlock);
3605 		return (EINVAL);
3606 	}
3607 
3608 	portn = nxgep->mac.portnum;
3609 	if (portn == XMAC_PORT_0 || portn == XMAC_PORT_1)
3610 		addrn = (uint8_t)slot - 1;
3611 	else
3612 		addrn = (uint8_t)slot;
3613 
3614 	if (mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED) {
3615 		if (npi_mac_altaddr_disable(nxgep->npi_handle, portn, addrn)
3616 				== NPI_SUCCESS) {
3617 			mmac_info->naddrfree++;
3618 			mmac_info->mac_pool[slot].flags &= ~MMAC_SLOT_USED;
3619 			/*
3620 			 * Regardless if the MAC we just stopped filtering
3621 			 * is a user addr or a facory addr, we must set
3622 			 * the MMAC_VENDOR_ADDR flag if this slot has an
3623 			 * associated factory MAC to indicate that a factory
3624 			 * MAC is available.
3625 			 */
3626 			if (slot <= mmac_info->num_factory_mmac) {
3627 				mmac_info->mac_pool[slot].flags
3628 					|= MMAC_VENDOR_ADDR;
3629 			}
3630 			/*
3631 			 * Clear mac_pool[slot].addr so that kstat shows 0
3632 			 * alternate MAC address if the slot is not used.
3633 			 * (But nxge_m_mmac_get returns the factory MAC even
3634 			 * when the slot is not used!)
3635 			 */
3636 			bzero(mmac_info->mac_pool[slot].addr, ETHERADDRL);
3637 			nxge_mmac_kstat_update(nxgep, slot, B_FALSE);
3638 		} else {
3639 			err = EIO;
3640 		}
3641 	} else {
3642 		err = EINVAL;
3643 	}
3644 
3645 	mutex_exit(nxgep->genlock);
3646 	return (err);
3647 }
3648 
3649 
3650 /*
3651  * Modify a mac address added by nxge_m_mmac_add or nxge_m_mmac_reserve().
3652  */
3653 static int
3654 nxge_m_mmac_modify(void *arg, mac_multi_addr_t *maddr)
3655 {
3656 	p_nxge_t nxgep = arg;
3657 	mac_addr_slot_t slot;
3658 	nxge_mmac_t *mmac_info;
3659 	int err = 0;
3660 	nxge_status_t status;
3661 
3662 	if (!mac_unicst_verify(nxgep->mach, maddr->mma_addr,
3663 			maddr->mma_addrlen))
3664 		return (EINVAL);
3665 
3666 	slot = maddr->mma_slot;
3667 
3668 	mutex_enter(nxgep->genlock);
3669 
3670 	/*
3671 	 * Make sure that nxge is initialized, if _start() has
3672 	 * not been called.
3673 	 */
3674 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
3675 		status = nxge_init(nxgep);
3676 		if (status != NXGE_OK) {
3677 			mutex_exit(nxgep->genlock);
3678 			return (ENXIO);
3679 		}
3680 	}
3681 
3682 	mmac_info = &nxgep->nxge_mmac_info;
3683 	if (slot < 1 || slot > mmac_info->num_mmac) {
3684 		mutex_exit(nxgep->genlock);
3685 		return (EINVAL);
3686 	}
3687 	if (mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED) {
3688 		if ((err = nxge_altmac_set(nxgep, maddr->mma_addr, slot))
3689 			!= 0) {
3690 			bcopy(maddr->mma_addr, mmac_info->mac_pool[slot].addr,
3691 				ETHERADDRL);
3692 			/*
3693 			 * Assume that the MAC passed down from the caller
3694 			 * is not a factory MAC address (The user should
3695 			 * call mmac_remove followed by mmac_reserve if
3696 			 * he wants to use the factory MAC for this slot).
3697 			 */
3698 			mmac_info->mac_pool[slot].flags &= ~MMAC_VENDOR_ADDR;
3699 			nxge_mmac_kstat_update(nxgep, slot, B_FALSE);
3700 		}
3701 	} else {
3702 		err = EINVAL;
3703 	}
3704 	mutex_exit(nxgep->genlock);
3705 	return (err);
3706 }
3707 
3708 /*
3709  * nxge_m_mmac_get() - Get the MAC address and other information
3710  * related to the slot.  mma_flags should be set to 0 in the call.
3711  * Note: although kstat shows MAC address as zero when a slot is
3712  * not used, Crossbow expects nxge_m_mmac_get to copy factory MAC
3713  * to the caller as long as the slot is not using a user MAC address.
3714  * The following table shows the rules,
3715  *
3716  *				   USED    VENDOR    mma_addr
3717  * ------------------------------------------------------------
3718  * (1) Slot uses a user MAC:        yes      no     user MAC
3719  * (2) Slot uses a factory MAC:     yes      yes    factory MAC
3720  * (3) Slot is not used but is
3721  *     factory MAC capable:         no       yes    factory MAC
3722  * (4) Slot is not used and is
3723  *     not factory MAC capable:     no       no        0
3724  * ------------------------------------------------------------
3725  */
3726 static int
3727 nxge_m_mmac_get(void *arg, mac_multi_addr_t *maddr)
3728 {
3729 	nxge_t *nxgep = arg;
3730 	mac_addr_slot_t slot;
3731 	nxge_mmac_t *mmac_info;
3732 	nxge_status_t status;
3733 
3734 	slot = maddr->mma_slot;
3735 
3736 	mutex_enter(nxgep->genlock);
3737 
3738 	/*
3739 	 * Make sure that nxge is initialized, if _start() has
3740 	 * not been called.
3741 	 */
3742 	if (!(nxgep->drv_state & STATE_HW_INITIALIZED)) {
3743 		status = nxge_init(nxgep);
3744 		if (status != NXGE_OK) {
3745 			mutex_exit(nxgep->genlock);
3746 			return (ENXIO);
3747 		}
3748 	}
3749 
3750 	mmac_info = &nxgep->nxge_mmac_info;
3751 
3752 	if (slot < 1 || slot > mmac_info->num_mmac) {
3753 		mutex_exit(nxgep->genlock);
3754 		return (EINVAL);
3755 	}
3756 	maddr->mma_flags = 0;
3757 	if (mmac_info->mac_pool[slot].flags & MMAC_SLOT_USED)
3758 		maddr->mma_flags |= MMAC_SLOT_USED;
3759 
3760 	if (mmac_info->mac_pool[slot].flags & MMAC_VENDOR_ADDR) {
3761 		maddr->mma_flags |= MMAC_VENDOR_ADDR;
3762 		bcopy(mmac_info->factory_mac_pool[slot],
3763 			maddr->mma_addr, ETHERADDRL);
3764 		maddr->mma_addrlen = ETHERADDRL;
3765 	} else {
3766 		if (maddr->mma_flags & MMAC_SLOT_USED) {
3767 			bcopy(mmac_info->mac_pool[slot].addr,
3768 				maddr->mma_addr, ETHERADDRL);
3769 			maddr->mma_addrlen = ETHERADDRL;
3770 		} else {
3771 			bzero(maddr->mma_addr, ETHERADDRL);
3772 			maddr->mma_addrlen = 0;
3773 		}
3774 	}
3775 	mutex_exit(nxgep->genlock);
3776 	return (0);
3777 }
3778 
3779 
3780 static boolean_t
3781 nxge_m_getcapab(void *arg, mac_capab_t cap, void *cap_data)
3782 {
3783 	nxge_t *nxgep = arg;
3784 	uint32_t *txflags = cap_data;
3785 	multiaddress_capab_t *mmacp = cap_data;
3786 
3787 	switch (cap) {
3788 	case MAC_CAPAB_HCKSUM:
3789 		*txflags = HCKSUM_INET_PARTIAL;
3790 		break;
3791 	case MAC_CAPAB_POLL:
3792 		/*
3793 		 * There's nothing for us to fill in, simply returning
3794 		 * B_TRUE stating that we support polling is sufficient.
3795 		 */
3796 		break;
3797 
3798 	case MAC_CAPAB_MULTIADDRESS:
3799 		mutex_enter(nxgep->genlock);
3800 
3801 		mmacp->maddr_naddr = nxgep->nxge_mmac_info.num_mmac;
3802 		mmacp->maddr_naddrfree = nxgep->nxge_mmac_info.naddrfree;
3803 		mmacp->maddr_flag = 0; /* 0 is requried by PSARC2006/265 */
3804 		/*
3805 		 * maddr_handle is driver's private data, passed back to
3806 		 * entry point functions as arg.
3807 		 */
3808 		mmacp->maddr_handle	= nxgep;
3809 		mmacp->maddr_add	= nxge_m_mmac_add;
3810 		mmacp->maddr_remove	= nxge_m_mmac_remove;
3811 		mmacp->maddr_modify	= nxge_m_mmac_modify;
3812 		mmacp->maddr_get	= nxge_m_mmac_get;
3813 		mmacp->maddr_reserve	= nxge_m_mmac_reserve;
3814 
3815 		mutex_exit(nxgep->genlock);
3816 		break;
3817 	case MAC_CAPAB_LSO: {
3818 		mac_capab_lso_t *cap_lso = cap_data;
3819 
3820 		if (nxgep->soft_lso_enable) {
3821 			cap_lso->lso_flags = LSO_TX_BASIC_TCP_IPV4;
3822 			if (nxge_lso_max > NXGE_LSO_MAXLEN) {
3823 				nxge_lso_max = NXGE_LSO_MAXLEN;
3824 			}
3825 			cap_lso->lso_basic_tcp_ipv4.lso_max = nxge_lso_max;
3826 			break;
3827 		} else {
3828 			return (B_FALSE);
3829 		}
3830 	}
3831 
3832 	default:
3833 		return (B_FALSE);
3834 	}
3835 	return (B_TRUE);
3836 }
3837 
3838 /*
3839  * Module loading and removing entry points.
3840  */
3841 
3842 static	struct cb_ops 	nxge_cb_ops = {
3843 	nodev,			/* cb_open */
3844 	nodev,			/* cb_close */
3845 	nodev,			/* cb_strategy */
3846 	nodev,			/* cb_print */
3847 	nodev,			/* cb_dump */
3848 	nodev,			/* cb_read */
3849 	nodev,			/* cb_write */
3850 	nodev,			/* cb_ioctl */
3851 	nodev,			/* cb_devmap */
3852 	nodev,			/* cb_mmap */
3853 	nodev,			/* cb_segmap */
3854 	nochpoll,		/* cb_chpoll */
3855 	ddi_prop_op,		/* cb_prop_op */
3856 	NULL,
3857 	D_MP, 			/* cb_flag */
3858 	CB_REV,			/* rev */
3859 	nodev,			/* int (*cb_aread)() */
3860 	nodev			/* int (*cb_awrite)() */
3861 };
3862 
3863 static struct dev_ops nxge_dev_ops = {
3864 	DEVO_REV,		/* devo_rev */
3865 	0,			/* devo_refcnt */
3866 	nulldev,
3867 	nulldev,		/* devo_identify */
3868 	nulldev,		/* devo_probe */
3869 	nxge_attach,		/* devo_attach */
3870 	nxge_detach,		/* devo_detach */
3871 	nodev,			/* devo_reset */
3872 	&nxge_cb_ops,		/* devo_cb_ops */
3873 	(struct bus_ops *)NULL, /* devo_bus_ops	*/
3874 	ddi_power		/* devo_power */
3875 };
3876 
3877 extern	struct	mod_ops	mod_driverops;
3878 
3879 #define	NXGE_DESC_VER		"Sun NIU 10Gb Ethernet"
3880 
3881 /*
3882  * Module linkage information for the kernel.
3883  */
3884 static struct modldrv 	nxge_modldrv = {
3885 	&mod_driverops,
3886 	NXGE_DESC_VER,
3887 	&nxge_dev_ops
3888 };
3889 
3890 static struct modlinkage modlinkage = {
3891 	MODREV_1, (void *) &nxge_modldrv, NULL
3892 };
3893 
3894 int
3895 _init(void)
3896 {
3897 	int		status;
3898 
3899 	NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _init"));
3900 	mac_init_ops(&nxge_dev_ops, "nxge");
3901 	status = ddi_soft_state_init(&nxge_list, sizeof (nxge_t), 0);
3902 	if (status != 0) {
3903 		NXGE_ERROR_MSG((NULL, NXGE_ERR_CTL,
3904 			"failed to init device soft state"));
3905 		goto _init_exit;
3906 	}
3907 	status = mod_install(&modlinkage);
3908 	if (status != 0) {
3909 		ddi_soft_state_fini(&nxge_list);
3910 		NXGE_ERROR_MSG((NULL, NXGE_ERR_CTL, "Mod install failed"));
3911 		goto _init_exit;
3912 	}
3913 
3914 	MUTEX_INIT(&nxge_common_lock, NULL, MUTEX_DRIVER, NULL);
3915 
3916 _init_exit:
3917 	NXGE_DEBUG_MSG((NULL, MOD_CTL, "_init status = 0x%X", status));
3918 
3919 	return (status);
3920 }
3921 
3922 int
3923 _fini(void)
3924 {
3925 	int		status;
3926 
3927 	NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _fini"));
3928 
3929 	NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _fini: mod_remove"));
3930 
3931 	if (nxge_mblks_pending)
3932 		return (EBUSY);
3933 
3934 	status = mod_remove(&modlinkage);
3935 	if (status != DDI_SUCCESS) {
3936 		NXGE_DEBUG_MSG((NULL, MOD_CTL,
3937 			    "Module removal failed 0x%08x",
3938 			    status));
3939 		goto _fini_exit;
3940 	}
3941 
3942 	mac_fini_ops(&nxge_dev_ops);
3943 
3944 	ddi_soft_state_fini(&nxge_list);
3945 
3946 	MUTEX_DESTROY(&nxge_common_lock);
3947 _fini_exit:
3948 	NXGE_DEBUG_MSG((NULL, MOD_CTL, "_fini status = 0x%08x", status));
3949 
3950 	return (status);
3951 }
3952 
3953 int
3954 _info(struct modinfo *modinfop)
3955 {
3956 	int		status;
3957 
3958 	NXGE_DEBUG_MSG((NULL, MOD_CTL, "==> _info"));
3959 	status = mod_info(&modlinkage, modinfop);
3960 	NXGE_DEBUG_MSG((NULL, MOD_CTL, " _info status = 0x%X", status));
3961 
3962 	return (status);
3963 }
3964 
3965 /*ARGSUSED*/
3966 static nxge_status_t
3967 nxge_add_intrs(p_nxge_t nxgep)
3968 {
3969 
3970 	int		intr_types;
3971 	int		type = 0;
3972 	int		ddi_status = DDI_SUCCESS;
3973 	nxge_status_t	status = NXGE_OK;
3974 
3975 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs"));
3976 
3977 	nxgep->nxge_intr_type.intr_registered = B_FALSE;
3978 	nxgep->nxge_intr_type.intr_enabled = B_FALSE;
3979 	nxgep->nxge_intr_type.msi_intx_cnt = 0;
3980 	nxgep->nxge_intr_type.intr_added = 0;
3981 	nxgep->nxge_intr_type.niu_msi_enable = B_FALSE;
3982 	nxgep->nxge_intr_type.intr_type = 0;
3983 
3984 	if (nxgep->niu_type == N2_NIU) {
3985 		nxgep->nxge_intr_type.niu_msi_enable = B_TRUE;
3986 	} else if (nxge_msi_enable) {
3987 		nxgep->nxge_intr_type.niu_msi_enable = B_TRUE;
3988 	}
3989 
3990 	/* Get the supported interrupt types */
3991 	if ((ddi_status = ddi_intr_get_supported_types(nxgep->dip, &intr_types))
3992 			!= DDI_SUCCESS) {
3993 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "<== nxge_add_intrs: "
3994 			"ddi_intr_get_supported_types failed: status 0x%08x",
3995 			ddi_status));
3996 		return (NXGE_ERROR | NXGE_DDI_FAILED);
3997 	}
3998 	nxgep->nxge_intr_type.intr_types = intr_types;
3999 
4000 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: "
4001 		"ddi_intr_get_supported_types: 0x%08x", intr_types));
4002 
4003 	/*
4004 	 * Solaris MSIX is not supported yet. use MSI for now.
4005 	 * nxge_msi_enable (1):
4006 	 *	1 - MSI		2 - MSI-X	others - FIXED
4007 	 */
4008 	switch (nxge_msi_enable) {
4009 	default:
4010 		type = DDI_INTR_TYPE_FIXED;
4011 		NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: "
4012 			"use fixed (intx emulation) type %08x",
4013 			type));
4014 		break;
4015 
4016 	case 2:
4017 		NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: "
4018 			"ddi_intr_get_supported_types: 0x%08x", intr_types));
4019 		if (intr_types & DDI_INTR_TYPE_MSIX) {
4020 			type = DDI_INTR_TYPE_MSIX;
4021 			NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: "
4022 				"ddi_intr_get_supported_types: MSIX 0x%08x",
4023 				type));
4024 		} else if (intr_types & DDI_INTR_TYPE_MSI) {
4025 			type = DDI_INTR_TYPE_MSI;
4026 			NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: "
4027 				"ddi_intr_get_supported_types: MSI 0x%08x",
4028 				type));
4029 		} else if (intr_types & DDI_INTR_TYPE_FIXED) {
4030 			type = DDI_INTR_TYPE_FIXED;
4031 			NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: "
4032 				"ddi_intr_get_supported_types: MSXED0x%08x",
4033 				type));
4034 		}
4035 		break;
4036 
4037 	case 1:
4038 		if (intr_types & DDI_INTR_TYPE_MSI) {
4039 			type = DDI_INTR_TYPE_MSI;
4040 			NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs: "
4041 				"ddi_intr_get_supported_types: MSI 0x%08x",
4042 				type));
4043 		} else if (intr_types & DDI_INTR_TYPE_MSIX) {
4044 			type = DDI_INTR_TYPE_MSIX;
4045 			NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: "
4046 				"ddi_intr_get_supported_types: MSIX 0x%08x",
4047 				type));
4048 		} else if (intr_types & DDI_INTR_TYPE_FIXED) {
4049 			type = DDI_INTR_TYPE_FIXED;
4050 			NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: "
4051 				"ddi_intr_get_supported_types: MSXED0x%08x",
4052 				type));
4053 		}
4054 	}
4055 
4056 	nxgep->nxge_intr_type.intr_type = type;
4057 	if ((type == DDI_INTR_TYPE_MSIX || type == DDI_INTR_TYPE_MSI ||
4058 		type == DDI_INTR_TYPE_FIXED) &&
4059 			nxgep->nxge_intr_type.niu_msi_enable) {
4060 		if ((status = nxge_add_intrs_adv(nxgep)) != DDI_SUCCESS) {
4061 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4062 				    " nxge_add_intrs: "
4063 				    " nxge_add_intrs_adv failed: status 0x%08x",
4064 				    status));
4065 			return (status);
4066 		} else {
4067 			NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs: "
4068 			"interrupts registered : type %d", type));
4069 			nxgep->nxge_intr_type.intr_registered = B_TRUE;
4070 
4071 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
4072 				"\nAdded advanced nxge add_intr_adv "
4073 					"intr type 0x%x\n", type));
4074 
4075 			return (status);
4076 		}
4077 	}
4078 
4079 	if (!nxgep->nxge_intr_type.intr_registered) {
4080 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "==> nxge_add_intrs: "
4081 			"failed to register interrupts"));
4082 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4083 	}
4084 
4085 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_add_intrs"));
4086 	return (status);
4087 }
4088 
4089 /*ARGSUSED*/
4090 static nxge_status_t
4091 nxge_add_soft_intrs(p_nxge_t nxgep)
4092 {
4093 
4094 	int		ddi_status = DDI_SUCCESS;
4095 	nxge_status_t	status = NXGE_OK;
4096 
4097 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_soft_intrs"));
4098 
4099 	nxgep->resched_id = NULL;
4100 	nxgep->resched_running = B_FALSE;
4101 	ddi_status = ddi_add_softintr(nxgep->dip, DDI_SOFTINT_LOW,
4102 			&nxgep->resched_id,
4103 		NULL, NULL, nxge_reschedule, (caddr_t)nxgep);
4104 	if (ddi_status != DDI_SUCCESS) {
4105 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "<== nxge_add_soft_intrs: "
4106 			"ddi_add_softintrs failed: status 0x%08x",
4107 			ddi_status));
4108 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4109 	}
4110 
4111 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_ddi_add_soft_intrs"));
4112 
4113 	return (status);
4114 }
4115 
4116 static nxge_status_t
4117 nxge_add_intrs_adv(p_nxge_t nxgep)
4118 {
4119 	int		intr_type;
4120 	p_nxge_intr_t	intrp;
4121 
4122 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs_adv"));
4123 
4124 	intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type;
4125 	intr_type = intrp->intr_type;
4126 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_add_intrs_adv: type 0x%x",
4127 		intr_type));
4128 
4129 	switch (intr_type) {
4130 	case DDI_INTR_TYPE_MSI: /* 0x2 */
4131 	case DDI_INTR_TYPE_MSIX: /* 0x4 */
4132 		return (nxge_add_intrs_adv_type(nxgep, intr_type));
4133 
4134 	case DDI_INTR_TYPE_FIXED: /* 0x1 */
4135 		return (nxge_add_intrs_adv_type_fix(nxgep, intr_type));
4136 
4137 	default:
4138 		return (NXGE_ERROR);
4139 	}
4140 }
4141 
4142 
4143 /*ARGSUSED*/
4144 static nxge_status_t
4145 nxge_add_intrs_adv_type(p_nxge_t nxgep, uint32_t int_type)
4146 {
4147 	dev_info_t		*dip = nxgep->dip;
4148 	p_nxge_ldg_t		ldgp;
4149 	p_nxge_intr_t		intrp;
4150 	uint_t			*inthandler;
4151 	void			*arg1, *arg2;
4152 	int			behavior;
4153 	int			nintrs, navail, nrequest;
4154 	int			nactual, nrequired;
4155 	int			inum = 0;
4156 	int			x, y;
4157 	int			ddi_status = DDI_SUCCESS;
4158 	nxge_status_t		status = NXGE_OK;
4159 
4160 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs_adv_type"));
4161 	intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type;
4162 	intrp->start_inum = 0;
4163 
4164 	ddi_status = ddi_intr_get_nintrs(dip, int_type, &nintrs);
4165 	if ((ddi_status != DDI_SUCCESS) || (nintrs == 0)) {
4166 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4167 			"ddi_intr_get_nintrs() failed, status: 0x%x%, "
4168 			    "nintrs: %d", ddi_status, nintrs));
4169 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4170 	}
4171 
4172 	ddi_status = ddi_intr_get_navail(dip, int_type, &navail);
4173 	if ((ddi_status != DDI_SUCCESS) || (navail == 0)) {
4174 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4175 			"ddi_intr_get_navail() failed, status: 0x%x%, "
4176 			    "nintrs: %d", ddi_status, navail));
4177 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4178 	}
4179 
4180 	NXGE_DEBUG_MSG((nxgep, INT_CTL,
4181 		"ddi_intr_get_navail() returned: nintrs %d, navail %d",
4182 		    nintrs, navail));
4183 
4184 	/* PSARC/2007/453 MSI-X interrupt limit override */
4185 	if (int_type == DDI_INTR_TYPE_MSIX) {
4186 		nrequest = nxge_create_msi_property(nxgep);
4187 		if (nrequest < navail) {
4188 			navail = nrequest;
4189 			NXGE_DEBUG_MSG((nxgep, INT_CTL,
4190 			    "nxge_add_intrs_adv_type: nintrs %d "
4191 			    "navail %d (nrequest %d)",
4192 			    nintrs, navail, nrequest));
4193 		}
4194 	}
4195 
4196 	if (int_type == DDI_INTR_TYPE_MSI && !ISP2(navail)) {
4197 		/* MSI must be power of 2 */
4198 		if ((navail & 16) == 16) {
4199 			navail = 16;
4200 		} else if ((navail & 8) == 8) {
4201 			navail = 8;
4202 		} else if ((navail & 4) == 4) {
4203 			navail = 4;
4204 		} else if ((navail & 2) == 2) {
4205 			navail = 2;
4206 		} else {
4207 			navail = 1;
4208 		}
4209 		NXGE_DEBUG_MSG((nxgep, INT_CTL,
4210 			"ddi_intr_get_navail(): (msi power of 2) nintrs %d, "
4211 			"navail %d", nintrs, navail));
4212 	}
4213 
4214 	behavior = ((int_type == DDI_INTR_TYPE_FIXED) ? DDI_INTR_ALLOC_STRICT :
4215 			DDI_INTR_ALLOC_NORMAL);
4216 	intrp->intr_size = navail * sizeof (ddi_intr_handle_t);
4217 	intrp->htable = kmem_alloc(intrp->intr_size, KM_SLEEP);
4218 	ddi_status = ddi_intr_alloc(dip, intrp->htable, int_type, inum,
4219 		    navail, &nactual, behavior);
4220 	if (ddi_status != DDI_SUCCESS || nactual == 0) {
4221 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4222 				    " ddi_intr_alloc() failed: %d",
4223 				    ddi_status));
4224 		kmem_free(intrp->htable, intrp->intr_size);
4225 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4226 	}
4227 
4228 	if ((ddi_status = ddi_intr_get_pri(intrp->htable[0],
4229 			(uint_t *)&intrp->pri)) != DDI_SUCCESS) {
4230 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4231 				    " ddi_intr_get_pri() failed: %d",
4232 				    ddi_status));
4233 		/* Free already allocated interrupts */
4234 		for (y = 0; y < nactual; y++) {
4235 			(void) ddi_intr_free(intrp->htable[y]);
4236 		}
4237 
4238 		kmem_free(intrp->htable, intrp->intr_size);
4239 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4240 	}
4241 
4242 	nrequired = 0;
4243 	switch (nxgep->niu_type) {
4244 	default:
4245 		status = nxge_ldgv_init(nxgep, &nactual, &nrequired);
4246 		break;
4247 
4248 	case N2_NIU:
4249 		status = nxge_ldgv_init_n2(nxgep, &nactual, &nrequired);
4250 		break;
4251 	}
4252 
4253 	if (status != NXGE_OK) {
4254 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4255 			"nxge_add_intrs_adv_typ:nxge_ldgv_init "
4256 			"failed: 0x%x", status));
4257 		/* Free already allocated interrupts */
4258 		for (y = 0; y < nactual; y++) {
4259 			(void) ddi_intr_free(intrp->htable[y]);
4260 		}
4261 
4262 		kmem_free(intrp->htable, intrp->intr_size);
4263 		return (status);
4264 	}
4265 
4266 	ldgp = nxgep->ldgvp->ldgp;
4267 	for (x = 0; x < nrequired; x++, ldgp++) {
4268 		ldgp->vector = (uint8_t)x;
4269 		ldgp->intdata = SID_DATA(ldgp->func, x);
4270 		arg1 = ldgp->ldvp;
4271 		arg2 = nxgep;
4272 		if (ldgp->nldvs == 1) {
4273 			inthandler = (uint_t *)ldgp->ldvp->ldv_intr_handler;
4274 			NXGE_DEBUG_MSG((nxgep, INT_CTL,
4275 				"nxge_add_intrs_adv_type: "
4276 				"arg1 0x%x arg2 0x%x: "
4277 				"1-1 int handler (entry %d intdata 0x%x)\n",
4278 				arg1, arg2,
4279 				x, ldgp->intdata));
4280 		} else if (ldgp->nldvs > 1) {
4281 			inthandler = (uint_t *)ldgp->sys_intr_handler;
4282 			NXGE_DEBUG_MSG((nxgep, INT_CTL,
4283 				"nxge_add_intrs_adv_type: "
4284 				"arg1 0x%x arg2 0x%x: "
4285 				"nldevs %d int handler "
4286 				"(entry %d intdata 0x%x)\n",
4287 				arg1, arg2,
4288 				ldgp->nldvs, x, ldgp->intdata));
4289 		}
4290 
4291 		NXGE_DEBUG_MSG((nxgep, INT_CTL,
4292 			"==> nxge_add_intrs_adv_type: ddi_add_intr(inum) #%d "
4293 			"htable 0x%llx", x, intrp->htable[x]));
4294 
4295 		if ((ddi_status = ddi_intr_add_handler(intrp->htable[x],
4296 			(ddi_intr_handler_t *)inthandler, arg1, arg2))
4297 				!= DDI_SUCCESS) {
4298 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4299 				"==> nxge_add_intrs_adv_type: failed #%d "
4300 				"status 0x%x", x, ddi_status));
4301 			for (y = 0; y < intrp->intr_added; y++) {
4302 				(void) ddi_intr_remove_handler(
4303 						intrp->htable[y]);
4304 			}
4305 			/* Free already allocated intr */
4306 			for (y = 0; y < nactual; y++) {
4307 				(void) ddi_intr_free(intrp->htable[y]);
4308 			}
4309 			kmem_free(intrp->htable, intrp->intr_size);
4310 
4311 			(void) nxge_ldgv_uninit(nxgep);
4312 
4313 			return (NXGE_ERROR | NXGE_DDI_FAILED);
4314 		}
4315 		intrp->intr_added++;
4316 	}
4317 
4318 	intrp->msi_intx_cnt = nactual;
4319 
4320 	NXGE_DEBUG_MSG((nxgep, DDI_CTL,
4321 		"Requested: %d, Allowed: %d msi_intx_cnt %d intr_added %d",
4322 		navail, nactual,
4323 		intrp->msi_intx_cnt,
4324 		intrp->intr_added));
4325 
4326 	(void) ddi_intr_get_cap(intrp->htable[0], &intrp->intr_cap);
4327 
4328 	(void) nxge_intr_ldgv_init(nxgep);
4329 
4330 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_add_intrs_adv_type"));
4331 
4332 	return (status);
4333 }
4334 
4335 /*ARGSUSED*/
4336 static nxge_status_t
4337 nxge_add_intrs_adv_type_fix(p_nxge_t nxgep, uint32_t int_type)
4338 {
4339 	dev_info_t		*dip = nxgep->dip;
4340 	p_nxge_ldg_t		ldgp;
4341 	p_nxge_intr_t		intrp;
4342 	uint_t			*inthandler;
4343 	void			*arg1, *arg2;
4344 	int			behavior;
4345 	int			nintrs, navail;
4346 	int			nactual, nrequired;
4347 	int			inum = 0;
4348 	int			x, y;
4349 	int			ddi_status = DDI_SUCCESS;
4350 	nxge_status_t		status = NXGE_OK;
4351 
4352 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_add_intrs_adv_type_fix"));
4353 	intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type;
4354 	intrp->start_inum = 0;
4355 
4356 	ddi_status = ddi_intr_get_nintrs(dip, int_type, &nintrs);
4357 	if ((ddi_status != DDI_SUCCESS) || (nintrs == 0)) {
4358 		NXGE_DEBUG_MSG((nxgep, INT_CTL,
4359 			"ddi_intr_get_nintrs() failed, status: 0x%x%, "
4360 			    "nintrs: %d", status, nintrs));
4361 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4362 	}
4363 
4364 	ddi_status = ddi_intr_get_navail(dip, int_type, &navail);
4365 	if ((ddi_status != DDI_SUCCESS) || (navail == 0)) {
4366 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4367 			"ddi_intr_get_navail() failed, status: 0x%x%, "
4368 			    "nintrs: %d", ddi_status, navail));
4369 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4370 	}
4371 
4372 	NXGE_DEBUG_MSG((nxgep, INT_CTL,
4373 		"ddi_intr_get_navail() returned: nintrs %d, naavail %d",
4374 		    nintrs, navail));
4375 
4376 	behavior = ((int_type == DDI_INTR_TYPE_FIXED) ? DDI_INTR_ALLOC_STRICT :
4377 			DDI_INTR_ALLOC_NORMAL);
4378 	intrp->intr_size = navail * sizeof (ddi_intr_handle_t);
4379 	intrp->htable = kmem_alloc(intrp->intr_size, KM_SLEEP);
4380 	ddi_status = ddi_intr_alloc(dip, intrp->htable, int_type, inum,
4381 		    navail, &nactual, behavior);
4382 	if (ddi_status != DDI_SUCCESS || nactual == 0) {
4383 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4384 			    " ddi_intr_alloc() failed: %d",
4385 			    ddi_status));
4386 		kmem_free(intrp->htable, intrp->intr_size);
4387 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4388 	}
4389 
4390 	if ((ddi_status = ddi_intr_get_pri(intrp->htable[0],
4391 			(uint_t *)&intrp->pri)) != DDI_SUCCESS) {
4392 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4393 				    " ddi_intr_get_pri() failed: %d",
4394 				    ddi_status));
4395 		/* Free already allocated interrupts */
4396 		for (y = 0; y < nactual; y++) {
4397 			(void) ddi_intr_free(intrp->htable[y]);
4398 		}
4399 
4400 		kmem_free(intrp->htable, intrp->intr_size);
4401 		return (NXGE_ERROR | NXGE_DDI_FAILED);
4402 	}
4403 
4404 	nrequired = 0;
4405 	switch (nxgep->niu_type) {
4406 	default:
4407 		status = nxge_ldgv_init(nxgep, &nactual, &nrequired);
4408 		break;
4409 
4410 	case N2_NIU:
4411 		status = nxge_ldgv_init_n2(nxgep, &nactual, &nrequired);
4412 		break;
4413 	}
4414 
4415 	if (status != NXGE_OK) {
4416 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4417 			"nxge_add_intrs_adv_type_fix:nxge_ldgv_init "
4418 			"failed: 0x%x", status));
4419 		/* Free already allocated interrupts */
4420 		for (y = 0; y < nactual; y++) {
4421 			(void) ddi_intr_free(intrp->htable[y]);
4422 		}
4423 
4424 		kmem_free(intrp->htable, intrp->intr_size);
4425 		return (status);
4426 	}
4427 
4428 	ldgp = nxgep->ldgvp->ldgp;
4429 	for (x = 0; x < nrequired; x++, ldgp++) {
4430 		ldgp->vector = (uint8_t)x;
4431 		if (nxgep->niu_type != N2_NIU) {
4432 			ldgp->intdata = SID_DATA(ldgp->func, x);
4433 		}
4434 
4435 		arg1 = ldgp->ldvp;
4436 		arg2 = nxgep;
4437 		if (ldgp->nldvs == 1) {
4438 			inthandler = (uint_t *)ldgp->ldvp->ldv_intr_handler;
4439 			NXGE_DEBUG_MSG((nxgep, INT_CTL,
4440 				"nxge_add_intrs_adv_type_fix: "
4441 				"1-1 int handler(%d) ldg %d ldv %d "
4442 				"arg1 $%p arg2 $%p\n",
4443 				x, ldgp->ldg, ldgp->ldvp->ldv,
4444 				arg1, arg2));
4445 		} else if (ldgp->nldvs > 1) {
4446 			inthandler = (uint_t *)ldgp->sys_intr_handler;
4447 			NXGE_DEBUG_MSG((nxgep, INT_CTL,
4448 				"nxge_add_intrs_adv_type_fix: "
4449 				"shared ldv %d int handler(%d) ldv %d ldg %d"
4450 				"arg1 0x%016llx arg2 0x%016llx\n",
4451 				x, ldgp->nldvs, ldgp->ldg, ldgp->ldvp->ldv,
4452 				arg1, arg2));
4453 		}
4454 
4455 		if ((ddi_status = ddi_intr_add_handler(intrp->htable[x],
4456 			(ddi_intr_handler_t *)inthandler, arg1, arg2))
4457 				!= DDI_SUCCESS) {
4458 			NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL,
4459 				"==> nxge_add_intrs_adv_type_fix: failed #%d "
4460 				"status 0x%x", x, ddi_status));
4461 			for (y = 0; y < intrp->intr_added; y++) {
4462 				(void) ddi_intr_remove_handler(
4463 						intrp->htable[y]);
4464 			}
4465 			for (y = 0; y < nactual; y++) {
4466 				(void) ddi_intr_free(intrp->htable[y]);
4467 			}
4468 			/* Free already allocated intr */
4469 			kmem_free(intrp->htable, intrp->intr_size);
4470 
4471 			(void) nxge_ldgv_uninit(nxgep);
4472 
4473 			return (NXGE_ERROR | NXGE_DDI_FAILED);
4474 		}
4475 		intrp->intr_added++;
4476 	}
4477 
4478 	intrp->msi_intx_cnt = nactual;
4479 
4480 	(void) ddi_intr_get_cap(intrp->htable[0], &intrp->intr_cap);
4481 
4482 	status = nxge_intr_ldgv_init(nxgep);
4483 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_add_intrs_adv_type_fix"));
4484 
4485 	return (status);
4486 }
4487 
4488 static void
4489 nxge_remove_intrs(p_nxge_t nxgep)
4490 {
4491 	int		i, inum;
4492 	p_nxge_intr_t	intrp;
4493 
4494 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_remove_intrs"));
4495 	intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type;
4496 	if (!intrp->intr_registered) {
4497 		NXGE_DEBUG_MSG((nxgep, INT_CTL,
4498 			"<== nxge_remove_intrs: interrupts not registered"));
4499 		return;
4500 	}
4501 
4502 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_remove_intrs:advanced"));
4503 
4504 	if (intrp->intr_cap & DDI_INTR_FLAG_BLOCK) {
4505 		(void) ddi_intr_block_disable(intrp->htable,
4506 			intrp->intr_added);
4507 	} else {
4508 		for (i = 0; i < intrp->intr_added; i++) {
4509 			(void) ddi_intr_disable(intrp->htable[i]);
4510 		}
4511 	}
4512 
4513 	for (inum = 0; inum < intrp->intr_added; inum++) {
4514 		if (intrp->htable[inum]) {
4515 			(void) ddi_intr_remove_handler(intrp->htable[inum]);
4516 		}
4517 	}
4518 
4519 	for (inum = 0; inum < intrp->msi_intx_cnt; inum++) {
4520 		if (intrp->htable[inum]) {
4521 			NXGE_DEBUG_MSG((nxgep, DDI_CTL,
4522 				"nxge_remove_intrs: ddi_intr_free inum %d "
4523 				"msi_intx_cnt %d intr_added %d",
4524 				inum,
4525 				intrp->msi_intx_cnt,
4526 				intrp->intr_added));
4527 
4528 			(void) ddi_intr_free(intrp->htable[inum]);
4529 		}
4530 	}
4531 
4532 	kmem_free(intrp->htable, intrp->intr_size);
4533 	intrp->intr_registered = B_FALSE;
4534 	intrp->intr_enabled = B_FALSE;
4535 	intrp->msi_intx_cnt = 0;
4536 	intrp->intr_added = 0;
4537 
4538 	(void) nxge_ldgv_uninit(nxgep);
4539 
4540 	(void) ddi_prop_remove(DDI_DEV_T_NONE, nxgep->dip,
4541 	    "#msix-request");
4542 
4543 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_remove_intrs"));
4544 }
4545 
4546 /*ARGSUSED*/
4547 static void
4548 nxge_remove_soft_intrs(p_nxge_t nxgep)
4549 {
4550 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_remove_soft_intrs"));
4551 	if (nxgep->resched_id) {
4552 		ddi_remove_softintr(nxgep->resched_id);
4553 		NXGE_DEBUG_MSG((nxgep, INT_CTL,
4554 			"==> nxge_remove_soft_intrs: removed"));
4555 		nxgep->resched_id = NULL;
4556 	}
4557 
4558 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_remove_soft_intrs"));
4559 }
4560 
4561 /*ARGSUSED*/
4562 static void
4563 nxge_intrs_enable(p_nxge_t nxgep)
4564 {
4565 	p_nxge_intr_t	intrp;
4566 	int		i;
4567 	int		status;
4568 
4569 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_enable"));
4570 
4571 	intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type;
4572 
4573 	if (!intrp->intr_registered) {
4574 		NXGE_ERROR_MSG((nxgep, NXGE_ERR_CTL, "<== nxge_intrs_enable: "
4575 			"interrupts are not registered"));
4576 		return;
4577 	}
4578 
4579 	if (intrp->intr_enabled) {
4580 		NXGE_DEBUG_MSG((nxgep, INT_CTL,
4581 			"<== nxge_intrs_enable: already enabled"));
4582 		return;
4583 	}
4584 
4585 	if (intrp->intr_cap & DDI_INTR_FLAG_BLOCK) {
4586 		status = ddi_intr_block_enable(intrp->htable,
4587 			intrp->intr_added);
4588 		NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_enable "
4589 			"block enable - status 0x%x total inums #%d\n",
4590 			status, intrp->intr_added));
4591 	} else {
4592 		for (i = 0; i < intrp->intr_added; i++) {
4593 			status = ddi_intr_enable(intrp->htable[i]);
4594 			NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_enable "
4595 				"ddi_intr_enable:enable - status 0x%x "
4596 				"total inums %d enable inum #%d\n",
4597 				status, intrp->intr_added, i));
4598 			if (status == DDI_SUCCESS) {
4599 				intrp->intr_enabled = B_TRUE;
4600 			}
4601 		}
4602 	}
4603 
4604 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_intrs_enable"));
4605 }
4606 
4607 /*ARGSUSED*/
4608 static void
4609 nxge_intrs_disable(p_nxge_t nxgep)
4610 {
4611 	p_nxge_intr_t	intrp;
4612 	int		i;
4613 
4614 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "==> nxge_intrs_disable"));
4615 
4616 	intrp = (p_nxge_intr_t)&nxgep->nxge_intr_type;
4617 
4618 	if (!intrp->intr_registered) {
4619 		NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_intrs_disable: "
4620 			"interrupts are not registered"));
4621 		return;
4622 	}
4623 
4624 	if (intrp->intr_cap & DDI_INTR_FLAG_BLOCK) {
4625 		(void) ddi_intr_block_disable(intrp->htable,
4626 			intrp->intr_added);
4627 	} else {
4628 		for (i = 0; i < intrp->intr_added; i++) {
4629 			(void) ddi_intr_disable(intrp->htable[i]);
4630 		}
4631 	}
4632 
4633 	intrp->intr_enabled = B_FALSE;
4634 	NXGE_DEBUG_MSG((nxgep, INT_CTL, "<== nxge_intrs_disable"));
4635 }
4636 
4637 static nxge_status_t
4638 nxge_mac_register(p_nxge_t nxgep)
4639 {
4640 	mac_register_t *macp;
4641 	int		status;
4642 
4643 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "==> nxge_mac_register"));
4644 
4645 	if ((macp = mac_alloc(MAC_VERSION)) == NULL)
4646 		return (NXGE_ERROR);
4647 
4648 	macp->m_type_ident = MAC_PLUGIN_IDENT_ETHER;
4649 	macp->m_driver = nxgep;
4650 	macp->m_dip = nxgep->dip;
4651 	macp->m_src_addr = nxgep->ouraddr.ether_addr_octet;
4652 	macp->m_callbacks = &nxge_m_callbacks;
4653 	macp->m_min_sdu = 0;
4654 	macp->m_max_sdu = nxgep->mac.maxframesize -
4655 		sizeof (struct ether_header) - ETHERFCSL - 4;
4656 	macp->m_margin = VLAN_TAGSZ;
4657 
4658 	status = mac_register(macp, &nxgep->mach);
4659 	mac_free(macp);
4660 
4661 	if (status != 0) {
4662 		cmn_err(CE_WARN,
4663 			"!nxge_mac_register failed (status %d instance %d)",
4664 			status, nxgep->instance);
4665 		return (NXGE_ERROR);
4666 	}
4667 
4668 	NXGE_DEBUG_MSG((nxgep, DDI_CTL, "<== nxge_mac_register success "
4669 		"(instance %d)", nxgep->instance));
4670 
4671 	return (NXGE_OK);
4672 }
4673 
4674 void
4675 nxge_err_inject(p_nxge_t nxgep, queue_t *wq, mblk_t *mp)
4676 {
4677 	ssize_t		size;
4678 	mblk_t		*nmp;
4679 	uint8_t		blk_id;
4680 	uint8_t		chan;
4681 	uint32_t	err_id;
4682 	err_inject_t	*eip;
4683 
4684 	NXGE_DEBUG_MSG((nxgep, STR_CTL, "==> nxge_err_inject"));
4685 
4686 	size = 1024;
4687 	nmp = mp->b_cont;
4688 	eip = (err_inject_t *)nmp->b_rptr;
4689 	blk_id = eip->blk_id;
4690 	err_id = eip->err_id;
4691 	chan = eip->chan;
4692 	cmn_err(CE_NOTE, "!blk_id = 0x%x\n", blk_id);
4693 	cmn_err(CE_NOTE, "!err_id = 0x%x\n", err_id);
4694 	cmn_err(CE_NOTE, "!chan = 0x%x\n", chan);
4695 	switch (blk_id) {
4696 	case MAC_BLK_ID:
4697 		break;
4698 	case TXMAC_BLK_ID:
4699 		break;
4700 	case RXMAC_BLK_ID:
4701 		break;
4702 	case MIF_BLK_ID:
4703 		break;
4704 	case IPP_BLK_ID:
4705 		nxge_ipp_inject_err(nxgep, err_id);
4706 		break;
4707 	case TXC_BLK_ID:
4708 		nxge_txc_inject_err(nxgep, err_id);
4709 		break;
4710 	case TXDMA_BLK_ID:
4711 		nxge_txdma_inject_err(nxgep, err_id, chan);
4712 		break;
4713 	case RXDMA_BLK_ID:
4714 		nxge_rxdma_inject_err(nxgep, err_id, chan);
4715 		break;
4716 	case ZCP_BLK_ID:
4717 		nxge_zcp_inject_err(nxgep, err_id);
4718 		break;
4719 	case ESPC_BLK_ID:
4720 		break;
4721 	case FFLP_BLK_ID:
4722 		break;
4723 	case PHY_BLK_ID:
4724 		break;
4725 	case ETHER_SERDES_BLK_ID:
4726 		break;
4727 	case PCIE_SERDES_BLK_ID:
4728 		break;
4729 	case VIR_BLK_ID:
4730 		break;
4731 	}
4732 
4733 	nmp->b_wptr = nmp->b_rptr + size;
4734 	NXGE_DEBUG_MSG((nxgep, STR_CTL, "<== nxge_err_inject"));
4735 
4736 	miocack(wq, mp, (int)size, 0);
4737 }
4738 
4739 static int
4740 nxge_init_common_dev(p_nxge_t nxgep)
4741 {
4742 	p_nxge_hw_list_t	hw_p;
4743 	dev_info_t 		*p_dip;
4744 
4745 	NXGE_DEBUG_MSG((nxgep, MOD_CTL, "==> nxge_init_common_device"));
4746 
4747 	p_dip = nxgep->p_dip;
4748 	MUTEX_ENTER(&nxge_common_lock);
4749 	NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4750 		"==> nxge_init_common_dev:func # %d",
4751 			nxgep->function_num));
4752 	/*
4753 	 * Loop through existing per neptune hardware list.
4754 	 */
4755 	for (hw_p = nxge_hw_list; hw_p; hw_p = hw_p->next) {
4756 		NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4757 			"==> nxge_init_common_device:func # %d "
4758 			"hw_p $%p parent dip $%p",
4759 			nxgep->function_num,
4760 			hw_p,
4761 			p_dip));
4762 		if (hw_p->parent_devp == p_dip) {
4763 			nxgep->nxge_hw_p = hw_p;
4764 			hw_p->ndevs++;
4765 			hw_p->nxge_p[nxgep->function_num] = nxgep;
4766 			NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4767 				"==> nxge_init_common_device:func # %d "
4768 				"hw_p $%p parent dip $%p "
4769 				"ndevs %d (found)",
4770 				nxgep->function_num,
4771 				hw_p,
4772 				p_dip,
4773 				hw_p->ndevs));
4774 			break;
4775 		}
4776 	}
4777 
4778 	if (hw_p == NULL) {
4779 		NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4780 			"==> nxge_init_common_device:func # %d "
4781 			"parent dip $%p (new)",
4782 			nxgep->function_num,
4783 			p_dip));
4784 		hw_p = kmem_zalloc(sizeof (nxge_hw_list_t), KM_SLEEP);
4785 		hw_p->parent_devp = p_dip;
4786 		hw_p->magic = NXGE_NEPTUNE_MAGIC;
4787 		nxgep->nxge_hw_p = hw_p;
4788 		hw_p->ndevs++;
4789 		hw_p->nxge_p[nxgep->function_num] = nxgep;
4790 		hw_p->next = nxge_hw_list;
4791 		if (nxgep->niu_type == N2_NIU) {
4792 			hw_p->niu_type = N2_NIU;
4793 			hw_p->platform_type = P_NEPTUNE_NIU;
4794 		} else {
4795 			hw_p->niu_type = NIU_TYPE_NONE;
4796 			hw_p->platform_type = P_NEPTUNE_NONE;
4797 		}
4798 
4799 		MUTEX_INIT(&hw_p->nxge_cfg_lock, NULL, MUTEX_DRIVER, NULL);
4800 		MUTEX_INIT(&hw_p->nxge_tcam_lock, NULL, MUTEX_DRIVER, NULL);
4801 		MUTEX_INIT(&hw_p->nxge_vlan_lock, NULL, MUTEX_DRIVER, NULL);
4802 		MUTEX_INIT(&hw_p->nxge_mdio_lock, NULL, MUTEX_DRIVER, NULL);
4803 
4804 		nxge_hw_list = hw_p;
4805 
4806 		(void) nxge_scan_ports_phy(nxgep, nxge_hw_list);
4807 	}
4808 
4809 	MUTEX_EXIT(&nxge_common_lock);
4810 
4811 	nxgep->platform_type = hw_p->platform_type;
4812 	if (nxgep->niu_type != N2_NIU) {
4813 		nxgep->niu_type = hw_p->niu_type;
4814 	}
4815 
4816 	NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4817 		"==> nxge_init_common_device (nxge_hw_list) $%p",
4818 		nxge_hw_list));
4819 	NXGE_DEBUG_MSG((nxgep, MOD_CTL, "<== nxge_init_common_device"));
4820 
4821 	return (NXGE_OK);
4822 }
4823 
4824 static void
4825 nxge_uninit_common_dev(p_nxge_t nxgep)
4826 {
4827 	p_nxge_hw_list_t	hw_p, h_hw_p;
4828 	dev_info_t 		*p_dip;
4829 
4830 	NXGE_DEBUG_MSG((nxgep, MOD_CTL, "==> nxge_uninit_common_device"));
4831 	if (nxgep->nxge_hw_p == NULL) {
4832 		NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4833 			"<== nxge_uninit_common_device (no common)"));
4834 		return;
4835 	}
4836 
4837 	MUTEX_ENTER(&nxge_common_lock);
4838 	h_hw_p = nxge_hw_list;
4839 	for (hw_p = nxge_hw_list; hw_p; hw_p = hw_p->next) {
4840 		p_dip = hw_p->parent_devp;
4841 		if (nxgep->nxge_hw_p == hw_p &&
4842 			p_dip == nxgep->p_dip &&
4843 			nxgep->nxge_hw_p->magic == NXGE_NEPTUNE_MAGIC &&
4844 			hw_p->magic == NXGE_NEPTUNE_MAGIC) {
4845 
4846 			NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4847 				"==> nxge_uninit_common_device:func # %d "
4848 				"hw_p $%p parent dip $%p "
4849 				"ndevs %d (found)",
4850 				nxgep->function_num,
4851 				hw_p,
4852 				p_dip,
4853 				hw_p->ndevs));
4854 
4855 			nxgep->nxge_hw_p = NULL;
4856 			if (hw_p->ndevs) {
4857 				hw_p->ndevs--;
4858 			}
4859 			hw_p->nxge_p[nxgep->function_num] = NULL;
4860 			if (!hw_p->ndevs) {
4861 				MUTEX_DESTROY(&hw_p->nxge_vlan_lock);
4862 				MUTEX_DESTROY(&hw_p->nxge_tcam_lock);
4863 				MUTEX_DESTROY(&hw_p->nxge_cfg_lock);
4864 				MUTEX_DESTROY(&hw_p->nxge_mdio_lock);
4865 				NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4866 					"==> nxge_uninit_common_device: "
4867 					"func # %d "
4868 					"hw_p $%p parent dip $%p "
4869 					"ndevs %d (last)",
4870 					nxgep->function_num,
4871 					hw_p,
4872 					p_dip,
4873 					hw_p->ndevs));
4874 
4875 				if (hw_p == nxge_hw_list) {
4876 					NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4877 						"==> nxge_uninit_common_device:"
4878 						"remove head func # %d "
4879 						"hw_p $%p parent dip $%p "
4880 						"ndevs %d (head)",
4881 						nxgep->function_num,
4882 						hw_p,
4883 						p_dip,
4884 						hw_p->ndevs));
4885 					nxge_hw_list = hw_p->next;
4886 				} else {
4887 					NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4888 						"==> nxge_uninit_common_device:"
4889 						"remove middle func # %d "
4890 						"hw_p $%p parent dip $%p "
4891 						"ndevs %d (middle)",
4892 						nxgep->function_num,
4893 						hw_p,
4894 						p_dip,
4895 						hw_p->ndevs));
4896 					h_hw_p->next = hw_p->next;
4897 				}
4898 
4899 				KMEM_FREE(hw_p, sizeof (nxge_hw_list_t));
4900 			}
4901 			break;
4902 		} else {
4903 			h_hw_p = hw_p;
4904 		}
4905 	}
4906 
4907 	MUTEX_EXIT(&nxge_common_lock);
4908 	NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4909 		"==> nxge_uninit_common_device (nxge_hw_list) $%p",
4910 		nxge_hw_list));
4911 
4912 	NXGE_DEBUG_MSG((nxgep, MOD_CTL, "<= nxge_uninit_common_device"));
4913 }
4914 
4915 /*
4916  * Determines the number of ports from the niu_type or the platform type.
4917  * Returns the number of ports, or returns zero on failure.
4918  */
4919 
4920 int
4921 nxge_get_nports(p_nxge_t nxgep)
4922 {
4923 	int	nports = 0;
4924 
4925 	switch (nxgep->niu_type) {
4926 	case N2_NIU:
4927 	case NEPTUNE_2_10GF:
4928 		nports = 2;
4929 		break;
4930 	case NEPTUNE_4_1GC:
4931 	case NEPTUNE_2_10GF_2_1GC:
4932 	case NEPTUNE_1_10GF_3_1GC:
4933 	case NEPTUNE_1_1GC_1_10GF_2_1GC:
4934 		nports = 4;
4935 		break;
4936 	default:
4937 		switch (nxgep->platform_type) {
4938 		case P_NEPTUNE_NIU:
4939 		case P_NEPTUNE_ATLAS_2PORT:
4940 			nports = 2;
4941 			break;
4942 		case P_NEPTUNE_ATLAS_4PORT:
4943 		case P_NEPTUNE_MARAMBA_P0:
4944 		case P_NEPTUNE_MARAMBA_P1:
4945 		case P_NEPTUNE_ALONSO:
4946 			nports = 4;
4947 			break;
4948 		default:
4949 			break;
4950 		}
4951 		break;
4952 	}
4953 
4954 	return (nports);
4955 }
4956 
4957 /*
4958  * The following two functions are to support
4959  * PSARC/2007/453 MSI-X interrupt limit override.
4960  */
4961 static int
4962 nxge_create_msi_property(p_nxge_t nxgep)
4963 {
4964 	int	nmsi;
4965 	extern	int ncpus;
4966 
4967 	NXGE_DEBUG_MSG((nxgep, MOD_CTL, "==>nxge_create_msi_property"));
4968 
4969 	switch (nxgep->mac.portmode) {
4970 	case PORT_10G_COPPER:
4971 	case PORT_10G_FIBER:
4972 		(void) ddi_prop_create(DDI_DEV_T_NONE, nxgep->dip,
4973 		    DDI_PROP_CANSLEEP, "#msix-request", NULL, 0);
4974 		/*
4975 		 * The maximum MSI-X requested will be 8.
4976 		 * If the # of CPUs is less than 8, we will reqeust
4977 		 * # MSI-X based on the # of CPUs.
4978 		 */
4979 		if (ncpus >= NXGE_MSIX_REQUEST_10G) {
4980 			nmsi = NXGE_MSIX_REQUEST_10G;
4981 		} else {
4982 			nmsi = ncpus;
4983 		}
4984 		NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4985 		    "==>nxge_create_msi_property(10G): exists 0x%x (nmsi %d)",
4986 		    ddi_prop_exists(DDI_DEV_T_NONE, nxgep->dip,
4987 		    DDI_PROP_CANSLEEP, "#msix-request"), nmsi));
4988 		break;
4989 
4990 	default:
4991 		nmsi = NXGE_MSIX_REQUEST_1G;
4992 		NXGE_DEBUG_MSG((nxgep, MOD_CTL,
4993 		    "==>nxge_create_msi_property(1G): exists 0x%x (nmsi %d)",
4994 		    ddi_prop_exists(DDI_DEV_T_NONE, nxgep->dip,
4995 		    DDI_PROP_CANSLEEP, "#msix-request"), nmsi));
4996 		break;
4997 	}
4998 
4999 	NXGE_DEBUG_MSG((nxgep, MOD_CTL, "<==nxge_create_msi_property"));
5000 	return (nmsi);
5001 }
5002