xref: /illumos-gate/usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c (revision 33c72b7598992897b94815b1f47b7b8077e53808)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2010 QLogic Corporation. All rights reserved.
24  */
25 
26 /*
27  * Copyright (c) 2018, Joyent, Inc.
28  */
29 
30 #include <qlge.h>
31 #include <sys/atomic.h>
32 #include <sys/strsubr.h>
33 #include <sys/pattr.h>
34 #include <netinet/in.h>
35 #include <netinet/ip.h>
36 #include <netinet/ip6.h>
37 #include <netinet/tcp.h>
38 #include <netinet/udp.h>
39 #include <inet/ip.h>
40 
41 
42 
43 /*
44  * Local variables
45  */
46 static struct ether_addr ql_ether_broadcast_addr =
47 	{0xff, 0xff, 0xff, 0xff, 0xff, 0xff};
48 static char version[] = "GLDv3 QLogic 81XX " VERSIONSTR;
49 
50 /*
51  * Local function prototypes
52  */
53 static void ql_free_resources(qlge_t *);
54 static void ql_fini_kstats(qlge_t *);
55 static uint32_t ql_get_link_state(qlge_t *);
56 static void ql_read_conf(qlge_t *);
57 static int ql_alloc_phys(dev_info_t *, ddi_dma_handle_t *,
58     ddi_device_acc_attr_t *, uint_t, ddi_acc_handle_t *,
59     size_t, size_t, caddr_t *, ddi_dma_cookie_t *);
60 static int ql_alloc_phys_rbuf(dev_info_t *, ddi_dma_handle_t *,
61     ddi_device_acc_attr_t *, uint_t, ddi_acc_handle_t *,
62     size_t, size_t, caddr_t *, ddi_dma_cookie_t *);
63 static void ql_free_phys(ddi_dma_handle_t *, ddi_acc_handle_t *);
64 static int ql_set_routing_reg(qlge_t *, uint32_t, uint32_t, int);
65 static int ql_attach(dev_info_t *, ddi_attach_cmd_t);
66 static int ql_detach(dev_info_t *, ddi_detach_cmd_t);
67 static int ql_bringdown_adapter(qlge_t *);
68 static int ql_bringup_adapter(qlge_t *);
69 static int ql_asic_reset(qlge_t *);
70 static void ql_wake_mpi_reset_soft_intr(qlge_t *);
71 static void ql_stop_timer(qlge_t *qlge);
72 static void ql_fm_fini(qlge_t *qlge);
73 int ql_clean_outbound_rx_ring(struct rx_ring *rx_ring);
74 
75 /*
76  * TX dma maping handlers allow multiple sscatter-gather lists
77  */
78 ddi_dma_attr_t  tx_mapping_dma_attr = {
79 	DMA_ATTR_V0,			/* dma_attr_version */
80 	QL_DMA_LOW_ADDRESS,		/* low DMA address range */
81 	QL_DMA_HIGH_64BIT_ADDRESS,	/* high DMA address range */
82 	QL_DMA_XFER_COUNTER,		/* DMA counter register */
83 	QL_DMA_ADDRESS_ALIGNMENT,	/* DMA address alignment, default - 8 */
84 	QL_DMA_BURSTSIZES,		/* DMA burstsizes */
85 	QL_DMA_MIN_XFER_SIZE,		/* min effective DMA size */
86 	QL_DMA_MAX_XFER_SIZE,		/* max DMA xfer size */
87 	QL_DMA_SEGMENT_BOUNDARY,	/* segment boundary */
88 	QL_MAX_TX_DMA_HANDLES,		/* s/g list length */
89 	QL_DMA_GRANULARITY,		/* granularity of device */
90 	DDI_DMA_RELAXED_ORDERING	/* DMA transfer flags */
91 };
92 
93 /*
94  * Receive buffers and Request/Response queues do not allow scatter-gather lists
95  */
96 ddi_dma_attr_t  dma_attr = {
97 	DMA_ATTR_V0,			/* dma_attr_version */
98 	QL_DMA_LOW_ADDRESS,		/* low DMA address range */
99 	QL_DMA_HIGH_64BIT_ADDRESS,	/* high DMA address range */
100 	QL_DMA_XFER_COUNTER,		/* DMA counter register */
101 	QL_DMA_ADDRESS_ALIGNMENT,	/* DMA address alignment, default - 8 */
102 	QL_DMA_BURSTSIZES,		/* DMA burstsizes */
103 	QL_DMA_MIN_XFER_SIZE,		/* min effective DMA size */
104 	QL_DMA_MAX_XFER_SIZE,		/* max DMA xfer size */
105 	QL_DMA_SEGMENT_BOUNDARY,	/* segment boundary */
106 	1,				/* s/g list length, i.e no sg list */
107 	QL_DMA_GRANULARITY,		/* granularity of device */
108 	QL_DMA_XFER_FLAGS		/* DMA transfer flags */
109 };
110 /*
111  * Receive buffers do not allow scatter-gather lists
112  */
113 ddi_dma_attr_t  dma_attr_rbuf = {
114 	DMA_ATTR_V0,			/* dma_attr_version */
115 	QL_DMA_LOW_ADDRESS,		/* low DMA address range */
116 	QL_DMA_HIGH_64BIT_ADDRESS,	/* high DMA address range */
117 	QL_DMA_XFER_COUNTER,		/* DMA counter register */
118 	0x1,				/* DMA address alignment, default - 8 */
119 	QL_DMA_BURSTSIZES,		/* DMA burstsizes */
120 	QL_DMA_MIN_XFER_SIZE,		/* min effective DMA size */
121 	QL_DMA_MAX_XFER_SIZE,		/* max DMA xfer size */
122 	QL_DMA_SEGMENT_BOUNDARY,	/* segment boundary */
123 	1,				/* s/g list length, i.e no sg list */
124 	QL_DMA_GRANULARITY,		/* granularity of device */
125 	DDI_DMA_RELAXED_ORDERING	/* DMA transfer flags */
126 };
127 /*
128  * DMA access attribute structure.
129  */
130 /* device register access from host */
131 ddi_device_acc_attr_t ql_dev_acc_attr = {
132 	DDI_DEVICE_ATTR_V0,
133 	DDI_STRUCTURE_LE_ACC,
134 	DDI_STRICTORDER_ACC
135 };
136 
137 /* host ring descriptors */
138 ddi_device_acc_attr_t ql_desc_acc_attr = {
139 	DDI_DEVICE_ATTR_V0,
140 	DDI_NEVERSWAP_ACC,
141 	DDI_STRICTORDER_ACC
142 };
143 
144 /* host ring buffer */
145 ddi_device_acc_attr_t ql_buf_acc_attr = {
146 	DDI_DEVICE_ATTR_V0,
147 	DDI_NEVERSWAP_ACC,
148 	DDI_STRICTORDER_ACC
149 };
150 
151 /*
152  * Hash key table for Receive Side Scaling (RSS) support
153  */
154 const uint8_t key_data[] = {
155 	0x23, 0x64, 0xa1, 0xaa, 0x37, 0xc0, 0xed, 0x05, 0x2b, 0x36,
156 	0x50, 0x5c, 0x45, 0x1e, 0x7e, 0xc8, 0x5d, 0x2a, 0x54, 0x2f,
157 	0xe4, 0x3d, 0x0f, 0xbb, 0x91, 0xd9, 0x25, 0x60, 0xd4, 0xf8,
158 	0x12, 0xa0, 0x59, 0x4b, 0x9e, 0x8a, 0x51, 0xda, 0xcd, 0x49};
159 
160 /*
161  * Shadow Registers:
162  * Outbound queues have a consumer index that is maintained by the chip.
163  * Inbound queues have a producer index that is maintained by the chip.
164  * For lower overhead, these registers are "shadowed" to host memory
165  * which allows the device driver to track the queue progress without
166  * PCI reads. When an entry is placed on an inbound queue, the chip will
167  * update the relevant index register and then copy the value to the
168  * shadow register in host memory.
169  * Currently, ql_read_sh_reg only read Inbound queues'producer index.
170  */
171 
172 static inline unsigned int
173 ql_read_sh_reg(qlge_t *qlge, struct rx_ring *rx_ring)
174 {
175 	uint32_t rtn;
176 
177 	/* re-synchronize shadow prod index dma buffer before reading */
178 	(void) ddi_dma_sync(qlge->host_copy_shadow_dma_attr.dma_handle,
179 	    rx_ring->prod_idx_sh_reg_offset,
180 	    sizeof (uint32_t), DDI_DMA_SYNC_FORKERNEL);
181 
182 	rtn = ddi_get32(qlge->host_copy_shadow_dma_attr.acc_handle,
183 	    (uint32_t *)rx_ring->prod_idx_sh_reg);
184 
185 	return (rtn);
186 }
187 
188 /*
189  * Read 32 bit atomically
190  */
191 uint32_t
192 ql_atomic_read_32(volatile uint32_t *target)
193 {
194 	/*
195 	 * atomic_add_32_nv returns the new value after the add,
196 	 * we are adding 0 so we should get the original value
197 	 */
198 	return (atomic_add_32_nv(target, 0));
199 }
200 
201 /*
202  * Set 32 bit atomically
203  */
204 void
205 ql_atomic_set_32(volatile uint32_t *target, uint32_t newval)
206 {
207 	(void) atomic_swap_32(target, newval);
208 }
209 
210 
211 /*
212  * Setup device PCI configuration registers.
213  * Kernel context.
214  */
215 static void
216 ql_pci_config(qlge_t *qlge)
217 {
218 	uint16_t w;
219 
220 	qlge->vendor_id = (uint16_t)pci_config_get16(qlge->pci_handle,
221 	    PCI_CONF_VENID);
222 	qlge->device_id = (uint16_t)pci_config_get16(qlge->pci_handle,
223 	    PCI_CONF_DEVID);
224 
225 	/*
226 	 * we want to respect framework's setting of PCI
227 	 * configuration space command register and also
228 	 * want to make sure that all bits of interest to us
229 	 * are properly set in PCI Command register(0x04).
230 	 * PCI_COMM_IO		0x1	 I/O access enable
231 	 * PCI_COMM_MAE		0x2	 Memory access enable
232 	 * PCI_COMM_ME		0x4	 bus master enable
233 	 * PCI_COMM_MEMWR_INVAL	0x10	 memory write and invalidate enable.
234 	 */
235 	w = (uint16_t)pci_config_get16(qlge->pci_handle, PCI_CONF_COMM);
236 	w = (uint16_t)(w & (~PCI_COMM_IO));
237 	w = (uint16_t)(w | PCI_COMM_MAE | PCI_COMM_ME |
238 	    /* PCI_COMM_MEMWR_INVAL | */
239 	    PCI_COMM_PARITY_DETECT | PCI_COMM_SERR_ENABLE);
240 
241 	pci_config_put16(qlge->pci_handle, PCI_CONF_COMM, w);
242 
243 	w = pci_config_get16(qlge->pci_handle, 0x54);
244 	w = (uint16_t)(w & (~0x7000));
245 	w = (uint16_t)(w | 0x5000);
246 	pci_config_put16(qlge->pci_handle, 0x54, w);
247 
248 	ql_dump_pci_config(qlge);
249 }
250 
251 /*
252  * This routine parforms the neccessary steps to set GLD mac information
253  * such as Function number, xgmac mask and shift bits
254  */
255 static int
256 ql_set_mac_info(qlge_t *qlge)
257 {
258 	uint32_t value;
259 	int rval = DDI_FAILURE;
260 	uint32_t fn0_net, fn1_net;
261 
262 	/* set default value */
263 	qlge->fn0_net = FN0_NET;
264 	qlge->fn1_net = FN1_NET;
265 
266 	if (ql_read_processor_data(qlge, MPI_REG, &value) != DDI_SUCCESS) {
267 		cmn_err(CE_WARN, "%s(%d) read MPI register failed",
268 		    __func__, qlge->instance);
269 		goto exit;
270 	} else {
271 		fn0_net = (value >> 1) & 0x07;
272 		fn1_net = (value >> 5) & 0x07;
273 		if ((fn0_net > 4) || (fn1_net > 4) || (fn0_net == fn1_net)) {
274 			cmn_err(CE_WARN, "%s(%d) bad mpi register value %x, \n"
275 			    "nic0 function number %d,"
276 			    "nic1 function number %d "
277 			    "use default\n",
278 			    __func__, qlge->instance, value, fn0_net, fn1_net);
279 			goto exit;
280 		} else {
281 			qlge->fn0_net = fn0_net;
282 			qlge->fn1_net = fn1_net;
283 		}
284 	}
285 
286 	/* Get the function number that the driver is associated with */
287 	value = ql_read_reg(qlge, REG_STATUS);
288 	qlge->func_number = (uint8_t)((value >> 6) & 0x03);
289 	QL_PRINT(DBG_INIT, ("status register is:%x, func_number: %d\n",
290 	    value, qlge->func_number));
291 
292 	/* The driver is loaded on a non-NIC function? */
293 	if ((qlge->func_number != qlge->fn0_net) &&
294 	    (qlge->func_number != qlge->fn1_net)) {
295 		cmn_err(CE_WARN,
296 		    "Invalid function number = 0x%x\n", qlge->func_number);
297 		goto exit;
298 	}
299 	/* network port 0? */
300 	if (qlge->func_number == qlge->fn0_net) {
301 		qlge->xgmac_sem_mask = QL_PORT0_XGMAC_SEM_MASK;
302 		qlge->xgmac_sem_bits = QL_PORT0_XGMAC_SEM_BITS;
303 	} else {
304 		qlge->xgmac_sem_mask = QL_PORT1_XGMAC_SEM_MASK;
305 		qlge->xgmac_sem_bits = QL_PORT1_XGMAC_SEM_BITS;
306 	}
307 	rval = DDI_SUCCESS;
308 exit:
309 	return (rval);
310 
311 }
312 
313 /*
314  * write to doorbell register
315  */
316 void
317 ql_write_doorbell_reg(qlge_t *qlge, uint32_t *addr, uint32_t data)
318 {
319 	ddi_put32(qlge->dev_doorbell_reg_handle, addr, data);
320 }
321 
322 /*
323  * read from doorbell register
324  */
325 uint32_t
326 ql_read_doorbell_reg(qlge_t *qlge, uint32_t *addr)
327 {
328 	uint32_t ret;
329 
330 	ret = ddi_get32(qlge->dev_doorbell_reg_handle, addr);
331 
332 	return	(ret);
333 }
334 
335 /*
336  * This function waits for a specific bit to come ready
337  * in a given register.  It is used mostly by the initialize
338  * process, but is also used in kernel thread API such as
339  * netdev->set_multi, netdev->set_mac_address, netdev->vlan_rx_add_vid.
340  */
341 static int
342 ql_wait_reg_rdy(qlge_t *qlge, uint32_t reg, uint32_t bit, uint32_t err_bit)
343 {
344 	uint32_t temp;
345 	int count = UDELAY_COUNT;
346 
347 	while (count) {
348 		temp = ql_read_reg(qlge, reg);
349 
350 		/* check for errors */
351 		if ((temp & err_bit) != 0) {
352 			break;
353 		} else if ((temp & bit) != 0)
354 			return (DDI_SUCCESS);
355 		qlge_delay(UDELAY_DELAY);
356 		count--;
357 	}
358 	cmn_err(CE_WARN,
359 	    "Waiting for reg %x to come ready failed.", reg);
360 	if (qlge->fm_enable) {
361 		ql_fm_ereport(qlge, DDI_FM_DEVICE_NO_RESPONSE);
362 		atomic_or_32(&qlge->flags, ADAPTER_ERROR);
363 	}
364 	return (DDI_FAILURE);
365 }
366 
367 /*
368  * The CFG register is used to download TX and RX control blocks
369  * to the chip. This function waits for an operation to complete.
370  */
371 static int
372 ql_wait_cfg(qlge_t *qlge, uint32_t bit)
373 {
374 	return (ql_wait_reg_bit(qlge, REG_CONFIGURATION, bit, BIT_RESET, 0));
375 }
376 
377 
378 /*
379  * Used to issue init control blocks to hw. Maps control block,
380  * sets address, triggers download, waits for completion.
381  */
382 static int
383 ql_write_cfg(qlge_t *qlge, uint32_t bit, uint64_t phy_addr, uint16_t q_id)
384 {
385 	int status = DDI_SUCCESS;
386 	uint32_t mask;
387 	uint32_t value;
388 
389 	status = ql_sem_spinlock(qlge, SEM_ICB_MASK);
390 	if (status != DDI_SUCCESS) {
391 		goto exit;
392 	}
393 	status = ql_wait_cfg(qlge, bit);
394 	if (status != DDI_SUCCESS) {
395 		goto exit;
396 	}
397 
398 	ql_write_reg(qlge, REG_ICB_ACCESS_ADDRESS_LOWER, LS_64BITS(phy_addr));
399 	ql_write_reg(qlge, REG_ICB_ACCESS_ADDRESS_UPPER, MS_64BITS(phy_addr));
400 
401 	mask = CFG_Q_MASK | (bit << 16);
402 	value = bit | (q_id << CFG_Q_SHIFT);
403 	ql_write_reg(qlge, REG_CONFIGURATION, (mask | value));
404 
405 	/*
406 	 * Wait for the bit to clear after signaling hw.
407 	 */
408 	status = ql_wait_cfg(qlge, bit);
409 	ql_sem_unlock(qlge, SEM_ICB_MASK); /* does flush too */
410 
411 exit:
412 	return (status);
413 }
414 
415 /*
416  * Initialize adapter instance
417  */
418 static int
419 ql_init_instance(qlge_t *qlge)
420 {
421 	int i;
422 
423 	/* Default value */
424 	qlge->mac_flags = QL_MAC_INIT;
425 	qlge->mtu = ETHERMTU;		/* set normal size as default */
426 	qlge->page_size = VM_PAGE_SIZE;	/* default page size */
427 
428 	for (i = 0; i < MAX_RX_RINGS; i++) {
429 		qlge->rx_polls[i] = 0;
430 		qlge->rx_interrupts[i] = 0;
431 	}
432 
433 	/*
434 	 * Set up the operating parameters.
435 	 */
436 	qlge->multicast_list_count = 0;
437 
438 	/*
439 	 * Set up the max number of unicast list
440 	 */
441 	qlge->unicst_total = MAX_UNICAST_LIST_SIZE;
442 	qlge->unicst_avail = MAX_UNICAST_LIST_SIZE;
443 
444 	/*
445 	 * read user defined properties in .conf file
446 	 */
447 	ql_read_conf(qlge); /* mtu, pause, LSO etc */
448 	qlge->rx_ring_count = qlge->tx_ring_count + qlge->rss_ring_count;
449 
450 	QL_PRINT(DBG_INIT, ("mtu is %d \n", qlge->mtu));
451 
452 	/* choose Memory Space mapping and get Vendor Id, Device ID etc */
453 	ql_pci_config(qlge);
454 	qlge->ip_hdr_offset = 0;
455 
456 	if (qlge->device_id == 0x8000) {
457 		/* Schultz card */
458 		qlge->cfg_flags |= CFG_CHIP_8100;
459 		/* enable just ipv4 chksum offload for Schultz */
460 		qlge->cfg_flags |= CFG_CKSUM_FULL_IPv4;
461 		/*
462 		 * Schultz firmware does not do pseduo IP header checksum
463 		 * calculation, needed to be done by driver
464 		 */
465 		qlge->cfg_flags |= CFG_HW_UNABLE_PSEUDO_HDR_CKSUM;
466 		if (qlge->lso_enable)
467 			qlge->cfg_flags |= CFG_LSO;
468 		qlge->cfg_flags |= CFG_SUPPORT_SCATTER_GATHER;
469 		/* Schultz must split packet header */
470 		qlge->cfg_flags |= CFG_ENABLE_SPLIT_HEADER;
471 		qlge->max_read_mbx = 5;
472 		qlge->ip_hdr_offset = 2;
473 	}
474 
475 	/* Set Function Number and some of the iocb mac information */
476 	if (ql_set_mac_info(qlge) != DDI_SUCCESS)
477 		return (DDI_FAILURE);
478 
479 	/* Read network settings from NVRAM */
480 	/* After nvram is read successfully, update dev_addr */
481 	if (ql_get_flash_params(qlge) == DDI_SUCCESS) {
482 		QL_PRINT(DBG_INIT, ("mac%d address is \n", qlge->func_number));
483 		for (i = 0; i < ETHERADDRL; i++) {
484 			qlge->dev_addr.ether_addr_octet[i] =
485 			    qlge->nic_config.factory_MAC[i];
486 		}
487 	} else {
488 		cmn_err(CE_WARN, "%s(%d): Failed to read flash memory",
489 		    __func__, qlge->instance);
490 		return (DDI_FAILURE);
491 	}
492 
493 	bcopy(qlge->dev_addr.ether_addr_octet,
494 	    qlge->unicst_addr[0].addr.ether_addr_octet,
495 	    ETHERADDRL);
496 	QL_DUMP(DBG_INIT, "\t flash mac address dump:\n",
497 	    &qlge->dev_addr.ether_addr_octet[0], 8, ETHERADDRL);
498 
499 	qlge->port_link_state = LS_DOWN;
500 
501 	return (DDI_SUCCESS);
502 }
503 
504 
505 /*
506  * This hardware semaphore provides the mechanism for exclusive access to
507  * resources shared between the NIC driver, MPI firmware,
508  * FCOE firmware and the FC driver.
509  */
510 static int
511 ql_sem_trylock(qlge_t *qlge, uint32_t sem_mask)
512 {
513 	uint32_t sem_bits = 0;
514 
515 	switch (sem_mask) {
516 	case SEM_XGMAC0_MASK:
517 		sem_bits = SEM_SET << SEM_XGMAC0_SHIFT;
518 		break;
519 	case SEM_XGMAC1_MASK:
520 		sem_bits = SEM_SET << SEM_XGMAC1_SHIFT;
521 		break;
522 	case SEM_ICB_MASK:
523 		sem_bits = SEM_SET << SEM_ICB_SHIFT;
524 		break;
525 	case SEM_MAC_ADDR_MASK:
526 		sem_bits = SEM_SET << SEM_MAC_ADDR_SHIFT;
527 		break;
528 	case SEM_FLASH_MASK:
529 		sem_bits = SEM_SET << SEM_FLASH_SHIFT;
530 		break;
531 	case SEM_PROBE_MASK:
532 		sem_bits = SEM_SET << SEM_PROBE_SHIFT;
533 		break;
534 	case SEM_RT_IDX_MASK:
535 		sem_bits = SEM_SET << SEM_RT_IDX_SHIFT;
536 		break;
537 	case SEM_PROC_REG_MASK:
538 		sem_bits = SEM_SET << SEM_PROC_REG_SHIFT;
539 		break;
540 	default:
541 		cmn_err(CE_WARN, "Bad Semaphore mask!.");
542 		return (DDI_FAILURE);
543 	}
544 
545 	ql_write_reg(qlge, REG_SEMAPHORE, sem_bits | sem_mask);
546 	return (!(ql_read_reg(qlge, REG_SEMAPHORE) & sem_bits));
547 }
548 
549 /*
550  * Lock a specific bit of Semaphore register to gain
551  * access to a particular shared register
552  */
553 int
554 ql_sem_spinlock(qlge_t *qlge, uint32_t sem_mask)
555 {
556 	unsigned int wait_count = 30;
557 
558 	while (wait_count) {
559 		if (!ql_sem_trylock(qlge, sem_mask))
560 			return (DDI_SUCCESS);
561 		qlge_delay(100);
562 		wait_count--;
563 	}
564 	cmn_err(CE_WARN, "%s(%d) sem_mask 0x%x lock timeout ",
565 	    __func__, qlge->instance, sem_mask);
566 	return (DDI_FAILURE);
567 }
568 
569 /*
570  * Unock a specific bit of Semaphore register to release
571  * access to a particular shared register
572  */
573 void
574 ql_sem_unlock(qlge_t *qlge, uint32_t sem_mask)
575 {
576 	ql_write_reg(qlge, REG_SEMAPHORE, sem_mask);
577 	(void) ql_read_reg(qlge, REG_SEMAPHORE);	/* flush */
578 }
579 
580 /*
581  * Get property value from configuration file.
582  *
583  * string = property string pointer.
584  *
585  * Returns:
586  * 0xFFFFFFFF = no property else property value.
587  */
588 static uint32_t
589 ql_get_prop(qlge_t *qlge, char *string)
590 {
591 	char buf[256];
592 	uint32_t data;
593 
594 	/* Get adapter instance parameter. */
595 	(void) sprintf(buf, "hba%d-%s", qlge->instance, string);
596 	data = (uint32_t)ddi_prop_get_int(DDI_DEV_T_ANY, qlge->dip, 0, buf,
597 	    (int)0xffffffff);
598 
599 	/* Adapter instance parameter found? */
600 	if (data == 0xffffffff) {
601 		/* No, get default parameter. */
602 		data = (uint32_t)ddi_prop_get_int(DDI_DEV_T_ANY, qlge->dip, 0,
603 		    string, (int)0xffffffff);
604 	}
605 
606 	return (data);
607 }
608 
609 /*
610  * Read user setting from configuration file.
611  */
612 static void
613 ql_read_conf(qlge_t *qlge)
614 {
615 	uint32_t data;
616 
617 	/* clear configuration flags */
618 	qlge->cfg_flags = 0;
619 
620 	/* Set up the default ring sizes. */
621 	qlge->tx_ring_size = NUM_TX_RING_ENTRIES;
622 	data = ql_get_prop(qlge, "tx_ring_size");
623 	/* if data is valid */
624 	if ((data != 0xffffffff) && data) {
625 		if (qlge->tx_ring_size != data) {
626 			qlge->tx_ring_size = (uint16_t)data;
627 		}
628 	}
629 
630 	qlge->rx_ring_size = NUM_RX_RING_ENTRIES;
631 	data = ql_get_prop(qlge, "rx_ring_size");
632 	/* if data is valid */
633 	if ((data != 0xffffffff) && data) {
634 		if (qlge->rx_ring_size != data) {
635 			qlge->rx_ring_size = (uint16_t)data;
636 		}
637 	}
638 
639 	qlge->tx_ring_count = 8;
640 	data = ql_get_prop(qlge, "tx_ring_count");
641 	/* if data is valid */
642 	if ((data != 0xffffffff) && data) {
643 		if (qlge->tx_ring_count != data) {
644 			qlge->tx_ring_count = (uint16_t)data;
645 		}
646 	}
647 
648 	qlge->rss_ring_count = 8;
649 	data = ql_get_prop(qlge, "rss_ring_count");
650 	/* if data is valid */
651 	if ((data != 0xffffffff) && data) {
652 		if (qlge->rss_ring_count != data) {
653 			qlge->rss_ring_count = (uint16_t)data;
654 		}
655 	}
656 
657 	/* Get default rx_copy enable/disable. */
658 	if ((data = ql_get_prop(qlge, "force-rx-copy")) == 0xffffffff ||
659 	    data == 0) {
660 		qlge->rx_copy = B_FALSE;
661 		QL_PRINT(DBG_INIT, ("rx copy mode disabled\n"));
662 	} else if (data == 1) {
663 		qlge->rx_copy = B_TRUE;
664 		QL_PRINT(DBG_INIT, ("rx copy mode enabled\n"));
665 	}
666 
667 	qlge->rx_copy_threshold = qlge->rx_ring_size / 4;
668 	data = ql_get_prop(qlge, "rx_copy_threshold");
669 	if ((data != 0xffffffff) && (data != 0)) {
670 		qlge->rx_copy_threshold = data;
671 		cmn_err(CE_NOTE, "!new rx_copy_threshold %d \n",
672 		    qlge->rx_copy_threshold);
673 	}
674 
675 	/* Get mtu packet size. */
676 	data = ql_get_prop(qlge, "mtu");
677 	if ((data == ETHERMTU) || (data == JUMBO_MTU)) {
678 		if (qlge->mtu != data) {
679 			qlge->mtu = data;
680 			cmn_err(CE_NOTE, "new mtu is %d\n", qlge->mtu);
681 		}
682 	}
683 
684 	if (qlge->mtu == JUMBO_MTU) {
685 		qlge->rx_coalesce_usecs = DFLT_RX_COALESCE_WAIT_JUMBO;
686 		qlge->tx_coalesce_usecs = DFLT_TX_COALESCE_WAIT_JUMBO;
687 		qlge->rx_max_coalesced_frames = DFLT_RX_INTER_FRAME_WAIT_JUMBO;
688 		qlge->tx_max_coalesced_frames = DFLT_TX_INTER_FRAME_WAIT_JUMBO;
689 	}
690 
691 
692 	/* Get pause mode, default is Per Priority mode. */
693 	qlge->pause = PAUSE_MODE_PER_PRIORITY;
694 	data = ql_get_prop(qlge, "pause");
695 	if (data <= PAUSE_MODE_PER_PRIORITY) {
696 		if (qlge->pause != data) {
697 			qlge->pause = data;
698 			cmn_err(CE_NOTE, "new pause mode %d\n", qlge->pause);
699 		}
700 	}
701 	/* Receive interrupt delay */
702 	qlge->rx_coalesce_usecs = DFLT_RX_COALESCE_WAIT;
703 	data = ql_get_prop(qlge, "rx_intr_delay");
704 	/* if data is valid */
705 	if ((data != 0xffffffff) && data) {
706 		if (qlge->rx_coalesce_usecs != data) {
707 			qlge->rx_coalesce_usecs = (uint16_t)data;
708 		}
709 	}
710 	/* Rx inter-packet delay. */
711 	qlge->rx_max_coalesced_frames = DFLT_RX_INTER_FRAME_WAIT;
712 	data = ql_get_prop(qlge, "rx_ipkt_delay");
713 	/* if data is valid */
714 	if ((data != 0xffffffff) && data) {
715 		if (qlge->rx_max_coalesced_frames != data) {
716 			qlge->rx_max_coalesced_frames = (uint16_t)data;
717 		}
718 	}
719 	/* Transmit interrupt delay */
720 	qlge->tx_coalesce_usecs = DFLT_TX_COALESCE_WAIT;
721 	data = ql_get_prop(qlge, "tx_intr_delay");
722 	/* if data is valid */
723 	if ((data != 0xffffffff) && data) {
724 		if (qlge->tx_coalesce_usecs != data) {
725 			qlge->tx_coalesce_usecs = (uint16_t)data;
726 		}
727 	}
728 	/* Tx inter-packet delay. */
729 	qlge->tx_max_coalesced_frames = DFLT_TX_INTER_FRAME_WAIT;
730 	data = ql_get_prop(qlge, "tx_ipkt_delay");
731 	/* if data is valid */
732 	if ((data != 0xffffffff) && data) {
733 		if (qlge->tx_max_coalesced_frames != data) {
734 			qlge->tx_max_coalesced_frames = (uint16_t)data;
735 		}
736 	}
737 
738 	/* Get split header payload_copy_thresh. */
739 	qlge->payload_copy_thresh = DFLT_PAYLOAD_COPY_THRESH;
740 	data = ql_get_prop(qlge, "payload_copy_thresh");
741 	/* if data is valid */
742 	if ((data != 0xffffffff) && (data != 0)) {
743 		if (qlge->payload_copy_thresh != data) {
744 			qlge->payload_copy_thresh = data;
745 		}
746 	}
747 
748 	/* large send offload (LSO) capability. */
749 	qlge->lso_enable = 1;
750 	data = ql_get_prop(qlge, "lso_enable");
751 	/* if data is valid */
752 	if ((data == 0) || (data == 1)) {
753 		if (qlge->lso_enable != data) {
754 			qlge->lso_enable = (uint16_t)data;
755 		}
756 	}
757 
758 	/* dcbx capability. */
759 	qlge->dcbx_enable = 1;
760 	data = ql_get_prop(qlge, "dcbx_enable");
761 	/* if data is valid */
762 	if ((data == 0) || (data == 1)) {
763 		if (qlge->dcbx_enable != data) {
764 			qlge->dcbx_enable = (uint16_t)data;
765 		}
766 	}
767 	/* fault management enable */
768 	qlge->fm_enable = B_TRUE;
769 	data = ql_get_prop(qlge, "fm-enable");
770 	if ((data == 0x1) || (data == 0)) {
771 		qlge->fm_enable = (boolean_t)data;
772 	}
773 
774 }
775 
776 /*
777  * Enable global interrupt
778  */
779 static void
780 ql_enable_global_interrupt(qlge_t *qlge)
781 {
782 	ql_write_reg(qlge, REG_INTERRUPT_ENABLE,
783 	    (INTR_EN_EI << 16) | INTR_EN_EI);
784 	qlge->flags |= INTERRUPTS_ENABLED;
785 }
786 
787 /*
788  * Disable global interrupt
789  */
790 static void
791 ql_disable_global_interrupt(qlge_t *qlge)
792 {
793 	ql_write_reg(qlge, REG_INTERRUPT_ENABLE, (INTR_EN_EI << 16));
794 	qlge->flags &= ~INTERRUPTS_ENABLED;
795 }
796 
797 /*
798  * Enable one ring interrupt
799  */
800 void
801 ql_enable_completion_interrupt(qlge_t *qlge, uint32_t intr)
802 {
803 	struct intr_ctx *ctx = qlge->intr_ctx + intr;
804 
805 	QL_PRINT(DBG_INTR, ("%s(%d): To enable intr %d, irq_cnt %d \n",
806 	    __func__, qlge->instance, intr, ctx->irq_cnt));
807 
808 	if ((qlge->intr_type == DDI_INTR_TYPE_MSIX) && intr) {
809 		/*
810 		 * Always enable if we're MSIX multi interrupts and
811 		 * it's not the default (zeroeth) interrupt.
812 		 */
813 		ql_write_reg(qlge, REG_INTERRUPT_ENABLE, ctx->intr_en_mask);
814 		return;
815 	}
816 
817 	if (!atomic_dec_32_nv(&ctx->irq_cnt)) {
818 		mutex_enter(&qlge->hw_mutex);
819 		ql_write_reg(qlge, REG_INTERRUPT_ENABLE, ctx->intr_en_mask);
820 		mutex_exit(&qlge->hw_mutex);
821 		QL_PRINT(DBG_INTR,
822 		    ("%s(%d): write %x to intr enable register \n",
823 		    __func__, qlge->instance, ctx->intr_en_mask));
824 	}
825 }
826 
827 /*
828  * ql_forced_disable_completion_interrupt
829  * Used by call from OS, may be called without
830  * a pending interrupt so force the disable
831  */
832 uint32_t
833 ql_forced_disable_completion_interrupt(qlge_t *qlge, uint32_t intr)
834 {
835 	uint32_t var = 0;
836 	struct intr_ctx *ctx = qlge->intr_ctx + intr;
837 
838 	QL_PRINT(DBG_INTR, ("%s(%d): To disable intr %d, irq_cnt %d \n",
839 	    __func__, qlge->instance, intr, ctx->irq_cnt));
840 
841 	if ((qlge->intr_type == DDI_INTR_TYPE_MSIX) && intr) {
842 		ql_write_reg(qlge, REG_INTERRUPT_ENABLE, ctx->intr_dis_mask);
843 		var = ql_read_reg(qlge, REG_STATUS);
844 		return (var);
845 	}
846 
847 	mutex_enter(&qlge->hw_mutex);
848 	ql_write_reg(qlge, REG_INTERRUPT_ENABLE, ctx->intr_dis_mask);
849 	var = ql_read_reg(qlge, REG_STATUS);
850 	mutex_exit(&qlge->hw_mutex);
851 
852 	return (var);
853 }
854 
855 /*
856  * Disable a completion interrupt
857  */
858 void
859 ql_disable_completion_interrupt(qlge_t *qlge, uint32_t intr)
860 {
861 	struct intr_ctx *ctx;
862 
863 	ctx = qlge->intr_ctx + intr;
864 	QL_PRINT(DBG_INTR, ("%s(%d): To disable intr %d, irq_cnt %d \n",
865 	    __func__, qlge->instance, intr, ctx->irq_cnt));
866 	/*
867 	 * HW disables for us if we're MSIX multi interrupts and
868 	 * it's not the default (zeroeth) interrupt.
869 	 */
870 	if ((qlge->intr_type == DDI_INTR_TYPE_MSIX) && (intr != 0))
871 		return;
872 
873 	if (ql_atomic_read_32(&ctx->irq_cnt) == 0) {
874 		mutex_enter(&qlge->hw_mutex);
875 		ql_write_reg(qlge, REG_INTERRUPT_ENABLE, ctx->intr_dis_mask);
876 		mutex_exit(&qlge->hw_mutex);
877 	}
878 	atomic_inc_32(&ctx->irq_cnt);
879 }
880 
881 /*
882  * Enable all completion interrupts
883  */
884 static void
885 ql_enable_all_completion_interrupts(qlge_t *qlge)
886 {
887 	int i;
888 	uint32_t value = 1;
889 
890 	for (i = 0; i < qlge->intr_cnt; i++) {
891 		/*
892 		 * Set the count to 1 for Legacy / MSI interrupts or for the
893 		 * default interrupt (0)
894 		 */
895 		if ((qlge->intr_type != DDI_INTR_TYPE_MSIX) || i == 0) {
896 			ql_atomic_set_32(&qlge->intr_ctx[i].irq_cnt, value);
897 		}
898 		ql_enable_completion_interrupt(qlge, i);
899 	}
900 }
901 
902 /*
903  * Disable all completion interrupts
904  */
905 static void
906 ql_disable_all_completion_interrupts(qlge_t *qlge)
907 {
908 	int i;
909 	uint32_t value = 0;
910 
911 	for (i = 0; i < qlge->intr_cnt; i++) {
912 
913 		/*
914 		 * Set the count to 0 for Legacy / MSI interrupts or for the
915 		 * default interrupt (0)
916 		 */
917 		if ((qlge->intr_type != DDI_INTR_TYPE_MSIX) || i == 0)
918 			ql_atomic_set_32(&qlge->intr_ctx[i].irq_cnt, value);
919 
920 		ql_disable_completion_interrupt(qlge, i);
921 	}
922 }
923 
924 /*
925  * Update small buffer queue producer index
926  */
927 static void
928 ql_update_sbq_prod_idx(qlge_t *qlge, struct rx_ring *rx_ring)
929 {
930 	/* Update the buffer producer index */
931 	QL_PRINT(DBG_RX, ("sbq: updating prod idx = %d.\n",
932 	    rx_ring->sbq_prod_idx));
933 	ql_write_doorbell_reg(qlge, rx_ring->sbq_prod_idx_db_reg,
934 	    rx_ring->sbq_prod_idx);
935 }
936 
937 /*
938  * Update large buffer queue producer index
939  */
940 static void
941 ql_update_lbq_prod_idx(qlge_t *qlge, struct rx_ring *rx_ring)
942 {
943 	/* Update the buffer producer index */
944 	QL_PRINT(DBG_RX, ("lbq: updating prod idx = %d.\n",
945 	    rx_ring->lbq_prod_idx));
946 	ql_write_doorbell_reg(qlge, rx_ring->lbq_prod_idx_db_reg,
947 	    rx_ring->lbq_prod_idx);
948 }
949 
950 /*
951  * Adds a small buffer descriptor to end of its in use list,
952  * assumes sbq_lock is already taken
953  */
954 static void
955 ql_add_sbuf_to_in_use_list(struct rx_ring *rx_ring,
956     struct bq_desc *sbq_desc)
957 {
958 	uint32_t inuse_idx = rx_ring->sbq_use_tail;
959 
960 	rx_ring->sbuf_in_use[inuse_idx] = sbq_desc;
961 	inuse_idx++;
962 	if (inuse_idx >= rx_ring->sbq_len)
963 		inuse_idx = 0;
964 	rx_ring->sbq_use_tail = inuse_idx;
965 	atomic_inc_32(&rx_ring->sbuf_in_use_count);
966 	ASSERT(rx_ring->sbuf_in_use_count <= rx_ring->sbq_len);
967 }
968 
969 /*
970  * Get a small buffer descriptor from its in use list
971  */
972 static struct bq_desc *
973 ql_get_sbuf_from_in_use_list(struct rx_ring *rx_ring)
974 {
975 	struct bq_desc *sbq_desc = NULL;
976 	uint32_t inuse_idx;
977 
978 	/* Pick from head of in use list */
979 	inuse_idx = rx_ring->sbq_use_head;
980 	sbq_desc = rx_ring->sbuf_in_use[inuse_idx];
981 	rx_ring->sbuf_in_use[inuse_idx] = NULL;
982 
983 	if (sbq_desc != NULL) {
984 		inuse_idx++;
985 		if (inuse_idx >= rx_ring->sbq_len)
986 			inuse_idx = 0;
987 		rx_ring->sbq_use_head = inuse_idx;
988 		atomic_dec_32(&rx_ring->sbuf_in_use_count);
989 		atomic_inc_32(&rx_ring->rx_indicate);
990 		sbq_desc->upl_inuse = 1;
991 		/* if mp is NULL */
992 		if (sbq_desc->mp == NULL) {
993 			/* try to remap mp again */
994 			sbq_desc->mp =
995 			    desballoc((unsigned char *)(sbq_desc->bd_dma.vaddr),
996 			    rx_ring->sbq_buf_size, 0, &sbq_desc->rx_recycle);
997 		}
998 	}
999 
1000 	return (sbq_desc);
1001 }
1002 
1003 /*
1004  * Add a small buffer descriptor to its free list
1005  */
1006 static void
1007 ql_add_sbuf_to_free_list(struct rx_ring *rx_ring,
1008     struct bq_desc *sbq_desc)
1009 {
1010 	uint32_t free_idx;
1011 
1012 	/* Add to the end of free list */
1013 	free_idx = rx_ring->sbq_free_tail;
1014 	rx_ring->sbuf_free[free_idx] = sbq_desc;
1015 	ASSERT(rx_ring->sbuf_free_count <= rx_ring->sbq_len);
1016 	free_idx++;
1017 	if (free_idx >= rx_ring->sbq_len)
1018 		free_idx = 0;
1019 	rx_ring->sbq_free_tail = free_idx;
1020 	atomic_inc_32(&rx_ring->sbuf_free_count);
1021 }
1022 
1023 /*
1024  * Get a small buffer descriptor from its free list
1025  */
1026 static struct bq_desc *
1027 ql_get_sbuf_from_free_list(struct rx_ring *rx_ring)
1028 {
1029 	struct bq_desc *sbq_desc;
1030 	uint32_t free_idx;
1031 
1032 	free_idx = rx_ring->sbq_free_head;
1033 	/* Pick from top of free list */
1034 	sbq_desc = rx_ring->sbuf_free[free_idx];
1035 	rx_ring->sbuf_free[free_idx] = NULL;
1036 	if (sbq_desc != NULL) {
1037 		free_idx++;
1038 		if (free_idx >= rx_ring->sbq_len)
1039 			free_idx = 0;
1040 		rx_ring->sbq_free_head = free_idx;
1041 		atomic_dec_32(&rx_ring->sbuf_free_count);
1042 	}
1043 	return (sbq_desc);
1044 }
1045 
1046 /*
1047  * Add a large buffer descriptor to its in use list
1048  */
1049 static void
1050 ql_add_lbuf_to_in_use_list(struct rx_ring *rx_ring,
1051     struct bq_desc *lbq_desc)
1052 {
1053 	uint32_t inuse_idx;
1054 
1055 	inuse_idx = rx_ring->lbq_use_tail;
1056 
1057 	rx_ring->lbuf_in_use[inuse_idx] = lbq_desc;
1058 	inuse_idx++;
1059 	if (inuse_idx >= rx_ring->lbq_len)
1060 		inuse_idx = 0;
1061 	rx_ring->lbq_use_tail = inuse_idx;
1062 	atomic_inc_32(&rx_ring->lbuf_in_use_count);
1063 }
1064 
1065 /*
1066  * Get a large buffer descriptor from in use list
1067  */
1068 static struct bq_desc *
1069 ql_get_lbuf_from_in_use_list(struct rx_ring *rx_ring)
1070 {
1071 	struct bq_desc *lbq_desc;
1072 	uint32_t inuse_idx;
1073 
1074 	/* Pick from head of in use list */
1075 	inuse_idx = rx_ring->lbq_use_head;
1076 	lbq_desc = rx_ring->lbuf_in_use[inuse_idx];
1077 	rx_ring->lbuf_in_use[inuse_idx] = NULL;
1078 
1079 	if (lbq_desc != NULL) {
1080 		inuse_idx++;
1081 		if (inuse_idx >= rx_ring->lbq_len)
1082 			inuse_idx = 0;
1083 		rx_ring->lbq_use_head = inuse_idx;
1084 		atomic_dec_32(&rx_ring->lbuf_in_use_count);
1085 		atomic_inc_32(&rx_ring->rx_indicate);
1086 		lbq_desc->upl_inuse = 1;
1087 
1088 		/* if mp is NULL */
1089 		if (lbq_desc->mp == NULL) {
1090 			/* try to remap mp again */
1091 			lbq_desc->mp =
1092 			    desballoc((unsigned char *)(lbq_desc->bd_dma.vaddr),
1093 			    rx_ring->lbq_buf_size, 0, &lbq_desc->rx_recycle);
1094 		}
1095 	}
1096 	return (lbq_desc);
1097 }
1098 
1099 /*
1100  * Add a large buffer descriptor to free list
1101  */
1102 static void
1103 ql_add_lbuf_to_free_list(struct rx_ring *rx_ring,
1104     struct bq_desc *lbq_desc)
1105 {
1106 	uint32_t free_idx;
1107 
1108 	/* Add to the end of free list */
1109 	free_idx = rx_ring->lbq_free_tail;
1110 	rx_ring->lbuf_free[free_idx] = lbq_desc;
1111 	free_idx++;
1112 	if (free_idx >= rx_ring->lbq_len)
1113 		free_idx = 0;
1114 	rx_ring->lbq_free_tail = free_idx;
1115 	atomic_inc_32(&rx_ring->lbuf_free_count);
1116 	ASSERT(rx_ring->lbuf_free_count <= rx_ring->lbq_len);
1117 }
1118 
1119 /*
1120  * Get a large buffer descriptor from its free list
1121  */
1122 static struct bq_desc *
1123 ql_get_lbuf_from_free_list(struct rx_ring *rx_ring)
1124 {
1125 	struct bq_desc *lbq_desc;
1126 	uint32_t free_idx;
1127 
1128 	free_idx = rx_ring->lbq_free_head;
1129 	/* Pick from head of free list */
1130 	lbq_desc = rx_ring->lbuf_free[free_idx];
1131 	rx_ring->lbuf_free[free_idx] = NULL;
1132 
1133 	if (lbq_desc != NULL) {
1134 		free_idx++;
1135 		if (free_idx >= rx_ring->lbq_len)
1136 			free_idx = 0;
1137 		rx_ring->lbq_free_head = free_idx;
1138 		atomic_dec_32(&rx_ring->lbuf_free_count);
1139 	}
1140 	return (lbq_desc);
1141 }
1142 
1143 /*
1144  * Add a small buffer descriptor to free list
1145  */
1146 static void
1147 ql_refill_sbuf_free_list(struct bq_desc *sbq_desc, boolean_t alloc_memory)
1148 {
1149 	struct rx_ring *rx_ring = sbq_desc->rx_ring;
1150 	uint64_t *sbq_entry;
1151 	qlge_t *qlge = (qlge_t *)rx_ring->qlge;
1152 	/*
1153 	 * Sync access
1154 	 */
1155 	mutex_enter(&rx_ring->sbq_lock);
1156 
1157 	sbq_desc->upl_inuse = 0;
1158 
1159 	/*
1160 	 * If we are freeing the buffers as a result of adapter unload, get out
1161 	 */
1162 	if ((sbq_desc->free_buf != 0) ||
1163 	    (qlge->mac_flags == QL_MAC_DETACH)) {
1164 		if (sbq_desc->free_buf == 0)
1165 			atomic_dec_32(&rx_ring->rx_indicate);
1166 		mutex_exit(&rx_ring->sbq_lock);
1167 		return;
1168 	}
1169 #ifdef QLGE_LOAD_UNLOAD
1170 	if (rx_ring->rx_indicate == 0)
1171 		cmn_err(CE_WARN, "sbq: indicate wrong");
1172 #endif
1173 #ifdef QLGE_TRACK_BUFFER_USAGE
1174 	uint32_t sb_consumer_idx;
1175 	uint32_t sb_producer_idx;
1176 	uint32_t num_free_buffers;
1177 	uint32_t temp;
1178 
1179 	temp = ql_read_doorbell_reg(qlge, rx_ring->sbq_prod_idx_db_reg);
1180 	sb_producer_idx = temp & 0x0000ffff;
1181 	sb_consumer_idx = (temp >> 16);
1182 
1183 	if (sb_consumer_idx > sb_producer_idx)
1184 		num_free_buffers = NUM_SMALL_BUFFERS -
1185 		    (sb_consumer_idx - sb_producer_idx);
1186 	else
1187 		num_free_buffers = sb_producer_idx - sb_consumer_idx;
1188 
1189 	if (num_free_buffers < qlge->rx_sb_low_count[rx_ring->cq_id])
1190 		qlge->rx_sb_low_count[rx_ring->cq_id] = num_free_buffers;
1191 
1192 #endif
1193 
1194 #ifdef QLGE_LOAD_UNLOAD
1195 	if (rx_ring->rx_indicate > 0xFF000000)
1196 		cmn_err(CE_WARN, "sbq: indicate(%d) wrong: %d mac_flags %d,"
1197 		    " sbq_desc index %d.",
1198 		    rx_ring->cq_id, rx_ring->rx_indicate, rx_ring->mac_flags,
1199 		    sbq_desc->index);
1200 #endif
1201 	if (alloc_memory) {
1202 		sbq_desc->mp =
1203 		    desballoc((unsigned char *)(sbq_desc->bd_dma.vaddr),
1204 		    rx_ring->sbq_buf_size, 0, &sbq_desc->rx_recycle);
1205 		if (sbq_desc->mp == NULL) {
1206 			rx_ring->rx_failed_sbq_allocs++;
1207 		}
1208 	}
1209 
1210 	/* Got the packet from the stack decrement rx_indicate count */
1211 	atomic_dec_32(&rx_ring->rx_indicate);
1212 
1213 	ql_add_sbuf_to_free_list(rx_ring, sbq_desc);
1214 
1215 	/* Rearm if possible */
1216 	if ((rx_ring->sbuf_free_count >= MIN_BUFFERS_FREE_COUNT) &&
1217 	    (qlge->mac_flags == QL_MAC_STARTED)) {
1218 		sbq_entry = rx_ring->sbq_dma.vaddr;
1219 		sbq_entry += rx_ring->sbq_prod_idx;
1220 
1221 		while (rx_ring->sbuf_free_count > MIN_BUFFERS_ARM_COUNT) {
1222 			/* Get first one from free list */
1223 			sbq_desc = ql_get_sbuf_from_free_list(rx_ring);
1224 
1225 			*sbq_entry = cpu_to_le64(sbq_desc->bd_dma.dma_addr);
1226 			sbq_entry++;
1227 			rx_ring->sbq_prod_idx++;
1228 			if (rx_ring->sbq_prod_idx >= rx_ring->sbq_len) {
1229 				rx_ring->sbq_prod_idx = 0;
1230 				sbq_entry = rx_ring->sbq_dma.vaddr;
1231 			}
1232 			/* Add to end of in use list */
1233 			ql_add_sbuf_to_in_use_list(rx_ring, sbq_desc);
1234 		}
1235 
1236 		/* Update small buffer queue producer index */
1237 		ql_update_sbq_prod_idx(qlge, rx_ring);
1238 	}
1239 
1240 	mutex_exit(&rx_ring->sbq_lock);
1241 	QL_PRINT(DBG_RX_RING, ("%s(%d) exited, sbuf_free_count %d\n",
1242 	    __func__, qlge->instance, rx_ring->sbuf_free_count));
1243 }
1244 
1245 /*
1246  * rx recycle call back function
1247  */
1248 static void
1249 ql_release_to_sbuf_free_list(caddr_t p)
1250 {
1251 	struct bq_desc *sbq_desc = (struct bq_desc *)(void *)p;
1252 
1253 	if (sbq_desc == NULL)
1254 		return;
1255 	ql_refill_sbuf_free_list(sbq_desc, B_TRUE);
1256 }
1257 
1258 /*
1259  * Add a large buffer descriptor to free list
1260  */
1261 static void
1262 ql_refill_lbuf_free_list(struct bq_desc *lbq_desc, boolean_t alloc_memory)
1263 {
1264 	struct rx_ring *rx_ring = lbq_desc->rx_ring;
1265 	uint64_t *lbq_entry;
1266 	qlge_t *qlge = rx_ring->qlge;
1267 
1268 	/* Sync access */
1269 	mutex_enter(&rx_ring->lbq_lock);
1270 
1271 	lbq_desc->upl_inuse = 0;
1272 	/*
1273 	 * If we are freeing the buffers as a result of adapter unload, get out
1274 	 */
1275 	if ((lbq_desc->free_buf != 0) ||
1276 	    (qlge->mac_flags == QL_MAC_DETACH)) {
1277 		if (lbq_desc->free_buf == 0)
1278 			atomic_dec_32(&rx_ring->rx_indicate);
1279 		mutex_exit(&rx_ring->lbq_lock);
1280 		return;
1281 	}
1282 #ifdef QLGE_LOAD_UNLOAD
1283 	if (rx_ring->rx_indicate == 0)
1284 		cmn_err(CE_WARN, "lbq: indicate wrong");
1285 #endif
1286 #ifdef QLGE_TRACK_BUFFER_USAGE
1287 	uint32_t lb_consumer_idx;
1288 	uint32_t lb_producer_idx;
1289 	uint32_t num_free_buffers;
1290 	uint32_t temp;
1291 
1292 	temp = ql_read_doorbell_reg(qlge, rx_ring->lbq_prod_idx_db_reg);
1293 
1294 	lb_producer_idx = temp & 0x0000ffff;
1295 	lb_consumer_idx = (temp >> 16);
1296 
1297 	if (lb_consumer_idx > lb_producer_idx)
1298 		num_free_buffers = NUM_LARGE_BUFFERS -
1299 		    (lb_consumer_idx - lb_producer_idx);
1300 	else
1301 		num_free_buffers = lb_producer_idx - lb_consumer_idx;
1302 
1303 	if (num_free_buffers < qlge->rx_lb_low_count[rx_ring->cq_id]) {
1304 		qlge->rx_lb_low_count[rx_ring->cq_id] = num_free_buffers;
1305 	}
1306 #endif
1307 
1308 #ifdef QLGE_LOAD_UNLOAD
1309 	if (rx_ring->rx_indicate > 0xFF000000)
1310 		cmn_err(CE_WARN, "lbq: indicate(%d) wrong: %d mac_flags %d,"
1311 		    "lbq_desc index %d",
1312 		    rx_ring->cq_id, rx_ring->rx_indicate, rx_ring->mac_flags,
1313 		    lbq_desc->index);
1314 #endif
1315 	if (alloc_memory) {
1316 		lbq_desc->mp =
1317 		    desballoc((unsigned char *)(lbq_desc->bd_dma.vaddr),
1318 		    rx_ring->lbq_buf_size, 0, &lbq_desc->rx_recycle);
1319 		if (lbq_desc->mp == NULL) {
1320 			rx_ring->rx_failed_lbq_allocs++;
1321 		}
1322 	}
1323 
1324 	/* Got the packet from the stack decrement rx_indicate count */
1325 	atomic_dec_32(&rx_ring->rx_indicate);
1326 
1327 	ql_add_lbuf_to_free_list(rx_ring, lbq_desc);
1328 
1329 	/* Rearm if possible */
1330 	if ((rx_ring->lbuf_free_count >= MIN_BUFFERS_FREE_COUNT) &&
1331 	    (qlge->mac_flags == QL_MAC_STARTED)) {
1332 		lbq_entry = rx_ring->lbq_dma.vaddr;
1333 		lbq_entry += rx_ring->lbq_prod_idx;
1334 		while (rx_ring->lbuf_free_count > MIN_BUFFERS_ARM_COUNT) {
1335 			/* Get first one from free list */
1336 			lbq_desc = ql_get_lbuf_from_free_list(rx_ring);
1337 
1338 			*lbq_entry = cpu_to_le64(lbq_desc->bd_dma.dma_addr);
1339 			lbq_entry++;
1340 			rx_ring->lbq_prod_idx++;
1341 			if (rx_ring->lbq_prod_idx >= rx_ring->lbq_len) {
1342 				rx_ring->lbq_prod_idx = 0;
1343 				lbq_entry = rx_ring->lbq_dma.vaddr;
1344 			}
1345 
1346 			/* Add to end of in use list */
1347 			ql_add_lbuf_to_in_use_list(rx_ring, lbq_desc);
1348 		}
1349 
1350 		/* Update large buffer queue producer index */
1351 		ql_update_lbq_prod_idx(rx_ring->qlge, rx_ring);
1352 	}
1353 
1354 	mutex_exit(&rx_ring->lbq_lock);
1355 	QL_PRINT(DBG_RX_RING, ("%s exitd, lbuf_free_count %d\n",
1356 	    __func__, rx_ring->lbuf_free_count));
1357 }
1358 /*
1359  * rx recycle call back function
1360  */
1361 static void
1362 ql_release_to_lbuf_free_list(caddr_t p)
1363 {
1364 	struct bq_desc *lbq_desc = (struct bq_desc *)(void *)p;
1365 
1366 	if (lbq_desc == NULL)
1367 		return;
1368 	ql_refill_lbuf_free_list(lbq_desc, B_TRUE);
1369 }
1370 
1371 /*
1372  * free small buffer queue buffers
1373  */
1374 static void
1375 ql_free_sbq_buffers(struct rx_ring *rx_ring)
1376 {
1377 	struct bq_desc *sbq_desc;
1378 	uint32_t i;
1379 	uint32_t j = rx_ring->sbq_free_head;
1380 	int  force_cnt = 0;
1381 
1382 	for (i = 0; i < rx_ring->sbuf_free_count; i++) {
1383 		sbq_desc = rx_ring->sbuf_free[j];
1384 		sbq_desc->free_buf = 1;
1385 		j++;
1386 		if (j >= rx_ring->sbq_len) {
1387 			j = 0;
1388 		}
1389 		if (sbq_desc->mp != NULL) {
1390 			freemsg(sbq_desc->mp);
1391 			sbq_desc->mp = NULL;
1392 		}
1393 	}
1394 	rx_ring->sbuf_free_count = 0;
1395 
1396 	j = rx_ring->sbq_use_head;
1397 	for (i = 0; i < rx_ring->sbuf_in_use_count; i++) {
1398 		sbq_desc = rx_ring->sbuf_in_use[j];
1399 		sbq_desc->free_buf = 1;
1400 		j++;
1401 		if (j >= rx_ring->sbq_len) {
1402 			j = 0;
1403 		}
1404 		if (sbq_desc->mp != NULL) {
1405 			freemsg(sbq_desc->mp);
1406 			sbq_desc->mp = NULL;
1407 		}
1408 	}
1409 	rx_ring->sbuf_in_use_count = 0;
1410 
1411 	sbq_desc = &rx_ring->sbq_desc[0];
1412 	for (i = 0; i < rx_ring->sbq_len; i++, sbq_desc++) {
1413 		/*
1414 		 * Set flag so that the callback does not allocate a new buffer
1415 		 */
1416 		sbq_desc->free_buf = 1;
1417 		if (sbq_desc->upl_inuse != 0) {
1418 			force_cnt++;
1419 		}
1420 		if (sbq_desc->bd_dma.dma_handle != NULL) {
1421 			ql_free_phys(&sbq_desc->bd_dma.dma_handle,
1422 			    &sbq_desc->bd_dma.acc_handle);
1423 			sbq_desc->bd_dma.dma_handle = NULL;
1424 			sbq_desc->bd_dma.acc_handle = NULL;
1425 		}
1426 	}
1427 #ifdef QLGE_LOAD_UNLOAD
1428 	cmn_err(CE_NOTE, "sbq: free %d inuse %d force %d\n",
1429 	    rx_ring->sbuf_free_count, rx_ring->sbuf_in_use_count, force_cnt);
1430 #endif
1431 	if (rx_ring->sbuf_in_use != NULL) {
1432 		kmem_free(rx_ring->sbuf_in_use, (rx_ring->sbq_len *
1433 		    sizeof (struct bq_desc *)));
1434 		rx_ring->sbuf_in_use = NULL;
1435 	}
1436 
1437 	if (rx_ring->sbuf_free != NULL) {
1438 		kmem_free(rx_ring->sbuf_free, (rx_ring->sbq_len *
1439 		    sizeof (struct bq_desc *)));
1440 		rx_ring->sbuf_free = NULL;
1441 	}
1442 }
1443 
1444 /* Allocate small buffers */
1445 static int
1446 ql_alloc_sbufs(qlge_t *qlge, struct rx_ring *rx_ring)
1447 {
1448 	struct bq_desc *sbq_desc;
1449 	int i;
1450 	ddi_dma_cookie_t dma_cookie;
1451 
1452 	rx_ring->sbq_use_head = 0;
1453 	rx_ring->sbq_use_tail = 0;
1454 	rx_ring->sbuf_in_use_count = 0;
1455 	rx_ring->sbq_free_head = 0;
1456 	rx_ring->sbq_free_tail = 0;
1457 	rx_ring->sbuf_free_count = 0;
1458 	rx_ring->sbuf_free = kmem_zalloc(rx_ring->sbq_len *
1459 	    sizeof (struct bq_desc *), KM_NOSLEEP);
1460 	if (rx_ring->sbuf_free == NULL) {
1461 		cmn_err(CE_WARN,
1462 		    "!%s: sbuf_free_list alloc: failed",
1463 		    __func__);
1464 		goto alloc_sbuf_err;
1465 	}
1466 
1467 	rx_ring->sbuf_in_use = kmem_zalloc(rx_ring->sbq_len *
1468 	    sizeof (struct bq_desc *), KM_NOSLEEP);
1469 	if (rx_ring->sbuf_in_use == NULL) {
1470 		cmn_err(CE_WARN,
1471 		    "!%s: sbuf_inuse_list alloc: failed",
1472 		    __func__);
1473 		goto alloc_sbuf_err;
1474 	}
1475 
1476 	sbq_desc = &rx_ring->sbq_desc[0];
1477 
1478 	for (i = 0; i < rx_ring->sbq_len; i++, sbq_desc++) {
1479 		/* Allocate buffer */
1480 		if (ql_alloc_phys_rbuf(qlge->dip, &sbq_desc->bd_dma.dma_handle,
1481 		    &ql_buf_acc_attr,
1482 		    DDI_DMA_READ | DDI_DMA_STREAMING,
1483 		    &sbq_desc->bd_dma.acc_handle,
1484 		    (size_t)rx_ring->sbq_buf_size,	/* mem size */
1485 		    (size_t)0,				/* default alignment */
1486 		    (caddr_t *)&sbq_desc->bd_dma.vaddr,
1487 		    &dma_cookie) != 0) {
1488 			cmn_err(CE_WARN,
1489 			    "!%s: ddi_dma_alloc_handle: failed",
1490 			    __func__);
1491 			goto alloc_sbuf_err;
1492 		}
1493 
1494 		/* Set context for Return buffer callback */
1495 		sbq_desc->bd_dma.dma_addr = dma_cookie.dmac_laddress;
1496 		sbq_desc->rx_recycle.free_func = ql_release_to_sbuf_free_list;
1497 		sbq_desc->rx_recycle.free_arg  = (caddr_t)sbq_desc;
1498 		sbq_desc->rx_ring = rx_ring;
1499 		sbq_desc->upl_inuse = 0;
1500 		sbq_desc->free_buf = 0;
1501 
1502 		sbq_desc->mp =
1503 		    desballoc((unsigned char *)(sbq_desc->bd_dma.vaddr),
1504 		    rx_ring->sbq_buf_size, 0, &sbq_desc->rx_recycle);
1505 		if (sbq_desc->mp == NULL) {
1506 			cmn_err(CE_WARN, "%s: desballoc() failed", __func__);
1507 			goto alloc_sbuf_err;
1508 		}
1509 		ql_add_sbuf_to_free_list(rx_ring, sbq_desc);
1510 	}
1511 
1512 	return (DDI_SUCCESS);
1513 
1514 alloc_sbuf_err:
1515 	ql_free_sbq_buffers(rx_ring);
1516 	return (DDI_FAILURE);
1517 }
1518 
1519 static void
1520 ql_free_lbq_buffers(struct rx_ring *rx_ring)
1521 {
1522 	struct bq_desc *lbq_desc;
1523 	uint32_t i, j;
1524 	int force_cnt = 0;
1525 
1526 	j = rx_ring->lbq_free_head;
1527 	for (i = 0; i < rx_ring->lbuf_free_count; i++) {
1528 		lbq_desc = rx_ring->lbuf_free[j];
1529 		lbq_desc->free_buf = 1;
1530 		j++;
1531 		if (j >= rx_ring->lbq_len)
1532 			j = 0;
1533 		if (lbq_desc->mp != NULL) {
1534 			freemsg(lbq_desc->mp);
1535 			lbq_desc->mp = NULL;
1536 		}
1537 	}
1538 	rx_ring->lbuf_free_count = 0;
1539 
1540 	j = rx_ring->lbq_use_head;
1541 	for (i = 0; i < rx_ring->lbuf_in_use_count; i++) {
1542 		lbq_desc = rx_ring->lbuf_in_use[j];
1543 		lbq_desc->free_buf = 1;
1544 		j++;
1545 		if (j >= rx_ring->lbq_len) {
1546 			j = 0;
1547 		}
1548 		if (lbq_desc->mp != NULL) {
1549 			freemsg(lbq_desc->mp);
1550 			lbq_desc->mp = NULL;
1551 		}
1552 	}
1553 	rx_ring->lbuf_in_use_count = 0;
1554 
1555 	lbq_desc = &rx_ring->lbq_desc[0];
1556 	for (i = 0; i < rx_ring->lbq_len; i++, lbq_desc++) {
1557 		/* Set flag so that callback will not allocate a new buffer */
1558 		lbq_desc->free_buf = 1;
1559 		if (lbq_desc->upl_inuse != 0) {
1560 			force_cnt++;
1561 		}
1562 		if (lbq_desc->bd_dma.dma_handle != NULL) {
1563 			ql_free_phys(&lbq_desc->bd_dma.dma_handle,
1564 			    &lbq_desc->bd_dma.acc_handle);
1565 			lbq_desc->bd_dma.dma_handle = NULL;
1566 			lbq_desc->bd_dma.acc_handle = NULL;
1567 		}
1568 	}
1569 #ifdef QLGE_LOAD_UNLOAD
1570 	if (force_cnt) {
1571 		cmn_err(CE_WARN, "lbq: free %d inuse %d force %d",
1572 		    rx_ring->lbuf_free_count, rx_ring->lbuf_in_use_count,
1573 		    force_cnt);
1574 	}
1575 #endif
1576 	if (rx_ring->lbuf_in_use != NULL) {
1577 		kmem_free(rx_ring->lbuf_in_use, (rx_ring->lbq_len *
1578 		    sizeof (struct bq_desc *)));
1579 		rx_ring->lbuf_in_use = NULL;
1580 	}
1581 
1582 	if (rx_ring->lbuf_free != NULL) {
1583 		kmem_free(rx_ring->lbuf_free, (rx_ring->lbq_len *
1584 		    sizeof (struct bq_desc *)));
1585 		rx_ring->lbuf_free = NULL;
1586 	}
1587 }
1588 
1589 /* Allocate large buffers */
1590 static int
1591 ql_alloc_lbufs(qlge_t *qlge, struct rx_ring *rx_ring)
1592 {
1593 	struct bq_desc *lbq_desc;
1594 	ddi_dma_cookie_t dma_cookie;
1595 	int i;
1596 	uint32_t lbq_buf_size;
1597 
1598 	rx_ring->lbq_use_head = 0;
1599 	rx_ring->lbq_use_tail = 0;
1600 	rx_ring->lbuf_in_use_count = 0;
1601 	rx_ring->lbq_free_head = 0;
1602 	rx_ring->lbq_free_tail = 0;
1603 	rx_ring->lbuf_free_count = 0;
1604 	rx_ring->lbuf_free = kmem_zalloc(rx_ring->lbq_len *
1605 	    sizeof (struct bq_desc *), KM_NOSLEEP);
1606 	if (rx_ring->lbuf_free == NULL) {
1607 		cmn_err(CE_WARN,
1608 		    "!%s: lbuf_free_list alloc: failed",
1609 		    __func__);
1610 		goto alloc_lbuf_err;
1611 	}
1612 
1613 	rx_ring->lbuf_in_use = kmem_zalloc(rx_ring->lbq_len *
1614 	    sizeof (struct bq_desc *), KM_NOSLEEP);
1615 
1616 	if (rx_ring->lbuf_in_use == NULL) {
1617 		cmn_err(CE_WARN,
1618 		    "!%s: lbuf_inuse_list alloc: failed",
1619 		    __func__);
1620 		goto alloc_lbuf_err;
1621 	}
1622 
1623 	lbq_buf_size = (qlge->mtu == ETHERMTU) ?
1624 	    LRG_BUF_NORMAL_SIZE : LRG_BUF_JUMBO_SIZE;
1625 
1626 	lbq_desc = &rx_ring->lbq_desc[0];
1627 	for (i = 0; i < rx_ring->lbq_len; i++, lbq_desc++) {
1628 		rx_ring->lbq_buf_size = lbq_buf_size;
1629 		/* Allocate buffer */
1630 		if (ql_alloc_phys_rbuf(qlge->dip, &lbq_desc->bd_dma.dma_handle,
1631 		    &ql_buf_acc_attr,
1632 		    DDI_DMA_READ | DDI_DMA_STREAMING,
1633 		    &lbq_desc->bd_dma.acc_handle,
1634 		    (size_t)rx_ring->lbq_buf_size,  /* mem size */
1635 		    (size_t)0, /* default alignment */
1636 		    (caddr_t *)&lbq_desc->bd_dma.vaddr,
1637 		    &dma_cookie) != 0) {
1638 			cmn_err(CE_WARN,
1639 			    "!%s: ddi_dma_alloc_handle: failed",
1640 			    __func__);
1641 			goto alloc_lbuf_err;
1642 		}
1643 
1644 		/* Set context for Return buffer callback */
1645 		lbq_desc->bd_dma.dma_addr = dma_cookie.dmac_laddress;
1646 		lbq_desc->rx_recycle.free_func = ql_release_to_lbuf_free_list;
1647 		lbq_desc->rx_recycle.free_arg  = (caddr_t)lbq_desc;
1648 		lbq_desc->rx_ring = rx_ring;
1649 		lbq_desc->upl_inuse = 0;
1650 		lbq_desc->free_buf = 0;
1651 
1652 		lbq_desc->mp =
1653 		    desballoc((unsigned char *)(lbq_desc->bd_dma.vaddr),
1654 		    rx_ring->lbq_buf_size, 0, &lbq_desc->rx_recycle);
1655 		if (lbq_desc->mp == NULL) {
1656 			cmn_err(CE_WARN, "%s: desballoc() failed", __func__);
1657 			goto alloc_lbuf_err;
1658 		}
1659 		ql_add_lbuf_to_free_list(rx_ring, lbq_desc);
1660 	} /* For all large buffers */
1661 
1662 	return (DDI_SUCCESS);
1663 
1664 alloc_lbuf_err:
1665 	ql_free_lbq_buffers(rx_ring);
1666 	return (DDI_FAILURE);
1667 }
1668 
1669 /*
1670  * Free rx buffers
1671  */
1672 static void
1673 ql_free_rx_buffers(qlge_t *qlge)
1674 {
1675 	int i;
1676 	struct rx_ring *rx_ring;
1677 
1678 	for (i = 0; i < qlge->rx_ring_count; i++) {
1679 		rx_ring = &qlge->rx_ring[i];
1680 		if (rx_ring->type != TX_Q) {
1681 			ql_free_lbq_buffers(rx_ring);
1682 			ql_free_sbq_buffers(rx_ring);
1683 		}
1684 	}
1685 }
1686 
1687 /*
1688  * Allocate rx buffers
1689  */
1690 static int
1691 ql_alloc_rx_buffers(qlge_t *qlge)
1692 {
1693 	struct rx_ring *rx_ring;
1694 	int i;
1695 
1696 	for (i = 0; i < qlge->rx_ring_count; i++) {
1697 		rx_ring = &qlge->rx_ring[i];
1698 		if (rx_ring->type != TX_Q) {
1699 			if (ql_alloc_sbufs(qlge, rx_ring) != DDI_SUCCESS)
1700 				goto alloc_err;
1701 			if (ql_alloc_lbufs(qlge, rx_ring) != DDI_SUCCESS)
1702 				goto alloc_err;
1703 		}
1704 	}
1705 #ifdef QLGE_TRACK_BUFFER_USAGE
1706 	for (i = 0; i < qlge->rx_ring_count; i++) {
1707 		if (qlge->rx_ring[i].type == RX_Q) {
1708 			qlge->rx_sb_low_count[i] = NUM_SMALL_BUFFERS;
1709 			qlge->rx_lb_low_count[i] = NUM_LARGE_BUFFERS;
1710 		}
1711 		qlge->cq_low_count[i] = NUM_RX_RING_ENTRIES;
1712 	}
1713 #endif
1714 	return (DDI_SUCCESS);
1715 
1716 alloc_err:
1717 	ql_free_rx_buffers(qlge);
1718 	return (DDI_FAILURE);
1719 }
1720 
1721 /*
1722  * Initialize large buffer queue ring
1723  */
1724 static void
1725 ql_init_lbq_ring(struct rx_ring *rx_ring)
1726 {
1727 	uint16_t i;
1728 	struct bq_desc *lbq_desc;
1729 
1730 	bzero(rx_ring->lbq_desc, rx_ring->lbq_len * sizeof (struct bq_desc));
1731 	for (i = 0; i < rx_ring->lbq_len; i++) {
1732 		lbq_desc = &rx_ring->lbq_desc[i];
1733 		lbq_desc->index = i;
1734 	}
1735 }
1736 
1737 /*
1738  * Initialize small buffer queue ring
1739  */
1740 static void
1741 ql_init_sbq_ring(struct rx_ring *rx_ring)
1742 {
1743 	uint16_t i;
1744 	struct bq_desc *sbq_desc;
1745 
1746 	bzero(rx_ring->sbq_desc, rx_ring->sbq_len * sizeof (struct bq_desc));
1747 	for (i = 0; i < rx_ring->sbq_len; i++) {
1748 		sbq_desc = &rx_ring->sbq_desc[i];
1749 		sbq_desc->index = i;
1750 	}
1751 }
1752 
1753 /*
1754  * Calculate the pseudo-header checksum if hardware can not do
1755  */
1756 static void
1757 ql_pseudo_cksum(uint8_t *buf)
1758 {
1759 	uint32_t cksum;
1760 	uint16_t iphl;
1761 	uint16_t proto;
1762 
1763 	iphl = (uint16_t)(4 * (buf[0] & 0xF));
1764 	cksum = (((uint16_t)buf[2])<<8) + buf[3] - iphl;
1765 	cksum += proto = buf[9];
1766 	cksum += (((uint16_t)buf[12])<<8) + buf[13];
1767 	cksum += (((uint16_t)buf[14])<<8) + buf[15];
1768 	cksum += (((uint16_t)buf[16])<<8) + buf[17];
1769 	cksum += (((uint16_t)buf[18])<<8) + buf[19];
1770 	cksum = (cksum>>16) + (cksum & 0xFFFF);
1771 	cksum = (cksum>>16) + (cksum & 0xFFFF);
1772 
1773 	/*
1774 	 * Point it to the TCP/UDP header, and
1775 	 * update the checksum field.
1776 	 */
1777 	buf += iphl + ((proto == IPPROTO_TCP) ?
1778 	    TCP_CKSUM_OFFSET : UDP_CKSUM_OFFSET);
1779 
1780 	*(uint16_t *)(void *)buf = (uint16_t)htons((uint16_t)cksum);
1781 
1782 }
1783 
1784 /*
1785  * Transmit an incoming packet.
1786  */
1787 mblk_t *
1788 ql_ring_tx(void *arg, mblk_t *mp)
1789 {
1790 	struct tx_ring *tx_ring = (struct tx_ring *)arg;
1791 	qlge_t *qlge = tx_ring->qlge;
1792 	mblk_t *next;
1793 	int rval;
1794 	uint32_t tx_count = 0;
1795 
1796 	if (qlge->port_link_state == LS_DOWN) {
1797 		/* can not send message while link is down */
1798 		mblk_t *tp;
1799 
1800 		while (mp != NULL) {
1801 			tp = mp->b_next;
1802 			mp->b_next = NULL;
1803 			freemsg(mp);
1804 			mp = tp;
1805 		}
1806 		goto exit;
1807 	}
1808 
1809 	mutex_enter(&tx_ring->tx_lock);
1810 	/* if mac is not started, driver is not ready, can not send */
1811 	if (tx_ring->mac_flags != QL_MAC_STARTED) {
1812 		cmn_err(CE_WARN, "%s(%d)ring not started, mode %d "
1813 		    " return packets",
1814 		    __func__, qlge->instance, tx_ring->mac_flags);
1815 		mutex_exit(&tx_ring->tx_lock);
1816 		goto exit;
1817 	}
1818 
1819 	/* we must try to send all */
1820 	while (mp != NULL) {
1821 		/*
1822 		 * if number of available slots is less than a threshold,
1823 		 * then quit
1824 		 */
1825 		if (tx_ring->tx_free_count <= TX_STOP_THRESHOLD) {
1826 			tx_ring->queue_stopped = 1;
1827 			rval = DDI_FAILURE;
1828 #ifdef QLGE_LOAD_UNLOAD
1829 			cmn_err(CE_WARN, "%s(%d) no resources",
1830 			    __func__, qlge->instance);
1831 #endif
1832 			tx_ring->defer++;
1833 			/*
1834 			 * If we return the buffer back we are expected to call
1835 			 * mac_tx_ring_update() when resources are available
1836 			 */
1837 			break;
1838 		}
1839 
1840 		next = mp->b_next;
1841 		mp->b_next = NULL;
1842 
1843 		rval = ql_send_common(tx_ring, mp);
1844 
1845 		if (rval != DDI_SUCCESS) {
1846 			mp->b_next = next;
1847 			break;
1848 		}
1849 		tx_count++;
1850 		mp = next;
1851 	}
1852 
1853 	/*
1854 	 * After all msg blocks are mapped or copied to tx buffer,
1855 	 * trigger the hardware to send!
1856 	 */
1857 	if (tx_count > 0) {
1858 		ql_write_doorbell_reg(tx_ring->qlge, tx_ring->prod_idx_db_reg,
1859 		    tx_ring->prod_idx);
1860 	}
1861 
1862 	mutex_exit(&tx_ring->tx_lock);
1863 exit:
1864 	return (mp);
1865 }
1866 
1867 
1868 /*
1869  * This function builds an mblk list for the given inbound
1870  * completion.
1871  */
1872 
1873 static mblk_t *
1874 ql_build_rx_mp(qlge_t *qlge, struct rx_ring *rx_ring,
1875     struct ib_mac_iocb_rsp *ib_mac_rsp)
1876 {
1877 	mblk_t *mp = NULL;
1878 	mblk_t *mp1 = NULL;	/* packet header */
1879 	mblk_t *mp2 = NULL;	/* packet content */
1880 	struct bq_desc *lbq_desc;
1881 	struct bq_desc *sbq_desc;
1882 	uint32_t err_flag = (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_ERR_MASK);
1883 	uint32_t payload_len = le32_to_cpu(ib_mac_rsp->data_len);
1884 	uint32_t header_len = le32_to_cpu(ib_mac_rsp->hdr_len);
1885 	uint32_t pkt_len = payload_len + header_len;
1886 	uint32_t done;
1887 	uint64_t *curr_ial_ptr;
1888 	uint32_t ial_data_addr_low;
1889 	uint32_t actual_data_addr_low;
1890 	mblk_t *mp_ial = NULL;	/* ial chained packets */
1891 	uint32_t size;
1892 	uint32_t cp_offset;
1893 	boolean_t rx_copy = B_FALSE;
1894 	mblk_t *tp = NULL;
1895 
1896 	/*
1897 	 * Check if error flags are set
1898 	 */
1899 	if (err_flag != 0) {
1900 		if ((err_flag & IB_MAC_IOCB_RSP_ERR_OVERSIZE) != 0)
1901 			rx_ring->frame_too_long++;
1902 		if ((err_flag & IB_MAC_IOCB_RSP_ERR_UNDERSIZE) != 0)
1903 			rx_ring->frame_too_short++;
1904 		if ((err_flag & IB_MAC_IOCB_RSP_ERR_CRC) != 0)
1905 			rx_ring->fcs_err++;
1906 #ifdef QLGE_LOAD_UNLOAD
1907 		cmn_err(CE_WARN, "bad packet, type 0x%x", err_flag);
1908 #endif
1909 		QL_DUMP(DBG_RX, "qlge_ring_rx: bad response iocb dump\n",
1910 		    (uint8_t *)ib_mac_rsp, 8,
1911 		    (size_t)sizeof (struct ib_mac_iocb_rsp));
1912 	}
1913 
1914 	/* header should not be in large buffer */
1915 	if (ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HL) {
1916 		cmn_err(CE_WARN, "header in large buffer or invalid!");
1917 		err_flag |= 1;
1918 	}
1919 	/* if whole packet is too big than rx buffer size */
1920 	if (pkt_len > qlge->max_frame_size) {
1921 		cmn_err(CE_WARN, "ql_build_rx_mpframe too long(%d)!", pkt_len);
1922 		err_flag |= 1;
1923 	}
1924 	if (qlge->rx_copy ||
1925 	    (rx_ring->sbuf_in_use_count <= qlge->rx_copy_threshold) ||
1926 	    (rx_ring->lbuf_in_use_count <= qlge->rx_copy_threshold)) {
1927 		rx_copy = B_TRUE;
1928 	}
1929 
1930 	/* if using rx copy mode, we need to allocate a big enough buffer */
1931 	if (rx_copy) {
1932 		qlge->stats.norcvbuf++;
1933 		tp = allocb(payload_len + header_len + qlge->ip_hdr_offset,
1934 		    BPRI_MED);
1935 		if (tp == NULL) {
1936 			cmn_err(CE_WARN, "rx copy failed to allocate memory");
1937 		} else {
1938 			tp->b_rptr += qlge->ip_hdr_offset;
1939 		}
1940 	}
1941 	/*
1942 	 * Handle the header buffer if present.
1943 	 * packet header must be valid and saved in one small buffer
1944 	 * broadcast/multicast packets' headers not splitted
1945 	 */
1946 	if ((ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HV) &&
1947 	    (ib_mac_rsp->flags4 & IB_MAC_IOCB_RSP_HS)) {
1948 		QL_PRINT(DBG_RX, ("Header of %d bytes in small buffer.\n",
1949 		    header_len));
1950 		/* Sync access */
1951 		sbq_desc = ql_get_sbuf_from_in_use_list(rx_ring);
1952 
1953 		ASSERT(sbq_desc != NULL);
1954 
1955 		/*
1956 		 * Validate addresses from the ASIC with the
1957 		 * expected sbuf address
1958 		 */
1959 		if (cpu_to_le64(sbq_desc->bd_dma.dma_addr)
1960 		    != ib_mac_rsp->hdr_addr) {
1961 			/* Small buffer address mismatch */
1962 			cmn_err(CE_WARN, "%s(%d) ring%d packet saved"
1963 			    " in wrong small buffer",
1964 			    __func__, qlge->instance, rx_ring->cq_id);
1965 			goto fatal_error;
1966 		}
1967 		/* get this packet */
1968 		mp1 = sbq_desc->mp;
1969 		/* Flush DMA'd data */
1970 		(void) ddi_dma_sync(sbq_desc->bd_dma.dma_handle,
1971 		    0, header_len, DDI_DMA_SYNC_FORKERNEL);
1972 
1973 		if ((err_flag != 0)|| (mp1 == NULL)) {
1974 			/* failed on this packet, put it back for re-arming */
1975 #ifdef QLGE_LOAD_UNLOAD
1976 			cmn_err(CE_WARN, "get header from small buffer fail");
1977 #endif
1978 			ql_refill_sbuf_free_list(sbq_desc, B_FALSE);
1979 			mp1 = NULL;
1980 		} else if (rx_copy) {
1981 			if (tp != NULL) {
1982 				bcopy(sbq_desc->bd_dma.vaddr, tp->b_rptr,
1983 				    header_len);
1984 			}
1985 			ql_refill_sbuf_free_list(sbq_desc, B_FALSE);
1986 			mp1 = NULL;
1987 		} else {
1988 			if ((qlge->ip_hdr_offset != 0)&&
1989 			    (header_len < SMALL_BUFFER_SIZE)) {
1990 				/*
1991 				 * copy entire header to a 2 bytes boundary
1992 				 * address for 8100 adapters so that the IP
1993 				 * header can be on a 4 byte boundary address
1994 				 */
1995 				bcopy(mp1->b_rptr,
1996 				    (mp1->b_rptr + SMALL_BUFFER_SIZE +
1997 				    qlge->ip_hdr_offset),
1998 				    header_len);
1999 				mp1->b_rptr += SMALL_BUFFER_SIZE +
2000 				    qlge->ip_hdr_offset;
2001 			}
2002 
2003 			/*
2004 			 * Adjust the mp payload_len to match
2005 			 * the packet header payload_len
2006 			 */
2007 			mp1->b_wptr = mp1->b_rptr + header_len;
2008 			mp1->b_next = mp1->b_cont = NULL;
2009 			QL_DUMP(DBG_RX, "\t RX packet header dump:\n",
2010 			    (uint8_t *)mp1->b_rptr, 8, header_len);
2011 		}
2012 	}
2013 
2014 	/*
2015 	 * packet data or whole packet can be in small or one or
2016 	 * several large buffer(s)
2017 	 */
2018 	if (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_DS) {
2019 		/*
2020 		 * The data is in a single small buffer.
2021 		 */
2022 		sbq_desc = ql_get_sbuf_from_in_use_list(rx_ring);
2023 
2024 		ASSERT(sbq_desc != NULL);
2025 
2026 		QL_PRINT(DBG_RX,
2027 		    ("%d bytes in a single small buffer, sbq_desc = %p, "
2028 		    "sbq_desc->bd_dma.dma_addr = %x,"
2029 		    " ib_mac_rsp->data_addr = %x, mp = %p\n",
2030 		    payload_len, sbq_desc, sbq_desc->bd_dma.dma_addr,
2031 		    ib_mac_rsp->data_addr, sbq_desc->mp));
2032 
2033 		/*
2034 		 * Validate  addresses from the ASIC with the
2035 		 * expected sbuf address
2036 		 */
2037 		if (cpu_to_le64(sbq_desc->bd_dma.dma_addr)
2038 		    != ib_mac_rsp->data_addr) {
2039 			/* Small buffer address mismatch */
2040 			cmn_err(CE_WARN, "%s(%d) ring%d packet saved"
2041 			    " in wrong small buffer",
2042 			    __func__, qlge->instance, rx_ring->cq_id);
2043 			goto fatal_error;
2044 		}
2045 		/* get this packet */
2046 		mp2 = sbq_desc->mp;
2047 		(void) ddi_dma_sync(sbq_desc->bd_dma.dma_handle,
2048 		    0, payload_len, DDI_DMA_SYNC_FORKERNEL);
2049 		if ((err_flag != 0) || (mp2 == NULL)) {
2050 #ifdef QLGE_LOAD_UNLOAD
2051 			/* failed on this packet, put it back for re-arming */
2052 			cmn_err(CE_WARN, "ignore bad data from small buffer");
2053 #endif
2054 			ql_refill_sbuf_free_list(sbq_desc, B_FALSE);
2055 			mp2 = NULL;
2056 		} else if (rx_copy) {
2057 			if (tp != NULL) {
2058 				bcopy(sbq_desc->bd_dma.vaddr,
2059 				    tp->b_rptr + header_len, payload_len);
2060 				tp->b_wptr =
2061 				    tp->b_rptr + header_len + payload_len;
2062 			}
2063 			ql_refill_sbuf_free_list(sbq_desc, B_FALSE);
2064 			mp2 = NULL;
2065 		} else {
2066 			/* Adjust the buffer length to match the payload_len */
2067 			mp2->b_wptr = mp2->b_rptr + payload_len;
2068 			mp2->b_next = mp2->b_cont = NULL;
2069 			/* Flush DMA'd data */
2070 			QL_DUMP(DBG_RX, "\t RX packet payload dump:\n",
2071 			    (uint8_t *)mp2->b_rptr, 8, payload_len);
2072 			/*
2073 			 * if payload is too small , copy to
2074 			 * the end of packet header
2075 			 */
2076 			if ((mp1 != NULL) &&
2077 			    (payload_len <= qlge->payload_copy_thresh) &&
2078 			    (pkt_len <
2079 			    (SMALL_BUFFER_SIZE - qlge->ip_hdr_offset))) {
2080 				bcopy(mp2->b_rptr, mp1->b_wptr, payload_len);
2081 				mp1->b_wptr += payload_len;
2082 				freemsg(mp2);
2083 				mp2 = NULL;
2084 			}
2085 		}
2086 	} else if (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_DL) {
2087 		/*
2088 		 * The data is in a single large buffer.
2089 		 */
2090 		lbq_desc = ql_get_lbuf_from_in_use_list(rx_ring);
2091 
2092 		QL_PRINT(DBG_RX,
2093 		    ("%d bytes in a single large buffer, lbq_desc = %p, "
2094 		    "lbq_desc->bd_dma.dma_addr = %x,"
2095 		    " ib_mac_rsp->data_addr = %x, mp = %p\n",
2096 		    payload_len, lbq_desc, lbq_desc->bd_dma.dma_addr,
2097 		    ib_mac_rsp->data_addr, lbq_desc->mp));
2098 
2099 		ASSERT(lbq_desc != NULL);
2100 
2101 		/*
2102 		 * Validate  addresses from the ASIC with
2103 		 * the expected lbuf address
2104 		 */
2105 		if (cpu_to_le64(lbq_desc->bd_dma.dma_addr)
2106 		    != ib_mac_rsp->data_addr) {
2107 			/* Large buffer address mismatch */
2108 			cmn_err(CE_WARN, "%s(%d) ring%d packet saved"
2109 			    " in wrong large buffer",
2110 			    __func__, qlge->instance, rx_ring->cq_id);
2111 			goto fatal_error;
2112 		}
2113 		mp2 = lbq_desc->mp;
2114 		/* Flush DMA'd data */
2115 		(void) ddi_dma_sync(lbq_desc->bd_dma.dma_handle,
2116 		    0, payload_len, DDI_DMA_SYNC_FORKERNEL);
2117 		if ((err_flag != 0) || (mp2 == NULL)) {
2118 #ifdef QLGE_LOAD_UNLOAD
2119 			cmn_err(CE_WARN, "ignore bad data from large buffer");
2120 #endif
2121 			/* failed on this packet, put it back for re-arming */
2122 			ql_refill_lbuf_free_list(lbq_desc, B_FALSE);
2123 			mp2 = NULL;
2124 		} else if (rx_copy) {
2125 			if (tp != NULL) {
2126 				bcopy(lbq_desc->bd_dma.vaddr,
2127 				    tp->b_rptr + header_len, payload_len);
2128 				tp->b_wptr =
2129 				    tp->b_rptr + header_len + payload_len;
2130 			}
2131 			ql_refill_lbuf_free_list(lbq_desc, B_FALSE);
2132 			mp2 = NULL;
2133 		} else {
2134 			/*
2135 			 * Adjust the buffer length to match
2136 			 * the packet payload_len
2137 			 */
2138 			mp2->b_wptr = mp2->b_rptr + payload_len;
2139 			mp2->b_next = mp2->b_cont = NULL;
2140 			QL_DUMP(DBG_RX, "\t RX packet payload dump:\n",
2141 			    (uint8_t *)mp2->b_rptr, 8, payload_len);
2142 			/*
2143 			 * if payload is too small , copy to
2144 			 * the end of packet header
2145 			 */
2146 			if ((mp1 != NULL) &&
2147 			    (payload_len <= qlge->payload_copy_thresh) &&
2148 			    (pkt_len<
2149 			    (SMALL_BUFFER_SIZE - qlge->ip_hdr_offset))) {
2150 				bcopy(mp2->b_rptr, mp1->b_wptr, payload_len);
2151 				mp1->b_wptr += payload_len;
2152 				freemsg(mp2);
2153 				mp2 = NULL;
2154 			}
2155 		}
2156 	} else if (payload_len) { /* ial case */
2157 		/*
2158 		 * payload available but not in sml nor lrg buffer,
2159 		 * so, it is saved in IAL
2160 		 */
2161 #ifdef QLGE_LOAD_UNLOAD
2162 		cmn_err(CE_NOTE, "packet chained in IAL \n");
2163 #endif
2164 		/* lrg buf addresses are saved in one small buffer */
2165 		sbq_desc = ql_get_sbuf_from_in_use_list(rx_ring);
2166 		curr_ial_ptr = (uint64_t *)sbq_desc->bd_dma.vaddr;
2167 		done = 0;
2168 		cp_offset = 0;
2169 
2170 		while (!done) {
2171 			ial_data_addr_low =
2172 			    (uint32_t)(le64_to_cpu(*curr_ial_ptr) &
2173 			    0xFFFFFFFE);
2174 			/* check if this is the last packet fragment */
2175 			done = (uint32_t)(le64_to_cpu(*curr_ial_ptr) & 1);
2176 			curr_ial_ptr++;
2177 			/*
2178 			 * The data is in one or several large buffer(s).
2179 			 */
2180 			lbq_desc = ql_get_lbuf_from_in_use_list(rx_ring);
2181 			actual_data_addr_low =
2182 			    (uint32_t)(lbq_desc->bd_dma.dma_addr &
2183 			    0xFFFFFFFE);
2184 			if (ial_data_addr_low != actual_data_addr_low) {
2185 				cmn_err(CE_WARN,
2186 				    "packet saved in wrong ial lrg buffer"
2187 				    " expected %x, actual %lx",
2188 				    ial_data_addr_low,
2189 				    (uintptr_t)lbq_desc->bd_dma.dma_addr);
2190 				goto fatal_error;
2191 			}
2192 
2193 			size = (payload_len < rx_ring->lbq_buf_size)?
2194 			    payload_len : rx_ring->lbq_buf_size;
2195 			payload_len -= size;
2196 			mp2 = lbq_desc->mp;
2197 			if ((err_flag != 0) || (mp2 == NULL)) {
2198 #ifdef QLGE_LOAD_UNLOAD
2199 				cmn_err(CE_WARN,
2200 				    "ignore bad data from large buffer");
2201 #endif
2202 				ql_refill_lbuf_free_list(lbq_desc, B_FALSE);
2203 				mp2 = NULL;
2204 			} else if (rx_copy) {
2205 				if (tp != NULL) {
2206 					(void) ddi_dma_sync(
2207 					    lbq_desc->bd_dma.dma_handle,
2208 					    0, size, DDI_DMA_SYNC_FORKERNEL);
2209 					bcopy(lbq_desc->bd_dma.vaddr,
2210 					    tp->b_rptr + header_len + cp_offset,
2211 					    size);
2212 					tp->b_wptr =
2213 					    tp->b_rptr + size + cp_offset +
2214 					    header_len;
2215 					cp_offset += size;
2216 				}
2217 				ql_refill_lbuf_free_list(lbq_desc, B_FALSE);
2218 				mp2 = NULL;
2219 			} else {
2220 				if (mp_ial == NULL) {
2221 					mp_ial = mp2;
2222 				} else {
2223 					linkb(mp_ial, mp2);
2224 				}
2225 
2226 				mp2->b_next = NULL;
2227 				mp2->b_cont = NULL;
2228 				mp2->b_wptr = mp2->b_rptr + size;
2229 				/* Flush DMA'd data */
2230 				(void) ddi_dma_sync(lbq_desc->bd_dma.dma_handle,
2231 				    0, size, DDI_DMA_SYNC_FORKERNEL);
2232 				QL_PRINT(DBG_RX, ("ial %d payload received \n",
2233 				    size));
2234 				QL_DUMP(DBG_RX, "\t Mac data dump:\n",
2235 				    (uint8_t *)mp2->b_rptr, 8, size);
2236 			}
2237 		}
2238 		if (err_flag != 0) {
2239 #ifdef QLGE_LOAD_UNLOAD
2240 			/* failed on this packet, put it back for re-arming */
2241 			cmn_err(CE_WARN, "ignore bad data from small buffer");
2242 #endif
2243 			ql_refill_sbuf_free_list(sbq_desc, B_FALSE);
2244 		} else {
2245 			mp2 = mp_ial;
2246 			freemsg(sbq_desc->mp);
2247 		}
2248 	}
2249 	/*
2250 	 * some packets' hdr not split, then send mp2 upstream, otherwise,
2251 	 * concatenate message block mp2 to the tail of message header, mp1
2252 	 */
2253 	if (!err_flag) {
2254 		if (rx_copy) {
2255 			if (tp != NULL) {
2256 				tp->b_next = NULL;
2257 				tp->b_cont = NULL;
2258 				tp->b_wptr = tp->b_rptr +
2259 				    header_len + payload_len;
2260 			}
2261 			mp = tp;
2262 		} else {
2263 			if (mp1) {
2264 				if (mp2) {
2265 					QL_PRINT(DBG_RX,
2266 					    ("packet in mp1 and mp2\n"));
2267 					/* mp1->b_cont = mp2; */
2268 					linkb(mp1, mp2);
2269 					mp = mp1;
2270 				} else {
2271 					QL_PRINT(DBG_RX,
2272 					    ("packet in mp1 only\n"));
2273 					mp = mp1;
2274 				}
2275 			} else if (mp2) {
2276 				QL_PRINT(DBG_RX, ("packet in mp2 only\n"));
2277 				mp = mp2;
2278 			}
2279 		}
2280 	}
2281 	return (mp);
2282 
2283 fatal_error:
2284 	/* fatal Error! */
2285 	if (qlge->fm_enable) {
2286 		ddi_fm_service_impact(qlge->dip, DDI_SERVICE_DEGRADED);
2287 		ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
2288 		atomic_or_32(&qlge->flags, ADAPTER_ERROR);
2289 	}
2290 	if (tp) {
2291 		freemsg(tp);
2292 	}
2293 
2294 	/* *mp->b_wptr = 0; */
2295 	ql_wake_asic_reset_soft_intr(qlge);
2296 	return (NULL);
2297 
2298 }
2299 
2300 /*
2301  * Bump completion queue consumer index.
2302  */
2303 static void
2304 ql_update_cq(struct rx_ring *rx_ring)
2305 {
2306 	rx_ring->cnsmr_idx++;
2307 	rx_ring->curr_entry++;
2308 	if (rx_ring->cnsmr_idx >= rx_ring->cq_len) {
2309 		rx_ring->cnsmr_idx = 0;
2310 		rx_ring->curr_entry = rx_ring->cq_dma.vaddr;
2311 	}
2312 }
2313 
2314 /*
2315  * Update completion queue consumer index.
2316  */
2317 static void
2318 ql_write_cq_idx(struct rx_ring *rx_ring)
2319 {
2320 	qlge_t *qlge = rx_ring->qlge;
2321 
2322 	ql_write_doorbell_reg(qlge, rx_ring->cnsmr_idx_db_reg,
2323 	    rx_ring->cnsmr_idx);
2324 }
2325 
2326 /*
2327  * Processes a SYS-Chip Event Notification Completion Event.
2328  * The incoming notification event that describes a link up/down
2329  * or some sorts of error happens.
2330  */
2331 static void
2332 ql_process_chip_ae_intr(qlge_t *qlge,
2333     struct ib_sys_event_iocb_rsp *ib_sys_event_rsp_ptr)
2334 {
2335 	uint8_t eventType = ib_sys_event_rsp_ptr->event_type;
2336 	uint32_t soft_req = 0;
2337 
2338 	switch (eventType) {
2339 		case SYS_EVENT_PORT_LINK_UP:	/* 0x0h */
2340 			QL_PRINT(DBG_MBX, ("Port Link Up\n"));
2341 			break;
2342 
2343 		case SYS_EVENT_PORT_LINK_DOWN:	/* 0x1h */
2344 			QL_PRINT(DBG_MBX, ("Port Link Down\n"));
2345 			break;
2346 
2347 		case SYS_EVENT_MULTIPLE_CAM_HITS : /* 0x6h */
2348 			cmn_err(CE_WARN, "A multiple CAM hits look up error "
2349 			    "occurred");
2350 			soft_req |= NEED_HW_RESET;
2351 			break;
2352 
2353 		case SYS_EVENT_SOFT_ECC_ERR:	/* 0x7h */
2354 			cmn_err(CE_WARN, "Soft ECC error detected");
2355 			soft_req |= NEED_HW_RESET;
2356 			break;
2357 
2358 		case SYS_EVENT_MGMT_FATAL_ERR:	/* 0x8h */
2359 			cmn_err(CE_WARN, "Management (MPI) Processor fatal"
2360 			    " error occured");
2361 			soft_req |= NEED_MPI_RESET;
2362 			break;
2363 
2364 		case SYS_EVENT_MAC_INTERRUPT:	/* 0x9h */
2365 			QL_PRINT(DBG_MBX, ("MAC Interrupt"));
2366 			break;
2367 
2368 		case SYS_EVENT_PCI_ERR_READING_SML_LRG_BUF:	/* 0x40h */
2369 			cmn_err(CE_WARN, "PCI Error reading small/large "
2370 			    "buffers occured");
2371 			soft_req |= NEED_HW_RESET;
2372 			break;
2373 
2374 		default:
2375 			QL_PRINT(DBG_RX, ("%s(%d) unknown Sys Event: "
2376 			    "type 0x%x occured",
2377 			    __func__, qlge->instance, eventType));
2378 			break;
2379 	}
2380 
2381 	if ((soft_req & NEED_MPI_RESET) != 0) {
2382 		ql_wake_mpi_reset_soft_intr(qlge);
2383 		if (qlge->fm_enable) {
2384 			ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
2385 			ddi_fm_service_impact(qlge->dip, DDI_SERVICE_DEGRADED);
2386 		}
2387 	} else if ((soft_req & NEED_HW_RESET) != 0) {
2388 		ql_wake_asic_reset_soft_intr(qlge);
2389 		if (qlge->fm_enable) {
2390 			ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
2391 			ddi_fm_service_impact(qlge->dip, DDI_SERVICE_DEGRADED);
2392 		}
2393 	}
2394 }
2395 
2396 /*
2397  * set received packet checksum flag
2398  */
2399 void
2400 ql_set_rx_cksum(mblk_t *mp, struct ib_mac_iocb_rsp *net_rsp)
2401 {
2402 	uint32_t flags;
2403 
2404 	/* Not TCP or UDP packet? nothing more to do */
2405 	if (((net_rsp->flags2 & IB_MAC_IOCB_RSP_T) == 0) &&
2406 	    ((net_rsp->flags2 & IB_MAC_IOCB_RSP_U) == 0))
2407 		return;
2408 
2409 	/* No CKO support for IPv6 */
2410 	if ((net_rsp->flags3 & IB_MAC_IOCB_RSP_V6) != 0)
2411 		return;
2412 
2413 	/*
2414 	 * If checksum error, don't set flags; stack will calculate
2415 	 * checksum, detect the error and update statistics
2416 	 */
2417 	if (((net_rsp->flags1 & IB_MAC_IOCB_RSP_TE) != 0) ||
2418 	    ((net_rsp->flags1 & IB_MAC_IOCB_RSP_IE) != 0))
2419 		return;
2420 
2421 	/* TCP or UDP packet and checksum valid */
2422 	if (((net_rsp->flags2 & IB_MAC_IOCB_RSP_T) != 0) &&
2423 	    ((net_rsp->flags1 & IB_MAC_IOCB_RSP_NU) == 0)) {
2424 		flags = HCK_FULLCKSUM_OK;
2425 		mac_hcksum_set(mp, 0, 0, 0, 0, flags);
2426 	}
2427 	if (((net_rsp->flags2 & IB_MAC_IOCB_RSP_U) != 0) &&
2428 	    ((net_rsp->flags1 & IB_MAC_IOCB_RSP_NU) == 0)) {
2429 		flags = HCK_FULLCKSUM_OK;
2430 		mac_hcksum_set(mp, 0, 0, 0, 0, flags);
2431 	}
2432 }
2433 
2434 /*
2435  * This function goes through h/w descriptor in one specified rx ring,
2436  * receives the data if the descriptor status shows the data is ready.
2437  * It returns a chain of mblks containing the received data, to be
2438  * passed up to mac_rx_ring().
2439  */
2440 mblk_t *
2441 ql_ring_rx(struct rx_ring *rx_ring, int poll_bytes)
2442 {
2443 	qlge_t *qlge = rx_ring->qlge;
2444 	uint32_t prod = ql_read_sh_reg(qlge, rx_ring);
2445 	struct ib_mac_iocb_rsp *net_rsp;
2446 	mblk_t *mp;
2447 	mblk_t *mblk_head;
2448 	mblk_t **mblk_tail;
2449 	uint32_t received_bytes = 0;
2450 	uint32_t length;
2451 #ifdef QLGE_PERFORMANCE
2452 	uint32_t pkt_ct = 0;
2453 #endif
2454 
2455 #ifdef QLGE_TRACK_BUFFER_USAGE
2456 	uint32_t consumer_idx;
2457 	uint32_t producer_idx;
2458 	uint32_t num_free_entries;
2459 	uint32_t temp;
2460 
2461 	temp = ql_read_doorbell_reg(qlge, rx_ring->cnsmr_idx_db_reg);
2462 	consumer_idx = temp & 0x0000ffff;
2463 	producer_idx = (temp >> 16);
2464 
2465 	if (consumer_idx > producer_idx)
2466 		num_free_entries = (consumer_idx - producer_idx);
2467 	else
2468 		num_free_entries = NUM_RX_RING_ENTRIES - (
2469 		    producer_idx - consumer_idx);
2470 
2471 	if (num_free_entries < qlge->cq_low_count[rx_ring->cq_id])
2472 		qlge->cq_low_count[rx_ring->cq_id] = num_free_entries;
2473 
2474 #endif
2475 	mblk_head = NULL;
2476 	mblk_tail = &mblk_head;
2477 
2478 	while ((prod != rx_ring->cnsmr_idx)) {
2479 		QL_PRINT(DBG_RX,
2480 		    ("%s cq_id = %d, prod = %d, cnsmr = %d.\n",
2481 		    __func__, rx_ring->cq_id, prod, rx_ring->cnsmr_idx));
2482 
2483 		net_rsp = (struct ib_mac_iocb_rsp *)rx_ring->curr_entry;
2484 		(void) ddi_dma_sync(rx_ring->cq_dma.dma_handle,
2485 		    (off_t)((uintptr_t)net_rsp -
2486 		    (uintptr_t)rx_ring->cq_dma.vaddr),
2487 		    (size_t)sizeof (*net_rsp), DDI_DMA_SYNC_FORKERNEL);
2488 		QL_DUMP(DBG_RX, "qlge_ring_rx: rx completion iocb\n",
2489 		    rx_ring->curr_entry, 8, (size_t)sizeof (*net_rsp));
2490 
2491 		switch (net_rsp->opcode) {
2492 
2493 		case OPCODE_IB_MAC_IOCB:
2494 			/* Adding length of pkt header and payload */
2495 			length = le32_to_cpu(net_rsp->data_len) +
2496 			    le32_to_cpu(net_rsp->hdr_len);
2497 			if ((poll_bytes != QLGE_POLL_ALL) &&
2498 			    ((received_bytes + length) > poll_bytes)) {
2499 				continue;
2500 			}
2501 			received_bytes += length;
2502 
2503 #ifdef QLGE_PERFORMANCE
2504 			pkt_ct++;
2505 #endif
2506 			mp = ql_build_rx_mp(qlge, rx_ring, net_rsp);
2507 			if (mp != NULL) {
2508 				if (rx_ring->mac_flags != QL_MAC_STARTED) {
2509 					/*
2510 					 * Increment number of packets we have
2511 					 * indicated to the stack, should be
2512 					 * decremented when we get it back
2513 					 * or when freemsg is called
2514 					 */
2515 					ASSERT(rx_ring->rx_indicate
2516 					    <= rx_ring->cq_len);
2517 #ifdef QLGE_LOAD_UNLOAD
2518 					cmn_err(CE_WARN, "%s do not send to OS,"
2519 					    " mac_flags %d, indicate %d",
2520 					    __func__, rx_ring->mac_flags,
2521 					    rx_ring->rx_indicate);
2522 #endif
2523 					QL_PRINT(DBG_RX,
2524 					    ("cq_id = %d, packet "
2525 					    "dropped, mac not "
2526 					    "enabled.\n",
2527 					    rx_ring->cq_id));
2528 					rx_ring->rx_pkt_dropped_mac_unenabled++;
2529 
2530 					/* rx_lock is expected to be held */
2531 					mutex_exit(&rx_ring->rx_lock);
2532 					freemsg(mp);
2533 					mutex_enter(&rx_ring->rx_lock);
2534 					mp = NULL;
2535 				}
2536 
2537 				if (mp != NULL) {
2538 					/*
2539 					 * IP full packet has been
2540 					 * successfully verified by
2541 					 * H/W and is correct
2542 					 */
2543 					ql_set_rx_cksum(mp, net_rsp);
2544 
2545 					rx_ring->rx_packets++;
2546 					rx_ring->rx_bytes += length;
2547 					*mblk_tail = mp;
2548 					mblk_tail = &mp->b_next;
2549 				}
2550 			} else {
2551 				QL_PRINT(DBG_RX,
2552 				    ("cq_id = %d, packet dropped\n",
2553 				    rx_ring->cq_id));
2554 				rx_ring->rx_packets_dropped_no_buffer++;
2555 			}
2556 			break;
2557 
2558 		case OPCODE_IB_SYS_EVENT_IOCB:
2559 			ql_process_chip_ae_intr(qlge,
2560 			    (struct ib_sys_event_iocb_rsp *)
2561 			    net_rsp);
2562 			break;
2563 
2564 		default:
2565 			cmn_err(CE_WARN,
2566 			    "%s Ring(%d)Hit default case, not handled!"
2567 			    " dropping the packet, "
2568 			    "opcode = %x.", __func__, rx_ring->cq_id,
2569 			    net_rsp->opcode);
2570 			break;
2571 		}
2572 		/* increment cnsmr_idx and curr_entry */
2573 		ql_update_cq(rx_ring);
2574 		prod = ql_read_sh_reg(qlge, rx_ring);
2575 
2576 	}
2577 
2578 #ifdef QLGE_PERFORMANCE
2579 	if (pkt_ct >= 7)
2580 		rx_ring->hist[7]++;
2581 	else if (pkt_ct == 6)
2582 		rx_ring->hist[6]++;
2583 	else if (pkt_ct == 5)
2584 		rx_ring->hist[5]++;
2585 	else if (pkt_ct == 4)
2586 		rx_ring->hist[4]++;
2587 	else if (pkt_ct == 3)
2588 		rx_ring->hist[3]++;
2589 	else if (pkt_ct == 2)
2590 		rx_ring->hist[2]++;
2591 	else if (pkt_ct == 1)
2592 		rx_ring->hist[1]++;
2593 	else if (pkt_ct == 0)
2594 		rx_ring->hist[0]++;
2595 #endif
2596 
2597 	/* update cnsmr_idx */
2598 	ql_write_cq_idx(rx_ring);
2599 	/* do not enable interrupt for polling mode */
2600 	if (poll_bytes == QLGE_POLL_ALL)
2601 		ql_enable_completion_interrupt(rx_ring->qlge, rx_ring->irq);
2602 	return (mblk_head);
2603 }
2604 
2605 /* Process an outbound completion from an rx ring. */
2606 static void
2607 ql_process_mac_tx_intr(qlge_t *qlge, struct ob_mac_iocb_rsp *mac_rsp)
2608 {
2609 	struct tx_ring *tx_ring;
2610 	struct tx_ring_desc *tx_ring_desc;
2611 	int j;
2612 
2613 	tx_ring = &qlge->tx_ring[mac_rsp->txq_idx];
2614 	tx_ring_desc = tx_ring->wq_desc;
2615 	tx_ring_desc += mac_rsp->tid;
2616 
2617 	if (tx_ring_desc->tx_type == USE_DMA) {
2618 		QL_PRINT(DBG_TX, ("%s(%d): tx type USE_DMA\n",
2619 		    __func__, qlge->instance));
2620 
2621 		/*
2622 		 * Release the DMA resource that is used for
2623 		 * DMA binding.
2624 		 */
2625 		for (j = 0; j < tx_ring_desc->tx_dma_handle_used; j++) {
2626 			(void) ddi_dma_unbind_handle(
2627 			    tx_ring_desc->tx_dma_handle[j]);
2628 		}
2629 
2630 		tx_ring_desc->tx_dma_handle_used = 0;
2631 		/*
2632 		 * Free the mblk after sending completed
2633 		 */
2634 		if (tx_ring_desc->mp != NULL) {
2635 			freemsg(tx_ring_desc->mp);
2636 			tx_ring_desc->mp = NULL;
2637 		}
2638 	}
2639 
2640 	tx_ring->obytes += tx_ring_desc->tx_bytes;
2641 	tx_ring->opackets++;
2642 
2643 	if (mac_rsp->flags1 & (OB_MAC_IOCB_RSP_E | OB_MAC_IOCB_RSP_S |
2644 	    OB_MAC_IOCB_RSP_L | OB_MAC_IOCB_RSP_B)) {
2645 		tx_ring->errxmt++;
2646 		if (mac_rsp->flags1 & OB_MAC_IOCB_RSP_E) {
2647 			/* EMPTY */
2648 			QL_PRINT(DBG_TX,
2649 			    ("Total descriptor length did not match "
2650 			    "transfer length.\n"));
2651 		}
2652 		if (mac_rsp->flags1 & OB_MAC_IOCB_RSP_S) {
2653 			/* EMPTY */
2654 			QL_PRINT(DBG_TX,
2655 			    ("Frame too short to be legal, not sent.\n"));
2656 		}
2657 		if (mac_rsp->flags1 & OB_MAC_IOCB_RSP_L) {
2658 			/* EMPTY */
2659 			QL_PRINT(DBG_TX,
2660 			    ("Frame too long, but sent anyway.\n"));
2661 		}
2662 		if (mac_rsp->flags3 & OB_MAC_IOCB_RSP_B) {
2663 			/* EMPTY */
2664 			QL_PRINT(DBG_TX,
2665 			    ("PCI backplane error. Frame not sent.\n"));
2666 		}
2667 	}
2668 	atomic_inc_32(&tx_ring->tx_free_count);
2669 }
2670 
2671 /*
2672  * clean up tx completion iocbs
2673  */
2674 int
2675 ql_clean_outbound_rx_ring(struct rx_ring *rx_ring)
2676 {
2677 	qlge_t *qlge = rx_ring->qlge;
2678 	uint32_t prod = ql_read_sh_reg(qlge, rx_ring);
2679 	struct ob_mac_iocb_rsp *net_rsp = NULL;
2680 	int count = 0;
2681 	struct tx_ring *tx_ring;
2682 	boolean_t resume_tx = B_FALSE;
2683 
2684 	mutex_enter(&rx_ring->rx_lock);
2685 #ifdef QLGE_TRACK_BUFFER_USAGE
2686 	{
2687 	uint32_t consumer_idx;
2688 	uint32_t producer_idx;
2689 	uint32_t num_free_entries;
2690 	uint32_t temp;
2691 
2692 	temp = ql_read_doorbell_reg(qlge, rx_ring->cnsmr_idx_db_reg);
2693 	consumer_idx = temp & 0x0000ffff;
2694 	producer_idx = (temp >> 16);
2695 
2696 	if (consumer_idx > producer_idx)
2697 		num_free_entries = (consumer_idx - producer_idx);
2698 	else
2699 		num_free_entries = NUM_RX_RING_ENTRIES -
2700 		    (producer_idx - consumer_idx);
2701 
2702 	if (num_free_entries < qlge->cq_low_count[rx_ring->cq_id])
2703 		qlge->cq_low_count[rx_ring->cq_id] = num_free_entries;
2704 
2705 	}
2706 #endif
2707 	/* While there are entries in the completion queue. */
2708 	while (prod != rx_ring->cnsmr_idx) {
2709 
2710 		QL_PRINT(DBG_RX,
2711 		    ("%s cq_id = %d, prod = %d, cnsmr = %d.\n", __func__,
2712 		    rx_ring->cq_id, prod, rx_ring->cnsmr_idx));
2713 
2714 		net_rsp = (struct ob_mac_iocb_rsp *)rx_ring->curr_entry;
2715 		(void) ddi_dma_sync(rx_ring->cq_dma.dma_handle,
2716 		    (off_t)((uintptr_t)net_rsp -
2717 		    (uintptr_t)rx_ring->cq_dma.vaddr),
2718 		    (size_t)sizeof (*net_rsp), DDI_DMA_SYNC_FORKERNEL);
2719 
2720 		QL_DUMP(DBG_RX, "ql_clean_outbound_rx_ring: "
2721 		    "response packet data\n",
2722 		    rx_ring->curr_entry, 8,
2723 		    (size_t)sizeof (*net_rsp));
2724 
2725 		switch (net_rsp->opcode) {
2726 
2727 		case OPCODE_OB_MAC_OFFLOAD_IOCB:
2728 		case OPCODE_OB_MAC_IOCB:
2729 			ql_process_mac_tx_intr(qlge, net_rsp);
2730 			break;
2731 
2732 		default:
2733 			cmn_err(CE_WARN,
2734 			    "%s Hit default case, not handled! "
2735 			    "dropping the packet,"
2736 			    " opcode = %x.",
2737 			    __func__, net_rsp->opcode);
2738 			break;
2739 		}
2740 		count++;
2741 		ql_update_cq(rx_ring);
2742 		prod = ql_read_sh_reg(qlge, rx_ring);
2743 	}
2744 	ql_write_cq_idx(rx_ring);
2745 
2746 	mutex_exit(&rx_ring->rx_lock);
2747 
2748 	net_rsp = (struct ob_mac_iocb_rsp *)rx_ring->curr_entry;
2749 	tx_ring = &qlge->tx_ring[net_rsp->txq_idx];
2750 
2751 	mutex_enter(&tx_ring->tx_lock);
2752 
2753 	if (tx_ring->queue_stopped &&
2754 	    (tx_ring->tx_free_count > TX_RESUME_THRESHOLD)) {
2755 		/*
2756 		 * The queue got stopped because the tx_ring was full.
2757 		 * Wake it up, because it's now at least 25% empty.
2758 		 */
2759 		tx_ring->queue_stopped = 0;
2760 		resume_tx = B_TRUE;
2761 	}
2762 
2763 	mutex_exit(&tx_ring->tx_lock);
2764 	/* Don't hold the lock during OS callback */
2765 	if (resume_tx)
2766 		RESUME_TX(tx_ring);
2767 	return (count);
2768 }
2769 
2770 /*
2771  * reset asic when error happens
2772  */
2773 /* ARGSUSED */
2774 static uint_t
2775 ql_asic_reset_work(caddr_t arg1, caddr_t arg2)
2776 {
2777 	qlge_t *qlge = (qlge_t *)((void *)arg1);
2778 	int status;
2779 
2780 	mutex_enter(&qlge->gen_mutex);
2781 	(void) ql_do_stop(qlge);
2782 	/*
2783 	 * Write default ethernet address to chip register Mac
2784 	 * Address slot 0 and Enable Primary Mac Function.
2785 	 */
2786 	mutex_enter(&qlge->hw_mutex);
2787 	(void) ql_unicst_set(qlge,
2788 	    (uint8_t *)qlge->unicst_addr[0].addr.ether_addr_octet, 0);
2789 	mutex_exit(&qlge->hw_mutex);
2790 	qlge->mac_flags = QL_MAC_INIT;
2791 	status = ql_do_start(qlge);
2792 	if (status != DDI_SUCCESS)
2793 		goto error;
2794 	qlge->mac_flags = QL_MAC_STARTED;
2795 	mutex_exit(&qlge->gen_mutex);
2796 	ddi_fm_service_impact(qlge->dip, DDI_SERVICE_RESTORED);
2797 
2798 	return (DDI_INTR_CLAIMED);
2799 
2800 error:
2801 	mutex_exit(&qlge->gen_mutex);
2802 	cmn_err(CE_WARN,
2803 	    "qlge up/down cycle failed, closing device");
2804 	if (qlge->fm_enable) {
2805 		ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
2806 		ddi_fm_service_impact(qlge->dip, DDI_SERVICE_LOST);
2807 		atomic_or_32(&qlge->flags, ADAPTER_ERROR);
2808 	}
2809 	return (DDI_INTR_CLAIMED);
2810 }
2811 
2812 /*
2813  * Reset MPI
2814  */
2815 /* ARGSUSED */
2816 static uint_t
2817 ql_mpi_reset_work(caddr_t arg1, caddr_t arg2)
2818 {
2819 	qlge_t *qlge = (qlge_t *)((void *)arg1);
2820 
2821 	(void) ql_reset_mpi_risc(qlge);
2822 	return (DDI_INTR_CLAIMED);
2823 }
2824 
2825 /*
2826  * Process MPI mailbox messages
2827  */
2828 /* ARGSUSED */
2829 static uint_t
2830 ql_mpi_event_work(caddr_t arg1, caddr_t arg2)
2831 {
2832 	qlge_t *qlge = (qlge_t *)((void *)arg1);
2833 
2834 	ql_do_mpi_intr(qlge);
2835 	return (DDI_INTR_CLAIMED);
2836 }
2837 
2838 /* Fire up a handler to reset the MPI processor. */
2839 void
2840 ql_wake_asic_reset_soft_intr(qlge_t *qlge)
2841 {
2842 	(void) ddi_intr_trigger_softint(qlge->asic_reset_intr_hdl, NULL);
2843 }
2844 
2845 static void
2846 ql_wake_mpi_reset_soft_intr(qlge_t *qlge)
2847 {
2848 	(void) ddi_intr_trigger_softint(qlge->mpi_reset_intr_hdl, NULL);
2849 }
2850 
2851 static void
2852 ql_wake_mpi_event_soft_intr(qlge_t *qlge)
2853 {
2854 	(void) ddi_intr_trigger_softint(qlge->mpi_event_intr_hdl, NULL);
2855 }
2856 
2857 /*
2858  * This handles a fatal error, MPI activity, and the default
2859  * rx_ring in an MSI-X multiple interrupt vector environment.
2860  * In MSI/Legacy environment it also process the rest of
2861  * the rx_rings.
2862  */
2863 /* ARGSUSED */
2864 static uint_t
2865 ql_isr(caddr_t arg1, caddr_t arg2)
2866 {
2867 	struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
2868 	struct rx_ring *ob_ring;
2869 	qlge_t *qlge = rx_ring->qlge;
2870 	struct intr_ctx *intr_ctx = &qlge->intr_ctx[0];
2871 	uint32_t var, prod;
2872 	int i;
2873 	int work_done = 0;
2874 
2875 	mblk_t *mp;
2876 
2877 	_NOTE(ARGUNUSED(arg2));
2878 
2879 	++qlge->rx_interrupts[rx_ring->cq_id];
2880 
2881 	if (ql_atomic_read_32(&qlge->intr_ctx[0].irq_cnt)) {
2882 		ql_write_reg(qlge, REG_RSVD7, 0xfeed0002);
2883 		var = ql_read_reg(qlge, REG_ERROR_STATUS);
2884 		var = ql_read_reg(qlge, REG_STATUS);
2885 		var = ql_read_reg(qlge, REG_INTERRUPT_STATUS_1);
2886 		return (DDI_INTR_CLAIMED);
2887 	}
2888 
2889 	ql_disable_completion_interrupt(qlge, intr_ctx->intr);
2890 
2891 	/*
2892 	 * process send completes on first stride tx ring if available
2893 	 */
2894 	if (qlge->isr_stride) {
2895 		ob_ring = &qlge->rx_ring[qlge->isr_stride];
2896 		if (ql_read_sh_reg(qlge, ob_ring) !=
2897 		    ob_ring->cnsmr_idx) {
2898 			(void) ql_clean_outbound_rx_ring(ob_ring);
2899 		}
2900 	}
2901 	/*
2902 	 * Check the default queue and wake handler if active.
2903 	 */
2904 	rx_ring = &qlge->rx_ring[0];
2905 	prod = ql_read_sh_reg(qlge, rx_ring);
2906 	QL_PRINT(DBG_INTR, ("rx-ring[0] prod index 0x%x, consumer 0x%x ",
2907 	    prod, rx_ring->cnsmr_idx));
2908 	/* check if interrupt is due to incoming packet */
2909 	if (prod != rx_ring->cnsmr_idx) {
2910 		QL_PRINT(DBG_INTR, ("Waking handler for rx_ring[0].\n"));
2911 		ql_disable_completion_interrupt(qlge, intr_ctx->intr);
2912 		mutex_enter(&rx_ring->rx_lock);
2913 		mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
2914 		mutex_exit(&rx_ring->rx_lock);
2915 
2916 		if (mp != NULL)
2917 			RX_UPSTREAM(rx_ring, mp);
2918 		work_done++;
2919 	} else {
2920 		/*
2921 		 * If interrupt is not due to incoming packet, read status
2922 		 * register to see if error happens or mailbox interrupt.
2923 		 */
2924 		var = ql_read_reg(qlge, REG_STATUS);
2925 		if ((var & STATUS_FE) != 0) {
2926 			ql_write_reg(qlge, REG_RSVD7, 0xfeed0003);
2927 			if (qlge->fm_enable) {
2928 				atomic_or_32(&qlge->flags, ADAPTER_ERROR);
2929 				ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
2930 				ddi_fm_service_impact(qlge->dip,
2931 				    DDI_SERVICE_LOST);
2932 			}
2933 			cmn_err(CE_WARN, "Got fatal error, STS = %x.", var);
2934 			var = ql_read_reg(qlge, REG_ERROR_STATUS);
2935 			cmn_err(CE_WARN,
2936 			    "Resetting chip. Error Status Register = 0x%x",
2937 			    var);
2938 			ql_wake_asic_reset_soft_intr(qlge);
2939 			return (DDI_INTR_CLAIMED);
2940 		}
2941 
2942 		/*
2943 		 * Check MPI processor activity.
2944 		 */
2945 		if ((var & STATUS_PI) != 0) {
2946 			/*
2947 			 * We've got an async event or mailbox completion.
2948 			 * Handle it and clear the source of the interrupt.
2949 			 */
2950 			ql_write_reg(qlge, REG_RSVD7, 0xfeed0004);
2951 
2952 			QL_PRINT(DBG_INTR, ("Got MPI processor interrupt.\n"));
2953 			ql_disable_completion_interrupt(qlge, intr_ctx->intr);
2954 			ql_wake_mpi_event_soft_intr(qlge);
2955 			work_done++;
2956 		}
2957 	}
2958 
2959 
2960 	if (qlge->intr_type != DDI_INTR_TYPE_MSIX) {
2961 		/*
2962 		 * Start the DPC for each active queue.
2963 		 */
2964 		for (i = 1; i < qlge->rx_ring_count; i++) {
2965 			rx_ring = &qlge->rx_ring[i];
2966 
2967 			if (ql_read_sh_reg(qlge, rx_ring) !=
2968 			    rx_ring->cnsmr_idx) {
2969 				QL_PRINT(DBG_INTR,
2970 				    ("Waking handler for rx_ring[%d].\n", i));
2971 
2972 				ql_disable_completion_interrupt(qlge,
2973 				    rx_ring->irq);
2974 				if (rx_ring->type == TX_Q) {
2975 					(void) ql_clean_outbound_rx_ring(
2976 					    rx_ring);
2977 					ql_enable_completion_interrupt(
2978 					    rx_ring->qlge, rx_ring->irq);
2979 				} else {
2980 					mutex_enter(&rx_ring->rx_lock);
2981 					mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
2982 					mutex_exit(&rx_ring->rx_lock);
2983 					if (mp != NULL)
2984 						RX_UPSTREAM(rx_ring, mp);
2985 #ifdef QLGE_LOAD_UNLOAD
2986 					if (rx_ring->mac_flags ==
2987 					    QL_MAC_STOPPED)
2988 						cmn_err(CE_NOTE,
2989 						    "%s rx_indicate(%d) %d\n",
2990 						    __func__, i,
2991 						    rx_ring->rx_indicate);
2992 #endif
2993 				}
2994 				work_done++;
2995 			}
2996 		}
2997 	}
2998 
2999 	ql_enable_completion_interrupt(qlge, intr_ctx->intr);
3000 
3001 	return (work_done ? DDI_INTR_CLAIMED : DDI_INTR_UNCLAIMED);
3002 }
3003 
3004 /*
3005  * MSI-X Multiple Vector Interrupt Handler for outbound (TX) completions.
3006  */
3007 /* ARGSUSED */
3008 static uint_t
3009 ql_msix_tx_isr(caddr_t arg1, caddr_t arg2)
3010 {
3011 	struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
3012 	qlge_t *qlge = rx_ring->qlge;
3013 	_NOTE(ARGUNUSED(arg2));
3014 
3015 	++qlge->rx_interrupts[rx_ring->cq_id];
3016 	(void) ql_clean_outbound_rx_ring(rx_ring);
3017 	ql_enable_completion_interrupt(rx_ring->qlge, rx_ring->irq);
3018 
3019 	return (DDI_INTR_CLAIMED);
3020 }
3021 
3022 /*
3023  * MSI-X Multiple Vector Interrupt Handler
3024  */
3025 /* ARGSUSED */
3026 static uint_t
3027 ql_msix_isr(caddr_t arg1, caddr_t arg2)
3028 {
3029 	struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
3030 	struct rx_ring *ob_ring;
3031 	qlge_t *qlge = rx_ring->qlge;
3032 	mblk_t *mp;
3033 	_NOTE(ARGUNUSED(arg2));
3034 
3035 	QL_PRINT(DBG_INTR, ("%s for ring %d\n", __func__, rx_ring->cq_id));
3036 
3037 	ql_disable_completion_interrupt(qlge, rx_ring->irq);
3038 
3039 	/*
3040 	 * process send completes on stride tx ring if available
3041 	 */
3042 	if (qlge->isr_stride) {
3043 		ob_ring = rx_ring + qlge->isr_stride;
3044 		if (ql_read_sh_reg(qlge, ob_ring) !=
3045 		    ob_ring->cnsmr_idx) {
3046 			++qlge->rx_interrupts[ob_ring->cq_id];
3047 			(void) ql_clean_outbound_rx_ring(ob_ring);
3048 		}
3049 	}
3050 
3051 	++qlge->rx_interrupts[rx_ring->cq_id];
3052 
3053 	mutex_enter(&rx_ring->rx_lock);
3054 	mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
3055 	mutex_exit(&rx_ring->rx_lock);
3056 
3057 	if (mp != NULL)
3058 		RX_UPSTREAM(rx_ring, mp);
3059 
3060 	return (DDI_INTR_CLAIMED);
3061 }
3062 
3063 /*
3064  * Poll n_bytes of chained incoming packets
3065  */
3066 mblk_t *
3067 ql_ring_rx_poll(void *arg, int n_bytes)
3068 {
3069 	struct rx_ring *rx_ring = (struct rx_ring *)arg;
3070 	qlge_t *qlge = rx_ring->qlge;
3071 	mblk_t *mp = NULL;
3072 	uint32_t var;
3073 
3074 	ASSERT(n_bytes >= 0);
3075 	QL_PRINT(DBG_GLD, ("%s for ring(%d) to read max %d bytes\n",
3076 	    __func__, rx_ring->cq_id, n_bytes));
3077 
3078 	++qlge->rx_polls[rx_ring->cq_id];
3079 
3080 	if (n_bytes == 0)
3081 		return (mp);
3082 	mutex_enter(&rx_ring->rx_lock);
3083 	mp = ql_ring_rx(rx_ring, n_bytes);
3084 	mutex_exit(&rx_ring->rx_lock);
3085 
3086 	if ((rx_ring->cq_id == 0) && (mp == NULL)) {
3087 		var = ql_read_reg(qlge, REG_STATUS);
3088 		/*
3089 		 * Check for fatal error.
3090 		 */
3091 		if ((var & STATUS_FE) != 0) {
3092 			ql_write_reg(qlge, REG_RSVD7, 0xfeed0003);
3093 			var = ql_read_reg(qlge, REG_ERROR_STATUS);
3094 			cmn_err(CE_WARN, "Got fatal error %x.", var);
3095 			ql_wake_asic_reset_soft_intr(qlge);
3096 			if (qlge->fm_enable) {
3097 				atomic_or_32(&qlge->flags, ADAPTER_ERROR);
3098 				ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
3099 				ddi_fm_service_impact(qlge->dip,
3100 				    DDI_SERVICE_LOST);
3101 			}
3102 		}
3103 		/*
3104 		 * Check MPI processor activity.
3105 		 */
3106 		if ((var & STATUS_PI) != 0) {
3107 			/*
3108 			 * We've got an async event or mailbox completion.
3109 			 * Handle it and clear the source of the interrupt.
3110 			 */
3111 			ql_write_reg(qlge, REG_RSVD7, 0xfeed0004);
3112 			ql_do_mpi_intr(qlge);
3113 		}
3114 	}
3115 
3116 	return (mp);
3117 }
3118 
3119 /*
3120  * MSI-X Multiple Vector Interrupt Handler for inbound (RX) completions.
3121  */
3122 /* ARGSUSED */
3123 static uint_t
3124 ql_msix_rx_isr(caddr_t arg1, caddr_t arg2)
3125 {
3126 	struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
3127 	qlge_t *qlge = rx_ring->qlge;
3128 	mblk_t *mp;
3129 	_NOTE(ARGUNUSED(arg2));
3130 
3131 	QL_PRINT(DBG_INTR, ("%s for ring %d\n", __func__, rx_ring->cq_id));
3132 
3133 	++qlge->rx_interrupts[rx_ring->cq_id];
3134 
3135 	mutex_enter(&rx_ring->rx_lock);
3136 	mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
3137 	mutex_exit(&rx_ring->rx_lock);
3138 
3139 	if (mp != NULL)
3140 		RX_UPSTREAM(rx_ring, mp);
3141 
3142 	return (DDI_INTR_CLAIMED);
3143 }
3144 
3145 
3146 /*
3147  *
3148  * Allocate DMA Buffer for ioctl service
3149  *
3150  */
3151 static int
3152 ql_alloc_ioctl_dma_buf(qlge_t *qlge)
3153 {
3154 	uint64_t phy_addr;
3155 	uint64_t alloc_size;
3156 	ddi_dma_cookie_t dma_cookie;
3157 
3158 	alloc_size = qlge->ioctl_buf_dma_attr.mem_len =
3159 	    max(WCS_MPI_CODE_RAM_LENGTH, MEMC_MPI_RAM_LENGTH);
3160 	if (ql_alloc_phys(qlge->dip, &qlge->ioctl_buf_dma_attr.dma_handle,
3161 	    &ql_buf_acc_attr,
3162 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3163 	    &qlge->ioctl_buf_dma_attr.acc_handle,
3164 	    (size_t)alloc_size,  /* mem size */
3165 	    (size_t)0,  /* alignment */
3166 	    (caddr_t *)&qlge->ioctl_buf_dma_attr.vaddr,
3167 	    &dma_cookie) != 0) {
3168 		cmn_err(CE_WARN, "%s(%d): ioctl DMA allocation failed.",
3169 		    __func__, qlge->instance);
3170 		return (DDI_FAILURE);
3171 	}
3172 
3173 	phy_addr = dma_cookie.dmac_laddress;
3174 
3175 	if (qlge->ioctl_buf_dma_attr.vaddr == NULL) {
3176 		cmn_err(CE_WARN, "%s(%d): failed.", __func__, qlge->instance);
3177 		return (DDI_FAILURE);
3178 	}
3179 
3180 	qlge->ioctl_buf_dma_attr.dma_addr = phy_addr;
3181 
3182 	QL_PRINT(DBG_MBX, ("%s: ioctl_dma_buf_virt_addr = 0x%lx, "
3183 	    "phy_addr = 0x%lx\n",
3184 	    __func__, qlge->ioctl_buf_dma_attr.vaddr, phy_addr));
3185 
3186 	return (DDI_SUCCESS);
3187 }
3188 
3189 
3190 /*
3191  * Function to free physical memory.
3192  */
3193 static void
3194 ql_free_phys(ddi_dma_handle_t *dma_handle, ddi_acc_handle_t *acc_handle)
3195 {
3196 	if (*dma_handle != NULL) {
3197 		(void) ddi_dma_unbind_handle(*dma_handle);
3198 		if (*acc_handle != NULL)
3199 			ddi_dma_mem_free(acc_handle);
3200 		ddi_dma_free_handle(dma_handle);
3201 		*acc_handle = NULL;
3202 		*dma_handle = NULL;
3203 	}
3204 }
3205 
3206 /*
3207  * Function to free ioctl dma buffer.
3208  */
3209 static void
3210 ql_free_ioctl_dma_buf(qlge_t *qlge)
3211 {
3212 	if (qlge->ioctl_buf_dma_attr.dma_handle != NULL) {
3213 		ql_free_phys(&qlge->ioctl_buf_dma_attr.dma_handle,
3214 		    &qlge->ioctl_buf_dma_attr.acc_handle);
3215 
3216 		qlge->ioctl_buf_dma_attr.vaddr = NULL;
3217 		qlge->ioctl_buf_dma_attr.dma_handle = NULL;
3218 	}
3219 }
3220 
3221 /*
3222  * Free shadow register space used for request and completion queues
3223  */
3224 static void
3225 ql_free_shadow_space(qlge_t *qlge)
3226 {
3227 	if (qlge->host_copy_shadow_dma_attr.dma_handle != NULL) {
3228 		ql_free_phys(&qlge->host_copy_shadow_dma_attr.dma_handle,
3229 		    &qlge->host_copy_shadow_dma_attr.acc_handle);
3230 		bzero(&qlge->host_copy_shadow_dma_attr,
3231 		    sizeof (qlge->host_copy_shadow_dma_attr));
3232 	}
3233 
3234 	if (qlge->buf_q_ptr_base_addr_dma_attr.dma_handle != NULL) {
3235 		ql_free_phys(&qlge->buf_q_ptr_base_addr_dma_attr.dma_handle,
3236 		    &qlge->buf_q_ptr_base_addr_dma_attr.acc_handle);
3237 		bzero(&qlge->buf_q_ptr_base_addr_dma_attr,
3238 		    sizeof (qlge->buf_q_ptr_base_addr_dma_attr));
3239 	}
3240 }
3241 
3242 /*
3243  * Allocate shadow register space for request and completion queues
3244  */
3245 static int
3246 ql_alloc_shadow_space(qlge_t *qlge)
3247 {
3248 	ddi_dma_cookie_t dma_cookie;
3249 
3250 	if (ql_alloc_phys(qlge->dip,
3251 	    &qlge->host_copy_shadow_dma_attr.dma_handle,
3252 	    &ql_dev_acc_attr,
3253 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3254 	    &qlge->host_copy_shadow_dma_attr.acc_handle,
3255 	    (size_t)VM_PAGE_SIZE,  /* mem size */
3256 	    (size_t)4, /* 4 bytes alignment */
3257 	    (caddr_t *)&qlge->host_copy_shadow_dma_attr.vaddr,
3258 	    &dma_cookie) != 0) {
3259 		bzero(&qlge->host_copy_shadow_dma_attr,
3260 		    sizeof (qlge->host_copy_shadow_dma_attr));
3261 
3262 		cmn_err(CE_WARN, "%s(%d): Unable to allocate DMA memory for "
3263 		    "response shadow registers", __func__, qlge->instance);
3264 		return (DDI_FAILURE);
3265 	}
3266 
3267 	qlge->host_copy_shadow_dma_attr.dma_addr = dma_cookie.dmac_laddress;
3268 
3269 	if (ql_alloc_phys(qlge->dip,
3270 	    &qlge->buf_q_ptr_base_addr_dma_attr.dma_handle,
3271 	    &ql_desc_acc_attr,
3272 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3273 	    &qlge->buf_q_ptr_base_addr_dma_attr.acc_handle,
3274 	    (size_t)VM_PAGE_SIZE,  /* mem size */
3275 	    (size_t)4, /* 4 bytes alignment */
3276 	    (caddr_t *)&qlge->buf_q_ptr_base_addr_dma_attr.vaddr,
3277 	    &dma_cookie) != 0) {
3278 		bzero(&qlge->buf_q_ptr_base_addr_dma_attr,
3279 		    sizeof (qlge->buf_q_ptr_base_addr_dma_attr));
3280 
3281 		cmn_err(CE_WARN, "%s(%d): Unable to allocate DMA memory "
3282 		    "for request shadow registers",
3283 		    __func__, qlge->instance);
3284 		goto err_wqp_sh_area;
3285 	}
3286 	qlge->buf_q_ptr_base_addr_dma_attr.dma_addr = dma_cookie.dmac_laddress;
3287 
3288 	return (DDI_SUCCESS);
3289 
3290 err_wqp_sh_area:
3291 	ql_free_phys(&qlge->host_copy_shadow_dma_attr.dma_handle,
3292 	    &qlge->host_copy_shadow_dma_attr.acc_handle);
3293 	bzero(&qlge->host_copy_shadow_dma_attr,
3294 	    sizeof (qlge->host_copy_shadow_dma_attr));
3295 
3296 	return (DDI_FAILURE);
3297 }
3298 
3299 /*
3300  * Initialize a tx ring
3301  */
3302 static void
3303 ql_init_tx_ring(struct tx_ring *tx_ring)
3304 {
3305 	int i;
3306 	struct ob_mac_iocb_req *mac_iocb_ptr = tx_ring->wq_dma.vaddr;
3307 	struct tx_ring_desc *tx_ring_desc = tx_ring->wq_desc;
3308 
3309 	for (i = 0; i < tx_ring->wq_len; i++) {
3310 		tx_ring_desc->index = i;
3311 		tx_ring_desc->queue_entry = mac_iocb_ptr;
3312 		mac_iocb_ptr++;
3313 		tx_ring_desc++;
3314 	}
3315 	tx_ring->tx_free_count = tx_ring->wq_len;
3316 	tx_ring->queue_stopped = 0;
3317 }
3318 
3319 /*
3320  * Free one tx ring resources
3321  */
3322 static void
3323 ql_free_tx_resources(struct tx_ring *tx_ring)
3324 {
3325 	struct tx_ring_desc *tx_ring_desc;
3326 	int i, j;
3327 
3328 	if (tx_ring->wq_dma.dma_handle != NULL) {
3329 		ql_free_phys(&tx_ring->wq_dma.dma_handle,
3330 		    &tx_ring->wq_dma.acc_handle);
3331 		bzero(&tx_ring->wq_dma, sizeof (tx_ring->wq_dma));
3332 	}
3333 	if (tx_ring->wq_desc != NULL) {
3334 		tx_ring_desc = tx_ring->wq_desc;
3335 		for (i = 0; i < tx_ring->wq_len; i++, tx_ring_desc++) {
3336 			for (j = 0; j < QL_MAX_TX_DMA_HANDLES; j++) {
3337 				if (tx_ring_desc->tx_dma_handle[j]) {
3338 					/*
3339 					 * The unbinding will happen in tx
3340 					 * completion, here we just free the
3341 					 * handles
3342 					 */
3343 					ddi_dma_free_handle(
3344 					    &(tx_ring_desc->tx_dma_handle[j]));
3345 					tx_ring_desc->tx_dma_handle[j] = NULL;
3346 				}
3347 			}
3348 			if (tx_ring_desc->oal != NULL) {
3349 				tx_ring_desc->oal_dma_addr = 0;
3350 				tx_ring_desc->oal = NULL;
3351 				tx_ring_desc->copy_buffer = NULL;
3352 				tx_ring_desc->copy_buffer_dma_addr = 0;
3353 
3354 				ql_free_phys(&tx_ring_desc->oal_dma.dma_handle,
3355 				    &tx_ring_desc->oal_dma.acc_handle);
3356 			}
3357 		}
3358 		kmem_free(tx_ring->wq_desc,
3359 		    tx_ring->wq_len * sizeof (struct tx_ring_desc));
3360 		tx_ring->wq_desc = NULL;
3361 	}
3362 	/* free the wqicb struct */
3363 	if (tx_ring->wqicb_dma.dma_handle) {
3364 		ql_free_phys(&tx_ring->wqicb_dma.dma_handle,
3365 		    &tx_ring->wqicb_dma.acc_handle);
3366 		bzero(&tx_ring->wqicb_dma, sizeof (tx_ring->wqicb_dma));
3367 	}
3368 }
3369 
3370 /*
3371  * Allocate work (request) queue memory and transmit
3372  * descriptors for this transmit ring
3373  */
3374 static int
3375 ql_alloc_tx_resources(qlge_t *qlge, struct tx_ring *tx_ring)
3376 {
3377 	ddi_dma_cookie_t dma_cookie;
3378 	struct tx_ring_desc *tx_ring_desc;
3379 	int i, j;
3380 	uint32_t length;
3381 
3382 	/* allocate dma buffers for obiocbs */
3383 	if (ql_alloc_phys(qlge->dip, &tx_ring->wq_dma.dma_handle,
3384 	    &ql_desc_acc_attr,
3385 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3386 	    &tx_ring->wq_dma.acc_handle,
3387 	    (size_t)tx_ring->wq_size,	/* mem size */
3388 	    (size_t)128, /* alignment:128 bytes boundary */
3389 	    (caddr_t *)&tx_ring->wq_dma.vaddr,
3390 	    &dma_cookie) != 0) {
3391 		bzero(&tx_ring->wq_dma, sizeof (&tx_ring->wq_dma));
3392 		cmn_err(CE_WARN, "%s(%d): reqQ allocation failed.",
3393 		    __func__, qlge->instance);
3394 		return (DDI_FAILURE);
3395 	}
3396 	tx_ring->wq_dma.dma_addr = dma_cookie.dmac_laddress;
3397 
3398 	tx_ring->wq_desc =
3399 	    kmem_zalloc(tx_ring->wq_len * sizeof (struct tx_ring_desc),
3400 	    KM_NOSLEEP);
3401 	if (tx_ring->wq_desc == NULL) {
3402 		goto err;
3403 	} else {
3404 		tx_ring_desc = tx_ring->wq_desc;
3405 		/*
3406 		 * Allocate a large enough structure to hold the following
3407 		 * 1. oal buffer MAX_SGELEMENTS * sizeof (oal_entry) bytes
3408 		 * 2. copy buffer of QL_MAX_COPY_LENGTH bytes
3409 		 */
3410 		length = (sizeof (struct oal_entry) * MAX_SG_ELEMENTS)
3411 		    + QL_MAX_COPY_LENGTH;
3412 		for (i = 0; i < tx_ring->wq_len; i++, tx_ring_desc++) {
3413 
3414 			if (ql_alloc_phys(qlge->dip,
3415 			    &tx_ring_desc->oal_dma.dma_handle,
3416 			    &ql_desc_acc_attr,
3417 			    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3418 			    &tx_ring_desc->oal_dma.acc_handle,
3419 			    (size_t)length,	/* mem size */
3420 			    (size_t)0, /* default alignment:8 bytes boundary */
3421 			    (caddr_t *)&tx_ring_desc->oal_dma.vaddr,
3422 			    &dma_cookie) != 0) {
3423 				bzero(&tx_ring_desc->oal_dma,
3424 				    sizeof (tx_ring_desc->oal_dma));
3425 				cmn_err(CE_WARN, "%s(%d): reqQ tx buf &"
3426 				    "oal alloc failed.",
3427 				    __func__, qlge->instance);
3428 				goto err;
3429 			}
3430 
3431 			tx_ring_desc->oal = tx_ring_desc->oal_dma.vaddr;
3432 			tx_ring_desc->oal_dma_addr = dma_cookie.dmac_laddress;
3433 			tx_ring_desc->copy_buffer =
3434 			    (caddr_t)((uint8_t *)tx_ring_desc->oal
3435 			    + (sizeof (struct oal_entry) * MAX_SG_ELEMENTS));
3436 			tx_ring_desc->copy_buffer_dma_addr =
3437 			    (tx_ring_desc->oal_dma_addr
3438 			    + (sizeof (struct oal_entry) * MAX_SG_ELEMENTS));
3439 
3440 			/* Allocate dma handles for transmit buffers */
3441 			for (j = 0; j < QL_MAX_TX_DMA_HANDLES; j++) {
3442 				if (ddi_dma_alloc_handle(qlge->dip,
3443 				    &tx_mapping_dma_attr,
3444 				    DDI_DMA_DONTWAIT,
3445 				    0, &tx_ring_desc->tx_dma_handle[j])
3446 				    != DDI_SUCCESS) {
3447 					tx_ring_desc->tx_dma_handle[j] = NULL;
3448 					cmn_err(CE_WARN,
3449 					    "!%s: ddi_dma_alloc_handle: "
3450 					    "tx_dma_handle "
3451 					    "alloc failed", __func__);
3452 					ql_free_phys(
3453 					    &tx_ring_desc->oal_dma.dma_handle,
3454 					    &tx_ring_desc->oal_dma.acc_handle);
3455 					goto err;
3456 				}
3457 			}
3458 		}
3459 	}
3460 	/* alloc a wqicb control block to load this tx ring to hw */
3461 	if (ql_alloc_phys(qlge->dip, &tx_ring->wqicb_dma.dma_handle,
3462 	    &ql_desc_acc_attr,
3463 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3464 	    &tx_ring->wqicb_dma.acc_handle,
3465 	    (size_t)sizeof (struct wqicb_t),	/* mem size */
3466 	    (size_t)0, /* alignment:128 bytes boundary */
3467 	    (caddr_t *)&tx_ring->wqicb_dma.vaddr,
3468 	    &dma_cookie) != 0) {
3469 		bzero(&tx_ring->wqicb_dma, sizeof (tx_ring->wqicb_dma));
3470 		cmn_err(CE_WARN, "%s(%d): wqicb allocation failed.",
3471 		    __func__, qlge->instance);
3472 		goto err;
3473 	}
3474 	tx_ring->wqicb_dma.dma_addr = dma_cookie.dmac_laddress;
3475 
3476 	return (DDI_SUCCESS);
3477 
3478 err:
3479 	ql_free_tx_resources(tx_ring);
3480 	return (DDI_FAILURE);
3481 }
3482 
3483 /*
3484  * Free one rx ring resources
3485  */
3486 static void
3487 ql_free_rx_resources(struct rx_ring *rx_ring)
3488 {
3489 	/* Free the small buffer queue. */
3490 	if (rx_ring->sbq_dma.dma_handle) {
3491 		ql_free_phys(&rx_ring->sbq_dma.dma_handle,
3492 		    &rx_ring->sbq_dma.acc_handle);
3493 		bzero(&rx_ring->sbq_dma, sizeof (rx_ring->sbq_dma));
3494 	}
3495 
3496 	/* Free the small buffer queue control blocks. */
3497 	if (rx_ring->sbq_desc != NULL) {
3498 		kmem_free(rx_ring->sbq_desc, rx_ring->sbq_len *
3499 		    sizeof (struct bq_desc));
3500 		rx_ring->sbq_desc = NULL;
3501 	}
3502 
3503 	/* Free the large buffer queue. */
3504 	if (rx_ring->lbq_dma.dma_handle) {
3505 		ql_free_phys(&rx_ring->lbq_dma.dma_handle,
3506 		    &rx_ring->lbq_dma.acc_handle);
3507 		bzero(&rx_ring->lbq_dma, sizeof (rx_ring->lbq_dma));
3508 	}
3509 
3510 	/* Free the large buffer queue control blocks. */
3511 	if (rx_ring->lbq_desc != NULL) {
3512 		kmem_free(rx_ring->lbq_desc, rx_ring->lbq_len *
3513 		    sizeof (struct bq_desc));
3514 		rx_ring->lbq_desc = NULL;
3515 	}
3516 
3517 	/* Free cqicb struct */
3518 	if (rx_ring->cqicb_dma.dma_handle) {
3519 		ql_free_phys(&rx_ring->cqicb_dma.dma_handle,
3520 		    &rx_ring->cqicb_dma.acc_handle);
3521 		bzero(&rx_ring->cqicb_dma, sizeof (rx_ring->cqicb_dma));
3522 	}
3523 	/* Free the rx queue. */
3524 	if (rx_ring->cq_dma.dma_handle) {
3525 		ql_free_phys(&rx_ring->cq_dma.dma_handle,
3526 		    &rx_ring->cq_dma.acc_handle);
3527 		bzero(&rx_ring->cq_dma, sizeof (rx_ring->cq_dma));
3528 	}
3529 }
3530 
3531 /*
3532  * Allocate queues and buffers for this completions queue based
3533  * on the values in the parameter structure.
3534  */
3535 static int
3536 ql_alloc_rx_resources(qlge_t *qlge, struct rx_ring *rx_ring)
3537 {
3538 	ddi_dma_cookie_t dma_cookie;
3539 
3540 	if (ql_alloc_phys(qlge->dip, &rx_ring->cq_dma.dma_handle,
3541 	    &ql_desc_acc_attr,
3542 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3543 	    &rx_ring->cq_dma.acc_handle,
3544 	    (size_t)rx_ring->cq_size,  /* mem size */
3545 	    (size_t)128, /* alignment:128 bytes boundary */
3546 	    (caddr_t *)&rx_ring->cq_dma.vaddr,
3547 	    &dma_cookie) != 0)	{
3548 		bzero(&rx_ring->cq_dma, sizeof (rx_ring->cq_dma));
3549 		cmn_err(CE_WARN, "%s(%d): rspQ allocation failed.",
3550 		    __func__, qlge->instance);
3551 		return (DDI_FAILURE);
3552 	}
3553 	rx_ring->cq_dma.dma_addr = dma_cookie.dmac_laddress;
3554 
3555 	if (rx_ring->sbq_len != 0) {
3556 		/*
3557 		 * Allocate small buffer queue.
3558 		 */
3559 		if (ql_alloc_phys(qlge->dip, &rx_ring->sbq_dma.dma_handle,
3560 		    &ql_desc_acc_attr,
3561 		    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3562 		    &rx_ring->sbq_dma.acc_handle,
3563 		    (size_t)rx_ring->sbq_size,  /* mem size */
3564 		    (size_t)128, /* alignment:128 bytes boundary */
3565 		    (caddr_t *)&rx_ring->sbq_dma.vaddr,
3566 		    &dma_cookie) != 0) {
3567 			bzero(&rx_ring->sbq_dma, sizeof (rx_ring->sbq_dma));
3568 			cmn_err(CE_WARN,
3569 			    "%s(%d): small buffer queue allocation failed.",
3570 			    __func__, qlge->instance);
3571 			goto err_mem;
3572 		}
3573 		rx_ring->sbq_dma.dma_addr = dma_cookie.dmac_laddress;
3574 
3575 		/*
3576 		 * Allocate small buffer queue control blocks.
3577 		 */
3578 		rx_ring->sbq_desc =
3579 		    kmem_zalloc(rx_ring->sbq_len * sizeof (struct bq_desc),
3580 		    KM_NOSLEEP);
3581 		if (rx_ring->sbq_desc == NULL) {
3582 			cmn_err(CE_WARN,
3583 			    "sbq control block allocation failed.");
3584 			goto err_mem;
3585 		}
3586 
3587 		ql_init_sbq_ring(rx_ring);
3588 	}
3589 
3590 	if (rx_ring->lbq_len != 0) {
3591 		/*
3592 		 * Allocate large buffer queue.
3593 		 */
3594 		if (ql_alloc_phys(qlge->dip, &rx_ring->lbq_dma.dma_handle,
3595 		    &ql_desc_acc_attr,
3596 		    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3597 		    &rx_ring->lbq_dma.acc_handle,
3598 		    (size_t)rx_ring->lbq_size,  /* mem size */
3599 		    (size_t)128, /* alignment:128 bytes boundary */
3600 		    (caddr_t *)&rx_ring->lbq_dma.vaddr,
3601 		    &dma_cookie) != 0) {
3602 			bzero(&rx_ring->lbq_dma, sizeof (rx_ring->lbq_dma));
3603 			cmn_err(CE_WARN, "%s(%d): lbq allocation failed.",
3604 			    __func__, qlge->instance);
3605 			goto err_mem;
3606 		}
3607 		rx_ring->lbq_dma.dma_addr = dma_cookie.dmac_laddress;
3608 
3609 		/*
3610 		 * Allocate large buffer queue control blocks.
3611 		 */
3612 		rx_ring->lbq_desc =
3613 		    kmem_zalloc(rx_ring->lbq_len * sizeof (struct bq_desc),
3614 		    KM_NOSLEEP);
3615 		if (rx_ring->lbq_desc == NULL) {
3616 			cmn_err(CE_WARN,
3617 			    "Large buffer queue control block allocation "
3618 			    "failed.");
3619 			goto err_mem;
3620 		}
3621 		ql_init_lbq_ring(rx_ring);
3622 	}
3623 
3624 	if (ql_alloc_phys(qlge->dip, &rx_ring->cqicb_dma.dma_handle,
3625 	    &ql_desc_acc_attr,
3626 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3627 	    &rx_ring->cqicb_dma.acc_handle,
3628 	    (size_t)sizeof (struct cqicb_t),  /* mem size */
3629 	    (size_t)0, /* alignment:128 bytes boundary */
3630 	    (caddr_t *)&rx_ring->cqicb_dma.vaddr,
3631 	    &dma_cookie) != 0) {
3632 		bzero(&rx_ring->cqicb_dma, sizeof (rx_ring->cqicb_dma));
3633 		cmn_err(CE_WARN, "%s(%d): cqicb allocation failed.",
3634 		    __func__, qlge->instance);
3635 		goto err_mem;
3636 	}
3637 	rx_ring->cqicb_dma.dma_addr = dma_cookie.dmac_laddress;
3638 
3639 	return (DDI_SUCCESS);
3640 
3641 err_mem:
3642 	ql_free_rx_resources(rx_ring);
3643 	return (DDI_FAILURE);
3644 }
3645 
3646 /*
3647  * Frees tx/rx queues memory resources
3648  */
3649 static void
3650 ql_free_mem_resources(qlge_t *qlge)
3651 {
3652 	int i;
3653 
3654 	if (qlge->ricb_dma.dma_handle) {
3655 		/* free the ricb struct */
3656 		ql_free_phys(&qlge->ricb_dma.dma_handle,
3657 		    &qlge->ricb_dma.acc_handle);
3658 		bzero(&qlge->ricb_dma, sizeof (qlge->ricb_dma));
3659 	}
3660 
3661 	ql_free_rx_buffers(qlge);
3662 
3663 	ql_free_ioctl_dma_buf(qlge);
3664 
3665 	for (i = 0; i < qlge->tx_ring_count; i++)
3666 		ql_free_tx_resources(&qlge->tx_ring[i]);
3667 
3668 	for (i = 0; i < qlge->rx_ring_count; i++)
3669 		ql_free_rx_resources(&qlge->rx_ring[i]);
3670 
3671 	ql_free_shadow_space(qlge);
3672 }
3673 
3674 /*
3675  * Allocate buffer queues, large buffers and small buffers etc
3676  *
3677  * This API is called in the gld_attach member function. It is called
3678  * only once.  Later reset,reboot should not re-allocate all rings and
3679  * buffers.
3680  */
3681 static int
3682 ql_alloc_mem_resources(qlge_t *qlge)
3683 {
3684 	int i;
3685 	ddi_dma_cookie_t dma_cookie;
3686 
3687 	/* Allocate space for our shadow registers */
3688 	if (ql_alloc_shadow_space(qlge))
3689 		return (DDI_FAILURE);
3690 
3691 	for (i = 0; i < qlge->rx_ring_count; i++) {
3692 		if (ql_alloc_rx_resources(qlge, &qlge->rx_ring[i]) != 0) {
3693 			cmn_err(CE_WARN, "RX resource allocation failed.");
3694 			goto err_mem;
3695 		}
3696 	}
3697 	/* Allocate tx queue resources */
3698 	for (i = 0; i < qlge->tx_ring_count; i++) {
3699 		if (ql_alloc_tx_resources(qlge, &qlge->tx_ring[i]) != 0) {
3700 			cmn_err(CE_WARN, "Tx resource allocation failed.");
3701 			goto err_mem;
3702 		}
3703 	}
3704 
3705 	if (ql_alloc_ioctl_dma_buf(qlge) != DDI_SUCCESS) {
3706 		goto err_mem;
3707 	}
3708 
3709 	if (ql_alloc_rx_buffers(qlge) != DDI_SUCCESS) {
3710 		cmn_err(CE_WARN, "?%s(%d): ql_alloc_rx_buffers failed",
3711 		    __func__, qlge->instance);
3712 		goto err_mem;
3713 	}
3714 
3715 	qlge->sequence |= INIT_ALLOC_RX_BUF;
3716 
3717 	if (ql_alloc_phys(qlge->dip, &qlge->ricb_dma.dma_handle,
3718 	    &ql_desc_acc_attr,
3719 	    DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
3720 	    &qlge->ricb_dma.acc_handle,
3721 	    (size_t)sizeof (struct ricb),  /* mem size */
3722 	    (size_t)0, /* alignment:128 bytes boundary */
3723 	    (caddr_t *)&qlge->ricb_dma.vaddr,
3724 	    &dma_cookie) != 0) {
3725 		bzero(&qlge->ricb_dma, sizeof (qlge->ricb_dma));
3726 		cmn_err(CE_WARN, "%s(%d): ricb allocation failed.",
3727 		    __func__, qlge->instance);
3728 		goto err_mem;
3729 	}
3730 	qlge->ricb_dma.dma_addr = dma_cookie.dmac_laddress;
3731 
3732 	return (DDI_SUCCESS);
3733 
3734 err_mem:
3735 	ql_free_mem_resources(qlge);
3736 	return (DDI_FAILURE);
3737 }
3738 
3739 
3740 /*
3741  * Function used to allocate physical memory and zero it.
3742  */
3743 
3744 static int
3745 ql_alloc_phys_rbuf(dev_info_t *dip, ddi_dma_handle_t *dma_handle,
3746     ddi_device_acc_attr_t *device_acc_attr,
3747     uint_t dma_flags,
3748     ddi_acc_handle_t *acc_handle,
3749     size_t size,
3750     size_t alignment,
3751     caddr_t *vaddr,
3752     ddi_dma_cookie_t *dma_cookie)
3753 {
3754 	size_t rlen;
3755 	uint_t cnt;
3756 
3757 	/*
3758 	 * Workaround for SUN XMITS buffer must end and start on 8 byte
3759 	 * boundary. Else, hardware will overrun the buffer. Simple fix is
3760 	 * to make sure buffer has enough room for overrun.
3761 	 */
3762 	if (size & 7) {
3763 		size += 8 - (size & 7);
3764 	}
3765 
3766 	/* Adjust the alignment if requested */
3767 	if (alignment) {
3768 		dma_attr.dma_attr_align = alignment;
3769 	}
3770 
3771 	/*
3772 	 * Allocate DMA handle
3773 	 */
3774 	if (ddi_dma_alloc_handle(dip, &dma_attr_rbuf, DDI_DMA_DONTWAIT, NULL,
3775 	    dma_handle) != DDI_SUCCESS) {
3776 		cmn_err(CE_WARN, QL_BANG "%s:  ddi_dma_alloc_handle FAILED",
3777 		    __func__);
3778 		*dma_handle = NULL;
3779 		return (QL_ERROR);
3780 	}
3781 	/*
3782 	 * Allocate DMA memory
3783 	 */
3784 	if (ddi_dma_mem_alloc(*dma_handle, size, device_acc_attr,
3785 	    dma_flags & (DDI_DMA_CONSISTENT|DDI_DMA_STREAMING),
3786 	    DDI_DMA_DONTWAIT,
3787 	    NULL, vaddr, &rlen, acc_handle) != DDI_SUCCESS) {
3788 		cmn_err(CE_WARN, "alloc_phys: DMA Memory alloc Failed");
3789 		ddi_dma_free_handle(dma_handle);
3790 		*acc_handle = NULL;
3791 		*dma_handle = NULL;
3792 		return (QL_ERROR);
3793 	}
3794 
3795 	if (ddi_dma_addr_bind_handle(*dma_handle, NULL, *vaddr, rlen,
3796 	    dma_flags, DDI_DMA_DONTWAIT, NULL,
3797 	    dma_cookie, &cnt) != DDI_DMA_MAPPED) {
3798 		ddi_dma_mem_free(acc_handle);
3799 
3800 		ddi_dma_free_handle(dma_handle);
3801 		cmn_err(CE_WARN, "%s ddi_dma_addr_bind_handle FAILED",
3802 		    __func__);
3803 		*acc_handle = NULL;
3804 		*dma_handle = NULL;
3805 		return (QL_ERROR);
3806 	}
3807 
3808 	if (cnt != 1) {
3809 
3810 		ql_free_phys(dma_handle, acc_handle);
3811 
3812 		cmn_err(CE_WARN, "%s: cnt != 1; Failed segment count",
3813 		    __func__);
3814 		return (QL_ERROR);
3815 	}
3816 
3817 	bzero((caddr_t)*vaddr, rlen);
3818 
3819 	return (0);
3820 }
3821 
3822 /*
3823  * Function used to allocate physical memory and zero it.
3824  */
3825 static int
3826 ql_alloc_phys(dev_info_t *dip, ddi_dma_handle_t *dma_handle,
3827     ddi_device_acc_attr_t *device_acc_attr,
3828     uint_t dma_flags,
3829     ddi_acc_handle_t *acc_handle,
3830     size_t size,
3831     size_t alignment,
3832     caddr_t *vaddr,
3833     ddi_dma_cookie_t *dma_cookie)
3834 {
3835 	size_t rlen;
3836 	uint_t cnt;
3837 
3838 	/*
3839 	 * Workaround for SUN XMITS buffer must end and start on 8 byte
3840 	 * boundary. Else, hardware will overrun the buffer. Simple fix is
3841 	 * to make sure buffer has enough room for overrun.
3842 	 */
3843 	if (size & 7) {
3844 		size += 8 - (size & 7);
3845 	}
3846 
3847 	/* Adjust the alignment if requested */
3848 	if (alignment) {
3849 		dma_attr.dma_attr_align = alignment;
3850 	}
3851 
3852 	/*
3853 	 * Allocate DMA handle
3854 	 */
3855 	if (ddi_dma_alloc_handle(dip, &dma_attr, DDI_DMA_DONTWAIT, NULL,
3856 	    dma_handle) != DDI_SUCCESS) {
3857 		cmn_err(CE_WARN, QL_BANG "%s:  ddi_dma_alloc_handle FAILED",
3858 		    __func__);
3859 		*dma_handle = NULL;
3860 		return (QL_ERROR);
3861 	}
3862 	/*
3863 	 * Allocate DMA memory
3864 	 */
3865 	if (ddi_dma_mem_alloc(*dma_handle, size, device_acc_attr,
3866 	    dma_flags & (DDI_DMA_CONSISTENT|DDI_DMA_STREAMING),
3867 	    DDI_DMA_DONTWAIT,
3868 	    NULL, vaddr, &rlen, acc_handle) != DDI_SUCCESS) {
3869 		cmn_err(CE_WARN, "alloc_phys: DMA Memory alloc Failed");
3870 		ddi_dma_free_handle(dma_handle);
3871 		*acc_handle = NULL;
3872 		*dma_handle = NULL;
3873 		return (QL_ERROR);
3874 	}
3875 
3876 	if (ddi_dma_addr_bind_handle(*dma_handle, NULL, *vaddr, rlen,
3877 	    dma_flags, DDI_DMA_DONTWAIT, NULL,
3878 	    dma_cookie, &cnt) != DDI_DMA_MAPPED) {
3879 		ddi_dma_mem_free(acc_handle);
3880 		ddi_dma_free_handle(dma_handle);
3881 		cmn_err(CE_WARN, "%s ddi_dma_addr_bind_handle FAILED",
3882 		    __func__);
3883 		*acc_handle = NULL;
3884 		*dma_handle = NULL;
3885 		return (QL_ERROR);
3886 	}
3887 
3888 	if (cnt != 1) {
3889 
3890 		ql_free_phys(dma_handle, acc_handle);
3891 
3892 		cmn_err(CE_WARN, "%s: cnt != 1; Failed segment count",
3893 		    __func__);
3894 		return (QL_ERROR);
3895 	}
3896 
3897 	bzero((caddr_t)*vaddr, rlen);
3898 
3899 	return (0);
3900 }
3901 
3902 /*
3903  * Add interrupt handlers based on the interrupt type.
3904  * Before adding the interrupt handlers, the interrupt vectors should
3905  * have been allocated, and the rx/tx rings have also been allocated.
3906  */
3907 static int
3908 ql_add_intr_handlers(qlge_t *qlge)
3909 {
3910 	int vector = 0;
3911 	int rc, i;
3912 	uint32_t value;
3913 	struct intr_ctx *intr_ctx = &qlge->intr_ctx[0];
3914 
3915 	switch (qlge->intr_type) {
3916 	case DDI_INTR_TYPE_MSIX:
3917 		/*
3918 		 * Add interrupt handler for rx and tx rings: vector[0 -
3919 		 * (qlge->intr_cnt -1)].
3920 		 */
3921 		value = 0;
3922 		for (vector = 0; vector < qlge->intr_cnt; vector++) {
3923 			ql_atomic_set_32(&intr_ctx->irq_cnt, value);
3924 
3925 			/*
3926 			 * associate interrupt vector with interrupt handler
3927 			 */
3928 			rc = ddi_intr_add_handler(qlge->htable[vector],
3929 			    (ddi_intr_handler_t *)intr_ctx->handler,
3930 			    (void *)&qlge->rx_ring[vector], NULL);
3931 
3932 			QL_PRINT(DBG_INIT, ("rx_ring[%d] 0x%p\n",
3933 			    vector, &qlge->rx_ring[vector]));
3934 			if (rc != DDI_SUCCESS) {
3935 				QL_PRINT(DBG_INIT,
3936 				    ("Add rx interrupt handler failed. "
3937 				    "return: %d, vector: %d", rc, vector));
3938 				for (vector--; vector >= 0; vector--) {
3939 					(void) ddi_intr_remove_handler(
3940 					    qlge->htable[vector]);
3941 				}
3942 				return (DDI_FAILURE);
3943 			}
3944 			intr_ctx++;
3945 		}
3946 		break;
3947 
3948 	case DDI_INTR_TYPE_MSI:
3949 		/*
3950 		 * Add interrupt handlers for the only vector
3951 		 */
3952 		ql_atomic_set_32(&intr_ctx->irq_cnt, value);
3953 
3954 		rc = ddi_intr_add_handler(qlge->htable[vector],
3955 		    ql_isr,
3956 		    (caddr_t)&qlge->rx_ring[0], NULL);
3957 
3958 		if (rc != DDI_SUCCESS) {
3959 			QL_PRINT(DBG_INIT,
3960 			    ("Add MSI interrupt handler failed: %d\n", rc));
3961 			return (DDI_FAILURE);
3962 		}
3963 		break;
3964 
3965 	case DDI_INTR_TYPE_FIXED:
3966 		/*
3967 		 * Add interrupt handlers for the only vector
3968 		 */
3969 		ql_atomic_set_32(&intr_ctx->irq_cnt, value);
3970 
3971 		rc = ddi_intr_add_handler(qlge->htable[vector],
3972 		    ql_isr,
3973 		    (caddr_t)&qlge->rx_ring[0], NULL);
3974 
3975 		if (rc != DDI_SUCCESS) {
3976 			QL_PRINT(DBG_INIT,
3977 			    ("Add legacy interrupt handler failed: %d\n", rc));
3978 			return (DDI_FAILURE);
3979 		}
3980 		break;
3981 
3982 	default:
3983 		return (DDI_FAILURE);
3984 	}
3985 
3986 	/* Enable interrupts */
3987 	/* Block enable */
3988 	if (qlge->intr_cap & DDI_INTR_FLAG_BLOCK) {
3989 		QL_PRINT(DBG_INIT, ("Block enabling %d interrupt(s)\n",
3990 		    qlge->intr_cnt));
3991 		(void) ddi_intr_block_enable(qlge->htable, qlge->intr_cnt);
3992 	} else { /* Non block enable */
3993 		for (i = 0; i < qlge->intr_cnt; i++) {
3994 			QL_PRINT(DBG_INIT, ("Non Block Enabling interrupt %d "
3995 			    "handle 0x%x\n", i, qlge->htable[i]));
3996 			(void) ddi_intr_enable(qlge->htable[i]);
3997 		}
3998 	}
3999 	qlge->sequence |= INIT_INTR_ENABLED;
4000 
4001 	return (DDI_SUCCESS);
4002 }
4003 
4004 /*
4005  * Here we build the intr_ctx structures based on
4006  * our rx_ring count and intr vector count.
4007  * The intr_ctx structure is used to hook each vector
4008  * to possibly different handlers.
4009  */
4010 static void
4011 ql_resolve_queues_to_irqs(qlge_t *qlge)
4012 {
4013 	int i = 0;
4014 	struct intr_ctx *intr_ctx = &qlge->intr_ctx[0];
4015 
4016 	if (qlge->intr_type == DDI_INTR_TYPE_MSIX) {
4017 		/*
4018 		 * Each rx_ring has its own intr_ctx since we
4019 		 * have separate vectors for each queue.
4020 		 * This only true when MSI-X is enabled.
4021 		 */
4022 		for (i = 0; i < qlge->intr_cnt; i++, intr_ctx++) {
4023 			qlge->rx_ring[i].irq = i;
4024 			intr_ctx->intr = i;
4025 			intr_ctx->qlge = qlge;
4026 
4027 			/*
4028 			 * We set up each vectors enable/disable/read bits so
4029 			 * there's no bit/mask calculations in critical path.
4030 			 */
4031 			intr_ctx->intr_en_mask =
4032 			    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4033 			    INTR_EN_TYPE_ENABLE | INTR_EN_IHD_MASK |
4034 			    INTR_EN_IHD | i;
4035 			intr_ctx->intr_dis_mask =
4036 			    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4037 			    INTR_EN_TYPE_DISABLE | INTR_EN_IHD_MASK |
4038 			    INTR_EN_IHD | i;
4039 			intr_ctx->intr_read_mask =
4040 			    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4041 			    INTR_EN_TYPE_READ | INTR_EN_IHD_MASK | INTR_EN_IHD
4042 			    | i;
4043 
4044 			if (i == 0) {
4045 				/*
4046 				 * Default queue handles bcast/mcast plus
4047 				 * async events.
4048 				 */
4049 				intr_ctx->handler = ql_isr;
4050 			} else if (qlge->rx_ring[i].type == TX_Q) {
4051 				/*
4052 				 * Outbound queue is for outbound completions
4053 				 * only.
4054 				 */
4055 				if (qlge->isr_stride)
4056 					intr_ctx->handler = ql_msix_isr;
4057 				else
4058 					intr_ctx->handler = ql_msix_tx_isr;
4059 			} else {
4060 				/*
4061 				 * Inbound queues handle unicast frames only.
4062 				 */
4063 				if (qlge->isr_stride)
4064 					intr_ctx->handler = ql_msix_isr;
4065 				else
4066 					intr_ctx->handler = ql_msix_rx_isr;
4067 			}
4068 		}
4069 		i = qlge->intr_cnt;
4070 		for (; i < qlge->rx_ring_count; i++, intr_ctx++) {
4071 			int iv = i - qlge->isr_stride;
4072 			qlge->rx_ring[i].irq = iv;
4073 			intr_ctx->intr = iv;
4074 			intr_ctx->qlge = qlge;
4075 
4076 			/*
4077 			 * We set up each vectors enable/disable/read bits so
4078 			 * there's no bit/mask calculations in critical path.
4079 			 */
4080 			intr_ctx->intr_en_mask =
4081 			    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4082 			    INTR_EN_TYPE_ENABLE | INTR_EN_IHD_MASK |
4083 			    INTR_EN_IHD | iv;
4084 			intr_ctx->intr_dis_mask =
4085 			    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4086 			    INTR_EN_TYPE_DISABLE | INTR_EN_IHD_MASK |
4087 			    INTR_EN_IHD | iv;
4088 			intr_ctx->intr_read_mask =
4089 			    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4090 			    INTR_EN_TYPE_READ | INTR_EN_IHD_MASK | INTR_EN_IHD
4091 			    | iv;
4092 
4093 			if (qlge->rx_ring[i].type == TX_Q) {
4094 				/*
4095 				 * Outbound queue is for outbound completions
4096 				 * only.
4097 				 */
4098 				intr_ctx->handler = ql_msix_isr;
4099 			} else {
4100 				/*
4101 				 * Inbound queues handle unicast frames only.
4102 				 */
4103 				intr_ctx->handler = ql_msix_rx_isr;
4104 			}
4105 		}
4106 	} else {
4107 		/*
4108 		 * All rx_rings use the same intr_ctx since
4109 		 * there is only one vector.
4110 		 */
4111 		intr_ctx->intr = 0;
4112 		intr_ctx->qlge = qlge;
4113 		/*
4114 		 * We set up each vectors enable/disable/read bits so
4115 		 * there's no bit/mask calculations in the critical path.
4116 		 */
4117 		intr_ctx->intr_en_mask =
4118 		    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4119 		    INTR_EN_TYPE_ENABLE;
4120 		intr_ctx->intr_dis_mask =
4121 		    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4122 		    INTR_EN_TYPE_DISABLE;
4123 		intr_ctx->intr_read_mask =
4124 		    INTR_EN_TYPE_MASK | INTR_EN_INTR_MASK |
4125 		    INTR_EN_TYPE_READ;
4126 		/*
4127 		 * Single interrupt means one handler for all rings.
4128 		 */
4129 		intr_ctx->handler = ql_isr;
4130 		for (i = 0; i < qlge->rx_ring_count; i++)
4131 			qlge->rx_ring[i].irq = 0;
4132 	}
4133 }
4134 
4135 
4136 /*
4137  * Free allocated interrupts.
4138  */
4139 static void
4140 ql_free_irq_vectors(qlge_t *qlge)
4141 {
4142 	int i;
4143 	int rc;
4144 
4145 	if (qlge->sequence & INIT_INTR_ENABLED) {
4146 		/* Disable all interrupts */
4147 		if (qlge->intr_cap & DDI_INTR_FLAG_BLOCK) {
4148 			/* Call ddi_intr_block_disable() */
4149 			(void) ddi_intr_block_disable(qlge->htable,
4150 			    qlge->intr_cnt);
4151 		} else {
4152 			for (i = 0; i < qlge->intr_cnt; i++) {
4153 				(void) ddi_intr_disable(qlge->htable[i]);
4154 			}
4155 		}
4156 
4157 		qlge->sequence &= ~INIT_INTR_ENABLED;
4158 	}
4159 
4160 	for (i = 0; i < qlge->intr_cnt; i++) {
4161 
4162 		if (qlge->sequence & INIT_ADD_INTERRUPT)
4163 			(void) ddi_intr_remove_handler(qlge->htable[i]);
4164 
4165 		if (qlge->sequence & INIT_INTR_ALLOC) {
4166 			rc = ddi_intr_free(qlge->htable[i]);
4167 			if (rc != DDI_SUCCESS) {
4168 				/* EMPTY */
4169 				QL_PRINT(DBG_INIT, ("Free intr failed: %d",
4170 				    rc));
4171 			}
4172 		}
4173 	}
4174 	if (qlge->sequence & INIT_INTR_ALLOC)
4175 		qlge->sequence &= ~INIT_INTR_ALLOC;
4176 
4177 	if (qlge->sequence & INIT_ADD_INTERRUPT)
4178 		qlge->sequence &= ~INIT_ADD_INTERRUPT;
4179 
4180 	if (qlge->htable) {
4181 		kmem_free(qlge->htable, qlge->intr_size);
4182 		qlge->htable = NULL;
4183 	}
4184 }
4185 
4186 /*
4187  * Allocate interrupt vectors
4188  * For legacy and MSI, only 1 handle is needed.
4189  * For MSI-X, if fewer than 2 vectors are available, return failure.
4190  * Upon success, this maps the vectors to rx and tx rings for
4191  * interrupts.
4192  */
4193 static int
4194 ql_request_irq_vectors(qlge_t *qlge, int intr_type)
4195 {
4196 	dev_info_t *devinfo;
4197 	uint32_t request, orig;
4198 	int count, avail, actual;
4199 	int minimum;
4200 	int rc;
4201 
4202 	devinfo = qlge->dip;
4203 
4204 	switch (intr_type) {
4205 	case DDI_INTR_TYPE_FIXED:
4206 		request = 1;	/* Request 1 legacy interrupt handle */
4207 		minimum = 1;
4208 		QL_PRINT(DBG_INIT, ("interrupt type: legacy\n"));
4209 		break;
4210 
4211 	case DDI_INTR_TYPE_MSI:
4212 		request = 1;	/* Request 1 MSI interrupt handle */
4213 		minimum = 1;
4214 		QL_PRINT(DBG_INIT, ("interrupt type: MSI\n"));
4215 		break;
4216 
4217 	case DDI_INTR_TYPE_MSIX:
4218 		/*
4219 		 * Ideal number of vectors for the adapter is
4220 		 * # rss rings + tx completion rings for default completion
4221 		 * queue.
4222 		 */
4223 		request = qlge->rx_ring_count;
4224 
4225 		orig = request;
4226 		if (request > (MAX_RX_RINGS))
4227 			request = MAX_RX_RINGS;
4228 		minimum = 2;
4229 		QL_PRINT(DBG_INIT, ("interrupt type: MSI-X\n"));
4230 		break;
4231 
4232 	default:
4233 		QL_PRINT(DBG_INIT, ("Invalid parameter\n"));
4234 		return (DDI_FAILURE);
4235 	}
4236 
4237 	QL_PRINT(DBG_INIT, ("interrupt handles requested: %d  minimum: %d\n",
4238 	    request, minimum));
4239 
4240 	/*
4241 	 * Get number of supported interrupts
4242 	 */
4243 	rc = ddi_intr_get_nintrs(devinfo, intr_type, &count);
4244 	if ((rc != DDI_SUCCESS) || (count < minimum)) {
4245 		QL_PRINT(DBG_INIT, ("Get interrupt number failed. Return: %d, "
4246 		    "count: %d\n", rc, count));
4247 		return (DDI_FAILURE);
4248 	}
4249 	QL_PRINT(DBG_INIT, ("interrupts supported: %d\n", count));
4250 
4251 	/*
4252 	 * Get number of available interrupts
4253 	 */
4254 	rc = ddi_intr_get_navail(devinfo, intr_type, &avail);
4255 	if ((rc != DDI_SUCCESS) || (avail < minimum)) {
4256 		QL_PRINT(DBG_INIT,
4257 		    ("Get interrupt available number failed. Return:"
4258 		    " %d, available: %d\n", rc, avail));
4259 		return (DDI_FAILURE);
4260 	}
4261 	QL_PRINT(DBG_INIT, ("interrupts available: %d\n", avail));
4262 
4263 	if (avail < request) {
4264 		QL_PRINT(DBG_INIT, ("Request %d handles, %d available\n",
4265 		    request, avail));
4266 		request = avail;
4267 	}
4268 
4269 	actual = 0;
4270 	qlge->intr_cnt = 0;
4271 
4272 	/*
4273 	 * Allocate an array of interrupt handles
4274 	 */
4275 	qlge->intr_size = (size_t)(request * sizeof (ddi_intr_handle_t));
4276 	qlge->htable = kmem_alloc(qlge->intr_size, KM_SLEEP);
4277 
4278 	rc = ddi_intr_alloc(devinfo, qlge->htable, intr_type, 0,
4279 	    (int)request, &actual, DDI_INTR_ALLOC_NORMAL);
4280 	if (rc != DDI_SUCCESS) {
4281 		cmn_err(CE_WARN, "%s(%d) Allocate interrupts failed. return:"
4282 		    " %d, request: %d, actual: %d",
4283 		    __func__, qlge->instance, rc, request, actual);
4284 		goto ql_intr_alloc_fail;
4285 	}
4286 	qlge->intr_cnt = actual;
4287 
4288 	qlge->sequence |= INIT_INTR_ALLOC;
4289 
4290 	/*
4291 	 * If the actual number of vectors is less than the minumum
4292 	 * then fail.
4293 	 */
4294 	if (actual < minimum) {
4295 		cmn_err(CE_WARN,
4296 		    "Insufficient interrupt handles available: %d", actual);
4297 		goto ql_intr_alloc_fail;
4298 	}
4299 
4300 	/*
4301 	 * For MSI-X, actual might force us to reduce number of tx & rx rings
4302 	 */
4303 	if ((intr_type == DDI_INTR_TYPE_MSIX) && (orig > actual)) {
4304 		if (actual >= (orig / 2)) {
4305 			count = orig / 2;
4306 			qlge->rss_ring_count = count;
4307 			qlge->tx_ring_count = count;
4308 			qlge->isr_stride = count;
4309 		} else if (actual >= (orig / 4)) {
4310 			count = orig / 4;
4311 			qlge->rss_ring_count = count;
4312 			qlge->tx_ring_count = count;
4313 			qlge->isr_stride = count;
4314 		} else if (actual >= (orig / 8)) {
4315 			count = orig / 8;
4316 			qlge->rss_ring_count = count;
4317 			qlge->tx_ring_count = count;
4318 			qlge->isr_stride = count;
4319 		} else if (actual < MAX_RX_RINGS) {
4320 			qlge->tx_ring_count = 1;
4321 			qlge->rss_ring_count = actual - 1;
4322 		}
4323 		qlge->intr_cnt = count;
4324 		qlge->rx_ring_count = qlge->tx_ring_count +
4325 		    qlge->rss_ring_count;
4326 	}
4327 	cmn_err(CE_NOTE, "!qlge(%d) tx %d, rss %d, stride %d\n", qlge->instance,
4328 	    qlge->tx_ring_count, qlge->rss_ring_count, qlge->isr_stride);
4329 
4330 	/*
4331 	 * Get priority for first vector, assume remaining are all the same
4332 	 */
4333 	rc = ddi_intr_get_pri(qlge->htable[0], &qlge->intr_pri);
4334 	if (rc != DDI_SUCCESS) {
4335 		QL_PRINT(DBG_INIT, ("Get interrupt priority failed: %d\n", rc));
4336 		goto ql_intr_alloc_fail;
4337 	}
4338 
4339 	rc = ddi_intr_get_cap(qlge->htable[0], &qlge->intr_cap);
4340 	if (rc != DDI_SUCCESS) {
4341 		QL_PRINT(DBG_INIT, ("Get interrupt cap failed: %d\n", rc));
4342 		goto ql_intr_alloc_fail;
4343 	}
4344 
4345 	qlge->intr_type = intr_type;
4346 
4347 	return (DDI_SUCCESS);
4348 
4349 ql_intr_alloc_fail:
4350 	ql_free_irq_vectors(qlge);
4351 
4352 	return (DDI_FAILURE);
4353 }
4354 
4355 /*
4356  * Allocate interrupt vector(s) for one of the following interrupt types, MSI-X,
4357  * MSI or Legacy. In MSI and Legacy modes we only support a single receive and
4358  * transmit queue.
4359  */
4360 int
4361 ql_alloc_irqs(qlge_t *qlge)
4362 {
4363 	int intr_types;
4364 	int rval;
4365 
4366 	/*
4367 	 * Get supported interrupt types
4368 	 */
4369 	if (ddi_intr_get_supported_types(qlge->dip, &intr_types)
4370 	    != DDI_SUCCESS) {
4371 		cmn_err(CE_WARN, "%s(%d):ddi_intr_get_supported_types failed",
4372 		    __func__, qlge->instance);
4373 
4374 		return (DDI_FAILURE);
4375 	}
4376 
4377 	QL_PRINT(DBG_INIT, ("%s(%d) Interrupt types supported %d\n",
4378 	    __func__, qlge->instance, intr_types));
4379 
4380 	/* Install MSI-X interrupts */
4381 	if ((intr_types & DDI_INTR_TYPE_MSIX) != 0) {
4382 		QL_PRINT(DBG_INIT, ("%s(%d) MSI-X interrupt supported %d\n",
4383 		    __func__, qlge->instance, intr_types));
4384 		rval = ql_request_irq_vectors(qlge, DDI_INTR_TYPE_MSIX);
4385 		if (rval == DDI_SUCCESS) {
4386 			return (rval);
4387 		}
4388 		QL_PRINT(DBG_INIT, ("%s(%d) MSI-X interrupt allocation failed,"
4389 		    " trying MSI interrupts ...\n", __func__, qlge->instance));
4390 	}
4391 
4392 	/*
4393 	 * We will have 2 completion queues in MSI / Legacy mode,
4394 	 * Queue 0 for default completions
4395 	 * Queue 1 for transmit completions
4396 	 */
4397 	qlge->rss_ring_count = 1; /* Default completion queue (0) for all */
4398 	qlge->tx_ring_count = 1; /* Single tx completion queue */
4399 	qlge->rx_ring_count = qlge->tx_ring_count + qlge->rss_ring_count;
4400 
4401 	QL_PRINT(DBG_INIT, ("%s(%d) Falling back to single completion queue \n",
4402 	    __func__, qlge->instance));
4403 	/*
4404 	 * Add the h/w interrupt handler and initialise mutexes
4405 	 */
4406 	rval = DDI_FAILURE;
4407 
4408 	/*
4409 	 * If OS supports MSIX interrupt but fails to allocate, then try
4410 	 * MSI interrupt. If MSI interrupt allocation fails also, then roll
4411 	 * back to fixed interrupt.
4412 	 */
4413 	if (intr_types & DDI_INTR_TYPE_MSI) {
4414 		rval = ql_request_irq_vectors(qlge, DDI_INTR_TYPE_MSI);
4415 		if (rval == DDI_SUCCESS) {
4416 			qlge->intr_type = DDI_INTR_TYPE_MSI;
4417 			QL_PRINT(DBG_INIT, ("%s(%d) use MSI Interrupt \n",
4418 			    __func__, qlge->instance));
4419 		}
4420 	}
4421 
4422 	/* Try Fixed interrupt Legacy mode */
4423 	if (rval != DDI_SUCCESS) {
4424 		rval = ql_request_irq_vectors(qlge, DDI_INTR_TYPE_FIXED);
4425 		if (rval != DDI_SUCCESS) {
4426 			cmn_err(CE_WARN, "%s(%d):Legacy mode interrupt "
4427 			    "allocation failed",
4428 			    __func__, qlge->instance);
4429 		} else {
4430 			qlge->intr_type = DDI_INTR_TYPE_FIXED;
4431 			QL_PRINT(DBG_INIT, ("%s(%d) use Fixed Interrupt \n",
4432 			    __func__, qlge->instance));
4433 		}
4434 	}
4435 
4436 	return (rval);
4437 }
4438 
4439 static void
4440 ql_free_rx_tx_locks(qlge_t *qlge)
4441 {
4442 	int i;
4443 	struct rx_ring *rx_ring;
4444 	struct tx_ring *tx_ring;
4445 
4446 	for (i = 0; i < qlge->tx_ring_count; i++) {
4447 		tx_ring = &qlge->tx_ring[i];
4448 		mutex_destroy(&tx_ring->tx_lock);
4449 	}
4450 
4451 	for (i = 0; i < qlge->rx_ring_count; i++) {
4452 		rx_ring = &qlge->rx_ring[i];
4453 		mutex_destroy(&rx_ring->rx_lock);
4454 		mutex_destroy(&rx_ring->sbq_lock);
4455 		mutex_destroy(&rx_ring->lbq_lock);
4456 	}
4457 }
4458 
4459 /*
4460  * Frees all resources allocated during attach.
4461  *
4462  * Input:
4463  * dip = pointer to device information structure.
4464  * sequence = bits indicating resources to free.
4465  *
4466  * Context:
4467  * Kernel context.
4468  */
4469 static void
4470 ql_free_resources(qlge_t *qlge)
4471 {
4472 
4473 	/* Disable driver timer */
4474 	ql_stop_timer(qlge);
4475 
4476 	if (qlge->sequence & INIT_MAC_REGISTERED) {
4477 		(void) mac_unregister(qlge->mh);
4478 		qlge->sequence &= ~INIT_MAC_REGISTERED;
4479 	}
4480 
4481 	if (qlge->sequence & INIT_MAC_ALLOC) {
4482 		/* Nothing to do, macp is already freed */
4483 		qlge->sequence &= ~INIT_MAC_ALLOC;
4484 	}
4485 
4486 	if (qlge->sequence & INIT_PCI_CONFIG_SETUP) {
4487 		pci_config_teardown(&qlge->pci_handle);
4488 		qlge->sequence &= ~INIT_PCI_CONFIG_SETUP;
4489 	}
4490 
4491 	if (qlge->sequence & INIT_INTR_ALLOC) {
4492 		ql_free_irq_vectors(qlge);
4493 		qlge->sequence &= ~INIT_ADD_INTERRUPT;
4494 	}
4495 
4496 	if (qlge->sequence & INIT_ADD_SOFT_INTERRUPT) {
4497 		(void) ddi_intr_remove_softint(qlge->mpi_event_intr_hdl);
4498 		(void) ddi_intr_remove_softint(qlge->mpi_reset_intr_hdl);
4499 		(void) ddi_intr_remove_softint(qlge->asic_reset_intr_hdl);
4500 		qlge->sequence &= ~INIT_ADD_SOFT_INTERRUPT;
4501 	}
4502 
4503 	if (qlge->sequence & INIT_KSTATS) {
4504 		ql_fini_kstats(qlge);
4505 		qlge->sequence &= ~INIT_KSTATS;
4506 	}
4507 
4508 	if (qlge->sequence & INIT_MUTEX) {
4509 		mutex_destroy(&qlge->gen_mutex);
4510 		mutex_destroy(&qlge->hw_mutex);
4511 		mutex_destroy(&qlge->mbx_mutex);
4512 		cv_destroy(&qlge->cv_mbx_intr);
4513 		qlge->sequence &= ~INIT_MUTEX;
4514 	}
4515 
4516 	if (qlge->sequence & INIT_LOCKS_CREATED) {
4517 		ql_free_rx_tx_locks(qlge);
4518 		qlge->sequence &= ~INIT_LOCKS_CREATED;
4519 	}
4520 
4521 	if (qlge->sequence & INIT_MEMORY_ALLOC) {
4522 		ql_free_mem_resources(qlge);
4523 		qlge->sequence &= ~INIT_MEMORY_ALLOC;
4524 	}
4525 
4526 	if (qlge->sequence & INIT_REGS_SETUP) {
4527 		ddi_regs_map_free(&qlge->dev_handle);
4528 		qlge->sequence &= ~INIT_REGS_SETUP;
4529 	}
4530 
4531 	if (qlge->sequence & INIT_DOORBELL_REGS_SETUP) {
4532 		ddi_regs_map_free(&qlge->dev_doorbell_reg_handle);
4533 		qlge->sequence &= ~INIT_DOORBELL_REGS_SETUP;
4534 	}
4535 
4536 	/*
4537 	 * free flash flt table that allocated in attach stage
4538 	 */
4539 	if ((qlge->flt.ql_flt_entry_ptr != NULL)&&
4540 	    (qlge->flt.header.length != 0)) {
4541 		kmem_free(qlge->flt.ql_flt_entry_ptr, qlge->flt.header.length);
4542 		qlge->flt.ql_flt_entry_ptr = NULL;
4543 	}
4544 
4545 	if (qlge->sequence & INIT_FM) {
4546 		ql_fm_fini(qlge);
4547 		qlge->sequence &= ~INIT_FM;
4548 	}
4549 
4550 	ddi_prop_remove_all(qlge->dip);
4551 	ddi_set_driver_private(qlge->dip, NULL);
4552 
4553 	/* finally, free qlge structure */
4554 	if (qlge->sequence & INIT_SOFTSTATE_ALLOC) {
4555 		kmem_free(qlge, sizeof (qlge_t));
4556 	}
4557 }
4558 
4559 /*
4560  * Set promiscuous mode of the driver
4561  * Caller must catch HW_LOCK
4562  */
4563 void
4564 ql_set_promiscuous(qlge_t *qlge, int mode)
4565 {
4566 	if (mode) {
4567 		(void) ql_set_routing_reg(qlge, RT_IDX_PROMISCUOUS_SLOT,
4568 		    RT_IDX_VALID, 1);
4569 	} else {
4570 		(void) ql_set_routing_reg(qlge, RT_IDX_PROMISCUOUS_SLOT,
4571 		    RT_IDX_VALID, 0);
4572 	}
4573 }
4574 /*
4575  * Write 'data1' to Mac Protocol Address Index Register and
4576  * 'data2' to Mac Protocol Address Data Register
4577  *  Assuming that the Mac Protocol semaphore lock has been acquired.
4578  */
4579 static int
4580 ql_write_mac_proto_regs(qlge_t *qlge, uint32_t data1, uint32_t data2)
4581 {
4582 	int return_value = DDI_SUCCESS;
4583 
4584 	if (ql_wait_reg_bit(qlge, REG_MAC_PROTOCOL_ADDRESS_INDEX,
4585 	    MAC_PROTOCOL_ADDRESS_INDEX_MW, BIT_SET, 5) != DDI_SUCCESS) {
4586 		cmn_err(CE_WARN, "Wait for MAC_PROTOCOL Address Register "
4587 		    "timeout.");
4588 		return_value = DDI_FAILURE;
4589 		goto out;
4590 	}
4591 	ql_write_reg(qlge, REG_MAC_PROTOCOL_ADDRESS_INDEX /* A8 */, data1);
4592 	ql_write_reg(qlge, REG_MAC_PROTOCOL_DATA /* 0xAC */, data2);
4593 out:
4594 	return (return_value);
4595 }
4596 /*
4597  * Enable the 'index'ed multicast address in the host memory's multicast_list
4598  */
4599 int
4600 ql_add_multicast_address(qlge_t *qlge, int index)
4601 {
4602 	int rtn_val = DDI_FAILURE;
4603 	uint32_t offset;
4604 	uint32_t value1, value2;
4605 
4606 	/* Acquire the required semaphore */
4607 	if (ql_sem_spinlock(qlge, QL_MAC_PROTOCOL_SEM_MASK) != DDI_SUCCESS) {
4608 		return (rtn_val);
4609 	}
4610 
4611 	/* Program Offset0 - lower 32 bits of the MAC address */
4612 	offset = 0;
4613 	value1 = MAC_PROTOCOL_ADDRESS_ENABLE | MAC_PROTOCOL_TYPE_MULTICAST |
4614 	    (index << 4) | offset;
4615 	value2 = ((qlge->multicast_list[index].addr.ether_addr_octet[2] << 24)
4616 	    |(qlge->multicast_list[index].addr.ether_addr_octet[3] << 16)
4617 	    |(qlge->multicast_list[index].addr.ether_addr_octet[4] << 8)
4618 	    |(qlge->multicast_list[index].addr.ether_addr_octet[5]));
4619 	if (ql_write_mac_proto_regs(qlge, value1, value2) != DDI_SUCCESS)
4620 		goto out;
4621 
4622 	/* Program offset1: upper 16 bits of the MAC address */
4623 	offset = 1;
4624 	value1 = MAC_PROTOCOL_ADDRESS_ENABLE | MAC_PROTOCOL_TYPE_MULTICAST |
4625 	    (index<<4) | offset;
4626 	value2 = ((qlge->multicast_list[index].addr.ether_addr_octet[0] << 8)
4627 	    |qlge->multicast_list[index].addr.ether_addr_octet[1]);
4628 	if (ql_write_mac_proto_regs(qlge, value1, value2) != DDI_SUCCESS) {
4629 		goto out;
4630 	}
4631 	rtn_val = DDI_SUCCESS;
4632 out:
4633 	ql_sem_unlock(qlge, QL_MAC_PROTOCOL_SEM_MASK);
4634 	return (rtn_val);
4635 }
4636 
4637 /*
4638  * Disable the 'index'ed multicast address in the host memory's multicast_list
4639  */
4640 int
4641 ql_remove_multicast_address(qlge_t *qlge, int index)
4642 {
4643 	int rtn_val = DDI_FAILURE;
4644 	uint32_t offset;
4645 	uint32_t value1, value2;
4646 
4647 	/* Acquire the required semaphore */
4648 	if (ql_sem_spinlock(qlge, QL_MAC_PROTOCOL_SEM_MASK) != DDI_SUCCESS) {
4649 		return (rtn_val);
4650 	}
4651 	/* Program Offset0 - lower 32 bits of the MAC address */
4652 	offset = 0;
4653 	value1 = (MAC_PROTOCOL_TYPE_MULTICAST | offset)|(index<<4);
4654 	value2 =
4655 	    ((qlge->multicast_list[index].addr.ether_addr_octet[2] << 24)
4656 	    |(qlge->multicast_list[index].addr.ether_addr_octet[3] << 16)
4657 	    |(qlge->multicast_list[index].addr.ether_addr_octet[4] << 8)
4658 	    |(qlge->multicast_list[index].addr.ether_addr_octet[5]));
4659 	if (ql_write_mac_proto_regs(qlge, value1, value2) != DDI_SUCCESS) {
4660 		goto out;
4661 	}
4662 	/* Program offset1: upper 16 bits of the MAC address */
4663 	offset = 1;
4664 	value1 = (MAC_PROTOCOL_TYPE_MULTICAST | offset)|(index<<4);
4665 	value2 = 0;
4666 	if (ql_write_mac_proto_regs(qlge, value1, value2) != DDI_SUCCESS) {
4667 		goto out;
4668 	}
4669 	rtn_val = DDI_SUCCESS;
4670 out:
4671 	ql_sem_unlock(qlge, QL_MAC_PROTOCOL_SEM_MASK);
4672 	return (rtn_val);
4673 }
4674 
4675 /*
4676  * Add a new multicast address to the list of supported list
4677  * This API is called after OS called gld_set_multicast (GLDv2)
4678  * or m_multicst (GLDv3)
4679  *
4680  * Restriction:
4681  * The number of maximum multicast address is limited by hardware.
4682  */
4683 int
4684 ql_add_to_multicast_list(qlge_t *qlge, uint8_t *ep)
4685 {
4686 	uint32_t index = qlge->multicast_list_count;
4687 	int rval = DDI_SUCCESS;
4688 	int status;
4689 
4690 	if ((ep[0] & 01) == 0) {
4691 		rval = EINVAL;
4692 		goto exit;
4693 	}
4694 
4695 	/* if there is an availabe space in multicast_list, then add it */
4696 	if (index < MAX_MULTICAST_LIST_SIZE) {
4697 		bcopy(ep, qlge->multicast_list[index].addr.ether_addr_octet,
4698 		    ETHERADDRL);
4699 		/* increment the total number of addresses in multicast list */
4700 		(void) ql_add_multicast_address(qlge, index);
4701 		qlge->multicast_list_count++;
4702 		QL_PRINT(DBG_GLD,
4703 		    ("%s(%d): added to index of multicast list= 0x%x, "
4704 		    "total %d\n", __func__, qlge->instance, index,
4705 		    qlge->multicast_list_count));
4706 
4707 		if (index > MAX_MULTICAST_HW_SIZE) {
4708 			if (!qlge->multicast_promisc) {
4709 				status = ql_set_routing_reg(qlge,
4710 				    RT_IDX_ALLMULTI_SLOT,
4711 				    RT_IDX_MCAST, 1);
4712 				if (status) {
4713 					cmn_err(CE_WARN,
4714 					    "Failed to init routing reg "
4715 					    "for mcast promisc mode.");
4716 					rval = ENOENT;
4717 					goto exit;
4718 				}
4719 				qlge->multicast_promisc = B_TRUE;
4720 			}
4721 		}
4722 	} else {
4723 		rval = ENOENT;
4724 	}
4725 exit:
4726 	return (rval);
4727 }
4728 
4729 /*
4730  * Remove an old multicast address from the list of supported multicast
4731  * addresses. This API is called after OS called gld_set_multicast (GLDv2)
4732  * or m_multicst (GLDv3)
4733  * The number of maximum multicast address is limited by hardware.
4734  */
4735 int
4736 ql_remove_from_multicast_list(qlge_t *qlge, uint8_t *ep)
4737 {
4738 	uint32_t total = qlge->multicast_list_count;
4739 	int i = 0;
4740 	int rmv_index = 0;
4741 	size_t length = sizeof (ql_multicast_addr);
4742 	int status;
4743 
4744 	for (i = 0; i < total; i++) {
4745 		if (bcmp(ep, &qlge->multicast_list[i].addr, ETHERADDRL) != 0) {
4746 			continue;
4747 		}
4748 
4749 		rmv_index = i;
4750 		/* block move the reset of other multicast address forward */
4751 		length = ((total -1) -i) * sizeof (ql_multicast_addr);
4752 		if (length > 0) {
4753 			bcopy(&qlge->multicast_list[i+1],
4754 			    &qlge->multicast_list[i], length);
4755 		}
4756 		qlge->multicast_list_count--;
4757 		if (qlge->multicast_list_count <= MAX_MULTICAST_HW_SIZE) {
4758 			/*
4759 			 * there is a deletion in multicast list table,
4760 			 * re-enable them
4761 			 */
4762 			for (i = rmv_index; i < qlge->multicast_list_count;
4763 			    i++) {
4764 				(void) ql_add_multicast_address(qlge, i);
4765 			}
4766 			/* and disable the last one */
4767 			(void) ql_remove_multicast_address(qlge, i);
4768 
4769 			/* disable multicast promiscuous mode */
4770 			if (qlge->multicast_promisc) {
4771 				status = ql_set_routing_reg(qlge,
4772 				    RT_IDX_ALLMULTI_SLOT,
4773 				    RT_IDX_MCAST, 0);
4774 				if (status) {
4775 					cmn_err(CE_WARN,
4776 					    "Failed to init routing reg for "
4777 					    "mcast promisc mode.");
4778 					goto exit;
4779 				}
4780 				/* write to config register */
4781 				qlge->multicast_promisc = B_FALSE;
4782 			}
4783 		}
4784 		break;
4785 	}
4786 exit:
4787 	return (DDI_SUCCESS);
4788 }
4789 
4790 /*
4791  * Read a XGMAC register
4792  */
4793 int
4794 ql_read_xgmac_reg(qlge_t *qlge, uint32_t addr, uint32_t *val)
4795 {
4796 	int rtn_val = DDI_FAILURE;
4797 
4798 	/* wait for XGMAC Address register RDY bit set */
4799 	if (ql_wait_reg_bit(qlge, REG_XGMAC_ADDRESS, XGMAC_ADDRESS_RDY,
4800 	    BIT_SET, 10) != DDI_SUCCESS) {
4801 		goto out;
4802 	}
4803 	/* start rx transaction */
4804 	ql_write_reg(qlge, REG_XGMAC_ADDRESS, addr|XGMAC_ADDRESS_READ_TRANSACT);
4805 
4806 	/*
4807 	 * wait for XGMAC Address register RDY bit set,
4808 	 * which indicates data is ready
4809 	 */
4810 	if (ql_wait_reg_bit(qlge, REG_XGMAC_ADDRESS, XGMAC_ADDRESS_RDY,
4811 	    BIT_SET, 10) != DDI_SUCCESS) {
4812 		goto out;
4813 	}
4814 	/* read data from XGAMC_DATA register */
4815 	*val = ql_read_reg(qlge, REG_XGMAC_DATA);
4816 	rtn_val = DDI_SUCCESS;
4817 out:
4818 	return (rtn_val);
4819 }
4820 
4821 /*
4822  * Implement checksum offload for IPv4 IP packets
4823  */
4824 static void
4825 ql_hw_csum_setup(qlge_t *qlge, uint32_t pflags, caddr_t bp,
4826     struct ob_mac_iocb_req *mac_iocb_ptr)
4827 {
4828 	struct ip *iphdr = NULL;
4829 	struct ether_header *ethhdr;
4830 	struct ether_vlan_header *ethvhdr;
4831 	struct tcphdr *tcp_hdr;
4832 	uint32_t etherType;
4833 	int mac_hdr_len, ip_hdr_len, tcp_udp_hdr_len;
4834 	int ip_hdr_off, tcp_udp_hdr_off, hdr_off;
4835 
4836 	ethhdr  = (struct ether_header *)((void *)bp);
4837 	ethvhdr = (struct ether_vlan_header *)((void *)bp);
4838 	/* Is this vlan packet? */
4839 	if (ntohs(ethvhdr->ether_tpid) == ETHERTYPE_VLAN) {
4840 		mac_hdr_len = sizeof (struct ether_vlan_header);
4841 		etherType = ntohs(ethvhdr->ether_type);
4842 	} else {
4843 		mac_hdr_len = sizeof (struct ether_header);
4844 		etherType = ntohs(ethhdr->ether_type);
4845 	}
4846 	/* Is this IPv4 or IPv6 packet? */
4847 	if (IPH_HDR_VERSION((ipha_t *)(void *)(bp+mac_hdr_len)) ==
4848 	    IPV4_VERSION) {
4849 		if (etherType == ETHERTYPE_IP /* 0800 */) {
4850 			iphdr = (struct ip *)(void *)(bp+mac_hdr_len);
4851 		} else {
4852 			/* EMPTY */
4853 			QL_PRINT(DBG_TX,
4854 			    ("%s(%d) : IPv4 None IP packet type 0x%x\n",
4855 			    __func__, qlge->instance, etherType));
4856 		}
4857 	}
4858 	/* ipV4 packets */
4859 	if (iphdr != NULL) {
4860 
4861 		ip_hdr_len = IPH_HDR_LENGTH(iphdr);
4862 		QL_PRINT(DBG_TX,
4863 		    ("%s(%d) : IPv4 header length using IPH_HDR_LENGTH:"
4864 		    " %d bytes \n", __func__, qlge->instance, ip_hdr_len));
4865 
4866 		ip_hdr_off = mac_hdr_len;
4867 		QL_PRINT(DBG_TX, ("%s(%d) : ip_hdr_len=%d\n",
4868 		    __func__, qlge->instance, ip_hdr_len));
4869 
4870 		mac_iocb_ptr->flag0 = (uint8_t)(mac_iocb_ptr->flag0 |
4871 		    OB_MAC_IOCB_REQ_IPv4);
4872 
4873 		if (pflags & HCK_IPV4_HDRCKSUM) {
4874 			QL_PRINT(DBG_TX, ("%s(%d) : Do IPv4 header checksum\n",
4875 			    __func__, qlge->instance));
4876 			mac_iocb_ptr->opcode = OPCODE_OB_MAC_OFFLOAD_IOCB;
4877 			mac_iocb_ptr->flag2 = (uint8_t)(mac_iocb_ptr->flag2 |
4878 			    OB_MAC_IOCB_REQ_IC);
4879 			iphdr->ip_sum = 0;
4880 			mac_iocb_ptr->hdr_off = (uint16_t)
4881 			    cpu_to_le16(ip_hdr_off);
4882 		}
4883 		if (pflags & HCK_FULLCKSUM) {
4884 			if (iphdr->ip_p == IPPROTO_TCP) {
4885 				tcp_hdr =
4886 				    (struct tcphdr *)(void *)
4887 				    ((uint8_t *)(void *)iphdr + ip_hdr_len);
4888 				QL_PRINT(DBG_TX, ("%s(%d) : Do TCP checksum\n",
4889 				    __func__, qlge->instance));
4890 				mac_iocb_ptr->opcode =
4891 				    OPCODE_OB_MAC_OFFLOAD_IOCB;
4892 				mac_iocb_ptr->flag1 =
4893 				    (uint8_t)(mac_iocb_ptr->flag1 |
4894 				    OB_MAC_IOCB_REQ_TC);
4895 				mac_iocb_ptr->flag2 =
4896 				    (uint8_t)(mac_iocb_ptr->flag2 |
4897 				    OB_MAC_IOCB_REQ_IC);
4898 				iphdr->ip_sum = 0;
4899 				tcp_udp_hdr_off = mac_hdr_len+ip_hdr_len;
4900 				tcp_udp_hdr_len = tcp_hdr->th_off*4;
4901 				QL_PRINT(DBG_TX, ("%s(%d): tcp header len:%d\n",
4902 				    __func__, qlge->instance, tcp_udp_hdr_len));
4903 				hdr_off = ip_hdr_off;
4904 				tcp_udp_hdr_off <<= 6;
4905 				hdr_off |= tcp_udp_hdr_off;
4906 				mac_iocb_ptr->hdr_off =
4907 				    (uint16_t)cpu_to_le16(hdr_off);
4908 				mac_iocb_ptr->protocol_hdr_len = (uint16_t)
4909 				    cpu_to_le16(mac_hdr_len + ip_hdr_len +
4910 				    tcp_udp_hdr_len);
4911 
4912 				/*
4913 				 * if the chip is unable to do pseudo header
4914 				 * cksum calculation, do it in then put the
4915 				 * result to the data passed to the chip
4916 				 */
4917 				if (qlge->cfg_flags &
4918 				    CFG_HW_UNABLE_PSEUDO_HDR_CKSUM) {
4919 					ql_pseudo_cksum((uint8_t *)iphdr);
4920 				}
4921 			} else if (iphdr->ip_p == IPPROTO_UDP) {
4922 				QL_PRINT(DBG_TX, ("%s(%d) : Do UDP checksum\n",
4923 				    __func__, qlge->instance));
4924 				mac_iocb_ptr->opcode =
4925 				    OPCODE_OB_MAC_OFFLOAD_IOCB;
4926 				mac_iocb_ptr->flag1 =
4927 				    (uint8_t)(mac_iocb_ptr->flag1 |
4928 				    OB_MAC_IOCB_REQ_UC);
4929 				mac_iocb_ptr->flag2 =
4930 				    (uint8_t)(mac_iocb_ptr->flag2 |
4931 				    OB_MAC_IOCB_REQ_IC);
4932 				iphdr->ip_sum = 0;
4933 				tcp_udp_hdr_off = mac_hdr_len + ip_hdr_len;
4934 				tcp_udp_hdr_len = sizeof (struct udphdr);
4935 				QL_PRINT(DBG_TX, ("%s(%d):udp header len:%d\n",
4936 				    __func__, qlge->instance, tcp_udp_hdr_len));
4937 				hdr_off = ip_hdr_off;
4938 				tcp_udp_hdr_off <<= 6;
4939 				hdr_off |= tcp_udp_hdr_off;
4940 				mac_iocb_ptr->hdr_off =
4941 				    (uint16_t)cpu_to_le16(hdr_off);
4942 				mac_iocb_ptr->protocol_hdr_len = (uint16_t)
4943 				    cpu_to_le16(mac_hdr_len + ip_hdr_len
4944 				    + tcp_udp_hdr_len);
4945 
4946 				/*
4947 				 * if the chip is unable to calculate pseudo
4948 				 * hdr cksum,do it in then put the result to
4949 				 * the data passed to the chip
4950 				 */
4951 				if (qlge->cfg_flags &
4952 				    CFG_HW_UNABLE_PSEUDO_HDR_CKSUM) {
4953 					ql_pseudo_cksum((uint8_t *)iphdr);
4954 				}
4955 			}
4956 		}
4957 	}
4958 }
4959 
4960 /*
4961  * For TSO/LSO:
4962  * MAC frame transmission with TCP large segment offload is performed in the
4963  * same way as the MAC frame transmission with checksum offload with the
4964  * exception that the maximum TCP segment size (MSS) must be specified to
4965  * allow the chip to segment the data into legal sized frames.
4966  * The host also needs to calculate a pseudo-header checksum over the
4967  * following fields:
4968  * Source IP Address, Destination IP Address, and the Protocol.
4969  * The TCP length is not included in the pseudo-header calculation.
4970  * The pseudo-header checksum is place in the TCP checksum field of the
4971  * prototype header.
4972  */
4973 static void
4974 ql_lso_pseudo_cksum(uint8_t *buf)
4975 {
4976 	uint32_t cksum;
4977 	uint16_t iphl;
4978 	uint16_t proto;
4979 
4980 	/*
4981 	 * Calculate the LSO pseudo-header checksum.
4982 	 */
4983 	iphl = (uint16_t)(4 * (buf[0] & 0xF));
4984 	cksum = proto = buf[9];
4985 	cksum += (((uint16_t)buf[12])<<8) + buf[13];
4986 	cksum += (((uint16_t)buf[14])<<8) + buf[15];
4987 	cksum += (((uint16_t)buf[16])<<8) + buf[17];
4988 	cksum += (((uint16_t)buf[18])<<8) + buf[19];
4989 	cksum = (cksum>>16) + (cksum & 0xFFFF);
4990 	cksum = (cksum>>16) + (cksum & 0xFFFF);
4991 
4992 	/*
4993 	 * Point it to the TCP/UDP header, and
4994 	 * update the checksum field.
4995 	 */
4996 	buf += iphl + ((proto == IPPROTO_TCP) ?
4997 	    TCP_CKSUM_OFFSET : UDP_CKSUM_OFFSET);
4998 
4999 	*(uint16_t *)(void *)buf = (uint16_t)htons((uint16_t)cksum);
5000 }
5001 
5002 /*
5003  * For IPv4 IP packets, distribute the tx packets evenly among tx rings
5004  */
5005 typedef	uint32_t	ub4; /* unsigned 4-byte quantities */
5006 typedef	uint8_t		ub1;
5007 
5008 #define	hashsize(n)	((ub4)1<<(n))
5009 #define	hashmask(n)	(hashsize(n)-1)
5010 
5011 #define	mix(a, b, c) \
5012 { \
5013 	a -= b; a -= c; a ^= (c>>13); \
5014 	b -= c; b -= a; b ^= (a<<8); \
5015 	c -= a; c -= b; c ^= (b>>13); \
5016 	a -= b; a -= c; a ^= (c>>12);  \
5017 	b -= c; b -= a; b ^= (a<<16); \
5018 	c -= a; c -= b; c ^= (b>>5); \
5019 	a -= b; a -= c; a ^= (c>>3);  \
5020 	b -= c; b -= a; b ^= (a<<10); \
5021 	c -= a; c -= b; c ^= (b>>15); \
5022 }
5023 
5024 ub4
5025 hash(ub1 *k, ub4 length, ub4 initval)
5026 {
5027 	ub4 a, b, c, len;
5028 
5029 	/* Set up the internal state */
5030 	len = length;
5031 	a = b = 0x9e3779b9;	/* the golden ratio; an arbitrary value */
5032 	c = initval;		/* the previous hash value */
5033 
5034 	/* handle most of the key */
5035 	while (len >= 12) {
5036 		a += (k[0] +((ub4)k[1]<<8) +((ub4)k[2]<<16) +((ub4)k[3]<<24));
5037 		b += (k[4] +((ub4)k[5]<<8) +((ub4)k[6]<<16) +((ub4)k[7]<<24));
5038 		c += (k[8] +((ub4)k[9]<<8) +((ub4)k[10]<<16)+((ub4)k[11]<<24));
5039 		mix(a, b, c);
5040 		k += 12;
5041 		len -= 12;
5042 	}
5043 
5044 	/* handle the last 11 bytes */
5045 	c += length;
5046 	/* all the case statements fall through */
5047 	switch (len) {
5048 		/* FALLTHRU */
5049 	case 11: c += ((ub4)k[10]<<24);
5050 		/* FALLTHRU */
5051 	case 10: c += ((ub4)k[9]<<16);
5052 		/* FALLTHRU */
5053 	case 9 : c += ((ub4)k[8]<<8);
5054 	/* the first byte of c is reserved for the length */
5055 		/* FALLTHRU */
5056 	case 8 : b += ((ub4)k[7]<<24);
5057 		/* FALLTHRU */
5058 	case 7 : b += ((ub4)k[6]<<16);
5059 		/* FALLTHRU */
5060 	case 6 : b += ((ub4)k[5]<<8);
5061 		/* FALLTHRU */
5062 	case 5 : b += k[4];
5063 		/* FALLTHRU */
5064 	case 4 : a += ((ub4)k[3]<<24);
5065 		/* FALLTHRU */
5066 	case 3 : a += ((ub4)k[2]<<16);
5067 		/* FALLTHRU */
5068 	case 2 : a += ((ub4)k[1]<<8);
5069 		/* FALLTHRU */
5070 	case 1 : a += k[0];
5071 	/* case 0: nothing left to add */
5072 	}
5073 	mix(a, b, c);
5074 	/* report the result */
5075 	return (c);
5076 }
5077 
5078 uint8_t
5079 ql_tx_hashing(qlge_t *qlge, caddr_t bp)
5080 {
5081 	struct ip *iphdr = NULL;
5082 	struct ether_header *ethhdr;
5083 	struct ether_vlan_header *ethvhdr;
5084 	struct tcphdr *tcp_hdr;
5085 	struct udphdr *udp_hdr;
5086 	uint32_t etherType;
5087 	int mac_hdr_len, ip_hdr_len;
5088 	uint32_t h = 0; /* 0 by default */
5089 	uint8_t tx_ring_id = 0;
5090 	uint32_t ip_src_addr = 0;
5091 	uint32_t ip_desc_addr = 0;
5092 	uint16_t src_port = 0;
5093 	uint16_t dest_port = 0;
5094 	uint8_t key[12];
5095 	QL_PRINT(DBG_TX, ("%s(%d) entered \n", __func__, qlge->instance));
5096 
5097 	ethhdr = (struct ether_header *)((void *)bp);
5098 	ethvhdr = (struct ether_vlan_header *)((void *)bp);
5099 
5100 	if (qlge->tx_ring_count == 1)
5101 		return (tx_ring_id);
5102 
5103 	/* Is this vlan packet? */
5104 	if (ntohs(ethvhdr->ether_tpid) == ETHERTYPE_VLAN) {
5105 		mac_hdr_len = sizeof (struct ether_vlan_header);
5106 		etherType = ntohs(ethvhdr->ether_type);
5107 	} else {
5108 		mac_hdr_len = sizeof (struct ether_header);
5109 		etherType = ntohs(ethhdr->ether_type);
5110 	}
5111 	/* Is this IPv4 or IPv6 packet? */
5112 	if (etherType == ETHERTYPE_IP /* 0800 */) {
5113 		if (IPH_HDR_VERSION((ipha_t *)(void *)(bp+mac_hdr_len))
5114 		    == IPV4_VERSION) {
5115 			iphdr = (struct ip *)(void *)(bp+mac_hdr_len);
5116 		}
5117 		if (((unsigned long)iphdr) & 0x3) {
5118 			/*  IP hdr not 4-byte aligned */
5119 			return (tx_ring_id);
5120 		}
5121 	}
5122 	/* ipV4 packets */
5123 	if (iphdr) {
5124 
5125 		ip_hdr_len = IPH_HDR_LENGTH(iphdr);
5126 		ip_src_addr = iphdr->ip_src.s_addr;
5127 		ip_desc_addr = iphdr->ip_dst.s_addr;
5128 
5129 		if (iphdr->ip_p == IPPROTO_TCP) {
5130 			tcp_hdr = (struct tcphdr *)(void *)
5131 			    ((uint8_t *)iphdr + ip_hdr_len);
5132 			src_port = tcp_hdr->th_sport;
5133 			dest_port = tcp_hdr->th_dport;
5134 		} else if (iphdr->ip_p == IPPROTO_UDP) {
5135 			udp_hdr = (struct udphdr *)(void *)
5136 			    ((uint8_t *)iphdr + ip_hdr_len);
5137 			src_port = udp_hdr->uh_sport;
5138 			dest_port = udp_hdr->uh_dport;
5139 		}
5140 		key[0] = (uint8_t)((ip_src_addr) &0xFF);
5141 		key[1] = (uint8_t)((ip_src_addr >> 8) &0xFF);
5142 		key[2] = (uint8_t)((ip_src_addr >> 16) &0xFF);
5143 		key[3] = (uint8_t)((ip_src_addr >> 24) &0xFF);
5144 		key[4] = (uint8_t)((ip_desc_addr) &0xFF);
5145 		key[5] = (uint8_t)((ip_desc_addr >> 8) &0xFF);
5146 		key[6] = (uint8_t)((ip_desc_addr >> 16) &0xFF);
5147 		key[7] = (uint8_t)((ip_desc_addr >> 24) &0xFF);
5148 		key[8] = (uint8_t)((src_port) &0xFF);
5149 		key[9] = (uint8_t)((src_port >> 8) &0xFF);
5150 		key[10] = (uint8_t)((dest_port) &0xFF);
5151 		key[11] = (uint8_t)((dest_port >> 8) &0xFF);
5152 		h = hash(key, 12, 0); /* return 32 bit */
5153 		tx_ring_id = (h & (qlge->tx_ring_count - 1));
5154 		if (tx_ring_id >= qlge->tx_ring_count) {
5155 			cmn_err(CE_WARN, "%s bad tx_ring_id %d\n",
5156 			    __func__, tx_ring_id);
5157 			tx_ring_id = 0;
5158 		}
5159 	}
5160 	return (tx_ring_id);
5161 }
5162 
5163 /*
5164  * Tell the hardware to do Large Send Offload (LSO)
5165  *
5166  * Some fields in ob_mac_iocb need to be set so hardware can know what is
5167  * the incoming packet, TCP or UDP, whether a VLAN tag needs to be inserted
5168  * in the right place of the packet etc, thus, hardware can process the
5169  * packet correctly.
5170  */
5171 static void
5172 ql_hw_lso_setup(qlge_t *qlge, uint32_t mss, caddr_t bp,
5173     struct ob_mac_iocb_req *mac_iocb_ptr)
5174 {
5175 	struct ip *iphdr = NULL;
5176 	struct ether_header *ethhdr;
5177 	struct ether_vlan_header *ethvhdr;
5178 	struct tcphdr *tcp_hdr;
5179 	struct udphdr *udp_hdr;
5180 	uint32_t etherType;
5181 	uint16_t mac_hdr_len, ip_hdr_len, tcp_udp_hdr_len;
5182 	uint16_t ip_hdr_off, tcp_udp_hdr_off, hdr_off;
5183 
5184 	ethhdr = (struct ether_header *)(void *)bp;
5185 	ethvhdr = (struct ether_vlan_header *)(void *)bp;
5186 
5187 	/* Is this vlan packet? */
5188 	if (ntohs(ethvhdr->ether_tpid) == ETHERTYPE_VLAN) {
5189 		mac_hdr_len = sizeof (struct ether_vlan_header);
5190 		etherType = ntohs(ethvhdr->ether_type);
5191 	} else {
5192 		mac_hdr_len = sizeof (struct ether_header);
5193 		etherType = ntohs(ethhdr->ether_type);
5194 	}
5195 	/* Is this IPv4 or IPv6 packet? */
5196 	if (IPH_HDR_VERSION((ipha_t *)(void *)(bp + mac_hdr_len)) ==
5197 	    IPV4_VERSION) {
5198 		if (etherType == ETHERTYPE_IP /* 0800 */) {
5199 			iphdr = (struct ip *)(void *)(bp+mac_hdr_len);
5200 		} else {
5201 			/* EMPTY */
5202 			QL_PRINT(DBG_TX, ("%s(%d) : IPv4 None IP packet"
5203 			    " type 0x%x\n",
5204 			    __func__, qlge->instance, etherType));
5205 		}
5206 	}
5207 
5208 	if (iphdr != NULL) { /* ipV4 packets */
5209 		ip_hdr_len = (uint16_t)IPH_HDR_LENGTH(iphdr);
5210 		QL_PRINT(DBG_TX,
5211 		    ("%s(%d) : IPv4 header length using IPH_HDR_LENGTH: %d"
5212 		    " bytes \n", __func__, qlge->instance, ip_hdr_len));
5213 
5214 		ip_hdr_off = mac_hdr_len;
5215 		QL_PRINT(DBG_TX, ("%s(%d) : ip_hdr_len=%d\n",
5216 		    __func__, qlge->instance, ip_hdr_len));
5217 
5218 		mac_iocb_ptr->flag0 = (uint8_t)(mac_iocb_ptr->flag0 |
5219 		    OB_MAC_IOCB_REQ_IPv4);
5220 		if (qlge->cfg_flags & CFG_CKSUM_FULL_IPv4) {
5221 			if (iphdr->ip_p == IPPROTO_TCP) {
5222 				tcp_hdr = (struct tcphdr *)(void *)
5223 				    ((uint8_t *)(void *)iphdr +
5224 				    ip_hdr_len);
5225 				QL_PRINT(DBG_TX, ("%s(%d) : Do TSO on TCP "
5226 				    "packet\n",
5227 				    __func__, qlge->instance));
5228 				mac_iocb_ptr->opcode =
5229 				    OPCODE_OB_MAC_OFFLOAD_IOCB;
5230 				mac_iocb_ptr->flag1 =
5231 				    (uint8_t)(mac_iocb_ptr->flag1 |
5232 				    OB_MAC_IOCB_REQ_LSO);
5233 				iphdr->ip_sum = 0;
5234 				tcp_udp_hdr_off =
5235 				    (uint16_t)(mac_hdr_len+ip_hdr_len);
5236 				tcp_udp_hdr_len =
5237 				    (uint16_t)(tcp_hdr->th_off*4);
5238 				QL_PRINT(DBG_TX, ("%s(%d): tcp header len:%d\n",
5239 				    __func__, qlge->instance, tcp_udp_hdr_len));
5240 				hdr_off = ip_hdr_off;
5241 				tcp_udp_hdr_off <<= 6;
5242 				hdr_off |= tcp_udp_hdr_off;
5243 				mac_iocb_ptr->hdr_off =
5244 				    (uint16_t)cpu_to_le16(hdr_off);
5245 				mac_iocb_ptr->protocol_hdr_len = (uint16_t)
5246 				    cpu_to_le16(mac_hdr_len + ip_hdr_len +
5247 				    tcp_udp_hdr_len);
5248 				mac_iocb_ptr->mss = (uint16_t)cpu_to_le16(mss);
5249 
5250 				/*
5251 				 * if the chip is unable to calculate pseudo
5252 				 * header checksum, do it in then put the result
5253 				 * to the data passed to the chip
5254 				 */
5255 				if (qlge->cfg_flags &
5256 				    CFG_HW_UNABLE_PSEUDO_HDR_CKSUM)
5257 					ql_lso_pseudo_cksum((uint8_t *)iphdr);
5258 			} else if (iphdr->ip_p == IPPROTO_UDP) {
5259 				udp_hdr = (struct udphdr *)(void *)
5260 				    ((uint8_t *)(void *)iphdr
5261 				    + ip_hdr_len);
5262 				QL_PRINT(DBG_TX, ("%s(%d) : Do TSO on UDP "
5263 				    "packet\n",
5264 				    __func__, qlge->instance));
5265 				mac_iocb_ptr->opcode =
5266 				    OPCODE_OB_MAC_OFFLOAD_IOCB;
5267 				mac_iocb_ptr->flag1 =
5268 				    (uint8_t)(mac_iocb_ptr->flag1 |
5269 				    OB_MAC_IOCB_REQ_LSO);
5270 				iphdr->ip_sum = 0;
5271 				tcp_udp_hdr_off =
5272 				    (uint16_t)(mac_hdr_len+ip_hdr_len);
5273 				tcp_udp_hdr_len =
5274 				    (uint16_t)(udp_hdr->uh_ulen*4);
5275 				QL_PRINT(DBG_TX, ("%s(%d):udp header len:%d\n",
5276 				    __func__, qlge->instance, tcp_udp_hdr_len));
5277 				hdr_off = ip_hdr_off;
5278 				tcp_udp_hdr_off <<= 6;
5279 				hdr_off |= tcp_udp_hdr_off;
5280 				mac_iocb_ptr->hdr_off =
5281 				    (uint16_t)cpu_to_le16(hdr_off);
5282 				mac_iocb_ptr->protocol_hdr_len = (uint16_t)
5283 				    cpu_to_le16(mac_hdr_len + ip_hdr_len +
5284 				    tcp_udp_hdr_len);
5285 				mac_iocb_ptr->mss = (uint16_t)cpu_to_le16(mss);
5286 
5287 				/*
5288 				 * if the chip is unable to do pseudo header
5289 				 * checksum calculation, do it here then put the
5290 				 * result to the data passed to the chip
5291 				 */
5292 				if (qlge->cfg_flags &
5293 				    CFG_HW_UNABLE_PSEUDO_HDR_CKSUM)
5294 					ql_lso_pseudo_cksum((uint8_t *)iphdr);
5295 			}
5296 		}
5297 	}
5298 }
5299 
5300 /*
5301  * Generic packet sending function which is used to send one packet.
5302  */
5303 int
5304 ql_send_common(struct tx_ring *tx_ring, mblk_t *mp)
5305 {
5306 	struct tx_ring_desc *tx_cb;
5307 	struct ob_mac_iocb_req *mac_iocb_ptr;
5308 	mblk_t *tp;
5309 	size_t msg_len = 0;
5310 	size_t off;
5311 	caddr_t bp;
5312 	size_t nbyte, total_len;
5313 	uint_t i = 0;
5314 	int j = 0, frags = 0;
5315 	uint32_t phy_addr_low, phy_addr_high;
5316 	uint64_t phys_addr;
5317 	clock_t now;
5318 	uint32_t pflags = 0;
5319 	uint32_t mss = 0;
5320 	enum tx_mode_t tx_mode;
5321 	struct oal_entry *oal_entry;
5322 	int status;
5323 	uint_t ncookies, oal_entries, max_oal_entries;
5324 	size_t max_seg_len = 0;
5325 	boolean_t use_lso = B_FALSE;
5326 	struct oal_entry *tx_entry = NULL;
5327 	struct oal_entry *last_oal_entry;
5328 	qlge_t *qlge = tx_ring->qlge;
5329 	ddi_dma_cookie_t dma_cookie;
5330 	size_t tx_buf_len = QL_MAX_COPY_LENGTH;
5331 	int force_pullup = 0;
5332 
5333 	tp = mp;
5334 	total_len = msg_len = 0;
5335 	max_oal_entries = TX_DESC_PER_IOCB + MAX_SG_ELEMENTS-1;
5336 
5337 	/* Calculate number of data and segments in the incoming message */
5338 	for (tp = mp; tp != NULL; tp = tp->b_cont) {
5339 		nbyte = MBLKL(tp);
5340 		total_len += nbyte;
5341 		max_seg_len = max(nbyte, max_seg_len);
5342 		QL_PRINT(DBG_TX, ("Requested sending data in %d segments, "
5343 		    "total length: %d\n", frags, nbyte));
5344 		frags++;
5345 	}
5346 
5347 	if (total_len >= QL_LSO_MAX) {
5348 		freemsg(mp);
5349 #ifdef QLGE_LOAD_UNLOAD
5350 		cmn_err(CE_NOTE, "%s: quit, packet oversize %d\n",
5351 		    __func__, (int)total_len);
5352 #endif
5353 		return (0);
5354 	}
5355 
5356 	bp = (caddr_t)mp->b_rptr;
5357 	if (bp[0] & 1) {
5358 		if (bcmp(bp, ql_ether_broadcast_addr.ether_addr_octet,
5359 		    ETHERADDRL) == 0) {
5360 			QL_PRINT(DBG_TX, ("Broadcast packet\n"));
5361 			tx_ring->brdcstxmt++;
5362 		} else {
5363 			QL_PRINT(DBG_TX, ("multicast packet\n"));
5364 			tx_ring->multixmt++;
5365 		}
5366 	}
5367 
5368 	tx_ring->obytes += total_len;
5369 	tx_ring->opackets ++;
5370 
5371 	QL_PRINT(DBG_TX, ("total requested sending data length: %d, in %d segs,"
5372 	    " max seg len: %d\n", total_len, frags, max_seg_len));
5373 
5374 	/* claim a free slot in tx ring */
5375 	tx_cb = &tx_ring->wq_desc[tx_ring->prod_idx];
5376 
5377 	/* get the tx descriptor */
5378 	mac_iocb_ptr = tx_cb->queue_entry;
5379 
5380 	bzero((void *)mac_iocb_ptr, 20);
5381 
5382 	ASSERT(tx_cb->mp == NULL);
5383 
5384 	/*
5385 	 * Decide to use DMA map or copy mode.
5386 	 * DMA map mode must be used when the total msg length is more than the
5387 	 * tx buffer length.
5388 	 */
5389 
5390 	if (total_len > tx_buf_len)
5391 		tx_mode = USE_DMA;
5392 	else if	(max_seg_len > QL_MAX_COPY_LENGTH)
5393 		tx_mode = USE_DMA;
5394 	else
5395 		tx_mode = USE_COPY;
5396 
5397 	if (qlge->chksum_cap) {
5398 		mac_hcksum_get(mp, NULL, NULL, NULL, NULL, &pflags);
5399 		QL_PRINT(DBG_TX, ("checksum flag is :0x%x, card capability "
5400 		    "is 0x%x \n", pflags, qlge->chksum_cap));
5401 		if (qlge->lso_enable) {
5402 			uint32_t lso_flags = 0;
5403 			mac_lso_get(mp, &mss, &lso_flags);
5404 			use_lso = (lso_flags == HW_LSO);
5405 		}
5406 		QL_PRINT(DBG_TX, ("mss :%d, use_lso %x \n",
5407 		    mss, use_lso));
5408 	}
5409 
5410 do_pullup:
5411 
5412 	/* concatenate all frags into one large packet if too fragmented */
5413 	if (((tx_mode == USE_DMA)&&(frags > QL_MAX_TX_DMA_HANDLES)) ||
5414 	    force_pullup) {
5415 		mblk_t *mp1;
5416 		if ((mp1 = msgpullup(mp, -1)) != NULL) {
5417 			freemsg(mp);
5418 			mp = mp1;
5419 			frags = 1;
5420 		} else {
5421 			tx_ring->tx_fail_dma_bind++;
5422 			goto bad;
5423 		}
5424 	}
5425 
5426 	tx_cb->tx_bytes = (uint32_t)total_len;
5427 	tx_cb->mp = mp;
5428 	tx_cb->tx_dma_handle_used = 0;
5429 
5430 	if (tx_mode == USE_DMA) {
5431 		msg_len = total_len;
5432 
5433 		mac_iocb_ptr->opcode = OPCODE_OB_MAC_IOCB;
5434 		mac_iocb_ptr->tid = tx_ring->prod_idx;
5435 		mac_iocb_ptr->frame_len = (uint32_t)cpu_to_le32(msg_len);
5436 		mac_iocb_ptr->txq_idx = tx_ring->wq_id;
5437 
5438 		tx_entry = &mac_iocb_ptr->oal_entry[0];
5439 		oal_entry = NULL;
5440 
5441 		for (tp = mp, oal_entries = j = 0; tp != NULL;
5442 		    tp = tp->b_cont) {
5443 			/* if too many tx dma handles needed */
5444 			if (j >= QL_MAX_TX_DMA_HANDLES) {
5445 				tx_ring->tx_no_dma_handle++;
5446 				if (!force_pullup) {
5447 					force_pullup = 1;
5448 					goto do_pullup;
5449 				} else {
5450 					goto bad;
5451 				}
5452 			}
5453 			nbyte = (uint16_t)MBLKL(tp);
5454 			if (nbyte == 0)
5455 				continue;
5456 
5457 			status = ddi_dma_addr_bind_handle(
5458 			    tx_cb->tx_dma_handle[j], NULL,
5459 			    (caddr_t)tp->b_rptr, nbyte,
5460 			    DDI_DMA_WRITE | DDI_DMA_STREAMING, DDI_DMA_DONTWAIT,
5461 			    0, &dma_cookie, &ncookies);
5462 
5463 			QL_PRINT(DBG_TX, ("map sending data segment: %d, "
5464 			    "length: %d, spans in %d cookies\n",
5465 			    j, nbyte, ncookies));
5466 
5467 			if (status != DDI_DMA_MAPPED) {
5468 				goto bad;
5469 			}
5470 			/*
5471 			 * Each fragment can span several cookies. One cookie
5472 			 * will use one tx descriptor to transmit.
5473 			 */
5474 			for (i = ncookies; i > 0; i--, tx_entry++,
5475 			    oal_entries++) {
5476 				/*
5477 				 * The number of TX descriptors that can be
5478 				 *  saved in tx iocb and oal list is limited
5479 				 */
5480 				if (oal_entries > max_oal_entries) {
5481 					tx_ring->tx_no_dma_cookie++;
5482 					if (!force_pullup) {
5483 						force_pullup = 1;
5484 						goto do_pullup;
5485 					} else {
5486 						goto bad;
5487 					}
5488 				}
5489 
5490 				if ((oal_entries == TX_DESC_PER_IOCB) &&
5491 				    !oal_entry) {
5492 					/*
5493 					 * Time to switch to an oal list
5494 					 * The last entry should be copied
5495 					 * to first entry in the oal list
5496 					 */
5497 					oal_entry = tx_cb->oal;
5498 					tx_entry =
5499 					    &mac_iocb_ptr->oal_entry[
5500 					    TX_DESC_PER_IOCB-1];
5501 					bcopy(tx_entry, oal_entry,
5502 					    sizeof (*oal_entry));
5503 
5504 					/*
5505 					 * last entry should be updated to
5506 					 * point to the extended oal list itself
5507 					 */
5508 					tx_entry->buf_addr_low =
5509 					    cpu_to_le32(
5510 					    LS_64BITS(tx_cb->oal_dma_addr));
5511 					tx_entry->buf_addr_high =
5512 					    cpu_to_le32(
5513 					    MS_64BITS(tx_cb->oal_dma_addr));
5514 					/*
5515 					 * Point tx_entry to the oal list
5516 					 * second entry
5517 					 */
5518 					tx_entry = &oal_entry[1];
5519 				}
5520 
5521 				tx_entry->buf_len =
5522 				    (uint32_t)cpu_to_le32(dma_cookie.dmac_size);
5523 				phys_addr = dma_cookie.dmac_laddress;
5524 				tx_entry->buf_addr_low =
5525 				    cpu_to_le32(LS_64BITS(phys_addr));
5526 				tx_entry->buf_addr_high =
5527 				    cpu_to_le32(MS_64BITS(phys_addr));
5528 
5529 				last_oal_entry = tx_entry;
5530 
5531 				if (i > 1)
5532 					ddi_dma_nextcookie(
5533 					    tx_cb->tx_dma_handle[j],
5534 					    &dma_cookie);
5535 			}
5536 			j++;
5537 		}
5538 		/*
5539 		 * if OAL is used, the last oal entry in tx iocb indicates
5540 		 * number of additional address/len pairs in OAL
5541 		 */
5542 		if (oal_entries > TX_DESC_PER_IOCB) {
5543 			tx_entry = &mac_iocb_ptr->oal_entry[TX_DESC_PER_IOCB-1];
5544 			tx_entry->buf_len = (uint32_t)
5545 			    (cpu_to_le32((sizeof (struct oal_entry) *
5546 			    (oal_entries -TX_DESC_PER_IOCB+1))|OAL_CONT_ENTRY));
5547 		}
5548 		last_oal_entry->buf_len = cpu_to_le32(
5549 		    le32_to_cpu(last_oal_entry->buf_len)|OAL_LAST_ENTRY);
5550 
5551 		tx_cb->tx_dma_handle_used = j;
5552 		QL_PRINT(DBG_TX, ("total tx_dma_handle_used %d cookies %d \n",
5553 		    j, oal_entries));
5554 
5555 		bp = (caddr_t)mp->b_rptr;
5556 	}
5557 	if (tx_mode == USE_COPY) {
5558 		bp = tx_cb->copy_buffer;
5559 		off = 0;
5560 		nbyte = 0;
5561 		frags = 0;
5562 		/*
5563 		 * Copy up to tx_buf_len of the transmit data
5564 		 * from mp to tx buffer
5565 		 */
5566 		for (tp = mp; tp != NULL; tp = tp->b_cont) {
5567 			nbyte = MBLKL(tp);
5568 			if ((off + nbyte) <= tx_buf_len) {
5569 				bcopy(tp->b_rptr, &bp[off], nbyte);
5570 				off += nbyte;
5571 				frags ++;
5572 			}
5573 		}
5574 
5575 		msg_len = off;
5576 
5577 		mac_iocb_ptr->opcode = OPCODE_OB_MAC_IOCB;
5578 		mac_iocb_ptr->tid = tx_ring->prod_idx;
5579 		mac_iocb_ptr->frame_len = (uint32_t)cpu_to_le32(msg_len);
5580 		mac_iocb_ptr->txq_idx = tx_ring->wq_id;
5581 
5582 		QL_PRINT(DBG_TX, ("Copy Mode:actual sent data length is: %d, "
5583 		    "from %d segaments\n", msg_len, frags));
5584 
5585 		phys_addr = tx_cb->copy_buffer_dma_addr;
5586 		phy_addr_low = cpu_to_le32(LS_64BITS(phys_addr));
5587 		phy_addr_high = cpu_to_le32(MS_64BITS(phys_addr));
5588 
5589 		QL_DUMP(DBG_TX, "\t requested sending data:\n",
5590 		    (uint8_t *)tx_cb->copy_buffer, 8, total_len);
5591 
5592 		mac_iocb_ptr->oal_entry[0].buf_len = (uint32_t)
5593 		    cpu_to_le32(msg_len | OAL_LAST_ENTRY);
5594 		mac_iocb_ptr->oal_entry[0].buf_addr_low  = phy_addr_low;
5595 		mac_iocb_ptr->oal_entry[0].buf_addr_high = phy_addr_high;
5596 
5597 		freemsg(mp); /* no need, we have copied */
5598 		tx_cb->mp = NULL;
5599 	} /* End of Copy Mode */
5600 
5601 	/* Do TSO/LSO on TCP packet? */
5602 	if (use_lso && mss) {
5603 		ql_hw_lso_setup(qlge, mss, bp, mac_iocb_ptr);
5604 	} else if (pflags & qlge->chksum_cap) {
5605 		/* Do checksum offloading */
5606 		ql_hw_csum_setup(qlge, pflags, bp, mac_iocb_ptr);
5607 	}
5608 
5609 	/* let device know the latest outbound IOCB */
5610 	(void) ddi_dma_sync(tx_ring->wq_dma.dma_handle,
5611 	    (off_t)((uintptr_t)mac_iocb_ptr - (uintptr_t)tx_ring->wq_dma.vaddr),
5612 	    (size_t)sizeof (*mac_iocb_ptr), DDI_DMA_SYNC_FORDEV);
5613 
5614 	if (tx_mode == USE_DMA) {
5615 		/* let device know the latest outbound OAL if necessary */
5616 		if (oal_entries > TX_DESC_PER_IOCB) {
5617 			(void) ddi_dma_sync(tx_cb->oal_dma.dma_handle,
5618 			    (off_t)0,
5619 			    (sizeof (struct oal_entry) *
5620 			    (oal_entries -TX_DESC_PER_IOCB+1)),
5621 			    DDI_DMA_SYNC_FORDEV);
5622 		}
5623 	} else { /* for USE_COPY mode, tx buffer has changed */
5624 		/* let device know the latest change */
5625 		(void) ddi_dma_sync(tx_cb->oal_dma.dma_handle,
5626 		/* copy buf offset */
5627 		    (off_t)(sizeof (oal_entry) * MAX_SG_ELEMENTS),
5628 		    msg_len, DDI_DMA_SYNC_FORDEV);
5629 	}
5630 
5631 	/* save how the packet was sent */
5632 	tx_cb->tx_type = tx_mode;
5633 
5634 	QL_DUMP_REQ_PKT(qlge, mac_iocb_ptr, tx_cb->oal, oal_entries);
5635 	/* reduce the number of available tx slot */
5636 	atomic_dec_32(&tx_ring->tx_free_count);
5637 
5638 	tx_ring->prod_idx++;
5639 	if (tx_ring->prod_idx >= tx_ring->wq_len)
5640 		tx_ring->prod_idx = 0;
5641 
5642 	now = ddi_get_lbolt();
5643 	qlge->last_tx_time = now;
5644 
5645 	return (DDI_SUCCESS);
5646 
5647 bad:
5648 	/*
5649 	 * if for any reason driver can not send, delete
5650 	 * the message pointer, mp
5651 	 */
5652 	now = ddi_get_lbolt();
5653 	freemsg(mp);
5654 	mp = NULL;
5655 	tx_cb->mp = NULL;
5656 	for (i = 0; i < j; i++)
5657 		(void) ddi_dma_unbind_handle(tx_cb->tx_dma_handle[i]);
5658 
5659 	QL_PRINT(DBG_TX, ("%s(%d) failed at 0x%x",
5660 	    __func__, qlge->instance, (int)now));
5661 
5662 	return (DDI_SUCCESS);
5663 }
5664 
5665 
5666 /*
5667  * Initializes hardware and driver software flags before the driver
5668  * is finally ready to work.
5669  */
5670 int
5671 ql_do_start(qlge_t *qlge)
5672 {
5673 	int i;
5674 	struct rx_ring *rx_ring;
5675 	uint16_t lbq_buf_size;
5676 	int rings_done;
5677 
5678 	ASSERT(qlge != NULL);
5679 
5680 	mutex_enter(&qlge->hw_mutex);
5681 
5682 	/* Reset adapter */
5683 	(void) ql_asic_reset(qlge);
5684 
5685 	lbq_buf_size = (uint16_t)
5686 	    ((qlge->mtu == ETHERMTU)? LRG_BUF_NORMAL_SIZE : LRG_BUF_JUMBO_SIZE);
5687 	if (qlge->rx_ring[0].lbq_buf_size != lbq_buf_size) {
5688 #ifdef QLGE_LOAD_UNLOAD
5689 		cmn_err(CE_NOTE, "realloc buffers old: %d new: %d\n",
5690 		    qlge->rx_ring[0].lbq_buf_size, lbq_buf_size);
5691 #endif
5692 		/*
5693 		 * Check if any ring has buffers still with upper layers
5694 		 * If buffers are pending with upper layers, we use the
5695 		 * existing buffers and don't reallocate new ones
5696 		 * Unfortunately there is no way to evict buffers from
5697 		 * upper layers. Using buffers with the current size may
5698 		 * cause slightly sub-optimal performance, but that seems
5699 		 * to be the easiest way to handle this situation.
5700 		 */
5701 		rings_done = 0;
5702 		for (i = 0; i < qlge->rx_ring_count; i++) {
5703 			rx_ring = &qlge->rx_ring[i];
5704 			if (rx_ring->rx_indicate == 0)
5705 				rings_done++;
5706 			else
5707 				break;
5708 		}
5709 		/*
5710 		 * No buffers pending with upper layers;
5711 		 * reallocte them for new MTU size
5712 		 */
5713 		if (rings_done >= qlge->rx_ring_count) {
5714 			/* free large buffer pool */
5715 			for (i = 0; i < qlge->rx_ring_count; i++) {
5716 				rx_ring = &qlge->rx_ring[i];
5717 				if (rx_ring->type != TX_Q) {
5718 					ql_free_sbq_buffers(rx_ring);
5719 					ql_free_lbq_buffers(rx_ring);
5720 				}
5721 			}
5722 			/* reallocate large buffer pool */
5723 			for (i = 0; i < qlge->rx_ring_count; i++) {
5724 				rx_ring = &qlge->rx_ring[i];
5725 				if (rx_ring->type != TX_Q) {
5726 					(void) ql_alloc_sbufs(qlge, rx_ring);
5727 					(void) ql_alloc_lbufs(qlge, rx_ring);
5728 				}
5729 			}
5730 		}
5731 	}
5732 
5733 	if (ql_bringup_adapter(qlge) != DDI_SUCCESS) {
5734 		cmn_err(CE_WARN, "qlge bringup adapter failed");
5735 		mutex_exit(&qlge->hw_mutex);
5736 		if (qlge->fm_enable) {
5737 			atomic_or_32(&qlge->flags, ADAPTER_ERROR);
5738 			ddi_fm_service_impact(qlge->dip, DDI_SERVICE_LOST);
5739 		}
5740 		return (DDI_FAILURE);
5741 	}
5742 
5743 	mutex_exit(&qlge->hw_mutex);
5744 	/* if adapter is up successfully but was bad before */
5745 	if (qlge->flags & ADAPTER_ERROR) {
5746 		atomic_and_32(&qlge->flags, ~ADAPTER_ERROR);
5747 		if (qlge->fm_enable) {
5748 			ddi_fm_service_impact(qlge->dip, DDI_SERVICE_RESTORED);
5749 		}
5750 	}
5751 
5752 	/* Get current link state */
5753 	qlge->port_link_state = ql_get_link_state(qlge);
5754 
5755 	if (qlge->port_link_state == LS_UP) {
5756 		QL_PRINT(DBG_GLD, ("%s(%d) Link UP !!\n",
5757 		    __func__, qlge->instance));
5758 		/* If driver detects a carrier on */
5759 		CARRIER_ON(qlge);
5760 	} else {
5761 		QL_PRINT(DBG_GLD, ("%s(%d) Link down\n",
5762 		    __func__, qlge->instance));
5763 		/* If driver detects a lack of carrier */
5764 		CARRIER_OFF(qlge);
5765 	}
5766 	qlge->mac_flags = QL_MAC_STARTED;
5767 	return (DDI_SUCCESS);
5768 }
5769 
5770 /*
5771  * Stop currently running driver
5772  * Driver needs to stop routing new packets to driver and wait until
5773  * all pending tx/rx buffers to be free-ed.
5774  */
5775 int
5776 ql_do_stop(qlge_t *qlge)
5777 {
5778 	int rc = DDI_FAILURE;
5779 	uint32_t i, j, k;
5780 	struct bq_desc *sbq_desc, *lbq_desc;
5781 	struct rx_ring *rx_ring;
5782 
5783 	ASSERT(qlge != NULL);
5784 
5785 	CARRIER_OFF(qlge);
5786 
5787 	rc = ql_bringdown_adapter(qlge);
5788 	if (rc != DDI_SUCCESS) {
5789 		cmn_err(CE_WARN, "qlge bringdown adapter failed.");
5790 	} else
5791 		rc = DDI_SUCCESS;
5792 
5793 	for (k = 0; k < qlge->rx_ring_count; k++) {
5794 		rx_ring = &qlge->rx_ring[k];
5795 		if (rx_ring->type != TX_Q) {
5796 			j = rx_ring->lbq_use_head;
5797 #ifdef QLGE_LOAD_UNLOAD
5798 			cmn_err(CE_NOTE, "ring %d: move %d lbufs in use list"
5799 			    " to free list %d\n total %d\n",
5800 			    k, rx_ring->lbuf_in_use_count,
5801 			    rx_ring->lbuf_free_count,
5802 			    rx_ring->lbuf_in_use_count +
5803 			    rx_ring->lbuf_free_count);
5804 #endif
5805 			for (i = 0; i < rx_ring->lbuf_in_use_count; i++) {
5806 				lbq_desc = rx_ring->lbuf_in_use[j];
5807 				j++;
5808 				if (j >= rx_ring->lbq_len) {
5809 					j = 0;
5810 				}
5811 				if (lbq_desc->mp) {
5812 					atomic_inc_32(&rx_ring->rx_indicate);
5813 					freemsg(lbq_desc->mp);
5814 				}
5815 			}
5816 			rx_ring->lbq_use_head = j;
5817 			rx_ring->lbq_use_tail = j;
5818 			rx_ring->lbuf_in_use_count = 0;
5819 			j = rx_ring->sbq_use_head;
5820 #ifdef QLGE_LOAD_UNLOAD
5821 			cmn_err(CE_NOTE, "ring %d: move %d sbufs in use list,"
5822 			    " to free list %d\n total %d \n",
5823 			    k, rx_ring->sbuf_in_use_count,
5824 			    rx_ring->sbuf_free_count,
5825 			    rx_ring->sbuf_in_use_count +
5826 			    rx_ring->sbuf_free_count);
5827 #endif
5828 			for (i = 0; i < rx_ring->sbuf_in_use_count; i++) {
5829 				sbq_desc = rx_ring->sbuf_in_use[j];
5830 				j++;
5831 				if (j >= rx_ring->sbq_len) {
5832 					j = 0;
5833 				}
5834 				if (sbq_desc->mp) {
5835 					atomic_inc_32(&rx_ring->rx_indicate);
5836 					freemsg(sbq_desc->mp);
5837 				}
5838 			}
5839 			rx_ring->sbq_use_head = j;
5840 			rx_ring->sbq_use_tail = j;
5841 			rx_ring->sbuf_in_use_count = 0;
5842 		}
5843 	}
5844 
5845 	qlge->mac_flags = QL_MAC_STOPPED;
5846 
5847 	return (rc);
5848 }
5849 
5850 /*
5851  * Support
5852  */
5853 
5854 void
5855 ql_disable_isr(qlge_t *qlge)
5856 {
5857 	/*
5858 	 * disable the hardware interrupt
5859 	 */
5860 	ISP_DISABLE_GLOBAL_INTRS(qlge);
5861 
5862 	qlge->flags &= ~INTERRUPTS_ENABLED;
5863 }
5864 
5865 
5866 
5867 /*
5868  * busy wait for 'usecs' microseconds.
5869  */
5870 void
5871 qlge_delay(clock_t usecs)
5872 {
5873 	drv_usecwait(usecs);
5874 }
5875 
5876 /*
5877  * retrieve firmware details.
5878  */
5879 
5880 pci_cfg_t *
5881 ql_get_pci_config(qlge_t *qlge)
5882 {
5883 	return (&(qlge->pci_cfg));
5884 }
5885 
5886 /*
5887  * Get current Link status
5888  */
5889 static uint32_t
5890 ql_get_link_state(qlge_t *qlge)
5891 {
5892 	uint32_t bitToCheck = 0;
5893 	uint32_t temp, linkState;
5894 
5895 	if (qlge->func_number == qlge->fn0_net) {
5896 		bitToCheck = STS_PL0;
5897 	} else {
5898 		bitToCheck = STS_PL1;
5899 	}
5900 	temp = ql_read_reg(qlge, REG_STATUS);
5901 	QL_PRINT(DBG_GLD, ("%s(%d) chip status reg: 0x%x\n",
5902 	    __func__, qlge->instance, temp));
5903 
5904 	if (temp & bitToCheck) {
5905 		linkState = LS_UP;
5906 	} else {
5907 		linkState = LS_DOWN;
5908 	}
5909 	if (CFG_IST(qlge, CFG_CHIP_8100)) {
5910 		/* for Schultz, link Speed is fixed to 10G, full duplex */
5911 		qlge->speed  = SPEED_10G;
5912 		qlge->duplex = 1;
5913 	}
5914 	return (linkState);
5915 }
5916 /*
5917  * Get current link status and report to OS
5918  */
5919 static void
5920 ql_get_and_report_link_state(qlge_t *qlge)
5921 {
5922 	uint32_t cur_link_state;
5923 
5924 	/* Get current link state */
5925 	cur_link_state = ql_get_link_state(qlge);
5926 	/* if link state has changed */
5927 	if (cur_link_state != qlge->port_link_state) {
5928 
5929 		qlge->port_link_state = cur_link_state;
5930 
5931 		if (qlge->port_link_state == LS_UP) {
5932 			QL_PRINT(DBG_GLD, ("%s(%d) Link UP !!\n",
5933 			    __func__, qlge->instance));
5934 			/* If driver detects a carrier on */
5935 			CARRIER_ON(qlge);
5936 		} else {
5937 			QL_PRINT(DBG_GLD, ("%s(%d) Link down\n",
5938 			    __func__, qlge->instance));
5939 			/* If driver detects a lack of carrier */
5940 			CARRIER_OFF(qlge);
5941 		}
5942 	}
5943 }
5944 
5945 /*
5946  * timer callback function executed after timer expires
5947  */
5948 static void
5949 ql_timer(void* arg)
5950 {
5951 	ql_get_and_report_link_state((qlge_t *)arg);
5952 }
5953 
5954 /*
5955  * stop the running timer if activated
5956  */
5957 static void
5958 ql_stop_timer(qlge_t *qlge)
5959 {
5960 	timeout_id_t timer_id;
5961 	/* Disable driver timer */
5962 	if (qlge->ql_timer_timeout_id != NULL) {
5963 		timer_id = qlge->ql_timer_timeout_id;
5964 		qlge->ql_timer_timeout_id = NULL;
5965 		(void) untimeout(timer_id);
5966 	}
5967 }
5968 
5969 /*
5970  * stop then restart timer
5971  */
5972 void
5973 ql_restart_timer(qlge_t *qlge)
5974 {
5975 	ql_stop_timer(qlge);
5976 	qlge->ql_timer_ticks = TICKS_PER_SEC / 4;
5977 	qlge->ql_timer_timeout_id = timeout(ql_timer,
5978 	    (void *)qlge, qlge->ql_timer_ticks);
5979 }
5980 
5981 /* ************************************************************************* */
5982 /*
5983  *		Hardware K-Stats Data Structures and Subroutines
5984  */
5985 /* ************************************************************************* */
5986 static const ql_ksindex_t ql_kstats_hw[] = {
5987 	/* PCI related hardware information */
5988 	{ 0, "Vendor Id"			},
5989 	{ 1, "Device Id"			},
5990 	{ 2, "Command"				},
5991 	{ 3, "Status"				},
5992 	{ 4, "Revision Id"			},
5993 	{ 5, "Cache Line Size"			},
5994 	{ 6, "Latency Timer"			},
5995 	{ 7, "Header Type"			},
5996 	{ 9, "I/O base addr"			},
5997 	{ 10, "Control Reg Base addr low"	},
5998 	{ 11, "Control Reg Base addr high"	},
5999 	{ 12, "Doorbell Reg Base addr low"	},
6000 	{ 13, "Doorbell Reg Base addr high"	},
6001 	{ 14, "Subsystem Vendor Id"		},
6002 	{ 15, "Subsystem Device ID"		},
6003 	{ 16, "PCIe Device Control"		},
6004 	{ 17, "PCIe Link Status"		},
6005 
6006 	{ -1,	NULL				},
6007 };
6008 
6009 /*
6010  * kstat update function for PCI registers
6011  */
6012 static int
6013 ql_kstats_get_pci_regs(kstat_t *ksp, int flag)
6014 {
6015 	qlge_t *qlge;
6016 	kstat_named_t *knp;
6017 
6018 	if (flag != KSTAT_READ)
6019 		return (EACCES);
6020 
6021 	qlge = ksp->ks_private;
6022 	knp = ksp->ks_data;
6023 	(knp++)->value.ui32 = qlge->pci_cfg.vendor_id;
6024 	(knp++)->value.ui32 = qlge->pci_cfg.device_id;
6025 	(knp++)->value.ui32 = qlge->pci_cfg.command;
6026 	(knp++)->value.ui32 = qlge->pci_cfg.status;
6027 	(knp++)->value.ui32 = qlge->pci_cfg.revision;
6028 	(knp++)->value.ui32 = qlge->pci_cfg.cache_line_size;
6029 	(knp++)->value.ui32 = qlge->pci_cfg.latency_timer;
6030 	(knp++)->value.ui32 = qlge->pci_cfg.header_type;
6031 	(knp++)->value.ui32 = qlge->pci_cfg.io_base_address;
6032 	(knp++)->value.ui32 =
6033 	    qlge->pci_cfg.pci_cntl_reg_set_mem_base_address_lower;
6034 	(knp++)->value.ui32 =
6035 	    qlge->pci_cfg.pci_cntl_reg_set_mem_base_address_upper;
6036 	(knp++)->value.ui32 =
6037 	    qlge->pci_cfg.pci_doorbell_mem_base_address_lower;
6038 	(knp++)->value.ui32 =
6039 	    qlge->pci_cfg.pci_doorbell_mem_base_address_upper;
6040 	(knp++)->value.ui32 = qlge->pci_cfg.sub_vendor_id;
6041 	(knp++)->value.ui32 = qlge->pci_cfg.sub_device_id;
6042 	(knp++)->value.ui32 = qlge->pci_cfg.pcie_device_control;
6043 	(knp++)->value.ui32 = qlge->pci_cfg.link_status;
6044 
6045 	return (0);
6046 }
6047 
6048 static const ql_ksindex_t ql_kstats_mii[] = {
6049 	/* MAC/MII related hardware information */
6050 	{ 0, "mtu"},
6051 
6052 	{ -1, NULL},
6053 };
6054 
6055 
6056 /*
6057  * kstat update function for MII related information.
6058  */
6059 static int
6060 ql_kstats_mii_update(kstat_t *ksp, int flag)
6061 {
6062 	qlge_t *qlge;
6063 	kstat_named_t *knp;
6064 
6065 	if (flag != KSTAT_READ)
6066 		return (EACCES);
6067 
6068 	qlge = ksp->ks_private;
6069 	knp = ksp->ks_data;
6070 
6071 	(knp++)->value.ui32 = qlge->mtu;
6072 
6073 	return (0);
6074 }
6075 
6076 static const ql_ksindex_t ql_kstats_reg[] = {
6077 	/* Register information */
6078 	{ 0, "System (0x08)"			},
6079 	{ 1, "Reset/Fail Over(0x0Ch"		},
6080 	{ 2, "Function Specific Control(0x10)"	},
6081 	{ 3, "Status (0x30)"			},
6082 	{ 4, "Intr Enable (0x34)"		},
6083 	{ 5, "Intr Status1 (0x3C)"		},
6084 	{ 6, "Error Status (0x54)"		},
6085 	{ 7, "XGMAC Flow Control(0x11C)"	},
6086 	{ 8, "XGMAC Tx Pause Frames(0x230)"	},
6087 	{ 9, "XGMAC Rx Pause Frames(0x388)"	},
6088 	{ 10, "XGMAC Rx FIFO Drop Count(0x5B8)"	},
6089 	{ 11, "interrupts actually allocated"	},
6090 	{ 12, "interrupts on rx ring 0"		},
6091 	{ 13, "interrupts on rx ring 1"		},
6092 	{ 14, "interrupts on rx ring 2"		},
6093 	{ 15, "interrupts on rx ring 3"		},
6094 	{ 16, "interrupts on rx ring 4"		},
6095 	{ 17, "interrupts on rx ring 5"		},
6096 	{ 18, "interrupts on rx ring 6"		},
6097 	{ 19, "interrupts on rx ring 7"		},
6098 	{ 20, "polls on rx ring 0"		},
6099 	{ 21, "polls on rx ring 1"		},
6100 	{ 22, "polls on rx ring 2"		},
6101 	{ 23, "polls on rx ring 3"		},
6102 	{ 24, "polls on rx ring 4"		},
6103 	{ 25, "polls on rx ring 5"		},
6104 	{ 26, "polls on rx ring 6"		},
6105 	{ 27, "polls on rx ring 7"		},
6106 	{ 28, "tx no resource on ring 0"	},
6107 	{ 29, "tx dma bind fail on ring 0"	},
6108 	{ 30, "tx dma no handle on ring 0"	},
6109 	{ 31, "tx dma no cookie on ring 0"	},
6110 	{ 32, "MPI firmware major version"	},
6111 	{ 33, "MPI firmware minor version"	},
6112 	{ 34, "MPI firmware sub version"	},
6113 	{ 35, "rx no resource"			},
6114 
6115 	{ -1, NULL},
6116 };
6117 
6118 
6119 /*
6120  * kstat update function for device register set
6121  */
6122 static int
6123 ql_kstats_get_reg_and_dev_stats(kstat_t *ksp, int flag)
6124 {
6125 	qlge_t *qlge;
6126 	kstat_named_t *knp;
6127 	uint32_t val32;
6128 	int i = 0;
6129 	struct tx_ring *tx_ring;
6130 	struct rx_ring *rx_ring;
6131 
6132 	if (flag != KSTAT_READ)
6133 		return (EACCES);
6134 
6135 	qlge = ksp->ks_private;
6136 	knp = ksp->ks_data;
6137 
6138 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_SYSTEM);
6139 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_RESET_FAILOVER);
6140 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_FUNCTION_SPECIFIC_CONTROL);
6141 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_STATUS);
6142 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_INTERRUPT_ENABLE);
6143 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_INTERRUPT_STATUS_1);
6144 	(knp++)->value.ui32 = ql_read_reg(qlge, REG_ERROR_STATUS);
6145 
6146 	if (ql_sem_spinlock(qlge, qlge->xgmac_sem_mask)) {
6147 		return (0);
6148 	}
6149 	(void) ql_read_xgmac_reg(qlge, REG_XGMAC_FLOW_CONTROL, &val32);
6150 	(knp++)->value.ui32 = val32;
6151 
6152 	(void) ql_read_xgmac_reg(qlge, REG_XGMAC_MAC_TX_PAUSE_PKTS, &val32);
6153 	(knp++)->value.ui32 = val32;
6154 
6155 	(void) ql_read_xgmac_reg(qlge, REG_XGMAC_MAC_RX_PAUSE_PKTS, &val32);
6156 	(knp++)->value.ui32 = val32;
6157 
6158 	(void) ql_read_xgmac_reg(qlge, REG_XGMAC_MAC_RX_FIFO_DROPS, &val32);
6159 	(knp++)->value.ui32 = val32;
6160 
6161 	ql_sem_unlock(qlge, qlge->xgmac_sem_mask);
6162 
6163 	(knp++)->value.ui32 = qlge->intr_cnt;
6164 
6165 	for (i = 0; i < 8; i++) {
6166 		(knp++)->value.ui32 = qlge->rx_interrupts[i];
6167 	}
6168 
6169 	for (i = 0; i < 8; i++) {
6170 		(knp++)->value.ui32 = qlge->rx_polls[i];
6171 	}
6172 
6173 	tx_ring = &qlge->tx_ring[0];
6174 	(knp++)->value.ui32 = tx_ring->defer;
6175 	(knp++)->value.ui32 = tx_ring->tx_fail_dma_bind;
6176 	(knp++)->value.ui32 = tx_ring->tx_no_dma_handle;
6177 	(knp++)->value.ui32 = tx_ring->tx_no_dma_cookie;
6178 
6179 	(knp++)->value.ui32 = qlge->fw_version_info.major_version;
6180 	(knp++)->value.ui32 = qlge->fw_version_info.minor_version;
6181 	(knp++)->value.ui32 = qlge->fw_version_info.sub_minor_version;
6182 
6183 	for (i = 0; i < qlge->rx_ring_count; i++) {
6184 		rx_ring = &qlge->rx_ring[i];
6185 		val32 += rx_ring->rx_packets_dropped_no_buffer;
6186 	}
6187 	(knp++)->value.ui32 = val32;
6188 
6189 	return (0);
6190 }
6191 
6192 
6193 static kstat_t *
6194 ql_setup_named_kstat(qlge_t *qlge, int instance, char *name,
6195     const ql_ksindex_t *ksip, size_t size, int (*update)(kstat_t *, int))
6196 {
6197 	kstat_t *ksp;
6198 	kstat_named_t *knp;
6199 	char *np;
6200 	int type;
6201 
6202 	size /= sizeof (ql_ksindex_t);
6203 	ksp = kstat_create(ADAPTER_NAME, instance, name, "net",
6204 	    KSTAT_TYPE_NAMED, ((uint32_t)size) - 1, KSTAT_FLAG_PERSISTENT);
6205 	if (ksp == NULL)
6206 		return (NULL);
6207 
6208 	ksp->ks_private = qlge;
6209 	ksp->ks_update = update;
6210 	for (knp = ksp->ks_data; (np = ksip->name) != NULL; ++knp, ++ksip) {
6211 		switch (*np) {
6212 		default:
6213 			type = KSTAT_DATA_UINT32;
6214 			break;
6215 		case '&':
6216 			np += 1;
6217 			type = KSTAT_DATA_CHAR;
6218 			break;
6219 		}
6220 		kstat_named_init(knp, np, (uint8_t)type);
6221 	}
6222 	kstat_install(ksp);
6223 
6224 	return (ksp);
6225 }
6226 
6227 /*
6228  * Setup various kstat
6229  */
6230 int
6231 ql_init_kstats(qlge_t *qlge)
6232 {
6233 	/* Hardware KStats */
6234 	qlge->ql_kstats[QL_KSTAT_CHIP] = ql_setup_named_kstat(qlge,
6235 	    qlge->instance, "chip", ql_kstats_hw,
6236 	    sizeof (ql_kstats_hw), ql_kstats_get_pci_regs);
6237 	if (qlge->ql_kstats[QL_KSTAT_CHIP] == NULL) {
6238 		return (DDI_FAILURE);
6239 	}
6240 
6241 	/* MII KStats */
6242 	qlge->ql_kstats[QL_KSTAT_LINK] = ql_setup_named_kstat(qlge,
6243 	    qlge->instance, "mii", ql_kstats_mii,
6244 	    sizeof (ql_kstats_mii), ql_kstats_mii_update);
6245 	if (qlge->ql_kstats[QL_KSTAT_LINK] == NULL) {
6246 		return (DDI_FAILURE);
6247 	}
6248 
6249 	/* REG KStats */
6250 	qlge->ql_kstats[QL_KSTAT_REG] = ql_setup_named_kstat(qlge,
6251 	    qlge->instance, "reg", ql_kstats_reg,
6252 	    sizeof (ql_kstats_reg), ql_kstats_get_reg_and_dev_stats);
6253 	if (qlge->ql_kstats[QL_KSTAT_REG] == NULL) {
6254 		return (DDI_FAILURE);
6255 	}
6256 	return (DDI_SUCCESS);
6257 }
6258 
6259 /*
6260  * delete all kstat
6261  */
6262 void
6263 ql_fini_kstats(qlge_t *qlge)
6264 {
6265 	int i;
6266 
6267 	for (i = 0; i < QL_KSTAT_COUNT; i++) {
6268 		if (qlge->ql_kstats[i] != NULL)
6269 			kstat_delete(qlge->ql_kstats[i]);
6270 	}
6271 }
6272 
6273 /* ************************************************************************* */
6274 /*
6275  *                                 kstat end
6276  */
6277 /* ************************************************************************* */
6278 
6279 /*
6280  * Setup the parameters for receive and transmit rings including buffer sizes
6281  * and completion queue sizes
6282  */
6283 static int
6284 ql_setup_rings(qlge_t *qlge)
6285 {
6286 	uint8_t i;
6287 	struct rx_ring *rx_ring;
6288 	struct tx_ring *tx_ring;
6289 	uint16_t lbq_buf_size;
6290 
6291 	lbq_buf_size = (uint16_t)
6292 	    ((qlge->mtu == ETHERMTU)? LRG_BUF_NORMAL_SIZE : LRG_BUF_JUMBO_SIZE);
6293 
6294 	/*
6295 	 * rx_ring[0] is always the default queue.
6296 	 */
6297 	/*
6298 	 * qlge->rx_ring_count:
6299 	 * Total number of rx_rings. This includes a number
6300 	 * of outbound completion handler rx_rings, and a
6301 	 * number of inbound completion handler rx_rings.
6302 	 * rss is only enabled if we have more than 1 rx completion
6303 	 * queue. If we have a single rx completion queue
6304 	 * then all rx completions go to this queue and
6305 	 * the last completion queue
6306 	 */
6307 
6308 	qlge->tx_ring_first_cq_id = qlge->rss_ring_count;
6309 
6310 	for (i = 0; i < qlge->tx_ring_count; i++) {
6311 		tx_ring = &qlge->tx_ring[i];
6312 		bzero((void *)tx_ring, sizeof (*tx_ring));
6313 		tx_ring->qlge = qlge;
6314 		tx_ring->wq_id = i;
6315 		tx_ring->wq_len = qlge->tx_ring_size;
6316 		tx_ring->wq_size = (uint32_t)(
6317 		    tx_ring->wq_len * sizeof (struct ob_mac_iocb_req));
6318 
6319 		/*
6320 		 * The completion queue ID for the tx rings start
6321 		 * immediately after the last rss completion queue.
6322 		 */
6323 		tx_ring->cq_id = (uint16_t)(i + qlge->tx_ring_first_cq_id);
6324 	}
6325 
6326 	for (i = 0; i < qlge->rx_ring_count; i++) {
6327 		rx_ring = &qlge->rx_ring[i];
6328 		bzero((void *)rx_ring, sizeof (*rx_ring));
6329 		rx_ring->qlge = qlge;
6330 		rx_ring->cq_id = i;
6331 		if (i != 0)
6332 			rx_ring->cpu = (i) % qlge->rx_ring_count;
6333 		else
6334 			rx_ring->cpu = 0;
6335 
6336 		if (i < qlge->rss_ring_count) {
6337 			/*
6338 			 * Inbound completions (RSS) queues
6339 			 * Default queue is queue 0 which handles
6340 			 * unicast plus bcast/mcast and async events.
6341 			 * Other inbound queues handle unicast frames only.
6342 			 */
6343 			rx_ring->cq_len = qlge->rx_ring_size;
6344 			rx_ring->cq_size = (uint32_t)
6345 			    (rx_ring->cq_len * sizeof (struct net_rsp_iocb));
6346 			rx_ring->lbq_len = NUM_LARGE_BUFFERS;
6347 			rx_ring->lbq_size = (uint32_t)
6348 			    (rx_ring->lbq_len * sizeof (uint64_t));
6349 			rx_ring->lbq_buf_size = lbq_buf_size;
6350 			rx_ring->sbq_len = NUM_SMALL_BUFFERS;
6351 			rx_ring->sbq_size = (uint32_t)
6352 			    (rx_ring->sbq_len * sizeof (uint64_t));
6353 			rx_ring->sbq_buf_size = SMALL_BUFFER_SIZE * 2;
6354 			rx_ring->type = RX_Q;
6355 
6356 			QL_PRINT(DBG_GLD,
6357 			    ("%s(%d)Allocating rss completion queue %d "
6358 			    "on cpu %d\n", __func__, qlge->instance,
6359 			    rx_ring->cq_id, rx_ring->cpu));
6360 		} else {
6361 			/*
6362 			 * Outbound queue handles outbound completions only
6363 			 */
6364 			/* outbound cq is same size as tx_ring it services. */
6365 			QL_PRINT(DBG_INIT, ("rx_ring 0x%p i %d\n", rx_ring, i));
6366 			rx_ring->cq_len = qlge->tx_ring_size;
6367 			rx_ring->cq_size = (uint32_t)
6368 			    (rx_ring->cq_len * sizeof (struct net_rsp_iocb));
6369 			rx_ring->lbq_len = 0;
6370 			rx_ring->lbq_size = 0;
6371 			rx_ring->lbq_buf_size = 0;
6372 			rx_ring->sbq_len = 0;
6373 			rx_ring->sbq_size = 0;
6374 			rx_ring->sbq_buf_size = 0;
6375 			rx_ring->type = TX_Q;
6376 
6377 			QL_PRINT(DBG_GLD,
6378 			    ("%s(%d)Allocating TX completion queue %d on"
6379 			    " cpu %d\n", __func__, qlge->instance,
6380 			    rx_ring->cq_id, rx_ring->cpu));
6381 		}
6382 	}
6383 
6384 	return (DDI_SUCCESS);
6385 }
6386 
6387 static int
6388 ql_start_rx_ring(qlge_t *qlge, struct rx_ring *rx_ring)
6389 {
6390 	struct cqicb_t *cqicb = (struct cqicb_t *)rx_ring->cqicb_dma.vaddr;
6391 	void *shadow_reg = (uint8_t *)qlge->host_copy_shadow_dma_attr.vaddr +
6392 	    (rx_ring->cq_id * sizeof (uint64_t) * RX_TX_RING_SHADOW_SPACE)
6393 	/* first shadow area is used by wqicb's host copy of consumer index */
6394 	    + sizeof (uint64_t);
6395 	uint64_t shadow_reg_dma = qlge->host_copy_shadow_dma_attr.dma_addr +
6396 	    (rx_ring->cq_id * sizeof (uint64_t) * RX_TX_RING_SHADOW_SPACE)
6397 	    + sizeof (uint64_t);
6398 	/* lrg/sml bufq pointers */
6399 	uint8_t *buf_q_base_reg =
6400 	    (uint8_t *)qlge->buf_q_ptr_base_addr_dma_attr.vaddr +
6401 	    (rx_ring->cq_id * sizeof (uint64_t) * BUF_Q_PTR_SPACE);
6402 	uint64_t buf_q_base_reg_dma =
6403 	    qlge->buf_q_ptr_base_addr_dma_attr.dma_addr +
6404 	    (rx_ring->cq_id * sizeof (uint64_t) * BUF_Q_PTR_SPACE);
6405 	caddr_t doorbell_area =
6406 	    qlge->doorbell_reg_iobase + (VM_PAGE_SIZE * (128 + rx_ring->cq_id));
6407 	int err = 0;
6408 	uint16_t bq_len;
6409 	uint64_t tmp;
6410 	uint64_t *base_indirect_ptr;
6411 	int page_entries;
6412 
6413 	/* Set up the shadow registers for this ring. */
6414 	rx_ring->prod_idx_sh_reg = shadow_reg;
6415 	rx_ring->prod_idx_sh_reg_dma = shadow_reg_dma;
6416 	rx_ring->prod_idx_sh_reg_offset = (off_t)(((rx_ring->cq_id *
6417 	    sizeof (uint64_t) * RX_TX_RING_SHADOW_SPACE) + sizeof (uint64_t)));
6418 
6419 	rx_ring->lbq_base_indirect = (uint64_t *)(void *)buf_q_base_reg;
6420 	rx_ring->lbq_base_indirect_dma = buf_q_base_reg_dma;
6421 
6422 	QL_PRINT(DBG_INIT, ("%s rx ring(%d): prod_idx virtual addr = 0x%lx,"
6423 	    " phys_addr 0x%lx\n", __func__, rx_ring->cq_id,
6424 	    rx_ring->prod_idx_sh_reg, rx_ring->prod_idx_sh_reg_dma));
6425 
6426 	buf_q_base_reg += ((BUF_Q_PTR_SPACE / 2) * sizeof (uint64_t));
6427 	buf_q_base_reg_dma += ((BUF_Q_PTR_SPACE / 2) * sizeof (uint64_t));
6428 	rx_ring->sbq_base_indirect = (uint64_t *)(void *)buf_q_base_reg;
6429 	rx_ring->sbq_base_indirect_dma = buf_q_base_reg_dma;
6430 
6431 	/* PCI doorbell mem area + 0x00 for consumer index register */
6432 	rx_ring->cnsmr_idx_db_reg = (uint32_t *)(void *)doorbell_area;
6433 	rx_ring->cnsmr_idx = 0;
6434 	*rx_ring->prod_idx_sh_reg = 0;
6435 	rx_ring->curr_entry = rx_ring->cq_dma.vaddr;
6436 
6437 	/* PCI doorbell mem area + 0x04 for valid register */
6438 	rx_ring->valid_db_reg = (uint32_t *)(void *)
6439 	    ((uint8_t *)(void *)doorbell_area + 0x04);
6440 
6441 	/* PCI doorbell mem area + 0x18 for large buffer consumer */
6442 	rx_ring->lbq_prod_idx_db_reg = (uint32_t *)(void *)
6443 	    ((uint8_t *)(void *)doorbell_area + 0x18);
6444 
6445 	/* PCI doorbell mem area + 0x1c */
6446 	rx_ring->sbq_prod_idx_db_reg = (uint32_t *)(void *)
6447 	    ((uint8_t *)(void *)doorbell_area + 0x1c);
6448 
6449 	bzero((void *)cqicb, sizeof (*cqicb));
6450 
6451 	cqicb->msix_vect = (uint8_t)rx_ring->irq;
6452 
6453 	bq_len = (uint16_t)((rx_ring->cq_len == 65536) ?
6454 	    (uint16_t)0 : (uint16_t)rx_ring->cq_len);
6455 	cqicb->len = (uint16_t)cpu_to_le16(bq_len | LEN_V | LEN_CPP_CONT);
6456 
6457 	cqicb->cq_base_addr_lo =
6458 	    cpu_to_le32(LS_64BITS(rx_ring->cq_dma.dma_addr));
6459 	cqicb->cq_base_addr_hi =
6460 	    cpu_to_le32(MS_64BITS(rx_ring->cq_dma.dma_addr));
6461 
6462 	cqicb->prod_idx_addr_lo =
6463 	    cpu_to_le32(LS_64BITS(rx_ring->prod_idx_sh_reg_dma));
6464 	cqicb->prod_idx_addr_hi =
6465 	    cpu_to_le32(MS_64BITS(rx_ring->prod_idx_sh_reg_dma));
6466 
6467 	/*
6468 	 * Set up the control block load flags.
6469 	 */
6470 	cqicb->flags = FLAGS_LC | /* Load queue base address */
6471 	    FLAGS_LV | /* Load MSI-X vector */
6472 	    FLAGS_LI;  /* Load irq delay values */
6473 	if (rx_ring->lbq_len) {
6474 		/* Load lbq values */
6475 		cqicb->flags = (uint8_t)(cqicb->flags | FLAGS_LL);
6476 		tmp = (uint64_t)rx_ring->lbq_dma.dma_addr;
6477 		base_indirect_ptr = (uint64_t *)rx_ring->lbq_base_indirect;
6478 		page_entries = 0;
6479 		do {
6480 			*base_indirect_ptr = cpu_to_le64(tmp);
6481 			tmp += VM_PAGE_SIZE;
6482 			base_indirect_ptr++;
6483 			page_entries++;
6484 		} while (page_entries < (int)(
6485 		    ((rx_ring->lbq_len * sizeof (uint64_t)) / VM_PAGE_SIZE)));
6486 
6487 		cqicb->lbq_addr_lo =
6488 		    cpu_to_le32(LS_64BITS(rx_ring->lbq_base_indirect_dma));
6489 		cqicb->lbq_addr_hi =
6490 		    cpu_to_le32(MS_64BITS(rx_ring->lbq_base_indirect_dma));
6491 		bq_len = (uint16_t)((rx_ring->lbq_buf_size == 65536) ?
6492 		    (uint16_t)0 : (uint16_t)rx_ring->lbq_buf_size);
6493 		cqicb->lbq_buf_size = (uint16_t)cpu_to_le16(bq_len);
6494 		bq_len = (uint16_t)((rx_ring->lbq_len == 65536) ? (uint16_t)0 :
6495 		    (uint16_t)rx_ring->lbq_len);
6496 		cqicb->lbq_len = (uint16_t)cpu_to_le16(bq_len);
6497 		rx_ring->lbq_prod_idx = 0;
6498 		rx_ring->lbq_curr_idx = 0;
6499 	}
6500 	if (rx_ring->sbq_len) {
6501 		/* Load sbq values */
6502 		cqicb->flags = (uint8_t)(cqicb->flags | FLAGS_LS);
6503 		tmp = (uint64_t)rx_ring->sbq_dma.dma_addr;
6504 		base_indirect_ptr = (uint64_t *)rx_ring->sbq_base_indirect;
6505 		page_entries = 0;
6506 
6507 		do {
6508 			*base_indirect_ptr = cpu_to_le64(tmp);
6509 			tmp += VM_PAGE_SIZE;
6510 			base_indirect_ptr++;
6511 			page_entries++;
6512 		} while (page_entries < (uint32_t)
6513 		    (((rx_ring->sbq_len * sizeof (uint64_t)) / VM_PAGE_SIZE)));
6514 
6515 		cqicb->sbq_addr_lo =
6516 		    cpu_to_le32(LS_64BITS(rx_ring->sbq_base_indirect_dma));
6517 		cqicb->sbq_addr_hi =
6518 		    cpu_to_le32(MS_64BITS(rx_ring->sbq_base_indirect_dma));
6519 		cqicb->sbq_buf_size = (uint16_t)
6520 		    cpu_to_le16((uint16_t)(rx_ring->sbq_buf_size/2));
6521 		bq_len = (uint16_t)((rx_ring->sbq_len == 65536) ?
6522 		    (uint16_t)0 : (uint16_t)rx_ring->sbq_len);
6523 		cqicb->sbq_len = (uint16_t)cpu_to_le16(bq_len);
6524 		rx_ring->sbq_prod_idx = 0;
6525 		rx_ring->sbq_curr_idx = 0;
6526 	}
6527 	switch (rx_ring->type) {
6528 	case TX_Q:
6529 		cqicb->irq_delay = (uint16_t)
6530 		    cpu_to_le16(qlge->tx_coalesce_usecs);
6531 		cqicb->pkt_delay = (uint16_t)
6532 		    cpu_to_le16(qlge->tx_max_coalesced_frames);
6533 		break;
6534 
6535 	case DEFAULT_Q:
6536 		cqicb->irq_delay = (uint16_t)
6537 		    cpu_to_le16(qlge->rx_coalesce_usecs);
6538 		cqicb->pkt_delay = (uint16_t)
6539 		    cpu_to_le16(qlge->rx_max_coalesced_frames);
6540 		break;
6541 
6542 	case RX_Q:
6543 		/*
6544 		 * Inbound completion handling rx_rings run in
6545 		 * separate NAPI contexts.
6546 		 */
6547 		cqicb->irq_delay = (uint16_t)
6548 		    cpu_to_le16(qlge->rx_coalesce_usecs);
6549 		cqicb->pkt_delay = (uint16_t)
6550 		    cpu_to_le16(qlge->rx_max_coalesced_frames);
6551 		break;
6552 	default:
6553 		cmn_err(CE_WARN, "Invalid rx_ring->type = %d.",
6554 		    rx_ring->type);
6555 	}
6556 	QL_PRINT(DBG_INIT, ("Initializing rx completion queue %d.\n",
6557 	    rx_ring->cq_id));
6558 	/* QL_DUMP_CQICB(qlge, cqicb); */
6559 	err = ql_write_cfg(qlge, CFG_LCQ, rx_ring->cqicb_dma.dma_addr,
6560 	    rx_ring->cq_id);
6561 	if (err) {
6562 		cmn_err(CE_WARN, "Failed to load CQICB.");
6563 		return (err);
6564 	}
6565 
6566 	rx_ring->rx_packets_dropped_no_buffer = 0;
6567 	rx_ring->rx_pkt_dropped_mac_unenabled = 0;
6568 	rx_ring->rx_failed_sbq_allocs = 0;
6569 	rx_ring->rx_failed_lbq_allocs = 0;
6570 	rx_ring->rx_packets = 0;
6571 	rx_ring->rx_bytes = 0;
6572 	rx_ring->frame_too_long = 0;
6573 	rx_ring->frame_too_short = 0;
6574 	rx_ring->fcs_err = 0;
6575 
6576 	return (err);
6577 }
6578 
6579 /*
6580  * start RSS
6581  */
6582 static int
6583 ql_start_rss(qlge_t *qlge)
6584 {
6585 	struct ricb *ricb = (struct ricb *)qlge->ricb_dma.vaddr;
6586 	int status = 0;
6587 	int i;
6588 	uint8_t *hash_id = (uint8_t *)ricb->hash_cq_id;
6589 
6590 	bzero((void *)ricb, sizeof (*ricb));
6591 
6592 	ricb->base_cq = RSS_L4K;
6593 	ricb->flags =
6594 	    (RSS_L6K | RSS_LI | RSS_LB | RSS_LM | RSS_RI4 | RSS_RI6 | RSS_RT4 |
6595 	    RSS_RT6);
6596 	ricb->mask = (uint16_t)cpu_to_le16(RSS_HASH_CQ_ID_MAX - 1);
6597 
6598 	/*
6599 	 * Fill out the Indirection Table.
6600 	 */
6601 	for (i = 0; i < RSS_HASH_CQ_ID_MAX; i++)
6602 		hash_id[i] = (uint8_t)(i & (qlge->rss_ring_count - 1));
6603 
6604 	(void) memcpy(&ricb->ipv6_hash_key[0], key_data, 40);
6605 	(void) memcpy(&ricb->ipv4_hash_key[0], key_data, 16);
6606 
6607 	QL_PRINT(DBG_INIT, ("Initializing RSS.\n"));
6608 
6609 	status = ql_write_cfg(qlge, CFG_LR, qlge->ricb_dma.dma_addr, 0);
6610 	if (status) {
6611 		cmn_err(CE_WARN, "Failed to load RICB.");
6612 		return (status);
6613 	}
6614 
6615 	return (status);
6616 }
6617 
6618 /*
6619  * load a tx ring control block to hw and start this ring
6620  */
6621 static int
6622 ql_start_tx_ring(qlge_t *qlge, struct tx_ring *tx_ring)
6623 {
6624 	struct wqicb_t *wqicb = (struct wqicb_t *)tx_ring->wqicb_dma.vaddr;
6625 	caddr_t doorbell_area =
6626 	    qlge->doorbell_reg_iobase + (VM_PAGE_SIZE * tx_ring->wq_id);
6627 	void *shadow_reg = (uint8_t *)qlge->host_copy_shadow_dma_attr.vaddr +
6628 	    (tx_ring->wq_id * sizeof (uint64_t)) * RX_TX_RING_SHADOW_SPACE;
6629 	uint64_t shadow_reg_dma = qlge->host_copy_shadow_dma_attr.dma_addr +
6630 	    (tx_ring->wq_id * sizeof (uint64_t)) * RX_TX_RING_SHADOW_SPACE;
6631 	int err = 0;
6632 
6633 	/*
6634 	 * Assign doorbell registers for this tx_ring.
6635 	 */
6636 
6637 	/* TX PCI doorbell mem area for tx producer index */
6638 	tx_ring->prod_idx_db_reg = (uint32_t *)(void *)doorbell_area;
6639 	tx_ring->prod_idx = 0;
6640 	/* TX PCI doorbell mem area + 0x04 */
6641 	tx_ring->valid_db_reg = (uint32_t *)(void *)
6642 	    ((uint8_t *)(void *)doorbell_area + 0x04);
6643 
6644 	/*
6645 	 * Assign shadow registers for this tx_ring.
6646 	 */
6647 	tx_ring->cnsmr_idx_sh_reg = shadow_reg;
6648 	tx_ring->cnsmr_idx_sh_reg_dma = shadow_reg_dma;
6649 	*tx_ring->cnsmr_idx_sh_reg = 0;
6650 
6651 	QL_PRINT(DBG_INIT, ("%s tx ring(%d): cnsmr_idx virtual addr = 0x%lx,"
6652 	    " phys_addr 0x%lx\n",
6653 	    __func__, tx_ring->wq_id, tx_ring->cnsmr_idx_sh_reg,
6654 	    tx_ring->cnsmr_idx_sh_reg_dma));
6655 
6656 	wqicb->len =
6657 	    (uint16_t)cpu_to_le16(tx_ring->wq_len | Q_LEN_V | Q_LEN_CPP_CONT);
6658 	wqicb->flags = cpu_to_le16(Q_FLAGS_LC |
6659 	    Q_FLAGS_LB | Q_FLAGS_LI | Q_FLAGS_LO);
6660 	wqicb->cq_id_rss = (uint16_t)cpu_to_le16(tx_ring->cq_id);
6661 	wqicb->rid = 0;
6662 	wqicb->wq_addr_lo = cpu_to_le32(LS_64BITS(tx_ring->wq_dma.dma_addr));
6663 	wqicb->wq_addr_hi = cpu_to_le32(MS_64BITS(tx_ring->wq_dma.dma_addr));
6664 	wqicb->cnsmr_idx_addr_lo =
6665 	    cpu_to_le32(LS_64BITS(tx_ring->cnsmr_idx_sh_reg_dma));
6666 	wqicb->cnsmr_idx_addr_hi =
6667 	    cpu_to_le32(MS_64BITS(tx_ring->cnsmr_idx_sh_reg_dma));
6668 
6669 	ql_init_tx_ring(tx_ring);
6670 	/* QL_DUMP_WQICB(qlge, wqicb); */
6671 	err = ql_write_cfg(qlge, CFG_LRQ, tx_ring->wqicb_dma.dma_addr,
6672 	    tx_ring->wq_id);
6673 
6674 	if (err) {
6675 		cmn_err(CE_WARN, "Failed to load WQICB.");
6676 		return (err);
6677 	}
6678 	return (err);
6679 }
6680 
6681 /*
6682  * Set up a MAC, multicast or VLAN address for the
6683  * inbound frame matching.
6684  */
6685 int
6686 ql_set_mac_addr_reg(qlge_t *qlge, uint8_t *addr, uint32_t type,
6687     uint16_t index)
6688 {
6689 	uint32_t offset = 0;
6690 	int status = DDI_SUCCESS;
6691 
6692 	switch (type) {
6693 	case MAC_ADDR_TYPE_MULTI_MAC:
6694 	case MAC_ADDR_TYPE_CAM_MAC: {
6695 		uint32_t cam_output;
6696 		uint32_t upper = (addr[0] << 8) | addr[1];
6697 		uint32_t lower =
6698 		    (addr[2] << 24) | (addr[3] << 16) | (addr[4] << 8) |
6699 		    (addr[5]);
6700 
6701 		QL_PRINT(DBG_INIT, ("Adding %s ", (type ==
6702 		    MAC_ADDR_TYPE_MULTI_MAC) ?
6703 		    "MULTICAST" : "UNICAST"));
6704 		QL_PRINT(DBG_INIT,
6705 		    ("addr %02x %02x %02x %02x %02x %02x at index %d in "
6706 		    "the CAM.\n",
6707 		    addr[0], addr[1], addr[2], addr[3], addr[4],
6708 		    addr[5], index));
6709 
6710 		status = ql_wait_reg_rdy(qlge,
6711 		    REG_MAC_PROTOCOL_ADDRESS_INDEX, MAC_ADDR_MW, 0);
6712 		if (status)
6713 			goto exit;
6714 		/* offset 0 - lower 32 bits of the MAC address */
6715 		ql_write_reg(qlge, REG_MAC_PROTOCOL_ADDRESS_INDEX,
6716 		    (offset++) |
6717 		    (index << MAC_ADDR_IDX_SHIFT) | /* index */
6718 		    type);	/* type */
6719 		ql_write_reg(qlge, REG_MAC_PROTOCOL_DATA, lower);
6720 		status = ql_wait_reg_rdy(qlge,
6721 		    REG_MAC_PROTOCOL_ADDRESS_INDEX, MAC_ADDR_MW, 0);
6722 		if (status)
6723 			goto exit;
6724 		/* offset 1 - upper 16 bits of the MAC address */
6725 		ql_write_reg(qlge, REG_MAC_PROTOCOL_ADDRESS_INDEX,
6726 		    (offset++) |
6727 		    (index << MAC_ADDR_IDX_SHIFT) | /* index */
6728 		    type);	/* type */
6729 		ql_write_reg(qlge, REG_MAC_PROTOCOL_DATA, upper);
6730 		status = ql_wait_reg_rdy(qlge,
6731 		    REG_MAC_PROTOCOL_ADDRESS_INDEX, MAC_ADDR_MW, 0);
6732 		if (status)
6733 			goto exit;
6734 		/* offset 2 - CQ ID associated with this MAC address */
6735 		ql_write_reg(qlge, REG_MAC_PROTOCOL_ADDRESS_INDEX,
6736 		    (offset) | (index << MAC_ADDR_IDX_SHIFT) |	/* index */
6737 		    type);	/* type */
6738 		/*
6739 		 * This field should also include the queue id
6740 		 * and possibly the function id.  Right now we hardcode
6741 		 * the route field to NIC core.
6742 		 */
6743 		if (type == MAC_ADDR_TYPE_CAM_MAC) {
6744 			cam_output = (CAM_OUT_ROUTE_NIC |
6745 			    (qlge->func_number << CAM_OUT_FUNC_SHIFT) |
6746 			    (0 <<
6747 			    CAM_OUT_CQ_ID_SHIFT));
6748 
6749 			/* route to NIC core */
6750 			ql_write_reg(qlge, REG_MAC_PROTOCOL_DATA,
6751 			    cam_output);
6752 			}
6753 		break;
6754 		}
6755 	default:
6756 		cmn_err(CE_WARN,
6757 		    "Address type %d not yet supported.", type);
6758 		status = DDI_FAILURE;
6759 	}
6760 exit:
6761 	return (status);
6762 }
6763 
6764 /*
6765  * The NIC function for this chip has 16 routing indexes.  Each one can be used
6766  * to route different frame types to various inbound queues.  We send broadcast
6767  * multicast/error frames to the default queue for slow handling,
6768  * and CAM hit/RSS frames to the fast handling queues.
6769  */
6770 static int
6771 ql_set_routing_reg(qlge_t *qlge, uint32_t index, uint32_t mask, int enable)
6772 {
6773 	int status;
6774 	uint32_t value = 0;
6775 
6776 	QL_PRINT(DBG_INIT,
6777 	    ("%s %s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s mask %s the routing reg.\n",
6778 	    (enable ? "Adding" : "Removing"),
6779 	    ((index == RT_IDX_ALL_ERR_SLOT) ? "MAC ERROR/ALL ERROR" : ""),
6780 	    ((index == RT_IDX_IP_CSUM_ERR_SLOT) ? "IP CSUM ERROR" : ""),
6781 	    ((index ==
6782 	    RT_IDX_TCP_UDP_CSUM_ERR_SLOT) ? "TCP/UDP CSUM ERROR" : ""),
6783 	    ((index == RT_IDX_BCAST_SLOT) ? "BROADCAST" : ""),
6784 	    ((index == RT_IDX_MCAST_MATCH_SLOT) ? "MULTICAST MATCH" : ""),
6785 	    ((index == RT_IDX_ALLMULTI_SLOT) ? "ALL MULTICAST MATCH" : ""),
6786 	    ((index == RT_IDX_UNUSED6_SLOT) ? "UNUSED6" : ""),
6787 	    ((index == RT_IDX_UNUSED7_SLOT) ? "UNUSED7" : ""),
6788 	    ((index == RT_IDX_RSS_MATCH_SLOT) ? "RSS ALL/IPV4 MATCH" : ""),
6789 	    ((index == RT_IDX_RSS_IPV6_SLOT) ? "RSS IPV6" : ""),
6790 	    ((index == RT_IDX_RSS_TCP4_SLOT) ? "RSS TCP4" : ""),
6791 	    ((index == RT_IDX_RSS_TCP6_SLOT) ? "RSS TCP6" : ""),
6792 	    ((index == RT_IDX_CAM_HIT_SLOT) ? "CAM HIT" : ""),
6793 	    ((index == RT_IDX_UNUSED013) ? "UNUSED13" : ""),
6794 	    ((index == RT_IDX_UNUSED014) ? "UNUSED14" : ""),
6795 	    ((index == RT_IDX_PROMISCUOUS_SLOT) ? "PROMISCUOUS" : ""),
6796 	    (enable ? "to" : "from")));
6797 
6798 	switch (mask) {
6799 	case RT_IDX_CAM_HIT:
6800 		value = RT_IDX_DST_CAM_Q | /* dest */
6801 		    RT_IDX_TYPE_NICQ | /* type */
6802 		    (RT_IDX_CAM_HIT_SLOT << RT_IDX_IDX_SHIFT); /* index */
6803 		break;
6804 
6805 	case RT_IDX_VALID: /* Promiscuous Mode frames. */
6806 		value = RT_IDX_DST_DFLT_Q |	/* dest */
6807 		    RT_IDX_TYPE_NICQ |	/* type */
6808 		    (RT_IDX_PROMISCUOUS_SLOT << RT_IDX_IDX_SHIFT); /* index */
6809 		break;
6810 
6811 	case RT_IDX_ERR:	/* Pass up MAC,IP,TCP/UDP error frames. */
6812 		value = RT_IDX_DST_DFLT_Q |	/* dest */
6813 		    RT_IDX_TYPE_NICQ |	/* type */
6814 		    (RT_IDX_ALL_ERR_SLOT << RT_IDX_IDX_SHIFT); /* index */
6815 		break;
6816 
6817 	case RT_IDX_BCAST:	/* Pass up Broadcast frames to default Q. */
6818 		value = RT_IDX_DST_DFLT_Q |	/* dest */
6819 		    RT_IDX_TYPE_NICQ |	/* type */
6820 		    (RT_IDX_BCAST_SLOT << RT_IDX_IDX_SHIFT); /* index */
6821 		break;
6822 
6823 	case RT_IDX_MCAST:	/* Pass up All Multicast frames. */
6824 		value = RT_IDX_DST_CAM_Q |	/* dest */
6825 		    RT_IDX_TYPE_NICQ |	/* type */
6826 		    (RT_IDX_ALLMULTI_SLOT << RT_IDX_IDX_SHIFT); /* index */
6827 		break;
6828 
6829 	case RT_IDX_MCAST_MATCH:	/* Pass up matched Multicast frames. */
6830 		value = RT_IDX_DST_CAM_Q |	/* dest */
6831 		    RT_IDX_TYPE_NICQ |	/* type */
6832 		    (RT_IDX_MCAST_MATCH_SLOT << RT_IDX_IDX_SHIFT); /* index */
6833 		break;
6834 
6835 	case RT_IDX_RSS_MATCH:	/* Pass up matched RSS frames. */
6836 		value = RT_IDX_DST_RSS |	/* dest */
6837 		    RT_IDX_TYPE_NICQ |	/* type */
6838 		    (RT_IDX_RSS_MATCH_SLOT << RT_IDX_IDX_SHIFT); /* index */
6839 		break;
6840 
6841 	case 0:	/* Clear the E-bit on an entry. */
6842 		value = RT_IDX_DST_DFLT_Q |	/* dest */
6843 		    RT_IDX_TYPE_NICQ |	/* type */
6844 		    (index << RT_IDX_IDX_SHIFT); /* index */
6845 		break;
6846 
6847 	default:
6848 		cmn_err(CE_WARN, "Mask type %d not yet supported.",
6849 		    mask);
6850 		status = -EPERM;
6851 		goto exit;
6852 	}
6853 
6854 	if (value != 0) {
6855 		status = ql_wait_reg_rdy(qlge, REG_ROUTING_INDEX, RT_IDX_MW, 0);
6856 		if (status)
6857 			goto exit;
6858 		value |= (enable ? RT_IDX_E : 0);
6859 		ql_write_reg(qlge, REG_ROUTING_INDEX, value);
6860 		ql_write_reg(qlge, REG_ROUTING_DATA, enable ? mask : 0);
6861 	}
6862 
6863 exit:
6864 	return (status);
6865 }
6866 
6867 /*
6868  * Clear all the entries in the routing table.
6869  * Caller must get semaphore in advance.
6870  */
6871 
6872 static int
6873 ql_stop_routing(qlge_t *qlge)
6874 {
6875 	int status = 0;
6876 	int i;
6877 	/* Clear all the entries in the routing table. */
6878 	for (i = 0; i < 16; i++) {
6879 		status = ql_set_routing_reg(qlge, i, 0, 0);
6880 		if (status) {
6881 			cmn_err(CE_WARN, "Stop routing failed. ");
6882 		}
6883 	}
6884 	return (status);
6885 }
6886 
6887 /* Initialize the frame-to-queue routing. */
6888 int
6889 ql_route_initialize(qlge_t *qlge)
6890 {
6891 	int status = 0;
6892 
6893 	status = ql_sem_spinlock(qlge, SEM_RT_IDX_MASK);
6894 	if (status != DDI_SUCCESS)
6895 		return (status);
6896 
6897 	/* Clear all the entries in the routing table. */
6898 	status = ql_stop_routing(qlge);
6899 	if (status) {
6900 		goto exit;
6901 	}
6902 	status = ql_set_routing_reg(qlge, RT_IDX_BCAST_SLOT, RT_IDX_BCAST, 1);
6903 	if (status) {
6904 		cmn_err(CE_WARN,
6905 		    "Failed to init routing register for broadcast packets.");
6906 		goto exit;
6907 	}
6908 	/*
6909 	 * If we have more than one inbound queue, then turn on RSS in the
6910 	 * routing block.
6911 	 */
6912 	if (qlge->rss_ring_count > 1) {
6913 		status = ql_set_routing_reg(qlge, RT_IDX_RSS_MATCH_SLOT,
6914 		    RT_IDX_RSS_MATCH, 1);
6915 		if (status) {
6916 			cmn_err(CE_WARN,
6917 			    "Failed to init routing register for MATCH RSS "
6918 			    "packets.");
6919 			goto exit;
6920 		}
6921 	}
6922 
6923 	status = ql_set_routing_reg(qlge, RT_IDX_CAM_HIT_SLOT,
6924 	    RT_IDX_CAM_HIT, 1);
6925 	if (status) {
6926 		cmn_err(CE_WARN,
6927 		    "Failed to init routing register for CAM packets.");
6928 		goto exit;
6929 	}
6930 
6931 	status = ql_set_routing_reg(qlge, RT_IDX_MCAST_MATCH_SLOT,
6932 	    RT_IDX_MCAST_MATCH, 1);
6933 	if (status) {
6934 		cmn_err(CE_WARN,
6935 		    "Failed to init routing register for Multicast "
6936 		    "packets.");
6937 	}
6938 
6939 exit:
6940 	ql_sem_unlock(qlge, SEM_RT_IDX_MASK);
6941 	return (status);
6942 }
6943 
6944 /*
6945  * Initialize hardware
6946  */
6947 static int
6948 ql_device_initialize(qlge_t *qlge)
6949 {
6950 	uint32_t value, mask;
6951 	int i;
6952 	int status = 0;
6953 	uint16_t pause = PAUSE_MODE_DISABLED;
6954 	boolean_t update_port_config = B_FALSE;
6955 	uint32_t pause_bit_mask;
6956 	boolean_t dcbx_enable = B_FALSE;
6957 	uint32_t dcbx_bit_mask = 0x10;
6958 	/*
6959 	 * Set up the System register to halt on errors.
6960 	 */
6961 	value = SYS_EFE | SYS_FAE;
6962 	mask = value << 16;
6963 	ql_write_reg(qlge, REG_SYSTEM, mask | value);
6964 
6965 	/* Set the default queue. */
6966 	value = NIC_RCV_CFG_DFQ;
6967 	mask = NIC_RCV_CFG_DFQ_MASK;
6968 
6969 	ql_write_reg(qlge, REG_NIC_RECEIVE_CONFIGURATION, mask | value);
6970 
6971 	/* Enable the MPI interrupt. */
6972 	ql_write_reg(qlge, REG_INTERRUPT_MASK, (INTR_MASK_PI << 16)
6973 	    | INTR_MASK_PI);
6974 	/* Enable the function, set pagesize, enable error checking. */
6975 	value = FSC_FE | FSC_EPC_INBOUND | FSC_EPC_OUTBOUND |
6976 	    FSC_EC | FSC_VM_PAGE_4K | FSC_DBRST_1024;
6977 	/* Set/clear header splitting. */
6978 	if (CFG_IST(qlge, CFG_ENABLE_SPLIT_HEADER)) {
6979 		value |= FSC_SH;
6980 		ql_write_reg(qlge, REG_SPLIT_HEADER, SMALL_BUFFER_SIZE);
6981 	}
6982 	mask = FSC_VM_PAGESIZE_MASK |
6983 	    FSC_DBL_MASK | FSC_DBRST_MASK | (value << 16);
6984 	ql_write_reg(qlge, REG_FUNCTION_SPECIFIC_CONTROL, mask | value);
6985 	/*
6986 	 * check current port max frame size, if different from OS setting,
6987 	 * then we need to change
6988 	 */
6989 	qlge->max_frame_size =
6990 	    (qlge->mtu == ETHERMTU)? NORMAL_FRAME_SIZE : JUMBO_FRAME_SIZE;
6991 
6992 	mutex_enter(&qlge->mbx_mutex);
6993 	status = ql_get_port_cfg(qlge);
6994 	mutex_exit(&qlge->mbx_mutex);
6995 
6996 	if (status == DDI_SUCCESS) {
6997 		/* if current frame size is smaller than required size */
6998 		if (qlge->port_cfg_info.max_frame_size <
6999 		    qlge->max_frame_size) {
7000 			QL_PRINT(DBG_MBX,
7001 			    ("update frame size, current %d, new %d\n",
7002 			    qlge->port_cfg_info.max_frame_size,
7003 			    qlge->max_frame_size));
7004 			qlge->port_cfg_info.max_frame_size =
7005 			    qlge->max_frame_size;
7006 			qlge->port_cfg_info.link_cfg |= ENABLE_JUMBO;
7007 			update_port_config = B_TRUE;
7008 		}
7009 
7010 		if (qlge->port_cfg_info.link_cfg & STD_PAUSE)
7011 			pause = PAUSE_MODE_STANDARD;
7012 		else if (qlge->port_cfg_info.link_cfg & PP_PAUSE)
7013 			pause = PAUSE_MODE_PER_PRIORITY;
7014 
7015 		if (pause != qlge->pause) {
7016 			pause_bit_mask = 0x60;	/* bit 5-6 */
7017 			/* clear pause bits */
7018 			qlge->port_cfg_info.link_cfg &= ~pause_bit_mask;
7019 			if (qlge->pause == PAUSE_MODE_STANDARD)
7020 				qlge->port_cfg_info.link_cfg |= STD_PAUSE;
7021 			else if (qlge->pause == PAUSE_MODE_PER_PRIORITY)
7022 				qlge->port_cfg_info.link_cfg |= PP_PAUSE;
7023 			update_port_config = B_TRUE;
7024 		}
7025 
7026 		if (qlge->port_cfg_info.link_cfg & DCBX_ENABLE)
7027 			dcbx_enable = B_TRUE;
7028 		if (dcbx_enable != qlge->dcbx_enable) {
7029 			qlge->port_cfg_info.link_cfg &= ~dcbx_bit_mask;
7030 			if (qlge->dcbx_enable)
7031 				qlge->port_cfg_info.link_cfg |= DCBX_ENABLE;
7032 		}
7033 
7034 		update_port_config = B_TRUE;
7035 
7036 		/* if need to update port configuration */
7037 		if (update_port_config) {
7038 			mutex_enter(&qlge->mbx_mutex);
7039 			(void) ql_set_mpi_port_config(qlge,
7040 			    qlge->port_cfg_info);
7041 			mutex_exit(&qlge->mbx_mutex);
7042 		}
7043 	} else
7044 		cmn_err(CE_WARN, "ql_get_port_cfg failed");
7045 
7046 	/* Start up the rx queues. */
7047 	for (i = 0; i < qlge->rx_ring_count; i++) {
7048 		status = ql_start_rx_ring(qlge, &qlge->rx_ring[i]);
7049 		if (status) {
7050 			cmn_err(CE_WARN,
7051 			    "Failed to start rx ring[%d]", i);
7052 			return (status);
7053 		}
7054 	}
7055 
7056 	/*
7057 	 * If there is more than one inbound completion queue
7058 	 * then download a RICB to configure RSS.
7059 	 */
7060 	if (qlge->rss_ring_count > 1) {
7061 		status = ql_start_rss(qlge);
7062 		if (status) {
7063 			cmn_err(CE_WARN, "Failed to start RSS.");
7064 			return (status);
7065 		}
7066 	}
7067 
7068 	/* Start up the tx queues. */
7069 	for (i = 0; i < qlge->tx_ring_count; i++) {
7070 		status = ql_start_tx_ring(qlge, &qlge->tx_ring[i]);
7071 		if (status) {
7072 			cmn_err(CE_WARN,
7073 			    "Failed to start tx ring[%d]", i);
7074 			return (status);
7075 		}
7076 	}
7077 	qlge->selected_tx_ring = 0;
7078 	/* Set the frame routing filter. */
7079 	status = ql_route_initialize(qlge);
7080 	if (status) {
7081 		cmn_err(CE_WARN,
7082 		    "Failed to init CAM/Routing tables.");
7083 		return (status);
7084 	}
7085 
7086 	return (status);
7087 }
7088 /*
7089  * Issue soft reset to chip.
7090  */
7091 static int
7092 ql_asic_reset(qlge_t *qlge)
7093 {
7094 	int status = DDI_SUCCESS;
7095 
7096 	ql_write_reg(qlge, REG_RESET_FAILOVER, FUNCTION_RESET_MASK
7097 	    |FUNCTION_RESET);
7098 
7099 	if (ql_wait_reg_bit(qlge, REG_RESET_FAILOVER, FUNCTION_RESET,
7100 	    BIT_RESET, 0) != DDI_SUCCESS) {
7101 		cmn_err(CE_WARN,
7102 		    "TIMEOUT!!! errored out of resetting the chip!");
7103 		status = DDI_FAILURE;
7104 	}
7105 
7106 	return (status);
7107 }
7108 
7109 /*
7110  * If there are more than MIN_BUFFERS_ARM_COUNT small buffer descriptors in
7111  * its free list, move xMIN_BUFFERS_ARM_COUNT descriptors to its in use list
7112  * to be used by hardware.
7113  */
7114 static void
7115 ql_arm_sbuf(qlge_t *qlge, struct rx_ring *rx_ring)
7116 {
7117 	struct bq_desc *sbq_desc;
7118 	int i;
7119 	uint64_t *sbq_entry = rx_ring->sbq_dma.vaddr;
7120 	uint32_t arm_count;
7121 
7122 	if (rx_ring->sbuf_free_count > rx_ring->sbq_len-MIN_BUFFERS_ARM_COUNT)
7123 		arm_count = (rx_ring->sbq_len-MIN_BUFFERS_ARM_COUNT);
7124 	else {
7125 		/* Adjust to a multiple of 16 */
7126 		arm_count = (rx_ring->sbuf_free_count / 16) * 16;
7127 #ifdef QLGE_LOAD_UNLOAD
7128 		cmn_err(CE_NOTE, "adjust sbuf arm_count %d\n", arm_count);
7129 #endif
7130 	}
7131 	for (i = 0; i < arm_count; i++) {
7132 		sbq_desc = ql_get_sbuf_from_free_list(rx_ring);
7133 		if (sbq_desc == NULL)
7134 			break;
7135 		/* Arm asic */
7136 		*sbq_entry = cpu_to_le64(sbq_desc->bd_dma.dma_addr);
7137 		sbq_entry++;
7138 
7139 		/* link the descriptors to in_use_list */
7140 		ql_add_sbuf_to_in_use_list(rx_ring, sbq_desc);
7141 		rx_ring->sbq_prod_idx++;
7142 	}
7143 	ql_update_sbq_prod_idx(qlge, rx_ring);
7144 }
7145 
7146 /*
7147  * If there are more than MIN_BUFFERS_ARM_COUNT large buffer descriptors in
7148  * its free list, move xMIN_BUFFERS_ARM_COUNT descriptors to its in use list
7149  * to be used by hardware.
7150  */
7151 static void
7152 ql_arm_lbuf(qlge_t *qlge, struct rx_ring *rx_ring)
7153 {
7154 	struct bq_desc *lbq_desc;
7155 	int i;
7156 	uint64_t *lbq_entry = rx_ring->lbq_dma.vaddr;
7157 	uint32_t arm_count;
7158 
7159 	if (rx_ring->lbuf_free_count > rx_ring->lbq_len-MIN_BUFFERS_ARM_COUNT)
7160 		arm_count = (rx_ring->lbq_len-MIN_BUFFERS_ARM_COUNT);
7161 	else {
7162 		/* Adjust to a multiple of 16 */
7163 		arm_count = (rx_ring->lbuf_free_count / 16) * 16;
7164 #ifdef QLGE_LOAD_UNLOAD
7165 		cmn_err(CE_NOTE, "adjust lbuf arm_count %d\n", arm_count);
7166 #endif
7167 	}
7168 	for (i = 0; i < arm_count; i++) {
7169 		lbq_desc = ql_get_lbuf_from_free_list(rx_ring);
7170 		if (lbq_desc == NULL)
7171 			break;
7172 		/* Arm asic */
7173 		*lbq_entry = cpu_to_le64(lbq_desc->bd_dma.dma_addr);
7174 		lbq_entry++;
7175 
7176 		/* link the descriptors to in_use_list */
7177 		ql_add_lbuf_to_in_use_list(rx_ring, lbq_desc);
7178 		rx_ring->lbq_prod_idx++;
7179 	}
7180 	ql_update_lbq_prod_idx(qlge, rx_ring);
7181 }
7182 
7183 
7184 /*
7185  * Initializes the adapter by configuring request and response queues,
7186  * allocates and ARMs small and large receive buffers to the
7187  * hardware
7188  */
7189 static int
7190 ql_bringup_adapter(qlge_t *qlge)
7191 {
7192 	int i;
7193 
7194 	if (ql_device_initialize(qlge) != DDI_SUCCESS) {
7195 		cmn_err(CE_WARN, "?%s(%d): ql_device_initialize failed",
7196 		    __func__, qlge->instance);
7197 		goto err_bringup;
7198 	}
7199 	qlge->sequence |= INIT_ADAPTER_UP;
7200 
7201 #ifdef QLGE_TRACK_BUFFER_USAGE
7202 	for (i = 0; i < qlge->rx_ring_count; i++) {
7203 		if (qlge->rx_ring[i].type != TX_Q) {
7204 			qlge->rx_sb_low_count[i] = NUM_SMALL_BUFFERS;
7205 			qlge->rx_lb_low_count[i] = NUM_LARGE_BUFFERS;
7206 		}
7207 		qlge->cq_low_count[i] = NUM_RX_RING_ENTRIES;
7208 	}
7209 #endif
7210 	/* Arm buffers */
7211 	for (i = 0; i < qlge->rx_ring_count; i++) {
7212 		if (qlge->rx_ring[i].type != TX_Q) {
7213 			ql_arm_sbuf(qlge, &qlge->rx_ring[i]);
7214 			ql_arm_lbuf(qlge, &qlge->rx_ring[i]);
7215 		}
7216 	}
7217 
7218 	/* Enable work/request queues */
7219 	for (i = 0; i < qlge->tx_ring_count; i++) {
7220 		if (qlge->tx_ring[i].valid_db_reg)
7221 			ql_write_doorbell_reg(qlge,
7222 			    qlge->tx_ring[i].valid_db_reg,
7223 			    REQ_Q_VALID);
7224 	}
7225 
7226 	/* Enable completion queues */
7227 	for (i = 0; i < qlge->rx_ring_count; i++) {
7228 		if (qlge->rx_ring[i].valid_db_reg)
7229 			ql_write_doorbell_reg(qlge,
7230 			    qlge->rx_ring[i].valid_db_reg,
7231 			    RSP_Q_VALID);
7232 	}
7233 
7234 	for (i = 0; i < qlge->tx_ring_count; i++) {
7235 		mutex_enter(&qlge->tx_ring[i].tx_lock);
7236 		qlge->tx_ring[i].mac_flags = QL_MAC_STARTED;
7237 		mutex_exit(&qlge->tx_ring[i].tx_lock);
7238 	}
7239 
7240 	for (i = 0; i < qlge->rx_ring_count; i++) {
7241 		mutex_enter(&qlge->rx_ring[i].rx_lock);
7242 		qlge->rx_ring[i].mac_flags = QL_MAC_STARTED;
7243 		mutex_exit(&qlge->rx_ring[i].rx_lock);
7244 	}
7245 
7246 	/* This mutex will get re-acquired in enable_completion interrupt */
7247 	mutex_exit(&qlge->hw_mutex);
7248 	/* Traffic can start flowing now */
7249 	ql_enable_all_completion_interrupts(qlge);
7250 	mutex_enter(&qlge->hw_mutex);
7251 
7252 	ql_enable_global_interrupt(qlge);
7253 
7254 	qlge->sequence |= ADAPTER_INIT;
7255 	return (DDI_SUCCESS);
7256 
7257 err_bringup:
7258 	(void) ql_asic_reset(qlge);
7259 	return (DDI_FAILURE);
7260 }
7261 
7262 /*
7263  * Initialize mutexes of each rx/tx rings
7264  */
7265 static int
7266 ql_init_rx_tx_locks(qlge_t *qlge)
7267 {
7268 	struct tx_ring *tx_ring;
7269 	struct rx_ring *rx_ring;
7270 	int i;
7271 
7272 	for (i = 0; i < qlge->tx_ring_count; i++) {
7273 		tx_ring = &qlge->tx_ring[i];
7274 		mutex_init(&tx_ring->tx_lock, NULL, MUTEX_DRIVER,
7275 		    DDI_INTR_PRI(qlge->intr_pri));
7276 	}
7277 
7278 	for (i = 0; i < qlge->rx_ring_count; i++) {
7279 		rx_ring = &qlge->rx_ring[i];
7280 		mutex_init(&rx_ring->rx_lock, NULL, MUTEX_DRIVER,
7281 		    DDI_INTR_PRI(qlge->intr_pri));
7282 		mutex_init(&rx_ring->sbq_lock, NULL, MUTEX_DRIVER,
7283 		    DDI_INTR_PRI(qlge->intr_pri));
7284 		mutex_init(&rx_ring->lbq_lock, NULL, MUTEX_DRIVER,
7285 		    DDI_INTR_PRI(qlge->intr_pri));
7286 	}
7287 
7288 	return (DDI_SUCCESS);
7289 }
7290 
7291 /*ARGSUSED*/
7292 /*
7293  * Simply call pci_ereport_post which generates ereports for errors
7294  * that occur in the PCI local bus configuration status registers.
7295  */
7296 static int
7297 ql_fm_error_cb(dev_info_t *dip, ddi_fm_error_t *err, const void *impl_data)
7298 {
7299 	pci_ereport_post(dip, err, NULL);
7300 	return (err->fme_status);
7301 }
7302 
7303 static void
7304 ql_fm_init(qlge_t *qlge)
7305 {
7306 	ddi_iblock_cookie_t iblk;
7307 
7308 	QL_PRINT(DBG_INIT, ("ql_fm_init(%d) entered, FMA capability %x\n",
7309 	    qlge->instance, qlge->fm_capabilities));
7310 	/*
7311 	 * Register capabilities with IO Fault Services. The capabilities
7312 	 * set above may not be supported by the parent nexus, in that case
7313 	 * some capability bits may be cleared.
7314 	 */
7315 	if (qlge->fm_capabilities)
7316 		ddi_fm_init(qlge->dip, &qlge->fm_capabilities, &iblk);
7317 
7318 	/*
7319 	 * Initialize pci ereport capabilities if ereport capable
7320 	 */
7321 	if (DDI_FM_EREPORT_CAP(qlge->fm_capabilities) ||
7322 	    DDI_FM_ERRCB_CAP(qlge->fm_capabilities)) {
7323 		pci_ereport_setup(qlge->dip);
7324 	}
7325 
7326 	/* Register error callback if error callback capable */
7327 	if (DDI_FM_ERRCB_CAP(qlge->fm_capabilities)) {
7328 		ddi_fm_handler_register(qlge->dip,
7329 		    ql_fm_error_cb, (void*) qlge);
7330 	}
7331 
7332 	/*
7333 	 * DDI_FLGERR_ACC indicates:
7334 	 *  Driver will check its access handle(s) for faults on
7335 	 *   a regular basis by calling ddi_fm_acc_err_get
7336 	 *  Driver is able to cope with incorrect results of I/O
7337 	 *   operations resulted from an I/O fault
7338 	 */
7339 	if (DDI_FM_ACC_ERR_CAP(qlge->fm_capabilities)) {
7340 		ql_dev_acc_attr.devacc_attr_access = DDI_FLAGERR_ACC;
7341 	}
7342 
7343 	/*
7344 	 * DDI_DMA_FLAGERR indicates:
7345 	 *  Driver will check its DMA handle(s) for faults on a
7346 	 *   regular basis using ddi_fm_dma_err_get
7347 	 *  Driver is able to cope with incorrect results of DMA
7348 	 *   operations resulted from an I/O fault
7349 	 */
7350 	if (DDI_FM_DMA_ERR_CAP(qlge->fm_capabilities)) {
7351 		tx_mapping_dma_attr.dma_attr_flags = DDI_DMA_FLAGERR;
7352 		dma_attr.dma_attr_flags = DDI_DMA_FLAGERR;
7353 	}
7354 	QL_PRINT(DBG_INIT, ("ql_fm_init(%d) done\n",
7355 	    qlge->instance));
7356 }
7357 
7358 static void
7359 ql_fm_fini(qlge_t *qlge)
7360 {
7361 	QL_PRINT(DBG_INIT, ("ql_fm_fini(%d) entered\n",
7362 	    qlge->instance));
7363 	/* Only unregister FMA capabilities if we registered some */
7364 	if (qlge->fm_capabilities) {
7365 
7366 		/*
7367 		 * Release any resources allocated by pci_ereport_setup()
7368 		 */
7369 		if (DDI_FM_EREPORT_CAP(qlge->fm_capabilities) ||
7370 		    DDI_FM_ERRCB_CAP(qlge->fm_capabilities))
7371 			pci_ereport_teardown(qlge->dip);
7372 
7373 		/*
7374 		 * Un-register error callback if error callback capable
7375 		 */
7376 		if (DDI_FM_ERRCB_CAP(qlge->fm_capabilities))
7377 			ddi_fm_handler_unregister(qlge->dip);
7378 
7379 		/* Unregister from IO Fault Services */
7380 		ddi_fm_fini(qlge->dip);
7381 	}
7382 	QL_PRINT(DBG_INIT, ("ql_fm_fini(%d) done\n",
7383 	    qlge->instance));
7384 }
7385 /*
7386  * ql_attach - Driver attach.
7387  */
7388 static int
7389 ql_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
7390 {
7391 	int instance;
7392 	qlge_t *qlge = NULL;
7393 	int rval;
7394 	uint16_t w;
7395 	mac_register_t *macp = NULL;
7396 	uint32_t data;
7397 
7398 	rval = DDI_FAILURE;
7399 
7400 	/* first get the instance */
7401 	instance = ddi_get_instance(dip);
7402 
7403 	switch (cmd) {
7404 	case DDI_ATTACH:
7405 		/*
7406 		 * Allocate our per-device-instance structure
7407 		 */
7408 		qlge = (qlge_t *)kmem_zalloc(sizeof (*qlge), KM_SLEEP);
7409 		ASSERT(qlge != NULL);
7410 		qlge->sequence |= INIT_SOFTSTATE_ALLOC;
7411 
7412 		qlge->dip = dip;
7413 		qlge->instance = instance;
7414 		/* Set up the coalescing parameters. */
7415 		qlge->ql_dbgprnt = 0;
7416 #if QL_DEBUG
7417 		qlge->ql_dbgprnt = QL_DEBUG;
7418 #endif /* QL_DEBUG */
7419 
7420 		/*
7421 		 * Initialize for fma support
7422 		 */
7423 		/* fault management (fm) capabilities. */
7424 		qlge->fm_capabilities =
7425 		    DDI_FM_EREPORT_CAPABLE | DDI_FM_ERRCB_CAPABLE;
7426 		data = ql_get_prop(qlge, "fm-capable");
7427 		if (data <= 0xf) {
7428 			qlge->fm_capabilities = data;
7429 		}
7430 		ql_fm_init(qlge);
7431 		qlge->sequence |= INIT_FM;
7432 		QL_PRINT(DBG_INIT, ("ql_attach(%d): fma init done\n",
7433 		    qlge->instance));
7434 
7435 		/*
7436 		 * Setup the ISP8x00 registers address mapping to be
7437 		 * accessed by this particular driver.
7438 		 * 0x0   Configuration Space
7439 		 * 0x1   I/O Space
7440 		 * 0x2   1st Memory Space address - Control Register Set
7441 		 * 0x3   2nd Memory Space address - Doorbell Memory Space
7442 		 */
7443 		w = 2;
7444 		if (ddi_regs_map_setup(dip, w, (caddr_t *)&qlge->iobase, 0,
7445 		    sizeof (dev_reg_t), &ql_dev_acc_attr,
7446 		    &qlge->dev_handle) != DDI_SUCCESS) {
7447 			cmn_err(CE_WARN, "%s(%d): Unable to map device "
7448 			    "registers", ADAPTER_NAME, instance);
7449 			break;
7450 		}
7451 		QL_PRINT(DBG_GLD, ("ql_attach: I/O base = 0x%x\n",
7452 		    qlge->iobase));
7453 		qlge->sequence |= INIT_REGS_SETUP;
7454 
7455 		/* map Doorbell memory space */
7456 		w = 3;
7457 		if (ddi_regs_map_setup(dip, w,
7458 		    (caddr_t *)&qlge->doorbell_reg_iobase, 0,
7459 		    0x100000 /* sizeof (dev_doorbell_reg_t) */,
7460 		    &ql_dev_acc_attr,
7461 		    &qlge->dev_doorbell_reg_handle) != DDI_SUCCESS) {
7462 			cmn_err(CE_WARN, "%s(%d): Unable to map Doorbell "
7463 			    "registers",
7464 			    ADAPTER_NAME, instance);
7465 			break;
7466 		}
7467 		QL_PRINT(DBG_GLD, ("ql_attach: Doorbell I/O base = 0x%x\n",
7468 		    qlge->doorbell_reg_iobase));
7469 		qlge->sequence |= INIT_DOORBELL_REGS_SETUP;
7470 
7471 		/*
7472 		 * Allocate a macinfo structure for this instance
7473 		 */
7474 		if ((macp = mac_alloc(MAC_VERSION)) == NULL) {
7475 			cmn_err(CE_WARN, "%s(%d): mac_alloc failed",
7476 			    __func__, instance);
7477 			break;
7478 		}
7479 		/* save adapter status to dip private data */
7480 		ddi_set_driver_private(dip, qlge);
7481 		QL_PRINT(DBG_INIT, ("%s(%d): Allocate macinfo structure done\n",
7482 		    ADAPTER_NAME, instance));
7483 		qlge->sequence |= INIT_MAC_ALLOC;
7484 
7485 		/*
7486 		 * Attach this instance of the device
7487 		 */
7488 		/* Setup PCI Local Bus Configuration resource. */
7489 		if (pci_config_setup(dip, &qlge->pci_handle) != DDI_SUCCESS) {
7490 			cmn_err(CE_WARN, "%s(%d):Unable to get PCI resources",
7491 			    ADAPTER_NAME, instance);
7492 			if (qlge->fm_enable) {
7493 				ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
7494 				ddi_fm_service_impact(qlge->dip,
7495 				    DDI_SERVICE_LOST);
7496 			}
7497 			break;
7498 		}
7499 		qlge->sequence |= INIT_PCI_CONFIG_SETUP;
7500 		QL_PRINT(DBG_GLD, ("ql_attach(%d): pci_config_setup done\n",
7501 		    instance));
7502 
7503 		if (ql_init_instance(qlge) != DDI_SUCCESS) {
7504 			cmn_err(CE_WARN, "%s(%d): Unable to initialize device "
7505 			    "instance", ADAPTER_NAME, instance);
7506 			if (qlge->fm_enable) {
7507 				ql_fm_ereport(qlge, DDI_FM_DEVICE_INVAL_STATE);
7508 				ddi_fm_service_impact(qlge->dip,
7509 				    DDI_SERVICE_LOST);
7510 			}
7511 			break;
7512 		}
7513 		QL_PRINT(DBG_GLD, ("ql_attach(%d): ql_init_instance done\n",
7514 		    instance));
7515 
7516 		/* Setup interrupt vectors */
7517 		if (ql_alloc_irqs(qlge) != DDI_SUCCESS) {
7518 			break;
7519 		}
7520 		qlge->sequence |= INIT_INTR_ALLOC;
7521 		QL_PRINT(DBG_GLD, ("ql_attach(%d): ql_alloc_irqs done\n",
7522 		    instance));
7523 
7524 		/* Configure queues */
7525 		if (ql_setup_rings(qlge) != DDI_SUCCESS) {
7526 			break;
7527 		}
7528 		qlge->sequence |= INIT_SETUP_RINGS;
7529 		QL_PRINT(DBG_GLD, ("ql_attach(%d): setup rings done\n",
7530 		    instance));
7531 
7532 		/*
7533 		 * Allocate memory resources
7534 		 */
7535 		if (ql_alloc_mem_resources(qlge) != DDI_SUCCESS) {
7536 			cmn_err(CE_WARN, "%s(%d): memory allocation failed",
7537 			    __func__, qlge->instance);
7538 			break;
7539 		}
7540 		qlge->sequence |= INIT_MEMORY_ALLOC;
7541 		QL_PRINT(DBG_GLD, ("ql_alloc_mem_resources(%d) done\n",
7542 		    instance));
7543 
7544 		/*
7545 		 * Map queues to interrupt vectors
7546 		 */
7547 		ql_resolve_queues_to_irqs(qlge);
7548 
7549 		/* Initialize mutex, need the interrupt priority */
7550 		(void) ql_init_rx_tx_locks(qlge);
7551 		qlge->sequence |= INIT_LOCKS_CREATED;
7552 		QL_PRINT(DBG_INIT, ("%s(%d): ql_init_rx_tx_locks done\n",
7553 		    ADAPTER_NAME, instance));
7554 
7555 		/*
7556 		 * Use a soft interrupt to do something that we do not want
7557 		 * to do in regular network functions or with mutexs being held
7558 		 */
7559 		if (ddi_intr_add_softint(qlge->dip, &qlge->mpi_event_intr_hdl,
7560 		    DDI_INTR_SOFTPRI_MIN, ql_mpi_event_work, (caddr_t)qlge)
7561 		    != DDI_SUCCESS) {
7562 			break;
7563 		}
7564 
7565 		if (ddi_intr_add_softint(qlge->dip, &qlge->asic_reset_intr_hdl,
7566 		    DDI_INTR_SOFTPRI_MIN, ql_asic_reset_work, (caddr_t)qlge)
7567 		    != DDI_SUCCESS) {
7568 			break;
7569 		}
7570 
7571 		if (ddi_intr_add_softint(qlge->dip, &qlge->mpi_reset_intr_hdl,
7572 		    DDI_INTR_SOFTPRI_MIN, ql_mpi_reset_work, (caddr_t)qlge)
7573 		    != DDI_SUCCESS) {
7574 			break;
7575 		}
7576 		qlge->sequence |= INIT_ADD_SOFT_INTERRUPT;
7577 		QL_PRINT(DBG_INIT, ("%s(%d): ddi_intr_add_softint done\n",
7578 		    ADAPTER_NAME, instance));
7579 
7580 		/*
7581 		 * mutex to protect the adapter state structure.
7582 		 * initialize mutexes according to the interrupt priority
7583 		 */
7584 		mutex_init(&qlge->gen_mutex, NULL, MUTEX_DRIVER,
7585 		    DDI_INTR_PRI(qlge->intr_pri));
7586 		mutex_init(&qlge->hw_mutex, NULL, MUTEX_DRIVER,
7587 		    DDI_INTR_PRI(qlge->intr_pri));
7588 		mutex_init(&qlge->mbx_mutex, NULL, MUTEX_DRIVER,
7589 		    DDI_INTR_PRI(qlge->intr_pri));
7590 
7591 		/* Mailbox wait and interrupt conditional variable. */
7592 		cv_init(&qlge->cv_mbx_intr, NULL, CV_DRIVER, NULL);
7593 		qlge->sequence |= INIT_MUTEX;
7594 		QL_PRINT(DBG_INIT, ("%s(%d): mutex_init done\n",
7595 		    ADAPTER_NAME, instance));
7596 
7597 		/*
7598 		 * KStats
7599 		 */
7600 		if (ql_init_kstats(qlge) != DDI_SUCCESS) {
7601 			cmn_err(CE_WARN, "%s(%d): KState initialization failed",
7602 			    ADAPTER_NAME, instance);
7603 			break;
7604 		}
7605 		qlge->sequence |= INIT_KSTATS;
7606 		QL_PRINT(DBG_INIT, ("%s(%d): ql_init_kstats done\n",
7607 		    ADAPTER_NAME, instance));
7608 
7609 		/*
7610 		 * Initialize gld macinfo structure
7611 		 */
7612 		ql_gld3_init(qlge, macp);
7613 		/*
7614 		 * Add interrupt handlers
7615 		 */
7616 		if (ql_add_intr_handlers(qlge) != DDI_SUCCESS) {
7617 			cmn_err(CE_WARN, "Failed to add interrupt "
7618 			    "handlers");
7619 			break;
7620 		}
7621 		qlge->sequence |= INIT_ADD_INTERRUPT;
7622 		QL_PRINT(DBG_INIT, ("%s(%d): Add interrupt handler done\n",
7623 		    ADAPTER_NAME, instance));
7624 
7625 		/*
7626 		 * MAC Register
7627 		 */
7628 		if (mac_register(macp, &qlge->mh) != DDI_SUCCESS) {
7629 			cmn_err(CE_WARN, "%s(%d): mac_register failed",
7630 			    __func__, instance);
7631 			break;
7632 		}
7633 		qlge->sequence |= INIT_MAC_REGISTERED;
7634 		QL_PRINT(DBG_GLD, ("%s(%d): mac_register done\n",
7635 		    ADAPTER_NAME, instance));
7636 
7637 		mac_free(macp);
7638 		macp = NULL;
7639 
7640 		qlge->mac_flags = QL_MAC_ATTACHED;
7641 
7642 		ddi_report_dev(dip);
7643 
7644 		rval = DDI_SUCCESS;
7645 
7646 	break;
7647 /*
7648  * DDI_RESUME
7649  * When called  with  cmd  set  to  DDI_RESUME,  attach()  must
7650  * restore  the hardware state of a device (power may have been
7651  * removed from the device), allow  pending  requests  to  con-
7652  * tinue,  and  service  new requests. In this case, the driver
7653  * must not  make  any  assumptions  about  the  state  of  the
7654  * hardware,  but  must  restore the state of the device except
7655  * for the power level of components.
7656  *
7657  */
7658 	case DDI_RESUME:
7659 
7660 		if ((qlge = (qlge_t *)QL_GET_DEV(dip)) == NULL)
7661 			return (DDI_FAILURE);
7662 
7663 		QL_PRINT(DBG_GLD, ("%s(%d)-DDI_RESUME\n",
7664 		    __func__, qlge->instance));
7665 
7666 		mutex_enter(&qlge->gen_mutex);
7667 		rval = ql_do_start(qlge);
7668 		mutex_exit(&qlge->gen_mutex);
7669 		break;
7670 
7671 	default:
7672 		break;
7673 	}
7674 
7675 	/* if failed to attach */
7676 	if ((cmd == DDI_ATTACH) && (rval != DDI_SUCCESS) && (qlge != NULL)) {
7677 		cmn_err(CE_WARN, "qlge driver attach failed, sequence %x",
7678 		    qlge->sequence);
7679 		ql_free_resources(qlge);
7680 	}
7681 
7682 	return (rval);
7683 }
7684 
7685 /*
7686  * Unbind all pending tx dma handles during driver bring down
7687  */
7688 static void
7689 ql_unbind_pending_tx_dma_handle(struct tx_ring *tx_ring)
7690 {
7691 	struct tx_ring_desc *tx_ring_desc;
7692 	int i, j;
7693 
7694 	if (tx_ring->wq_desc) {
7695 		tx_ring_desc = tx_ring->wq_desc;
7696 		for (i = 0; i < tx_ring->wq_len; i++, tx_ring_desc++) {
7697 			for (j = 0; j < tx_ring_desc->tx_dma_handle_used; j++) {
7698 				if (tx_ring_desc->tx_dma_handle[j]) {
7699 					(void) ddi_dma_unbind_handle(
7700 					    tx_ring_desc->tx_dma_handle[j]);
7701 				}
7702 			}
7703 			tx_ring_desc->tx_dma_handle_used = 0;
7704 		} /* end of for loop */
7705 	}
7706 }
7707 /*
7708  * Wait for all the packets sent to the chip to finish transmission
7709  * to prevent buffers to be unmapped before or during a transmit operation
7710  */
7711 static int
7712 ql_wait_tx_quiesce(qlge_t *qlge)
7713 {
7714 	int count = MAX_TX_WAIT_COUNT, i;
7715 	int rings_done;
7716 	volatile struct tx_ring *tx_ring;
7717 	uint32_t consumer_idx;
7718 	uint32_t producer_idx;
7719 	uint32_t temp;
7720 	int done = 0;
7721 	int rval = DDI_FAILURE;
7722 
7723 	while (!done) {
7724 		rings_done = 0;
7725 
7726 		for (i = 0; i < qlge->tx_ring_count; i++) {
7727 			tx_ring = &qlge->tx_ring[i];
7728 			temp = ql_read_doorbell_reg(qlge,
7729 			    tx_ring->prod_idx_db_reg);
7730 			producer_idx = temp & 0x0000ffff;
7731 			consumer_idx = (temp >> 16);
7732 
7733 			if (qlge->isr_stride) {
7734 				struct rx_ring *ob_ring;
7735 				ob_ring = &qlge->rx_ring[tx_ring->cq_id];
7736 				if (producer_idx != ob_ring->cnsmr_idx) {
7737 					cmn_err(CE_NOTE, " force clean \n");
7738 					(void) ql_clean_outbound_rx_ring(
7739 					    ob_ring);
7740 				}
7741 			}
7742 			/*
7743 			 * Get the pending iocb count, ones which have not been
7744 			 * pulled down by the chip
7745 			 */
7746 			if (producer_idx >= consumer_idx)
7747 				temp = (producer_idx - consumer_idx);
7748 			else
7749 				temp = (tx_ring->wq_len - consumer_idx) +
7750 				    producer_idx;
7751 
7752 			if ((tx_ring->tx_free_count + temp) >= tx_ring->wq_len)
7753 				rings_done++;
7754 			else {
7755 				done = 1;
7756 				break;
7757 			}
7758 		}
7759 
7760 		/* If all the rings are done */
7761 		if (rings_done >= qlge->tx_ring_count) {
7762 #ifdef QLGE_LOAD_UNLOAD
7763 			cmn_err(CE_NOTE, "%s(%d) done successfully \n",
7764 			    __func__, qlge->instance);
7765 #endif
7766 			rval = DDI_SUCCESS;
7767 			break;
7768 		}
7769 
7770 		qlge_delay(100);
7771 
7772 		count--;
7773 		if (!count) {
7774 
7775 			count = MAX_TX_WAIT_COUNT;
7776 #ifdef QLGE_LOAD_UNLOAD
7777 			volatile struct rx_ring *rx_ring;
7778 			cmn_err(CE_NOTE, "%s(%d): Waiting for %d pending"
7779 			    " Transmits on queue %d to complete .\n",
7780 			    __func__, qlge->instance,
7781 			    (qlge->tx_ring[i].wq_len -
7782 			    qlge->tx_ring[i].tx_free_count),
7783 			    i);
7784 
7785 			rx_ring = &qlge->rx_ring[i+1];
7786 			temp = ql_read_doorbell_reg(qlge,
7787 			    rx_ring->cnsmr_idx_db_reg);
7788 			consumer_idx = temp & 0x0000ffff;
7789 			producer_idx = (temp >> 16);
7790 			cmn_err(CE_NOTE, "%s(%d): Transmit completion queue %d,"
7791 			    " Producer %d, Consumer %d\n",
7792 			    __func__, qlge->instance,
7793 			    i+1,
7794 			    producer_idx, consumer_idx);
7795 
7796 			temp = ql_read_doorbell_reg(qlge,
7797 			    tx_ring->prod_idx_db_reg);
7798 			producer_idx = temp & 0x0000ffff;
7799 			consumer_idx = (temp >> 16);
7800 			cmn_err(CE_NOTE, "%s(%d): Transmit request queue %d,"
7801 			    " Producer %d, Consumer %d\n",
7802 			    __func__, qlge->instance, i,
7803 			    producer_idx, consumer_idx);
7804 #endif
7805 
7806 			/* For now move on */
7807 			break;
7808 		}
7809 	}
7810 	/* Stop the request queue */
7811 	mutex_enter(&qlge->hw_mutex);
7812 	for (i = 0; i < qlge->tx_ring_count; i++) {
7813 		if (qlge->tx_ring[i].valid_db_reg) {
7814 			ql_write_doorbell_reg(qlge,
7815 			    qlge->tx_ring[i].valid_db_reg, 0);
7816 		}
7817 	}
7818 	mutex_exit(&qlge->hw_mutex);
7819 	return (rval);
7820 }
7821 
7822 /*
7823  * Wait for all the receives indicated to the stack to come back
7824  */
7825 static int
7826 ql_wait_rx_complete(qlge_t *qlge)
7827 {
7828 	int i;
7829 	/* Disable all the completion queues */
7830 	mutex_enter(&qlge->hw_mutex);
7831 	for (i = 0; i < qlge->rx_ring_count; i++) {
7832 		if (qlge->rx_ring[i].valid_db_reg) {
7833 			ql_write_doorbell_reg(qlge,
7834 			    qlge->rx_ring[i].valid_db_reg, 0);
7835 		}
7836 	}
7837 	mutex_exit(&qlge->hw_mutex);
7838 
7839 	/* Wait for OS to return all rx buffers */
7840 	qlge_delay(QL_ONE_SEC_DELAY);
7841 	return (DDI_SUCCESS);
7842 }
7843 
7844 /*
7845  * stop the driver
7846  */
7847 static int
7848 ql_bringdown_adapter(qlge_t *qlge)
7849 {
7850 	int i;
7851 	int status = DDI_SUCCESS;
7852 
7853 	qlge->mac_flags = QL_MAC_BRINGDOWN;
7854 	if (qlge->sequence & ADAPTER_INIT) {
7855 		/* stop forwarding external packets to driver */
7856 		status = ql_sem_spinlock(qlge, SEM_RT_IDX_MASK);
7857 		if (status)
7858 			return (status);
7859 		(void) ql_stop_routing(qlge);
7860 		ql_sem_unlock(qlge, SEM_RT_IDX_MASK);
7861 		/*
7862 		 * Set the flag for receive and transmit
7863 		 * operations to cease
7864 		 */
7865 		for (i = 0; i < qlge->tx_ring_count; i++) {
7866 			mutex_enter(&qlge->tx_ring[i].tx_lock);
7867 			qlge->tx_ring[i].mac_flags = QL_MAC_STOPPED;
7868 			mutex_exit(&qlge->tx_ring[i].tx_lock);
7869 		}
7870 
7871 		for (i = 0; i < qlge->rx_ring_count; i++) {
7872 			mutex_enter(&qlge->rx_ring[i].rx_lock);
7873 			qlge->rx_ring[i].mac_flags = QL_MAC_STOPPED;
7874 			mutex_exit(&qlge->rx_ring[i].rx_lock);
7875 		}
7876 
7877 		/*
7878 		 * Need interrupts to be running while the transmit
7879 		 * completions are cleared. Wait for the packets
7880 		 * queued to the chip to be sent out
7881 		 */
7882 		(void) ql_wait_tx_quiesce(qlge);
7883 		/* Interrupts not needed from now */
7884 		ql_disable_all_completion_interrupts(qlge);
7885 
7886 		mutex_enter(&qlge->hw_mutex);
7887 		/* Disable Global interrupt */
7888 		ql_disable_global_interrupt(qlge);
7889 		mutex_exit(&qlge->hw_mutex);
7890 
7891 		/* Wait for all the indicated packets to come back */
7892 		status = ql_wait_rx_complete(qlge);
7893 
7894 		mutex_enter(&qlge->hw_mutex);
7895 		/* Reset adapter */
7896 		(void) ql_asic_reset(qlge);
7897 		/*
7898 		 * Unbind all tx dma handles to prevent pending tx descriptors'
7899 		 * dma handles from being re-used.
7900 		 */
7901 		for (i = 0; i < qlge->tx_ring_count; i++) {
7902 			ql_unbind_pending_tx_dma_handle(&qlge->tx_ring[i]);
7903 		}
7904 
7905 		qlge->sequence &= ~ADAPTER_INIT;
7906 
7907 		mutex_exit(&qlge->hw_mutex);
7908 	}
7909 	return (status);
7910 }
7911 
7912 /*
7913  * ql_detach
7914  * Used to remove all the states associated with a given
7915  * instances of a device node prior to the removal of that
7916  * instance from the system.
7917  */
7918 static int
7919 ql_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
7920 {
7921 	qlge_t *qlge;
7922 	int rval;
7923 
7924 	rval = DDI_SUCCESS;
7925 
7926 	switch (cmd) {
7927 	case DDI_DETACH:
7928 
7929 		if ((qlge = QL_GET_DEV(dip)) == NULL)
7930 			return (DDI_FAILURE);
7931 		rval = ql_bringdown_adapter(qlge);
7932 		if (rval != DDI_SUCCESS)
7933 			break;
7934 
7935 		qlge->mac_flags = QL_MAC_DETACH;
7936 
7937 		/* free memory resources */
7938 		if (qlge->sequence & INIT_MEMORY_ALLOC) {
7939 			ql_free_mem_resources(qlge);
7940 			qlge->sequence &= ~INIT_MEMORY_ALLOC;
7941 		}
7942 		ql_free_resources(qlge);
7943 
7944 		break;
7945 
7946 	case DDI_SUSPEND:
7947 		if ((qlge = QL_GET_DEV(dip)) == NULL)
7948 			return (DDI_FAILURE);
7949 
7950 		mutex_enter(&qlge->gen_mutex);
7951 		if ((qlge->mac_flags == QL_MAC_ATTACHED) ||
7952 		    (qlge->mac_flags == QL_MAC_STARTED)) {
7953 			(void) ql_do_stop(qlge);
7954 		}
7955 		qlge->mac_flags = QL_MAC_SUSPENDED;
7956 		mutex_exit(&qlge->gen_mutex);
7957 
7958 		break;
7959 	default:
7960 		rval = DDI_FAILURE;
7961 		break;
7962 	}
7963 
7964 	return (rval);
7965 }
7966 
7967 /*
7968  * quiesce(9E) entry point.
7969  *
7970  * This function is called when the system is single-threaded at high
7971  * PIL with preemption disabled. Therefore, this function must not be
7972  * blocked.
7973  *
7974  * This function returns DDI_SUCCESS on success, or DDI_FAILURE on failure.
7975  */
7976 int
7977 ql_quiesce(dev_info_t *dip)
7978 {
7979 	qlge_t *qlge;
7980 	int i;
7981 
7982 	if ((qlge = QL_GET_DEV(dip)) == NULL)
7983 		return (DDI_FAILURE);
7984 
7985 	if (CFG_IST(qlge, CFG_CHIP_8100)) {
7986 		/* stop forwarding external packets to driver */
7987 		(void) ql_sem_spinlock(qlge, SEM_RT_IDX_MASK);
7988 		(void) ql_stop_routing(qlge);
7989 		ql_sem_unlock(qlge, SEM_RT_IDX_MASK);
7990 		/* Stop all the request queues */
7991 		for (i = 0; i < qlge->tx_ring_count; i++) {
7992 			if (qlge->tx_ring[i].valid_db_reg) {
7993 				ql_write_doorbell_reg(qlge,
7994 				    qlge->tx_ring[i].valid_db_reg, 0);
7995 			}
7996 		}
7997 		qlge_delay(QL_ONE_SEC_DELAY/4);
7998 		/* Interrupts not needed from now */
7999 		/* Disable MPI interrupt */
8000 		ql_write_reg(qlge, REG_INTERRUPT_MASK,
8001 		    (INTR_MASK_PI << 16));
8002 		ql_disable_global_interrupt(qlge);
8003 
8004 		/* Disable all the rx completion queues */
8005 		for (i = 0; i < qlge->rx_ring_count; i++) {
8006 			if (qlge->rx_ring[i].valid_db_reg) {
8007 				ql_write_doorbell_reg(qlge,
8008 				    qlge->rx_ring[i].valid_db_reg, 0);
8009 			}
8010 		}
8011 		qlge_delay(QL_ONE_SEC_DELAY/4);
8012 		qlge->mac_flags = QL_MAC_STOPPED;
8013 		/* Reset adapter */
8014 		(void) ql_asic_reset(qlge);
8015 		qlge_delay(100);
8016 	}
8017 
8018 	return (DDI_SUCCESS);
8019 }
8020 
8021 QL_STREAM_OPS(ql_ops, ql_attach, ql_detach);
8022 
8023 /*
8024  * Loadable Driver Interface Structures.
8025  * Declare and initialize the module configuration section...
8026  */
8027 static struct modldrv modldrv = {
8028 	&mod_driverops,		/* type of module: driver */
8029 	version,		/* name of module */
8030 	&ql_ops			/* driver dev_ops */
8031 };
8032 
8033 static struct modlinkage modlinkage = {
8034 	MODREV_1,	&modldrv,	NULL
8035 };
8036 
8037 /*
8038  * Loadable Module Routines
8039  */
8040 
8041 /*
8042  * _init
8043  * Initializes a loadable module. It is called before any other
8044  * routine in a loadable module.
8045  */
8046 int
8047 _init(void)
8048 {
8049 	int rval;
8050 
8051 	mac_init_ops(&ql_ops, ADAPTER_NAME);
8052 	rval = mod_install(&modlinkage);
8053 	if (rval != DDI_SUCCESS) {
8054 		mac_fini_ops(&ql_ops);
8055 		cmn_err(CE_WARN, "?Unable to install/attach driver '%s'",
8056 		    ADAPTER_NAME);
8057 	}
8058 
8059 	return (rval);
8060 }
8061 
8062 /*
8063  * _fini
8064  * Prepares a module for unloading. It is called when the system
8065  * wants to unload a module. If the module determines that it can
8066  * be unloaded, then _fini() returns the value returned by
8067  * mod_remove(). Upon successful return from _fini() no other
8068  * routine in the module will be called before _init() is called.
8069  */
8070 int
8071 _fini(void)
8072 {
8073 	int rval;
8074 
8075 	rval = mod_remove(&modlinkage);
8076 	if (rval == DDI_SUCCESS) {
8077 		mac_fini_ops(&ql_ops);
8078 	}
8079 
8080 	return (rval);
8081 }
8082 
8083 /*
8084  * _info
8085  * Returns information about loadable module.
8086  */
8087 int
8088 _info(struct modinfo *modinfop)
8089 {
8090 	return (mod_info(&modlinkage, modinfop));
8091 }
8092