xref: /freebsd/sys/dev/qlxgbe/ql_hw.c (revision 04803ab2177f1c57467b12f497a7ab8eabc3e8d2)
1 /*
2  * Copyright (c) 2013-2014 Qlogic Corporation
3  * All rights reserved.
4  *
5  *  Redistribution and use in source and binary forms, with or without
6  *  modification, are permitted provided that the following conditions
7  *  are met:
8  *
9  *  1. Redistributions of source code must retain the above copyright
10  *     notice, this list of conditions and the following disclaimer.
11  *  2. Redistributions in binary form must reproduce the above copyright
12  *     notice, this list of conditions and the following disclaimer in the
13  *     documentation and/or other materials provided with the distribution.
14  *
15  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16  *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19  *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20  *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21  *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22  *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23  *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24  *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25  *  POSSIBILITY OF SUCH DAMAGE.
26  */
27 
28 /*
29  * File: ql_hw.c
30  * Author : David C Somayajulu, Qlogic Corporation, Aliso Viejo, CA 92656.
31  * Content: Contains Hardware dependant functions
32  */
33 
34 #include <sys/cdefs.h>
35 __FBSDID("$FreeBSD$");
36 
37 #include "ql_os.h"
38 #include "ql_hw.h"
39 #include "ql_def.h"
40 #include "ql_inline.h"
41 #include "ql_ver.h"
42 #include "ql_glbl.h"
43 #include "ql_dbg.h"
44 
45 /*
46  * Static Functions
47  */
48 
49 static void qla_del_rcv_cntxt(qla_host_t *ha);
50 static int qla_init_rcv_cntxt(qla_host_t *ha);
51 static void qla_del_xmt_cntxt(qla_host_t *ha);
52 static int qla_init_xmt_cntxt(qla_host_t *ha);
53 static void qla_hw_tx_done_locked(qla_host_t *ha, uint32_t txr_idx);
54 static int qla_mbx_cmd(qla_host_t *ha, uint32_t *h_mbox, uint32_t n_hmbox,
55 	uint32_t *fw_mbox, uint32_t n_fwmbox, uint32_t no_pause);
56 static int qla_config_intr_cntxt(qla_host_t *ha, uint32_t num_intrs,
57 	uint32_t create);
58 static int qla_get_nic_partition(qla_host_t *ha);
59 static int qla_config_rss(qla_host_t *ha, uint16_t cntxt_id);
60 static int qla_config_intr_coalesce(qla_host_t *ha, uint16_t cntxt_id,
61 	int tenable);
62 static int qla_set_mac_rcv_mode(qla_host_t *ha, uint32_t mode);
63 static int qla_link_event_req(qla_host_t *ha, uint16_t cntxt_id);
64 
65 static int qla_tx_tso(qla_host_t *ha, struct mbuf *mp, q80_tx_cmd_t *tx_cmd,
66 		uint8_t *hdr);
67 static int qla_hw_add_all_mcast(qla_host_t *ha);
68 static int qla_hw_del_all_mcast(qla_host_t *ha);
69 static int qla_add_rcv_rings(qla_host_t *ha, uint32_t sds_idx);
70 
71 static int qla_minidump_init(qla_host_t *ha);
72 static void qla_minidump_free(qla_host_t *ha);
73 
74 
75 static int
76 qla_sysctl_get_drvr_stats(SYSCTL_HANDLER_ARGS)
77 {
78         int err = 0, ret;
79         qla_host_t *ha;
80 	uint32_t i;
81 
82         err = sysctl_handle_int(oidp, &ret, 0, req);
83 
84         if (err || !req->newptr)
85                 return (err);
86 
87         if (ret == 1) {
88 
89                 ha = (qla_host_t *)arg1;
90 
91 		for (i = 0; i < ha->hw.num_sds_rings; i++)
92 			device_printf(ha->pci_dev,
93 				"%s: sds_ring[%d] = %p\n", __func__,i,
94 				(void *)ha->hw.sds[i].intr_count);
95 
96 		for (i = 0; i < ha->hw.num_tx_rings; i++)
97 			device_printf(ha->pci_dev,
98 				"%s: tx[%d] = %p\n", __func__,i,
99 				(void *)ha->tx_ring[i].count);
100 
101 		for (i = 0; i < ha->hw.num_rds_rings; i++)
102 			device_printf(ha->pci_dev,
103 				"%s: rds_ring[%d] = %p\n", __func__,i,
104 				(void *)ha->hw.rds[i].count);
105 
106 		device_printf(ha->pci_dev, "%s: lro_pkt_count = %p\n", __func__,
107 			(void *)ha->lro_pkt_count);
108 
109 		device_printf(ha->pci_dev, "%s: lro_bytes = %p\n", __func__,
110 			(void *)ha->lro_bytes);
111 	}
112 	return (err);
113 }
114 
115 #ifdef QL_DBG
116 
117 static void
118 qla_stop_pegs(qla_host_t *ha)
119 {
120         uint32_t val = 1;
121 
122         ql_rdwr_indreg32(ha, Q8_CRB_PEG_0, &val, 0);
123         ql_rdwr_indreg32(ha, Q8_CRB_PEG_1, &val, 0);
124         ql_rdwr_indreg32(ha, Q8_CRB_PEG_2, &val, 0);
125         ql_rdwr_indreg32(ha, Q8_CRB_PEG_3, &val, 0);
126         ql_rdwr_indreg32(ha, Q8_CRB_PEG_4, &val, 0);
127         device_printf(ha->pci_dev, "%s PEGS HALTED!!!!!\n", __func__);
128 }
129 
130 static int
131 qla_sysctl_stop_pegs(SYSCTL_HANDLER_ARGS)
132 {
133 	int err, ret = 0;
134 	qla_host_t *ha;
135 
136 	err = sysctl_handle_int(oidp, &ret, 0, req);
137 
138 
139 	if (err || !req->newptr)
140 		return (err);
141 
142 	if (ret == 1) {
143 		ha = (qla_host_t *)arg1;
144 		(void)QLA_LOCK(ha, __func__, 0);
145 		qla_stop_pegs(ha);
146 		QLA_UNLOCK(ha, __func__);
147 	}
148 
149 	return err;
150 }
151 #endif /* #ifdef QL_DBG */
152 
153 /*
154  * Name: ql_hw_add_sysctls
155  * Function: Add P3Plus specific sysctls
156  */
157 void
158 ql_hw_add_sysctls(qla_host_t *ha)
159 {
160         device_t	dev;
161 
162         dev = ha->pci_dev;
163 
164 	ha->hw.num_sds_rings = MAX_SDS_RINGS;
165 	ha->hw.num_rds_rings = MAX_RDS_RINGS;
166 	ha->hw.num_tx_rings = NUM_TX_RINGS;
167 
168 	SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
169 		SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
170 		OID_AUTO, "num_rds_rings", CTLFLAG_RD, &ha->hw.num_rds_rings,
171 		ha->hw.num_rds_rings, "Number of Rcv Descriptor Rings");
172 
173         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
174                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
175                 OID_AUTO, "num_sds_rings", CTLFLAG_RD, &ha->hw.num_sds_rings,
176 		ha->hw.num_sds_rings, "Number of Status Descriptor Rings");
177 
178         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
179                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
180                 OID_AUTO, "num_tx_rings", CTLFLAG_RD, &ha->hw.num_tx_rings,
181 		ha->hw.num_tx_rings, "Number of Transmit Rings");
182 
183         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
184                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
185                 OID_AUTO, "tx_ring_index", CTLFLAG_RW, &ha->txr_idx,
186 		ha->txr_idx, "Tx Ring Used");
187 
188 	SYSCTL_ADD_PROC(device_get_sysctl_ctx(dev),
189 		SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
190 		OID_AUTO, "drvr_stats", CTLTYPE_INT | CTLFLAG_RW,
191 		(void *)ha, 0,
192 		qla_sysctl_get_drvr_stats, "I", "Driver Maintained Statistics");
193 
194         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
195                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
196                 OID_AUTO, "max_tx_segs", CTLFLAG_RD, &ha->hw.max_tx_segs,
197 		ha->hw.max_tx_segs, "Max # of Segments in a non-TSO pkt");
198 
199 	ha->hw.sds_cidx_thres = 32;
200         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
201                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
202                 OID_AUTO, "sds_cidx_thres", CTLFLAG_RW, &ha->hw.sds_cidx_thres,
203 		ha->hw.sds_cidx_thres,
204 		"Number of SDS entries to process before updating"
205 		" SDS Ring Consumer Index");
206 
207 	ha->hw.rds_pidx_thres = 32;
208         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
209                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
210                 OID_AUTO, "rds_pidx_thres", CTLFLAG_RW, &ha->hw.rds_pidx_thres,
211 		ha->hw.rds_pidx_thres,
212 		"Number of Rcv Rings Entries to post before updating"
213 		" RDS Ring Producer Index");
214 
215 	ha->hw.min_lro_pkt_size = 512;
216 	SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
217 		SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
218 		OID_AUTO, "min_lro_pkt_size", CTLFLAG_RD, &ha->hw.min_lro_pkt_size,
219 		ha->hw.min_lro_pkt_size, "minimum packet size to trigger lro");
220 
221 	ha->hw.mdump_active = 0;
222         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
223                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
224                 OID_AUTO, "minidump_active", CTLFLAG_RW, &ha->hw.mdump_active,
225 		ha->hw.mdump_active,
226 		"Minidump Utility is Active \n"
227 		"\t 0 = Minidump Utility is not active\n"
228 		"\t 1 = Minidump Utility is retrieved on this port\n"
229 		"\t 2 = Minidump Utility is retrieved on the other port\n");
230 
231 	ha->hw.mdump_start = 0;
232         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
233                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
234                 OID_AUTO, "minidump_start", CTLFLAG_RW,
235 		&ha->hw.mdump_start, ha->hw.mdump_start,
236 		"Minidump Utility can start minidump process");
237 #ifdef QL_DBG
238 
239         SYSCTL_ADD_UINT(device_get_sysctl_ctx(dev),
240                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
241                 OID_AUTO, "err_inject",
242                 CTLFLAG_RW, &ha->err_inject, ha->err_inject,
243                 "Error to be injected\n"
244                 "\t\t\t 0: No Errors\n"
245                 "\t\t\t 1: rcv: rxb struct invalid\n"
246                 "\t\t\t 2: rcv: mp == NULL\n"
247                 "\t\t\t 3: lro: rxb struct invalid\n"
248                 "\t\t\t 4: lro: mp == NULL\n"
249                 "\t\t\t 5: rcv: num handles invalid\n"
250                 "\t\t\t 6: reg: indirect reg rd_wr failure\n"
251                 "\t\t\t 7: ocm: offchip memory rd_wr failure\n"
252                 "\t\t\t 8: mbx: mailbox command failure\n"
253                 "\t\t\t 9: heartbeat failure\n"
254                 "\t\t\t A: temperature failure\n" );
255 
256 	SYSCTL_ADD_PROC(device_get_sysctl_ctx(dev),
257                 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)),
258                 OID_AUTO, "peg_stop", CTLTYPE_INT | CTLFLAG_RW,
259                 (void *)ha, 0,
260                 qla_sysctl_stop_pegs, "I", "Peg Stop");
261 
262 #endif /* #ifdef QL_DBG */
263 
264 }
265 
266 void
267 ql_hw_link_status(qla_host_t *ha)
268 {
269 	device_printf(ha->pci_dev, "cable_oui\t\t 0x%08x\n", ha->hw.cable_oui);
270 
271 	if (ha->hw.link_up) {
272 		device_printf(ha->pci_dev, "link Up\n");
273 	} else {
274 		device_printf(ha->pci_dev, "link Down\n");
275 	}
276 
277 	if (ha->hw.flags.fduplex) {
278 		device_printf(ha->pci_dev, "Full Duplex\n");
279 	} else {
280 		device_printf(ha->pci_dev, "Half Duplex\n");
281 	}
282 
283 	if (ha->hw.flags.autoneg) {
284 		device_printf(ha->pci_dev, "Auto Negotiation Enabled\n");
285 	} else {
286 		device_printf(ha->pci_dev, "Auto Negotiation Disabled\n");
287 	}
288 
289 	switch (ha->hw.link_speed) {
290 	case 0x710:
291 		device_printf(ha->pci_dev, "link speed\t\t 10Gps\n");
292 		break;
293 
294 	case 0x3E8:
295 		device_printf(ha->pci_dev, "link speed\t\t 1Gps\n");
296 		break;
297 
298 	case 0x64:
299 		device_printf(ha->pci_dev, "link speed\t\t 100Mbps\n");
300 		break;
301 
302 	default:
303 		device_printf(ha->pci_dev, "link speed\t\t Unknown\n");
304 		break;
305 	}
306 
307 	switch (ha->hw.module_type) {
308 
309 	case 0x01:
310 		device_printf(ha->pci_dev, "Module Type 10GBase-LRM\n");
311 		break;
312 
313 	case 0x02:
314 		device_printf(ha->pci_dev, "Module Type 10GBase-LR\n");
315 		break;
316 
317 	case 0x03:
318 		device_printf(ha->pci_dev, "Module Type 10GBase-SR\n");
319 		break;
320 
321 	case 0x04:
322 		device_printf(ha->pci_dev,
323 			"Module Type 10GE Passive Copper(Compliant)[%d m]\n",
324 			ha->hw.cable_length);
325 		break;
326 
327 	case 0x05:
328 		device_printf(ha->pci_dev, "Module Type 10GE Active"
329 			" Limiting Copper(Compliant)[%d m]\n",
330 			ha->hw.cable_length);
331 		break;
332 
333 	case 0x06:
334 		device_printf(ha->pci_dev,
335 			"Module Type 10GE Passive Copper"
336 			" (Legacy, Best Effort)[%d m]\n",
337 			ha->hw.cable_length);
338 		break;
339 
340 	case 0x07:
341 		device_printf(ha->pci_dev, "Module Type 1000Base-SX\n");
342 		break;
343 
344 	case 0x08:
345 		device_printf(ha->pci_dev, "Module Type 1000Base-LX\n");
346 		break;
347 
348 	case 0x09:
349 		device_printf(ha->pci_dev, "Module Type 1000Base-CX\n");
350 		break;
351 
352 	case 0x0A:
353 		device_printf(ha->pci_dev, "Module Type 1000Base-T\n");
354 		break;
355 
356 	case 0x0B:
357 		device_printf(ha->pci_dev, "Module Type 1GE Passive Copper"
358 			"(Legacy, Best Effort)\n");
359 		break;
360 
361 	default:
362 		device_printf(ha->pci_dev, "Unknown Module Type 0x%x\n",
363 			ha->hw.module_type);
364 		break;
365 	}
366 
367 	if (ha->hw.link_faults == 1)
368 		device_printf(ha->pci_dev, "SFP Power Fault\n");
369 }
370 
371 /*
372  * Name: ql_free_dma
373  * Function: Frees the DMA'able memory allocated in ql_alloc_dma()
374  */
375 void
376 ql_free_dma(qla_host_t *ha)
377 {
378 	uint32_t i;
379 
380         if (ha->hw.dma_buf.flags.sds_ring) {
381 		for (i = 0; i < ha->hw.num_sds_rings; i++) {
382 			ql_free_dmabuf(ha, &ha->hw.dma_buf.sds_ring[i]);
383 		}
384         	ha->hw.dma_buf.flags.sds_ring = 0;
385 	}
386 
387         if (ha->hw.dma_buf.flags.rds_ring) {
388 		for (i = 0; i < ha->hw.num_rds_rings; i++) {
389 			ql_free_dmabuf(ha, &ha->hw.dma_buf.rds_ring[i]);
390 		}
391         	ha->hw.dma_buf.flags.rds_ring = 0;
392 	}
393 
394         if (ha->hw.dma_buf.flags.tx_ring) {
395 		ql_free_dmabuf(ha, &ha->hw.dma_buf.tx_ring);
396         	ha->hw.dma_buf.flags.tx_ring = 0;
397 	}
398 	qla_minidump_free(ha);
399 }
400 
401 /*
402  * Name: ql_alloc_dma
403  * Function: Allocates DMA'able memory for Tx/Rx Rings, Tx/Rx Contexts.
404  */
405 int
406 ql_alloc_dma(qla_host_t *ha)
407 {
408         device_t                dev;
409 	uint32_t		i, j, size, tx_ring_size;
410 	qla_hw_t		*hw;
411 	qla_hw_tx_cntxt_t	*tx_cntxt;
412 	uint8_t			*vaddr;
413 	bus_addr_t		paddr;
414 
415         dev = ha->pci_dev;
416 
417         QL_DPRINT2(ha, (dev, "%s: enter\n", __func__));
418 
419 	hw = &ha->hw;
420 	/*
421 	 * Allocate Transmit Ring
422 	 */
423 	tx_ring_size = (sizeof(q80_tx_cmd_t) * NUM_TX_DESCRIPTORS);
424 	size = (tx_ring_size * ha->hw.num_tx_rings);
425 
426 	hw->dma_buf.tx_ring.alignment = 8;
427 	hw->dma_buf.tx_ring.size = size + PAGE_SIZE;
428 
429         if (ql_alloc_dmabuf(ha, &hw->dma_buf.tx_ring)) {
430                 device_printf(dev, "%s: tx ring alloc failed\n", __func__);
431                 goto ql_alloc_dma_exit;
432         }
433 
434 	vaddr = (uint8_t *)hw->dma_buf.tx_ring.dma_b;
435 	paddr = hw->dma_buf.tx_ring.dma_addr;
436 
437 	for (i = 0; i < ha->hw.num_tx_rings; i++) {
438 		tx_cntxt = (qla_hw_tx_cntxt_t *)&hw->tx_cntxt[i];
439 
440 		tx_cntxt->tx_ring_base = (q80_tx_cmd_t *)vaddr;
441 		tx_cntxt->tx_ring_paddr = paddr;
442 
443 		vaddr += tx_ring_size;
444 		paddr += tx_ring_size;
445 	}
446 
447 	for (i = 0; i < ha->hw.num_tx_rings; i++) {
448 		tx_cntxt = (qla_hw_tx_cntxt_t *)&hw->tx_cntxt[i];
449 
450 		tx_cntxt->tx_cons = (uint32_t *)vaddr;
451 		tx_cntxt->tx_cons_paddr = paddr;
452 
453 		vaddr += sizeof (uint32_t);
454 		paddr += sizeof (uint32_t);
455 	}
456 
457         ha->hw.dma_buf.flags.tx_ring = 1;
458 
459 	QL_DPRINT2(ha, (dev, "%s: tx_ring phys %p virt %p\n",
460 		__func__, (void *)(hw->dma_buf.tx_ring.dma_addr),
461 		hw->dma_buf.tx_ring.dma_b));
462 	/*
463 	 * Allocate Receive Descriptor Rings
464 	 */
465 
466 	for (i = 0; i < hw->num_rds_rings; i++) {
467 
468 		hw->dma_buf.rds_ring[i].alignment = 8;
469 		hw->dma_buf.rds_ring[i].size =
470 			(sizeof(q80_recv_desc_t)) * NUM_RX_DESCRIPTORS;
471 
472 		if (ql_alloc_dmabuf(ha, &hw->dma_buf.rds_ring[i])) {
473 			device_printf(dev, "%s: rds ring[%d] alloc failed\n",
474 				__func__, i);
475 
476 			for (j = 0; j < i; j++)
477 				ql_free_dmabuf(ha, &hw->dma_buf.rds_ring[j]);
478 
479 			goto ql_alloc_dma_exit;
480 		}
481 		QL_DPRINT4(ha, (dev, "%s: rx_ring[%d] phys %p virt %p\n",
482 			__func__, i, (void *)(hw->dma_buf.rds_ring[i].dma_addr),
483 			hw->dma_buf.rds_ring[i].dma_b));
484 	}
485 
486 	hw->dma_buf.flags.rds_ring = 1;
487 
488 	/*
489 	 * Allocate Status Descriptor Rings
490 	 */
491 
492 	for (i = 0; i < hw->num_sds_rings; i++) {
493 		hw->dma_buf.sds_ring[i].alignment = 8;
494 		hw->dma_buf.sds_ring[i].size =
495 			(sizeof(q80_stat_desc_t)) * NUM_STATUS_DESCRIPTORS;
496 
497 		if (ql_alloc_dmabuf(ha, &hw->dma_buf.sds_ring[i])) {
498 			device_printf(dev, "%s: sds ring alloc failed\n",
499 				__func__);
500 
501 			for (j = 0; j < i; j++)
502 				ql_free_dmabuf(ha, &hw->dma_buf.sds_ring[j]);
503 
504 			goto ql_alloc_dma_exit;
505 		}
506 		QL_DPRINT4(ha, (dev, "%s: sds_ring[%d] phys %p virt %p\n",
507 			__func__, i,
508 			(void *)(hw->dma_buf.sds_ring[i].dma_addr),
509 			hw->dma_buf.sds_ring[i].dma_b));
510 	}
511 	for (i = 0; i < hw->num_sds_rings; i++) {
512 		hw->sds[i].sds_ring_base =
513 			(q80_stat_desc_t *)hw->dma_buf.sds_ring[i].dma_b;
514 	}
515 
516 	hw->dma_buf.flags.sds_ring = 1;
517 
518 	return 0;
519 
520 ql_alloc_dma_exit:
521 	ql_free_dma(ha);
522 	return -1;
523 }
524 
525 #define Q8_MBX_MSEC_DELAY	5000
526 
527 static int
528 qla_mbx_cmd(qla_host_t *ha, uint32_t *h_mbox, uint32_t n_hmbox,
529 	uint32_t *fw_mbox, uint32_t n_fwmbox, uint32_t no_pause)
530 {
531 	uint32_t i;
532 	uint32_t data;
533 	int ret = 0;
534 
535 	if (QL_ERR_INJECT(ha, INJCT_MBX_CMD_FAILURE)) {
536 		ret = -3;
537 		ha->qla_initiate_recovery = 1;
538 		goto exit_qla_mbx_cmd;
539 	}
540 
541 	if (no_pause)
542 		i = 1000;
543 	else
544 		i = Q8_MBX_MSEC_DELAY;
545 
546 	while (i) {
547 		data = READ_REG32(ha, Q8_HOST_MBOX_CNTRL);
548 		if (data == 0)
549 			break;
550 		if (no_pause) {
551 			DELAY(1000);
552 		} else {
553 			qla_mdelay(__func__, 1);
554 		}
555 		i--;
556 	}
557 
558 	if (i == 0) {
559 		device_printf(ha->pci_dev, "%s: host_mbx_cntrl 0x%08x\n",
560 			__func__, data);
561 		ret = -1;
562 		ha->qla_initiate_recovery = 1;
563 		goto exit_qla_mbx_cmd;
564 	}
565 
566 	for (i = 0; i < n_hmbox; i++) {
567 		WRITE_REG32(ha, (Q8_HOST_MBOX0 + (i << 2)), *h_mbox);
568 		h_mbox++;
569 	}
570 
571 	WRITE_REG32(ha, Q8_HOST_MBOX_CNTRL, 0x1);
572 
573 
574 	i = Q8_MBX_MSEC_DELAY;
575 	while (i) {
576 		data = READ_REG32(ha, Q8_FW_MBOX_CNTRL);
577 
578 		if ((data & 0x3) == 1) {
579 			data = READ_REG32(ha, Q8_FW_MBOX0);
580 			if ((data & 0xF000) != 0x8000)
581 				break;
582 		}
583 		if (no_pause) {
584 			DELAY(1000);
585 		} else {
586 			qla_mdelay(__func__, 1);
587 		}
588 		i--;
589 	}
590 	if (i == 0) {
591 		device_printf(ha->pci_dev, "%s: fw_mbx_cntrl 0x%08x\n",
592 			__func__, data);
593 		ret = -2;
594 		ha->qla_initiate_recovery = 1;
595 		goto exit_qla_mbx_cmd;
596 	}
597 
598 	for (i = 0; i < n_fwmbox; i++) {
599 		*fw_mbox++ = READ_REG32(ha, (Q8_FW_MBOX0 + (i << 2)));
600 	}
601 
602 	WRITE_REG32(ha, Q8_FW_MBOX_CNTRL, 0x0);
603 	WRITE_REG32(ha, ha->hw.mbx_intr_mask_offset, 0x0);
604 
605 exit_qla_mbx_cmd:
606 	return (ret);
607 }
608 
609 static int
610 qla_get_nic_partition(qla_host_t *ha)
611 {
612 	uint32_t *mbox, err;
613 	device_t dev = ha->pci_dev;
614 
615 	bzero(ha->hw.mbox, (sizeof (uint32_t) * Q8_NUM_MBOX));
616 
617 	mbox = ha->hw.mbox;
618 
619 	mbox[0] = Q8_MBX_GET_NIC_PARTITION | (0x2 << 16) | (0x2 << 29);
620 
621 	if (qla_mbx_cmd(ha, mbox, 2, mbox, 19, 0)) {
622 		device_printf(dev, "%s: failed0\n", __func__);
623 		return (-1);
624 	}
625 	err = mbox[0] >> 25;
626 
627 	if ((err != 1) && (err != 0)) {
628 		device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
629 		return (-1);
630 	}
631 	return 0;
632 }
633 
634 static int
635 qla_config_intr_cntxt(qla_host_t *ha, uint32_t num_intrs, uint32_t create)
636 {
637 	uint32_t i, err;
638 	device_t dev = ha->pci_dev;
639 	q80_config_intr_t *c_intr;
640 	q80_config_intr_rsp_t *c_intr_rsp;
641 
642 	c_intr = (q80_config_intr_t *)ha->hw.mbox;
643 	bzero(c_intr, (sizeof (q80_config_intr_t)));
644 
645 	c_intr->opcode = Q8_MBX_CONFIG_INTR;
646 
647 	c_intr->count_version = (sizeof (q80_config_intr_t) >> 2);
648 	c_intr->count_version |= Q8_MBX_CMD_VERSION;
649 
650 	c_intr->nentries = num_intrs;
651 
652 	for (i = 0; i < num_intrs; i++) {
653 		if (create) {
654 			c_intr->intr[i].cmd_type = Q8_MBX_CONFIG_INTR_CREATE;
655 			c_intr->intr[i].msix_index = i + 1;
656 		} else {
657 			c_intr->intr[i].cmd_type = Q8_MBX_CONFIG_INTR_DELETE;
658 			c_intr->intr[i].msix_index = ha->hw.intr_id[i];
659 		}
660 
661 		c_intr->intr[i].cmd_type |= Q8_MBX_CONFIG_INTR_TYPE_MSI_X;
662 	}
663 
664 	if (qla_mbx_cmd(ha, (uint32_t *)c_intr,
665 		(sizeof (q80_config_intr_t) >> 2),
666 		ha->hw.mbox, (sizeof (q80_config_intr_rsp_t) >> 2), 0)) {
667 		device_printf(dev, "%s: failed0\n", __func__);
668 		return (-1);
669 	}
670 
671 	c_intr_rsp = (q80_config_intr_rsp_t *)ha->hw.mbox;
672 
673 	err = Q8_MBX_RSP_STATUS(c_intr_rsp->regcnt_status);
674 
675 	if (err) {
676 		device_printf(dev, "%s: failed1 [0x%08x, %d]\n", __func__, err,
677 			c_intr_rsp->nentries);
678 
679 		for (i = 0; i < c_intr_rsp->nentries; i++) {
680 			device_printf(dev, "%s: [%d]:[0x%x 0x%x 0x%x]\n",
681 				__func__, i,
682 				c_intr_rsp->intr[i].status,
683 				c_intr_rsp->intr[i].intr_id,
684 				c_intr_rsp->intr[i].intr_src);
685 		}
686 
687 		return (-1);
688 	}
689 
690 	for (i = 0; ((i < num_intrs) && create); i++) {
691 		if (!c_intr_rsp->intr[i].status) {
692 			ha->hw.intr_id[i] = c_intr_rsp->intr[i].intr_id;
693 			ha->hw.intr_src[i] = c_intr_rsp->intr[i].intr_src;
694 		}
695 	}
696 
697 	return (0);
698 }
699 
700 /*
701  * Name: qla_config_rss
702  * Function: Configure RSS for the context/interface.
703  */
704 static const uint64_t rss_key[] = { 0xbeac01fa6a42b73bULL,
705 			0x8030f20c77cb2da3ULL,
706 			0xae7b30b4d0ca2bcbULL, 0x43a38fb04167253dULL,
707 			0x255b0ec26d5a56daULL };
708 
709 static int
710 qla_config_rss(qla_host_t *ha, uint16_t cntxt_id)
711 {
712 	q80_config_rss_t	*c_rss;
713 	q80_config_rss_rsp_t	*c_rss_rsp;
714 	uint32_t		err, i;
715 	device_t		dev = ha->pci_dev;
716 
717 	c_rss = (q80_config_rss_t *)ha->hw.mbox;
718 	bzero(c_rss, (sizeof (q80_config_rss_t)));
719 
720 	c_rss->opcode = Q8_MBX_CONFIG_RSS;
721 
722 	c_rss->count_version = (sizeof (q80_config_rss_t) >> 2);
723 	c_rss->count_version |= Q8_MBX_CMD_VERSION;
724 
725 	c_rss->hash_type = (Q8_MBX_RSS_HASH_TYPE_IPV4_TCP_IP |
726 				Q8_MBX_RSS_HASH_TYPE_IPV6_TCP_IP);
727 
728 	c_rss->flags = Q8_MBX_RSS_FLAGS_ENABLE_RSS;
729 	c_rss->flags |= Q8_MBX_RSS_FLAGS_USE_IND_TABLE;
730 
731 	c_rss->indtbl_mask = Q8_MBX_RSS_INDTBL_MASK;
732 
733 	c_rss->indtbl_mask |= Q8_MBX_RSS_FLAGS_MULTI_RSS_VALID;
734 	c_rss->flags |= Q8_MBX_RSS_FLAGS_TYPE_CRSS;
735 
736 	c_rss->cntxt_id = cntxt_id;
737 
738 	for (i = 0; i < 5; i++) {
739 		c_rss->rss_key[i] = rss_key[i];
740 	}
741 
742 	if (qla_mbx_cmd(ha, (uint32_t *)c_rss,
743 		(sizeof (q80_config_rss_t) >> 2),
744 		ha->hw.mbox, (sizeof(q80_config_rss_rsp_t) >> 2), 0)) {
745 		device_printf(dev, "%s: failed0\n", __func__);
746 		return (-1);
747 	}
748 	c_rss_rsp = (q80_config_rss_rsp_t *)ha->hw.mbox;
749 
750 	err = Q8_MBX_RSP_STATUS(c_rss_rsp->regcnt_status);
751 
752 	if (err) {
753 		device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
754 		return (-1);
755 	}
756 	return 0;
757 }
758 
759 static uint8_t rss_ind_default_table[Q8_RSS_IND_TBL_SIZE];
760 
761 static int
762 qla_set_rss_ind_table(qla_host_t *ha, uint32_t start_idx, uint32_t count,
763         uint16_t cntxt_id, uint8_t *ind_table)
764 {
765         q80_config_rss_ind_table_t      *c_rss_ind;
766         q80_config_rss_ind_table_rsp_t  *c_rss_ind_rsp;
767         uint32_t                        err;
768         device_t                        dev = ha->pci_dev;
769 
770 	if ((count > Q8_RSS_IND_TBL_SIZE) ||
771 		((start_idx + count - 1) > Q8_RSS_IND_TBL_MAX_IDX)) {
772 		device_printf(dev, "%s: illegal count [%d, %d]\n", __func__,
773 			start_idx, count);
774 		return (-1);
775 	}
776 
777         c_rss_ind = (q80_config_rss_ind_table_t *)ha->hw.mbox;
778         bzero(c_rss_ind, sizeof (q80_config_rss_ind_table_t));
779 
780         c_rss_ind->opcode = Q8_MBX_CONFIG_RSS_TABLE;
781         c_rss_ind->count_version = (sizeof (q80_config_rss_ind_table_t) >> 2);
782         c_rss_ind->count_version |= Q8_MBX_CMD_VERSION;
783 
784 	c_rss_ind->start_idx = start_idx;
785 	c_rss_ind->end_idx = start_idx + count - 1;
786 	c_rss_ind->cntxt_id = cntxt_id;
787 	bcopy(ind_table, c_rss_ind->ind_table, count);
788 
789 	if (qla_mbx_cmd(ha, (uint32_t *)c_rss_ind,
790 		(sizeof (q80_config_rss_ind_table_t) >> 2), ha->hw.mbox,
791 		(sizeof(q80_config_rss_ind_table_rsp_t) >> 2), 0)) {
792 		device_printf(dev, "%s: failed0\n", __func__);
793 		return (-1);
794 	}
795 
796 	c_rss_ind_rsp = (q80_config_rss_ind_table_rsp_t *)ha->hw.mbox;
797 	err = Q8_MBX_RSP_STATUS(c_rss_ind_rsp->regcnt_status);
798 
799 	if (err) {
800 		device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
801 		return (-1);
802 	}
803 	return 0;
804 }
805 
806 /*
807  * Name: qla_config_intr_coalesce
808  * Function: Configure Interrupt Coalescing.
809  */
810 static int
811 qla_config_intr_coalesce(qla_host_t *ha, uint16_t cntxt_id, int tenable)
812 {
813 	q80_config_intr_coalesc_t	*intrc;
814 	q80_config_intr_coalesc_rsp_t	*intrc_rsp;
815 	uint32_t			err, i;
816 	device_t			dev = ha->pci_dev;
817 
818 	intrc = (q80_config_intr_coalesc_t *)ha->hw.mbox;
819 	bzero(intrc, (sizeof (q80_config_intr_coalesc_t)));
820 
821 	intrc->opcode = Q8_MBX_CONFIG_INTR_COALESCE;
822 	intrc->count_version = (sizeof (q80_config_intr_coalesc_t) >> 2);
823 	intrc->count_version |= Q8_MBX_CMD_VERSION;
824 
825 	intrc->flags = Q8_MBX_INTRC_FLAGS_RCV;
826 	intrc->cntxt_id = cntxt_id;
827 
828 	intrc->max_pkts = 256;
829 	intrc->max_mswait = 3;
830 
831 	if (tenable) {
832 		intrc->flags |= Q8_MBX_INTRC_FLAGS_PERIODIC;
833 		intrc->timer_type = Q8_MBX_INTRC_TIMER_PERIODIC;
834 
835 		for (i = 0; i < ha->hw.num_sds_rings; i++) {
836 			intrc->sds_ring_mask |= (1 << i);
837 		}
838 		intrc->ms_timeout = 1000;
839 	}
840 
841 	if (qla_mbx_cmd(ha, (uint32_t *)intrc,
842 		(sizeof (q80_config_intr_coalesc_t) >> 2),
843 		ha->hw.mbox, (sizeof(q80_config_intr_coalesc_rsp_t) >> 2), 0)) {
844 		device_printf(dev, "%s: failed0\n", __func__);
845 		return (-1);
846 	}
847 	intrc_rsp = (q80_config_intr_coalesc_rsp_t *)ha->hw.mbox;
848 
849 	err = Q8_MBX_RSP_STATUS(intrc_rsp->regcnt_status);
850 
851 	if (err) {
852 		device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
853 		return (-1);
854 	}
855 
856 	return 0;
857 }
858 
859 
860 /*
861  * Name: qla_config_mac_addr
862  * Function: binds a MAC address to the context/interface.
863  *	Can be unicast, multicast or broadcast.
864  */
865 static int
866 qla_config_mac_addr(qla_host_t *ha, uint8_t *mac_addr, uint32_t add_mac)
867 {
868 	q80_config_mac_addr_t		*cmac;
869 	q80_config_mac_addr_rsp_t	*cmac_rsp;
870 	uint32_t			err;
871 	device_t			dev = ha->pci_dev;
872 
873 	cmac = (q80_config_mac_addr_t *)ha->hw.mbox;
874 	bzero(cmac, (sizeof (q80_config_mac_addr_t)));
875 
876 	cmac->opcode = Q8_MBX_CONFIG_MAC_ADDR;
877 	cmac->count_version = sizeof (q80_config_mac_addr_t) >> 2;
878 	cmac->count_version |= Q8_MBX_CMD_VERSION;
879 
880 	if (add_mac)
881 		cmac->cmd = Q8_MBX_CMAC_CMD_ADD_MAC_ADDR;
882 	else
883 		cmac->cmd = Q8_MBX_CMAC_CMD_DEL_MAC_ADDR;
884 
885 	cmac->cmd |= Q8_MBX_CMAC_CMD_CAM_INGRESS;
886 
887 	cmac->nmac_entries = 1;
888 	cmac->cntxt_id = ha->hw.rcv_cntxt_id;
889 	bcopy(mac_addr, cmac->mac_addr[0].addr, 6);
890 
891 	if (qla_mbx_cmd(ha, (uint32_t *)cmac,
892 		(sizeof (q80_config_mac_addr_t) >> 2),
893 		ha->hw.mbox, (sizeof(q80_config_mac_addr_rsp_t) >> 2), 1)) {
894 		device_printf(dev, "%s: %s failed0\n", __func__,
895 			(add_mac ? "Add" : "Del"));
896 		return (-1);
897 	}
898 	cmac_rsp = (q80_config_mac_addr_rsp_t *)ha->hw.mbox;
899 
900 	err = Q8_MBX_RSP_STATUS(cmac_rsp->regcnt_status);
901 
902 	if (err) {
903 		device_printf(dev, "%s: %s "
904 			"%02x:%02x:%02x:%02x:%02x:%02x failed1 [0x%08x]\n",
905 			__func__, (add_mac ? "Add" : "Del"),
906 			mac_addr[0], mac_addr[1], mac_addr[2],
907 			mac_addr[3], mac_addr[4], mac_addr[5], err);
908 		return (-1);
909 	}
910 
911 	return 0;
912 }
913 
914 
915 /*
916  * Name: qla_set_mac_rcv_mode
917  * Function: Enable/Disable AllMulticast and Promiscous Modes.
918  */
919 static int
920 qla_set_mac_rcv_mode(qla_host_t *ha, uint32_t mode)
921 {
922 	q80_config_mac_rcv_mode_t	*rcv_mode;
923 	uint32_t			err;
924 	q80_config_mac_rcv_mode_rsp_t	*rcv_mode_rsp;
925 	device_t			dev = ha->pci_dev;
926 
927 	rcv_mode = (q80_config_mac_rcv_mode_t *)ha->hw.mbox;
928 	bzero(rcv_mode, (sizeof (q80_config_mac_rcv_mode_t)));
929 
930 	rcv_mode->opcode = Q8_MBX_CONFIG_MAC_RX_MODE;
931 	rcv_mode->count_version = sizeof (q80_config_mac_rcv_mode_t) >> 2;
932 	rcv_mode->count_version |= Q8_MBX_CMD_VERSION;
933 
934 	rcv_mode->mode = mode;
935 
936 	rcv_mode->cntxt_id = ha->hw.rcv_cntxt_id;
937 
938 	if (qla_mbx_cmd(ha, (uint32_t *)rcv_mode,
939 		(sizeof (q80_config_mac_rcv_mode_t) >> 2),
940 		ha->hw.mbox, (sizeof(q80_config_mac_rcv_mode_rsp_t) >> 2), 1)) {
941 		device_printf(dev, "%s: failed0\n", __func__);
942 		return (-1);
943 	}
944 	rcv_mode_rsp = (q80_config_mac_rcv_mode_rsp_t *)ha->hw.mbox;
945 
946 	err = Q8_MBX_RSP_STATUS(rcv_mode_rsp->regcnt_status);
947 
948 	if (err) {
949 		device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
950 		return (-1);
951 	}
952 
953 	return 0;
954 }
955 
956 int
957 ql_set_promisc(qla_host_t *ha)
958 {
959 	int ret;
960 
961 	ha->hw.mac_rcv_mode |= Q8_MBX_MAC_RCV_PROMISC_ENABLE;
962 	ret = qla_set_mac_rcv_mode(ha, ha->hw.mac_rcv_mode);
963 	return (ret);
964 }
965 
966 int
967 ql_set_allmulti(qla_host_t *ha)
968 {
969 	int ret;
970 
971 	ha->hw.mac_rcv_mode |= Q8_MBX_MAC_ALL_MULTI_ENABLE;
972 	ret = qla_set_mac_rcv_mode(ha, ha->hw.mac_rcv_mode);
973 	return (ret);
974 }
975 
976 
977 /*
978  * Name: ql_set_max_mtu
979  * Function:
980  *	Sets the maximum transfer unit size for the specified rcv context.
981  */
982 int
983 ql_set_max_mtu(qla_host_t *ha, uint32_t mtu, uint16_t cntxt_id)
984 {
985 	device_t		dev;
986 	q80_set_max_mtu_t	*max_mtu;
987 	q80_set_max_mtu_rsp_t	*max_mtu_rsp;
988 	uint32_t		err;
989 
990 	dev = ha->pci_dev;
991 
992 	max_mtu = (q80_set_max_mtu_t *)ha->hw.mbox;
993 	bzero(max_mtu, (sizeof (q80_set_max_mtu_t)));
994 
995 	max_mtu->opcode = Q8_MBX_SET_MAX_MTU;
996 	max_mtu->count_version = (sizeof (q80_set_max_mtu_t) >> 2);
997 	max_mtu->count_version |= Q8_MBX_CMD_VERSION;
998 
999 	max_mtu->cntxt_id = cntxt_id;
1000 	max_mtu->mtu = mtu;
1001 
1002         if (qla_mbx_cmd(ha, (uint32_t *)max_mtu,
1003 		(sizeof (q80_set_max_mtu_t) >> 2),
1004                 ha->hw.mbox, (sizeof (q80_set_max_mtu_rsp_t) >> 2), 1)) {
1005                 device_printf(dev, "%s: failed\n", __func__);
1006                 return -1;
1007         }
1008 
1009 	max_mtu_rsp = (q80_set_max_mtu_rsp_t *)ha->hw.mbox;
1010 
1011         err = Q8_MBX_RSP_STATUS(max_mtu_rsp->regcnt_status);
1012 
1013         if (err) {
1014                 device_printf(dev, "%s: failed [0x%08x]\n", __func__, err);
1015         }
1016 
1017 	return 0;
1018 }
1019 
1020 static int
1021 qla_link_event_req(qla_host_t *ha, uint16_t cntxt_id)
1022 {
1023 	device_t		dev;
1024 	q80_link_event_t	*lnk;
1025 	q80_link_event_rsp_t	*lnk_rsp;
1026 	uint32_t		err;
1027 
1028 	dev = ha->pci_dev;
1029 
1030 	lnk = (q80_link_event_t *)ha->hw.mbox;
1031 	bzero(lnk, (sizeof (q80_link_event_t)));
1032 
1033 	lnk->opcode = Q8_MBX_LINK_EVENT_REQ;
1034 	lnk->count_version = (sizeof (q80_link_event_t) >> 2);
1035 	lnk->count_version |= Q8_MBX_CMD_VERSION;
1036 
1037 	lnk->cntxt_id = cntxt_id;
1038 	lnk->cmd = Q8_LINK_EVENT_CMD_ENABLE_ASYNC;
1039 
1040         if (qla_mbx_cmd(ha, (uint32_t *)lnk, (sizeof (q80_link_event_t) >> 2),
1041                 ha->hw.mbox, (sizeof (q80_link_event_rsp_t) >> 2), 0)) {
1042                 device_printf(dev, "%s: failed\n", __func__);
1043                 return -1;
1044         }
1045 
1046 	lnk_rsp = (q80_link_event_rsp_t *)ha->hw.mbox;
1047 
1048         err = Q8_MBX_RSP_STATUS(lnk_rsp->regcnt_status);
1049 
1050         if (err) {
1051                 device_printf(dev, "%s: failed [0x%08x]\n", __func__, err);
1052         }
1053 
1054 	return 0;
1055 }
1056 
1057 static int
1058 qla_config_fw_lro(qla_host_t *ha, uint16_t cntxt_id)
1059 {
1060 	device_t		dev;
1061 	q80_config_fw_lro_t	*fw_lro;
1062 	q80_config_fw_lro_rsp_t	*fw_lro_rsp;
1063 	uint32_t		err;
1064 
1065 	dev = ha->pci_dev;
1066 
1067 	fw_lro = (q80_config_fw_lro_t *)ha->hw.mbox;
1068 	bzero(fw_lro, sizeof(q80_config_fw_lro_t));
1069 
1070 	fw_lro->opcode = Q8_MBX_CONFIG_FW_LRO;
1071 	fw_lro->count_version = (sizeof (q80_config_fw_lro_t) >> 2);
1072 	fw_lro->count_version |= Q8_MBX_CMD_VERSION;
1073 
1074 	fw_lro->flags |= Q8_MBX_FW_LRO_IPV4 | Q8_MBX_FW_LRO_IPV4_WO_DST_IP_CHK;
1075 
1076 	fw_lro->cntxt_id = cntxt_id;
1077 
1078 	if (ha->hw.min_lro_pkt_size) {
1079 		fw_lro->flags |= Q8_MBX_FW_LRO_LOW_THRESHOLD;
1080 		fw_lro->low_threshold = ha->hw.min_lro_pkt_size;
1081 	}
1082 
1083 	if (qla_mbx_cmd(ha, (uint32_t *)fw_lro,
1084 		(sizeof (q80_config_fw_lro_t) >> 2),
1085 		ha->hw.mbox, (sizeof (q80_config_fw_lro_rsp_t) >> 2), 0)) {
1086 		device_printf(dev, "%s: failed\n", __func__);
1087 		return -1;
1088 	}
1089 
1090 	fw_lro_rsp = (q80_config_fw_lro_rsp_t *)ha->hw.mbox;
1091 
1092 	err = Q8_MBX_RSP_STATUS(fw_lro_rsp->regcnt_status);
1093 
1094 	if (err) {
1095 		device_printf(dev, "%s: failed [0x%08x]\n", __func__, err);
1096 	}
1097 
1098 	return 0;
1099 }
1100 
1101 static void
1102 qla_xmt_stats(qla_host_t *ha, q80_xmt_stats_t *xstat)
1103 {
1104 	device_t dev = ha->pci_dev;
1105 
1106 	device_printf(dev, "%s: total_bytes\t\t%" PRIu64 "\n", __func__,
1107 		xstat->total_bytes);
1108 	device_printf(dev, "%s: total_pkts\t\t%" PRIu64 "\n", __func__,
1109 		xstat->total_pkts);
1110 	device_printf(dev, "%s: errors\t\t%" PRIu64 "\n", __func__,
1111 		xstat->errors);
1112 	device_printf(dev, "%s: pkts_dropped\t%" PRIu64 "\n", __func__,
1113 		xstat->pkts_dropped);
1114 	device_printf(dev, "%s: switch_pkts\t\t%" PRIu64 "\n", __func__,
1115 		xstat->switch_pkts);
1116 	device_printf(dev, "%s: num_buffers\t\t%" PRIu64 "\n", __func__,
1117 		xstat->num_buffers);
1118 }
1119 
1120 static void
1121 qla_rcv_stats(qla_host_t *ha, q80_rcv_stats_t *rstat)
1122 {
1123 	device_t dev = ha->pci_dev;
1124 
1125 	device_printf(dev, "%s: total_bytes\t\t\t%" PRIu64 "\n", __func__,
1126 		rstat->total_bytes);
1127 	device_printf(dev, "%s: total_pkts\t\t\t%" PRIu64 "\n", __func__,
1128 		rstat->total_pkts);
1129 	device_printf(dev, "%s: lro_pkt_count\t\t%" PRIu64 "\n", __func__,
1130 		rstat->lro_pkt_count);
1131 	device_printf(dev, "%s: sw_pkt_count\t\t%" PRIu64 "\n", __func__,
1132 		rstat->sw_pkt_count);
1133 	device_printf(dev, "%s: ip_chksum_err\t\t%" PRIu64 "\n", __func__,
1134 		rstat->ip_chksum_err);
1135 	device_printf(dev, "%s: pkts_wo_acntxts\t\t%" PRIu64 "\n", __func__,
1136 		rstat->pkts_wo_acntxts);
1137 	device_printf(dev, "%s: pkts_dropped_no_sds_card\t%" PRIu64 "\n",
1138 		__func__, rstat->pkts_dropped_no_sds_card);
1139 	device_printf(dev, "%s: pkts_dropped_no_sds_host\t%" PRIu64 "\n",
1140 		__func__, rstat->pkts_dropped_no_sds_host);
1141 	device_printf(dev, "%s: oversized_pkts\t\t%" PRIu64 "\n", __func__,
1142 		rstat->oversized_pkts);
1143 	device_printf(dev, "%s: pkts_dropped_no_rds\t\t%" PRIu64 "\n",
1144 		__func__, rstat->pkts_dropped_no_rds);
1145 	device_printf(dev, "%s: unxpctd_mcast_pkts\t\t%" PRIu64 "\n",
1146 		__func__, rstat->unxpctd_mcast_pkts);
1147 	device_printf(dev, "%s: re1_fbq_error\t\t%" PRIu64 "\n", __func__,
1148 		rstat->re1_fbq_error);
1149 	device_printf(dev, "%s: invalid_mac_addr\t\t%" PRIu64 "\n", __func__,
1150 		rstat->invalid_mac_addr);
1151 	device_printf(dev, "%s: rds_prime_trys\t\t%" PRIu64 "\n", __func__,
1152 		rstat->rds_prime_trys);
1153 	device_printf(dev, "%s: rds_prime_success\t\t%" PRIu64 "\n", __func__,
1154 		rstat->rds_prime_success);
1155 	device_printf(dev, "%s: lro_flows_added\t\t%" PRIu64 "\n", __func__,
1156 		rstat->lro_flows_added);
1157 	device_printf(dev, "%s: lro_flows_deleted\t\t%" PRIu64 "\n", __func__,
1158 		rstat->lro_flows_deleted);
1159 	device_printf(dev, "%s: lro_flows_active\t\t%" PRIu64 "\n", __func__,
1160 		rstat->lro_flows_active);
1161 	device_printf(dev, "%s: pkts_droped_unknown\t\t%" PRIu64 "\n",
1162 		__func__, rstat->pkts_droped_unknown);
1163 }
1164 
1165 static void
1166 qla_mac_stats(qla_host_t *ha, q80_mac_stats_t *mstat)
1167 {
1168 	device_t dev = ha->pci_dev;
1169 
1170 	device_printf(dev, "%s: xmt_frames\t\t\t%" PRIu64 "\n", __func__,
1171 		mstat->xmt_frames);
1172 	device_printf(dev, "%s: xmt_bytes\t\t\t%" PRIu64 "\n", __func__,
1173 		mstat->xmt_bytes);
1174 	device_printf(dev, "%s: xmt_mcast_pkts\t\t%" PRIu64 "\n", __func__,
1175 		mstat->xmt_mcast_pkts);
1176 	device_printf(dev, "%s: xmt_bcast_pkts\t\t%" PRIu64 "\n", __func__,
1177 		mstat->xmt_bcast_pkts);
1178 	device_printf(dev, "%s: xmt_pause_frames\t\t%" PRIu64 "\n", __func__,
1179 		mstat->xmt_pause_frames);
1180 	device_printf(dev, "%s: xmt_cntrl_pkts\t\t%" PRIu64 "\n", __func__,
1181 		mstat->xmt_cntrl_pkts);
1182 	device_printf(dev, "%s: xmt_pkt_lt_64bytes\t\t%" PRIu64 "\n",
1183 		__func__, mstat->xmt_pkt_lt_64bytes);
1184 	device_printf(dev, "%s: xmt_pkt_lt_127bytes\t\t%" PRIu64 "\n",
1185 		__func__, mstat->xmt_pkt_lt_127bytes);
1186 	device_printf(dev, "%s: xmt_pkt_lt_255bytes\t\t%" PRIu64 "\n",
1187 		__func__, mstat->xmt_pkt_lt_255bytes);
1188 	device_printf(dev, "%s: xmt_pkt_lt_511bytes\t\t%" PRIu64 "\n",
1189 		__func__, mstat->xmt_pkt_lt_511bytes);
1190 	device_printf(dev, "%s: xmt_pkt_lt_1023bytes\t%" PRIu64 "\n",
1191 		__func__, mstat->xmt_pkt_lt_1023bytes);
1192 	device_printf(dev, "%s: xmt_pkt_lt_1518bytes\t%" PRIu64 "\n",
1193 		__func__, mstat->xmt_pkt_lt_1518bytes);
1194 	device_printf(dev, "%s: xmt_pkt_gt_1518bytes\t%" PRIu64 "\n",
1195 		__func__, mstat->xmt_pkt_gt_1518bytes);
1196 
1197 	device_printf(dev, "%s: rcv_frames\t\t\t%" PRIu64 "\n", __func__,
1198 		mstat->rcv_frames);
1199 	device_printf(dev, "%s: rcv_bytes\t\t\t%" PRIu64 "\n", __func__,
1200 		mstat->rcv_bytes);
1201 	device_printf(dev, "%s: rcv_mcast_pkts\t\t%" PRIu64 "\n", __func__,
1202 		mstat->rcv_mcast_pkts);
1203 	device_printf(dev, "%s: rcv_bcast_pkts\t\t%" PRIu64 "\n", __func__,
1204 		mstat->rcv_bcast_pkts);
1205 	device_printf(dev, "%s: rcv_pause_frames\t\t%" PRIu64 "\n", __func__,
1206 		mstat->rcv_pause_frames);
1207 	device_printf(dev, "%s: rcv_cntrl_pkts\t\t%" PRIu64 "\n", __func__,
1208 		mstat->rcv_cntrl_pkts);
1209 	device_printf(dev, "%s: rcv_pkt_lt_64bytes\t\t%" PRIu64 "\n",
1210 		__func__, mstat->rcv_pkt_lt_64bytes);
1211 	device_printf(dev, "%s: rcv_pkt_lt_127bytes\t\t%" PRIu64 "\n",
1212 		__func__, mstat->rcv_pkt_lt_127bytes);
1213 	device_printf(dev, "%s: rcv_pkt_lt_255bytes\t\t%" PRIu64 "\n",
1214 		__func__, mstat->rcv_pkt_lt_255bytes);
1215 	device_printf(dev, "%s: rcv_pkt_lt_511bytes\t\t%" PRIu64 "\n",
1216 		__func__, mstat->rcv_pkt_lt_511bytes);
1217 	device_printf(dev, "%s: rcv_pkt_lt_1023bytes\t%" PRIu64 "\n",
1218 		__func__, mstat->rcv_pkt_lt_1023bytes);
1219 	device_printf(dev, "%s: rcv_pkt_lt_1518bytes\t%" PRIu64 "\n",
1220 		__func__, mstat->rcv_pkt_lt_1518bytes);
1221 	device_printf(dev, "%s: rcv_pkt_gt_1518bytes\t%" PRIu64 "\n",
1222 		__func__, mstat->rcv_pkt_gt_1518bytes);
1223 
1224 	device_printf(dev, "%s: rcv_len_error\t\t%" PRIu64 "\n", __func__,
1225 		mstat->rcv_len_error);
1226 	device_printf(dev, "%s: rcv_len_small\t\t%" PRIu64 "\n", __func__,
1227 		mstat->rcv_len_small);
1228 	device_printf(dev, "%s: rcv_len_large\t\t%" PRIu64 "\n", __func__,
1229 		mstat->rcv_len_large);
1230 	device_printf(dev, "%s: rcv_jabber\t\t\t%" PRIu64 "\n", __func__,
1231 		mstat->rcv_jabber);
1232 	device_printf(dev, "%s: rcv_dropped\t\t\t%" PRIu64 "\n", __func__,
1233 		mstat->rcv_dropped);
1234 	device_printf(dev, "%s: fcs_error\t\t\t%" PRIu64 "\n", __func__,
1235 		mstat->fcs_error);
1236 	device_printf(dev, "%s: align_error\t\t\t%" PRIu64 "\n", __func__,
1237 		mstat->align_error);
1238 }
1239 
1240 
1241 static int
1242 qla_get_hw_stats(qla_host_t *ha, uint32_t cmd)
1243 {
1244 	device_t		dev;
1245 	q80_get_stats_t		*stat;
1246 	q80_get_stats_rsp_t	*stat_rsp;
1247 	uint32_t		err;
1248 
1249 	dev = ha->pci_dev;
1250 
1251 	stat = (q80_get_stats_t *)ha->hw.mbox;
1252 	bzero(stat, (sizeof (q80_get_stats_t)));
1253 
1254 	stat->opcode = Q8_MBX_GET_STATS;
1255 	stat->count_version = 2;
1256 	stat->count_version |= Q8_MBX_CMD_VERSION;
1257 
1258 	stat->cmd = cmd;
1259 
1260         if (qla_mbx_cmd(ha, (uint32_t *)stat, 2,
1261                 ha->hw.mbox, (sizeof (q80_get_stats_rsp_t) >> 2), 0)) {
1262                 device_printf(dev, "%s: failed\n", __func__);
1263                 return -1;
1264         }
1265 
1266 	stat_rsp = (q80_get_stats_rsp_t *)ha->hw.mbox;
1267 
1268         err = Q8_MBX_RSP_STATUS(stat_rsp->regcnt_status);
1269 
1270         if (err) {
1271                 return -1;
1272         }
1273 
1274 	return 0;
1275 }
1276 
1277 void
1278 ql_get_stats(qla_host_t *ha)
1279 {
1280 	q80_get_stats_rsp_t	*stat_rsp;
1281 	q80_mac_stats_t		*mstat;
1282 	q80_xmt_stats_t		*xstat;
1283 	q80_rcv_stats_t		*rstat;
1284 	uint32_t		cmd;
1285 
1286 	stat_rsp = (q80_get_stats_rsp_t *)ha->hw.mbox;
1287 	/*
1288 	 * Get MAC Statistics
1289 	 */
1290 	cmd = Q8_GET_STATS_CMD_TYPE_MAC;
1291 
1292 	cmd |= ((ha->pci_func & 0x1) << 16);
1293 
1294 	if (qla_get_hw_stats(ha, cmd) == 0) {
1295 		mstat = (q80_mac_stats_t *)&stat_rsp->u.mac;
1296 		qla_mac_stats(ha, mstat);
1297 	} else {
1298                 device_printf(ha->pci_dev, "%s: mac failed [0x%08x]\n",
1299 			__func__, ha->hw.mbox[0]);
1300 	}
1301 	/*
1302 	 * Get RCV Statistics
1303 	 */
1304 	cmd = Q8_GET_STATS_CMD_RCV | Q8_GET_STATS_CMD_TYPE_CNTXT;
1305 	cmd |= (ha->hw.rcv_cntxt_id << 16);
1306 
1307 	if (qla_get_hw_stats(ha, cmd) == 0) {
1308 		rstat = (q80_rcv_stats_t *)&stat_rsp->u.rcv;
1309 		qla_rcv_stats(ha, rstat);
1310 	} else {
1311                 device_printf(ha->pci_dev, "%s: rcv failed [0x%08x]\n",
1312 			__func__, ha->hw.mbox[0]);
1313 	}
1314 	/*
1315 	 * Get XMT Statistics
1316 	 */
1317 	cmd = Q8_GET_STATS_CMD_XMT | Q8_GET_STATS_CMD_TYPE_CNTXT;
1318 	cmd |= (ha->hw.tx_cntxt[ha->txr_idx].tx_cntxt_id << 16);
1319 
1320 
1321 	if (qla_get_hw_stats(ha, cmd) == 0) {
1322 		xstat = (q80_xmt_stats_t *)&stat_rsp->u.xmt;
1323 		qla_xmt_stats(ha, xstat);
1324 	} else {
1325                 device_printf(ha->pci_dev, "%s: xmt failed [0x%08x]\n",
1326 			__func__, ha->hw.mbox[0]);
1327 	}
1328 }
1329 
1330 /*
1331  * Name: qla_tx_tso
1332  * Function: Checks if the packet to be transmitted is a candidate for
1333  *	Large TCP Segment Offload. If yes, the appropriate fields in the Tx
1334  *	Ring Structure are plugged in.
1335  */
1336 static int
1337 qla_tx_tso(qla_host_t *ha, struct mbuf *mp, q80_tx_cmd_t *tx_cmd, uint8_t *hdr)
1338 {
1339 	struct ether_vlan_header *eh;
1340 	struct ip *ip = NULL;
1341 	struct ip6_hdr *ip6 = NULL;
1342 	struct tcphdr *th = NULL;
1343 	uint32_t ehdrlen,  hdrlen, ip_hlen, tcp_hlen, tcp_opt_off;
1344 	uint16_t etype, opcode, offload = 1;
1345 	device_t dev;
1346 
1347 	dev = ha->pci_dev;
1348 
1349 
1350 	eh = mtod(mp, struct ether_vlan_header *);
1351 
1352 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
1353 		ehdrlen = ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN;
1354 		etype = ntohs(eh->evl_proto);
1355 	} else {
1356 		ehdrlen = ETHER_HDR_LEN;
1357 		etype = ntohs(eh->evl_encap_proto);
1358 	}
1359 
1360 	hdrlen = 0;
1361 
1362 	switch (etype) {
1363 		case ETHERTYPE_IP:
1364 
1365 			tcp_opt_off = ehdrlen + sizeof(struct ip) +
1366 					sizeof(struct tcphdr);
1367 
1368 			if (mp->m_len < tcp_opt_off) {
1369 				m_copydata(mp, 0, tcp_opt_off, hdr);
1370 				ip = (struct ip *)(hdr + ehdrlen);
1371 			} else {
1372 				ip = (struct ip *)(mp->m_data + ehdrlen);
1373 			}
1374 
1375 			ip_hlen = ip->ip_hl << 2;
1376 			opcode = Q8_TX_CMD_OP_XMT_TCP_LSO;
1377 
1378 
1379 			if ((ip->ip_p != IPPROTO_TCP) ||
1380 				(ip_hlen != sizeof (struct ip))){
1381 				/* IP Options are not supported */
1382 
1383 				offload = 0;
1384 			} else
1385 				th = (struct tcphdr *)((caddr_t)ip + ip_hlen);
1386 
1387 		break;
1388 
1389 		case ETHERTYPE_IPV6:
1390 
1391 			tcp_opt_off = ehdrlen + sizeof(struct ip6_hdr) +
1392 					sizeof (struct tcphdr);
1393 
1394 			if (mp->m_len < tcp_opt_off) {
1395 				m_copydata(mp, 0, tcp_opt_off, hdr);
1396 				ip6 = (struct ip6_hdr *)(hdr + ehdrlen);
1397 			} else {
1398 				ip6 = (struct ip6_hdr *)(mp->m_data + ehdrlen);
1399 			}
1400 
1401 			ip_hlen = sizeof(struct ip6_hdr);
1402 			opcode = Q8_TX_CMD_OP_XMT_TCP_LSO_IPV6;
1403 
1404 			if (ip6->ip6_nxt != IPPROTO_TCP) {
1405 				//device_printf(dev, "%s: ipv6\n", __func__);
1406 				offload = 0;
1407 			} else
1408 				th = (struct tcphdr *)((caddr_t)ip6 + ip_hlen);
1409 		break;
1410 
1411 		default:
1412 			QL_DPRINT8(ha, (dev, "%s: type!=ip\n", __func__));
1413 			offload = 0;
1414 		break;
1415 	}
1416 
1417 	if (!offload)
1418 		return (-1);
1419 
1420 	tcp_hlen = th->th_off << 2;
1421 	hdrlen = ehdrlen + ip_hlen + tcp_hlen;
1422 
1423         if (mp->m_len < hdrlen) {
1424                 if (mp->m_len < tcp_opt_off) {
1425                         if (tcp_hlen > sizeof(struct tcphdr)) {
1426                                 m_copydata(mp, tcp_opt_off,
1427                                         (tcp_hlen - sizeof(struct tcphdr)),
1428                                         &hdr[tcp_opt_off]);
1429                         }
1430                 } else {
1431                         m_copydata(mp, 0, hdrlen, hdr);
1432                 }
1433         }
1434 
1435 	tx_cmd->mss = mp->m_pkthdr.tso_segsz;
1436 
1437 	tx_cmd->flags_opcode = opcode ;
1438 	tx_cmd->tcp_hdr_off = ip_hlen + ehdrlen;
1439 	tx_cmd->total_hdr_len = hdrlen;
1440 
1441 	/* Check for Multicast least significant bit of MSB == 1 */
1442 	if (eh->evl_dhost[0] & 0x01) {
1443 		tx_cmd->flags_opcode |= Q8_TX_CMD_FLAGS_MULTICAST;
1444 	}
1445 
1446 	if (mp->m_len < hdrlen) {
1447 		printf("%d\n", hdrlen);
1448 		return (1);
1449 	}
1450 
1451 	return (0);
1452 }
1453 
1454 /*
1455  * Name: qla_tx_chksum
1456  * Function: Checks if the packet to be transmitted is a candidate for
1457  *	TCP/UDP Checksum offload. If yes, the appropriate fields in the Tx
1458  *	Ring Structure are plugged in.
1459  */
1460 static int
1461 qla_tx_chksum(qla_host_t *ha, struct mbuf *mp, uint32_t *op_code,
1462 	uint32_t *tcp_hdr_off)
1463 {
1464 	struct ether_vlan_header *eh;
1465 	struct ip *ip;
1466 	struct ip6_hdr *ip6;
1467 	uint32_t ehdrlen, ip_hlen;
1468 	uint16_t etype, opcode, offload = 1;
1469 	device_t dev;
1470 	uint8_t buf[sizeof(struct ip6_hdr)];
1471 
1472 	dev = ha->pci_dev;
1473 
1474 	*op_code = 0;
1475 
1476 	if ((mp->m_pkthdr.csum_flags & (CSUM_TCP|CSUM_UDP)) == 0)
1477 		return (-1);
1478 
1479 	eh = mtod(mp, struct ether_vlan_header *);
1480 
1481 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
1482 		ehdrlen = ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN;
1483 		etype = ntohs(eh->evl_proto);
1484 	} else {
1485 		ehdrlen = ETHER_HDR_LEN;
1486 		etype = ntohs(eh->evl_encap_proto);
1487 	}
1488 
1489 
1490 	switch (etype) {
1491 		case ETHERTYPE_IP:
1492 			ip = (struct ip *)(mp->m_data + ehdrlen);
1493 
1494 			ip_hlen = sizeof (struct ip);
1495 
1496 			if (mp->m_len < (ehdrlen + ip_hlen)) {
1497 				m_copydata(mp, ehdrlen, sizeof(struct ip), buf);
1498 				ip = (struct ip *)buf;
1499 			}
1500 
1501 			if (ip->ip_p == IPPROTO_TCP)
1502 				opcode = Q8_TX_CMD_OP_XMT_TCP_CHKSUM;
1503 			else if (ip->ip_p == IPPROTO_UDP)
1504 				opcode = Q8_TX_CMD_OP_XMT_UDP_CHKSUM;
1505 			else {
1506 				//device_printf(dev, "%s: ipv4\n", __func__);
1507 				offload = 0;
1508 			}
1509 		break;
1510 
1511 		case ETHERTYPE_IPV6:
1512 			ip6 = (struct ip6_hdr *)(mp->m_data + ehdrlen);
1513 
1514 			ip_hlen = sizeof(struct ip6_hdr);
1515 
1516 			if (mp->m_len < (ehdrlen + ip_hlen)) {
1517 				m_copydata(mp, ehdrlen, sizeof (struct ip6_hdr),
1518 					buf);
1519 				ip6 = (struct ip6_hdr *)buf;
1520 			}
1521 
1522 			if (ip6->ip6_nxt == IPPROTO_TCP)
1523 				opcode = Q8_TX_CMD_OP_XMT_TCP_CHKSUM_IPV6;
1524 			else if (ip6->ip6_nxt == IPPROTO_UDP)
1525 				opcode = Q8_TX_CMD_OP_XMT_UDP_CHKSUM_IPV6;
1526 			else {
1527 				//device_printf(dev, "%s: ipv6\n", __func__);
1528 				offload = 0;
1529 			}
1530 		break;
1531 
1532 		default:
1533 			offload = 0;
1534 		break;
1535 	}
1536 	if (!offload)
1537 		return (-1);
1538 
1539 	*op_code = opcode;
1540 	*tcp_hdr_off = (ip_hlen + ehdrlen);
1541 
1542 	return (0);
1543 }
1544 
1545 #define QLA_TX_MIN_FREE 2
1546 /*
1547  * Name: ql_hw_send
1548  * Function: Transmits a packet. It first checks if the packet is a
1549  *	candidate for Large TCP Segment Offload and then for UDP/TCP checksum
1550  *	offload. If either of these creteria are not met, it is transmitted
1551  *	as a regular ethernet frame.
1552  */
1553 int
1554 ql_hw_send(qla_host_t *ha, bus_dma_segment_t *segs, int nsegs,
1555 	uint32_t tx_idx,  struct mbuf *mp, uint32_t txr_idx)
1556 {
1557 	struct ether_vlan_header *eh;
1558 	qla_hw_t *hw = &ha->hw;
1559 	q80_tx_cmd_t *tx_cmd, tso_cmd;
1560 	bus_dma_segment_t *c_seg;
1561 	uint32_t num_tx_cmds, hdr_len = 0;
1562 	uint32_t total_length = 0, bytes, tx_cmd_count = 0, txr_next;
1563 	device_t dev;
1564 	int i, ret;
1565 	uint8_t *src = NULL, *dst = NULL;
1566 	uint8_t frame_hdr[QL_FRAME_HDR_SIZE];
1567 	uint32_t op_code = 0;
1568 	uint32_t tcp_hdr_off = 0;
1569 
1570 	dev = ha->pci_dev;
1571 
1572 	/*
1573 	 * Always make sure there is atleast one empty slot in the tx_ring
1574 	 * tx_ring is considered full when there only one entry available
1575 	 */
1576         num_tx_cmds = (nsegs + (Q8_TX_CMD_MAX_SEGMENTS - 1)) >> 2;
1577 
1578 	total_length = mp->m_pkthdr.len;
1579 	if (total_length > QLA_MAX_TSO_FRAME_SIZE) {
1580 		device_printf(dev, "%s: total length exceeds maxlen(%d)\n",
1581 			__func__, total_length);
1582 		return (-1);
1583 	}
1584 	eh = mtod(mp, struct ether_vlan_header *);
1585 
1586 	if (mp->m_pkthdr.csum_flags & CSUM_TSO) {
1587 
1588 		bzero((void *)&tso_cmd, sizeof(q80_tx_cmd_t));
1589 
1590 		src = frame_hdr;
1591 		ret = qla_tx_tso(ha, mp, &tso_cmd, src);
1592 
1593 		if (!(ret & ~1)) {
1594 			/* find the additional tx_cmd descriptors required */
1595 
1596 			if (mp->m_flags & M_VLANTAG)
1597 				tso_cmd.total_hdr_len += ETHER_VLAN_ENCAP_LEN;
1598 
1599 			hdr_len = tso_cmd.total_hdr_len;
1600 
1601 			bytes = sizeof(q80_tx_cmd_t) - Q8_TX_CMD_TSO_ALIGN;
1602 			bytes = QL_MIN(bytes, hdr_len);
1603 
1604 			num_tx_cmds++;
1605 			hdr_len -= bytes;
1606 
1607 			while (hdr_len) {
1608 				bytes = QL_MIN((sizeof(q80_tx_cmd_t)), hdr_len);
1609 				hdr_len -= bytes;
1610 				num_tx_cmds++;
1611 			}
1612 			hdr_len = tso_cmd.total_hdr_len;
1613 
1614 			if (ret == 0)
1615 				src = (uint8_t *)eh;
1616 		} else
1617 			return (EINVAL);
1618 	} else {
1619 		(void)qla_tx_chksum(ha, mp, &op_code, &tcp_hdr_off);
1620 	}
1621 
1622 	if (hw->tx_cntxt[txr_idx].txr_free <= (num_tx_cmds + QLA_TX_MIN_FREE)) {
1623 		qla_hw_tx_done_locked(ha, txr_idx);
1624 		if (hw->tx_cntxt[txr_idx].txr_free <=
1625 				(num_tx_cmds + QLA_TX_MIN_FREE)) {
1626         		QL_DPRINT8(ha, (dev, "%s: (hw->txr_free <= "
1627 				"(num_tx_cmds + QLA_TX_MIN_FREE))\n",
1628 				__func__));
1629 			return (-1);
1630 		}
1631 	}
1632 
1633 	tx_cmd = &hw->tx_cntxt[txr_idx].tx_ring_base[tx_idx];
1634 
1635         if (!(mp->m_pkthdr.csum_flags & CSUM_TSO)) {
1636 
1637                 if (nsegs > ha->hw.max_tx_segs)
1638                         ha->hw.max_tx_segs = nsegs;
1639 
1640                 bzero((void *)tx_cmd, sizeof(q80_tx_cmd_t));
1641 
1642                 if (op_code) {
1643                         tx_cmd->flags_opcode = op_code;
1644                         tx_cmd->tcp_hdr_off = tcp_hdr_off;
1645 
1646                 } else {
1647                         tx_cmd->flags_opcode = Q8_TX_CMD_OP_XMT_ETHER;
1648                 }
1649 	} else {
1650 		bcopy(&tso_cmd, tx_cmd, sizeof(q80_tx_cmd_t));
1651 		ha->tx_tso_frames++;
1652 	}
1653 
1654 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
1655         	tx_cmd->flags_opcode |= Q8_TX_CMD_FLAGS_VLAN_TAGGED;
1656 	} else if (mp->m_flags & M_VLANTAG) {
1657 
1658 		if (hdr_len) { /* TSO */
1659 			tx_cmd->flags_opcode |= (Q8_TX_CMD_FLAGS_VLAN_TAGGED |
1660 						Q8_TX_CMD_FLAGS_HW_VLAN_ID);
1661 			tx_cmd->tcp_hdr_off += ETHER_VLAN_ENCAP_LEN;
1662 		} else
1663 			tx_cmd->flags_opcode |= Q8_TX_CMD_FLAGS_HW_VLAN_ID;
1664 
1665 		ha->hw_vlan_tx_frames++;
1666 		tx_cmd->vlan_tci = mp->m_pkthdr.ether_vtag;
1667 	}
1668 
1669 
1670         tx_cmd->n_bufs = (uint8_t)nsegs;
1671         tx_cmd->data_len_lo = (uint8_t)(total_length & 0xFF);
1672         tx_cmd->data_len_hi = qla_host_to_le16(((uint16_t)(total_length >> 8)));
1673 	tx_cmd->cntxtid = Q8_TX_CMD_PORT_CNXTID(ha->pci_func);
1674 
1675 	c_seg = segs;
1676 
1677 	while (1) {
1678 		for (i = 0; ((i < Q8_TX_CMD_MAX_SEGMENTS) && nsegs); i++) {
1679 
1680 			switch (i) {
1681 			case 0:
1682 				tx_cmd->buf1_addr = c_seg->ds_addr;
1683 				tx_cmd->buf1_len = c_seg->ds_len;
1684 				break;
1685 
1686 			case 1:
1687 				tx_cmd->buf2_addr = c_seg->ds_addr;
1688 				tx_cmd->buf2_len = c_seg->ds_len;
1689 				break;
1690 
1691 			case 2:
1692 				tx_cmd->buf3_addr = c_seg->ds_addr;
1693 				tx_cmd->buf3_len = c_seg->ds_len;
1694 				break;
1695 
1696 			case 3:
1697 				tx_cmd->buf4_addr = c_seg->ds_addr;
1698 				tx_cmd->buf4_len = c_seg->ds_len;
1699 				break;
1700 			}
1701 
1702 			c_seg++;
1703 			nsegs--;
1704 		}
1705 
1706 		txr_next = hw->tx_cntxt[txr_idx].txr_next =
1707 			(hw->tx_cntxt[txr_idx].txr_next + 1) &
1708 				(NUM_TX_DESCRIPTORS - 1);
1709 		tx_cmd_count++;
1710 
1711 		if (!nsegs)
1712 			break;
1713 
1714 		tx_cmd = &hw->tx_cntxt[txr_idx].tx_ring_base[txr_next];
1715 		bzero((void *)tx_cmd, sizeof(q80_tx_cmd_t));
1716 	}
1717 
1718 	if (mp->m_pkthdr.csum_flags & CSUM_TSO) {
1719 
1720 		/* TSO : Copy the header in the following tx cmd descriptors */
1721 
1722 		txr_next = hw->tx_cntxt[txr_idx].txr_next;
1723 
1724 		tx_cmd = &hw->tx_cntxt[txr_idx].tx_ring_base[txr_next];
1725 		bzero((void *)tx_cmd, sizeof(q80_tx_cmd_t));
1726 
1727 		bytes = sizeof(q80_tx_cmd_t) - Q8_TX_CMD_TSO_ALIGN;
1728 		bytes = QL_MIN(bytes, hdr_len);
1729 
1730 		dst = (uint8_t *)tx_cmd + Q8_TX_CMD_TSO_ALIGN;
1731 
1732 		if (mp->m_flags & M_VLANTAG) {
1733 			/* first copy the src/dst MAC addresses */
1734 			bcopy(src, dst, (ETHER_ADDR_LEN * 2));
1735 			dst += (ETHER_ADDR_LEN * 2);
1736 			src += (ETHER_ADDR_LEN * 2);
1737 
1738 			*((uint16_t *)dst) = htons(ETHERTYPE_VLAN);
1739 			dst += 2;
1740 			*((uint16_t *)dst) = htons(mp->m_pkthdr.ether_vtag);
1741 			dst += 2;
1742 
1743 			/* bytes left in src header */
1744 			hdr_len -= ((ETHER_ADDR_LEN * 2) +
1745 					ETHER_VLAN_ENCAP_LEN);
1746 
1747 			/* bytes left in TxCmd Entry */
1748 			bytes -= ((ETHER_ADDR_LEN * 2) + ETHER_VLAN_ENCAP_LEN);
1749 
1750 
1751 			bcopy(src, dst, bytes);
1752 			src += bytes;
1753 			hdr_len -= bytes;
1754 		} else {
1755 			bcopy(src, dst, bytes);
1756 			src += bytes;
1757 			hdr_len -= bytes;
1758 		}
1759 
1760 		txr_next = hw->tx_cntxt[txr_idx].txr_next =
1761 				(hw->tx_cntxt[txr_idx].txr_next + 1) &
1762 					(NUM_TX_DESCRIPTORS - 1);
1763 		tx_cmd_count++;
1764 
1765 		while (hdr_len) {
1766 			tx_cmd = &hw->tx_cntxt[txr_idx].tx_ring_base[txr_next];
1767 			bzero((void *)tx_cmd, sizeof(q80_tx_cmd_t));
1768 
1769 			bytes = QL_MIN((sizeof(q80_tx_cmd_t)), hdr_len);
1770 
1771 			bcopy(src, tx_cmd, bytes);
1772 			src += bytes;
1773 			hdr_len -= bytes;
1774 
1775 			txr_next = hw->tx_cntxt[txr_idx].txr_next =
1776 				(hw->tx_cntxt[txr_idx].txr_next + 1) &
1777 					(NUM_TX_DESCRIPTORS - 1);
1778 			tx_cmd_count++;
1779 		}
1780 	}
1781 
1782 	hw->tx_cntxt[txr_idx].txr_free =
1783 		hw->tx_cntxt[txr_idx].txr_free - tx_cmd_count;
1784 
1785 	QL_UPDATE_TX_PRODUCER_INDEX(ha, hw->tx_cntxt[txr_idx].txr_next,\
1786 		txr_idx);
1787        	QL_DPRINT8(ha, (dev, "%s: return\n", __func__));
1788 
1789 	return (0);
1790 }
1791 
1792 
1793 static int
1794 qla_config_rss_ind_table(qla_host_t *ha)
1795 {
1796 	uint32_t i, count;
1797 	uint8_t rss_ind_tbl[16];
1798 
1799 	bzero(rss_ind_default_table, sizeof(rss_ind_default_table));
1800 
1801 
1802 	for (i = 0; i < 16; i++) {
1803 		rss_ind_tbl[i] = i % ha->hw.num_sds_rings;
1804 	}
1805 
1806 	for (i = 0; i <= Q8_RSS_IND_TBL_MAX_IDX ; i = i + 16) {
1807 
1808 		if ((i + 16) > Q8_RSS_IND_TBL_MAX_IDX) {
1809 			count = Q8_RSS_IND_TBL_MAX_IDX - i + 1;
1810 		} else {
1811 			count = 16;
1812 		}
1813 
1814 		if (qla_set_rss_ind_table(ha, i, count, ha->hw.rcv_cntxt_id,
1815 			rss_ind_tbl))
1816 			return (-1);
1817 	}
1818 
1819 	return (0);
1820 }
1821 
1822 /*
1823  * Name: ql_del_hw_if
1824  * Function: Destroys the hardware specific entities corresponding to an
1825  *	Ethernet Interface
1826  */
1827 void
1828 ql_del_hw_if(qla_host_t *ha)
1829 {
1830 
1831 	qla_del_rcv_cntxt(ha);
1832 	qla_del_xmt_cntxt(ha);
1833 
1834 	if (ha->hw.flags.init_intr_cnxt) {
1835 		qla_config_intr_cntxt(ha, ha->hw.num_sds_rings, 0);
1836 		ha->hw.flags.init_intr_cnxt = 0;
1837 	}
1838 }
1839 
1840 /*
1841  * Name: ql_init_hw_if
1842  * Function: Creates the hardware specific entities corresponding to an
1843  *	Ethernet Interface - Transmit and Receive Contexts. Sets the MAC Address
1844  *	corresponding to the interface. Enables LRO if allowed.
1845  */
1846 int
1847 ql_init_hw_if(qla_host_t *ha)
1848 {
1849 	device_t	dev;
1850 	uint32_t	i;
1851 	uint8_t		bcast_mac[6];
1852 	qla_rdesc_t	*rdesc;
1853 
1854 	dev = ha->pci_dev;
1855 
1856 	for (i = 0; i < ha->hw.num_sds_rings; i++) {
1857 		bzero(ha->hw.dma_buf.sds_ring[i].dma_b,
1858 			ha->hw.dma_buf.sds_ring[i].size);
1859 	}
1860 	ha->hw.mbx_intr_mask_offset = READ_REG32(ha, Q8_MBOX_INT_MASK_MSIX);
1861 
1862 	/* Use MSI-X vector 0; Enable Firmware Mailbox Interrupt */
1863 	WRITE_REG32(ha, Q8_MBOX_INT_ENABLE, BIT_2);
1864 	WRITE_REG32(ha, ha->hw.mbx_intr_mask_offset, 0x0);
1865 
1866 	qla_get_nic_partition(ha);
1867 
1868 	if (qla_config_intr_cntxt(ha, ha->hw.num_sds_rings, 1) == 0) {
1869 		ha->hw.flags.init_intr_cnxt = 1;
1870 	} else
1871 		return (-1);
1872 
1873 	if (ha->hw.mdump_init == 0) {
1874 		qla_minidump_init(ha);
1875 	}
1876 
1877 	/*
1878 	 * Create Receive Context
1879 	 */
1880 	if (qla_init_rcv_cntxt(ha)) {
1881 		return (-1);
1882 	}
1883 
1884 	for (i = 0; i < ha->hw.num_rds_rings; i++) {
1885 		rdesc = &ha->hw.rds[i];
1886 		rdesc->rx_next = NUM_RX_DESCRIPTORS - 2;
1887 		rdesc->rx_in = 0;
1888 		/* Update the RDS Producer Indices */
1889 		QL_UPDATE_RDS_PRODUCER_INDEX(ha, rdesc->prod_std,\
1890 			rdesc->rx_next);
1891 	}
1892 
1893 
1894 	/*
1895 	 * Create Transmit Context
1896 	 */
1897 	if (qla_init_xmt_cntxt(ha)) {
1898 		qla_del_rcv_cntxt(ha);
1899 		return (-1);
1900 	}
1901 	ha->hw.max_tx_segs = 0;
1902 
1903 	if (qla_config_mac_addr(ha, ha->hw.mac_addr, 1))
1904 		return(-1);
1905 
1906 	ha->hw.flags.unicast_mac = 1;
1907 
1908 	bcast_mac[0] = 0xFF; bcast_mac[1] = 0xFF; bcast_mac[2] = 0xFF;
1909 	bcast_mac[3] = 0xFF; bcast_mac[4] = 0xFF; bcast_mac[5] = 0xFF;
1910 
1911 	if (qla_config_mac_addr(ha, bcast_mac, 1))
1912 		return (-1);
1913 
1914 	ha->hw.flags.bcast_mac = 1;
1915 
1916 	/*
1917 	 * program any cached multicast addresses
1918 	 */
1919 	if (qla_hw_add_all_mcast(ha))
1920 		return (-1);
1921 
1922 	if (qla_config_rss(ha, ha->hw.rcv_cntxt_id))
1923 		return (-1);
1924 
1925 	if (qla_config_rss_ind_table(ha))
1926 		return (-1);
1927 
1928 	if (qla_config_intr_coalesce(ha, ha->hw.rcv_cntxt_id, 0))
1929 		return (-1);
1930 
1931 	if (qla_link_event_req(ha, ha->hw.rcv_cntxt_id))
1932 		return (-1);
1933 
1934 	if (qla_config_fw_lro(ha, ha->hw.rcv_cntxt_id))
1935 		return (-1);
1936 
1937 	for (i = 0; i < ha->hw.num_sds_rings; i++)
1938 		QL_ENABLE_INTERRUPTS(ha, i);
1939 
1940 	return (0);
1941 }
1942 
1943 static int
1944 qla_map_sds_to_rds(qla_host_t *ha)
1945 {
1946         device_t                dev = ha->pci_dev;
1947         q80_rq_map_sds_to_rds_t *map_rings;
1948         q80_rsp_add_rcv_rings_t *map_rings_rsp;
1949         uint32_t                i, err;
1950         qla_hw_t                *hw = &ha->hw;
1951 
1952         map_rings = (q80_rq_map_sds_to_rds_t *)ha->hw.mbox;
1953         bzero(map_rings, sizeof(q80_rq_map_sds_to_rds_t));
1954 
1955         map_rings->opcode = Q8_MBX_MAP_SDS_TO_RDS;
1956         map_rings->count_version = (sizeof (q80_rq_map_sds_to_rds_t) >> 2);
1957         map_rings->count_version |= Q8_MBX_CMD_VERSION;
1958 
1959         map_rings->cntxt_id = hw->rcv_cntxt_id;
1960         map_rings->num_rings = hw->num_sds_rings;
1961 
1962         for (i = 0; i < hw->num_sds_rings; i++) {
1963                 map_rings->sds_rds[i].sds_ring = i;
1964                 map_rings->sds_rds[i].rds_ring = i;
1965         }
1966 
1967         if (qla_mbx_cmd(ha, (uint32_t *)map_rings,
1968                 (sizeof (q80_rq_map_sds_to_rds_t) >> 2),
1969                 ha->hw.mbox, (sizeof(q80_rsp_add_rcv_rings_t) >> 2), 0)) {
1970                 device_printf(dev, "%s: failed0\n", __func__);
1971                 return (-1);
1972         }
1973 
1974         map_rings_rsp = (q80_rsp_add_rcv_rings_t *)ha->hw.mbox;
1975 
1976         err = Q8_MBX_RSP_STATUS(map_rings_rsp->regcnt_status);
1977 
1978         if (err) {
1979                 device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
1980                 return (-1);
1981         }
1982 
1983         return (0);
1984 }
1985 
1986 /*
1987  * Name: qla_init_rcv_cntxt
1988  * Function: Creates the Receive Context.
1989  */
1990 static int
1991 qla_init_rcv_cntxt(qla_host_t *ha)
1992 {
1993 	q80_rq_rcv_cntxt_t	*rcntxt;
1994 	q80_rsp_rcv_cntxt_t	*rcntxt_rsp;
1995 	q80_stat_desc_t		*sdesc;
1996 	int			i, j;
1997         qla_hw_t		*hw = &ha->hw;
1998 	device_t		dev;
1999 	uint32_t		err;
2000 	uint32_t		rcntxt_sds_rings;
2001 	uint32_t		rcntxt_rds_rings;
2002 
2003 	dev = ha->pci_dev;
2004 
2005 	/*
2006 	 * Create Receive Context
2007 	 */
2008 
2009 	for (i = 0; i < hw->num_sds_rings; i++) {
2010 		sdesc = (q80_stat_desc_t *)&hw->sds[i].sds_ring_base[0];
2011 
2012 		for (j = 0; j < NUM_STATUS_DESCRIPTORS; j++) {
2013 			sdesc->data[0] = 1ULL;
2014 			sdesc->data[1] = 1ULL;
2015 		}
2016 	}
2017 
2018 	rcntxt_sds_rings = hw->num_sds_rings;
2019 	if (hw->num_sds_rings > MAX_RCNTXT_SDS_RINGS)
2020 		rcntxt_sds_rings = MAX_RCNTXT_SDS_RINGS;
2021 
2022 	rcntxt_rds_rings = hw->num_rds_rings;
2023 
2024 	if (hw->num_rds_rings > MAX_RDS_RING_SETS)
2025 		rcntxt_rds_rings = MAX_RDS_RING_SETS;
2026 
2027 	rcntxt = (q80_rq_rcv_cntxt_t *)ha->hw.mbox;
2028 	bzero(rcntxt, (sizeof (q80_rq_rcv_cntxt_t)));
2029 
2030 	rcntxt->opcode = Q8_MBX_CREATE_RX_CNTXT;
2031 	rcntxt->count_version = (sizeof (q80_rq_rcv_cntxt_t) >> 2);
2032 	rcntxt->count_version |= Q8_MBX_CMD_VERSION;
2033 
2034 	rcntxt->cap0 = Q8_RCV_CNTXT_CAP0_BASEFW |
2035 			Q8_RCV_CNTXT_CAP0_LRO |
2036 			Q8_RCV_CNTXT_CAP0_HW_LRO |
2037 			Q8_RCV_CNTXT_CAP0_RSS |
2038 			Q8_RCV_CNTXT_CAP0_SGL_JUMBO |
2039 			Q8_RCV_CNTXT_CAP0_SGL_LRO;
2040 
2041 	if (ha->hw.num_rds_rings > 1) {
2042 		rcntxt->nrds_sets_rings = rcntxt_rds_rings | (1 << 5);
2043 		rcntxt->cap0 |= Q8_RCV_CNTXT_CAP0_MULTI_RDS;
2044 	} else
2045 		rcntxt->nrds_sets_rings = 0x1 | (1 << 5);
2046 
2047 	rcntxt->nsds_rings = rcntxt_sds_rings;
2048 
2049 	rcntxt->rds_producer_mode = Q8_RCV_CNTXT_RDS_PROD_MODE_UNIQUE;
2050 
2051 	rcntxt->rcv_vpid = 0;
2052 
2053 	for (i = 0; i <  rcntxt_sds_rings; i++) {
2054 		rcntxt->sds[i].paddr =
2055 			qla_host_to_le64(hw->dma_buf.sds_ring[i].dma_addr);
2056 		rcntxt->sds[i].size =
2057 			qla_host_to_le32(NUM_STATUS_DESCRIPTORS);
2058 		if (ha->msix_count == 2) {
2059 			rcntxt->sds[i].intr_id =
2060 				qla_host_to_le16(hw->intr_id[0]);
2061 			rcntxt->sds[i].intr_src_bit = qla_host_to_le16((i));
2062 		} else {
2063 			rcntxt->sds[i].intr_id =
2064 				qla_host_to_le16(hw->intr_id[i]);
2065 			rcntxt->sds[i].intr_src_bit = qla_host_to_le16(0);
2066 		}
2067 	}
2068 
2069 	for (i = 0; i <  rcntxt_rds_rings; i++) {
2070 		rcntxt->rds[i].paddr_std =
2071 			qla_host_to_le64(hw->dma_buf.rds_ring[i].dma_addr);
2072 		rcntxt->rds[i].std_bsize = qla_host_to_le64(MCLBYTES);
2073 		rcntxt->rds[i].std_nentries =
2074 			qla_host_to_le32(NUM_RX_DESCRIPTORS);
2075 	}
2076 
2077         if (qla_mbx_cmd(ha, (uint32_t *)rcntxt,
2078 		(sizeof (q80_rq_rcv_cntxt_t) >> 2),
2079                 ha->hw.mbox, (sizeof(q80_rsp_rcv_cntxt_t) >> 2), 0)) {
2080                 device_printf(dev, "%s: failed0\n", __func__);
2081                 return (-1);
2082         }
2083 
2084         rcntxt_rsp = (q80_rsp_rcv_cntxt_t *)ha->hw.mbox;
2085 
2086         err = Q8_MBX_RSP_STATUS(rcntxt_rsp->regcnt_status);
2087 
2088         if (err) {
2089                 device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
2090                 return (-1);
2091         }
2092 
2093 	for (i = 0; i <  rcntxt_sds_rings; i++) {
2094 		hw->sds[i].sds_consumer = rcntxt_rsp->sds_cons[i];
2095 	}
2096 
2097 	for (i = 0; i <  rcntxt_rds_rings; i++) {
2098 		hw->rds[i].prod_std = rcntxt_rsp->rds[i].prod_std;
2099 	}
2100 
2101 	hw->rcv_cntxt_id = rcntxt_rsp->cntxt_id;
2102 
2103 	ha->hw.flags.init_rx_cnxt = 1;
2104 
2105 	if (hw->num_sds_rings > MAX_RCNTXT_SDS_RINGS) {
2106 		err = qla_add_rcv_rings(ha, MAX_RCNTXT_SDS_RINGS);
2107 		if (err)
2108 			return -1;
2109 	}
2110 
2111         if (hw->num_rds_rings > 1) {
2112 		err = qla_map_sds_to_rds(ha);
2113 		if (err)
2114 			return -1;
2115 	}
2116 
2117 	return (0);
2118 }
2119 
2120 static int
2121 qla_add_rcv_rings(qla_host_t *ha, uint32_t sds_idx)
2122 {
2123 	device_t		dev = ha->pci_dev;
2124 	q80_rq_add_rcv_rings_t	*add_rcv;
2125 	q80_rsp_add_rcv_rings_t	*add_rcv_rsp;
2126 	uint32_t		i,j, err;
2127 	uint8_t			nsds;
2128         qla_hw_t		*hw = &ha->hw;
2129 
2130 	nsds = hw->num_sds_rings - MAX_RCNTXT_SDS_RINGS;
2131 
2132 	add_rcv = (q80_rq_add_rcv_rings_t *)ha->hw.mbox;
2133 	bzero(add_rcv, sizeof (q80_rq_add_rcv_rings_t));
2134 
2135 	add_rcv->opcode = Q8_MBX_ADD_RX_RINGS;
2136 	add_rcv->count_version = (sizeof (q80_rq_add_rcv_rings_t) >> 2);
2137 	add_rcv->count_version |= Q8_MBX_CMD_VERSION;
2138 
2139         if (hw->num_rds_rings > 1)
2140                 add_rcv->nrds_sets_rings = nsds | (1 << 5);
2141         else
2142                 add_rcv->nrds_sets_rings = 0;
2143 
2144 	add_rcv->nsds_rings = nsds;
2145 	add_rcv->cntxt_id = hw->rcv_cntxt_id;
2146 
2147         for (i = 0; i <  nsds; i++) {
2148 
2149 		j = i + sds_idx;
2150 
2151                 add_rcv->sds[i].paddr =
2152                         qla_host_to_le64(hw->dma_buf.sds_ring[j].dma_addr);
2153 
2154                 add_rcv->sds[i].size =
2155                         qla_host_to_le32(NUM_STATUS_DESCRIPTORS);
2156 
2157                 if (ha->msix_count == 2) {
2158                         add_rcv->sds[i].intr_id =
2159                                 qla_host_to_le16(hw->intr_id[0]);
2160                         add_rcv->sds[i].intr_src_bit = qla_host_to_le16(j);
2161                 } else {
2162                         add_rcv->sds[i].intr_id =
2163                                 qla_host_to_le16(hw->intr_id[j]);
2164                         add_rcv->sds[i].intr_src_bit = qla_host_to_le16(0);
2165                 }
2166 
2167         }
2168         for (i = 0; ((i <  nsds) && (hw->num_rds_rings > 1)); i++) {
2169                 j = i + sds_idx;
2170                 add_rcv->rds[i].paddr_std =
2171                         qla_host_to_le64(hw->dma_buf.rds_ring[j].dma_addr);
2172                 add_rcv->rds[i].std_bsize = qla_host_to_le64(MCLBYTES);
2173                 add_rcv->rds[i].std_nentries =
2174                         qla_host_to_le32(NUM_RX_DESCRIPTORS);
2175         }
2176 
2177 
2178         if (qla_mbx_cmd(ha, (uint32_t *)add_rcv,
2179 		(sizeof (q80_rq_add_rcv_rings_t) >> 2),
2180                 ha->hw.mbox, (sizeof(q80_rsp_add_rcv_rings_t) >> 2), 0)) {
2181                 device_printf(dev, "%s: failed0\n", __func__);
2182                 return (-1);
2183         }
2184 
2185         add_rcv_rsp = (q80_rsp_add_rcv_rings_t *)ha->hw.mbox;
2186 
2187         err = Q8_MBX_RSP_STATUS(add_rcv_rsp->regcnt_status);
2188 
2189         if (err) {
2190                 device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
2191                 return (-1);
2192         }
2193 
2194 	for (i = sds_idx; i < hw->num_sds_rings; i++) {
2195 		hw->sds[i].sds_consumer = add_rcv_rsp->sds_cons[(i - sds_idx)];
2196 	}
2197 	for (i = sds_idx; i < hw->num_rds_rings; i++) {
2198 		hw->rds[i].prod_std = add_rcv_rsp->rds[(i - sds_idx)].prod_std;
2199 	}
2200 	return (0);
2201 }
2202 
2203 /*
2204  * Name: qla_del_rcv_cntxt
2205  * Function: Destroys the Receive Context.
2206  */
2207 static void
2208 qla_del_rcv_cntxt(qla_host_t *ha)
2209 {
2210 	device_t			dev = ha->pci_dev;
2211 	q80_rcv_cntxt_destroy_t		*rcntxt;
2212 	q80_rcv_cntxt_destroy_rsp_t	*rcntxt_rsp;
2213 	uint32_t			err;
2214 	uint8_t				bcast_mac[6];
2215 
2216 	if (!ha->hw.flags.init_rx_cnxt)
2217 		return;
2218 
2219 	if (qla_hw_del_all_mcast(ha))
2220 		return;
2221 
2222 	if (ha->hw.flags.bcast_mac) {
2223 
2224 		bcast_mac[0] = 0xFF; bcast_mac[1] = 0xFF; bcast_mac[2] = 0xFF;
2225 		bcast_mac[3] = 0xFF; bcast_mac[4] = 0xFF; bcast_mac[5] = 0xFF;
2226 
2227 		if (qla_config_mac_addr(ha, bcast_mac, 0))
2228 			return;
2229 		ha->hw.flags.bcast_mac = 0;
2230 
2231 	}
2232 
2233 	if (ha->hw.flags.unicast_mac) {
2234 		if (qla_config_mac_addr(ha, ha->hw.mac_addr, 0))
2235 			return;
2236 		ha->hw.flags.unicast_mac = 0;
2237 	}
2238 
2239 	rcntxt = (q80_rcv_cntxt_destroy_t *)ha->hw.mbox;
2240 	bzero(rcntxt, (sizeof (q80_rcv_cntxt_destroy_t)));
2241 
2242 	rcntxt->opcode = Q8_MBX_DESTROY_RX_CNTXT;
2243 	rcntxt->count_version = (sizeof (q80_rcv_cntxt_destroy_t) >> 2);
2244 	rcntxt->count_version |= Q8_MBX_CMD_VERSION;
2245 
2246 	rcntxt->cntxt_id = ha->hw.rcv_cntxt_id;
2247 
2248         if (qla_mbx_cmd(ha, (uint32_t *)rcntxt,
2249 		(sizeof (q80_rcv_cntxt_destroy_t) >> 2),
2250                 ha->hw.mbox, (sizeof(q80_rcv_cntxt_destroy_rsp_t) >> 2), 0)) {
2251                 device_printf(dev, "%s: failed0\n", __func__);
2252                 return;
2253         }
2254         rcntxt_rsp = (q80_rcv_cntxt_destroy_rsp_t *)ha->hw.mbox;
2255 
2256         err = Q8_MBX_RSP_STATUS(rcntxt_rsp->regcnt_status);
2257 
2258         if (err) {
2259                 device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
2260         }
2261 
2262 	ha->hw.flags.init_rx_cnxt = 0;
2263 	return;
2264 }
2265 
2266 /*
2267  * Name: qla_init_xmt_cntxt
2268  * Function: Creates the Transmit Context.
2269  */
2270 static int
2271 qla_init_xmt_cntxt_i(qla_host_t *ha, uint32_t txr_idx)
2272 {
2273 	device_t		dev;
2274         qla_hw_t		*hw = &ha->hw;
2275 	q80_rq_tx_cntxt_t	*tcntxt;
2276 	q80_rsp_tx_cntxt_t	*tcntxt_rsp;
2277 	uint32_t		err;
2278 	qla_hw_tx_cntxt_t       *hw_tx_cntxt;
2279 
2280 	hw_tx_cntxt = &hw->tx_cntxt[txr_idx];
2281 
2282 	dev = ha->pci_dev;
2283 
2284 	/*
2285 	 * Create Transmit Context
2286 	 */
2287 	tcntxt = (q80_rq_tx_cntxt_t *)ha->hw.mbox;
2288 	bzero(tcntxt, (sizeof (q80_rq_tx_cntxt_t)));
2289 
2290 	tcntxt->opcode = Q8_MBX_CREATE_TX_CNTXT;
2291 	tcntxt->count_version = (sizeof (q80_rq_tx_cntxt_t) >> 2);
2292 	tcntxt->count_version |= Q8_MBX_CMD_VERSION;
2293 
2294 	tcntxt->cap0 = Q8_TX_CNTXT_CAP0_BASEFW | Q8_TX_CNTXT_CAP0_LSO;
2295 
2296 	tcntxt->ntx_rings = 1;
2297 
2298 	tcntxt->tx_ring[0].paddr =
2299 		qla_host_to_le64(hw_tx_cntxt->tx_ring_paddr);
2300 	tcntxt->tx_ring[0].tx_consumer =
2301 		qla_host_to_le64(hw_tx_cntxt->tx_cons_paddr);
2302 	tcntxt->tx_ring[0].nentries = qla_host_to_le16(NUM_TX_DESCRIPTORS);
2303 
2304 	tcntxt->tx_ring[0].intr_id = qla_host_to_le16(hw->intr_id[0]);
2305 	tcntxt->tx_ring[0].intr_src_bit = qla_host_to_le16(0);
2306 
2307 
2308 	hw_tx_cntxt->txr_free = NUM_TX_DESCRIPTORS;
2309 	hw_tx_cntxt->txr_next = hw_tx_cntxt->txr_comp = 0;
2310 
2311         if (qla_mbx_cmd(ha, (uint32_t *)tcntxt,
2312 		(sizeof (q80_rq_tx_cntxt_t) >> 2),
2313                 ha->hw.mbox,
2314 		(sizeof(q80_rsp_tx_cntxt_t) >> 2), 0)) {
2315                 device_printf(dev, "%s: failed0\n", __func__);
2316                 return (-1);
2317         }
2318         tcntxt_rsp = (q80_rsp_tx_cntxt_t *)ha->hw.mbox;
2319 
2320         err = Q8_MBX_RSP_STATUS(tcntxt_rsp->regcnt_status);
2321 
2322         if (err) {
2323                 device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
2324 		return -1;
2325         }
2326 
2327 	hw_tx_cntxt->tx_prod_reg = tcntxt_rsp->tx_ring[0].prod_index;
2328 	hw_tx_cntxt->tx_cntxt_id = tcntxt_rsp->tx_ring[0].cntxt_id;
2329 
2330 	return (0);
2331 }
2332 
2333 
2334 /*
2335  * Name: qla_del_xmt_cntxt
2336  * Function: Destroys the Transmit Context.
2337  */
2338 static int
2339 qla_del_xmt_cntxt_i(qla_host_t *ha, uint32_t txr_idx)
2340 {
2341 	device_t			dev = ha->pci_dev;
2342 	q80_tx_cntxt_destroy_t		*tcntxt;
2343 	q80_tx_cntxt_destroy_rsp_t	*tcntxt_rsp;
2344 	uint32_t			err;
2345 
2346 	tcntxt = (q80_tx_cntxt_destroy_t *)ha->hw.mbox;
2347 	bzero(tcntxt, (sizeof (q80_tx_cntxt_destroy_t)));
2348 
2349 	tcntxt->opcode = Q8_MBX_DESTROY_TX_CNTXT;
2350 	tcntxt->count_version = (sizeof (q80_tx_cntxt_destroy_t) >> 2);
2351 	tcntxt->count_version |= Q8_MBX_CMD_VERSION;
2352 
2353 	tcntxt->cntxt_id = ha->hw.tx_cntxt[txr_idx].tx_cntxt_id;
2354 
2355         if (qla_mbx_cmd(ha, (uint32_t *)tcntxt,
2356 		(sizeof (q80_tx_cntxt_destroy_t) >> 2),
2357                 ha->hw.mbox, (sizeof (q80_tx_cntxt_destroy_rsp_t) >> 2), 0)) {
2358                 device_printf(dev, "%s: failed0\n", __func__);
2359                 return (-1);
2360         }
2361         tcntxt_rsp = (q80_tx_cntxt_destroy_rsp_t *)ha->hw.mbox;
2362 
2363         err = Q8_MBX_RSP_STATUS(tcntxt_rsp->regcnt_status);
2364 
2365         if (err) {
2366                 device_printf(dev, "%s: failed1 [0x%08x]\n", __func__, err);
2367 		return (-1);
2368         }
2369 
2370 	return (0);
2371 }
2372 static void
2373 qla_del_xmt_cntxt(qla_host_t *ha)
2374 {
2375 	uint32_t i;
2376 
2377 	if (!ha->hw.flags.init_tx_cnxt)
2378 		return;
2379 
2380 	for (i = 0; i < ha->hw.num_tx_rings; i++) {
2381 		if (qla_del_xmt_cntxt_i(ha, i))
2382 			break;
2383 	}
2384 	ha->hw.flags.init_tx_cnxt = 0;
2385 }
2386 
2387 static int
2388 qla_init_xmt_cntxt(qla_host_t *ha)
2389 {
2390 	uint32_t i, j;
2391 
2392 	for (i = 0; i < ha->hw.num_tx_rings; i++) {
2393 		if (qla_init_xmt_cntxt_i(ha, i) != 0) {
2394 			for (j = 0; j < i; j++)
2395 				qla_del_xmt_cntxt_i(ha, j);
2396 			return (-1);
2397 		}
2398 	}
2399 	ha->hw.flags.init_tx_cnxt = 1;
2400 	return (0);
2401 }
2402 
2403 static int
2404 qla_hw_add_all_mcast(qla_host_t *ha)
2405 {
2406 	int i, nmcast;
2407 
2408 	nmcast = ha->hw.nmcast;
2409 
2410 	for (i = 0 ; ((i < Q8_MAX_NUM_MULTICAST_ADDRS) && nmcast); i++) {
2411 		if ((ha->hw.mcast[i].addr[0] != 0) ||
2412 			(ha->hw.mcast[i].addr[1] != 0) ||
2413 			(ha->hw.mcast[i].addr[2] != 0) ||
2414 			(ha->hw.mcast[i].addr[3] != 0) ||
2415 			(ha->hw.mcast[i].addr[4] != 0) ||
2416 			(ha->hw.mcast[i].addr[5] != 0)) {
2417 
2418 			if (qla_config_mac_addr(ha, ha->hw.mcast[i].addr, 1)) {
2419                 		device_printf(ha->pci_dev, "%s: failed\n",
2420 					__func__);
2421 				return (-1);
2422 			}
2423 
2424 			nmcast--;
2425 		}
2426 	}
2427 	return 0;
2428 }
2429 
2430 static int
2431 qla_hw_del_all_mcast(qla_host_t *ha)
2432 {
2433 	int i, nmcast;
2434 
2435 	nmcast = ha->hw.nmcast;
2436 
2437 	for (i = 0 ; ((i < Q8_MAX_NUM_MULTICAST_ADDRS) && nmcast); i++) {
2438 		if ((ha->hw.mcast[i].addr[0] != 0) ||
2439 			(ha->hw.mcast[i].addr[1] != 0) ||
2440 			(ha->hw.mcast[i].addr[2] != 0) ||
2441 			(ha->hw.mcast[i].addr[3] != 0) ||
2442 			(ha->hw.mcast[i].addr[4] != 0) ||
2443 			(ha->hw.mcast[i].addr[5] != 0)) {
2444 
2445 			if (qla_config_mac_addr(ha, ha->hw.mcast[i].addr, 0))
2446 				return (-1);
2447 
2448 			nmcast--;
2449 		}
2450 	}
2451 	return 0;
2452 }
2453 
2454 static int
2455 qla_hw_add_mcast(qla_host_t *ha, uint8_t *mta)
2456 {
2457 	int i;
2458 
2459 	for (i = 0; i < Q8_MAX_NUM_MULTICAST_ADDRS; i++) {
2460 
2461 		if (QL_MAC_CMP(ha->hw.mcast[i].addr, mta) == 0)
2462 			return 0; /* its been already added */
2463 	}
2464 
2465 	for (i = 0; i < Q8_MAX_NUM_MULTICAST_ADDRS; i++) {
2466 
2467 		if ((ha->hw.mcast[i].addr[0] == 0) &&
2468 			(ha->hw.mcast[i].addr[1] == 0) &&
2469 			(ha->hw.mcast[i].addr[2] == 0) &&
2470 			(ha->hw.mcast[i].addr[3] == 0) &&
2471 			(ha->hw.mcast[i].addr[4] == 0) &&
2472 			(ha->hw.mcast[i].addr[5] == 0)) {
2473 
2474 			if (qla_config_mac_addr(ha, mta, 1))
2475 				return (-1);
2476 
2477 			bcopy(mta, ha->hw.mcast[i].addr, Q8_MAC_ADDR_LEN);
2478 			ha->hw.nmcast++;
2479 
2480 			return 0;
2481 		}
2482 	}
2483 	return 0;
2484 }
2485 
2486 static int
2487 qla_hw_del_mcast(qla_host_t *ha, uint8_t *mta)
2488 {
2489 	int i;
2490 
2491 	for (i = 0; i < Q8_MAX_NUM_MULTICAST_ADDRS; i++) {
2492 		if (QL_MAC_CMP(ha->hw.mcast[i].addr, mta) == 0) {
2493 
2494 			if (qla_config_mac_addr(ha, mta, 0))
2495 				return (-1);
2496 
2497 			ha->hw.mcast[i].addr[0] = 0;
2498 			ha->hw.mcast[i].addr[1] = 0;
2499 			ha->hw.mcast[i].addr[2] = 0;
2500 			ha->hw.mcast[i].addr[3] = 0;
2501 			ha->hw.mcast[i].addr[4] = 0;
2502 			ha->hw.mcast[i].addr[5] = 0;
2503 
2504 			ha->hw.nmcast--;
2505 
2506 			return 0;
2507 		}
2508 	}
2509 	return 0;
2510 }
2511 
2512 /*
2513  * Name: ql_hw_set_multi
2514  * Function: Sets the Multicast Addresses provided the host O.S into the
2515  *	hardware (for the given interface)
2516  */
2517 int
2518 ql_hw_set_multi(qla_host_t *ha, uint8_t *mcast, uint32_t mcnt,
2519 	uint32_t add_mac)
2520 {
2521 	int i;
2522 	uint8_t *mta = mcast;
2523 	int ret = 0;
2524 
2525 	for (i = 0; i < mcnt; i++) {
2526 		if (add_mac) {
2527 			ret = qla_hw_add_mcast(ha, mta);
2528 			if (ret)
2529 				break;
2530 		} else {
2531 			ret = qla_hw_del_mcast(ha, mta);
2532 			if (ret)
2533 				break;
2534 		}
2535 
2536 		mta += Q8_MAC_ADDR_LEN;
2537 	}
2538 	return (ret);
2539 }
2540 
2541 /*
2542  * Name: qla_hw_tx_done_locked
2543  * Function: Handle Transmit Completions
2544  */
2545 static void
2546 qla_hw_tx_done_locked(qla_host_t *ha, uint32_t txr_idx)
2547 {
2548 	qla_tx_buf_t *txb;
2549         qla_hw_t *hw = &ha->hw;
2550 	uint32_t comp_idx, comp_count = 0;
2551 	qla_hw_tx_cntxt_t *hw_tx_cntxt;
2552 
2553 	hw_tx_cntxt = &hw->tx_cntxt[txr_idx];
2554 
2555 	/* retrieve index of last entry in tx ring completed */
2556 	comp_idx = qla_le32_to_host(*(hw_tx_cntxt->tx_cons));
2557 
2558 	while (comp_idx != hw_tx_cntxt->txr_comp) {
2559 
2560 		txb = &ha->tx_ring[txr_idx].tx_buf[hw_tx_cntxt->txr_comp];
2561 
2562 		hw_tx_cntxt->txr_comp++;
2563 		if (hw_tx_cntxt->txr_comp == NUM_TX_DESCRIPTORS)
2564 			hw_tx_cntxt->txr_comp = 0;
2565 
2566 		comp_count++;
2567 
2568 		if (txb->m_head) {
2569 			if_inc_counter(ha->ifp, IFCOUNTER_OPACKETS, 1);
2570 
2571 			bus_dmamap_sync(ha->tx_tag, txb->map,
2572 				BUS_DMASYNC_POSTWRITE);
2573 			bus_dmamap_unload(ha->tx_tag, txb->map);
2574 			m_freem(txb->m_head);
2575 
2576 			txb->m_head = NULL;
2577 		}
2578 	}
2579 
2580 	hw_tx_cntxt->txr_free += comp_count;
2581 	return;
2582 }
2583 
2584 /*
2585  * Name: ql_hw_tx_done
2586  * Function: Handle Transmit Completions
2587  */
2588 void
2589 ql_hw_tx_done(qla_host_t *ha)
2590 {
2591 	int i;
2592 	uint32_t flag = 0;
2593 
2594 	if (!mtx_trylock(&ha->tx_lock)) {
2595        		QL_DPRINT8(ha, (ha->pci_dev,
2596 			"%s: !mtx_trylock(&ha->tx_lock)\n", __func__));
2597 		return;
2598 	}
2599 	for (i = 0; i < ha->hw.num_tx_rings; i++) {
2600 		qla_hw_tx_done_locked(ha, i);
2601 		if (ha->hw.tx_cntxt[i].txr_free <= (NUM_TX_DESCRIPTORS >> 1))
2602 			flag = 1;
2603 	}
2604 
2605 	if (!flag)
2606 		ha->ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
2607 
2608 	QLA_TX_UNLOCK(ha);
2609 	return;
2610 }
2611 
2612 void
2613 ql_update_link_state(qla_host_t *ha)
2614 {
2615 	uint32_t link_state;
2616 	uint32_t prev_link_state;
2617 
2618 	if (!(ha->ifp->if_drv_flags & IFF_DRV_RUNNING)) {
2619 		ha->hw.link_up = 0;
2620 		return;
2621 	}
2622 	link_state = READ_REG32(ha, Q8_LINK_STATE);
2623 
2624 	prev_link_state =  ha->hw.link_up;
2625 
2626 	if (ha->pci_func == 0)
2627 		ha->hw.link_up = (((link_state & 0xF) == 1)? 1 : 0);
2628 	else
2629 		ha->hw.link_up = ((((link_state >> 4)& 0xF) == 1)? 1 : 0);
2630 
2631 	if (prev_link_state !=  ha->hw.link_up) {
2632 		if (ha->hw.link_up) {
2633 			if_link_state_change(ha->ifp, LINK_STATE_UP);
2634 		} else {
2635 			if_link_state_change(ha->ifp, LINK_STATE_DOWN);
2636 		}
2637 	}
2638 	return;
2639 }
2640 
2641 void
2642 ql_hw_stop_rcv(qla_host_t *ha)
2643 {
2644 	int i, done, count = 100;
2645 
2646 	while (count--) {
2647 		done = 1;
2648 		for (i = 0; i < ha->hw.num_sds_rings; i++) {
2649 			if (ha->hw.sds[i].rcv_active)
2650 				done = 0;
2651 		}
2652 		if (done)
2653 			break;
2654 		else
2655 			qla_mdelay(__func__, 10);
2656 	}
2657 	if (!count)
2658 		device_printf(ha->pci_dev, "%s: Counter expired.\n", __func__);
2659 
2660 	return;
2661 }
2662 
2663 int
2664 ql_hw_check_health(qla_host_t *ha)
2665 {
2666 	uint32_t val;
2667 
2668 	ha->hw.health_count++;
2669 
2670 	if (ha->hw.health_count < 1000)
2671 		return 0;
2672 
2673 	ha->hw.health_count = 0;
2674 
2675 	val = READ_REG32(ha, Q8_ASIC_TEMPERATURE);
2676 
2677 	if (((val & 0xFFFF) == 2) || ((val & 0xFFFF) == 3) ||
2678 		(QL_ERR_INJECT(ha, INJCT_TEMPERATURE_FAILURE))) {
2679 		device_printf(ha->pci_dev, "%s: Temperature Alert [0x%08x]\n",
2680 			__func__, val);
2681 		return -1;
2682 	}
2683 
2684 	val = READ_REG32(ha, Q8_FIRMWARE_HEARTBEAT);
2685 
2686 	if ((val != ha->hw.hbeat_value) &&
2687 		(!(QL_ERR_INJECT(ha, INJCT_TEMPERATURE_FAILURE)))) {
2688 		ha->hw.hbeat_value = val;
2689 		return 0;
2690 	}
2691 	device_printf(ha->pci_dev, "%s: Heartbeat Failue [0x%08x]\n",
2692 		__func__, val);
2693 
2694 	return -1;
2695 }
2696 
2697 static int
2698 qla_get_minidump_tmplt_size(qla_host_t *ha, uint32_t *size)
2699 {
2700 	uint32_t			err;
2701 	device_t			dev = ha->pci_dev;
2702 	q80_config_md_templ_size_t	*md_size;
2703 	q80_config_md_templ_size_rsp_t	*md_size_rsp;
2704 
2705 	md_size = (q80_config_md_templ_size_t *) ha->hw.mbox;
2706 	bzero(md_size, sizeof(q80_config_md_templ_size_t));
2707 
2708 	md_size->opcode = Q8_MBX_GET_MINIDUMP_TMPLT_SIZE;
2709 	md_size->count_version = (sizeof (q80_config_md_templ_size_t) >> 2);
2710 	md_size->count_version |= Q8_MBX_CMD_VERSION;
2711 
2712 	if (qla_mbx_cmd(ha, (uint32_t *) md_size,
2713 		(sizeof(q80_config_md_templ_size_t) >> 2), ha->hw.mbox,
2714 		(sizeof(q80_config_md_templ_size_rsp_t) >> 2), 0)) {
2715 
2716 		device_printf(dev, "%s: failed\n", __func__);
2717 
2718 		return (-1);
2719 	}
2720 
2721 	md_size_rsp = (q80_config_md_templ_size_rsp_t *) ha->hw.mbox;
2722 
2723 	err = Q8_MBX_RSP_STATUS(md_size_rsp->regcnt_status);
2724 
2725         if (err) {
2726 		device_printf(dev, "%s: failed [0x%08x]\n", __func__, err);
2727 		return(-1);
2728         }
2729 
2730 	*size = md_size_rsp->templ_size;
2731 
2732 	return (0);
2733 }
2734 
2735 static int
2736 qla_get_minidump_template(qla_host_t *ha)
2737 {
2738 	uint32_t			err;
2739 	device_t			dev = ha->pci_dev;
2740 	q80_config_md_templ_cmd_t	*md_templ;
2741 	q80_config_md_templ_cmd_rsp_t	*md_templ_rsp;
2742 
2743 	md_templ = (q80_config_md_templ_cmd_t *) ha->hw.mbox;
2744 	bzero(md_templ, (sizeof (q80_config_md_templ_cmd_t)));
2745 
2746 	md_templ->opcode = Q8_MBX_GET_MINIDUMP_TMPLT;
2747 	md_templ->count_version = ( sizeof(q80_config_md_templ_cmd_t) >> 2);
2748 	md_templ->count_version |= Q8_MBX_CMD_VERSION;
2749 
2750 	md_templ->buf_addr = ha->hw.dma_buf.minidump.dma_addr;
2751 	md_templ->buff_size = ha->hw.dma_buf.minidump.size;
2752 
2753 	if (qla_mbx_cmd(ha, (uint32_t *) md_templ,
2754 		(sizeof(q80_config_md_templ_cmd_t) >> 2),
2755 		 ha->hw.mbox,
2756 		(sizeof(q80_config_md_templ_cmd_rsp_t) >> 2), 0)) {
2757 
2758 		device_printf(dev, "%s: failed\n", __func__);
2759 
2760 		return (-1);
2761 	}
2762 
2763 	md_templ_rsp = (q80_config_md_templ_cmd_rsp_t *) ha->hw.mbox;
2764 
2765 	err = Q8_MBX_RSP_STATUS(md_templ_rsp->regcnt_status);
2766 
2767 	if (err) {
2768 		device_printf(dev, "%s: failed [0x%08x]\n", __func__, err);
2769 		return (-1);
2770 	}
2771 
2772 	return (0);
2773 
2774 }
2775 
2776 static int
2777 qla_minidump_init(qla_host_t *ha)
2778 {
2779 	int		ret;
2780 	uint32_t	template_size = 0;
2781 	device_t	dev = ha->pci_dev;
2782 
2783 	/*
2784 	 * Get Minidump Template Size
2785  	 */
2786 	ret = qla_get_minidump_tmplt_size(ha, &template_size);
2787 
2788 	if (ret || (template_size == 0)) {
2789 		device_printf(dev, "%s: failed [%d, %d]\n", __func__, ret,
2790 			template_size);
2791 		return (-1);
2792 	}
2793 
2794 	/*
2795 	 * Allocate Memory for Minidump Template
2796 	 */
2797 
2798 	ha->hw.dma_buf.minidump.alignment = 8;
2799 	ha->hw.dma_buf.minidump.size = template_size;
2800 
2801 	if (ql_alloc_dmabuf(ha, &ha->hw.dma_buf.minidump)) {
2802 
2803 		device_printf(dev, "%s: minidump dma alloc failed\n", __func__);
2804 
2805 		return (-1);
2806 	}
2807 	ha->hw.dma_buf.flags.minidump = 1;
2808 
2809 	/*
2810 	 * Retrieve Minidump Template
2811 	 */
2812 	ret = qla_get_minidump_template(ha);
2813 
2814 	if (ret) {
2815 		qla_minidump_free(ha);
2816 	} else {
2817 		ha->hw.mdump_init = 1;
2818 	}
2819 
2820 	return (ret);
2821 }
2822 
2823 
2824 static void
2825 qla_minidump_free(qla_host_t *ha)
2826 {
2827 	ha->hw.mdump_init = 0;
2828 	if (ha->hw.dma_buf.flags.minidump) {
2829 		ha->hw.dma_buf.flags.minidump = 0;
2830 		ql_free_dmabuf(ha, &ha->hw.dma_buf.minidump);
2831 	}
2832 	return;
2833 }
2834 
2835 void
2836 ql_minidump(qla_host_t *ha)
2837 {
2838 	uint32_t delay = 6000;
2839 
2840 	if (!ha->hw.mdump_init)
2841 		return;
2842 
2843 	if (!ha->hw.mdump_active)
2844 		return;
2845 
2846 	if (ha->hw.mdump_active == 1) {
2847 		ha->hw.mdump_start_seq_index = ql_stop_sequence(ha);
2848 		ha->hw.mdump_start = 1;
2849 	}
2850 
2851 	while (delay-- && ha->hw.mdump_active) {
2852 		qla_mdelay(__func__, 100);
2853 	}
2854 	ha->hw.mdump_start = 0;
2855 	ql_start_sequence(ha, ha->hw.mdump_start_seq_index);
2856 
2857 	return;
2858 }
2859