xref: /freebsd/sys/dev/qlxgbe/ql_isr.c (revision f4b37ed0f8b307b1f3f0f630ca725d68f1dff30d)
1 /*
2  * Copyright (c) 2013-2016 Qlogic Corporation
3  * All rights reserved.
4  *
5  *  Redistribution and use in source and binary forms, with or without
6  *  modification, are permitted provided that the following conditions
7  *  are met:
8  *
9  *  1. Redistributions of source code must retain the above copyright
10  *     notice, this list of conditions and the following disclaimer.
11  *  2. Redistributions in binary form must reproduce the above copyright
12  *     notice, this list of conditions and the following disclaimer in the
13  *     documentation and/or other materials provided with the distribution.
14  *
15  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16  *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19  *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20  *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21  *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22  *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23  *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24  *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25  *  POSSIBILITY OF SUCH DAMAGE.
26  */
27 
28 /*
29  * File: ql_isr.c
30  * Author : David C Somayajulu, Qlogic Corporation, Aliso Viejo, CA 92656.
31  */
32 
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
35 
36 
37 #include "ql_os.h"
38 #include "ql_hw.h"
39 #include "ql_def.h"
40 #include "ql_inline.h"
41 #include "ql_ver.h"
42 #include "ql_glbl.h"
43 #include "ql_dbg.h"
44 
45 static void qla_replenish_normal_rx(qla_host_t *ha, qla_sds_t *sdsp,
46 		uint32_t r_idx);
47 
48 static void
49 qla_rcv_error(qla_host_t *ha)
50 {
51 	ha->flags.stop_rcv = 1;
52 	ha->qla_initiate_recovery = 1;
53 }
54 
55 
56 /*
57  * Name: qla_rx_intr
58  * Function: Handles normal ethernet frames received
59  */
60 static void
61 qla_rx_intr(qla_host_t *ha, qla_sgl_rcv_t *sgc, uint32_t sds_idx)
62 {
63 	qla_rx_buf_t		*rxb;
64 	struct mbuf		*mp = NULL, *mpf = NULL, *mpl = NULL;
65 	struct ifnet		*ifp = ha->ifp;
66 	qla_sds_t		*sdsp;
67 	struct ether_vlan_header *eh;
68 	uint32_t		i, rem_len = 0;
69 	uint32_t		r_idx = 0;
70 	qla_rx_ring_t		*rx_ring;
71 
72 	if (ha->hw.num_rds_rings > 1)
73 		r_idx = sds_idx;
74 
75 	ha->hw.rds[r_idx].count++;
76 
77 	sdsp = &ha->hw.sds[sds_idx];
78 	rx_ring = &ha->rx_ring[r_idx];
79 
80 	for (i = 0; i < sgc->num_handles; i++) {
81 		rxb = &rx_ring->rx_buf[sgc->handle[i] & 0x7FFF];
82 
83 		QL_ASSERT(ha, (rxb != NULL),
84 			("%s: [sds_idx]=[%d] rxb != NULL\n", __func__,\
85 			sds_idx));
86 
87 		if ((rxb == NULL) || QL_ERR_INJECT(ha, INJCT_RX_RXB_INVAL)) {
88 			/* log the error */
89 			device_printf(ha->pci_dev,
90 				"%s invalid rxb[%d, %d, 0x%04x]\n",
91 				__func__, sds_idx, i, sgc->handle[i]);
92 			qla_rcv_error(ha);
93 			return;
94 		}
95 
96 		mp = rxb->m_head;
97 		if (i == 0)
98 			mpf = mp;
99 
100 		QL_ASSERT(ha, (mp != NULL),
101 			("%s: [sds_idx]=[%d] mp != NULL\n", __func__,\
102 			sds_idx));
103 
104 		bus_dmamap_sync(ha->rx_tag, rxb->map, BUS_DMASYNC_POSTREAD);
105 
106 		rxb->m_head = NULL;
107 		rxb->next = sdsp->rxb_free;
108 		sdsp->rxb_free = rxb;
109 		sdsp->rx_free++;
110 
111 		if ((mp == NULL) || QL_ERR_INJECT(ha, INJCT_RX_MP_NULL)) {
112 			/* log the error */
113 			device_printf(ha->pci_dev,
114 				"%s mp  == NULL [%d, %d, 0x%04x]\n",
115 				__func__, sds_idx, i, sgc->handle[i]);
116 			qla_rcv_error(ha);
117 			return;
118 		}
119 
120 		if (i == 0) {
121 			mpl = mpf = mp;
122 			mp->m_flags |= M_PKTHDR;
123 			mp->m_pkthdr.len = sgc->pkt_length;
124 			mp->m_pkthdr.rcvif = ifp;
125 			rem_len = mp->m_pkthdr.len;
126 		} else {
127 			mp->m_flags &= ~M_PKTHDR;
128 			mpl->m_next = mp;
129 			mpl = mp;
130 			rem_len = rem_len - mp->m_len;
131 		}
132 	}
133 
134 	mpl->m_len = rem_len;
135 
136 	eh = mtod(mpf, struct ether_vlan_header *);
137 
138 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
139 		uint32_t *data = (uint32_t *)eh;
140 
141 		mpf->m_pkthdr.ether_vtag = ntohs(eh->evl_tag);
142 		mpf->m_flags |= M_VLANTAG;
143 
144 		*(data + 3) = *(data + 2);
145 		*(data + 2) = *(data + 1);
146 		*(data + 1) = *data;
147 
148 		m_adj(mpf, ETHER_VLAN_ENCAP_LEN);
149 	}
150 
151 	if (sgc->chksum_status == Q8_STAT_DESC_STATUS_CHKSUM_OK) {
152 		mpf->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID |
153 			CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
154 		mpf->m_pkthdr.csum_data = 0xFFFF;
155 	} else {
156 		mpf->m_pkthdr.csum_flags = 0;
157 	}
158 
159 	if_inc_counter(ifp, IFCOUNTER_IPACKETS, 1);
160 
161 	mpf->m_pkthdr.flowid = sgc->rss_hash;
162 	M_HASHTYPE_SET(mpf, M_HASHTYPE_OPAQUE);
163 
164 	(*ifp->if_input)(ifp, mpf);
165 
166 	if (sdsp->rx_free > ha->std_replenish)
167 		qla_replenish_normal_rx(ha, sdsp, r_idx);
168 
169 	return;
170 }
171 
172 #define QLA_TCP_HDR_SIZE        20
173 #define QLA_TCP_TS_OPTION_SIZE  12
174 
175 /*
176  * Name: qla_lro_intr
177  * Function: Handles normal ethernet frames received
178  */
179 static int
180 qla_lro_intr(qla_host_t *ha, qla_sgl_lro_t *sgc, uint32_t sds_idx)
181 {
182 	qla_rx_buf_t *rxb;
183 	struct mbuf *mp = NULL, *mpf = NULL, *mpl = NULL;
184 	struct ifnet *ifp = ha->ifp;
185 	qla_sds_t *sdsp;
186 	struct ether_vlan_header *eh;
187 	uint32_t i, rem_len = 0, pkt_length, iplen;
188 	struct tcphdr *th;
189 	struct ip *ip = NULL;
190 	struct ip6_hdr *ip6 = NULL;
191 	uint16_t etype;
192 	uint32_t r_idx = 0;
193 	qla_rx_ring_t *rx_ring;
194 
195 	if (ha->hw.num_rds_rings > 1)
196 		r_idx = sds_idx;
197 
198 	ha->hw.rds[r_idx].count++;
199 
200 	rx_ring = &ha->rx_ring[r_idx];
201 
202 	ha->lro_pkt_count++;
203 
204 	sdsp = &ha->hw.sds[sds_idx];
205 
206 	pkt_length = sgc->payload_length + sgc->l4_offset;
207 
208 	if (sgc->flags & Q8_LRO_COMP_TS) {
209 		pkt_length += QLA_TCP_HDR_SIZE + QLA_TCP_TS_OPTION_SIZE;
210 	} else {
211 		pkt_length += QLA_TCP_HDR_SIZE;
212 	}
213 	ha->lro_bytes += pkt_length;
214 
215 	for (i = 0; i < sgc->num_handles; i++) {
216 		rxb = &rx_ring->rx_buf[sgc->handle[i] & 0x7FFF];
217 
218 		QL_ASSERT(ha, (rxb != NULL),
219 			("%s: [sds_idx]=[%d] rxb != NULL\n", __func__,\
220 			sds_idx));
221 
222 		if ((rxb == NULL) || QL_ERR_INJECT(ha, INJCT_LRO_RXB_INVAL)) {
223 			/* log the error */
224 			device_printf(ha->pci_dev,
225 				"%s invalid rxb[%d, %d, 0x%04x]\n",
226 				__func__, sds_idx, i, sgc->handle[i]);
227 			qla_rcv_error(ha);
228 			return (0);
229 		}
230 
231 		mp = rxb->m_head;
232 		if (i == 0)
233 			mpf = mp;
234 
235 		QL_ASSERT(ha, (mp != NULL),
236 			("%s: [sds_idx]=[%d] mp != NULL\n", __func__,\
237 			sds_idx));
238 
239 		bus_dmamap_sync(ha->rx_tag, rxb->map, BUS_DMASYNC_POSTREAD);
240 
241 		rxb->m_head = NULL;
242 		rxb->next = sdsp->rxb_free;
243 		sdsp->rxb_free = rxb;
244 		sdsp->rx_free++;
245 
246 		if ((mp == NULL) || QL_ERR_INJECT(ha, INJCT_LRO_MP_NULL)) {
247 			/* log the error */
248 			device_printf(ha->pci_dev,
249 				"%s mp  == NULL [%d, %d, 0x%04x]\n",
250 				__func__, sds_idx, i, sgc->handle[i]);
251 			qla_rcv_error(ha);
252 			return (0);
253 		}
254 
255 		if (i == 0) {
256 			mpl = mpf = mp;
257 			mp->m_flags |= M_PKTHDR;
258 			mp->m_pkthdr.len = pkt_length;
259 			mp->m_pkthdr.rcvif = ifp;
260 			rem_len = mp->m_pkthdr.len;
261 		} else {
262 			mp->m_flags &= ~M_PKTHDR;
263 			mpl->m_next = mp;
264 			mpl = mp;
265 			rem_len = rem_len - mp->m_len;
266 		}
267 	}
268 
269 	mpl->m_len = rem_len;
270 
271 	th = (struct tcphdr *)(mpf->m_data + sgc->l4_offset);
272 
273 	if (sgc->flags & Q8_LRO_COMP_PUSH_BIT)
274 		th->th_flags |= TH_PUSH;
275 
276 	m_adj(mpf, sgc->l2_offset);
277 
278 	eh = mtod(mpf, struct ether_vlan_header *);
279 
280 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
281 		uint32_t *data = (uint32_t *)eh;
282 
283 		mpf->m_pkthdr.ether_vtag = ntohs(eh->evl_tag);
284 		mpf->m_flags |= M_VLANTAG;
285 
286 		*(data + 3) = *(data + 2);
287 		*(data + 2) = *(data + 1);
288 		*(data + 1) = *data;
289 
290 		m_adj(mpf, ETHER_VLAN_ENCAP_LEN);
291 
292 		etype = ntohs(eh->evl_proto);
293 	} else {
294 		etype = ntohs(eh->evl_encap_proto);
295 	}
296 
297 	if (etype == ETHERTYPE_IP) {
298 		ip = (struct ip *)(mpf->m_data + ETHER_HDR_LEN);
299 
300 		iplen = (ip->ip_hl << 2) + (th->th_off << 2) +
301 				sgc->payload_length;
302 
303                 ip->ip_len = htons(iplen);
304 
305 		ha->ipv4_lro++;
306 	} else if (etype == ETHERTYPE_IPV6) {
307 		ip6 = (struct ip6_hdr *)(mpf->m_data + ETHER_HDR_LEN);
308 
309 		iplen = (th->th_off << 2) + sgc->payload_length;
310 
311 		ip6->ip6_plen = htons(iplen);
312 
313 		ha->ipv6_lro++;
314 	} else {
315 		m_freem(mpf);
316 
317 		if (sdsp->rx_free > ha->std_replenish)
318 			qla_replenish_normal_rx(ha, sdsp, r_idx);
319 		return 0;
320 	}
321 
322 	mpf->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID |
323 					CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
324 	mpf->m_pkthdr.csum_data = 0xFFFF;
325 
326 	mpf->m_pkthdr.flowid = sgc->rss_hash;
327 	M_HASHTYPE_SET(mpf, M_HASHTYPE_OPAQUE);
328 
329 	if_inc_counter(ifp, IFCOUNTER_IPACKETS, 1);
330 
331 	(*ifp->if_input)(ifp, mpf);
332 
333 	if (sdsp->rx_free > ha->std_replenish)
334 		qla_replenish_normal_rx(ha, sdsp, r_idx);
335 
336 	return (0);
337 }
338 
339 static int
340 qla_rcv_cont_sds(qla_host_t *ha, uint32_t sds_idx, uint32_t comp_idx,
341 	uint32_t dcount, uint16_t *handle, uint16_t *nhandles)
342 {
343 	uint32_t i;
344 	uint16_t num_handles;
345 	q80_stat_desc_t *sdesc;
346 	uint32_t opcode;
347 
348 	*nhandles = 0;
349 	dcount--;
350 
351 	for (i = 0; i < dcount; i++) {
352 		comp_idx = (comp_idx + 1) & (NUM_STATUS_DESCRIPTORS-1);
353 		sdesc = (q80_stat_desc_t *)
354 				&ha->hw.sds[sds_idx].sds_ring_base[comp_idx];
355 
356 		opcode = Q8_STAT_DESC_OPCODE((sdesc->data[1]));
357 
358 		if (!opcode) {
359 			device_printf(ha->pci_dev, "%s: opcode=0 %p %p\n",
360 				__func__, (void *)sdesc->data[0],
361 				(void *)sdesc->data[1]);
362 			return -1;
363 		}
364 
365 		num_handles = Q8_SGL_STAT_DESC_NUM_HANDLES((sdesc->data[1]));
366 		if (!num_handles) {
367 			device_printf(ha->pci_dev, "%s: opcode=0 %p %p\n",
368 				__func__, (void *)sdesc->data[0],
369 				(void *)sdesc->data[1]);
370 			return -1;
371 		}
372 
373 		if (QL_ERR_INJECT(ha, INJCT_NUM_HNDLE_INVALID))
374 			num_handles = -1;
375 
376 		switch (num_handles) {
377 
378 		case 1:
379 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
380 			break;
381 
382 		case 2:
383 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
384 			*handle++ = Q8_SGL_STAT_DESC_HANDLE2((sdesc->data[0]));
385 			break;
386 
387 		case 3:
388 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
389 			*handle++ = Q8_SGL_STAT_DESC_HANDLE2((sdesc->data[0]));
390 			*handle++ = Q8_SGL_STAT_DESC_HANDLE3((sdesc->data[0]));
391 			break;
392 
393 		case 4:
394 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
395 			*handle++ = Q8_SGL_STAT_DESC_HANDLE2((sdesc->data[0]));
396 			*handle++ = Q8_SGL_STAT_DESC_HANDLE3((sdesc->data[0]));
397 			*handle++ = Q8_SGL_STAT_DESC_HANDLE4((sdesc->data[0]));
398 			break;
399 
400 		case 5:
401 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
402 			*handle++ = Q8_SGL_STAT_DESC_HANDLE2((sdesc->data[0]));
403 			*handle++ = Q8_SGL_STAT_DESC_HANDLE3((sdesc->data[0]));
404 			*handle++ = Q8_SGL_STAT_DESC_HANDLE4((sdesc->data[0]));
405 			*handle++ = Q8_SGL_STAT_DESC_HANDLE5((sdesc->data[1]));
406 			break;
407 
408 		case 6:
409 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
410 			*handle++ = Q8_SGL_STAT_DESC_HANDLE2((sdesc->data[0]));
411 			*handle++ = Q8_SGL_STAT_DESC_HANDLE3((sdesc->data[0]));
412 			*handle++ = Q8_SGL_STAT_DESC_HANDLE4((sdesc->data[0]));
413 			*handle++ = Q8_SGL_STAT_DESC_HANDLE5((sdesc->data[1]));
414 			*handle++ = Q8_SGL_STAT_DESC_HANDLE6((sdesc->data[1]));
415 			break;
416 
417 		case 7:
418 			*handle++ = Q8_SGL_STAT_DESC_HANDLE1((sdesc->data[0]));
419 			*handle++ = Q8_SGL_STAT_DESC_HANDLE2((sdesc->data[0]));
420 			*handle++ = Q8_SGL_STAT_DESC_HANDLE3((sdesc->data[0]));
421 			*handle++ = Q8_SGL_STAT_DESC_HANDLE4((sdesc->data[0]));
422 			*handle++ = Q8_SGL_STAT_DESC_HANDLE5((sdesc->data[1]));
423 			*handle++ = Q8_SGL_STAT_DESC_HANDLE6((sdesc->data[1]));
424 			*handle++ = Q8_SGL_STAT_DESC_HANDLE7((sdesc->data[1]));
425 			break;
426 
427 		default:
428 			device_printf(ha->pci_dev,
429 				"%s: invalid num handles %p %p\n",
430 				__func__, (void *)sdesc->data[0],
431 				(void *)sdesc->data[1]);
432 
433 			QL_ASSERT(ha, (0),\
434 			("%s: %s [nh, sds, d0, d1]=[%d, %d, %p, %p]\n",
435 			__func__, "invalid num handles", sds_idx, num_handles,
436 			(void *)sdesc->data[0],(void *)sdesc->data[1]));
437 
438 			qla_rcv_error(ha);
439 			return 0;
440 		}
441 		*nhandles = *nhandles + num_handles;
442 	}
443 	return 0;
444 }
445 
446 /*
447  * Name: qla_rcv_isr
448  * Function: Main Interrupt Service Routine
449  */
450 static uint32_t
451 qla_rcv_isr(qla_host_t *ha, uint32_t sds_idx, uint32_t count)
452 {
453 	device_t dev;
454 	qla_hw_t *hw;
455 	uint32_t comp_idx, c_idx = 0, desc_count = 0, opcode;
456 	volatile q80_stat_desc_t *sdesc, *sdesc0 = NULL;
457 	uint32_t ret = 0;
458 	qla_sgl_comp_t sgc;
459 	uint16_t nhandles;
460 	uint32_t sds_replenish_threshold = 0;
461 
462 	dev = ha->pci_dev;
463 	hw = &ha->hw;
464 
465 	hw->sds[sds_idx].rcv_active = 1;
466 	if (ha->flags.stop_rcv) {
467 		hw->sds[sds_idx].rcv_active = 0;
468 		return 0;
469 	}
470 
471 	QL_DPRINT2(ha, (dev, "%s: [%d]enter\n", __func__, sds_idx));
472 
473 	/*
474 	 * receive interrupts
475 	 */
476 	comp_idx = hw->sds[sds_idx].sdsr_next;
477 
478 	while (count-- && !ha->flags.stop_rcv) {
479 
480 		sdesc = (q80_stat_desc_t *)
481 				&hw->sds[sds_idx].sds_ring_base[comp_idx];
482 
483 		opcode = Q8_STAT_DESC_OPCODE((sdesc->data[1]));
484 
485 		if (!opcode)
486 			break;
487 
488 		hw->sds[sds_idx].intr_count++;
489 		switch (opcode) {
490 
491 		case Q8_STAT_DESC_OPCODE_RCV_PKT:
492 
493 			desc_count = 1;
494 
495 			bzero(&sgc, sizeof(qla_sgl_comp_t));
496 
497 			sgc.rcv.pkt_length =
498 				Q8_STAT_DESC_TOTAL_LENGTH((sdesc->data[0]));
499 			sgc.rcv.num_handles = 1;
500 			sgc.rcv.handle[0] =
501 				Q8_STAT_DESC_HANDLE((sdesc->data[0]));
502 			sgc.rcv.chksum_status =
503 				Q8_STAT_DESC_STATUS((sdesc->data[1]));
504 
505 			sgc.rcv.rss_hash =
506 				Q8_STAT_DESC_RSS_HASH((sdesc->data[0]));
507 
508 			if (Q8_STAT_DESC_VLAN((sdesc->data[1]))) {
509 				sgc.rcv.vlan_tag =
510 					Q8_STAT_DESC_VLAN_ID((sdesc->data[1]));
511 			}
512 			qla_rx_intr(ha, &sgc.rcv, sds_idx);
513 			break;
514 
515 		case Q8_STAT_DESC_OPCODE_SGL_RCV:
516 
517 			desc_count =
518 				Q8_STAT_DESC_COUNT_SGL_RCV((sdesc->data[1]));
519 
520 			if (desc_count > 1) {
521 				c_idx = (comp_idx + desc_count -1) &
522 						(NUM_STATUS_DESCRIPTORS-1);
523 				sdesc0 = (q80_stat_desc_t *)
524 					&hw->sds[sds_idx].sds_ring_base[c_idx];
525 
526 				if (Q8_STAT_DESC_OPCODE((sdesc0->data[1])) !=
527 						Q8_STAT_DESC_OPCODE_CONT) {
528 					desc_count = 0;
529 					break;
530 				}
531 			}
532 
533 			bzero(&sgc, sizeof(qla_sgl_comp_t));
534 
535 			sgc.rcv.pkt_length =
536 				Q8_STAT_DESC_TOTAL_LENGTH_SGL_RCV(\
537 					(sdesc->data[0]));
538 			sgc.rcv.chksum_status =
539 				Q8_STAT_DESC_STATUS((sdesc->data[1]));
540 
541 			sgc.rcv.rss_hash =
542 				Q8_STAT_DESC_RSS_HASH((sdesc->data[0]));
543 
544 			if (Q8_STAT_DESC_VLAN((sdesc->data[1]))) {
545 				sgc.rcv.vlan_tag =
546 					Q8_STAT_DESC_VLAN_ID((sdesc->data[1]));
547 			}
548 
549 			QL_ASSERT(ha, (desc_count <= 2) ,\
550 				("%s: [sds_idx, data0, data1]="\
551 				"%d, %p, %p]\n", __func__, sds_idx,\
552 				(void *)sdesc->data[0],\
553 				(void *)sdesc->data[1]));
554 
555 			sgc.rcv.num_handles = 1;
556 			sgc.rcv.handle[0] =
557 				Q8_STAT_DESC_HANDLE((sdesc->data[0]));
558 
559 			if (qla_rcv_cont_sds(ha, sds_idx, comp_idx, desc_count,
560 				&sgc.rcv.handle[1], &nhandles)) {
561 				device_printf(dev,
562 					"%s: [sds_idx, dcount, data0, data1]="
563 					 "[%d, %d, 0x%llx, 0x%llx]\n",
564 					__func__, sds_idx, desc_count,
565 					(long long unsigned int)sdesc->data[0],
566 					(long long unsigned int)sdesc->data[1]);
567 				desc_count = 0;
568 				break;
569 			}
570 
571 			sgc.rcv.num_handles += nhandles;
572 
573 			qla_rx_intr(ha, &sgc.rcv, sds_idx);
574 
575 			break;
576 
577 		case Q8_STAT_DESC_OPCODE_SGL_LRO:
578 
579 			desc_count =
580 				Q8_STAT_DESC_COUNT_SGL_LRO((sdesc->data[1]));
581 
582 			if (desc_count > 1) {
583 				c_idx = (comp_idx + desc_count -1) &
584 						(NUM_STATUS_DESCRIPTORS-1);
585 				sdesc0 = (q80_stat_desc_t *)
586 					&hw->sds[sds_idx].sds_ring_base[c_idx];
587 
588 				if (Q8_STAT_DESC_OPCODE((sdesc0->data[1])) !=
589 						Q8_STAT_DESC_OPCODE_CONT) {
590 					desc_count = 0;
591 					break;
592 				}
593 			}
594 			bzero(&sgc, sizeof(qla_sgl_comp_t));
595 
596 			sgc.lro.payload_length =
597 			Q8_STAT_DESC_TOTAL_LENGTH_SGL_RCV((sdesc->data[0]));
598 
599 			sgc.lro.rss_hash =
600 				Q8_STAT_DESC_RSS_HASH((sdesc->data[0]));
601 
602 			sgc.lro.num_handles = 1;
603 			sgc.lro.handle[0] =
604 				Q8_STAT_DESC_HANDLE((sdesc->data[0]));
605 
606 			if (Q8_SGL_LRO_STAT_TS((sdesc->data[1])))
607 				sgc.lro.flags |= Q8_LRO_COMP_TS;
608 
609 			if (Q8_SGL_LRO_STAT_PUSH_BIT((sdesc->data[1])))
610 				sgc.lro.flags |= Q8_LRO_COMP_PUSH_BIT;
611 
612 			sgc.lro.l2_offset =
613 				Q8_SGL_LRO_STAT_L2_OFFSET((sdesc->data[1]));
614 			sgc.lro.l4_offset =
615 				Q8_SGL_LRO_STAT_L4_OFFSET((sdesc->data[1]));
616 
617 			if (Q8_STAT_DESC_VLAN((sdesc->data[1]))) {
618 				sgc.lro.vlan_tag =
619 					Q8_STAT_DESC_VLAN_ID((sdesc->data[1]));
620 			}
621 
622 			QL_ASSERT(ha, (desc_count <= 7) ,\
623 				("%s: [sds_idx, data0, data1]="\
624 				 "[%d, 0x%llx, 0x%llx]\n",\
625 				__func__, sds_idx,\
626 				(long long unsigned int)sdesc->data[0],\
627 				(long long unsigned int)sdesc->data[1]));
628 
629 			if (qla_rcv_cont_sds(ha, sds_idx, comp_idx,
630 				desc_count, &sgc.lro.handle[1], &nhandles)) {
631 				device_printf(dev,
632 				"%s: [sds_idx, data0, data1]="\
633 				 "[%d, 0x%llx, 0x%llx]\n",\
634 				__func__, sds_idx,\
635 				(long long unsigned int)sdesc->data[0],\
636 				(long long unsigned int)sdesc->data[1]);
637 
638 				desc_count = 0;
639 				break;
640 			}
641 
642 			sgc.lro.num_handles += nhandles;
643 
644 			if (qla_lro_intr(ha, &sgc.lro, sds_idx)) {
645 				device_printf(dev,
646 				"%s: [sds_idx, data0, data1]="\
647 				 "[%d, 0x%llx, 0x%llx]\n",\
648 				__func__, sds_idx,\
649 				(long long unsigned int)sdesc->data[0],\
650 				(long long unsigned int)sdesc->data[1]);
651 				device_printf(dev,
652 				"%s: [comp_idx, c_idx, dcount, nhndls]="\
653 				 "[%d, %d, %d, %d]\n",\
654 				__func__, comp_idx, c_idx, desc_count,
655 				sgc.lro.num_handles);
656 				if (desc_count > 1) {
657 				device_printf(dev,
658 				"%s: [sds_idx, data0, data1]="\
659 				 "[%d, 0x%llx, 0x%llx]\n",\
660 				__func__, sds_idx,\
661 				(long long unsigned int)sdesc0->data[0],\
662 				(long long unsigned int)sdesc0->data[1]);
663 				}
664 			}
665 
666 			break;
667 
668 		default:
669 			device_printf(dev, "%s: default 0x%llx!\n", __func__,
670 					(long long unsigned int)sdesc->data[0]);
671 			break;
672 		}
673 
674 		if (desc_count == 0)
675 			break;
676 
677 		sds_replenish_threshold += desc_count;
678 
679 
680 		while (desc_count--) {
681 			sdesc->data[0] = 0ULL;
682 			sdesc->data[1] = 0ULL;
683 			comp_idx = (comp_idx + 1) & (NUM_STATUS_DESCRIPTORS-1);
684 			sdesc = (q80_stat_desc_t *)
685 				&hw->sds[sds_idx].sds_ring_base[comp_idx];
686 		}
687 
688 		if (sds_replenish_threshold > ha->hw.sds_cidx_thres) {
689 			sds_replenish_threshold = 0;
690 			if (hw->sds[sds_idx].sdsr_next != comp_idx) {
691 				QL_UPDATE_SDS_CONSUMER_INDEX(ha, sds_idx,\
692 					comp_idx);
693 			}
694 			hw->sds[sds_idx].sdsr_next = comp_idx;
695 		}
696 	}
697 
698 	if (ha->flags.stop_rcv)
699 		goto qla_rcv_isr_exit;
700 
701 	if (hw->sds[sds_idx].sdsr_next != comp_idx) {
702 		QL_UPDATE_SDS_CONSUMER_INDEX(ha, sds_idx, comp_idx);
703 	}
704 	hw->sds[sds_idx].sdsr_next = comp_idx;
705 
706 	sdesc = (q80_stat_desc_t *)&hw->sds[sds_idx].sds_ring_base[comp_idx];
707 	opcode = Q8_STAT_DESC_OPCODE((sdesc->data[1]));
708 
709 	if (opcode)
710 		ret = -1;
711 
712 qla_rcv_isr_exit:
713 	hw->sds[sds_idx].rcv_active = 0;
714 
715 	return (ret);
716 }
717 
718 void
719 ql_mbx_isr(void *arg)
720 {
721 	qla_host_t *ha;
722 	uint32_t data;
723 	uint32_t prev_link_state;
724 
725 	ha = arg;
726 
727 	if (ha == NULL) {
728 		device_printf(ha->pci_dev, "%s: arg == NULL\n", __func__);
729 		return;
730 	}
731 
732 	data = READ_REG32(ha, Q8_FW_MBOX_CNTRL);
733 	if ((data & 0x3) != 0x1) {
734 		WRITE_REG32(ha, ha->hw.mbx_intr_mask_offset, 0);
735 		return;
736 	}
737 
738 	data = READ_REG32(ha, Q8_FW_MBOX0);
739 
740 	if ((data & 0xF000) != 0x8000)
741 		return;
742 
743 	data = data & 0xFFFF;
744 
745 	switch (data) {
746 
747 	case 0x8001:  /* It's an AEN */
748 
749 		ha->hw.cable_oui = READ_REG32(ha, (Q8_FW_MBOX0 + 4));
750 
751 		data = READ_REG32(ha, (Q8_FW_MBOX0 + 8));
752 		ha->hw.cable_length = data & 0xFFFF;
753 
754 		data = data >> 16;
755 		ha->hw.link_speed = data & 0xFFF;
756 
757 		data = READ_REG32(ha, (Q8_FW_MBOX0 + 12));
758 
759 		prev_link_state =  ha->hw.link_up;
760 		ha->hw.link_up = (((data & 0xFF) == 0) ? 0 : 1);
761 
762 		if (prev_link_state !=  ha->hw.link_up) {
763 			if (ha->hw.link_up)
764 				if_link_state_change(ha->ifp, LINK_STATE_UP);
765 			else
766 				if_link_state_change(ha->ifp, LINK_STATE_DOWN);
767 		}
768 
769 
770 		ha->hw.module_type = ((data >> 8) & 0xFF);
771 		ha->hw.flags.fduplex = (((data & 0xFF0000) == 0) ? 0 : 1);
772 		ha->hw.flags.autoneg = (((data & 0xFF000000) == 0) ? 0 : 1);
773 
774 		data = READ_REG32(ha, (Q8_FW_MBOX0 + 16));
775 		ha->hw.flags.loopback_mode = data & 0x03;
776 
777 		ha->hw.link_faults = (data >> 3) & 0xFF;
778 
779 		break;
780 
781         case 0x8100:
782 		ha->hw.imd_compl=1;
783 		break;
784 
785         case 0x8101:
786                 ha->async_event = 1;
787                 ha->hw.aen_mb0 = 0x8101;
788                 ha->hw.aen_mb1 = READ_REG32(ha, (Q8_FW_MBOX0 + 4));
789                 ha->hw.aen_mb2 = READ_REG32(ha, (Q8_FW_MBOX0 + 8));
790                 ha->hw.aen_mb3 = READ_REG32(ha, (Q8_FW_MBOX0 + 12));
791                 ha->hw.aen_mb4 = READ_REG32(ha, (Q8_FW_MBOX0 + 16));
792                 break;
793 
794         case 0x8110:
795                 /* for now just dump the registers */
796                 {
797                         uint32_t ombx[5];
798 
799                         ombx[0] = READ_REG32(ha, (Q8_FW_MBOX0 + 4));
800                         ombx[1] = READ_REG32(ha, (Q8_FW_MBOX0 + 8));
801                         ombx[2] = READ_REG32(ha, (Q8_FW_MBOX0 + 12));
802                         ombx[3] = READ_REG32(ha, (Q8_FW_MBOX0 + 16));
803                         ombx[4] = READ_REG32(ha, (Q8_FW_MBOX0 + 20));
804 
805                         device_printf(ha->pci_dev, "%s: "
806                                 "0x%08x 0x%08x 0x%08x 0x%08x 0x%08x 0x%08x\n",
807                                 __func__, data, ombx[0], ombx[1], ombx[2],
808                                 ombx[3], ombx[4]);
809                 }
810 
811                 break;
812 
813         case 0x8130:
814                 /* sfp insertion aen */
815                 device_printf(ha->pci_dev, "%s: sfp inserted [0x%08x]\n",
816                         __func__, READ_REG32(ha, (Q8_FW_MBOX0 + 4)));
817                 break;
818 
819         case 0x8131:
820                 /* sfp removal aen */
821                 device_printf(ha->pci_dev, "%s: sfp removed]\n", __func__);
822                 break;
823 
824 	default:
825 		device_printf(ha->pci_dev, "%s: AEN[0x%08x]\n", __func__, data);
826 		break;
827 	}
828 	WRITE_REG32(ha, Q8_FW_MBOX_CNTRL, 0x0);
829 	WRITE_REG32(ha, ha->hw.mbx_intr_mask_offset, 0x0);
830 	return;
831 }
832 
833 
834 static void
835 qla_replenish_normal_rx(qla_host_t *ha, qla_sds_t *sdsp, uint32_t r_idx)
836 {
837 	qla_rx_buf_t *rxb;
838 	int count = sdsp->rx_free;
839 	uint32_t rx_next;
840 	qla_rdesc_t *rdesc;
841 
842 	/* we can play with this value via a sysctl */
843 	uint32_t replenish_thresh = ha->hw.rds_pidx_thres;
844 
845 	rdesc = &ha->hw.rds[r_idx];
846 
847 	rx_next = rdesc->rx_next;
848 
849 	while (count--) {
850 		rxb = sdsp->rxb_free;
851 
852 		if (rxb == NULL)
853 			break;
854 
855 		sdsp->rxb_free = rxb->next;
856 		sdsp->rx_free--;
857 
858 		if (ql_get_mbuf(ha, rxb, NULL) == 0) {
859 			qla_set_hw_rcv_desc(ha, r_idx, rdesc->rx_in,
860 				rxb->handle,
861 				rxb->paddr, (rxb->m_head)->m_pkthdr.len);
862 			rdesc->rx_in++;
863 			if (rdesc->rx_in == NUM_RX_DESCRIPTORS)
864 				rdesc->rx_in = 0;
865 			rdesc->rx_next++;
866 			if (rdesc->rx_next == NUM_RX_DESCRIPTORS)
867 				rdesc->rx_next = 0;
868 		} else {
869 			device_printf(ha->pci_dev,
870 				"%s: ql_get_mbuf [0,(%d),(%d)] failed\n",
871 				__func__, rdesc->rx_in, rxb->handle);
872 
873 			rxb->m_head = NULL;
874 			rxb->next = sdsp->rxb_free;
875 			sdsp->rxb_free = rxb;
876 			sdsp->rx_free++;
877 
878 			break;
879 		}
880 		if (replenish_thresh-- == 0) {
881 			QL_UPDATE_RDS_PRODUCER_INDEX(ha, rdesc->prod_std,
882 				rdesc->rx_next);
883 			rx_next = rdesc->rx_next;
884 			replenish_thresh = ha->hw.rds_pidx_thres;
885 		}
886 	}
887 
888 	if (rx_next != rdesc->rx_next) {
889 		QL_UPDATE_RDS_PRODUCER_INDEX(ha, rdesc->prod_std,
890 			rdesc->rx_next);
891 	}
892 }
893 
894 void
895 ql_isr(void *arg)
896 {
897 	qla_ivec_t *ivec = arg;
898 	qla_host_t *ha ;
899 	int idx;
900 	qla_hw_t *hw;
901 	struct ifnet *ifp;
902 	uint32_t ret = 0;
903 
904 	ha = ivec->ha;
905 	hw = &ha->hw;
906 	ifp = ha->ifp;
907 
908 	if ((idx = ivec->sds_idx) >= ha->hw.num_sds_rings)
909 		return;
910 
911 	if (idx == 0)
912 		taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
913 
914 	ret = qla_rcv_isr(ha, idx, -1);
915 
916 	if (idx == 0)
917 		taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
918 
919 	if (!ha->flags.stop_rcv) {
920 		QL_ENABLE_INTERRUPTS(ha, idx);
921 	}
922 	return;
923 }
924 
925