xref: /freebsd/sys/dev/oce/oce_if.c (revision ce3adf4362fcca6a43e500b2531f0038adbfbd21)
1 /*-
2  * Copyright (C) 2013 Emulex
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions are met:
7  *
8  * 1. Redistributions of source code must retain the above copyright notice,
9  *    this list of conditions and the following disclaimer.
10  *
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * 3. Neither the name of the Emulex Corporation nor the names of its
16  *    contributors may be used to endorse or promote products derived from
17  *    this software without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
20  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
23  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29  * POSSIBILITY OF SUCH DAMAGE.
30  *
31  * Contact Information:
32  * freebsd-drivers@emulex.com
33  *
34  * Emulex
35  * 3333 Susan Street
36  * Costa Mesa, CA 92626
37  */
38 
39 
40 /* $FreeBSD$ */
41 
42 #include "opt_inet6.h"
43 #include "opt_inet.h"
44 
45 #include "oce_if.h"
46 
47 
48 /* Driver entry points prototypes */
49 static int  oce_probe(device_t dev);
50 static int  oce_attach(device_t dev);
51 static int  oce_detach(device_t dev);
52 static int  oce_shutdown(device_t dev);
53 static int  oce_ioctl(struct ifnet *ifp, u_long command, caddr_t data);
54 static void oce_init(void *xsc);
55 static int  oce_multiq_start(struct ifnet *ifp, struct mbuf *m);
56 static void oce_multiq_flush(struct ifnet *ifp);
57 
58 /* Driver interrupt routines protypes */
59 static void oce_intr(void *arg, int pending);
60 static int  oce_setup_intr(POCE_SOFTC sc);
61 static int  oce_fast_isr(void *arg);
62 static int  oce_alloc_intr(POCE_SOFTC sc, int vector,
63 			  void (*isr) (void *arg, int pending));
64 
65 /* Media callbacks prototypes */
66 static void oce_media_status(struct ifnet *ifp, struct ifmediareq *req);
67 static int  oce_media_change(struct ifnet *ifp);
68 
69 /* Transmit routines prototypes */
70 static int  oce_tx(POCE_SOFTC sc, struct mbuf **mpp, int wq_index);
71 static void oce_tx_restart(POCE_SOFTC sc, struct oce_wq *wq);
72 static void oce_tx_complete(struct oce_wq *wq, uint32_t wqe_idx,
73 					uint32_t status);
74 static int  oce_multiq_transmit(struct ifnet *ifp, struct mbuf *m,
75 				 struct oce_wq *wq);
76 
77 /* Receive routines prototypes */
78 static void oce_discard_rx_comp(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe);
79 static int  oce_cqe_vtp_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe);
80 static int  oce_cqe_portid_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe);
81 static void oce_rx(struct oce_rq *rq, uint32_t rqe_idx,
82 						struct oce_nic_rx_cqe *cqe);
83 
84 /* Helper function prototypes in this file */
85 static int  oce_attach_ifp(POCE_SOFTC sc);
86 static void oce_add_vlan(void *arg, struct ifnet *ifp, uint16_t vtag);
87 static void oce_del_vlan(void *arg, struct ifnet *ifp, uint16_t vtag);
88 static int  oce_vid_config(POCE_SOFTC sc);
89 static void oce_mac_addr_set(POCE_SOFTC sc);
90 static int  oce_handle_passthrough(struct ifnet *ifp, caddr_t data);
91 static void oce_local_timer(void *arg);
92 static void oce_if_deactivate(POCE_SOFTC sc);
93 static void oce_if_activate(POCE_SOFTC sc);
94 static void setup_max_queues_want(POCE_SOFTC sc);
95 static void update_queues_got(POCE_SOFTC sc);
96 static void process_link_state(POCE_SOFTC sc,
97 		 struct oce_async_cqe_link_state *acqe);
98 static int oce_tx_asic_stall_verify(POCE_SOFTC sc, struct mbuf *m);
99 static void oce_get_config(POCE_SOFTC sc);
100 static struct mbuf *oce_insert_vlan_tag(POCE_SOFTC sc, struct mbuf *m, boolean_t *complete);
101 
102 /* IP specific */
103 #if defined(INET6) || defined(INET)
104 static int  oce_init_lro(POCE_SOFTC sc);
105 static void oce_rx_flush_lro(struct oce_rq *rq);
106 static struct mbuf * oce_tso_setup(POCE_SOFTC sc, struct mbuf **mpp);
107 #endif
108 
109 static device_method_t oce_dispatch[] = {
110 	DEVMETHOD(device_probe, oce_probe),
111 	DEVMETHOD(device_attach, oce_attach),
112 	DEVMETHOD(device_detach, oce_detach),
113 	DEVMETHOD(device_shutdown, oce_shutdown),
114 
115 	DEVMETHOD_END
116 };
117 
118 static driver_t oce_driver = {
119 	"oce",
120 	oce_dispatch,
121 	sizeof(OCE_SOFTC)
122 };
123 static devclass_t oce_devclass;
124 
125 
126 DRIVER_MODULE(oce, pci, oce_driver, oce_devclass, 0, 0);
127 MODULE_DEPEND(oce, pci, 1, 1, 1);
128 MODULE_DEPEND(oce, ether, 1, 1, 1);
129 MODULE_VERSION(oce, 1);
130 
131 
132 /* global vars */
133 const char component_revision[32] = {"///" COMPONENT_REVISION "///"};
134 
135 /* Module capabilites and parameters */
136 uint32_t oce_max_rsp_handled = OCE_MAX_RSP_HANDLED;
137 uint32_t oce_enable_rss = OCE_MODCAP_RSS;
138 
139 
140 TUNABLE_INT("hw.oce.max_rsp_handled", &oce_max_rsp_handled);
141 TUNABLE_INT("hw.oce.enable_rss", &oce_enable_rss);
142 
143 
144 /* Supported devices table */
145 static uint32_t supportedDevices[] =  {
146 	(PCI_VENDOR_SERVERENGINES << 16) | PCI_PRODUCT_BE2,
147 	(PCI_VENDOR_SERVERENGINES << 16) | PCI_PRODUCT_BE3,
148 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_BE3,
149 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_XE201,
150 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_XE201_VF,
151 	(PCI_VENDOR_EMULEX << 16) | PCI_PRODUCT_SH
152 };
153 
154 
155 
156 
157 /*****************************************************************************
158  *			Driver entry points functions                        *
159  *****************************************************************************/
160 
161 static int
162 oce_probe(device_t dev)
163 {
164 	uint16_t vendor = 0;
165 	uint16_t device = 0;
166 	int i = 0;
167 	char str[256] = {0};
168 	POCE_SOFTC sc;
169 
170 	sc = device_get_softc(dev);
171 	bzero(sc, sizeof(OCE_SOFTC));
172 	sc->dev = dev;
173 
174 	vendor = pci_get_vendor(dev);
175 	device = pci_get_device(dev);
176 
177 	for (i = 0; i < (sizeof(supportedDevices) / sizeof(uint32_t)); i++) {
178 		if (vendor == ((supportedDevices[i] >> 16) & 0xffff)) {
179 			if (device == (supportedDevices[i] & 0xffff)) {
180 				sprintf(str, "%s:%s", "Emulex CNA NIC function",
181 					component_revision);
182 				device_set_desc_copy(dev, str);
183 
184 				switch (device) {
185 				case PCI_PRODUCT_BE2:
186 					sc->flags |= OCE_FLAGS_BE2;
187 					break;
188 				case PCI_PRODUCT_BE3:
189 					sc->flags |= OCE_FLAGS_BE3;
190 					break;
191 				case PCI_PRODUCT_XE201:
192 				case PCI_PRODUCT_XE201_VF:
193 					sc->flags |= OCE_FLAGS_XE201;
194 					break;
195 				case PCI_PRODUCT_SH:
196 					sc->flags |= OCE_FLAGS_SH;
197 					break;
198 				default:
199 					return ENXIO;
200 				}
201 				return BUS_PROBE_DEFAULT;
202 			}
203 		}
204 	}
205 
206 	return ENXIO;
207 }
208 
209 
210 static int
211 oce_attach(device_t dev)
212 {
213 	POCE_SOFTC sc;
214 	int rc = 0;
215 
216 	sc = device_get_softc(dev);
217 
218 	rc = oce_hw_pci_alloc(sc);
219 	if (rc)
220 		return rc;
221 
222 	sc->tx_ring_size = OCE_TX_RING_SIZE;
223 	sc->rx_ring_size = OCE_RX_RING_SIZE;
224 	sc->rq_frag_size = OCE_RQ_BUF_SIZE;
225 	sc->flow_control = OCE_DEFAULT_FLOW_CONTROL;
226 	sc->promisc	 = OCE_DEFAULT_PROMISCUOUS;
227 
228 	LOCK_CREATE(&sc->bmbx_lock, "Mailbox_lock");
229 	LOCK_CREATE(&sc->dev_lock,  "Device_lock");
230 
231 	/* initialise the hardware */
232 	rc = oce_hw_init(sc);
233 	if (rc)
234 		goto pci_res_free;
235 
236 	oce_get_config(sc);
237 
238 	setup_max_queues_want(sc);
239 
240 	rc = oce_setup_intr(sc);
241 	if (rc)
242 		goto mbox_free;
243 
244 	rc = oce_queue_init_all(sc);
245 	if (rc)
246 		goto intr_free;
247 
248 	rc = oce_attach_ifp(sc);
249 	if (rc)
250 		goto queues_free;
251 
252 #if defined(INET6) || defined(INET)
253 	rc = oce_init_lro(sc);
254 	if (rc)
255 		goto ifp_free;
256 #endif
257 
258 	rc = oce_hw_start(sc);
259 	if (rc)
260 		goto lro_free;
261 
262 	sc->vlan_attach = EVENTHANDLER_REGISTER(vlan_config,
263 				oce_add_vlan, sc, EVENTHANDLER_PRI_FIRST);
264 	sc->vlan_detach = EVENTHANDLER_REGISTER(vlan_unconfig,
265 				oce_del_vlan, sc, EVENTHANDLER_PRI_FIRST);
266 
267 	rc = oce_stats_init(sc);
268 	if (rc)
269 		goto vlan_free;
270 
271 	oce_add_sysctls(sc);
272 
273 	callout_init(&sc->timer, CALLOUT_MPSAFE);
274 	rc = callout_reset(&sc->timer, 2 * hz, oce_local_timer, sc);
275 	if (rc)
276 		goto stats_free;
277 
278 	return 0;
279 
280 stats_free:
281 	callout_drain(&sc->timer);
282 	oce_stats_free(sc);
283 vlan_free:
284 	if (sc->vlan_attach)
285 		EVENTHANDLER_DEREGISTER(vlan_config, sc->vlan_attach);
286 	if (sc->vlan_detach)
287 		EVENTHANDLER_DEREGISTER(vlan_unconfig, sc->vlan_detach);
288 	oce_hw_intr_disable(sc);
289 lro_free:
290 #if defined(INET6) || defined(INET)
291 	oce_free_lro(sc);
292 ifp_free:
293 #endif
294 	ether_ifdetach(sc->ifp);
295 	if_free(sc->ifp);
296 queues_free:
297 	oce_queue_release_all(sc);
298 intr_free:
299 	oce_intr_free(sc);
300 mbox_free:
301 	oce_dma_free(sc, &sc->bsmbx);
302 pci_res_free:
303 	oce_hw_pci_free(sc);
304 	LOCK_DESTROY(&sc->dev_lock);
305 	LOCK_DESTROY(&sc->bmbx_lock);
306 	return rc;
307 
308 }
309 
310 
311 static int
312 oce_detach(device_t dev)
313 {
314 	POCE_SOFTC sc = device_get_softc(dev);
315 
316 	LOCK(&sc->dev_lock);
317 	oce_if_deactivate(sc);
318 	UNLOCK(&sc->dev_lock);
319 
320 	callout_drain(&sc->timer);
321 
322 	if (sc->vlan_attach != NULL)
323 		EVENTHANDLER_DEREGISTER(vlan_config, sc->vlan_attach);
324 	if (sc->vlan_detach != NULL)
325 		EVENTHANDLER_DEREGISTER(vlan_unconfig, sc->vlan_detach);
326 
327 	ether_ifdetach(sc->ifp);
328 
329 	if_free(sc->ifp);
330 
331 	oce_hw_shutdown(sc);
332 
333 	bus_generic_detach(dev);
334 
335 	return 0;
336 }
337 
338 
339 static int
340 oce_shutdown(device_t dev)
341 {
342 	int rc;
343 
344 	rc = oce_detach(dev);
345 
346 	return rc;
347 }
348 
349 
350 static int
351 oce_ioctl(struct ifnet *ifp, u_long command, caddr_t data)
352 {
353 	struct ifreq *ifr = (struct ifreq *)data;
354 	POCE_SOFTC sc = ifp->if_softc;
355 	int rc = 0;
356 	uint32_t u;
357 
358 	switch (command) {
359 
360 	case SIOCGIFMEDIA:
361 		rc = ifmedia_ioctl(ifp, ifr, &sc->media, command);
362 		break;
363 
364 	case SIOCSIFMTU:
365 		if (ifr->ifr_mtu > OCE_MAX_MTU)
366 			rc = EINVAL;
367 		else
368 			ifp->if_mtu = ifr->ifr_mtu;
369 		break;
370 
371 	case SIOCSIFFLAGS:
372 		if (ifp->if_flags & IFF_UP) {
373 			if (!(ifp->if_drv_flags & IFF_DRV_RUNNING)) {
374 				sc->ifp->if_drv_flags |= IFF_DRV_RUNNING;
375 				oce_init(sc);
376 			}
377 			device_printf(sc->dev, "Interface Up\n");
378 		} else {
379 			LOCK(&sc->dev_lock);
380 
381 			sc->ifp->if_drv_flags &=
382 			    ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
383 			oce_if_deactivate(sc);
384 
385 			UNLOCK(&sc->dev_lock);
386 
387 			device_printf(sc->dev, "Interface Down\n");
388 		}
389 
390 		if ((ifp->if_flags & IFF_PROMISC) && !sc->promisc) {
391 			sc->promisc = TRUE;
392 			oce_rxf_set_promiscuous(sc, sc->promisc);
393 		} else if (!(ifp->if_flags & IFF_PROMISC) && sc->promisc) {
394 			sc->promisc = FALSE;
395 			oce_rxf_set_promiscuous(sc, sc->promisc);
396 		}
397 
398 		break;
399 
400 	case SIOCADDMULTI:
401 	case SIOCDELMULTI:
402 		rc = oce_hw_update_multicast(sc);
403 		if (rc)
404 			device_printf(sc->dev,
405 				"Update multicast address failed\n");
406 		break;
407 
408 	case SIOCSIFCAP:
409 		u = ifr->ifr_reqcap ^ ifp->if_capenable;
410 
411 		if (u & IFCAP_TXCSUM) {
412 			ifp->if_capenable ^= IFCAP_TXCSUM;
413 			ifp->if_hwassist ^= (CSUM_TCP | CSUM_UDP | CSUM_IP);
414 
415 			if (IFCAP_TSO & ifp->if_capenable &&
416 			    !(IFCAP_TXCSUM & ifp->if_capenable)) {
417 				ifp->if_capenable &= ~IFCAP_TSO;
418 				ifp->if_hwassist &= ~CSUM_TSO;
419 				if_printf(ifp,
420 					 "TSO disabled due to -txcsum.\n");
421 			}
422 		}
423 
424 		if (u & IFCAP_RXCSUM)
425 			ifp->if_capenable ^= IFCAP_RXCSUM;
426 
427 		if (u & IFCAP_TSO4) {
428 			ifp->if_capenable ^= IFCAP_TSO4;
429 
430 			if (IFCAP_TSO & ifp->if_capenable) {
431 				if (IFCAP_TXCSUM & ifp->if_capenable)
432 					ifp->if_hwassist |= CSUM_TSO;
433 				else {
434 					ifp->if_capenable &= ~IFCAP_TSO;
435 					ifp->if_hwassist &= ~CSUM_TSO;
436 					if_printf(ifp,
437 					    "Enable txcsum first.\n");
438 					rc = EAGAIN;
439 				}
440 			} else
441 				ifp->if_hwassist &= ~CSUM_TSO;
442 		}
443 
444 		if (u & IFCAP_VLAN_HWTAGGING)
445 			ifp->if_capenable ^= IFCAP_VLAN_HWTAGGING;
446 
447 		if (u & IFCAP_VLAN_HWFILTER) {
448 			ifp->if_capenable ^= IFCAP_VLAN_HWFILTER;
449 			oce_vid_config(sc);
450 		}
451 #if defined(INET6) || defined(INET)
452 		if (u & IFCAP_LRO)
453 			ifp->if_capenable ^= IFCAP_LRO;
454 #endif
455 
456 		break;
457 
458 	case SIOCGPRIVATE_0:
459 		rc = oce_handle_passthrough(ifp, data);
460 		break;
461 	default:
462 		rc = ether_ioctl(ifp, command, data);
463 		break;
464 	}
465 
466 	return rc;
467 }
468 
469 
470 static void
471 oce_init(void *arg)
472 {
473 	POCE_SOFTC sc = arg;
474 
475 	LOCK(&sc->dev_lock);
476 
477 	if (sc->ifp->if_flags & IFF_UP) {
478 		oce_if_deactivate(sc);
479 		oce_if_activate(sc);
480 	}
481 
482 	UNLOCK(&sc->dev_lock);
483 
484 }
485 
486 
487 static int
488 oce_multiq_start(struct ifnet *ifp, struct mbuf *m)
489 {
490 	POCE_SOFTC sc = ifp->if_softc;
491 	struct oce_wq *wq = NULL;
492 	int queue_index = 0;
493 	int status = 0;
494 
495 	if (!sc->link_status)
496 		return ENXIO;
497 
498 	if ((m->m_flags & M_FLOWID) != 0)
499 		queue_index = m->m_pkthdr.flowid % sc->nwqs;
500 
501 	wq = sc->wq[queue_index];
502 
503 	LOCK(&wq->tx_lock);
504 	status = oce_multiq_transmit(ifp, m, wq);
505 	UNLOCK(&wq->tx_lock);
506 
507 	return status;
508 
509 }
510 
511 
512 static void
513 oce_multiq_flush(struct ifnet *ifp)
514 {
515 	POCE_SOFTC sc = ifp->if_softc;
516 	struct mbuf     *m;
517 	int i = 0;
518 
519 	for (i = 0; i < sc->nwqs; i++) {
520 		while ((m = buf_ring_dequeue_sc(sc->wq[i]->br)) != NULL)
521 			m_freem(m);
522 	}
523 	if_qflush(ifp);
524 }
525 
526 
527 
528 /*****************************************************************************
529  *                   Driver interrupt routines functions                     *
530  *****************************************************************************/
531 
532 static void
533 oce_intr(void *arg, int pending)
534 {
535 
536 	POCE_INTR_INFO ii = (POCE_INTR_INFO) arg;
537 	POCE_SOFTC sc = ii->sc;
538 	struct oce_eq *eq = ii->eq;
539 	struct oce_eqe *eqe;
540 	struct oce_cq *cq = NULL;
541 	int i, num_eqes = 0;
542 
543 
544 	bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
545 				 BUS_DMASYNC_POSTWRITE);
546 	do {
547 		eqe = RING_GET_CONSUMER_ITEM_VA(eq->ring, struct oce_eqe);
548 		if (eqe->evnt == 0)
549 			break;
550 		eqe->evnt = 0;
551 		bus_dmamap_sync(eq->ring->dma.tag, eq->ring->dma.map,
552 					BUS_DMASYNC_POSTWRITE);
553 		RING_GET(eq->ring, 1);
554 		num_eqes++;
555 
556 	} while (TRUE);
557 
558 	if (!num_eqes)
559 		goto eq_arm; /* Spurious */
560 
561  	/* Clear EQ entries, but dont arm */
562 	oce_arm_eq(sc, eq->eq_id, num_eqes, FALSE, FALSE);
563 
564 	/* Process TX, RX and MCC. But dont arm CQ*/
565 	for (i = 0; i < eq->cq_valid; i++) {
566 		cq = eq->cq[i];
567 		(*cq->cq_handler)(cq->cb_arg);
568 	}
569 
570 	/* Arm all cqs connected to this EQ */
571 	for (i = 0; i < eq->cq_valid; i++) {
572 		cq = eq->cq[i];
573 		oce_arm_cq(sc, cq->cq_id, 0, TRUE);
574 	}
575 
576 eq_arm:
577 	oce_arm_eq(sc, eq->eq_id, 0, TRUE, FALSE);
578 
579 	return;
580 }
581 
582 
583 static int
584 oce_setup_intr(POCE_SOFTC sc)
585 {
586 	int rc = 0, use_intx = 0;
587 	int vector = 0, req_vectors = 0;
588 
589 	if (is_rss_enabled(sc))
590 		req_vectors = MAX((sc->nrqs - 1), sc->nwqs);
591 	else
592 		req_vectors = 1;
593 
594 	if (sc->flags & OCE_FLAGS_MSIX_CAPABLE) {
595 		sc->intr_count = req_vectors;
596 		rc = pci_alloc_msix(sc->dev, &sc->intr_count);
597 		if (rc != 0) {
598 			use_intx = 1;
599 			pci_release_msi(sc->dev);
600 		} else
601 			sc->flags |= OCE_FLAGS_USING_MSIX;
602 	} else
603 		use_intx = 1;
604 
605 	if (use_intx)
606 		sc->intr_count = 1;
607 
608 	/* Scale number of queues based on intr we got */
609 	update_queues_got(sc);
610 
611 	if (use_intx) {
612 		device_printf(sc->dev, "Using legacy interrupt\n");
613 		rc = oce_alloc_intr(sc, vector, oce_intr);
614 		if (rc)
615 			goto error;
616 	} else {
617 		for (; vector < sc->intr_count; vector++) {
618 			rc = oce_alloc_intr(sc, vector, oce_intr);
619 			if (rc)
620 				goto error;
621 		}
622 	}
623 
624 	return 0;
625 error:
626 	oce_intr_free(sc);
627 	return rc;
628 }
629 
630 
631 static int
632 oce_fast_isr(void *arg)
633 {
634 	POCE_INTR_INFO ii = (POCE_INTR_INFO) arg;
635 	POCE_SOFTC sc = ii->sc;
636 
637 	if (ii->eq == NULL)
638 		return FILTER_STRAY;
639 
640 	oce_arm_eq(sc, ii->eq->eq_id, 0, FALSE, TRUE);
641 
642 	taskqueue_enqueue_fast(ii->tq, &ii->task);
643 
644  	ii->eq->intr++;
645 
646 	return FILTER_HANDLED;
647 }
648 
649 
650 static int
651 oce_alloc_intr(POCE_SOFTC sc, int vector, void (*isr) (void *arg, int pending))
652 {
653 	POCE_INTR_INFO ii = &sc->intrs[vector];
654 	int rc = 0, rr;
655 
656 	if (vector >= OCE_MAX_EQ)
657 		return (EINVAL);
658 
659 	/* Set the resource id for the interrupt.
660 	 * MSIx is vector + 1 for the resource id,
661 	 * INTx is 0 for the resource id.
662 	 */
663 	if (sc->flags & OCE_FLAGS_USING_MSIX)
664 		rr = vector + 1;
665 	else
666 		rr = 0;
667 	ii->intr_res = bus_alloc_resource_any(sc->dev,
668 					      SYS_RES_IRQ,
669 					      &rr, RF_ACTIVE|RF_SHAREABLE);
670 	ii->irq_rr = rr;
671 	if (ii->intr_res == NULL) {
672 		device_printf(sc->dev,
673 			  "Could not allocate interrupt\n");
674 		rc = ENXIO;
675 		return rc;
676 	}
677 
678 	TASK_INIT(&ii->task, 0, isr, ii);
679 	ii->vector = vector;
680 	sprintf(ii->task_name, "oce_task[%d]", ii->vector);
681 	ii->tq = taskqueue_create_fast(ii->task_name,
682 			M_NOWAIT,
683 			taskqueue_thread_enqueue,
684 			&ii->tq);
685 	taskqueue_start_threads(&ii->tq, 1, PI_NET, "%s taskq",
686 			device_get_nameunit(sc->dev));
687 
688 	ii->sc = sc;
689 	rc = bus_setup_intr(sc->dev,
690 			ii->intr_res,
691 			INTR_TYPE_NET,
692 			oce_fast_isr, NULL, ii, &ii->tag);
693 	return rc;
694 
695 }
696 
697 
698 void
699 oce_intr_free(POCE_SOFTC sc)
700 {
701 	int i = 0;
702 
703 	for (i = 0; i < sc->intr_count; i++) {
704 
705 		if (sc->intrs[i].tag != NULL)
706 			bus_teardown_intr(sc->dev, sc->intrs[i].intr_res,
707 						sc->intrs[i].tag);
708 		if (sc->intrs[i].tq != NULL)
709 			taskqueue_free(sc->intrs[i].tq);
710 
711 		if (sc->intrs[i].intr_res != NULL)
712 			bus_release_resource(sc->dev, SYS_RES_IRQ,
713 						sc->intrs[i].irq_rr,
714 						sc->intrs[i].intr_res);
715 		sc->intrs[i].tag = NULL;
716 		sc->intrs[i].intr_res = NULL;
717 	}
718 
719 	if (sc->flags & OCE_FLAGS_USING_MSIX)
720 		pci_release_msi(sc->dev);
721 
722 }
723 
724 
725 
726 /******************************************************************************
727 *			  Media callbacks functions 			      *
728 ******************************************************************************/
729 
730 static void
731 oce_media_status(struct ifnet *ifp, struct ifmediareq *req)
732 {
733 	POCE_SOFTC sc = (POCE_SOFTC) ifp->if_softc;
734 
735 
736 	req->ifm_status = IFM_AVALID;
737 	req->ifm_active = IFM_ETHER;
738 
739 	if (sc->link_status == 1)
740 		req->ifm_status |= IFM_ACTIVE;
741 	else
742 		return;
743 
744 	switch (sc->link_speed) {
745 	case 1: /* 10 Mbps */
746 		req->ifm_active |= IFM_10_T | IFM_FDX;
747 		sc->speed = 10;
748 		break;
749 	case 2: /* 100 Mbps */
750 		req->ifm_active |= IFM_100_TX | IFM_FDX;
751 		sc->speed = 100;
752 		break;
753 	case 3: /* 1 Gbps */
754 		req->ifm_active |= IFM_1000_T | IFM_FDX;
755 		sc->speed = 1000;
756 		break;
757 	case 4: /* 10 Gbps */
758 		req->ifm_active |= IFM_10G_SR | IFM_FDX;
759 		sc->speed = 10000;
760 		break;
761 	}
762 
763 	return;
764 }
765 
766 
767 int
768 oce_media_change(struct ifnet *ifp)
769 {
770 	return 0;
771 }
772 
773 
774 
775 
776 /*****************************************************************************
777  *			  Transmit routines functions			     *
778  *****************************************************************************/
779 
780 static int
781 oce_tx(POCE_SOFTC sc, struct mbuf **mpp, int wq_index)
782 {
783 	int rc = 0, i, retry_cnt = 0;
784 	bus_dma_segment_t segs[OCE_MAX_TX_ELEMENTS];
785 	struct mbuf *m, *m_temp;
786 	struct oce_wq *wq = sc->wq[wq_index];
787 	struct oce_packet_desc *pd;
788 	struct oce_nic_hdr_wqe *nichdr;
789 	struct oce_nic_frag_wqe *nicfrag;
790 	int num_wqes;
791 	uint32_t reg_value;
792 	boolean_t complete = TRUE;
793 
794 	m = *mpp;
795 	if (!m)
796 		return EINVAL;
797 
798 	if (!(m->m_flags & M_PKTHDR)) {
799 		rc = ENXIO;
800 		goto free_ret;
801 	}
802 
803 	if(oce_tx_asic_stall_verify(sc, m)) {
804 		m = oce_insert_vlan_tag(sc, m, &complete);
805 		if(!m) {
806 			device_printf(sc->dev, "Insertion unsuccessful\n");
807 			return 0;
808 		}
809 
810 	}
811 
812 	if (m->m_pkthdr.csum_flags & CSUM_TSO) {
813 		/* consolidate packet buffers for TSO/LSO segment offload */
814 #if defined(INET6) || defined(INET)
815 		m = oce_tso_setup(sc, mpp);
816 #else
817 		m = NULL;
818 #endif
819 		if (m == NULL) {
820 			rc = ENXIO;
821 			goto free_ret;
822 		}
823 	}
824 
825 	pd = &wq->pckts[wq->pkt_desc_head];
826 retry:
827 	rc = bus_dmamap_load_mbuf_sg(wq->tag,
828 				     pd->map,
829 				     m, segs, &pd->nsegs, BUS_DMA_NOWAIT);
830 	if (rc == 0) {
831 		num_wqes = pd->nsegs + 1;
832 		if (IS_BE(sc) || IS_SH(sc)) {
833 			/*Dummy required only for BE3.*/
834 			if (num_wqes & 1)
835 				num_wqes++;
836 		}
837 		if (num_wqes >= RING_NUM_FREE(wq->ring)) {
838 			bus_dmamap_unload(wq->tag, pd->map);
839 			return EBUSY;
840 		}
841 		atomic_store_rel_int(&wq->pkt_desc_head,
842 				     (wq->pkt_desc_head + 1) % \
843 				      OCE_WQ_PACKET_ARRAY_SIZE);
844 		bus_dmamap_sync(wq->tag, pd->map, BUS_DMASYNC_PREWRITE);
845 		pd->mbuf = m;
846 
847 		nichdr =
848 		    RING_GET_PRODUCER_ITEM_VA(wq->ring, struct oce_nic_hdr_wqe);
849 		nichdr->u0.dw[0] = 0;
850 		nichdr->u0.dw[1] = 0;
851 		nichdr->u0.dw[2] = 0;
852 		nichdr->u0.dw[3] = 0;
853 
854 		nichdr->u0.s.complete = complete;
855 		nichdr->u0.s.event = 1;
856 		nichdr->u0.s.crc = 1;
857 		nichdr->u0.s.forward = 0;
858 		nichdr->u0.s.ipcs = (m->m_pkthdr.csum_flags & CSUM_IP) ? 1 : 0;
859 		nichdr->u0.s.udpcs =
860 			(m->m_pkthdr.csum_flags & CSUM_UDP) ? 1 : 0;
861 		nichdr->u0.s.tcpcs =
862 			(m->m_pkthdr.csum_flags & CSUM_TCP) ? 1 : 0;
863 		nichdr->u0.s.num_wqe = num_wqes;
864 		nichdr->u0.s.total_length = m->m_pkthdr.len;
865 		if (m->m_flags & M_VLANTAG) {
866 			nichdr->u0.s.vlan = 1; /*Vlan present*/
867 			nichdr->u0.s.vlan_tag = m->m_pkthdr.ether_vtag;
868 		}
869 		if (m->m_pkthdr.csum_flags & CSUM_TSO) {
870 			if (m->m_pkthdr.tso_segsz) {
871 				nichdr->u0.s.lso = 1;
872 				nichdr->u0.s.lso_mss  = m->m_pkthdr.tso_segsz;
873 			}
874 			if (!IS_BE(sc) || !IS_SH(sc))
875 				nichdr->u0.s.ipcs = 1;
876 		}
877 
878 		RING_PUT(wq->ring, 1);
879 		atomic_add_int(&wq->ring->num_used, 1);
880 
881 		for (i = 0; i < pd->nsegs; i++) {
882 			nicfrag =
883 			    RING_GET_PRODUCER_ITEM_VA(wq->ring,
884 						      struct oce_nic_frag_wqe);
885 			nicfrag->u0.s.rsvd0 = 0;
886 			nicfrag->u0.s.frag_pa_hi = ADDR_HI(segs[i].ds_addr);
887 			nicfrag->u0.s.frag_pa_lo = ADDR_LO(segs[i].ds_addr);
888 			nicfrag->u0.s.frag_len = segs[i].ds_len;
889 			pd->wqe_idx = wq->ring->pidx;
890 			RING_PUT(wq->ring, 1);
891 			atomic_add_int(&wq->ring->num_used, 1);
892 		}
893 		if (num_wqes > (pd->nsegs + 1)) {
894 			nicfrag =
895 			    RING_GET_PRODUCER_ITEM_VA(wq->ring,
896 						      struct oce_nic_frag_wqe);
897 			nicfrag->u0.dw[0] = 0;
898 			nicfrag->u0.dw[1] = 0;
899 			nicfrag->u0.dw[2] = 0;
900 			nicfrag->u0.dw[3] = 0;
901 			pd->wqe_idx = wq->ring->pidx;
902 			RING_PUT(wq->ring, 1);
903 			atomic_add_int(&wq->ring->num_used, 1);
904 			pd->nsegs++;
905 		}
906 
907 		sc->ifp->if_opackets++;
908 		wq->tx_stats.tx_reqs++;
909 		wq->tx_stats.tx_wrbs += num_wqes;
910 		wq->tx_stats.tx_bytes += m->m_pkthdr.len;
911 		wq->tx_stats.tx_pkts++;
912 
913 		bus_dmamap_sync(wq->ring->dma.tag, wq->ring->dma.map,
914 				BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
915 		reg_value = (num_wqes << 16) | wq->wq_id;
916 		OCE_WRITE_REG32(sc, db, wq->db_offset, reg_value);
917 
918 	} else if (rc == EFBIG)	{
919 		if (retry_cnt == 0) {
920 			m_temp = m_defrag(m, M_NOWAIT);
921 			if (m_temp == NULL)
922 				goto free_ret;
923 			m = m_temp;
924 			*mpp = m_temp;
925 			retry_cnt = retry_cnt + 1;
926 			goto retry;
927 		} else
928 			goto free_ret;
929 	} else if (rc == ENOMEM)
930 		return rc;
931 	else
932 		goto free_ret;
933 
934 	return 0;
935 
936 free_ret:
937 	m_freem(*mpp);
938 	*mpp = NULL;
939 	return rc;
940 }
941 
942 
943 static void
944 oce_tx_complete(struct oce_wq *wq, uint32_t wqe_idx, uint32_t status)
945 {
946 	struct oce_packet_desc *pd;
947 	POCE_SOFTC sc = (POCE_SOFTC) wq->parent;
948 	struct mbuf *m;
949 
950 	pd = &wq->pckts[wq->pkt_desc_tail];
951 	atomic_store_rel_int(&wq->pkt_desc_tail,
952 			     (wq->pkt_desc_tail + 1) % OCE_WQ_PACKET_ARRAY_SIZE);
953 	atomic_subtract_int(&wq->ring->num_used, pd->nsegs + 1);
954 	bus_dmamap_sync(wq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
955 	bus_dmamap_unload(wq->tag, pd->map);
956 
957 	m = pd->mbuf;
958 	m_freem(m);
959 	pd->mbuf = NULL;
960 
961 
962 	if (sc->ifp->if_drv_flags & IFF_DRV_OACTIVE) {
963 		if (wq->ring->num_used < (wq->ring->num_items / 2)) {
964 			sc->ifp->if_drv_flags &= ~(IFF_DRV_OACTIVE);
965 			oce_tx_restart(sc, wq);
966 		}
967 	}
968 }
969 
970 
971 static void
972 oce_tx_restart(POCE_SOFTC sc, struct oce_wq *wq)
973 {
974 
975 	if ((sc->ifp->if_drv_flags & IFF_DRV_RUNNING) != IFF_DRV_RUNNING)
976 		return;
977 
978 #if __FreeBSD_version >= 800000
979 	if (!drbr_empty(sc->ifp, wq->br))
980 #else
981 	if (!IFQ_DRV_IS_EMPTY(&sc->ifp->if_snd))
982 #endif
983 		taskqueue_enqueue_fast(taskqueue_swi, &wq->txtask);
984 
985 }
986 
987 
988 #if defined(INET6) || defined(INET)
989 static struct mbuf *
990 oce_tso_setup(POCE_SOFTC sc, struct mbuf **mpp)
991 {
992 	struct mbuf *m;
993 #ifdef INET
994 	struct ip *ip;
995 #endif
996 #ifdef INET6
997 	struct ip6_hdr *ip6;
998 #endif
999 	struct ether_vlan_header *eh;
1000 	struct tcphdr *th;
1001 	uint16_t etype;
1002 	int total_len = 0, ehdrlen = 0;
1003 
1004 	m = *mpp;
1005 
1006 	if (M_WRITABLE(m) == 0) {
1007 		m = m_dup(*mpp, M_NOWAIT);
1008 		if (!m)
1009 			return NULL;
1010 		m_freem(*mpp);
1011 		*mpp = m;
1012 	}
1013 
1014 	eh = mtod(m, struct ether_vlan_header *);
1015 	if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
1016 		etype = ntohs(eh->evl_proto);
1017 		ehdrlen = ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN;
1018 	} else {
1019 		etype = ntohs(eh->evl_encap_proto);
1020 		ehdrlen = ETHER_HDR_LEN;
1021 	}
1022 
1023 	switch (etype) {
1024 #ifdef INET
1025 	case ETHERTYPE_IP:
1026 		ip = (struct ip *)(m->m_data + ehdrlen);
1027 		if (ip->ip_p != IPPROTO_TCP)
1028 			return NULL;
1029 		th = (struct tcphdr *)((caddr_t)ip + (ip->ip_hl << 2));
1030 
1031 		total_len = ehdrlen + (ip->ip_hl << 2) + (th->th_off << 2);
1032 		break;
1033 #endif
1034 #ifdef INET6
1035 	case ETHERTYPE_IPV6:
1036 		ip6 = (struct ip6_hdr *)(m->m_data + ehdrlen);
1037 		if (ip6->ip6_nxt != IPPROTO_TCP)
1038 			return NULL;
1039 		th = (struct tcphdr *)((caddr_t)ip6 + sizeof(struct ip6_hdr));
1040 
1041 		total_len = ehdrlen + sizeof(struct ip6_hdr) + (th->th_off << 2);
1042 		break;
1043 #endif
1044 	default:
1045 		return NULL;
1046 	}
1047 
1048 	m = m_pullup(m, total_len);
1049 	if (!m)
1050 		return NULL;
1051 	*mpp = m;
1052 	return m;
1053 
1054 }
1055 #endif /* INET6 || INET */
1056 
1057 void
1058 oce_tx_task(void *arg, int npending)
1059 {
1060 	struct oce_wq *wq = arg;
1061 	POCE_SOFTC sc = wq->parent;
1062 	struct ifnet *ifp = sc->ifp;
1063 	int rc = 0;
1064 
1065 #if __FreeBSD_version >= 800000
1066 	LOCK(&wq->tx_lock);
1067 	rc = oce_multiq_transmit(ifp, NULL, wq);
1068 	if (rc) {
1069 		device_printf(sc->dev,
1070 				"TX[%d] restart failed\n", wq->queue_index);
1071 	}
1072 	UNLOCK(&wq->tx_lock);
1073 #else
1074 	oce_start(ifp);
1075 #endif
1076 
1077 }
1078 
1079 
1080 void
1081 oce_start(struct ifnet *ifp)
1082 {
1083 	POCE_SOFTC sc = ifp->if_softc;
1084 	struct mbuf *m;
1085 	int rc = 0;
1086 	int def_q = 0; /* Defualt tx queue is 0*/
1087 
1088 	if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
1089 			IFF_DRV_RUNNING)
1090 		return;
1091 
1092 	if (!sc->link_status)
1093 		return;
1094 
1095 	do {
1096 		IF_DEQUEUE(&sc->ifp->if_snd, m);
1097 		if (m == NULL)
1098 			break;
1099 
1100 		LOCK(&sc->wq[def_q]->tx_lock);
1101 		rc = oce_tx(sc, &m, def_q);
1102 		UNLOCK(&sc->wq[def_q]->tx_lock);
1103 		if (rc) {
1104 			if (m != NULL) {
1105 				sc->wq[def_q]->tx_stats.tx_stops ++;
1106 				ifp->if_drv_flags |= IFF_DRV_OACTIVE;
1107 				IFQ_DRV_PREPEND(&ifp->if_snd, m);
1108 				m = NULL;
1109 			}
1110 			break;
1111 		}
1112 		if (m != NULL)
1113 			ETHER_BPF_MTAP(ifp, m);
1114 
1115 	} while (TRUE);
1116 
1117 	return;
1118 }
1119 
1120 
1121 /* Handle the Completion Queue for transmit */
1122 uint16_t
1123 oce_wq_handler(void *arg)
1124 {
1125 	struct oce_wq *wq = (struct oce_wq *)arg;
1126 	POCE_SOFTC sc = wq->parent;
1127 	struct oce_cq *cq = wq->cq;
1128 	struct oce_nic_tx_cqe *cqe;
1129 	int num_cqes = 0;
1130 
1131 	bus_dmamap_sync(cq->ring->dma.tag,
1132 			cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1133 	cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_tx_cqe);
1134 	while (cqe->u0.dw[3]) {
1135 		DW_SWAP((uint32_t *) cqe, sizeof(oce_wq_cqe));
1136 
1137 		wq->ring->cidx = cqe->u0.s.wqe_index + 1;
1138 		if (wq->ring->cidx >= wq->ring->num_items)
1139 			wq->ring->cidx -= wq->ring->num_items;
1140 
1141 		oce_tx_complete(wq, cqe->u0.s.wqe_index, cqe->u0.s.status);
1142 		wq->tx_stats.tx_compl++;
1143 		cqe->u0.dw[3] = 0;
1144 		RING_GET(cq->ring, 1);
1145 		bus_dmamap_sync(cq->ring->dma.tag,
1146 				cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1147 		cqe =
1148 		    RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_tx_cqe);
1149 		num_cqes++;
1150 	}
1151 
1152 	if (num_cqes)
1153 		oce_arm_cq(sc, cq->cq_id, num_cqes, FALSE);
1154 
1155 	return 0;
1156 }
1157 
1158 
1159 static int
1160 oce_multiq_transmit(struct ifnet *ifp, struct mbuf *m, struct oce_wq *wq)
1161 {
1162 	POCE_SOFTC sc = ifp->if_softc;
1163 	int status = 0, queue_index = 0;
1164 	struct mbuf *next = NULL;
1165 	struct buf_ring *br = NULL;
1166 
1167 	br  = wq->br;
1168 	queue_index = wq->queue_index;
1169 
1170 	if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
1171 		IFF_DRV_RUNNING) {
1172 		if (m != NULL)
1173 			status = drbr_enqueue(ifp, br, m);
1174 		return status;
1175 	}
1176 
1177 	 if (m != NULL) {
1178 		if ((status = drbr_enqueue(ifp, br, m)) != 0)
1179 			return status;
1180 	}
1181 	while ((next = drbr_peek(ifp, br)) != NULL) {
1182 		if (oce_tx(sc, &next, queue_index)) {
1183 			if (next == NULL) {
1184 				drbr_advance(ifp, br);
1185 			} else {
1186 				drbr_putback(ifp, br, next);
1187 				wq->tx_stats.tx_stops ++;
1188 				ifp->if_drv_flags |= IFF_DRV_OACTIVE;
1189 				status = drbr_enqueue(ifp, br, next);
1190 			}
1191 			break;
1192 		}
1193 		drbr_advance(ifp, br);
1194 		ifp->if_obytes += next->m_pkthdr.len;
1195 		if (next->m_flags & M_MCAST)
1196 			ifp->if_omcasts++;
1197 		ETHER_BPF_MTAP(ifp, next);
1198 	}
1199 
1200 	return status;
1201 }
1202 
1203 
1204 
1205 
1206 /*****************************************************************************
1207  *			    Receive  routines functions 		     *
1208  *****************************************************************************/
1209 
1210 static void
1211 oce_rx(struct oce_rq *rq, uint32_t rqe_idx, struct oce_nic_rx_cqe *cqe)
1212 {
1213 	uint32_t out;
1214 	struct oce_packet_desc *pd;
1215 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1216 	int i, len, frag_len;
1217 	struct mbuf *m = NULL, *tail = NULL;
1218 	uint16_t vtag;
1219 
1220 	len = cqe->u0.s.pkt_size;
1221 	if (!len) {
1222 		/*partial DMA workaround for Lancer*/
1223 		oce_discard_rx_comp(rq, cqe);
1224 		goto exit;
1225 	}
1226 
1227 	 /* Get vlan_tag value */
1228 	if(IS_BE(sc) || IS_SH(sc))
1229 		vtag = BSWAP_16(cqe->u0.s.vlan_tag);
1230 	else
1231 		vtag = cqe->u0.s.vlan_tag;
1232 
1233 
1234 	for (i = 0; i < cqe->u0.s.num_fragments; i++) {
1235 
1236 		if (rq->packets_out == rq->packets_in) {
1237 			device_printf(sc->dev,
1238 				  "RQ transmit descriptor missing\n");
1239 		}
1240 		out = rq->packets_out + 1;
1241 		if (out == OCE_RQ_PACKET_ARRAY_SIZE)
1242 			out = 0;
1243 		pd = &rq->pckts[rq->packets_out];
1244 		rq->packets_out = out;
1245 
1246 		bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
1247 		bus_dmamap_unload(rq->tag, pd->map);
1248 		rq->pending--;
1249 
1250 		frag_len = (len > rq->cfg.frag_size) ? rq->cfg.frag_size : len;
1251 		pd->mbuf->m_len = frag_len;
1252 
1253 		if (tail != NULL) {
1254 			/* additional fragments */
1255 			pd->mbuf->m_flags &= ~M_PKTHDR;
1256 			tail->m_next = pd->mbuf;
1257 			tail = pd->mbuf;
1258 		} else {
1259 			/* first fragment, fill out much of the packet header */
1260 			pd->mbuf->m_pkthdr.len = len;
1261 			pd->mbuf->m_pkthdr.csum_flags = 0;
1262 			if (IF_CSUM_ENABLED(sc)) {
1263 				if (cqe->u0.s.l4_cksum_pass) {
1264 					pd->mbuf->m_pkthdr.csum_flags |=
1265 					    (CSUM_DATA_VALID | CSUM_PSEUDO_HDR);
1266 					pd->mbuf->m_pkthdr.csum_data = 0xffff;
1267 				}
1268 				if (cqe->u0.s.ip_cksum_pass) {
1269 					if (!cqe->u0.s.ip_ver) { /* IPV4 */
1270 						pd->mbuf->m_pkthdr.csum_flags |=
1271 						(CSUM_IP_CHECKED|CSUM_IP_VALID);
1272 					}
1273 				}
1274 			}
1275 			m = tail = pd->mbuf;
1276 		}
1277 		pd->mbuf = NULL;
1278 		len -= frag_len;
1279 	}
1280 
1281 	if (m) {
1282 		if (!oce_cqe_portid_valid(sc, cqe)) {
1283 			 m_freem(m);
1284 			 goto exit;
1285 		}
1286 
1287 		m->m_pkthdr.rcvif = sc->ifp;
1288 #if __FreeBSD_version >= 800000
1289 		if (rq->queue_index)
1290 			m->m_pkthdr.flowid = (rq->queue_index - 1);
1291 		else
1292 			m->m_pkthdr.flowid = rq->queue_index;
1293 		m->m_flags |= M_FLOWID;
1294 #endif
1295 		/* This deternies if vlan tag is Valid */
1296 		if (oce_cqe_vtp_valid(sc, cqe)) {
1297 			if (sc->function_mode & FNM_FLEX10_MODE) {
1298 				/* FLEX10. If QnQ is not set, neglect VLAN */
1299 				if (cqe->u0.s.qnq) {
1300 					m->m_pkthdr.ether_vtag = vtag;
1301 					m->m_flags |= M_VLANTAG;
1302 				}
1303 			} else if (sc->pvid != (vtag & VLAN_VID_MASK))  {
1304 				/* In UMC mode generally pvid will be striped by
1305 				   hw. But in some cases we have seen it comes
1306 				   with pvid. So if pvid == vlan, neglect vlan.
1307 				*/
1308 				m->m_pkthdr.ether_vtag = vtag;
1309 				m->m_flags |= M_VLANTAG;
1310 			}
1311 		}
1312 
1313 		sc->ifp->if_ipackets++;
1314 #if defined(INET6) || defined(INET)
1315 		/* Try to queue to LRO */
1316 		if (IF_LRO_ENABLED(sc) &&
1317 		    (cqe->u0.s.ip_cksum_pass) &&
1318 		    (cqe->u0.s.l4_cksum_pass) &&
1319 		    (!cqe->u0.s.ip_ver)       &&
1320 		    (rq->lro.lro_cnt != 0)) {
1321 
1322 			if (tcp_lro_rx(&rq->lro, m, 0) == 0) {
1323 				rq->lro_pkts_queued ++;
1324 				goto post_done;
1325 			}
1326 			/* If LRO posting fails then try to post to STACK */
1327 		}
1328 #endif
1329 
1330 		(*sc->ifp->if_input) (sc->ifp, m);
1331 #if defined(INET6) || defined(INET)
1332 post_done:
1333 #endif
1334 		/* Update rx stats per queue */
1335 		rq->rx_stats.rx_pkts++;
1336 		rq->rx_stats.rx_bytes += cqe->u0.s.pkt_size;
1337 		rq->rx_stats.rx_frags += cqe->u0.s.num_fragments;
1338 		if (cqe->u0.s.pkt_type == OCE_MULTICAST_PACKET)
1339 			rq->rx_stats.rx_mcast_pkts++;
1340 		if (cqe->u0.s.pkt_type == OCE_UNICAST_PACKET)
1341 			rq->rx_stats.rx_ucast_pkts++;
1342 	}
1343 exit:
1344 	return;
1345 }
1346 
1347 
1348 static void
1349 oce_discard_rx_comp(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe)
1350 {
1351 	uint32_t out, i = 0;
1352 	struct oce_packet_desc *pd;
1353 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1354 	int num_frags = cqe->u0.s.num_fragments;
1355 
1356 	for (i = 0; i < num_frags; i++) {
1357 		if (rq->packets_out == rq->packets_in) {
1358 			device_printf(sc->dev,
1359 				"RQ transmit descriptor missing\n");
1360 		}
1361 		out = rq->packets_out + 1;
1362 		if (out == OCE_RQ_PACKET_ARRAY_SIZE)
1363 			out = 0;
1364 		pd = &rq->pckts[rq->packets_out];
1365 		rq->packets_out = out;
1366 
1367 		bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE);
1368 		bus_dmamap_unload(rq->tag, pd->map);
1369 		rq->pending--;
1370 		m_freem(pd->mbuf);
1371 	}
1372 
1373 }
1374 
1375 
1376 static int
1377 oce_cqe_vtp_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe)
1378 {
1379 	struct oce_nic_rx_cqe_v1 *cqe_v1;
1380 	int vtp = 0;
1381 
1382 	if (sc->be3_native) {
1383 		cqe_v1 = (struct oce_nic_rx_cqe_v1 *)cqe;
1384 		vtp =  cqe_v1->u0.s.vlan_tag_present;
1385 	} else
1386 		vtp = cqe->u0.s.vlan_tag_present;
1387 
1388 	return vtp;
1389 
1390 }
1391 
1392 
1393 static int
1394 oce_cqe_portid_valid(POCE_SOFTC sc, struct oce_nic_rx_cqe *cqe)
1395 {
1396 	struct oce_nic_rx_cqe_v1 *cqe_v1;
1397 	int port_id = 0;
1398 
1399 	if (sc->be3_native && (IS_BE(sc) || IS_SH(sc))) {
1400 		cqe_v1 = (struct oce_nic_rx_cqe_v1 *)cqe;
1401 		port_id =  cqe_v1->u0.s.port;
1402 		if (sc->port_id != port_id)
1403 			return 0;
1404 	} else
1405 		;/* For BE3 legacy and Lancer this is dummy */
1406 
1407 	return 1;
1408 
1409 }
1410 
1411 #if defined(INET6) || defined(INET)
1412 static void
1413 oce_rx_flush_lro(struct oce_rq *rq)
1414 {
1415 	struct lro_ctrl	*lro = &rq->lro;
1416 	struct lro_entry *queued;
1417 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1418 
1419 	if (!IF_LRO_ENABLED(sc))
1420 		return;
1421 
1422 	while ((queued = SLIST_FIRST(&lro->lro_active)) != NULL) {
1423 		SLIST_REMOVE_HEAD(&lro->lro_active, next);
1424 		tcp_lro_flush(lro, queued);
1425 	}
1426 	rq->lro_pkts_queued = 0;
1427 
1428 	return;
1429 }
1430 
1431 
1432 static int
1433 oce_init_lro(POCE_SOFTC sc)
1434 {
1435 	struct lro_ctrl *lro = NULL;
1436 	int i = 0, rc = 0;
1437 
1438 	for (i = 0; i < sc->nrqs; i++) {
1439 		lro = &sc->rq[i]->lro;
1440 		rc = tcp_lro_init(lro);
1441 		if (rc != 0) {
1442 			device_printf(sc->dev, "LRO init failed\n");
1443 			return rc;
1444 		}
1445 		lro->ifp = sc->ifp;
1446 	}
1447 
1448 	return rc;
1449 }
1450 
1451 
1452 void
1453 oce_free_lro(POCE_SOFTC sc)
1454 {
1455 	struct lro_ctrl *lro = NULL;
1456 	int i = 0;
1457 
1458 	for (i = 0; i < sc->nrqs; i++) {
1459 		lro = &sc->rq[i]->lro;
1460 		if (lro)
1461 			tcp_lro_free(lro);
1462 	}
1463 }
1464 #endif
1465 
1466 int
1467 oce_alloc_rx_bufs(struct oce_rq *rq, int count)
1468 {
1469 	POCE_SOFTC sc = (POCE_SOFTC) rq->parent;
1470 	int i, in, rc;
1471 	struct oce_packet_desc *pd;
1472 	bus_dma_segment_t segs[6];
1473 	int nsegs, added = 0;
1474 	struct oce_nic_rqe *rqe;
1475 	pd_rxulp_db_t rxdb_reg;
1476 
1477 	bzero(&rxdb_reg, sizeof(pd_rxulp_db_t));
1478 	for (i = 0; i < count; i++) {
1479 		in = rq->packets_in + 1;
1480 		if (in == OCE_RQ_PACKET_ARRAY_SIZE)
1481 			in = 0;
1482 		if (in == rq->packets_out)
1483 			break;	/* no more room */
1484 
1485 		pd = &rq->pckts[rq->packets_in];
1486 		pd->mbuf = m_getcl(M_NOWAIT, MT_DATA, M_PKTHDR);
1487 		if (pd->mbuf == NULL)
1488 			break;
1489 
1490 		pd->mbuf->m_len = pd->mbuf->m_pkthdr.len = MCLBYTES;
1491 		rc = bus_dmamap_load_mbuf_sg(rq->tag,
1492 					     pd->map,
1493 					     pd->mbuf,
1494 					     segs, &nsegs, BUS_DMA_NOWAIT);
1495 		if (rc) {
1496 			m_free(pd->mbuf);
1497 			break;
1498 		}
1499 
1500 		if (nsegs != 1) {
1501 			i--;
1502 			continue;
1503 		}
1504 
1505 		rq->packets_in = in;
1506 		bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_PREREAD);
1507 
1508 		rqe = RING_GET_PRODUCER_ITEM_VA(rq->ring, struct oce_nic_rqe);
1509 		rqe->u0.s.frag_pa_hi = ADDR_HI(segs[0].ds_addr);
1510 		rqe->u0.s.frag_pa_lo = ADDR_LO(segs[0].ds_addr);
1511 		DW_SWAP(u32ptr(rqe), sizeof(struct oce_nic_rqe));
1512 		RING_PUT(rq->ring, 1);
1513 		added++;
1514 		rq->pending++;
1515 	}
1516 	if (added != 0) {
1517 		for (i = added / OCE_MAX_RQ_POSTS; i > 0; i--) {
1518 			rxdb_reg.bits.num_posted = OCE_MAX_RQ_POSTS;
1519 			rxdb_reg.bits.qid = rq->rq_id;
1520 			OCE_WRITE_REG32(sc, db, PD_RXULP_DB, rxdb_reg.dw0);
1521 			added -= OCE_MAX_RQ_POSTS;
1522 		}
1523 		if (added > 0) {
1524 			rxdb_reg.bits.qid = rq->rq_id;
1525 			rxdb_reg.bits.num_posted = added;
1526 			OCE_WRITE_REG32(sc, db, PD_RXULP_DB, rxdb_reg.dw0);
1527 		}
1528 	}
1529 
1530 	return 0;
1531 }
1532 
1533 
1534 /* Handle the Completion Queue for receive */
1535 uint16_t
1536 oce_rq_handler(void *arg)
1537 {
1538 	struct oce_rq *rq = (struct oce_rq *)arg;
1539 	struct oce_cq *cq = rq->cq;
1540 	POCE_SOFTC sc = rq->parent;
1541 	struct oce_nic_rx_cqe *cqe;
1542 	int num_cqes = 0, rq_buffers_used = 0;
1543 
1544 
1545 	bus_dmamap_sync(cq->ring->dma.tag,
1546 			cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1547 	cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_rx_cqe);
1548 	while (cqe->u0.dw[2]) {
1549 		DW_SWAP((uint32_t *) cqe, sizeof(oce_rq_cqe));
1550 
1551 		RING_GET(rq->ring, 1);
1552 		if (cqe->u0.s.error == 0) {
1553 			oce_rx(rq, cqe->u0.s.frag_index, cqe);
1554 		} else {
1555 			rq->rx_stats.rxcp_err++;
1556 			sc->ifp->if_ierrors++;
1557 			/* Post L3/L4 errors to stack.*/
1558 			oce_rx(rq, cqe->u0.s.frag_index, cqe);
1559 		}
1560 		rq->rx_stats.rx_compl++;
1561 		cqe->u0.dw[2] = 0;
1562 
1563 #if defined(INET6) || defined(INET)
1564 		if (IF_LRO_ENABLED(sc) && rq->lro_pkts_queued >= 16) {
1565 			oce_rx_flush_lro(rq);
1566 		}
1567 #endif
1568 
1569 		RING_GET(cq->ring, 1);
1570 		bus_dmamap_sync(cq->ring->dma.tag,
1571 				cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
1572 		cqe =
1573 		    RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_rx_cqe);
1574 		num_cqes++;
1575 		if (num_cqes >= (IS_XE201(sc) ? 8 : oce_max_rsp_handled))
1576 			break;
1577 	}
1578 
1579 #if defined(INET6) || defined(INET)
1580 	if (IF_LRO_ENABLED(sc))
1581 		oce_rx_flush_lro(rq);
1582 #endif
1583 
1584 	if (num_cqes) {
1585 		oce_arm_cq(sc, cq->cq_id, num_cqes, FALSE);
1586 		rq_buffers_used = OCE_RQ_PACKET_ARRAY_SIZE - rq->pending;
1587 		if (rq_buffers_used > 1)
1588 			oce_alloc_rx_bufs(rq, (rq_buffers_used - 1));
1589 	}
1590 
1591 	return 0;
1592 
1593 }
1594 
1595 
1596 
1597 
1598 /*****************************************************************************
1599  *		   Helper function prototypes in this file 		     *
1600  *****************************************************************************/
1601 
1602 static int
1603 oce_attach_ifp(POCE_SOFTC sc)
1604 {
1605 
1606 	sc->ifp = if_alloc(IFT_ETHER);
1607 	if (!sc->ifp)
1608 		return ENOMEM;
1609 
1610 	ifmedia_init(&sc->media, IFM_IMASK, oce_media_change, oce_media_status);
1611 	ifmedia_add(&sc->media, IFM_ETHER | IFM_AUTO, 0, NULL);
1612 	ifmedia_set(&sc->media, IFM_ETHER | IFM_AUTO);
1613 
1614 	sc->ifp->if_flags = IFF_BROADCAST | IFF_MULTICAST;
1615 	sc->ifp->if_ioctl = oce_ioctl;
1616 	sc->ifp->if_start = oce_start;
1617 	sc->ifp->if_init = oce_init;
1618 	sc->ifp->if_mtu = ETHERMTU;
1619 	sc->ifp->if_softc = sc;
1620 #if __FreeBSD_version >= 800000
1621 	sc->ifp->if_transmit = oce_multiq_start;
1622 	sc->ifp->if_qflush = oce_multiq_flush;
1623 #endif
1624 
1625 	if_initname(sc->ifp,
1626 		    device_get_name(sc->dev), device_get_unit(sc->dev));
1627 
1628 	sc->ifp->if_snd.ifq_drv_maxlen = OCE_MAX_TX_DESC - 1;
1629 	IFQ_SET_MAXLEN(&sc->ifp->if_snd, sc->ifp->if_snd.ifq_drv_maxlen);
1630 	IFQ_SET_READY(&sc->ifp->if_snd);
1631 
1632 	sc->ifp->if_hwassist = OCE_IF_HWASSIST;
1633 	sc->ifp->if_hwassist |= CSUM_TSO;
1634 	sc->ifp->if_hwassist |= (CSUM_IP | CSUM_TCP | CSUM_UDP);
1635 
1636 	sc->ifp->if_capabilities = OCE_IF_CAPABILITIES;
1637 	sc->ifp->if_capabilities |= IFCAP_HWCSUM;
1638 	sc->ifp->if_capabilities |= IFCAP_VLAN_HWFILTER;
1639 
1640 #if defined(INET6) || defined(INET)
1641 	sc->ifp->if_capabilities |= IFCAP_TSO;
1642 	sc->ifp->if_capabilities |= IFCAP_LRO;
1643 	sc->ifp->if_capabilities |= IFCAP_VLAN_HWTSO;
1644 #endif
1645 
1646 	sc->ifp->if_capenable = sc->ifp->if_capabilities;
1647 	if_initbaudrate(sc->ifp, IF_Gbps(10));
1648 
1649 	ether_ifattach(sc->ifp, sc->macaddr.mac_addr);
1650 
1651 	return 0;
1652 }
1653 
1654 
1655 static void
1656 oce_add_vlan(void *arg, struct ifnet *ifp, uint16_t vtag)
1657 {
1658 	POCE_SOFTC sc = ifp->if_softc;
1659 
1660 	if (ifp->if_softc !=  arg)
1661 		return;
1662 	if ((vtag == 0) || (vtag > 4095))
1663 		return;
1664 
1665 	sc->vlan_tag[vtag] = 1;
1666 	sc->vlans_added++;
1667 	oce_vid_config(sc);
1668 }
1669 
1670 
1671 static void
1672 oce_del_vlan(void *arg, struct ifnet *ifp, uint16_t vtag)
1673 {
1674 	POCE_SOFTC sc = ifp->if_softc;
1675 
1676 	if (ifp->if_softc !=  arg)
1677 		return;
1678 	if ((vtag == 0) || (vtag > 4095))
1679 		return;
1680 
1681 	sc->vlan_tag[vtag] = 0;
1682 	sc->vlans_added--;
1683 	oce_vid_config(sc);
1684 }
1685 
1686 
1687 /*
1688  * A max of 64 vlans can be configured in BE. If the user configures
1689  * more, place the card in vlan promiscuous mode.
1690  */
1691 static int
1692 oce_vid_config(POCE_SOFTC sc)
1693 {
1694 	struct normal_vlan vtags[MAX_VLANFILTER_SIZE];
1695 	uint16_t ntags = 0, i;
1696 	int status = 0;
1697 
1698 	if ((sc->vlans_added <= MAX_VLANFILTER_SIZE) &&
1699 			(sc->ifp->if_capenable & IFCAP_VLAN_HWFILTER)) {
1700 		for (i = 0; i < MAX_VLANS; i++) {
1701 			if (sc->vlan_tag[i]) {
1702 				vtags[ntags].vtag = i;
1703 				ntags++;
1704 			}
1705 		}
1706 		if (ntags)
1707 			status = oce_config_vlan(sc, (uint8_t) sc->if_id,
1708 						vtags, ntags, 1, 0);
1709 	} else
1710 		status = oce_config_vlan(sc, (uint8_t) sc->if_id,
1711 					 	NULL, 0, 1, 1);
1712 	return status;
1713 }
1714 
1715 
1716 static void
1717 oce_mac_addr_set(POCE_SOFTC sc)
1718 {
1719 	uint32_t old_pmac_id = sc->pmac_id;
1720 	int status = 0;
1721 
1722 
1723 	status = bcmp((IF_LLADDR(sc->ifp)), sc->macaddr.mac_addr,
1724 			 sc->macaddr.size_of_struct);
1725 	if (!status)
1726 		return;
1727 
1728 	status = oce_mbox_macaddr_add(sc, (uint8_t *)(IF_LLADDR(sc->ifp)),
1729 					sc->if_id, &sc->pmac_id);
1730 	if (!status) {
1731 		status = oce_mbox_macaddr_del(sc, sc->if_id, old_pmac_id);
1732 		bcopy((IF_LLADDR(sc->ifp)), sc->macaddr.mac_addr,
1733 				 sc->macaddr.size_of_struct);
1734 	}
1735 	if (status)
1736 		device_printf(sc->dev, "Failed update macaddress\n");
1737 
1738 }
1739 
1740 
1741 static int
1742 oce_handle_passthrough(struct ifnet *ifp, caddr_t data)
1743 {
1744 	POCE_SOFTC sc = ifp->if_softc;
1745 	struct ifreq *ifr = (struct ifreq *)data;
1746 	int rc = ENXIO;
1747 	char cookie[32] = {0};
1748 	void *priv_data = (void *)ifr->ifr_data;
1749 	void *ioctl_ptr;
1750 	uint32_t req_size;
1751 	struct mbx_hdr req;
1752 	OCE_DMA_MEM dma_mem;
1753 	struct mbx_common_get_cntl_attr *fw_cmd;
1754 
1755 	if (copyin(priv_data, cookie, strlen(IOCTL_COOKIE)))
1756 		return EFAULT;
1757 
1758 	if (memcmp(cookie, IOCTL_COOKIE, strlen(IOCTL_COOKIE)))
1759 		return EINVAL;
1760 
1761 	ioctl_ptr = (char *)priv_data + strlen(IOCTL_COOKIE);
1762 	if (copyin(ioctl_ptr, &req, sizeof(struct mbx_hdr)))
1763 		return EFAULT;
1764 
1765 	req_size = le32toh(req.u0.req.request_length);
1766 	if (req_size > 65536)
1767 		return EINVAL;
1768 
1769 	req_size += sizeof(struct mbx_hdr);
1770 	rc = oce_dma_alloc(sc, req_size, &dma_mem, 0);
1771 	if (rc)
1772 		return ENOMEM;
1773 
1774 	if (copyin(ioctl_ptr, OCE_DMAPTR(&dma_mem,char), req_size)) {
1775 		rc = EFAULT;
1776 		goto dma_free;
1777 	}
1778 
1779 	rc = oce_pass_through_mbox(sc, &dma_mem, req_size);
1780 	if (rc) {
1781 		rc = EIO;
1782 		goto dma_free;
1783 	}
1784 
1785 	if (copyout(OCE_DMAPTR(&dma_mem,char), ioctl_ptr, req_size))
1786 		rc =  EFAULT;
1787 
1788 	/*
1789 	   firmware is filling all the attributes for this ioctl except
1790 	   the driver version..so fill it
1791 	 */
1792 	if(req.u0.rsp.opcode == OPCODE_COMMON_GET_CNTL_ATTRIBUTES) {
1793 		fw_cmd = (struct mbx_common_get_cntl_attr *) ioctl_ptr;
1794 		strncpy(fw_cmd->params.rsp.cntl_attr_info.hba_attr.drv_ver_str,
1795 			COMPONENT_REVISION, strlen(COMPONENT_REVISION));
1796 	}
1797 
1798 dma_free:
1799 	oce_dma_free(sc, &dma_mem);
1800 	return rc;
1801 
1802 }
1803 
1804 static void
1805 oce_eqd_set_periodic(POCE_SOFTC sc)
1806 {
1807 	struct oce_set_eqd set_eqd[OCE_MAX_EQ];
1808 	struct oce_aic_obj *aic;
1809 	struct oce_eq *eqo;
1810 	uint64_t now = 0, delta;
1811 	int eqd, i, num = 0;
1812 	uint32_t ips = 0;
1813 	int tps;
1814 
1815 	for (i = 0 ; i < sc->neqs; i++) {
1816 		eqo = sc->eq[i];
1817 		aic = &sc->aic_obj[i];
1818 		/* When setting the static eq delay from the user space */
1819 		if (!aic->enable) {
1820 			eqd = aic->et_eqd;
1821 			goto modify_eqd;
1822 		}
1823 
1824 		now = ticks;
1825 
1826 		/* Over flow check */
1827 		if ((now < aic->ticks) || (eqo->intr < aic->intr_prev))
1828 			goto done;
1829 
1830 		delta = now - aic->ticks;
1831 		tps = delta/hz;
1832 
1833 		/* Interrupt rate based on elapsed ticks */
1834 		if(tps)
1835 			ips = (uint32_t)(eqo->intr - aic->intr_prev) / tps;
1836 
1837 		if (ips > INTR_RATE_HWM)
1838 			eqd = aic->cur_eqd + 20;
1839 		else if (ips < INTR_RATE_LWM)
1840 			eqd = aic->cur_eqd / 2;
1841 		else
1842 			goto done;
1843 
1844 		if (eqd < 10)
1845 			eqd = 0;
1846 
1847 		/* Make sure that the eq delay is in the known range */
1848 		eqd = min(eqd, aic->max_eqd);
1849 		eqd = max(eqd, aic->min_eqd);
1850 
1851 modify_eqd:
1852 		if (eqd != aic->cur_eqd) {
1853 			set_eqd[num].delay_multiplier = (eqd * 65)/100;
1854 			set_eqd[num].eq_id = eqo->eq_id;
1855 			aic->cur_eqd = eqd;
1856 			num++;
1857 		}
1858 done:
1859 		aic->intr_prev = eqo->intr;
1860 		aic->ticks = now;
1861 	}
1862 
1863 	/* Is there atleast one eq that needs to be modified? */
1864 	if(num)
1865 		oce_mbox_eqd_modify_periodic(sc, set_eqd, num);
1866 
1867 }
1868 
1869 static void
1870 oce_local_timer(void *arg)
1871 {
1872 	POCE_SOFTC sc = arg;
1873 	int i = 0;
1874 
1875 	oce_refresh_nic_stats(sc);
1876 	oce_refresh_queue_stats(sc);
1877 	oce_mac_addr_set(sc);
1878 
1879 	/* TX Watch Dog*/
1880 	for (i = 0; i < sc->nwqs; i++)
1881 		oce_tx_restart(sc, sc->wq[i]);
1882 
1883 	/* calculate and set the eq delay for optimal interrupt rate */
1884 	if (IS_BE(sc) || IS_SH(sc))
1885 		oce_eqd_set_periodic(sc);
1886 
1887 	callout_reset(&sc->timer, hz, oce_local_timer, sc);
1888 }
1889 
1890 
1891 /* NOTE : This should only be called holding
1892  *        DEVICE_LOCK.
1893 */
1894 static void
1895 oce_if_deactivate(POCE_SOFTC sc)
1896 {
1897 	int i, mtime = 0;
1898 	int wait_req = 0;
1899 	struct oce_rq *rq;
1900 	struct oce_wq *wq;
1901 	struct oce_eq *eq;
1902 
1903 	sc->ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
1904 
1905 	/*Wait for max of 400ms for TX completions to be done */
1906 	while (mtime < 400) {
1907 		wait_req = 0;
1908 		for_all_wq_queues(sc, wq, i) {
1909 			if (wq->ring->num_used) {
1910 				wait_req = 1;
1911 				DELAY(1);
1912 				break;
1913 			}
1914 		}
1915 		mtime += 1;
1916 		if (!wait_req)
1917 			break;
1918 	}
1919 
1920 	/* Stop intrs and finish any bottom halves pending */
1921 	oce_hw_intr_disable(sc);
1922 
1923 	/* Since taskqueue_drain takes a Gaint Lock, We should not acquire
1924 	   any other lock. So unlock device lock and require after
1925 	   completing taskqueue_drain.
1926 	*/
1927 	UNLOCK(&sc->dev_lock);
1928 	for (i = 0; i < sc->intr_count; i++) {
1929 		if (sc->intrs[i].tq != NULL) {
1930 			taskqueue_drain(sc->intrs[i].tq, &sc->intrs[i].task);
1931 		}
1932 	}
1933 	LOCK(&sc->dev_lock);
1934 
1935 	/* Delete RX queue in card with flush param */
1936 	oce_stop_rx(sc);
1937 
1938 	/* Invalidate any pending cq and eq entries*/
1939 	for_all_evnt_queues(sc, eq, i)
1940 		oce_drain_eq(eq);
1941 	for_all_rq_queues(sc, rq, i)
1942 		oce_drain_rq_cq(rq);
1943 	for_all_wq_queues(sc, wq, i)
1944 		oce_drain_wq_cq(wq);
1945 
1946 	/* But still we need to get MCC aync events.
1947 	   So enable intrs and also arm first EQ
1948 	*/
1949 	oce_hw_intr_enable(sc);
1950 	oce_arm_eq(sc, sc->eq[0]->eq_id, 0, TRUE, FALSE);
1951 
1952 	DELAY(10);
1953 }
1954 
1955 
1956 static void
1957 oce_if_activate(POCE_SOFTC sc)
1958 {
1959 	struct oce_eq *eq;
1960 	struct oce_rq *rq;
1961 	struct oce_wq *wq;
1962 	int i, rc = 0;
1963 
1964 	sc->ifp->if_drv_flags |= IFF_DRV_RUNNING;
1965 
1966 	oce_hw_intr_disable(sc);
1967 
1968 	oce_start_rx(sc);
1969 
1970 	for_all_rq_queues(sc, rq, i) {
1971 		rc = oce_start_rq(rq);
1972 		if (rc)
1973 			device_printf(sc->dev, "Unable to start RX\n");
1974 	}
1975 
1976 	for_all_wq_queues(sc, wq, i) {
1977 		rc = oce_start_wq(wq);
1978 		if (rc)
1979 			device_printf(sc->dev, "Unable to start TX\n");
1980 	}
1981 
1982 
1983 	for_all_evnt_queues(sc, eq, i)
1984 		oce_arm_eq(sc, eq->eq_id, 0, TRUE, FALSE);
1985 
1986 	oce_hw_intr_enable(sc);
1987 
1988 }
1989 
1990 static void
1991 process_link_state(POCE_SOFTC sc, struct oce_async_cqe_link_state *acqe)
1992 {
1993 	/* Update Link status */
1994 	if ((acqe->u0.s.link_status & ~ASYNC_EVENT_LOGICAL) ==
1995 	     ASYNC_EVENT_LINK_UP) {
1996 		sc->link_status = ASYNC_EVENT_LINK_UP;
1997 		if_link_state_change(sc->ifp, LINK_STATE_UP);
1998 	} else {
1999 		sc->link_status = ASYNC_EVENT_LINK_DOWN;
2000 		if_link_state_change(sc->ifp, LINK_STATE_DOWN);
2001 	}
2002 
2003 	/* Update speed */
2004 	sc->link_speed = acqe->u0.s.speed;
2005 	sc->qos_link_speed = (uint32_t) acqe->u0.s.qos_link_speed * 10;
2006 
2007 }
2008 
2009 
2010 /* Handle the Completion Queue for the Mailbox/Async notifications */
2011 uint16_t
2012 oce_mq_handler(void *arg)
2013 {
2014 	struct oce_mq *mq = (struct oce_mq *)arg;
2015 	POCE_SOFTC sc = mq->parent;
2016 	struct oce_cq *cq = mq->cq;
2017 	int num_cqes = 0, evt_type = 0, optype = 0;
2018 	struct oce_mq_cqe *cqe;
2019 	struct oce_async_cqe_link_state *acqe;
2020 	struct oce_async_event_grp5_pvid_state *gcqe;
2021 	struct oce_async_event_qnq *dbgcqe;
2022 
2023 
2024 	bus_dmamap_sync(cq->ring->dma.tag,
2025 			cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2026 	cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_mq_cqe);
2027 
2028 	while (cqe->u0.dw[3]) {
2029 		DW_SWAP((uint32_t *) cqe, sizeof(oce_mq_cqe));
2030 		if (cqe->u0.s.async_event) {
2031 			evt_type = cqe->u0.s.event_type;
2032 			optype = cqe->u0.s.async_type;
2033 			if (evt_type  == ASYNC_EVENT_CODE_LINK_STATE) {
2034 				/* Link status evt */
2035 				acqe = (struct oce_async_cqe_link_state *)cqe;
2036 				process_link_state(sc, acqe);
2037 			} else if ((evt_type == ASYNC_EVENT_GRP5) &&
2038 				   (optype == ASYNC_EVENT_PVID_STATE)) {
2039 				/* GRP5 PVID */
2040 				gcqe =
2041 				(struct oce_async_event_grp5_pvid_state *)cqe;
2042 				if (gcqe->enabled)
2043 					sc->pvid = gcqe->tag & VLAN_VID_MASK;
2044 				else
2045 					sc->pvid = 0;
2046 
2047 			}
2048 			else if(evt_type == ASYNC_EVENT_CODE_DEBUG &&
2049 				optype == ASYNC_EVENT_DEBUG_QNQ) {
2050 				dbgcqe =
2051 				(struct oce_async_event_qnq *)cqe;
2052 				if(dbgcqe->valid)
2053 					sc->qnqid = dbgcqe->vlan_tag;
2054 				sc->qnq_debug_event = TRUE;
2055 			}
2056 		}
2057 		cqe->u0.dw[3] = 0;
2058 		RING_GET(cq->ring, 1);
2059 		bus_dmamap_sync(cq->ring->dma.tag,
2060 				cq->ring->dma.map, BUS_DMASYNC_POSTWRITE);
2061 		cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_mq_cqe);
2062 		num_cqes++;
2063 	}
2064 
2065 	if (num_cqes)
2066 		oce_arm_cq(sc, cq->cq_id, num_cqes, FALSE);
2067 
2068 	return 0;
2069 }
2070 
2071 
2072 static void
2073 setup_max_queues_want(POCE_SOFTC sc)
2074 {
2075 	/* Check if it is FLEX machine. Is so dont use RSS */
2076 	if ((sc->function_mode & FNM_FLEX10_MODE) ||
2077 	    (sc->function_mode & FNM_UMC_MODE)    ||
2078 	    (sc->function_mode & FNM_VNIC_MODE)	  ||
2079 	    (!is_rss_enabled(sc))		  ||
2080 	    (sc->flags & OCE_FLAGS_BE2)) {
2081 		sc->nrqs = 1;
2082 		sc->nwqs = 1;
2083 	}
2084 }
2085 
2086 
2087 static void
2088 update_queues_got(POCE_SOFTC sc)
2089 {
2090 	if (is_rss_enabled(sc)) {
2091 		sc->nrqs = sc->intr_count + 1;
2092 		sc->nwqs = sc->intr_count;
2093 	} else {
2094 		sc->nrqs = 1;
2095 		sc->nwqs = 1;
2096 	}
2097 }
2098 
2099 static int
2100 oce_check_ipv6_ext_hdr(struct mbuf *m)
2101 {
2102 	struct ether_header *eh = mtod(m, struct ether_header *);
2103 	caddr_t m_datatemp = m->m_data;
2104 
2105 	if (eh->ether_type == htons(ETHERTYPE_IPV6)) {
2106 		m->m_data += sizeof(struct ether_header);
2107 		struct ip6_hdr *ip6 = mtod(m, struct ip6_hdr *);
2108 
2109 		if((ip6->ip6_nxt != IPPROTO_TCP) && \
2110 				(ip6->ip6_nxt != IPPROTO_UDP)){
2111 			struct ip6_ext *ip6e = NULL;
2112 			m->m_data += sizeof(struct ip6_hdr);
2113 
2114 			ip6e = (struct ip6_ext *) mtod(m, struct ip6_ext *);
2115 			if(ip6e->ip6e_len == 0xff) {
2116 				m->m_data = m_datatemp;
2117 				return TRUE;
2118 			}
2119 		}
2120 		m->m_data = m_datatemp;
2121 	}
2122 	return FALSE;
2123 }
2124 
2125 static int
2126 is_be3_a1(POCE_SOFTC sc)
2127 {
2128 	if((sc->flags & OCE_FLAGS_BE3)  && ((sc->asic_revision & 0xFF) < 2)) {
2129 		return TRUE;
2130 	}
2131 	return FALSE;
2132 }
2133 
2134 static struct mbuf *
2135 oce_insert_vlan_tag(POCE_SOFTC sc, struct mbuf *m, boolean_t *complete)
2136 {
2137 	uint16_t vlan_tag = 0;
2138 
2139 	if(!M_WRITABLE(m))
2140 		return NULL;
2141 
2142 	/* Embed vlan tag in the packet if it is not part of it */
2143 	if(m->m_flags & M_VLANTAG) {
2144 		vlan_tag = EVL_VLANOFTAG(m->m_pkthdr.ether_vtag);
2145 		m->m_flags &= ~M_VLANTAG;
2146 	}
2147 
2148 	/* if UMC, ignore vlan tag insertion and instead insert pvid */
2149 	if(sc->pvid) {
2150 		if(!vlan_tag)
2151 			vlan_tag = sc->pvid;
2152 		*complete = FALSE;
2153 	}
2154 
2155 	if(vlan_tag) {
2156 		m = ether_vlanencap(m, vlan_tag);
2157 	}
2158 
2159 	if(sc->qnqid) {
2160 		m = ether_vlanencap(m, sc->qnqid);
2161 		*complete = FALSE;
2162 	}
2163 	return m;
2164 }
2165 
2166 static int
2167 oce_tx_asic_stall_verify(POCE_SOFTC sc, struct mbuf *m)
2168 {
2169 	if(is_be3_a1(sc) && IS_QNQ_OR_UMC(sc) && \
2170 			oce_check_ipv6_ext_hdr(m)) {
2171 		return TRUE;
2172 	}
2173 	return FALSE;
2174 }
2175 
2176 static void
2177 oce_get_config(POCE_SOFTC sc)
2178 {
2179 	int rc = 0;
2180 	uint32_t max_rss = 0;
2181 
2182 	if ((IS_BE(sc) || IS_SH(sc)) && (!sc->be3_native))
2183 		max_rss = OCE_LEGACY_MODE_RSS;
2184 	else
2185 		max_rss = OCE_MAX_RSS;
2186 
2187 	if (!IS_BE(sc)) {
2188 		rc = oce_get_func_config(sc);
2189 		if (rc) {
2190 			sc->nwqs = OCE_MAX_WQ;
2191 			sc->nrssqs = max_rss;
2192 			sc->nrqs = sc->nrssqs + 1;
2193 		}
2194 	}
2195 	else {
2196 		rc = oce_get_profile_config(sc);
2197 		sc->nrssqs = max_rss;
2198 		sc->nrqs = sc->nrssqs + 1;
2199 		if (rc)
2200 			sc->nwqs = OCE_MAX_WQ;
2201 	}
2202 }
2203