xref: /freebsd/sys/dev/ixgbe/if_ixv.c (revision 97cb52fa9aefd90fad38790fded50905aeeb9b9e)
1 /******************************************************************************
2 
3   Copyright (c) 2001-2017, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 
36 #ifndef IXGBE_STANDALONE_BUILD
37 #include "opt_inet.h"
38 #include "opt_inet6.h"
39 #endif
40 
41 #include "ixgbe.h"
42 
43 /************************************************************************
44  * Driver version
45  ************************************************************************/
46 char ixv_driver_version[] = "1.5.13-k";
47 
48 /************************************************************************
49  * PCI Device ID Table
50  *
51  *   Used by probe to select devices to load on
52  *   Last field stores an index into ixv_strings
53  *   Last entry must be all 0s
54  *
55  *   { Vendor ID, Device ID, SubVendor ID, SubDevice ID, String Index }
56  ************************************************************************/
57 static ixgbe_vendor_info_t ixv_vendor_info_array[] =
58 {
59 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_VF, 0, 0, 0},
60 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X540_VF, 0, 0, 0},
61 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550_VF, 0, 0, 0},
62 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_VF, 0, 0, 0},
63 	{IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_VF, 0, 0, 0},
64 	/* required last entry */
65 	{0, 0, 0, 0, 0}
66 };
67 
68 /************************************************************************
69  * Table of branding strings
70  ************************************************************************/
71 static char *ixv_strings[] = {
72 	"Intel(R) PRO/10GbE Virtual Function Network Driver"
73 };
74 
75 /************************************************************************
76  * Function prototypes
77  ************************************************************************/
78 static int      ixv_probe(device_t);
79 static int      ixv_attach(device_t);
80 static int      ixv_detach(device_t);
81 static int      ixv_shutdown(device_t);
82 static int      ixv_ioctl(struct ifnet *, u_long, caddr_t);
83 static void     ixv_init(void *);
84 static void     ixv_init_locked(struct adapter *);
85 static void     ixv_stop(void *);
86 static uint64_t ixv_get_counter(struct ifnet *, ift_counter);
87 static void     ixv_init_device_features(struct adapter *);
88 static void     ixv_media_status(struct ifnet *, struct ifmediareq *);
89 static int      ixv_media_change(struct ifnet *);
90 static int      ixv_allocate_pci_resources(struct adapter *);
91 static int      ixv_allocate_msix(struct adapter *);
92 static int      ixv_configure_interrupts(struct adapter *);
93 static void     ixv_free_pci_resources(struct adapter *);
94 static void     ixv_local_timer(void *);
95 static void     ixv_setup_interface(device_t, struct adapter *);
96 static int      ixv_negotiate_api(struct adapter *);
97 
98 static void     ixv_initialize_transmit_units(struct adapter *);
99 static void     ixv_initialize_receive_units(struct adapter *);
100 static void     ixv_initialize_rss_mapping(struct adapter *);
101 static void     ixv_check_link(struct adapter *);
102 
103 static void     ixv_enable_intr(struct adapter *);
104 static void     ixv_disable_intr(struct adapter *);
105 static void     ixv_set_multi(struct adapter *);
106 static void     ixv_update_link_status(struct adapter *);
107 static int      ixv_sysctl_debug(SYSCTL_HANDLER_ARGS);
108 static void     ixv_set_ivar(struct adapter *, u8, u8, s8);
109 static void     ixv_configure_ivars(struct adapter *);
110 static u8       *ixv_mc_array_itr(struct ixgbe_hw *, u8 **, u32 *);
111 
112 static void     ixv_setup_vlan_support(struct adapter *);
113 static void     ixv_register_vlan(void *, struct ifnet *, u16);
114 static void     ixv_unregister_vlan(void *, struct ifnet *, u16);
115 
116 static void     ixv_save_stats(struct adapter *);
117 static void     ixv_init_stats(struct adapter *);
118 static void     ixv_update_stats(struct adapter *);
119 static void     ixv_add_stats_sysctls(struct adapter *);
120 static void     ixv_set_sysctl_value(struct adapter *, const char *,
121                                      const char *, int *, int);
122 
123 /* The MSI-X Interrupt handlers */
124 static void     ixv_msix_que(void *);
125 static void     ixv_msix_mbx(void *);
126 
127 /* Deferred interrupt tasklets */
128 static void     ixv_handle_que(void *, int);
129 static void     ixv_handle_link(void *, int);
130 
131 /************************************************************************
132  * FreeBSD Device Interface Entry Points
133  ************************************************************************/
134 static device_method_t ixv_methods[] = {
135 	/* Device interface */
136 	DEVMETHOD(device_probe, ixv_probe),
137 	DEVMETHOD(device_attach, ixv_attach),
138 	DEVMETHOD(device_detach, ixv_detach),
139 	DEVMETHOD(device_shutdown, ixv_shutdown),
140 	DEVMETHOD_END
141 };
142 
143 static driver_t ixv_driver = {
144 	"ixv", ixv_methods, sizeof(struct adapter),
145 };
146 
147 devclass_t ixv_devclass;
148 DRIVER_MODULE(ixv, pci, ixv_driver, ixv_devclass, 0, 0);
149 MODULE_DEPEND(ixv, pci, 1, 1, 1);
150 MODULE_DEPEND(ixv, ether, 1, 1, 1);
151 MODULE_DEPEND(ixv, netmap, 1, 1, 1);
152 
153 /*
154  * TUNEABLE PARAMETERS:
155  */
156 
157 /* Number of Queues - do not exceed MSI-X vectors - 1 */
158 static int ixv_num_queues = 1;
159 TUNABLE_INT("hw.ixv.num_queues", &ixv_num_queues);
160 
161 /*
162  * AIM: Adaptive Interrupt Moderation
163  * which means that the interrupt rate
164  * is varied over time based on the
165  * traffic for that interrupt vector
166  */
167 static int ixv_enable_aim = FALSE;
168 TUNABLE_INT("hw.ixv.enable_aim", &ixv_enable_aim);
169 
170 /* How many packets rxeof tries to clean at a time */
171 static int ixv_rx_process_limit = 256;
172 TUNABLE_INT("hw.ixv.rx_process_limit", &ixv_rx_process_limit);
173 
174 /* How many packets txeof tries to clean at a time */
175 static int ixv_tx_process_limit = 256;
176 TUNABLE_INT("hw.ixv.tx_process_limit", &ixv_tx_process_limit);
177 
178 /* Flow control setting, default to full */
179 static int ixv_flow_control = ixgbe_fc_full;
180 TUNABLE_INT("hw.ixv.flow_control", &ixv_flow_control);
181 
182 /*
183  * Header split: this causes the hardware to DMA
184  * the header into a separate mbuf from the payload,
185  * it can be a performance win in some workloads, but
186  * in others it actually hurts, its off by default.
187  */
188 static int ixv_header_split = FALSE;
189 TUNABLE_INT("hw.ixv.hdr_split", &ixv_header_split);
190 
191 /*
192  * Number of TX descriptors per ring,
193  * setting higher than RX as this seems
194  * the better performing choice.
195  */
196 static int ixv_txd = DEFAULT_TXD;
197 TUNABLE_INT("hw.ixv.txd", &ixv_txd);
198 
199 /* Number of RX descriptors per ring */
200 static int ixv_rxd = DEFAULT_RXD;
201 TUNABLE_INT("hw.ixv.rxd", &ixv_rxd);
202 
203 /* Legacy Transmit (single queue) */
204 static int ixv_enable_legacy_tx = 0;
205 TUNABLE_INT("hw.ixv.enable_legacy_tx", &ixv_enable_legacy_tx);
206 
207 /*
208  * Shadow VFTA table, this is needed because
209  * the real filter table gets cleared during
210  * a soft reset and we need to repopulate it.
211  */
212 static u32 ixv_shadow_vfta[IXGBE_VFTA_SIZE];
213 
214 static int (*ixv_start_locked)(struct ifnet *, struct tx_ring *);
215 static int (*ixv_ring_empty)(struct ifnet *, struct buf_ring *);
216 
217 /************************************************************************
218  * ixv_probe - Device identification routine
219  *
220  *   Determines if the driver should be loaded on
221  *   adapter based on its PCI vendor/device ID.
222  *
223  *   return BUS_PROBE_DEFAULT on success, positive on failure
224  ************************************************************************/
225 static int
226 ixv_probe(device_t dev)
227 {
228 	ixgbe_vendor_info_t *ent;
229 	u16                 pci_vendor_id = 0;
230 	u16                 pci_device_id = 0;
231 	u16                 pci_subvendor_id = 0;
232 	u16                 pci_subdevice_id = 0;
233 	char                adapter_name[256];
234 
235 
236 	pci_vendor_id = pci_get_vendor(dev);
237 	if (pci_vendor_id != IXGBE_INTEL_VENDOR_ID)
238 		return (ENXIO);
239 
240 	pci_device_id = pci_get_device(dev);
241 	pci_subvendor_id = pci_get_subvendor(dev);
242 	pci_subdevice_id = pci_get_subdevice(dev);
243 
244 	ent = ixv_vendor_info_array;
245 	while (ent->vendor_id != 0) {
246 		if ((pci_vendor_id == ent->vendor_id) &&
247 		    (pci_device_id == ent->device_id) &&
248 		    ((pci_subvendor_id == ent->subvendor_id) ||
249 		     (ent->subvendor_id == 0)) &&
250 		    ((pci_subdevice_id == ent->subdevice_id) ||
251 		     (ent->subdevice_id == 0))) {
252 			sprintf(adapter_name, "%s, Version - %s",
253 			    ixv_strings[ent->index], ixv_driver_version);
254 			device_set_desc_copy(dev, adapter_name);
255 			return (BUS_PROBE_DEFAULT);
256 		}
257 		ent++;
258 	}
259 
260 	return (ENXIO);
261 } /* ixv_probe */
262 
263 /************************************************************************
264  * ixv_attach - Device initialization routine
265  *
266  *   Called when the driver is being loaded.
267  *   Identifies the type of hardware, allocates all resources
268  *   and initializes the hardware.
269  *
270  *   return 0 on success, positive on failure
271  ************************************************************************/
272 static int
273 ixv_attach(device_t dev)
274 {
275 	struct adapter  *adapter;
276 	struct ixgbe_hw *hw;
277 	int             error = 0;
278 
279 	INIT_DEBUGOUT("ixv_attach: begin");
280 
281 	/*
282 	 * Make sure BUSMASTER is set, on a VM under
283 	 * KVM it may not be and will break things.
284 	 */
285 	pci_enable_busmaster(dev);
286 
287 	/* Allocate, clear, and link in our adapter structure */
288 	adapter = device_get_softc(dev);
289 	adapter->dev = dev;
290 	adapter->hw.back = adapter;
291 	hw = &adapter->hw;
292 
293 	adapter->init_locked = ixv_init_locked;
294 	adapter->stop_locked = ixv_stop;
295 
296 	/* Core Lock Init*/
297 	IXGBE_CORE_LOCK_INIT(adapter, device_get_nameunit(dev));
298 
299 	/* Do base PCI setup - map BAR0 */
300 	if (ixv_allocate_pci_resources(adapter)) {
301 		device_printf(dev, "ixv_allocate_pci_resources() failed!\n");
302 		error = ENXIO;
303 		goto err_out;
304 	}
305 
306 	/* SYSCTL APIs */
307 	SYSCTL_ADD_PROC(device_get_sysctl_ctx(dev),
308 	    SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), OID_AUTO, "debug",
309 	    CTLTYPE_INT | CTLFLAG_RW, adapter, 0, ixv_sysctl_debug, "I",
310 	    "Debug Info");
311 
312 	SYSCTL_ADD_INT(device_get_sysctl_ctx(dev),
313 	    SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), OID_AUTO,
314 	    "enable_aim", CTLFLAG_RW, &ixv_enable_aim, 1,
315 	    "Interrupt Moderation");
316 
317 	/* Set up the timer callout */
318 	callout_init_mtx(&adapter->timer, &adapter->core_mtx, 0);
319 
320 	/* Save off the information about this board */
321 	hw->vendor_id = pci_get_vendor(dev);
322 	hw->device_id = pci_get_device(dev);
323 	hw->revision_id = pci_get_revid(dev);
324 	hw->subsystem_vendor_id = pci_get_subvendor(dev);
325 	hw->subsystem_device_id = pci_get_subdevice(dev);
326 
327 	/* A subset of set_mac_type */
328 	switch (hw->device_id) {
329 	case IXGBE_DEV_ID_82599_VF:
330 		hw->mac.type = ixgbe_mac_82599_vf;
331 		break;
332 	case IXGBE_DEV_ID_X540_VF:
333 		hw->mac.type = ixgbe_mac_X540_vf;
334 		break;
335 	case IXGBE_DEV_ID_X550_VF:
336 		hw->mac.type = ixgbe_mac_X550_vf;
337 		break;
338 	case IXGBE_DEV_ID_X550EM_X_VF:
339 		hw->mac.type = ixgbe_mac_X550EM_x_vf;
340 		break;
341 	case IXGBE_DEV_ID_X550EM_A_VF:
342 		hw->mac.type = ixgbe_mac_X550EM_a_vf;
343 		break;
344 	default:
345 		/* Shouldn't get here since probe succeeded */
346 		device_printf(dev, "Unknown device ID!\n");
347 		error = ENXIO;
348 		goto err_out;
349 		break;
350 	}
351 
352 	ixv_init_device_features(adapter);
353 
354 	/* Initialize the shared code */
355 	error = ixgbe_init_ops_vf(hw);
356 	if (error) {
357 		device_printf(dev, "ixgbe_init_ops_vf() failed!\n");
358 		error = EIO;
359 		goto err_out;
360 	}
361 
362 	/* Setup the mailbox */
363 	ixgbe_init_mbx_params_vf(hw);
364 
365 	/* Set the right number of segments */
366 	adapter->num_segs = IXGBE_82599_SCATTER;
367 
368 	error = hw->mac.ops.reset_hw(hw);
369 	if (error == IXGBE_ERR_RESET_FAILED)
370 		device_printf(dev, "...reset_hw() failure: Reset Failed!\n");
371 	else if (error)
372 		device_printf(dev, "...reset_hw() failed with error %d\n",
373 		    error);
374 	if (error) {
375 		error = EIO;
376 		goto err_out;
377 	}
378 
379 	error = hw->mac.ops.init_hw(hw);
380 	if (error) {
381 		device_printf(dev, "...init_hw() failed with error %d\n",
382 		    error);
383 		error = EIO;
384 		goto err_out;
385 	}
386 
387 	/* Negotiate mailbox API version */
388 	error = ixv_negotiate_api(adapter);
389 	if (error) {
390 		device_printf(dev,
391 		    "Mailbox API negotiation failed during attach!\n");
392 		goto err_out;
393 	}
394 
395 	/* If no mac address was assigned, make a random one */
396 	if (!ixv_check_ether_addr(hw->mac.addr)) {
397 		u8 addr[ETHER_ADDR_LEN];
398 		arc4rand(&addr, sizeof(addr), 0);
399 		addr[0] &= 0xFE;
400 		addr[0] |= 0x02;
401 		bcopy(addr, hw->mac.addr, sizeof(addr));
402 		bcopy(addr, hw->mac.perm_addr, sizeof(addr));
403 	}
404 
405 	/* Register for VLAN events */
406 	adapter->vlan_attach = EVENTHANDLER_REGISTER(vlan_config,
407 	    ixv_register_vlan, adapter, EVENTHANDLER_PRI_FIRST);
408 	adapter->vlan_detach = EVENTHANDLER_REGISTER(vlan_unconfig,
409 	    ixv_unregister_vlan, adapter, EVENTHANDLER_PRI_FIRST);
410 
411 	/* Sysctls for limiting the amount of work done in the taskqueues */
412 	ixv_set_sysctl_value(adapter, "rx_processing_limit",
413 	    "max number of rx packets to process",
414 	    &adapter->rx_process_limit, ixv_rx_process_limit);
415 
416 	ixv_set_sysctl_value(adapter, "tx_processing_limit",
417 	    "max number of tx packets to process",
418 	    &adapter->tx_process_limit, ixv_tx_process_limit);
419 
420 	/* Do descriptor calc and sanity checks */
421 	if (((ixv_txd * sizeof(union ixgbe_adv_tx_desc)) % DBA_ALIGN) != 0 ||
422 	    ixv_txd < MIN_TXD || ixv_txd > MAX_TXD) {
423 		device_printf(dev, "TXD config issue, using default!\n");
424 		adapter->num_tx_desc = DEFAULT_TXD;
425 	} else
426 		adapter->num_tx_desc = ixv_txd;
427 
428 	if (((ixv_rxd * sizeof(union ixgbe_adv_rx_desc)) % DBA_ALIGN) != 0 ||
429 	    ixv_rxd < MIN_RXD || ixv_rxd > MAX_RXD) {
430 		device_printf(dev, "RXD config issue, using default!\n");
431 		adapter->num_rx_desc = DEFAULT_RXD;
432 	} else
433 		adapter->num_rx_desc = ixv_rxd;
434 
435 	/* Setup MSI-X */
436 	error = ixv_configure_interrupts(adapter);
437 	if (error)
438 		goto err_out;
439 
440 	/* Allocate our TX/RX Queues */
441 	if (ixgbe_allocate_queues(adapter)) {
442 		device_printf(dev, "ixgbe_allocate_queues() failed!\n");
443 		error = ENOMEM;
444 		goto err_out;
445 	}
446 
447 	/* Setup OS specific network interface */
448 	ixv_setup_interface(dev, adapter);
449 
450 	error = ixv_allocate_msix(adapter);
451 	if (error) {
452 		device_printf(dev, "ixv_allocate_msix() failed!\n");
453 		goto err_late;
454 	}
455 
456 	/* Do the stats setup */
457 	ixv_save_stats(adapter);
458 	ixv_init_stats(adapter);
459 	ixv_add_stats_sysctls(adapter);
460 
461 	if (adapter->feat_en & IXGBE_FEATURE_NETMAP)
462 		ixgbe_netmap_attach(adapter);
463 
464 	INIT_DEBUGOUT("ixv_attach: end");
465 
466 	return (0);
467 
468 err_late:
469 	ixgbe_free_transmit_structures(adapter);
470 	ixgbe_free_receive_structures(adapter);
471 	free(adapter->queues, M_DEVBUF);
472 err_out:
473 	ixv_free_pci_resources(adapter);
474 	IXGBE_CORE_LOCK_DESTROY(adapter);
475 
476 	return (error);
477 } /* ixv_attach */
478 
479 /************************************************************************
480  * ixv_detach - Device removal routine
481  *
482  *   Called when the driver is being removed.
483  *   Stops the adapter and deallocates all the resources
484  *   that were allocated for driver operation.
485  *
486  *   return 0 on success, positive on failure
487  ************************************************************************/
488 static int
489 ixv_detach(device_t dev)
490 {
491 	struct adapter  *adapter = device_get_softc(dev);
492 	struct ix_queue *que = adapter->queues;
493 
494 	INIT_DEBUGOUT("ixv_detach: begin");
495 
496 	/* Make sure VLANS are not using driver */
497 	if (adapter->ifp->if_vlantrunk != NULL) {
498 		device_printf(dev, "Vlan in use, detach first\n");
499 		return (EBUSY);
500 	}
501 
502 	ether_ifdetach(adapter->ifp);
503 	IXGBE_CORE_LOCK(adapter);
504 	ixv_stop(adapter);
505 	IXGBE_CORE_UNLOCK(adapter);
506 
507 	for (int i = 0; i < adapter->num_queues; i++, que++) {
508 		if (que->tq) {
509 			struct tx_ring  *txr = que->txr;
510 			taskqueue_drain(que->tq, &txr->txq_task);
511 			taskqueue_drain(que->tq, &que->que_task);
512 			taskqueue_free(que->tq);
513 		}
514 	}
515 
516 	/* Drain the Mailbox(link) queue */
517 	if (adapter->tq) {
518 		taskqueue_drain(adapter->tq, &adapter->link_task);
519 		taskqueue_free(adapter->tq);
520 	}
521 
522 	/* Unregister VLAN events */
523 	if (adapter->vlan_attach != NULL)
524 		EVENTHANDLER_DEREGISTER(vlan_config, adapter->vlan_attach);
525 	if (adapter->vlan_detach != NULL)
526 		EVENTHANDLER_DEREGISTER(vlan_unconfig, adapter->vlan_detach);
527 
528 	callout_drain(&adapter->timer);
529 
530 	if (adapter->feat_en & IXGBE_FEATURE_NETMAP)
531 		netmap_detach(adapter->ifp);
532 
533 	ixv_free_pci_resources(adapter);
534 	bus_generic_detach(dev);
535 	if_free(adapter->ifp);
536 
537 	ixgbe_free_transmit_structures(adapter);
538 	ixgbe_free_receive_structures(adapter);
539 	free(adapter->queues, M_DEVBUF);
540 
541 	IXGBE_CORE_LOCK_DESTROY(adapter);
542 
543 	return (0);
544 } /* ixv_detach */
545 
546 /************************************************************************
547  * ixv_init_locked - Init entry point
548  *
549  *   Used in two ways: It is used by the stack as an init entry
550  *   point in network interface structure. It is also used
551  *   by the driver as a hw/sw initialization routine to get
552  *   to a consistent state.
553  *
554  *   return 0 on success, positive on failure
555  ************************************************************************/
556 void
557 ixv_init_locked(struct adapter *adapter)
558 {
559 	struct ifnet    *ifp = adapter->ifp;
560 	device_t        dev = adapter->dev;
561 	struct ixgbe_hw *hw = &adapter->hw;
562 	int             error = 0;
563 
564 	INIT_DEBUGOUT("ixv_init_locked: begin");
565 	mtx_assert(&adapter->core_mtx, MA_OWNED);
566 	hw->adapter_stopped = FALSE;
567 	hw->mac.ops.stop_adapter(hw);
568 	callout_stop(&adapter->timer);
569 
570 	/* reprogram the RAR[0] in case user changed it. */
571 	hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
572 
573 	/* Get the latest mac address, User can use a LAA */
574 	bcopy(IF_LLADDR(adapter->ifp), hw->mac.addr,
575 	    IXGBE_ETH_LENGTH_OF_ADDRESS);
576 	hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, 1);
577 
578 	/* Prepare transmit descriptors and buffers */
579 	if (ixgbe_setup_transmit_structures(adapter)) {
580 		device_printf(dev, "Could not setup transmit structures\n");
581 		ixv_stop(adapter);
582 		return;
583 	}
584 
585 	/* Reset VF and renegotiate mailbox API version */
586 	hw->mac.ops.reset_hw(hw);
587 	error = ixv_negotiate_api(adapter);
588 	if (error) {
589 		device_printf(dev,
590 		    "Mailbox API negotiation failed in init_locked!\n");
591 		return;
592 	}
593 
594 	ixv_initialize_transmit_units(adapter);
595 
596 	/* Setup Multicast table */
597 	ixv_set_multi(adapter);
598 
599 	/*
600 	 * Determine the correct mbuf pool
601 	 * for doing jumbo/headersplit
602 	 */
603 	if (ifp->if_mtu > ETHERMTU)
604 		adapter->rx_mbuf_sz = MJUMPAGESIZE;
605 	else
606 		adapter->rx_mbuf_sz = MCLBYTES;
607 
608 	/* Prepare receive descriptors and buffers */
609 	if (ixgbe_setup_receive_structures(adapter)) {
610 		device_printf(dev, "Could not setup receive structures\n");
611 		ixv_stop(adapter);
612 		return;
613 	}
614 
615 	/* Configure RX settings */
616 	ixv_initialize_receive_units(adapter);
617 
618 	/* Set the various hardware offload abilities */
619 	ifp->if_hwassist = 0;
620 	if (ifp->if_capenable & IFCAP_TSO4)
621 		ifp->if_hwassist |= CSUM_TSO;
622 	if (ifp->if_capenable & IFCAP_TXCSUM) {
623 		ifp->if_hwassist |= (CSUM_TCP | CSUM_UDP);
624 #if __FreeBSD_version >= 800000
625 		ifp->if_hwassist |= CSUM_SCTP;
626 #endif
627 	}
628 
629 	/* Set up VLAN offload and filter */
630 	ixv_setup_vlan_support(adapter);
631 
632 	/* Set up MSI-X routing */
633 	ixv_configure_ivars(adapter);
634 
635 	/* Set up auto-mask */
636 	IXGBE_WRITE_REG(hw, IXGBE_VTEIAM, IXGBE_EICS_RTX_QUEUE);
637 
638 	/* Set moderation on the Link interrupt */
639 	IXGBE_WRITE_REG(hw, IXGBE_VTEITR(adapter->vector), IXGBE_LINK_ITR);
640 
641 	/* Stats init */
642 	ixv_init_stats(adapter);
643 
644 	/* Config/Enable Link */
645 	hw->mac.ops.check_link(hw, &adapter->link_speed, &adapter->link_up,
646 	    FALSE);
647 
648 	/* Start watchdog */
649 	callout_reset(&adapter->timer, hz, ixv_local_timer, adapter);
650 
651 	/* And now turn on interrupts */
652 	ixv_enable_intr(adapter);
653 
654 	/* Now inform the stack we're ready */
655 	ifp->if_drv_flags |= IFF_DRV_RUNNING;
656 	ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
657 
658 	return;
659 } /* ixv_init_locked */
660 
661 /*
662  * MSI-X Interrupt Handlers and Tasklets
663  */
664 
665 static inline void
666 ixv_enable_queue(struct adapter *adapter, u32 vector)
667 {
668 	struct ixgbe_hw *hw = &adapter->hw;
669 	u32             queue = 1 << vector;
670 	u32             mask;
671 
672 	mask = (IXGBE_EIMS_RTX_QUEUE & queue);
673 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, mask);
674 } /* ixv_enable_queue */
675 
676 static inline void
677 ixv_disable_queue(struct adapter *adapter, u32 vector)
678 {
679 	struct ixgbe_hw *hw = &adapter->hw;
680 	u64             queue = (u64)(1 << vector);
681 	u32             mask;
682 
683 	mask = (IXGBE_EIMS_RTX_QUEUE & queue);
684 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMC, mask);
685 } /* ixv_disable_queue */
686 
687 static inline void
688 ixv_rearm_queues(struct adapter *adapter, u64 queues)
689 {
690 	u32 mask = (IXGBE_EIMS_RTX_QUEUE & queues);
691 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEICS, mask);
692 } /* ixv_rearm_queues */
693 
694 
695 /************************************************************************
696  * ixv_msix_que - MSI Queue Interrupt Service routine
697  ************************************************************************/
698 void
699 ixv_msix_que(void *arg)
700 {
701 	struct ix_queue *que = arg;
702 	struct adapter  *adapter = que->adapter;
703 	struct ifnet    *ifp = adapter->ifp;
704 	struct tx_ring  *txr = que->txr;
705 	struct rx_ring  *rxr = que->rxr;
706 	bool            more;
707 	u32             newitr = 0;
708 
709 	ixv_disable_queue(adapter, que->msix);
710 	++que->irqs;
711 
712 	more = ixgbe_rxeof(que);
713 
714 	IXGBE_TX_LOCK(txr);
715 	ixgbe_txeof(txr);
716 	/*
717 	 * Make certain that if the stack
718 	 * has anything queued the task gets
719 	 * scheduled to handle it.
720 	 */
721 	if (!ixv_ring_empty(adapter->ifp, txr->br))
722 		ixv_start_locked(ifp, txr);
723 	IXGBE_TX_UNLOCK(txr);
724 
725 	/* Do AIM now? */
726 
727 	if (ixv_enable_aim == FALSE)
728 		goto no_calc;
729 	/*
730 	 * Do Adaptive Interrupt Moderation:
731 	 *  - Write out last calculated setting
732 	 *  - Calculate based on average size over
733 	 *    the last interval.
734 	 */
735 	if (que->eitr_setting)
736 		IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEITR(que->msix),
737 		    que->eitr_setting);
738 
739 	que->eitr_setting = 0;
740 
741 	/* Idle, do nothing */
742 	if ((txr->bytes == 0) && (rxr->bytes == 0))
743 		goto no_calc;
744 
745 	if ((txr->bytes) && (txr->packets))
746 		newitr = txr->bytes/txr->packets;
747 	if ((rxr->bytes) && (rxr->packets))
748 		newitr = max(newitr, (rxr->bytes / rxr->packets));
749 	newitr += 24; /* account for hardware frame, crc */
750 
751 	/* set an upper boundary */
752 	newitr = min(newitr, 3000);
753 
754 	/* Be nice to the mid range */
755 	if ((newitr > 300) && (newitr < 1200))
756 		newitr = (newitr / 3);
757 	else
758 		newitr = (newitr / 2);
759 
760 	newitr |= newitr << 16;
761 
762 	/* save for next interrupt */
763 	que->eitr_setting = newitr;
764 
765 	/* Reset state */
766 	txr->bytes = 0;
767 	txr->packets = 0;
768 	rxr->bytes = 0;
769 	rxr->packets = 0;
770 
771 no_calc:
772 	if (more)
773 		taskqueue_enqueue(que->tq, &que->que_task);
774 	else /* Re-enable this interrupt */
775 		ixv_enable_queue(adapter, que->msix);
776 
777 	return;
778 } /* ixv_msix_que */
779 
780 /************************************************************************
781  * ixv_msix_mbx
782  ************************************************************************/
783 static void
784 ixv_msix_mbx(void *arg)
785 {
786 	struct adapter  *adapter = arg;
787 	struct ixgbe_hw *hw = &adapter->hw;
788 	u32             reg;
789 
790 	++adapter->link_irq;
791 
792 	/* First get the cause */
793 	reg = IXGBE_READ_REG(hw, IXGBE_VTEICS);
794 	/* Clear interrupt with write */
795 	IXGBE_WRITE_REG(hw, IXGBE_VTEICR, reg);
796 
797 	/* Link status change */
798 	if (reg & IXGBE_EICR_LSC)
799 		taskqueue_enqueue(adapter->tq, &adapter->link_task);
800 
801 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, IXGBE_EIMS_OTHER);
802 
803 	return;
804 } /* ixv_msix_mbx */
805 
806 /************************************************************************
807  * ixv_media_status - Media Ioctl callback
808  *
809  *   Called whenever the user queries the status of
810  *   the interface using ifconfig.
811  ************************************************************************/
812 static void
813 ixv_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
814 {
815 	struct adapter *adapter = ifp->if_softc;
816 
817 	INIT_DEBUGOUT("ixv_media_status: begin");
818 	IXGBE_CORE_LOCK(adapter);
819 	ixv_update_link_status(adapter);
820 
821 	ifmr->ifm_status = IFM_AVALID;
822 	ifmr->ifm_active = IFM_ETHER;
823 
824 	if (!adapter->link_active) {
825 		IXGBE_CORE_UNLOCK(adapter);
826 		return;
827 	}
828 
829 	ifmr->ifm_status |= IFM_ACTIVE;
830 
831 	switch (adapter->link_speed) {
832 		case IXGBE_LINK_SPEED_1GB_FULL:
833 			ifmr->ifm_active |= IFM_1000_T | IFM_FDX;
834 			break;
835 		case IXGBE_LINK_SPEED_10GB_FULL:
836 			ifmr->ifm_active |= IFM_10G_T | IFM_FDX;
837 			break;
838 		case IXGBE_LINK_SPEED_100_FULL:
839 			ifmr->ifm_active |= IFM_100_TX | IFM_FDX;
840 			break;
841 		case IXGBE_LINK_SPEED_10_FULL:
842 			ifmr->ifm_active |= IFM_10_T | IFM_FDX;
843 			break;
844 	}
845 
846 	IXGBE_CORE_UNLOCK(adapter);
847 
848 	return;
849 } /* ixv_media_status */
850 
851 /************************************************************************
852  * ixv_media_change - Media Ioctl callback
853  *
854  *   Called when the user changes speed/duplex using
855  *   media/mediopt option with ifconfig.
856  ************************************************************************/
857 static int
858 ixv_media_change(struct ifnet *ifp)
859 {
860 	struct adapter *adapter = ifp->if_softc;
861 	struct ifmedia *ifm = &adapter->media;
862 
863 	INIT_DEBUGOUT("ixv_media_change: begin");
864 
865 	if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
866 		return (EINVAL);
867 
868 	switch (IFM_SUBTYPE(ifm->ifm_media)) {
869 	case IFM_AUTO:
870 		break;
871 	default:
872 		device_printf(adapter->dev, "Only auto media type\n");
873 		return (EINVAL);
874 	}
875 
876 	return (0);
877 } /* ixv_media_change */
878 
879 
880 /************************************************************************
881  * ixv_negotiate_api
882  *
883  *   Negotiate the Mailbox API with the PF;
884  *   start with the most featured API first.
885  ************************************************************************/
886 static int
887 ixv_negotiate_api(struct adapter *adapter)
888 {
889 	struct ixgbe_hw *hw = &adapter->hw;
890 	int             mbx_api[] = { ixgbe_mbox_api_11,
891 	                              ixgbe_mbox_api_10,
892 	                              ixgbe_mbox_api_unknown };
893 	int             i = 0;
894 
895 	while (mbx_api[i] != ixgbe_mbox_api_unknown) {
896 		if (ixgbevf_negotiate_api_version(hw, mbx_api[i]) == 0)
897 			return (0);
898 		i++;
899 	}
900 
901 	return (EINVAL);
902 } /* ixv_negotiate_api */
903 
904 
905 /************************************************************************
906  * ixv_set_multi - Multicast Update
907  *
908  *   Called whenever multicast address list is updated.
909  ************************************************************************/
910 static void
911 ixv_set_multi(struct adapter *adapter)
912 {
913 	u8       mta[MAX_NUM_MULTICAST_ADDRESSES * IXGBE_ETH_LENGTH_OF_ADDRESS];
914 	u8                 *update_ptr;
915 	struct ifmultiaddr *ifma;
916 	struct ifnet       *ifp = adapter->ifp;
917 	int                mcnt = 0;
918 
919 	IOCTL_DEBUGOUT("ixv_set_multi: begin");
920 
921 #if __FreeBSD_version < 800000
922 	IF_ADDR_LOCK(ifp);
923 #else
924 	if_maddr_rlock(ifp);
925 #endif
926 	TAILQ_FOREACH(ifma, &ifp->if_multiaddrs, ifma_link) {
927 		if (ifma->ifma_addr->sa_family != AF_LINK)
928 			continue;
929 		bcopy(LLADDR((struct sockaddr_dl *)ifma->ifma_addr),
930 		    &mta[mcnt * IXGBE_ETH_LENGTH_OF_ADDRESS],
931 		    IXGBE_ETH_LENGTH_OF_ADDRESS);
932 		mcnt++;
933 	}
934 #if __FreeBSD_version < 800000
935 	IF_ADDR_UNLOCK(ifp);
936 #else
937 	if_maddr_runlock(ifp);
938 #endif
939 
940 	update_ptr = mta;
941 
942 	adapter->hw.mac.ops.update_mc_addr_list(&adapter->hw, update_ptr, mcnt,
943 	    ixv_mc_array_itr, TRUE);
944 
945 	return;
946 } /* ixv_set_multi */
947 
948 /************************************************************************
949  * ixv_mc_array_itr
950  *
951  *   An iterator function needed by the multicast shared code.
952  *   It feeds the shared code routine the addresses in the
953  *   array of ixv_set_multi() one by one.
954  ************************************************************************/
955 static u8 *
956 ixv_mc_array_itr(struct ixgbe_hw *hw, u8 **update_ptr, u32 *vmdq)
957 {
958 	u8 *addr = *update_ptr;
959 	u8 *newptr;
960 	*vmdq = 0;
961 
962 	newptr = addr + IXGBE_ETH_LENGTH_OF_ADDRESS;
963 	*update_ptr = newptr;
964 
965 	return addr;
966 } /* ixv_mc_array_itr */
967 
968 /************************************************************************
969  * ixv_local_timer - Timer routine
970  *
971  *   Checks for link status, updates statistics,
972  *   and runs the watchdog check.
973  ************************************************************************/
974 static void
975 ixv_local_timer(void *arg)
976 {
977 	struct adapter  *adapter = arg;
978 	device_t        dev = adapter->dev;
979 	struct ix_queue *que = adapter->queues;
980 	u64             queues = 0;
981 	int             hung = 0;
982 
983 	mtx_assert(&adapter->core_mtx, MA_OWNED);
984 
985 	ixv_check_link(adapter);
986 
987 	/* Stats Update */
988 	ixv_update_stats(adapter);
989 
990 	/*
991 	 * Check the TX queues status
992 	 *      - mark hung queues so we don't schedule on them
993 	 *      - watchdog only if all queues show hung
994 	 */
995 	for (int i = 0; i < adapter->num_queues; i++, que++) {
996 		/* Keep track of queues with work for soft irq */
997 		if (que->txr->busy)
998 			queues |= ((u64)1 << que->me);
999 		/*
1000 		 * Each time txeof runs without cleaning, but there
1001 		 * are uncleaned descriptors it increments busy. If
1002 		 * we get to the MAX we declare it hung.
1003 		 */
1004 		if (que->busy == IXGBE_QUEUE_HUNG) {
1005 			++hung;
1006 			/* Mark the queue as inactive */
1007 			adapter->active_queues &= ~((u64)1 << que->me);
1008 			continue;
1009 		} else {
1010 			/* Check if we've come back from hung */
1011 			if ((adapter->active_queues & ((u64)1 << que->me)) == 0)
1012 				adapter->active_queues |= ((u64)1 << que->me);
1013 		}
1014 		if (que->busy >= IXGBE_MAX_TX_BUSY) {
1015 			device_printf(dev,
1016 			    "Warning queue %d appears to be hung!\n", i);
1017 			que->txr->busy = IXGBE_QUEUE_HUNG;
1018 			++hung;
1019 		}
1020 
1021 	}
1022 
1023 	/* Only truly watchdog if all queues show hung */
1024 	if (hung == adapter->num_queues)
1025 		goto watchdog;
1026 	else if (queues != 0) { /* Force an IRQ on queues with work */
1027 		ixv_rearm_queues(adapter, queues);
1028 	}
1029 
1030 	callout_reset(&adapter->timer, hz, ixv_local_timer, adapter);
1031 
1032 	return;
1033 
1034 watchdog:
1035 
1036 	device_printf(adapter->dev, "Watchdog timeout -- resetting\n");
1037 	adapter->ifp->if_drv_flags &= ~IFF_DRV_RUNNING;
1038 	adapter->watchdog_events++;
1039 	ixv_init_locked(adapter);
1040 } /* ixv_local_timer */
1041 
1042 /************************************************************************
1043  * ixv_update_link_status - Update OS on link state
1044  *
1045  * Note: Only updates the OS on the cached link state.
1046  *       The real check of the hardware only happens with
1047  *       a link interrupt.
1048  ************************************************************************/
1049 static void
1050 ixv_update_link_status(struct adapter *adapter)
1051 {
1052 	struct ifnet *ifp = adapter->ifp;
1053 	device_t     dev = adapter->dev;
1054 
1055 	if (adapter->link_up) {
1056 		if (adapter->link_active == FALSE) {
1057 			if (bootverbose)
1058 				device_printf(dev,"Link is up %d Gbps %s \n",
1059 				    ((adapter->link_speed == 128) ? 10 : 1),
1060 				    "Full Duplex");
1061 			adapter->link_active = TRUE;
1062 			if_link_state_change(ifp, LINK_STATE_UP);
1063 		}
1064 	} else { /* Link down */
1065 		if (adapter->link_active == TRUE) {
1066 			if (bootverbose)
1067 				device_printf(dev,"Link is Down\n");
1068 			if_link_state_change(ifp, LINK_STATE_DOWN);
1069 			adapter->link_active = FALSE;
1070 		}
1071 	}
1072 
1073 	return;
1074 } /* ixv_update_link_status */
1075 
1076 
1077 /************************************************************************
1078  * ixv_stop - Stop the hardware
1079  *
1080  *   Disables all traffic on the adapter by issuing a
1081  *   global reset on the MAC and deallocates TX/RX buffers.
1082  ************************************************************************/
1083 static void
1084 ixv_stop(void *arg)
1085 {
1086 	struct ifnet    *ifp;
1087 	struct adapter  *adapter = arg;
1088 	struct ixgbe_hw *hw = &adapter->hw;
1089 
1090 	ifp = adapter->ifp;
1091 
1092 	mtx_assert(&adapter->core_mtx, MA_OWNED);
1093 
1094 	INIT_DEBUGOUT("ixv_stop: begin\n");
1095 	ixv_disable_intr(adapter);
1096 
1097 	/* Tell the stack that the interface is no longer active */
1098 	ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
1099 
1100 	hw->mac.ops.reset_hw(hw);
1101 	adapter->hw.adapter_stopped = FALSE;
1102 	hw->mac.ops.stop_adapter(hw);
1103 	callout_stop(&adapter->timer);
1104 
1105 	/* reprogram the RAR[0] in case user changed it. */
1106 	hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
1107 
1108 	return;
1109 } /* ixv_stop */
1110 
1111 
1112 /************************************************************************
1113  * ixv_allocate_pci_resources
1114  ************************************************************************/
1115 static int
1116 ixv_allocate_pci_resources(struct adapter *adapter)
1117 {
1118 	device_t dev = adapter->dev;
1119 	int      rid;
1120 
1121 	rid = PCIR_BAR(0);
1122 	adapter->pci_mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid,
1123 	    RF_ACTIVE);
1124 
1125 	if (!(adapter->pci_mem)) {
1126 		device_printf(dev, "Unable to allocate bus resource: memory\n");
1127 		return (ENXIO);
1128 	}
1129 
1130 	adapter->osdep.mem_bus_space_tag = rman_get_bustag(adapter->pci_mem);
1131 	adapter->osdep.mem_bus_space_handle =
1132 	    rman_get_bushandle(adapter->pci_mem);
1133 	adapter->hw.hw_addr = (u8 *)&adapter->osdep.mem_bus_space_handle;
1134 
1135 	/* Pick up the tuneable queues */
1136 	adapter->num_queues = ixv_num_queues;
1137 
1138 	return (0);
1139 } /* ixv_allocate_pci_resources */
1140 
1141 /************************************************************************
1142  * ixv_free_pci_resources
1143  ************************************************************************/
1144 static void
1145 ixv_free_pci_resources(struct adapter * adapter)
1146 {
1147 	struct ix_queue *que = adapter->queues;
1148 	device_t        dev = adapter->dev;
1149 	int             rid, memrid;
1150 
1151 	memrid = PCIR_BAR(MSIX_82598_BAR);
1152 
1153 	/*
1154 	 * There is a slight possibility of a failure mode
1155 	 * in attach that will result in entering this function
1156 	 * before interrupt resources have been initialized, and
1157 	 * in that case we do not want to execute the loops below
1158 	 * We can detect this reliably by the state of the adapter
1159 	 * res pointer.
1160 	 */
1161 	if (adapter->res == NULL)
1162 		goto mem;
1163 
1164 	/*
1165 	 *  Release all msix queue resources:
1166 	 */
1167 	for (int i = 0; i < adapter->num_queues; i++, que++) {
1168 		rid = que->msix + 1;
1169 		if (que->tag != NULL) {
1170 			bus_teardown_intr(dev, que->res, que->tag);
1171 			que->tag = NULL;
1172 		}
1173 		if (que->res != NULL)
1174 			bus_release_resource(dev, SYS_RES_IRQ, rid, que->res);
1175 	}
1176 
1177 
1178 	/* Clean the Mailbox interrupt last */
1179 	rid = adapter->vector + 1;
1180 
1181 	if (adapter->tag != NULL) {
1182 		bus_teardown_intr(dev, adapter->res, adapter->tag);
1183 		adapter->tag = NULL;
1184 	}
1185 	if (adapter->res != NULL)
1186 		bus_release_resource(dev, SYS_RES_IRQ, rid, adapter->res);
1187 
1188 mem:
1189 	pci_release_msi(dev);
1190 
1191 	if (adapter->msix_mem != NULL)
1192 		bus_release_resource(dev, SYS_RES_MEMORY, memrid,
1193 		    adapter->msix_mem);
1194 
1195 	if (adapter->pci_mem != NULL)
1196 		bus_release_resource(dev, SYS_RES_MEMORY, PCIR_BAR(0),
1197 		    adapter->pci_mem);
1198 
1199 	return;
1200 } /* ixv_free_pci_resources */
1201 
1202 /************************************************************************
1203  * ixv_setup_interface
1204  *
1205  *   Setup networking device structure and register an interface.
1206  ************************************************************************/
1207 static void
1208 ixv_setup_interface(device_t dev, struct adapter *adapter)
1209 {
1210 	struct ifnet *ifp;
1211 
1212 	INIT_DEBUGOUT("ixv_setup_interface: begin");
1213 
1214 	ifp = adapter->ifp = if_alloc(IFT_ETHER);
1215 	if (ifp == NULL)
1216 		panic("%s: can not if_alloc()\n", device_get_nameunit(dev));
1217 	if_initname(ifp, device_get_name(dev), device_get_unit(dev));
1218 	ifp->if_baudrate = 1000000000;
1219 	ifp->if_init = ixv_init;
1220 	ifp->if_softc = adapter;
1221 	ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
1222 	ifp->if_ioctl = ixv_ioctl;
1223 	if_setgetcounterfn(ifp, ixv_get_counter);
1224 	/* TSO parameters */
1225 	ifp->if_hw_tsomax = 65518;
1226 	ifp->if_hw_tsomaxsegcount = IXGBE_82599_SCATTER;
1227 	ifp->if_hw_tsomaxsegsize = 2048;
1228 	if (adapter->feat_en & IXGBE_FEATURE_LEGACY_TX) {
1229 		ifp->if_start = ixgbe_legacy_start;
1230 		ixv_start_locked = ixgbe_legacy_start_locked;
1231 		ixv_ring_empty = ixgbe_legacy_ring_empty;
1232 	} else {
1233 		ifp->if_transmit = ixgbe_mq_start;
1234 		ifp->if_qflush = ixgbe_qflush;
1235 		ixv_start_locked = ixgbe_mq_start_locked;
1236 		ixv_ring_empty = drbr_empty;
1237 	}
1238 	IFQ_SET_MAXLEN(&ifp->if_snd, adapter->num_tx_desc - 2);
1239 
1240 	ether_ifattach(ifp, adapter->hw.mac.addr);
1241 
1242 	adapter->max_frame_size = ifp->if_mtu + IXGBE_MTU_HDR;
1243 
1244 	/*
1245 	 * Tell the upper layer(s) we support long frames.
1246 	 */
1247 	ifp->if_hdrlen = sizeof(struct ether_vlan_header);
1248 
1249 	/* Set capability flags */
1250 	ifp->if_capabilities |= IFCAP_HWCSUM
1251 	                     |  IFCAP_HWCSUM_IPV6
1252 	                     |  IFCAP_TSO
1253 	                     |  IFCAP_LRO
1254 	                     |  IFCAP_VLAN_HWTAGGING
1255 	                     |  IFCAP_VLAN_HWTSO
1256 	                     |  IFCAP_VLAN_HWCSUM
1257 	                     |  IFCAP_JUMBO_MTU
1258 	                     |  IFCAP_VLAN_MTU;
1259 
1260 	/* Enable the above capabilities by default */
1261 	ifp->if_capenable = ifp->if_capabilities;
1262 
1263 	/*
1264 	 * Specify the media types supported by this adapter and register
1265 	 * callbacks to update media and link information
1266 	 */
1267 	ifmedia_init(&adapter->media, IFM_IMASK, ixv_media_change,
1268 	    ixv_media_status);
1269 	ifmedia_add(&adapter->media, IFM_ETHER | IFM_AUTO, 0, NULL);
1270 	ifmedia_set(&adapter->media, IFM_ETHER | IFM_AUTO);
1271 
1272 	return;
1273 } /* ixv_setup_interface */
1274 
1275 
1276 /************************************************************************
1277  * ixv_initialize_transmit_units - Enable transmit unit.
1278  ************************************************************************/
1279 static void
1280 ixv_initialize_transmit_units(struct adapter *adapter)
1281 {
1282 	struct tx_ring  *txr = adapter->tx_rings;
1283 	struct ixgbe_hw *hw = &adapter->hw;
1284 
1285 
1286 	for (int i = 0; i < adapter->num_queues; i++, txr++) {
1287 		u64 tdba = txr->txdma.dma_paddr;
1288 		u32 txctrl, txdctl;
1289 
1290 		/* Set WTHRESH to 8, burst writeback */
1291 		txdctl = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(i));
1292 		txdctl |= (8 << 16);
1293 		IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(i), txdctl);
1294 
1295 		/* Set the HW Tx Head and Tail indices */
1296 		IXGBE_WRITE_REG(&adapter->hw, IXGBE_VFTDH(i), 0);
1297 		IXGBE_WRITE_REG(&adapter->hw, IXGBE_VFTDT(i), 0);
1298 
1299 		/* Set Tx Tail register */
1300 		txr->tail = IXGBE_VFTDT(i);
1301 
1302 		/* Set Ring parameters */
1303 		IXGBE_WRITE_REG(hw, IXGBE_VFTDBAL(i),
1304 		    (tdba & 0x00000000ffffffffULL));
1305 		IXGBE_WRITE_REG(hw, IXGBE_VFTDBAH(i), (tdba >> 32));
1306 		IXGBE_WRITE_REG(hw, IXGBE_VFTDLEN(i),
1307 		    adapter->num_tx_desc * sizeof(struct ixgbe_legacy_tx_desc));
1308 		txctrl = IXGBE_READ_REG(hw, IXGBE_VFDCA_TXCTRL(i));
1309 		txctrl &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1310 		IXGBE_WRITE_REG(hw, IXGBE_VFDCA_TXCTRL(i), txctrl);
1311 
1312 		/* Now enable */
1313 		txdctl = IXGBE_READ_REG(hw, IXGBE_VFTXDCTL(i));
1314 		txdctl |= IXGBE_TXDCTL_ENABLE;
1315 		IXGBE_WRITE_REG(hw, IXGBE_VFTXDCTL(i), txdctl);
1316 	}
1317 
1318 	return;
1319 } /* ixv_initialize_transmit_units */
1320 
1321 
1322 /************************************************************************
1323  * ixv_initialize_rss_mapping
1324  ************************************************************************/
1325 static void
1326 ixv_initialize_rss_mapping(struct adapter *adapter)
1327 {
1328 	struct ixgbe_hw *hw = &adapter->hw;
1329 	u32             reta = 0, mrqc, rss_key[10];
1330 	int             queue_id;
1331 	int             i, j;
1332 	u32             rss_hash_config;
1333 
1334 	if (adapter->feat_en & IXGBE_FEATURE_RSS) {
1335 		/* Fetch the configured RSS key */
1336 		rss_getkey((uint8_t *)&rss_key);
1337 	} else {
1338 		/* set up random bits */
1339 		arc4rand(&rss_key, sizeof(rss_key), 0);
1340 	}
1341 
1342 	/* Now fill out hash function seeds */
1343 	for (i = 0; i < 10; i++)
1344 		IXGBE_WRITE_REG(hw, IXGBE_VFRSSRK(i), rss_key[i]);
1345 
1346 	/* Set up the redirection table */
1347 	for (i = 0, j = 0; i < 64; i++, j++) {
1348 		if (j == adapter->num_queues)
1349 			j = 0;
1350 
1351 		if (adapter->feat_en & IXGBE_FEATURE_RSS) {
1352 			/*
1353 			 * Fetch the RSS bucket id for the given indirection
1354 			 * entry. Cap it at the number of configured buckets
1355 			 * (which is num_queues.)
1356 			 */
1357 			queue_id = rss_get_indirection_to_bucket(i);
1358 			queue_id = queue_id % adapter->num_queues;
1359 		} else
1360 			queue_id = j;
1361 
1362 		/*
1363 		 * The low 8 bits are for hash value (n+0);
1364 		 * The next 8 bits are for hash value (n+1), etc.
1365 		 */
1366 		reta >>= 8;
1367 		reta |= ((uint32_t)queue_id) << 24;
1368 		if ((i & 3) == 3) {
1369 			IXGBE_WRITE_REG(hw, IXGBE_VFRETA(i >> 2), reta);
1370 			reta = 0;
1371 		}
1372 	}
1373 
1374 	/* Perform hash on these packet types */
1375 	if (adapter->feat_en & IXGBE_FEATURE_RSS)
1376 		rss_hash_config = rss_gethashconfig();
1377 	else {
1378 		/*
1379 		 * Disable UDP - IP fragments aren't currently being handled
1380 		 * and so we end up with a mix of 2-tuple and 4-tuple
1381 		 * traffic.
1382 		 */
1383 		rss_hash_config = RSS_HASHTYPE_RSS_IPV4
1384 		                | RSS_HASHTYPE_RSS_TCP_IPV4
1385 		                | RSS_HASHTYPE_RSS_IPV6
1386 		                | RSS_HASHTYPE_RSS_TCP_IPV6;
1387 	}
1388 
1389 	mrqc = IXGBE_MRQC_RSSEN;
1390 	if (rss_hash_config & RSS_HASHTYPE_RSS_IPV4)
1391 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4;
1392 	if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV4)
1393 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_TCP;
1394 	if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6)
1395 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6;
1396 	if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6)
1397 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_TCP;
1398 	if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6_EX)
1399 		device_printf(adapter->dev, "%s: RSS_HASHTYPE_RSS_IPV6_EX defined, but not supported\n",
1400 		    __func__);
1401 	if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6_EX)
1402 		device_printf(adapter->dev, "%s: RSS_HASHTYPE_RSS_TCP_IPV6_EX defined, but not supported\n",
1403 		    __func__);
1404 	if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV4)
1405 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_UDP;
1406 	if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6)
1407 		mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_UDP;
1408 	if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6_EX)
1409 		device_printf(adapter->dev, "%s: RSS_HASHTYPE_RSS_UDP_IPV6_EX defined, but not supported\n",
1410 		    __func__);
1411 	IXGBE_WRITE_REG(hw, IXGBE_VFMRQC, mrqc);
1412 } /* ixv_initialize_rss_mapping */
1413 
1414 
1415 /************************************************************************
1416  * ixv_initialize_receive_units - Setup receive registers and features.
1417  ************************************************************************/
1418 static void
1419 ixv_initialize_receive_units(struct adapter *adapter)
1420 {
1421 	struct rx_ring  *rxr = adapter->rx_rings;
1422 	struct ixgbe_hw *hw = &adapter->hw;
1423 	struct ifnet    *ifp = adapter->ifp;
1424 	u32             bufsz, rxcsum, psrtype;
1425 
1426 	if (ifp->if_mtu > ETHERMTU)
1427 		bufsz = 4096 >> IXGBE_SRRCTL_BSIZEPKT_SHIFT;
1428 	else
1429 		bufsz = 2048 >> IXGBE_SRRCTL_BSIZEPKT_SHIFT;
1430 
1431 	psrtype = IXGBE_PSRTYPE_TCPHDR
1432 	        | IXGBE_PSRTYPE_UDPHDR
1433 	        | IXGBE_PSRTYPE_IPV4HDR
1434 	        | IXGBE_PSRTYPE_IPV6HDR
1435 	        | IXGBE_PSRTYPE_L2HDR;
1436 
1437 	if (adapter->num_queues > 1)
1438 		psrtype |= 1 << 29;
1439 
1440 	IXGBE_WRITE_REG(hw, IXGBE_VFPSRTYPE, psrtype);
1441 
1442 	/* Tell PF our max_frame size */
1443 	if (ixgbevf_rlpml_set_vf(hw, adapter->max_frame_size) != 0) {
1444 		device_printf(adapter->dev, "There is a problem with the PF setup.  It is likely the receive unit for this VF will not function correctly.\n");
1445 	}
1446 
1447 	for (int i = 0; i < adapter->num_queues; i++, rxr++) {
1448 		u64 rdba = rxr->rxdma.dma_paddr;
1449 		u32 reg, rxdctl;
1450 
1451 		/* Disable the queue */
1452 		rxdctl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(i));
1453 		rxdctl &= ~IXGBE_RXDCTL_ENABLE;
1454 		IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(i), rxdctl);
1455 		for (int j = 0; j < 10; j++) {
1456 			if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(i)) &
1457 			    IXGBE_RXDCTL_ENABLE)
1458 				msec_delay(1);
1459 			else
1460 				break;
1461 		}
1462 		wmb();
1463 		/* Setup the Base and Length of the Rx Descriptor Ring */
1464 		IXGBE_WRITE_REG(hw, IXGBE_VFRDBAL(i),
1465 		    (rdba & 0x00000000ffffffffULL));
1466 		IXGBE_WRITE_REG(hw, IXGBE_VFRDBAH(i), (rdba >> 32));
1467 		IXGBE_WRITE_REG(hw, IXGBE_VFRDLEN(i),
1468 		    adapter->num_rx_desc * sizeof(union ixgbe_adv_rx_desc));
1469 
1470 		/* Reset the ring indices */
1471 		IXGBE_WRITE_REG(hw, IXGBE_VFRDH(rxr->me), 0);
1472 		IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me), 0);
1473 
1474 		/* Set up the SRRCTL register */
1475 		reg = IXGBE_READ_REG(hw, IXGBE_VFSRRCTL(i));
1476 		reg &= ~IXGBE_SRRCTL_BSIZEHDR_MASK;
1477 		reg &= ~IXGBE_SRRCTL_BSIZEPKT_MASK;
1478 		reg |= bufsz;
1479 		reg |= IXGBE_SRRCTL_DESCTYPE_ADV_ONEBUF;
1480 		IXGBE_WRITE_REG(hw, IXGBE_VFSRRCTL(i), reg);
1481 
1482 		/* Capture Rx Tail index */
1483 		rxr->tail = IXGBE_VFRDT(rxr->me);
1484 
1485 		/* Do the queue enabling last */
1486 		rxdctl |= IXGBE_RXDCTL_ENABLE | IXGBE_RXDCTL_VME;
1487 		IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(i), rxdctl);
1488 		for (int k = 0; k < 10; k++) {
1489 			if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(i)) &
1490 			    IXGBE_RXDCTL_ENABLE)
1491 				break;
1492 			msec_delay(1);
1493 		}
1494 		wmb();
1495 
1496 		/* Set the Tail Pointer */
1497 		/*
1498 		 * In netmap mode, we must preserve the buffers made
1499 		 * available to userspace before the if_init()
1500 		 * (this is true by default on the TX side, because
1501 		 * init makes all buffers available to userspace).
1502 		 *
1503 		 * netmap_reset() and the device specific routines
1504 		 * (e.g. ixgbe_setup_receive_rings()) map these
1505 		 * buffers at the end of the NIC ring, so here we
1506 		 * must set the RDT (tail) register to make sure
1507 		 * they are not overwritten.
1508 		 *
1509 		 * In this driver the NIC ring starts at RDH = 0,
1510 		 * RDT points to the last slot available for reception (?),
1511 		 * so RDT = num_rx_desc - 1 means the whole ring is available.
1512 		 */
1513 #ifdef DEV_NETMAP
1514 		if ((adapter->feat_en & IXGBE_FEATURE_NETMAP) &&
1515 		    (ifp->if_capenable & IFCAP_NETMAP)) {
1516 			struct netmap_adapter *na = NA(adapter->ifp);
1517 			struct netmap_kring *kring = &na->rx_rings[i];
1518 			int t = na->num_rx_desc - 1 - nm_kr_rxspace(kring);
1519 
1520 			IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me), t);
1521 		} else
1522 #endif /* DEV_NETMAP */
1523 			IXGBE_WRITE_REG(hw, IXGBE_VFRDT(rxr->me),
1524 			    adapter->num_rx_desc - 1);
1525 	}
1526 
1527 	rxcsum = IXGBE_READ_REG(hw, IXGBE_RXCSUM);
1528 
1529 	ixv_initialize_rss_mapping(adapter);
1530 
1531 	if (adapter->num_queues > 1) {
1532 		/* RSS and RX IPP Checksum are mutually exclusive */
1533 		rxcsum |= IXGBE_RXCSUM_PCSD;
1534 	}
1535 
1536 	if (ifp->if_capenable & IFCAP_RXCSUM)
1537 		rxcsum |= IXGBE_RXCSUM_PCSD;
1538 
1539 	if (!(rxcsum & IXGBE_RXCSUM_PCSD))
1540 		rxcsum |= IXGBE_RXCSUM_IPPCSE;
1541 
1542 	IXGBE_WRITE_REG(hw, IXGBE_RXCSUM, rxcsum);
1543 
1544 	return;
1545 } /* ixv_initialize_receive_units */
1546 
1547 /************************************************************************
1548  * ixv_setup_vlan_support
1549  ************************************************************************/
1550 static void
1551 ixv_setup_vlan_support(struct adapter *adapter)
1552 {
1553 	struct ixgbe_hw *hw = &adapter->hw;
1554 	u32             ctrl, vid, vfta, retry;
1555 
1556 	/*
1557 	 * We get here thru init_locked, meaning
1558 	 * a soft reset, this has already cleared
1559 	 * the VFTA and other state, so if there
1560 	 * have been no vlan's registered do nothing.
1561 	 */
1562 	if (adapter->num_vlans == 0)
1563 		return;
1564 
1565 	/* Enable the queues */
1566 	for (int i = 0; i < adapter->num_queues; i++) {
1567 		ctrl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(i));
1568 		ctrl |= IXGBE_RXDCTL_VME;
1569 		IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(i), ctrl);
1570 		/*
1571 		 * Let Rx path know that it needs to store VLAN tag
1572 		 * as part of extra mbuf info.
1573 		 */
1574 		adapter->rx_rings[i].vtag_strip = TRUE;
1575 	}
1576 
1577 	/*
1578 	 * A soft reset zero's out the VFTA, so
1579 	 * we need to repopulate it now.
1580 	 */
1581 	for (int i = 0; i < IXGBE_VFTA_SIZE; i++) {
1582 		if (ixv_shadow_vfta[i] == 0)
1583 			continue;
1584 		vfta = ixv_shadow_vfta[i];
1585 		/*
1586 		 * Reconstruct the vlan id's
1587 		 * based on the bits set in each
1588 		 * of the array ints.
1589 		 */
1590 		for (int j = 0; j < 32; j++) {
1591 			retry = 0;
1592 			if ((vfta & (1 << j)) == 0)
1593 				continue;
1594 			vid = (i * 32) + j;
1595 			/* Call the shared code mailbox routine */
1596 			while (hw->mac.ops.set_vfta(hw, vid, 0, TRUE, FALSE)) {
1597 				if (++retry > 5)
1598 					break;
1599 			}
1600 		}
1601 	}
1602 } /* ixv_setup_vlan_support */
1603 
1604 /************************************************************************
1605  * ixv_register_vlan
1606  *
1607  *   Run via a vlan config EVENT, it enables us to use the
1608  *   HW Filter table since we can get the vlan id. This just
1609  *   creates the entry in the soft version of the VFTA, init
1610  *   will repopulate the real table.
1611  ************************************************************************/
1612 static void
1613 ixv_register_vlan(void *arg, struct ifnet *ifp, u16 vtag)
1614 {
1615 	struct adapter *adapter = ifp->if_softc;
1616 	u16            index, bit;
1617 
1618 	if (ifp->if_softc != arg) /* Not our event */
1619 		return;
1620 
1621 	if ((vtag == 0) || (vtag > 4095)) /* Invalid */
1622 		return;
1623 
1624 	IXGBE_CORE_LOCK(adapter);
1625 	index = (vtag >> 5) & 0x7F;
1626 	bit = vtag & 0x1F;
1627 	ixv_shadow_vfta[index] |= (1 << bit);
1628 	++adapter->num_vlans;
1629 	/* Re-init to load the changes */
1630 	ixv_init_locked(adapter);
1631 	IXGBE_CORE_UNLOCK(adapter);
1632 } /* ixv_register_vlan */
1633 
1634 /************************************************************************
1635  * ixv_unregister_vlan
1636  *
1637  *   Run via a vlan unconfig EVENT, remove our entry
1638  *   in the soft vfta.
1639  ************************************************************************/
1640 static void
1641 ixv_unregister_vlan(void *arg, struct ifnet *ifp, u16 vtag)
1642 {
1643 	struct adapter *adapter = ifp->if_softc;
1644 	u16            index, bit;
1645 
1646 	if (ifp->if_softc !=  arg)
1647 		return;
1648 
1649 	if ((vtag == 0) || (vtag > 4095))  /* Invalid */
1650 		return;
1651 
1652 	IXGBE_CORE_LOCK(adapter);
1653 	index = (vtag >> 5) & 0x7F;
1654 	bit = vtag & 0x1F;
1655 	ixv_shadow_vfta[index] &= ~(1 << bit);
1656 	--adapter->num_vlans;
1657 	/* Re-init to load the changes */
1658 	ixv_init_locked(adapter);
1659 	IXGBE_CORE_UNLOCK(adapter);
1660 } /* ixv_unregister_vlan */
1661 
1662 /************************************************************************
1663  * ixv_enable_intr
1664  ************************************************************************/
1665 static void
1666 ixv_enable_intr(struct adapter *adapter)
1667 {
1668 	struct ixgbe_hw *hw = &adapter->hw;
1669 	struct ix_queue *que = adapter->queues;
1670 	u32             mask = (IXGBE_EIMS_ENABLE_MASK & ~IXGBE_EIMS_RTX_QUEUE);
1671 
1672 
1673 	IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, mask);
1674 
1675 	mask = IXGBE_EIMS_ENABLE_MASK;
1676 	mask &= ~(IXGBE_EIMS_OTHER | IXGBE_EIMS_LSC);
1677 	IXGBE_WRITE_REG(hw, IXGBE_VTEIAC, mask);
1678 
1679 	for (int i = 0; i < adapter->num_queues; i++, que++)
1680 		ixv_enable_queue(adapter, que->msix);
1681 
1682 	IXGBE_WRITE_FLUSH(hw);
1683 
1684 	return;
1685 } /* ixv_enable_intr */
1686 
1687 /************************************************************************
1688  * ixv_disable_intr
1689  ************************************************************************/
1690 static void
1691 ixv_disable_intr(struct adapter *adapter)
1692 {
1693 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEIAC, 0);
1694 	IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEIMC, ~0);
1695 	IXGBE_WRITE_FLUSH(&adapter->hw);
1696 
1697 	return;
1698 } /* ixv_disable_intr */
1699 
1700 /************************************************************************
1701  * ixv_set_ivar
1702  *
1703  *   Setup the correct IVAR register for a particular MSI-X interrupt
1704  *    - entry is the register array entry
1705  *    - vector is the MSI-X vector for this queue
1706  *    - type is RX/TX/MISC
1707  ************************************************************************/
1708 static void
1709 ixv_set_ivar(struct adapter *adapter, u8 entry, u8 vector, s8 type)
1710 {
1711 	struct ixgbe_hw *hw = &adapter->hw;
1712 	u32             ivar, index;
1713 
1714 	vector |= IXGBE_IVAR_ALLOC_VAL;
1715 
1716 	if (type == -1) { /* MISC IVAR */
1717 		ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR_MISC);
1718 		ivar &= ~0xFF;
1719 		ivar |= vector;
1720 		IXGBE_WRITE_REG(hw, IXGBE_VTIVAR_MISC, ivar);
1721 	} else {          /* RX/TX IVARS */
1722 		index = (16 * (entry & 1)) + (8 * type);
1723 		ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR(entry >> 1));
1724 		ivar &= ~(0xFF << index);
1725 		ivar |= (vector << index);
1726 		IXGBE_WRITE_REG(hw, IXGBE_VTIVAR(entry >> 1), ivar);
1727 	}
1728 } /* ixv_set_ivar */
1729 
1730 /************************************************************************
1731  * ixv_configure_ivars
1732  ************************************************************************/
1733 static void
1734 ixv_configure_ivars(struct adapter *adapter)
1735 {
1736 	struct ix_queue *que = adapter->queues;
1737 
1738 	for (int i = 0; i < adapter->num_queues; i++, que++) {
1739 		/* First the RX queue entry */
1740 		ixv_set_ivar(adapter, i, que->msix, 0);
1741 		/* ... and the TX */
1742 		ixv_set_ivar(adapter, i, que->msix, 1);
1743 		/* Set an initial value in EITR */
1744 		IXGBE_WRITE_REG(&adapter->hw, IXGBE_VTEITR(que->msix),
1745 		    IXGBE_EITR_DEFAULT);
1746 	}
1747 
1748 	/* For the mailbox interrupt */
1749 	ixv_set_ivar(adapter, 1, adapter->vector, -1);
1750 } /* ixv_configure_ivars */
1751 
1752 
1753 /************************************************************************
1754  * ixv_get_counter
1755  ************************************************************************/
1756 static uint64_t
1757 ixv_get_counter(struct ifnet *ifp, ift_counter cnt)
1758 {
1759 	struct adapter *adapter;
1760 
1761 	adapter = if_getsoftc(ifp);
1762 
1763 	switch (cnt) {
1764 	case IFCOUNTER_IPACKETS:
1765 		return (adapter->ipackets);
1766 	case IFCOUNTER_OPACKETS:
1767 		return (adapter->opackets);
1768 	case IFCOUNTER_IBYTES:
1769 		return (adapter->ibytes);
1770 	case IFCOUNTER_OBYTES:
1771 		return (adapter->obytes);
1772 	case IFCOUNTER_IMCASTS:
1773 		return (adapter->imcasts);
1774 	default:
1775 		return (if_get_counter_default(ifp, cnt));
1776 	}
1777 } /* ixv_get_counter */
1778 
1779 /************************************************************************
1780  * ixv_save_stats
1781  *
1782  *   The VF stats registers never have a truly virgin
1783  *   starting point, so this routine tries to make an
1784  *   artificial one, marking ground zero on attach as
1785  *   it were.
1786  ************************************************************************/
1787 static void
1788 ixv_save_stats(struct adapter *adapter)
1789 {
1790 	if (adapter->stats.vf.vfgprc || adapter->stats.vf.vfgptc) {
1791 		adapter->stats.vf.saved_reset_vfgprc +=
1792 		    adapter->stats.vf.vfgprc - adapter->stats.vf.base_vfgprc;
1793 		adapter->stats.vf.saved_reset_vfgptc +=
1794 		    adapter->stats.vf.vfgptc - adapter->stats.vf.base_vfgptc;
1795 		adapter->stats.vf.saved_reset_vfgorc +=
1796 		    adapter->stats.vf.vfgorc - adapter->stats.vf.base_vfgorc;
1797 		adapter->stats.vf.saved_reset_vfgotc +=
1798 		    adapter->stats.vf.vfgotc - adapter->stats.vf.base_vfgotc;
1799 		adapter->stats.vf.saved_reset_vfmprc +=
1800 		    adapter->stats.vf.vfmprc - adapter->stats.vf.base_vfmprc;
1801 	}
1802 } /* ixv_save_stats */
1803 
1804 /************************************************************************
1805  * ixv_init_stats
1806  ************************************************************************/
1807 static void
1808 ixv_init_stats(struct adapter *adapter)
1809 {
1810 	struct ixgbe_hw *hw = &adapter->hw;
1811 
1812 	adapter->stats.vf.last_vfgprc = IXGBE_READ_REG(hw, IXGBE_VFGPRC);
1813 	adapter->stats.vf.last_vfgorc = IXGBE_READ_REG(hw, IXGBE_VFGORC_LSB);
1814 	adapter->stats.vf.last_vfgorc |=
1815 	    (((u64)(IXGBE_READ_REG(hw, IXGBE_VFGORC_MSB))) << 32);
1816 
1817 	adapter->stats.vf.last_vfgptc = IXGBE_READ_REG(hw, IXGBE_VFGPTC);
1818 	adapter->stats.vf.last_vfgotc = IXGBE_READ_REG(hw, IXGBE_VFGOTC_LSB);
1819 	adapter->stats.vf.last_vfgotc |=
1820 	    (((u64)(IXGBE_READ_REG(hw, IXGBE_VFGOTC_MSB))) << 32);
1821 
1822 	adapter->stats.vf.last_vfmprc = IXGBE_READ_REG(hw, IXGBE_VFMPRC);
1823 
1824 	adapter->stats.vf.base_vfgprc = adapter->stats.vf.last_vfgprc;
1825 	adapter->stats.vf.base_vfgorc = adapter->stats.vf.last_vfgorc;
1826 	adapter->stats.vf.base_vfgptc = adapter->stats.vf.last_vfgptc;
1827 	adapter->stats.vf.base_vfgotc = adapter->stats.vf.last_vfgotc;
1828 	adapter->stats.vf.base_vfmprc = adapter->stats.vf.last_vfmprc;
1829 } /* ixv_init_stats */
1830 
1831 #define UPDATE_STAT_32(reg, last, count)                \
1832 {                                                       \
1833 	u32 current = IXGBE_READ_REG(hw, reg);          \
1834 	if (current < last)                             \
1835 		count += 0x100000000LL;                 \
1836 	last = current;                                 \
1837 	count &= 0xFFFFFFFF00000000LL;                  \
1838 	count |= current;                               \
1839 }
1840 
1841 #define UPDATE_STAT_36(lsb, msb, last, count)           \
1842 {                                                       \
1843 	u64 cur_lsb = IXGBE_READ_REG(hw, lsb);          \
1844 	u64 cur_msb = IXGBE_READ_REG(hw, msb);          \
1845 	u64 current = ((cur_msb << 32) | cur_lsb);      \
1846 	if (current < last)                             \
1847 		count += 0x1000000000LL;                \
1848 	last = current;                                 \
1849 	count &= 0xFFFFFFF000000000LL;                  \
1850 	count |= current;                               \
1851 }
1852 
1853 /************************************************************************
1854  * ixv_update_stats - Update the board statistics counters.
1855  ************************************************************************/
1856 void
1857 ixv_update_stats(struct adapter *adapter)
1858 {
1859 	struct ixgbe_hw *hw = &adapter->hw;
1860 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
1861 
1862         UPDATE_STAT_32(IXGBE_VFGPRC, adapter->stats.vf.last_vfgprc,
1863 	    adapter->stats.vf.vfgprc);
1864         UPDATE_STAT_32(IXGBE_VFGPTC, adapter->stats.vf.last_vfgptc,
1865 	    adapter->stats.vf.vfgptc);
1866         UPDATE_STAT_36(IXGBE_VFGORC_LSB, IXGBE_VFGORC_MSB,
1867 	    adapter->stats.vf.last_vfgorc, adapter->stats.vf.vfgorc);
1868         UPDATE_STAT_36(IXGBE_VFGOTC_LSB, IXGBE_VFGOTC_MSB,
1869 	    adapter->stats.vf.last_vfgotc, adapter->stats.vf.vfgotc);
1870         UPDATE_STAT_32(IXGBE_VFMPRC, adapter->stats.vf.last_vfmprc,
1871 	    adapter->stats.vf.vfmprc);
1872 
1873 	/* Fill out the OS statistics structure */
1874 	IXGBE_SET_IPACKETS(adapter, stats->vfgprc);
1875 	IXGBE_SET_OPACKETS(adapter, stats->vfgptc);
1876 	IXGBE_SET_IBYTES(adapter, stats->vfgorc);
1877 	IXGBE_SET_OBYTES(adapter, stats->vfgotc);
1878 	IXGBE_SET_IMCASTS(adapter, stats->vfmprc);
1879 } /* ixv_update_stats */
1880 
1881 /************************************************************************
1882  * ixv_add_stats_sysctls - Add statistic sysctls for the VF.
1883  ************************************************************************/
1884 static void
1885 ixv_add_stats_sysctls(struct adapter *adapter)
1886 {
1887 	device_t                dev = adapter->dev;
1888 	struct tx_ring          *txr = adapter->tx_rings;
1889 	struct rx_ring          *rxr = adapter->rx_rings;
1890 	struct sysctl_ctx_list  *ctx = device_get_sysctl_ctx(dev);
1891 	struct sysctl_oid       *tree = device_get_sysctl_tree(dev);
1892 	struct sysctl_oid_list  *child = SYSCTL_CHILDREN(tree);
1893 	struct ixgbevf_hw_stats *stats = &adapter->stats.vf;
1894 	struct sysctl_oid       *stat_node, *queue_node;
1895 	struct sysctl_oid_list  *stat_list, *queue_list;
1896 
1897 #define QUEUE_NAME_LEN 32
1898 	char                    namebuf[QUEUE_NAME_LEN];
1899 
1900 	/* Driver Statistics */
1901 	SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "dropped",
1902 	    CTLFLAG_RD, &adapter->dropped_pkts, "Driver dropped packets");
1903 	SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "mbuf_defrag_failed",
1904 	    CTLFLAG_RD, &adapter->mbuf_defrag_failed, "m_defrag() failed");
1905 	SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "watchdog_events",
1906 	    CTLFLAG_RD, &adapter->watchdog_events, "Watchdog timeouts");
1907 	SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "link_irq",
1908 	    CTLFLAG_RD, &adapter->link_irq, "Link MSI-X IRQ Handled");
1909 
1910 	for (int i = 0; i < adapter->num_queues; i++, txr++) {
1911 		snprintf(namebuf, QUEUE_NAME_LEN, "queue%d", i);
1912 		queue_node = SYSCTL_ADD_NODE(ctx, child, OID_AUTO, namebuf,
1913 		    CTLFLAG_RD, NULL, "Queue Name");
1914 		queue_list = SYSCTL_CHILDREN(queue_node);
1915 
1916 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "irqs",
1917 		    CTLFLAG_RD, &(adapter->queues[i].irqs), "IRQs on queue");
1918 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "no_tx_dma_setup",
1919 		    CTLFLAG_RD, &(txr->no_tx_dma_setup),
1920 		    "Driver Tx DMA failure in Tx");
1921 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "tx_no_desc",
1922 		    CTLFLAG_RD, &(txr->no_desc_avail),
1923 		    "Not-enough-descriptors count: TX");
1924 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "tx_packets",
1925 		    CTLFLAG_RD, &(txr->total_packets), "TX Packets");
1926 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "br_drops",
1927 		    CTLFLAG_RD, &(txr->br->br_drops),
1928 		    "Packets dropped in buf_ring");
1929 	}
1930 
1931 	for (int i = 0; i < adapter->num_queues; i++, rxr++) {
1932 		snprintf(namebuf, QUEUE_NAME_LEN, "queue%d", i);
1933 		queue_node = SYSCTL_ADD_NODE(ctx, child, OID_AUTO, namebuf,
1934 		    CTLFLAG_RD, NULL, "Queue Name");
1935 		queue_list = SYSCTL_CHILDREN(queue_node);
1936 
1937 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_packets",
1938 		    CTLFLAG_RD, &(rxr->rx_packets), "RX packets");
1939 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_bytes",
1940 		    CTLFLAG_RD, &(rxr->rx_bytes), "RX bytes");
1941 		SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_discarded",
1942 		    CTLFLAG_RD, &(rxr->rx_discarded), "Discarded RX packets");
1943 	}
1944 
1945 	stat_node = SYSCTL_ADD_NODE(ctx, child, OID_AUTO, "mac",
1946 	    CTLFLAG_RD, NULL, "VF Statistics (read from HW registers)");
1947 	stat_list = SYSCTL_CHILDREN(stat_node);
1948 
1949 	SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_pkts_rcvd",
1950 	    CTLFLAG_RD, &stats->vfgprc, "Good Packets Received");
1951 	SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_octets_rcvd",
1952 	    CTLFLAG_RD, &stats->vfgorc, "Good Octets Received");
1953 	SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "mcast_pkts_rcvd",
1954 	    CTLFLAG_RD, &stats->vfmprc, "Multicast Packets Received");
1955 	SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_pkts_txd",
1956 	    CTLFLAG_RD, &stats->vfgptc, "Good Packets Transmitted");
1957 	SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_octets_txd",
1958 	    CTLFLAG_RD, &stats->vfgotc, "Good Octets Transmitted");
1959 } /* ixv_add_stats_sysctls */
1960 
1961 /************************************************************************
1962  * ixv_set_sysctl_value
1963  ************************************************************************/
1964 static void
1965 ixv_set_sysctl_value(struct adapter *adapter, const char *name,
1966 	const char *description, int *limit, int value)
1967 {
1968 	*limit = value;
1969 	SYSCTL_ADD_INT(device_get_sysctl_ctx(adapter->dev),
1970 	    SYSCTL_CHILDREN(device_get_sysctl_tree(adapter->dev)),
1971 	    OID_AUTO, name, CTLFLAG_RW, limit, value, description);
1972 } /* ixv_set_sysctl_value */
1973 
1974 /************************************************************************
1975  * ixv_print_debug_info
1976  *
1977  *   Called only when em_display_debug_stats is enabled.
1978  *   Provides a way to take a look at important statistics
1979  *   maintained by the driver and hardware.
1980  ************************************************************************/
1981 static void
1982 ixv_print_debug_info(struct adapter *adapter)
1983 {
1984 	device_t        dev = adapter->dev;
1985 	struct ixgbe_hw *hw = &adapter->hw;
1986 	struct ix_queue *que = adapter->queues;
1987 	struct rx_ring  *rxr;
1988 	struct tx_ring  *txr;
1989 	struct lro_ctrl *lro;
1990 
1991 	device_printf(dev, "Error Byte Count = %u \n",
1992 	    IXGBE_READ_REG(hw, IXGBE_ERRBC));
1993 
1994 	for (int i = 0; i < adapter->num_queues; i++, que++) {
1995 		txr = que->txr;
1996 		rxr = que->rxr;
1997 		lro = &rxr->lro;
1998 		device_printf(dev, "QUE(%d) IRQs Handled: %lu\n",
1999 		    que->msix, (long)que->irqs);
2000 		device_printf(dev, "RX(%d) Packets Received: %lld\n",
2001 		    rxr->me, (long long)rxr->rx_packets);
2002 		device_printf(dev, "RX(%d) Bytes Received: %lu\n",
2003 		    rxr->me, (long)rxr->rx_bytes);
2004 		device_printf(dev, "RX(%d) LRO Queued= %lld\n",
2005 		    rxr->me, (long long)lro->lro_queued);
2006 		device_printf(dev, "RX(%d) LRO Flushed= %lld\n",
2007 		    rxr->me, (long long)lro->lro_flushed);
2008 		device_printf(dev, "TX(%d) Packets Sent: %lu\n",
2009 		    txr->me, (long)txr->total_packets);
2010 		device_printf(dev, "TX(%d) NO Desc Avail: %lu\n",
2011 		    txr->me, (long)txr->no_desc_avail);
2012 	}
2013 
2014 	device_printf(dev, "MBX IRQ Handled: %lu\n", (long)adapter->link_irq);
2015 } /* ixv_print_debug_info */
2016 
2017 /************************************************************************
2018  * ixv_sysctl_debug
2019  ************************************************************************/
2020 static int
2021 ixv_sysctl_debug(SYSCTL_HANDLER_ARGS)
2022 {
2023 	struct adapter *adapter;
2024 	int            error, result;
2025 
2026 	result = -1;
2027 	error = sysctl_handle_int(oidp, &result, 0, req);
2028 
2029 	if (error || !req->newptr)
2030 		return (error);
2031 
2032 	if (result == 1) {
2033 		adapter = (struct adapter *)arg1;
2034 		ixv_print_debug_info(adapter);
2035 	}
2036 
2037 	return error;
2038 } /* ixv_sysctl_debug */
2039 
2040 /************************************************************************
2041  * ixv_init_device_features
2042  ************************************************************************/
2043 static void
2044 ixv_init_device_features(struct adapter *adapter)
2045 {
2046 	adapter->feat_cap = IXGBE_FEATURE_NETMAP
2047 	                  | IXGBE_FEATURE_VF
2048 	                  | IXGBE_FEATURE_RSS
2049 	                  | IXGBE_FEATURE_LEGACY_TX;
2050 
2051 	/* A tad short on feature flags for VFs, atm. */
2052 	switch (adapter->hw.mac.type) {
2053 	case ixgbe_mac_82599_vf:
2054 		break;
2055 	case ixgbe_mac_X540_vf:
2056 		break;
2057 	case ixgbe_mac_X550_vf:
2058 	case ixgbe_mac_X550EM_x_vf:
2059 	case ixgbe_mac_X550EM_a_vf:
2060 		adapter->feat_cap |= IXGBE_FEATURE_NEEDS_CTXD;
2061 		break;
2062 	default:
2063 		break;
2064 	}
2065 
2066 	/* Enabled by default... */
2067 	/* Is a virtual function (VF) */
2068 	if (adapter->feat_cap & IXGBE_FEATURE_VF)
2069 		adapter->feat_en |= IXGBE_FEATURE_VF;
2070 	/* Netmap */
2071 	if (adapter->feat_cap & IXGBE_FEATURE_NETMAP)
2072 		adapter->feat_en |= IXGBE_FEATURE_NETMAP;
2073 	/* Receive-Side Scaling (RSS) */
2074 	if (adapter->feat_cap & IXGBE_FEATURE_RSS)
2075 		adapter->feat_en |= IXGBE_FEATURE_RSS;
2076 	/* Needs advanced context descriptor regardless of offloads req'd */
2077 	if (adapter->feat_cap & IXGBE_FEATURE_NEEDS_CTXD)
2078 		adapter->feat_en |= IXGBE_FEATURE_NEEDS_CTXD;
2079 
2080 	/* Enabled via sysctl... */
2081 	/* Legacy (single queue) transmit */
2082 	if ((adapter->feat_cap & IXGBE_FEATURE_LEGACY_TX) &&
2083 	    ixv_enable_legacy_tx)
2084 		adapter->feat_en |= IXGBE_FEATURE_LEGACY_TX;
2085 } /* ixv_init_device_features */
2086 
2087 /************************************************************************
2088  * ixv_shutdown - Shutdown entry point
2089  ************************************************************************/
2090 static int
2091 ixv_shutdown(device_t dev)
2092 {
2093 	struct adapter *adapter = device_get_softc(dev);
2094 	IXGBE_CORE_LOCK(adapter);
2095 	ixv_stop(adapter);
2096 	IXGBE_CORE_UNLOCK(adapter);
2097 
2098 	return (0);
2099 } /* ixv_shutdown */
2100 
2101 
2102 /************************************************************************
2103  * ixv_ioctl - Ioctl entry point
2104  *
2105  *   Called when the user wants to configure the interface.
2106  *
2107  *   return 0 on success, positive on failure
2108  ************************************************************************/
2109 static int
2110 ixv_ioctl(struct ifnet *ifp, u_long command, caddr_t data)
2111 {
2112 	struct adapter *adapter = ifp->if_softc;
2113 	struct ifreq   *ifr = (struct ifreq *)data;
2114 #if defined(INET) || defined(INET6)
2115 	struct ifaddr  *ifa = (struct ifaddr *)data;
2116 	bool           avoid_reset = FALSE;
2117 #endif
2118 	int            error = 0;
2119 
2120 	switch (command) {
2121 
2122 	case SIOCSIFADDR:
2123 #ifdef INET
2124 		if (ifa->ifa_addr->sa_family == AF_INET)
2125 			avoid_reset = TRUE;
2126 #endif
2127 #ifdef INET6
2128 		if (ifa->ifa_addr->sa_family == AF_INET6)
2129 			avoid_reset = TRUE;
2130 #endif
2131 #if defined(INET) || defined(INET6)
2132 		/*
2133 		 * Calling init results in link renegotiation,
2134 		 * so we avoid doing it when possible.
2135 		 */
2136 		if (avoid_reset) {
2137 			ifp->if_flags |= IFF_UP;
2138 			if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
2139 				ixv_init(adapter);
2140 			if (!(ifp->if_flags & IFF_NOARP))
2141 				arp_ifinit(ifp, ifa);
2142 		} else
2143 			error = ether_ioctl(ifp, command, data);
2144 		break;
2145 #endif
2146 	case SIOCSIFMTU:
2147 		IOCTL_DEBUGOUT("ioctl: SIOCSIFMTU (Set Interface MTU)");
2148 		if (ifr->ifr_mtu > IXGBE_MAX_MTU) {
2149 			error = EINVAL;
2150 		} else {
2151 			IXGBE_CORE_LOCK(adapter);
2152 			ifp->if_mtu = ifr->ifr_mtu;
2153 			adapter->max_frame_size = ifp->if_mtu + IXGBE_MTU_HDR;
2154 			if (ifp->if_drv_flags & IFF_DRV_RUNNING)
2155 				ixv_init_locked(adapter);
2156 			IXGBE_CORE_UNLOCK(adapter);
2157 		}
2158 		break;
2159 	case SIOCSIFFLAGS:
2160 		IOCTL_DEBUGOUT("ioctl: SIOCSIFFLAGS (Set Interface Flags)");
2161 		IXGBE_CORE_LOCK(adapter);
2162 		if (ifp->if_flags & IFF_UP) {
2163 			if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0)
2164 				ixv_init_locked(adapter);
2165 		} else
2166 			if (ifp->if_drv_flags & IFF_DRV_RUNNING)
2167 				ixv_stop(adapter);
2168 		adapter->if_flags = ifp->if_flags;
2169 		IXGBE_CORE_UNLOCK(adapter);
2170 		break;
2171 	case SIOCADDMULTI:
2172 	case SIOCDELMULTI:
2173 		IOCTL_DEBUGOUT("ioctl: SIOC(ADD|DEL)MULTI");
2174 		if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
2175 			IXGBE_CORE_LOCK(adapter);
2176 			ixv_disable_intr(adapter);
2177 			ixv_set_multi(adapter);
2178 			ixv_enable_intr(adapter);
2179 			IXGBE_CORE_UNLOCK(adapter);
2180 		}
2181 		break;
2182 	case SIOCSIFMEDIA:
2183 	case SIOCGIFMEDIA:
2184 		IOCTL_DEBUGOUT("ioctl: SIOCxIFMEDIA (Get/Set Interface Media)");
2185 		error = ifmedia_ioctl(ifp, ifr, &adapter->media, command);
2186 		break;
2187 	case SIOCSIFCAP:
2188 	{
2189 		int mask = ifr->ifr_reqcap ^ ifp->if_capenable;
2190 		IOCTL_DEBUGOUT("ioctl: SIOCSIFCAP (Set Capabilities)");
2191 		if (mask & IFCAP_HWCSUM)
2192 			ifp->if_capenable ^= IFCAP_HWCSUM;
2193 		if (mask & IFCAP_TSO4)
2194 			ifp->if_capenable ^= IFCAP_TSO4;
2195 		if (mask & IFCAP_LRO)
2196 			ifp->if_capenable ^= IFCAP_LRO;
2197 		if (mask & IFCAP_VLAN_HWTAGGING)
2198 			ifp->if_capenable ^= IFCAP_VLAN_HWTAGGING;
2199 		if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
2200 			IXGBE_CORE_LOCK(adapter);
2201 			ixv_init_locked(adapter);
2202 			IXGBE_CORE_UNLOCK(adapter);
2203 		}
2204 		VLAN_CAPABILITIES(ifp);
2205 		break;
2206 	}
2207 
2208 	default:
2209 		IOCTL_DEBUGOUT1("ioctl: UNKNOWN (0x%X)\n", (int)command);
2210 		error = ether_ioctl(ifp, command, data);
2211 		break;
2212 	}
2213 
2214 	return (error);
2215 } /* ixv_ioctl */
2216 
2217 /************************************************************************
2218  * ixv_init
2219  ************************************************************************/
2220 static void
2221 ixv_init(void *arg)
2222 {
2223 	struct adapter *adapter = arg;
2224 
2225 	IXGBE_CORE_LOCK(adapter);
2226 	ixv_init_locked(adapter);
2227 	IXGBE_CORE_UNLOCK(adapter);
2228 
2229 	return;
2230 } /* ixv_init */
2231 
2232 
2233 /************************************************************************
2234  * ixv_handle_que
2235  ************************************************************************/
2236 static void
2237 ixv_handle_que(void *context, int pending)
2238 {
2239 	struct ix_queue *que = context;
2240 	struct adapter  *adapter = que->adapter;
2241 	struct tx_ring  *txr = que->txr;
2242 	struct ifnet    *ifp = adapter->ifp;
2243 	bool            more;
2244 
2245 	if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
2246 		more = ixgbe_rxeof(que);
2247 		IXGBE_TX_LOCK(txr);
2248 		ixgbe_txeof(txr);
2249 		if (!ixv_ring_empty(ifp, txr->br))
2250 			ixv_start_locked(ifp, txr);
2251 		IXGBE_TX_UNLOCK(txr);
2252 		if (more) {
2253 			taskqueue_enqueue(que->tq, &que->que_task);
2254 			return;
2255 		}
2256 	}
2257 
2258 	/* Re-enable this interrupt */
2259 	ixv_enable_queue(adapter, que->msix);
2260 
2261 	return;
2262 } /* ixv_handle_que */
2263 
2264 /************************************************************************
2265  * ixv_allocate_msix - Setup MSI-X Interrupt resources and handlers
2266  ************************************************************************/
2267 static int
2268 ixv_allocate_msix(struct adapter *adapter)
2269 {
2270 	device_t        dev = adapter->dev;
2271 	struct ix_queue *que = adapter->queues;
2272 	struct tx_ring  *txr = adapter->tx_rings;
2273 	int             error, msix_ctrl, rid, vector = 0;
2274 
2275 	for (int i = 0; i < adapter->num_queues; i++, vector++, que++, txr++) {
2276 		rid = vector + 1;
2277 		que->res = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid,
2278 		    RF_SHAREABLE | RF_ACTIVE);
2279 		if (que->res == NULL) {
2280 			device_printf(dev, "Unable to allocate bus resource: que interrupt [%d]\n",
2281 			    vector);
2282 			return (ENXIO);
2283 		}
2284 		/* Set the handler function */
2285 		error = bus_setup_intr(dev, que->res,
2286 		    INTR_TYPE_NET | INTR_MPSAFE, NULL,
2287 		    ixv_msix_que, que, &que->tag);
2288 		if (error) {
2289 			que->res = NULL;
2290 			device_printf(dev, "Failed to register QUE handler");
2291 			return (error);
2292 		}
2293 #if __FreeBSD_version >= 800504
2294 		bus_describe_intr(dev, que->res, que->tag, "que %d", i);
2295 #endif
2296 		que->msix = vector;
2297 		adapter->active_queues |= (u64)(1 << que->msix);
2298 		/*
2299 		 * Bind the MSI-X vector, and thus the
2300 		 * ring to the corresponding CPU.
2301 		 */
2302 		if (adapter->num_queues > 1)
2303 			bus_bind_intr(dev, que->res, i);
2304 		TASK_INIT(&txr->txq_task, 0, ixgbe_deferred_mq_start, txr);
2305 		TASK_INIT(&que->que_task, 0, ixv_handle_que, que);
2306 		que->tq = taskqueue_create_fast("ixv_que", M_NOWAIT,
2307 		    taskqueue_thread_enqueue, &que->tq);
2308 		taskqueue_start_threads(&que->tq, 1, PI_NET, "%s que",
2309 		    device_get_nameunit(adapter->dev));
2310 	}
2311 
2312 	/* and Mailbox */
2313 	rid = vector + 1;
2314 	adapter->res = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid,
2315 	    RF_SHAREABLE | RF_ACTIVE);
2316 	if (!adapter->res) {
2317 		device_printf(dev,
2318 		    "Unable to allocate bus resource: MBX interrupt [%d]\n",
2319 		    rid);
2320 		return (ENXIO);
2321 	}
2322 	/* Set the mbx handler function */
2323 	error = bus_setup_intr(dev, adapter->res, INTR_TYPE_NET | INTR_MPSAFE,
2324 	    NULL, ixv_msix_mbx, adapter, &adapter->tag);
2325 	if (error) {
2326 		adapter->res = NULL;
2327 		device_printf(dev, "Failed to register LINK handler");
2328 		return (error);
2329 	}
2330 #if __FreeBSD_version >= 800504
2331 	bus_describe_intr(dev, adapter->res, adapter->tag, "mbx");
2332 #endif
2333 	adapter->vector = vector;
2334 	/* Tasklets for Mailbox */
2335 	TASK_INIT(&adapter->link_task, 0, ixv_handle_link, adapter);
2336 	adapter->tq = taskqueue_create_fast("ixv_mbx", M_NOWAIT,
2337 	    taskqueue_thread_enqueue, &adapter->tq);
2338 	taskqueue_start_threads(&adapter->tq, 1, PI_NET, "%s mbxq",
2339 	    device_get_nameunit(adapter->dev));
2340 	/*
2341 	 * Due to a broken design QEMU will fail to properly
2342 	 * enable the guest for MSI-X unless the vectors in
2343 	 * the table are all set up, so we must rewrite the
2344 	 * ENABLE in the MSI-X control register again at this
2345 	 * point to cause it to successfully initialize us.
2346 	 */
2347 	if (adapter->hw.mac.type == ixgbe_mac_82599_vf) {
2348 		pci_find_cap(dev, PCIY_MSIX, &rid);
2349 		rid += PCIR_MSIX_CTRL;
2350 		msix_ctrl = pci_read_config(dev, rid, 2);
2351 		msix_ctrl |= PCIM_MSIXCTRL_MSIX_ENABLE;
2352 		pci_write_config(dev, rid, msix_ctrl, 2);
2353 	}
2354 
2355 	return (0);
2356 } /* ixv_allocate_msix */
2357 
2358 /************************************************************************
2359  * ixv_configure_interrupts - Setup MSI-X resources
2360  *
2361  *   Note: The VF device MUST use MSI-X, there is no fallback.
2362  ************************************************************************/
2363 static int
2364 ixv_configure_interrupts(struct adapter *adapter)
2365 {
2366 	device_t dev = adapter->dev;
2367 	int      rid, want, msgs;
2368 
2369 	/* Must have at least 2 MSI-X vectors */
2370 	msgs = pci_msix_count(dev);
2371 	if (msgs < 2)
2372 		goto out;
2373 	rid = PCIR_BAR(3);
2374 	adapter->msix_mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid,
2375 	    RF_ACTIVE);
2376 	if (adapter->msix_mem == NULL) {
2377 		device_printf(adapter->dev, "Unable to map MSI-X table \n");
2378 		goto out;
2379 	}
2380 
2381 	/*
2382 	 * Want vectors for the queues,
2383 	 * plus an additional for mailbox.
2384 	 */
2385 	want = adapter->num_queues + 1;
2386 	if (want > msgs) {
2387 		want = msgs;
2388 		adapter->num_queues = msgs - 1;
2389 	} else
2390 		msgs = want;
2391 	if ((pci_alloc_msix(dev, &msgs) == 0) && (msgs == want)) {
2392 		device_printf(adapter->dev,
2393 		    "Using MSI-X interrupts with %d vectors\n", want);
2394 		/* reflect correct sysctl value */
2395 		ixv_num_queues = adapter->num_queues;
2396 
2397 		return (0);
2398 	}
2399 	/* Release in case alloc was insufficient */
2400 	pci_release_msi(dev);
2401 out:
2402 	if (adapter->msix_mem != NULL) {
2403 		bus_release_resource(dev, SYS_RES_MEMORY, rid,
2404 		    adapter->msix_mem);
2405 		adapter->msix_mem = NULL;
2406 	}
2407 	device_printf(adapter->dev, "MSI-X config error\n");
2408 
2409 	return (ENXIO);
2410 } /* ixv_configure_interrupts */
2411 
2412 
2413 /************************************************************************
2414  * ixv_handle_link - Tasklet handler for MSI-X MBX interrupts
2415  *
2416  *   Done outside of interrupt context since the driver might sleep
2417  ************************************************************************/
2418 static void
2419 ixv_handle_link(void *context, int pending)
2420 {
2421 	struct adapter *adapter = context;
2422 
2423 	adapter->hw.mac.ops.check_link(&adapter->hw, &adapter->link_speed,
2424 	    &adapter->link_up, FALSE);
2425 	ixv_update_link_status(adapter);
2426 } /* ixv_handle_link */
2427 
2428 /************************************************************************
2429  * ixv_check_link - Used in the local timer to poll for link changes
2430  ************************************************************************/
2431 static void
2432 ixv_check_link(struct adapter *adapter)
2433 {
2434 	adapter->hw.mac.get_link_status = TRUE;
2435 
2436 	adapter->hw.mac.ops.check_link(&adapter->hw, &adapter->link_speed,
2437 	    &adapter->link_up, FALSE);
2438 	ixv_update_link_status(adapter);
2439 } /* ixv_check_link */
2440 
2441