xref: /freebsd/sys/dev/vnic/nicvf_main.c (revision 3416500aef140042c64bc149cb1ec6620483bc44)
1 /*
2  * Copyright (C) 2015 Cavium Inc.
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  *
26  * $FreeBSD$
27  *
28  */
29 #include <sys/cdefs.h>
30 __FBSDID("$FreeBSD$");
31 
32 #include "opt_inet.h"
33 #include "opt_inet6.h"
34 
35 #include <sys/param.h>
36 #include <sys/systm.h>
37 #include <sys/bitset.h>
38 #include <sys/bitstring.h>
39 #include <sys/bus.h>
40 #include <sys/endian.h>
41 #include <sys/kernel.h>
42 #include <sys/malloc.h>
43 #include <sys/mbuf.h>
44 #include <sys/module.h>
45 #include <sys/rman.h>
46 #include <sys/pciio.h>
47 #include <sys/pcpu.h>
48 #include <sys/proc.h>
49 #include <sys/socket.h>
50 #include <sys/sockio.h>
51 #include <sys/stdatomic.h>
52 #include <sys/cpuset.h>
53 #include <sys/lock.h>
54 #include <sys/mutex.h>
55 #include <sys/smp.h>
56 #include <sys/taskqueue.h>
57 
58 #include <net/bpf.h>
59 #include <net/ethernet.h>
60 #include <net/if.h>
61 #include <net/if_var.h>
62 #include <net/if_arp.h>
63 #include <net/if_dl.h>
64 #include <net/if_media.h>
65 #include <net/if_types.h>
66 #include <net/if_vlan_var.h>
67 
68 #include <netinet/in.h>
69 #include <netinet/ip.h>
70 #include <netinet/if_ether.h>
71 #include <netinet/tcp_lro.h>
72 
73 #include <dev/pci/pcireg.h>
74 #include <dev/pci/pcivar.h>
75 
76 #include <sys/dnv.h>
77 #include <sys/nv.h>
78 #include <sys/iov_schema.h>
79 
80 #include <machine/bus.h>
81 
82 #include "thunder_bgx.h"
83 #include "nic_reg.h"
84 #include "nic.h"
85 #include "nicvf_queues.h"
86 
87 #define	VNIC_VF_DEVSTR		"Cavium Thunder NIC Virtual Function Driver"
88 
89 #define	VNIC_VF_REG_RID		PCIR_BAR(PCI_CFG_REG_BAR_NUM)
90 
91 /* Lock for core interface settings */
92 #define	NICVF_CORE_LOCK_INIT(nic)				\
93     sx_init(&(nic)->core_sx, device_get_nameunit((nic)->dev))
94 
95 #define	NICVF_CORE_LOCK_DESTROY(nic)				\
96     sx_destroy(&(nic)->core_sx)
97 
98 #define	NICVF_CORE_LOCK(nic)		sx_xlock(&(nic)->core_sx)
99 #define	NICVF_CORE_UNLOCK(nic)		sx_xunlock(&(nic)->core_sx)
100 
101 #define	NICVF_CORE_LOCK_ASSERT(nic)	sx_assert(&(nic)->core_sx, SA_XLOCKED)
102 
103 #define	SPEED_10	10
104 #define	SPEED_100	100
105 #define	SPEED_1000	1000
106 #define	SPEED_10000	10000
107 #define	SPEED_40000	40000
108 
109 MALLOC_DEFINE(M_NICVF, "nicvf", "ThunderX VNIC VF dynamic memory");
110 
111 static int nicvf_probe(device_t);
112 static int nicvf_attach(device_t);
113 static int nicvf_detach(device_t);
114 
115 static device_method_t nicvf_methods[] = {
116 	/* Device interface */
117 	DEVMETHOD(device_probe,		nicvf_probe),
118 	DEVMETHOD(device_attach,	nicvf_attach),
119 	DEVMETHOD(device_detach,	nicvf_detach),
120 
121 	DEVMETHOD_END,
122 };
123 
124 static driver_t nicvf_driver = {
125 	"vnic",
126 	nicvf_methods,
127 	sizeof(struct nicvf),
128 };
129 
130 static devclass_t nicvf_devclass;
131 
132 DRIVER_MODULE(vnicvf, pci, nicvf_driver, nicvf_devclass, 0, 0);
133 MODULE_VERSION(vnicvf, 1);
134 MODULE_DEPEND(vnicvf, pci, 1, 1, 1);
135 MODULE_DEPEND(vnicvf, ether, 1, 1, 1);
136 MODULE_DEPEND(vnicvf, vnicpf, 1, 1, 1);
137 
138 static int nicvf_allocate_misc_interrupt(struct nicvf *);
139 static int nicvf_enable_misc_interrupt(struct nicvf *);
140 static int nicvf_allocate_net_interrupts(struct nicvf *);
141 static void nicvf_release_all_interrupts(struct nicvf *);
142 static int nicvf_update_hw_max_frs(struct nicvf *, int);
143 static int nicvf_hw_set_mac_addr(struct nicvf *, uint8_t *);
144 static void nicvf_config_cpi(struct nicvf *);
145 static int nicvf_rss_init(struct nicvf *);
146 static int nicvf_init_resources(struct nicvf *);
147 
148 static int nicvf_setup_ifnet(struct nicvf *);
149 static int nicvf_setup_ifmedia(struct nicvf *);
150 static void nicvf_hw_addr_random(uint8_t *);
151 
152 static int nicvf_if_ioctl(struct ifnet *, u_long, caddr_t);
153 static void nicvf_if_init(void *);
154 static void nicvf_if_init_locked(struct nicvf *);
155 static int nicvf_if_transmit(struct ifnet *, struct mbuf *);
156 static void nicvf_if_qflush(struct ifnet *);
157 static uint64_t nicvf_if_getcounter(struct ifnet *, ift_counter);
158 
159 static int nicvf_stop_locked(struct nicvf *);
160 
161 static void nicvf_media_status(struct ifnet *, struct ifmediareq *);
162 static int nicvf_media_change(struct ifnet *);
163 
164 static void nicvf_tick_stats(void *);
165 
166 static int
167 nicvf_probe(device_t dev)
168 {
169 	uint16_t vendor_id;
170 	uint16_t device_id;
171 
172 	vendor_id = pci_get_vendor(dev);
173 	device_id = pci_get_device(dev);
174 
175 	if (vendor_id != PCI_VENDOR_ID_CAVIUM)
176 		return (ENXIO);
177 
178 	if (device_id == PCI_DEVICE_ID_THUNDER_NIC_VF ||
179 	    device_id == PCI_DEVICE_ID_THUNDER_PASS1_NIC_VF) {
180 		device_set_desc(dev, VNIC_VF_DEVSTR);
181 		return (BUS_PROBE_DEFAULT);
182 	}
183 
184 	return (ENXIO);
185 }
186 
187 static int
188 nicvf_attach(device_t dev)
189 {
190 	struct nicvf *nic;
191 	int rid, qcount;
192 	int err = 0;
193 	uint8_t hwaddr[ETHER_ADDR_LEN];
194 	uint8_t zeromac[] = {[0 ... (ETHER_ADDR_LEN - 1)] = 0};
195 
196 	nic = device_get_softc(dev);
197 	nic->dev = dev;
198 	nic->pnicvf = nic;
199 
200 	NICVF_CORE_LOCK_INIT(nic);
201 	/* Enable HW TSO on Pass2 */
202 	if (!pass1_silicon(dev))
203 		nic->hw_tso = TRUE;
204 
205 	rid = VNIC_VF_REG_RID;
206 	nic->reg_base = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid,
207 	    RF_ACTIVE);
208 	if (nic->reg_base == NULL) {
209 		device_printf(dev, "Could not allocate registers memory\n");
210 		return (ENXIO);
211 	}
212 
213 	qcount = MAX_CMP_QUEUES_PER_QS;
214 	nic->max_queues = qcount;
215 
216 	err = nicvf_set_qset_resources(nic);
217 	if (err != 0)
218 		goto err_free_res;
219 
220 	/* Check if PF is alive and get MAC address for this VF */
221 	err = nicvf_allocate_misc_interrupt(nic);
222 	if (err != 0)
223 		goto err_free_res;
224 
225 	NICVF_CORE_LOCK(nic);
226 	err = nicvf_enable_misc_interrupt(nic);
227 	NICVF_CORE_UNLOCK(nic);
228 	if (err != 0)
229 		goto err_release_intr;
230 
231 	err = nicvf_allocate_net_interrupts(nic);
232 	if (err != 0) {
233 		device_printf(dev,
234 		    "Could not allocate network interface interrupts\n");
235 		goto err_free_ifnet;
236 	}
237 
238 	/* If no MAC address was obtained we generate random one */
239 	if (memcmp(nic->hwaddr, zeromac, ETHER_ADDR_LEN) == 0) {
240 		nicvf_hw_addr_random(hwaddr);
241 		memcpy(nic->hwaddr, hwaddr, ETHER_ADDR_LEN);
242 		NICVF_CORE_LOCK(nic);
243 		nicvf_hw_set_mac_addr(nic, hwaddr);
244 		NICVF_CORE_UNLOCK(nic);
245 	}
246 
247 	/* Configure CPI alorithm */
248 	nic->cpi_alg = CPI_ALG_NONE;
249 	NICVF_CORE_LOCK(nic);
250 	nicvf_config_cpi(nic);
251 	/* Configure receive side scaling */
252 	if (nic->qs->rq_cnt > 1)
253 		nicvf_rss_init(nic);
254 	NICVF_CORE_UNLOCK(nic);
255 
256 	err = nicvf_setup_ifnet(nic);
257 	if (err != 0) {
258 		device_printf(dev, "Could not set-up ifnet\n");
259 		goto err_release_intr;
260 	}
261 
262 	err = nicvf_setup_ifmedia(nic);
263 	if (err != 0) {
264 		device_printf(dev, "Could not set-up ifmedia\n");
265 		goto err_free_ifnet;
266 	}
267 
268 	mtx_init(&nic->stats_mtx, "VNIC stats", NULL, MTX_DEF);
269 	callout_init_mtx(&nic->stats_callout, &nic->stats_mtx, 0);
270 
271 	ether_ifattach(nic->ifp, nic->hwaddr);
272 
273 	return (0);
274 
275 err_free_ifnet:
276 	if_free(nic->ifp);
277 err_release_intr:
278 	nicvf_release_all_interrupts(nic);
279 err_free_res:
280 	bus_release_resource(dev, SYS_RES_MEMORY, rman_get_rid(nic->reg_base),
281 	    nic->reg_base);
282 
283 	return (err);
284 }
285 
286 static int
287 nicvf_detach(device_t dev)
288 {
289 	struct nicvf *nic;
290 
291 	nic = device_get_softc(dev);
292 
293 	NICVF_CORE_LOCK(nic);
294 	/* Shut down the port and release ring resources */
295 	nicvf_stop_locked(nic);
296 	/* Release stats lock */
297 	mtx_destroy(&nic->stats_mtx);
298 	/* Release interrupts */
299 	nicvf_release_all_interrupts(nic);
300 	/* Release memory resource */
301 	if (nic->reg_base != NULL) {
302 		bus_release_resource(dev, SYS_RES_MEMORY,
303 		    rman_get_rid(nic->reg_base), nic->reg_base);
304 	}
305 
306 	/* Remove all ifmedia configurations */
307 	ifmedia_removeall(&nic->if_media);
308 	/* Free this ifnet */
309 	if_free(nic->ifp);
310 	NICVF_CORE_UNLOCK(nic);
311 	/* Finally destroy the lock */
312 	NICVF_CORE_LOCK_DESTROY(nic);
313 
314 	return (0);
315 }
316 
317 static void
318 nicvf_hw_addr_random(uint8_t *hwaddr)
319 {
320 	uint32_t rnd;
321 	uint8_t addr[ETHER_ADDR_LEN];
322 
323 	/*
324 	 * Create randomized MAC address.
325 	 * Set 'bsd' + random 24 low-order bits.
326 	 */
327 	rnd = arc4random() & 0x00ffffff;
328 	addr[0] = 'b';
329 	addr[1] = 's';
330 	addr[2] = 'd';
331 	addr[3] = rnd >> 16;
332 	addr[4] = rnd >> 8;
333 	addr[5] = rnd >> 0;
334 
335 	memcpy(hwaddr, addr, ETHER_ADDR_LEN);
336 }
337 
338 static int
339 nicvf_setup_ifnet(struct nicvf *nic)
340 {
341 	struct ifnet *ifp;
342 
343 	ifp = if_alloc(IFT_ETHER);
344 	if (ifp == NULL) {
345 		device_printf(nic->dev, "Could not allocate ifnet structure\n");
346 		return (ENOMEM);
347 	}
348 
349 	nic->ifp = ifp;
350 
351 	if_setsoftc(ifp, nic);
352 	if_initname(ifp, device_get_name(nic->dev), device_get_unit(nic->dev));
353 	if_setflags(ifp, IFF_BROADCAST | IFF_SIMPLEX);
354 
355 	if_settransmitfn(ifp, nicvf_if_transmit);
356 	if_setqflushfn(ifp, nicvf_if_qflush);
357 	if_setioctlfn(ifp, nicvf_if_ioctl);
358 	if_setinitfn(ifp, nicvf_if_init);
359 	if_setgetcounterfn(ifp, nicvf_if_getcounter);
360 
361 	if_setmtu(ifp, ETHERMTU);
362 
363 	/* Reset caps */
364 	if_setcapabilities(ifp, 0);
365 
366 	/* Set the default values */
367 	if_setcapabilitiesbit(ifp, IFCAP_VLAN_MTU | IFCAP_JUMBO_MTU, 0);
368 	if_setcapabilitiesbit(ifp, IFCAP_LRO, 0);
369 	if (nic->hw_tso) {
370 		/* TSO */
371 		if_setcapabilitiesbit(ifp, IFCAP_TSO4, 0);
372 		/* TSO parameters */
373 		if_sethwtsomax(ifp, NICVF_TSO_MAXSIZE);
374 		if_sethwtsomaxsegcount(ifp, NICVF_TSO_NSEGS);
375 		if_sethwtsomaxsegsize(ifp, MCLBYTES);
376 	}
377 	/* IP/TCP/UDP HW checksums */
378 	if_setcapabilitiesbit(ifp, IFCAP_HWCSUM, 0);
379 	if_setcapabilitiesbit(ifp, IFCAP_HWSTATS, 0);
380 	/*
381 	 * HW offload enable
382 	 */
383 	if_clearhwassist(ifp);
384 	if_sethwassistbits(ifp, (CSUM_IP | CSUM_TCP | CSUM_UDP | CSUM_SCTP), 0);
385 	if (nic->hw_tso)
386 		if_sethwassistbits(ifp, (CSUM_TSO), 0);
387 	if_setcapenable(ifp, if_getcapabilities(ifp));
388 
389 	return (0);
390 }
391 
392 static int
393 nicvf_setup_ifmedia(struct nicvf *nic)
394 {
395 
396 	ifmedia_init(&nic->if_media, IFM_IMASK, nicvf_media_change,
397 	    nicvf_media_status);
398 
399 	/*
400 	 * Advertise availability of all possible connection types,
401 	 * even though not all are possible at the same time.
402 	 */
403 
404 	ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_10_T | IFM_FDX),
405 	    0, NULL);
406 	ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_100_TX | IFM_FDX),
407 	    0, NULL);
408 	ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_1000_T | IFM_FDX),
409 	    0, NULL);
410 	ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_10G_SR | IFM_FDX),
411 	    0, NULL);
412 	ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_40G_CR4 | IFM_FDX),
413 	    0, NULL);
414 	ifmedia_add(&nic->if_media, (IFM_ETHER | IFM_AUTO | IFM_FDX),
415 	    0, NULL);
416 
417 	ifmedia_set(&nic->if_media, (IFM_ETHER | IFM_AUTO | IFM_FDX));
418 
419 	return (0);
420 }
421 
422 static int
423 nicvf_if_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
424 {
425 	struct nicvf *nic;
426 	struct rcv_queue *rq;
427 	struct ifreq *ifr;
428 	uint32_t flags;
429 	int mask, err;
430 	int rq_idx;
431 #if defined(INET) || defined(INET6)
432 	struct ifaddr *ifa;
433 	boolean_t avoid_reset = FALSE;
434 #endif
435 
436 	nic = if_getsoftc(ifp);
437 	ifr = (struct ifreq *)data;
438 #if defined(INET) || defined(INET6)
439 	ifa = (struct ifaddr *)data;
440 #endif
441 	err = 0;
442 	switch (cmd) {
443 	case SIOCSIFADDR:
444 #ifdef INET
445 		if (ifa->ifa_addr->sa_family == AF_INET)
446 			avoid_reset = TRUE;
447 #endif
448 #ifdef INET6
449 		if (ifa->ifa_addr->sa_family == AF_INET6)
450 			avoid_reset = TRUE;
451 #endif
452 
453 #if defined(INET) || defined(INET6)
454 		/* Avoid reinitialization unless it's necessary */
455 		if (avoid_reset) {
456 			if_setflagbits(ifp, IFF_UP, 0);
457 			if (!(if_getdrvflags(ifp) & IFF_DRV_RUNNING))
458 				nicvf_if_init(nic);
459 #ifdef INET
460 			if (!(if_getflags(ifp) & IFF_NOARP))
461 				arp_ifinit(ifp, ifa);
462 #endif
463 
464 			return (0);
465 		}
466 #endif
467 		err = ether_ioctl(ifp, cmd, data);
468 		break;
469 	case SIOCSIFMTU:
470 		if (ifr->ifr_mtu < NIC_HW_MIN_FRS ||
471 		    ifr->ifr_mtu > NIC_HW_MAX_FRS) {
472 			err = EINVAL;
473 		} else {
474 			NICVF_CORE_LOCK(nic);
475 			err = nicvf_update_hw_max_frs(nic, ifr->ifr_mtu);
476 			if (err == 0)
477 				if_setmtu(ifp, ifr->ifr_mtu);
478 			NICVF_CORE_UNLOCK(nic);
479 		}
480 		break;
481 	case SIOCSIFFLAGS:
482 		NICVF_CORE_LOCK(nic);
483 		if (if_getflags(ifp) & IFF_UP) {
484 			if (if_getdrvflags(ifp) & IFF_DRV_RUNNING) {
485 				flags = if_getflags(ifp) ^ nic->if_flags;
486 				if ((nic->if_flags & if_getflags(ifp)) &
487 				    IFF_PROMISC) {
488 					/* Change promiscous mode */
489 #if 0
490 					/* ARM64TODO */
491 					nicvf_set_promiscous(nic);
492 #endif
493 				}
494 
495 				if ((nic->if_flags ^ if_getflags(ifp)) &
496 				    IFF_ALLMULTI) {
497 					/* Change multicasting settings */
498 #if 0
499 					/* ARM64TODO */
500 					nicvf_set_multicast(nic);
501 #endif
502 				}
503 			} else {
504 				nicvf_if_init_locked(nic);
505 			}
506 		} else if (if_getdrvflags(ifp) & IFF_DRV_RUNNING)
507 			nicvf_stop_locked(nic);
508 
509 		nic->if_flags = if_getflags(ifp);
510 		NICVF_CORE_UNLOCK(nic);
511 		break;
512 
513 	case SIOCADDMULTI:
514 	case SIOCDELMULTI:
515 		if (if_getdrvflags(ifp) & IFF_DRV_RUNNING) {
516 #if 0
517 			NICVF_CORE_LOCK(nic);
518 			/* ARM64TODO */
519 			nicvf_set_multicast(nic);
520 			NICVF_CORE_UNLOCK(nic);
521 #endif
522 		}
523 		break;
524 
525 	case SIOCSIFMEDIA:
526 	case SIOCGIFMEDIA:
527 		err = ifmedia_ioctl(ifp, ifr, &nic->if_media, cmd);
528 		break;
529 
530 	case SIOCSIFCAP:
531 		mask = if_getcapenable(ifp) ^ ifr->ifr_reqcap;
532 		if (mask & IFCAP_VLAN_MTU) {
533 			/* No work to do except acknowledge the change took. */
534 			if_togglecapenable(ifp, IFCAP_VLAN_MTU);
535 		}
536 		if (mask & IFCAP_TXCSUM)
537 			if_togglecapenable(ifp, IFCAP_TXCSUM);
538 		if (mask & IFCAP_RXCSUM)
539 			if_togglecapenable(ifp, IFCAP_RXCSUM);
540 		if ((mask & IFCAP_TSO4) && nic->hw_tso)
541 			if_togglecapenable(ifp, IFCAP_TSO4);
542 		if (mask & IFCAP_LRO) {
543 			/*
544 			 * Lock the driver for a moment to avoid
545 			 * mismatch in per-queue settings.
546 			 */
547 			NICVF_CORE_LOCK(nic);
548 			if_togglecapenable(ifp, IFCAP_LRO);
549 			if ((if_getdrvflags(nic->ifp) & IFF_DRV_RUNNING) != 0) {
550 				/*
551 				 * Now disable LRO for subsequent packets.
552 				 * Atomicity of this change is not necessary
553 				 * as we don't need precise toggle of this
554 				 * feature for all threads processing the
555 				 * completion queue.
556 				 */
557 				for (rq_idx = 0;
558 				    rq_idx < nic->qs->rq_cnt; rq_idx++) {
559 					rq = &nic->qs->rq[rq_idx];
560 					rq->lro_enabled = !rq->lro_enabled;
561 				}
562 			}
563 			NICVF_CORE_UNLOCK(nic);
564 		}
565 
566 		break;
567 
568 	default:
569 		err = ether_ioctl(ifp, cmd, data);
570 		break;
571 	}
572 
573 	return (err);
574 }
575 
576 static void
577 nicvf_if_init_locked(struct nicvf *nic)
578 {
579 	struct queue_set *qs = nic->qs;
580 	struct ifnet *ifp;
581 	int qidx;
582 	int err;
583 	caddr_t if_addr;
584 
585 	NICVF_CORE_LOCK_ASSERT(nic);
586 	ifp = nic->ifp;
587 
588 	if ((if_getdrvflags(ifp) & IFF_DRV_RUNNING) != 0)
589 		nicvf_stop_locked(nic);
590 
591 	err = nicvf_enable_misc_interrupt(nic);
592 	if (err != 0) {
593 		if_printf(ifp, "Could not reenable Mbox interrupt\n");
594 		return;
595 	}
596 
597 	/* Get the latest MAC address */
598 	if_addr = if_getlladdr(ifp);
599 	/* Update MAC address if changed */
600 	if (memcmp(nic->hwaddr, if_addr, ETHER_ADDR_LEN) != 0) {
601 		memcpy(nic->hwaddr, if_addr, ETHER_ADDR_LEN);
602 		nicvf_hw_set_mac_addr(nic, if_addr);
603 	}
604 
605 	/* Initialize the queues */
606 	err = nicvf_init_resources(nic);
607 	if (err != 0)
608 		goto error;
609 
610 	/* Make sure queue initialization is written */
611 	wmb();
612 
613 	nicvf_reg_write(nic, NIC_VF_INT, ~0UL);
614 	/* Enable Qset err interrupt */
615 	nicvf_enable_intr(nic, NICVF_INTR_QS_ERR, 0);
616 
617 	/* Enable completion queue interrupt */
618 	for (qidx = 0; qidx < qs->cq_cnt; qidx++)
619 		nicvf_enable_intr(nic, NICVF_INTR_CQ, qidx);
620 
621 	/* Enable RBDR threshold interrupt */
622 	for (qidx = 0; qidx < qs->rbdr_cnt; qidx++)
623 		nicvf_enable_intr(nic, NICVF_INTR_RBDR, qidx);
624 
625 	nic->drv_stats.txq_stop = 0;
626 	nic->drv_stats.txq_wake = 0;
627 
628 	/* Activate network interface */
629 	if_setdrvflagbits(ifp, IFF_DRV_RUNNING, IFF_DRV_OACTIVE);
630 
631 	/* Schedule callout to update stats */
632 	callout_reset(&nic->stats_callout, hz, nicvf_tick_stats, nic);
633 
634 	return;
635 
636 error:
637 	/* Something went very wrong. Disable this ifnet for good */
638 	if_setdrvflagbits(ifp, IFF_DRV_OACTIVE, IFF_DRV_RUNNING);
639 }
640 
641 static void
642 nicvf_if_init(void *if_softc)
643 {
644 	struct nicvf *nic = if_softc;
645 
646 	NICVF_CORE_LOCK(nic);
647 	nicvf_if_init_locked(nic);
648 	NICVF_CORE_UNLOCK(nic);
649 }
650 
651 static int
652 nicvf_if_transmit(struct ifnet *ifp, struct mbuf *mbuf)
653 {
654 	struct nicvf *nic = if_getsoftc(ifp);
655 	struct queue_set *qs = nic->qs;
656 	struct snd_queue *sq;
657 	struct mbuf *mtmp;
658 	int qidx;
659 	int err = 0;
660 
661 
662 	if (__predict_false(qs == NULL)) {
663 		panic("%s: missing queue set for %s", __func__,
664 		    device_get_nameunit(nic->dev));
665 	}
666 
667 	/* Select queue */
668 	if (M_HASHTYPE_GET(mbuf) != M_HASHTYPE_NONE)
669 		qidx = mbuf->m_pkthdr.flowid % qs->sq_cnt;
670 	else
671 		qidx = curcpu % qs->sq_cnt;
672 
673 	sq = &qs->sq[qidx];
674 
675 	if (mbuf->m_next != NULL &&
676 	    (mbuf->m_pkthdr.csum_flags &
677 	    (CSUM_IP | CSUM_TCP | CSUM_UDP | CSUM_SCTP)) != 0) {
678 		if (M_WRITABLE(mbuf) == 0) {
679 			mtmp = m_dup(mbuf, M_NOWAIT);
680 			m_freem(mbuf);
681 			if (mtmp == NULL)
682 				return (ENOBUFS);
683 			mbuf = mtmp;
684 		}
685 	}
686 
687 	err = drbr_enqueue(ifp, sq->br, mbuf);
688 	if (((if_getdrvflags(ifp) & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
689 	    IFF_DRV_RUNNING) || !nic->link_up || (err != 0)) {
690 		/*
691 		 * Try to enqueue packet to the ring buffer.
692 		 * If the driver is not active, link down or enqueue operation
693 		 * failed, return with the appropriate error code.
694 		 */
695 		return (err);
696 	}
697 
698 	if (NICVF_TX_TRYLOCK(sq) != 0) {
699 		err = nicvf_xmit_locked(sq);
700 		NICVF_TX_UNLOCK(sq);
701 		return (err);
702 	} else
703 		taskqueue_enqueue(sq->snd_taskq, &sq->snd_task);
704 
705 	return (0);
706 }
707 
708 static void
709 nicvf_if_qflush(struct ifnet *ifp)
710 {
711 	struct nicvf *nic;
712 	struct queue_set *qs;
713 	struct snd_queue *sq;
714 	struct mbuf *mbuf;
715 	size_t idx;
716 
717 	nic = if_getsoftc(ifp);
718 	qs = nic->qs;
719 
720 	for (idx = 0; idx < qs->sq_cnt; idx++) {
721 		sq = &qs->sq[idx];
722 		NICVF_TX_LOCK(sq);
723 		while ((mbuf = buf_ring_dequeue_sc(sq->br)) != NULL)
724 			m_freem(mbuf);
725 		NICVF_TX_UNLOCK(sq);
726 	}
727 	if_qflush(ifp);
728 }
729 
730 static uint64_t
731 nicvf_if_getcounter(struct ifnet *ifp, ift_counter cnt)
732 {
733 	struct nicvf *nic;
734 	struct nicvf_hw_stats *hw_stats;
735 	struct nicvf_drv_stats *drv_stats;
736 
737 	nic = if_getsoftc(ifp);
738 	hw_stats = &nic->hw_stats;
739 	drv_stats = &nic->drv_stats;
740 
741 	switch (cnt) {
742 	case IFCOUNTER_IPACKETS:
743 		return (drv_stats->rx_frames_ok);
744 	case IFCOUNTER_OPACKETS:
745 		return (drv_stats->tx_frames_ok);
746 	case IFCOUNTER_IBYTES:
747 		return (hw_stats->rx_bytes);
748 	case IFCOUNTER_OBYTES:
749 		return (hw_stats->tx_bytes_ok);
750 	case IFCOUNTER_IMCASTS:
751 		return (hw_stats->rx_mcast_frames);
752 	case IFCOUNTER_COLLISIONS:
753 		return (0);
754 	case IFCOUNTER_IQDROPS:
755 		return (drv_stats->rx_drops);
756 	case IFCOUNTER_OQDROPS:
757 		return (drv_stats->tx_drops);
758 	default:
759 		return (if_get_counter_default(ifp, cnt));
760 	}
761 
762 }
763 
764 static void
765 nicvf_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
766 {
767 	struct nicvf *nic = if_getsoftc(ifp);
768 
769 	NICVF_CORE_LOCK(nic);
770 
771 	ifmr->ifm_status = IFM_AVALID;
772 	ifmr->ifm_active = IFM_ETHER;
773 
774 	if (nic->link_up) {
775 		/* Device attached to working network */
776 		ifmr->ifm_status |= IFM_ACTIVE;
777 	}
778 
779 	switch (nic->speed) {
780 	case SPEED_10:
781 		ifmr->ifm_active |= IFM_10_T;
782 		break;
783 	case SPEED_100:
784 		ifmr->ifm_active |= IFM_100_TX;
785 		break;
786 	case SPEED_1000:
787 		ifmr->ifm_active |= IFM_1000_T;
788 		break;
789 	case SPEED_10000:
790 		ifmr->ifm_active |= IFM_10G_SR;
791 		break;
792 	case SPEED_40000:
793 		ifmr->ifm_active |= IFM_40G_CR4;
794 		break;
795 	default:
796 		ifmr->ifm_active |= IFM_AUTO;
797 		break;
798 	}
799 
800 	if (nic->duplex)
801 		ifmr->ifm_active |= IFM_FDX;
802 	else
803 		ifmr->ifm_active |= IFM_HDX;
804 
805 	NICVF_CORE_UNLOCK(nic);
806 }
807 
808 static int
809 nicvf_media_change(struct ifnet *ifp __unused)
810 {
811 
812 	return (0);
813 }
814 
815 /* Register read/write APIs */
816 void
817 nicvf_reg_write(struct nicvf *nic, bus_space_handle_t offset, uint64_t val)
818 {
819 
820 	bus_write_8(nic->reg_base, offset, val);
821 }
822 
823 uint64_t
824 nicvf_reg_read(struct nicvf *nic, uint64_t offset)
825 {
826 
827 	return (bus_read_8(nic->reg_base, offset));
828 }
829 
830 void
831 nicvf_queue_reg_write(struct nicvf *nic, bus_space_handle_t offset,
832     uint64_t qidx, uint64_t val)
833 {
834 
835 	bus_write_8(nic->reg_base, offset + (qidx << NIC_Q_NUM_SHIFT), val);
836 }
837 
838 uint64_t
839 nicvf_queue_reg_read(struct nicvf *nic, bus_space_handle_t offset,
840     uint64_t qidx)
841 {
842 
843 	return (bus_read_8(nic->reg_base, offset + (qidx << NIC_Q_NUM_SHIFT)));
844 }
845 
846 /* VF -> PF mailbox communication */
847 static void
848 nicvf_write_to_mbx(struct nicvf *nic, union nic_mbx *mbx)
849 {
850 	uint64_t *msg = (uint64_t *)mbx;
851 
852 	nicvf_reg_write(nic, NIC_VF_PF_MAILBOX_0_1 + 0, msg[0]);
853 	nicvf_reg_write(nic, NIC_VF_PF_MAILBOX_0_1 + 8, msg[1]);
854 }
855 
856 int
857 nicvf_send_msg_to_pf(struct nicvf *nic, union nic_mbx *mbx)
858 {
859 	int timeout = NIC_MBOX_MSG_TIMEOUT * 10;
860 	int sleep = 2;
861 
862 	NICVF_CORE_LOCK_ASSERT(nic);
863 
864 	nic->pf_acked = FALSE;
865 	nic->pf_nacked = FALSE;
866 
867 	nicvf_write_to_mbx(nic, mbx);
868 
869 	/* Wait for previous message to be acked, timeout 2sec */
870 	while (!nic->pf_acked) {
871 		if (nic->pf_nacked)
872 			return (EINVAL);
873 
874 		DELAY(sleep * 1000);
875 
876 		if (nic->pf_acked)
877 			break;
878 		timeout -= sleep;
879 		if (!timeout) {
880 			device_printf(nic->dev,
881 				   "PF didn't ack to mbox msg %d from VF%d\n",
882 				   (mbx->msg.msg & 0xFF), nic->vf_id);
883 
884 			return (EBUSY);
885 		}
886 	}
887 	return (0);
888 }
889 
890 /*
891  * Checks if VF is able to comminicate with PF
892  * and also gets the VNIC number this VF is associated to.
893  */
894 static int
895 nicvf_check_pf_ready(struct nicvf *nic)
896 {
897 	union nic_mbx mbx = {};
898 
899 	mbx.msg.msg = NIC_MBOX_MSG_READY;
900 	if (nicvf_send_msg_to_pf(nic, &mbx)) {
901 		device_printf(nic->dev,
902 			   "PF didn't respond to READY msg\n");
903 		return 0;
904 	}
905 
906 	return 1;
907 }
908 
909 static void
910 nicvf_read_bgx_stats(struct nicvf *nic, struct bgx_stats_msg *bgx)
911 {
912 
913 	if (bgx->rx)
914 		nic->bgx_stats.rx_stats[bgx->idx] = bgx->stats;
915 	else
916 		nic->bgx_stats.tx_stats[bgx->idx] = bgx->stats;
917 }
918 
919 static void
920 nicvf_handle_mbx_intr(struct nicvf *nic)
921 {
922 	union nic_mbx mbx = {};
923 	uint64_t *mbx_data;
924 	uint64_t mbx_addr;
925 	int i;
926 
927 	mbx_addr = NIC_VF_PF_MAILBOX_0_1;
928 	mbx_data = (uint64_t *)&mbx;
929 
930 	for (i = 0; i < NIC_PF_VF_MAILBOX_SIZE; i++) {
931 		*mbx_data = nicvf_reg_read(nic, mbx_addr);
932 		mbx_data++;
933 		mbx_addr += sizeof(uint64_t);
934 	}
935 
936 	switch (mbx.msg.msg) {
937 	case NIC_MBOX_MSG_READY:
938 		nic->pf_acked = TRUE;
939 		nic->vf_id = mbx.nic_cfg.vf_id & 0x7F;
940 		nic->tns_mode = mbx.nic_cfg.tns_mode & 0x7F;
941 		nic->node = mbx.nic_cfg.node_id;
942 		memcpy(nic->hwaddr, mbx.nic_cfg.mac_addr, ETHER_ADDR_LEN);
943 		nic->loopback_supported = mbx.nic_cfg.loopback_supported;
944 		nic->link_up = FALSE;
945 		nic->duplex = 0;
946 		nic->speed = 0;
947 		break;
948 	case NIC_MBOX_MSG_ACK:
949 		nic->pf_acked = TRUE;
950 		break;
951 	case NIC_MBOX_MSG_NACK:
952 		nic->pf_nacked = TRUE;
953 		break;
954 	case NIC_MBOX_MSG_RSS_SIZE:
955 		nic->rss_info.rss_size = mbx.rss_size.ind_tbl_size;
956 		nic->pf_acked = TRUE;
957 		break;
958 	case NIC_MBOX_MSG_BGX_STATS:
959 		nicvf_read_bgx_stats(nic, &mbx.bgx_stats);
960 		nic->pf_acked = TRUE;
961 		break;
962 	case NIC_MBOX_MSG_BGX_LINK_CHANGE:
963 		nic->pf_acked = TRUE;
964 		nic->link_up = mbx.link_status.link_up;
965 		nic->duplex = mbx.link_status.duplex;
966 		nic->speed = mbx.link_status.speed;
967 		if (nic->link_up) {
968 			if_setbaudrate(nic->ifp, nic->speed * 1000000);
969 			if_link_state_change(nic->ifp, LINK_STATE_UP);
970 		} else {
971 			if_setbaudrate(nic->ifp, 0);
972 			if_link_state_change(nic->ifp, LINK_STATE_DOWN);
973 		}
974 		break;
975 	default:
976 		device_printf(nic->dev,
977 			   "Invalid message from PF, msg 0x%x\n", mbx.msg.msg);
978 		break;
979 	}
980 	nicvf_clear_intr(nic, NICVF_INTR_MBOX, 0);
981 }
982 
983 static int
984 nicvf_update_hw_max_frs(struct nicvf *nic, int mtu)
985 {
986 	union nic_mbx mbx = {};
987 
988 	mbx.frs.msg = NIC_MBOX_MSG_SET_MAX_FRS;
989 	mbx.frs.max_frs = mtu;
990 	mbx.frs.vf_id = nic->vf_id;
991 
992 	return nicvf_send_msg_to_pf(nic, &mbx);
993 }
994 
995 static int
996 nicvf_hw_set_mac_addr(struct nicvf *nic, uint8_t *hwaddr)
997 {
998 	union nic_mbx mbx = {};
999 
1000 	mbx.mac.msg = NIC_MBOX_MSG_SET_MAC;
1001 	mbx.mac.vf_id = nic->vf_id;
1002 	memcpy(mbx.mac.mac_addr, hwaddr, ETHER_ADDR_LEN);
1003 
1004 	return (nicvf_send_msg_to_pf(nic, &mbx));
1005 }
1006 
1007 static void
1008 nicvf_config_cpi(struct nicvf *nic)
1009 {
1010 	union nic_mbx mbx = {};
1011 
1012 	mbx.cpi_cfg.msg = NIC_MBOX_MSG_CPI_CFG;
1013 	mbx.cpi_cfg.vf_id = nic->vf_id;
1014 	mbx.cpi_cfg.cpi_alg = nic->cpi_alg;
1015 	mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt;
1016 
1017 	nicvf_send_msg_to_pf(nic, &mbx);
1018 }
1019 
1020 static void
1021 nicvf_get_rss_size(struct nicvf *nic)
1022 {
1023 	union nic_mbx mbx = {};
1024 
1025 	mbx.rss_size.msg = NIC_MBOX_MSG_RSS_SIZE;
1026 	mbx.rss_size.vf_id = nic->vf_id;
1027 	nicvf_send_msg_to_pf(nic, &mbx);
1028 }
1029 
1030 static void
1031 nicvf_config_rss(struct nicvf *nic)
1032 {
1033 	union nic_mbx mbx = {};
1034 	struct nicvf_rss_info *rss;
1035 	int ind_tbl_len;
1036 	int i, nextq;
1037 
1038 	rss = &nic->rss_info;
1039 	ind_tbl_len = rss->rss_size;
1040 	nextq = 0;
1041 
1042 	mbx.rss_cfg.vf_id = nic->vf_id;
1043 	mbx.rss_cfg.hash_bits = rss->hash_bits;
1044 	while (ind_tbl_len != 0) {
1045 		mbx.rss_cfg.tbl_offset = nextq;
1046 		mbx.rss_cfg.tbl_len = MIN(ind_tbl_len,
1047 		    RSS_IND_TBL_LEN_PER_MBX_MSG);
1048 		mbx.rss_cfg.msg = mbx.rss_cfg.tbl_offset ?
1049 		    NIC_MBOX_MSG_RSS_CFG_CONT : NIC_MBOX_MSG_RSS_CFG;
1050 
1051 		for (i = 0; i < mbx.rss_cfg.tbl_len; i++)
1052 			mbx.rss_cfg.ind_tbl[i] = rss->ind_tbl[nextq++];
1053 
1054 		nicvf_send_msg_to_pf(nic, &mbx);
1055 
1056 		ind_tbl_len -= mbx.rss_cfg.tbl_len;
1057 	}
1058 }
1059 
1060 static void
1061 nicvf_set_rss_key(struct nicvf *nic)
1062 {
1063 	struct nicvf_rss_info *rss;
1064 	uint64_t key_addr;
1065 	int idx;
1066 
1067 	rss = &nic->rss_info;
1068 	key_addr = NIC_VNIC_RSS_KEY_0_4;
1069 
1070 	for (idx = 0; idx < RSS_HASH_KEY_SIZE; idx++) {
1071 		nicvf_reg_write(nic, key_addr, rss->key[idx]);
1072 		key_addr += sizeof(uint64_t);
1073 	}
1074 }
1075 
1076 static int
1077 nicvf_rss_init(struct nicvf *nic)
1078 {
1079 	struct nicvf_rss_info *rss;
1080 	int idx;
1081 
1082 	nicvf_get_rss_size(nic);
1083 
1084 	rss = &nic->rss_info;
1085 	if (nic->cpi_alg != CPI_ALG_NONE) {
1086 		rss->enable = FALSE;
1087 		rss->hash_bits = 0;
1088 		return (ENXIO);
1089 	}
1090 
1091 	rss->enable = TRUE;
1092 
1093 	/* Using the HW reset value for now */
1094 	rss->key[0] = 0xFEED0BADFEED0BADUL;
1095 	rss->key[1] = 0xFEED0BADFEED0BADUL;
1096 	rss->key[2] = 0xFEED0BADFEED0BADUL;
1097 	rss->key[3] = 0xFEED0BADFEED0BADUL;
1098 	rss->key[4] = 0xFEED0BADFEED0BADUL;
1099 
1100 	nicvf_set_rss_key(nic);
1101 
1102 	rss->cfg = RSS_IP_HASH_ENA | RSS_TCP_HASH_ENA | RSS_UDP_HASH_ENA;
1103 	nicvf_reg_write(nic, NIC_VNIC_RSS_CFG, rss->cfg);
1104 
1105 	rss->hash_bits = fls(rss->rss_size) - 1;
1106 	for (idx = 0; idx < rss->rss_size; idx++)
1107 		rss->ind_tbl[idx] = idx % nic->rx_queues;
1108 
1109 	nicvf_config_rss(nic);
1110 
1111 	return (0);
1112 }
1113 
1114 static int
1115 nicvf_init_resources(struct nicvf *nic)
1116 {
1117 	int err;
1118 	union nic_mbx mbx = {};
1119 
1120 	mbx.msg.msg = NIC_MBOX_MSG_CFG_DONE;
1121 
1122 	/* Enable Qset */
1123 	nicvf_qset_config(nic, TRUE);
1124 
1125 	/* Initialize queues and HW for data transfer */
1126 	err = nicvf_config_data_transfer(nic, TRUE);
1127 	if (err) {
1128 		device_printf(nic->dev,
1129 		    "Failed to alloc/config VF's QSet resources\n");
1130 		return (err);
1131 	}
1132 
1133 	/* Send VF config done msg to PF */
1134 	nicvf_write_to_mbx(nic, &mbx);
1135 
1136 	return (0);
1137 }
1138 
1139 static void
1140 nicvf_misc_intr_handler(void *arg)
1141 {
1142 	struct nicvf *nic = (struct nicvf *)arg;
1143 	uint64_t intr;
1144 
1145 	intr = nicvf_reg_read(nic, NIC_VF_INT);
1146 	/* Check for spurious interrupt */
1147 	if (!(intr & NICVF_INTR_MBOX_MASK))
1148 		return;
1149 
1150 	nicvf_handle_mbx_intr(nic);
1151 }
1152 
1153 static int
1154 nicvf_intr_handler(void *arg)
1155 {
1156 	struct nicvf *nic;
1157 	struct cmp_queue *cq;
1158 	int qidx;
1159 
1160 	cq = (struct cmp_queue *)arg;
1161 	nic = cq->nic;
1162 	qidx = cq->idx;
1163 
1164 	/* Disable interrupts */
1165 	nicvf_disable_intr(nic, NICVF_INTR_CQ, qidx);
1166 
1167 	taskqueue_enqueue(cq->cmp_taskq, &cq->cmp_task);
1168 
1169 	/* Clear interrupt */
1170 	nicvf_clear_intr(nic, NICVF_INTR_CQ, qidx);
1171 
1172 	return (FILTER_HANDLED);
1173 }
1174 
1175 static void
1176 nicvf_rbdr_intr_handler(void *arg)
1177 {
1178 	struct nicvf *nic;
1179 	struct queue_set *qs;
1180 	struct rbdr *rbdr;
1181 	int qidx;
1182 
1183 	nic = (struct nicvf *)arg;
1184 
1185 	/* Disable RBDR interrupt and schedule softirq */
1186 	for (qidx = 0; qidx < nic->qs->rbdr_cnt; qidx++) {
1187 		if (!nicvf_is_intr_enabled(nic, NICVF_INTR_RBDR, qidx))
1188 			continue;
1189 		nicvf_disable_intr(nic, NICVF_INTR_RBDR, qidx);
1190 
1191 		qs = nic->qs;
1192 		rbdr = &qs->rbdr[qidx];
1193 		taskqueue_enqueue(rbdr->rbdr_taskq, &rbdr->rbdr_task_nowait);
1194 		/* Clear interrupt */
1195 		nicvf_clear_intr(nic, NICVF_INTR_RBDR, qidx);
1196 	}
1197 }
1198 
1199 static void
1200 nicvf_qs_err_intr_handler(void *arg)
1201 {
1202 	struct nicvf *nic = (struct nicvf *)arg;
1203 	struct queue_set *qs = nic->qs;
1204 
1205 	/* Disable Qset err interrupt and schedule softirq */
1206 	nicvf_disable_intr(nic, NICVF_INTR_QS_ERR, 0);
1207 	taskqueue_enqueue(qs->qs_err_taskq, &qs->qs_err_task);
1208 	nicvf_clear_intr(nic, NICVF_INTR_QS_ERR, 0);
1209 
1210 }
1211 
1212 static int
1213 nicvf_enable_msix(struct nicvf *nic)
1214 {
1215 	struct pci_devinfo *dinfo;
1216 	int rid, count;
1217 	int ret;
1218 
1219 	dinfo = device_get_ivars(nic->dev);
1220 	rid = dinfo->cfg.msix.msix_table_bar;
1221 	nic->msix_table_res =
1222 	    bus_alloc_resource_any(nic->dev, SYS_RES_MEMORY, &rid, RF_ACTIVE);
1223 	if (nic->msix_table_res == NULL) {
1224 		device_printf(nic->dev,
1225 		    "Could not allocate memory for MSI-X table\n");
1226 		return (ENXIO);
1227 	}
1228 
1229 	count = nic->num_vec = NIC_VF_MSIX_VECTORS;
1230 
1231 	ret = pci_alloc_msix(nic->dev, &count);
1232 	if ((ret != 0) || (count != nic->num_vec)) {
1233 		device_printf(nic->dev,
1234 		    "Request for #%d msix vectors failed, error: %d\n",
1235 		    nic->num_vec, ret);
1236 		return (ret);
1237 	}
1238 
1239 	nic->msix_enabled = 1;
1240 	return (0);
1241 }
1242 
1243 static void
1244 nicvf_disable_msix(struct nicvf *nic)
1245 {
1246 
1247 	if (nic->msix_enabled) {
1248 		pci_release_msi(nic->dev);
1249 		nic->msix_enabled = 0;
1250 		nic->num_vec = 0;
1251 	}
1252 }
1253 
1254 static void
1255 nicvf_release_all_interrupts(struct nicvf *nic)
1256 {
1257 	struct resource *res;
1258 	int irq;
1259 	int err;
1260 
1261 	/* Free registered interrupts */
1262 	for (irq = 0; irq < nic->num_vec; irq++) {
1263 		res = nic->msix_entries[irq].irq_res;
1264 		if (res == NULL)
1265 			continue;
1266 		/* Teardown interrupt first */
1267 		if (nic->msix_entries[irq].handle != NULL) {
1268 			err = bus_teardown_intr(nic->dev,
1269 			    nic->msix_entries[irq].irq_res,
1270 			    nic->msix_entries[irq].handle);
1271 			KASSERT(err == 0,
1272 			    ("ERROR: Unable to teardown interrupt %d", irq));
1273 			nic->msix_entries[irq].handle = NULL;
1274 		}
1275 
1276 		bus_release_resource(nic->dev, SYS_RES_IRQ,
1277 			    rman_get_rid(res), nic->msix_entries[irq].irq_res);
1278 		nic->msix_entries[irq].irq_res = NULL;
1279 	}
1280 	/* Disable MSI-X */
1281 	nicvf_disable_msix(nic);
1282 }
1283 
1284 /*
1285  * Initialize MSIX vectors and register MISC interrupt.
1286  * Send READY message to PF to check if its alive
1287  */
1288 static int
1289 nicvf_allocate_misc_interrupt(struct nicvf *nic)
1290 {
1291 	struct resource *res;
1292 	int irq, rid;
1293 	int ret = 0;
1294 
1295 	/* Return if mailbox interrupt is already registered */
1296 	if (nic->msix_enabled)
1297 		return (0);
1298 
1299 	/* Enable MSI-X */
1300 	if (nicvf_enable_msix(nic) != 0)
1301 		return (ENXIO);
1302 
1303 	irq = NICVF_INTR_ID_MISC;
1304 	rid = irq + 1;
1305 	nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1306 	    SYS_RES_IRQ, &rid, (RF_SHAREABLE | RF_ACTIVE));
1307 	if (nic->msix_entries[irq].irq_res == NULL) {
1308 		device_printf(nic->dev,
1309 		    "Could not allocate Mbox interrupt for VF%d\n",
1310 		    device_get_unit(nic->dev));
1311 		return (ENXIO);
1312 	}
1313 
1314 	ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1315 	    (INTR_MPSAFE | INTR_TYPE_MISC), NULL, nicvf_misc_intr_handler, nic,
1316 	    &nic->msix_entries[irq].handle);
1317 	if (ret != 0) {
1318 		res = nic->msix_entries[irq].irq_res;
1319 		bus_release_resource(nic->dev, SYS_RES_IRQ,
1320 			    rman_get_rid(res), res);
1321 		nic->msix_entries[irq].irq_res = NULL;
1322 		return (ret);
1323 	}
1324 
1325 	return (0);
1326 }
1327 
1328 static int
1329 nicvf_enable_misc_interrupt(struct nicvf *nic)
1330 {
1331 
1332 	/* Enable mailbox interrupt */
1333 	nicvf_enable_intr(nic, NICVF_INTR_MBOX, 0);
1334 
1335 	/* Check if VF is able to communicate with PF */
1336 	if (!nicvf_check_pf_ready(nic)) {
1337 		nicvf_disable_intr(nic, NICVF_INTR_MBOX, 0);
1338 		return (ENXIO);
1339 	}
1340 
1341 	return (0);
1342 }
1343 
1344 static void
1345 nicvf_release_net_interrupts(struct nicvf *nic)
1346 {
1347 	struct resource *res;
1348 	int irq;
1349 	int err;
1350 
1351 	for_each_cq_irq(irq) {
1352 		res = nic->msix_entries[irq].irq_res;
1353 		if (res == NULL)
1354 			continue;
1355 		/* Teardown active interrupts first */
1356 		if (nic->msix_entries[irq].handle != NULL) {
1357 			err = bus_teardown_intr(nic->dev,
1358 			    nic->msix_entries[irq].irq_res,
1359 			    nic->msix_entries[irq].handle);
1360 			KASSERT(err == 0,
1361 			    ("ERROR: Unable to teardown CQ interrupt %d",
1362 			    (irq - NICVF_INTR_ID_CQ)));
1363 			if (err != 0)
1364 				continue;
1365 		}
1366 
1367 		/* Release resource */
1368 		bus_release_resource(nic->dev, SYS_RES_IRQ, rman_get_rid(res),
1369 		    res);
1370 		nic->msix_entries[irq].irq_res = NULL;
1371 	}
1372 
1373 	for_each_rbdr_irq(irq) {
1374 		res = nic->msix_entries[irq].irq_res;
1375 		if (res == NULL)
1376 			continue;
1377 		/* Teardown active interrupts first */
1378 		if (nic->msix_entries[irq].handle != NULL) {
1379 			err = bus_teardown_intr(nic->dev,
1380 			    nic->msix_entries[irq].irq_res,
1381 			    nic->msix_entries[irq].handle);
1382 			KASSERT(err == 0,
1383 			    ("ERROR: Unable to teardown RDBR interrupt %d",
1384 			    (irq - NICVF_INTR_ID_RBDR)));
1385 			if (err != 0)
1386 				continue;
1387 		}
1388 
1389 		/* Release resource */
1390 		bus_release_resource(nic->dev, SYS_RES_IRQ, rman_get_rid(res),
1391 		    res);
1392 		nic->msix_entries[irq].irq_res = NULL;
1393 	}
1394 
1395 	irq = NICVF_INTR_ID_QS_ERR;
1396 	res = nic->msix_entries[irq].irq_res;
1397 	if (res != NULL) {
1398 		/* Teardown active interrupts first */
1399 		if (nic->msix_entries[irq].handle != NULL) {
1400 			err = bus_teardown_intr(nic->dev,
1401 			    nic->msix_entries[irq].irq_res,
1402 			    nic->msix_entries[irq].handle);
1403 			KASSERT(err == 0,
1404 			    ("ERROR: Unable to teardown QS Error interrupt %d",
1405 			    irq));
1406 			if (err != 0)
1407 				return;
1408 		}
1409 
1410 		/* Release resource */
1411 		bus_release_resource(nic->dev, SYS_RES_IRQ, rman_get_rid(res),
1412 		    res);
1413 		nic->msix_entries[irq].irq_res = NULL;
1414 	}
1415 }
1416 
1417 static int
1418 nicvf_allocate_net_interrupts(struct nicvf *nic)
1419 {
1420 	u_int cpuid;
1421 	int irq, rid;
1422 	int qidx;
1423 	int ret = 0;
1424 
1425 	/* MSI-X must be configured by now */
1426 	if (!nic->msix_enabled) {
1427 		device_printf(nic->dev, "Cannot alloacte queue interrups. "
1428 		    "MSI-X interrupts disabled.\n");
1429 		return (ENXIO);
1430 	}
1431 
1432 	/* Register CQ interrupts */
1433 	for_each_cq_irq(irq) {
1434 		if (irq >= (NICVF_INTR_ID_CQ + nic->qs->cq_cnt))
1435 			break;
1436 
1437 		qidx = irq - NICVF_INTR_ID_CQ;
1438 		rid = irq + 1;
1439 		nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1440 		    SYS_RES_IRQ, &rid, (RF_SHAREABLE | RF_ACTIVE));
1441 		if (nic->msix_entries[irq].irq_res == NULL) {
1442 			device_printf(nic->dev,
1443 			    "Could not allocate CQ interrupt %d for VF%d\n",
1444 			    (irq - NICVF_INTR_ID_CQ), device_get_unit(nic->dev));
1445 			ret = ENXIO;
1446 			goto error;
1447 		}
1448 		ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1449 		    (INTR_MPSAFE | INTR_TYPE_NET), nicvf_intr_handler,
1450 		    NULL, &nic->qs->cq[qidx], &nic->msix_entries[irq].handle);
1451 		if (ret != 0) {
1452 			device_printf(nic->dev,
1453 			    "Could not setup CQ interrupt %d for VF%d\n",
1454 			    (irq - NICVF_INTR_ID_CQ), device_get_unit(nic->dev));
1455 			goto error;
1456 		}
1457 		cpuid = (device_get_unit(nic->dev) * CMP_QUEUE_CNT) + qidx;
1458 		cpuid %= mp_ncpus;
1459 		/*
1460 		 * Save CPU ID for later use when system-wide RSS is enabled.
1461 		 * It will be used to pit the CQ task to the same CPU that got
1462 		 * interrupted.
1463 		 */
1464 		nic->qs->cq[qidx].cmp_cpuid = cpuid;
1465 		if (bootverbose) {
1466 			device_printf(nic->dev, "bind CQ%d IRQ to CPU%d\n",
1467 			    qidx, cpuid);
1468 		}
1469 		/* Bind interrupts to the given CPU */
1470 		bus_bind_intr(nic->dev, nic->msix_entries[irq].irq_res, cpuid);
1471 	}
1472 
1473 	/* Register RBDR interrupt */
1474 	for_each_rbdr_irq(irq) {
1475 		if (irq >= (NICVF_INTR_ID_RBDR + nic->qs->rbdr_cnt))
1476 			break;
1477 
1478 		rid = irq + 1;
1479 		nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1480 		    SYS_RES_IRQ, &rid, (RF_SHAREABLE | RF_ACTIVE));
1481 		if (nic->msix_entries[irq].irq_res == NULL) {
1482 			device_printf(nic->dev,
1483 			    "Could not allocate RBDR interrupt %d for VF%d\n",
1484 			    (irq - NICVF_INTR_ID_RBDR),
1485 			    device_get_unit(nic->dev));
1486 			ret = ENXIO;
1487 			goto error;
1488 		}
1489 		ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1490 		    (INTR_MPSAFE | INTR_TYPE_NET), NULL,
1491 		    nicvf_rbdr_intr_handler, nic,
1492 		    &nic->msix_entries[irq].handle);
1493 		if (ret != 0) {
1494 			device_printf(nic->dev,
1495 			    "Could not setup RBDR interrupt %d for VF%d\n",
1496 			    (irq - NICVF_INTR_ID_RBDR),
1497 			    device_get_unit(nic->dev));
1498 			goto error;
1499 		}
1500 	}
1501 
1502 	/* Register QS error interrupt */
1503 	irq = NICVF_INTR_ID_QS_ERR;
1504 	rid = irq + 1;
1505 	nic->msix_entries[irq].irq_res = bus_alloc_resource_any(nic->dev,
1506 	    SYS_RES_IRQ, &rid, (RF_SHAREABLE | RF_ACTIVE));
1507 	if (nic->msix_entries[irq].irq_res == NULL) {
1508 		device_printf(nic->dev,
1509 		    "Could not allocate QS Error interrupt for VF%d\n",
1510 		    device_get_unit(nic->dev));
1511 		ret = ENXIO;
1512 		goto error;
1513 	}
1514 	ret = bus_setup_intr(nic->dev, nic->msix_entries[irq].irq_res,
1515 	    (INTR_MPSAFE | INTR_TYPE_NET), NULL, nicvf_qs_err_intr_handler,
1516 	    nic, &nic->msix_entries[irq].handle);
1517 	if (ret != 0) {
1518 		device_printf(nic->dev,
1519 		    "Could not setup QS Error interrupt for VF%d\n",
1520 		    device_get_unit(nic->dev));
1521 		goto error;
1522 	}
1523 
1524 	return (0);
1525 error:
1526 	nicvf_release_net_interrupts(nic);
1527 	return (ret);
1528 }
1529 
1530 static int
1531 nicvf_stop_locked(struct nicvf *nic)
1532 {
1533 	struct ifnet *ifp;
1534 	int qidx;
1535 	struct queue_set *qs = nic->qs;
1536 	union nic_mbx mbx = {};
1537 
1538 	NICVF_CORE_LOCK_ASSERT(nic);
1539 	/* Stop callout. Can block here since holding SX lock */
1540 	callout_drain(&nic->stats_callout);
1541 
1542 	ifp = nic->ifp;
1543 
1544 	mbx.msg.msg = NIC_MBOX_MSG_SHUTDOWN;
1545 	nicvf_send_msg_to_pf(nic, &mbx);
1546 
1547 	/* Disable RBDR & QS error interrupts */
1548 	for (qidx = 0; qidx < qs->rbdr_cnt; qidx++) {
1549 		nicvf_disable_intr(nic, NICVF_INTR_RBDR, qidx);
1550 		nicvf_clear_intr(nic, NICVF_INTR_RBDR, qidx);
1551 	}
1552 	nicvf_disable_intr(nic, NICVF_INTR_QS_ERR, 0);
1553 	nicvf_clear_intr(nic, NICVF_INTR_QS_ERR, 0);
1554 
1555 	/* Deactivate network interface */
1556 	if_setdrvflagbits(ifp, IFF_DRV_OACTIVE, IFF_DRV_RUNNING);
1557 
1558 	/* Free resources */
1559 	nicvf_config_data_transfer(nic, FALSE);
1560 
1561 	/* Disable HW Qset */
1562 	nicvf_qset_config(nic, FALSE);
1563 
1564 	/* disable mailbox interrupt */
1565 	nicvf_disable_intr(nic, NICVF_INTR_MBOX, 0);
1566 
1567 	return (0);
1568 }
1569 
1570 static void
1571 nicvf_update_stats(struct nicvf *nic)
1572 {
1573 	int qidx;
1574 	struct nicvf_hw_stats *stats = &nic->hw_stats;
1575 	struct nicvf_drv_stats *drv_stats = &nic->drv_stats;
1576 	struct queue_set *qs = nic->qs;
1577 
1578 #define	GET_RX_STATS(reg) \
1579     nicvf_reg_read(nic, NIC_VNIC_RX_STAT_0_13 | ((reg) << 3))
1580 #define GET_TX_STATS(reg) \
1581     nicvf_reg_read(nic, NIC_VNIC_TX_STAT_0_4 | ((reg) << 3))
1582 
1583 	stats->rx_bytes = GET_RX_STATS(RX_OCTS);
1584 	stats->rx_ucast_frames = GET_RX_STATS(RX_UCAST);
1585 	stats->rx_bcast_frames = GET_RX_STATS(RX_BCAST);
1586 	stats->rx_mcast_frames = GET_RX_STATS(RX_MCAST);
1587 	stats->rx_fcs_errors = GET_RX_STATS(RX_FCS);
1588 	stats->rx_l2_errors = GET_RX_STATS(RX_L2ERR);
1589 	stats->rx_drop_red = GET_RX_STATS(RX_RED);
1590 	stats->rx_drop_red_bytes = GET_RX_STATS(RX_RED_OCTS);
1591 	stats->rx_drop_overrun = GET_RX_STATS(RX_ORUN);
1592 	stats->rx_drop_overrun_bytes = GET_RX_STATS(RX_ORUN_OCTS);
1593 	stats->rx_drop_bcast = GET_RX_STATS(RX_DRP_BCAST);
1594 	stats->rx_drop_mcast = GET_RX_STATS(RX_DRP_MCAST);
1595 	stats->rx_drop_l3_bcast = GET_RX_STATS(RX_DRP_L3BCAST);
1596 	stats->rx_drop_l3_mcast = GET_RX_STATS(RX_DRP_L3MCAST);
1597 
1598 	stats->tx_bytes_ok = GET_TX_STATS(TX_OCTS);
1599 	stats->tx_ucast_frames_ok = GET_TX_STATS(TX_UCAST);
1600 	stats->tx_bcast_frames_ok = GET_TX_STATS(TX_BCAST);
1601 	stats->tx_mcast_frames_ok = GET_TX_STATS(TX_MCAST);
1602 	stats->tx_drops = GET_TX_STATS(TX_DROP);
1603 
1604 	drv_stats->tx_frames_ok = stats->tx_ucast_frames_ok +
1605 	    stats->tx_bcast_frames_ok + stats->tx_mcast_frames_ok;
1606 	drv_stats->rx_drops = stats->rx_drop_red + stats->rx_drop_overrun;
1607 	drv_stats->tx_drops = stats->tx_drops;
1608 
1609 	/* Update RQ and SQ stats */
1610 	for (qidx = 0; qidx < qs->rq_cnt; qidx++)
1611 		nicvf_update_rq_stats(nic, qidx);
1612 	for (qidx = 0; qidx < qs->sq_cnt; qidx++)
1613 		nicvf_update_sq_stats(nic, qidx);
1614 }
1615 
1616 static void
1617 nicvf_tick_stats(void *arg)
1618 {
1619 	struct nicvf *nic;
1620 
1621 	nic = (struct nicvf *)arg;
1622 
1623 	/* Read the statistics */
1624 	nicvf_update_stats(nic);
1625 
1626 	callout_reset(&nic->stats_callout, hz, nicvf_tick_stats, nic);
1627 }
1628