xref: /freebsd/sys/dev/netmap/netmap_bdg.c (revision 038405f32f71ad8ba0280ae066417f986ede79db)
1 /*
2  * Copyright (C) 2013-2016 Universita` di Pisa
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  *   1. Redistributions of source code must retain the above copyright
9  *      notice, this list of conditions and the following disclaimer.
10  *   2. Redistributions in binary form must reproduce the above copyright
11  *      notice, this list of conditions and the following disclaimer in the
12  *      documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26 
27 
28 /*
29  * This module implements the VALE switch for netmap
30 
31 --- VALE SWITCH ---
32 
33 NMG_LOCK() serializes all modifications to switches and ports.
34 A switch cannot be deleted until all ports are gone.
35 
36 For each switch, an SX lock (RWlock on linux) protects
37 deletion of ports. When configuring or deleting a new port, the
38 lock is acquired in exclusive mode (after holding NMG_LOCK).
39 When forwarding, the lock is acquired in shared mode (without NMG_LOCK).
40 The lock is held throughout the entire forwarding cycle,
41 during which the thread may incur in a page fault.
42 Hence it is important that sleepable shared locks are used.
43 
44 On the rx ring, the per-port lock is grabbed initially to reserve
45 a number of slot in the ring, then the lock is released,
46 packets are copied from source to destination, and then
47 the lock is acquired again and the receive ring is updated.
48 (A similar thing is done on the tx ring for NIC and host stack
49 ports attached to the switch)
50 
51  */
52 
53 /*
54  * OS-specific code that is used only within this file.
55  * Other OS-specific code that must be accessed by drivers
56  * is present in netmap_kern.h
57  */
58 
59 #if defined(__FreeBSD__)
60 #include <sys/cdefs.h> /* prerequisite */
61 __FBSDID("$FreeBSD$");
62 
63 #include <sys/types.h>
64 #include <sys/errno.h>
65 #include <sys/param.h>	/* defines used in kernel.h */
66 #include <sys/kernel.h>	/* types used in module initialization */
67 #include <sys/conf.h>	/* cdevsw struct, UID, GID */
68 #include <sys/sockio.h>
69 #include <sys/socketvar.h>	/* struct socket */
70 #include <sys/malloc.h>
71 #include <sys/poll.h>
72 #include <sys/rwlock.h>
73 #include <sys/socket.h> /* sockaddrs */
74 #include <sys/selinfo.h>
75 #include <sys/sysctl.h>
76 #include <net/if.h>
77 #include <net/if_var.h>
78 #include <net/bpf.h>		/* BIOCIMMEDIATE */
79 #include <machine/bus.h>	/* bus_dmamap_* */
80 #include <sys/endian.h>
81 #include <sys/refcount.h>
82 #include <sys/smp.h>
83 
84 
85 #elif defined(linux)
86 
87 #include "bsd_glue.h"
88 
89 #elif defined(__APPLE__)
90 
91 #warning OSX support is only partial
92 #include "osx_glue.h"
93 
94 #elif defined(_WIN32)
95 #include "win_glue.h"
96 
97 #else
98 
99 #error	Unsupported platform
100 
101 #endif /* unsupported */
102 
103 /*
104  * common headers
105  */
106 
107 #include <net/netmap.h>
108 #include <dev/netmap/netmap_kern.h>
109 #include <dev/netmap/netmap_mem2.h>
110 
111 #include <dev/netmap/netmap_bdg.h>
112 
113 const char*
114 netmap_bdg_name(struct netmap_vp_adapter *vp)
115 {
116 	struct nm_bridge *b = vp->na_bdg;
117 	if (b == NULL)
118 		return NULL;
119 	return b->bdg_basename;
120 }
121 
122 
123 #ifndef CONFIG_NET_NS
124 /*
125  * XXX in principle nm_bridges could be created dynamically
126  * Right now we have a static array and deletions are protected
127  * by an exclusive lock.
128  */
129 struct nm_bridge *nm_bridges;
130 #endif /* !CONFIG_NET_NS */
131 
132 
133 static int
134 nm_is_id_char(const char c)
135 {
136 	return (c >= 'a' && c <= 'z') ||
137 	       (c >= 'A' && c <= 'Z') ||
138 	       (c >= '0' && c <= '9') ||
139 	       (c == '_');
140 }
141 
142 /* Validate the name of a bdg port and return the
143  * position of the ":" character. */
144 static int
145 nm_bdg_name_validate(const char *name, size_t prefixlen)
146 {
147 	int colon_pos = -1;
148 	int i;
149 
150 	if (!name || strlen(name) < prefixlen) {
151 		return -1;
152 	}
153 
154 	for (i = 0; i < NM_BDG_IFNAMSIZ && name[i]; i++) {
155 		if (name[i] == ':') {
156 			colon_pos = i;
157 			break;
158 		} else if (!nm_is_id_char(name[i])) {
159 			return -1;
160 		}
161 	}
162 
163 	if (strlen(name) - colon_pos > IFNAMSIZ) {
164 		/* interface name too long */
165 		return -1;
166 	}
167 
168 	return colon_pos;
169 }
170 
171 /*
172  * locate a bridge among the existing ones.
173  * MUST BE CALLED WITH NMG_LOCK()
174  *
175  * a ':' in the name terminates the bridge name. Otherwise, just NM_NAME.
176  * We assume that this is called with a name of at least NM_NAME chars.
177  */
178 struct nm_bridge *
179 nm_find_bridge(const char *name, int create, struct netmap_bdg_ops *ops)
180 {
181 	int i, namelen;
182 	struct nm_bridge *b = NULL, *bridges;
183 	u_int num_bridges;
184 
185 	NMG_LOCK_ASSERT();
186 
187 	netmap_bns_getbridges(&bridges, &num_bridges);
188 
189 	namelen = nm_bdg_name_validate(name,
190 			(ops != NULL ? strlen(ops->name) : 0));
191 	if (namelen < 0) {
192 		nm_prerr("invalid bridge name %s", name ? name : NULL);
193 		return NULL;
194 	}
195 
196 	/* lookup the name, remember empty slot if there is one */
197 	for (i = 0; i < num_bridges; i++) {
198 		struct nm_bridge *x = bridges + i;
199 
200 		if ((x->bdg_flags & NM_BDG_ACTIVE) + x->bdg_active_ports == 0) {
201 			if (create && b == NULL)
202 				b = x;	/* record empty slot */
203 		} else if (x->bdg_namelen != namelen) {
204 			continue;
205 		} else if (strncmp(name, x->bdg_basename, namelen) == 0) {
206 			nm_prdis("found '%.*s' at %d", namelen, name, i);
207 			b = x;
208 			break;
209 		}
210 	}
211 	if (i == num_bridges && b) { /* name not found, can create entry */
212 		/* initialize the bridge */
213 		nm_prdis("create new bridge %s with ports %d", b->bdg_basename,
214 			b->bdg_active_ports);
215 		b->ht = nm_os_malloc(sizeof(struct nm_hash_ent) * NM_BDG_HASH);
216 		if (b->ht == NULL) {
217 			nm_prerr("failed to allocate hash table");
218 			return NULL;
219 		}
220 		strncpy(b->bdg_basename, name, namelen);
221 		b->bdg_namelen = namelen;
222 		b->bdg_active_ports = 0;
223 		for (i = 0; i < NM_BDG_MAXPORTS; i++)
224 			b->bdg_port_index[i] = i;
225 		/* set the default function */
226 		b->bdg_ops = b->bdg_saved_ops = *ops;
227 		b->private_data = b->ht;
228 		b->bdg_flags = 0;
229 		NM_BNS_GET(b);
230 	}
231 	return b;
232 }
233 
234 
235 int
236 netmap_bdg_free(struct nm_bridge *b)
237 {
238 	if ((b->bdg_flags & NM_BDG_ACTIVE) + b->bdg_active_ports != 0) {
239 		return EBUSY;
240 	}
241 
242 	nm_prdis("marking bridge %s as free", b->bdg_basename);
243 	nm_os_free(b->ht);
244 	memset(&b->bdg_ops, 0, sizeof(b->bdg_ops));
245 	memset(&b->bdg_saved_ops, 0, sizeof(b->bdg_saved_ops));
246 	b->bdg_flags = 0;
247 	NM_BNS_PUT(b);
248 	return 0;
249 }
250 
251 /* Called by external kernel modules (e.g., Openvswitch).
252  * to modify the private data previously given to regops().
253  * 'name' may be just bridge's name (including ':' if it
254  * is not just NM_BDG_NAME).
255  * Called without NMG_LOCK.
256  */
257 int
258 netmap_bdg_update_private_data(const char *name, bdg_update_private_data_fn_t callback,
259 	void *callback_data, void *auth_token)
260 {
261 	void *private_data = NULL;
262 	struct nm_bridge *b;
263 	int error = 0;
264 
265 	NMG_LOCK();
266 	b = nm_find_bridge(name, 0 /* don't create */, NULL);
267 	if (!b) {
268 		error = EINVAL;
269 		goto unlock_update_priv;
270 	}
271 	if (!nm_bdg_valid_auth_token(b, auth_token)) {
272 		error = EACCES;
273 		goto unlock_update_priv;
274 	}
275 	BDG_WLOCK(b);
276 	private_data = callback(b->private_data, callback_data, &error);
277 	b->private_data = private_data;
278 	BDG_WUNLOCK(b);
279 
280 unlock_update_priv:
281 	NMG_UNLOCK();
282 	return error;
283 }
284 
285 
286 
287 /* remove from bridge b the ports in slots hw and sw
288  * (sw can be -1 if not needed)
289  */
290 void
291 netmap_bdg_detach_common(struct nm_bridge *b, int hw, int sw)
292 {
293 	int s_hw = hw, s_sw = sw;
294 	int i, lim =b->bdg_active_ports;
295 	uint32_t *tmp = b->tmp_bdg_port_index;
296 
297 	/*
298 	New algorithm:
299 	make a copy of bdg_port_index;
300 	lookup NA(ifp)->bdg_port and SWNA(ifp)->bdg_port
301 	in the array of bdg_port_index, replacing them with
302 	entries from the bottom of the array;
303 	decrement bdg_active_ports;
304 	acquire BDG_WLOCK() and copy back the array.
305 	 */
306 
307 	if (netmap_debug & NM_DEBUG_BDG)
308 		nm_prinf("detach %d and %d (lim %d)", hw, sw, lim);
309 	/* make a copy of the list of active ports, update it,
310 	 * and then copy back within BDG_WLOCK().
311 	 */
312 	memcpy(b->tmp_bdg_port_index, b->bdg_port_index, sizeof(b->tmp_bdg_port_index));
313 	for (i = 0; (hw >= 0 || sw >= 0) && i < lim; ) {
314 		if (hw >= 0 && tmp[i] == hw) {
315 			nm_prdis("detach hw %d at %d", hw, i);
316 			lim--; /* point to last active port */
317 			tmp[i] = tmp[lim]; /* swap with i */
318 			tmp[lim] = hw;	/* now this is inactive */
319 			hw = -1;
320 		} else if (sw >= 0 && tmp[i] == sw) {
321 			nm_prdis("detach sw %d at %d", sw, i);
322 			lim--;
323 			tmp[i] = tmp[lim];
324 			tmp[lim] = sw;
325 			sw = -1;
326 		} else {
327 			i++;
328 		}
329 	}
330 	if (hw >= 0 || sw >= 0) {
331 		nm_prerr("delete failed hw %d sw %d, should panic...", hw, sw);
332 	}
333 
334 	BDG_WLOCK(b);
335 	if (b->bdg_ops.dtor)
336 		b->bdg_ops.dtor(b->bdg_ports[s_hw]);
337 	b->bdg_ports[s_hw] = NULL;
338 	if (s_sw >= 0) {
339 		b->bdg_ports[s_sw] = NULL;
340 	}
341 	memcpy(b->bdg_port_index, b->tmp_bdg_port_index, sizeof(b->tmp_bdg_port_index));
342 	b->bdg_active_ports = lim;
343 	BDG_WUNLOCK(b);
344 
345 	nm_prdis("now %d active ports", lim);
346 	netmap_bdg_free(b);
347 }
348 
349 
350 /* nm_bdg_ctl callback for VALE ports */
351 int
352 netmap_vp_bdg_ctl(struct nmreq_header *hdr, struct netmap_adapter *na)
353 {
354 	struct netmap_vp_adapter *vpna = (struct netmap_vp_adapter *)na;
355 	struct nm_bridge *b = vpna->na_bdg;
356 
357 	if (hdr->nr_reqtype == NETMAP_REQ_VALE_ATTACH) {
358 		return 0; /* nothing to do */
359 	}
360 	if (b) {
361 		netmap_set_all_rings(na, 0 /* disable */);
362 		netmap_bdg_detach_common(b, vpna->bdg_port, -1);
363 		vpna->na_bdg = NULL;
364 		netmap_set_all_rings(na, 1 /* enable */);
365 	}
366 	/* I have took reference just for attach */
367 	netmap_adapter_put(na);
368 	return 0;
369 }
370 
371 int
372 netmap_default_bdg_attach(const char *name, struct netmap_adapter *na,
373 		struct nm_bridge *b)
374 {
375 	return NM_NEED_BWRAP;
376 }
377 
378 /* Try to get a reference to a netmap adapter attached to a VALE switch.
379  * If the adapter is found (or is created), this function returns 0, a
380  * non NULL pointer is returned into *na, and the caller holds a
381  * reference to the adapter.
382  * If an adapter is not found, then no reference is grabbed and the
383  * function returns an error code, or 0 if there is just a VALE prefix
384  * mismatch. Therefore the caller holds a reference when
385  * (*na != NULL && return == 0).
386  */
387 int
388 netmap_get_bdg_na(struct nmreq_header *hdr, struct netmap_adapter **na,
389 	struct netmap_mem_d *nmd, int create, struct netmap_bdg_ops *ops)
390 {
391 	char *nr_name = hdr->nr_name;
392 	const char *ifname;
393 	struct ifnet *ifp = NULL;
394 	int error = 0;
395 	struct netmap_vp_adapter *vpna, *hostna = NULL;
396 	struct nm_bridge *b;
397 	uint32_t i, j;
398 	uint32_t cand = NM_BDG_NOPORT, cand2 = NM_BDG_NOPORT;
399 	int needed;
400 
401 	*na = NULL;     /* default return value */
402 
403 	/* first try to see if this is a bridge port. */
404 	NMG_LOCK_ASSERT();
405 	if (strncmp(nr_name, ops->name, strlen(ops->name) - 1)) {
406 		return 0;  /* no error, but no VALE prefix */
407 	}
408 
409 	b = nm_find_bridge(nr_name, create, ops);
410 	if (b == NULL) {
411 		nm_prdis("no bridges available for '%s'", nr_name);
412 		return (create ? ENOMEM : ENXIO);
413 	}
414 	if (strlen(nr_name) < b->bdg_namelen) /* impossible */
415 		panic("x");
416 
417 	/* Now we are sure that name starts with the bridge's name,
418 	 * lookup the port in the bridge. We need to scan the entire
419 	 * list. It is not important to hold a WLOCK on the bridge
420 	 * during the search because NMG_LOCK already guarantees
421 	 * that there are no other possible writers.
422 	 */
423 
424 	/* lookup in the local list of ports */
425 	for (j = 0; j < b->bdg_active_ports; j++) {
426 		i = b->bdg_port_index[j];
427 		vpna = b->bdg_ports[i];
428 		nm_prdis("checking %s", vpna->up.name);
429 		if (!strcmp(vpna->up.name, nr_name)) {
430 			netmap_adapter_get(&vpna->up);
431 			nm_prdis("found existing if %s refs %d", nr_name)
432 			*na = &vpna->up;
433 			return 0;
434 		}
435 	}
436 	/* not found, should we create it? */
437 	if (!create)
438 		return ENXIO;
439 	/* yes we should, see if we have space to attach entries */
440 	needed = 2; /* in some cases we only need 1 */
441 	if (b->bdg_active_ports + needed >= NM_BDG_MAXPORTS) {
442 		nm_prerr("bridge full %d, cannot create new port", b->bdg_active_ports);
443 		return ENOMEM;
444 	}
445 	/* record the next two ports available, but do not allocate yet */
446 	cand = b->bdg_port_index[b->bdg_active_ports];
447 	cand2 = b->bdg_port_index[b->bdg_active_ports + 1];
448 	nm_prdis("+++ bridge %s port %s used %d avail %d %d",
449 		b->bdg_basename, ifname, b->bdg_active_ports, cand, cand2);
450 
451 	/*
452 	 * try see if there is a matching NIC with this name
453 	 * (after the bridge's name)
454 	 */
455 	ifname = nr_name + b->bdg_namelen + 1;
456 	ifp = ifunit_ref(ifname);
457 	if (!ifp) {
458 		/* Create an ephemeral virtual port.
459 		 * This block contains all the ephemeral-specific logic.
460 		 */
461 
462 		if (hdr->nr_reqtype != NETMAP_REQ_REGISTER) {
463 			error = EINVAL;
464 			goto out;
465 		}
466 
467 		/* bdg_netmap_attach creates a struct netmap_adapter */
468 		error = b->bdg_ops.vp_create(hdr, NULL, nmd, &vpna);
469 		if (error) {
470 			if (netmap_debug & NM_DEBUG_BDG)
471 				nm_prerr("error %d", error);
472 			goto out;
473 		}
474 		/* shortcut - we can skip get_hw_na(),
475 		 * ownership check and nm_bdg_attach()
476 		 */
477 
478 	} else {
479 		struct netmap_adapter *hw;
480 
481 		/* the vale:nic syntax is only valid for some commands */
482 		switch (hdr->nr_reqtype) {
483 		case NETMAP_REQ_VALE_ATTACH:
484 		case NETMAP_REQ_VALE_DETACH:
485 		case NETMAP_REQ_VALE_POLLING_ENABLE:
486 		case NETMAP_REQ_VALE_POLLING_DISABLE:
487 			break; /* ok */
488 		default:
489 			error = EINVAL;
490 			goto out;
491 		}
492 
493 		error = netmap_get_hw_na(ifp, nmd, &hw);
494 		if (error || hw == NULL)
495 			goto out;
496 
497 		/* host adapter might not be created */
498 		error = hw->nm_bdg_attach(nr_name, hw, b);
499 		if (error == NM_NEED_BWRAP) {
500 			error = b->bdg_ops.bwrap_attach(nr_name, hw);
501 		}
502 		if (error)
503 			goto out;
504 		vpna = hw->na_vp;
505 		hostna = hw->na_hostvp;
506 		if (hdr->nr_reqtype == NETMAP_REQ_VALE_ATTACH) {
507 			/* Check if we need to skip the host rings. */
508 			struct nmreq_vale_attach *areq =
509 				(struct nmreq_vale_attach *)(uintptr_t)hdr->nr_body;
510 			if (areq->reg.nr_mode != NR_REG_NIC_SW) {
511 				hostna = NULL;
512 			}
513 		}
514 	}
515 
516 	BDG_WLOCK(b);
517 	vpna->bdg_port = cand;
518 	nm_prdis("NIC  %p to bridge port %d", vpna, cand);
519 	/* bind the port to the bridge (virtual ports are not active) */
520 	b->bdg_ports[cand] = vpna;
521 	vpna->na_bdg = b;
522 	b->bdg_active_ports++;
523 	if (hostna != NULL) {
524 		/* also bind the host stack to the bridge */
525 		b->bdg_ports[cand2] = hostna;
526 		hostna->bdg_port = cand2;
527 		hostna->na_bdg = b;
528 		b->bdg_active_ports++;
529 		nm_prdis("host %p to bridge port %d", hostna, cand2);
530 	}
531 	nm_prdis("if %s refs %d", ifname, vpna->up.na_refcount);
532 	BDG_WUNLOCK(b);
533 	*na = &vpna->up;
534 	netmap_adapter_get(*na);
535 
536 out:
537 	if (ifp)
538 		if_rele(ifp);
539 
540 	return error;
541 }
542 
543 /* Process NETMAP_REQ_VALE_ATTACH.
544  */
545 int
546 netmap_bdg_attach(struct nmreq_header *hdr, void *auth_token)
547 {
548 	struct nmreq_vale_attach *req =
549 		(struct nmreq_vale_attach *)(uintptr_t)hdr->nr_body;
550 	struct netmap_vp_adapter * vpna;
551 	struct netmap_adapter *na = NULL;
552 	struct netmap_mem_d *nmd = NULL;
553 	struct nm_bridge *b = NULL;
554 	int error;
555 
556 	NMG_LOCK();
557 	/* permission check for modified bridges */
558 	b = nm_find_bridge(hdr->nr_name, 0 /* don't create */, NULL);
559 	if (b && !nm_bdg_valid_auth_token(b, auth_token)) {
560 		error = EACCES;
561 		goto unlock_exit;
562 	}
563 
564 	if (req->reg.nr_mem_id) {
565 		nmd = netmap_mem_find(req->reg.nr_mem_id);
566 		if (nmd == NULL) {
567 			error = EINVAL;
568 			goto unlock_exit;
569 		}
570 	}
571 
572 	/* check for existing one */
573 	error = netmap_get_vale_na(hdr, &na, nmd, 0);
574 	if (na) {
575 		error = EBUSY;
576 		goto unref_exit;
577 	}
578 	error = netmap_get_vale_na(hdr, &na,
579 				nmd, 1 /* create if not exists */);
580 	if (error) { /* no device */
581 		goto unlock_exit;
582 	}
583 
584 	if (na == NULL) { /* VALE prefix missing */
585 		error = EINVAL;
586 		goto unlock_exit;
587 	}
588 
589 	if (NETMAP_OWNED_BY_ANY(na)) {
590 		error = EBUSY;
591 		goto unref_exit;
592 	}
593 
594 	if (na->nm_bdg_ctl) {
595 		/* nop for VALE ports. The bwrap needs to put the hwna
596 		 * in netmap mode (see netmap_bwrap_bdg_ctl)
597 		 */
598 		error = na->nm_bdg_ctl(hdr, na);
599 		if (error)
600 			goto unref_exit;
601 		nm_prdis("registered %s to netmap-mode", na->name);
602 	}
603 	vpna = (struct netmap_vp_adapter *)na;
604 	req->port_index = vpna->bdg_port;
605 
606 	if (nmd)
607 		netmap_mem_put(nmd);
608 
609 	NMG_UNLOCK();
610 	return 0;
611 
612 unref_exit:
613 	netmap_adapter_put(na);
614 unlock_exit:
615 	if (nmd)
616 		netmap_mem_put(nmd);
617 
618 	NMG_UNLOCK();
619 	return error;
620 }
621 
622 
623 int
624 nm_is_bwrap(struct netmap_adapter *na)
625 {
626 	return na->nm_register == netmap_bwrap_reg;
627 }
628 
629 /* Process NETMAP_REQ_VALE_DETACH.
630  */
631 int
632 netmap_bdg_detach(struct nmreq_header *hdr, void *auth_token)
633 {
634 	int error;
635 
636 	NMG_LOCK();
637 	error = netmap_bdg_detach_locked(hdr, auth_token);
638 	NMG_UNLOCK();
639 	return error;
640 }
641 
642 int
643 netmap_bdg_detach_locked(struct nmreq_header *hdr, void *auth_token)
644 {
645 	struct nmreq_vale_detach *nmreq_det = (void *)(uintptr_t)hdr->nr_body;
646 	struct netmap_vp_adapter *vpna;
647 	struct netmap_adapter *na;
648 	struct nm_bridge *b = NULL;
649 	int error;
650 
651 	/* permission check for modified bridges */
652 	b = nm_find_bridge(hdr->nr_name, 0 /* don't create */, NULL);
653 	if (b && !nm_bdg_valid_auth_token(b, auth_token)) {
654 		error = EACCES;
655 		goto error_exit;
656 	}
657 
658 	error = netmap_get_vale_na(hdr, &na, NULL, 0 /* don't create */);
659 	if (error) { /* no device, or another bridge or user owns the device */
660 		goto error_exit;
661 	}
662 
663 	if (na == NULL) { /* VALE prefix missing */
664 		error = EINVAL;
665 		goto error_exit;
666 	} else if (nm_is_bwrap(na) &&
667 		   ((struct netmap_bwrap_adapter *)na)->na_polling_state) {
668 		/* Don't detach a NIC with polling */
669 		error = EBUSY;
670 		goto unref_exit;
671 	}
672 
673 	vpna = (struct netmap_vp_adapter *)na;
674 	if (na->na_vp != vpna) {
675 		/* trying to detach first attach of VALE persistent port attached
676 		 * to 2 bridges
677 		 */
678 		error = EBUSY;
679 		goto unref_exit;
680 	}
681 	nmreq_det->port_index = vpna->bdg_port;
682 
683 	if (na->nm_bdg_ctl) {
684 		/* remove the port from bridge. The bwrap
685 		 * also needs to put the hwna in normal mode
686 		 */
687 		error = na->nm_bdg_ctl(hdr, na);
688 	}
689 
690 unref_exit:
691 	netmap_adapter_put(na);
692 error_exit:
693 	return error;
694 
695 }
696 
697 
698 struct nm_bdg_polling_state;
699 struct
700 nm_bdg_kthread {
701 	struct nm_kctx *nmk;
702 	u_int qfirst;
703 	u_int qlast;
704 	struct nm_bdg_polling_state *bps;
705 };
706 
707 struct nm_bdg_polling_state {
708 	bool configured;
709 	bool stopped;
710 	struct netmap_bwrap_adapter *bna;
711 	uint32_t mode;
712 	u_int qfirst;
713 	u_int qlast;
714 	u_int cpu_from;
715 	u_int ncpus;
716 	struct nm_bdg_kthread *kthreads;
717 };
718 
719 static void
720 netmap_bwrap_polling(void *data)
721 {
722 	struct nm_bdg_kthread *nbk = data;
723 	struct netmap_bwrap_adapter *bna;
724 	u_int qfirst, qlast, i;
725 	struct netmap_kring **kring0, *kring;
726 
727 	if (!nbk)
728 		return;
729 	qfirst = nbk->qfirst;
730 	qlast = nbk->qlast;
731 	bna = nbk->bps->bna;
732 	kring0 = NMR(bna->hwna, NR_RX);
733 
734 	for (i = qfirst; i < qlast; i++) {
735 		kring = kring0[i];
736 		kring->nm_notify(kring, 0);
737 	}
738 }
739 
740 static int
741 nm_bdg_create_kthreads(struct nm_bdg_polling_state *bps)
742 {
743 	struct nm_kctx_cfg kcfg;
744 	int i, j;
745 
746 	bps->kthreads = nm_os_malloc(sizeof(struct nm_bdg_kthread) * bps->ncpus);
747 	if (bps->kthreads == NULL)
748 		return ENOMEM;
749 
750 	bzero(&kcfg, sizeof(kcfg));
751 	kcfg.worker_fn = netmap_bwrap_polling;
752 	for (i = 0; i < bps->ncpus; i++) {
753 		struct nm_bdg_kthread *t = bps->kthreads + i;
754 		int all = (bps->ncpus == 1 &&
755 			bps->mode == NETMAP_POLLING_MODE_SINGLE_CPU);
756 		int affinity = bps->cpu_from + i;
757 
758 		t->bps = bps;
759 		t->qfirst = all ? bps->qfirst /* must be 0 */: affinity;
760 		t->qlast = all ? bps->qlast : t->qfirst + 1;
761 		if (netmap_verbose)
762 			nm_prinf("kthread %d a:%u qf:%u ql:%u", i, affinity, t->qfirst,
763 				t->qlast);
764 
765 		kcfg.type = i;
766 		kcfg.worker_private = t;
767 		t->nmk = nm_os_kctx_create(&kcfg, NULL);
768 		if (t->nmk == NULL) {
769 			goto cleanup;
770 		}
771 		nm_os_kctx_worker_setaff(t->nmk, affinity);
772 	}
773 	return 0;
774 
775 cleanup:
776 	for (j = 0; j < i; j++) {
777 		struct nm_bdg_kthread *t = bps->kthreads + i;
778 		nm_os_kctx_destroy(t->nmk);
779 	}
780 	nm_os_free(bps->kthreads);
781 	return EFAULT;
782 }
783 
784 /* A variant of ptnetmap_start_kthreads() */
785 static int
786 nm_bdg_polling_start_kthreads(struct nm_bdg_polling_state *bps)
787 {
788 	int error, i, j;
789 
790 	if (!bps) {
791 		nm_prerr("polling is not configured");
792 		return EFAULT;
793 	}
794 	bps->stopped = false;
795 
796 	for (i = 0; i < bps->ncpus; i++) {
797 		struct nm_bdg_kthread *t = bps->kthreads + i;
798 		error = nm_os_kctx_worker_start(t->nmk);
799 		if (error) {
800 			nm_prerr("error in nm_kthread_start(): %d", error);
801 			goto cleanup;
802 		}
803 	}
804 	return 0;
805 
806 cleanup:
807 	for (j = 0; j < i; j++) {
808 		struct nm_bdg_kthread *t = bps->kthreads + i;
809 		nm_os_kctx_worker_stop(t->nmk);
810 	}
811 	bps->stopped = true;
812 	return error;
813 }
814 
815 static void
816 nm_bdg_polling_stop_delete_kthreads(struct nm_bdg_polling_state *bps)
817 {
818 	int i;
819 
820 	if (!bps)
821 		return;
822 
823 	for (i = 0; i < bps->ncpus; i++) {
824 		struct nm_bdg_kthread *t = bps->kthreads + i;
825 		nm_os_kctx_worker_stop(t->nmk);
826 		nm_os_kctx_destroy(t->nmk);
827 	}
828 	bps->stopped = true;
829 }
830 
831 static int
832 get_polling_cfg(struct nmreq_vale_polling *req, struct netmap_adapter *na,
833 		struct nm_bdg_polling_state *bps)
834 {
835 	unsigned int avail_cpus, core_from;
836 	unsigned int qfirst, qlast;
837 	uint32_t i = req->nr_first_cpu_id;
838 	uint32_t req_cpus = req->nr_num_polling_cpus;
839 
840 	avail_cpus = nm_os_ncpus();
841 
842 	if (req_cpus == 0) {
843 		nm_prerr("req_cpus must be > 0");
844 		return EINVAL;
845 	} else if (req_cpus >= avail_cpus) {
846 		nm_prerr("Cannot use all the CPUs in the system");
847 		return EINVAL;
848 	}
849 
850 	if (req->nr_mode == NETMAP_POLLING_MODE_MULTI_CPU) {
851 		/* Use a separate core for each ring. If nr_num_polling_cpus>1
852 		 * more consecutive rings are polled.
853 		 * For example, if nr_first_cpu_id=2 and nr_num_polling_cpus=2,
854 		 * ring 2 and 3 are polled by core 2 and 3, respectively. */
855 		if (i + req_cpus > nma_get_nrings(na, NR_RX)) {
856 			nm_prerr("Rings %u-%u not in range (have %d rings)",
857 				i, i + req_cpus, nma_get_nrings(na, NR_RX));
858 			return EINVAL;
859 		}
860 		qfirst = i;
861 		qlast = qfirst + req_cpus;
862 		core_from = qfirst;
863 
864 	} else if (req->nr_mode == NETMAP_POLLING_MODE_SINGLE_CPU) {
865 		/* Poll all the rings using a core specified by nr_first_cpu_id.
866 		 * the number of cores must be 1. */
867 		if (req_cpus != 1) {
868 			nm_prerr("ncpus must be 1 for NETMAP_POLLING_MODE_SINGLE_CPU "
869 				"(was %d)", req_cpus);
870 			return EINVAL;
871 		}
872 		qfirst = 0;
873 		qlast = nma_get_nrings(na, NR_RX);
874 		core_from = i;
875 	} else {
876 		nm_prerr("Invalid polling mode");
877 		return EINVAL;
878 	}
879 
880 	bps->mode = req->nr_mode;
881 	bps->qfirst = qfirst;
882 	bps->qlast = qlast;
883 	bps->cpu_from = core_from;
884 	bps->ncpus = req_cpus;
885 	nm_prinf("%s qfirst %u qlast %u cpu_from %u ncpus %u",
886 		req->nr_mode == NETMAP_POLLING_MODE_MULTI_CPU ?
887 		"MULTI" : "SINGLE",
888 		qfirst, qlast, core_from, req_cpus);
889 	return 0;
890 }
891 
892 static int
893 nm_bdg_ctl_polling_start(struct nmreq_vale_polling *req, struct netmap_adapter *na)
894 {
895 	struct nm_bdg_polling_state *bps;
896 	struct netmap_bwrap_adapter *bna;
897 	int error;
898 
899 	bna = (struct netmap_bwrap_adapter *)na;
900 	if (bna->na_polling_state) {
901 		nm_prerr("ERROR adapter already in polling mode");
902 		return EFAULT;
903 	}
904 
905 	bps = nm_os_malloc(sizeof(*bps));
906 	if (!bps)
907 		return ENOMEM;
908 	bps->configured = false;
909 	bps->stopped = true;
910 
911 	if (get_polling_cfg(req, na, bps)) {
912 		nm_os_free(bps);
913 		return EINVAL;
914 	}
915 
916 	if (nm_bdg_create_kthreads(bps)) {
917 		nm_os_free(bps);
918 		return EFAULT;
919 	}
920 
921 	bps->configured = true;
922 	bna->na_polling_state = bps;
923 	bps->bna = bna;
924 
925 	/* disable interrupts if possible */
926 	nma_intr_enable(bna->hwna, 0);
927 	/* start kthread now */
928 	error = nm_bdg_polling_start_kthreads(bps);
929 	if (error) {
930 		nm_prerr("ERROR nm_bdg_polling_start_kthread()");
931 		nm_os_free(bps->kthreads);
932 		nm_os_free(bps);
933 		bna->na_polling_state = NULL;
934 		nma_intr_enable(bna->hwna, 1);
935 	}
936 	return error;
937 }
938 
939 static int
940 nm_bdg_ctl_polling_stop(struct netmap_adapter *na)
941 {
942 	struct netmap_bwrap_adapter *bna = (struct netmap_bwrap_adapter *)na;
943 	struct nm_bdg_polling_state *bps;
944 
945 	if (!bna->na_polling_state) {
946 		nm_prerr("ERROR adapter is not in polling mode");
947 		return EFAULT;
948 	}
949 	bps = bna->na_polling_state;
950 	nm_bdg_polling_stop_delete_kthreads(bna->na_polling_state);
951 	bps->configured = false;
952 	nm_os_free(bps);
953 	bna->na_polling_state = NULL;
954 	/* re-enable interrupts */
955 	nma_intr_enable(bna->hwna, 1);
956 	return 0;
957 }
958 
959 int
960 nm_bdg_polling(struct nmreq_header *hdr)
961 {
962 	struct nmreq_vale_polling *req =
963 		(struct nmreq_vale_polling *)(uintptr_t)hdr->nr_body;
964 	struct netmap_adapter *na = NULL;
965 	int error = 0;
966 
967 	NMG_LOCK();
968 	error = netmap_get_vale_na(hdr, &na, NULL, /*create=*/0);
969 	if (na && !error) {
970 		if (!nm_is_bwrap(na)) {
971 			error = EOPNOTSUPP;
972 		} else if (hdr->nr_reqtype == NETMAP_BDG_POLLING_ON) {
973 			error = nm_bdg_ctl_polling_start(req, na);
974 			if (!error)
975 				netmap_adapter_get(na);
976 		} else {
977 			error = nm_bdg_ctl_polling_stop(na);
978 			if (!error)
979 				netmap_adapter_put(na);
980 		}
981 		netmap_adapter_put(na);
982 	} else if (!na && !error) {
983 		/* Not VALE port. */
984 		error = EINVAL;
985 	}
986 	NMG_UNLOCK();
987 
988 	return error;
989 }
990 
991 /* Called by external kernel modules (e.g., Openvswitch).
992  * to set configure/lookup/dtor functions of a VALE instance.
993  * Register callbacks to the given bridge. 'name' may be just
994  * bridge's name (including ':' if it is not just NM_BDG_NAME).
995  *
996  * Called without NMG_LOCK.
997  */
998 
999 int
1000 netmap_bdg_regops(const char *name, struct netmap_bdg_ops *bdg_ops, void *private_data, void *auth_token)
1001 {
1002 	struct nm_bridge *b;
1003 	int error = 0;
1004 
1005 	NMG_LOCK();
1006 	b = nm_find_bridge(name, 0 /* don't create */, NULL);
1007 	if (!b) {
1008 		error = ENXIO;
1009 		goto unlock_regops;
1010 	}
1011 	if (!nm_bdg_valid_auth_token(b, auth_token)) {
1012 		error = EACCES;
1013 		goto unlock_regops;
1014 	}
1015 
1016 	BDG_WLOCK(b);
1017 	if (!bdg_ops) {
1018 		/* resetting the bridge */
1019 		bzero(b->ht, sizeof(struct nm_hash_ent) * NM_BDG_HASH);
1020 		b->bdg_ops = b->bdg_saved_ops;
1021 		b->private_data = b->ht;
1022 	} else {
1023 		/* modifying the bridge */
1024 		b->private_data = private_data;
1025 #define nm_bdg_override(m) if (bdg_ops->m) b->bdg_ops.m = bdg_ops->m
1026 		nm_bdg_override(lookup);
1027 		nm_bdg_override(config);
1028 		nm_bdg_override(dtor);
1029 		nm_bdg_override(vp_create);
1030 		nm_bdg_override(bwrap_attach);
1031 #undef nm_bdg_override
1032 
1033 	}
1034 	BDG_WUNLOCK(b);
1035 
1036 unlock_regops:
1037 	NMG_UNLOCK();
1038 	return error;
1039 }
1040 
1041 
1042 int
1043 netmap_bdg_config(struct nm_ifreq *nr)
1044 {
1045 	struct nm_bridge *b;
1046 	int error = EINVAL;
1047 
1048 	NMG_LOCK();
1049 	b = nm_find_bridge(nr->nifr_name, 0, NULL);
1050 	if (!b) {
1051 		NMG_UNLOCK();
1052 		return error;
1053 	}
1054 	NMG_UNLOCK();
1055 	/* Don't call config() with NMG_LOCK() held */
1056 	BDG_RLOCK(b);
1057 	if (b->bdg_ops.config != NULL)
1058 		error = b->bdg_ops.config(nr);
1059 	BDG_RUNLOCK(b);
1060 	return error;
1061 }
1062 
1063 
1064 /* nm_register callback for VALE ports */
1065 int
1066 netmap_vp_reg(struct netmap_adapter *na, int onoff)
1067 {
1068 	struct netmap_vp_adapter *vpna =
1069 		(struct netmap_vp_adapter*)na;
1070 
1071 	/* persistent ports may be put in netmap mode
1072 	 * before being attached to a bridge
1073 	 */
1074 	if (vpna->na_bdg)
1075 		BDG_WLOCK(vpna->na_bdg);
1076 	if (onoff) {
1077 		netmap_krings_mode_commit(na, onoff);
1078 		if (na->active_fds == 0)
1079 			na->na_flags |= NAF_NETMAP_ON;
1080 		 /* XXX on FreeBSD, persistent VALE ports should also
1081 		 * toggle IFCAP_NETMAP in na->ifp (2014-03-16)
1082 		 */
1083 	} else {
1084 		if (na->active_fds == 0)
1085 			na->na_flags &= ~NAF_NETMAP_ON;
1086 		netmap_krings_mode_commit(na, onoff);
1087 	}
1088 	if (vpna->na_bdg)
1089 		BDG_WUNLOCK(vpna->na_bdg);
1090 	return 0;
1091 }
1092 
1093 
1094 /* rxsync code used by VALE ports nm_rxsync callback and also
1095  * internally by the brwap
1096  */
1097 static int
1098 netmap_vp_rxsync_locked(struct netmap_kring *kring, int flags)
1099 {
1100 	struct netmap_adapter *na = kring->na;
1101 	struct netmap_ring *ring = kring->ring;
1102 	u_int nm_i, lim = kring->nkr_num_slots - 1;
1103 	u_int head = kring->rhead;
1104 	int n;
1105 
1106 	if (head > lim) {
1107 		nm_prerr("ouch dangerous reset!!!");
1108 		n = netmap_ring_reinit(kring);
1109 		goto done;
1110 	}
1111 
1112 	/* First part, import newly received packets. */
1113 	/* actually nothing to do here, they are already in the kring */
1114 
1115 	/* Second part, skip past packets that userspace has released. */
1116 	nm_i = kring->nr_hwcur;
1117 	if (nm_i != head) {
1118 		/* consistency check, but nothing really important here */
1119 		for (n = 0; likely(nm_i != head); n++) {
1120 			struct netmap_slot *slot = &ring->slot[nm_i];
1121 			void *addr = NMB(na, slot);
1122 
1123 			if (addr == NETMAP_BUF_BASE(kring->na)) { /* bad buf */
1124 				nm_prerr("bad buffer index %d, ignore ?",
1125 					slot->buf_idx);
1126 			}
1127 			slot->flags &= ~NS_BUF_CHANGED;
1128 			nm_i = nm_next(nm_i, lim);
1129 		}
1130 		kring->nr_hwcur = head;
1131 	}
1132 
1133 	n = 0;
1134 done:
1135 	return n;
1136 }
1137 
1138 /*
1139  * nm_rxsync callback for VALE ports
1140  * user process reading from a VALE switch.
1141  * Already protected against concurrent calls from userspace,
1142  * but we must acquire the queue's lock to protect against
1143  * writers on the same queue.
1144  */
1145 int
1146 netmap_vp_rxsync(struct netmap_kring *kring, int flags)
1147 {
1148 	int n;
1149 
1150 	mtx_lock(&kring->q_lock);
1151 	n = netmap_vp_rxsync_locked(kring, flags);
1152 	mtx_unlock(&kring->q_lock);
1153 	return n;
1154 }
1155 
1156 int
1157 netmap_bwrap_attach(const char *nr_name, struct netmap_adapter *hwna,
1158 		struct netmap_bdg_ops *ops)
1159 {
1160 	return ops->bwrap_attach(nr_name, hwna);
1161 }
1162 
1163 
1164 /* Bridge wrapper code (bwrap).
1165  * This is used to connect a non-VALE-port netmap_adapter (hwna) to a
1166  * VALE switch.
1167  * The main task is to swap the meaning of tx and rx rings to match the
1168  * expectations of the VALE switch code (see nm_bdg_flush).
1169  *
1170  * The bwrap works by interposing a netmap_bwrap_adapter between the
1171  * rest of the system and the hwna. The netmap_bwrap_adapter looks like
1172  * a netmap_vp_adapter to the rest the system, but, internally, it
1173  * translates all callbacks to what the hwna expects.
1174  *
1175  * Note that we have to intercept callbacks coming from two sides:
1176  *
1177  *  - callbacks coming from the netmap module are intercepted by
1178  *    passing around the netmap_bwrap_adapter instead of the hwna
1179  *
1180  *  - callbacks coming from outside of the netmap module only know
1181  *    about the hwna. This, however, only happens in interrupt
1182  *    handlers, where only the hwna->nm_notify callback is called.
1183  *    What the bwrap does is to overwrite the hwna->nm_notify callback
1184  *    with its own netmap_bwrap_intr_notify.
1185  *    XXX This assumes that the hwna->nm_notify callback was the
1186  *    standard netmap_notify(), as it is the case for nic adapters.
1187  *    Any additional action performed by hwna->nm_notify will not be
1188  *    performed by netmap_bwrap_intr_notify.
1189  *
1190  * Additionally, the bwrap can optionally attach the host rings pair
1191  * of the wrapped adapter to a different port of the switch.
1192  */
1193 
1194 
1195 static void
1196 netmap_bwrap_dtor(struct netmap_adapter *na)
1197 {
1198 	struct netmap_bwrap_adapter *bna = (struct netmap_bwrap_adapter*)na;
1199 	struct netmap_adapter *hwna = bna->hwna;
1200 	struct nm_bridge *b = bna->up.na_bdg,
1201 		*bh = bna->host.na_bdg;
1202 
1203 	if (bna->host.up.nm_mem)
1204 		netmap_mem_put(bna->host.up.nm_mem);
1205 
1206 	if (b) {
1207 		netmap_bdg_detach_common(b, bna->up.bdg_port,
1208 			    (bh ? bna->host.bdg_port : -1));
1209 	}
1210 
1211 	nm_prdis("na %p", na);
1212 	na->ifp = NULL;
1213 	bna->host.up.ifp = NULL;
1214 	hwna->na_vp = bna->saved_na_vp;
1215 	hwna->na_hostvp = NULL;
1216 	hwna->na_private = NULL;
1217 	hwna->na_flags &= ~NAF_BUSY;
1218 	netmap_adapter_put(hwna);
1219 
1220 }
1221 
1222 
1223 /*
1224  * Intr callback for NICs connected to a bridge.
1225  * Simply ignore tx interrupts (maybe we could try to recover space ?)
1226  * and pass received packets from nic to the bridge.
1227  *
1228  * XXX TODO check locking: this is called from the interrupt
1229  * handler so we should make sure that the interface is not
1230  * disconnected while passing down an interrupt.
1231  *
1232  * Note, no user process can access this NIC or the host stack.
1233  * The only part of the ring that is significant are the slots,
1234  * and head/cur/tail are set from the kring as needed
1235  * (part as a receive ring, part as a transmit ring).
1236  *
1237  * callback that overwrites the hwna notify callback.
1238  * Packets come from the outside or from the host stack and are put on an
1239  * hwna rx ring.
1240  * The bridge wrapper then sends the packets through the bridge.
1241  */
1242 int
1243 netmap_bwrap_intr_notify(struct netmap_kring *kring, int flags)
1244 {
1245 	struct netmap_adapter *na = kring->na;
1246 	struct netmap_bwrap_adapter *bna = na->na_private;
1247 	struct netmap_kring *bkring;
1248 	struct netmap_vp_adapter *vpna = &bna->up;
1249 	u_int ring_nr = kring->ring_id;
1250 	int ret = NM_IRQ_COMPLETED;
1251 	int error;
1252 
1253 	if (netmap_debug & NM_DEBUG_RXINTR)
1254 	    nm_prinf("%s %s 0x%x", na->name, kring->name, flags);
1255 
1256 	bkring = vpna->up.tx_rings[ring_nr];
1257 
1258 	/* make sure the ring is not disabled */
1259 	if (nm_kr_tryget(kring, 0 /* can't sleep */, NULL)) {
1260 		return EIO;
1261 	}
1262 
1263 	if (netmap_debug & NM_DEBUG_RXINTR)
1264 	    nm_prinf("%s head %d cur %d tail %d",  na->name,
1265 		kring->rhead, kring->rcur, kring->rtail);
1266 
1267 	/* simulate a user wakeup on the rx ring
1268 	 * fetch packets that have arrived.
1269 	 */
1270 	error = kring->nm_sync(kring, 0);
1271 	if (error)
1272 		goto put_out;
1273 	if (kring->nr_hwcur == kring->nr_hwtail) {
1274 		if (netmap_verbose)
1275 			nm_prlim(1, "interrupt with no packets on %s",
1276 				kring->name);
1277 		goto put_out;
1278 	}
1279 
1280 	/* new packets are kring->rcur to kring->nr_hwtail, and the bkring
1281 	 * had hwcur == bkring->rhead. So advance bkring->rhead to kring->nr_hwtail
1282 	 * to push all packets out.
1283 	 */
1284 	bkring->rhead = bkring->rcur = kring->nr_hwtail;
1285 
1286 	bkring->nm_sync(bkring, flags);
1287 
1288 	/* mark all buffers as released on this ring */
1289 	kring->rhead = kring->rcur = kring->rtail = kring->nr_hwtail;
1290 	/* another call to actually release the buffers */
1291 	error = kring->nm_sync(kring, 0);
1292 
1293 	/* The second rxsync may have further advanced hwtail. If this happens,
1294 	 *  return NM_IRQ_RESCHED, otherwise just return NM_IRQ_COMPLETED. */
1295 	if (kring->rcur != kring->nr_hwtail) {
1296 		ret = NM_IRQ_RESCHED;
1297 	}
1298 put_out:
1299 	nm_kr_put(kring);
1300 
1301 	return error ? error : ret;
1302 }
1303 
1304 
1305 /* nm_register callback for bwrap */
1306 int
1307 netmap_bwrap_reg(struct netmap_adapter *na, int onoff)
1308 {
1309 	struct netmap_bwrap_adapter *bna =
1310 		(struct netmap_bwrap_adapter *)na;
1311 	struct netmap_adapter *hwna = bna->hwna;
1312 	struct netmap_vp_adapter *hostna = &bna->host;
1313 	int error, i;
1314 	enum txrx t;
1315 
1316 	nm_prdis("%s %s", na->name, onoff ? "on" : "off");
1317 
1318 	if (onoff) {
1319 		/* netmap_do_regif has been called on the bwrap na.
1320 		 * We need to pass the information about the
1321 		 * memory allocator down to the hwna before
1322 		 * putting it in netmap mode
1323 		 */
1324 		hwna->na_lut = na->na_lut;
1325 
1326 		if (hostna->na_bdg) {
1327 			/* if the host rings have been attached to switch,
1328 			 * we need to copy the memory allocator information
1329 			 * in the hostna also
1330 			 */
1331 			hostna->up.na_lut = na->na_lut;
1332 		}
1333 
1334 	}
1335 
1336 	/* pass down the pending ring state information */
1337 	for_rx_tx(t) {
1338 		for (i = 0; i < netmap_all_rings(na, t); i++) {
1339 			NMR(hwna, nm_txrx_swap(t))[i]->nr_pending_mode =
1340 				NMR(na, t)[i]->nr_pending_mode;
1341 		}
1342 	}
1343 
1344 	/* forward the request to the hwna */
1345 	error = hwna->nm_register(hwna, onoff);
1346 	if (error)
1347 		return error;
1348 
1349 	/* copy up the current ring state information */
1350 	for_rx_tx(t) {
1351 		for (i = 0; i < netmap_all_rings(na, t); i++) {
1352 			struct netmap_kring *kring = NMR(hwna, nm_txrx_swap(t))[i];
1353 			NMR(na, t)[i]->nr_mode = kring->nr_mode;
1354 		}
1355 	}
1356 
1357 	/* impersonate a netmap_vp_adapter */
1358 	netmap_vp_reg(na, onoff);
1359 	if (hostna->na_bdg)
1360 		netmap_vp_reg(&hostna->up, onoff);
1361 
1362 	if (onoff) {
1363 		u_int i;
1364 		/* intercept the hwna nm_nofify callback on the hw rings */
1365 		for (i = 0; i < hwna->num_rx_rings; i++) {
1366 			hwna->rx_rings[i]->save_notify = hwna->rx_rings[i]->nm_notify;
1367 			hwna->rx_rings[i]->nm_notify = bna->nm_intr_notify;
1368 		}
1369 		i = hwna->num_rx_rings; /* for safety */
1370 		/* save the host ring notify unconditionally */
1371 		for (; i < netmap_real_rings(hwna, NR_RX); i++) {
1372 			hwna->rx_rings[i]->save_notify =
1373 				hwna->rx_rings[i]->nm_notify;
1374 			if (hostna->na_bdg) {
1375 				/* also intercept the host ring notify */
1376 				hwna->rx_rings[i]->nm_notify =
1377 					netmap_bwrap_intr_notify;
1378 				na->tx_rings[i]->nm_sync = na->nm_txsync;
1379 			}
1380 		}
1381 		if (na->active_fds == 0)
1382 			na->na_flags |= NAF_NETMAP_ON;
1383 	} else {
1384 		u_int i;
1385 
1386 		if (na->active_fds == 0)
1387 			na->na_flags &= ~NAF_NETMAP_ON;
1388 
1389 		/* reset all notify callbacks (including host ring) */
1390 		for (i = 0; i < netmap_all_rings(hwna, NR_RX); i++) {
1391 			hwna->rx_rings[i]->nm_notify =
1392 				hwna->rx_rings[i]->save_notify;
1393 			hwna->rx_rings[i]->save_notify = NULL;
1394 		}
1395 		hwna->na_lut.lut = NULL;
1396 		hwna->na_lut.plut = NULL;
1397 		hwna->na_lut.objtotal = 0;
1398 		hwna->na_lut.objsize = 0;
1399 
1400 		/* reset the number of host rings to default */
1401 		for_rx_tx(t) {
1402 			nma_set_host_nrings(hwna, t, 1);
1403 		}
1404 
1405 	}
1406 
1407 	return 0;
1408 }
1409 
1410 /* nm_config callback for bwrap */
1411 static int
1412 netmap_bwrap_config(struct netmap_adapter *na, struct nm_config_info *info)
1413 {
1414 	struct netmap_bwrap_adapter *bna =
1415 		(struct netmap_bwrap_adapter *)na;
1416 	struct netmap_adapter *hwna = bna->hwna;
1417 	int error;
1418 
1419 	/* cache the lut in the embedded host adapter */
1420 	error = netmap_mem_get_lut(hwna->nm_mem, &bna->host.up.na_lut);
1421 	if (error)
1422 		return error;
1423 
1424 	/* Forward the request to the hwna. It may happen that nobody
1425 	 * registered hwna yet, so netmap_mem_get_lut() may have not
1426 	 * been called yet. */
1427 	error = netmap_mem_get_lut(hwna->nm_mem, &hwna->na_lut);
1428 	if (error)
1429 		return error;
1430 	netmap_update_config(hwna);
1431 	/* swap the results and propagate */
1432 	info->num_tx_rings = hwna->num_rx_rings;
1433 	info->num_tx_descs = hwna->num_rx_desc;
1434 	info->num_rx_rings = hwna->num_tx_rings;
1435 	info->num_rx_descs = hwna->num_tx_desc;
1436 	info->rx_buf_maxsize = hwna->rx_buf_maxsize;
1437 
1438 	if (na->na_flags & NAF_HOST_RINGS) {
1439 		struct netmap_adapter *hostna = &bna->host.up;
1440 		enum txrx t;
1441 
1442 		/* limit the number of host rings to that of hw */
1443 		if (na->na_flags & NAF_HOST_ALL) {
1444 			hostna->num_tx_rings = nma_get_nrings(hwna, NR_RX);
1445 			hostna->num_rx_rings = nma_get_nrings(hwna, NR_TX);
1446 		} else {
1447 			nm_bound_var(&hostna->num_tx_rings, 1, 1,
1448 				nma_get_nrings(hwna, NR_TX), NULL);
1449 			nm_bound_var(&hostna->num_rx_rings, 1, 1,
1450 				nma_get_nrings(hwna, NR_RX), NULL);
1451 		}
1452 		for_rx_tx(t) {
1453 			enum txrx r = nm_txrx_swap(t);
1454 			u_int nr = nma_get_nrings(hostna, t);
1455 
1456 			nma_set_host_nrings(na, t, nr);
1457 			if (nma_get_host_nrings(hwna, t) < nr) {
1458 				nma_set_host_nrings(hwna, t, nr);
1459 			}
1460 			nma_set_ndesc(hostna, t, nma_get_ndesc(hwna, r));
1461 		}
1462 	}
1463 
1464 	return 0;
1465 }
1466 
1467 /* nm_bufcfg callback for bwrap */
1468 static int
1469 netmap_bwrap_bufcfg(struct netmap_kring *kring, uint64_t target)
1470 {
1471 	struct netmap_adapter *na = kring->na;
1472 	struct netmap_bwrap_adapter *bna =
1473 		(struct netmap_bwrap_adapter *)na;
1474 	struct netmap_adapter *hwna = bna->hwna;
1475 	struct netmap_kring *hwkring;
1476 	enum txrx r;
1477 	int error;
1478 
1479 	/* we need the hw kring that corresponds to the bwrap one:
1480 	 * remember that rx and tx are swapped
1481 	 */
1482 	r = nm_txrx_swap(kring->tx);
1483 	hwkring = NMR(hwna, r)[kring->ring_id];
1484 
1485 	/* copy down the offset information, forward the request
1486 	 * and copy up the results
1487 	 */
1488 	hwkring->offset_mask = kring->offset_mask;
1489 	hwkring->offset_max  = kring->offset_max;
1490 	hwkring->offset_gap  = kring->offset_gap;
1491 
1492 	error = hwkring->nm_bufcfg(hwkring, target);
1493 	if (error)
1494 		return error;
1495 
1496 	kring->hwbuf_len = hwkring->hwbuf_len;
1497 	kring->buf_align = hwkring->buf_align;
1498 
1499 	return 0;
1500 }
1501 
1502 /* nm_krings_create callback for bwrap */
1503 int
1504 netmap_bwrap_krings_create_common(struct netmap_adapter *na)
1505 {
1506 	struct netmap_bwrap_adapter *bna =
1507 		(struct netmap_bwrap_adapter *)na;
1508 	struct netmap_adapter *hwna = bna->hwna;
1509 	struct netmap_adapter *hostna = &bna->host.up;
1510 	int i, error = 0;
1511 	enum txrx t;
1512 
1513 	/* also create the hwna krings */
1514 	error = hwna->nm_krings_create(hwna);
1515 	if (error) {
1516 		return error;
1517 	}
1518 
1519 	/* increment the usage counter for all the hwna krings */
1520 	for_rx_tx(t) {
1521 		for (i = 0; i < netmap_all_rings(hwna, t); i++) {
1522 			NMR(hwna, t)[i]->users++;
1523 			/* this to prevent deletion of the rings through
1524 			 * our krings, instead of through the hwna ones */
1525 			NMR(na, t)[i]->nr_kflags |= NKR_NEEDRING;
1526 		}
1527 	}
1528 
1529 	/* now create the actual rings */
1530 	error = netmap_mem_rings_create(hwna);
1531 	if (error) {
1532 		goto err_dec_users;
1533 	}
1534 
1535 	/* cross-link the netmap rings
1536 	 * The original number of rings comes from hwna,
1537 	 * rx rings on one side equals tx rings on the other.
1538 	 */
1539 	for_rx_tx(t) {
1540 		enum txrx r = nm_txrx_swap(t); /* swap NR_TX <-> NR_RX */
1541 		for (i = 0; i < netmap_all_rings(hwna, r); i++) {
1542 			NMR(na, t)[i]->nkr_num_slots = NMR(hwna, r)[i]->nkr_num_slots;
1543 			NMR(na, t)[i]->ring = NMR(hwna, r)[i]->ring;
1544 		}
1545 	}
1546 
1547 	if (na->na_flags & NAF_HOST_RINGS) {
1548 		/* the hostna rings are the host rings of the bwrap.
1549 		 * The corresponding krings must point back to the
1550 		 * hostna
1551 		 */
1552 		hostna->tx_rings = &na->tx_rings[na->num_tx_rings];
1553 		hostna->rx_rings = &na->rx_rings[na->num_rx_rings];
1554 		for_rx_tx(t) {
1555 			for (i = 0; i < nma_get_nrings(hostna, t); i++) {
1556 				NMR(hostna, t)[i]->na = hostna;
1557 			}
1558 		}
1559 	}
1560 
1561 	return 0;
1562 
1563 err_dec_users:
1564 	for_rx_tx(t) {
1565 		for (i = 0; i < netmap_all_rings(hwna, t); i++) {
1566 			NMR(hwna, t)[i]->users--;
1567 			NMR(na, t)[i]->users--;
1568 		}
1569 	}
1570 	hwna->nm_krings_delete(hwna);
1571 	return error;
1572 }
1573 
1574 
1575 void
1576 netmap_bwrap_krings_delete_common(struct netmap_adapter *na)
1577 {
1578 	struct netmap_bwrap_adapter *bna =
1579 		(struct netmap_bwrap_adapter *)na;
1580 	struct netmap_adapter *hwna = bna->hwna;
1581 	enum txrx t;
1582 	int i;
1583 
1584 	nm_prdis("%s", na->name);
1585 
1586 	/* decrement the usage counter for all the hwna krings */
1587 	for_rx_tx(t) {
1588 		for (i = 0; i < netmap_all_rings(hwna, t); i++) {
1589 			NMR(hwna, t)[i]->users--;
1590 			NMR(na, t)[i]->users--;
1591 		}
1592 	}
1593 
1594 	/* delete any netmap rings that are no longer needed */
1595 	netmap_mem_rings_delete(hwna);
1596 	hwna->nm_krings_delete(hwna);
1597 }
1598 
1599 
1600 /* notify method for the bridge-->hwna direction */
1601 int
1602 netmap_bwrap_notify(struct netmap_kring *kring, int flags)
1603 {
1604 	struct netmap_adapter *na = kring->na;
1605 	struct netmap_bwrap_adapter *bna = na->na_private;
1606 	struct netmap_adapter *hwna = bna->hwna;
1607 	u_int ring_n = kring->ring_id;
1608 	u_int lim = kring->nkr_num_slots - 1;
1609 	struct netmap_kring *hw_kring;
1610 	int error;
1611 
1612 	nm_prdis("%s: na %s hwna %s",
1613 			(kring ? kring->name : "NULL!"),
1614 			(na ? na->name : "NULL!"),
1615 			(hwna ? hwna->name : "NULL!"));
1616 	hw_kring = hwna->tx_rings[ring_n];
1617 
1618 	if (nm_kr_tryget(hw_kring, 0, NULL)) {
1619 		return ENXIO;
1620 	}
1621 
1622 	/* first step: simulate a user wakeup on the rx ring */
1623 	netmap_vp_rxsync(kring, flags);
1624 	nm_prdis("%s[%d] PRE rx(c%3d t%3d l%3d) ring(h%3d c%3d t%3d) tx(c%3d ht%3d t%3d)",
1625 		na->name, ring_n,
1626 		kring->nr_hwcur, kring->nr_hwtail, kring->nkr_hwlease,
1627 		kring->rhead, kring->rcur, kring->rtail,
1628 		hw_kring->nr_hwcur, hw_kring->nr_hwtail, hw_kring->rtail);
1629 	/* second step: the new packets are sent on the tx ring
1630 	 * (which is actually the same ring)
1631 	 */
1632 	hw_kring->rhead = hw_kring->rcur = kring->nr_hwtail;
1633 	error = hw_kring->nm_sync(hw_kring, flags);
1634 	if (error)
1635 		goto put_out;
1636 
1637 	/* third step: now we are back the rx ring */
1638 	/* claim ownership on all hw owned bufs */
1639 	kring->rhead = kring->rcur = nm_next(hw_kring->nr_hwtail, lim); /* skip past reserved slot */
1640 
1641 	/* fourth step: the user goes to sleep again, causing another rxsync */
1642 	netmap_vp_rxsync(kring, flags);
1643 	nm_prdis("%s[%d] PST rx(c%3d t%3d l%3d) ring(h%3d c%3d t%3d) tx(c%3d ht%3d t%3d)",
1644 		na->name, ring_n,
1645 		kring->nr_hwcur, kring->nr_hwtail, kring->nkr_hwlease,
1646 		kring->rhead, kring->rcur, kring->rtail,
1647 		hw_kring->nr_hwcur, hw_kring->nr_hwtail, hw_kring->rtail);
1648 put_out:
1649 	nm_kr_put(hw_kring);
1650 
1651 	return error ? error : NM_IRQ_COMPLETED;
1652 }
1653 
1654 
1655 /* nm_bdg_ctl callback for the bwrap.
1656  * Called on bridge-attach and detach, as an effect of valectl -[ahd].
1657  * On attach, it needs to provide a fake netmap_priv_d structure and
1658  * perform a netmap_do_regif() on the bwrap. This will put both the
1659  * bwrap and the hwna in netmap mode, with the netmap rings shared
1660  * and cross linked. Moroever, it will start intercepting interrupts
1661  * directed to hwna.
1662  */
1663 static int
1664 netmap_bwrap_bdg_ctl(struct nmreq_header *hdr, struct netmap_adapter *na)
1665 {
1666 	struct netmap_priv_d *npriv;
1667 	struct netmap_bwrap_adapter *bna = (struct netmap_bwrap_adapter*)na;
1668 	int error = 0;
1669 
1670 	if (hdr->nr_reqtype == NETMAP_REQ_VALE_ATTACH) {
1671 		struct nmreq_vale_attach *req =
1672 			(struct nmreq_vale_attach *)(uintptr_t)hdr->nr_body;
1673 		if (req->reg.nr_ringid != 0 ||
1674 			(req->reg.nr_mode != NR_REG_ALL_NIC &&
1675 				req->reg.nr_mode != NR_REG_NIC_SW)) {
1676 			/* We only support attaching all the NIC rings
1677 			 * and/or the host stack. */
1678 			return EINVAL;
1679 		}
1680 		if (NETMAP_OWNED_BY_ANY(na)) {
1681 			return EBUSY;
1682 		}
1683 		if (bna->na_kpriv) {
1684 			/* nothing to do */
1685 			return 0;
1686 		}
1687 		npriv = netmap_priv_new();
1688 		if (npriv == NULL)
1689 			return ENOMEM;
1690 		npriv->np_ifp = na->ifp; /* let the priv destructor release the ref */
1691 		error = netmap_do_regif(npriv, na, hdr);
1692 		if (error) {
1693 			netmap_priv_delete(npriv);
1694 			netmap_mem_restore(bna->hwna);
1695 			return error;
1696 		}
1697 		bna->na_kpriv = npriv;
1698 		na->na_flags |= NAF_BUSY;
1699 	} else {
1700 		if (na->active_fds == 0) /* not registered */
1701 			return EINVAL;
1702 		netmap_priv_delete(bna->na_kpriv);
1703 		bna->na_kpriv = NULL;
1704 		na->na_flags &= ~NAF_BUSY;
1705 		netmap_mem_restore(bna->hwna);
1706 	}
1707 
1708 	return error;
1709 }
1710 
1711 /* attach a bridge wrapper to the 'real' device */
1712 int
1713 netmap_bwrap_attach_common(struct netmap_adapter *na,
1714 		struct netmap_adapter *hwna)
1715 {
1716 	struct netmap_bwrap_adapter *bna;
1717 	struct netmap_adapter *hostna = NULL;
1718 	int error = 0;
1719 	enum txrx t;
1720 
1721 	/* make sure the NIC is not already in use */
1722 	if (NETMAP_OWNED_BY_ANY(hwna)) {
1723 		nm_prerr("NIC %s busy, cannot attach to bridge", hwna->name);
1724 		return EBUSY;
1725 	}
1726 
1727 	bna = (struct netmap_bwrap_adapter *)na;
1728 	/* make bwrap ifp point to the real ifp */
1729 	na->ifp = hwna->ifp;
1730 	if_ref(na->ifp);
1731 	na->na_private = bna;
1732 	/* fill the ring data for the bwrap adapter with rx/tx meanings
1733 	 * swapped. The real cross-linking will be done during register,
1734 	 * when all the krings will have been created.
1735 	 */
1736 	for_rx_tx(t) {
1737 		enum txrx r = nm_txrx_swap(t); /* swap NR_TX <-> NR_RX */
1738 		nma_set_nrings(na, t, nma_get_nrings(hwna, r));
1739 		nma_set_ndesc(na, t, nma_get_ndesc(hwna, r));
1740 	}
1741 	na->nm_dtor = netmap_bwrap_dtor;
1742 	na->nm_config = netmap_bwrap_config;
1743 	na->nm_bufcfg = netmap_bwrap_bufcfg;
1744 	na->nm_bdg_ctl = netmap_bwrap_bdg_ctl;
1745 	na->pdev = hwna->pdev;
1746 	na->nm_mem = netmap_mem_get(hwna->nm_mem);
1747 	na->virt_hdr_len = hwna->virt_hdr_len;
1748 	na->rx_buf_maxsize = hwna->rx_buf_maxsize;
1749 
1750 	bna->hwna = hwna;
1751 	netmap_adapter_get(hwna);
1752 	hwna->na_private = bna; /* weak reference */
1753 	bna->saved_na_vp = hwna->na_vp;
1754 	hwna->na_vp = &bna->up;
1755 	bna->up.up.na_vp = &(bna->up);
1756 
1757 	if (hwna->na_flags & NAF_HOST_RINGS) {
1758 		if (hwna->na_flags & NAF_SW_ONLY)
1759 			na->na_flags |= NAF_SW_ONLY;
1760 		na->na_flags |= NAF_HOST_RINGS;
1761 		hostna = &bna->host.up;
1762 
1763 		snprintf(hostna->name, sizeof(hostna->name), "%s^", na->name);
1764 		hostna->ifp = hwna->ifp;
1765 		// hostna->nm_txsync = netmap_bwrap_host_txsync;
1766 		// hostna->nm_rxsync = netmap_bwrap_host_rxsync;
1767 		hostna->nm_mem = netmap_mem_get(na->nm_mem);
1768 		hostna->na_private = bna;
1769 		hostna->na_vp = &bna->up;
1770 		na->na_hostvp = hwna->na_hostvp =
1771 			hostna->na_hostvp = &bna->host;
1772 		hostna->na_flags = NAF_BUSY; /* prevent NIOCREGIF */
1773 		hostna->rx_buf_maxsize = hwna->rx_buf_maxsize;
1774 		/* bwrap_config() will determine the number of host rings */
1775 	}
1776 	if (hwna->na_flags & NAF_MOREFRAG)
1777 		na->na_flags |= NAF_MOREFRAG;
1778 
1779 	nm_prdis("%s<->%s txr %d txd %d rxr %d rxd %d",
1780 		na->name, ifp->if_xname,
1781 		na->num_tx_rings, na->num_tx_desc,
1782 		na->num_rx_rings, na->num_rx_desc);
1783 
1784 	error = netmap_attach_common(na);
1785 	if (error) {
1786 		goto err_put;
1787 	}
1788 	hwna->na_flags |= NAF_BUSY;
1789 	return 0;
1790 
1791 err_put:
1792 	hwna->na_vp = hwna->na_hostvp = NULL;
1793 	netmap_adapter_put(hwna);
1794 	return error;
1795 
1796 }
1797 
1798 struct nm_bridge *
1799 netmap_init_bridges2(u_int n)
1800 {
1801 	int i;
1802 	struct nm_bridge *b;
1803 
1804 	b = nm_os_malloc(sizeof(struct nm_bridge) * n);
1805 	if (b == NULL)
1806 		return NULL;
1807 	for (i = 0; i < n; i++)
1808 		BDG_RWINIT(&b[i]);
1809 	return b;
1810 }
1811 
1812 void
1813 netmap_uninit_bridges2(struct nm_bridge *b, u_int n)
1814 {
1815 	int i;
1816 
1817 	if (b == NULL)
1818 		return;
1819 
1820 	for (i = 0; i < n; i++)
1821 		BDG_RWDESTROY(&b[i]);
1822 	nm_os_free(b);
1823 }
1824 
1825 int
1826 netmap_init_bridges(void)
1827 {
1828 #ifdef CONFIG_NET_NS
1829 	return netmap_bns_register();
1830 #else
1831 	nm_bridges = netmap_init_bridges2(vale_max_bridges);
1832 	if (nm_bridges == NULL)
1833 		return ENOMEM;
1834 	return 0;
1835 #endif
1836 }
1837 
1838 void
1839 netmap_uninit_bridges(void)
1840 {
1841 #ifdef CONFIG_NET_NS
1842 	netmap_bns_unregister();
1843 #else
1844 	netmap_uninit_bridges2(nm_bridges, vale_max_bridges);
1845 #endif
1846 }
1847