xref: /linux/net/bridge/br_fdb.c (revision c532de5a67a70f8533d495f8f2aaa9a0491c3ad0)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *	Forwarding database
4  *	Linux ethernet bridge
5  *
6  *	Authors:
7  *	Lennert Buytenhek		<buytenh@gnu.org>
8  */
9 
10 #include <linux/kernel.h>
11 #include <linux/init.h>
12 #include <linux/rculist.h>
13 #include <linux/spinlock.h>
14 #include <linux/times.h>
15 #include <linux/netdevice.h>
16 #include <linux/etherdevice.h>
17 #include <linux/jhash.h>
18 #include <linux/random.h>
19 #include <linux/slab.h>
20 #include <linux/atomic.h>
21 #include <linux/unaligned.h>
22 #include <linux/if_vlan.h>
23 #include <net/switchdev.h>
24 #include <trace/events/bridge.h>
25 #include "br_private.h"
26 
27 static const struct rhashtable_params br_fdb_rht_params = {
28 	.head_offset = offsetof(struct net_bridge_fdb_entry, rhnode),
29 	.key_offset = offsetof(struct net_bridge_fdb_entry, key),
30 	.key_len = sizeof(struct net_bridge_fdb_key),
31 	.automatic_shrinking = true,
32 };
33 
34 static struct kmem_cache *br_fdb_cache __read_mostly;
35 
36 int __init br_fdb_init(void)
37 {
38 	br_fdb_cache = KMEM_CACHE(net_bridge_fdb_entry, SLAB_HWCACHE_ALIGN);
39 	if (!br_fdb_cache)
40 		return -ENOMEM;
41 
42 	return 0;
43 }
44 
45 void br_fdb_fini(void)
46 {
47 	kmem_cache_destroy(br_fdb_cache);
48 }
49 
50 int br_fdb_hash_init(struct net_bridge *br)
51 {
52 	return rhashtable_init(&br->fdb_hash_tbl, &br_fdb_rht_params);
53 }
54 
55 void br_fdb_hash_fini(struct net_bridge *br)
56 {
57 	rhashtable_destroy(&br->fdb_hash_tbl);
58 }
59 
60 /* if topology_changing then use forward_delay (default 15 sec)
61  * otherwise keep longer (default 5 minutes)
62  */
63 static inline unsigned long hold_time(const struct net_bridge *br)
64 {
65 	return br->topology_change ? br->forward_delay : br->ageing_time;
66 }
67 
68 static inline int has_expired(const struct net_bridge *br,
69 				  const struct net_bridge_fdb_entry *fdb)
70 {
71 	return !test_bit(BR_FDB_STATIC, &fdb->flags) &&
72 	       !test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &fdb->flags) &&
73 	       time_before_eq(fdb->updated + hold_time(br), jiffies);
74 }
75 
76 static void fdb_rcu_free(struct rcu_head *head)
77 {
78 	struct net_bridge_fdb_entry *ent
79 		= container_of(head, struct net_bridge_fdb_entry, rcu);
80 	kmem_cache_free(br_fdb_cache, ent);
81 }
82 
83 static int fdb_to_nud(const struct net_bridge *br,
84 		      const struct net_bridge_fdb_entry *fdb)
85 {
86 	if (test_bit(BR_FDB_LOCAL, &fdb->flags))
87 		return NUD_PERMANENT;
88 	else if (test_bit(BR_FDB_STATIC, &fdb->flags))
89 		return NUD_NOARP;
90 	else if (has_expired(br, fdb))
91 		return NUD_STALE;
92 	else
93 		return NUD_REACHABLE;
94 }
95 
96 static int fdb_fill_info(struct sk_buff *skb, const struct net_bridge *br,
97 			 const struct net_bridge_fdb_entry *fdb,
98 			 u32 portid, u32 seq, int type, unsigned int flags)
99 {
100 	const struct net_bridge_port *dst = READ_ONCE(fdb->dst);
101 	unsigned long now = jiffies;
102 	struct nda_cacheinfo ci;
103 	struct nlmsghdr *nlh;
104 	struct ndmsg *ndm;
105 	u32 ext_flags = 0;
106 
107 	nlh = nlmsg_put(skb, portid, seq, type, sizeof(*ndm), flags);
108 	if (nlh == NULL)
109 		return -EMSGSIZE;
110 
111 	ndm = nlmsg_data(nlh);
112 	ndm->ndm_family	 = AF_BRIDGE;
113 	ndm->ndm_pad1    = 0;
114 	ndm->ndm_pad2    = 0;
115 	ndm->ndm_flags	 = 0;
116 	ndm->ndm_type	 = 0;
117 	ndm->ndm_ifindex = dst ? dst->dev->ifindex : br->dev->ifindex;
118 	ndm->ndm_state   = fdb_to_nud(br, fdb);
119 
120 	if (test_bit(BR_FDB_OFFLOADED, &fdb->flags))
121 		ndm->ndm_flags |= NTF_OFFLOADED;
122 	if (test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &fdb->flags))
123 		ndm->ndm_flags |= NTF_EXT_LEARNED;
124 	if (test_bit(BR_FDB_STICKY, &fdb->flags))
125 		ndm->ndm_flags |= NTF_STICKY;
126 	if (test_bit(BR_FDB_LOCKED, &fdb->flags))
127 		ext_flags |= NTF_EXT_LOCKED;
128 
129 	if (nla_put(skb, NDA_LLADDR, ETH_ALEN, &fdb->key.addr))
130 		goto nla_put_failure;
131 	if (nla_put_u32(skb, NDA_MASTER, br->dev->ifindex))
132 		goto nla_put_failure;
133 	if (nla_put_u32(skb, NDA_FLAGS_EXT, ext_flags))
134 		goto nla_put_failure;
135 
136 	ci.ndm_used	 = jiffies_to_clock_t(now - fdb->used);
137 	ci.ndm_confirmed = 0;
138 	ci.ndm_updated	 = jiffies_to_clock_t(now - fdb->updated);
139 	ci.ndm_refcnt	 = 0;
140 	if (nla_put(skb, NDA_CACHEINFO, sizeof(ci), &ci))
141 		goto nla_put_failure;
142 
143 	if (fdb->key.vlan_id && nla_put(skb, NDA_VLAN, sizeof(u16),
144 					&fdb->key.vlan_id))
145 		goto nla_put_failure;
146 
147 	if (test_bit(BR_FDB_NOTIFY, &fdb->flags)) {
148 		struct nlattr *nest = nla_nest_start(skb, NDA_FDB_EXT_ATTRS);
149 		u8 notify_bits = FDB_NOTIFY_BIT;
150 
151 		if (!nest)
152 			goto nla_put_failure;
153 		if (test_bit(BR_FDB_NOTIFY_INACTIVE, &fdb->flags))
154 			notify_bits |= FDB_NOTIFY_INACTIVE_BIT;
155 
156 		if (nla_put_u8(skb, NFEA_ACTIVITY_NOTIFY, notify_bits)) {
157 			nla_nest_cancel(skb, nest);
158 			goto nla_put_failure;
159 		}
160 
161 		nla_nest_end(skb, nest);
162 	}
163 
164 	nlmsg_end(skb, nlh);
165 	return 0;
166 
167 nla_put_failure:
168 	nlmsg_cancel(skb, nlh);
169 	return -EMSGSIZE;
170 }
171 
172 static inline size_t fdb_nlmsg_size(void)
173 {
174 	return NLMSG_ALIGN(sizeof(struct ndmsg))
175 		+ nla_total_size(ETH_ALEN) /* NDA_LLADDR */
176 		+ nla_total_size(sizeof(u32)) /* NDA_MASTER */
177 		+ nla_total_size(sizeof(u32)) /* NDA_FLAGS_EXT */
178 		+ nla_total_size(sizeof(u16)) /* NDA_VLAN */
179 		+ nla_total_size(sizeof(struct nda_cacheinfo))
180 		+ nla_total_size(0) /* NDA_FDB_EXT_ATTRS */
181 		+ nla_total_size(sizeof(u8)); /* NFEA_ACTIVITY_NOTIFY */
182 }
183 
184 static void fdb_notify(struct net_bridge *br,
185 		       const struct net_bridge_fdb_entry *fdb, int type,
186 		       bool swdev_notify)
187 {
188 	struct net *net = dev_net(br->dev);
189 	struct sk_buff *skb;
190 	int err = -ENOBUFS;
191 
192 	if (swdev_notify)
193 		br_switchdev_fdb_notify(br, fdb, type);
194 
195 	skb = nlmsg_new(fdb_nlmsg_size(), GFP_ATOMIC);
196 	if (skb == NULL)
197 		goto errout;
198 
199 	err = fdb_fill_info(skb, br, fdb, 0, 0, type, 0);
200 	if (err < 0) {
201 		/* -EMSGSIZE implies BUG in fdb_nlmsg_size() */
202 		WARN_ON(err == -EMSGSIZE);
203 		kfree_skb(skb);
204 		goto errout;
205 	}
206 	rtnl_notify(skb, net, 0, RTNLGRP_NEIGH, NULL, GFP_ATOMIC);
207 	return;
208 errout:
209 	rtnl_set_sk_err(net, RTNLGRP_NEIGH, err);
210 }
211 
212 static struct net_bridge_fdb_entry *fdb_find_rcu(struct rhashtable *tbl,
213 						 const unsigned char *addr,
214 						 __u16 vid)
215 {
216 	struct net_bridge_fdb_key key;
217 
218 	WARN_ON_ONCE(!rcu_read_lock_held());
219 
220 	key.vlan_id = vid;
221 	memcpy(key.addr.addr, addr, sizeof(key.addr.addr));
222 
223 	return rhashtable_lookup(tbl, &key, br_fdb_rht_params);
224 }
225 
226 /* requires bridge hash_lock */
227 static struct net_bridge_fdb_entry *br_fdb_find(struct net_bridge *br,
228 						const unsigned char *addr,
229 						__u16 vid)
230 {
231 	struct net_bridge_fdb_entry *fdb;
232 
233 	lockdep_assert_held_once(&br->hash_lock);
234 
235 	rcu_read_lock();
236 	fdb = fdb_find_rcu(&br->fdb_hash_tbl, addr, vid);
237 	rcu_read_unlock();
238 
239 	return fdb;
240 }
241 
242 struct net_device *br_fdb_find_port(const struct net_device *br_dev,
243 				    const unsigned char *addr,
244 				    __u16 vid)
245 {
246 	struct net_bridge_fdb_entry *f;
247 	struct net_device *dev = NULL;
248 	struct net_bridge *br;
249 
250 	ASSERT_RTNL();
251 
252 	if (!netif_is_bridge_master(br_dev))
253 		return NULL;
254 
255 	br = netdev_priv(br_dev);
256 	rcu_read_lock();
257 	f = br_fdb_find_rcu(br, addr, vid);
258 	if (f && f->dst)
259 		dev = f->dst->dev;
260 	rcu_read_unlock();
261 
262 	return dev;
263 }
264 EXPORT_SYMBOL_GPL(br_fdb_find_port);
265 
266 struct net_bridge_fdb_entry *br_fdb_find_rcu(struct net_bridge *br,
267 					     const unsigned char *addr,
268 					     __u16 vid)
269 {
270 	return fdb_find_rcu(&br->fdb_hash_tbl, addr, vid);
271 }
272 
273 /* When a static FDB entry is added, the mac address from the entry is
274  * added to the bridge private HW address list and all required ports
275  * are then updated with the new information.
276  * Called under RTNL.
277  */
278 static void fdb_add_hw_addr(struct net_bridge *br, const unsigned char *addr)
279 {
280 	int err;
281 	struct net_bridge_port *p;
282 
283 	ASSERT_RTNL();
284 
285 	list_for_each_entry(p, &br->port_list, list) {
286 		if (!br_promisc_port(p)) {
287 			err = dev_uc_add(p->dev, addr);
288 			if (err)
289 				goto undo;
290 		}
291 	}
292 
293 	return;
294 undo:
295 	list_for_each_entry_continue_reverse(p, &br->port_list, list) {
296 		if (!br_promisc_port(p))
297 			dev_uc_del(p->dev, addr);
298 	}
299 }
300 
301 /* When a static FDB entry is deleted, the HW address from that entry is
302  * also removed from the bridge private HW address list and updates all
303  * the ports with needed information.
304  * Called under RTNL.
305  */
306 static void fdb_del_hw_addr(struct net_bridge *br, const unsigned char *addr)
307 {
308 	struct net_bridge_port *p;
309 
310 	ASSERT_RTNL();
311 
312 	list_for_each_entry(p, &br->port_list, list) {
313 		if (!br_promisc_port(p))
314 			dev_uc_del(p->dev, addr);
315 	}
316 }
317 
318 static void fdb_delete(struct net_bridge *br, struct net_bridge_fdb_entry *f,
319 		       bool swdev_notify)
320 {
321 	trace_fdb_delete(br, f);
322 
323 	if (test_bit(BR_FDB_STATIC, &f->flags))
324 		fdb_del_hw_addr(br, f->key.addr.addr);
325 
326 	hlist_del_init_rcu(&f->fdb_node);
327 	rhashtable_remove_fast(&br->fdb_hash_tbl, &f->rhnode,
328 			       br_fdb_rht_params);
329 	if (test_and_clear_bit(BR_FDB_DYNAMIC_LEARNED, &f->flags))
330 		atomic_dec(&br->fdb_n_learned);
331 	fdb_notify(br, f, RTM_DELNEIGH, swdev_notify);
332 	call_rcu(&f->rcu, fdb_rcu_free);
333 }
334 
335 /* Delete a local entry if no other port had the same address.
336  *
337  * This function should only be called on entries with BR_FDB_LOCAL set,
338  * so even with BR_FDB_ADDED_BY_USER cleared we never need to increase
339  * the accounting for dynamically learned entries again.
340  */
341 static void fdb_delete_local(struct net_bridge *br,
342 			     const struct net_bridge_port *p,
343 			     struct net_bridge_fdb_entry *f)
344 {
345 	const unsigned char *addr = f->key.addr.addr;
346 	struct net_bridge_vlan_group *vg;
347 	const struct net_bridge_vlan *v;
348 	struct net_bridge_port *op;
349 	u16 vid = f->key.vlan_id;
350 
351 	/* Maybe another port has same hw addr? */
352 	list_for_each_entry(op, &br->port_list, list) {
353 		vg = nbp_vlan_group(op);
354 		if (op != p && ether_addr_equal(op->dev->dev_addr, addr) &&
355 		    (!vid || br_vlan_find(vg, vid))) {
356 			f->dst = op;
357 			clear_bit(BR_FDB_ADDED_BY_USER, &f->flags);
358 			return;
359 		}
360 	}
361 
362 	vg = br_vlan_group(br);
363 	v = br_vlan_find(vg, vid);
364 	/* Maybe bridge device has same hw addr? */
365 	if (p && ether_addr_equal(br->dev->dev_addr, addr) &&
366 	    (!vid || (v && br_vlan_should_use(v)))) {
367 		f->dst = NULL;
368 		clear_bit(BR_FDB_ADDED_BY_USER, &f->flags);
369 		return;
370 	}
371 
372 	fdb_delete(br, f, true);
373 }
374 
375 void br_fdb_find_delete_local(struct net_bridge *br,
376 			      const struct net_bridge_port *p,
377 			      const unsigned char *addr, u16 vid)
378 {
379 	struct net_bridge_fdb_entry *f;
380 
381 	spin_lock_bh(&br->hash_lock);
382 	f = br_fdb_find(br, addr, vid);
383 	if (f && test_bit(BR_FDB_LOCAL, &f->flags) &&
384 	    !test_bit(BR_FDB_ADDED_BY_USER, &f->flags) && f->dst == p)
385 		fdb_delete_local(br, p, f);
386 	spin_unlock_bh(&br->hash_lock);
387 }
388 
389 static struct net_bridge_fdb_entry *fdb_create(struct net_bridge *br,
390 					       struct net_bridge_port *source,
391 					       const unsigned char *addr,
392 					       __u16 vid,
393 					       unsigned long flags)
394 {
395 	bool learned = !test_bit(BR_FDB_ADDED_BY_USER, &flags) &&
396 		       !test_bit(BR_FDB_LOCAL, &flags);
397 	u32 max_learned = READ_ONCE(br->fdb_max_learned);
398 	struct net_bridge_fdb_entry *fdb;
399 	int err;
400 
401 	if (likely(learned)) {
402 		int n_learned = atomic_read(&br->fdb_n_learned);
403 
404 		if (unlikely(max_learned && n_learned >= max_learned))
405 			return NULL;
406 		__set_bit(BR_FDB_DYNAMIC_LEARNED, &flags);
407 	}
408 
409 	fdb = kmem_cache_alloc(br_fdb_cache, GFP_ATOMIC);
410 	if (!fdb)
411 		return NULL;
412 
413 	memcpy(fdb->key.addr.addr, addr, ETH_ALEN);
414 	WRITE_ONCE(fdb->dst, source);
415 	fdb->key.vlan_id = vid;
416 	fdb->flags = flags;
417 	fdb->updated = fdb->used = jiffies;
418 	err = rhashtable_lookup_insert_fast(&br->fdb_hash_tbl, &fdb->rhnode,
419 					    br_fdb_rht_params);
420 	if (err) {
421 		kmem_cache_free(br_fdb_cache, fdb);
422 		return NULL;
423 	}
424 
425 	if (likely(learned))
426 		atomic_inc(&br->fdb_n_learned);
427 
428 	hlist_add_head_rcu(&fdb->fdb_node, &br->fdb_list);
429 
430 	return fdb;
431 }
432 
433 static int fdb_add_local(struct net_bridge *br, struct net_bridge_port *source,
434 			 const unsigned char *addr, u16 vid)
435 {
436 	struct net_bridge_fdb_entry *fdb;
437 
438 	if (!is_valid_ether_addr(addr))
439 		return -EINVAL;
440 
441 	fdb = br_fdb_find(br, addr, vid);
442 	if (fdb) {
443 		/* it is okay to have multiple ports with same
444 		 * address, just use the first one.
445 		 */
446 		if (test_bit(BR_FDB_LOCAL, &fdb->flags))
447 			return 0;
448 		br_warn(br, "adding interface %s with same address as a received packet (addr:%pM, vlan:%u)\n",
449 			source ? source->dev->name : br->dev->name, addr, vid);
450 		fdb_delete(br, fdb, true);
451 	}
452 
453 	fdb = fdb_create(br, source, addr, vid,
454 			 BIT(BR_FDB_LOCAL) | BIT(BR_FDB_STATIC));
455 	if (!fdb)
456 		return -ENOMEM;
457 
458 	fdb_add_hw_addr(br, addr);
459 	fdb_notify(br, fdb, RTM_NEWNEIGH, true);
460 	return 0;
461 }
462 
463 void br_fdb_changeaddr(struct net_bridge_port *p, const unsigned char *newaddr)
464 {
465 	struct net_bridge_vlan_group *vg;
466 	struct net_bridge_fdb_entry *f;
467 	struct net_bridge *br = p->br;
468 	struct net_bridge_vlan *v;
469 
470 	spin_lock_bh(&br->hash_lock);
471 	vg = nbp_vlan_group(p);
472 	hlist_for_each_entry(f, &br->fdb_list, fdb_node) {
473 		if (f->dst == p && test_bit(BR_FDB_LOCAL, &f->flags) &&
474 		    !test_bit(BR_FDB_ADDED_BY_USER, &f->flags)) {
475 			/* delete old one */
476 			fdb_delete_local(br, p, f);
477 
478 			/* if this port has no vlan information
479 			 * configured, we can safely be done at
480 			 * this point.
481 			 */
482 			if (!vg || !vg->num_vlans)
483 				goto insert;
484 		}
485 	}
486 
487 insert:
488 	/* insert new address,  may fail if invalid address or dup. */
489 	fdb_add_local(br, p, newaddr, 0);
490 
491 	if (!vg || !vg->num_vlans)
492 		goto done;
493 
494 	/* Now add entries for every VLAN configured on the port.
495 	 * This function runs under RTNL so the bitmap will not change
496 	 * from under us.
497 	 */
498 	list_for_each_entry(v, &vg->vlan_list, vlist)
499 		fdb_add_local(br, p, newaddr, v->vid);
500 
501 done:
502 	spin_unlock_bh(&br->hash_lock);
503 }
504 
505 void br_fdb_change_mac_address(struct net_bridge *br, const u8 *newaddr)
506 {
507 	struct net_bridge_vlan_group *vg;
508 	struct net_bridge_fdb_entry *f;
509 	struct net_bridge_vlan *v;
510 
511 	spin_lock_bh(&br->hash_lock);
512 
513 	/* If old entry was unassociated with any port, then delete it. */
514 	f = br_fdb_find(br, br->dev->dev_addr, 0);
515 	if (f && test_bit(BR_FDB_LOCAL, &f->flags) &&
516 	    !f->dst && !test_bit(BR_FDB_ADDED_BY_USER, &f->flags))
517 		fdb_delete_local(br, NULL, f);
518 
519 	fdb_add_local(br, NULL, newaddr, 0);
520 	vg = br_vlan_group(br);
521 	if (!vg || !vg->num_vlans)
522 		goto out;
523 	/* Now remove and add entries for every VLAN configured on the
524 	 * bridge.  This function runs under RTNL so the bitmap will not
525 	 * change from under us.
526 	 */
527 	list_for_each_entry(v, &vg->vlan_list, vlist) {
528 		if (!br_vlan_should_use(v))
529 			continue;
530 		f = br_fdb_find(br, br->dev->dev_addr, v->vid);
531 		if (f && test_bit(BR_FDB_LOCAL, &f->flags) &&
532 		    !f->dst && !test_bit(BR_FDB_ADDED_BY_USER, &f->flags))
533 			fdb_delete_local(br, NULL, f);
534 		fdb_add_local(br, NULL, newaddr, v->vid);
535 	}
536 out:
537 	spin_unlock_bh(&br->hash_lock);
538 }
539 
540 void br_fdb_cleanup(struct work_struct *work)
541 {
542 	struct net_bridge *br = container_of(work, struct net_bridge,
543 					     gc_work.work);
544 	struct net_bridge_fdb_entry *f = NULL;
545 	unsigned long delay = hold_time(br);
546 	unsigned long work_delay = delay;
547 	unsigned long now = jiffies;
548 
549 	/* this part is tricky, in order to avoid blocking learning and
550 	 * consequently forwarding, we rely on rcu to delete objects with
551 	 * delayed freeing allowing us to continue traversing
552 	 */
553 	rcu_read_lock();
554 	hlist_for_each_entry_rcu(f, &br->fdb_list, fdb_node) {
555 		unsigned long this_timer = f->updated + delay;
556 
557 		if (test_bit(BR_FDB_STATIC, &f->flags) ||
558 		    test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &f->flags)) {
559 			if (test_bit(BR_FDB_NOTIFY, &f->flags)) {
560 				if (time_after(this_timer, now))
561 					work_delay = min(work_delay,
562 							 this_timer - now);
563 				else if (!test_and_set_bit(BR_FDB_NOTIFY_INACTIVE,
564 							   &f->flags))
565 					fdb_notify(br, f, RTM_NEWNEIGH, false);
566 			}
567 			continue;
568 		}
569 
570 		if (time_after(this_timer, now)) {
571 			work_delay = min(work_delay, this_timer - now);
572 		} else {
573 			spin_lock_bh(&br->hash_lock);
574 			if (!hlist_unhashed(&f->fdb_node))
575 				fdb_delete(br, f, true);
576 			spin_unlock_bh(&br->hash_lock);
577 		}
578 	}
579 	rcu_read_unlock();
580 
581 	/* Cleanup minimum 10 milliseconds apart */
582 	work_delay = max_t(unsigned long, work_delay, msecs_to_jiffies(10));
583 	mod_delayed_work(system_long_wq, &br->gc_work, work_delay);
584 }
585 
586 static bool __fdb_flush_matches(const struct net_bridge *br,
587 				const struct net_bridge_fdb_entry *f,
588 				const struct net_bridge_fdb_flush_desc *desc)
589 {
590 	const struct net_bridge_port *dst = READ_ONCE(f->dst);
591 	int port_ifidx = dst ? dst->dev->ifindex : br->dev->ifindex;
592 
593 	if (desc->vlan_id && desc->vlan_id != f->key.vlan_id)
594 		return false;
595 	if (desc->port_ifindex && desc->port_ifindex != port_ifidx)
596 		return false;
597 	if (desc->flags_mask && (f->flags & desc->flags_mask) != desc->flags)
598 		return false;
599 
600 	return true;
601 }
602 
603 /* Flush forwarding database entries matching the description */
604 void br_fdb_flush(struct net_bridge *br,
605 		  const struct net_bridge_fdb_flush_desc *desc)
606 {
607 	struct net_bridge_fdb_entry *f;
608 
609 	rcu_read_lock();
610 	hlist_for_each_entry_rcu(f, &br->fdb_list, fdb_node) {
611 		if (!__fdb_flush_matches(br, f, desc))
612 			continue;
613 
614 		spin_lock_bh(&br->hash_lock);
615 		if (!hlist_unhashed(&f->fdb_node))
616 			fdb_delete(br, f, true);
617 		spin_unlock_bh(&br->hash_lock);
618 	}
619 	rcu_read_unlock();
620 }
621 
622 static unsigned long __ndm_state_to_fdb_flags(u16 ndm_state)
623 {
624 	unsigned long flags = 0;
625 
626 	if (ndm_state & NUD_PERMANENT)
627 		__set_bit(BR_FDB_LOCAL, &flags);
628 	if (ndm_state & NUD_NOARP)
629 		__set_bit(BR_FDB_STATIC, &flags);
630 
631 	return flags;
632 }
633 
634 static unsigned long __ndm_flags_to_fdb_flags(u8 ndm_flags)
635 {
636 	unsigned long flags = 0;
637 
638 	if (ndm_flags & NTF_USE)
639 		__set_bit(BR_FDB_ADDED_BY_USER, &flags);
640 	if (ndm_flags & NTF_EXT_LEARNED)
641 		__set_bit(BR_FDB_ADDED_BY_EXT_LEARN, &flags);
642 	if (ndm_flags & NTF_OFFLOADED)
643 		__set_bit(BR_FDB_OFFLOADED, &flags);
644 	if (ndm_flags & NTF_STICKY)
645 		__set_bit(BR_FDB_STICKY, &flags);
646 
647 	return flags;
648 }
649 
650 static int __fdb_flush_validate_ifindex(const struct net_bridge *br,
651 					int ifindex,
652 					struct netlink_ext_ack *extack)
653 {
654 	const struct net_device *dev;
655 
656 	dev = __dev_get_by_index(dev_net(br->dev), ifindex);
657 	if (!dev) {
658 		NL_SET_ERR_MSG_MOD(extack, "Unknown flush device ifindex");
659 		return -ENODEV;
660 	}
661 	if (!netif_is_bridge_master(dev) && !netif_is_bridge_port(dev)) {
662 		NL_SET_ERR_MSG_MOD(extack, "Flush device is not a bridge or bridge port");
663 		return -EINVAL;
664 	}
665 	if (netif_is_bridge_master(dev) && dev != br->dev) {
666 		NL_SET_ERR_MSG_MOD(extack,
667 				   "Flush bridge device does not match target bridge device");
668 		return -EINVAL;
669 	}
670 	if (netif_is_bridge_port(dev)) {
671 		struct net_bridge_port *p = br_port_get_rtnl(dev);
672 
673 		if (p->br != br) {
674 			NL_SET_ERR_MSG_MOD(extack, "Port belongs to a different bridge device");
675 			return -EINVAL;
676 		}
677 	}
678 
679 	return 0;
680 }
681 
682 static const struct nla_policy br_fdb_del_bulk_policy[NDA_MAX + 1] = {
683 	[NDA_VLAN]	= NLA_POLICY_RANGE(NLA_U16, 1, VLAN_N_VID - 2),
684 	[NDA_IFINDEX]	= NLA_POLICY_MIN(NLA_S32, 1),
685 	[NDA_NDM_STATE_MASK]	= { .type = NLA_U16 },
686 	[NDA_NDM_FLAGS_MASK]	= { .type = NLA_U8 },
687 };
688 
689 int br_fdb_delete_bulk(struct nlmsghdr *nlh, struct net_device *dev,
690 		       struct netlink_ext_ack *extack)
691 {
692 	struct net_bridge_fdb_flush_desc desc = {};
693 	struct ndmsg *ndm = nlmsg_data(nlh);
694 	struct net_bridge_port *p = NULL;
695 	struct nlattr *tb[NDA_MAX + 1];
696 	struct net_bridge *br;
697 	u8 ndm_flags;
698 	int err;
699 
700 	ndm_flags = ndm->ndm_flags & ~FDB_FLUSH_IGNORED_NDM_FLAGS;
701 
702 	err = nlmsg_parse(nlh, sizeof(*ndm), tb, NDA_MAX,
703 			  br_fdb_del_bulk_policy, extack);
704 	if (err)
705 		return err;
706 
707 	if (netif_is_bridge_master(dev)) {
708 		br = netdev_priv(dev);
709 	} else {
710 		p = br_port_get_rtnl(dev);
711 		if (!p) {
712 			NL_SET_ERR_MSG_MOD(extack, "Device is not a bridge port");
713 			return -EINVAL;
714 		}
715 		br = p->br;
716 	}
717 
718 	if (tb[NDA_VLAN])
719 		desc.vlan_id = nla_get_u16(tb[NDA_VLAN]);
720 
721 	if (ndm_flags & ~FDB_FLUSH_ALLOWED_NDM_FLAGS) {
722 		NL_SET_ERR_MSG(extack, "Unsupported fdb flush ndm flag bits set");
723 		return -EINVAL;
724 	}
725 	if (ndm->ndm_state & ~FDB_FLUSH_ALLOWED_NDM_STATES) {
726 		NL_SET_ERR_MSG(extack, "Unsupported fdb flush ndm state bits set");
727 		return -EINVAL;
728 	}
729 
730 	desc.flags |= __ndm_state_to_fdb_flags(ndm->ndm_state);
731 	desc.flags |= __ndm_flags_to_fdb_flags(ndm_flags);
732 	if (tb[NDA_NDM_STATE_MASK]) {
733 		u16 ndm_state_mask = nla_get_u16(tb[NDA_NDM_STATE_MASK]);
734 
735 		desc.flags_mask |= __ndm_state_to_fdb_flags(ndm_state_mask);
736 	}
737 	if (tb[NDA_NDM_FLAGS_MASK]) {
738 		u8 ndm_flags_mask = nla_get_u8(tb[NDA_NDM_FLAGS_MASK]);
739 
740 		desc.flags_mask |= __ndm_flags_to_fdb_flags(ndm_flags_mask);
741 	}
742 	if (tb[NDA_IFINDEX]) {
743 		int ifidx = nla_get_s32(tb[NDA_IFINDEX]);
744 
745 		err = __fdb_flush_validate_ifindex(br, ifidx, extack);
746 		if (err)
747 			return err;
748 		desc.port_ifindex = ifidx;
749 	} else if (p) {
750 		/* flush was invoked with port device and NTF_MASTER */
751 		desc.port_ifindex = p->dev->ifindex;
752 	}
753 
754 	br_debug(br, "flushing port ifindex: %d vlan id: %u flags: 0x%lx flags mask: 0x%lx\n",
755 		 desc.port_ifindex, desc.vlan_id, desc.flags, desc.flags_mask);
756 
757 	br_fdb_flush(br, &desc);
758 
759 	return 0;
760 }
761 
762 /* Flush all entries referring to a specific port.
763  * if do_all is set also flush static entries
764  * if vid is set delete all entries that match the vlan_id
765  */
766 void br_fdb_delete_by_port(struct net_bridge *br,
767 			   const struct net_bridge_port *p,
768 			   u16 vid,
769 			   int do_all)
770 {
771 	struct net_bridge_fdb_entry *f;
772 	struct hlist_node *tmp;
773 
774 	spin_lock_bh(&br->hash_lock);
775 	hlist_for_each_entry_safe(f, tmp, &br->fdb_list, fdb_node) {
776 		if (f->dst != p)
777 			continue;
778 
779 		if (!do_all)
780 			if (test_bit(BR_FDB_STATIC, &f->flags) ||
781 			    (test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &f->flags) &&
782 			     !test_bit(BR_FDB_OFFLOADED, &f->flags)) ||
783 			    (vid && f->key.vlan_id != vid))
784 				continue;
785 
786 		if (test_bit(BR_FDB_LOCAL, &f->flags))
787 			fdb_delete_local(br, p, f);
788 		else
789 			fdb_delete(br, f, true);
790 	}
791 	spin_unlock_bh(&br->hash_lock);
792 }
793 
794 #if IS_ENABLED(CONFIG_ATM_LANE)
795 /* Interface used by ATM LANE hook to test
796  * if an addr is on some other bridge port */
797 int br_fdb_test_addr(struct net_device *dev, unsigned char *addr)
798 {
799 	struct net_bridge_fdb_entry *fdb;
800 	struct net_bridge_port *port;
801 	int ret;
802 
803 	rcu_read_lock();
804 	port = br_port_get_rcu(dev);
805 	if (!port)
806 		ret = 0;
807 	else {
808 		const struct net_bridge_port *dst = NULL;
809 
810 		fdb = br_fdb_find_rcu(port->br, addr, 0);
811 		if (fdb)
812 			dst = READ_ONCE(fdb->dst);
813 
814 		ret = dst && dst->dev != dev &&
815 		      dst->state == BR_STATE_FORWARDING;
816 	}
817 	rcu_read_unlock();
818 
819 	return ret;
820 }
821 #endif /* CONFIG_ATM_LANE */
822 
823 /*
824  * Fill buffer with forwarding table records in
825  * the API format.
826  */
827 int br_fdb_fillbuf(struct net_bridge *br, void *buf,
828 		   unsigned long maxnum, unsigned long skip)
829 {
830 	struct net_bridge_fdb_entry *f;
831 	struct __fdb_entry *fe = buf;
832 	int num = 0;
833 
834 	memset(buf, 0, maxnum*sizeof(struct __fdb_entry));
835 
836 	rcu_read_lock();
837 	hlist_for_each_entry_rcu(f, &br->fdb_list, fdb_node) {
838 		if (num >= maxnum)
839 			break;
840 
841 		if (has_expired(br, f))
842 			continue;
843 
844 		/* ignore pseudo entry for local MAC address */
845 		if (!f->dst)
846 			continue;
847 
848 		if (skip) {
849 			--skip;
850 			continue;
851 		}
852 
853 		/* convert from internal format to API */
854 		memcpy(fe->mac_addr, f->key.addr.addr, ETH_ALEN);
855 
856 		/* due to ABI compat need to split into hi/lo */
857 		fe->port_no = f->dst->port_no;
858 		fe->port_hi = f->dst->port_no >> 8;
859 
860 		fe->is_local = test_bit(BR_FDB_LOCAL, &f->flags);
861 		if (!test_bit(BR_FDB_STATIC, &f->flags))
862 			fe->ageing_timer_value = jiffies_delta_to_clock_t(jiffies - f->updated);
863 		++fe;
864 		++num;
865 	}
866 	rcu_read_unlock();
867 
868 	return num;
869 }
870 
871 /* Add entry for local address of interface */
872 int br_fdb_add_local(struct net_bridge *br, struct net_bridge_port *source,
873 		     const unsigned char *addr, u16 vid)
874 {
875 	int ret;
876 
877 	spin_lock_bh(&br->hash_lock);
878 	ret = fdb_add_local(br, source, addr, vid);
879 	spin_unlock_bh(&br->hash_lock);
880 	return ret;
881 }
882 
883 /* returns true if the fdb was modified */
884 static bool __fdb_mark_active(struct net_bridge_fdb_entry *fdb)
885 {
886 	return !!(test_bit(BR_FDB_NOTIFY_INACTIVE, &fdb->flags) &&
887 		  test_and_clear_bit(BR_FDB_NOTIFY_INACTIVE, &fdb->flags));
888 }
889 
890 void br_fdb_update(struct net_bridge *br, struct net_bridge_port *source,
891 		   const unsigned char *addr, u16 vid, unsigned long flags)
892 {
893 	struct net_bridge_fdb_entry *fdb;
894 
895 	/* some users want to always flood. */
896 	if (hold_time(br) == 0)
897 		return;
898 
899 	fdb = fdb_find_rcu(&br->fdb_hash_tbl, addr, vid);
900 	if (likely(fdb)) {
901 		/* attempt to update an entry for a local interface */
902 		if (unlikely(test_bit(BR_FDB_LOCAL, &fdb->flags))) {
903 			if (net_ratelimit())
904 				br_warn(br, "received packet on %s with own address as source address (addr:%pM, vlan:%u)\n",
905 					source->dev->name, addr, vid);
906 		} else {
907 			unsigned long now = jiffies;
908 			bool fdb_modified = false;
909 
910 			if (now != fdb->updated) {
911 				fdb->updated = now;
912 				fdb_modified = __fdb_mark_active(fdb);
913 			}
914 
915 			/* fastpath: update of existing entry */
916 			if (unlikely(source != READ_ONCE(fdb->dst) &&
917 				     !test_bit(BR_FDB_STICKY, &fdb->flags))) {
918 				br_switchdev_fdb_notify(br, fdb, RTM_DELNEIGH);
919 				WRITE_ONCE(fdb->dst, source);
920 				fdb_modified = true;
921 				/* Take over HW learned entry */
922 				if (unlikely(test_bit(BR_FDB_ADDED_BY_EXT_LEARN,
923 						      &fdb->flags)))
924 					clear_bit(BR_FDB_ADDED_BY_EXT_LEARN,
925 						  &fdb->flags);
926 				/* Clear locked flag when roaming to an
927 				 * unlocked port.
928 				 */
929 				if (unlikely(test_bit(BR_FDB_LOCKED, &fdb->flags)))
930 					clear_bit(BR_FDB_LOCKED, &fdb->flags);
931 			}
932 
933 			if (unlikely(test_bit(BR_FDB_ADDED_BY_USER, &flags))) {
934 				set_bit(BR_FDB_ADDED_BY_USER, &fdb->flags);
935 				if (test_and_clear_bit(BR_FDB_DYNAMIC_LEARNED,
936 						       &fdb->flags))
937 					atomic_dec(&br->fdb_n_learned);
938 			}
939 			if (unlikely(fdb_modified)) {
940 				trace_br_fdb_update(br, source, addr, vid, flags);
941 				fdb_notify(br, fdb, RTM_NEWNEIGH, true);
942 			}
943 		}
944 	} else {
945 		spin_lock(&br->hash_lock);
946 		fdb = fdb_create(br, source, addr, vid, flags);
947 		if (fdb) {
948 			trace_br_fdb_update(br, source, addr, vid, flags);
949 			fdb_notify(br, fdb, RTM_NEWNEIGH, true);
950 		}
951 		/* else  we lose race and someone else inserts
952 		 * it first, don't bother updating
953 		 */
954 		spin_unlock(&br->hash_lock);
955 	}
956 }
957 
958 /* Dump information about entries, in response to GETNEIGH */
959 int br_fdb_dump(struct sk_buff *skb,
960 		struct netlink_callback *cb,
961 		struct net_device *dev,
962 		struct net_device *filter_dev,
963 		int *idx)
964 {
965 	struct net_bridge *br = netdev_priv(dev);
966 	struct net_bridge_fdb_entry *f;
967 	int err = 0;
968 
969 	if (!netif_is_bridge_master(dev))
970 		return err;
971 
972 	if (!filter_dev) {
973 		err = ndo_dflt_fdb_dump(skb, cb, dev, NULL, idx);
974 		if (err < 0)
975 			return err;
976 	}
977 
978 	rcu_read_lock();
979 	hlist_for_each_entry_rcu(f, &br->fdb_list, fdb_node) {
980 		if (*idx < cb->args[2])
981 			goto skip;
982 		if (filter_dev && (!f->dst || f->dst->dev != filter_dev)) {
983 			if (filter_dev != dev)
984 				goto skip;
985 			/* !f->dst is a special case for bridge
986 			 * It means the MAC belongs to the bridge
987 			 * Therefore need a little more filtering
988 			 * we only want to dump the !f->dst case
989 			 */
990 			if (f->dst)
991 				goto skip;
992 		}
993 		if (!filter_dev && f->dst)
994 			goto skip;
995 
996 		err = fdb_fill_info(skb, br, f,
997 				    NETLINK_CB(cb->skb).portid,
998 				    cb->nlh->nlmsg_seq,
999 				    RTM_NEWNEIGH,
1000 				    NLM_F_MULTI);
1001 		if (err < 0)
1002 			break;
1003 skip:
1004 		*idx += 1;
1005 	}
1006 	rcu_read_unlock();
1007 
1008 	return err;
1009 }
1010 
1011 int br_fdb_get(struct sk_buff *skb,
1012 	       struct nlattr *tb[],
1013 	       struct net_device *dev,
1014 	       const unsigned char *addr,
1015 	       u16 vid, u32 portid, u32 seq,
1016 	       struct netlink_ext_ack *extack)
1017 {
1018 	struct net_bridge *br = netdev_priv(dev);
1019 	struct net_bridge_fdb_entry *f;
1020 	int err = 0;
1021 
1022 	rcu_read_lock();
1023 	f = br_fdb_find_rcu(br, addr, vid);
1024 	if (!f) {
1025 		NL_SET_ERR_MSG(extack, "Fdb entry not found");
1026 		err = -ENOENT;
1027 		goto errout;
1028 	}
1029 
1030 	err = fdb_fill_info(skb, br, f, portid, seq,
1031 			    RTM_NEWNEIGH, 0);
1032 errout:
1033 	rcu_read_unlock();
1034 	return err;
1035 }
1036 
1037 /* returns true if the fdb is modified */
1038 static bool fdb_handle_notify(struct net_bridge_fdb_entry *fdb, u8 notify)
1039 {
1040 	bool modified = false;
1041 
1042 	/* allow to mark an entry as inactive, usually done on creation */
1043 	if ((notify & FDB_NOTIFY_INACTIVE_BIT) &&
1044 	    !test_and_set_bit(BR_FDB_NOTIFY_INACTIVE, &fdb->flags))
1045 		modified = true;
1046 
1047 	if ((notify & FDB_NOTIFY_BIT) &&
1048 	    !test_and_set_bit(BR_FDB_NOTIFY, &fdb->flags)) {
1049 		/* enabled activity tracking */
1050 		modified = true;
1051 	} else if (!(notify & FDB_NOTIFY_BIT) &&
1052 		   test_and_clear_bit(BR_FDB_NOTIFY, &fdb->flags)) {
1053 		/* disabled activity tracking, clear notify state */
1054 		clear_bit(BR_FDB_NOTIFY_INACTIVE, &fdb->flags);
1055 		modified = true;
1056 	}
1057 
1058 	return modified;
1059 }
1060 
1061 /* Update (create or replace) forwarding database entry */
1062 static int fdb_add_entry(struct net_bridge *br, struct net_bridge_port *source,
1063 			 const u8 *addr, struct ndmsg *ndm, u16 flags, u16 vid,
1064 			 struct nlattr *nfea_tb[])
1065 {
1066 	bool is_sticky = !!(ndm->ndm_flags & NTF_STICKY);
1067 	bool refresh = !nfea_tb[NFEA_DONT_REFRESH];
1068 	struct net_bridge_fdb_entry *fdb;
1069 	u16 state = ndm->ndm_state;
1070 	bool modified = false;
1071 	u8 notify = 0;
1072 
1073 	/* If the port cannot learn allow only local and static entries */
1074 	if (source && !(state & NUD_PERMANENT) && !(state & NUD_NOARP) &&
1075 	    !(source->state == BR_STATE_LEARNING ||
1076 	      source->state == BR_STATE_FORWARDING))
1077 		return -EPERM;
1078 
1079 	if (!source && !(state & NUD_PERMANENT)) {
1080 		pr_info("bridge: RTM_NEWNEIGH %s without NUD_PERMANENT\n",
1081 			br->dev->name);
1082 		return -EINVAL;
1083 	}
1084 
1085 	if (is_sticky && (state & NUD_PERMANENT))
1086 		return -EINVAL;
1087 
1088 	if (nfea_tb[NFEA_ACTIVITY_NOTIFY]) {
1089 		notify = nla_get_u8(nfea_tb[NFEA_ACTIVITY_NOTIFY]);
1090 		if ((notify & ~BR_FDB_NOTIFY_SETTABLE_BITS) ||
1091 		    (notify & BR_FDB_NOTIFY_SETTABLE_BITS) == FDB_NOTIFY_INACTIVE_BIT)
1092 			return -EINVAL;
1093 	}
1094 
1095 	fdb = br_fdb_find(br, addr, vid);
1096 	if (fdb == NULL) {
1097 		if (!(flags & NLM_F_CREATE))
1098 			return -ENOENT;
1099 
1100 		fdb = fdb_create(br, source, addr, vid,
1101 				 BIT(BR_FDB_ADDED_BY_USER));
1102 		if (!fdb)
1103 			return -ENOMEM;
1104 
1105 		modified = true;
1106 	} else {
1107 		if (flags & NLM_F_EXCL)
1108 			return -EEXIST;
1109 
1110 		if (READ_ONCE(fdb->dst) != source) {
1111 			WRITE_ONCE(fdb->dst, source);
1112 			modified = true;
1113 		}
1114 
1115 		set_bit(BR_FDB_ADDED_BY_USER, &fdb->flags);
1116 		if (test_and_clear_bit(BR_FDB_DYNAMIC_LEARNED, &fdb->flags))
1117 			atomic_dec(&br->fdb_n_learned);
1118 	}
1119 
1120 	if (fdb_to_nud(br, fdb) != state) {
1121 		if (state & NUD_PERMANENT) {
1122 			set_bit(BR_FDB_LOCAL, &fdb->flags);
1123 			if (!test_and_set_bit(BR_FDB_STATIC, &fdb->flags))
1124 				fdb_add_hw_addr(br, addr);
1125 		} else if (state & NUD_NOARP) {
1126 			clear_bit(BR_FDB_LOCAL, &fdb->flags);
1127 			if (!test_and_set_bit(BR_FDB_STATIC, &fdb->flags))
1128 				fdb_add_hw_addr(br, addr);
1129 		} else {
1130 			clear_bit(BR_FDB_LOCAL, &fdb->flags);
1131 			if (test_and_clear_bit(BR_FDB_STATIC, &fdb->flags))
1132 				fdb_del_hw_addr(br, addr);
1133 		}
1134 
1135 		modified = true;
1136 	}
1137 
1138 	if (is_sticky != test_bit(BR_FDB_STICKY, &fdb->flags)) {
1139 		change_bit(BR_FDB_STICKY, &fdb->flags);
1140 		modified = true;
1141 	}
1142 
1143 	if (test_and_clear_bit(BR_FDB_LOCKED, &fdb->flags))
1144 		modified = true;
1145 
1146 	if (fdb_handle_notify(fdb, notify))
1147 		modified = true;
1148 
1149 	fdb->used = jiffies;
1150 	if (modified) {
1151 		if (refresh)
1152 			fdb->updated = jiffies;
1153 		fdb_notify(br, fdb, RTM_NEWNEIGH, true);
1154 	}
1155 
1156 	return 0;
1157 }
1158 
1159 static int __br_fdb_add(struct ndmsg *ndm, struct net_bridge *br,
1160 			struct net_bridge_port *p, const unsigned char *addr,
1161 			u16 nlh_flags, u16 vid, struct nlattr *nfea_tb[],
1162 			struct netlink_ext_ack *extack)
1163 {
1164 	int err = 0;
1165 
1166 	if (ndm->ndm_flags & NTF_USE) {
1167 		if (!p) {
1168 			pr_info("bridge: RTM_NEWNEIGH %s with NTF_USE is not supported\n",
1169 				br->dev->name);
1170 			return -EINVAL;
1171 		}
1172 		if (!nbp_state_should_learn(p))
1173 			return 0;
1174 
1175 		local_bh_disable();
1176 		rcu_read_lock();
1177 		br_fdb_update(br, p, addr, vid, BIT(BR_FDB_ADDED_BY_USER));
1178 		rcu_read_unlock();
1179 		local_bh_enable();
1180 	} else if (ndm->ndm_flags & NTF_EXT_LEARNED) {
1181 		if (!p && !(ndm->ndm_state & NUD_PERMANENT)) {
1182 			NL_SET_ERR_MSG_MOD(extack,
1183 					   "FDB entry towards bridge must be permanent");
1184 			return -EINVAL;
1185 		}
1186 		err = br_fdb_external_learn_add(br, p, addr, vid, false, true);
1187 	} else {
1188 		spin_lock_bh(&br->hash_lock);
1189 		err = fdb_add_entry(br, p, addr, ndm, nlh_flags, vid, nfea_tb);
1190 		spin_unlock_bh(&br->hash_lock);
1191 	}
1192 
1193 	return err;
1194 }
1195 
1196 static const struct nla_policy br_nda_fdb_pol[NFEA_MAX + 1] = {
1197 	[NFEA_ACTIVITY_NOTIFY]	= { .type = NLA_U8 },
1198 	[NFEA_DONT_REFRESH]	= { .type = NLA_FLAG },
1199 };
1200 
1201 /* Add new permanent fdb entry with RTM_NEWNEIGH */
1202 int br_fdb_add(struct ndmsg *ndm, struct nlattr *tb[],
1203 	       struct net_device *dev,
1204 	       const unsigned char *addr, u16 vid, u16 nlh_flags,
1205 	       struct netlink_ext_ack *extack)
1206 {
1207 	struct nlattr *nfea_tb[NFEA_MAX + 1], *attr;
1208 	struct net_bridge_vlan_group *vg;
1209 	struct net_bridge_port *p = NULL;
1210 	struct net_bridge_vlan *v;
1211 	struct net_bridge *br = NULL;
1212 	u32 ext_flags = 0;
1213 	int err = 0;
1214 
1215 	trace_br_fdb_add(ndm, dev, addr, vid, nlh_flags);
1216 
1217 	if (!(ndm->ndm_state & (NUD_PERMANENT|NUD_NOARP|NUD_REACHABLE))) {
1218 		pr_info("bridge: RTM_NEWNEIGH with invalid state %#x\n", ndm->ndm_state);
1219 		return -EINVAL;
1220 	}
1221 
1222 	if (is_zero_ether_addr(addr)) {
1223 		pr_info("bridge: RTM_NEWNEIGH with invalid ether address\n");
1224 		return -EINVAL;
1225 	}
1226 
1227 	if (netif_is_bridge_master(dev)) {
1228 		br = netdev_priv(dev);
1229 		vg = br_vlan_group(br);
1230 	} else {
1231 		p = br_port_get_rtnl(dev);
1232 		if (!p) {
1233 			pr_info("bridge: RTM_NEWNEIGH %s not a bridge port\n",
1234 				dev->name);
1235 			return -EINVAL;
1236 		}
1237 		br = p->br;
1238 		vg = nbp_vlan_group(p);
1239 	}
1240 
1241 	if (tb[NDA_FLAGS_EXT])
1242 		ext_flags = nla_get_u32(tb[NDA_FLAGS_EXT]);
1243 
1244 	if (ext_flags & NTF_EXT_LOCKED) {
1245 		NL_SET_ERR_MSG_MOD(extack, "Cannot add FDB entry with \"locked\" flag set");
1246 		return -EINVAL;
1247 	}
1248 
1249 	if (tb[NDA_FDB_EXT_ATTRS]) {
1250 		attr = tb[NDA_FDB_EXT_ATTRS];
1251 		err = nla_parse_nested(nfea_tb, NFEA_MAX, attr,
1252 				       br_nda_fdb_pol, extack);
1253 		if (err)
1254 			return err;
1255 	} else {
1256 		memset(nfea_tb, 0, sizeof(struct nlattr *) * (NFEA_MAX + 1));
1257 	}
1258 
1259 	if (vid) {
1260 		v = br_vlan_find(vg, vid);
1261 		if (!v || !br_vlan_should_use(v)) {
1262 			pr_info("bridge: RTM_NEWNEIGH with unconfigured vlan %d on %s\n", vid, dev->name);
1263 			return -EINVAL;
1264 		}
1265 
1266 		/* VID was specified, so use it. */
1267 		err = __br_fdb_add(ndm, br, p, addr, nlh_flags, vid, nfea_tb,
1268 				   extack);
1269 	} else {
1270 		err = __br_fdb_add(ndm, br, p, addr, nlh_flags, 0, nfea_tb,
1271 				   extack);
1272 		if (err || !vg || !vg->num_vlans)
1273 			goto out;
1274 
1275 		/* We have vlans configured on this port and user didn't
1276 		 * specify a VLAN.  To be nice, add/update entry for every
1277 		 * vlan on this port.
1278 		 */
1279 		list_for_each_entry(v, &vg->vlan_list, vlist) {
1280 			if (!br_vlan_should_use(v))
1281 				continue;
1282 			err = __br_fdb_add(ndm, br, p, addr, nlh_flags, v->vid,
1283 					   nfea_tb, extack);
1284 			if (err)
1285 				goto out;
1286 		}
1287 	}
1288 
1289 out:
1290 	return err;
1291 }
1292 
1293 static int fdb_delete_by_addr_and_port(struct net_bridge *br,
1294 				       const struct net_bridge_port *p,
1295 				       const u8 *addr, u16 vlan)
1296 {
1297 	struct net_bridge_fdb_entry *fdb;
1298 
1299 	fdb = br_fdb_find(br, addr, vlan);
1300 	if (!fdb || READ_ONCE(fdb->dst) != p)
1301 		return -ENOENT;
1302 
1303 	fdb_delete(br, fdb, true);
1304 
1305 	return 0;
1306 }
1307 
1308 static int __br_fdb_delete(struct net_bridge *br,
1309 			   const struct net_bridge_port *p,
1310 			   const unsigned char *addr, u16 vid)
1311 {
1312 	int err;
1313 
1314 	spin_lock_bh(&br->hash_lock);
1315 	err = fdb_delete_by_addr_and_port(br, p, addr, vid);
1316 	spin_unlock_bh(&br->hash_lock);
1317 
1318 	return err;
1319 }
1320 
1321 /* Remove neighbor entry with RTM_DELNEIGH */
1322 int br_fdb_delete(struct ndmsg *ndm, struct nlattr *tb[],
1323 		  struct net_device *dev,
1324 		  const unsigned char *addr, u16 vid,
1325 		  struct netlink_ext_ack *extack)
1326 {
1327 	struct net_bridge_vlan_group *vg;
1328 	struct net_bridge_port *p = NULL;
1329 	struct net_bridge_vlan *v;
1330 	struct net_bridge *br;
1331 	int err;
1332 
1333 	if (netif_is_bridge_master(dev)) {
1334 		br = netdev_priv(dev);
1335 		vg = br_vlan_group(br);
1336 	} else {
1337 		p = br_port_get_rtnl(dev);
1338 		if (!p) {
1339 			pr_info("bridge: RTM_DELNEIGH %s not a bridge port\n",
1340 				dev->name);
1341 			return -EINVAL;
1342 		}
1343 		vg = nbp_vlan_group(p);
1344 		br = p->br;
1345 	}
1346 
1347 	if (vid) {
1348 		v = br_vlan_find(vg, vid);
1349 		if (!v) {
1350 			pr_info("bridge: RTM_DELNEIGH with unconfigured vlan %d on %s\n", vid, dev->name);
1351 			return -EINVAL;
1352 		}
1353 
1354 		err = __br_fdb_delete(br, p, addr, vid);
1355 	} else {
1356 		err = -ENOENT;
1357 		err &= __br_fdb_delete(br, p, addr, 0);
1358 		if (!vg || !vg->num_vlans)
1359 			return err;
1360 
1361 		list_for_each_entry(v, &vg->vlan_list, vlist) {
1362 			if (!br_vlan_should_use(v))
1363 				continue;
1364 			err &= __br_fdb_delete(br, p, addr, v->vid);
1365 		}
1366 	}
1367 
1368 	return err;
1369 }
1370 
1371 int br_fdb_sync_static(struct net_bridge *br, struct net_bridge_port *p)
1372 {
1373 	struct net_bridge_fdb_entry *f, *tmp;
1374 	int err = 0;
1375 
1376 	ASSERT_RTNL();
1377 
1378 	/* the key here is that static entries change only under rtnl */
1379 	rcu_read_lock();
1380 	hlist_for_each_entry_rcu(f, &br->fdb_list, fdb_node) {
1381 		/* We only care for static entries */
1382 		if (!test_bit(BR_FDB_STATIC, &f->flags))
1383 			continue;
1384 		err = dev_uc_add(p->dev, f->key.addr.addr);
1385 		if (err)
1386 			goto rollback;
1387 	}
1388 done:
1389 	rcu_read_unlock();
1390 
1391 	return err;
1392 
1393 rollback:
1394 	hlist_for_each_entry_rcu(tmp, &br->fdb_list, fdb_node) {
1395 		/* We only care for static entries */
1396 		if (!test_bit(BR_FDB_STATIC, &tmp->flags))
1397 			continue;
1398 		if (tmp == f)
1399 			break;
1400 		dev_uc_del(p->dev, tmp->key.addr.addr);
1401 	}
1402 
1403 	goto done;
1404 }
1405 
1406 void br_fdb_unsync_static(struct net_bridge *br, struct net_bridge_port *p)
1407 {
1408 	struct net_bridge_fdb_entry *f;
1409 
1410 	ASSERT_RTNL();
1411 
1412 	rcu_read_lock();
1413 	hlist_for_each_entry_rcu(f, &br->fdb_list, fdb_node) {
1414 		/* We only care for static entries */
1415 		if (!test_bit(BR_FDB_STATIC, &f->flags))
1416 			continue;
1417 
1418 		dev_uc_del(p->dev, f->key.addr.addr);
1419 	}
1420 	rcu_read_unlock();
1421 }
1422 
1423 int br_fdb_external_learn_add(struct net_bridge *br, struct net_bridge_port *p,
1424 			      const unsigned char *addr, u16 vid, bool locked,
1425 			      bool swdev_notify)
1426 {
1427 	struct net_bridge_fdb_entry *fdb;
1428 	bool modified = false;
1429 	int err = 0;
1430 
1431 	trace_br_fdb_external_learn_add(br, p, addr, vid);
1432 
1433 	if (locked && (!p || !(p->flags & BR_PORT_MAB)))
1434 		return -EINVAL;
1435 
1436 	spin_lock_bh(&br->hash_lock);
1437 
1438 	fdb = br_fdb_find(br, addr, vid);
1439 	if (!fdb) {
1440 		unsigned long flags = BIT(BR_FDB_ADDED_BY_EXT_LEARN);
1441 
1442 		if (swdev_notify)
1443 			flags |= BIT(BR_FDB_ADDED_BY_USER);
1444 
1445 		if (!p)
1446 			flags |= BIT(BR_FDB_LOCAL);
1447 
1448 		if (locked)
1449 			flags |= BIT(BR_FDB_LOCKED);
1450 
1451 		fdb = fdb_create(br, p, addr, vid, flags);
1452 		if (!fdb) {
1453 			err = -ENOMEM;
1454 			goto err_unlock;
1455 		}
1456 		fdb_notify(br, fdb, RTM_NEWNEIGH, swdev_notify);
1457 	} else {
1458 		if (locked &&
1459 		    (!test_bit(BR_FDB_LOCKED, &fdb->flags) ||
1460 		     READ_ONCE(fdb->dst) != p)) {
1461 			err = -EINVAL;
1462 			goto err_unlock;
1463 		}
1464 
1465 		fdb->updated = jiffies;
1466 
1467 		if (READ_ONCE(fdb->dst) != p) {
1468 			WRITE_ONCE(fdb->dst, p);
1469 			modified = true;
1470 		}
1471 
1472 		if (test_and_set_bit(BR_FDB_ADDED_BY_EXT_LEARN, &fdb->flags)) {
1473 			/* Refresh entry */
1474 			fdb->used = jiffies;
1475 		} else {
1476 			modified = true;
1477 		}
1478 
1479 		if (locked != test_bit(BR_FDB_LOCKED, &fdb->flags)) {
1480 			change_bit(BR_FDB_LOCKED, &fdb->flags);
1481 			modified = true;
1482 		}
1483 
1484 		if (swdev_notify)
1485 			set_bit(BR_FDB_ADDED_BY_USER, &fdb->flags);
1486 
1487 		if (!p)
1488 			set_bit(BR_FDB_LOCAL, &fdb->flags);
1489 
1490 		if ((swdev_notify || !p) &&
1491 		    test_and_clear_bit(BR_FDB_DYNAMIC_LEARNED, &fdb->flags))
1492 			atomic_dec(&br->fdb_n_learned);
1493 
1494 		if (modified)
1495 			fdb_notify(br, fdb, RTM_NEWNEIGH, swdev_notify);
1496 	}
1497 
1498 err_unlock:
1499 	spin_unlock_bh(&br->hash_lock);
1500 
1501 	return err;
1502 }
1503 
1504 int br_fdb_external_learn_del(struct net_bridge *br, struct net_bridge_port *p,
1505 			      const unsigned char *addr, u16 vid,
1506 			      bool swdev_notify)
1507 {
1508 	struct net_bridge_fdb_entry *fdb;
1509 	int err = 0;
1510 
1511 	spin_lock_bh(&br->hash_lock);
1512 
1513 	fdb = br_fdb_find(br, addr, vid);
1514 	if (fdb && test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &fdb->flags))
1515 		fdb_delete(br, fdb, swdev_notify);
1516 	else
1517 		err = -ENOENT;
1518 
1519 	spin_unlock_bh(&br->hash_lock);
1520 
1521 	return err;
1522 }
1523 
1524 void br_fdb_offloaded_set(struct net_bridge *br, struct net_bridge_port *p,
1525 			  const unsigned char *addr, u16 vid, bool offloaded)
1526 {
1527 	struct net_bridge_fdb_entry *fdb;
1528 
1529 	spin_lock_bh(&br->hash_lock);
1530 
1531 	fdb = br_fdb_find(br, addr, vid);
1532 	if (fdb && offloaded != test_bit(BR_FDB_OFFLOADED, &fdb->flags))
1533 		change_bit(BR_FDB_OFFLOADED, &fdb->flags);
1534 
1535 	spin_unlock_bh(&br->hash_lock);
1536 }
1537 
1538 void br_fdb_clear_offload(const struct net_device *dev, u16 vid)
1539 {
1540 	struct net_bridge_fdb_entry *f;
1541 	struct net_bridge_port *p;
1542 
1543 	ASSERT_RTNL();
1544 
1545 	p = br_port_get_rtnl(dev);
1546 	if (!p)
1547 		return;
1548 
1549 	spin_lock_bh(&p->br->hash_lock);
1550 	hlist_for_each_entry(f, &p->br->fdb_list, fdb_node) {
1551 		if (f->dst == p && f->key.vlan_id == vid)
1552 			clear_bit(BR_FDB_OFFLOADED, &f->flags);
1553 	}
1554 	spin_unlock_bh(&p->br->hash_lock);
1555 }
1556 EXPORT_SYMBOL_GPL(br_fdb_clear_offload);
1557