xref: /linux/net/bridge/br_mdb.c (revision be58f7103700a68d5c7ca60a2bc0b309907599ab)
1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/err.h>
3 #include <linux/igmp.h>
4 #include <linux/kernel.h>
5 #include <linux/netdevice.h>
6 #include <linux/rculist.h>
7 #include <linux/skbuff.h>
8 #include <linux/if_ether.h>
9 #include <net/ip.h>
10 #include <net/netlink.h>
11 #include <net/switchdev.h>
12 #if IS_ENABLED(CONFIG_IPV6)
13 #include <net/ipv6.h>
14 #include <net/addrconf.h>
15 #endif
16 
17 #include "br_private.h"
18 
19 static bool
20 br_ip4_rports_get_timer(struct net_bridge_mcast_port *pmctx,
21 			unsigned long *timer)
22 {
23 	*timer = br_timer_value(&pmctx->ip4_mc_router_timer);
24 	return !hlist_unhashed(&pmctx->ip4_rlist);
25 }
26 
27 static bool
28 br_ip6_rports_get_timer(struct net_bridge_mcast_port *pmctx,
29 			unsigned long *timer)
30 {
31 #if IS_ENABLED(CONFIG_IPV6)
32 	*timer = br_timer_value(&pmctx->ip6_mc_router_timer);
33 	return !hlist_unhashed(&pmctx->ip6_rlist);
34 #else
35 	*timer = 0;
36 	return false;
37 #endif
38 }
39 
40 static size_t __br_rports_one_size(void)
41 {
42 	return nla_total_size(sizeof(u32)) + /* MDBA_ROUTER_PORT */
43 	       nla_total_size(sizeof(u32)) + /* MDBA_ROUTER_PATTR_TIMER */
44 	       nla_total_size(sizeof(u8)) +  /* MDBA_ROUTER_PATTR_TYPE */
45 	       nla_total_size(sizeof(u32)) + /* MDBA_ROUTER_PATTR_INET_TIMER */
46 	       nla_total_size(sizeof(u32)) + /* MDBA_ROUTER_PATTR_INET6_TIMER */
47 	       nla_total_size(sizeof(u32));  /* MDBA_ROUTER_PATTR_VID */
48 }
49 
50 size_t br_rports_size(const struct net_bridge_mcast *brmctx)
51 {
52 	struct net_bridge_mcast_port *pmctx;
53 	size_t size = nla_total_size(0); /* MDBA_ROUTER */
54 
55 	rcu_read_lock();
56 	hlist_for_each_entry_rcu(pmctx, &brmctx->ip4_mc_router_list,
57 				 ip4_rlist)
58 		size += __br_rports_one_size();
59 
60 #if IS_ENABLED(CONFIG_IPV6)
61 	hlist_for_each_entry_rcu(pmctx, &brmctx->ip6_mc_router_list,
62 				 ip6_rlist)
63 		size += __br_rports_one_size();
64 #endif
65 	rcu_read_unlock();
66 
67 	return size;
68 }
69 
70 int br_rports_fill_info(struct sk_buff *skb,
71 			const struct net_bridge_mcast *brmctx)
72 {
73 	u16 vid = brmctx->vlan ? brmctx->vlan->vid : 0;
74 	bool have_ip4_mc_rtr, have_ip6_mc_rtr;
75 	unsigned long ip4_timer, ip6_timer;
76 	struct nlattr *nest, *port_nest;
77 	struct net_bridge_port *p;
78 
79 	if (!brmctx->multicast_router || !br_rports_have_mc_router(brmctx))
80 		return 0;
81 
82 	nest = nla_nest_start_noflag(skb, MDBA_ROUTER);
83 	if (nest == NULL)
84 		return -EMSGSIZE;
85 
86 	list_for_each_entry_rcu(p, &brmctx->br->port_list, list) {
87 		struct net_bridge_mcast_port *pmctx;
88 
89 		if (vid) {
90 			struct net_bridge_vlan *v;
91 
92 			v = br_vlan_find(nbp_vlan_group(p), vid);
93 			if (!v)
94 				continue;
95 			pmctx = &v->port_mcast_ctx;
96 		} else {
97 			pmctx = &p->multicast_ctx;
98 		}
99 
100 		have_ip4_mc_rtr = br_ip4_rports_get_timer(pmctx, &ip4_timer);
101 		have_ip6_mc_rtr = br_ip6_rports_get_timer(pmctx, &ip6_timer);
102 
103 		if (!have_ip4_mc_rtr && !have_ip6_mc_rtr)
104 			continue;
105 
106 		port_nest = nla_nest_start_noflag(skb, MDBA_ROUTER_PORT);
107 		if (!port_nest)
108 			goto fail;
109 
110 		if (nla_put_nohdr(skb, sizeof(u32), &p->dev->ifindex) ||
111 		    nla_put_u32(skb, MDBA_ROUTER_PATTR_TIMER,
112 				max(ip4_timer, ip6_timer)) ||
113 		    nla_put_u8(skb, MDBA_ROUTER_PATTR_TYPE,
114 			       p->multicast_ctx.multicast_router) ||
115 		    (have_ip4_mc_rtr &&
116 		     nla_put_u32(skb, MDBA_ROUTER_PATTR_INET_TIMER,
117 				 ip4_timer)) ||
118 		    (have_ip6_mc_rtr &&
119 		     nla_put_u32(skb, MDBA_ROUTER_PATTR_INET6_TIMER,
120 				 ip6_timer)) ||
121 		    (vid && nla_put_u16(skb, MDBA_ROUTER_PATTR_VID, vid))) {
122 			nla_nest_cancel(skb, port_nest);
123 			goto fail;
124 		}
125 		nla_nest_end(skb, port_nest);
126 	}
127 
128 	nla_nest_end(skb, nest);
129 	return 0;
130 fail:
131 	nla_nest_cancel(skb, nest);
132 	return -EMSGSIZE;
133 }
134 
135 static void __mdb_entry_fill_flags(struct br_mdb_entry *e, unsigned char flags)
136 {
137 	e->state = flags & MDB_PG_FLAGS_PERMANENT;
138 	e->flags = 0;
139 	if (flags & MDB_PG_FLAGS_OFFLOAD)
140 		e->flags |= MDB_FLAGS_OFFLOAD;
141 	if (flags & MDB_PG_FLAGS_FAST_LEAVE)
142 		e->flags |= MDB_FLAGS_FAST_LEAVE;
143 	if (flags & MDB_PG_FLAGS_STAR_EXCL)
144 		e->flags |= MDB_FLAGS_STAR_EXCL;
145 	if (flags & MDB_PG_FLAGS_BLOCKED)
146 		e->flags |= MDB_FLAGS_BLOCKED;
147 }
148 
149 static void __mdb_entry_to_br_ip(struct br_mdb_entry *entry, struct br_ip *ip,
150 				 struct nlattr **mdb_attrs)
151 {
152 	memset(ip, 0, sizeof(struct br_ip));
153 	ip->vid = entry->vid;
154 	ip->proto = entry->addr.proto;
155 	switch (ip->proto) {
156 	case htons(ETH_P_IP):
157 		ip->dst.ip4 = entry->addr.u.ip4;
158 		if (mdb_attrs && mdb_attrs[MDBE_ATTR_SOURCE])
159 			ip->src.ip4 = nla_get_in_addr(mdb_attrs[MDBE_ATTR_SOURCE]);
160 		break;
161 #if IS_ENABLED(CONFIG_IPV6)
162 	case htons(ETH_P_IPV6):
163 		ip->dst.ip6 = entry->addr.u.ip6;
164 		if (mdb_attrs && mdb_attrs[MDBE_ATTR_SOURCE])
165 			ip->src.ip6 = nla_get_in6_addr(mdb_attrs[MDBE_ATTR_SOURCE]);
166 		break;
167 #endif
168 	default:
169 		ether_addr_copy(ip->dst.mac_addr, entry->addr.u.mac_addr);
170 	}
171 
172 }
173 
174 static int __mdb_fill_srcs(struct sk_buff *skb,
175 			   struct net_bridge_port_group *p)
176 {
177 	struct net_bridge_group_src *ent;
178 	struct nlattr *nest, *nest_ent;
179 
180 	if (hlist_empty(&p->src_list))
181 		return 0;
182 
183 	nest = nla_nest_start(skb, MDBA_MDB_EATTR_SRC_LIST);
184 	if (!nest)
185 		return -EMSGSIZE;
186 
187 	hlist_for_each_entry_rcu(ent, &p->src_list, node,
188 				 lockdep_is_held(&p->key.port->br->multicast_lock)) {
189 		nest_ent = nla_nest_start(skb, MDBA_MDB_SRCLIST_ENTRY);
190 		if (!nest_ent)
191 			goto out_cancel_err;
192 		switch (ent->addr.proto) {
193 		case htons(ETH_P_IP):
194 			if (nla_put_in_addr(skb, MDBA_MDB_SRCATTR_ADDRESS,
195 					    ent->addr.src.ip4)) {
196 				nla_nest_cancel(skb, nest_ent);
197 				goto out_cancel_err;
198 			}
199 			break;
200 #if IS_ENABLED(CONFIG_IPV6)
201 		case htons(ETH_P_IPV6):
202 			if (nla_put_in6_addr(skb, MDBA_MDB_SRCATTR_ADDRESS,
203 					     &ent->addr.src.ip6)) {
204 				nla_nest_cancel(skb, nest_ent);
205 				goto out_cancel_err;
206 			}
207 			break;
208 #endif
209 		default:
210 			nla_nest_cancel(skb, nest_ent);
211 			continue;
212 		}
213 		if (nla_put_u32(skb, MDBA_MDB_SRCATTR_TIMER,
214 				br_timer_value(&ent->timer))) {
215 			nla_nest_cancel(skb, nest_ent);
216 			goto out_cancel_err;
217 		}
218 		nla_nest_end(skb, nest_ent);
219 	}
220 
221 	nla_nest_end(skb, nest);
222 
223 	return 0;
224 
225 out_cancel_err:
226 	nla_nest_cancel(skb, nest);
227 	return -EMSGSIZE;
228 }
229 
230 static int __mdb_fill_info(struct sk_buff *skb,
231 			   struct net_bridge_mdb_entry *mp,
232 			   struct net_bridge_port_group *p)
233 {
234 	bool dump_srcs_mode = false;
235 	struct timer_list *mtimer;
236 	struct nlattr *nest_ent;
237 	struct br_mdb_entry e;
238 	u8 flags = 0;
239 	int ifindex;
240 
241 	memset(&e, 0, sizeof(e));
242 	if (p) {
243 		ifindex = p->key.port->dev->ifindex;
244 		mtimer = &p->timer;
245 		flags = p->flags;
246 	} else {
247 		ifindex = mp->br->dev->ifindex;
248 		mtimer = &mp->timer;
249 	}
250 
251 	__mdb_entry_fill_flags(&e, flags);
252 	e.ifindex = ifindex;
253 	e.vid = mp->addr.vid;
254 	if (mp->addr.proto == htons(ETH_P_IP))
255 		e.addr.u.ip4 = mp->addr.dst.ip4;
256 #if IS_ENABLED(CONFIG_IPV6)
257 	else if (mp->addr.proto == htons(ETH_P_IPV6))
258 		e.addr.u.ip6 = mp->addr.dst.ip6;
259 #endif
260 	else
261 		ether_addr_copy(e.addr.u.mac_addr, mp->addr.dst.mac_addr);
262 	e.addr.proto = mp->addr.proto;
263 	nest_ent = nla_nest_start_noflag(skb,
264 					 MDBA_MDB_ENTRY_INFO);
265 	if (!nest_ent)
266 		return -EMSGSIZE;
267 
268 	if (nla_put_nohdr(skb, sizeof(e), &e) ||
269 	    nla_put_u32(skb,
270 			MDBA_MDB_EATTR_TIMER,
271 			br_timer_value(mtimer)))
272 		goto nest_err;
273 
274 	switch (mp->addr.proto) {
275 	case htons(ETH_P_IP):
276 		dump_srcs_mode = !!(mp->br->multicast_ctx.multicast_igmp_version == 3);
277 		if (mp->addr.src.ip4) {
278 			if (nla_put_in_addr(skb, MDBA_MDB_EATTR_SOURCE,
279 					    mp->addr.src.ip4))
280 				goto nest_err;
281 			break;
282 		}
283 		break;
284 #if IS_ENABLED(CONFIG_IPV6)
285 	case htons(ETH_P_IPV6):
286 		dump_srcs_mode = !!(mp->br->multicast_ctx.multicast_mld_version == 2);
287 		if (!ipv6_addr_any(&mp->addr.src.ip6)) {
288 			if (nla_put_in6_addr(skb, MDBA_MDB_EATTR_SOURCE,
289 					     &mp->addr.src.ip6))
290 				goto nest_err;
291 			break;
292 		}
293 		break;
294 #endif
295 	default:
296 		ether_addr_copy(e.addr.u.mac_addr, mp->addr.dst.mac_addr);
297 	}
298 	if (p) {
299 		if (nla_put_u8(skb, MDBA_MDB_EATTR_RTPROT, p->rt_protocol))
300 			goto nest_err;
301 		if (dump_srcs_mode &&
302 		    (__mdb_fill_srcs(skb, p) ||
303 		     nla_put_u8(skb, MDBA_MDB_EATTR_GROUP_MODE,
304 				p->filter_mode)))
305 			goto nest_err;
306 	}
307 	nla_nest_end(skb, nest_ent);
308 
309 	return 0;
310 
311 nest_err:
312 	nla_nest_cancel(skb, nest_ent);
313 	return -EMSGSIZE;
314 }
315 
316 static int br_mdb_fill_info(struct sk_buff *skb, struct netlink_callback *cb,
317 			    struct net_device *dev)
318 {
319 	int idx = 0, s_idx = cb->args[1], err = 0, pidx = 0, s_pidx = cb->args[2];
320 	struct net_bridge *br = netdev_priv(dev);
321 	struct net_bridge_mdb_entry *mp;
322 	struct nlattr *nest, *nest2;
323 
324 	if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
325 		return 0;
326 
327 	nest = nla_nest_start_noflag(skb, MDBA_MDB);
328 	if (nest == NULL)
329 		return -EMSGSIZE;
330 
331 	hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
332 		struct net_bridge_port_group *p;
333 		struct net_bridge_port_group __rcu **pp;
334 
335 		if (idx < s_idx)
336 			goto skip;
337 
338 		nest2 = nla_nest_start_noflag(skb, MDBA_MDB_ENTRY);
339 		if (!nest2) {
340 			err = -EMSGSIZE;
341 			break;
342 		}
343 
344 		if (!s_pidx && mp->host_joined) {
345 			err = __mdb_fill_info(skb, mp, NULL);
346 			if (err) {
347 				nla_nest_cancel(skb, nest2);
348 				break;
349 			}
350 		}
351 
352 		for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
353 		      pp = &p->next) {
354 			if (!p->key.port)
355 				continue;
356 			if (pidx < s_pidx)
357 				goto skip_pg;
358 
359 			err = __mdb_fill_info(skb, mp, p);
360 			if (err) {
361 				nla_nest_end(skb, nest2);
362 				goto out;
363 			}
364 skip_pg:
365 			pidx++;
366 		}
367 		pidx = 0;
368 		s_pidx = 0;
369 		nla_nest_end(skb, nest2);
370 skip:
371 		idx++;
372 	}
373 
374 out:
375 	cb->args[1] = idx;
376 	cb->args[2] = pidx;
377 	nla_nest_end(skb, nest);
378 	return err;
379 }
380 
381 static int br_mdb_valid_dump_req(const struct nlmsghdr *nlh,
382 				 struct netlink_ext_ack *extack)
383 {
384 	struct br_port_msg *bpm;
385 
386 	if (nlh->nlmsg_len < nlmsg_msg_size(sizeof(*bpm))) {
387 		NL_SET_ERR_MSG_MOD(extack, "Invalid header for mdb dump request");
388 		return -EINVAL;
389 	}
390 
391 	bpm = nlmsg_data(nlh);
392 	if (bpm->ifindex) {
393 		NL_SET_ERR_MSG_MOD(extack, "Filtering by device index is not supported for mdb dump request");
394 		return -EINVAL;
395 	}
396 	if (nlmsg_attrlen(nlh, sizeof(*bpm))) {
397 		NL_SET_ERR_MSG(extack, "Invalid data after header in mdb dump request");
398 		return -EINVAL;
399 	}
400 
401 	return 0;
402 }
403 
404 static int br_mdb_dump(struct sk_buff *skb, struct netlink_callback *cb)
405 {
406 	struct net_device *dev;
407 	struct net *net = sock_net(skb->sk);
408 	struct nlmsghdr *nlh = NULL;
409 	int idx = 0, s_idx;
410 
411 	if (cb->strict_check) {
412 		int err = br_mdb_valid_dump_req(cb->nlh, cb->extack);
413 
414 		if (err < 0)
415 			return err;
416 	}
417 
418 	s_idx = cb->args[0];
419 
420 	rcu_read_lock();
421 
422 	cb->seq = net->dev_base_seq;
423 
424 	for_each_netdev_rcu(net, dev) {
425 		if (dev->priv_flags & IFF_EBRIDGE) {
426 			struct net_bridge *br = netdev_priv(dev);
427 			struct br_port_msg *bpm;
428 
429 			if (idx < s_idx)
430 				goto skip;
431 
432 			nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid,
433 					cb->nlh->nlmsg_seq, RTM_GETMDB,
434 					sizeof(*bpm), NLM_F_MULTI);
435 			if (nlh == NULL)
436 				break;
437 
438 			bpm = nlmsg_data(nlh);
439 			memset(bpm, 0, sizeof(*bpm));
440 			bpm->ifindex = dev->ifindex;
441 			if (br_mdb_fill_info(skb, cb, dev) < 0)
442 				goto out;
443 			if (br_rports_fill_info(skb, &br->multicast_ctx) < 0)
444 				goto out;
445 
446 			cb->args[1] = 0;
447 			nlmsg_end(skb, nlh);
448 		skip:
449 			idx++;
450 		}
451 	}
452 
453 out:
454 	if (nlh)
455 		nlmsg_end(skb, nlh);
456 	rcu_read_unlock();
457 	cb->args[0] = idx;
458 	return skb->len;
459 }
460 
461 static int nlmsg_populate_mdb_fill(struct sk_buff *skb,
462 				   struct net_device *dev,
463 				   struct net_bridge_mdb_entry *mp,
464 				   struct net_bridge_port_group *pg,
465 				   int type)
466 {
467 	struct nlmsghdr *nlh;
468 	struct br_port_msg *bpm;
469 	struct nlattr *nest, *nest2;
470 
471 	nlh = nlmsg_put(skb, 0, 0, type, sizeof(*bpm), 0);
472 	if (!nlh)
473 		return -EMSGSIZE;
474 
475 	bpm = nlmsg_data(nlh);
476 	memset(bpm, 0, sizeof(*bpm));
477 	bpm->family  = AF_BRIDGE;
478 	bpm->ifindex = dev->ifindex;
479 	nest = nla_nest_start_noflag(skb, MDBA_MDB);
480 	if (nest == NULL)
481 		goto cancel;
482 	nest2 = nla_nest_start_noflag(skb, MDBA_MDB_ENTRY);
483 	if (nest2 == NULL)
484 		goto end;
485 
486 	if (__mdb_fill_info(skb, mp, pg))
487 		goto end;
488 
489 	nla_nest_end(skb, nest2);
490 	nla_nest_end(skb, nest);
491 	nlmsg_end(skb, nlh);
492 	return 0;
493 
494 end:
495 	nla_nest_end(skb, nest);
496 cancel:
497 	nlmsg_cancel(skb, nlh);
498 	return -EMSGSIZE;
499 }
500 
501 static size_t rtnl_mdb_nlmsg_size(struct net_bridge_port_group *pg)
502 {
503 	size_t nlmsg_size = NLMSG_ALIGN(sizeof(struct br_port_msg)) +
504 			    nla_total_size(sizeof(struct br_mdb_entry)) +
505 			    nla_total_size(sizeof(u32));
506 	struct net_bridge_group_src *ent;
507 	size_t addr_size = 0;
508 
509 	if (!pg)
510 		goto out;
511 
512 	/* MDBA_MDB_EATTR_RTPROT */
513 	nlmsg_size += nla_total_size(sizeof(u8));
514 
515 	switch (pg->key.addr.proto) {
516 	case htons(ETH_P_IP):
517 		/* MDBA_MDB_EATTR_SOURCE */
518 		if (pg->key.addr.src.ip4)
519 			nlmsg_size += nla_total_size(sizeof(__be32));
520 		if (pg->key.port->br->multicast_ctx.multicast_igmp_version == 2)
521 			goto out;
522 		addr_size = sizeof(__be32);
523 		break;
524 #if IS_ENABLED(CONFIG_IPV6)
525 	case htons(ETH_P_IPV6):
526 		/* MDBA_MDB_EATTR_SOURCE */
527 		if (!ipv6_addr_any(&pg->key.addr.src.ip6))
528 			nlmsg_size += nla_total_size(sizeof(struct in6_addr));
529 		if (pg->key.port->br->multicast_ctx.multicast_mld_version == 1)
530 			goto out;
531 		addr_size = sizeof(struct in6_addr);
532 		break;
533 #endif
534 	}
535 
536 	/* MDBA_MDB_EATTR_GROUP_MODE */
537 	nlmsg_size += nla_total_size(sizeof(u8));
538 
539 	/* MDBA_MDB_EATTR_SRC_LIST nested attr */
540 	if (!hlist_empty(&pg->src_list))
541 		nlmsg_size += nla_total_size(0);
542 
543 	hlist_for_each_entry(ent, &pg->src_list, node) {
544 		/* MDBA_MDB_SRCLIST_ENTRY nested attr +
545 		 * MDBA_MDB_SRCATTR_ADDRESS + MDBA_MDB_SRCATTR_TIMER
546 		 */
547 		nlmsg_size += nla_total_size(0) +
548 			      nla_total_size(addr_size) +
549 			      nla_total_size(sizeof(u32));
550 	}
551 out:
552 	return nlmsg_size;
553 }
554 
555 struct br_mdb_complete_info {
556 	struct net_bridge_port *port;
557 	struct br_ip ip;
558 };
559 
560 static void br_mdb_complete(struct net_device *dev, int err, void *priv)
561 {
562 	struct br_mdb_complete_info *data = priv;
563 	struct net_bridge_port_group __rcu **pp;
564 	struct net_bridge_port_group *p;
565 	struct net_bridge_mdb_entry *mp;
566 	struct net_bridge_port *port = data->port;
567 	struct net_bridge *br = port->br;
568 
569 	if (err)
570 		goto err;
571 
572 	spin_lock_bh(&br->multicast_lock);
573 	mp = br_mdb_ip_get(br, &data->ip);
574 	if (!mp)
575 		goto out;
576 	for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
577 	     pp = &p->next) {
578 		if (p->key.port != port)
579 			continue;
580 		p->flags |= MDB_PG_FLAGS_OFFLOAD;
581 	}
582 out:
583 	spin_unlock_bh(&br->multicast_lock);
584 err:
585 	kfree(priv);
586 }
587 
588 static void br_switchdev_mdb_populate(struct switchdev_obj_port_mdb *mdb,
589 				      const struct net_bridge_mdb_entry *mp)
590 {
591 	if (mp->addr.proto == htons(ETH_P_IP))
592 		ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
593 #if IS_ENABLED(CONFIG_IPV6)
594 	else if (mp->addr.proto == htons(ETH_P_IPV6))
595 		ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
596 #endif
597 	else
598 		ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
599 
600 	mdb->vid = mp->addr.vid;
601 }
602 
603 static int br_mdb_replay_one(struct notifier_block *nb, struct net_device *dev,
604 			     const struct switchdev_obj_port_mdb *mdb,
605 			     unsigned long action, const void *ctx,
606 			     struct netlink_ext_ack *extack)
607 {
608 	struct switchdev_notifier_port_obj_info obj_info = {
609 		.info = {
610 			.dev = dev,
611 			.extack = extack,
612 			.ctx = ctx,
613 		},
614 		.obj = &mdb->obj,
615 	};
616 	int err;
617 
618 	err = nb->notifier_call(nb, action, &obj_info);
619 	return notifier_to_errno(err);
620 }
621 
622 static int br_mdb_queue_one(struct list_head *mdb_list,
623 			    enum switchdev_obj_id id,
624 			    const struct net_bridge_mdb_entry *mp,
625 			    struct net_device *orig_dev)
626 {
627 	struct switchdev_obj_port_mdb *mdb;
628 
629 	mdb = kzalloc(sizeof(*mdb), GFP_ATOMIC);
630 	if (!mdb)
631 		return -ENOMEM;
632 
633 	mdb->obj.id = id;
634 	mdb->obj.orig_dev = orig_dev;
635 	br_switchdev_mdb_populate(mdb, mp);
636 	list_add_tail(&mdb->obj.list, mdb_list);
637 
638 	return 0;
639 }
640 
641 int br_mdb_replay(struct net_device *br_dev, struct net_device *dev,
642 		  const void *ctx, bool adding, struct notifier_block *nb,
643 		  struct netlink_ext_ack *extack)
644 {
645 	const struct net_bridge_mdb_entry *mp;
646 	struct switchdev_obj *obj, *tmp;
647 	struct net_bridge *br;
648 	unsigned long action;
649 	LIST_HEAD(mdb_list);
650 	int err = 0;
651 
652 	ASSERT_RTNL();
653 
654 	if (!nb)
655 		return 0;
656 
657 	if (!netif_is_bridge_master(br_dev) || !netif_is_bridge_port(dev))
658 		return -EINVAL;
659 
660 	br = netdev_priv(br_dev);
661 
662 	if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
663 		return 0;
664 
665 	/* We cannot walk over br->mdb_list protected just by the rtnl_mutex,
666 	 * because the write-side protection is br->multicast_lock. But we
667 	 * need to emulate the [ blocking ] calling context of a regular
668 	 * switchdev event, so since both br->multicast_lock and RCU read side
669 	 * critical sections are atomic, we have no choice but to pick the RCU
670 	 * read side lock, queue up all our events, leave the critical section
671 	 * and notify switchdev from blocking context.
672 	 */
673 	rcu_read_lock();
674 
675 	hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
676 		struct net_bridge_port_group __rcu * const *pp;
677 		const struct net_bridge_port_group *p;
678 
679 		if (mp->host_joined) {
680 			err = br_mdb_queue_one(&mdb_list,
681 					       SWITCHDEV_OBJ_ID_HOST_MDB,
682 					       mp, br_dev);
683 			if (err) {
684 				rcu_read_unlock();
685 				goto out_free_mdb;
686 			}
687 		}
688 
689 		for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
690 		     pp = &p->next) {
691 			if (p->key.port->dev != dev)
692 				continue;
693 
694 			err = br_mdb_queue_one(&mdb_list,
695 					       SWITCHDEV_OBJ_ID_PORT_MDB,
696 					       mp, dev);
697 			if (err) {
698 				rcu_read_unlock();
699 				goto out_free_mdb;
700 			}
701 		}
702 	}
703 
704 	rcu_read_unlock();
705 
706 	if (adding)
707 		action = SWITCHDEV_PORT_OBJ_ADD;
708 	else
709 		action = SWITCHDEV_PORT_OBJ_DEL;
710 
711 	list_for_each_entry(obj, &mdb_list, list) {
712 		err = br_mdb_replay_one(nb, dev, SWITCHDEV_OBJ_PORT_MDB(obj),
713 					action, ctx, extack);
714 		if (err)
715 			goto out_free_mdb;
716 	}
717 
718 out_free_mdb:
719 	list_for_each_entry_safe(obj, tmp, &mdb_list, list) {
720 		list_del(&obj->list);
721 		kfree(SWITCHDEV_OBJ_PORT_MDB(obj));
722 	}
723 
724 	return err;
725 }
726 
727 static void br_mdb_switchdev_host_port(struct net_device *dev,
728 				       struct net_device *lower_dev,
729 				       struct net_bridge_mdb_entry *mp,
730 				       int type)
731 {
732 	struct switchdev_obj_port_mdb mdb = {
733 		.obj = {
734 			.id = SWITCHDEV_OBJ_ID_HOST_MDB,
735 			.flags = SWITCHDEV_F_DEFER,
736 			.orig_dev = dev,
737 		},
738 	};
739 
740 	br_switchdev_mdb_populate(&mdb, mp);
741 
742 	switch (type) {
743 	case RTM_NEWMDB:
744 		switchdev_port_obj_add(lower_dev, &mdb.obj, NULL);
745 		break;
746 	case RTM_DELMDB:
747 		switchdev_port_obj_del(lower_dev, &mdb.obj);
748 		break;
749 	}
750 }
751 
752 static void br_mdb_switchdev_host(struct net_device *dev,
753 				  struct net_bridge_mdb_entry *mp, int type)
754 {
755 	struct net_device *lower_dev;
756 	struct list_head *iter;
757 
758 	netdev_for_each_lower_dev(dev, lower_dev, iter)
759 		br_mdb_switchdev_host_port(dev, lower_dev, mp, type);
760 }
761 
762 void br_mdb_notify(struct net_device *dev,
763 		   struct net_bridge_mdb_entry *mp,
764 		   struct net_bridge_port_group *pg,
765 		   int type)
766 {
767 	struct br_mdb_complete_info *complete_info;
768 	struct switchdev_obj_port_mdb mdb = {
769 		.obj = {
770 			.id = SWITCHDEV_OBJ_ID_PORT_MDB,
771 			.flags = SWITCHDEV_F_DEFER,
772 		},
773 	};
774 	struct net *net = dev_net(dev);
775 	struct sk_buff *skb;
776 	int err = -ENOBUFS;
777 
778 	if (pg) {
779 		br_switchdev_mdb_populate(&mdb, mp);
780 
781 		mdb.obj.orig_dev = pg->key.port->dev;
782 		switch (type) {
783 		case RTM_NEWMDB:
784 			complete_info = kmalloc(sizeof(*complete_info), GFP_ATOMIC);
785 			if (!complete_info)
786 				break;
787 			complete_info->port = pg->key.port;
788 			complete_info->ip = mp->addr;
789 			mdb.obj.complete_priv = complete_info;
790 			mdb.obj.complete = br_mdb_complete;
791 			if (switchdev_port_obj_add(pg->key.port->dev, &mdb.obj, NULL))
792 				kfree(complete_info);
793 			break;
794 		case RTM_DELMDB:
795 			switchdev_port_obj_del(pg->key.port->dev, &mdb.obj);
796 			break;
797 		}
798 	} else {
799 		br_mdb_switchdev_host(dev, mp, type);
800 	}
801 
802 	skb = nlmsg_new(rtnl_mdb_nlmsg_size(pg), GFP_ATOMIC);
803 	if (!skb)
804 		goto errout;
805 
806 	err = nlmsg_populate_mdb_fill(skb, dev, mp, pg, type);
807 	if (err < 0) {
808 		kfree_skb(skb);
809 		goto errout;
810 	}
811 
812 	rtnl_notify(skb, net, 0, RTNLGRP_MDB, NULL, GFP_ATOMIC);
813 	return;
814 errout:
815 	rtnl_set_sk_err(net, RTNLGRP_MDB, err);
816 }
817 
818 static int nlmsg_populate_rtr_fill(struct sk_buff *skb,
819 				   struct net_device *dev,
820 				   int ifindex, u16 vid, u32 pid,
821 				   u32 seq, int type, unsigned int flags)
822 {
823 	struct nlattr *nest, *port_nest;
824 	struct br_port_msg *bpm;
825 	struct nlmsghdr *nlh;
826 
827 	nlh = nlmsg_put(skb, pid, seq, type, sizeof(*bpm), 0);
828 	if (!nlh)
829 		return -EMSGSIZE;
830 
831 	bpm = nlmsg_data(nlh);
832 	memset(bpm, 0, sizeof(*bpm));
833 	bpm->family = AF_BRIDGE;
834 	bpm->ifindex = dev->ifindex;
835 	nest = nla_nest_start_noflag(skb, MDBA_ROUTER);
836 	if (!nest)
837 		goto cancel;
838 
839 	port_nest = nla_nest_start_noflag(skb, MDBA_ROUTER_PORT);
840 	if (!port_nest)
841 		goto end;
842 	if (nla_put_nohdr(skb, sizeof(u32), &ifindex)) {
843 		nla_nest_cancel(skb, port_nest);
844 		goto end;
845 	}
846 	if (vid && nla_put_u16(skb, MDBA_ROUTER_PATTR_VID, vid)) {
847 		nla_nest_cancel(skb, port_nest);
848 		goto end;
849 	}
850 	nla_nest_end(skb, port_nest);
851 
852 	nla_nest_end(skb, nest);
853 	nlmsg_end(skb, nlh);
854 	return 0;
855 
856 end:
857 	nla_nest_end(skb, nest);
858 cancel:
859 	nlmsg_cancel(skb, nlh);
860 	return -EMSGSIZE;
861 }
862 
863 static inline size_t rtnl_rtr_nlmsg_size(void)
864 {
865 	return NLMSG_ALIGN(sizeof(struct br_port_msg))
866 		+ nla_total_size(sizeof(__u32))
867 		+ nla_total_size(sizeof(u16));
868 }
869 
870 void br_rtr_notify(struct net_device *dev, struct net_bridge_mcast_port *pmctx,
871 		   int type)
872 {
873 	struct net *net = dev_net(dev);
874 	struct sk_buff *skb;
875 	int err = -ENOBUFS;
876 	int ifindex;
877 	u16 vid;
878 
879 	ifindex = pmctx ? pmctx->port->dev->ifindex : 0;
880 	vid = pmctx && br_multicast_port_ctx_is_vlan(pmctx) ? pmctx->vlan->vid :
881 							      0;
882 	skb = nlmsg_new(rtnl_rtr_nlmsg_size(), GFP_ATOMIC);
883 	if (!skb)
884 		goto errout;
885 
886 	err = nlmsg_populate_rtr_fill(skb, dev, ifindex, vid, 0, 0, type,
887 				      NTF_SELF);
888 	if (err < 0) {
889 		kfree_skb(skb);
890 		goto errout;
891 	}
892 
893 	rtnl_notify(skb, net, 0, RTNLGRP_MDB, NULL, GFP_ATOMIC);
894 	return;
895 
896 errout:
897 	rtnl_set_sk_err(net, RTNLGRP_MDB, err);
898 }
899 
900 static bool is_valid_mdb_entry(struct br_mdb_entry *entry,
901 			       struct netlink_ext_ack *extack)
902 {
903 	if (entry->ifindex == 0) {
904 		NL_SET_ERR_MSG_MOD(extack, "Zero entry ifindex is not allowed");
905 		return false;
906 	}
907 
908 	if (entry->addr.proto == htons(ETH_P_IP)) {
909 		if (!ipv4_is_multicast(entry->addr.u.ip4)) {
910 			NL_SET_ERR_MSG_MOD(extack, "IPv4 entry group address is not multicast");
911 			return false;
912 		}
913 		if (ipv4_is_local_multicast(entry->addr.u.ip4)) {
914 			NL_SET_ERR_MSG_MOD(extack, "IPv4 entry group address is local multicast");
915 			return false;
916 		}
917 #if IS_ENABLED(CONFIG_IPV6)
918 	} else if (entry->addr.proto == htons(ETH_P_IPV6)) {
919 		if (ipv6_addr_is_ll_all_nodes(&entry->addr.u.ip6)) {
920 			NL_SET_ERR_MSG_MOD(extack, "IPv6 entry group address is link-local all nodes");
921 			return false;
922 		}
923 #endif
924 	} else if (entry->addr.proto == 0) {
925 		/* L2 mdb */
926 		if (!is_multicast_ether_addr(entry->addr.u.mac_addr)) {
927 			NL_SET_ERR_MSG_MOD(extack, "L2 entry group is not multicast");
928 			return false;
929 		}
930 	} else {
931 		NL_SET_ERR_MSG_MOD(extack, "Unknown entry protocol");
932 		return false;
933 	}
934 
935 	if (entry->state != MDB_PERMANENT && entry->state != MDB_TEMPORARY) {
936 		NL_SET_ERR_MSG_MOD(extack, "Unknown entry state");
937 		return false;
938 	}
939 	if (entry->vid >= VLAN_VID_MASK) {
940 		NL_SET_ERR_MSG_MOD(extack, "Invalid entry VLAN id");
941 		return false;
942 	}
943 
944 	return true;
945 }
946 
947 static bool is_valid_mdb_source(struct nlattr *attr, __be16 proto,
948 				struct netlink_ext_ack *extack)
949 {
950 	switch (proto) {
951 	case htons(ETH_P_IP):
952 		if (nla_len(attr) != sizeof(struct in_addr)) {
953 			NL_SET_ERR_MSG_MOD(extack, "IPv4 invalid source address length");
954 			return false;
955 		}
956 		if (ipv4_is_multicast(nla_get_in_addr(attr))) {
957 			NL_SET_ERR_MSG_MOD(extack, "IPv4 multicast source address is not allowed");
958 			return false;
959 		}
960 		break;
961 #if IS_ENABLED(CONFIG_IPV6)
962 	case htons(ETH_P_IPV6): {
963 		struct in6_addr src;
964 
965 		if (nla_len(attr) != sizeof(struct in6_addr)) {
966 			NL_SET_ERR_MSG_MOD(extack, "IPv6 invalid source address length");
967 			return false;
968 		}
969 		src = nla_get_in6_addr(attr);
970 		if (ipv6_addr_is_multicast(&src)) {
971 			NL_SET_ERR_MSG_MOD(extack, "IPv6 multicast source address is not allowed");
972 			return false;
973 		}
974 		break;
975 	}
976 #endif
977 	default:
978 		NL_SET_ERR_MSG_MOD(extack, "Invalid protocol used with source address");
979 		return false;
980 	}
981 
982 	return true;
983 }
984 
985 static const struct nla_policy br_mdbe_attrs_pol[MDBE_ATTR_MAX + 1] = {
986 	[MDBE_ATTR_SOURCE] = NLA_POLICY_RANGE(NLA_BINARY,
987 					      sizeof(struct in_addr),
988 					      sizeof(struct in6_addr)),
989 };
990 
991 static int br_mdb_parse(struct sk_buff *skb, struct nlmsghdr *nlh,
992 			struct net_device **pdev, struct br_mdb_entry **pentry,
993 			struct nlattr **mdb_attrs, struct netlink_ext_ack *extack)
994 {
995 	struct net *net = sock_net(skb->sk);
996 	struct br_mdb_entry *entry;
997 	struct br_port_msg *bpm;
998 	struct nlattr *tb[MDBA_SET_ENTRY_MAX+1];
999 	struct net_device *dev;
1000 	int err;
1001 
1002 	err = nlmsg_parse_deprecated(nlh, sizeof(*bpm), tb,
1003 				     MDBA_SET_ENTRY_MAX, NULL, NULL);
1004 	if (err < 0)
1005 		return err;
1006 
1007 	bpm = nlmsg_data(nlh);
1008 	if (bpm->ifindex == 0) {
1009 		NL_SET_ERR_MSG_MOD(extack, "Invalid bridge ifindex");
1010 		return -EINVAL;
1011 	}
1012 
1013 	dev = __dev_get_by_index(net, bpm->ifindex);
1014 	if (dev == NULL) {
1015 		NL_SET_ERR_MSG_MOD(extack, "Bridge device doesn't exist");
1016 		return -ENODEV;
1017 	}
1018 
1019 	if (!(dev->priv_flags & IFF_EBRIDGE)) {
1020 		NL_SET_ERR_MSG_MOD(extack, "Device is not a bridge");
1021 		return -EOPNOTSUPP;
1022 	}
1023 
1024 	*pdev = dev;
1025 
1026 	if (!tb[MDBA_SET_ENTRY]) {
1027 		NL_SET_ERR_MSG_MOD(extack, "Missing MDBA_SET_ENTRY attribute");
1028 		return -EINVAL;
1029 	}
1030 	if (nla_len(tb[MDBA_SET_ENTRY]) != sizeof(struct br_mdb_entry)) {
1031 		NL_SET_ERR_MSG_MOD(extack, "Invalid MDBA_SET_ENTRY attribute length");
1032 		return -EINVAL;
1033 	}
1034 
1035 	entry = nla_data(tb[MDBA_SET_ENTRY]);
1036 	if (!is_valid_mdb_entry(entry, extack))
1037 		return -EINVAL;
1038 	*pentry = entry;
1039 
1040 	if (tb[MDBA_SET_ENTRY_ATTRS]) {
1041 		err = nla_parse_nested(mdb_attrs, MDBE_ATTR_MAX,
1042 				       tb[MDBA_SET_ENTRY_ATTRS],
1043 				       br_mdbe_attrs_pol, extack);
1044 		if (err)
1045 			return err;
1046 		if (mdb_attrs[MDBE_ATTR_SOURCE] &&
1047 		    !is_valid_mdb_source(mdb_attrs[MDBE_ATTR_SOURCE],
1048 					 entry->addr.proto, extack))
1049 			return -EINVAL;
1050 	} else {
1051 		memset(mdb_attrs, 0,
1052 		       sizeof(struct nlattr *) * (MDBE_ATTR_MAX + 1));
1053 	}
1054 
1055 	return 0;
1056 }
1057 
1058 static struct net_bridge_mcast *
1059 __br_mdb_choose_context(struct net_bridge *br,
1060 			const struct br_mdb_entry *entry,
1061 			struct netlink_ext_ack *extack)
1062 {
1063 	struct net_bridge_mcast *brmctx = NULL;
1064 	struct net_bridge_vlan *v;
1065 
1066 	if (!br_opt_get(br, BROPT_MCAST_VLAN_SNOOPING_ENABLED)) {
1067 		brmctx = &br->multicast_ctx;
1068 		goto out;
1069 	}
1070 
1071 	if (!entry->vid) {
1072 		NL_SET_ERR_MSG_MOD(extack, "Cannot add an entry without a vlan when vlan snooping is enabled");
1073 		goto out;
1074 	}
1075 
1076 	v = br_vlan_find(br_vlan_group(br), entry->vid);
1077 	if (!v) {
1078 		NL_SET_ERR_MSG_MOD(extack, "Vlan is not configured");
1079 		goto out;
1080 	}
1081 	if (br_multicast_ctx_vlan_global_disabled(&v->br_mcast_ctx)) {
1082 		NL_SET_ERR_MSG_MOD(extack, "Vlan's multicast processing is disabled");
1083 		goto out;
1084 	}
1085 	brmctx = &v->br_mcast_ctx;
1086 out:
1087 	return brmctx;
1088 }
1089 
1090 static int br_mdb_add_group(struct net_bridge *br, struct net_bridge_port *port,
1091 			    struct br_mdb_entry *entry,
1092 			    struct nlattr **mdb_attrs,
1093 			    struct netlink_ext_ack *extack)
1094 {
1095 	struct net_bridge_mdb_entry *mp, *star_mp;
1096 	struct net_bridge_port_group __rcu **pp;
1097 	struct net_bridge_port_group *p;
1098 	struct net_bridge_mcast *brmctx;
1099 	struct br_ip group, star_group;
1100 	unsigned long now = jiffies;
1101 	unsigned char flags = 0;
1102 	u8 filter_mode;
1103 	int err;
1104 
1105 	__mdb_entry_to_br_ip(entry, &group, mdb_attrs);
1106 
1107 	brmctx = __br_mdb_choose_context(br, entry, extack);
1108 	if (!brmctx)
1109 		return -EINVAL;
1110 
1111 	/* host join errors which can happen before creating the group */
1112 	if (!port) {
1113 		/* don't allow any flags for host-joined groups */
1114 		if (entry->state) {
1115 			NL_SET_ERR_MSG_MOD(extack, "Flags are not allowed for host groups");
1116 			return -EINVAL;
1117 		}
1118 		if (!br_multicast_is_star_g(&group)) {
1119 			NL_SET_ERR_MSG_MOD(extack, "Groups with sources cannot be manually host joined");
1120 			return -EINVAL;
1121 		}
1122 	}
1123 
1124 	if (br_group_is_l2(&group) && entry->state != MDB_PERMANENT) {
1125 		NL_SET_ERR_MSG_MOD(extack, "Only permanent L2 entries allowed");
1126 		return -EINVAL;
1127 	}
1128 
1129 	mp = br_mdb_ip_get(br, &group);
1130 	if (!mp) {
1131 		mp = br_multicast_new_group(br, &group);
1132 		err = PTR_ERR_OR_ZERO(mp);
1133 		if (err)
1134 			return err;
1135 	}
1136 
1137 	/* host join */
1138 	if (!port) {
1139 		if (mp->host_joined) {
1140 			NL_SET_ERR_MSG_MOD(extack, "Group is already joined by host");
1141 			return -EEXIST;
1142 		}
1143 
1144 		br_multicast_host_join(brmctx, mp, false);
1145 		br_mdb_notify(br->dev, mp, NULL, RTM_NEWMDB);
1146 
1147 		return 0;
1148 	}
1149 
1150 	for (pp = &mp->ports;
1151 	     (p = mlock_dereference(*pp, br)) != NULL;
1152 	     pp = &p->next) {
1153 		if (p->key.port == port) {
1154 			NL_SET_ERR_MSG_MOD(extack, "Group is already joined by port");
1155 			return -EEXIST;
1156 		}
1157 		if ((unsigned long)p->key.port < (unsigned long)port)
1158 			break;
1159 	}
1160 
1161 	filter_mode = br_multicast_is_star_g(&group) ? MCAST_EXCLUDE :
1162 						       MCAST_INCLUDE;
1163 
1164 	if (entry->state == MDB_PERMANENT)
1165 		flags |= MDB_PG_FLAGS_PERMANENT;
1166 
1167 	p = br_multicast_new_port_group(port, &group, *pp, flags, NULL,
1168 					filter_mode, RTPROT_STATIC);
1169 	if (unlikely(!p)) {
1170 		NL_SET_ERR_MSG_MOD(extack, "Couldn't allocate new port group");
1171 		return -ENOMEM;
1172 	}
1173 	rcu_assign_pointer(*pp, p);
1174 	if (entry->state == MDB_TEMPORARY)
1175 		mod_timer(&p->timer,
1176 			  now + brmctx->multicast_membership_interval);
1177 	br_mdb_notify(br->dev, mp, p, RTM_NEWMDB);
1178 	/* if we are adding a new EXCLUDE port group (*,G) it needs to be also
1179 	 * added to all S,G entries for proper replication, if we are adding
1180 	 * a new INCLUDE port (S,G) then all of *,G EXCLUDE ports need to be
1181 	 * added to it for proper replication
1182 	 */
1183 	if (br_multicast_should_handle_mode(brmctx, group.proto)) {
1184 		switch (filter_mode) {
1185 		case MCAST_EXCLUDE:
1186 			br_multicast_star_g_handle_mode(p, MCAST_EXCLUDE);
1187 			break;
1188 		case MCAST_INCLUDE:
1189 			star_group = p->key.addr;
1190 			memset(&star_group.src, 0, sizeof(star_group.src));
1191 			star_mp = br_mdb_ip_get(br, &star_group);
1192 			if (star_mp)
1193 				br_multicast_sg_add_exclude_ports(star_mp, p);
1194 			break;
1195 		}
1196 	}
1197 
1198 	return 0;
1199 }
1200 
1201 static int __br_mdb_add(struct net *net, struct net_bridge *br,
1202 			struct net_bridge_port *p,
1203 			struct br_mdb_entry *entry,
1204 			struct nlattr **mdb_attrs,
1205 			struct netlink_ext_ack *extack)
1206 {
1207 	int ret;
1208 
1209 	spin_lock_bh(&br->multicast_lock);
1210 	ret = br_mdb_add_group(br, p, entry, mdb_attrs, extack);
1211 	spin_unlock_bh(&br->multicast_lock);
1212 
1213 	return ret;
1214 }
1215 
1216 static int br_mdb_add(struct sk_buff *skb, struct nlmsghdr *nlh,
1217 		      struct netlink_ext_ack *extack)
1218 {
1219 	struct nlattr *mdb_attrs[MDBE_ATTR_MAX + 1];
1220 	struct net *net = sock_net(skb->sk);
1221 	struct net_bridge_vlan_group *vg;
1222 	struct net_bridge_port *p = NULL;
1223 	struct net_device *dev, *pdev;
1224 	struct br_mdb_entry *entry;
1225 	struct net_bridge_vlan *v;
1226 	struct net_bridge *br;
1227 	int err;
1228 
1229 	err = br_mdb_parse(skb, nlh, &dev, &entry, mdb_attrs, extack);
1230 	if (err < 0)
1231 		return err;
1232 
1233 	br = netdev_priv(dev);
1234 
1235 	if (!netif_running(br->dev)) {
1236 		NL_SET_ERR_MSG_MOD(extack, "Bridge device is not running");
1237 		return -EINVAL;
1238 	}
1239 
1240 	if (!br_opt_get(br, BROPT_MULTICAST_ENABLED)) {
1241 		NL_SET_ERR_MSG_MOD(extack, "Bridge's multicast processing is disabled");
1242 		return -EINVAL;
1243 	}
1244 
1245 	if (entry->ifindex != br->dev->ifindex) {
1246 		pdev = __dev_get_by_index(net, entry->ifindex);
1247 		if (!pdev) {
1248 			NL_SET_ERR_MSG_MOD(extack, "Port net device doesn't exist");
1249 			return -ENODEV;
1250 		}
1251 
1252 		p = br_port_get_rtnl(pdev);
1253 		if (!p) {
1254 			NL_SET_ERR_MSG_MOD(extack, "Net device is not a bridge port");
1255 			return -EINVAL;
1256 		}
1257 
1258 		if (p->br != br) {
1259 			NL_SET_ERR_MSG_MOD(extack, "Port belongs to a different bridge device");
1260 			return -EINVAL;
1261 		}
1262 		if (p->state == BR_STATE_DISABLED) {
1263 			NL_SET_ERR_MSG_MOD(extack, "Port is in disabled state");
1264 			return -EINVAL;
1265 		}
1266 		vg = nbp_vlan_group(p);
1267 	} else {
1268 		vg = br_vlan_group(br);
1269 	}
1270 
1271 	/* If vlan filtering is enabled and VLAN is not specified
1272 	 * install mdb entry on all vlans configured on the port.
1273 	 */
1274 	if (br_vlan_enabled(br->dev) && vg && entry->vid == 0) {
1275 		list_for_each_entry(v, &vg->vlan_list, vlist) {
1276 			entry->vid = v->vid;
1277 			err = __br_mdb_add(net, br, p, entry, mdb_attrs, extack);
1278 			if (err)
1279 				break;
1280 		}
1281 	} else {
1282 		err = __br_mdb_add(net, br, p, entry, mdb_attrs, extack);
1283 	}
1284 
1285 	return err;
1286 }
1287 
1288 static int __br_mdb_del(struct net_bridge *br, struct br_mdb_entry *entry,
1289 			struct nlattr **mdb_attrs)
1290 {
1291 	struct net_bridge_mdb_entry *mp;
1292 	struct net_bridge_port_group *p;
1293 	struct net_bridge_port_group __rcu **pp;
1294 	struct br_ip ip;
1295 	int err = -EINVAL;
1296 
1297 	if (!netif_running(br->dev) || !br_opt_get(br, BROPT_MULTICAST_ENABLED))
1298 		return -EINVAL;
1299 
1300 	__mdb_entry_to_br_ip(entry, &ip, mdb_attrs);
1301 
1302 	spin_lock_bh(&br->multicast_lock);
1303 	mp = br_mdb_ip_get(br, &ip);
1304 	if (!mp)
1305 		goto unlock;
1306 
1307 	/* host leave */
1308 	if (entry->ifindex == mp->br->dev->ifindex && mp->host_joined) {
1309 		br_multicast_host_leave(mp, false);
1310 		err = 0;
1311 		br_mdb_notify(br->dev, mp, NULL, RTM_DELMDB);
1312 		if (!mp->ports && netif_running(br->dev))
1313 			mod_timer(&mp->timer, jiffies);
1314 		goto unlock;
1315 	}
1316 
1317 	for (pp = &mp->ports;
1318 	     (p = mlock_dereference(*pp, br)) != NULL;
1319 	     pp = &p->next) {
1320 		if (!p->key.port || p->key.port->dev->ifindex != entry->ifindex)
1321 			continue;
1322 
1323 		if (p->key.port->state == BR_STATE_DISABLED)
1324 			goto unlock;
1325 
1326 		br_multicast_del_pg(mp, p, pp);
1327 		err = 0;
1328 		break;
1329 	}
1330 
1331 unlock:
1332 	spin_unlock_bh(&br->multicast_lock);
1333 	return err;
1334 }
1335 
1336 static int br_mdb_del(struct sk_buff *skb, struct nlmsghdr *nlh,
1337 		      struct netlink_ext_ack *extack)
1338 {
1339 	struct nlattr *mdb_attrs[MDBE_ATTR_MAX + 1];
1340 	struct net *net = sock_net(skb->sk);
1341 	struct net_bridge_vlan_group *vg;
1342 	struct net_bridge_port *p = NULL;
1343 	struct net_device *dev, *pdev;
1344 	struct br_mdb_entry *entry;
1345 	struct net_bridge_vlan *v;
1346 	struct net_bridge *br;
1347 	int err;
1348 
1349 	err = br_mdb_parse(skb, nlh, &dev, &entry, mdb_attrs, extack);
1350 	if (err < 0)
1351 		return err;
1352 
1353 	br = netdev_priv(dev);
1354 
1355 	if (entry->ifindex != br->dev->ifindex) {
1356 		pdev = __dev_get_by_index(net, entry->ifindex);
1357 		if (!pdev)
1358 			return -ENODEV;
1359 
1360 		p = br_port_get_rtnl(pdev);
1361 		if (!p || p->br != br || p->state == BR_STATE_DISABLED)
1362 			return -EINVAL;
1363 		vg = nbp_vlan_group(p);
1364 	} else {
1365 		vg = br_vlan_group(br);
1366 	}
1367 
1368 	/* If vlan filtering is enabled and VLAN is not specified
1369 	 * delete mdb entry on all vlans configured on the port.
1370 	 */
1371 	if (br_vlan_enabled(br->dev) && vg && entry->vid == 0) {
1372 		list_for_each_entry(v, &vg->vlan_list, vlist) {
1373 			entry->vid = v->vid;
1374 			err = __br_mdb_del(br, entry, mdb_attrs);
1375 		}
1376 	} else {
1377 		err = __br_mdb_del(br, entry, mdb_attrs);
1378 	}
1379 
1380 	return err;
1381 }
1382 
1383 void br_mdb_init(void)
1384 {
1385 	rtnl_register_module(THIS_MODULE, PF_BRIDGE, RTM_GETMDB, NULL, br_mdb_dump, 0);
1386 	rtnl_register_module(THIS_MODULE, PF_BRIDGE, RTM_NEWMDB, br_mdb_add, NULL, 0);
1387 	rtnl_register_module(THIS_MODULE, PF_BRIDGE, RTM_DELMDB, br_mdb_del, NULL, 0);
1388 }
1389 
1390 void br_mdb_uninit(void)
1391 {
1392 	rtnl_unregister(PF_BRIDGE, RTM_GETMDB);
1393 	rtnl_unregister(PF_BRIDGE, RTM_NEWMDB);
1394 	rtnl_unregister(PF_BRIDGE, RTM_DELMDB);
1395 }
1396