1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright (C) B.A.T.M.A.N. contributors: 3 * 4 * Linus Lüssing 5 */ 6 7 #include "multicast.h" 8 #include "main.h" 9 10 #include <linux/atomic.h> 11 #include <linux/bitops.h> 12 #include <linux/bug.h> 13 #include <linux/byteorder/generic.h> 14 #include <linux/container_of.h> 15 #include <linux/errno.h> 16 #include <linux/etherdevice.h> 17 #include <linux/gfp.h> 18 #include <linux/icmpv6.h> 19 #include <linux/if_bridge.h> 20 #include <linux/if_ether.h> 21 #include <linux/igmp.h> 22 #include <linux/in.h> 23 #include <linux/in6.h> 24 #include <linux/inetdevice.h> 25 #include <linux/ip.h> 26 #include <linux/ipv6.h> 27 #include <linux/jiffies.h> 28 #include <linux/list.h> 29 #include <linux/lockdep.h> 30 #include <linux/netdevice.h> 31 #include <linux/netlink.h> 32 #include <linux/printk.h> 33 #include <linux/rculist.h> 34 #include <linux/rcupdate.h> 35 #include <linux/skbuff.h> 36 #include <linux/slab.h> 37 #include <linux/spinlock.h> 38 #include <linux/sprintf.h> 39 #include <linux/stddef.h> 40 #include <linux/string.h> 41 #include <linux/types.h> 42 #include <linux/workqueue.h> 43 #include <net/addrconf.h> 44 #include <net/genetlink.h> 45 #include <net/if_inet6.h> 46 #include <net/ip.h> 47 #include <net/ipv6.h> 48 #include <net/netlink.h> 49 #include <net/sock.h> 50 #include <uapi/linux/batadv_packet.h> 51 #include <uapi/linux/batman_adv.h> 52 53 #include "bridge_loop_avoidance.h" 54 #include "hard-interface.h" 55 #include "hash.h" 56 #include "log.h" 57 #include "netlink.h" 58 #include "send.h" 59 #include "soft-interface.h" 60 #include "translation-table.h" 61 #include "tvlv.h" 62 63 static void batadv_mcast_mla_update(struct work_struct *work); 64 65 /** 66 * batadv_mcast_start_timer() - schedule the multicast periodic worker 67 * @bat_priv: the bat priv with all the soft interface information 68 */ 69 static void batadv_mcast_start_timer(struct batadv_priv *bat_priv) 70 { 71 queue_delayed_work(batadv_event_workqueue, &bat_priv->mcast.work, 72 msecs_to_jiffies(BATADV_MCAST_WORK_PERIOD)); 73 } 74 75 /** 76 * batadv_mcast_get_bridge() - get the bridge on top of the softif if it exists 77 * @soft_iface: netdev struct of the mesh interface 78 * 79 * If the given soft interface has a bridge on top then the refcount 80 * of the according net device is increased. 81 * 82 * Return: NULL if no such bridge exists. Otherwise the net device of the 83 * bridge. 84 */ 85 static struct net_device *batadv_mcast_get_bridge(struct net_device *soft_iface) 86 { 87 struct net_device *upper = soft_iface; 88 89 rcu_read_lock(); 90 do { 91 upper = netdev_master_upper_dev_get_rcu(upper); 92 } while (upper && !netif_is_bridge_master(upper)); 93 94 dev_hold(upper); 95 rcu_read_unlock(); 96 97 return upper; 98 } 99 100 /** 101 * batadv_mcast_mla_rtr_flags_softif_get_ipv4() - get mcast router flags from 102 * node for IPv4 103 * @dev: the interface to check 104 * 105 * Checks the presence of an IPv4 multicast router on this node. 106 * 107 * Caller needs to hold rcu read lock. 108 * 109 * Return: BATADV_NO_FLAGS if present, BATADV_MCAST_WANT_NO_RTR4 otherwise. 110 */ 111 static u8 batadv_mcast_mla_rtr_flags_softif_get_ipv4(struct net_device *dev) 112 { 113 struct in_device *in_dev = __in_dev_get_rcu(dev); 114 115 if (in_dev && IN_DEV_MFORWARD(in_dev)) 116 return BATADV_NO_FLAGS; 117 else 118 return BATADV_MCAST_WANT_NO_RTR4; 119 } 120 121 /** 122 * batadv_mcast_mla_rtr_flags_softif_get_ipv6() - get mcast router flags from 123 * node for IPv6 124 * @dev: the interface to check 125 * 126 * Checks the presence of an IPv6 multicast router on this node. 127 * 128 * Caller needs to hold rcu read lock. 129 * 130 * Return: BATADV_NO_FLAGS if present, BATADV_MCAST_WANT_NO_RTR6 otherwise. 131 */ 132 #if IS_ENABLED(CONFIG_IPV6_MROUTE) 133 static u8 batadv_mcast_mla_rtr_flags_softif_get_ipv6(struct net_device *dev) 134 { 135 struct inet6_dev *in6_dev = __in6_dev_get(dev); 136 137 if (in6_dev && atomic_read(&in6_dev->cnf.mc_forwarding)) 138 return BATADV_NO_FLAGS; 139 else 140 return BATADV_MCAST_WANT_NO_RTR6; 141 } 142 #else 143 static inline u8 144 batadv_mcast_mla_rtr_flags_softif_get_ipv6(struct net_device *dev) 145 { 146 return BATADV_MCAST_WANT_NO_RTR6; 147 } 148 #endif 149 150 /** 151 * batadv_mcast_mla_rtr_flags_softif_get() - get mcast router flags from node 152 * @bat_priv: the bat priv with all the soft interface information 153 * @bridge: bridge interface on top of the soft_iface if present, 154 * otherwise pass NULL 155 * 156 * Checks the presence of IPv4 and IPv6 multicast routers on this 157 * node. 158 * 159 * Return: 160 * BATADV_NO_FLAGS: Both an IPv4 and IPv6 multicast router is present 161 * BATADV_MCAST_WANT_NO_RTR4: No IPv4 multicast router is present 162 * BATADV_MCAST_WANT_NO_RTR6: No IPv6 multicast router is present 163 * The former two OR'd: no multicast router is present 164 */ 165 static u8 batadv_mcast_mla_rtr_flags_softif_get(struct batadv_priv *bat_priv, 166 struct net_device *bridge) 167 { 168 struct net_device *dev = bridge ? bridge : bat_priv->soft_iface; 169 u8 flags = BATADV_NO_FLAGS; 170 171 rcu_read_lock(); 172 173 flags |= batadv_mcast_mla_rtr_flags_softif_get_ipv4(dev); 174 flags |= batadv_mcast_mla_rtr_flags_softif_get_ipv6(dev); 175 176 rcu_read_unlock(); 177 178 return flags; 179 } 180 181 /** 182 * batadv_mcast_mla_rtr_flags_bridge_get() - get mcast router flags from bridge 183 * @bat_priv: the bat priv with all the soft interface information 184 * @bridge: bridge interface on top of the soft_iface if present, 185 * otherwise pass NULL 186 * 187 * Checks the presence of IPv4 and IPv6 multicast routers behind a bridge. 188 * 189 * Return: 190 * BATADV_NO_FLAGS: Both an IPv4 and IPv6 multicast router is present 191 * BATADV_MCAST_WANT_NO_RTR4: No IPv4 multicast router is present 192 * BATADV_MCAST_WANT_NO_RTR6: No IPv6 multicast router is present 193 * The former two OR'd: no multicast router is present 194 */ 195 static u8 batadv_mcast_mla_rtr_flags_bridge_get(struct batadv_priv *bat_priv, 196 struct net_device *bridge) 197 { 198 struct net_device *dev = bat_priv->soft_iface; 199 u8 flags = BATADV_NO_FLAGS; 200 201 if (!bridge) 202 return BATADV_MCAST_WANT_NO_RTR4 | BATADV_MCAST_WANT_NO_RTR6; 203 204 if (!br_multicast_has_router_adjacent(dev, ETH_P_IP)) 205 flags |= BATADV_MCAST_WANT_NO_RTR4; 206 if (!br_multicast_has_router_adjacent(dev, ETH_P_IPV6)) 207 flags |= BATADV_MCAST_WANT_NO_RTR6; 208 209 return flags; 210 } 211 212 /** 213 * batadv_mcast_mla_rtr_flags_get() - get multicast router flags 214 * @bat_priv: the bat priv with all the soft interface information 215 * @bridge: bridge interface on top of the soft_iface if present, 216 * otherwise pass NULL 217 * 218 * Checks the presence of IPv4 and IPv6 multicast routers on this 219 * node or behind its bridge. 220 * 221 * Return: 222 * BATADV_NO_FLAGS: Both an IPv4 and IPv6 multicast router is present 223 * BATADV_MCAST_WANT_NO_RTR4: No IPv4 multicast router is present 224 * BATADV_MCAST_WANT_NO_RTR6: No IPv6 multicast router is present 225 * The former two OR'd: no multicast router is present 226 */ 227 static u8 batadv_mcast_mla_rtr_flags_get(struct batadv_priv *bat_priv, 228 struct net_device *bridge) 229 { 230 u8 flags = BATADV_MCAST_WANT_NO_RTR4 | BATADV_MCAST_WANT_NO_RTR6; 231 232 flags &= batadv_mcast_mla_rtr_flags_softif_get(bat_priv, bridge); 233 flags &= batadv_mcast_mla_rtr_flags_bridge_get(bat_priv, bridge); 234 235 return flags; 236 } 237 238 /** 239 * batadv_mcast_mla_forw_flags_get() - get multicast forwarding flags 240 * @bat_priv: the bat priv with all the soft interface information 241 * 242 * Checks if all active hard interfaces have an MTU larger or equal to 1280 243 * bytes (IPv6 minimum MTU). 244 * 245 * Return: BATADV_MCAST_HAVE_MC_PTYPE_CAPA if yes, BATADV_NO_FLAGS otherwise. 246 */ 247 static u8 batadv_mcast_mla_forw_flags_get(struct batadv_priv *bat_priv) 248 { 249 const struct batadv_hard_iface *hard_iface; 250 251 rcu_read_lock(); 252 list_for_each_entry_rcu(hard_iface, &batadv_hardif_list, list) { 253 if (hard_iface->if_status != BATADV_IF_ACTIVE) 254 continue; 255 256 if (hard_iface->soft_iface != bat_priv->soft_iface) 257 continue; 258 259 if (hard_iface->net_dev->mtu < IPV6_MIN_MTU) { 260 rcu_read_unlock(); 261 return BATADV_NO_FLAGS; 262 } 263 } 264 rcu_read_unlock(); 265 266 return BATADV_MCAST_HAVE_MC_PTYPE_CAPA; 267 } 268 269 /** 270 * batadv_mcast_mla_flags_get() - get the new multicast flags 271 * @bat_priv: the bat priv with all the soft interface information 272 * 273 * Return: A set of flags for the current/next TVLV, querier and 274 * bridge state. 275 */ 276 static struct batadv_mcast_mla_flags 277 batadv_mcast_mla_flags_get(struct batadv_priv *bat_priv) 278 { 279 struct net_device *dev = bat_priv->soft_iface; 280 struct batadv_mcast_querier_state *qr4, *qr6; 281 struct batadv_mcast_mla_flags mla_flags; 282 struct net_device *bridge; 283 284 bridge = batadv_mcast_get_bridge(dev); 285 286 memset(&mla_flags, 0, sizeof(mla_flags)); 287 mla_flags.enabled = 1; 288 mla_flags.tvlv_flags |= batadv_mcast_mla_rtr_flags_get(bat_priv, 289 bridge); 290 mla_flags.tvlv_flags |= batadv_mcast_mla_forw_flags_get(bat_priv); 291 292 if (!bridge) 293 return mla_flags; 294 295 dev_put(bridge); 296 297 mla_flags.bridged = 1; 298 qr4 = &mla_flags.querier_ipv4; 299 qr6 = &mla_flags.querier_ipv6; 300 301 if (!IS_ENABLED(CONFIG_BRIDGE_IGMP_SNOOPING)) 302 pr_warn_once("No bridge IGMP snooping compiled - multicast optimizations disabled\n"); 303 304 qr4->exists = br_multicast_has_querier_anywhere(dev, ETH_P_IP); 305 qr4->shadowing = br_multicast_has_querier_adjacent(dev, ETH_P_IP); 306 307 qr6->exists = br_multicast_has_querier_anywhere(dev, ETH_P_IPV6); 308 qr6->shadowing = br_multicast_has_querier_adjacent(dev, ETH_P_IPV6); 309 310 mla_flags.tvlv_flags |= BATADV_MCAST_WANT_ALL_UNSNOOPABLES; 311 312 /* 1) If no querier exists at all, then multicast listeners on 313 * our local TT clients behind the bridge will keep silent. 314 * 2) If the selected querier is on one of our local TT clients, 315 * behind the bridge, then this querier might shadow multicast 316 * listeners on our local TT clients, behind this bridge. 317 * 318 * In both cases, we will signalize other batman nodes that 319 * we need all multicast traffic of the according protocol. 320 */ 321 if (!qr4->exists || qr4->shadowing) { 322 mla_flags.tvlv_flags |= BATADV_MCAST_WANT_ALL_IPV4; 323 mla_flags.tvlv_flags &= ~BATADV_MCAST_WANT_NO_RTR4; 324 } 325 326 if (!qr6->exists || qr6->shadowing) { 327 mla_flags.tvlv_flags |= BATADV_MCAST_WANT_ALL_IPV6; 328 mla_flags.tvlv_flags &= ~BATADV_MCAST_WANT_NO_RTR6; 329 } 330 331 return mla_flags; 332 } 333 334 /** 335 * batadv_mcast_mla_is_duplicate() - check whether an address is in a list 336 * @mcast_addr: the multicast address to check 337 * @mcast_list: the list with multicast addresses to search in 338 * 339 * Return: true if the given address is already in the given list. 340 * Otherwise returns false. 341 */ 342 static bool batadv_mcast_mla_is_duplicate(u8 *mcast_addr, 343 struct hlist_head *mcast_list) 344 { 345 struct batadv_hw_addr *mcast_entry; 346 347 hlist_for_each_entry(mcast_entry, mcast_list, list) 348 if (batadv_compare_eth(mcast_entry->addr, mcast_addr)) 349 return true; 350 351 return false; 352 } 353 354 /** 355 * batadv_mcast_mla_softif_get_ipv4() - get softif IPv4 multicast listeners 356 * @dev: the device to collect multicast addresses from 357 * @mcast_list: a list to put found addresses into 358 * @flags: flags indicating the new multicast state 359 * 360 * Collects multicast addresses of IPv4 multicast listeners residing 361 * on this kernel on the given soft interface, dev, in 362 * the given mcast_list. In general, multicast listeners provided by 363 * your multicast receiving applications run directly on this node. 364 * 365 * Return: -ENOMEM on memory allocation error or the number of 366 * items added to the mcast_list otherwise. 367 */ 368 static int 369 batadv_mcast_mla_softif_get_ipv4(struct net_device *dev, 370 struct hlist_head *mcast_list, 371 struct batadv_mcast_mla_flags *flags) 372 { 373 struct batadv_hw_addr *new; 374 struct in_device *in_dev; 375 u8 mcast_addr[ETH_ALEN]; 376 struct ip_mc_list *pmc; 377 int ret = 0; 378 379 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_IPV4) 380 return 0; 381 382 rcu_read_lock(); 383 384 in_dev = __in_dev_get_rcu(dev); 385 if (!in_dev) { 386 rcu_read_unlock(); 387 return 0; 388 } 389 390 for (pmc = rcu_dereference(in_dev->mc_list); pmc; 391 pmc = rcu_dereference(pmc->next_rcu)) { 392 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 393 ipv4_is_local_multicast(pmc->multiaddr)) 394 continue; 395 396 if (!(flags->tvlv_flags & BATADV_MCAST_WANT_NO_RTR4) && 397 !ipv4_is_local_multicast(pmc->multiaddr)) 398 continue; 399 400 ip_eth_mc_map(pmc->multiaddr, mcast_addr); 401 402 if (batadv_mcast_mla_is_duplicate(mcast_addr, mcast_list)) 403 continue; 404 405 new = kmalloc(sizeof(*new), GFP_ATOMIC); 406 if (!new) { 407 ret = -ENOMEM; 408 break; 409 } 410 411 ether_addr_copy(new->addr, mcast_addr); 412 hlist_add_head(&new->list, mcast_list); 413 ret++; 414 } 415 rcu_read_unlock(); 416 417 return ret; 418 } 419 420 /** 421 * batadv_mcast_mla_softif_get_ipv6() - get softif IPv6 multicast listeners 422 * @dev: the device to collect multicast addresses from 423 * @mcast_list: a list to put found addresses into 424 * @flags: flags indicating the new multicast state 425 * 426 * Collects multicast addresses of IPv6 multicast listeners residing 427 * on this kernel on the given soft interface, dev, in 428 * the given mcast_list. In general, multicast listeners provided by 429 * your multicast receiving applications run directly on this node. 430 * 431 * Return: -ENOMEM on memory allocation error or the number of 432 * items added to the mcast_list otherwise. 433 */ 434 #if IS_ENABLED(CONFIG_IPV6) 435 static int 436 batadv_mcast_mla_softif_get_ipv6(struct net_device *dev, 437 struct hlist_head *mcast_list, 438 struct batadv_mcast_mla_flags *flags) 439 { 440 struct batadv_hw_addr *new; 441 struct inet6_dev *in6_dev; 442 u8 mcast_addr[ETH_ALEN]; 443 struct ifmcaddr6 *pmc6; 444 int ret = 0; 445 446 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_IPV6) 447 return 0; 448 449 rcu_read_lock(); 450 451 in6_dev = __in6_dev_get(dev); 452 if (!in6_dev) { 453 rcu_read_unlock(); 454 return 0; 455 } 456 457 for (pmc6 = rcu_dereference(in6_dev->mc_list); 458 pmc6; 459 pmc6 = rcu_dereference(pmc6->next)) { 460 if (IPV6_ADDR_MC_SCOPE(&pmc6->mca_addr) < 461 IPV6_ADDR_SCOPE_LINKLOCAL) 462 continue; 463 464 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 465 ipv6_addr_is_ll_all_nodes(&pmc6->mca_addr)) 466 continue; 467 468 if (!(flags->tvlv_flags & BATADV_MCAST_WANT_NO_RTR6) && 469 IPV6_ADDR_MC_SCOPE(&pmc6->mca_addr) > 470 IPV6_ADDR_SCOPE_LINKLOCAL) 471 continue; 472 473 ipv6_eth_mc_map(&pmc6->mca_addr, mcast_addr); 474 475 if (batadv_mcast_mla_is_duplicate(mcast_addr, mcast_list)) 476 continue; 477 478 new = kmalloc(sizeof(*new), GFP_ATOMIC); 479 if (!new) { 480 ret = -ENOMEM; 481 break; 482 } 483 484 ether_addr_copy(new->addr, mcast_addr); 485 hlist_add_head(&new->list, mcast_list); 486 ret++; 487 } 488 rcu_read_unlock(); 489 490 return ret; 491 } 492 #else 493 static inline int 494 batadv_mcast_mla_softif_get_ipv6(struct net_device *dev, 495 struct hlist_head *mcast_list, 496 struct batadv_mcast_mla_flags *flags) 497 { 498 return 0; 499 } 500 #endif 501 502 /** 503 * batadv_mcast_mla_softif_get() - get softif multicast listeners 504 * @dev: the device to collect multicast addresses from 505 * @mcast_list: a list to put found addresses into 506 * @flags: flags indicating the new multicast state 507 * 508 * Collects multicast addresses of multicast listeners residing 509 * on this kernel on the given soft interface, dev, in 510 * the given mcast_list. In general, multicast listeners provided by 511 * your multicast receiving applications run directly on this node. 512 * 513 * If there is a bridge interface on top of dev, collect from that one 514 * instead. Just like with IP addresses and routes, multicast listeners 515 * will(/should) register to the bridge interface instead of an 516 * enslaved bat0. 517 * 518 * Return: -ENOMEM on memory allocation error or the number of 519 * items added to the mcast_list otherwise. 520 */ 521 static int 522 batadv_mcast_mla_softif_get(struct net_device *dev, 523 struct hlist_head *mcast_list, 524 struct batadv_mcast_mla_flags *flags) 525 { 526 struct net_device *bridge = batadv_mcast_get_bridge(dev); 527 int ret4, ret6 = 0; 528 529 if (bridge) 530 dev = bridge; 531 532 ret4 = batadv_mcast_mla_softif_get_ipv4(dev, mcast_list, flags); 533 if (ret4 < 0) 534 goto out; 535 536 ret6 = batadv_mcast_mla_softif_get_ipv6(dev, mcast_list, flags); 537 if (ret6 < 0) { 538 ret4 = 0; 539 goto out; 540 } 541 542 out: 543 dev_put(bridge); 544 545 return ret4 + ret6; 546 } 547 548 /** 549 * batadv_mcast_mla_br_addr_cpy() - copy a bridge multicast address 550 * @dst: destination to write to - a multicast MAC address 551 * @src: source to read from - a multicast IP address 552 * 553 * Converts a given multicast IPv4/IPv6 address from a bridge 554 * to its matching multicast MAC address and copies it into the given 555 * destination buffer. 556 * 557 * Caller needs to make sure the destination buffer can hold 558 * at least ETH_ALEN bytes. 559 */ 560 static void batadv_mcast_mla_br_addr_cpy(char *dst, const struct br_ip *src) 561 { 562 if (src->proto == htons(ETH_P_IP)) 563 ip_eth_mc_map(src->dst.ip4, dst); 564 #if IS_ENABLED(CONFIG_IPV6) 565 else if (src->proto == htons(ETH_P_IPV6)) 566 ipv6_eth_mc_map(&src->dst.ip6, dst); 567 #endif 568 else 569 eth_zero_addr(dst); 570 } 571 572 /** 573 * batadv_mcast_mla_bridge_get() - get bridged-in multicast listeners 574 * @dev: a bridge slave whose bridge to collect multicast addresses from 575 * @mcast_list: a list to put found addresses into 576 * @flags: flags indicating the new multicast state 577 * 578 * Collects multicast addresses of multicast listeners residing 579 * on foreign, non-mesh devices which we gave access to our mesh via 580 * a bridge on top of the given soft interface, dev, in the given 581 * mcast_list. 582 * 583 * Return: -ENOMEM on memory allocation error or the number of 584 * items added to the mcast_list otherwise. 585 */ 586 static int batadv_mcast_mla_bridge_get(struct net_device *dev, 587 struct hlist_head *mcast_list, 588 struct batadv_mcast_mla_flags *flags) 589 { 590 struct list_head bridge_mcast_list = LIST_HEAD_INIT(bridge_mcast_list); 591 struct br_ip_list *br_ip_entry, *tmp; 592 u8 tvlv_flags = flags->tvlv_flags; 593 struct batadv_hw_addr *new; 594 u8 mcast_addr[ETH_ALEN]; 595 int ret; 596 597 /* we don't need to detect these devices/listeners, the IGMP/MLD 598 * snooping code of the Linux bridge already does that for us 599 */ 600 ret = br_multicast_list_adjacent(dev, &bridge_mcast_list); 601 if (ret < 0) 602 goto out; 603 604 list_for_each_entry(br_ip_entry, &bridge_mcast_list, list) { 605 if (br_ip_entry->addr.proto == htons(ETH_P_IP)) { 606 if (tvlv_flags & BATADV_MCAST_WANT_ALL_IPV4) 607 continue; 608 609 if (tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 610 ipv4_is_local_multicast(br_ip_entry->addr.dst.ip4)) 611 continue; 612 613 if (!(tvlv_flags & BATADV_MCAST_WANT_NO_RTR4) && 614 !ipv4_is_local_multicast(br_ip_entry->addr.dst.ip4)) 615 continue; 616 } 617 618 #if IS_ENABLED(CONFIG_IPV6) 619 if (br_ip_entry->addr.proto == htons(ETH_P_IPV6)) { 620 if (tvlv_flags & BATADV_MCAST_WANT_ALL_IPV6) 621 continue; 622 623 if (tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 624 ipv6_addr_is_ll_all_nodes(&br_ip_entry->addr.dst.ip6)) 625 continue; 626 627 if (!(tvlv_flags & BATADV_MCAST_WANT_NO_RTR6) && 628 IPV6_ADDR_MC_SCOPE(&br_ip_entry->addr.dst.ip6) > 629 IPV6_ADDR_SCOPE_LINKLOCAL) 630 continue; 631 } 632 #endif 633 634 batadv_mcast_mla_br_addr_cpy(mcast_addr, &br_ip_entry->addr); 635 if (batadv_mcast_mla_is_duplicate(mcast_addr, mcast_list)) 636 continue; 637 638 new = kmalloc(sizeof(*new), GFP_ATOMIC); 639 if (!new) { 640 ret = -ENOMEM; 641 break; 642 } 643 644 ether_addr_copy(new->addr, mcast_addr); 645 hlist_add_head(&new->list, mcast_list); 646 } 647 648 out: 649 list_for_each_entry_safe(br_ip_entry, tmp, &bridge_mcast_list, list) { 650 list_del(&br_ip_entry->list); 651 kfree(br_ip_entry); 652 } 653 654 return ret; 655 } 656 657 /** 658 * batadv_mcast_mla_list_free() - free a list of multicast addresses 659 * @mcast_list: the list to free 660 * 661 * Removes and frees all items in the given mcast_list. 662 */ 663 static void batadv_mcast_mla_list_free(struct hlist_head *mcast_list) 664 { 665 struct batadv_hw_addr *mcast_entry; 666 struct hlist_node *tmp; 667 668 hlist_for_each_entry_safe(mcast_entry, tmp, mcast_list, list) { 669 hlist_del(&mcast_entry->list); 670 kfree(mcast_entry); 671 } 672 } 673 674 /** 675 * batadv_mcast_mla_tt_retract() - clean up multicast listener announcements 676 * @bat_priv: the bat priv with all the soft interface information 677 * @mcast_list: a list of addresses which should _not_ be removed 678 * 679 * Retracts the announcement of any multicast listener from the 680 * translation table except the ones listed in the given mcast_list. 681 * 682 * If mcast_list is NULL then all are retracted. 683 */ 684 static void batadv_mcast_mla_tt_retract(struct batadv_priv *bat_priv, 685 struct hlist_head *mcast_list) 686 { 687 struct batadv_hw_addr *mcast_entry; 688 struct hlist_node *tmp; 689 690 hlist_for_each_entry_safe(mcast_entry, tmp, &bat_priv->mcast.mla_list, 691 list) { 692 if (mcast_list && 693 batadv_mcast_mla_is_duplicate(mcast_entry->addr, 694 mcast_list)) 695 continue; 696 697 batadv_tt_local_remove(bat_priv, mcast_entry->addr, 698 BATADV_NO_FLAGS, 699 "mcast TT outdated", false); 700 701 hlist_del(&mcast_entry->list); 702 kfree(mcast_entry); 703 } 704 } 705 706 /** 707 * batadv_mcast_mla_tt_add() - add multicast listener announcements 708 * @bat_priv: the bat priv with all the soft interface information 709 * @mcast_list: a list of addresses which are going to get added 710 * 711 * Adds multicast listener announcements from the given mcast_list to the 712 * translation table if they have not been added yet. 713 */ 714 static void batadv_mcast_mla_tt_add(struct batadv_priv *bat_priv, 715 struct hlist_head *mcast_list) 716 { 717 struct batadv_hw_addr *mcast_entry; 718 struct hlist_node *tmp; 719 720 if (!mcast_list) 721 return; 722 723 hlist_for_each_entry_safe(mcast_entry, tmp, mcast_list, list) { 724 if (batadv_mcast_mla_is_duplicate(mcast_entry->addr, 725 &bat_priv->mcast.mla_list)) 726 continue; 727 728 if (!batadv_tt_local_add(bat_priv->soft_iface, 729 mcast_entry->addr, BATADV_NO_FLAGS, 730 BATADV_NULL_IFINDEX, BATADV_NO_MARK)) 731 continue; 732 733 hlist_del(&mcast_entry->list); 734 hlist_add_head(&mcast_entry->list, &bat_priv->mcast.mla_list); 735 } 736 } 737 738 /** 739 * batadv_mcast_querier_log() - debug output regarding the querier status on 740 * link 741 * @bat_priv: the bat priv with all the soft interface information 742 * @str_proto: a string for the querier protocol (e.g. "IGMP" or "MLD") 743 * @old_state: the previous querier state on our link 744 * @new_state: the new querier state on our link 745 * 746 * Outputs debug messages to the logging facility with log level 'mcast' 747 * regarding changes to the querier status on the link which are relevant 748 * to our multicast optimizations. 749 * 750 * Usually this is about whether a querier appeared or vanished in 751 * our mesh or whether the querier is in the suboptimal position of being 752 * behind our local bridge segment: Snooping switches will directly 753 * forward listener reports to the querier, therefore batman-adv and 754 * the bridge will potentially not see these listeners - the querier is 755 * potentially shadowing listeners from us then. 756 * 757 * This is only interesting for nodes with a bridge on top of their 758 * soft interface. 759 */ 760 static void 761 batadv_mcast_querier_log(struct batadv_priv *bat_priv, char *str_proto, 762 struct batadv_mcast_querier_state *old_state, 763 struct batadv_mcast_querier_state *new_state) 764 { 765 if (!old_state->exists && new_state->exists) 766 batadv_info(bat_priv->soft_iface, "%s Querier appeared\n", 767 str_proto); 768 else if (old_state->exists && !new_state->exists) 769 batadv_info(bat_priv->soft_iface, 770 "%s Querier disappeared - multicast optimizations disabled\n", 771 str_proto); 772 else if (!bat_priv->mcast.mla_flags.bridged && !new_state->exists) 773 batadv_info(bat_priv->soft_iface, 774 "No %s Querier present - multicast optimizations disabled\n", 775 str_proto); 776 777 if (new_state->exists) { 778 if ((!old_state->shadowing && new_state->shadowing) || 779 (!old_state->exists && new_state->shadowing)) 780 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 781 "%s Querier is behind our bridged segment: Might shadow listeners\n", 782 str_proto); 783 else if (old_state->shadowing && !new_state->shadowing) 784 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 785 "%s Querier is not behind our bridged segment\n", 786 str_proto); 787 } 788 } 789 790 /** 791 * batadv_mcast_bridge_log() - debug output for topology changes in bridged 792 * setups 793 * @bat_priv: the bat priv with all the soft interface information 794 * @new_flags: flags indicating the new multicast state 795 * 796 * If no bridges are ever used on this node, then this function does nothing. 797 * 798 * Otherwise this function outputs debug information to the 'mcast' log level 799 * which might be relevant to our multicast optimizations. 800 * 801 * More precisely, it outputs information when a bridge interface is added or 802 * removed from a soft interface. And when a bridge is present, it further 803 * outputs information about the querier state which is relevant for the 804 * multicast flags this node is going to set. 805 */ 806 static void 807 batadv_mcast_bridge_log(struct batadv_priv *bat_priv, 808 struct batadv_mcast_mla_flags *new_flags) 809 { 810 struct batadv_mcast_mla_flags *old_flags = &bat_priv->mcast.mla_flags; 811 812 if (!old_flags->bridged && new_flags->bridged) 813 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 814 "Bridge added: Setting Unsnoopables(U)-flag\n"); 815 else if (old_flags->bridged && !new_flags->bridged) 816 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 817 "Bridge removed: Unsetting Unsnoopables(U)-flag\n"); 818 819 if (new_flags->bridged) { 820 batadv_mcast_querier_log(bat_priv, "IGMP", 821 &old_flags->querier_ipv4, 822 &new_flags->querier_ipv4); 823 batadv_mcast_querier_log(bat_priv, "MLD", 824 &old_flags->querier_ipv6, 825 &new_flags->querier_ipv6); 826 } 827 } 828 829 /** 830 * batadv_mcast_flags_log() - output debug information about mcast flag changes 831 * @bat_priv: the bat priv with all the soft interface information 832 * @flags: TVLV flags indicating the new multicast state 833 * 834 * Whenever the multicast TVLV flags this node announces change, this function 835 * should be used to notify userspace about the change. 836 */ 837 static void batadv_mcast_flags_log(struct batadv_priv *bat_priv, u8 flags) 838 { 839 bool old_enabled = bat_priv->mcast.mla_flags.enabled; 840 u8 old_flags = bat_priv->mcast.mla_flags.tvlv_flags; 841 char str_old_flags[] = "[.... . .]"; 842 843 sprintf(str_old_flags, "[%c%c%c%s%s%c]", 844 (old_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) ? 'U' : '.', 845 (old_flags & BATADV_MCAST_WANT_ALL_IPV4) ? '4' : '.', 846 (old_flags & BATADV_MCAST_WANT_ALL_IPV6) ? '6' : '.', 847 !(old_flags & BATADV_MCAST_WANT_NO_RTR4) ? "R4" : ". ", 848 !(old_flags & BATADV_MCAST_WANT_NO_RTR6) ? "R6" : ". ", 849 !(old_flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA) ? 'P' : '.'); 850 851 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 852 "Changing multicast flags from '%s' to '[%c%c%c%s%s%c]'\n", 853 old_enabled ? str_old_flags : "<undefined>", 854 (flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) ? 'U' : '.', 855 (flags & BATADV_MCAST_WANT_ALL_IPV4) ? '4' : '.', 856 (flags & BATADV_MCAST_WANT_ALL_IPV6) ? '6' : '.', 857 !(flags & BATADV_MCAST_WANT_NO_RTR4) ? "R4" : ". ", 858 !(flags & BATADV_MCAST_WANT_NO_RTR6) ? "R6" : ". ", 859 !(flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA) ? 'P' : '.'); 860 } 861 862 /** 863 * batadv_mcast_mla_flags_update() - update multicast flags 864 * @bat_priv: the bat priv with all the soft interface information 865 * @flags: flags indicating the new multicast state 866 * 867 * Updates the own multicast tvlv with our current multicast related settings, 868 * capabilities and inabilities. 869 */ 870 static void 871 batadv_mcast_mla_flags_update(struct batadv_priv *bat_priv, 872 struct batadv_mcast_mla_flags *flags) 873 { 874 struct batadv_tvlv_mcast_data mcast_data; 875 876 if (!memcmp(flags, &bat_priv->mcast.mla_flags, sizeof(*flags))) 877 return; 878 879 batadv_mcast_bridge_log(bat_priv, flags); 880 batadv_mcast_flags_log(bat_priv, flags->tvlv_flags); 881 882 mcast_data.flags = flags->tvlv_flags; 883 memset(mcast_data.reserved, 0, sizeof(mcast_data.reserved)); 884 885 batadv_tvlv_container_register(bat_priv, BATADV_TVLV_MCAST, 2, 886 &mcast_data, sizeof(mcast_data)); 887 888 bat_priv->mcast.mla_flags = *flags; 889 } 890 891 /** 892 * __batadv_mcast_mla_update() - update the own MLAs 893 * @bat_priv: the bat priv with all the soft interface information 894 * 895 * Updates the own multicast listener announcements in the translation 896 * table as well as the own, announced multicast tvlv container. 897 * 898 * Note that non-conflicting reads and writes to bat_priv->mcast.mla_list 899 * in batadv_mcast_mla_tt_retract() and batadv_mcast_mla_tt_add() are 900 * ensured by the non-parallel execution of the worker this function 901 * belongs to. 902 */ 903 static void __batadv_mcast_mla_update(struct batadv_priv *bat_priv) 904 { 905 struct net_device *soft_iface = bat_priv->soft_iface; 906 struct hlist_head mcast_list = HLIST_HEAD_INIT; 907 struct batadv_mcast_mla_flags flags; 908 int ret; 909 910 flags = batadv_mcast_mla_flags_get(bat_priv); 911 912 ret = batadv_mcast_mla_softif_get(soft_iface, &mcast_list, &flags); 913 if (ret < 0) 914 goto out; 915 916 ret = batadv_mcast_mla_bridge_get(soft_iface, &mcast_list, &flags); 917 if (ret < 0) 918 goto out; 919 920 spin_lock(&bat_priv->mcast.mla_lock); 921 batadv_mcast_mla_tt_retract(bat_priv, &mcast_list); 922 batadv_mcast_mla_tt_add(bat_priv, &mcast_list); 923 batadv_mcast_mla_flags_update(bat_priv, &flags); 924 spin_unlock(&bat_priv->mcast.mla_lock); 925 926 out: 927 batadv_mcast_mla_list_free(&mcast_list); 928 } 929 930 /** 931 * batadv_mcast_mla_update() - update the own MLAs 932 * @work: kernel work struct 933 * 934 * Updates the own multicast listener announcements in the translation 935 * table as well as the own, announced multicast tvlv container. 936 * 937 * In the end, reschedules the work timer. 938 */ 939 static void batadv_mcast_mla_update(struct work_struct *work) 940 { 941 struct delayed_work *delayed_work; 942 struct batadv_priv_mcast *priv_mcast; 943 struct batadv_priv *bat_priv; 944 945 delayed_work = to_delayed_work(work); 946 priv_mcast = container_of(delayed_work, struct batadv_priv_mcast, work); 947 bat_priv = container_of(priv_mcast, struct batadv_priv, mcast); 948 949 __batadv_mcast_mla_update(bat_priv); 950 batadv_mcast_start_timer(bat_priv); 951 } 952 953 /** 954 * batadv_mcast_is_report_ipv4() - check for IGMP reports 955 * @skb: the ethernet frame destined for the mesh 956 * 957 * This call might reallocate skb data. 958 * 959 * Checks whether the given frame is a valid IGMP report. 960 * 961 * Return: If so then true, otherwise false. 962 */ 963 static bool batadv_mcast_is_report_ipv4(struct sk_buff *skb) 964 { 965 if (ip_mc_check_igmp(skb) < 0) 966 return false; 967 968 switch (igmp_hdr(skb)->type) { 969 case IGMP_HOST_MEMBERSHIP_REPORT: 970 case IGMPV2_HOST_MEMBERSHIP_REPORT: 971 case IGMPV3_HOST_MEMBERSHIP_REPORT: 972 return true; 973 } 974 975 return false; 976 } 977 978 /** 979 * batadv_mcast_forw_mode_check_ipv4() - check for optimized forwarding 980 * potential 981 * @bat_priv: the bat priv with all the soft interface information 982 * @skb: the IPv4 packet to check 983 * @is_unsnoopable: stores whether the destination is snoopable 984 * @is_routable: stores whether the destination is routable 985 * 986 * Checks whether the given IPv4 packet has the potential to be forwarded with a 987 * mode more optimal than classic flooding. 988 * 989 * Return: If so then 0. Otherwise -EINVAL or -ENOMEM in case of memory 990 * allocation failure. 991 */ 992 static int batadv_mcast_forw_mode_check_ipv4(struct batadv_priv *bat_priv, 993 struct sk_buff *skb, 994 bool *is_unsnoopable, 995 int *is_routable) 996 { 997 struct iphdr *iphdr; 998 999 /* We might fail due to out-of-memory -> drop it */ 1000 if (!pskb_may_pull(skb, sizeof(struct ethhdr) + sizeof(*iphdr))) 1001 return -ENOMEM; 1002 1003 if (batadv_mcast_is_report_ipv4(skb)) 1004 return -EINVAL; 1005 1006 iphdr = ip_hdr(skb); 1007 1008 /* link-local multicast listeners behind a bridge are 1009 * not snoopable (see RFC4541, section 2.1.2.2) 1010 */ 1011 if (ipv4_is_local_multicast(iphdr->daddr)) 1012 *is_unsnoopable = true; 1013 else 1014 *is_routable = ETH_P_IP; 1015 1016 return 0; 1017 } 1018 1019 /** 1020 * batadv_mcast_is_report_ipv6() - check for MLD reports 1021 * @skb: the ethernet frame destined for the mesh 1022 * 1023 * This call might reallocate skb data. 1024 * 1025 * Checks whether the given frame is a valid MLD report. 1026 * 1027 * Return: If so then true, otherwise false. 1028 */ 1029 static bool batadv_mcast_is_report_ipv6(struct sk_buff *skb) 1030 { 1031 if (ipv6_mc_check_mld(skb) < 0) 1032 return false; 1033 1034 switch (icmp6_hdr(skb)->icmp6_type) { 1035 case ICMPV6_MGM_REPORT: 1036 case ICMPV6_MLD2_REPORT: 1037 return true; 1038 } 1039 1040 return false; 1041 } 1042 1043 /** 1044 * batadv_mcast_forw_mode_check_ipv6() - check for optimized forwarding 1045 * potential 1046 * @bat_priv: the bat priv with all the soft interface information 1047 * @skb: the IPv6 packet to check 1048 * @is_unsnoopable: stores whether the destination is snoopable 1049 * @is_routable: stores whether the destination is routable 1050 * 1051 * Checks whether the given IPv6 packet has the potential to be forwarded with a 1052 * mode more optimal than classic flooding. 1053 * 1054 * Return: If so then 0. Otherwise -EINVAL is or -ENOMEM if we are out of memory 1055 */ 1056 static int batadv_mcast_forw_mode_check_ipv6(struct batadv_priv *bat_priv, 1057 struct sk_buff *skb, 1058 bool *is_unsnoopable, 1059 int *is_routable) 1060 { 1061 struct ipv6hdr *ip6hdr; 1062 1063 /* We might fail due to out-of-memory -> drop it */ 1064 if (!pskb_may_pull(skb, sizeof(struct ethhdr) + sizeof(*ip6hdr))) 1065 return -ENOMEM; 1066 1067 if (batadv_mcast_is_report_ipv6(skb)) 1068 return -EINVAL; 1069 1070 ip6hdr = ipv6_hdr(skb); 1071 1072 if (IPV6_ADDR_MC_SCOPE(&ip6hdr->daddr) < IPV6_ADDR_SCOPE_LINKLOCAL) 1073 return -EINVAL; 1074 1075 /* link-local-all-nodes multicast listeners behind a bridge are 1076 * not snoopable (see RFC4541, section 3, paragraph 3) 1077 */ 1078 if (ipv6_addr_is_ll_all_nodes(&ip6hdr->daddr)) 1079 *is_unsnoopable = true; 1080 else if (IPV6_ADDR_MC_SCOPE(&ip6hdr->daddr) > IPV6_ADDR_SCOPE_LINKLOCAL) 1081 *is_routable = ETH_P_IPV6; 1082 1083 return 0; 1084 } 1085 1086 /** 1087 * batadv_mcast_forw_mode_check() - check for optimized forwarding potential 1088 * @bat_priv: the bat priv with all the soft interface information 1089 * @skb: the multicast frame to check 1090 * @is_unsnoopable: stores whether the destination is snoopable 1091 * @is_routable: stores whether the destination is routable 1092 * 1093 * Checks whether the given multicast ethernet frame has the potential to be 1094 * forwarded with a mode more optimal than classic flooding. 1095 * 1096 * Return: If so then 0. Otherwise -EINVAL is or -ENOMEM if we are out of memory 1097 */ 1098 static int batadv_mcast_forw_mode_check(struct batadv_priv *bat_priv, 1099 struct sk_buff *skb, 1100 bool *is_unsnoopable, 1101 int *is_routable) 1102 { 1103 struct ethhdr *ethhdr = eth_hdr(skb); 1104 1105 if (!atomic_read(&bat_priv->multicast_mode)) 1106 return -EINVAL; 1107 1108 switch (ntohs(ethhdr->h_proto)) { 1109 case ETH_P_IP: 1110 return batadv_mcast_forw_mode_check_ipv4(bat_priv, skb, 1111 is_unsnoopable, 1112 is_routable); 1113 case ETH_P_IPV6: 1114 if (!IS_ENABLED(CONFIG_IPV6)) 1115 return -EINVAL; 1116 1117 return batadv_mcast_forw_mode_check_ipv6(bat_priv, skb, 1118 is_unsnoopable, 1119 is_routable); 1120 default: 1121 return -EINVAL; 1122 } 1123 } 1124 1125 /** 1126 * batadv_mcast_forw_want_all_ip_count() - count nodes with unspecific mcast 1127 * interest 1128 * @bat_priv: the bat priv with all the soft interface information 1129 * @ethhdr: ethernet header of a packet 1130 * 1131 * Return: the number of nodes which want all IPv4 multicast traffic if the 1132 * given ethhdr is from an IPv4 packet or the number of nodes which want all 1133 * IPv6 traffic if it matches an IPv6 packet. 1134 */ 1135 static int batadv_mcast_forw_want_all_ip_count(struct batadv_priv *bat_priv, 1136 struct ethhdr *ethhdr) 1137 { 1138 switch (ntohs(ethhdr->h_proto)) { 1139 case ETH_P_IP: 1140 return atomic_read(&bat_priv->mcast.num_want_all_ipv4); 1141 case ETH_P_IPV6: 1142 return atomic_read(&bat_priv->mcast.num_want_all_ipv6); 1143 default: 1144 /* we shouldn't be here... */ 1145 return 0; 1146 } 1147 } 1148 1149 /** 1150 * batadv_mcast_forw_rtr_count() - count nodes with a multicast router 1151 * @bat_priv: the bat priv with all the soft interface information 1152 * @protocol: the ethernet protocol type to count multicast routers for 1153 * 1154 * Return: the number of nodes which want all routable IPv4 multicast traffic 1155 * if the protocol is ETH_P_IP or the number of nodes which want all routable 1156 * IPv6 traffic if the protocol is ETH_P_IPV6. Otherwise returns 0. 1157 */ 1158 1159 static int batadv_mcast_forw_rtr_count(struct batadv_priv *bat_priv, 1160 int protocol) 1161 { 1162 switch (protocol) { 1163 case ETH_P_IP: 1164 return atomic_read(&bat_priv->mcast.num_want_all_rtr4); 1165 case ETH_P_IPV6: 1166 return atomic_read(&bat_priv->mcast.num_want_all_rtr6); 1167 default: 1168 return 0; 1169 } 1170 } 1171 1172 /** 1173 * batadv_mcast_forw_mode_by_count() - get forwarding mode by count 1174 * @bat_priv: the bat priv with all the soft interface information 1175 * @skb: the multicast packet to check 1176 * @vid: the vlan identifier 1177 * @is_routable: stores whether the destination is routable 1178 * @count: the number of originators the multicast packet need to be sent to 1179 * 1180 * For a multicast packet with multiple destination originators, checks which 1181 * mode to use. For BATADV_FORW_MCAST it also encapsulates the packet with a 1182 * complete batman-adv multicast header. 1183 * 1184 * Return: 1185 * BATADV_FORW_MCAST: If all nodes have multicast packet routing 1186 * capabilities and an MTU >= 1280 on all hard interfaces (including us) 1187 * and the encapsulated multicast packet with all destination addresses 1188 * would still fit into an 1280 bytes batman-adv multicast packet 1189 * (excluding the outer ethernet frame) and we could successfully push 1190 * the full batman-adv multicast packet header. 1191 * BATADV_FORW_UCASTS: If the packet cannot be sent in a batman-adv 1192 * multicast packet and the amount of batman-adv unicast packets needed 1193 * is smaller or equal to the configured multicast fanout. 1194 * BATADV_FORW_BCAST: Otherwise. 1195 */ 1196 static enum batadv_forw_mode 1197 batadv_mcast_forw_mode_by_count(struct batadv_priv *bat_priv, 1198 struct sk_buff *skb, unsigned short vid, 1199 int is_routable, int count) 1200 { 1201 unsigned int mcast_hdrlen = batadv_mcast_forw_packet_hdrlen(count); 1202 u8 own_tvlv_flags = bat_priv->mcast.mla_flags.tvlv_flags; 1203 1204 if (!atomic_read(&bat_priv->mcast.num_no_mc_ptype_capa) && 1205 own_tvlv_flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA && 1206 skb->len + mcast_hdrlen <= IPV6_MIN_MTU && 1207 batadv_mcast_forw_push(bat_priv, skb, vid, is_routable, count)) 1208 return BATADV_FORW_MCAST; 1209 1210 if (count <= atomic_read(&bat_priv->multicast_fanout)) 1211 return BATADV_FORW_UCASTS; 1212 1213 return BATADV_FORW_BCAST; 1214 } 1215 1216 /** 1217 * batadv_mcast_forw_mode() - check on how to forward a multicast packet 1218 * @bat_priv: the bat priv with all the soft interface information 1219 * @skb: the multicast packet to check 1220 * @vid: the vlan identifier 1221 * @is_routable: stores whether the destination is routable 1222 * 1223 * Return: The forwarding mode as enum batadv_forw_mode. 1224 */ 1225 enum batadv_forw_mode 1226 batadv_mcast_forw_mode(struct batadv_priv *bat_priv, struct sk_buff *skb, 1227 unsigned short vid, int *is_routable) 1228 { 1229 int ret, tt_count, ip_count, unsnoop_count, total_count; 1230 bool is_unsnoopable = false; 1231 struct ethhdr *ethhdr; 1232 int rtr_count = 0; 1233 1234 ret = batadv_mcast_forw_mode_check(bat_priv, skb, &is_unsnoopable, 1235 is_routable); 1236 if (ret == -ENOMEM) 1237 return BATADV_FORW_NONE; 1238 else if (ret < 0) 1239 return BATADV_FORW_BCAST; 1240 1241 ethhdr = eth_hdr(skb); 1242 1243 tt_count = batadv_tt_global_hash_count(bat_priv, ethhdr->h_dest, 1244 BATADV_NO_FLAGS); 1245 ip_count = batadv_mcast_forw_want_all_ip_count(bat_priv, ethhdr); 1246 unsnoop_count = !is_unsnoopable ? 0 : 1247 atomic_read(&bat_priv->mcast.num_want_all_unsnoopables); 1248 rtr_count = batadv_mcast_forw_rtr_count(bat_priv, *is_routable); 1249 1250 total_count = tt_count + ip_count + unsnoop_count + rtr_count; 1251 1252 if (!total_count) 1253 return BATADV_FORW_NONE; 1254 else if (unsnoop_count) 1255 return BATADV_FORW_BCAST; 1256 1257 return batadv_mcast_forw_mode_by_count(bat_priv, skb, vid, *is_routable, 1258 total_count); 1259 } 1260 1261 /** 1262 * batadv_mcast_forw_send_orig() - send a multicast packet to an originator 1263 * @bat_priv: the bat priv with all the soft interface information 1264 * @skb: the multicast packet to send 1265 * @vid: the vlan identifier 1266 * @orig_node: the originator to send the packet to 1267 * 1268 * Return: NET_XMIT_DROP in case of error or NET_XMIT_SUCCESS otherwise. 1269 */ 1270 static int batadv_mcast_forw_send_orig(struct batadv_priv *bat_priv, 1271 struct sk_buff *skb, 1272 unsigned short vid, 1273 struct batadv_orig_node *orig_node) 1274 { 1275 /* Avoid sending multicast-in-unicast packets to other BLA 1276 * gateways - they already got the frame from the LAN side 1277 * we share with them. 1278 * TODO: Refactor to take BLA into account earlier, to avoid 1279 * reducing the mcast_fanout count. 1280 */ 1281 if (batadv_bla_is_backbone_gw_orig(bat_priv, orig_node->orig, vid)) { 1282 dev_kfree_skb(skb); 1283 return NET_XMIT_SUCCESS; 1284 } 1285 1286 return batadv_send_skb_unicast(bat_priv, skb, BATADV_UNICAST, 0, 1287 orig_node, vid); 1288 } 1289 1290 /** 1291 * batadv_mcast_forw_tt() - forwards a packet to multicast listeners 1292 * @bat_priv: the bat priv with all the soft interface information 1293 * @skb: the multicast packet to transmit 1294 * @vid: the vlan identifier 1295 * 1296 * Sends copies of a frame with multicast destination to any multicast 1297 * listener registered in the translation table. A transmission is performed 1298 * via a batman-adv unicast packet for each such destination node. 1299 * 1300 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1301 * otherwise. 1302 */ 1303 static int 1304 batadv_mcast_forw_tt(struct batadv_priv *bat_priv, struct sk_buff *skb, 1305 unsigned short vid) 1306 { 1307 int ret = NET_XMIT_SUCCESS; 1308 struct sk_buff *newskb; 1309 1310 struct batadv_tt_orig_list_entry *orig_entry; 1311 1312 struct batadv_tt_global_entry *tt_global; 1313 const u8 *addr = eth_hdr(skb)->h_dest; 1314 1315 tt_global = batadv_tt_global_hash_find(bat_priv, addr, vid); 1316 if (!tt_global) 1317 goto out; 1318 1319 rcu_read_lock(); 1320 hlist_for_each_entry_rcu(orig_entry, &tt_global->orig_list, list) { 1321 newskb = skb_copy(skb, GFP_ATOMIC); 1322 if (!newskb) { 1323 ret = NET_XMIT_DROP; 1324 break; 1325 } 1326 1327 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, 1328 orig_entry->orig_node); 1329 } 1330 rcu_read_unlock(); 1331 1332 batadv_tt_global_entry_put(tt_global); 1333 1334 out: 1335 return ret; 1336 } 1337 1338 /** 1339 * batadv_mcast_forw_want_all_ipv4() - forward to nodes with want-all-ipv4 1340 * @bat_priv: the bat priv with all the soft interface information 1341 * @skb: the multicast packet to transmit 1342 * @vid: the vlan identifier 1343 * 1344 * Sends copies of a frame with multicast destination to any node with a 1345 * BATADV_MCAST_WANT_ALL_IPV4 flag set. A transmission is performed via a 1346 * batman-adv unicast packet for each such destination node. 1347 * 1348 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1349 * otherwise. 1350 */ 1351 static int 1352 batadv_mcast_forw_want_all_ipv4(struct batadv_priv *bat_priv, 1353 struct sk_buff *skb, unsigned short vid) 1354 { 1355 struct batadv_orig_node *orig_node; 1356 int ret = NET_XMIT_SUCCESS; 1357 struct sk_buff *newskb; 1358 1359 rcu_read_lock(); 1360 hlist_for_each_entry_rcu(orig_node, 1361 &bat_priv->mcast.want_all_ipv4_list, 1362 mcast_want_all_ipv4_node) { 1363 newskb = skb_copy(skb, GFP_ATOMIC); 1364 if (!newskb) { 1365 ret = NET_XMIT_DROP; 1366 break; 1367 } 1368 1369 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1370 } 1371 rcu_read_unlock(); 1372 return ret; 1373 } 1374 1375 /** 1376 * batadv_mcast_forw_want_all_ipv6() - forward to nodes with want-all-ipv6 1377 * @bat_priv: the bat priv with all the soft interface information 1378 * @skb: The multicast packet to transmit 1379 * @vid: the vlan identifier 1380 * 1381 * Sends copies of a frame with multicast destination to any node with a 1382 * BATADV_MCAST_WANT_ALL_IPV6 flag set. A transmission is performed via a 1383 * batman-adv unicast packet for each such destination node. 1384 * 1385 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1386 * otherwise. 1387 */ 1388 static int 1389 batadv_mcast_forw_want_all_ipv6(struct batadv_priv *bat_priv, 1390 struct sk_buff *skb, unsigned short vid) 1391 { 1392 struct batadv_orig_node *orig_node; 1393 int ret = NET_XMIT_SUCCESS; 1394 struct sk_buff *newskb; 1395 1396 rcu_read_lock(); 1397 hlist_for_each_entry_rcu(orig_node, 1398 &bat_priv->mcast.want_all_ipv6_list, 1399 mcast_want_all_ipv6_node) { 1400 newskb = skb_copy(skb, GFP_ATOMIC); 1401 if (!newskb) { 1402 ret = NET_XMIT_DROP; 1403 break; 1404 } 1405 1406 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1407 } 1408 rcu_read_unlock(); 1409 return ret; 1410 } 1411 1412 /** 1413 * batadv_mcast_forw_want_all() - forward packet to nodes in a want-all list 1414 * @bat_priv: the bat priv with all the soft interface information 1415 * @skb: the multicast packet to transmit 1416 * @vid: the vlan identifier 1417 * 1418 * Sends copies of a frame with multicast destination to any node with a 1419 * BATADV_MCAST_WANT_ALL_IPV4 or BATADV_MCAST_WANT_ALL_IPV6 flag set. A 1420 * transmission is performed via a batman-adv unicast packet for each such 1421 * destination node. 1422 * 1423 * Return: NET_XMIT_DROP on memory allocation failure or if the protocol family 1424 * is neither IPv4 nor IPv6. NET_XMIT_SUCCESS otherwise. 1425 */ 1426 static int 1427 batadv_mcast_forw_want_all(struct batadv_priv *bat_priv, 1428 struct sk_buff *skb, unsigned short vid) 1429 { 1430 switch (ntohs(eth_hdr(skb)->h_proto)) { 1431 case ETH_P_IP: 1432 return batadv_mcast_forw_want_all_ipv4(bat_priv, skb, vid); 1433 case ETH_P_IPV6: 1434 return batadv_mcast_forw_want_all_ipv6(bat_priv, skb, vid); 1435 default: 1436 /* we shouldn't be here... */ 1437 return NET_XMIT_DROP; 1438 } 1439 } 1440 1441 /** 1442 * batadv_mcast_forw_want_all_rtr4() - forward to nodes with want-all-rtr4 1443 * @bat_priv: the bat priv with all the soft interface information 1444 * @skb: the multicast packet to transmit 1445 * @vid: the vlan identifier 1446 * 1447 * Sends copies of a frame with multicast destination to any node with a 1448 * BATADV_MCAST_WANT_NO_RTR4 flag unset. A transmission is performed via a 1449 * batman-adv unicast packet for each such destination node. 1450 * 1451 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1452 * otherwise. 1453 */ 1454 static int 1455 batadv_mcast_forw_want_all_rtr4(struct batadv_priv *bat_priv, 1456 struct sk_buff *skb, unsigned short vid) 1457 { 1458 struct batadv_orig_node *orig_node; 1459 int ret = NET_XMIT_SUCCESS; 1460 struct sk_buff *newskb; 1461 1462 rcu_read_lock(); 1463 hlist_for_each_entry_rcu(orig_node, 1464 &bat_priv->mcast.want_all_rtr4_list, 1465 mcast_want_all_rtr4_node) { 1466 newskb = skb_copy(skb, GFP_ATOMIC); 1467 if (!newskb) { 1468 ret = NET_XMIT_DROP; 1469 break; 1470 } 1471 1472 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1473 } 1474 rcu_read_unlock(); 1475 return ret; 1476 } 1477 1478 /** 1479 * batadv_mcast_forw_want_all_rtr6() - forward to nodes with want-all-rtr6 1480 * @bat_priv: the bat priv with all the soft interface information 1481 * @skb: The multicast packet to transmit 1482 * @vid: the vlan identifier 1483 * 1484 * Sends copies of a frame with multicast destination to any node with a 1485 * BATADV_MCAST_WANT_NO_RTR6 flag unset. A transmission is performed via a 1486 * batman-adv unicast packet for each such destination node. 1487 * 1488 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1489 * otherwise. 1490 */ 1491 static int 1492 batadv_mcast_forw_want_all_rtr6(struct batadv_priv *bat_priv, 1493 struct sk_buff *skb, unsigned short vid) 1494 { 1495 struct batadv_orig_node *orig_node; 1496 int ret = NET_XMIT_SUCCESS; 1497 struct sk_buff *newskb; 1498 1499 rcu_read_lock(); 1500 hlist_for_each_entry_rcu(orig_node, 1501 &bat_priv->mcast.want_all_rtr6_list, 1502 mcast_want_all_rtr6_node) { 1503 newskb = skb_copy(skb, GFP_ATOMIC); 1504 if (!newskb) { 1505 ret = NET_XMIT_DROP; 1506 break; 1507 } 1508 1509 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1510 } 1511 rcu_read_unlock(); 1512 return ret; 1513 } 1514 1515 /** 1516 * batadv_mcast_forw_want_rtr() - forward packet to nodes in a want-all-rtr list 1517 * @bat_priv: the bat priv with all the soft interface information 1518 * @skb: the multicast packet to transmit 1519 * @vid: the vlan identifier 1520 * 1521 * Sends copies of a frame with multicast destination to any node with a 1522 * BATADV_MCAST_WANT_NO_RTR4 or BATADV_MCAST_WANT_NO_RTR6 flag unset. A 1523 * transmission is performed via a batman-adv unicast packet for each such 1524 * destination node. 1525 * 1526 * Return: NET_XMIT_DROP on memory allocation failure or if the protocol family 1527 * is neither IPv4 nor IPv6. NET_XMIT_SUCCESS otherwise. 1528 */ 1529 static int 1530 batadv_mcast_forw_want_rtr(struct batadv_priv *bat_priv, 1531 struct sk_buff *skb, unsigned short vid) 1532 { 1533 switch (ntohs(eth_hdr(skb)->h_proto)) { 1534 case ETH_P_IP: 1535 return batadv_mcast_forw_want_all_rtr4(bat_priv, skb, vid); 1536 case ETH_P_IPV6: 1537 return batadv_mcast_forw_want_all_rtr6(bat_priv, skb, vid); 1538 default: 1539 /* we shouldn't be here... */ 1540 return NET_XMIT_DROP; 1541 } 1542 } 1543 1544 /** 1545 * batadv_mcast_forw_send() - send packet to any detected multicast recipient 1546 * @bat_priv: the bat priv with all the soft interface information 1547 * @skb: the multicast packet to transmit 1548 * @vid: the vlan identifier 1549 * @is_routable: stores whether the destination is routable 1550 * 1551 * Sends copies of a frame with multicast destination to any node that signaled 1552 * interest in it, that is either via the translation table or the according 1553 * want-all flags. A transmission is performed via a batman-adv unicast packet 1554 * for each such destination node. 1555 * 1556 * The given skb is consumed/freed. 1557 * 1558 * Return: NET_XMIT_DROP on memory allocation failure or if the protocol family 1559 * is neither IPv4 nor IPv6. NET_XMIT_SUCCESS otherwise. 1560 */ 1561 int batadv_mcast_forw_send(struct batadv_priv *bat_priv, struct sk_buff *skb, 1562 unsigned short vid, int is_routable) 1563 { 1564 int ret; 1565 1566 ret = batadv_mcast_forw_tt(bat_priv, skb, vid); 1567 if (ret != NET_XMIT_SUCCESS) { 1568 kfree_skb(skb); 1569 return ret; 1570 } 1571 1572 ret = batadv_mcast_forw_want_all(bat_priv, skb, vid); 1573 if (ret != NET_XMIT_SUCCESS) { 1574 kfree_skb(skb); 1575 return ret; 1576 } 1577 1578 if (!is_routable) 1579 goto skip_mc_router; 1580 1581 ret = batadv_mcast_forw_want_rtr(bat_priv, skb, vid); 1582 if (ret != NET_XMIT_SUCCESS) { 1583 kfree_skb(skb); 1584 return ret; 1585 } 1586 1587 skip_mc_router: 1588 consume_skb(skb); 1589 return ret; 1590 } 1591 1592 /** 1593 * batadv_mcast_want_unsnoop_update() - update unsnoop counter and list 1594 * @bat_priv: the bat priv with all the soft interface information 1595 * @orig: the orig_node which multicast state might have changed of 1596 * @mcast_flags: flags indicating the new multicast state 1597 * 1598 * If the BATADV_MCAST_WANT_ALL_UNSNOOPABLES flag of this originator, 1599 * orig, has toggled then this method updates the counter and the list 1600 * accordingly. 1601 * 1602 * Caller needs to hold orig->mcast_handler_lock. 1603 */ 1604 static void batadv_mcast_want_unsnoop_update(struct batadv_priv *bat_priv, 1605 struct batadv_orig_node *orig, 1606 u8 mcast_flags) 1607 { 1608 struct hlist_node *node = &orig->mcast_want_all_unsnoopables_node; 1609 struct hlist_head *head = &bat_priv->mcast.want_all_unsnoopables_list; 1610 1611 lockdep_assert_held(&orig->mcast_handler_lock); 1612 1613 /* switched from flag unset to set */ 1614 if (mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 1615 !(orig->mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES)) { 1616 atomic_inc(&bat_priv->mcast.num_want_all_unsnoopables); 1617 1618 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1619 /* flag checks above + mcast_handler_lock prevents this */ 1620 WARN_ON(!hlist_unhashed(node)); 1621 1622 hlist_add_head_rcu(node, head); 1623 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1624 /* switched from flag set to unset */ 1625 } else if (!(mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) && 1626 orig->mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) { 1627 atomic_dec(&bat_priv->mcast.num_want_all_unsnoopables); 1628 1629 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1630 /* flag checks above + mcast_handler_lock prevents this */ 1631 WARN_ON(hlist_unhashed(node)); 1632 1633 hlist_del_init_rcu(node); 1634 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1635 } 1636 } 1637 1638 /** 1639 * batadv_mcast_want_ipv4_update() - update want-all-ipv4 counter and list 1640 * @bat_priv: the bat priv with all the soft interface information 1641 * @orig: the orig_node which multicast state might have changed of 1642 * @mcast_flags: flags indicating the new multicast state 1643 * 1644 * If the BATADV_MCAST_WANT_ALL_IPV4 flag of this originator, orig, has 1645 * toggled then this method updates the counter and the list accordingly. 1646 * 1647 * Caller needs to hold orig->mcast_handler_lock. 1648 */ 1649 static void batadv_mcast_want_ipv4_update(struct batadv_priv *bat_priv, 1650 struct batadv_orig_node *orig, 1651 u8 mcast_flags) 1652 { 1653 struct hlist_node *node = &orig->mcast_want_all_ipv4_node; 1654 struct hlist_head *head = &bat_priv->mcast.want_all_ipv4_list; 1655 1656 lockdep_assert_held(&orig->mcast_handler_lock); 1657 1658 /* switched from flag unset to set */ 1659 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV4 && 1660 !(orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV4)) { 1661 atomic_inc(&bat_priv->mcast.num_want_all_ipv4); 1662 1663 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1664 /* flag checks above + mcast_handler_lock prevents this */ 1665 WARN_ON(!hlist_unhashed(node)); 1666 1667 hlist_add_head_rcu(node, head); 1668 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1669 /* switched from flag set to unset */ 1670 } else if (!(mcast_flags & BATADV_MCAST_WANT_ALL_IPV4) && 1671 orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV4) { 1672 atomic_dec(&bat_priv->mcast.num_want_all_ipv4); 1673 1674 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1675 /* flag checks above + mcast_handler_lock prevents this */ 1676 WARN_ON(hlist_unhashed(node)); 1677 1678 hlist_del_init_rcu(node); 1679 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1680 } 1681 } 1682 1683 /** 1684 * batadv_mcast_want_ipv6_update() - update want-all-ipv6 counter and list 1685 * @bat_priv: the bat priv with all the soft interface information 1686 * @orig: the orig_node which multicast state might have changed of 1687 * @mcast_flags: flags indicating the new multicast state 1688 * 1689 * If the BATADV_MCAST_WANT_ALL_IPV6 flag of this originator, orig, has 1690 * toggled then this method updates the counter and the list accordingly. 1691 * 1692 * Caller needs to hold orig->mcast_handler_lock. 1693 */ 1694 static void batadv_mcast_want_ipv6_update(struct batadv_priv *bat_priv, 1695 struct batadv_orig_node *orig, 1696 u8 mcast_flags) 1697 { 1698 struct hlist_node *node = &orig->mcast_want_all_ipv6_node; 1699 struct hlist_head *head = &bat_priv->mcast.want_all_ipv6_list; 1700 1701 lockdep_assert_held(&orig->mcast_handler_lock); 1702 1703 /* switched from flag unset to set */ 1704 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV6 && 1705 !(orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV6)) { 1706 atomic_inc(&bat_priv->mcast.num_want_all_ipv6); 1707 1708 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1709 /* flag checks above + mcast_handler_lock prevents this */ 1710 WARN_ON(!hlist_unhashed(node)); 1711 1712 hlist_add_head_rcu(node, head); 1713 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1714 /* switched from flag set to unset */ 1715 } else if (!(mcast_flags & BATADV_MCAST_WANT_ALL_IPV6) && 1716 orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV6) { 1717 atomic_dec(&bat_priv->mcast.num_want_all_ipv6); 1718 1719 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1720 /* flag checks above + mcast_handler_lock prevents this */ 1721 WARN_ON(hlist_unhashed(node)); 1722 1723 hlist_del_init_rcu(node); 1724 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1725 } 1726 } 1727 1728 /** 1729 * batadv_mcast_want_rtr4_update() - update want-all-rtr4 counter and list 1730 * @bat_priv: the bat priv with all the soft interface information 1731 * @orig: the orig_node which multicast state might have changed of 1732 * @mcast_flags: flags indicating the new multicast state 1733 * 1734 * If the BATADV_MCAST_WANT_NO_RTR4 flag of this originator, orig, has 1735 * toggled then this method updates the counter and the list accordingly. 1736 * 1737 * Caller needs to hold orig->mcast_handler_lock. 1738 */ 1739 static void batadv_mcast_want_rtr4_update(struct batadv_priv *bat_priv, 1740 struct batadv_orig_node *orig, 1741 u8 mcast_flags) 1742 { 1743 struct hlist_node *node = &orig->mcast_want_all_rtr4_node; 1744 struct hlist_head *head = &bat_priv->mcast.want_all_rtr4_list; 1745 1746 lockdep_assert_held(&orig->mcast_handler_lock); 1747 1748 /* switched from flag set to unset */ 1749 if (!(mcast_flags & BATADV_MCAST_WANT_NO_RTR4) && 1750 orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR4) { 1751 atomic_inc(&bat_priv->mcast.num_want_all_rtr4); 1752 1753 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1754 /* flag checks above + mcast_handler_lock prevents this */ 1755 WARN_ON(!hlist_unhashed(node)); 1756 1757 hlist_add_head_rcu(node, head); 1758 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1759 /* switched from flag unset to set */ 1760 } else if (mcast_flags & BATADV_MCAST_WANT_NO_RTR4 && 1761 !(orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR4)) { 1762 atomic_dec(&bat_priv->mcast.num_want_all_rtr4); 1763 1764 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1765 /* flag checks above + mcast_handler_lock prevents this */ 1766 WARN_ON(hlist_unhashed(node)); 1767 1768 hlist_del_init_rcu(node); 1769 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1770 } 1771 } 1772 1773 /** 1774 * batadv_mcast_want_rtr6_update() - update want-all-rtr6 counter and list 1775 * @bat_priv: the bat priv with all the soft interface information 1776 * @orig: the orig_node which multicast state might have changed of 1777 * @mcast_flags: flags indicating the new multicast state 1778 * 1779 * If the BATADV_MCAST_WANT_NO_RTR6 flag of this originator, orig, has 1780 * toggled then this method updates the counter and the list accordingly. 1781 * 1782 * Caller needs to hold orig->mcast_handler_lock. 1783 */ 1784 static void batadv_mcast_want_rtr6_update(struct batadv_priv *bat_priv, 1785 struct batadv_orig_node *orig, 1786 u8 mcast_flags) 1787 { 1788 struct hlist_node *node = &orig->mcast_want_all_rtr6_node; 1789 struct hlist_head *head = &bat_priv->mcast.want_all_rtr6_list; 1790 1791 lockdep_assert_held(&orig->mcast_handler_lock); 1792 1793 /* switched from flag set to unset */ 1794 if (!(mcast_flags & BATADV_MCAST_WANT_NO_RTR6) && 1795 orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR6) { 1796 atomic_inc(&bat_priv->mcast.num_want_all_rtr6); 1797 1798 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1799 /* flag checks above + mcast_handler_lock prevents this */ 1800 WARN_ON(!hlist_unhashed(node)); 1801 1802 hlist_add_head_rcu(node, head); 1803 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1804 /* switched from flag unset to set */ 1805 } else if (mcast_flags & BATADV_MCAST_WANT_NO_RTR6 && 1806 !(orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR6)) { 1807 atomic_dec(&bat_priv->mcast.num_want_all_rtr6); 1808 1809 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1810 /* flag checks above + mcast_handler_lock prevents this */ 1811 WARN_ON(hlist_unhashed(node)); 1812 1813 hlist_del_init_rcu(node); 1814 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1815 } 1816 } 1817 1818 /** 1819 * batadv_mcast_have_mc_ptype_update() - update multicast packet type counter 1820 * @bat_priv: the bat priv with all the soft interface information 1821 * @orig: the orig_node which multicast state might have changed of 1822 * @mcast_flags: flags indicating the new multicast state 1823 * 1824 * If the BATADV_MCAST_HAVE_MC_PTYPE_CAPA flag of this originator, orig, has 1825 * toggled then this method updates the counter accordingly. 1826 */ 1827 static void batadv_mcast_have_mc_ptype_update(struct batadv_priv *bat_priv, 1828 struct batadv_orig_node *orig, 1829 u8 mcast_flags) 1830 { 1831 lockdep_assert_held(&orig->mcast_handler_lock); 1832 1833 /* switched from flag set to unset */ 1834 if (!(mcast_flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA) && 1835 orig->mcast_flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA) 1836 atomic_inc(&bat_priv->mcast.num_no_mc_ptype_capa); 1837 /* switched from flag unset to set */ 1838 else if (mcast_flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA && 1839 !(orig->mcast_flags & BATADV_MCAST_HAVE_MC_PTYPE_CAPA)) 1840 atomic_dec(&bat_priv->mcast.num_no_mc_ptype_capa); 1841 } 1842 1843 /** 1844 * batadv_mcast_tvlv_flags_get() - get multicast flags from an OGM TVLV 1845 * @enabled: whether the originator has multicast TVLV support enabled 1846 * @tvlv_value: tvlv buffer containing the multicast flags 1847 * @tvlv_value_len: tvlv buffer length 1848 * 1849 * Return: multicast flags for the given tvlv buffer 1850 */ 1851 static u8 1852 batadv_mcast_tvlv_flags_get(bool enabled, void *tvlv_value, u16 tvlv_value_len) 1853 { 1854 u8 mcast_flags = BATADV_NO_FLAGS; 1855 1856 if (enabled && tvlv_value && tvlv_value_len >= sizeof(mcast_flags)) 1857 mcast_flags = *(u8 *)tvlv_value; 1858 1859 if (!enabled) { 1860 mcast_flags |= BATADV_MCAST_WANT_ALL_IPV4; 1861 mcast_flags |= BATADV_MCAST_WANT_ALL_IPV6; 1862 } 1863 1864 /* remove redundant flags to avoid sending duplicate packets later */ 1865 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV4) 1866 mcast_flags |= BATADV_MCAST_WANT_NO_RTR4; 1867 1868 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV6) 1869 mcast_flags |= BATADV_MCAST_WANT_NO_RTR6; 1870 1871 return mcast_flags; 1872 } 1873 1874 /** 1875 * batadv_mcast_tvlv_ogm_handler() - process incoming multicast tvlv container 1876 * @bat_priv: the bat priv with all the soft interface information 1877 * @orig: the orig_node of the ogm 1878 * @flags: flags indicating the tvlv state (see batadv_tvlv_handler_flags) 1879 * @tvlv_value: tvlv buffer containing the multicast data 1880 * @tvlv_value_len: tvlv buffer length 1881 */ 1882 static void batadv_mcast_tvlv_ogm_handler(struct batadv_priv *bat_priv, 1883 struct batadv_orig_node *orig, 1884 u8 flags, 1885 void *tvlv_value, 1886 u16 tvlv_value_len) 1887 { 1888 bool orig_mcast_enabled = !(flags & BATADV_TVLV_HANDLER_OGM_CIFNOTFND); 1889 u8 mcast_flags; 1890 1891 mcast_flags = batadv_mcast_tvlv_flags_get(orig_mcast_enabled, 1892 tvlv_value, tvlv_value_len); 1893 1894 spin_lock_bh(&orig->mcast_handler_lock); 1895 1896 if (orig_mcast_enabled && 1897 !test_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities)) { 1898 set_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities); 1899 } else if (!orig_mcast_enabled && 1900 test_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities)) { 1901 clear_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities); 1902 } 1903 1904 set_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capa_initialized); 1905 1906 batadv_mcast_want_unsnoop_update(bat_priv, orig, mcast_flags); 1907 batadv_mcast_want_ipv4_update(bat_priv, orig, mcast_flags); 1908 batadv_mcast_want_ipv6_update(bat_priv, orig, mcast_flags); 1909 batadv_mcast_want_rtr4_update(bat_priv, orig, mcast_flags); 1910 batadv_mcast_want_rtr6_update(bat_priv, orig, mcast_flags); 1911 batadv_mcast_have_mc_ptype_update(bat_priv, orig, mcast_flags); 1912 1913 orig->mcast_flags = mcast_flags; 1914 spin_unlock_bh(&orig->mcast_handler_lock); 1915 } 1916 1917 /** 1918 * batadv_mcast_init() - initialize the multicast optimizations structures 1919 * @bat_priv: the bat priv with all the soft interface information 1920 */ 1921 void batadv_mcast_init(struct batadv_priv *bat_priv) 1922 { 1923 batadv_tvlv_handler_register(bat_priv, batadv_mcast_tvlv_ogm_handler, 1924 NULL, NULL, BATADV_TVLV_MCAST, 2, 1925 BATADV_TVLV_HANDLER_OGM_CIFNOTFND); 1926 batadv_tvlv_handler_register(bat_priv, NULL, NULL, 1927 batadv_mcast_forw_tracker_tvlv_handler, 1928 BATADV_TVLV_MCAST_TRACKER, 1, 1929 BATADV_TVLV_HANDLER_OGM_CIFNOTFND); 1930 1931 INIT_DELAYED_WORK(&bat_priv->mcast.work, batadv_mcast_mla_update); 1932 batadv_mcast_start_timer(bat_priv); 1933 } 1934 1935 /** 1936 * batadv_mcast_mesh_info_put() - put multicast info into a netlink message 1937 * @msg: buffer for the message 1938 * @bat_priv: the bat priv with all the soft interface information 1939 * 1940 * Return: 0 or error code. 1941 */ 1942 int batadv_mcast_mesh_info_put(struct sk_buff *msg, 1943 struct batadv_priv *bat_priv) 1944 { 1945 u32 flags = bat_priv->mcast.mla_flags.tvlv_flags; 1946 u32 flags_priv = BATADV_NO_FLAGS; 1947 1948 if (bat_priv->mcast.mla_flags.bridged) { 1949 flags_priv |= BATADV_MCAST_FLAGS_BRIDGED; 1950 1951 if (bat_priv->mcast.mla_flags.querier_ipv4.exists) 1952 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV4_EXISTS; 1953 if (bat_priv->mcast.mla_flags.querier_ipv6.exists) 1954 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV6_EXISTS; 1955 if (bat_priv->mcast.mla_flags.querier_ipv4.shadowing) 1956 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV4_SHADOWING; 1957 if (bat_priv->mcast.mla_flags.querier_ipv6.shadowing) 1958 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV6_SHADOWING; 1959 } 1960 1961 if (nla_put_u32(msg, BATADV_ATTR_MCAST_FLAGS, flags) || 1962 nla_put_u32(msg, BATADV_ATTR_MCAST_FLAGS_PRIV, flags_priv)) 1963 return -EMSGSIZE; 1964 1965 return 0; 1966 } 1967 1968 /** 1969 * batadv_mcast_flags_dump_entry() - dump one entry of the multicast flags table 1970 * to a netlink socket 1971 * @msg: buffer for the message 1972 * @portid: netlink port 1973 * @cb: Control block containing additional options 1974 * @orig_node: originator to dump the multicast flags of 1975 * 1976 * Return: 0 or error code. 1977 */ 1978 static int 1979 batadv_mcast_flags_dump_entry(struct sk_buff *msg, u32 portid, 1980 struct netlink_callback *cb, 1981 struct batadv_orig_node *orig_node) 1982 { 1983 void *hdr; 1984 1985 hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, 1986 &batadv_netlink_family, NLM_F_MULTI, 1987 BATADV_CMD_GET_MCAST_FLAGS); 1988 if (!hdr) 1989 return -ENOBUFS; 1990 1991 genl_dump_check_consistent(cb, hdr); 1992 1993 if (nla_put(msg, BATADV_ATTR_ORIG_ADDRESS, ETH_ALEN, 1994 orig_node->orig)) { 1995 genlmsg_cancel(msg, hdr); 1996 return -EMSGSIZE; 1997 } 1998 1999 if (test_bit(BATADV_ORIG_CAPA_HAS_MCAST, 2000 &orig_node->capabilities)) { 2001 if (nla_put_u32(msg, BATADV_ATTR_MCAST_FLAGS, 2002 orig_node->mcast_flags)) { 2003 genlmsg_cancel(msg, hdr); 2004 return -EMSGSIZE; 2005 } 2006 } 2007 2008 genlmsg_end(msg, hdr); 2009 return 0; 2010 } 2011 2012 /** 2013 * batadv_mcast_flags_dump_bucket() - dump one bucket of the multicast flags 2014 * table to a netlink socket 2015 * @msg: buffer for the message 2016 * @portid: netlink port 2017 * @cb: Control block containing additional options 2018 * @hash: hash to dump 2019 * @bucket: bucket index to dump 2020 * @idx_skip: How many entries to skip 2021 * 2022 * Return: 0 or error code. 2023 */ 2024 static int 2025 batadv_mcast_flags_dump_bucket(struct sk_buff *msg, u32 portid, 2026 struct netlink_callback *cb, 2027 struct batadv_hashtable *hash, 2028 unsigned int bucket, long *idx_skip) 2029 { 2030 struct batadv_orig_node *orig_node; 2031 long idx = 0; 2032 2033 spin_lock_bh(&hash->list_locks[bucket]); 2034 cb->seq = atomic_read(&hash->generation) << 1 | 1; 2035 2036 hlist_for_each_entry(orig_node, &hash->table[bucket], hash_entry) { 2037 if (!test_bit(BATADV_ORIG_CAPA_HAS_MCAST, 2038 &orig_node->capa_initialized)) 2039 continue; 2040 2041 if (idx < *idx_skip) 2042 goto skip; 2043 2044 if (batadv_mcast_flags_dump_entry(msg, portid, cb, orig_node)) { 2045 spin_unlock_bh(&hash->list_locks[bucket]); 2046 *idx_skip = idx; 2047 2048 return -EMSGSIZE; 2049 } 2050 2051 skip: 2052 idx++; 2053 } 2054 spin_unlock_bh(&hash->list_locks[bucket]); 2055 2056 return 0; 2057 } 2058 2059 /** 2060 * __batadv_mcast_flags_dump() - dump multicast flags table to a netlink socket 2061 * @msg: buffer for the message 2062 * @portid: netlink port 2063 * @cb: Control block containing additional options 2064 * @bat_priv: the bat priv with all the soft interface information 2065 * @bucket: current bucket to dump 2066 * @idx: index in current bucket to the next entry to dump 2067 * 2068 * Return: 0 or error code. 2069 */ 2070 static int 2071 __batadv_mcast_flags_dump(struct sk_buff *msg, u32 portid, 2072 struct netlink_callback *cb, 2073 struct batadv_priv *bat_priv, long *bucket, long *idx) 2074 { 2075 struct batadv_hashtable *hash = bat_priv->orig_hash; 2076 long bucket_tmp = *bucket; 2077 long idx_tmp = *idx; 2078 2079 while (bucket_tmp < hash->size) { 2080 if (batadv_mcast_flags_dump_bucket(msg, portid, cb, hash, 2081 bucket_tmp, &idx_tmp)) 2082 break; 2083 2084 bucket_tmp++; 2085 idx_tmp = 0; 2086 } 2087 2088 *bucket = bucket_tmp; 2089 *idx = idx_tmp; 2090 2091 return msg->len; 2092 } 2093 2094 /** 2095 * batadv_mcast_netlink_get_primary() - get primary interface from netlink 2096 * callback 2097 * @cb: netlink callback structure 2098 * @primary_if: the primary interface pointer to return the result in 2099 * 2100 * Return: 0 or error code. 2101 */ 2102 static int 2103 batadv_mcast_netlink_get_primary(struct netlink_callback *cb, 2104 struct batadv_hard_iface **primary_if) 2105 { 2106 struct batadv_hard_iface *hard_iface = NULL; 2107 struct net *net = sock_net(cb->skb->sk); 2108 struct net_device *soft_iface; 2109 struct batadv_priv *bat_priv; 2110 int ifindex; 2111 int ret = 0; 2112 2113 ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); 2114 if (!ifindex) 2115 return -EINVAL; 2116 2117 soft_iface = dev_get_by_index(net, ifindex); 2118 if (!soft_iface || !batadv_softif_is_valid(soft_iface)) { 2119 ret = -ENODEV; 2120 goto out; 2121 } 2122 2123 bat_priv = netdev_priv(soft_iface); 2124 2125 hard_iface = batadv_primary_if_get_selected(bat_priv); 2126 if (!hard_iface || hard_iface->if_status != BATADV_IF_ACTIVE) { 2127 ret = -ENOENT; 2128 goto out; 2129 } 2130 2131 out: 2132 dev_put(soft_iface); 2133 2134 if (!ret && primary_if) 2135 *primary_if = hard_iface; 2136 else 2137 batadv_hardif_put(hard_iface); 2138 2139 return ret; 2140 } 2141 2142 /** 2143 * batadv_mcast_flags_dump() - dump multicast flags table to a netlink socket 2144 * @msg: buffer for the message 2145 * @cb: callback structure containing arguments 2146 * 2147 * Return: message length. 2148 */ 2149 int batadv_mcast_flags_dump(struct sk_buff *msg, struct netlink_callback *cb) 2150 { 2151 struct batadv_hard_iface *primary_if = NULL; 2152 int portid = NETLINK_CB(cb->skb).portid; 2153 struct batadv_priv *bat_priv; 2154 long *bucket = &cb->args[0]; 2155 long *idx = &cb->args[1]; 2156 int ret; 2157 2158 ret = batadv_mcast_netlink_get_primary(cb, &primary_if); 2159 if (ret) 2160 return ret; 2161 2162 bat_priv = netdev_priv(primary_if->soft_iface); 2163 ret = __batadv_mcast_flags_dump(msg, portid, cb, bat_priv, bucket, idx); 2164 2165 batadv_hardif_put(primary_if); 2166 return ret; 2167 } 2168 2169 /** 2170 * batadv_mcast_free() - free the multicast optimizations structures 2171 * @bat_priv: the bat priv with all the soft interface information 2172 */ 2173 void batadv_mcast_free(struct batadv_priv *bat_priv) 2174 { 2175 cancel_delayed_work_sync(&bat_priv->mcast.work); 2176 2177 batadv_tvlv_container_unregister(bat_priv, BATADV_TVLV_MCAST, 2); 2178 batadv_tvlv_handler_unregister(bat_priv, BATADV_TVLV_MCAST, 2); 2179 2180 /* safely calling outside of worker, as worker was canceled above */ 2181 batadv_mcast_mla_tt_retract(bat_priv, NULL); 2182 } 2183 2184 /** 2185 * batadv_mcast_purge_orig() - reset originator global mcast state modifications 2186 * @orig: the originator which is going to get purged 2187 */ 2188 void batadv_mcast_purge_orig(struct batadv_orig_node *orig) 2189 { 2190 struct batadv_priv *bat_priv = orig->bat_priv; 2191 2192 spin_lock_bh(&orig->mcast_handler_lock); 2193 2194 batadv_mcast_want_unsnoop_update(bat_priv, orig, BATADV_NO_FLAGS); 2195 batadv_mcast_want_ipv4_update(bat_priv, orig, BATADV_NO_FLAGS); 2196 batadv_mcast_want_ipv6_update(bat_priv, orig, BATADV_NO_FLAGS); 2197 batadv_mcast_want_rtr4_update(bat_priv, orig, 2198 BATADV_MCAST_WANT_NO_RTR4); 2199 batadv_mcast_want_rtr6_update(bat_priv, orig, 2200 BATADV_MCAST_WANT_NO_RTR6); 2201 2202 spin_unlock_bh(&orig->mcast_handler_lock); 2203 } 2204