1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Multicast support for IPv6 4 * Linux INET6 implementation 5 * 6 * Authors: 7 * Pedro Roque <roque@di.fc.ul.pt> 8 * 9 * Based on linux/ipv4/igmp.c and linux/ipv4/ip_sockglue.c 10 */ 11 12 /* Changes: 13 * 14 * yoshfuji : fix format of router-alert option 15 * YOSHIFUJI Hideaki @USAGI: 16 * Fixed source address for MLD message based on 17 * <draft-ietf-magma-mld-source-05.txt>. 18 * YOSHIFUJI Hideaki @USAGI: 19 * - Ignore Queries for invalid addresses. 20 * - MLD for link-local addresses. 21 * David L Stevens <dlstevens@us.ibm.com>: 22 * - MLDv2 support 23 */ 24 25 #include <linux/module.h> 26 #include <linux/errno.h> 27 #include <linux/types.h> 28 #include <linux/string.h> 29 #include <linux/socket.h> 30 #include <linux/sockios.h> 31 #include <linux/jiffies.h> 32 #include <linux/net.h> 33 #include <linux/in.h> 34 #include <linux/in6.h> 35 #include <linux/netdevice.h> 36 #include <linux/if_addr.h> 37 #include <linux/if_arp.h> 38 #include <linux/route.h> 39 #include <linux/rtnetlink.h> 40 #include <linux/init.h> 41 #include <linux/proc_fs.h> 42 #include <linux/seq_file.h> 43 #include <linux/slab.h> 44 #include <linux/pkt_sched.h> 45 #include <net/mld.h> 46 #include <linux/workqueue.h> 47 48 #include <linux/netfilter.h> 49 #include <linux/netfilter_ipv6.h> 50 51 #include <net/net_namespace.h> 52 #include <net/netlink.h> 53 #include <net/sock.h> 54 #include <net/snmp.h> 55 56 #include <net/ipv6.h> 57 #include <net/protocol.h> 58 #include <net/if_inet6.h> 59 #include <net/ndisc.h> 60 #include <net/addrconf.h> 61 #include <net/ip6_route.h> 62 #include <net/inet_common.h> 63 64 #include <net/ip6_checksum.h> 65 66 /* Ensure that we have struct in6_addr aligned on 32bit word. */ 67 static int __mld2_query_bugs[] __attribute__((__unused__)) = { 68 BUILD_BUG_ON_ZERO(offsetof(struct mld2_query, mld2q_srcs) % 4), 69 BUILD_BUG_ON_ZERO(offsetof(struct mld2_report, mld2r_grec) % 4), 70 BUILD_BUG_ON_ZERO(offsetof(struct mld2_grec, grec_mca) % 4) 71 }; 72 73 static struct workqueue_struct *mld_wq; 74 static struct in6_addr mld2_all_mcr = MLD2_ALL_MCR_INIT; 75 76 static void igmp6_join_group(struct ifmcaddr6 *ma); 77 static void igmp6_leave_group(struct ifmcaddr6 *ma); 78 static void mld_mca_work(struct work_struct *work); 79 80 static void mld_ifc_event(struct inet6_dev *idev); 81 static bool mld_in_v1_mode(const struct inet6_dev *idev); 82 static int sf_setstate(struct ifmcaddr6 *pmc); 83 static void sf_markstate(struct ifmcaddr6 *pmc); 84 static void ip6_mc_clear_src(struct ifmcaddr6 *pmc); 85 static int ip6_mc_del_src(struct inet6_dev *idev, const struct in6_addr *pmca, 86 int sfmode, int sfcount, const struct in6_addr *psfsrc, 87 int delta); 88 static int ip6_mc_add_src(struct inet6_dev *idev, const struct in6_addr *pmca, 89 int sfmode, int sfcount, const struct in6_addr *psfsrc, 90 int delta); 91 static int ip6_mc_leave_src(struct sock *sk, struct ipv6_mc_socklist *iml, 92 struct inet6_dev *idev); 93 static int __ipv6_dev_mc_inc(struct net_device *dev, 94 const struct in6_addr *addr, unsigned int mode); 95 96 #define MLD_QRV_DEFAULT 2 97 /* RFC3810, 9.2. Query Interval */ 98 #define MLD_QI_DEFAULT (125 * HZ) 99 /* RFC3810, 9.3. Query Response Interval */ 100 #define MLD_QRI_DEFAULT (10 * HZ) 101 102 /* RFC3810, 8.1 Query Version Distinctions */ 103 #define MLD_V1_QUERY_LEN 24 104 #define MLD_V2_QUERY_LEN_MIN 28 105 106 #define IPV6_MLD_MAX_MSF 64 107 108 int sysctl_mld_max_msf __read_mostly = IPV6_MLD_MAX_MSF; 109 int sysctl_mld_qrv __read_mostly = MLD_QRV_DEFAULT; 110 111 #define mc_assert_locked(idev) \ 112 lockdep_assert_held(&(idev)->mc_lock) 113 114 #define mc_dereference(e, idev) \ 115 rcu_dereference_protected(e, lockdep_is_held(&(idev)->mc_lock)) 116 117 #define sock_dereference(e, sk) \ 118 rcu_dereference_protected(e, lockdep_sock_is_held(sk)) 119 120 #define for_each_pmc_socklock(np, sk, pmc) \ 121 for (pmc = sock_dereference((np)->ipv6_mc_list, sk); \ 122 pmc; \ 123 pmc = sock_dereference(pmc->next, sk)) 124 125 #define for_each_pmc_rcu(np, pmc) \ 126 for (pmc = rcu_dereference((np)->ipv6_mc_list); \ 127 pmc; \ 128 pmc = rcu_dereference(pmc->next)) 129 130 #define for_each_psf_mclock(mc, psf) \ 131 for (psf = mc_dereference((mc)->mca_sources, mc->idev); \ 132 psf; \ 133 psf = mc_dereference(psf->sf_next, mc->idev)) 134 135 #define for_each_psf_rcu(mc, psf) \ 136 for (psf = rcu_dereference((mc)->mca_sources); \ 137 psf; \ 138 psf = rcu_dereference(psf->sf_next)) 139 140 #define for_each_psf_tomb(mc, psf) \ 141 for (psf = mc_dereference((mc)->mca_tomb, mc->idev); \ 142 psf; \ 143 psf = mc_dereference(psf->sf_next, mc->idev)) 144 145 #define for_each_mc_mclock(idev, mc) \ 146 for (mc = mc_dereference((idev)->mc_list, idev); \ 147 mc; \ 148 mc = mc_dereference(mc->next, idev)) 149 150 #define for_each_mc_rcu(idev, mc) \ 151 for (mc = rcu_dereference((idev)->mc_list); \ 152 mc; \ 153 mc = rcu_dereference(mc->next)) 154 155 #define for_each_mc_tomb(idev, mc) \ 156 for (mc = mc_dereference((idev)->mc_tomb, idev); \ 157 mc; \ 158 mc = mc_dereference(mc->next, idev)) 159 160 static int unsolicited_report_interval(struct inet6_dev *idev) 161 { 162 int iv; 163 164 if (mld_in_v1_mode(idev)) 165 iv = READ_ONCE(idev->cnf.mldv1_unsolicited_report_interval); 166 else 167 iv = READ_ONCE(idev->cnf.mldv2_unsolicited_report_interval); 168 169 return iv > 0 ? iv : 1; 170 } 171 172 static struct net_device *ip6_mc_find_dev(struct net *net, 173 const struct in6_addr *group, 174 int ifindex) 175 { 176 struct net_device *dev = NULL; 177 struct rt6_info *rt; 178 179 if (ifindex == 0) { 180 rcu_read_lock(); 181 rt = rt6_lookup(net, group, NULL, 0, NULL, 0); 182 if (rt) { 183 dev = dst_dev_rcu(&rt->dst); 184 dev_hold(dev); 185 ip6_rt_put(rt); 186 } 187 rcu_read_unlock(); 188 } else { 189 dev = dev_get_by_index(net, ifindex); 190 } 191 192 return dev; 193 } 194 195 /* 196 * socket join on multicast group 197 */ 198 static int __ipv6_sock_mc_join(struct sock *sk, int ifindex, 199 const struct in6_addr *addr, unsigned int mode) 200 { 201 struct ipv6_pinfo *np = inet6_sk(sk); 202 struct ipv6_mc_socklist *mc_lst; 203 struct net *net = sock_net(sk); 204 struct net_device *dev = NULL; 205 int err; 206 207 if (!ipv6_addr_is_multicast(addr)) 208 return -EINVAL; 209 210 for_each_pmc_socklock(np, sk, mc_lst) { 211 if ((ifindex == 0 || mc_lst->ifindex == ifindex) && 212 ipv6_addr_equal(&mc_lst->addr, addr)) 213 return -EADDRINUSE; 214 } 215 216 mc_lst = sock_kmalloc(sk, sizeof(struct ipv6_mc_socklist), GFP_KERNEL); 217 if (!mc_lst) 218 return -ENOMEM; 219 220 mc_lst->next = NULL; 221 mc_lst->addr = *addr; 222 223 dev = ip6_mc_find_dev(net, addr, ifindex); 224 if (!dev) { 225 sock_kfree_s(sk, mc_lst, sizeof(*mc_lst)); 226 return -ENODEV; 227 } 228 229 mc_lst->ifindex = dev->ifindex; 230 mc_lst->sfmode = mode; 231 RCU_INIT_POINTER(mc_lst->sflist, NULL); 232 233 /* now add/increase the group membership on the device */ 234 err = __ipv6_dev_mc_inc(dev, addr, mode); 235 236 dev_put(dev); 237 238 if (err) { 239 sock_kfree_s(sk, mc_lst, sizeof(*mc_lst)); 240 return err; 241 } 242 243 mc_lst->next = np->ipv6_mc_list; 244 rcu_assign_pointer(np->ipv6_mc_list, mc_lst); 245 246 return 0; 247 } 248 249 int ipv6_sock_mc_join(struct sock *sk, int ifindex, const struct in6_addr *addr) 250 { 251 return __ipv6_sock_mc_join(sk, ifindex, addr, MCAST_EXCLUDE); 252 } 253 EXPORT_SYMBOL(ipv6_sock_mc_join); 254 255 int ipv6_sock_mc_join_ssm(struct sock *sk, int ifindex, 256 const struct in6_addr *addr, unsigned int mode) 257 { 258 return __ipv6_sock_mc_join(sk, ifindex, addr, mode); 259 } 260 261 /* 262 * socket leave on multicast group 263 */ 264 static void __ipv6_sock_mc_drop(struct sock *sk, struct ipv6_mc_socklist *mc_lst) 265 { 266 struct net *net = sock_net(sk); 267 struct net_device *dev; 268 269 dev = dev_get_by_index(net, mc_lst->ifindex); 270 if (dev) { 271 struct inet6_dev *idev = in6_dev_get(dev); 272 273 ip6_mc_leave_src(sk, mc_lst, idev); 274 275 if (idev) { 276 __ipv6_dev_mc_dec(idev, &mc_lst->addr); 277 in6_dev_put(idev); 278 } 279 280 dev_put(dev); 281 } else { 282 ip6_mc_leave_src(sk, mc_lst, NULL); 283 } 284 285 atomic_sub(sizeof(*mc_lst), &sk->sk_omem_alloc); 286 kfree_rcu(mc_lst, rcu); 287 } 288 289 int ipv6_sock_mc_drop(struct sock *sk, int ifindex, const struct in6_addr *addr) 290 { 291 struct ipv6_pinfo *np = inet6_sk(sk); 292 struct ipv6_mc_socklist __rcu **lnk; 293 struct ipv6_mc_socklist *mc_lst; 294 295 if (!ipv6_addr_is_multicast(addr)) 296 return -EINVAL; 297 298 for (lnk = &np->ipv6_mc_list; 299 (mc_lst = sock_dereference(*lnk, sk)) != NULL; 300 lnk = &mc_lst->next) { 301 if ((ifindex == 0 || mc_lst->ifindex == ifindex) && 302 ipv6_addr_equal(&mc_lst->addr, addr)) { 303 *lnk = mc_lst->next; 304 __ipv6_sock_mc_drop(sk, mc_lst); 305 return 0; 306 } 307 } 308 309 return -EADDRNOTAVAIL; 310 } 311 EXPORT_SYMBOL(ipv6_sock_mc_drop); 312 313 static struct inet6_dev *ip6_mc_find_idev(struct net *net, 314 const struct in6_addr *group, 315 int ifindex) 316 { 317 struct net_device *dev; 318 struct inet6_dev *idev; 319 320 dev = ip6_mc_find_dev(net, group, ifindex); 321 if (!dev) 322 return NULL; 323 324 idev = in6_dev_get(dev); 325 dev_put(dev); 326 327 return idev; 328 } 329 330 void __ipv6_sock_mc_close(struct sock *sk) 331 { 332 struct ipv6_pinfo *np = inet6_sk(sk); 333 struct ipv6_mc_socklist *mc_lst; 334 335 while ((mc_lst = sock_dereference(np->ipv6_mc_list, sk)) != NULL) { 336 np->ipv6_mc_list = mc_lst->next; 337 __ipv6_sock_mc_drop(sk, mc_lst); 338 } 339 } 340 341 void ipv6_sock_mc_close(struct sock *sk) 342 { 343 struct ipv6_pinfo *np = inet6_sk(sk); 344 345 if (!rcu_access_pointer(np->ipv6_mc_list)) 346 return; 347 348 lock_sock(sk); 349 __ipv6_sock_mc_close(sk); 350 release_sock(sk); 351 } 352 353 int ip6_mc_source(int add, int omode, struct sock *sk, 354 struct group_source_req *pgsr) 355 { 356 struct ipv6_pinfo *inet6 = inet6_sk(sk); 357 struct in6_addr *source, *group; 358 struct net *net = sock_net(sk); 359 struct ipv6_mc_socklist *pmc; 360 struct ip6_sf_socklist *psl; 361 struct inet6_dev *idev; 362 int leavegroup = 0; 363 int i, j, rv; 364 int err; 365 366 source = &((struct sockaddr_in6 *)&pgsr->gsr_source)->sin6_addr; 367 group = &((struct sockaddr_in6 *)&pgsr->gsr_group)->sin6_addr; 368 369 if (!ipv6_addr_is_multicast(group)) 370 return -EINVAL; 371 372 idev = ip6_mc_find_idev(net, group, pgsr->gsr_interface); 373 if (!idev) 374 return -ENODEV; 375 376 mutex_lock(&idev->mc_lock); 377 378 if (idev->dead) { 379 err = -ENODEV; 380 goto done; 381 } 382 383 err = -EADDRNOTAVAIL; 384 385 for_each_pmc_socklock(inet6, sk, pmc) { 386 if (pgsr->gsr_interface && pmc->ifindex != pgsr->gsr_interface) 387 continue; 388 if (ipv6_addr_equal(&pmc->addr, group)) 389 break; 390 } 391 if (!pmc) { /* must have a prior join */ 392 err = -EINVAL; 393 goto done; 394 } 395 /* if a source filter was set, must be the same mode as before */ 396 if (rcu_access_pointer(pmc->sflist)) { 397 if (pmc->sfmode != omode) { 398 err = -EINVAL; 399 goto done; 400 } 401 } else if (pmc->sfmode != omode) { 402 /* allow mode switches for empty-set filters */ 403 ip6_mc_add_src(idev, group, omode, 0, NULL, 0); 404 ip6_mc_del_src(idev, group, pmc->sfmode, 0, NULL, 0); 405 pmc->sfmode = omode; 406 } 407 408 psl = sock_dereference(pmc->sflist, sk); 409 if (!add) { 410 if (!psl) 411 goto done; /* err = -EADDRNOTAVAIL */ 412 rv = !0; 413 for (i = 0; i < psl->sl_count; i++) { 414 rv = !ipv6_addr_equal(&psl->sl_addr[i], source); 415 if (rv == 0) 416 break; 417 } 418 if (rv) /* source not found */ 419 goto done; /* err = -EADDRNOTAVAIL */ 420 421 /* special case - (INCLUDE, empty) == LEAVE_GROUP */ 422 if (psl->sl_count == 1 && omode == MCAST_INCLUDE) { 423 leavegroup = 1; 424 goto done; 425 } 426 427 /* update the interface filter */ 428 ip6_mc_del_src(idev, group, omode, 1, source, 1); 429 430 for (j = i+1; j < psl->sl_count; j++) 431 psl->sl_addr[j-1] = psl->sl_addr[j]; 432 psl->sl_count--; 433 err = 0; 434 goto done; 435 } 436 /* else, add a new source to the filter */ 437 438 if (psl && psl->sl_count >= sysctl_mld_max_msf) { 439 err = -ENOBUFS; 440 goto done; 441 } 442 if (!psl || psl->sl_count == psl->sl_max) { 443 struct ip6_sf_socklist *newpsl; 444 int count = IP6_SFBLOCK; 445 446 if (psl) 447 count += psl->sl_max; 448 newpsl = sock_kmalloc(sk, struct_size(newpsl, sl_addr, count), 449 GFP_KERNEL); 450 if (!newpsl) { 451 err = -ENOBUFS; 452 goto done; 453 } 454 newpsl->sl_max = count; 455 newpsl->sl_count = count - IP6_SFBLOCK; 456 if (psl) { 457 for (i = 0; i < psl->sl_count; i++) 458 newpsl->sl_addr[i] = psl->sl_addr[i]; 459 atomic_sub(struct_size(psl, sl_addr, psl->sl_max), 460 &sk->sk_omem_alloc); 461 } 462 rcu_assign_pointer(pmc->sflist, newpsl); 463 kfree_rcu(psl, rcu); 464 psl = newpsl; 465 } 466 rv = 1; /* > 0 for insert logic below if sl_count is 0 */ 467 for (i = 0; i < psl->sl_count; i++) { 468 rv = !ipv6_addr_equal(&psl->sl_addr[i], source); 469 if (rv == 0) /* There is an error in the address. */ 470 goto done; 471 } 472 for (j = psl->sl_count-1; j >= i; j--) 473 psl->sl_addr[j+1] = psl->sl_addr[j]; 474 psl->sl_addr[i] = *source; 475 psl->sl_count++; 476 err = 0; 477 /* update the interface list */ 478 ip6_mc_add_src(idev, group, omode, 1, source, 1); 479 done: 480 mutex_unlock(&idev->mc_lock); 481 in6_dev_put(idev); 482 if (leavegroup) 483 err = ipv6_sock_mc_drop(sk, pgsr->gsr_interface, group); 484 return err; 485 } 486 487 int ip6_mc_msfilter(struct sock *sk, struct group_filter *gsf, 488 struct sockaddr_storage *list) 489 { 490 struct ipv6_pinfo *inet6 = inet6_sk(sk); 491 struct ip6_sf_socklist *newpsl, *psl; 492 struct net *net = sock_net(sk); 493 const struct in6_addr *group; 494 struct ipv6_mc_socklist *pmc; 495 struct inet6_dev *idev; 496 int leavegroup = 0; 497 int i, err; 498 499 group = &((struct sockaddr_in6 *)&gsf->gf_group)->sin6_addr; 500 501 if (!ipv6_addr_is_multicast(group)) 502 return -EINVAL; 503 if (gsf->gf_fmode != MCAST_INCLUDE && 504 gsf->gf_fmode != MCAST_EXCLUDE) 505 return -EINVAL; 506 507 idev = ip6_mc_find_idev(net, group, gsf->gf_interface); 508 if (!idev) 509 return -ENODEV; 510 511 mutex_lock(&idev->mc_lock); 512 513 if (idev->dead) { 514 err = -ENODEV; 515 goto done; 516 } 517 518 err = 0; 519 520 if (gsf->gf_fmode == MCAST_INCLUDE && gsf->gf_numsrc == 0) { 521 leavegroup = 1; 522 goto done; 523 } 524 525 for_each_pmc_socklock(inet6, sk, pmc) { 526 if (pmc->ifindex != gsf->gf_interface) 527 continue; 528 if (ipv6_addr_equal(&pmc->addr, group)) 529 break; 530 } 531 if (!pmc) { /* must have a prior join */ 532 err = -EINVAL; 533 goto done; 534 } 535 if (gsf->gf_numsrc) { 536 newpsl = sock_kmalloc(sk, struct_size(newpsl, sl_addr, 537 gsf->gf_numsrc), 538 GFP_KERNEL); 539 if (!newpsl) { 540 err = -ENOBUFS; 541 goto done; 542 } 543 newpsl->sl_max = newpsl->sl_count = gsf->gf_numsrc; 544 for (i = 0; i < newpsl->sl_count; ++i, ++list) { 545 struct sockaddr_in6 *psin6; 546 547 psin6 = (struct sockaddr_in6 *)list; 548 newpsl->sl_addr[i] = psin6->sin6_addr; 549 } 550 551 err = ip6_mc_add_src(idev, group, gsf->gf_fmode, 552 newpsl->sl_count, newpsl->sl_addr, 0); 553 if (err) { 554 sock_kfree_s(sk, newpsl, struct_size(newpsl, sl_addr, 555 newpsl->sl_max)); 556 goto done; 557 } 558 } else { 559 newpsl = NULL; 560 ip6_mc_add_src(idev, group, gsf->gf_fmode, 0, NULL, 0); 561 } 562 563 psl = sock_dereference(pmc->sflist, sk); 564 if (psl) { 565 ip6_mc_del_src(idev, group, pmc->sfmode, 566 psl->sl_count, psl->sl_addr, 0); 567 atomic_sub(struct_size(psl, sl_addr, psl->sl_max), 568 &sk->sk_omem_alloc); 569 } else { 570 ip6_mc_del_src(idev, group, pmc->sfmode, 0, NULL, 0); 571 } 572 573 rcu_assign_pointer(pmc->sflist, newpsl); 574 kfree_rcu(psl, rcu); 575 pmc->sfmode = gsf->gf_fmode; 576 err = 0; 577 done: 578 mutex_unlock(&idev->mc_lock); 579 in6_dev_put(idev); 580 if (leavegroup) 581 err = ipv6_sock_mc_drop(sk, gsf->gf_interface, group); 582 return err; 583 } 584 585 int ip6_mc_msfget(struct sock *sk, struct group_filter *gsf, 586 sockptr_t optval, size_t ss_offset) 587 { 588 struct ipv6_pinfo *inet6 = inet6_sk(sk); 589 const struct in6_addr *group; 590 struct ipv6_mc_socklist *pmc; 591 struct ip6_sf_socklist *psl; 592 unsigned int count; 593 int i, copycount; 594 595 group = &((struct sockaddr_in6 *)&gsf->gf_group)->sin6_addr; 596 597 if (!ipv6_addr_is_multicast(group)) 598 return -EINVAL; 599 600 for_each_pmc_socklock(inet6, sk, pmc) { 601 if (pmc->ifindex != gsf->gf_interface) 602 continue; 603 if (ipv6_addr_equal(group, &pmc->addr)) 604 break; 605 } 606 if (!pmc) /* must have a prior join */ 607 return -EADDRNOTAVAIL; 608 609 gsf->gf_fmode = pmc->sfmode; 610 psl = sock_dereference(pmc->sflist, sk); 611 count = psl ? psl->sl_count : 0; 612 613 copycount = min(count, gsf->gf_numsrc); 614 gsf->gf_numsrc = count; 615 for (i = 0; i < copycount; i++) { 616 struct sockaddr_in6 *psin6; 617 struct sockaddr_storage ss; 618 619 psin6 = (struct sockaddr_in6 *)&ss; 620 memset(&ss, 0, sizeof(ss)); 621 psin6->sin6_family = AF_INET6; 622 psin6->sin6_addr = psl->sl_addr[i]; 623 if (copy_to_sockptr_offset(optval, ss_offset, &ss, sizeof(ss))) 624 return -EFAULT; 625 ss_offset += sizeof(ss); 626 } 627 return 0; 628 } 629 630 bool inet6_mc_check(const struct sock *sk, const struct in6_addr *mc_addr, 631 const struct in6_addr *src_addr) 632 { 633 const struct ipv6_pinfo *np = inet6_sk(sk); 634 const struct ipv6_mc_socklist *mc; 635 const struct ip6_sf_socklist *psl; 636 bool rv = true; 637 638 rcu_read_lock(); 639 for_each_pmc_rcu(np, mc) { 640 if (ipv6_addr_equal(&mc->addr, mc_addr)) 641 break; 642 } 643 if (!mc) { 644 rcu_read_unlock(); 645 return inet6_test_bit(MC6_ALL, sk); 646 } 647 psl = rcu_dereference(mc->sflist); 648 if (!psl) { 649 rv = mc->sfmode == MCAST_EXCLUDE; 650 } else { 651 int i; 652 653 for (i = 0; i < psl->sl_count; i++) { 654 if (ipv6_addr_equal(&psl->sl_addr[i], src_addr)) 655 break; 656 } 657 if (mc->sfmode == MCAST_INCLUDE && i >= psl->sl_count) 658 rv = false; 659 if (mc->sfmode == MCAST_EXCLUDE && i < psl->sl_count) 660 rv = false; 661 } 662 rcu_read_unlock(); 663 664 return rv; 665 } 666 667 static void igmp6_group_added(struct ifmcaddr6 *mc) 668 { 669 struct net_device *dev = mc->idev->dev; 670 char buf[MAX_ADDR_LEN]; 671 672 mc_assert_locked(mc->idev); 673 674 if (IPV6_ADDR_MC_SCOPE(&mc->mca_addr) < 675 IPV6_ADDR_SCOPE_LINKLOCAL) 676 return; 677 678 if (!(mc->mca_flags&MAF_LOADED)) { 679 mc->mca_flags |= MAF_LOADED; 680 if (ndisc_mc_map(&mc->mca_addr, buf, dev, 0) == 0) 681 dev_mc_add(dev, buf); 682 } 683 684 if (!(dev->flags & IFF_UP) || (mc->mca_flags & MAF_NOREPORT)) 685 return; 686 687 if (mld_in_v1_mode(mc->idev)) { 688 igmp6_join_group(mc); 689 return; 690 } 691 /* else v2 */ 692 693 /* Based on RFC3810 6.1, for newly added INCLUDE SSM, we 694 * should not send filter-mode change record as the mode 695 * should be from IN() to IN(A). 696 */ 697 if (mc->mca_sfmode == MCAST_EXCLUDE) 698 mc->mca_crcount = mc->idev->mc_qrv; 699 700 mld_ifc_event(mc->idev); 701 } 702 703 static void igmp6_group_dropped(struct ifmcaddr6 *mc) 704 { 705 struct net_device *dev = mc->idev->dev; 706 char buf[MAX_ADDR_LEN]; 707 708 mc_assert_locked(mc->idev); 709 710 if (IPV6_ADDR_MC_SCOPE(&mc->mca_addr) < 711 IPV6_ADDR_SCOPE_LINKLOCAL) 712 return; 713 714 if (mc->mca_flags&MAF_LOADED) { 715 mc->mca_flags &= ~MAF_LOADED; 716 if (ndisc_mc_map(&mc->mca_addr, buf, dev, 0) == 0) 717 dev_mc_del(dev, buf); 718 } 719 720 if (mc->mca_flags & MAF_NOREPORT) 721 return; 722 723 if (!mc->idev->dead) 724 igmp6_leave_group(mc); 725 726 if (cancel_delayed_work(&mc->mca_work)) 727 refcount_dec(&mc->mca_refcnt); 728 } 729 730 /* deleted ifmcaddr6 manipulation */ 731 static void mld_add_delrec(struct inet6_dev *idev, struct ifmcaddr6 *im) 732 { 733 struct ifmcaddr6 *pmc; 734 735 mc_assert_locked(idev); 736 737 /* this is an "ifmcaddr6" for convenience; only the fields below 738 * are actually used. In particular, the refcnt and users are not 739 * used for management of the delete list. Using the same structure 740 * for deleted items allows change reports to use common code with 741 * non-deleted or query-response MCA's. 742 */ 743 pmc = kzalloc(sizeof(*pmc), GFP_KERNEL); 744 if (!pmc) 745 return; 746 747 pmc->idev = im->idev; 748 in6_dev_hold(idev); 749 pmc->mca_addr = im->mca_addr; 750 pmc->mca_crcount = idev->mc_qrv; 751 pmc->mca_sfmode = im->mca_sfmode; 752 if (pmc->mca_sfmode == MCAST_INCLUDE) { 753 struct ip6_sf_list *psf; 754 755 rcu_assign_pointer(pmc->mca_tomb, 756 mc_dereference(im->mca_tomb, idev)); 757 rcu_assign_pointer(pmc->mca_sources, 758 mc_dereference(im->mca_sources, idev)); 759 RCU_INIT_POINTER(im->mca_tomb, NULL); 760 RCU_INIT_POINTER(im->mca_sources, NULL); 761 762 for_each_psf_mclock(pmc, psf) 763 psf->sf_crcount = pmc->mca_crcount; 764 } 765 766 rcu_assign_pointer(pmc->next, idev->mc_tomb); 767 rcu_assign_pointer(idev->mc_tomb, pmc); 768 } 769 770 static void mld_del_delrec(struct inet6_dev *idev, struct ifmcaddr6 *im) 771 { 772 struct ip6_sf_list *psf, *sources, *tomb; 773 struct in6_addr *pmca = &im->mca_addr; 774 struct ifmcaddr6 *pmc, *pmc_prev; 775 776 mc_assert_locked(idev); 777 778 pmc_prev = NULL; 779 for_each_mc_tomb(idev, pmc) { 780 if (ipv6_addr_equal(&pmc->mca_addr, pmca)) 781 break; 782 pmc_prev = pmc; 783 } 784 if (!pmc) 785 return; 786 if (pmc_prev) 787 rcu_assign_pointer(pmc_prev->next, pmc->next); 788 else 789 rcu_assign_pointer(idev->mc_tomb, pmc->next); 790 791 im->idev = pmc->idev; 792 if (im->mca_sfmode == MCAST_INCLUDE) { 793 tomb = rcu_replace_pointer(im->mca_tomb, 794 mc_dereference(pmc->mca_tomb, pmc->idev), 795 lockdep_is_held(&im->idev->mc_lock)); 796 rcu_assign_pointer(pmc->mca_tomb, tomb); 797 798 sources = rcu_replace_pointer(im->mca_sources, 799 mc_dereference(pmc->mca_sources, pmc->idev), 800 lockdep_is_held(&im->idev->mc_lock)); 801 rcu_assign_pointer(pmc->mca_sources, sources); 802 for_each_psf_mclock(im, psf) 803 psf->sf_crcount = idev->mc_qrv; 804 } else { 805 im->mca_crcount = idev->mc_qrv; 806 } 807 ip6_mc_clear_src(pmc); 808 in6_dev_put(pmc->idev); 809 kfree_rcu(pmc, rcu); 810 } 811 812 static void mld_clear_delrec(struct inet6_dev *idev) 813 { 814 struct ifmcaddr6 *pmc, *nextpmc; 815 816 mc_assert_locked(idev); 817 818 pmc = mc_dereference(idev->mc_tomb, idev); 819 RCU_INIT_POINTER(idev->mc_tomb, NULL); 820 821 for (; pmc; pmc = nextpmc) { 822 nextpmc = mc_dereference(pmc->next, idev); 823 ip6_mc_clear_src(pmc); 824 in6_dev_put(pmc->idev); 825 kfree_rcu(pmc, rcu); 826 } 827 828 /* clear dead sources, too */ 829 for_each_mc_mclock(idev, pmc) { 830 struct ip6_sf_list *psf, *psf_next; 831 832 psf = mc_dereference(pmc->mca_tomb, idev); 833 RCU_INIT_POINTER(pmc->mca_tomb, NULL); 834 for (; psf; psf = psf_next) { 835 psf_next = mc_dereference(psf->sf_next, idev); 836 kfree_rcu(psf, rcu); 837 } 838 } 839 } 840 841 static void mld_clear_query(struct inet6_dev *idev) 842 { 843 spin_lock_bh(&idev->mc_query_lock); 844 __skb_queue_purge(&idev->mc_query_queue); 845 spin_unlock_bh(&idev->mc_query_lock); 846 } 847 848 static void mld_clear_report(struct inet6_dev *idev) 849 { 850 spin_lock_bh(&idev->mc_report_lock); 851 __skb_queue_purge(&idev->mc_report_queue); 852 spin_unlock_bh(&idev->mc_report_lock); 853 } 854 855 static void ma_put(struct ifmcaddr6 *mc) 856 { 857 if (refcount_dec_and_test(&mc->mca_refcnt)) { 858 in6_dev_put(mc->idev); 859 kfree_rcu(mc, rcu); 860 } 861 } 862 863 static struct ifmcaddr6 *mca_alloc(struct inet6_dev *idev, 864 const struct in6_addr *addr, 865 unsigned int mode) 866 { 867 struct ifmcaddr6 *mc; 868 869 mc_assert_locked(idev); 870 871 mc = kzalloc(sizeof(*mc), GFP_KERNEL); 872 if (!mc) 873 return NULL; 874 875 INIT_DELAYED_WORK(&mc->mca_work, mld_mca_work); 876 877 mc->mca_addr = *addr; 878 mc->idev = idev; /* reference taken by caller */ 879 mc->mca_users = 1; 880 /* mca_stamp should be updated upon changes */ 881 mc->mca_cstamp = mc->mca_tstamp = jiffies; 882 refcount_set(&mc->mca_refcnt, 1); 883 884 mc->mca_sfmode = mode; 885 mc->mca_sfcount[mode] = 1; 886 887 if (ipv6_addr_is_ll_all_nodes(&mc->mca_addr) || 888 IPV6_ADDR_MC_SCOPE(&mc->mca_addr) < IPV6_ADDR_SCOPE_LINKLOCAL) 889 mc->mca_flags |= MAF_NOREPORT; 890 891 return mc; 892 } 893 894 static void inet6_ifmcaddr_notify(struct net_device *dev, 895 const struct ifmcaddr6 *ifmca, int event) 896 { 897 struct inet6_fill_args fillargs = { 898 .portid = 0, 899 .seq = 0, 900 .event = event, 901 .flags = 0, 902 .netnsid = -1, 903 .force_rt_scope_universe = true, 904 }; 905 struct net *net = dev_net(dev); 906 struct sk_buff *skb; 907 int err = -ENOMEM; 908 909 skb = nlmsg_new(NLMSG_ALIGN(sizeof(struct ifaddrmsg)) + 910 nla_total_size(sizeof(struct in6_addr)) + 911 nla_total_size(sizeof(struct ifa_cacheinfo)), 912 GFP_KERNEL); 913 if (!skb) 914 goto error; 915 916 err = inet6_fill_ifmcaddr(skb, ifmca, &fillargs); 917 if (err < 0) { 918 WARN_ON_ONCE(err == -EMSGSIZE); 919 nlmsg_free(skb); 920 goto error; 921 } 922 923 rtnl_notify(skb, net, 0, RTNLGRP_IPV6_MCADDR, NULL, GFP_KERNEL); 924 return; 925 error: 926 rtnl_set_sk_err(net, RTNLGRP_IPV6_MCADDR, err); 927 } 928 929 /* 930 * device multicast group inc (add if not found) 931 */ 932 static int __ipv6_dev_mc_inc(struct net_device *dev, 933 const struct in6_addr *addr, unsigned int mode) 934 { 935 struct inet6_dev *idev; 936 struct ifmcaddr6 *mc; 937 938 /* we need to take a reference on idev */ 939 idev = in6_dev_get(dev); 940 if (!idev) 941 return -EINVAL; 942 943 mutex_lock(&idev->mc_lock); 944 945 if (READ_ONCE(idev->dead)) { 946 mutex_unlock(&idev->mc_lock); 947 in6_dev_put(idev); 948 return -ENODEV; 949 } 950 951 for_each_mc_mclock(idev, mc) { 952 if (ipv6_addr_equal(&mc->mca_addr, addr)) { 953 mc->mca_users++; 954 ip6_mc_add_src(idev, &mc->mca_addr, mode, 0, NULL, 0); 955 mutex_unlock(&idev->mc_lock); 956 in6_dev_put(idev); 957 return 0; 958 } 959 } 960 961 mc = mca_alloc(idev, addr, mode); 962 if (!mc) { 963 mutex_unlock(&idev->mc_lock); 964 in6_dev_put(idev); 965 return -ENOMEM; 966 } 967 968 rcu_assign_pointer(mc->next, idev->mc_list); 969 rcu_assign_pointer(idev->mc_list, mc); 970 971 mld_del_delrec(idev, mc); 972 igmp6_group_added(mc); 973 inet6_ifmcaddr_notify(dev, mc, RTM_NEWMULTICAST); 974 mutex_unlock(&idev->mc_lock); 975 976 return 0; 977 } 978 979 int ipv6_dev_mc_inc(struct net_device *dev, const struct in6_addr *addr) 980 { 981 return __ipv6_dev_mc_inc(dev, addr, MCAST_EXCLUDE); 982 } 983 EXPORT_SYMBOL(ipv6_dev_mc_inc); 984 985 /* 986 * device multicast group del 987 */ 988 int __ipv6_dev_mc_dec(struct inet6_dev *idev, const struct in6_addr *addr) 989 { 990 struct ifmcaddr6 *ma, __rcu **map; 991 992 mutex_lock(&idev->mc_lock); 993 994 for (map = &idev->mc_list; 995 (ma = mc_dereference(*map, idev)); 996 map = &ma->next) { 997 if (ipv6_addr_equal(&ma->mca_addr, addr)) { 998 if (--ma->mca_users == 0) { 999 *map = ma->next; 1000 1001 igmp6_group_dropped(ma); 1002 inet6_ifmcaddr_notify(idev->dev, ma, 1003 RTM_DELMULTICAST); 1004 ip6_mc_clear_src(ma); 1005 mutex_unlock(&idev->mc_lock); 1006 1007 ma_put(ma); 1008 return 0; 1009 } 1010 mutex_unlock(&idev->mc_lock); 1011 return 0; 1012 } 1013 } 1014 1015 mutex_unlock(&idev->mc_lock); 1016 return -ENOENT; 1017 } 1018 1019 int ipv6_dev_mc_dec(struct net_device *dev, const struct in6_addr *addr) 1020 { 1021 struct inet6_dev *idev; 1022 int err; 1023 1024 idev = in6_dev_get(dev); 1025 if (!idev) 1026 return -ENODEV; 1027 1028 err = __ipv6_dev_mc_dec(idev, addr); 1029 in6_dev_put(idev); 1030 1031 return err; 1032 } 1033 EXPORT_SYMBOL(ipv6_dev_mc_dec); 1034 1035 /* 1036 * check if the interface/address pair is valid 1037 */ 1038 bool ipv6_chk_mcast_addr(struct net_device *dev, const struct in6_addr *group, 1039 const struct in6_addr *src_addr) 1040 { 1041 struct inet6_dev *idev; 1042 struct ifmcaddr6 *mc; 1043 bool rv = false; 1044 1045 rcu_read_lock(); 1046 idev = __in6_dev_get(dev); 1047 if (!idev) 1048 goto unlock; 1049 for_each_mc_rcu(idev, mc) { 1050 if (ipv6_addr_equal(&mc->mca_addr, group)) 1051 break; 1052 } 1053 if (!mc) 1054 goto unlock; 1055 if (src_addr && !ipv6_addr_any(src_addr)) { 1056 struct ip6_sf_list *psf; 1057 1058 for_each_psf_rcu(mc, psf) { 1059 if (ipv6_addr_equal(&psf->sf_addr, src_addr)) 1060 break; 1061 } 1062 if (psf) 1063 rv = READ_ONCE(psf->sf_count[MCAST_INCLUDE]) || 1064 READ_ONCE(psf->sf_count[MCAST_EXCLUDE]) != 1065 READ_ONCE(mc->mca_sfcount[MCAST_EXCLUDE]); 1066 else 1067 rv = READ_ONCE(mc->mca_sfcount[MCAST_EXCLUDE]) != 0; 1068 } else { 1069 rv = true; /* don't filter unspecified source */ 1070 } 1071 unlock: 1072 rcu_read_unlock(); 1073 return rv; 1074 } 1075 1076 static void mld_gq_start_work(struct inet6_dev *idev) 1077 { 1078 unsigned long tv = get_random_u32_below(idev->mc_maxdelay); 1079 1080 mc_assert_locked(idev); 1081 1082 idev->mc_gq_running = 1; 1083 if (!mod_delayed_work(mld_wq, &idev->mc_gq_work, tv + 2)) 1084 in6_dev_hold(idev); 1085 } 1086 1087 static void mld_gq_stop_work(struct inet6_dev *idev) 1088 { 1089 mc_assert_locked(idev); 1090 1091 idev->mc_gq_running = 0; 1092 if (cancel_delayed_work(&idev->mc_gq_work)) 1093 __in6_dev_put(idev); 1094 } 1095 1096 static void mld_ifc_start_work(struct inet6_dev *idev, unsigned long delay) 1097 { 1098 unsigned long tv = get_random_u32_below(delay); 1099 1100 mc_assert_locked(idev); 1101 1102 if (!mod_delayed_work(mld_wq, &idev->mc_ifc_work, tv + 2)) 1103 in6_dev_hold(idev); 1104 } 1105 1106 static void mld_ifc_stop_work(struct inet6_dev *idev) 1107 { 1108 mc_assert_locked(idev); 1109 1110 idev->mc_ifc_count = 0; 1111 if (cancel_delayed_work(&idev->mc_ifc_work)) 1112 __in6_dev_put(idev); 1113 } 1114 1115 static void mld_dad_start_work(struct inet6_dev *idev, unsigned long delay) 1116 { 1117 unsigned long tv = get_random_u32_below(delay); 1118 1119 mc_assert_locked(idev); 1120 1121 if (!mod_delayed_work(mld_wq, &idev->mc_dad_work, tv + 2)) 1122 in6_dev_hold(idev); 1123 } 1124 1125 static void mld_dad_stop_work(struct inet6_dev *idev) 1126 { 1127 if (cancel_delayed_work(&idev->mc_dad_work)) 1128 __in6_dev_put(idev); 1129 } 1130 1131 static void mld_query_stop_work(struct inet6_dev *idev) 1132 { 1133 spin_lock_bh(&idev->mc_query_lock); 1134 if (cancel_delayed_work(&idev->mc_query_work)) 1135 __in6_dev_put(idev); 1136 spin_unlock_bh(&idev->mc_query_lock); 1137 } 1138 1139 static void mld_report_stop_work(struct inet6_dev *idev) 1140 { 1141 if (cancel_delayed_work_sync(&idev->mc_report_work)) 1142 __in6_dev_put(idev); 1143 } 1144 1145 /* IGMP handling (alias multicast ICMPv6 messages) */ 1146 static void igmp6_group_queried(struct ifmcaddr6 *ma, unsigned long resptime) 1147 { 1148 unsigned long delay = resptime; 1149 1150 mc_assert_locked(ma->idev); 1151 1152 /* Do not start work for these addresses */ 1153 if (ipv6_addr_is_ll_all_nodes(&ma->mca_addr) || 1154 IPV6_ADDR_MC_SCOPE(&ma->mca_addr) < IPV6_ADDR_SCOPE_LINKLOCAL) 1155 return; 1156 1157 if (cancel_delayed_work(&ma->mca_work)) { 1158 refcount_dec(&ma->mca_refcnt); 1159 delay = ma->mca_work.timer.expires - jiffies; 1160 } 1161 1162 if (delay >= resptime) 1163 delay = get_random_u32_below(resptime); 1164 1165 if (!mod_delayed_work(mld_wq, &ma->mca_work, delay)) 1166 refcount_inc(&ma->mca_refcnt); 1167 ma->mca_flags |= MAF_TIMER_RUNNING; 1168 } 1169 1170 /* mark EXCLUDE-mode sources */ 1171 static bool mld_xmarksources(struct ifmcaddr6 *pmc, int nsrcs, 1172 const struct in6_addr *srcs) 1173 { 1174 struct ip6_sf_list *psf; 1175 int i, scount; 1176 1177 mc_assert_locked(pmc->idev); 1178 1179 scount = 0; 1180 for_each_psf_mclock(pmc, psf) { 1181 if (scount == nsrcs) 1182 break; 1183 for (i = 0; i < nsrcs; i++) { 1184 /* skip inactive filters */ 1185 if (psf->sf_count[MCAST_INCLUDE] || 1186 pmc->mca_sfcount[MCAST_EXCLUDE] != 1187 psf->sf_count[MCAST_EXCLUDE]) 1188 break; 1189 if (ipv6_addr_equal(&srcs[i], &psf->sf_addr)) { 1190 scount++; 1191 break; 1192 } 1193 } 1194 } 1195 pmc->mca_flags &= ~MAF_GSQUERY; 1196 if (scount == nsrcs) /* all sources excluded */ 1197 return false; 1198 return true; 1199 } 1200 1201 static bool mld_marksources(struct ifmcaddr6 *pmc, int nsrcs, 1202 const struct in6_addr *srcs) 1203 { 1204 struct ip6_sf_list *psf; 1205 int i, scount; 1206 1207 mc_assert_locked(pmc->idev); 1208 1209 if (pmc->mca_sfmode == MCAST_EXCLUDE) 1210 return mld_xmarksources(pmc, nsrcs, srcs); 1211 1212 /* mark INCLUDE-mode sources */ 1213 1214 scount = 0; 1215 for_each_psf_mclock(pmc, psf) { 1216 if (scount == nsrcs) 1217 break; 1218 for (i = 0; i < nsrcs; i++) { 1219 if (ipv6_addr_equal(&srcs[i], &psf->sf_addr)) { 1220 psf->sf_gsresp = 1; 1221 scount++; 1222 break; 1223 } 1224 } 1225 } 1226 if (!scount) { 1227 pmc->mca_flags &= ~MAF_GSQUERY; 1228 return false; 1229 } 1230 pmc->mca_flags |= MAF_GSQUERY; 1231 return true; 1232 } 1233 1234 static int mld_force_mld_version(const struct inet6_dev *idev) 1235 { 1236 const struct net *net = dev_net(idev->dev); 1237 int all_force; 1238 1239 all_force = READ_ONCE(net->ipv6.devconf_all->force_mld_version); 1240 /* Normally, both are 0 here. If enforcement to a particular is 1241 * being used, individual device enforcement will have a lower 1242 * precedence over 'all' device (.../conf/all/force_mld_version). 1243 */ 1244 return all_force ?: READ_ONCE(idev->cnf.force_mld_version); 1245 } 1246 1247 static bool mld_in_v2_mode_only(const struct inet6_dev *idev) 1248 { 1249 return mld_force_mld_version(idev) == 2; 1250 } 1251 1252 static bool mld_in_v1_mode_only(const struct inet6_dev *idev) 1253 { 1254 return mld_force_mld_version(idev) == 1; 1255 } 1256 1257 static bool mld_in_v1_mode(const struct inet6_dev *idev) 1258 { 1259 if (mld_in_v2_mode_only(idev)) 1260 return false; 1261 if (mld_in_v1_mode_only(idev)) 1262 return true; 1263 if (idev->mc_v1_seen && time_before(jiffies, idev->mc_v1_seen)) 1264 return true; 1265 1266 return false; 1267 } 1268 1269 static void mld_set_v1_mode(struct inet6_dev *idev) 1270 { 1271 /* RFC3810, relevant sections: 1272 * - 9.1. Robustness Variable 1273 * - 9.2. Query Interval 1274 * - 9.3. Query Response Interval 1275 * - 9.12. Older Version Querier Present Timeout 1276 */ 1277 unsigned long switchback; 1278 1279 switchback = (idev->mc_qrv * idev->mc_qi) + idev->mc_qri; 1280 1281 idev->mc_v1_seen = jiffies + switchback; 1282 } 1283 1284 static void mld_update_qrv(struct inet6_dev *idev, 1285 const struct mld2_query *mlh2) 1286 { 1287 /* RFC3810, relevant sections: 1288 * - 5.1.8. QRV (Querier's Robustness Variable) 1289 * - 9.1. Robustness Variable 1290 */ 1291 1292 /* The value of the Robustness Variable MUST NOT be zero, 1293 * and SHOULD NOT be one. Catch this here if we ever run 1294 * into such a case in future. 1295 */ 1296 const int min_qrv = min(MLD_QRV_DEFAULT, sysctl_mld_qrv); 1297 WARN_ON(idev->mc_qrv == 0); 1298 1299 if (mlh2->mld2q_qrv > 0) 1300 idev->mc_qrv = mlh2->mld2q_qrv; 1301 1302 if (unlikely(idev->mc_qrv < min_qrv)) { 1303 net_warn_ratelimited("IPv6: MLD: clamping QRV from %u to %u!\n", 1304 idev->mc_qrv, min_qrv); 1305 idev->mc_qrv = min_qrv; 1306 } 1307 } 1308 1309 static void mld_update_qi(struct inet6_dev *idev, 1310 const struct mld2_query *mlh2) 1311 { 1312 /* RFC3810, relevant sections: 1313 * - 5.1.9. QQIC (Querier's Query Interval Code) 1314 * - 9.2. Query Interval 1315 * - 9.12. Older Version Querier Present Timeout 1316 * (the [Query Interval] in the last Query received) 1317 */ 1318 unsigned long mc_qqi; 1319 1320 if (mlh2->mld2q_qqic < 128) { 1321 mc_qqi = mlh2->mld2q_qqic; 1322 } else { 1323 unsigned long mc_man, mc_exp; 1324 1325 mc_exp = MLDV2_QQIC_EXP(mlh2->mld2q_qqic); 1326 mc_man = MLDV2_QQIC_MAN(mlh2->mld2q_qqic); 1327 1328 mc_qqi = (mc_man | 0x10) << (mc_exp + 3); 1329 } 1330 1331 idev->mc_qi = mc_qqi * HZ; 1332 } 1333 1334 static void mld_update_qri(struct inet6_dev *idev, 1335 const struct mld2_query *mlh2) 1336 { 1337 /* RFC3810, relevant sections: 1338 * - 5.1.3. Maximum Response Code 1339 * - 9.3. Query Response Interval 1340 */ 1341 idev->mc_qri = msecs_to_jiffies(mldv2_mrc(mlh2)); 1342 } 1343 1344 static int mld_process_v1(struct inet6_dev *idev, struct mld_msg *mld, 1345 unsigned long *max_delay, bool v1_query) 1346 { 1347 unsigned long mldv1_md; 1348 1349 /* Ignore v1 queries */ 1350 if (mld_in_v2_mode_only(idev)) 1351 return -EINVAL; 1352 1353 mldv1_md = ntohs(mld->mld_maxdelay); 1354 1355 /* When in MLDv1 fallback and a MLDv2 router start-up being 1356 * unaware of current MLDv1 operation, the MRC == MRD mapping 1357 * only works when the exponential algorithm is not being 1358 * used (as MLDv1 is unaware of such things). 1359 * 1360 * According to the RFC author, the MLDv2 implementations 1361 * he's aware of all use a MRC < 32768 on start up queries. 1362 * 1363 * Thus, should we *ever* encounter something else larger 1364 * than that, just assume the maximum possible within our 1365 * reach. 1366 */ 1367 if (!v1_query) 1368 mldv1_md = min(mldv1_md, MLDV1_MRD_MAX_COMPAT); 1369 1370 *max_delay = max(msecs_to_jiffies(mldv1_md), 1UL); 1371 1372 /* MLDv1 router present: we need to go into v1 mode *only* 1373 * when an MLDv1 query is received as per section 9.12. of 1374 * RFC3810! And we know from RFC2710 section 3.7 that MLDv1 1375 * queries MUST be of exactly 24 octets. 1376 */ 1377 if (v1_query) 1378 mld_set_v1_mode(idev); 1379 1380 /* cancel MLDv2 report work */ 1381 mld_gq_stop_work(idev); 1382 /* cancel the interface change work */ 1383 mld_ifc_stop_work(idev); 1384 /* clear deleted report items */ 1385 mld_clear_delrec(idev); 1386 1387 return 0; 1388 } 1389 1390 static void mld_process_v2(struct inet6_dev *idev, struct mld2_query *mld, 1391 unsigned long *max_delay) 1392 { 1393 *max_delay = max(msecs_to_jiffies(mldv2_mrc(mld)), 1UL); 1394 1395 mld_update_qrv(idev, mld); 1396 mld_update_qi(idev, mld); 1397 mld_update_qri(idev, mld); 1398 1399 idev->mc_maxdelay = *max_delay; 1400 1401 return; 1402 } 1403 1404 /* called with rcu_read_lock() */ 1405 void igmp6_event_query(struct sk_buff *skb) 1406 { 1407 struct inet6_dev *idev = __in6_dev_get(skb->dev); 1408 1409 if (!idev || idev->dead) 1410 goto out; 1411 1412 spin_lock_bh(&idev->mc_query_lock); 1413 if (skb_queue_len(&idev->mc_query_queue) < MLD_MAX_SKBS) { 1414 __skb_queue_tail(&idev->mc_query_queue, skb); 1415 if (!mod_delayed_work(mld_wq, &idev->mc_query_work, 0)) 1416 in6_dev_hold(idev); 1417 skb = NULL; 1418 } 1419 spin_unlock_bh(&idev->mc_query_lock); 1420 out: 1421 kfree_skb(skb); 1422 } 1423 1424 static void __mld_query_work(struct sk_buff *skb) 1425 { 1426 struct mld2_query *mlh2 = NULL; 1427 const struct in6_addr *group; 1428 unsigned long max_delay; 1429 struct inet6_dev *idev; 1430 struct ifmcaddr6 *ma; 1431 struct mld_msg *mld; 1432 int group_type; 1433 int mark = 0; 1434 int len, err; 1435 1436 if (!pskb_may_pull(skb, sizeof(struct in6_addr))) 1437 goto kfree_skb; 1438 1439 /* compute payload length excluding extension headers */ 1440 len = ntohs(ipv6_hdr(skb)->payload_len) + sizeof(struct ipv6hdr); 1441 len -= skb_network_header_len(skb); 1442 1443 /* RFC3810 6.2 1444 * Upon reception of an MLD message that contains a Query, the node 1445 * checks if the source address of the message is a valid link-local 1446 * address, if the Hop Limit is set to 1, and if the Router Alert 1447 * option is present in the Hop-By-Hop Options header of the IPv6 1448 * packet. If any of these checks fails, the packet is dropped. 1449 */ 1450 if (!(ipv6_addr_type(&ipv6_hdr(skb)->saddr) & IPV6_ADDR_LINKLOCAL) || 1451 ipv6_hdr(skb)->hop_limit != 1 || 1452 !(IP6CB(skb)->flags & IP6SKB_ROUTERALERT) || 1453 IP6CB(skb)->ra != htons(IPV6_OPT_ROUTERALERT_MLD)) 1454 goto kfree_skb; 1455 1456 idev = in6_dev_get(skb->dev); 1457 if (!idev) 1458 goto kfree_skb; 1459 1460 mld = (struct mld_msg *)icmp6_hdr(skb); 1461 group = &mld->mld_mca; 1462 group_type = ipv6_addr_type(group); 1463 1464 if (group_type != IPV6_ADDR_ANY && 1465 !(group_type&IPV6_ADDR_MULTICAST)) 1466 goto out; 1467 1468 if (len < MLD_V1_QUERY_LEN) { 1469 goto out; 1470 } else if (len == MLD_V1_QUERY_LEN || mld_in_v1_mode(idev)) { 1471 err = mld_process_v1(idev, mld, &max_delay, 1472 len == MLD_V1_QUERY_LEN); 1473 if (err < 0) 1474 goto out; 1475 } else if (len >= MLD_V2_QUERY_LEN_MIN) { 1476 int srcs_offset = sizeof(struct mld2_query) - 1477 sizeof(struct icmp6hdr); 1478 1479 if (!pskb_may_pull(skb, srcs_offset)) 1480 goto out; 1481 1482 mlh2 = (struct mld2_query *)skb_transport_header(skb); 1483 1484 mld_process_v2(idev, mlh2, &max_delay); 1485 1486 if (group_type == IPV6_ADDR_ANY) { /* general query */ 1487 if (mlh2->mld2q_nsrcs) 1488 goto out; /* no sources allowed */ 1489 1490 mld_gq_start_work(idev); 1491 goto out; 1492 } 1493 /* mark sources to include, if group & source-specific */ 1494 if (mlh2->mld2q_nsrcs != 0) { 1495 if (!pskb_may_pull(skb, srcs_offset + 1496 ntohs(mlh2->mld2q_nsrcs) * sizeof(struct in6_addr))) 1497 goto out; 1498 1499 mlh2 = (struct mld2_query *)skb_transport_header(skb); 1500 mark = 1; 1501 } 1502 } else { 1503 goto out; 1504 } 1505 1506 if (group_type == IPV6_ADDR_ANY) { 1507 for_each_mc_mclock(idev, ma) { 1508 igmp6_group_queried(ma, max_delay); 1509 } 1510 } else { 1511 for_each_mc_mclock(idev, ma) { 1512 if (!ipv6_addr_equal(group, &ma->mca_addr)) 1513 continue; 1514 if (ma->mca_flags & MAF_TIMER_RUNNING) { 1515 /* gsquery <- gsquery && mark */ 1516 if (!mark) 1517 ma->mca_flags &= ~MAF_GSQUERY; 1518 } else { 1519 /* gsquery <- mark */ 1520 if (mark) 1521 ma->mca_flags |= MAF_GSQUERY; 1522 else 1523 ma->mca_flags &= ~MAF_GSQUERY; 1524 } 1525 if (!(ma->mca_flags & MAF_GSQUERY) || 1526 mld_marksources(ma, ntohs(mlh2->mld2q_nsrcs), mlh2->mld2q_srcs)) 1527 igmp6_group_queried(ma, max_delay); 1528 break; 1529 } 1530 } 1531 1532 out: 1533 in6_dev_put(idev); 1534 kfree_skb: 1535 consume_skb(skb); 1536 } 1537 1538 static void mld_query_work(struct work_struct *work) 1539 { 1540 struct inet6_dev *idev = container_of(to_delayed_work(work), 1541 struct inet6_dev, 1542 mc_query_work); 1543 struct sk_buff_head q; 1544 struct sk_buff *skb; 1545 bool rework = false; 1546 int cnt = 0; 1547 1548 skb_queue_head_init(&q); 1549 1550 spin_lock_bh(&idev->mc_query_lock); 1551 while ((skb = __skb_dequeue(&idev->mc_query_queue))) { 1552 __skb_queue_tail(&q, skb); 1553 1554 if (++cnt >= MLD_MAX_QUEUE) { 1555 rework = true; 1556 break; 1557 } 1558 } 1559 spin_unlock_bh(&idev->mc_query_lock); 1560 1561 mutex_lock(&idev->mc_lock); 1562 while ((skb = __skb_dequeue(&q))) 1563 __mld_query_work(skb); 1564 mutex_unlock(&idev->mc_lock); 1565 1566 if (rework && queue_delayed_work(mld_wq, &idev->mc_query_work, 0)) 1567 return; 1568 1569 in6_dev_put(idev); 1570 } 1571 1572 /* called with rcu_read_lock() */ 1573 void igmp6_event_report(struct sk_buff *skb) 1574 { 1575 struct inet6_dev *idev = __in6_dev_get(skb->dev); 1576 1577 if (!idev || idev->dead) 1578 goto out; 1579 1580 spin_lock_bh(&idev->mc_report_lock); 1581 if (skb_queue_len(&idev->mc_report_queue) < MLD_MAX_SKBS) { 1582 __skb_queue_tail(&idev->mc_report_queue, skb); 1583 if (!mod_delayed_work(mld_wq, &idev->mc_report_work, 0)) 1584 in6_dev_hold(idev); 1585 skb = NULL; 1586 } 1587 spin_unlock_bh(&idev->mc_report_lock); 1588 out: 1589 kfree_skb(skb); 1590 } 1591 1592 static void __mld_report_work(struct sk_buff *skb) 1593 { 1594 struct inet6_dev *idev; 1595 struct ifmcaddr6 *ma; 1596 struct mld_msg *mld; 1597 int addr_type; 1598 1599 /* Our own report looped back. Ignore it. */ 1600 if (skb->pkt_type == PACKET_LOOPBACK) 1601 goto kfree_skb; 1602 1603 /* send our report if the MC router may not have heard this report */ 1604 if (skb->pkt_type != PACKET_MULTICAST && 1605 skb->pkt_type != PACKET_BROADCAST) 1606 goto kfree_skb; 1607 1608 if (!pskb_may_pull(skb, sizeof(*mld) - sizeof(struct icmp6hdr))) 1609 goto kfree_skb; 1610 1611 mld = (struct mld_msg *)icmp6_hdr(skb); 1612 1613 /* Drop reports with not link local source */ 1614 addr_type = ipv6_addr_type(&ipv6_hdr(skb)->saddr); 1615 if (addr_type != IPV6_ADDR_ANY && 1616 !(addr_type&IPV6_ADDR_LINKLOCAL)) 1617 goto kfree_skb; 1618 1619 idev = in6_dev_get(skb->dev); 1620 if (!idev) 1621 goto kfree_skb; 1622 1623 /* 1624 * Cancel the work for this group 1625 */ 1626 1627 for_each_mc_mclock(idev, ma) { 1628 if (ipv6_addr_equal(&ma->mca_addr, &mld->mld_mca)) { 1629 if (cancel_delayed_work(&ma->mca_work)) 1630 refcount_dec(&ma->mca_refcnt); 1631 ma->mca_flags &= ~(MAF_LAST_REPORTER | 1632 MAF_TIMER_RUNNING); 1633 break; 1634 } 1635 } 1636 1637 in6_dev_put(idev); 1638 kfree_skb: 1639 consume_skb(skb); 1640 } 1641 1642 static void mld_report_work(struct work_struct *work) 1643 { 1644 struct inet6_dev *idev = container_of(to_delayed_work(work), 1645 struct inet6_dev, 1646 mc_report_work); 1647 struct sk_buff_head q; 1648 struct sk_buff *skb; 1649 bool rework = false; 1650 int cnt = 0; 1651 1652 skb_queue_head_init(&q); 1653 spin_lock_bh(&idev->mc_report_lock); 1654 while ((skb = __skb_dequeue(&idev->mc_report_queue))) { 1655 __skb_queue_tail(&q, skb); 1656 1657 if (++cnt >= MLD_MAX_QUEUE) { 1658 rework = true; 1659 break; 1660 } 1661 } 1662 spin_unlock_bh(&idev->mc_report_lock); 1663 1664 mutex_lock(&idev->mc_lock); 1665 while ((skb = __skb_dequeue(&q))) 1666 __mld_report_work(skb); 1667 mutex_unlock(&idev->mc_lock); 1668 1669 if (rework && queue_delayed_work(mld_wq, &idev->mc_report_work, 0)) 1670 return; 1671 1672 in6_dev_put(idev); 1673 } 1674 1675 static bool is_in(struct ifmcaddr6 *pmc, struct ip6_sf_list *psf, int type, 1676 int gdeleted, int sdeleted) 1677 { 1678 switch (type) { 1679 case MLD2_MODE_IS_INCLUDE: 1680 case MLD2_MODE_IS_EXCLUDE: 1681 if (gdeleted || sdeleted) 1682 return false; 1683 if (!((pmc->mca_flags & MAF_GSQUERY) && !psf->sf_gsresp)) { 1684 if (pmc->mca_sfmode == MCAST_INCLUDE) 1685 return true; 1686 /* don't include if this source is excluded 1687 * in all filters 1688 */ 1689 if (psf->sf_count[MCAST_INCLUDE]) 1690 return type == MLD2_MODE_IS_INCLUDE; 1691 return pmc->mca_sfcount[MCAST_EXCLUDE] == 1692 psf->sf_count[MCAST_EXCLUDE]; 1693 } 1694 return false; 1695 case MLD2_CHANGE_TO_INCLUDE: 1696 if (gdeleted || sdeleted) 1697 return false; 1698 return psf->sf_count[MCAST_INCLUDE] != 0; 1699 case MLD2_CHANGE_TO_EXCLUDE: 1700 if (gdeleted || sdeleted) 1701 return false; 1702 if (pmc->mca_sfcount[MCAST_EXCLUDE] == 0 || 1703 psf->sf_count[MCAST_INCLUDE]) 1704 return false; 1705 return pmc->mca_sfcount[MCAST_EXCLUDE] == 1706 psf->sf_count[MCAST_EXCLUDE]; 1707 case MLD2_ALLOW_NEW_SOURCES: 1708 if (gdeleted || !psf->sf_crcount) 1709 return false; 1710 return (pmc->mca_sfmode == MCAST_INCLUDE) ^ sdeleted; 1711 case MLD2_BLOCK_OLD_SOURCES: 1712 if (pmc->mca_sfmode == MCAST_INCLUDE) 1713 return gdeleted || (psf->sf_crcount && sdeleted); 1714 return psf->sf_crcount && !gdeleted && !sdeleted; 1715 } 1716 return false; 1717 } 1718 1719 static int 1720 mld_scount(struct ifmcaddr6 *pmc, int type, int gdeleted, int sdeleted) 1721 { 1722 struct ip6_sf_list *psf; 1723 int scount = 0; 1724 1725 for_each_psf_mclock(pmc, psf) { 1726 if (!is_in(pmc, psf, type, gdeleted, sdeleted)) 1727 continue; 1728 scount++; 1729 } 1730 return scount; 1731 } 1732 1733 static void ip6_mc_hdr(const struct sock *sk, struct sk_buff *skb, 1734 struct net_device *dev, const struct in6_addr *saddr, 1735 const struct in6_addr *daddr, int proto, int len) 1736 { 1737 struct ipv6hdr *hdr; 1738 1739 skb->protocol = htons(ETH_P_IPV6); 1740 skb->dev = dev; 1741 1742 skb_reset_network_header(skb); 1743 skb_put(skb, sizeof(struct ipv6hdr)); 1744 hdr = ipv6_hdr(skb); 1745 1746 ip6_flow_hdr(hdr, 0, 0); 1747 1748 hdr->payload_len = htons(len); 1749 hdr->nexthdr = proto; 1750 hdr->hop_limit = READ_ONCE(inet6_sk(sk)->hop_limit); 1751 1752 hdr->saddr = *saddr; 1753 hdr->daddr = *daddr; 1754 } 1755 1756 static struct sk_buff *mld_newpack(struct inet6_dev *idev, unsigned int mtu) 1757 { 1758 u8 ra[8] = { IPPROTO_ICMPV6, 0, IPV6_TLV_ROUTERALERT, 1759 2, 0, 0, IPV6_TLV_PADN, 0 }; 1760 struct net_device *dev = idev->dev; 1761 int hlen = LL_RESERVED_SPACE(dev); 1762 int tlen = dev->needed_tailroom; 1763 const struct in6_addr *saddr; 1764 struct in6_addr addr_buf; 1765 struct mld2_report *pmr; 1766 struct sk_buff *skb; 1767 unsigned int size; 1768 struct sock *sk; 1769 struct net *net; 1770 1771 /* we assume size > sizeof(ra) here 1772 * Also try to not allocate high-order pages for big MTU 1773 */ 1774 size = min_t(int, mtu, PAGE_SIZE / 2) + hlen + tlen; 1775 skb = alloc_skb(size, GFP_KERNEL); 1776 if (!skb) 1777 return NULL; 1778 1779 skb->priority = TC_PRIO_CONTROL; 1780 skb_reserve(skb, hlen); 1781 skb_tailroom_reserve(skb, mtu, tlen); 1782 1783 rcu_read_lock(); 1784 1785 net = dev_net_rcu(dev); 1786 sk = net->ipv6.igmp_sk; 1787 skb_set_owner_w(skb, sk); 1788 1789 if (ipv6_get_lladdr(dev, &addr_buf, IFA_F_TENTATIVE)) { 1790 /* <draft-ietf-magma-mld-source-05.txt>: 1791 * use unspecified address as the source address 1792 * when a valid link-local address is not available. 1793 */ 1794 saddr = &in6addr_any; 1795 } else 1796 saddr = &addr_buf; 1797 1798 ip6_mc_hdr(sk, skb, dev, saddr, &mld2_all_mcr, NEXTHDR_HOP, 0); 1799 1800 rcu_read_unlock(); 1801 1802 skb_put_data(skb, ra, sizeof(ra)); 1803 1804 skb_set_transport_header(skb, skb_tail_pointer(skb) - skb->data); 1805 skb_put(skb, sizeof(*pmr)); 1806 pmr = (struct mld2_report *)skb_transport_header(skb); 1807 pmr->mld2r_type = ICMPV6_MLD2_REPORT; 1808 pmr->mld2r_resv1 = 0; 1809 pmr->mld2r_cksum = 0; 1810 pmr->mld2r_resv2 = 0; 1811 pmr->mld2r_ngrec = 0; 1812 return skb; 1813 } 1814 1815 static void mld_sendpack(struct sk_buff *skb) 1816 { 1817 struct ipv6hdr *pip6 = ipv6_hdr(skb); 1818 struct mld2_report *pmr = 1819 (struct mld2_report *)skb_transport_header(skb); 1820 int payload_len, mldlen; 1821 struct inet6_dev *idev; 1822 struct net *net = dev_net(skb->dev); 1823 int err; 1824 struct flowi6 fl6; 1825 struct dst_entry *dst; 1826 1827 rcu_read_lock(); 1828 idev = __in6_dev_get(skb->dev); 1829 IP6_INC_STATS(net, idev, IPSTATS_MIB_OUTREQUESTS); 1830 1831 payload_len = (skb_tail_pointer(skb) - skb_network_header(skb)) - 1832 sizeof(*pip6); 1833 mldlen = skb_tail_pointer(skb) - skb_transport_header(skb); 1834 pip6->payload_len = htons(payload_len); 1835 1836 pmr->mld2r_cksum = csum_ipv6_magic(&pip6->saddr, &pip6->daddr, mldlen, 1837 IPPROTO_ICMPV6, 1838 csum_partial(skb_transport_header(skb), 1839 mldlen, 0)); 1840 1841 icmpv6_flow_init(net->ipv6.igmp_sk, &fl6, ICMPV6_MLD2_REPORT, 1842 &ipv6_hdr(skb)->saddr, &ipv6_hdr(skb)->daddr, 1843 skb->dev->ifindex); 1844 dst = icmp6_dst_alloc(skb->dev, &fl6); 1845 1846 err = 0; 1847 if (IS_ERR(dst)) { 1848 err = PTR_ERR(dst); 1849 dst = NULL; 1850 } 1851 skb_dst_set(skb, dst); 1852 if (err) 1853 goto err_out; 1854 1855 err = NF_HOOK(NFPROTO_IPV6, NF_INET_LOCAL_OUT, 1856 net, net->ipv6.igmp_sk, skb, NULL, skb->dev, 1857 dst_output); 1858 out: 1859 if (!err) { 1860 ICMP6MSGOUT_INC_STATS(net, idev, ICMPV6_MLD2_REPORT); 1861 ICMP6_INC_STATS(net, idev, ICMP6_MIB_OUTMSGS); 1862 } else { 1863 IP6_INC_STATS(net, idev, IPSTATS_MIB_OUTDISCARDS); 1864 } 1865 1866 rcu_read_unlock(); 1867 return; 1868 1869 err_out: 1870 kfree_skb(skb); 1871 goto out; 1872 } 1873 1874 static int grec_size(struct ifmcaddr6 *pmc, int type, int gdel, int sdel) 1875 { 1876 return sizeof(struct mld2_grec) + 16 * mld_scount(pmc,type,gdel,sdel); 1877 } 1878 1879 static struct sk_buff *add_grhead(struct sk_buff *skb, struct ifmcaddr6 *pmc, 1880 int type, struct mld2_grec **ppgr, unsigned int mtu) 1881 { 1882 struct mld2_report *pmr; 1883 struct mld2_grec *pgr; 1884 1885 if (!skb) { 1886 skb = mld_newpack(pmc->idev, mtu); 1887 if (!skb) 1888 return NULL; 1889 } 1890 pgr = skb_put(skb, sizeof(struct mld2_grec)); 1891 pgr->grec_type = type; 1892 pgr->grec_auxwords = 0; 1893 pgr->grec_nsrcs = 0; 1894 pgr->grec_mca = pmc->mca_addr; /* structure copy */ 1895 pmr = (struct mld2_report *)skb_transport_header(skb); 1896 pmr->mld2r_ngrec = htons(ntohs(pmr->mld2r_ngrec)+1); 1897 *ppgr = pgr; 1898 return skb; 1899 } 1900 1901 #define AVAILABLE(skb) ((skb) ? skb_availroom(skb) : 0) 1902 1903 static struct sk_buff *add_grec(struct sk_buff *skb, struct ifmcaddr6 *pmc, 1904 int type, int gdeleted, int sdeleted, 1905 int crsend) 1906 { 1907 struct ip6_sf_list *psf, *psf_prev, *psf_next; 1908 int scount, stotal, first, isquery, truncate; 1909 struct ip6_sf_list __rcu **psf_list; 1910 struct inet6_dev *idev = pmc->idev; 1911 struct net_device *dev = idev->dev; 1912 struct mld2_grec *pgr = NULL; 1913 struct mld2_report *pmr; 1914 unsigned int mtu; 1915 1916 mc_assert_locked(idev); 1917 1918 if (pmc->mca_flags & MAF_NOREPORT) 1919 return skb; 1920 1921 mtu = READ_ONCE(dev->mtu); 1922 if (mtu < IPV6_MIN_MTU) 1923 return skb; 1924 1925 isquery = type == MLD2_MODE_IS_INCLUDE || 1926 type == MLD2_MODE_IS_EXCLUDE; 1927 truncate = type == MLD2_MODE_IS_EXCLUDE || 1928 type == MLD2_CHANGE_TO_EXCLUDE; 1929 1930 stotal = scount = 0; 1931 1932 psf_list = sdeleted ? &pmc->mca_tomb : &pmc->mca_sources; 1933 1934 if (!rcu_access_pointer(*psf_list)) 1935 goto empty_source; 1936 1937 pmr = skb ? (struct mld2_report *)skb_transport_header(skb) : NULL; 1938 1939 /* EX and TO_EX get a fresh packet, if needed */ 1940 if (truncate) { 1941 if (pmr && pmr->mld2r_ngrec && 1942 AVAILABLE(skb) < grec_size(pmc, type, gdeleted, sdeleted)) { 1943 if (skb) 1944 mld_sendpack(skb); 1945 skb = mld_newpack(idev, mtu); 1946 } 1947 } 1948 first = 1; 1949 psf_prev = NULL; 1950 for (psf = mc_dereference(*psf_list, idev); 1951 psf; 1952 psf = psf_next) { 1953 struct in6_addr *psrc; 1954 1955 psf_next = mc_dereference(psf->sf_next, idev); 1956 1957 if (!is_in(pmc, psf, type, gdeleted, sdeleted) && !crsend) { 1958 psf_prev = psf; 1959 continue; 1960 } 1961 1962 /* Based on RFC3810 6.1. Should not send source-list change 1963 * records when there is a filter mode change. 1964 */ 1965 if (((gdeleted && pmc->mca_sfmode == MCAST_EXCLUDE) || 1966 (!gdeleted && pmc->mca_crcount)) && 1967 (type == MLD2_ALLOW_NEW_SOURCES || 1968 type == MLD2_BLOCK_OLD_SOURCES) && psf->sf_crcount) 1969 goto decrease_sf_crcount; 1970 1971 /* clear marks on query responses */ 1972 if (isquery) 1973 psf->sf_gsresp = 0; 1974 1975 if (AVAILABLE(skb) < sizeof(*psrc) + 1976 first*sizeof(struct mld2_grec)) { 1977 if (truncate && !first) 1978 break; /* truncate these */ 1979 if (pgr) 1980 pgr->grec_nsrcs = htons(scount); 1981 if (skb) 1982 mld_sendpack(skb); 1983 skb = mld_newpack(idev, mtu); 1984 first = 1; 1985 scount = 0; 1986 } 1987 if (first) { 1988 skb = add_grhead(skb, pmc, type, &pgr, mtu); 1989 first = 0; 1990 } 1991 if (!skb) 1992 return NULL; 1993 psrc = skb_put(skb, sizeof(*psrc)); 1994 *psrc = psf->sf_addr; 1995 scount++; stotal++; 1996 if ((type == MLD2_ALLOW_NEW_SOURCES || 1997 type == MLD2_BLOCK_OLD_SOURCES) && psf->sf_crcount) { 1998 decrease_sf_crcount: 1999 psf->sf_crcount--; 2000 if ((sdeleted || gdeleted) && psf->sf_crcount == 0) { 2001 if (psf_prev) 2002 rcu_assign_pointer(psf_prev->sf_next, 2003 mc_dereference(psf->sf_next, idev)); 2004 else 2005 rcu_assign_pointer(*psf_list, 2006 mc_dereference(psf->sf_next, idev)); 2007 kfree_rcu(psf, rcu); 2008 continue; 2009 } 2010 } 2011 psf_prev = psf; 2012 } 2013 2014 empty_source: 2015 if (!stotal) { 2016 if (type == MLD2_ALLOW_NEW_SOURCES || 2017 type == MLD2_BLOCK_OLD_SOURCES) 2018 return skb; 2019 if (pmc->mca_crcount || isquery || crsend) { 2020 /* make sure we have room for group header */ 2021 if (skb && AVAILABLE(skb) < sizeof(struct mld2_grec)) { 2022 mld_sendpack(skb); 2023 skb = NULL; /* add_grhead will get a new one */ 2024 } 2025 skb = add_grhead(skb, pmc, type, &pgr, mtu); 2026 } 2027 } 2028 if (pgr) 2029 pgr->grec_nsrcs = htons(scount); 2030 2031 if (isquery) 2032 pmc->mca_flags &= ~MAF_GSQUERY; /* clear query state */ 2033 return skb; 2034 } 2035 2036 static void mld_send_report(struct inet6_dev *idev, struct ifmcaddr6 *pmc) 2037 { 2038 struct sk_buff *skb = NULL; 2039 int type; 2040 2041 mc_assert_locked(idev); 2042 2043 if (!pmc) { 2044 for_each_mc_mclock(idev, pmc) { 2045 if (pmc->mca_flags & MAF_NOREPORT) 2046 continue; 2047 if (pmc->mca_sfcount[MCAST_EXCLUDE]) 2048 type = MLD2_MODE_IS_EXCLUDE; 2049 else 2050 type = MLD2_MODE_IS_INCLUDE; 2051 skb = add_grec(skb, pmc, type, 0, 0, 0); 2052 } 2053 } else { 2054 if (pmc->mca_sfcount[MCAST_EXCLUDE]) 2055 type = MLD2_MODE_IS_EXCLUDE; 2056 else 2057 type = MLD2_MODE_IS_INCLUDE; 2058 skb = add_grec(skb, pmc, type, 0, 0, 0); 2059 } 2060 if (skb) 2061 mld_sendpack(skb); 2062 } 2063 2064 /* remove zero-count source records from a source filter list */ 2065 static void mld_clear_zeros(struct ip6_sf_list __rcu **ppsf, struct inet6_dev *idev) 2066 { 2067 struct ip6_sf_list *psf_prev, *psf_next, *psf; 2068 2069 psf_prev = NULL; 2070 for (psf = mc_dereference(*ppsf, idev); 2071 psf; 2072 psf = psf_next) { 2073 psf_next = mc_dereference(psf->sf_next, idev); 2074 if (psf->sf_crcount == 0) { 2075 if (psf_prev) 2076 rcu_assign_pointer(psf_prev->sf_next, 2077 mc_dereference(psf->sf_next, idev)); 2078 else 2079 rcu_assign_pointer(*ppsf, 2080 mc_dereference(psf->sf_next, idev)); 2081 kfree_rcu(psf, rcu); 2082 } else { 2083 psf_prev = psf; 2084 } 2085 } 2086 } 2087 2088 static void mld_send_cr(struct inet6_dev *idev) 2089 { 2090 struct ifmcaddr6 *pmc, *pmc_prev, *pmc_next; 2091 struct sk_buff *skb = NULL; 2092 int type, dtype; 2093 2094 /* deleted MCA's */ 2095 pmc_prev = NULL; 2096 for (pmc = mc_dereference(idev->mc_tomb, idev); 2097 pmc; 2098 pmc = pmc_next) { 2099 pmc_next = mc_dereference(pmc->next, idev); 2100 if (pmc->mca_sfmode == MCAST_INCLUDE) { 2101 type = MLD2_BLOCK_OLD_SOURCES; 2102 dtype = MLD2_BLOCK_OLD_SOURCES; 2103 skb = add_grec(skb, pmc, type, 1, 0, 0); 2104 skb = add_grec(skb, pmc, dtype, 1, 1, 0); 2105 } 2106 if (pmc->mca_crcount) { 2107 if (pmc->mca_sfmode == MCAST_EXCLUDE) { 2108 type = MLD2_CHANGE_TO_INCLUDE; 2109 skb = add_grec(skb, pmc, type, 1, 0, 0); 2110 } 2111 pmc->mca_crcount--; 2112 if (pmc->mca_crcount == 0) { 2113 mld_clear_zeros(&pmc->mca_tomb, idev); 2114 mld_clear_zeros(&pmc->mca_sources, idev); 2115 } 2116 } 2117 if (pmc->mca_crcount == 0 && 2118 !rcu_access_pointer(pmc->mca_tomb) && 2119 !rcu_access_pointer(pmc->mca_sources)) { 2120 if (pmc_prev) 2121 rcu_assign_pointer(pmc_prev->next, pmc_next); 2122 else 2123 rcu_assign_pointer(idev->mc_tomb, pmc_next); 2124 in6_dev_put(pmc->idev); 2125 kfree_rcu(pmc, rcu); 2126 } else 2127 pmc_prev = pmc; 2128 } 2129 2130 /* change recs */ 2131 for_each_mc_mclock(idev, pmc) { 2132 if (pmc->mca_sfcount[MCAST_EXCLUDE]) { 2133 type = MLD2_BLOCK_OLD_SOURCES; 2134 dtype = MLD2_ALLOW_NEW_SOURCES; 2135 } else { 2136 type = MLD2_ALLOW_NEW_SOURCES; 2137 dtype = MLD2_BLOCK_OLD_SOURCES; 2138 } 2139 skb = add_grec(skb, pmc, type, 0, 0, 0); 2140 skb = add_grec(skb, pmc, dtype, 0, 1, 0); /* deleted sources */ 2141 2142 /* filter mode changes */ 2143 if (pmc->mca_crcount) { 2144 if (pmc->mca_sfmode == MCAST_EXCLUDE) 2145 type = MLD2_CHANGE_TO_EXCLUDE; 2146 else 2147 type = MLD2_CHANGE_TO_INCLUDE; 2148 skb = add_grec(skb, pmc, type, 0, 0, 0); 2149 pmc->mca_crcount--; 2150 } 2151 } 2152 if (!skb) 2153 return; 2154 (void) mld_sendpack(skb); 2155 } 2156 2157 static void igmp6_send(struct in6_addr *addr, struct net_device *dev, int type) 2158 { 2159 const struct in6_addr *snd_addr, *saddr; 2160 int err, len, payload_len, full_len; 2161 struct in6_addr addr_buf; 2162 struct inet6_dev *idev; 2163 struct sk_buff *skb; 2164 struct mld_msg *hdr; 2165 int hlen = LL_RESERVED_SPACE(dev); 2166 int tlen = dev->needed_tailroom; 2167 u8 ra[8] = { IPPROTO_ICMPV6, 0, 2168 IPV6_TLV_ROUTERALERT, 2, 0, 0, 2169 IPV6_TLV_PADN, 0 }; 2170 struct dst_entry *dst; 2171 struct flowi6 fl6; 2172 struct net *net; 2173 struct sock *sk; 2174 2175 if (type == ICMPV6_MGM_REDUCTION) 2176 snd_addr = &in6addr_linklocal_allrouters; 2177 else 2178 snd_addr = addr; 2179 2180 len = sizeof(struct icmp6hdr) + sizeof(struct in6_addr); 2181 payload_len = len + sizeof(ra); 2182 full_len = sizeof(struct ipv6hdr) + payload_len; 2183 2184 skb = alloc_skb(hlen + tlen + full_len, GFP_KERNEL); 2185 2186 rcu_read_lock(); 2187 2188 net = dev_net_rcu(dev); 2189 idev = __in6_dev_get(dev); 2190 IP6_INC_STATS(net, idev, IPSTATS_MIB_OUTREQUESTS); 2191 if (!skb) { 2192 IP6_INC_STATS(net, idev, IPSTATS_MIB_OUTDISCARDS); 2193 rcu_read_unlock(); 2194 return; 2195 } 2196 sk = net->ipv6.igmp_sk; 2197 skb_set_owner_w(skb, sk); 2198 2199 skb->priority = TC_PRIO_CONTROL; 2200 skb_reserve(skb, hlen); 2201 2202 if (ipv6_get_lladdr(dev, &addr_buf, IFA_F_TENTATIVE)) { 2203 /* <draft-ietf-magma-mld-source-05.txt>: 2204 * use unspecified address as the source address 2205 * when a valid link-local address is not available. 2206 */ 2207 saddr = &in6addr_any; 2208 } else 2209 saddr = &addr_buf; 2210 2211 ip6_mc_hdr(sk, skb, dev, saddr, snd_addr, NEXTHDR_HOP, payload_len); 2212 2213 skb_put_data(skb, ra, sizeof(ra)); 2214 2215 hdr = skb_put_zero(skb, sizeof(struct mld_msg)); 2216 hdr->mld_type = type; 2217 hdr->mld_mca = *addr; 2218 2219 hdr->mld_cksum = csum_ipv6_magic(saddr, snd_addr, len, 2220 IPPROTO_ICMPV6, 2221 csum_partial(hdr, len, 0)); 2222 2223 icmpv6_flow_init(sk, &fl6, type, 2224 &ipv6_hdr(skb)->saddr, &ipv6_hdr(skb)->daddr, 2225 skb->dev->ifindex); 2226 dst = icmp6_dst_alloc(skb->dev, &fl6); 2227 if (IS_ERR(dst)) { 2228 err = PTR_ERR(dst); 2229 goto err_out; 2230 } 2231 2232 skb_dst_set(skb, dst); 2233 err = NF_HOOK(NFPROTO_IPV6, NF_INET_LOCAL_OUT, 2234 net, sk, skb, NULL, skb->dev, 2235 dst_output); 2236 out: 2237 if (!err) { 2238 ICMP6MSGOUT_INC_STATS(net, idev, type); 2239 ICMP6_INC_STATS(net, idev, ICMP6_MIB_OUTMSGS); 2240 } else 2241 IP6_INC_STATS(net, idev, IPSTATS_MIB_OUTDISCARDS); 2242 2243 rcu_read_unlock(); 2244 return; 2245 2246 err_out: 2247 kfree_skb(skb); 2248 goto out; 2249 } 2250 2251 static void mld_send_initial_cr(struct inet6_dev *idev) 2252 { 2253 struct ifmcaddr6 *pmc; 2254 struct sk_buff *skb; 2255 int type; 2256 2257 mc_assert_locked(idev); 2258 2259 if (mld_in_v1_mode(idev)) 2260 return; 2261 2262 skb = NULL; 2263 for_each_mc_mclock(idev, pmc) { 2264 if (pmc->mca_sfcount[MCAST_EXCLUDE]) 2265 type = MLD2_CHANGE_TO_EXCLUDE; 2266 else 2267 type = MLD2_ALLOW_NEW_SOURCES; 2268 skb = add_grec(skb, pmc, type, 0, 0, 1); 2269 } 2270 if (skb) 2271 mld_sendpack(skb); 2272 } 2273 2274 void ipv6_mc_dad_complete(struct inet6_dev *idev) 2275 { 2276 mutex_lock(&idev->mc_lock); 2277 idev->mc_dad_count = idev->mc_qrv; 2278 if (idev->mc_dad_count) { 2279 mld_send_initial_cr(idev); 2280 idev->mc_dad_count--; 2281 if (idev->mc_dad_count) 2282 mld_dad_start_work(idev, 2283 unsolicited_report_interval(idev)); 2284 } 2285 mutex_unlock(&idev->mc_lock); 2286 } 2287 2288 static void mld_dad_work(struct work_struct *work) 2289 { 2290 struct inet6_dev *idev = container_of(to_delayed_work(work), 2291 struct inet6_dev, 2292 mc_dad_work); 2293 mutex_lock(&idev->mc_lock); 2294 mld_send_initial_cr(idev); 2295 if (idev->mc_dad_count) { 2296 idev->mc_dad_count--; 2297 if (idev->mc_dad_count) 2298 mld_dad_start_work(idev, 2299 unsolicited_report_interval(idev)); 2300 } 2301 mutex_unlock(&idev->mc_lock); 2302 in6_dev_put(idev); 2303 } 2304 2305 static int ip6_mc_del1_src(struct ifmcaddr6 *pmc, int sfmode, 2306 const struct in6_addr *psfsrc) 2307 { 2308 struct ip6_sf_list *psf, *psf_prev; 2309 int rv = 0; 2310 2311 mc_assert_locked(pmc->idev); 2312 2313 psf_prev = NULL; 2314 for_each_psf_mclock(pmc, psf) { 2315 if (ipv6_addr_equal(&psf->sf_addr, psfsrc)) 2316 break; 2317 psf_prev = psf; 2318 } 2319 if (!psf || psf->sf_count[sfmode] == 0) { 2320 /* source filter not found, or count wrong => bug */ 2321 return -ESRCH; 2322 } 2323 WRITE_ONCE(psf->sf_count[sfmode], psf->sf_count[sfmode] - 1); 2324 if (!psf->sf_count[MCAST_INCLUDE] && !psf->sf_count[MCAST_EXCLUDE]) { 2325 struct inet6_dev *idev = pmc->idev; 2326 2327 /* no more filters for this source */ 2328 if (psf_prev) 2329 rcu_assign_pointer(psf_prev->sf_next, 2330 mc_dereference(psf->sf_next, idev)); 2331 else 2332 rcu_assign_pointer(pmc->mca_sources, 2333 mc_dereference(psf->sf_next, idev)); 2334 2335 if (psf->sf_oldin && !(pmc->mca_flags & MAF_NOREPORT) && 2336 !mld_in_v1_mode(idev)) { 2337 psf->sf_crcount = idev->mc_qrv; 2338 rcu_assign_pointer(psf->sf_next, 2339 mc_dereference(pmc->mca_tomb, idev)); 2340 rcu_assign_pointer(pmc->mca_tomb, psf); 2341 rv = 1; 2342 } else { 2343 kfree_rcu(psf, rcu); 2344 } 2345 } 2346 return rv; 2347 } 2348 2349 static int ip6_mc_del_src(struct inet6_dev *idev, const struct in6_addr *pmca, 2350 int sfmode, int sfcount, const struct in6_addr *psfsrc, 2351 int delta) 2352 { 2353 struct ifmcaddr6 *pmc; 2354 int changerec = 0; 2355 int i, err; 2356 2357 if (!idev) 2358 return -ENODEV; 2359 2360 mc_assert_locked(idev); 2361 2362 for_each_mc_mclock(idev, pmc) { 2363 if (ipv6_addr_equal(pmca, &pmc->mca_addr)) 2364 break; 2365 } 2366 if (!pmc) 2367 return -ESRCH; 2368 2369 sf_markstate(pmc); 2370 if (!delta) { 2371 if (!pmc->mca_sfcount[sfmode]) 2372 return -EINVAL; 2373 2374 pmc->mca_sfcount[sfmode]--; 2375 } 2376 err = 0; 2377 for (i = 0; i < sfcount; i++) { 2378 int rv = ip6_mc_del1_src(pmc, sfmode, &psfsrc[i]); 2379 2380 changerec |= rv > 0; 2381 if (!err && rv < 0) 2382 err = rv; 2383 } 2384 if (pmc->mca_sfmode == MCAST_EXCLUDE && 2385 pmc->mca_sfcount[MCAST_EXCLUDE] == 0 && 2386 pmc->mca_sfcount[MCAST_INCLUDE]) { 2387 struct ip6_sf_list *psf; 2388 2389 /* filter mode change */ 2390 pmc->mca_sfmode = MCAST_INCLUDE; 2391 pmc->mca_crcount = idev->mc_qrv; 2392 idev->mc_ifc_count = pmc->mca_crcount; 2393 for_each_psf_mclock(pmc, psf) 2394 psf->sf_crcount = 0; 2395 mld_ifc_event(pmc->idev); 2396 } else if (sf_setstate(pmc) || changerec) { 2397 mld_ifc_event(pmc->idev); 2398 } 2399 2400 return err; 2401 } 2402 2403 /* Add multicast single-source filter to the interface list */ 2404 static int ip6_mc_add1_src(struct ifmcaddr6 *pmc, int sfmode, 2405 const struct in6_addr *psfsrc) 2406 { 2407 struct ip6_sf_list *psf, *psf_prev; 2408 2409 mc_assert_locked(pmc->idev); 2410 2411 psf_prev = NULL; 2412 for_each_psf_mclock(pmc, psf) { 2413 if (ipv6_addr_equal(&psf->sf_addr, psfsrc)) 2414 break; 2415 psf_prev = psf; 2416 } 2417 if (!psf) { 2418 psf = kzalloc(sizeof(*psf), GFP_KERNEL); 2419 if (!psf) 2420 return -ENOBUFS; 2421 2422 psf->sf_addr = *psfsrc; 2423 if (psf_prev) { 2424 rcu_assign_pointer(psf_prev->sf_next, psf); 2425 } else { 2426 rcu_assign_pointer(pmc->mca_sources, psf); 2427 } 2428 } 2429 WRITE_ONCE(psf->sf_count[sfmode], psf->sf_count[sfmode] + 1); 2430 return 0; 2431 } 2432 2433 static void sf_markstate(struct ifmcaddr6 *pmc) 2434 { 2435 int mca_xcount = pmc->mca_sfcount[MCAST_EXCLUDE]; 2436 struct ip6_sf_list *psf; 2437 2438 mc_assert_locked(pmc->idev); 2439 2440 for_each_psf_mclock(pmc, psf) { 2441 if (pmc->mca_sfcount[MCAST_EXCLUDE]) { 2442 psf->sf_oldin = mca_xcount == 2443 psf->sf_count[MCAST_EXCLUDE] && 2444 !psf->sf_count[MCAST_INCLUDE]; 2445 } else { 2446 psf->sf_oldin = psf->sf_count[MCAST_INCLUDE] != 0; 2447 } 2448 } 2449 } 2450 2451 static int sf_setstate(struct ifmcaddr6 *pmc) 2452 { 2453 int mca_xcount = pmc->mca_sfcount[MCAST_EXCLUDE]; 2454 struct ip6_sf_list *psf, *dpsf; 2455 int qrv = pmc->idev->mc_qrv; 2456 int new_in, rv; 2457 2458 mc_assert_locked(pmc->idev); 2459 2460 rv = 0; 2461 for_each_psf_mclock(pmc, psf) { 2462 if (pmc->mca_sfcount[MCAST_EXCLUDE]) { 2463 new_in = mca_xcount == psf->sf_count[MCAST_EXCLUDE] && 2464 !psf->sf_count[MCAST_INCLUDE]; 2465 } else 2466 new_in = psf->sf_count[MCAST_INCLUDE] != 0; 2467 if (new_in) { 2468 if (!psf->sf_oldin) { 2469 struct ip6_sf_list *prev = NULL; 2470 2471 for_each_psf_tomb(pmc, dpsf) { 2472 if (ipv6_addr_equal(&dpsf->sf_addr, 2473 &psf->sf_addr)) 2474 break; 2475 prev = dpsf; 2476 } 2477 if (dpsf) { 2478 if (prev) 2479 rcu_assign_pointer(prev->sf_next, 2480 mc_dereference(dpsf->sf_next, 2481 pmc->idev)); 2482 else 2483 rcu_assign_pointer(pmc->mca_tomb, 2484 mc_dereference(dpsf->sf_next, 2485 pmc->idev)); 2486 kfree_rcu(dpsf, rcu); 2487 } 2488 psf->sf_crcount = qrv; 2489 rv++; 2490 } 2491 } else if (psf->sf_oldin) { 2492 psf->sf_crcount = 0; 2493 /* 2494 * add or update "delete" records if an active filter 2495 * is now inactive 2496 */ 2497 2498 for_each_psf_tomb(pmc, dpsf) 2499 if (ipv6_addr_equal(&dpsf->sf_addr, 2500 &psf->sf_addr)) 2501 break; 2502 if (!dpsf) { 2503 dpsf = kmalloc(sizeof(*dpsf), GFP_KERNEL); 2504 if (!dpsf) 2505 continue; 2506 *dpsf = *psf; 2507 rcu_assign_pointer(dpsf->sf_next, 2508 mc_dereference(pmc->mca_tomb, pmc->idev)); 2509 rcu_assign_pointer(pmc->mca_tomb, dpsf); 2510 } 2511 dpsf->sf_crcount = qrv; 2512 rv++; 2513 } 2514 } 2515 return rv; 2516 } 2517 2518 /* Add multicast source filter list to the interface list */ 2519 static int ip6_mc_add_src(struct inet6_dev *idev, const struct in6_addr *pmca, 2520 int sfmode, int sfcount, const struct in6_addr *psfsrc, 2521 int delta) 2522 { 2523 struct ifmcaddr6 *pmc; 2524 int isexclude; 2525 int i, err; 2526 2527 if (!idev) 2528 return -ENODEV; 2529 2530 mc_assert_locked(idev); 2531 2532 for_each_mc_mclock(idev, pmc) { 2533 if (ipv6_addr_equal(pmca, &pmc->mca_addr)) 2534 break; 2535 } 2536 if (!pmc) 2537 return -ESRCH; 2538 2539 sf_markstate(pmc); 2540 isexclude = pmc->mca_sfmode == MCAST_EXCLUDE; 2541 if (!delta) 2542 WRITE_ONCE(pmc->mca_sfcount[sfmode], 2543 pmc->mca_sfcount[sfmode] + 1); 2544 err = 0; 2545 for (i = 0; i < sfcount; i++) { 2546 err = ip6_mc_add1_src(pmc, sfmode, &psfsrc[i]); 2547 if (err) 2548 break; 2549 } 2550 if (err) { 2551 int j; 2552 2553 if (!delta) 2554 WRITE_ONCE(pmc->mca_sfcount[sfmode], 2555 pmc->mca_sfcount[sfmode] - 1); 2556 for (j = 0; j < i; j++) 2557 ip6_mc_del1_src(pmc, sfmode, &psfsrc[j]); 2558 } else if (isexclude != (pmc->mca_sfcount[MCAST_EXCLUDE] != 0)) { 2559 struct ip6_sf_list *psf; 2560 2561 /* filter mode change */ 2562 if (pmc->mca_sfcount[MCAST_EXCLUDE]) 2563 pmc->mca_sfmode = MCAST_EXCLUDE; 2564 else if (pmc->mca_sfcount[MCAST_INCLUDE]) 2565 pmc->mca_sfmode = MCAST_INCLUDE; 2566 /* else no filters; keep old mode for reports */ 2567 2568 pmc->mca_crcount = idev->mc_qrv; 2569 idev->mc_ifc_count = pmc->mca_crcount; 2570 for_each_psf_mclock(pmc, psf) 2571 psf->sf_crcount = 0; 2572 mld_ifc_event(idev); 2573 } else if (sf_setstate(pmc)) { 2574 mld_ifc_event(idev); 2575 } 2576 return err; 2577 } 2578 2579 static void ip6_mc_clear_src(struct ifmcaddr6 *pmc) 2580 { 2581 struct ip6_sf_list *psf, *nextpsf; 2582 2583 mc_assert_locked(pmc->idev); 2584 2585 for (psf = mc_dereference(pmc->mca_tomb, pmc->idev); 2586 psf; 2587 psf = nextpsf) { 2588 nextpsf = mc_dereference(psf->sf_next, pmc->idev); 2589 kfree_rcu(psf, rcu); 2590 } 2591 RCU_INIT_POINTER(pmc->mca_tomb, NULL); 2592 for (psf = mc_dereference(pmc->mca_sources, pmc->idev); 2593 psf; 2594 psf = nextpsf) { 2595 nextpsf = mc_dereference(psf->sf_next, pmc->idev); 2596 kfree_rcu(psf, rcu); 2597 } 2598 RCU_INIT_POINTER(pmc->mca_sources, NULL); 2599 pmc->mca_sfmode = MCAST_EXCLUDE; 2600 pmc->mca_sfcount[MCAST_INCLUDE] = 0; 2601 /* Paired with the READ_ONCE() from ipv6_chk_mcast_addr() */ 2602 WRITE_ONCE(pmc->mca_sfcount[MCAST_EXCLUDE], 1); 2603 } 2604 2605 static void igmp6_join_group(struct ifmcaddr6 *ma) 2606 { 2607 unsigned long delay; 2608 2609 mc_assert_locked(ma->idev); 2610 2611 if (ma->mca_flags & MAF_NOREPORT) 2612 return; 2613 2614 igmp6_send(&ma->mca_addr, ma->idev->dev, ICMPV6_MGM_REPORT); 2615 2616 delay = get_random_u32_below(unsolicited_report_interval(ma->idev)); 2617 2618 if (cancel_delayed_work(&ma->mca_work)) { 2619 refcount_dec(&ma->mca_refcnt); 2620 delay = ma->mca_work.timer.expires - jiffies; 2621 } 2622 2623 if (!mod_delayed_work(mld_wq, &ma->mca_work, delay)) 2624 refcount_inc(&ma->mca_refcnt); 2625 ma->mca_flags |= MAF_TIMER_RUNNING | MAF_LAST_REPORTER; 2626 } 2627 2628 static int ip6_mc_leave_src(struct sock *sk, struct ipv6_mc_socklist *iml, 2629 struct inet6_dev *idev) 2630 { 2631 struct ip6_sf_socklist *psl; 2632 int err; 2633 2634 psl = sock_dereference(iml->sflist, sk); 2635 2636 if (idev) 2637 mutex_lock(&idev->mc_lock); 2638 2639 if (!psl) { 2640 /* any-source empty exclude case */ 2641 err = ip6_mc_del_src(idev, &iml->addr, iml->sfmode, 0, NULL, 0); 2642 } else { 2643 err = ip6_mc_del_src(idev, &iml->addr, iml->sfmode, 2644 psl->sl_count, psl->sl_addr, 0); 2645 RCU_INIT_POINTER(iml->sflist, NULL); 2646 atomic_sub(struct_size(psl, sl_addr, psl->sl_max), 2647 &sk->sk_omem_alloc); 2648 kfree_rcu(psl, rcu); 2649 } 2650 2651 if (idev) 2652 mutex_unlock(&idev->mc_lock); 2653 2654 return err; 2655 } 2656 2657 static void igmp6_leave_group(struct ifmcaddr6 *ma) 2658 { 2659 mc_assert_locked(ma->idev); 2660 2661 if (mld_in_v1_mode(ma->idev)) { 2662 if (ma->mca_flags & MAF_LAST_REPORTER) { 2663 igmp6_send(&ma->mca_addr, ma->idev->dev, 2664 ICMPV6_MGM_REDUCTION); 2665 } 2666 } else { 2667 mld_add_delrec(ma->idev, ma); 2668 mld_ifc_event(ma->idev); 2669 } 2670 } 2671 2672 static void mld_gq_work(struct work_struct *work) 2673 { 2674 struct inet6_dev *idev = container_of(to_delayed_work(work), 2675 struct inet6_dev, 2676 mc_gq_work); 2677 2678 mutex_lock(&idev->mc_lock); 2679 mld_send_report(idev, NULL); 2680 idev->mc_gq_running = 0; 2681 mutex_unlock(&idev->mc_lock); 2682 2683 in6_dev_put(idev); 2684 } 2685 2686 static void mld_ifc_work(struct work_struct *work) 2687 { 2688 struct inet6_dev *idev = container_of(to_delayed_work(work), 2689 struct inet6_dev, 2690 mc_ifc_work); 2691 2692 mutex_lock(&idev->mc_lock); 2693 mld_send_cr(idev); 2694 2695 if (idev->mc_ifc_count) { 2696 idev->mc_ifc_count--; 2697 if (idev->mc_ifc_count) 2698 mld_ifc_start_work(idev, 2699 unsolicited_report_interval(idev)); 2700 } 2701 mutex_unlock(&idev->mc_lock); 2702 in6_dev_put(idev); 2703 } 2704 2705 static void mld_ifc_event(struct inet6_dev *idev) 2706 { 2707 mc_assert_locked(idev); 2708 2709 if (mld_in_v1_mode(idev)) 2710 return; 2711 2712 idev->mc_ifc_count = idev->mc_qrv; 2713 mld_ifc_start_work(idev, 1); 2714 } 2715 2716 static void mld_mca_work(struct work_struct *work) 2717 { 2718 struct ifmcaddr6 *ma = container_of(to_delayed_work(work), 2719 struct ifmcaddr6, mca_work); 2720 2721 mutex_lock(&ma->idev->mc_lock); 2722 if (mld_in_v1_mode(ma->idev)) 2723 igmp6_send(&ma->mca_addr, ma->idev->dev, ICMPV6_MGM_REPORT); 2724 else 2725 mld_send_report(ma->idev, ma); 2726 ma->mca_flags |= MAF_LAST_REPORTER; 2727 ma->mca_flags &= ~MAF_TIMER_RUNNING; 2728 mutex_unlock(&ma->idev->mc_lock); 2729 2730 ma_put(ma); 2731 } 2732 2733 /* Device changing type */ 2734 2735 void ipv6_mc_unmap(struct inet6_dev *idev) 2736 { 2737 struct ifmcaddr6 *i; 2738 2739 /* Install multicast list, except for all-nodes (already installed) */ 2740 2741 mutex_lock(&idev->mc_lock); 2742 for_each_mc_mclock(idev, i) 2743 igmp6_group_dropped(i); 2744 mutex_unlock(&idev->mc_lock); 2745 } 2746 2747 void ipv6_mc_remap(struct inet6_dev *idev) 2748 { 2749 ipv6_mc_up(idev); 2750 } 2751 2752 /* Device going down */ 2753 void ipv6_mc_down(struct inet6_dev *idev) 2754 { 2755 struct ifmcaddr6 *i; 2756 2757 mutex_lock(&idev->mc_lock); 2758 /* Withdraw multicast list */ 2759 for_each_mc_mclock(idev, i) 2760 igmp6_group_dropped(i); 2761 mutex_unlock(&idev->mc_lock); 2762 2763 /* Should stop work after group drop. or we will 2764 * start work again in mld_ifc_event() 2765 */ 2766 mld_query_stop_work(idev); 2767 mld_report_stop_work(idev); 2768 2769 mutex_lock(&idev->mc_lock); 2770 mld_ifc_stop_work(idev); 2771 mld_gq_stop_work(idev); 2772 mutex_unlock(&idev->mc_lock); 2773 2774 mld_dad_stop_work(idev); 2775 } 2776 2777 static void ipv6_mc_reset(struct inet6_dev *idev) 2778 { 2779 idev->mc_qrv = sysctl_mld_qrv; 2780 idev->mc_qi = MLD_QI_DEFAULT; 2781 idev->mc_qri = MLD_QRI_DEFAULT; 2782 idev->mc_v1_seen = 0; 2783 idev->mc_maxdelay = unsolicited_report_interval(idev); 2784 } 2785 2786 /* Device going up */ 2787 2788 void ipv6_mc_up(struct inet6_dev *idev) 2789 { 2790 struct ifmcaddr6 *i; 2791 2792 /* Install multicast list, except for all-nodes (already installed) */ 2793 2794 ipv6_mc_reset(idev); 2795 mutex_lock(&idev->mc_lock); 2796 for_each_mc_mclock(idev, i) { 2797 mld_del_delrec(idev, i); 2798 igmp6_group_added(i); 2799 } 2800 mutex_unlock(&idev->mc_lock); 2801 } 2802 2803 /* IPv6 device initialization. */ 2804 2805 void ipv6_mc_init_dev(struct inet6_dev *idev) 2806 { 2807 idev->mc_gq_running = 0; 2808 INIT_DELAYED_WORK(&idev->mc_gq_work, mld_gq_work); 2809 RCU_INIT_POINTER(idev->mc_tomb, NULL); 2810 idev->mc_ifc_count = 0; 2811 INIT_DELAYED_WORK(&idev->mc_ifc_work, mld_ifc_work); 2812 INIT_DELAYED_WORK(&idev->mc_dad_work, mld_dad_work); 2813 INIT_DELAYED_WORK(&idev->mc_query_work, mld_query_work); 2814 INIT_DELAYED_WORK(&idev->mc_report_work, mld_report_work); 2815 skb_queue_head_init(&idev->mc_query_queue); 2816 skb_queue_head_init(&idev->mc_report_queue); 2817 spin_lock_init(&idev->mc_query_lock); 2818 spin_lock_init(&idev->mc_report_lock); 2819 mutex_init(&idev->mc_lock); 2820 ipv6_mc_reset(idev); 2821 } 2822 2823 /* 2824 * Device is about to be destroyed: clean up. 2825 */ 2826 2827 void ipv6_mc_destroy_dev(struct inet6_dev *idev) 2828 { 2829 struct ifmcaddr6 *i; 2830 2831 /* Deactivate works */ 2832 ipv6_mc_down(idev); 2833 mutex_lock(&idev->mc_lock); 2834 mld_clear_delrec(idev); 2835 mutex_unlock(&idev->mc_lock); 2836 mld_clear_query(idev); 2837 mld_clear_report(idev); 2838 2839 /* Delete all-nodes address. */ 2840 /* We cannot call ipv6_dev_mc_dec() directly, our caller in 2841 * addrconf.c has NULL'd out dev->ip6_ptr so in6_dev_get() will 2842 * fail. 2843 */ 2844 __ipv6_dev_mc_dec(idev, &in6addr_linklocal_allnodes); 2845 2846 if (idev->cnf.forwarding) 2847 __ipv6_dev_mc_dec(idev, &in6addr_linklocal_allrouters); 2848 2849 mutex_lock(&idev->mc_lock); 2850 while ((i = mc_dereference(idev->mc_list, idev))) { 2851 rcu_assign_pointer(idev->mc_list, mc_dereference(i->next, idev)); 2852 2853 ip6_mc_clear_src(i); 2854 ma_put(i); 2855 } 2856 mutex_unlock(&idev->mc_lock); 2857 } 2858 2859 static void ipv6_mc_rejoin_groups(struct inet6_dev *idev) 2860 { 2861 struct ifmcaddr6 *pmc; 2862 2863 mutex_lock(&idev->mc_lock); 2864 if (mld_in_v1_mode(idev)) { 2865 for_each_mc_mclock(idev, pmc) 2866 igmp6_join_group(pmc); 2867 } else { 2868 mld_send_report(idev, NULL); 2869 } 2870 mutex_unlock(&idev->mc_lock); 2871 } 2872 2873 static int ipv6_mc_netdev_event(struct notifier_block *this, 2874 unsigned long event, 2875 void *ptr) 2876 { 2877 struct net_device *dev = netdev_notifier_info_to_dev(ptr); 2878 struct inet6_dev *idev = __in6_dev_get(dev); 2879 2880 switch (event) { 2881 case NETDEV_RESEND_IGMP: 2882 if (idev) 2883 ipv6_mc_rejoin_groups(idev); 2884 break; 2885 default: 2886 break; 2887 } 2888 2889 return NOTIFY_DONE; 2890 } 2891 2892 static struct notifier_block igmp6_netdev_notifier = { 2893 .notifier_call = ipv6_mc_netdev_event, 2894 }; 2895 2896 #ifdef CONFIG_PROC_FS 2897 struct igmp6_mc_iter_state { 2898 struct seq_net_private p; 2899 struct net_device *dev; 2900 struct inet6_dev *idev; 2901 }; 2902 2903 #define igmp6_mc_seq_private(seq) ((struct igmp6_mc_iter_state *)(seq)->private) 2904 2905 static inline struct ifmcaddr6 *igmp6_mc_get_first(struct seq_file *seq) 2906 { 2907 struct ifmcaddr6 *im = NULL; 2908 struct igmp6_mc_iter_state *state = igmp6_mc_seq_private(seq); 2909 struct net *net = seq_file_net(seq); 2910 2911 state->idev = NULL; 2912 for_each_netdev_rcu(net, state->dev) { 2913 struct inet6_dev *idev; 2914 idev = __in6_dev_get(state->dev); 2915 if (!idev) 2916 continue; 2917 2918 im = rcu_dereference(idev->mc_list); 2919 if (im) { 2920 state->idev = idev; 2921 break; 2922 } 2923 } 2924 return im; 2925 } 2926 2927 static struct ifmcaddr6 *igmp6_mc_get_next(struct seq_file *seq, struct ifmcaddr6 *im) 2928 { 2929 struct igmp6_mc_iter_state *state = igmp6_mc_seq_private(seq); 2930 2931 im = rcu_dereference(im->next); 2932 while (!im) { 2933 state->dev = next_net_device_rcu(state->dev); 2934 if (!state->dev) { 2935 state->idev = NULL; 2936 break; 2937 } 2938 state->idev = __in6_dev_get(state->dev); 2939 if (!state->idev) 2940 continue; 2941 im = rcu_dereference(state->idev->mc_list); 2942 } 2943 return im; 2944 } 2945 2946 static struct ifmcaddr6 *igmp6_mc_get_idx(struct seq_file *seq, loff_t pos) 2947 { 2948 struct ifmcaddr6 *im = igmp6_mc_get_first(seq); 2949 if (im) 2950 while (pos && (im = igmp6_mc_get_next(seq, im)) != NULL) 2951 --pos; 2952 return pos ? NULL : im; 2953 } 2954 2955 static void *igmp6_mc_seq_start(struct seq_file *seq, loff_t *pos) 2956 __acquires(RCU) 2957 { 2958 rcu_read_lock(); 2959 return igmp6_mc_get_idx(seq, *pos); 2960 } 2961 2962 static void *igmp6_mc_seq_next(struct seq_file *seq, void *v, loff_t *pos) 2963 { 2964 struct ifmcaddr6 *im = igmp6_mc_get_next(seq, v); 2965 2966 ++*pos; 2967 return im; 2968 } 2969 2970 static void igmp6_mc_seq_stop(struct seq_file *seq, void *v) 2971 __releases(RCU) 2972 { 2973 struct igmp6_mc_iter_state *state = igmp6_mc_seq_private(seq); 2974 2975 if (likely(state->idev)) 2976 state->idev = NULL; 2977 state->dev = NULL; 2978 rcu_read_unlock(); 2979 } 2980 2981 static int igmp6_mc_seq_show(struct seq_file *seq, void *v) 2982 { 2983 struct ifmcaddr6 *im = (struct ifmcaddr6 *)v; 2984 struct igmp6_mc_iter_state *state = igmp6_mc_seq_private(seq); 2985 2986 seq_printf(seq, 2987 "%-4d %-15s %pi6 %5d %08X %ld\n", 2988 state->dev->ifindex, state->dev->name, 2989 &im->mca_addr, 2990 im->mca_users, im->mca_flags, 2991 (im->mca_flags & MAF_TIMER_RUNNING) ? 2992 jiffies_to_clock_t(im->mca_work.timer.expires - jiffies) : 0); 2993 return 0; 2994 } 2995 2996 static const struct seq_operations igmp6_mc_seq_ops = { 2997 .start = igmp6_mc_seq_start, 2998 .next = igmp6_mc_seq_next, 2999 .stop = igmp6_mc_seq_stop, 3000 .show = igmp6_mc_seq_show, 3001 }; 3002 3003 struct igmp6_mcf_iter_state { 3004 struct seq_net_private p; 3005 struct net_device *dev; 3006 struct inet6_dev *idev; 3007 struct ifmcaddr6 *im; 3008 }; 3009 3010 #define igmp6_mcf_seq_private(seq) ((struct igmp6_mcf_iter_state *)(seq)->private) 3011 3012 static inline struct ip6_sf_list *igmp6_mcf_get_first(struct seq_file *seq) 3013 { 3014 struct ip6_sf_list *psf = NULL; 3015 struct ifmcaddr6 *im = NULL; 3016 struct igmp6_mcf_iter_state *state = igmp6_mcf_seq_private(seq); 3017 struct net *net = seq_file_net(seq); 3018 3019 state->idev = NULL; 3020 state->im = NULL; 3021 for_each_netdev_rcu(net, state->dev) { 3022 struct inet6_dev *idev; 3023 idev = __in6_dev_get(state->dev); 3024 if (unlikely(idev == NULL)) 3025 continue; 3026 3027 im = rcu_dereference(idev->mc_list); 3028 if (likely(im)) { 3029 psf = rcu_dereference(im->mca_sources); 3030 if (likely(psf)) { 3031 state->im = im; 3032 state->idev = idev; 3033 break; 3034 } 3035 } 3036 } 3037 return psf; 3038 } 3039 3040 static struct ip6_sf_list *igmp6_mcf_get_next(struct seq_file *seq, struct ip6_sf_list *psf) 3041 { 3042 struct igmp6_mcf_iter_state *state = igmp6_mcf_seq_private(seq); 3043 3044 psf = rcu_dereference(psf->sf_next); 3045 while (!psf) { 3046 state->im = rcu_dereference(state->im->next); 3047 while (!state->im) { 3048 state->dev = next_net_device_rcu(state->dev); 3049 if (!state->dev) { 3050 state->idev = NULL; 3051 goto out; 3052 } 3053 state->idev = __in6_dev_get(state->dev); 3054 if (!state->idev) 3055 continue; 3056 state->im = rcu_dereference(state->idev->mc_list); 3057 } 3058 psf = rcu_dereference(state->im->mca_sources); 3059 } 3060 out: 3061 return psf; 3062 } 3063 3064 static struct ip6_sf_list *igmp6_mcf_get_idx(struct seq_file *seq, loff_t pos) 3065 { 3066 struct ip6_sf_list *psf = igmp6_mcf_get_first(seq); 3067 if (psf) 3068 while (pos && (psf = igmp6_mcf_get_next(seq, psf)) != NULL) 3069 --pos; 3070 return pos ? NULL : psf; 3071 } 3072 3073 static void *igmp6_mcf_seq_start(struct seq_file *seq, loff_t *pos) 3074 __acquires(RCU) 3075 { 3076 rcu_read_lock(); 3077 return *pos ? igmp6_mcf_get_idx(seq, *pos - 1) : SEQ_START_TOKEN; 3078 } 3079 3080 static void *igmp6_mcf_seq_next(struct seq_file *seq, void *v, loff_t *pos) 3081 { 3082 struct ip6_sf_list *psf; 3083 if (v == SEQ_START_TOKEN) 3084 psf = igmp6_mcf_get_first(seq); 3085 else 3086 psf = igmp6_mcf_get_next(seq, v); 3087 ++*pos; 3088 return psf; 3089 } 3090 3091 static void igmp6_mcf_seq_stop(struct seq_file *seq, void *v) 3092 __releases(RCU) 3093 { 3094 struct igmp6_mcf_iter_state *state = igmp6_mcf_seq_private(seq); 3095 3096 if (likely(state->im)) 3097 state->im = NULL; 3098 if (likely(state->idev)) 3099 state->idev = NULL; 3100 3101 state->dev = NULL; 3102 rcu_read_unlock(); 3103 } 3104 3105 static int igmp6_mcf_seq_show(struct seq_file *seq, void *v) 3106 { 3107 struct ip6_sf_list *psf = (struct ip6_sf_list *)v; 3108 struct igmp6_mcf_iter_state *state = igmp6_mcf_seq_private(seq); 3109 3110 if (v == SEQ_START_TOKEN) { 3111 seq_puts(seq, "Idx Device Multicast Address Source Address INC EXC\n"); 3112 } else { 3113 seq_printf(seq, 3114 "%3d %6.6s %pi6 %pi6 %6lu %6lu\n", 3115 state->dev->ifindex, state->dev->name, 3116 &state->im->mca_addr, 3117 &psf->sf_addr, 3118 READ_ONCE(psf->sf_count[MCAST_INCLUDE]), 3119 READ_ONCE(psf->sf_count[MCAST_EXCLUDE])); 3120 } 3121 return 0; 3122 } 3123 3124 static const struct seq_operations igmp6_mcf_seq_ops = { 3125 .start = igmp6_mcf_seq_start, 3126 .next = igmp6_mcf_seq_next, 3127 .stop = igmp6_mcf_seq_stop, 3128 .show = igmp6_mcf_seq_show, 3129 }; 3130 3131 static int __net_init igmp6_proc_init(struct net *net) 3132 { 3133 int err; 3134 3135 err = -ENOMEM; 3136 if (!proc_create_net("igmp6", 0444, net->proc_net, &igmp6_mc_seq_ops, 3137 sizeof(struct igmp6_mc_iter_state))) 3138 goto out; 3139 if (!proc_create_net("mcfilter6", 0444, net->proc_net, 3140 &igmp6_mcf_seq_ops, 3141 sizeof(struct igmp6_mcf_iter_state))) 3142 goto out_proc_net_igmp6; 3143 3144 err = 0; 3145 out: 3146 return err; 3147 3148 out_proc_net_igmp6: 3149 remove_proc_entry("igmp6", net->proc_net); 3150 goto out; 3151 } 3152 3153 static void __net_exit igmp6_proc_exit(struct net *net) 3154 { 3155 remove_proc_entry("mcfilter6", net->proc_net); 3156 remove_proc_entry("igmp6", net->proc_net); 3157 } 3158 #else 3159 static inline int igmp6_proc_init(struct net *net) 3160 { 3161 return 0; 3162 } 3163 static inline void igmp6_proc_exit(struct net *net) 3164 { 3165 } 3166 #endif 3167 3168 static int __net_init igmp6_net_init(struct net *net) 3169 { 3170 int err; 3171 3172 err = inet_ctl_sock_create(&net->ipv6.igmp_sk, PF_INET6, 3173 SOCK_RAW, IPPROTO_ICMPV6, net); 3174 if (err < 0) { 3175 pr_err("Failed to initialize the IGMP6 control socket (err %d)\n", 3176 err); 3177 goto out; 3178 } 3179 3180 inet6_sk(net->ipv6.igmp_sk)->hop_limit = 1; 3181 net->ipv6.igmp_sk->sk_allocation = GFP_KERNEL; 3182 3183 err = inet_ctl_sock_create(&net->ipv6.mc_autojoin_sk, PF_INET6, 3184 SOCK_RAW, IPPROTO_ICMPV6, net); 3185 if (err < 0) { 3186 pr_err("Failed to initialize the IGMP6 autojoin socket (err %d)\n", 3187 err); 3188 goto out_sock_create; 3189 } 3190 3191 err = igmp6_proc_init(net); 3192 if (err) 3193 goto out_sock_create_autojoin; 3194 3195 return 0; 3196 3197 out_sock_create_autojoin: 3198 inet_ctl_sock_destroy(net->ipv6.mc_autojoin_sk); 3199 out_sock_create: 3200 inet_ctl_sock_destroy(net->ipv6.igmp_sk); 3201 out: 3202 return err; 3203 } 3204 3205 static void __net_exit igmp6_net_exit(struct net *net) 3206 { 3207 inet_ctl_sock_destroy(net->ipv6.igmp_sk); 3208 inet_ctl_sock_destroy(net->ipv6.mc_autojoin_sk); 3209 igmp6_proc_exit(net); 3210 } 3211 3212 static struct pernet_operations igmp6_net_ops = { 3213 .init = igmp6_net_init, 3214 .exit = igmp6_net_exit, 3215 }; 3216 3217 int __init igmp6_init(void) 3218 { 3219 int err; 3220 3221 err = register_pernet_subsys(&igmp6_net_ops); 3222 if (err) 3223 return err; 3224 3225 mld_wq = create_workqueue("mld"); 3226 if (!mld_wq) { 3227 unregister_pernet_subsys(&igmp6_net_ops); 3228 return -ENOMEM; 3229 } 3230 3231 return err; 3232 } 3233 3234 int __init igmp6_late_init(void) 3235 { 3236 return register_netdevice_notifier(&igmp6_netdev_notifier); 3237 } 3238 3239 void igmp6_cleanup(void) 3240 { 3241 unregister_pernet_subsys(&igmp6_net_ops); 3242 destroy_workqueue(mld_wq); 3243 } 3244 3245 void igmp6_late_cleanup(void) 3246 { 3247 unregister_netdevice_notifier(&igmp6_netdev_notifier); 3248 } 3249