1 /* 2 * INET 802.1Q VLAN 3 * Ethernet-type device handling. 4 * 5 * Authors: Ben Greear <greearb@candelatech.com> 6 * Please send support related email to: netdev@vger.kernel.org 7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html 8 * 9 * Fixes: 10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>; 11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>; 12 * Correct all the locking - David S. Miller <davem@redhat.com>; 13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com> 14 * 15 * This program is free software; you can redistribute it and/or 16 * modify it under the terms of the GNU General Public License 17 * as published by the Free Software Foundation; either version 18 * 2 of the License, or (at your option) any later version. 19 */ 20 21 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 22 23 #include <linux/capability.h> 24 #include <linux/module.h> 25 #include <linux/netdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/slab.h> 28 #include <linux/init.h> 29 #include <linux/rculist.h> 30 #include <net/p8022.h> 31 #include <net/arp.h> 32 #include <linux/rtnetlink.h> 33 #include <linux/notifier.h> 34 #include <net/rtnetlink.h> 35 #include <net/net_namespace.h> 36 #include <net/netns/generic.h> 37 #include <asm/uaccess.h> 38 39 #include <linux/if_vlan.h> 40 #include "vlan.h" 41 #include "vlanproc.h" 42 43 #define DRV_VERSION "1.8" 44 45 /* Global VLAN variables */ 46 47 int vlan_net_id __read_mostly; 48 49 const char vlan_fullname[] = "802.1Q VLAN Support"; 50 const char vlan_version[] = DRV_VERSION; 51 52 /* End of global variables definitions. */ 53 54 static int vlan_group_prealloc_vid(struct vlan_group *vg, 55 __be16 vlan_proto, u16 vlan_id) 56 { 57 struct net_device **array; 58 unsigned int pidx, vidx; 59 unsigned int size; 60 61 ASSERT_RTNL(); 62 63 pidx = vlan_proto_idx(vlan_proto); 64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN; 65 array = vg->vlan_devices_arrays[pidx][vidx]; 66 if (array != NULL) 67 return 0; 68 69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN; 70 array = kzalloc(size, GFP_KERNEL); 71 if (array == NULL) 72 return -ENOBUFS; 73 74 vg->vlan_devices_arrays[pidx][vidx] = array; 75 return 0; 76 } 77 78 void unregister_vlan_dev(struct net_device *dev, struct list_head *head) 79 { 80 struct vlan_dev_priv *vlan = vlan_dev_priv(dev); 81 struct net_device *real_dev = vlan->real_dev; 82 struct vlan_info *vlan_info; 83 struct vlan_group *grp; 84 u16 vlan_id = vlan->vlan_id; 85 86 ASSERT_RTNL(); 87 88 vlan_info = rtnl_dereference(real_dev->vlan_info); 89 BUG_ON(!vlan_info); 90 91 grp = &vlan_info->grp; 92 93 grp->nr_vlan_devs--; 94 95 if (vlan->flags & VLAN_FLAG_MVRP) 96 vlan_mvrp_request_leave(dev); 97 if (vlan->flags & VLAN_FLAG_GVRP) 98 vlan_gvrp_request_leave(dev); 99 100 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL); 101 102 netdev_upper_dev_unlink(real_dev, dev); 103 /* Because unregister_netdevice_queue() makes sure at least one rcu 104 * grace period is respected before device freeing, 105 * we dont need to call synchronize_net() here. 106 */ 107 unregister_netdevice_queue(dev, head); 108 109 if (grp->nr_vlan_devs == 0) { 110 vlan_mvrp_uninit_applicant(real_dev); 111 vlan_gvrp_uninit_applicant(real_dev); 112 } 113 114 /* Take it out of our own structures, but be sure to interlock with 115 * HW accelerating devices or SW vlan input packet processing if 116 * VLAN is not 0 (leave it there for 802.1p). 117 */ 118 if (vlan_id) 119 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id); 120 121 /* Get rid of the vlan's reference to real_dev */ 122 dev_put(real_dev); 123 } 124 125 int vlan_check_real_dev(struct net_device *real_dev, 126 __be16 protocol, u16 vlan_id) 127 { 128 const char *name = real_dev->name; 129 130 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) { 131 pr_info("VLANs not supported on %s\n", name); 132 return -EOPNOTSUPP; 133 } 134 135 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL) 136 return -EEXIST; 137 138 return 0; 139 } 140 141 int register_vlan_dev(struct net_device *dev) 142 { 143 struct vlan_dev_priv *vlan = vlan_dev_priv(dev); 144 struct net_device *real_dev = vlan->real_dev; 145 u16 vlan_id = vlan->vlan_id; 146 struct vlan_info *vlan_info; 147 struct vlan_group *grp; 148 int err; 149 150 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id); 151 if (err) 152 return err; 153 154 vlan_info = rtnl_dereference(real_dev->vlan_info); 155 /* vlan_info should be there now. vlan_vid_add took care of it */ 156 BUG_ON(!vlan_info); 157 158 grp = &vlan_info->grp; 159 if (grp->nr_vlan_devs == 0) { 160 err = vlan_gvrp_init_applicant(real_dev); 161 if (err < 0) 162 goto out_vid_del; 163 err = vlan_mvrp_init_applicant(real_dev); 164 if (err < 0) 165 goto out_uninit_gvrp; 166 } 167 168 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id); 169 if (err < 0) 170 goto out_uninit_mvrp; 171 172 vlan->nest_level = dev_get_nest_level(real_dev, is_vlan_dev) + 1; 173 err = register_netdevice(dev); 174 if (err < 0) 175 goto out_uninit_mvrp; 176 177 err = netdev_upper_dev_link(real_dev, dev); 178 if (err) 179 goto out_unregister_netdev; 180 181 /* Account for reference in struct vlan_dev_priv */ 182 dev_hold(real_dev); 183 184 netif_stacked_transfer_operstate(real_dev, dev); 185 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */ 186 187 /* So, got the sucker initialized, now lets place 188 * it into our local structure. 189 */ 190 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev); 191 grp->nr_vlan_devs++; 192 193 return 0; 194 195 out_unregister_netdev: 196 unregister_netdevice(dev); 197 out_uninit_mvrp: 198 if (grp->nr_vlan_devs == 0) 199 vlan_mvrp_uninit_applicant(real_dev); 200 out_uninit_gvrp: 201 if (grp->nr_vlan_devs == 0) 202 vlan_gvrp_uninit_applicant(real_dev); 203 out_vid_del: 204 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id); 205 return err; 206 } 207 208 /* Attach a VLAN device to a mac address (ie Ethernet Card). 209 * Returns 0 if the device was created or a negative error code otherwise. 210 */ 211 static int register_vlan_device(struct net_device *real_dev, u16 vlan_id) 212 { 213 struct net_device *new_dev; 214 struct vlan_dev_priv *vlan; 215 struct net *net = dev_net(real_dev); 216 struct vlan_net *vn = net_generic(net, vlan_net_id); 217 char name[IFNAMSIZ]; 218 int err; 219 220 if (vlan_id >= VLAN_VID_MASK) 221 return -ERANGE; 222 223 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id); 224 if (err < 0) 225 return err; 226 227 /* Gotta set up the fields for the device. */ 228 switch (vn->name_type) { 229 case VLAN_NAME_TYPE_RAW_PLUS_VID: 230 /* name will look like: eth1.0005 */ 231 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id); 232 break; 233 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD: 234 /* Put our vlan.VID in the name. 235 * Name will look like: vlan5 236 */ 237 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id); 238 break; 239 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD: 240 /* Put our vlan.VID in the name. 241 * Name will look like: eth0.5 242 */ 243 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id); 244 break; 245 case VLAN_NAME_TYPE_PLUS_VID: 246 /* Put our vlan.VID in the name. 247 * Name will look like: vlan0005 248 */ 249 default: 250 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id); 251 } 252 253 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name, 254 NET_NAME_UNKNOWN, vlan_setup); 255 256 if (new_dev == NULL) 257 return -ENOBUFS; 258 259 dev_net_set(new_dev, net); 260 /* need 4 bytes for extra VLAN header info, 261 * hope the underlying device can handle it. 262 */ 263 new_dev->mtu = real_dev->mtu; 264 265 vlan = vlan_dev_priv(new_dev); 266 vlan->vlan_proto = htons(ETH_P_8021Q); 267 vlan->vlan_id = vlan_id; 268 vlan->real_dev = real_dev; 269 vlan->dent = NULL; 270 vlan->flags = VLAN_FLAG_REORDER_HDR; 271 272 new_dev->rtnl_link_ops = &vlan_link_ops; 273 err = register_vlan_dev(new_dev); 274 if (err < 0) 275 goto out_free_newdev; 276 277 return 0; 278 279 out_free_newdev: 280 free_netdev(new_dev); 281 return err; 282 } 283 284 static void vlan_sync_address(struct net_device *dev, 285 struct net_device *vlandev) 286 { 287 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev); 288 289 /* May be called without an actual change */ 290 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr)) 291 return; 292 293 /* vlan address was different from the old address and is equal to 294 * the new address */ 295 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) && 296 ether_addr_equal(vlandev->dev_addr, dev->dev_addr)) 297 dev_uc_del(dev, vlandev->dev_addr); 298 299 /* vlan address was equal to the old address and is different from 300 * the new address */ 301 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) && 302 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr)) 303 dev_uc_add(dev, vlandev->dev_addr); 304 305 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr); 306 } 307 308 static void vlan_transfer_features(struct net_device *dev, 309 struct net_device *vlandev) 310 { 311 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev); 312 313 vlandev->gso_max_size = dev->gso_max_size; 314 315 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto)) 316 vlandev->hard_header_len = dev->hard_header_len; 317 else 318 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN; 319 320 #if IS_ENABLED(CONFIG_FCOE) 321 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid; 322 #endif 323 324 netdev_update_features(vlandev); 325 } 326 327 static int __vlan_device_event(struct net_device *dev, unsigned long event) 328 { 329 int err = 0; 330 331 switch (event) { 332 case NETDEV_CHANGENAME: 333 vlan_proc_rem_dev(dev); 334 err = vlan_proc_add_dev(dev); 335 break; 336 case NETDEV_REGISTER: 337 err = vlan_proc_add_dev(dev); 338 break; 339 case NETDEV_UNREGISTER: 340 vlan_proc_rem_dev(dev); 341 break; 342 } 343 344 return err; 345 } 346 347 static int vlan_device_event(struct notifier_block *unused, unsigned long event, 348 void *ptr) 349 { 350 struct net_device *dev = netdev_notifier_info_to_dev(ptr); 351 struct vlan_group *grp; 352 struct vlan_info *vlan_info; 353 int i, flgs; 354 struct net_device *vlandev; 355 struct vlan_dev_priv *vlan; 356 bool last = false; 357 LIST_HEAD(list); 358 359 if (is_vlan_dev(dev)) { 360 int err = __vlan_device_event(dev, event); 361 362 if (err) 363 return notifier_from_errno(err); 364 } 365 366 if ((event == NETDEV_UP) && 367 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) { 368 pr_info("adding VLAN 0 to HW filter on device %s\n", 369 dev->name); 370 vlan_vid_add(dev, htons(ETH_P_8021Q), 0); 371 } 372 373 vlan_info = rtnl_dereference(dev->vlan_info); 374 if (!vlan_info) 375 goto out; 376 grp = &vlan_info->grp; 377 378 /* It is OK that we do not hold the group lock right now, 379 * as we run under the RTNL lock. 380 */ 381 382 switch (event) { 383 case NETDEV_CHANGE: 384 /* Propagate real device state to vlan devices */ 385 vlan_group_for_each_dev(grp, i, vlandev) 386 netif_stacked_transfer_operstate(dev, vlandev); 387 break; 388 389 case NETDEV_CHANGEADDR: 390 /* Adjust unicast filters on underlying device */ 391 vlan_group_for_each_dev(grp, i, vlandev) { 392 flgs = vlandev->flags; 393 if (!(flgs & IFF_UP)) 394 continue; 395 396 vlan_sync_address(dev, vlandev); 397 } 398 break; 399 400 case NETDEV_CHANGEMTU: 401 vlan_group_for_each_dev(grp, i, vlandev) { 402 if (vlandev->mtu <= dev->mtu) 403 continue; 404 405 dev_set_mtu(vlandev, dev->mtu); 406 } 407 break; 408 409 case NETDEV_FEAT_CHANGE: 410 /* Propagate device features to underlying device */ 411 vlan_group_for_each_dev(grp, i, vlandev) 412 vlan_transfer_features(dev, vlandev); 413 break; 414 415 case NETDEV_DOWN: { 416 struct net_device *tmp; 417 LIST_HEAD(close_list); 418 419 if (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER) 420 vlan_vid_del(dev, htons(ETH_P_8021Q), 0); 421 422 /* Put all VLANs for this dev in the down state too. */ 423 vlan_group_for_each_dev(grp, i, vlandev) { 424 flgs = vlandev->flags; 425 if (!(flgs & IFF_UP)) 426 continue; 427 428 vlan = vlan_dev_priv(vlandev); 429 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) 430 list_add(&vlandev->close_list, &close_list); 431 } 432 433 dev_close_many(&close_list, false); 434 435 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) { 436 netif_stacked_transfer_operstate(dev, vlandev); 437 list_del_init(&vlandev->close_list); 438 } 439 list_del(&close_list); 440 break; 441 } 442 case NETDEV_UP: 443 /* Put all VLANs for this dev in the up state too. */ 444 vlan_group_for_each_dev(grp, i, vlandev) { 445 flgs = dev_get_flags(vlandev); 446 if (flgs & IFF_UP) 447 continue; 448 449 vlan = vlan_dev_priv(vlandev); 450 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) 451 dev_change_flags(vlandev, flgs | IFF_UP); 452 netif_stacked_transfer_operstate(dev, vlandev); 453 } 454 break; 455 456 case NETDEV_UNREGISTER: 457 /* twiddle thumbs on netns device moves */ 458 if (dev->reg_state != NETREG_UNREGISTERING) 459 break; 460 461 vlan_group_for_each_dev(grp, i, vlandev) { 462 /* removal of last vid destroys vlan_info, abort 463 * afterwards */ 464 if (vlan_info->nr_vids == 1) 465 last = true; 466 467 unregister_vlan_dev(vlandev, &list); 468 if (last) 469 break; 470 } 471 unregister_netdevice_many(&list); 472 break; 473 474 case NETDEV_PRE_TYPE_CHANGE: 475 /* Forbid underlaying device to change its type. */ 476 if (vlan_uses_dev(dev)) 477 return NOTIFY_BAD; 478 break; 479 480 case NETDEV_NOTIFY_PEERS: 481 case NETDEV_BONDING_FAILOVER: 482 case NETDEV_RESEND_IGMP: 483 /* Propagate to vlan devices */ 484 vlan_group_for_each_dev(grp, i, vlandev) 485 call_netdevice_notifiers(event, vlandev); 486 break; 487 } 488 489 out: 490 return NOTIFY_DONE; 491 } 492 493 static struct notifier_block vlan_notifier_block __read_mostly = { 494 .notifier_call = vlan_device_event, 495 }; 496 497 /* 498 * VLAN IOCTL handler. 499 * o execute requested action or pass command to the device driver 500 * arg is really a struct vlan_ioctl_args __user *. 501 */ 502 static int vlan_ioctl_handler(struct net *net, void __user *arg) 503 { 504 int err; 505 struct vlan_ioctl_args args; 506 struct net_device *dev = NULL; 507 508 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args))) 509 return -EFAULT; 510 511 /* Null terminate this sucker, just in case. */ 512 args.device1[23] = 0; 513 args.u.device2[23] = 0; 514 515 rtnl_lock(); 516 517 switch (args.cmd) { 518 case SET_VLAN_INGRESS_PRIORITY_CMD: 519 case SET_VLAN_EGRESS_PRIORITY_CMD: 520 case SET_VLAN_FLAG_CMD: 521 case ADD_VLAN_CMD: 522 case DEL_VLAN_CMD: 523 case GET_VLAN_REALDEV_NAME_CMD: 524 case GET_VLAN_VID_CMD: 525 err = -ENODEV; 526 dev = __dev_get_by_name(net, args.device1); 527 if (!dev) 528 goto out; 529 530 err = -EINVAL; 531 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev)) 532 goto out; 533 } 534 535 switch (args.cmd) { 536 case SET_VLAN_INGRESS_PRIORITY_CMD: 537 err = -EPERM; 538 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 539 break; 540 vlan_dev_set_ingress_priority(dev, 541 args.u.skb_priority, 542 args.vlan_qos); 543 err = 0; 544 break; 545 546 case SET_VLAN_EGRESS_PRIORITY_CMD: 547 err = -EPERM; 548 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 549 break; 550 err = vlan_dev_set_egress_priority(dev, 551 args.u.skb_priority, 552 args.vlan_qos); 553 break; 554 555 case SET_VLAN_FLAG_CMD: 556 err = -EPERM; 557 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 558 break; 559 err = vlan_dev_change_flags(dev, 560 args.vlan_qos ? args.u.flag : 0, 561 args.u.flag); 562 break; 563 564 case SET_VLAN_NAME_TYPE_CMD: 565 err = -EPERM; 566 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 567 break; 568 if ((args.u.name_type >= 0) && 569 (args.u.name_type < VLAN_NAME_TYPE_HIGHEST)) { 570 struct vlan_net *vn; 571 572 vn = net_generic(net, vlan_net_id); 573 vn->name_type = args.u.name_type; 574 err = 0; 575 } else { 576 err = -EINVAL; 577 } 578 break; 579 580 case ADD_VLAN_CMD: 581 err = -EPERM; 582 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 583 break; 584 err = register_vlan_device(dev, args.u.VID); 585 break; 586 587 case DEL_VLAN_CMD: 588 err = -EPERM; 589 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 590 break; 591 unregister_vlan_dev(dev, NULL); 592 err = 0; 593 break; 594 595 case GET_VLAN_REALDEV_NAME_CMD: 596 err = 0; 597 vlan_dev_get_realdev_name(dev, args.u.device2); 598 if (copy_to_user(arg, &args, 599 sizeof(struct vlan_ioctl_args))) 600 err = -EFAULT; 601 break; 602 603 case GET_VLAN_VID_CMD: 604 err = 0; 605 args.u.VID = vlan_dev_vlan_id(dev); 606 if (copy_to_user(arg, &args, 607 sizeof(struct vlan_ioctl_args))) 608 err = -EFAULT; 609 break; 610 611 default: 612 err = -EOPNOTSUPP; 613 break; 614 } 615 out: 616 rtnl_unlock(); 617 return err; 618 } 619 620 static struct sk_buff **vlan_gro_receive(struct sk_buff **head, 621 struct sk_buff *skb) 622 { 623 struct sk_buff *p, **pp = NULL; 624 struct vlan_hdr *vhdr; 625 unsigned int hlen, off_vlan; 626 const struct packet_offload *ptype; 627 __be16 type; 628 int flush = 1; 629 630 off_vlan = skb_gro_offset(skb); 631 hlen = off_vlan + sizeof(*vhdr); 632 vhdr = skb_gro_header_fast(skb, off_vlan); 633 if (skb_gro_header_hard(skb, hlen)) { 634 vhdr = skb_gro_header_slow(skb, hlen, off_vlan); 635 if (unlikely(!vhdr)) 636 goto out; 637 } 638 639 type = vhdr->h_vlan_encapsulated_proto; 640 641 rcu_read_lock(); 642 ptype = gro_find_receive_by_type(type); 643 if (!ptype) 644 goto out_unlock; 645 646 flush = 0; 647 648 for (p = *head; p; p = p->next) { 649 struct vlan_hdr *vhdr2; 650 651 if (!NAPI_GRO_CB(p)->same_flow) 652 continue; 653 654 vhdr2 = (struct vlan_hdr *)(p->data + off_vlan); 655 if (compare_vlan_header(vhdr, vhdr2)) 656 NAPI_GRO_CB(p)->same_flow = 0; 657 } 658 659 skb_gro_pull(skb, sizeof(*vhdr)); 660 skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr)); 661 pp = ptype->callbacks.gro_receive(head, skb); 662 663 out_unlock: 664 rcu_read_unlock(); 665 out: 666 NAPI_GRO_CB(skb)->flush |= flush; 667 668 return pp; 669 } 670 671 static int vlan_gro_complete(struct sk_buff *skb, int nhoff) 672 { 673 struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff); 674 __be16 type = vhdr->h_vlan_encapsulated_proto; 675 struct packet_offload *ptype; 676 int err = -ENOENT; 677 678 rcu_read_lock(); 679 ptype = gro_find_complete_by_type(type); 680 if (ptype) 681 err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr)); 682 683 rcu_read_unlock(); 684 return err; 685 } 686 687 static struct packet_offload vlan_packet_offloads[] __read_mostly = { 688 { 689 .type = cpu_to_be16(ETH_P_8021Q), 690 .priority = 10, 691 .callbacks = { 692 .gro_receive = vlan_gro_receive, 693 .gro_complete = vlan_gro_complete, 694 }, 695 }, 696 { 697 .type = cpu_to_be16(ETH_P_8021AD), 698 .priority = 10, 699 .callbacks = { 700 .gro_receive = vlan_gro_receive, 701 .gro_complete = vlan_gro_complete, 702 }, 703 }, 704 }; 705 706 static int __net_init vlan_init_net(struct net *net) 707 { 708 struct vlan_net *vn = net_generic(net, vlan_net_id); 709 int err; 710 711 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD; 712 713 err = vlan_proc_init(net); 714 715 return err; 716 } 717 718 static void __net_exit vlan_exit_net(struct net *net) 719 { 720 vlan_proc_cleanup(net); 721 } 722 723 static struct pernet_operations vlan_net_ops = { 724 .init = vlan_init_net, 725 .exit = vlan_exit_net, 726 .id = &vlan_net_id, 727 .size = sizeof(struct vlan_net), 728 }; 729 730 static int __init vlan_proto_init(void) 731 { 732 int err; 733 unsigned int i; 734 735 pr_info("%s v%s\n", vlan_fullname, vlan_version); 736 737 err = register_pernet_subsys(&vlan_net_ops); 738 if (err < 0) 739 goto err0; 740 741 err = register_netdevice_notifier(&vlan_notifier_block); 742 if (err < 0) 743 goto err2; 744 745 err = vlan_gvrp_init(); 746 if (err < 0) 747 goto err3; 748 749 err = vlan_mvrp_init(); 750 if (err < 0) 751 goto err4; 752 753 err = vlan_netlink_init(); 754 if (err < 0) 755 goto err5; 756 757 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++) 758 dev_add_offload(&vlan_packet_offloads[i]); 759 760 vlan_ioctl_set(vlan_ioctl_handler); 761 return 0; 762 763 err5: 764 vlan_mvrp_uninit(); 765 err4: 766 vlan_gvrp_uninit(); 767 err3: 768 unregister_netdevice_notifier(&vlan_notifier_block); 769 err2: 770 unregister_pernet_subsys(&vlan_net_ops); 771 err0: 772 return err; 773 } 774 775 static void __exit vlan_cleanup_module(void) 776 { 777 unsigned int i; 778 779 vlan_ioctl_set(NULL); 780 781 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++) 782 dev_remove_offload(&vlan_packet_offloads[i]); 783 784 vlan_netlink_fini(); 785 786 unregister_netdevice_notifier(&vlan_notifier_block); 787 788 unregister_pernet_subsys(&vlan_net_ops); 789 rcu_barrier(); /* Wait for completion of call_rcu()'s */ 790 791 vlan_mvrp_uninit(); 792 vlan_gvrp_uninit(); 793 } 794 795 module_init(vlan_proto_init); 796 module_exit(vlan_cleanup_module); 797 798 MODULE_LICENSE("GPL"); 799 MODULE_VERSION(DRV_VERSION); 800