Loading...
1/*
2 * INET 802.1Q VLAN
3 * Ethernet-type device handling.
4 *
5 * Authors: Ben Greear <greearb@candelatech.com>
6 * Please send support related email to: netdev@vger.kernel.org
7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
8 *
9 * Fixes:
10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
12 * Correct all the locking - David S. Miller <davem@redhat.com>;
13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
14 *
15 * This program is free software; you can redistribute it and/or
16 * modify it under the terms of the GNU General Public License
17 * as published by the Free Software Foundation; either version
18 * 2 of the License, or (at your option) any later version.
19 */
20
21#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
22
23#include <linux/capability.h>
24#include <linux/module.h>
25#include <linux/netdevice.h>
26#include <linux/skbuff.h>
27#include <linux/slab.h>
28#include <linux/init.h>
29#include <linux/rculist.h>
30#include <net/p8022.h>
31#include <net/arp.h>
32#include <linux/rtnetlink.h>
33#include <linux/notifier.h>
34#include <net/rtnetlink.h>
35#include <net/net_namespace.h>
36#include <net/netns/generic.h>
37#include <asm/uaccess.h>
38
39#include <linux/if_vlan.h>
40#include "vlan.h"
41#include "vlanproc.h"
42
43#define DRV_VERSION "1.8"
44
45/* Global VLAN variables */
46
47int vlan_net_id __read_mostly;
48
49const char vlan_fullname[] = "802.1Q VLAN Support";
50const char vlan_version[] = DRV_VERSION;
51
52/* End of global variables definitions. */
53
54static void vlan_group_free(struct vlan_group *grp)
55{
56 int i;
57
58 for (i = 0; i < VLAN_GROUP_ARRAY_SPLIT_PARTS; i++)
59 kfree(grp->vlan_devices_arrays[i]);
60 kfree(grp);
61}
62
63static struct vlan_group *vlan_group_alloc(struct net_device *real_dev)
64{
65 struct vlan_group *grp;
66
67 grp = kzalloc(sizeof(struct vlan_group), GFP_KERNEL);
68 if (!grp)
69 return NULL;
70
71 grp->real_dev = real_dev;
72 return grp;
73}
74
75static int vlan_group_prealloc_vid(struct vlan_group *vg, u16 vlan_id)
76{
77 struct net_device **array;
78 unsigned int size;
79
80 ASSERT_RTNL();
81
82 array = vg->vlan_devices_arrays[vlan_id / VLAN_GROUP_ARRAY_PART_LEN];
83 if (array != NULL)
84 return 0;
85
86 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
87 array = kzalloc(size, GFP_KERNEL);
88 if (array == NULL)
89 return -ENOBUFS;
90
91 vg->vlan_devices_arrays[vlan_id / VLAN_GROUP_ARRAY_PART_LEN] = array;
92 return 0;
93}
94
95static void vlan_rcu_free(struct rcu_head *rcu)
96{
97 vlan_group_free(container_of(rcu, struct vlan_group, rcu));
98}
99
100void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
101{
102 struct vlan_dev_info *vlan = vlan_dev_info(dev);
103 struct net_device *real_dev = vlan->real_dev;
104 const struct net_device_ops *ops = real_dev->netdev_ops;
105 struct vlan_group *grp;
106 u16 vlan_id = vlan->vlan_id;
107
108 ASSERT_RTNL();
109
110 grp = rtnl_dereference(real_dev->vlgrp);
111 BUG_ON(!grp);
112
113 /* Take it out of our own structures, but be sure to interlock with
114 * HW accelerating devices or SW vlan input packet processing if
115 * VLAN is not 0 (leave it there for 802.1p).
116 */
117 if (vlan_id && (real_dev->features & NETIF_F_HW_VLAN_FILTER))
118 ops->ndo_vlan_rx_kill_vid(real_dev, vlan_id);
119
120 grp->nr_vlans--;
121
122 if (vlan->flags & VLAN_FLAG_GVRP)
123 vlan_gvrp_request_leave(dev);
124
125 vlan_group_set_device(grp, vlan_id, NULL);
126 /* Because unregister_netdevice_queue() makes sure at least one rcu
127 * grace period is respected before device freeing,
128 * we dont need to call synchronize_net() here.
129 */
130 unregister_netdevice_queue(dev, head);
131
132 /* If the group is now empty, kill off the group. */
133 if (grp->nr_vlans == 0) {
134 vlan_gvrp_uninit_applicant(real_dev);
135
136 rcu_assign_pointer(real_dev->vlgrp, NULL);
137
138 /* Free the group, after all cpu's are done. */
139 call_rcu(&grp->rcu, vlan_rcu_free);
140 }
141
142 /* Get rid of the vlan's reference to real_dev */
143 dev_put(real_dev);
144}
145
146int vlan_check_real_dev(struct net_device *real_dev, u16 vlan_id)
147{
148 const char *name = real_dev->name;
149 const struct net_device_ops *ops = real_dev->netdev_ops;
150
151 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) {
152 pr_info("VLANs not supported on %s\n", name);
153 return -EOPNOTSUPP;
154 }
155
156 if ((real_dev->features & NETIF_F_HW_VLAN_FILTER) &&
157 (!ops->ndo_vlan_rx_add_vid || !ops->ndo_vlan_rx_kill_vid)) {
158 pr_info("Device %s has buggy VLAN hw accel\n", name);
159 return -EOPNOTSUPP;
160 }
161
162 if (vlan_find_dev(real_dev, vlan_id) != NULL)
163 return -EEXIST;
164
165 return 0;
166}
167
168int register_vlan_dev(struct net_device *dev)
169{
170 struct vlan_dev_info *vlan = vlan_dev_info(dev);
171 struct net_device *real_dev = vlan->real_dev;
172 const struct net_device_ops *ops = real_dev->netdev_ops;
173 u16 vlan_id = vlan->vlan_id;
174 struct vlan_group *grp, *ngrp = NULL;
175 int err;
176
177 grp = rtnl_dereference(real_dev->vlgrp);
178 if (!grp) {
179 ngrp = grp = vlan_group_alloc(real_dev);
180 if (!grp)
181 return -ENOBUFS;
182 err = vlan_gvrp_init_applicant(real_dev);
183 if (err < 0)
184 goto out_free_group;
185 }
186
187 err = vlan_group_prealloc_vid(grp, vlan_id);
188 if (err < 0)
189 goto out_uninit_applicant;
190
191 err = register_netdevice(dev);
192 if (err < 0)
193 goto out_uninit_applicant;
194
195 /* Account for reference in struct vlan_dev_info */
196 dev_hold(real_dev);
197
198 netif_stacked_transfer_operstate(real_dev, dev);
199 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
200
201 /* So, got the sucker initialized, now lets place
202 * it into our local structure.
203 */
204 vlan_group_set_device(grp, vlan_id, dev);
205 grp->nr_vlans++;
206
207 if (ngrp) {
208 rcu_assign_pointer(real_dev->vlgrp, ngrp);
209 }
210 if (real_dev->features & NETIF_F_HW_VLAN_FILTER)
211 ops->ndo_vlan_rx_add_vid(real_dev, vlan_id);
212
213 return 0;
214
215out_uninit_applicant:
216 if (ngrp)
217 vlan_gvrp_uninit_applicant(real_dev);
218out_free_group:
219 if (ngrp) {
220 /* Free the group, after all cpu's are done. */
221 call_rcu(&ngrp->rcu, vlan_rcu_free);
222 }
223 return err;
224}
225
226/* Attach a VLAN device to a mac address (ie Ethernet Card).
227 * Returns 0 if the device was created or a negative error code otherwise.
228 */
229static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
230{
231 struct net_device *new_dev;
232 struct net *net = dev_net(real_dev);
233 struct vlan_net *vn = net_generic(net, vlan_net_id);
234 char name[IFNAMSIZ];
235 int err;
236
237 if (vlan_id >= VLAN_VID_MASK)
238 return -ERANGE;
239
240 err = vlan_check_real_dev(real_dev, vlan_id);
241 if (err < 0)
242 return err;
243
244 /* Gotta set up the fields for the device. */
245 switch (vn->name_type) {
246 case VLAN_NAME_TYPE_RAW_PLUS_VID:
247 /* name will look like: eth1.0005 */
248 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
249 break;
250 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
251 /* Put our vlan.VID in the name.
252 * Name will look like: vlan5
253 */
254 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
255 break;
256 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
257 /* Put our vlan.VID in the name.
258 * Name will look like: eth0.5
259 */
260 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
261 break;
262 case VLAN_NAME_TYPE_PLUS_VID:
263 /* Put our vlan.VID in the name.
264 * Name will look like: vlan0005
265 */
266 default:
267 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
268 }
269
270 new_dev = alloc_netdev(sizeof(struct vlan_dev_info), name, vlan_setup);
271
272 if (new_dev == NULL)
273 return -ENOBUFS;
274
275 dev_net_set(new_dev, net);
276 /* need 4 bytes for extra VLAN header info,
277 * hope the underlying device can handle it.
278 */
279 new_dev->mtu = real_dev->mtu;
280
281 vlan_dev_info(new_dev)->vlan_id = vlan_id;
282 vlan_dev_info(new_dev)->real_dev = real_dev;
283 vlan_dev_info(new_dev)->dent = NULL;
284 vlan_dev_info(new_dev)->flags = VLAN_FLAG_REORDER_HDR;
285
286 new_dev->rtnl_link_ops = &vlan_link_ops;
287 err = register_vlan_dev(new_dev);
288 if (err < 0)
289 goto out_free_newdev;
290
291 return 0;
292
293out_free_newdev:
294 free_netdev(new_dev);
295 return err;
296}
297
298static void vlan_sync_address(struct net_device *dev,
299 struct net_device *vlandev)
300{
301 struct vlan_dev_info *vlan = vlan_dev_info(vlandev);
302
303 /* May be called without an actual change */
304 if (!compare_ether_addr(vlan->real_dev_addr, dev->dev_addr))
305 return;
306
307 /* vlan address was different from the old address and is equal to
308 * the new address */
309 if (compare_ether_addr(vlandev->dev_addr, vlan->real_dev_addr) &&
310 !compare_ether_addr(vlandev->dev_addr, dev->dev_addr))
311 dev_uc_del(dev, vlandev->dev_addr);
312
313 /* vlan address was equal to the old address and is different from
314 * the new address */
315 if (!compare_ether_addr(vlandev->dev_addr, vlan->real_dev_addr) &&
316 compare_ether_addr(vlandev->dev_addr, dev->dev_addr))
317 dev_uc_add(dev, vlandev->dev_addr);
318
319 memcpy(vlan->real_dev_addr, dev->dev_addr, ETH_ALEN);
320}
321
322static void vlan_transfer_features(struct net_device *dev,
323 struct net_device *vlandev)
324{
325 vlandev->gso_max_size = dev->gso_max_size;
326
327 if (dev->features & NETIF_F_HW_VLAN_TX)
328 vlandev->hard_header_len = dev->hard_header_len;
329 else
330 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
331
332#if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE)
333 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
334#endif
335
336 netdev_update_features(vlandev);
337}
338
339static void __vlan_device_event(struct net_device *dev, unsigned long event)
340{
341 switch (event) {
342 case NETDEV_CHANGENAME:
343 vlan_proc_rem_dev(dev);
344 if (vlan_proc_add_dev(dev) < 0)
345 pr_warn("failed to change proc name for %s\n",
346 dev->name);
347 break;
348 case NETDEV_REGISTER:
349 if (vlan_proc_add_dev(dev) < 0)
350 pr_warn("failed to add proc entry for %s\n", dev->name);
351 break;
352 case NETDEV_UNREGISTER:
353 vlan_proc_rem_dev(dev);
354 break;
355 }
356}
357
358static int vlan_device_event(struct notifier_block *unused, unsigned long event,
359 void *ptr)
360{
361 struct net_device *dev = ptr;
362 struct vlan_group *grp;
363 int i, flgs;
364 struct net_device *vlandev;
365 struct vlan_dev_info *vlan;
366 LIST_HEAD(list);
367
368 if (is_vlan_dev(dev))
369 __vlan_device_event(dev, event);
370
371 if ((event == NETDEV_UP) &&
372 (dev->features & NETIF_F_HW_VLAN_FILTER) &&
373 dev->netdev_ops->ndo_vlan_rx_add_vid) {
374 pr_info("adding VLAN 0 to HW filter on device %s\n",
375 dev->name);
376 dev->netdev_ops->ndo_vlan_rx_add_vid(dev, 0);
377 }
378
379 grp = rtnl_dereference(dev->vlgrp);
380 if (!grp)
381 goto out;
382
383 /* It is OK that we do not hold the group lock right now,
384 * as we run under the RTNL lock.
385 */
386
387 switch (event) {
388 case NETDEV_CHANGE:
389 /* Propagate real device state to vlan devices */
390 for (i = 0; i < VLAN_N_VID; i++) {
391 vlandev = vlan_group_get_device(grp, i);
392 if (!vlandev)
393 continue;
394
395 netif_stacked_transfer_operstate(dev, vlandev);
396 }
397 break;
398
399 case NETDEV_CHANGEADDR:
400 /* Adjust unicast filters on underlying device */
401 for (i = 0; i < VLAN_N_VID; i++) {
402 vlandev = vlan_group_get_device(grp, i);
403 if (!vlandev)
404 continue;
405
406 flgs = vlandev->flags;
407 if (!(flgs & IFF_UP))
408 continue;
409
410 vlan_sync_address(dev, vlandev);
411 }
412 break;
413
414 case NETDEV_CHANGEMTU:
415 for (i = 0; i < VLAN_N_VID; i++) {
416 vlandev = vlan_group_get_device(grp, i);
417 if (!vlandev)
418 continue;
419
420 if (vlandev->mtu <= dev->mtu)
421 continue;
422
423 dev_set_mtu(vlandev, dev->mtu);
424 }
425 break;
426
427 case NETDEV_FEAT_CHANGE:
428 /* Propagate device features to underlying device */
429 for (i = 0; i < VLAN_N_VID; i++) {
430 vlandev = vlan_group_get_device(grp, i);
431 if (!vlandev)
432 continue;
433
434 vlan_transfer_features(dev, vlandev);
435 }
436
437 break;
438
439 case NETDEV_DOWN:
440 /* Put all VLANs for this dev in the down state too. */
441 for (i = 0; i < VLAN_N_VID; i++) {
442 vlandev = vlan_group_get_device(grp, i);
443 if (!vlandev)
444 continue;
445
446 flgs = vlandev->flags;
447 if (!(flgs & IFF_UP))
448 continue;
449
450 vlan = vlan_dev_info(vlandev);
451 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
452 dev_change_flags(vlandev, flgs & ~IFF_UP);
453 netif_stacked_transfer_operstate(dev, vlandev);
454 }
455 break;
456
457 case NETDEV_UP:
458 /* Put all VLANs for this dev in the up state too. */
459 for (i = 0; i < VLAN_N_VID; i++) {
460 vlandev = vlan_group_get_device(grp, i);
461 if (!vlandev)
462 continue;
463
464 flgs = vlandev->flags;
465 if (flgs & IFF_UP)
466 continue;
467
468 vlan = vlan_dev_info(vlandev);
469 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
470 dev_change_flags(vlandev, flgs | IFF_UP);
471 netif_stacked_transfer_operstate(dev, vlandev);
472 }
473 break;
474
475 case NETDEV_UNREGISTER:
476 /* twiddle thumbs on netns device moves */
477 if (dev->reg_state != NETREG_UNREGISTERING)
478 break;
479
480 for (i = 0; i < VLAN_N_VID; i++) {
481 vlandev = vlan_group_get_device(grp, i);
482 if (!vlandev)
483 continue;
484
485 /* unregistration of last vlan destroys group, abort
486 * afterwards */
487 if (grp->nr_vlans == 1)
488 i = VLAN_N_VID;
489
490 unregister_vlan_dev(vlandev, &list);
491 }
492 unregister_netdevice_many(&list);
493 break;
494
495 case NETDEV_PRE_TYPE_CHANGE:
496 /* Forbid underlaying device to change its type. */
497 return NOTIFY_BAD;
498
499 case NETDEV_NOTIFY_PEERS:
500 case NETDEV_BONDING_FAILOVER:
501 /* Propagate to vlan devices */
502 for (i = 0; i < VLAN_N_VID; i++) {
503 vlandev = vlan_group_get_device(grp, i);
504 if (!vlandev)
505 continue;
506
507 call_netdevice_notifiers(event, vlandev);
508 }
509 break;
510 }
511
512out:
513 return NOTIFY_DONE;
514}
515
516static struct notifier_block vlan_notifier_block __read_mostly = {
517 .notifier_call = vlan_device_event,
518};
519
520/*
521 * VLAN IOCTL handler.
522 * o execute requested action or pass command to the device driver
523 * arg is really a struct vlan_ioctl_args __user *.
524 */
525static int vlan_ioctl_handler(struct net *net, void __user *arg)
526{
527 int err;
528 struct vlan_ioctl_args args;
529 struct net_device *dev = NULL;
530
531 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
532 return -EFAULT;
533
534 /* Null terminate this sucker, just in case. */
535 args.device1[23] = 0;
536 args.u.device2[23] = 0;
537
538 rtnl_lock();
539
540 switch (args.cmd) {
541 case SET_VLAN_INGRESS_PRIORITY_CMD:
542 case SET_VLAN_EGRESS_PRIORITY_CMD:
543 case SET_VLAN_FLAG_CMD:
544 case ADD_VLAN_CMD:
545 case DEL_VLAN_CMD:
546 case GET_VLAN_REALDEV_NAME_CMD:
547 case GET_VLAN_VID_CMD:
548 err = -ENODEV;
549 dev = __dev_get_by_name(net, args.device1);
550 if (!dev)
551 goto out;
552
553 err = -EINVAL;
554 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
555 goto out;
556 }
557
558 switch (args.cmd) {
559 case SET_VLAN_INGRESS_PRIORITY_CMD:
560 err = -EPERM;
561 if (!capable(CAP_NET_ADMIN))
562 break;
563 vlan_dev_set_ingress_priority(dev,
564 args.u.skb_priority,
565 args.vlan_qos);
566 err = 0;
567 break;
568
569 case SET_VLAN_EGRESS_PRIORITY_CMD:
570 err = -EPERM;
571 if (!capable(CAP_NET_ADMIN))
572 break;
573 err = vlan_dev_set_egress_priority(dev,
574 args.u.skb_priority,
575 args.vlan_qos);
576 break;
577
578 case SET_VLAN_FLAG_CMD:
579 err = -EPERM;
580 if (!capable(CAP_NET_ADMIN))
581 break;
582 err = vlan_dev_change_flags(dev,
583 args.vlan_qos ? args.u.flag : 0,
584 args.u.flag);
585 break;
586
587 case SET_VLAN_NAME_TYPE_CMD:
588 err = -EPERM;
589 if (!capable(CAP_NET_ADMIN))
590 break;
591 if ((args.u.name_type >= 0) &&
592 (args.u.name_type < VLAN_NAME_TYPE_HIGHEST)) {
593 struct vlan_net *vn;
594
595 vn = net_generic(net, vlan_net_id);
596 vn->name_type = args.u.name_type;
597 err = 0;
598 } else {
599 err = -EINVAL;
600 }
601 break;
602
603 case ADD_VLAN_CMD:
604 err = -EPERM;
605 if (!capable(CAP_NET_ADMIN))
606 break;
607 err = register_vlan_device(dev, args.u.VID);
608 break;
609
610 case DEL_VLAN_CMD:
611 err = -EPERM;
612 if (!capable(CAP_NET_ADMIN))
613 break;
614 unregister_vlan_dev(dev, NULL);
615 err = 0;
616 break;
617
618 case GET_VLAN_REALDEV_NAME_CMD:
619 err = 0;
620 vlan_dev_get_realdev_name(dev, args.u.device2);
621 if (copy_to_user(arg, &args,
622 sizeof(struct vlan_ioctl_args)))
623 err = -EFAULT;
624 break;
625
626 case GET_VLAN_VID_CMD:
627 err = 0;
628 args.u.VID = vlan_dev_vlan_id(dev);
629 if (copy_to_user(arg, &args,
630 sizeof(struct vlan_ioctl_args)))
631 err = -EFAULT;
632 break;
633
634 default:
635 err = -EOPNOTSUPP;
636 break;
637 }
638out:
639 rtnl_unlock();
640 return err;
641}
642
643static int __net_init vlan_init_net(struct net *net)
644{
645 struct vlan_net *vn = net_generic(net, vlan_net_id);
646 int err;
647
648 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
649
650 err = vlan_proc_init(net);
651
652 return err;
653}
654
655static void __net_exit vlan_exit_net(struct net *net)
656{
657 vlan_proc_cleanup(net);
658}
659
660static struct pernet_operations vlan_net_ops = {
661 .init = vlan_init_net,
662 .exit = vlan_exit_net,
663 .id = &vlan_net_id,
664 .size = sizeof(struct vlan_net),
665};
666
667static int __init vlan_proto_init(void)
668{
669 int err;
670
671 pr_info("%s v%s\n", vlan_fullname, vlan_version);
672
673 err = register_pernet_subsys(&vlan_net_ops);
674 if (err < 0)
675 goto err0;
676
677 err = register_netdevice_notifier(&vlan_notifier_block);
678 if (err < 0)
679 goto err2;
680
681 err = vlan_gvrp_init();
682 if (err < 0)
683 goto err3;
684
685 err = vlan_netlink_init();
686 if (err < 0)
687 goto err4;
688
689 vlan_ioctl_set(vlan_ioctl_handler);
690 return 0;
691
692err4:
693 vlan_gvrp_uninit();
694err3:
695 unregister_netdevice_notifier(&vlan_notifier_block);
696err2:
697 unregister_pernet_subsys(&vlan_net_ops);
698err0:
699 return err;
700}
701
702static void __exit vlan_cleanup_module(void)
703{
704 vlan_ioctl_set(NULL);
705 vlan_netlink_fini();
706
707 unregister_netdevice_notifier(&vlan_notifier_block);
708
709 unregister_pernet_subsys(&vlan_net_ops);
710 rcu_barrier(); /* Wait for completion of call_rcu()'s */
711
712 vlan_gvrp_uninit();
713}
714
715module_init(vlan_proto_init);
716module_exit(vlan_cleanup_module);
717
718MODULE_LICENSE("GPL");
719MODULE_VERSION(DRV_VERSION);
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * INET 802.1Q VLAN
4 * Ethernet-type device handling.
5 *
6 * Authors: Ben Greear <greearb@candelatech.com>
7 * Please send support related email to: netdev@vger.kernel.org
8 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
9 *
10 * Fixes:
11 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
12 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
13 * Correct all the locking - David S. Miller <davem@redhat.com>;
14 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include <linux/capability.h>
20#include <linux/module.h>
21#include <linux/netdevice.h>
22#include <linux/skbuff.h>
23#include <linux/slab.h>
24#include <linux/init.h>
25#include <linux/rculist.h>
26#include <net/p8022.h>
27#include <net/arp.h>
28#include <linux/rtnetlink.h>
29#include <linux/notifier.h>
30#include <net/rtnetlink.h>
31#include <net/net_namespace.h>
32#include <net/netns/generic.h>
33#include <linux/uaccess.h>
34
35#include <linux/if_vlan.h>
36#include "vlan.h"
37#include "vlanproc.h"
38
39#define DRV_VERSION "1.8"
40
41/* Global VLAN variables */
42
43unsigned int vlan_net_id __read_mostly;
44
45const char vlan_fullname[] = "802.1Q VLAN Support";
46const char vlan_version[] = DRV_VERSION;
47
48/* End of global variables definitions. */
49
50static int vlan_group_prealloc_vid(struct vlan_group *vg,
51 __be16 vlan_proto, u16 vlan_id)
52{
53 struct net_device **array;
54 unsigned int vidx;
55 unsigned int size;
56 int pidx;
57
58 ASSERT_RTNL();
59
60 pidx = vlan_proto_idx(vlan_proto);
61 if (pidx < 0)
62 return -EINVAL;
63
64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65 array = vg->vlan_devices_arrays[pidx][vidx];
66 if (array != NULL)
67 return 0;
68
69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70 array = kzalloc(size, GFP_KERNEL_ACCOUNT);
71 if (array == NULL)
72 return -ENOBUFS;
73
74 /* paired with smp_rmb() in __vlan_group_get_device() */
75 smp_wmb();
76
77 vg->vlan_devices_arrays[pidx][vidx] = array;
78 return 0;
79}
80
81static void vlan_stacked_transfer_operstate(const struct net_device *rootdev,
82 struct net_device *dev,
83 struct vlan_dev_priv *vlan)
84{
85 if (!(vlan->flags & VLAN_FLAG_BRIDGE_BINDING))
86 netif_stacked_transfer_operstate(rootdev, dev);
87}
88
89void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
90{
91 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
92 struct net_device *real_dev = vlan->real_dev;
93 struct vlan_info *vlan_info;
94 struct vlan_group *grp;
95 u16 vlan_id = vlan->vlan_id;
96
97 ASSERT_RTNL();
98
99 vlan_info = rtnl_dereference(real_dev->vlan_info);
100 BUG_ON(!vlan_info);
101
102 grp = &vlan_info->grp;
103
104 grp->nr_vlan_devs--;
105
106 if (vlan->flags & VLAN_FLAG_MVRP)
107 vlan_mvrp_request_leave(dev);
108 if (vlan->flags & VLAN_FLAG_GVRP)
109 vlan_gvrp_request_leave(dev);
110
111 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
112
113 netdev_upper_dev_unlink(real_dev, dev);
114 /* Because unregister_netdevice_queue() makes sure at least one rcu
115 * grace period is respected before device freeing,
116 * we dont need to call synchronize_net() here.
117 */
118 unregister_netdevice_queue(dev, head);
119
120 if (grp->nr_vlan_devs == 0) {
121 vlan_mvrp_uninit_applicant(real_dev);
122 vlan_gvrp_uninit_applicant(real_dev);
123 }
124
125 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
126}
127
128int vlan_check_real_dev(struct net_device *real_dev,
129 __be16 protocol, u16 vlan_id,
130 struct netlink_ext_ack *extack)
131{
132 const char *name = real_dev->name;
133
134 if (real_dev->features & NETIF_F_VLAN_CHALLENGED ||
135 real_dev->type != ARPHRD_ETHER) {
136 pr_info("VLANs not supported on %s\n", name);
137 NL_SET_ERR_MSG_MOD(extack, "VLANs not supported on device");
138 return -EOPNOTSUPP;
139 }
140
141 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL) {
142 NL_SET_ERR_MSG_MOD(extack, "VLAN device already exists");
143 return -EEXIST;
144 }
145
146 return 0;
147}
148
149int register_vlan_dev(struct net_device *dev, struct netlink_ext_ack *extack)
150{
151 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
152 struct net_device *real_dev = vlan->real_dev;
153 u16 vlan_id = vlan->vlan_id;
154 struct vlan_info *vlan_info;
155 struct vlan_group *grp;
156 int err;
157
158 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
159 if (err)
160 return err;
161
162 vlan_info = rtnl_dereference(real_dev->vlan_info);
163 /* vlan_info should be there now. vlan_vid_add took care of it */
164 BUG_ON(!vlan_info);
165
166 grp = &vlan_info->grp;
167 if (grp->nr_vlan_devs == 0) {
168 err = vlan_gvrp_init_applicant(real_dev);
169 if (err < 0)
170 goto out_vid_del;
171 err = vlan_mvrp_init_applicant(real_dev);
172 if (err < 0)
173 goto out_uninit_gvrp;
174 }
175
176 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
177 if (err < 0)
178 goto out_uninit_mvrp;
179
180 err = register_netdevice(dev);
181 if (err < 0)
182 goto out_uninit_mvrp;
183
184 err = netdev_upper_dev_link(real_dev, dev, extack);
185 if (err)
186 goto out_unregister_netdev;
187
188 vlan_stacked_transfer_operstate(real_dev, dev, vlan);
189 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
190
191 /* So, got the sucker initialized, now lets place
192 * it into our local structure.
193 */
194 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
195 grp->nr_vlan_devs++;
196
197 return 0;
198
199out_unregister_netdev:
200 unregister_netdevice(dev);
201out_uninit_mvrp:
202 if (grp->nr_vlan_devs == 0)
203 vlan_mvrp_uninit_applicant(real_dev);
204out_uninit_gvrp:
205 if (grp->nr_vlan_devs == 0)
206 vlan_gvrp_uninit_applicant(real_dev);
207out_vid_del:
208 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
209 return err;
210}
211
212/* Attach a VLAN device to a mac address (ie Ethernet Card).
213 * Returns 0 if the device was created or a negative error code otherwise.
214 */
215static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
216{
217 struct net_device *new_dev;
218 struct vlan_dev_priv *vlan;
219 struct net *net = dev_net(real_dev);
220 struct vlan_net *vn = net_generic(net, vlan_net_id);
221 char name[IFNAMSIZ];
222 int err;
223
224 if (vlan_id >= VLAN_VID_MASK)
225 return -ERANGE;
226
227 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id,
228 NULL);
229 if (err < 0)
230 return err;
231
232 /* Gotta set up the fields for the device. */
233 switch (vn->name_type) {
234 case VLAN_NAME_TYPE_RAW_PLUS_VID:
235 /* name will look like: eth1.0005 */
236 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
237 break;
238 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
239 /* Put our vlan.VID in the name.
240 * Name will look like: vlan5
241 */
242 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
243 break;
244 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
245 /* Put our vlan.VID in the name.
246 * Name will look like: eth0.5
247 */
248 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
249 break;
250 case VLAN_NAME_TYPE_PLUS_VID:
251 /* Put our vlan.VID in the name.
252 * Name will look like: vlan0005
253 */
254 default:
255 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
256 }
257
258 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
259 NET_NAME_UNKNOWN, vlan_setup);
260
261 if (new_dev == NULL)
262 return -ENOBUFS;
263
264 dev_net_set(new_dev, net);
265 /* need 4 bytes for extra VLAN header info,
266 * hope the underlying device can handle it.
267 */
268 new_dev->mtu = real_dev->mtu;
269
270 vlan = vlan_dev_priv(new_dev);
271 vlan->vlan_proto = htons(ETH_P_8021Q);
272 vlan->vlan_id = vlan_id;
273 vlan->real_dev = real_dev;
274 vlan->dent = NULL;
275 vlan->flags = VLAN_FLAG_REORDER_HDR;
276
277 new_dev->rtnl_link_ops = &vlan_link_ops;
278 err = register_vlan_dev(new_dev, NULL);
279 if (err < 0)
280 goto out_free_newdev;
281
282 return 0;
283
284out_free_newdev:
285 free_netdev(new_dev);
286 return err;
287}
288
289static void vlan_sync_address(struct net_device *dev,
290 struct net_device *vlandev)
291{
292 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
293
294 /* May be called without an actual change */
295 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
296 return;
297
298 /* vlan continues to inherit address of lower device */
299 if (vlan_dev_inherit_address(vlandev, dev))
300 goto out;
301
302 /* vlan address was different from the old address and is equal to
303 * the new address */
304 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
305 ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
306 dev_uc_del(dev, vlandev->dev_addr);
307
308 /* vlan address was equal to the old address and is different from
309 * the new address */
310 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
311 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
312 dev_uc_add(dev, vlandev->dev_addr);
313
314out:
315 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
316}
317
318static void vlan_transfer_features(struct net_device *dev,
319 struct net_device *vlandev)
320{
321 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
322
323 netif_inherit_tso_max(vlandev, dev);
324
325 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
326 vlandev->hard_header_len = dev->hard_header_len;
327 else
328 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
329
330#if IS_ENABLED(CONFIG_FCOE)
331 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
332#endif
333
334 vlandev->priv_flags &= ~IFF_XMIT_DST_RELEASE;
335 vlandev->priv_flags |= (vlan->real_dev->priv_flags & IFF_XMIT_DST_RELEASE);
336 vlandev->hw_enc_features = vlan_tnl_features(vlan->real_dev);
337
338 netdev_update_features(vlandev);
339}
340
341static int __vlan_device_event(struct net_device *dev, unsigned long event)
342{
343 int err = 0;
344
345 switch (event) {
346 case NETDEV_CHANGENAME:
347 vlan_proc_rem_dev(dev);
348 err = vlan_proc_add_dev(dev);
349 break;
350 case NETDEV_REGISTER:
351 err = vlan_proc_add_dev(dev);
352 break;
353 case NETDEV_UNREGISTER:
354 vlan_proc_rem_dev(dev);
355 break;
356 }
357
358 return err;
359}
360
361static int vlan_device_event(struct notifier_block *unused, unsigned long event,
362 void *ptr)
363{
364 struct netlink_ext_ack *extack = netdev_notifier_info_to_extack(ptr);
365 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
366 struct vlan_group *grp;
367 struct vlan_info *vlan_info;
368 int i, flgs;
369 struct net_device *vlandev;
370 struct vlan_dev_priv *vlan;
371 bool last = false;
372 LIST_HEAD(list);
373 int err;
374
375 if (is_vlan_dev(dev)) {
376 int err = __vlan_device_event(dev, event);
377
378 if (err)
379 return notifier_from_errno(err);
380 }
381
382 if ((event == NETDEV_UP) &&
383 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
384 pr_info("adding VLAN 0 to HW filter on device %s\n",
385 dev->name);
386 vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
387 }
388 if (event == NETDEV_DOWN &&
389 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER))
390 vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
391
392 vlan_info = rtnl_dereference(dev->vlan_info);
393 if (!vlan_info)
394 goto out;
395 grp = &vlan_info->grp;
396
397 /* It is OK that we do not hold the group lock right now,
398 * as we run under the RTNL lock.
399 */
400
401 switch (event) {
402 case NETDEV_CHANGE:
403 /* Propagate real device state to vlan devices */
404 vlan_group_for_each_dev(grp, i, vlandev)
405 vlan_stacked_transfer_operstate(dev, vlandev,
406 vlan_dev_priv(vlandev));
407 break;
408
409 case NETDEV_CHANGEADDR:
410 /* Adjust unicast filters on underlying device */
411 vlan_group_for_each_dev(grp, i, vlandev) {
412 flgs = vlandev->flags;
413 if (!(flgs & IFF_UP))
414 continue;
415
416 vlan_sync_address(dev, vlandev);
417 }
418 break;
419
420 case NETDEV_CHANGEMTU:
421 vlan_group_for_each_dev(grp, i, vlandev) {
422 if (vlandev->mtu <= dev->mtu)
423 continue;
424
425 dev_set_mtu(vlandev, dev->mtu);
426 }
427 break;
428
429 case NETDEV_FEAT_CHANGE:
430 /* Propagate device features to underlying device */
431 vlan_group_for_each_dev(grp, i, vlandev)
432 vlan_transfer_features(dev, vlandev);
433 break;
434
435 case NETDEV_DOWN: {
436 struct net_device *tmp;
437 LIST_HEAD(close_list);
438
439 /* Put all VLANs for this dev in the down state too. */
440 vlan_group_for_each_dev(grp, i, vlandev) {
441 flgs = vlandev->flags;
442 if (!(flgs & IFF_UP))
443 continue;
444
445 vlan = vlan_dev_priv(vlandev);
446 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
447 list_add(&vlandev->close_list, &close_list);
448 }
449
450 dev_close_many(&close_list, false);
451
452 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
453 vlan_stacked_transfer_operstate(dev, vlandev,
454 vlan_dev_priv(vlandev));
455 list_del_init(&vlandev->close_list);
456 }
457 list_del(&close_list);
458 break;
459 }
460 case NETDEV_UP:
461 /* Put all VLANs for this dev in the up state too. */
462 vlan_group_for_each_dev(grp, i, vlandev) {
463 flgs = dev_get_flags(vlandev);
464 if (flgs & IFF_UP)
465 continue;
466
467 vlan = vlan_dev_priv(vlandev);
468 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
469 dev_change_flags(vlandev, flgs | IFF_UP,
470 extack);
471 vlan_stacked_transfer_operstate(dev, vlandev, vlan);
472 }
473 break;
474
475 case NETDEV_UNREGISTER:
476 /* twiddle thumbs on netns device moves */
477 if (dev->reg_state != NETREG_UNREGISTERING)
478 break;
479
480 vlan_group_for_each_dev(grp, i, vlandev) {
481 /* removal of last vid destroys vlan_info, abort
482 * afterwards */
483 if (vlan_info->nr_vids == 1)
484 last = true;
485
486 unregister_vlan_dev(vlandev, &list);
487 if (last)
488 break;
489 }
490 unregister_netdevice_many(&list);
491 break;
492
493 case NETDEV_PRE_TYPE_CHANGE:
494 /* Forbid underlaying device to change its type. */
495 if (vlan_uses_dev(dev))
496 return NOTIFY_BAD;
497 break;
498
499 case NETDEV_NOTIFY_PEERS:
500 case NETDEV_BONDING_FAILOVER:
501 case NETDEV_RESEND_IGMP:
502 /* Propagate to vlan devices */
503 vlan_group_for_each_dev(grp, i, vlandev)
504 call_netdevice_notifiers(event, vlandev);
505 break;
506
507 case NETDEV_CVLAN_FILTER_PUSH_INFO:
508 err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021Q));
509 if (err)
510 return notifier_from_errno(err);
511 break;
512
513 case NETDEV_CVLAN_FILTER_DROP_INFO:
514 vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021Q));
515 break;
516
517 case NETDEV_SVLAN_FILTER_PUSH_INFO:
518 err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021AD));
519 if (err)
520 return notifier_from_errno(err);
521 break;
522
523 case NETDEV_SVLAN_FILTER_DROP_INFO:
524 vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021AD));
525 break;
526 }
527
528out:
529 return NOTIFY_DONE;
530}
531
532static struct notifier_block vlan_notifier_block __read_mostly = {
533 .notifier_call = vlan_device_event,
534};
535
536/*
537 * VLAN IOCTL handler.
538 * o execute requested action or pass command to the device driver
539 * arg is really a struct vlan_ioctl_args __user *.
540 */
541static int vlan_ioctl_handler(struct net *net, void __user *arg)
542{
543 int err;
544 struct vlan_ioctl_args args;
545 struct net_device *dev = NULL;
546
547 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
548 return -EFAULT;
549
550 /* Null terminate this sucker, just in case. */
551 args.device1[sizeof(args.device1) - 1] = 0;
552 args.u.device2[sizeof(args.u.device2) - 1] = 0;
553
554 rtnl_lock();
555
556 switch (args.cmd) {
557 case SET_VLAN_INGRESS_PRIORITY_CMD:
558 case SET_VLAN_EGRESS_PRIORITY_CMD:
559 case SET_VLAN_FLAG_CMD:
560 case ADD_VLAN_CMD:
561 case DEL_VLAN_CMD:
562 case GET_VLAN_REALDEV_NAME_CMD:
563 case GET_VLAN_VID_CMD:
564 err = -ENODEV;
565 dev = __dev_get_by_name(net, args.device1);
566 if (!dev)
567 goto out;
568
569 err = -EINVAL;
570 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
571 goto out;
572 }
573
574 switch (args.cmd) {
575 case SET_VLAN_INGRESS_PRIORITY_CMD:
576 err = -EPERM;
577 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
578 break;
579 vlan_dev_set_ingress_priority(dev,
580 args.u.skb_priority,
581 args.vlan_qos);
582 err = 0;
583 break;
584
585 case SET_VLAN_EGRESS_PRIORITY_CMD:
586 err = -EPERM;
587 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
588 break;
589 err = vlan_dev_set_egress_priority(dev,
590 args.u.skb_priority,
591 args.vlan_qos);
592 break;
593
594 case SET_VLAN_FLAG_CMD:
595 err = -EPERM;
596 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
597 break;
598 err = vlan_dev_change_flags(dev,
599 args.vlan_qos ? args.u.flag : 0,
600 args.u.flag);
601 break;
602
603 case SET_VLAN_NAME_TYPE_CMD:
604 err = -EPERM;
605 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
606 break;
607 if (args.u.name_type < VLAN_NAME_TYPE_HIGHEST) {
608 struct vlan_net *vn;
609
610 vn = net_generic(net, vlan_net_id);
611 vn->name_type = args.u.name_type;
612 err = 0;
613 } else {
614 err = -EINVAL;
615 }
616 break;
617
618 case ADD_VLAN_CMD:
619 err = -EPERM;
620 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
621 break;
622 err = register_vlan_device(dev, args.u.VID);
623 break;
624
625 case DEL_VLAN_CMD:
626 err = -EPERM;
627 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
628 break;
629 unregister_vlan_dev(dev, NULL);
630 err = 0;
631 break;
632
633 case GET_VLAN_REALDEV_NAME_CMD:
634 err = 0;
635 vlan_dev_get_realdev_name(dev, args.u.device2,
636 sizeof(args.u.device2));
637 if (copy_to_user(arg, &args,
638 sizeof(struct vlan_ioctl_args)))
639 err = -EFAULT;
640 break;
641
642 case GET_VLAN_VID_CMD:
643 err = 0;
644 args.u.VID = vlan_dev_vlan_id(dev);
645 if (copy_to_user(arg, &args,
646 sizeof(struct vlan_ioctl_args)))
647 err = -EFAULT;
648 break;
649
650 default:
651 err = -EOPNOTSUPP;
652 break;
653 }
654out:
655 rtnl_unlock();
656 return err;
657}
658
659static int __net_init vlan_init_net(struct net *net)
660{
661 struct vlan_net *vn = net_generic(net, vlan_net_id);
662 int err;
663
664 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
665
666 err = vlan_proc_init(net);
667
668 return err;
669}
670
671static void __net_exit vlan_exit_net(struct net *net)
672{
673 vlan_proc_cleanup(net);
674}
675
676static struct pernet_operations vlan_net_ops = {
677 .init = vlan_init_net,
678 .exit = vlan_exit_net,
679 .id = &vlan_net_id,
680 .size = sizeof(struct vlan_net),
681};
682
683static int __init vlan_proto_init(void)
684{
685 int err;
686
687 pr_info("%s v%s\n", vlan_fullname, vlan_version);
688
689 err = register_pernet_subsys(&vlan_net_ops);
690 if (err < 0)
691 goto err0;
692
693 err = register_netdevice_notifier(&vlan_notifier_block);
694 if (err < 0)
695 goto err2;
696
697 err = vlan_gvrp_init();
698 if (err < 0)
699 goto err3;
700
701 err = vlan_mvrp_init();
702 if (err < 0)
703 goto err4;
704
705 err = vlan_netlink_init();
706 if (err < 0)
707 goto err5;
708
709 vlan_ioctl_set(vlan_ioctl_handler);
710 return 0;
711
712err5:
713 vlan_mvrp_uninit();
714err4:
715 vlan_gvrp_uninit();
716err3:
717 unregister_netdevice_notifier(&vlan_notifier_block);
718err2:
719 unregister_pernet_subsys(&vlan_net_ops);
720err0:
721 return err;
722}
723
724static void __exit vlan_cleanup_module(void)
725{
726 vlan_ioctl_set(NULL);
727
728 vlan_netlink_fini();
729
730 unregister_netdevice_notifier(&vlan_notifier_block);
731
732 unregister_pernet_subsys(&vlan_net_ops);
733 rcu_barrier(); /* Wait for completion of call_rcu()'s */
734
735 vlan_mvrp_uninit();
736 vlan_gvrp_uninit();
737}
738
739module_init(vlan_proto_init);
740module_exit(vlan_cleanup_module);
741
742MODULE_DESCRIPTION("802.1Q/802.1ad VLAN Protocol");
743MODULE_LICENSE("GPL");
744MODULE_VERSION(DRV_VERSION);