Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * INET 802.1Q VLAN
4 * Ethernet-type device handling.
5 *
6 * Authors: Ben Greear <greearb@candelatech.com>
7 * Please send support related email to: netdev@vger.kernel.org
8 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
9 *
10 * Fixes:
11 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
12 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
13 * Correct all the locking - David S. Miller <davem@redhat.com>;
14 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include <linux/capability.h>
20#include <linux/module.h>
21#include <linux/netdevice.h>
22#include <linux/skbuff.h>
23#include <linux/slab.h>
24#include <linux/init.h>
25#include <linux/rculist.h>
26#include <net/p8022.h>
27#include <net/arp.h>
28#include <linux/rtnetlink.h>
29#include <linux/notifier.h>
30#include <net/rtnetlink.h>
31#include <net/net_namespace.h>
32#include <net/netns/generic.h>
33#include <linux/uaccess.h>
34
35#include <linux/if_vlan.h>
36#include "vlan.h"
37#include "vlanproc.h"
38
39#define DRV_VERSION "1.8"
40
41/* Global VLAN variables */
42
43unsigned int vlan_net_id __read_mostly;
44
45const char vlan_fullname[] = "802.1Q VLAN Support";
46const char vlan_version[] = DRV_VERSION;
47
48/* End of global variables definitions. */
49
50static int vlan_group_prealloc_vid(struct vlan_group *vg,
51 __be16 vlan_proto, u16 vlan_id)
52{
53 struct net_device **array;
54 unsigned int vidx;
55 unsigned int size;
56 int pidx;
57
58 ASSERT_RTNL();
59
60 pidx = vlan_proto_idx(vlan_proto);
61 if (pidx < 0)
62 return -EINVAL;
63
64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65 array = vg->vlan_devices_arrays[pidx][vidx];
66 if (array != NULL)
67 return 0;
68
69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70 array = kzalloc(size, GFP_KERNEL_ACCOUNT);
71 if (array == NULL)
72 return -ENOBUFS;
73
74 /* paired with smp_rmb() in __vlan_group_get_device() */
75 smp_wmb();
76
77 vg->vlan_devices_arrays[pidx][vidx] = array;
78 return 0;
79}
80
81static void vlan_stacked_transfer_operstate(const struct net_device *rootdev,
82 struct net_device *dev,
83 struct vlan_dev_priv *vlan)
84{
85 if (!(vlan->flags & VLAN_FLAG_BRIDGE_BINDING))
86 netif_stacked_transfer_operstate(rootdev, dev);
87}
88
89void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
90{
91 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
92 struct net_device *real_dev = vlan->real_dev;
93 struct vlan_info *vlan_info;
94 struct vlan_group *grp;
95 u16 vlan_id = vlan->vlan_id;
96
97 ASSERT_RTNL();
98
99 vlan_info = rtnl_dereference(real_dev->vlan_info);
100 BUG_ON(!vlan_info);
101
102 grp = &vlan_info->grp;
103
104 grp->nr_vlan_devs--;
105
106 if (vlan->flags & VLAN_FLAG_MVRP)
107 vlan_mvrp_request_leave(dev);
108 if (vlan->flags & VLAN_FLAG_GVRP)
109 vlan_gvrp_request_leave(dev);
110
111 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
112
113 netdev_upper_dev_unlink(real_dev, dev);
114 /* Because unregister_netdevice_queue() makes sure at least one rcu
115 * grace period is respected before device freeing,
116 * we dont need to call synchronize_net() here.
117 */
118 unregister_netdevice_queue(dev, head);
119
120 if (grp->nr_vlan_devs == 0) {
121 vlan_mvrp_uninit_applicant(real_dev);
122 vlan_gvrp_uninit_applicant(real_dev);
123 }
124
125 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
126}
127
128int vlan_check_real_dev(struct net_device *real_dev,
129 __be16 protocol, u16 vlan_id,
130 struct netlink_ext_ack *extack)
131{
132 const char *name = real_dev->name;
133
134 if (real_dev->features & NETIF_F_VLAN_CHALLENGED ||
135 real_dev->type != ARPHRD_ETHER) {
136 pr_info("VLANs not supported on %s\n", name);
137 NL_SET_ERR_MSG_MOD(extack, "VLANs not supported on device");
138 return -EOPNOTSUPP;
139 }
140
141 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL) {
142 NL_SET_ERR_MSG_MOD(extack, "VLAN device already exists");
143 return -EEXIST;
144 }
145
146 return 0;
147}
148
149int register_vlan_dev(struct net_device *dev, struct netlink_ext_ack *extack)
150{
151 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
152 struct net_device *real_dev = vlan->real_dev;
153 u16 vlan_id = vlan->vlan_id;
154 struct vlan_info *vlan_info;
155 struct vlan_group *grp;
156 int err;
157
158 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
159 if (err)
160 return err;
161
162 vlan_info = rtnl_dereference(real_dev->vlan_info);
163 /* vlan_info should be there now. vlan_vid_add took care of it */
164 BUG_ON(!vlan_info);
165
166 grp = &vlan_info->grp;
167 if (grp->nr_vlan_devs == 0) {
168 err = vlan_gvrp_init_applicant(real_dev);
169 if (err < 0)
170 goto out_vid_del;
171 err = vlan_mvrp_init_applicant(real_dev);
172 if (err < 0)
173 goto out_uninit_gvrp;
174 }
175
176 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
177 if (err < 0)
178 goto out_uninit_mvrp;
179
180 err = register_netdevice(dev);
181 if (err < 0)
182 goto out_uninit_mvrp;
183
184 err = netdev_upper_dev_link(real_dev, dev, extack);
185 if (err)
186 goto out_unregister_netdev;
187
188 vlan_stacked_transfer_operstate(real_dev, dev, vlan);
189 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
190
191 /* So, got the sucker initialized, now lets place
192 * it into our local structure.
193 */
194 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
195 grp->nr_vlan_devs++;
196
197 return 0;
198
199out_unregister_netdev:
200 unregister_netdevice(dev);
201out_uninit_mvrp:
202 if (grp->nr_vlan_devs == 0)
203 vlan_mvrp_uninit_applicant(real_dev);
204out_uninit_gvrp:
205 if (grp->nr_vlan_devs == 0)
206 vlan_gvrp_uninit_applicant(real_dev);
207out_vid_del:
208 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
209 return err;
210}
211
212/* Attach a VLAN device to a mac address (ie Ethernet Card).
213 * Returns 0 if the device was created or a negative error code otherwise.
214 */
215static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
216{
217 struct net_device *new_dev;
218 struct vlan_dev_priv *vlan;
219 struct net *net = dev_net(real_dev);
220 struct vlan_net *vn = net_generic(net, vlan_net_id);
221 char name[IFNAMSIZ];
222 int err;
223
224 if (vlan_id >= VLAN_VID_MASK)
225 return -ERANGE;
226
227 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id,
228 NULL);
229 if (err < 0)
230 return err;
231
232 /* Gotta set up the fields for the device. */
233 switch (vn->name_type) {
234 case VLAN_NAME_TYPE_RAW_PLUS_VID:
235 /* name will look like: eth1.0005 */
236 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
237 break;
238 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
239 /* Put our vlan.VID in the name.
240 * Name will look like: vlan5
241 */
242 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
243 break;
244 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
245 /* Put our vlan.VID in the name.
246 * Name will look like: eth0.5
247 */
248 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
249 break;
250 case VLAN_NAME_TYPE_PLUS_VID:
251 /* Put our vlan.VID in the name.
252 * Name will look like: vlan0005
253 */
254 default:
255 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
256 }
257
258 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
259 NET_NAME_UNKNOWN, vlan_setup);
260
261 if (new_dev == NULL)
262 return -ENOBUFS;
263
264 dev_net_set(new_dev, net);
265 /* need 4 bytes for extra VLAN header info,
266 * hope the underlying device can handle it.
267 */
268 new_dev->mtu = real_dev->mtu;
269
270 vlan = vlan_dev_priv(new_dev);
271 vlan->vlan_proto = htons(ETH_P_8021Q);
272 vlan->vlan_id = vlan_id;
273 vlan->real_dev = real_dev;
274 vlan->dent = NULL;
275 vlan->flags = VLAN_FLAG_REORDER_HDR;
276
277 new_dev->rtnl_link_ops = &vlan_link_ops;
278 err = register_vlan_dev(new_dev, NULL);
279 if (err < 0)
280 goto out_free_newdev;
281
282 return 0;
283
284out_free_newdev:
285 free_netdev(new_dev);
286 return err;
287}
288
289static void vlan_sync_address(struct net_device *dev,
290 struct net_device *vlandev)
291{
292 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
293
294 /* May be called without an actual change */
295 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
296 return;
297
298 /* vlan continues to inherit address of lower device */
299 if (vlan_dev_inherit_address(vlandev, dev))
300 goto out;
301
302 /* vlan address was different from the old address and is equal to
303 * the new address */
304 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
305 ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
306 dev_uc_del(dev, vlandev->dev_addr);
307
308 /* vlan address was equal to the old address and is different from
309 * the new address */
310 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
311 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
312 dev_uc_add(dev, vlandev->dev_addr);
313
314out:
315 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
316}
317
318static void vlan_transfer_features(struct net_device *dev,
319 struct net_device *vlandev)
320{
321 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
322
323 netif_inherit_tso_max(vlandev, dev);
324
325 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
326 vlandev->hard_header_len = dev->hard_header_len;
327 else
328 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
329
330#if IS_ENABLED(CONFIG_FCOE)
331 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
332#endif
333
334 vlandev->priv_flags &= ~IFF_XMIT_DST_RELEASE;
335 vlandev->priv_flags |= (vlan->real_dev->priv_flags & IFF_XMIT_DST_RELEASE);
336 vlandev->hw_enc_features = vlan_tnl_features(vlan->real_dev);
337
338 netdev_update_features(vlandev);
339}
340
341static int __vlan_device_event(struct net_device *dev, unsigned long event)
342{
343 int err = 0;
344
345 switch (event) {
346 case NETDEV_CHANGENAME:
347 vlan_proc_rem_dev(dev);
348 err = vlan_proc_add_dev(dev);
349 break;
350 case NETDEV_REGISTER:
351 err = vlan_proc_add_dev(dev);
352 break;
353 case NETDEV_UNREGISTER:
354 vlan_proc_rem_dev(dev);
355 break;
356 }
357
358 return err;
359}
360
361static int vlan_device_event(struct notifier_block *unused, unsigned long event,
362 void *ptr)
363{
364 struct netlink_ext_ack *extack = netdev_notifier_info_to_extack(ptr);
365 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
366 struct vlan_group *grp;
367 struct vlan_info *vlan_info;
368 int i, flgs;
369 struct net_device *vlandev;
370 struct vlan_dev_priv *vlan;
371 bool last = false;
372 LIST_HEAD(list);
373 int err;
374
375 if (is_vlan_dev(dev)) {
376 int err = __vlan_device_event(dev, event);
377
378 if (err)
379 return notifier_from_errno(err);
380 }
381
382 if ((event == NETDEV_UP) &&
383 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
384 pr_info("adding VLAN 0 to HW filter on device %s\n",
385 dev->name);
386 vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
387 }
388 if (event == NETDEV_DOWN &&
389 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER))
390 vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
391
392 vlan_info = rtnl_dereference(dev->vlan_info);
393 if (!vlan_info)
394 goto out;
395 grp = &vlan_info->grp;
396
397 /* It is OK that we do not hold the group lock right now,
398 * as we run under the RTNL lock.
399 */
400
401 switch (event) {
402 case NETDEV_CHANGE:
403 /* Propagate real device state to vlan devices */
404 vlan_group_for_each_dev(grp, i, vlandev)
405 vlan_stacked_transfer_operstate(dev, vlandev,
406 vlan_dev_priv(vlandev));
407 break;
408
409 case NETDEV_CHANGEADDR:
410 /* Adjust unicast filters on underlying device */
411 vlan_group_for_each_dev(grp, i, vlandev) {
412 flgs = vlandev->flags;
413 if (!(flgs & IFF_UP))
414 continue;
415
416 vlan_sync_address(dev, vlandev);
417 }
418 break;
419
420 case NETDEV_CHANGEMTU:
421 vlan_group_for_each_dev(grp, i, vlandev) {
422 if (vlandev->mtu <= dev->mtu)
423 continue;
424
425 dev_set_mtu(vlandev, dev->mtu);
426 }
427 break;
428
429 case NETDEV_FEAT_CHANGE:
430 /* Propagate device features to underlying device */
431 vlan_group_for_each_dev(grp, i, vlandev)
432 vlan_transfer_features(dev, vlandev);
433 break;
434
435 case NETDEV_DOWN: {
436 struct net_device *tmp;
437 LIST_HEAD(close_list);
438
439 /* Put all VLANs for this dev in the down state too. */
440 vlan_group_for_each_dev(grp, i, vlandev) {
441 flgs = vlandev->flags;
442 if (!(flgs & IFF_UP))
443 continue;
444
445 vlan = vlan_dev_priv(vlandev);
446 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
447 list_add(&vlandev->close_list, &close_list);
448 }
449
450 dev_close_many(&close_list, false);
451
452 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
453 vlan_stacked_transfer_operstate(dev, vlandev,
454 vlan_dev_priv(vlandev));
455 list_del_init(&vlandev->close_list);
456 }
457 list_del(&close_list);
458 break;
459 }
460 case NETDEV_UP:
461 /* Put all VLANs for this dev in the up state too. */
462 vlan_group_for_each_dev(grp, i, vlandev) {
463 flgs = dev_get_flags(vlandev);
464 if (flgs & IFF_UP)
465 continue;
466
467 vlan = vlan_dev_priv(vlandev);
468 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
469 dev_change_flags(vlandev, flgs | IFF_UP,
470 extack);
471 vlan_stacked_transfer_operstate(dev, vlandev, vlan);
472 }
473 break;
474
475 case NETDEV_UNREGISTER:
476 /* twiddle thumbs on netns device moves */
477 if (dev->reg_state != NETREG_UNREGISTERING)
478 break;
479
480 vlan_group_for_each_dev(grp, i, vlandev) {
481 /* removal of last vid destroys vlan_info, abort
482 * afterwards */
483 if (vlan_info->nr_vids == 1)
484 last = true;
485
486 unregister_vlan_dev(vlandev, &list);
487 if (last)
488 break;
489 }
490 unregister_netdevice_many(&list);
491 break;
492
493 case NETDEV_PRE_TYPE_CHANGE:
494 /* Forbid underlaying device to change its type. */
495 if (vlan_uses_dev(dev))
496 return NOTIFY_BAD;
497 break;
498
499 case NETDEV_NOTIFY_PEERS:
500 case NETDEV_BONDING_FAILOVER:
501 case NETDEV_RESEND_IGMP:
502 /* Propagate to vlan devices */
503 vlan_group_for_each_dev(grp, i, vlandev)
504 call_netdevice_notifiers(event, vlandev);
505 break;
506
507 case NETDEV_CVLAN_FILTER_PUSH_INFO:
508 err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021Q));
509 if (err)
510 return notifier_from_errno(err);
511 break;
512
513 case NETDEV_CVLAN_FILTER_DROP_INFO:
514 vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021Q));
515 break;
516
517 case NETDEV_SVLAN_FILTER_PUSH_INFO:
518 err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021AD));
519 if (err)
520 return notifier_from_errno(err);
521 break;
522
523 case NETDEV_SVLAN_FILTER_DROP_INFO:
524 vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021AD));
525 break;
526 }
527
528out:
529 return NOTIFY_DONE;
530}
531
532static struct notifier_block vlan_notifier_block __read_mostly = {
533 .notifier_call = vlan_device_event,
534};
535
536/*
537 * VLAN IOCTL handler.
538 * o execute requested action or pass command to the device driver
539 * arg is really a struct vlan_ioctl_args __user *.
540 */
541static int vlan_ioctl_handler(struct net *net, void __user *arg)
542{
543 int err;
544 struct vlan_ioctl_args args;
545 struct net_device *dev = NULL;
546
547 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
548 return -EFAULT;
549
550 /* Null terminate this sucker, just in case. */
551 args.device1[sizeof(args.device1) - 1] = 0;
552 args.u.device2[sizeof(args.u.device2) - 1] = 0;
553
554 rtnl_lock();
555
556 switch (args.cmd) {
557 case SET_VLAN_INGRESS_PRIORITY_CMD:
558 case SET_VLAN_EGRESS_PRIORITY_CMD:
559 case SET_VLAN_FLAG_CMD:
560 case ADD_VLAN_CMD:
561 case DEL_VLAN_CMD:
562 case GET_VLAN_REALDEV_NAME_CMD:
563 case GET_VLAN_VID_CMD:
564 err = -ENODEV;
565 dev = __dev_get_by_name(net, args.device1);
566 if (!dev)
567 goto out;
568
569 err = -EINVAL;
570 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
571 goto out;
572 }
573
574 switch (args.cmd) {
575 case SET_VLAN_INGRESS_PRIORITY_CMD:
576 err = -EPERM;
577 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
578 break;
579 vlan_dev_set_ingress_priority(dev,
580 args.u.skb_priority,
581 args.vlan_qos);
582 err = 0;
583 break;
584
585 case SET_VLAN_EGRESS_PRIORITY_CMD:
586 err = -EPERM;
587 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
588 break;
589 err = vlan_dev_set_egress_priority(dev,
590 args.u.skb_priority,
591 args.vlan_qos);
592 break;
593
594 case SET_VLAN_FLAG_CMD:
595 err = -EPERM;
596 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
597 break;
598 err = vlan_dev_change_flags(dev,
599 args.vlan_qos ? args.u.flag : 0,
600 args.u.flag);
601 break;
602
603 case SET_VLAN_NAME_TYPE_CMD:
604 err = -EPERM;
605 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
606 break;
607 if (args.u.name_type < VLAN_NAME_TYPE_HIGHEST) {
608 struct vlan_net *vn;
609
610 vn = net_generic(net, vlan_net_id);
611 vn->name_type = args.u.name_type;
612 err = 0;
613 } else {
614 err = -EINVAL;
615 }
616 break;
617
618 case ADD_VLAN_CMD:
619 err = -EPERM;
620 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
621 break;
622 err = register_vlan_device(dev, args.u.VID);
623 break;
624
625 case DEL_VLAN_CMD:
626 err = -EPERM;
627 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
628 break;
629 unregister_vlan_dev(dev, NULL);
630 err = 0;
631 break;
632
633 case GET_VLAN_REALDEV_NAME_CMD:
634 err = 0;
635 vlan_dev_get_realdev_name(dev, args.u.device2,
636 sizeof(args.u.device2));
637 if (copy_to_user(arg, &args,
638 sizeof(struct vlan_ioctl_args)))
639 err = -EFAULT;
640 break;
641
642 case GET_VLAN_VID_CMD:
643 err = 0;
644 args.u.VID = vlan_dev_vlan_id(dev);
645 if (copy_to_user(arg, &args,
646 sizeof(struct vlan_ioctl_args)))
647 err = -EFAULT;
648 break;
649
650 default:
651 err = -EOPNOTSUPP;
652 break;
653 }
654out:
655 rtnl_unlock();
656 return err;
657}
658
659static int __net_init vlan_init_net(struct net *net)
660{
661 struct vlan_net *vn = net_generic(net, vlan_net_id);
662 int err;
663
664 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
665
666 err = vlan_proc_init(net);
667
668 return err;
669}
670
671static void __net_exit vlan_exit_net(struct net *net)
672{
673 vlan_proc_cleanup(net);
674}
675
676static struct pernet_operations vlan_net_ops = {
677 .init = vlan_init_net,
678 .exit = vlan_exit_net,
679 .id = &vlan_net_id,
680 .size = sizeof(struct vlan_net),
681};
682
683static int __init vlan_proto_init(void)
684{
685 int err;
686
687 pr_info("%s v%s\n", vlan_fullname, vlan_version);
688
689 err = register_pernet_subsys(&vlan_net_ops);
690 if (err < 0)
691 goto err0;
692
693 err = register_netdevice_notifier(&vlan_notifier_block);
694 if (err < 0)
695 goto err2;
696
697 err = vlan_gvrp_init();
698 if (err < 0)
699 goto err3;
700
701 err = vlan_mvrp_init();
702 if (err < 0)
703 goto err4;
704
705 err = vlan_netlink_init();
706 if (err < 0)
707 goto err5;
708
709 vlan_ioctl_set(vlan_ioctl_handler);
710 return 0;
711
712err5:
713 vlan_mvrp_uninit();
714err4:
715 vlan_gvrp_uninit();
716err3:
717 unregister_netdevice_notifier(&vlan_notifier_block);
718err2:
719 unregister_pernet_subsys(&vlan_net_ops);
720err0:
721 return err;
722}
723
724static void __exit vlan_cleanup_module(void)
725{
726 vlan_ioctl_set(NULL);
727
728 vlan_netlink_fini();
729
730 unregister_netdevice_notifier(&vlan_notifier_block);
731
732 unregister_pernet_subsys(&vlan_net_ops);
733 rcu_barrier(); /* Wait for completion of call_rcu()'s */
734
735 vlan_mvrp_uninit();
736 vlan_gvrp_uninit();
737}
738
739module_init(vlan_proto_init);
740module_exit(vlan_cleanup_module);
741
742MODULE_DESCRIPTION("802.1Q/802.1ad VLAN Protocol");
743MODULE_LICENSE("GPL");
744MODULE_VERSION(DRV_VERSION);
1/*
2 * INET 802.1Q VLAN
3 * Ethernet-type device handling.
4 *
5 * Authors: Ben Greear <greearb@candelatech.com>
6 * Please send support related email to: netdev@vger.kernel.org
7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
8 *
9 * Fixes:
10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
12 * Correct all the locking - David S. Miller <davem@redhat.com>;
13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
14 *
15 * This program is free software; you can redistribute it and/or
16 * modify it under the terms of the GNU General Public License
17 * as published by the Free Software Foundation; either version
18 * 2 of the License, or (at your option) any later version.
19 */
20
21#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
22
23#include <linux/capability.h>
24#include <linux/module.h>
25#include <linux/netdevice.h>
26#include <linux/skbuff.h>
27#include <linux/slab.h>
28#include <linux/init.h>
29#include <linux/rculist.h>
30#include <net/p8022.h>
31#include <net/arp.h>
32#include <linux/rtnetlink.h>
33#include <linux/notifier.h>
34#include <net/rtnetlink.h>
35#include <net/net_namespace.h>
36#include <net/netns/generic.h>
37#include <asm/uaccess.h>
38
39#include <linux/if_vlan.h>
40#include "vlan.h"
41#include "vlanproc.h"
42
43#define DRV_VERSION "1.8"
44
45/* Global VLAN variables */
46
47int vlan_net_id __read_mostly;
48
49const char vlan_fullname[] = "802.1Q VLAN Support";
50const char vlan_version[] = DRV_VERSION;
51
52/* End of global variables definitions. */
53
54static int vlan_group_prealloc_vid(struct vlan_group *vg,
55 __be16 vlan_proto, u16 vlan_id)
56{
57 struct net_device **array;
58 unsigned int pidx, vidx;
59 unsigned int size;
60
61 ASSERT_RTNL();
62
63 pidx = vlan_proto_idx(vlan_proto);
64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65 array = vg->vlan_devices_arrays[pidx][vidx];
66 if (array != NULL)
67 return 0;
68
69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70 array = kzalloc(size, GFP_KERNEL);
71 if (array == NULL)
72 return -ENOBUFS;
73
74 vg->vlan_devices_arrays[pidx][vidx] = array;
75 return 0;
76}
77
78void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
79{
80 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
81 struct net_device *real_dev = vlan->real_dev;
82 struct vlan_info *vlan_info;
83 struct vlan_group *grp;
84 u16 vlan_id = vlan->vlan_id;
85
86 ASSERT_RTNL();
87
88 vlan_info = rtnl_dereference(real_dev->vlan_info);
89 BUG_ON(!vlan_info);
90
91 grp = &vlan_info->grp;
92
93 grp->nr_vlan_devs--;
94
95 if (vlan->flags & VLAN_FLAG_MVRP)
96 vlan_mvrp_request_leave(dev);
97 if (vlan->flags & VLAN_FLAG_GVRP)
98 vlan_gvrp_request_leave(dev);
99
100 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
101
102 netdev_upper_dev_unlink(real_dev, dev);
103 /* Because unregister_netdevice_queue() makes sure at least one rcu
104 * grace period is respected before device freeing,
105 * we dont need to call synchronize_net() here.
106 */
107 unregister_netdevice_queue(dev, head);
108
109 if (grp->nr_vlan_devs == 0) {
110 vlan_mvrp_uninit_applicant(real_dev);
111 vlan_gvrp_uninit_applicant(real_dev);
112 }
113
114 /* Take it out of our own structures, but be sure to interlock with
115 * HW accelerating devices or SW vlan input packet processing if
116 * VLAN is not 0 (leave it there for 802.1p).
117 */
118 if (vlan_id)
119 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
120
121 /* Get rid of the vlan's reference to real_dev */
122 dev_put(real_dev);
123}
124
125int vlan_check_real_dev(struct net_device *real_dev,
126 __be16 protocol, u16 vlan_id)
127{
128 const char *name = real_dev->name;
129
130 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) {
131 pr_info("VLANs not supported on %s\n", name);
132 return -EOPNOTSUPP;
133 }
134
135 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL)
136 return -EEXIST;
137
138 return 0;
139}
140
141int register_vlan_dev(struct net_device *dev)
142{
143 struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
144 struct net_device *real_dev = vlan->real_dev;
145 u16 vlan_id = vlan->vlan_id;
146 struct vlan_info *vlan_info;
147 struct vlan_group *grp;
148 int err;
149
150 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
151 if (err)
152 return err;
153
154 vlan_info = rtnl_dereference(real_dev->vlan_info);
155 /* vlan_info should be there now. vlan_vid_add took care of it */
156 BUG_ON(!vlan_info);
157
158 grp = &vlan_info->grp;
159 if (grp->nr_vlan_devs == 0) {
160 err = vlan_gvrp_init_applicant(real_dev);
161 if (err < 0)
162 goto out_vid_del;
163 err = vlan_mvrp_init_applicant(real_dev);
164 if (err < 0)
165 goto out_uninit_gvrp;
166 }
167
168 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
169 if (err < 0)
170 goto out_uninit_mvrp;
171
172 vlan->nest_level = dev_get_nest_level(real_dev, is_vlan_dev) + 1;
173 err = register_netdevice(dev);
174 if (err < 0)
175 goto out_uninit_mvrp;
176
177 err = netdev_upper_dev_link(real_dev, dev);
178 if (err)
179 goto out_unregister_netdev;
180
181 /* Account for reference in struct vlan_dev_priv */
182 dev_hold(real_dev);
183
184 netif_stacked_transfer_operstate(real_dev, dev);
185 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
186
187 /* So, got the sucker initialized, now lets place
188 * it into our local structure.
189 */
190 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
191 grp->nr_vlan_devs++;
192
193 return 0;
194
195out_unregister_netdev:
196 unregister_netdevice(dev);
197out_uninit_mvrp:
198 if (grp->nr_vlan_devs == 0)
199 vlan_mvrp_uninit_applicant(real_dev);
200out_uninit_gvrp:
201 if (grp->nr_vlan_devs == 0)
202 vlan_gvrp_uninit_applicant(real_dev);
203out_vid_del:
204 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
205 return err;
206}
207
208/* Attach a VLAN device to a mac address (ie Ethernet Card).
209 * Returns 0 if the device was created or a negative error code otherwise.
210 */
211static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
212{
213 struct net_device *new_dev;
214 struct vlan_dev_priv *vlan;
215 struct net *net = dev_net(real_dev);
216 struct vlan_net *vn = net_generic(net, vlan_net_id);
217 char name[IFNAMSIZ];
218 int err;
219
220 if (vlan_id >= VLAN_VID_MASK)
221 return -ERANGE;
222
223 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id);
224 if (err < 0)
225 return err;
226
227 /* Gotta set up the fields for the device. */
228 switch (vn->name_type) {
229 case VLAN_NAME_TYPE_RAW_PLUS_VID:
230 /* name will look like: eth1.0005 */
231 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
232 break;
233 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
234 /* Put our vlan.VID in the name.
235 * Name will look like: vlan5
236 */
237 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
238 break;
239 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
240 /* Put our vlan.VID in the name.
241 * Name will look like: eth0.5
242 */
243 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
244 break;
245 case VLAN_NAME_TYPE_PLUS_VID:
246 /* Put our vlan.VID in the name.
247 * Name will look like: vlan0005
248 */
249 default:
250 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
251 }
252
253 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
254 NET_NAME_UNKNOWN, vlan_setup);
255
256 if (new_dev == NULL)
257 return -ENOBUFS;
258
259 dev_net_set(new_dev, net);
260 /* need 4 bytes for extra VLAN header info,
261 * hope the underlying device can handle it.
262 */
263 new_dev->mtu = real_dev->mtu;
264
265 vlan = vlan_dev_priv(new_dev);
266 vlan->vlan_proto = htons(ETH_P_8021Q);
267 vlan->vlan_id = vlan_id;
268 vlan->real_dev = real_dev;
269 vlan->dent = NULL;
270 vlan->flags = VLAN_FLAG_REORDER_HDR;
271
272 new_dev->rtnl_link_ops = &vlan_link_ops;
273 err = register_vlan_dev(new_dev);
274 if (err < 0)
275 goto out_free_newdev;
276
277 return 0;
278
279out_free_newdev:
280 free_netdev(new_dev);
281 return err;
282}
283
284static void vlan_sync_address(struct net_device *dev,
285 struct net_device *vlandev)
286{
287 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
288
289 /* May be called without an actual change */
290 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
291 return;
292
293 /* vlan address was different from the old address and is equal to
294 * the new address */
295 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
296 ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
297 dev_uc_del(dev, vlandev->dev_addr);
298
299 /* vlan address was equal to the old address and is different from
300 * the new address */
301 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
302 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
303 dev_uc_add(dev, vlandev->dev_addr);
304
305 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
306}
307
308static void vlan_transfer_features(struct net_device *dev,
309 struct net_device *vlandev)
310{
311 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
312
313 vlandev->gso_max_size = dev->gso_max_size;
314 vlandev->gso_max_segs = dev->gso_max_segs;
315
316 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
317 vlandev->hard_header_len = dev->hard_header_len;
318 else
319 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
320
321#if IS_ENABLED(CONFIG_FCOE)
322 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
323#endif
324
325 netdev_update_features(vlandev);
326}
327
328static int __vlan_device_event(struct net_device *dev, unsigned long event)
329{
330 int err = 0;
331
332 switch (event) {
333 case NETDEV_CHANGENAME:
334 vlan_proc_rem_dev(dev);
335 err = vlan_proc_add_dev(dev);
336 break;
337 case NETDEV_REGISTER:
338 err = vlan_proc_add_dev(dev);
339 break;
340 case NETDEV_UNREGISTER:
341 vlan_proc_rem_dev(dev);
342 break;
343 }
344
345 return err;
346}
347
348static int vlan_device_event(struct notifier_block *unused, unsigned long event,
349 void *ptr)
350{
351 struct net_device *dev = netdev_notifier_info_to_dev(ptr);
352 struct vlan_group *grp;
353 struct vlan_info *vlan_info;
354 int i, flgs;
355 struct net_device *vlandev;
356 struct vlan_dev_priv *vlan;
357 bool last = false;
358 LIST_HEAD(list);
359
360 if (is_vlan_dev(dev)) {
361 int err = __vlan_device_event(dev, event);
362
363 if (err)
364 return notifier_from_errno(err);
365 }
366
367 if ((event == NETDEV_UP) &&
368 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
369 pr_info("adding VLAN 0 to HW filter on device %s\n",
370 dev->name);
371 vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
372 }
373
374 vlan_info = rtnl_dereference(dev->vlan_info);
375 if (!vlan_info)
376 goto out;
377 grp = &vlan_info->grp;
378
379 /* It is OK that we do not hold the group lock right now,
380 * as we run under the RTNL lock.
381 */
382
383 switch (event) {
384 case NETDEV_CHANGE:
385 /* Propagate real device state to vlan devices */
386 vlan_group_for_each_dev(grp, i, vlandev)
387 netif_stacked_transfer_operstate(dev, vlandev);
388 break;
389
390 case NETDEV_CHANGEADDR:
391 /* Adjust unicast filters on underlying device */
392 vlan_group_for_each_dev(grp, i, vlandev) {
393 flgs = vlandev->flags;
394 if (!(flgs & IFF_UP))
395 continue;
396
397 vlan_sync_address(dev, vlandev);
398 }
399 break;
400
401 case NETDEV_CHANGEMTU:
402 vlan_group_for_each_dev(grp, i, vlandev) {
403 if (vlandev->mtu <= dev->mtu)
404 continue;
405
406 dev_set_mtu(vlandev, dev->mtu);
407 }
408 break;
409
410 case NETDEV_FEAT_CHANGE:
411 /* Propagate device features to underlying device */
412 vlan_group_for_each_dev(grp, i, vlandev)
413 vlan_transfer_features(dev, vlandev);
414 break;
415
416 case NETDEV_DOWN: {
417 struct net_device *tmp;
418 LIST_HEAD(close_list);
419
420 if (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)
421 vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
422
423 /* Put all VLANs for this dev in the down state too. */
424 vlan_group_for_each_dev(grp, i, vlandev) {
425 flgs = vlandev->flags;
426 if (!(flgs & IFF_UP))
427 continue;
428
429 vlan = vlan_dev_priv(vlandev);
430 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
431 list_add(&vlandev->close_list, &close_list);
432 }
433
434 dev_close_many(&close_list, false);
435
436 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
437 netif_stacked_transfer_operstate(dev, vlandev);
438 list_del_init(&vlandev->close_list);
439 }
440 list_del(&close_list);
441 break;
442 }
443 case NETDEV_UP:
444 /* Put all VLANs for this dev in the up state too. */
445 vlan_group_for_each_dev(grp, i, vlandev) {
446 flgs = dev_get_flags(vlandev);
447 if (flgs & IFF_UP)
448 continue;
449
450 vlan = vlan_dev_priv(vlandev);
451 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
452 dev_change_flags(vlandev, flgs | IFF_UP);
453 netif_stacked_transfer_operstate(dev, vlandev);
454 }
455 break;
456
457 case NETDEV_UNREGISTER:
458 /* twiddle thumbs on netns device moves */
459 if (dev->reg_state != NETREG_UNREGISTERING)
460 break;
461
462 vlan_group_for_each_dev(grp, i, vlandev) {
463 /* removal of last vid destroys vlan_info, abort
464 * afterwards */
465 if (vlan_info->nr_vids == 1)
466 last = true;
467
468 unregister_vlan_dev(vlandev, &list);
469 if (last)
470 break;
471 }
472 unregister_netdevice_many(&list);
473 break;
474
475 case NETDEV_PRE_TYPE_CHANGE:
476 /* Forbid underlaying device to change its type. */
477 if (vlan_uses_dev(dev))
478 return NOTIFY_BAD;
479 break;
480
481 case NETDEV_NOTIFY_PEERS:
482 case NETDEV_BONDING_FAILOVER:
483 case NETDEV_RESEND_IGMP:
484 /* Propagate to vlan devices */
485 vlan_group_for_each_dev(grp, i, vlandev)
486 call_netdevice_notifiers(event, vlandev);
487 break;
488 }
489
490out:
491 return NOTIFY_DONE;
492}
493
494static struct notifier_block vlan_notifier_block __read_mostly = {
495 .notifier_call = vlan_device_event,
496};
497
498/*
499 * VLAN IOCTL handler.
500 * o execute requested action or pass command to the device driver
501 * arg is really a struct vlan_ioctl_args __user *.
502 */
503static int vlan_ioctl_handler(struct net *net, void __user *arg)
504{
505 int err;
506 struct vlan_ioctl_args args;
507 struct net_device *dev = NULL;
508
509 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
510 return -EFAULT;
511
512 /* Null terminate this sucker, just in case. */
513 args.device1[23] = 0;
514 args.u.device2[23] = 0;
515
516 rtnl_lock();
517
518 switch (args.cmd) {
519 case SET_VLAN_INGRESS_PRIORITY_CMD:
520 case SET_VLAN_EGRESS_PRIORITY_CMD:
521 case SET_VLAN_FLAG_CMD:
522 case ADD_VLAN_CMD:
523 case DEL_VLAN_CMD:
524 case GET_VLAN_REALDEV_NAME_CMD:
525 case GET_VLAN_VID_CMD:
526 err = -ENODEV;
527 dev = __dev_get_by_name(net, args.device1);
528 if (!dev)
529 goto out;
530
531 err = -EINVAL;
532 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
533 goto out;
534 }
535
536 switch (args.cmd) {
537 case SET_VLAN_INGRESS_PRIORITY_CMD:
538 err = -EPERM;
539 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
540 break;
541 vlan_dev_set_ingress_priority(dev,
542 args.u.skb_priority,
543 args.vlan_qos);
544 err = 0;
545 break;
546
547 case SET_VLAN_EGRESS_PRIORITY_CMD:
548 err = -EPERM;
549 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
550 break;
551 err = vlan_dev_set_egress_priority(dev,
552 args.u.skb_priority,
553 args.vlan_qos);
554 break;
555
556 case SET_VLAN_FLAG_CMD:
557 err = -EPERM;
558 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
559 break;
560 err = vlan_dev_change_flags(dev,
561 args.vlan_qos ? args.u.flag : 0,
562 args.u.flag);
563 break;
564
565 case SET_VLAN_NAME_TYPE_CMD:
566 err = -EPERM;
567 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
568 break;
569 if ((args.u.name_type >= 0) &&
570 (args.u.name_type < VLAN_NAME_TYPE_HIGHEST)) {
571 struct vlan_net *vn;
572
573 vn = net_generic(net, vlan_net_id);
574 vn->name_type = args.u.name_type;
575 err = 0;
576 } else {
577 err = -EINVAL;
578 }
579 break;
580
581 case ADD_VLAN_CMD:
582 err = -EPERM;
583 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
584 break;
585 err = register_vlan_device(dev, args.u.VID);
586 break;
587
588 case DEL_VLAN_CMD:
589 err = -EPERM;
590 if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
591 break;
592 unregister_vlan_dev(dev, NULL);
593 err = 0;
594 break;
595
596 case GET_VLAN_REALDEV_NAME_CMD:
597 err = 0;
598 vlan_dev_get_realdev_name(dev, args.u.device2);
599 if (copy_to_user(arg, &args,
600 sizeof(struct vlan_ioctl_args)))
601 err = -EFAULT;
602 break;
603
604 case GET_VLAN_VID_CMD:
605 err = 0;
606 args.u.VID = vlan_dev_vlan_id(dev);
607 if (copy_to_user(arg, &args,
608 sizeof(struct vlan_ioctl_args)))
609 err = -EFAULT;
610 break;
611
612 default:
613 err = -EOPNOTSUPP;
614 break;
615 }
616out:
617 rtnl_unlock();
618 return err;
619}
620
621static struct sk_buff **vlan_gro_receive(struct sk_buff **head,
622 struct sk_buff *skb)
623{
624 struct sk_buff *p, **pp = NULL;
625 struct vlan_hdr *vhdr;
626 unsigned int hlen, off_vlan;
627 const struct packet_offload *ptype;
628 __be16 type;
629 int flush = 1;
630
631 off_vlan = skb_gro_offset(skb);
632 hlen = off_vlan + sizeof(*vhdr);
633 vhdr = skb_gro_header_fast(skb, off_vlan);
634 if (skb_gro_header_hard(skb, hlen)) {
635 vhdr = skb_gro_header_slow(skb, hlen, off_vlan);
636 if (unlikely(!vhdr))
637 goto out;
638 }
639
640 type = vhdr->h_vlan_encapsulated_proto;
641
642 rcu_read_lock();
643 ptype = gro_find_receive_by_type(type);
644 if (!ptype)
645 goto out_unlock;
646
647 flush = 0;
648
649 for (p = *head; p; p = p->next) {
650 struct vlan_hdr *vhdr2;
651
652 if (!NAPI_GRO_CB(p)->same_flow)
653 continue;
654
655 vhdr2 = (struct vlan_hdr *)(p->data + off_vlan);
656 if (compare_vlan_header(vhdr, vhdr2))
657 NAPI_GRO_CB(p)->same_flow = 0;
658 }
659
660 skb_gro_pull(skb, sizeof(*vhdr));
661 skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr));
662 pp = ptype->callbacks.gro_receive(head, skb);
663
664out_unlock:
665 rcu_read_unlock();
666out:
667 NAPI_GRO_CB(skb)->flush |= flush;
668
669 return pp;
670}
671
672static int vlan_gro_complete(struct sk_buff *skb, int nhoff)
673{
674 struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff);
675 __be16 type = vhdr->h_vlan_encapsulated_proto;
676 struct packet_offload *ptype;
677 int err = -ENOENT;
678
679 rcu_read_lock();
680 ptype = gro_find_complete_by_type(type);
681 if (ptype)
682 err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr));
683
684 rcu_read_unlock();
685 return err;
686}
687
688static struct packet_offload vlan_packet_offloads[] __read_mostly = {
689 {
690 .type = cpu_to_be16(ETH_P_8021Q),
691 .priority = 10,
692 .callbacks = {
693 .gro_receive = vlan_gro_receive,
694 .gro_complete = vlan_gro_complete,
695 },
696 },
697 {
698 .type = cpu_to_be16(ETH_P_8021AD),
699 .priority = 10,
700 .callbacks = {
701 .gro_receive = vlan_gro_receive,
702 .gro_complete = vlan_gro_complete,
703 },
704 },
705};
706
707static int __net_init vlan_init_net(struct net *net)
708{
709 struct vlan_net *vn = net_generic(net, vlan_net_id);
710 int err;
711
712 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
713
714 err = vlan_proc_init(net);
715
716 return err;
717}
718
719static void __net_exit vlan_exit_net(struct net *net)
720{
721 vlan_proc_cleanup(net);
722}
723
724static struct pernet_operations vlan_net_ops = {
725 .init = vlan_init_net,
726 .exit = vlan_exit_net,
727 .id = &vlan_net_id,
728 .size = sizeof(struct vlan_net),
729};
730
731static int __init vlan_proto_init(void)
732{
733 int err;
734 unsigned int i;
735
736 pr_info("%s v%s\n", vlan_fullname, vlan_version);
737
738 err = register_pernet_subsys(&vlan_net_ops);
739 if (err < 0)
740 goto err0;
741
742 err = register_netdevice_notifier(&vlan_notifier_block);
743 if (err < 0)
744 goto err2;
745
746 err = vlan_gvrp_init();
747 if (err < 0)
748 goto err3;
749
750 err = vlan_mvrp_init();
751 if (err < 0)
752 goto err4;
753
754 err = vlan_netlink_init();
755 if (err < 0)
756 goto err5;
757
758 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
759 dev_add_offload(&vlan_packet_offloads[i]);
760
761 vlan_ioctl_set(vlan_ioctl_handler);
762 return 0;
763
764err5:
765 vlan_mvrp_uninit();
766err4:
767 vlan_gvrp_uninit();
768err3:
769 unregister_netdevice_notifier(&vlan_notifier_block);
770err2:
771 unregister_pernet_subsys(&vlan_net_ops);
772err0:
773 return err;
774}
775
776static void __exit vlan_cleanup_module(void)
777{
778 unsigned int i;
779
780 vlan_ioctl_set(NULL);
781
782 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
783 dev_remove_offload(&vlan_packet_offloads[i]);
784
785 vlan_netlink_fini();
786
787 unregister_netdevice_notifier(&vlan_notifier_block);
788
789 unregister_pernet_subsys(&vlan_net_ops);
790 rcu_barrier(); /* Wait for completion of call_rcu()'s */
791
792 vlan_mvrp_uninit();
793 vlan_gvrp_uninit();
794}
795
796module_init(vlan_proto_init);
797module_exit(vlan_cleanup_module);
798
799MODULE_LICENSE("GPL");
800MODULE_VERSION(DRV_VERSION);