Linux Audio

Check our new training course

Loading...
v4.17
 
   1#include <linux/kernel.h>
   2#include <linux/netdevice.h>
   3#include <linux/rtnetlink.h>
   4#include <linux/slab.h>
   5#include <net/switchdev.h>
   6
   7#include "br_private.h"
   8#include "br_private_tunnel.h"
   9
 
 
  10static inline int br_vlan_cmp(struct rhashtable_compare_arg *arg,
  11			      const void *ptr)
  12{
  13	const struct net_bridge_vlan *vle = ptr;
  14	u16 vid = *(u16 *)arg->key;
  15
  16	return vle->vid != vid;
  17}
  18
  19static const struct rhashtable_params br_vlan_rht_params = {
  20	.head_offset = offsetof(struct net_bridge_vlan, vnode),
  21	.key_offset = offsetof(struct net_bridge_vlan, vid),
  22	.key_len = sizeof(u16),
  23	.nelem_hint = 3,
  24	.locks_mul = 1,
  25	.max_size = VLAN_N_VID,
  26	.obj_cmpfn = br_vlan_cmp,
  27	.automatic_shrinking = true,
  28};
  29
  30static struct net_bridge_vlan *br_vlan_lookup(struct rhashtable *tbl, u16 vid)
  31{
  32	return rhashtable_lookup_fast(tbl, &vid, br_vlan_rht_params);
  33}
  34
  35static bool __vlan_add_pvid(struct net_bridge_vlan_group *vg, u16 vid)
  36{
  37	if (vg->pvid == vid)
  38		return false;
  39
  40	smp_wmb();
  41	vg->pvid = vid;
  42
  43	return true;
  44}
  45
  46static bool __vlan_delete_pvid(struct net_bridge_vlan_group *vg, u16 vid)
  47{
  48	if (vg->pvid != vid)
  49		return false;
  50
  51	smp_wmb();
  52	vg->pvid = 0;
  53
  54	return true;
  55}
  56
  57/* return true if anything changed, false otherwise */
  58static bool __vlan_add_flags(struct net_bridge_vlan *v, u16 flags)
  59{
  60	struct net_bridge_vlan_group *vg;
  61	u16 old_flags = v->flags;
  62	bool ret;
  63
  64	if (br_vlan_is_master(v))
  65		vg = br_vlan_group(v->br);
  66	else
  67		vg = nbp_vlan_group(v->port);
  68
  69	if (flags & BRIDGE_VLAN_INFO_PVID)
  70		ret = __vlan_add_pvid(vg, v->vid);
  71	else
  72		ret = __vlan_delete_pvid(vg, v->vid);
  73
  74	if (flags & BRIDGE_VLAN_INFO_UNTAGGED)
  75		v->flags |= BRIDGE_VLAN_INFO_UNTAGGED;
  76	else
  77		v->flags &= ~BRIDGE_VLAN_INFO_UNTAGGED;
  78
  79	return ret || !!(old_flags ^ v->flags);
  80}
  81
  82static int __vlan_vid_add(struct net_device *dev, struct net_bridge *br,
  83			  u16 vid, u16 flags)
 
  84{
  85	struct switchdev_obj_port_vlan v = {
  86		.obj.orig_dev = dev,
  87		.obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
  88		.flags = flags,
  89		.vid_begin = vid,
  90		.vid_end = vid,
  91	};
  92	int err;
  93
  94	/* Try switchdev op first. In case it is not supported, fallback to
  95	 * 8021q add.
  96	 */
  97	err = switchdev_port_obj_add(dev, &v.obj);
  98	if (err == -EOPNOTSUPP)
  99		return vlan_vid_add(dev, br->vlan_proto, vid);
 
 100	return err;
 101}
 102
 103static void __vlan_add_list(struct net_bridge_vlan *v)
 104{
 105	struct net_bridge_vlan_group *vg;
 106	struct list_head *headp, *hpos;
 107	struct net_bridge_vlan *vent;
 108
 109	if (br_vlan_is_master(v))
 110		vg = br_vlan_group(v->br);
 111	else
 112		vg = nbp_vlan_group(v->port);
 113
 114	headp = &vg->vlan_list;
 115	list_for_each_prev(hpos, headp) {
 116		vent = list_entry(hpos, struct net_bridge_vlan, vlist);
 117		if (v->vid < vent->vid)
 118			continue;
 119		else
 120			break;
 121	}
 122	list_add_rcu(&v->vlist, hpos);
 123}
 124
 125static void __vlan_del_list(struct net_bridge_vlan *v)
 126{
 127	list_del_rcu(&v->vlist);
 128}
 129
 130static int __vlan_vid_del(struct net_device *dev, struct net_bridge *br,
 131			  u16 vid)
 132{
 133	struct switchdev_obj_port_vlan v = {
 134		.obj.orig_dev = dev,
 135		.obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
 136		.vid_begin = vid,
 137		.vid_end = vid,
 138	};
 139	int err;
 140
 141	/* Try switchdev op first. In case it is not supported, fallback to
 142	 * 8021q del.
 143	 */
 144	err = switchdev_port_obj_del(dev, &v.obj);
 145	if (err == -EOPNOTSUPP) {
 146		vlan_vid_del(dev, br->vlan_proto, vid);
 147		return 0;
 148	}
 149	return err;
 150}
 151
 152/* Returns a master vlan, if it didn't exist it gets created. In all cases a
 153 * a reference is taken to the master vlan before returning.
 154 */
 155static struct net_bridge_vlan *br_vlan_get_master(struct net_bridge *br, u16 vid)
 
 
 156{
 157	struct net_bridge_vlan_group *vg;
 158	struct net_bridge_vlan *masterv;
 159
 160	vg = br_vlan_group(br);
 161	masterv = br_vlan_find(vg, vid);
 162	if (!masterv) {
 163		bool changed;
 164
 165		/* missing global ctx, create it now */
 166		if (br_vlan_add(br, vid, 0, &changed))
 167			return NULL;
 168		masterv = br_vlan_find(vg, vid);
 169		if (WARN_ON(!masterv))
 170			return NULL;
 171		refcount_set(&masterv->refcnt, 1);
 172		return masterv;
 173	}
 174	refcount_inc(&masterv->refcnt);
 175
 176	return masterv;
 177}
 178
 179static void br_master_vlan_rcu_free(struct rcu_head *rcu)
 180{
 181	struct net_bridge_vlan *v;
 182
 183	v = container_of(rcu, struct net_bridge_vlan, rcu);
 184	WARN_ON(!br_vlan_is_master(v));
 185	free_percpu(v->stats);
 186	v->stats = NULL;
 187	kfree(v);
 188}
 189
 190static void br_vlan_put_master(struct net_bridge_vlan *masterv)
 191{
 192	struct net_bridge_vlan_group *vg;
 193
 194	if (!br_vlan_is_master(masterv))
 195		return;
 196
 197	vg = br_vlan_group(masterv->br);
 198	if (refcount_dec_and_test(&masterv->refcnt)) {
 199		rhashtable_remove_fast(&vg->vlan_hash,
 200				       &masterv->vnode, br_vlan_rht_params);
 201		__vlan_del_list(masterv);
 202		call_rcu(&masterv->rcu, br_master_vlan_rcu_free);
 203	}
 204}
 205
 
 
 
 
 
 
 
 
 
 
 
 
 
 206/* This is the shared VLAN add function which works for both ports and bridge
 207 * devices. There are four possible calls to this function in terms of the
 208 * vlan entry type:
 209 * 1. vlan is being added on a port (no master flags, global entry exists)
 210 * 2. vlan is being added on a bridge (both master and brentry flags)
 211 * 3. vlan is being added on a port, but a global entry didn't exist which
 212 *    is being created right now (master flag set, brentry flag unset), the
 213 *    global entry is used for global per-vlan features, but not for filtering
 214 * 4. same as 3 but with both master and brentry flags set so the entry
 215 *    will be used for filtering in both the port and the bridge
 216 */
 217static int __vlan_add(struct net_bridge_vlan *v, u16 flags)
 
 218{
 219	struct net_bridge_vlan *masterv = NULL;
 220	struct net_bridge_port *p = NULL;
 221	struct net_bridge_vlan_group *vg;
 222	struct net_device *dev;
 223	struct net_bridge *br;
 224	int err;
 225
 226	if (br_vlan_is_master(v)) {
 227		br = v->br;
 228		dev = br->dev;
 229		vg = br_vlan_group(br);
 230	} else {
 231		p = v->port;
 232		br = p->br;
 233		dev = p->dev;
 234		vg = nbp_vlan_group(p);
 235	}
 236
 237	if (p) {
 238		/* Add VLAN to the device filter if it is supported.
 239		 * This ensures tagged traffic enters the bridge when
 240		 * promiscuous mode is disabled by br_manage_promisc().
 241		 */
 242		err = __vlan_vid_add(dev, br, v->vid, flags);
 243		if (err)
 244			goto out;
 245
 246		/* need to work on the master vlan too */
 247		if (flags & BRIDGE_VLAN_INFO_MASTER) {
 248			bool changed;
 249
 250			err = br_vlan_add(br, v->vid,
 251					  flags | BRIDGE_VLAN_INFO_BRENTRY,
 252					  &changed);
 253			if (err)
 254				goto out_filt;
 255		}
 256
 257		masterv = br_vlan_get_master(br, v->vid);
 258		if (!masterv)
 259			goto out_filt;
 260		v->brvlan = masterv;
 261		v->stats = masterv->stats;
 
 
 
 
 
 
 
 
 
 
 
 
 
 262	}
 263
 264	/* Add the dev mac and count the vlan only if it's usable */
 265	if (br_vlan_should_use(v)) {
 266		err = br_fdb_insert(br, p, dev->dev_addr, v->vid);
 267		if (err) {
 268			br_err(br, "failed insert local address into bridge forwarding table\n");
 269			goto out_filt;
 270		}
 271		vg->num_vlans++;
 272	}
 273
 274	err = rhashtable_lookup_insert_fast(&vg->vlan_hash, &v->vnode,
 275					    br_vlan_rht_params);
 276	if (err)
 277		goto out_fdb_insert;
 278
 279	__vlan_add_list(v);
 280	__vlan_add_flags(v, flags);
 
 
 
 281out:
 282	return err;
 283
 284out_fdb_insert:
 285	if (br_vlan_should_use(v)) {
 286		br_fdb_find_delete_local(br, p, dev->dev_addr, v->vid);
 287		vg->num_vlans--;
 288	}
 289
 290out_filt:
 291	if (p) {
 292		__vlan_vid_del(dev, br, v->vid);
 293		if (masterv) {
 
 
 
 
 294			br_vlan_put_master(masterv);
 295			v->brvlan = NULL;
 296		}
 
 
 297	}
 298
 299	goto out;
 300}
 301
 302static int __vlan_del(struct net_bridge_vlan *v)
 303{
 304	struct net_bridge_vlan *masterv = v;
 305	struct net_bridge_vlan_group *vg;
 306	struct net_bridge_port *p = NULL;
 307	int err = 0;
 308
 309	if (br_vlan_is_master(v)) {
 310		vg = br_vlan_group(v->br);
 311	} else {
 312		p = v->port;
 313		vg = nbp_vlan_group(v->port);
 314		masterv = v->brvlan;
 315	}
 316
 317	__vlan_delete_pvid(vg, v->vid);
 318	if (p) {
 319		err = __vlan_vid_del(p->dev, p->br, v->vid);
 320		if (err)
 321			goto out;
 
 
 
 
 
 322	}
 323
 324	if (br_vlan_should_use(v)) {
 325		v->flags &= ~BRIDGE_VLAN_INFO_BRENTRY;
 326		vg->num_vlans--;
 327	}
 328
 329	if (masterv != v) {
 330		vlan_tunnel_info_del(vg, v);
 331		rhashtable_remove_fast(&vg->vlan_hash, &v->vnode,
 332				       br_vlan_rht_params);
 333		__vlan_del_list(v);
 334		kfree_rcu(v, rcu);
 
 335	}
 336
 337	br_vlan_put_master(masterv);
 338out:
 339	return err;
 340}
 341
 342static void __vlan_group_free(struct net_bridge_vlan_group *vg)
 343{
 344	WARN_ON(!list_empty(&vg->vlan_list));
 345	rhashtable_destroy(&vg->vlan_hash);
 346	vlan_tunnel_deinit(vg);
 347	kfree(vg);
 348}
 349
 350static void __vlan_flush(struct net_bridge_vlan_group *vg)
 351{
 352	struct net_bridge_vlan *vlan, *tmp;
 353
 354	__vlan_delete_pvid(vg, vg->pvid);
 355	list_for_each_entry_safe(vlan, tmp, &vg->vlan_list, vlist)
 356		__vlan_del(vlan);
 357}
 358
 359struct sk_buff *br_handle_vlan(struct net_bridge *br,
 360			       const struct net_bridge_port *p,
 361			       struct net_bridge_vlan_group *vg,
 362			       struct sk_buff *skb)
 363{
 364	struct br_vlan_stats *stats;
 365	struct net_bridge_vlan *v;
 366	u16 vid;
 367
 368	/* If this packet was not filtered at input, let it pass */
 369	if (!BR_INPUT_SKB_CB(skb)->vlan_filtered)
 370		goto out;
 371
 372	/* At this point, we know that the frame was filtered and contains
 373	 * a valid vlan id.  If the vlan id has untagged flag set,
 374	 * send untagged; otherwise, send tagged.
 375	 */
 376	br_vlan_get_tag(skb, &vid);
 377	v = br_vlan_find(vg, vid);
 378	/* Vlan entry must be configured at this point.  The
 379	 * only exception is the bridge is set in promisc mode and the
 380	 * packet is destined for the bridge device.  In this case
 381	 * pass the packet as is.
 382	 */
 383	if (!v || !br_vlan_should_use(v)) {
 384		if ((br->dev->flags & IFF_PROMISC) && skb->dev == br->dev) {
 385			goto out;
 386		} else {
 387			kfree_skb(skb);
 388			return NULL;
 389		}
 390	}
 391	if (br->vlan_stats_enabled) {
 392		stats = this_cpu_ptr(v->stats);
 393		u64_stats_update_begin(&stats->syncp);
 394		stats->tx_bytes += skb->len;
 395		stats->tx_packets++;
 396		u64_stats_update_end(&stats->syncp);
 397	}
 398
 399	if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
 400		skb->vlan_tci = 0;
 401
 402	if (p && (p->flags & BR_VLAN_TUNNEL) &&
 403	    br_handle_egress_vlan_tunnel(skb, v)) {
 404		kfree_skb(skb);
 405		return NULL;
 406	}
 407out:
 408	return skb;
 409}
 410
 411/* Called under RCU */
 412static bool __allowed_ingress(const struct net_bridge *br,
 413			      struct net_bridge_vlan_group *vg,
 414			      struct sk_buff *skb, u16 *vid)
 415{
 416	struct br_vlan_stats *stats;
 417	struct net_bridge_vlan *v;
 418	bool tagged;
 419
 420	BR_INPUT_SKB_CB(skb)->vlan_filtered = true;
 421	/* If vlan tx offload is disabled on bridge device and frame was
 422	 * sent from vlan device on the bridge device, it does not have
 423	 * HW accelerated vlan tag.
 424	 */
 425	if (unlikely(!skb_vlan_tag_present(skb) &&
 426		     skb->protocol == br->vlan_proto)) {
 427		skb = skb_vlan_untag(skb);
 428		if (unlikely(!skb))
 429			return false;
 430	}
 431
 432	if (!br_vlan_get_tag(skb, vid)) {
 433		/* Tagged frame */
 434		if (skb->vlan_proto != br->vlan_proto) {
 435			/* Protocol-mismatch, empty out vlan_tci for new tag */
 436			skb_push(skb, ETH_HLEN);
 437			skb = vlan_insert_tag_set_proto(skb, skb->vlan_proto,
 438							skb_vlan_tag_get(skb));
 439			if (unlikely(!skb))
 440				return false;
 441
 442			skb_pull(skb, ETH_HLEN);
 443			skb_reset_mac_len(skb);
 444			*vid = 0;
 445			tagged = false;
 446		} else {
 447			tagged = true;
 448		}
 449	} else {
 450		/* Untagged frame */
 451		tagged = false;
 452	}
 453
 454	if (!*vid) {
 455		u16 pvid = br_get_pvid(vg);
 456
 457		/* Frame had a tag with VID 0 or did not have a tag.
 458		 * See if pvid is set on this port.  That tells us which
 459		 * vlan untagged or priority-tagged traffic belongs to.
 460		 */
 461		if (!pvid)
 462			goto drop;
 463
 464		/* PVID is set on this port.  Any untagged or priority-tagged
 465		 * ingress frame is considered to belong to this vlan.
 466		 */
 467		*vid = pvid;
 468		if (likely(!tagged))
 469			/* Untagged Frame. */
 470			__vlan_hwaccel_put_tag(skb, br->vlan_proto, pvid);
 471		else
 472			/* Priority-tagged Frame.
 473			 * At this point, We know that skb->vlan_tci had
 474			 * VLAN_TAG_PRESENT bit and its VID field was 0x000.
 475			 * We update only VID field and preserve PCP field.
 476			 */
 477			skb->vlan_tci |= pvid;
 478
 479		/* if stats are disabled we can avoid the lookup */
 480		if (!br->vlan_stats_enabled)
 481			return true;
 482	}
 483	v = br_vlan_find(vg, *vid);
 484	if (!v || !br_vlan_should_use(v))
 485		goto drop;
 486
 487	if (br->vlan_stats_enabled) {
 488		stats = this_cpu_ptr(v->stats);
 489		u64_stats_update_begin(&stats->syncp);
 490		stats->rx_bytes += skb->len;
 491		stats->rx_packets++;
 492		u64_stats_update_end(&stats->syncp);
 493	}
 494
 495	return true;
 496
 497drop:
 498	kfree_skb(skb);
 499	return false;
 500}
 501
 502bool br_allowed_ingress(const struct net_bridge *br,
 503			struct net_bridge_vlan_group *vg, struct sk_buff *skb,
 504			u16 *vid)
 505{
 506	/* If VLAN filtering is disabled on the bridge, all packets are
 507	 * permitted.
 508	 */
 509	if (!br->vlan_enabled) {
 510		BR_INPUT_SKB_CB(skb)->vlan_filtered = false;
 511		return true;
 512	}
 513
 514	return __allowed_ingress(br, vg, skb, vid);
 515}
 516
 517/* Called under RCU. */
 518bool br_allowed_egress(struct net_bridge_vlan_group *vg,
 519		       const struct sk_buff *skb)
 520{
 521	const struct net_bridge_vlan *v;
 522	u16 vid;
 523
 524	/* If this packet was not filtered at input, let it pass */
 525	if (!BR_INPUT_SKB_CB(skb)->vlan_filtered)
 526		return true;
 527
 528	br_vlan_get_tag(skb, &vid);
 529	v = br_vlan_find(vg, vid);
 530	if (v && br_vlan_should_use(v))
 531		return true;
 532
 533	return false;
 534}
 535
 536/* Called under RCU */
 537bool br_should_learn(struct net_bridge_port *p, struct sk_buff *skb, u16 *vid)
 538{
 539	struct net_bridge_vlan_group *vg;
 540	struct net_bridge *br = p->br;
 541
 542	/* If filtering was disabled at input, let it pass. */
 543	if (!br->vlan_enabled)
 544		return true;
 545
 546	vg = nbp_vlan_group_rcu(p);
 547	if (!vg || !vg->num_vlans)
 548		return false;
 549
 550	if (!br_vlan_get_tag(skb, vid) && skb->vlan_proto != br->vlan_proto)
 551		*vid = 0;
 552
 553	if (!*vid) {
 554		*vid = br_get_pvid(vg);
 555		if (!*vid)
 556			return false;
 557
 558		return true;
 559	}
 560
 561	if (br_vlan_find(vg, *vid))
 562		return true;
 563
 564	return false;
 565}
 566
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 567/* Must be protected by RTNL.
 568 * Must be called with vid in range from 1 to 4094 inclusive.
 569 * changed must be true only if the vlan was created or updated
 570 */
 571int br_vlan_add(struct net_bridge *br, u16 vid, u16 flags, bool *changed)
 
 572{
 573	struct net_bridge_vlan_group *vg;
 574	struct net_bridge_vlan *vlan;
 575	int ret;
 576
 577	ASSERT_RTNL();
 578
 579	*changed = false;
 580	vg = br_vlan_group(br);
 581	vlan = br_vlan_find(vg, vid);
 582	if (vlan) {
 583		if (!br_vlan_is_brentry(vlan)) {
 584			/* Trying to change flags of non-existent bridge vlan */
 585			if (!(flags & BRIDGE_VLAN_INFO_BRENTRY))
 586				return -EINVAL;
 587			/* It was only kept for port vlans, now make it real */
 588			ret = br_fdb_insert(br, NULL, br->dev->dev_addr,
 589					    vlan->vid);
 590			if (ret) {
 591				br_err(br, "failed insert local address into bridge forwarding table\n");
 592				return ret;
 593			}
 594			refcount_inc(&vlan->refcnt);
 595			vlan->flags |= BRIDGE_VLAN_INFO_BRENTRY;
 596			vg->num_vlans++;
 597			*changed = true;
 598		}
 599		if (__vlan_add_flags(vlan, flags))
 600			*changed = true;
 601
 602		return 0;
 603	}
 604
 605	vlan = kzalloc(sizeof(*vlan), GFP_KERNEL);
 606	if (!vlan)
 607		return -ENOMEM;
 608
 609	vlan->stats = netdev_alloc_pcpu_stats(struct br_vlan_stats);
 610	if (!vlan->stats) {
 611		kfree(vlan);
 612		return -ENOMEM;
 613	}
 614	vlan->vid = vid;
 615	vlan->flags = flags | BRIDGE_VLAN_INFO_MASTER;
 616	vlan->flags &= ~BRIDGE_VLAN_INFO_PVID;
 617	vlan->br = br;
 618	if (flags & BRIDGE_VLAN_INFO_BRENTRY)
 619		refcount_set(&vlan->refcnt, 1);
 620	ret = __vlan_add(vlan, flags);
 621	if (ret) {
 622		free_percpu(vlan->stats);
 623		kfree(vlan);
 624	} else {
 625		*changed = true;
 626	}
 627
 628	return ret;
 629}
 630
 631/* Must be protected by RTNL.
 632 * Must be called with vid in range from 1 to 4094 inclusive.
 633 */
 634int br_vlan_delete(struct net_bridge *br, u16 vid)
 635{
 636	struct net_bridge_vlan_group *vg;
 637	struct net_bridge_vlan *v;
 638
 639	ASSERT_RTNL();
 640
 641	vg = br_vlan_group(br);
 642	v = br_vlan_find(vg, vid);
 643	if (!v || !br_vlan_is_brentry(v))
 644		return -ENOENT;
 645
 646	br_fdb_find_delete_local(br, NULL, br->dev->dev_addr, vid);
 647	br_fdb_delete_by_port(br, NULL, vid, 0);
 648
 649	vlan_tunnel_info_del(vg, v);
 650
 651	return __vlan_del(v);
 652}
 653
 654void br_vlan_flush(struct net_bridge *br)
 655{
 656	struct net_bridge_vlan_group *vg;
 657
 658	ASSERT_RTNL();
 659
 660	vg = br_vlan_group(br);
 661	__vlan_flush(vg);
 662	RCU_INIT_POINTER(br->vlgrp, NULL);
 663	synchronize_rcu();
 664	__vlan_group_free(vg);
 665}
 666
 667struct net_bridge_vlan *br_vlan_find(struct net_bridge_vlan_group *vg, u16 vid)
 668{
 669	if (!vg)
 670		return NULL;
 671
 672	return br_vlan_lookup(&vg->vlan_hash, vid);
 673}
 674
 675/* Must be protected by RTNL. */
 676static void recalculate_group_addr(struct net_bridge *br)
 677{
 678	if (br->group_addr_set)
 679		return;
 680
 681	spin_lock_bh(&br->lock);
 682	if (!br->vlan_enabled || br->vlan_proto == htons(ETH_P_8021Q)) {
 
 683		/* Bridge Group Address */
 684		br->group_addr[5] = 0x00;
 685	} else { /* vlan_enabled && ETH_P_8021AD */
 686		/* Provider Bridge Group Address */
 687		br->group_addr[5] = 0x08;
 688	}
 689	spin_unlock_bh(&br->lock);
 690}
 691
 692/* Must be protected by RTNL. */
 693void br_recalculate_fwd_mask(struct net_bridge *br)
 694{
 695	if (!br->vlan_enabled || br->vlan_proto == htons(ETH_P_8021Q))
 
 696		br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
 697	else /* vlan_enabled && ETH_P_8021AD */
 698		br->group_fwd_mask_required = BR_GROUPFWD_8021AD &
 699					      ~(1u << br->group_addr[5]);
 700}
 701
 702int __br_vlan_filter_toggle(struct net_bridge *br, unsigned long val)
 703{
 704	struct switchdev_attr attr = {
 705		.orig_dev = br->dev,
 706		.id = SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING,
 707		.flags = SWITCHDEV_F_SKIP_EOPNOTSUPP,
 708		.u.vlan_filtering = val,
 709	};
 710	int err;
 711
 712	if (br->vlan_enabled == val)
 713		return 0;
 714
 715	err = switchdev_port_attr_set(br->dev, &attr);
 716	if (err && err != -EOPNOTSUPP)
 717		return err;
 718
 719	br->vlan_enabled = val;
 720	br_manage_promisc(br);
 721	recalculate_group_addr(br);
 722	br_recalculate_fwd_mask(br);
 723
 724	return 0;
 725}
 726
 727int br_vlan_filter_toggle(struct net_bridge *br, unsigned long val)
 728{
 729	return __br_vlan_filter_toggle(br, val);
 730}
 731
 732bool br_vlan_enabled(const struct net_device *dev)
 733{
 734	struct net_bridge *br = netdev_priv(dev);
 735
 736	return !!br->vlan_enabled;
 737}
 738EXPORT_SYMBOL_GPL(br_vlan_enabled);
 739
 
 
 
 
 
 
 
 
 
 
 740int __br_vlan_set_proto(struct net_bridge *br, __be16 proto)
 741{
 742	int err = 0;
 743	struct net_bridge_port *p;
 744	struct net_bridge_vlan *vlan;
 745	struct net_bridge_vlan_group *vg;
 746	__be16 oldproto;
 747
 748	if (br->vlan_proto == proto)
 749		return 0;
 750
 751	/* Add VLANs for the new proto to the device filter. */
 752	list_for_each_entry(p, &br->port_list, list) {
 753		vg = nbp_vlan_group(p);
 754		list_for_each_entry(vlan, &vg->vlan_list, vlist) {
 755			err = vlan_vid_add(p->dev, proto, vlan->vid);
 756			if (err)
 757				goto err_filt;
 758		}
 759	}
 760
 761	oldproto = br->vlan_proto;
 762	br->vlan_proto = proto;
 763
 764	recalculate_group_addr(br);
 765	br_recalculate_fwd_mask(br);
 766
 767	/* Delete VLANs for the old proto from the device filter. */
 768	list_for_each_entry(p, &br->port_list, list) {
 769		vg = nbp_vlan_group(p);
 770		list_for_each_entry(vlan, &vg->vlan_list, vlist)
 771			vlan_vid_del(p->dev, oldproto, vlan->vid);
 772	}
 773
 774	return 0;
 775
 776err_filt:
 777	list_for_each_entry_continue_reverse(vlan, &vg->vlan_list, vlist)
 778		vlan_vid_del(p->dev, proto, vlan->vid);
 779
 780	list_for_each_entry_continue_reverse(p, &br->port_list, list) {
 781		vg = nbp_vlan_group(p);
 782		list_for_each_entry(vlan, &vg->vlan_list, vlist)
 783			vlan_vid_del(p->dev, proto, vlan->vid);
 784	}
 785
 786	return err;
 787}
 788
 789int br_vlan_set_proto(struct net_bridge *br, unsigned long val)
 790{
 791	if (val != ETH_P_8021Q && val != ETH_P_8021AD)
 792		return -EPROTONOSUPPORT;
 793
 794	return __br_vlan_set_proto(br, htons(val));
 795}
 796
 797int br_vlan_set_stats(struct net_bridge *br, unsigned long val)
 798{
 799	switch (val) {
 800	case 0:
 801	case 1:
 802		br->vlan_stats_enabled = val;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 803		break;
 804	default:
 805		return -EINVAL;
 806	}
 807
 808	return 0;
 809}
 810
 811static bool vlan_default_pvid(struct net_bridge_vlan_group *vg, u16 vid)
 812{
 813	struct net_bridge_vlan *v;
 814
 815	if (vid != vg->pvid)
 816		return false;
 817
 818	v = br_vlan_lookup(&vg->vlan_hash, vid);
 819	if (v && br_vlan_should_use(v) &&
 820	    (v->flags & BRIDGE_VLAN_INFO_UNTAGGED))
 821		return true;
 822
 823	return false;
 824}
 825
 826static void br_vlan_disable_default_pvid(struct net_bridge *br)
 827{
 828	struct net_bridge_port *p;
 829	u16 pvid = br->default_pvid;
 830
 831	/* Disable default_pvid on all ports where it is still
 832	 * configured.
 833	 */
 834	if (vlan_default_pvid(br_vlan_group(br), pvid))
 835		br_vlan_delete(br, pvid);
 836
 837	list_for_each_entry(p, &br->port_list, list) {
 838		if (vlan_default_pvid(nbp_vlan_group(p), pvid))
 839			nbp_vlan_delete(p, pvid);
 840	}
 841
 842	br->default_pvid = 0;
 843}
 844
 845int __br_vlan_set_default_pvid(struct net_bridge *br, u16 pvid)
 
 846{
 847	const struct net_bridge_vlan *pvent;
 848	struct net_bridge_vlan_group *vg;
 849	struct net_bridge_port *p;
 850	unsigned long *changed;
 851	bool vlchange;
 852	u16 old_pvid;
 853	int err = 0;
 854
 855	if (!pvid) {
 856		br_vlan_disable_default_pvid(br);
 857		return 0;
 858	}
 859
 860	changed = kcalloc(BITS_TO_LONGS(BR_MAX_PORTS), sizeof(unsigned long),
 861			  GFP_KERNEL);
 862	if (!changed)
 863		return -ENOMEM;
 864
 865	old_pvid = br->default_pvid;
 866
 867	/* Update default_pvid config only if we do not conflict with
 868	 * user configuration.
 869	 */
 870	vg = br_vlan_group(br);
 871	pvent = br_vlan_find(vg, pvid);
 872	if ((!old_pvid || vlan_default_pvid(vg, old_pvid)) &&
 873	    (!pvent || !br_vlan_should_use(pvent))) {
 874		err = br_vlan_add(br, pvid,
 875				  BRIDGE_VLAN_INFO_PVID |
 876				  BRIDGE_VLAN_INFO_UNTAGGED |
 877				  BRIDGE_VLAN_INFO_BRENTRY,
 878				  &vlchange);
 879		if (err)
 880			goto out;
 881		br_vlan_delete(br, old_pvid);
 882		set_bit(0, changed);
 883	}
 884
 885	list_for_each_entry(p, &br->port_list, list) {
 886		/* Update default_pvid config only if we do not conflict with
 887		 * user configuration.
 888		 */
 889		vg = nbp_vlan_group(p);
 890		if ((old_pvid &&
 891		     !vlan_default_pvid(vg, old_pvid)) ||
 892		    br_vlan_find(vg, pvid))
 893			continue;
 894
 895		err = nbp_vlan_add(p, pvid,
 896				   BRIDGE_VLAN_INFO_PVID |
 897				   BRIDGE_VLAN_INFO_UNTAGGED,
 898				   &vlchange);
 899		if (err)
 900			goto err_port;
 901		nbp_vlan_delete(p, old_pvid);
 902		set_bit(p->port_no, changed);
 903	}
 904
 905	br->default_pvid = pvid;
 906
 907out:
 908	kfree(changed);
 909	return err;
 910
 911err_port:
 912	list_for_each_entry_continue_reverse(p, &br->port_list, list) {
 913		if (!test_bit(p->port_no, changed))
 914			continue;
 915
 916		if (old_pvid)
 917			nbp_vlan_add(p, old_pvid,
 918				     BRIDGE_VLAN_INFO_PVID |
 919				     BRIDGE_VLAN_INFO_UNTAGGED,
 920				     &vlchange);
 921		nbp_vlan_delete(p, pvid);
 922	}
 923
 924	if (test_bit(0, changed)) {
 925		if (old_pvid)
 926			br_vlan_add(br, old_pvid,
 927				    BRIDGE_VLAN_INFO_PVID |
 928				    BRIDGE_VLAN_INFO_UNTAGGED |
 929				    BRIDGE_VLAN_INFO_BRENTRY,
 930				    &vlchange);
 931		br_vlan_delete(br, pvid);
 932	}
 933	goto out;
 934}
 935
 936int br_vlan_set_default_pvid(struct net_bridge *br, unsigned long val)
 937{
 938	u16 pvid = val;
 939	int err = 0;
 940
 941	if (val >= VLAN_VID_MASK)
 942		return -EINVAL;
 943
 944	if (pvid == br->default_pvid)
 945		goto out;
 946
 947	/* Only allow default pvid change when filtering is disabled */
 948	if (br->vlan_enabled) {
 949		pr_info_once("Please disable vlan filtering to change default_pvid\n");
 950		err = -EPERM;
 951		goto out;
 952	}
 953	err = __br_vlan_set_default_pvid(br, pvid);
 954out:
 955	return err;
 956}
 957
 958int br_vlan_init(struct net_bridge *br)
 959{
 960	struct net_bridge_vlan_group *vg;
 961	int ret = -ENOMEM;
 962	bool changed;
 963
 964	vg = kzalloc(sizeof(*vg), GFP_KERNEL);
 965	if (!vg)
 966		goto out;
 967	ret = rhashtable_init(&vg->vlan_hash, &br_vlan_rht_params);
 968	if (ret)
 969		goto err_rhtbl;
 970	ret = vlan_tunnel_init(vg);
 971	if (ret)
 972		goto err_tunnel_init;
 973	INIT_LIST_HEAD(&vg->vlan_list);
 974	br->vlan_proto = htons(ETH_P_8021Q);
 975	br->default_pvid = 1;
 976	rcu_assign_pointer(br->vlgrp, vg);
 977	ret = br_vlan_add(br, 1,
 978			  BRIDGE_VLAN_INFO_PVID | BRIDGE_VLAN_INFO_UNTAGGED |
 979			  BRIDGE_VLAN_INFO_BRENTRY, &changed);
 980	if (ret)
 981		goto err_vlan_add;
 982
 983out:
 984	return ret;
 985
 986err_vlan_add:
 987	vlan_tunnel_deinit(vg);
 988err_tunnel_init:
 989	rhashtable_destroy(&vg->vlan_hash);
 990err_rhtbl:
 991	kfree(vg);
 992
 993	goto out;
 994}
 995
 996int nbp_vlan_init(struct net_bridge_port *p)
 997{
 998	struct switchdev_attr attr = {
 999		.orig_dev = p->br->dev,
1000		.id = SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING,
1001		.flags = SWITCHDEV_F_SKIP_EOPNOTSUPP,
1002		.u.vlan_filtering = p->br->vlan_enabled,
1003	};
1004	struct net_bridge_vlan_group *vg;
1005	int ret = -ENOMEM;
1006
1007	vg = kzalloc(sizeof(struct net_bridge_vlan_group), GFP_KERNEL);
1008	if (!vg)
1009		goto out;
1010
1011	ret = switchdev_port_attr_set(p->dev, &attr);
1012	if (ret && ret != -EOPNOTSUPP)
1013		goto err_vlan_enabled;
1014
1015	ret = rhashtable_init(&vg->vlan_hash, &br_vlan_rht_params);
1016	if (ret)
1017		goto err_rhtbl;
1018	ret = vlan_tunnel_init(vg);
1019	if (ret)
1020		goto err_tunnel_init;
1021	INIT_LIST_HEAD(&vg->vlan_list);
1022	rcu_assign_pointer(p->vlgrp, vg);
1023	if (p->br->default_pvid) {
1024		bool changed;
1025
1026		ret = nbp_vlan_add(p, p->br->default_pvid,
1027				   BRIDGE_VLAN_INFO_PVID |
1028				   BRIDGE_VLAN_INFO_UNTAGGED,
1029				   &changed);
1030		if (ret)
1031			goto err_vlan_add;
1032	}
1033out:
1034	return ret;
1035
1036err_vlan_add:
1037	RCU_INIT_POINTER(p->vlgrp, NULL);
1038	synchronize_rcu();
1039	vlan_tunnel_deinit(vg);
1040err_tunnel_init:
1041	rhashtable_destroy(&vg->vlan_hash);
1042err_rhtbl:
1043err_vlan_enabled:
1044	kfree(vg);
1045
1046	goto out;
1047}
1048
1049/* Must be protected by RTNL.
1050 * Must be called with vid in range from 1 to 4094 inclusive.
1051 * changed must be true only if the vlan was created or updated
1052 */
1053int nbp_vlan_add(struct net_bridge_port *port, u16 vid, u16 flags,
1054		 bool *changed)
1055{
1056	struct switchdev_obj_port_vlan v = {
1057		.obj.orig_dev = port->dev,
1058		.obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
1059		.flags = flags,
1060		.vid_begin = vid,
1061		.vid_end = vid,
1062	};
1063	struct net_bridge_vlan *vlan;
1064	int ret;
1065
1066	ASSERT_RTNL();
1067
1068	*changed = false;
1069	vlan = br_vlan_find(nbp_vlan_group(port), vid);
1070	if (vlan) {
1071		/* Pass the flags to the hardware bridge */
1072		ret = switchdev_port_obj_add(port->dev, &v.obj);
1073		if (ret && ret != -EOPNOTSUPP)
1074			return ret;
1075		*changed = __vlan_add_flags(vlan, flags);
1076
1077		return 0;
1078	}
1079
1080	vlan = kzalloc(sizeof(*vlan), GFP_KERNEL);
1081	if (!vlan)
1082		return -ENOMEM;
1083
1084	vlan->vid = vid;
1085	vlan->port = port;
1086	ret = __vlan_add(vlan, flags);
1087	if (ret)
1088		kfree(vlan);
1089	else
1090		*changed = true;
1091
1092	return ret;
1093}
1094
1095/* Must be protected by RTNL.
1096 * Must be called with vid in range from 1 to 4094 inclusive.
1097 */
1098int nbp_vlan_delete(struct net_bridge_port *port, u16 vid)
1099{
1100	struct net_bridge_vlan *v;
1101
1102	ASSERT_RTNL();
1103
1104	v = br_vlan_find(nbp_vlan_group(port), vid);
1105	if (!v)
1106		return -ENOENT;
1107	br_fdb_find_delete_local(port->br, port, port->dev->dev_addr, vid);
1108	br_fdb_delete_by_port(port->br, port, vid, 0);
1109
1110	return __vlan_del(v);
1111}
1112
1113void nbp_vlan_flush(struct net_bridge_port *port)
1114{
1115	struct net_bridge_vlan_group *vg;
1116
1117	ASSERT_RTNL();
1118
1119	vg = nbp_vlan_group(port);
1120	__vlan_flush(vg);
1121	RCU_INIT_POINTER(port->vlgrp, NULL);
1122	synchronize_rcu();
1123	__vlan_group_free(vg);
1124}
1125
1126void br_vlan_get_stats(const struct net_bridge_vlan *v,
1127		       struct br_vlan_stats *stats)
1128{
1129	int i;
1130
1131	memset(stats, 0, sizeof(*stats));
1132	for_each_possible_cpu(i) {
1133		u64 rxpackets, rxbytes, txpackets, txbytes;
1134		struct br_vlan_stats *cpu_stats;
1135		unsigned int start;
1136
1137		cpu_stats = per_cpu_ptr(v->stats, i);
1138		do {
1139			start = u64_stats_fetch_begin_irq(&cpu_stats->syncp);
1140			rxpackets = cpu_stats->rx_packets;
1141			rxbytes = cpu_stats->rx_bytes;
1142			txbytes = cpu_stats->tx_bytes;
1143			txpackets = cpu_stats->tx_packets;
1144		} while (u64_stats_fetch_retry_irq(&cpu_stats->syncp, start));
1145
1146		stats->rx_packets += rxpackets;
1147		stats->rx_bytes += rxbytes;
1148		stats->tx_bytes += txbytes;
1149		stats->tx_packets += txpackets;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1150	}
1151}
v5.4
   1// SPDX-License-Identifier: GPL-2.0-only
   2#include <linux/kernel.h>
   3#include <linux/netdevice.h>
   4#include <linux/rtnetlink.h>
   5#include <linux/slab.h>
   6#include <net/switchdev.h>
   7
   8#include "br_private.h"
   9#include "br_private_tunnel.h"
  10
  11static void nbp_vlan_set_vlan_dev_state(struct net_bridge_port *p, u16 vid);
  12
  13static inline int br_vlan_cmp(struct rhashtable_compare_arg *arg,
  14			      const void *ptr)
  15{
  16	const struct net_bridge_vlan *vle = ptr;
  17	u16 vid = *(u16 *)arg->key;
  18
  19	return vle->vid != vid;
  20}
  21
  22static const struct rhashtable_params br_vlan_rht_params = {
  23	.head_offset = offsetof(struct net_bridge_vlan, vnode),
  24	.key_offset = offsetof(struct net_bridge_vlan, vid),
  25	.key_len = sizeof(u16),
  26	.nelem_hint = 3,
 
  27	.max_size = VLAN_N_VID,
  28	.obj_cmpfn = br_vlan_cmp,
  29	.automatic_shrinking = true,
  30};
  31
  32static struct net_bridge_vlan *br_vlan_lookup(struct rhashtable *tbl, u16 vid)
  33{
  34	return rhashtable_lookup_fast(tbl, &vid, br_vlan_rht_params);
  35}
  36
  37static bool __vlan_add_pvid(struct net_bridge_vlan_group *vg, u16 vid)
  38{
  39	if (vg->pvid == vid)
  40		return false;
  41
  42	smp_wmb();
  43	vg->pvid = vid;
  44
  45	return true;
  46}
  47
  48static bool __vlan_delete_pvid(struct net_bridge_vlan_group *vg, u16 vid)
  49{
  50	if (vg->pvid != vid)
  51		return false;
  52
  53	smp_wmb();
  54	vg->pvid = 0;
  55
  56	return true;
  57}
  58
  59/* return true if anything changed, false otherwise */
  60static bool __vlan_add_flags(struct net_bridge_vlan *v, u16 flags)
  61{
  62	struct net_bridge_vlan_group *vg;
  63	u16 old_flags = v->flags;
  64	bool ret;
  65
  66	if (br_vlan_is_master(v))
  67		vg = br_vlan_group(v->br);
  68	else
  69		vg = nbp_vlan_group(v->port);
  70
  71	if (flags & BRIDGE_VLAN_INFO_PVID)
  72		ret = __vlan_add_pvid(vg, v->vid);
  73	else
  74		ret = __vlan_delete_pvid(vg, v->vid);
  75
  76	if (flags & BRIDGE_VLAN_INFO_UNTAGGED)
  77		v->flags |= BRIDGE_VLAN_INFO_UNTAGGED;
  78	else
  79		v->flags &= ~BRIDGE_VLAN_INFO_UNTAGGED;
  80
  81	return ret || !!(old_flags ^ v->flags);
  82}
  83
  84static int __vlan_vid_add(struct net_device *dev, struct net_bridge *br,
  85			  struct net_bridge_vlan *v, u16 flags,
  86			  struct netlink_ext_ack *extack)
  87{
 
 
 
 
 
 
 
  88	int err;
  89
  90	/* Try switchdev op first. In case it is not supported, fallback to
  91	 * 8021q add.
  92	 */
  93	err = br_switchdev_port_vlan_add(dev, v->vid, flags, extack);
  94	if (err == -EOPNOTSUPP)
  95		return vlan_vid_add(dev, br->vlan_proto, v->vid);
  96	v->priv_flags |= BR_VLFLAG_ADDED_BY_SWITCHDEV;
  97	return err;
  98}
  99
 100static void __vlan_add_list(struct net_bridge_vlan *v)
 101{
 102	struct net_bridge_vlan_group *vg;
 103	struct list_head *headp, *hpos;
 104	struct net_bridge_vlan *vent;
 105
 106	if (br_vlan_is_master(v))
 107		vg = br_vlan_group(v->br);
 108	else
 109		vg = nbp_vlan_group(v->port);
 110
 111	headp = &vg->vlan_list;
 112	list_for_each_prev(hpos, headp) {
 113		vent = list_entry(hpos, struct net_bridge_vlan, vlist);
 114		if (v->vid < vent->vid)
 115			continue;
 116		else
 117			break;
 118	}
 119	list_add_rcu(&v->vlist, hpos);
 120}
 121
 122static void __vlan_del_list(struct net_bridge_vlan *v)
 123{
 124	list_del_rcu(&v->vlist);
 125}
 126
 127static int __vlan_vid_del(struct net_device *dev, struct net_bridge *br,
 128			  const struct net_bridge_vlan *v)
 129{
 
 
 
 
 
 
 130	int err;
 131
 132	/* Try switchdev op first. In case it is not supported, fallback to
 133	 * 8021q del.
 134	 */
 135	err = br_switchdev_port_vlan_del(dev, v->vid);
 136	if (!(v->priv_flags & BR_VLFLAG_ADDED_BY_SWITCHDEV))
 137		vlan_vid_del(dev, br->vlan_proto, v->vid);
 138	return err == -EOPNOTSUPP ? 0 : err;
 
 
 139}
 140
 141/* Returns a master vlan, if it didn't exist it gets created. In all cases a
 142 * a reference is taken to the master vlan before returning.
 143 */
 144static struct net_bridge_vlan *
 145br_vlan_get_master(struct net_bridge *br, u16 vid,
 146		   struct netlink_ext_ack *extack)
 147{
 148	struct net_bridge_vlan_group *vg;
 149	struct net_bridge_vlan *masterv;
 150
 151	vg = br_vlan_group(br);
 152	masterv = br_vlan_find(vg, vid);
 153	if (!masterv) {
 154		bool changed;
 155
 156		/* missing global ctx, create it now */
 157		if (br_vlan_add(br, vid, 0, &changed, extack))
 158			return NULL;
 159		masterv = br_vlan_find(vg, vid);
 160		if (WARN_ON(!masterv))
 161			return NULL;
 162		refcount_set(&masterv->refcnt, 1);
 163		return masterv;
 164	}
 165	refcount_inc(&masterv->refcnt);
 166
 167	return masterv;
 168}
 169
 170static void br_master_vlan_rcu_free(struct rcu_head *rcu)
 171{
 172	struct net_bridge_vlan *v;
 173
 174	v = container_of(rcu, struct net_bridge_vlan, rcu);
 175	WARN_ON(!br_vlan_is_master(v));
 176	free_percpu(v->stats);
 177	v->stats = NULL;
 178	kfree(v);
 179}
 180
 181static void br_vlan_put_master(struct net_bridge_vlan *masterv)
 182{
 183	struct net_bridge_vlan_group *vg;
 184
 185	if (!br_vlan_is_master(masterv))
 186		return;
 187
 188	vg = br_vlan_group(masterv->br);
 189	if (refcount_dec_and_test(&masterv->refcnt)) {
 190		rhashtable_remove_fast(&vg->vlan_hash,
 191				       &masterv->vnode, br_vlan_rht_params);
 192		__vlan_del_list(masterv);
 193		call_rcu(&masterv->rcu, br_master_vlan_rcu_free);
 194	}
 195}
 196
 197static void nbp_vlan_rcu_free(struct rcu_head *rcu)
 198{
 199	struct net_bridge_vlan *v;
 200
 201	v = container_of(rcu, struct net_bridge_vlan, rcu);
 202	WARN_ON(br_vlan_is_master(v));
 203	/* if we had per-port stats configured then free them here */
 204	if (v->priv_flags & BR_VLFLAG_PER_PORT_STATS)
 205		free_percpu(v->stats);
 206	v->stats = NULL;
 207	kfree(v);
 208}
 209
 210/* This is the shared VLAN add function which works for both ports and bridge
 211 * devices. There are four possible calls to this function in terms of the
 212 * vlan entry type:
 213 * 1. vlan is being added on a port (no master flags, global entry exists)
 214 * 2. vlan is being added on a bridge (both master and brentry flags)
 215 * 3. vlan is being added on a port, but a global entry didn't exist which
 216 *    is being created right now (master flag set, brentry flag unset), the
 217 *    global entry is used for global per-vlan features, but not for filtering
 218 * 4. same as 3 but with both master and brentry flags set so the entry
 219 *    will be used for filtering in both the port and the bridge
 220 */
 221static int __vlan_add(struct net_bridge_vlan *v, u16 flags,
 222		      struct netlink_ext_ack *extack)
 223{
 224	struct net_bridge_vlan *masterv = NULL;
 225	struct net_bridge_port *p = NULL;
 226	struct net_bridge_vlan_group *vg;
 227	struct net_device *dev;
 228	struct net_bridge *br;
 229	int err;
 230
 231	if (br_vlan_is_master(v)) {
 232		br = v->br;
 233		dev = br->dev;
 234		vg = br_vlan_group(br);
 235	} else {
 236		p = v->port;
 237		br = p->br;
 238		dev = p->dev;
 239		vg = nbp_vlan_group(p);
 240	}
 241
 242	if (p) {
 243		/* Add VLAN to the device filter if it is supported.
 244		 * This ensures tagged traffic enters the bridge when
 245		 * promiscuous mode is disabled by br_manage_promisc().
 246		 */
 247		err = __vlan_vid_add(dev, br, v, flags, extack);
 248		if (err)
 249			goto out;
 250
 251		/* need to work on the master vlan too */
 252		if (flags & BRIDGE_VLAN_INFO_MASTER) {
 253			bool changed;
 254
 255			err = br_vlan_add(br, v->vid,
 256					  flags | BRIDGE_VLAN_INFO_BRENTRY,
 257					  &changed, extack);
 258			if (err)
 259				goto out_filt;
 260		}
 261
 262		masterv = br_vlan_get_master(br, v->vid, extack);
 263		if (!masterv)
 264			goto out_filt;
 265		v->brvlan = masterv;
 266		if (br_opt_get(br, BROPT_VLAN_STATS_PER_PORT)) {
 267			v->stats = netdev_alloc_pcpu_stats(struct br_vlan_stats);
 268			if (!v->stats) {
 269				err = -ENOMEM;
 270				goto out_filt;
 271			}
 272			v->priv_flags |= BR_VLFLAG_PER_PORT_STATS;
 273		} else {
 274			v->stats = masterv->stats;
 275		}
 276	} else {
 277		err = br_switchdev_port_vlan_add(dev, v->vid, flags, extack);
 278		if (err && err != -EOPNOTSUPP)
 279			goto out;
 280	}
 281
 282	/* Add the dev mac and count the vlan only if it's usable */
 283	if (br_vlan_should_use(v)) {
 284		err = br_fdb_insert(br, p, dev->dev_addr, v->vid);
 285		if (err) {
 286			br_err(br, "failed insert local address into bridge forwarding table\n");
 287			goto out_filt;
 288		}
 289		vg->num_vlans++;
 290	}
 291
 292	err = rhashtable_lookup_insert_fast(&vg->vlan_hash, &v->vnode,
 293					    br_vlan_rht_params);
 294	if (err)
 295		goto out_fdb_insert;
 296
 297	__vlan_add_list(v);
 298	__vlan_add_flags(v, flags);
 299
 300	if (p)
 301		nbp_vlan_set_vlan_dev_state(p, v->vid);
 302out:
 303	return err;
 304
 305out_fdb_insert:
 306	if (br_vlan_should_use(v)) {
 307		br_fdb_find_delete_local(br, p, dev->dev_addr, v->vid);
 308		vg->num_vlans--;
 309	}
 310
 311out_filt:
 312	if (p) {
 313		__vlan_vid_del(dev, br, v);
 314		if (masterv) {
 315			if (v->stats && masterv->stats != v->stats)
 316				free_percpu(v->stats);
 317			v->stats = NULL;
 318
 319			br_vlan_put_master(masterv);
 320			v->brvlan = NULL;
 321		}
 322	} else {
 323		br_switchdev_port_vlan_del(dev, v->vid);
 324	}
 325
 326	goto out;
 327}
 328
 329static int __vlan_del(struct net_bridge_vlan *v)
 330{
 331	struct net_bridge_vlan *masterv = v;
 332	struct net_bridge_vlan_group *vg;
 333	struct net_bridge_port *p = NULL;
 334	int err = 0;
 335
 336	if (br_vlan_is_master(v)) {
 337		vg = br_vlan_group(v->br);
 338	} else {
 339		p = v->port;
 340		vg = nbp_vlan_group(v->port);
 341		masterv = v->brvlan;
 342	}
 343
 344	__vlan_delete_pvid(vg, v->vid);
 345	if (p) {
 346		err = __vlan_vid_del(p->dev, p->br, v);
 347		if (err)
 348			goto out;
 349	} else {
 350		err = br_switchdev_port_vlan_del(v->br->dev, v->vid);
 351		if (err && err != -EOPNOTSUPP)
 352			goto out;
 353		err = 0;
 354	}
 355
 356	if (br_vlan_should_use(v)) {
 357		v->flags &= ~BRIDGE_VLAN_INFO_BRENTRY;
 358		vg->num_vlans--;
 359	}
 360
 361	if (masterv != v) {
 362		vlan_tunnel_info_del(vg, v);
 363		rhashtable_remove_fast(&vg->vlan_hash, &v->vnode,
 364				       br_vlan_rht_params);
 365		__vlan_del_list(v);
 366		nbp_vlan_set_vlan_dev_state(p, v->vid);
 367		call_rcu(&v->rcu, nbp_vlan_rcu_free);
 368	}
 369
 370	br_vlan_put_master(masterv);
 371out:
 372	return err;
 373}
 374
 375static void __vlan_group_free(struct net_bridge_vlan_group *vg)
 376{
 377	WARN_ON(!list_empty(&vg->vlan_list));
 378	rhashtable_destroy(&vg->vlan_hash);
 379	vlan_tunnel_deinit(vg);
 380	kfree(vg);
 381}
 382
 383static void __vlan_flush(struct net_bridge_vlan_group *vg)
 384{
 385	struct net_bridge_vlan *vlan, *tmp;
 386
 387	__vlan_delete_pvid(vg, vg->pvid);
 388	list_for_each_entry_safe(vlan, tmp, &vg->vlan_list, vlist)
 389		__vlan_del(vlan);
 390}
 391
 392struct sk_buff *br_handle_vlan(struct net_bridge *br,
 393			       const struct net_bridge_port *p,
 394			       struct net_bridge_vlan_group *vg,
 395			       struct sk_buff *skb)
 396{
 397	struct br_vlan_stats *stats;
 398	struct net_bridge_vlan *v;
 399	u16 vid;
 400
 401	/* If this packet was not filtered at input, let it pass */
 402	if (!BR_INPUT_SKB_CB(skb)->vlan_filtered)
 403		goto out;
 404
 405	/* At this point, we know that the frame was filtered and contains
 406	 * a valid vlan id.  If the vlan id has untagged flag set,
 407	 * send untagged; otherwise, send tagged.
 408	 */
 409	br_vlan_get_tag(skb, &vid);
 410	v = br_vlan_find(vg, vid);
 411	/* Vlan entry must be configured at this point.  The
 412	 * only exception is the bridge is set in promisc mode and the
 413	 * packet is destined for the bridge device.  In this case
 414	 * pass the packet as is.
 415	 */
 416	if (!v || !br_vlan_should_use(v)) {
 417		if ((br->dev->flags & IFF_PROMISC) && skb->dev == br->dev) {
 418			goto out;
 419		} else {
 420			kfree_skb(skb);
 421			return NULL;
 422		}
 423	}
 424	if (br_opt_get(br, BROPT_VLAN_STATS_ENABLED)) {
 425		stats = this_cpu_ptr(v->stats);
 426		u64_stats_update_begin(&stats->syncp);
 427		stats->tx_bytes += skb->len;
 428		stats->tx_packets++;
 429		u64_stats_update_end(&stats->syncp);
 430	}
 431
 432	if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
 433		__vlan_hwaccel_clear_tag(skb);
 434
 435	if (p && (p->flags & BR_VLAN_TUNNEL) &&
 436	    br_handle_egress_vlan_tunnel(skb, v)) {
 437		kfree_skb(skb);
 438		return NULL;
 439	}
 440out:
 441	return skb;
 442}
 443
 444/* Called under RCU */
 445static bool __allowed_ingress(const struct net_bridge *br,
 446			      struct net_bridge_vlan_group *vg,
 447			      struct sk_buff *skb, u16 *vid)
 448{
 449	struct br_vlan_stats *stats;
 450	struct net_bridge_vlan *v;
 451	bool tagged;
 452
 453	BR_INPUT_SKB_CB(skb)->vlan_filtered = true;
 454	/* If vlan tx offload is disabled on bridge device and frame was
 455	 * sent from vlan device on the bridge device, it does not have
 456	 * HW accelerated vlan tag.
 457	 */
 458	if (unlikely(!skb_vlan_tag_present(skb) &&
 459		     skb->protocol == br->vlan_proto)) {
 460		skb = skb_vlan_untag(skb);
 461		if (unlikely(!skb))
 462			return false;
 463	}
 464
 465	if (!br_vlan_get_tag(skb, vid)) {
 466		/* Tagged frame */
 467		if (skb->vlan_proto != br->vlan_proto) {
 468			/* Protocol-mismatch, empty out vlan_tci for new tag */
 469			skb_push(skb, ETH_HLEN);
 470			skb = vlan_insert_tag_set_proto(skb, skb->vlan_proto,
 471							skb_vlan_tag_get(skb));
 472			if (unlikely(!skb))
 473				return false;
 474
 475			skb_pull(skb, ETH_HLEN);
 476			skb_reset_mac_len(skb);
 477			*vid = 0;
 478			tagged = false;
 479		} else {
 480			tagged = true;
 481		}
 482	} else {
 483		/* Untagged frame */
 484		tagged = false;
 485	}
 486
 487	if (!*vid) {
 488		u16 pvid = br_get_pvid(vg);
 489
 490		/* Frame had a tag with VID 0 or did not have a tag.
 491		 * See if pvid is set on this port.  That tells us which
 492		 * vlan untagged or priority-tagged traffic belongs to.
 493		 */
 494		if (!pvid)
 495			goto drop;
 496
 497		/* PVID is set on this port.  Any untagged or priority-tagged
 498		 * ingress frame is considered to belong to this vlan.
 499		 */
 500		*vid = pvid;
 501		if (likely(!tagged))
 502			/* Untagged Frame. */
 503			__vlan_hwaccel_put_tag(skb, br->vlan_proto, pvid);
 504		else
 505			/* Priority-tagged Frame.
 506			 * At this point, we know that skb->vlan_tci VID
 507			 * field was 0.
 508			 * We update only VID field and preserve PCP field.
 509			 */
 510			skb->vlan_tci |= pvid;
 511
 512		/* if stats are disabled we can avoid the lookup */
 513		if (!br_opt_get(br, BROPT_VLAN_STATS_ENABLED))
 514			return true;
 515	}
 516	v = br_vlan_find(vg, *vid);
 517	if (!v || !br_vlan_should_use(v))
 518		goto drop;
 519
 520	if (br_opt_get(br, BROPT_VLAN_STATS_ENABLED)) {
 521		stats = this_cpu_ptr(v->stats);
 522		u64_stats_update_begin(&stats->syncp);
 523		stats->rx_bytes += skb->len;
 524		stats->rx_packets++;
 525		u64_stats_update_end(&stats->syncp);
 526	}
 527
 528	return true;
 529
 530drop:
 531	kfree_skb(skb);
 532	return false;
 533}
 534
 535bool br_allowed_ingress(const struct net_bridge *br,
 536			struct net_bridge_vlan_group *vg, struct sk_buff *skb,
 537			u16 *vid)
 538{
 539	/* If VLAN filtering is disabled on the bridge, all packets are
 540	 * permitted.
 541	 */
 542	if (!br_opt_get(br, BROPT_VLAN_ENABLED)) {
 543		BR_INPUT_SKB_CB(skb)->vlan_filtered = false;
 544		return true;
 545	}
 546
 547	return __allowed_ingress(br, vg, skb, vid);
 548}
 549
 550/* Called under RCU. */
 551bool br_allowed_egress(struct net_bridge_vlan_group *vg,
 552		       const struct sk_buff *skb)
 553{
 554	const struct net_bridge_vlan *v;
 555	u16 vid;
 556
 557	/* If this packet was not filtered at input, let it pass */
 558	if (!BR_INPUT_SKB_CB(skb)->vlan_filtered)
 559		return true;
 560
 561	br_vlan_get_tag(skb, &vid);
 562	v = br_vlan_find(vg, vid);
 563	if (v && br_vlan_should_use(v))
 564		return true;
 565
 566	return false;
 567}
 568
 569/* Called under RCU */
 570bool br_should_learn(struct net_bridge_port *p, struct sk_buff *skb, u16 *vid)
 571{
 572	struct net_bridge_vlan_group *vg;
 573	struct net_bridge *br = p->br;
 574
 575	/* If filtering was disabled at input, let it pass. */
 576	if (!br_opt_get(br, BROPT_VLAN_ENABLED))
 577		return true;
 578
 579	vg = nbp_vlan_group_rcu(p);
 580	if (!vg || !vg->num_vlans)
 581		return false;
 582
 583	if (!br_vlan_get_tag(skb, vid) && skb->vlan_proto != br->vlan_proto)
 584		*vid = 0;
 585
 586	if (!*vid) {
 587		*vid = br_get_pvid(vg);
 588		if (!*vid)
 589			return false;
 590
 591		return true;
 592	}
 593
 594	if (br_vlan_find(vg, *vid))
 595		return true;
 596
 597	return false;
 598}
 599
 600static int br_vlan_add_existing(struct net_bridge *br,
 601				struct net_bridge_vlan_group *vg,
 602				struct net_bridge_vlan *vlan,
 603				u16 flags, bool *changed,
 604				struct netlink_ext_ack *extack)
 605{
 606	int err;
 607
 608	err = br_switchdev_port_vlan_add(br->dev, vlan->vid, flags, extack);
 609	if (err && err != -EOPNOTSUPP)
 610		return err;
 611
 612	if (!br_vlan_is_brentry(vlan)) {
 613		/* Trying to change flags of non-existent bridge vlan */
 614		if (!(flags & BRIDGE_VLAN_INFO_BRENTRY)) {
 615			err = -EINVAL;
 616			goto err_flags;
 617		}
 618		/* It was only kept for port vlans, now make it real */
 619		err = br_fdb_insert(br, NULL, br->dev->dev_addr,
 620				    vlan->vid);
 621		if (err) {
 622			br_err(br, "failed to insert local address into bridge forwarding table\n");
 623			goto err_fdb_insert;
 624		}
 625
 626		refcount_inc(&vlan->refcnt);
 627		vlan->flags |= BRIDGE_VLAN_INFO_BRENTRY;
 628		vg->num_vlans++;
 629		*changed = true;
 630	}
 631
 632	if (__vlan_add_flags(vlan, flags))
 633		*changed = true;
 634
 635	return 0;
 636
 637err_fdb_insert:
 638err_flags:
 639	br_switchdev_port_vlan_del(br->dev, vlan->vid);
 640	return err;
 641}
 642
 643/* Must be protected by RTNL.
 644 * Must be called with vid in range from 1 to 4094 inclusive.
 645 * changed must be true only if the vlan was created or updated
 646 */
 647int br_vlan_add(struct net_bridge *br, u16 vid, u16 flags, bool *changed,
 648		struct netlink_ext_ack *extack)
 649{
 650	struct net_bridge_vlan_group *vg;
 651	struct net_bridge_vlan *vlan;
 652	int ret;
 653
 654	ASSERT_RTNL();
 655
 656	*changed = false;
 657	vg = br_vlan_group(br);
 658	vlan = br_vlan_find(vg, vid);
 659	if (vlan)
 660		return br_vlan_add_existing(br, vg, vlan, flags, changed,
 661					    extack);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 662
 663	vlan = kzalloc(sizeof(*vlan), GFP_KERNEL);
 664	if (!vlan)
 665		return -ENOMEM;
 666
 667	vlan->stats = netdev_alloc_pcpu_stats(struct br_vlan_stats);
 668	if (!vlan->stats) {
 669		kfree(vlan);
 670		return -ENOMEM;
 671	}
 672	vlan->vid = vid;
 673	vlan->flags = flags | BRIDGE_VLAN_INFO_MASTER;
 674	vlan->flags &= ~BRIDGE_VLAN_INFO_PVID;
 675	vlan->br = br;
 676	if (flags & BRIDGE_VLAN_INFO_BRENTRY)
 677		refcount_set(&vlan->refcnt, 1);
 678	ret = __vlan_add(vlan, flags, extack);
 679	if (ret) {
 680		free_percpu(vlan->stats);
 681		kfree(vlan);
 682	} else {
 683		*changed = true;
 684	}
 685
 686	return ret;
 687}
 688
 689/* Must be protected by RTNL.
 690 * Must be called with vid in range from 1 to 4094 inclusive.
 691 */
 692int br_vlan_delete(struct net_bridge *br, u16 vid)
 693{
 694	struct net_bridge_vlan_group *vg;
 695	struct net_bridge_vlan *v;
 696
 697	ASSERT_RTNL();
 698
 699	vg = br_vlan_group(br);
 700	v = br_vlan_find(vg, vid);
 701	if (!v || !br_vlan_is_brentry(v))
 702		return -ENOENT;
 703
 704	br_fdb_find_delete_local(br, NULL, br->dev->dev_addr, vid);
 705	br_fdb_delete_by_port(br, NULL, vid, 0);
 706
 707	vlan_tunnel_info_del(vg, v);
 708
 709	return __vlan_del(v);
 710}
 711
 712void br_vlan_flush(struct net_bridge *br)
 713{
 714	struct net_bridge_vlan_group *vg;
 715
 716	ASSERT_RTNL();
 717
 718	vg = br_vlan_group(br);
 719	__vlan_flush(vg);
 720	RCU_INIT_POINTER(br->vlgrp, NULL);
 721	synchronize_rcu();
 722	__vlan_group_free(vg);
 723}
 724
 725struct net_bridge_vlan *br_vlan_find(struct net_bridge_vlan_group *vg, u16 vid)
 726{
 727	if (!vg)
 728		return NULL;
 729
 730	return br_vlan_lookup(&vg->vlan_hash, vid);
 731}
 732
 733/* Must be protected by RTNL. */
 734static void recalculate_group_addr(struct net_bridge *br)
 735{
 736	if (br_opt_get(br, BROPT_GROUP_ADDR_SET))
 737		return;
 738
 739	spin_lock_bh(&br->lock);
 740	if (!br_opt_get(br, BROPT_VLAN_ENABLED) ||
 741	    br->vlan_proto == htons(ETH_P_8021Q)) {
 742		/* Bridge Group Address */
 743		br->group_addr[5] = 0x00;
 744	} else { /* vlan_enabled && ETH_P_8021AD */
 745		/* Provider Bridge Group Address */
 746		br->group_addr[5] = 0x08;
 747	}
 748	spin_unlock_bh(&br->lock);
 749}
 750
 751/* Must be protected by RTNL. */
 752void br_recalculate_fwd_mask(struct net_bridge *br)
 753{
 754	if (!br_opt_get(br, BROPT_VLAN_ENABLED) ||
 755	    br->vlan_proto == htons(ETH_P_8021Q))
 756		br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
 757	else /* vlan_enabled && ETH_P_8021AD */
 758		br->group_fwd_mask_required = BR_GROUPFWD_8021AD &
 759					      ~(1u << br->group_addr[5]);
 760}
 761
 762int __br_vlan_filter_toggle(struct net_bridge *br, unsigned long val)
 763{
 764	struct switchdev_attr attr = {
 765		.orig_dev = br->dev,
 766		.id = SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING,
 767		.flags = SWITCHDEV_F_SKIP_EOPNOTSUPP,
 768		.u.vlan_filtering = val,
 769	};
 770	int err;
 771
 772	if (br_opt_get(br, BROPT_VLAN_ENABLED) == !!val)
 773		return 0;
 774
 775	err = switchdev_port_attr_set(br->dev, &attr);
 776	if (err && err != -EOPNOTSUPP)
 777		return err;
 778
 779	br_opt_toggle(br, BROPT_VLAN_ENABLED, !!val);
 780	br_manage_promisc(br);
 781	recalculate_group_addr(br);
 782	br_recalculate_fwd_mask(br);
 783
 784	return 0;
 785}
 786
 787int br_vlan_filter_toggle(struct net_bridge *br, unsigned long val)
 788{
 789	return __br_vlan_filter_toggle(br, val);
 790}
 791
 792bool br_vlan_enabled(const struct net_device *dev)
 793{
 794	struct net_bridge *br = netdev_priv(dev);
 795
 796	return br_opt_get(br, BROPT_VLAN_ENABLED);
 797}
 798EXPORT_SYMBOL_GPL(br_vlan_enabled);
 799
 800int br_vlan_get_proto(const struct net_device *dev, u16 *p_proto)
 801{
 802	struct net_bridge *br = netdev_priv(dev);
 803
 804	*p_proto = ntohs(br->vlan_proto);
 805
 806	return 0;
 807}
 808EXPORT_SYMBOL_GPL(br_vlan_get_proto);
 809
 810int __br_vlan_set_proto(struct net_bridge *br, __be16 proto)
 811{
 812	int err = 0;
 813	struct net_bridge_port *p;
 814	struct net_bridge_vlan *vlan;
 815	struct net_bridge_vlan_group *vg;
 816	__be16 oldproto;
 817
 818	if (br->vlan_proto == proto)
 819		return 0;
 820
 821	/* Add VLANs for the new proto to the device filter. */
 822	list_for_each_entry(p, &br->port_list, list) {
 823		vg = nbp_vlan_group(p);
 824		list_for_each_entry(vlan, &vg->vlan_list, vlist) {
 825			err = vlan_vid_add(p->dev, proto, vlan->vid);
 826			if (err)
 827				goto err_filt;
 828		}
 829	}
 830
 831	oldproto = br->vlan_proto;
 832	br->vlan_proto = proto;
 833
 834	recalculate_group_addr(br);
 835	br_recalculate_fwd_mask(br);
 836
 837	/* Delete VLANs for the old proto from the device filter. */
 838	list_for_each_entry(p, &br->port_list, list) {
 839		vg = nbp_vlan_group(p);
 840		list_for_each_entry(vlan, &vg->vlan_list, vlist)
 841			vlan_vid_del(p->dev, oldproto, vlan->vid);
 842	}
 843
 844	return 0;
 845
 846err_filt:
 847	list_for_each_entry_continue_reverse(vlan, &vg->vlan_list, vlist)
 848		vlan_vid_del(p->dev, proto, vlan->vid);
 849
 850	list_for_each_entry_continue_reverse(p, &br->port_list, list) {
 851		vg = nbp_vlan_group(p);
 852		list_for_each_entry(vlan, &vg->vlan_list, vlist)
 853			vlan_vid_del(p->dev, proto, vlan->vid);
 854	}
 855
 856	return err;
 857}
 858
 859int br_vlan_set_proto(struct net_bridge *br, unsigned long val)
 860{
 861	if (val != ETH_P_8021Q && val != ETH_P_8021AD)
 862		return -EPROTONOSUPPORT;
 863
 864	return __br_vlan_set_proto(br, htons(val));
 865}
 866
 867int br_vlan_set_stats(struct net_bridge *br, unsigned long val)
 868{
 869	switch (val) {
 870	case 0:
 871	case 1:
 872		br_opt_toggle(br, BROPT_VLAN_STATS_ENABLED, !!val);
 873		break;
 874	default:
 875		return -EINVAL;
 876	}
 877
 878	return 0;
 879}
 880
 881int br_vlan_set_stats_per_port(struct net_bridge *br, unsigned long val)
 882{
 883	struct net_bridge_port *p;
 884
 885	/* allow to change the option if there are no port vlans configured */
 886	list_for_each_entry(p, &br->port_list, list) {
 887		struct net_bridge_vlan_group *vg = nbp_vlan_group(p);
 888
 889		if (vg->num_vlans)
 890			return -EBUSY;
 891	}
 892
 893	switch (val) {
 894	case 0:
 895	case 1:
 896		br_opt_toggle(br, BROPT_VLAN_STATS_PER_PORT, !!val);
 897		break;
 898	default:
 899		return -EINVAL;
 900	}
 901
 902	return 0;
 903}
 904
 905static bool vlan_default_pvid(struct net_bridge_vlan_group *vg, u16 vid)
 906{
 907	struct net_bridge_vlan *v;
 908
 909	if (vid != vg->pvid)
 910		return false;
 911
 912	v = br_vlan_lookup(&vg->vlan_hash, vid);
 913	if (v && br_vlan_should_use(v) &&
 914	    (v->flags & BRIDGE_VLAN_INFO_UNTAGGED))
 915		return true;
 916
 917	return false;
 918}
 919
 920static void br_vlan_disable_default_pvid(struct net_bridge *br)
 921{
 922	struct net_bridge_port *p;
 923	u16 pvid = br->default_pvid;
 924
 925	/* Disable default_pvid on all ports where it is still
 926	 * configured.
 927	 */
 928	if (vlan_default_pvid(br_vlan_group(br), pvid))
 929		br_vlan_delete(br, pvid);
 930
 931	list_for_each_entry(p, &br->port_list, list) {
 932		if (vlan_default_pvid(nbp_vlan_group(p), pvid))
 933			nbp_vlan_delete(p, pvid);
 934	}
 935
 936	br->default_pvid = 0;
 937}
 938
 939int __br_vlan_set_default_pvid(struct net_bridge *br, u16 pvid,
 940			       struct netlink_ext_ack *extack)
 941{
 942	const struct net_bridge_vlan *pvent;
 943	struct net_bridge_vlan_group *vg;
 944	struct net_bridge_port *p;
 945	unsigned long *changed;
 946	bool vlchange;
 947	u16 old_pvid;
 948	int err = 0;
 949
 950	if (!pvid) {
 951		br_vlan_disable_default_pvid(br);
 952		return 0;
 953	}
 954
 955	changed = bitmap_zalloc(BR_MAX_PORTS, GFP_KERNEL);
 
 956	if (!changed)
 957		return -ENOMEM;
 958
 959	old_pvid = br->default_pvid;
 960
 961	/* Update default_pvid config only if we do not conflict with
 962	 * user configuration.
 963	 */
 964	vg = br_vlan_group(br);
 965	pvent = br_vlan_find(vg, pvid);
 966	if ((!old_pvid || vlan_default_pvid(vg, old_pvid)) &&
 967	    (!pvent || !br_vlan_should_use(pvent))) {
 968		err = br_vlan_add(br, pvid,
 969				  BRIDGE_VLAN_INFO_PVID |
 970				  BRIDGE_VLAN_INFO_UNTAGGED |
 971				  BRIDGE_VLAN_INFO_BRENTRY,
 972				  &vlchange, extack);
 973		if (err)
 974			goto out;
 975		br_vlan_delete(br, old_pvid);
 976		set_bit(0, changed);
 977	}
 978
 979	list_for_each_entry(p, &br->port_list, list) {
 980		/* Update default_pvid config only if we do not conflict with
 981		 * user configuration.
 982		 */
 983		vg = nbp_vlan_group(p);
 984		if ((old_pvid &&
 985		     !vlan_default_pvid(vg, old_pvid)) ||
 986		    br_vlan_find(vg, pvid))
 987			continue;
 988
 989		err = nbp_vlan_add(p, pvid,
 990				   BRIDGE_VLAN_INFO_PVID |
 991				   BRIDGE_VLAN_INFO_UNTAGGED,
 992				   &vlchange, extack);
 993		if (err)
 994			goto err_port;
 995		nbp_vlan_delete(p, old_pvid);
 996		set_bit(p->port_no, changed);
 997	}
 998
 999	br->default_pvid = pvid;
1000
1001out:
1002	bitmap_free(changed);
1003	return err;
1004
1005err_port:
1006	list_for_each_entry_continue_reverse(p, &br->port_list, list) {
1007		if (!test_bit(p->port_no, changed))
1008			continue;
1009
1010		if (old_pvid)
1011			nbp_vlan_add(p, old_pvid,
1012				     BRIDGE_VLAN_INFO_PVID |
1013				     BRIDGE_VLAN_INFO_UNTAGGED,
1014				     &vlchange, NULL);
1015		nbp_vlan_delete(p, pvid);
1016	}
1017
1018	if (test_bit(0, changed)) {
1019		if (old_pvid)
1020			br_vlan_add(br, old_pvid,
1021				    BRIDGE_VLAN_INFO_PVID |
1022				    BRIDGE_VLAN_INFO_UNTAGGED |
1023				    BRIDGE_VLAN_INFO_BRENTRY,
1024				    &vlchange, NULL);
1025		br_vlan_delete(br, pvid);
1026	}
1027	goto out;
1028}
1029
1030int br_vlan_set_default_pvid(struct net_bridge *br, unsigned long val)
1031{
1032	u16 pvid = val;
1033	int err = 0;
1034
1035	if (val >= VLAN_VID_MASK)
1036		return -EINVAL;
1037
1038	if (pvid == br->default_pvid)
1039		goto out;
1040
1041	/* Only allow default pvid change when filtering is disabled */
1042	if (br_opt_get(br, BROPT_VLAN_ENABLED)) {
1043		pr_info_once("Please disable vlan filtering to change default_pvid\n");
1044		err = -EPERM;
1045		goto out;
1046	}
1047	err = __br_vlan_set_default_pvid(br, pvid, NULL);
1048out:
1049	return err;
1050}
1051
1052int br_vlan_init(struct net_bridge *br)
1053{
1054	struct net_bridge_vlan_group *vg;
1055	int ret = -ENOMEM;
 
1056
1057	vg = kzalloc(sizeof(*vg), GFP_KERNEL);
1058	if (!vg)
1059		goto out;
1060	ret = rhashtable_init(&vg->vlan_hash, &br_vlan_rht_params);
1061	if (ret)
1062		goto err_rhtbl;
1063	ret = vlan_tunnel_init(vg);
1064	if (ret)
1065		goto err_tunnel_init;
1066	INIT_LIST_HEAD(&vg->vlan_list);
1067	br->vlan_proto = htons(ETH_P_8021Q);
1068	br->default_pvid = 1;
1069	rcu_assign_pointer(br->vlgrp, vg);
 
 
 
 
 
1070
1071out:
1072	return ret;
1073
 
 
1074err_tunnel_init:
1075	rhashtable_destroy(&vg->vlan_hash);
1076err_rhtbl:
1077	kfree(vg);
1078
1079	goto out;
1080}
1081
1082int nbp_vlan_init(struct net_bridge_port *p, struct netlink_ext_ack *extack)
1083{
1084	struct switchdev_attr attr = {
1085		.orig_dev = p->br->dev,
1086		.id = SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING,
1087		.flags = SWITCHDEV_F_SKIP_EOPNOTSUPP,
1088		.u.vlan_filtering = br_opt_get(p->br, BROPT_VLAN_ENABLED),
1089	};
1090	struct net_bridge_vlan_group *vg;
1091	int ret = -ENOMEM;
1092
1093	vg = kzalloc(sizeof(struct net_bridge_vlan_group), GFP_KERNEL);
1094	if (!vg)
1095		goto out;
1096
1097	ret = switchdev_port_attr_set(p->dev, &attr);
1098	if (ret && ret != -EOPNOTSUPP)
1099		goto err_vlan_enabled;
1100
1101	ret = rhashtable_init(&vg->vlan_hash, &br_vlan_rht_params);
1102	if (ret)
1103		goto err_rhtbl;
1104	ret = vlan_tunnel_init(vg);
1105	if (ret)
1106		goto err_tunnel_init;
1107	INIT_LIST_HEAD(&vg->vlan_list);
1108	rcu_assign_pointer(p->vlgrp, vg);
1109	if (p->br->default_pvid) {
1110		bool changed;
1111
1112		ret = nbp_vlan_add(p, p->br->default_pvid,
1113				   BRIDGE_VLAN_INFO_PVID |
1114				   BRIDGE_VLAN_INFO_UNTAGGED,
1115				   &changed, extack);
1116		if (ret)
1117			goto err_vlan_add;
1118	}
1119out:
1120	return ret;
1121
1122err_vlan_add:
1123	RCU_INIT_POINTER(p->vlgrp, NULL);
1124	synchronize_rcu();
1125	vlan_tunnel_deinit(vg);
1126err_tunnel_init:
1127	rhashtable_destroy(&vg->vlan_hash);
1128err_rhtbl:
1129err_vlan_enabled:
1130	kfree(vg);
1131
1132	goto out;
1133}
1134
1135/* Must be protected by RTNL.
1136 * Must be called with vid in range from 1 to 4094 inclusive.
1137 * changed must be true only if the vlan was created or updated
1138 */
1139int nbp_vlan_add(struct net_bridge_port *port, u16 vid, u16 flags,
1140		 bool *changed, struct netlink_ext_ack *extack)
1141{
 
 
 
 
 
 
 
1142	struct net_bridge_vlan *vlan;
1143	int ret;
1144
1145	ASSERT_RTNL();
1146
1147	*changed = false;
1148	vlan = br_vlan_find(nbp_vlan_group(port), vid);
1149	if (vlan) {
1150		/* Pass the flags to the hardware bridge */
1151		ret = br_switchdev_port_vlan_add(port->dev, vid, flags, extack);
1152		if (ret && ret != -EOPNOTSUPP)
1153			return ret;
1154		*changed = __vlan_add_flags(vlan, flags);
1155
1156		return 0;
1157	}
1158
1159	vlan = kzalloc(sizeof(*vlan), GFP_KERNEL);
1160	if (!vlan)
1161		return -ENOMEM;
1162
1163	vlan->vid = vid;
1164	vlan->port = port;
1165	ret = __vlan_add(vlan, flags, extack);
1166	if (ret)
1167		kfree(vlan);
1168	else
1169		*changed = true;
1170
1171	return ret;
1172}
1173
1174/* Must be protected by RTNL.
1175 * Must be called with vid in range from 1 to 4094 inclusive.
1176 */
1177int nbp_vlan_delete(struct net_bridge_port *port, u16 vid)
1178{
1179	struct net_bridge_vlan *v;
1180
1181	ASSERT_RTNL();
1182
1183	v = br_vlan_find(nbp_vlan_group(port), vid);
1184	if (!v)
1185		return -ENOENT;
1186	br_fdb_find_delete_local(port->br, port, port->dev->dev_addr, vid);
1187	br_fdb_delete_by_port(port->br, port, vid, 0);
1188
1189	return __vlan_del(v);
1190}
1191
1192void nbp_vlan_flush(struct net_bridge_port *port)
1193{
1194	struct net_bridge_vlan_group *vg;
1195
1196	ASSERT_RTNL();
1197
1198	vg = nbp_vlan_group(port);
1199	__vlan_flush(vg);
1200	RCU_INIT_POINTER(port->vlgrp, NULL);
1201	synchronize_rcu();
1202	__vlan_group_free(vg);
1203}
1204
1205void br_vlan_get_stats(const struct net_bridge_vlan *v,
1206		       struct br_vlan_stats *stats)
1207{
1208	int i;
1209
1210	memset(stats, 0, sizeof(*stats));
1211	for_each_possible_cpu(i) {
1212		u64 rxpackets, rxbytes, txpackets, txbytes;
1213		struct br_vlan_stats *cpu_stats;
1214		unsigned int start;
1215
1216		cpu_stats = per_cpu_ptr(v->stats, i);
1217		do {
1218			start = u64_stats_fetch_begin_irq(&cpu_stats->syncp);
1219			rxpackets = cpu_stats->rx_packets;
1220			rxbytes = cpu_stats->rx_bytes;
1221			txbytes = cpu_stats->tx_bytes;
1222			txpackets = cpu_stats->tx_packets;
1223		} while (u64_stats_fetch_retry_irq(&cpu_stats->syncp, start));
1224
1225		stats->rx_packets += rxpackets;
1226		stats->rx_bytes += rxbytes;
1227		stats->tx_bytes += txbytes;
1228		stats->tx_packets += txpackets;
1229	}
1230}
1231
1232static int __br_vlan_get_pvid(const struct net_device *dev,
1233			      struct net_bridge_port *p, u16 *p_pvid)
1234{
1235	struct net_bridge_vlan_group *vg;
1236
1237	if (p)
1238		vg = nbp_vlan_group(p);
1239	else if (netif_is_bridge_master(dev))
1240		vg = br_vlan_group(netdev_priv(dev));
1241	else
1242		return -EINVAL;
1243
1244	*p_pvid = br_get_pvid(vg);
1245	return 0;
1246}
1247
1248int br_vlan_get_pvid(const struct net_device *dev, u16 *p_pvid)
1249{
1250	ASSERT_RTNL();
1251
1252	return __br_vlan_get_pvid(dev, br_port_get_check_rtnl(dev), p_pvid);
1253}
1254EXPORT_SYMBOL_GPL(br_vlan_get_pvid);
1255
1256int br_vlan_get_pvid_rcu(const struct net_device *dev, u16 *p_pvid)
1257{
1258	return __br_vlan_get_pvid(dev, br_port_get_check_rcu(dev), p_pvid);
1259}
1260EXPORT_SYMBOL_GPL(br_vlan_get_pvid_rcu);
1261
1262int br_vlan_get_info(const struct net_device *dev, u16 vid,
1263		     struct bridge_vlan_info *p_vinfo)
1264{
1265	struct net_bridge_vlan_group *vg;
1266	struct net_bridge_vlan *v;
1267	struct net_bridge_port *p;
1268
1269	ASSERT_RTNL();
1270	p = br_port_get_check_rtnl(dev);
1271	if (p)
1272		vg = nbp_vlan_group(p);
1273	else if (netif_is_bridge_master(dev))
1274		vg = br_vlan_group(netdev_priv(dev));
1275	else
1276		return -EINVAL;
1277
1278	v = br_vlan_find(vg, vid);
1279	if (!v)
1280		return -ENOENT;
1281
1282	p_vinfo->vid = vid;
1283	p_vinfo->flags = v->flags;
1284	if (vid == br_get_pvid(vg))
1285		p_vinfo->flags |= BRIDGE_VLAN_INFO_PVID;
1286	return 0;
1287}
1288EXPORT_SYMBOL_GPL(br_vlan_get_info);
1289
1290static int br_vlan_is_bind_vlan_dev(const struct net_device *dev)
1291{
1292	return is_vlan_dev(dev) &&
1293		!!(vlan_dev_priv(dev)->flags & VLAN_FLAG_BRIDGE_BINDING);
1294}
1295
1296static int br_vlan_is_bind_vlan_dev_fn(struct net_device *dev,
1297				       __always_unused void *data)
1298{
1299	return br_vlan_is_bind_vlan_dev(dev);
1300}
1301
1302static bool br_vlan_has_upper_bind_vlan_dev(struct net_device *dev)
1303{
1304	int found;
1305
1306	rcu_read_lock();
1307	found = netdev_walk_all_upper_dev_rcu(dev, br_vlan_is_bind_vlan_dev_fn,
1308					      NULL);
1309	rcu_read_unlock();
1310
1311	return !!found;
1312}
1313
1314struct br_vlan_bind_walk_data {
1315	u16 vid;
1316	struct net_device *result;
1317};
1318
1319static int br_vlan_match_bind_vlan_dev_fn(struct net_device *dev,
1320					  void *data_in)
1321{
1322	struct br_vlan_bind_walk_data *data = data_in;
1323	int found = 0;
1324
1325	if (br_vlan_is_bind_vlan_dev(dev) &&
1326	    vlan_dev_priv(dev)->vlan_id == data->vid) {
1327		data->result = dev;
1328		found = 1;
1329	}
1330
1331	return found;
1332}
1333
1334static struct net_device *
1335br_vlan_get_upper_bind_vlan_dev(struct net_device *dev, u16 vid)
1336{
1337	struct br_vlan_bind_walk_data data = {
1338		.vid = vid,
1339	};
1340
1341	rcu_read_lock();
1342	netdev_walk_all_upper_dev_rcu(dev, br_vlan_match_bind_vlan_dev_fn,
1343				      &data);
1344	rcu_read_unlock();
1345
1346	return data.result;
1347}
1348
1349static bool br_vlan_is_dev_up(const struct net_device *dev)
1350{
1351	return  !!(dev->flags & IFF_UP) && netif_oper_up(dev);
1352}
1353
1354static void br_vlan_set_vlan_dev_state(const struct net_bridge *br,
1355				       struct net_device *vlan_dev)
1356{
1357	u16 vid = vlan_dev_priv(vlan_dev)->vlan_id;
1358	struct net_bridge_vlan_group *vg;
1359	struct net_bridge_port *p;
1360	bool has_carrier = false;
1361
1362	if (!netif_carrier_ok(br->dev)) {
1363		netif_carrier_off(vlan_dev);
1364		return;
1365	}
1366
1367	list_for_each_entry(p, &br->port_list, list) {
1368		vg = nbp_vlan_group(p);
1369		if (br_vlan_find(vg, vid) && br_vlan_is_dev_up(p->dev)) {
1370			has_carrier = true;
1371			break;
1372		}
1373	}
1374
1375	if (has_carrier)
1376		netif_carrier_on(vlan_dev);
1377	else
1378		netif_carrier_off(vlan_dev);
1379}
1380
1381static void br_vlan_set_all_vlan_dev_state(struct net_bridge_port *p)
1382{
1383	struct net_bridge_vlan_group *vg = nbp_vlan_group(p);
1384	struct net_bridge_vlan *vlan;
1385	struct net_device *vlan_dev;
1386
1387	list_for_each_entry(vlan, &vg->vlan_list, vlist) {
1388		vlan_dev = br_vlan_get_upper_bind_vlan_dev(p->br->dev,
1389							   vlan->vid);
1390		if (vlan_dev) {
1391			if (br_vlan_is_dev_up(p->dev)) {
1392				if (netif_carrier_ok(p->br->dev))
1393					netif_carrier_on(vlan_dev);
1394			} else {
1395				br_vlan_set_vlan_dev_state(p->br, vlan_dev);
1396			}
1397		}
1398	}
1399}
1400
1401static void br_vlan_upper_change(struct net_device *dev,
1402				 struct net_device *upper_dev,
1403				 bool linking)
1404{
1405	struct net_bridge *br = netdev_priv(dev);
1406
1407	if (!br_vlan_is_bind_vlan_dev(upper_dev))
1408		return;
1409
1410	if (linking) {
1411		br_vlan_set_vlan_dev_state(br, upper_dev);
1412		br_opt_toggle(br, BROPT_VLAN_BRIDGE_BINDING, true);
1413	} else {
1414		br_opt_toggle(br, BROPT_VLAN_BRIDGE_BINDING,
1415			      br_vlan_has_upper_bind_vlan_dev(dev));
1416	}
1417}
1418
1419struct br_vlan_link_state_walk_data {
1420	struct net_bridge *br;
1421};
1422
1423static int br_vlan_link_state_change_fn(struct net_device *vlan_dev,
1424					void *data_in)
1425{
1426	struct br_vlan_link_state_walk_data *data = data_in;
1427
1428	if (br_vlan_is_bind_vlan_dev(vlan_dev))
1429		br_vlan_set_vlan_dev_state(data->br, vlan_dev);
1430
1431	return 0;
1432}
1433
1434static void br_vlan_link_state_change(struct net_device *dev,
1435				      struct net_bridge *br)
1436{
1437	struct br_vlan_link_state_walk_data data = {
1438		.br = br
1439	};
1440
1441	rcu_read_lock();
1442	netdev_walk_all_upper_dev_rcu(dev, br_vlan_link_state_change_fn,
1443				      &data);
1444	rcu_read_unlock();
1445}
1446
1447/* Must be protected by RTNL. */
1448static void nbp_vlan_set_vlan_dev_state(struct net_bridge_port *p, u16 vid)
1449{
1450	struct net_device *vlan_dev;
1451
1452	if (!br_opt_get(p->br, BROPT_VLAN_BRIDGE_BINDING))
1453		return;
1454
1455	vlan_dev = br_vlan_get_upper_bind_vlan_dev(p->br->dev, vid);
1456	if (vlan_dev)
1457		br_vlan_set_vlan_dev_state(p->br, vlan_dev);
1458}
1459
1460/* Must be protected by RTNL. */
1461int br_vlan_bridge_event(struct net_device *dev, unsigned long event, void *ptr)
1462{
1463	struct netdev_notifier_changeupper_info *info;
1464	struct net_bridge *br = netdev_priv(dev);
1465	bool changed;
1466	int ret = 0;
1467
1468	switch (event) {
1469	case NETDEV_REGISTER:
1470		ret = br_vlan_add(br, br->default_pvid,
1471				  BRIDGE_VLAN_INFO_PVID |
1472				  BRIDGE_VLAN_INFO_UNTAGGED |
1473				  BRIDGE_VLAN_INFO_BRENTRY, &changed, NULL);
1474		break;
1475	case NETDEV_UNREGISTER:
1476		br_vlan_delete(br, br->default_pvid);
1477		break;
1478	case NETDEV_CHANGEUPPER:
1479		info = ptr;
1480		br_vlan_upper_change(dev, info->upper_dev, info->linking);
1481		break;
1482
1483	case NETDEV_CHANGE:
1484	case NETDEV_UP:
1485		if (!br_opt_get(br, BROPT_VLAN_BRIDGE_BINDING))
1486			break;
1487		br_vlan_link_state_change(dev, br);
1488		break;
1489	}
1490
1491	return ret;
1492}
1493
1494/* Must be protected by RTNL. */
1495void br_vlan_port_event(struct net_bridge_port *p, unsigned long event)
1496{
1497	if (!br_opt_get(p->br, BROPT_VLAN_BRIDGE_BINDING))
1498		return;
1499
1500	switch (event) {
1501	case NETDEV_CHANGE:
1502	case NETDEV_DOWN:
1503	case NETDEV_UP:
1504		br_vlan_set_all_vlan_dev_state(p);
1505		break;
1506	}
1507}