Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Device handling code
4 * Linux ethernet bridge
5 *
6 * Authors:
7 * Lennert Buytenhek <buytenh@gnu.org>
8 */
9
10#include <linux/kernel.h>
11#include <linux/netdevice.h>
12#include <linux/netpoll.h>
13#include <linux/etherdevice.h>
14#include <linux/ethtool.h>
15#include <linux/list.h>
16#include <linux/netfilter_bridge.h>
17
18#include <linux/uaccess.h>
19#include "br_private.h"
20
21#define COMMON_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HIGHDMA | \
22 NETIF_F_GSO_MASK | NETIF_F_HW_CSUM)
23
24const struct nf_br_ops __rcu *nf_br_ops __read_mostly;
25EXPORT_SYMBOL_GPL(nf_br_ops);
26
27/* net device transmit always called with BH disabled */
28netdev_tx_t br_dev_xmit(struct sk_buff *skb, struct net_device *dev)
29{
30 enum skb_drop_reason reason = pskb_may_pull_reason(skb, ETH_HLEN);
31 struct net_bridge_mcast_port *pmctx_null = NULL;
32 struct net_bridge *br = netdev_priv(dev);
33 struct net_bridge_mcast *brmctx = &br->multicast_ctx;
34 struct net_bridge_fdb_entry *dst;
35 struct net_bridge_mdb_entry *mdst;
36 const struct nf_br_ops *nf_ops;
37 u8 state = BR_STATE_FORWARDING;
38 struct net_bridge_vlan *vlan;
39 const unsigned char *dest;
40 u16 vid = 0;
41
42 if (unlikely(reason != SKB_NOT_DROPPED_YET)) {
43 kfree_skb_reason(skb, reason);
44 return NETDEV_TX_OK;
45 }
46
47 memset(skb->cb, 0, sizeof(struct br_input_skb_cb));
48 br_tc_skb_miss_set(skb, false);
49
50 rcu_read_lock();
51 nf_ops = rcu_dereference(nf_br_ops);
52 if (nf_ops && nf_ops->br_dev_xmit_hook(skb)) {
53 rcu_read_unlock();
54 return NETDEV_TX_OK;
55 }
56
57 dev_sw_netstats_tx_add(dev, 1, skb->len);
58
59 br_switchdev_frame_unmark(skb);
60 BR_INPUT_SKB_CB(skb)->brdev = dev;
61 BR_INPUT_SKB_CB(skb)->frag_max_size = 0;
62
63 skb_reset_mac_header(skb);
64 skb_pull(skb, ETH_HLEN);
65
66 if (!br_allowed_ingress(br, br_vlan_group_rcu(br), skb, &vid,
67 &state, &vlan))
68 goto out;
69
70 if (IS_ENABLED(CONFIG_INET) &&
71 (eth_hdr(skb)->h_proto == htons(ETH_P_ARP) ||
72 eth_hdr(skb)->h_proto == htons(ETH_P_RARP)) &&
73 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED)) {
74 br_do_proxy_suppress_arp(skb, br, vid, NULL);
75 } else if (IS_ENABLED(CONFIG_IPV6) &&
76 skb->protocol == htons(ETH_P_IPV6) &&
77 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED) &&
78 pskb_may_pull(skb, sizeof(struct ipv6hdr) +
79 sizeof(struct nd_msg)) &&
80 ipv6_hdr(skb)->nexthdr == IPPROTO_ICMPV6) {
81 struct nd_msg *msg, _msg;
82
83 msg = br_is_nd_neigh_msg(skb, &_msg);
84 if (msg)
85 br_do_suppress_nd(skb, br, vid, NULL, msg);
86 }
87
88 dest = eth_hdr(skb)->h_dest;
89 if (is_broadcast_ether_addr(dest)) {
90 br_flood(br, skb, BR_PKT_BROADCAST, false, true, vid);
91 } else if (is_multicast_ether_addr(dest)) {
92 if (unlikely(netpoll_tx_running(dev))) {
93 br_flood(br, skb, BR_PKT_MULTICAST, false, true, vid);
94 goto out;
95 }
96 if (br_multicast_rcv(&brmctx, &pmctx_null, vlan, skb, vid)) {
97 kfree_skb(skb);
98 goto out;
99 }
100
101 mdst = br_mdb_entry_skb_get(brmctx, skb, vid);
102 if ((mdst || BR_INPUT_SKB_CB_MROUTERS_ONLY(skb)) &&
103 br_multicast_querier_exists(brmctx, eth_hdr(skb), mdst))
104 br_multicast_flood(mdst, skb, brmctx, false, true);
105 else
106 br_flood(br, skb, BR_PKT_MULTICAST, false, true, vid);
107 } else if ((dst = br_fdb_find_rcu(br, dest, vid)) != NULL) {
108 br_forward(dst->dst, skb, false, true);
109 } else {
110 br_flood(br, skb, BR_PKT_UNICAST, false, true, vid);
111 }
112out:
113 rcu_read_unlock();
114 return NETDEV_TX_OK;
115}
116
117static int br_dev_init(struct net_device *dev)
118{
119 struct net_bridge *br = netdev_priv(dev);
120 int err;
121
122 err = br_fdb_hash_init(br);
123 if (err)
124 return err;
125
126 err = br_mdb_hash_init(br);
127 if (err) {
128 br_fdb_hash_fini(br);
129 return err;
130 }
131
132 err = br_vlan_init(br);
133 if (err) {
134 br_mdb_hash_fini(br);
135 br_fdb_hash_fini(br);
136 return err;
137 }
138
139 err = br_multicast_init_stats(br);
140 if (err) {
141 br_vlan_flush(br);
142 br_mdb_hash_fini(br);
143 br_fdb_hash_fini(br);
144 return err;
145 }
146
147 netdev_lockdep_set_classes(dev);
148 return 0;
149}
150
151static void br_dev_uninit(struct net_device *dev)
152{
153 struct net_bridge *br = netdev_priv(dev);
154
155 br_multicast_dev_del(br);
156 br_multicast_uninit_stats(br);
157 br_vlan_flush(br);
158 br_mdb_hash_fini(br);
159 br_fdb_hash_fini(br);
160}
161
162static int br_dev_open(struct net_device *dev)
163{
164 struct net_bridge *br = netdev_priv(dev);
165
166 netdev_update_features(dev);
167 netif_start_queue(dev);
168 br_stp_enable_bridge(br);
169 br_multicast_open(br);
170
171 if (br_opt_get(br, BROPT_MULTICAST_ENABLED))
172 br_multicast_join_snoopers(br);
173
174 return 0;
175}
176
177static void br_dev_set_multicast_list(struct net_device *dev)
178{
179}
180
181static void br_dev_change_rx_flags(struct net_device *dev, int change)
182{
183 if (change & IFF_PROMISC)
184 br_manage_promisc(netdev_priv(dev));
185}
186
187static int br_dev_stop(struct net_device *dev)
188{
189 struct net_bridge *br = netdev_priv(dev);
190
191 br_stp_disable_bridge(br);
192 br_multicast_stop(br);
193
194 if (br_opt_get(br, BROPT_MULTICAST_ENABLED))
195 br_multicast_leave_snoopers(br);
196
197 netif_stop_queue(dev);
198
199 return 0;
200}
201
202static int br_change_mtu(struct net_device *dev, int new_mtu)
203{
204 struct net_bridge *br = netdev_priv(dev);
205
206 WRITE_ONCE(dev->mtu, new_mtu);
207
208 /* this flag will be cleared if the MTU was automatically adjusted */
209 br_opt_toggle(br, BROPT_MTU_SET_BY_USER, true);
210#if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
211 /* remember the MTU in the rtable for PMTU */
212 dst_metric_set(&br->fake_rtable.dst, RTAX_MTU, new_mtu);
213#endif
214
215 return 0;
216}
217
218/* Allow setting mac address to any valid ethernet address. */
219static int br_set_mac_address(struct net_device *dev, void *p)
220{
221 struct net_bridge *br = netdev_priv(dev);
222 struct sockaddr *addr = p;
223
224 if (!is_valid_ether_addr(addr->sa_data))
225 return -EADDRNOTAVAIL;
226
227 /* dev_set_mac_addr() can be called by a master device on bridge's
228 * NETDEV_UNREGISTER, but since it's being destroyed do nothing
229 */
230 if (dev->reg_state != NETREG_REGISTERED)
231 return -EBUSY;
232
233 spin_lock_bh(&br->lock);
234 if (!ether_addr_equal(dev->dev_addr, addr->sa_data)) {
235 /* Mac address will be changed in br_stp_change_bridge_id(). */
236 br_stp_change_bridge_id(br, addr->sa_data);
237 }
238 spin_unlock_bh(&br->lock);
239
240 return 0;
241}
242
243static void br_getinfo(struct net_device *dev, struct ethtool_drvinfo *info)
244{
245 strscpy(info->driver, "bridge", sizeof(info->driver));
246 strscpy(info->version, BR_VERSION, sizeof(info->version));
247 strscpy(info->fw_version, "N/A", sizeof(info->fw_version));
248 strscpy(info->bus_info, "N/A", sizeof(info->bus_info));
249}
250
251static int br_get_link_ksettings(struct net_device *dev,
252 struct ethtool_link_ksettings *cmd)
253{
254 struct net_bridge *br = netdev_priv(dev);
255 struct net_bridge_port *p;
256
257 cmd->base.duplex = DUPLEX_UNKNOWN;
258 cmd->base.port = PORT_OTHER;
259 cmd->base.speed = SPEED_UNKNOWN;
260
261 list_for_each_entry(p, &br->port_list, list) {
262 struct ethtool_link_ksettings ecmd;
263 struct net_device *pdev = p->dev;
264
265 if (!netif_running(pdev) || !netif_oper_up(pdev))
266 continue;
267
268 if (__ethtool_get_link_ksettings(pdev, &ecmd))
269 continue;
270
271 if (ecmd.base.speed == (__u32)SPEED_UNKNOWN)
272 continue;
273
274 if (cmd->base.speed == (__u32)SPEED_UNKNOWN ||
275 cmd->base.speed < ecmd.base.speed)
276 cmd->base.speed = ecmd.base.speed;
277 }
278
279 return 0;
280}
281
282static netdev_features_t br_fix_features(struct net_device *dev,
283 netdev_features_t features)
284{
285 struct net_bridge *br = netdev_priv(dev);
286
287 return br_features_recompute(br, features);
288}
289
290#ifdef CONFIG_NET_POLL_CONTROLLER
291static void br_poll_controller(struct net_device *br_dev)
292{
293}
294
295static void br_netpoll_cleanup(struct net_device *dev)
296{
297 struct net_bridge *br = netdev_priv(dev);
298 struct net_bridge_port *p;
299
300 list_for_each_entry(p, &br->port_list, list)
301 br_netpoll_disable(p);
302}
303
304static int __br_netpoll_enable(struct net_bridge_port *p)
305{
306 struct netpoll *np;
307 int err;
308
309 np = kzalloc(sizeof(*p->np), GFP_KERNEL);
310 if (!np)
311 return -ENOMEM;
312
313 err = __netpoll_setup(np, p->dev);
314 if (err) {
315 kfree(np);
316 return err;
317 }
318
319 p->np = np;
320 return err;
321}
322
323int br_netpoll_enable(struct net_bridge_port *p)
324{
325 if (!p->br->dev->npinfo)
326 return 0;
327
328 return __br_netpoll_enable(p);
329}
330
331static int br_netpoll_setup(struct net_device *dev)
332{
333 struct net_bridge *br = netdev_priv(dev);
334 struct net_bridge_port *p;
335 int err = 0;
336
337 list_for_each_entry(p, &br->port_list, list) {
338 if (!p->dev)
339 continue;
340 err = __br_netpoll_enable(p);
341 if (err)
342 goto fail;
343 }
344
345out:
346 return err;
347
348fail:
349 br_netpoll_cleanup(dev);
350 goto out;
351}
352
353void br_netpoll_disable(struct net_bridge_port *p)
354{
355 struct netpoll *np = p->np;
356
357 if (!np)
358 return;
359
360 p->np = NULL;
361
362 __netpoll_free(np);
363}
364
365#endif
366
367static int br_add_slave(struct net_device *dev, struct net_device *slave_dev,
368 struct netlink_ext_ack *extack)
369
370{
371 struct net_bridge *br = netdev_priv(dev);
372
373 return br_add_if(br, slave_dev, extack);
374}
375
376static int br_del_slave(struct net_device *dev, struct net_device *slave_dev)
377{
378 struct net_bridge *br = netdev_priv(dev);
379
380 return br_del_if(br, slave_dev);
381}
382
383static int br_fill_forward_path(struct net_device_path_ctx *ctx,
384 struct net_device_path *path)
385{
386 struct net_bridge_fdb_entry *f;
387 struct net_bridge_port *dst;
388 struct net_bridge *br;
389
390 if (netif_is_bridge_port(ctx->dev))
391 return -1;
392
393 br = netdev_priv(ctx->dev);
394
395 br_vlan_fill_forward_path_pvid(br, ctx, path);
396
397 f = br_fdb_find_rcu(br, ctx->daddr, path->bridge.vlan_id);
398 if (!f)
399 return -1;
400
401 dst = READ_ONCE(f->dst);
402 if (!dst)
403 return -1;
404
405 if (br_vlan_fill_forward_path_mode(br, dst, path))
406 return -1;
407
408 path->type = DEV_PATH_BRIDGE;
409 path->dev = dst->br->dev;
410 ctx->dev = dst->dev;
411
412 switch (path->bridge.vlan_mode) {
413 case DEV_PATH_BR_VLAN_TAG:
414 if (ctx->num_vlans >= ARRAY_SIZE(ctx->vlan))
415 return -ENOSPC;
416 ctx->vlan[ctx->num_vlans].id = path->bridge.vlan_id;
417 ctx->vlan[ctx->num_vlans].proto = path->bridge.vlan_proto;
418 ctx->num_vlans++;
419 break;
420 case DEV_PATH_BR_VLAN_UNTAG_HW:
421 case DEV_PATH_BR_VLAN_UNTAG:
422 ctx->num_vlans--;
423 break;
424 case DEV_PATH_BR_VLAN_KEEP:
425 break;
426 }
427
428 return 0;
429}
430
431static const struct ethtool_ops br_ethtool_ops = {
432 .get_drvinfo = br_getinfo,
433 .get_link = ethtool_op_get_link,
434 .get_link_ksettings = br_get_link_ksettings,
435};
436
437static const struct net_device_ops br_netdev_ops = {
438 .ndo_open = br_dev_open,
439 .ndo_stop = br_dev_stop,
440 .ndo_init = br_dev_init,
441 .ndo_uninit = br_dev_uninit,
442 .ndo_start_xmit = br_dev_xmit,
443 .ndo_get_stats64 = dev_get_tstats64,
444 .ndo_set_mac_address = br_set_mac_address,
445 .ndo_set_rx_mode = br_dev_set_multicast_list,
446 .ndo_change_rx_flags = br_dev_change_rx_flags,
447 .ndo_change_mtu = br_change_mtu,
448 .ndo_siocdevprivate = br_dev_siocdevprivate,
449#ifdef CONFIG_NET_POLL_CONTROLLER
450 .ndo_netpoll_setup = br_netpoll_setup,
451 .ndo_netpoll_cleanup = br_netpoll_cleanup,
452 .ndo_poll_controller = br_poll_controller,
453#endif
454 .ndo_add_slave = br_add_slave,
455 .ndo_del_slave = br_del_slave,
456 .ndo_fix_features = br_fix_features,
457 .ndo_fdb_add = br_fdb_add,
458 .ndo_fdb_del = br_fdb_delete,
459 .ndo_fdb_del_bulk = br_fdb_delete_bulk,
460 .ndo_fdb_dump = br_fdb_dump,
461 .ndo_fdb_get = br_fdb_get,
462 .ndo_mdb_add = br_mdb_add,
463 .ndo_mdb_del = br_mdb_del,
464 .ndo_mdb_del_bulk = br_mdb_del_bulk,
465 .ndo_mdb_dump = br_mdb_dump,
466 .ndo_mdb_get = br_mdb_get,
467 .ndo_bridge_getlink = br_getlink,
468 .ndo_bridge_setlink = br_setlink,
469 .ndo_bridge_dellink = br_dellink,
470 .ndo_features_check = passthru_features_check,
471 .ndo_fill_forward_path = br_fill_forward_path,
472};
473
474static const struct device_type br_type = {
475 .name = "bridge",
476};
477
478void br_dev_setup(struct net_device *dev)
479{
480 struct net_bridge *br = netdev_priv(dev);
481
482 eth_hw_addr_random(dev);
483 ether_setup(dev);
484
485 dev->netdev_ops = &br_netdev_ops;
486 dev->needs_free_netdev = true;
487 dev->ethtool_ops = &br_ethtool_ops;
488 SET_NETDEV_DEVTYPE(dev, &br_type);
489 dev->priv_flags = IFF_EBRIDGE | IFF_NO_QUEUE;
490 dev->lltx = true;
491 dev->netns_local = true;
492
493 dev->features = COMMON_FEATURES | NETIF_F_HW_VLAN_CTAG_TX |
494 NETIF_F_HW_VLAN_STAG_TX;
495 dev->hw_features = COMMON_FEATURES | NETIF_F_HW_VLAN_CTAG_TX |
496 NETIF_F_HW_VLAN_STAG_TX;
497 dev->vlan_features = COMMON_FEATURES;
498 dev->pcpu_stat_type = NETDEV_PCPU_STAT_TSTATS;
499
500 br->dev = dev;
501 spin_lock_init(&br->lock);
502 INIT_LIST_HEAD(&br->port_list);
503 INIT_HLIST_HEAD(&br->fdb_list);
504 INIT_HLIST_HEAD(&br->frame_type_list);
505#if IS_ENABLED(CONFIG_BRIDGE_MRP)
506 INIT_HLIST_HEAD(&br->mrp_list);
507#endif
508#if IS_ENABLED(CONFIG_BRIDGE_CFM)
509 INIT_HLIST_HEAD(&br->mep_list);
510#endif
511 spin_lock_init(&br->hash_lock);
512
513 br->bridge_id.prio[0] = 0x80;
514 br->bridge_id.prio[1] = 0x00;
515
516 ether_addr_copy(br->group_addr, eth_stp_addr);
517
518 br->stp_enabled = BR_NO_STP;
519 br->group_fwd_mask = BR_GROUPFWD_DEFAULT;
520 br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
521
522 br->designated_root = br->bridge_id;
523 br->bridge_max_age = br->max_age = 20 * HZ;
524 br->bridge_hello_time = br->hello_time = 2 * HZ;
525 br->bridge_forward_delay = br->forward_delay = 15 * HZ;
526 br->bridge_ageing_time = br->ageing_time = BR_DEFAULT_AGEING_TIME;
527 dev->max_mtu = ETH_MAX_MTU;
528
529 br_netfilter_rtable_init(br);
530 br_stp_timer_init(br);
531 br_multicast_init(br);
532 INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup);
533}
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Device handling code
4 * Linux ethernet bridge
5 *
6 * Authors:
7 * Lennert Buytenhek <buytenh@gnu.org>
8 */
9
10#include <linux/kernel.h>
11#include <linux/netdevice.h>
12#include <linux/netpoll.h>
13#include <linux/etherdevice.h>
14#include <linux/ethtool.h>
15#include <linux/list.h>
16#include <linux/netfilter_bridge.h>
17
18#include <linux/uaccess.h>
19#include "br_private.h"
20
21#define COMMON_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HIGHDMA | \
22 NETIF_F_GSO_MASK | NETIF_F_HW_CSUM)
23
24const struct nf_br_ops __rcu *nf_br_ops __read_mostly;
25EXPORT_SYMBOL_GPL(nf_br_ops);
26
27/* net device transmit always called with BH disabled */
28netdev_tx_t br_dev_xmit(struct sk_buff *skb, struct net_device *dev)
29{
30 struct net_bridge *br = netdev_priv(dev);
31 struct net_bridge_fdb_entry *dst;
32 struct net_bridge_mdb_entry *mdst;
33 struct pcpu_sw_netstats *brstats = this_cpu_ptr(br->stats);
34 const struct nf_br_ops *nf_ops;
35 const unsigned char *dest;
36 struct ethhdr *eth;
37 u16 vid = 0;
38
39 rcu_read_lock();
40 nf_ops = rcu_dereference(nf_br_ops);
41 if (nf_ops && nf_ops->br_dev_xmit_hook(skb)) {
42 rcu_read_unlock();
43 return NETDEV_TX_OK;
44 }
45
46 u64_stats_update_begin(&brstats->syncp);
47 brstats->tx_packets++;
48 brstats->tx_bytes += skb->len;
49 u64_stats_update_end(&brstats->syncp);
50
51 br_switchdev_frame_unmark(skb);
52 BR_INPUT_SKB_CB(skb)->brdev = dev;
53 BR_INPUT_SKB_CB(skb)->frag_max_size = 0;
54
55 skb_reset_mac_header(skb);
56 eth = eth_hdr(skb);
57 skb_pull(skb, ETH_HLEN);
58
59 if (!br_allowed_ingress(br, br_vlan_group_rcu(br), skb, &vid))
60 goto out;
61
62 if (IS_ENABLED(CONFIG_INET) &&
63 (eth->h_proto == htons(ETH_P_ARP) ||
64 eth->h_proto == htons(ETH_P_RARP)) &&
65 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED)) {
66 br_do_proxy_suppress_arp(skb, br, vid, NULL);
67 } else if (IS_ENABLED(CONFIG_IPV6) &&
68 skb->protocol == htons(ETH_P_IPV6) &&
69 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED) &&
70 pskb_may_pull(skb, sizeof(struct ipv6hdr) +
71 sizeof(struct nd_msg)) &&
72 ipv6_hdr(skb)->nexthdr == IPPROTO_ICMPV6) {
73 struct nd_msg *msg, _msg;
74
75 msg = br_is_nd_neigh_msg(skb, &_msg);
76 if (msg)
77 br_do_suppress_nd(skb, br, vid, NULL, msg);
78 }
79
80 dest = eth_hdr(skb)->h_dest;
81 if (is_broadcast_ether_addr(dest)) {
82 br_flood(br, skb, BR_PKT_BROADCAST, false, true);
83 } else if (is_multicast_ether_addr(dest)) {
84 if (unlikely(netpoll_tx_running(dev))) {
85 br_flood(br, skb, BR_PKT_MULTICAST, false, true);
86 goto out;
87 }
88 if (br_multicast_rcv(br, NULL, skb, vid)) {
89 kfree_skb(skb);
90 goto out;
91 }
92
93 mdst = br_mdb_get(br, skb, vid);
94 if ((mdst || BR_INPUT_SKB_CB_MROUTERS_ONLY(skb)) &&
95 br_multicast_querier_exists(br, eth_hdr(skb)))
96 br_multicast_flood(mdst, skb, false, true);
97 else
98 br_flood(br, skb, BR_PKT_MULTICAST, false, true);
99 } else if ((dst = br_fdb_find_rcu(br, dest, vid)) != NULL) {
100 br_forward(dst->dst, skb, false, true);
101 } else {
102 br_flood(br, skb, BR_PKT_UNICAST, false, true);
103 }
104out:
105 rcu_read_unlock();
106 return NETDEV_TX_OK;
107}
108
109static int br_dev_init(struct net_device *dev)
110{
111 struct net_bridge *br = netdev_priv(dev);
112 int err;
113
114 br->stats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
115 if (!br->stats)
116 return -ENOMEM;
117
118 err = br_fdb_hash_init(br);
119 if (err) {
120 free_percpu(br->stats);
121 return err;
122 }
123
124 err = br_mdb_hash_init(br);
125 if (err) {
126 free_percpu(br->stats);
127 br_fdb_hash_fini(br);
128 return err;
129 }
130
131 err = br_vlan_init(br);
132 if (err) {
133 free_percpu(br->stats);
134 br_mdb_hash_fini(br);
135 br_fdb_hash_fini(br);
136 return err;
137 }
138
139 err = br_multicast_init_stats(br);
140 if (err) {
141 free_percpu(br->stats);
142 br_vlan_flush(br);
143 br_mdb_hash_fini(br);
144 br_fdb_hash_fini(br);
145 }
146
147 return err;
148}
149
150static void br_dev_uninit(struct net_device *dev)
151{
152 struct net_bridge *br = netdev_priv(dev);
153
154 br_multicast_dev_del(br);
155 br_multicast_uninit_stats(br);
156 br_vlan_flush(br);
157 br_mdb_hash_fini(br);
158 br_fdb_hash_fini(br);
159 free_percpu(br->stats);
160}
161
162static int br_dev_open(struct net_device *dev)
163{
164 struct net_bridge *br = netdev_priv(dev);
165
166 netdev_update_features(dev);
167 netif_start_queue(dev);
168 br_stp_enable_bridge(br);
169 br_multicast_open(br);
170
171 return 0;
172}
173
174static void br_dev_set_multicast_list(struct net_device *dev)
175{
176}
177
178static void br_dev_change_rx_flags(struct net_device *dev, int change)
179{
180 if (change & IFF_PROMISC)
181 br_manage_promisc(netdev_priv(dev));
182}
183
184static int br_dev_stop(struct net_device *dev)
185{
186 struct net_bridge *br = netdev_priv(dev);
187
188 br_stp_disable_bridge(br);
189 br_multicast_stop(br);
190
191 netif_stop_queue(dev);
192
193 return 0;
194}
195
196static void br_get_stats64(struct net_device *dev,
197 struct rtnl_link_stats64 *stats)
198{
199 struct net_bridge *br = netdev_priv(dev);
200 struct pcpu_sw_netstats tmp, sum = { 0 };
201 unsigned int cpu;
202
203 for_each_possible_cpu(cpu) {
204 unsigned int start;
205 const struct pcpu_sw_netstats *bstats
206 = per_cpu_ptr(br->stats, cpu);
207 do {
208 start = u64_stats_fetch_begin_irq(&bstats->syncp);
209 memcpy(&tmp, bstats, sizeof(tmp));
210 } while (u64_stats_fetch_retry_irq(&bstats->syncp, start));
211 sum.tx_bytes += tmp.tx_bytes;
212 sum.tx_packets += tmp.tx_packets;
213 sum.rx_bytes += tmp.rx_bytes;
214 sum.rx_packets += tmp.rx_packets;
215 }
216
217 stats->tx_bytes = sum.tx_bytes;
218 stats->tx_packets = sum.tx_packets;
219 stats->rx_bytes = sum.rx_bytes;
220 stats->rx_packets = sum.rx_packets;
221}
222
223static int br_change_mtu(struct net_device *dev, int new_mtu)
224{
225 struct net_bridge *br = netdev_priv(dev);
226
227 dev->mtu = new_mtu;
228
229 /* this flag will be cleared if the MTU was automatically adjusted */
230 br_opt_toggle(br, BROPT_MTU_SET_BY_USER, true);
231#if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
232 /* remember the MTU in the rtable for PMTU */
233 dst_metric_set(&br->fake_rtable.dst, RTAX_MTU, new_mtu);
234#endif
235
236 return 0;
237}
238
239/* Allow setting mac address to any valid ethernet address. */
240static int br_set_mac_address(struct net_device *dev, void *p)
241{
242 struct net_bridge *br = netdev_priv(dev);
243 struct sockaddr *addr = p;
244
245 if (!is_valid_ether_addr(addr->sa_data))
246 return -EADDRNOTAVAIL;
247
248 spin_lock_bh(&br->lock);
249 if (!ether_addr_equal(dev->dev_addr, addr->sa_data)) {
250 /* Mac address will be changed in br_stp_change_bridge_id(). */
251 br_stp_change_bridge_id(br, addr->sa_data);
252 }
253 spin_unlock_bh(&br->lock);
254
255 return 0;
256}
257
258static void br_getinfo(struct net_device *dev, struct ethtool_drvinfo *info)
259{
260 strlcpy(info->driver, "bridge", sizeof(info->driver));
261 strlcpy(info->version, BR_VERSION, sizeof(info->version));
262 strlcpy(info->fw_version, "N/A", sizeof(info->fw_version));
263 strlcpy(info->bus_info, "N/A", sizeof(info->bus_info));
264}
265
266static netdev_features_t br_fix_features(struct net_device *dev,
267 netdev_features_t features)
268{
269 struct net_bridge *br = netdev_priv(dev);
270
271 return br_features_recompute(br, features);
272}
273
274#ifdef CONFIG_NET_POLL_CONTROLLER
275static void br_poll_controller(struct net_device *br_dev)
276{
277}
278
279static void br_netpoll_cleanup(struct net_device *dev)
280{
281 struct net_bridge *br = netdev_priv(dev);
282 struct net_bridge_port *p;
283
284 list_for_each_entry(p, &br->port_list, list)
285 br_netpoll_disable(p);
286}
287
288static int __br_netpoll_enable(struct net_bridge_port *p)
289{
290 struct netpoll *np;
291 int err;
292
293 np = kzalloc(sizeof(*p->np), GFP_KERNEL);
294 if (!np)
295 return -ENOMEM;
296
297 err = __netpoll_setup(np, p->dev);
298 if (err) {
299 kfree(np);
300 return err;
301 }
302
303 p->np = np;
304 return err;
305}
306
307int br_netpoll_enable(struct net_bridge_port *p)
308{
309 if (!p->br->dev->npinfo)
310 return 0;
311
312 return __br_netpoll_enable(p);
313}
314
315static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni)
316{
317 struct net_bridge *br = netdev_priv(dev);
318 struct net_bridge_port *p;
319 int err = 0;
320
321 list_for_each_entry(p, &br->port_list, list) {
322 if (!p->dev)
323 continue;
324 err = __br_netpoll_enable(p);
325 if (err)
326 goto fail;
327 }
328
329out:
330 return err;
331
332fail:
333 br_netpoll_cleanup(dev);
334 goto out;
335}
336
337void br_netpoll_disable(struct net_bridge_port *p)
338{
339 struct netpoll *np = p->np;
340
341 if (!np)
342 return;
343
344 p->np = NULL;
345
346 __netpoll_free(np);
347}
348
349#endif
350
351static int br_add_slave(struct net_device *dev, struct net_device *slave_dev,
352 struct netlink_ext_ack *extack)
353
354{
355 struct net_bridge *br = netdev_priv(dev);
356
357 return br_add_if(br, slave_dev, extack);
358}
359
360static int br_del_slave(struct net_device *dev, struct net_device *slave_dev)
361{
362 struct net_bridge *br = netdev_priv(dev);
363
364 return br_del_if(br, slave_dev);
365}
366
367static const struct ethtool_ops br_ethtool_ops = {
368 .get_drvinfo = br_getinfo,
369 .get_link = ethtool_op_get_link,
370};
371
372static const struct net_device_ops br_netdev_ops = {
373 .ndo_open = br_dev_open,
374 .ndo_stop = br_dev_stop,
375 .ndo_init = br_dev_init,
376 .ndo_uninit = br_dev_uninit,
377 .ndo_start_xmit = br_dev_xmit,
378 .ndo_get_stats64 = br_get_stats64,
379 .ndo_set_mac_address = br_set_mac_address,
380 .ndo_set_rx_mode = br_dev_set_multicast_list,
381 .ndo_change_rx_flags = br_dev_change_rx_flags,
382 .ndo_change_mtu = br_change_mtu,
383 .ndo_do_ioctl = br_dev_ioctl,
384#ifdef CONFIG_NET_POLL_CONTROLLER
385 .ndo_netpoll_setup = br_netpoll_setup,
386 .ndo_netpoll_cleanup = br_netpoll_cleanup,
387 .ndo_poll_controller = br_poll_controller,
388#endif
389 .ndo_add_slave = br_add_slave,
390 .ndo_del_slave = br_del_slave,
391 .ndo_fix_features = br_fix_features,
392 .ndo_fdb_add = br_fdb_add,
393 .ndo_fdb_del = br_fdb_delete,
394 .ndo_fdb_dump = br_fdb_dump,
395 .ndo_fdb_get = br_fdb_get,
396 .ndo_bridge_getlink = br_getlink,
397 .ndo_bridge_setlink = br_setlink,
398 .ndo_bridge_dellink = br_dellink,
399 .ndo_features_check = passthru_features_check,
400};
401
402static struct device_type br_type = {
403 .name = "bridge",
404};
405
406void br_dev_setup(struct net_device *dev)
407{
408 struct net_bridge *br = netdev_priv(dev);
409
410 eth_hw_addr_random(dev);
411 ether_setup(dev);
412
413 dev->netdev_ops = &br_netdev_ops;
414 dev->needs_free_netdev = true;
415 dev->ethtool_ops = &br_ethtool_ops;
416 SET_NETDEV_DEVTYPE(dev, &br_type);
417 dev->priv_flags = IFF_EBRIDGE | IFF_NO_QUEUE;
418
419 dev->features = COMMON_FEATURES | NETIF_F_LLTX | NETIF_F_NETNS_LOCAL |
420 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_STAG_TX;
421 dev->hw_features = COMMON_FEATURES | NETIF_F_HW_VLAN_CTAG_TX |
422 NETIF_F_HW_VLAN_STAG_TX;
423 dev->vlan_features = COMMON_FEATURES;
424
425 br->dev = dev;
426 spin_lock_init(&br->lock);
427 INIT_LIST_HEAD(&br->port_list);
428 INIT_HLIST_HEAD(&br->fdb_list);
429 spin_lock_init(&br->hash_lock);
430
431 br->bridge_id.prio[0] = 0x80;
432 br->bridge_id.prio[1] = 0x00;
433
434 ether_addr_copy(br->group_addr, eth_stp_addr);
435
436 br->stp_enabled = BR_NO_STP;
437 br->group_fwd_mask = BR_GROUPFWD_DEFAULT;
438 br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
439
440 br->designated_root = br->bridge_id;
441 br->bridge_max_age = br->max_age = 20 * HZ;
442 br->bridge_hello_time = br->hello_time = 2 * HZ;
443 br->bridge_forward_delay = br->forward_delay = 15 * HZ;
444 br->bridge_ageing_time = br->ageing_time = BR_DEFAULT_AGEING_TIME;
445 dev->max_mtu = ETH_MAX_MTU;
446
447 br_netfilter_rtable_init(br);
448 br_stp_timer_init(br);
449 br_multicast_init(br);
450 INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup);
451}