Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Device handling code
4 * Linux ethernet bridge
5 *
6 * Authors:
7 * Lennert Buytenhek <buytenh@gnu.org>
8 */
9
10#include <linux/kernel.h>
11#include <linux/netdevice.h>
12#include <linux/netpoll.h>
13#include <linux/etherdevice.h>
14#include <linux/ethtool.h>
15#include <linux/list.h>
16#include <linux/netfilter_bridge.h>
17
18#include <linux/uaccess.h>
19#include "br_private.h"
20
21#define COMMON_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HIGHDMA | \
22 NETIF_F_GSO_MASK | NETIF_F_HW_CSUM)
23
24const struct nf_br_ops __rcu *nf_br_ops __read_mostly;
25EXPORT_SYMBOL_GPL(nf_br_ops);
26
27/* net device transmit always called with BH disabled */
28netdev_tx_t br_dev_xmit(struct sk_buff *skb, struct net_device *dev)
29{
30 struct net_bridge_mcast_port *pmctx_null = NULL;
31 struct net_bridge *br = netdev_priv(dev);
32 struct net_bridge_mcast *brmctx = &br->multicast_ctx;
33 struct net_bridge_fdb_entry *dst;
34 struct net_bridge_mdb_entry *mdst;
35 const struct nf_br_ops *nf_ops;
36 u8 state = BR_STATE_FORWARDING;
37 struct net_bridge_vlan *vlan;
38 const unsigned char *dest;
39 u16 vid = 0;
40
41 memset(skb->cb, 0, sizeof(struct br_input_skb_cb));
42 br_tc_skb_miss_set(skb, false);
43
44 rcu_read_lock();
45 nf_ops = rcu_dereference(nf_br_ops);
46 if (nf_ops && nf_ops->br_dev_xmit_hook(skb)) {
47 rcu_read_unlock();
48 return NETDEV_TX_OK;
49 }
50
51 dev_sw_netstats_tx_add(dev, 1, skb->len);
52
53 br_switchdev_frame_unmark(skb);
54 BR_INPUT_SKB_CB(skb)->brdev = dev;
55 BR_INPUT_SKB_CB(skb)->frag_max_size = 0;
56
57 skb_reset_mac_header(skb);
58 skb_pull(skb, ETH_HLEN);
59
60 if (!br_allowed_ingress(br, br_vlan_group_rcu(br), skb, &vid,
61 &state, &vlan))
62 goto out;
63
64 if (IS_ENABLED(CONFIG_INET) &&
65 (eth_hdr(skb)->h_proto == htons(ETH_P_ARP) ||
66 eth_hdr(skb)->h_proto == htons(ETH_P_RARP)) &&
67 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED)) {
68 br_do_proxy_suppress_arp(skb, br, vid, NULL);
69 } else if (IS_ENABLED(CONFIG_IPV6) &&
70 skb->protocol == htons(ETH_P_IPV6) &&
71 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED) &&
72 pskb_may_pull(skb, sizeof(struct ipv6hdr) +
73 sizeof(struct nd_msg)) &&
74 ipv6_hdr(skb)->nexthdr == IPPROTO_ICMPV6) {
75 struct nd_msg *msg, _msg;
76
77 msg = br_is_nd_neigh_msg(skb, &_msg);
78 if (msg)
79 br_do_suppress_nd(skb, br, vid, NULL, msg);
80 }
81
82 dest = eth_hdr(skb)->h_dest;
83 if (is_broadcast_ether_addr(dest)) {
84 br_flood(br, skb, BR_PKT_BROADCAST, false, true, vid);
85 } else if (is_multicast_ether_addr(dest)) {
86 if (unlikely(netpoll_tx_running(dev))) {
87 br_flood(br, skb, BR_PKT_MULTICAST, false, true, vid);
88 goto out;
89 }
90 if (br_multicast_rcv(&brmctx, &pmctx_null, vlan, skb, vid)) {
91 kfree_skb(skb);
92 goto out;
93 }
94
95 mdst = br_mdb_entry_skb_get(brmctx, skb, vid);
96 if ((mdst || BR_INPUT_SKB_CB_MROUTERS_ONLY(skb)) &&
97 br_multicast_querier_exists(brmctx, eth_hdr(skb), mdst))
98 br_multicast_flood(mdst, skb, brmctx, false, true);
99 else
100 br_flood(br, skb, BR_PKT_MULTICAST, false, true, vid);
101 } else if ((dst = br_fdb_find_rcu(br, dest, vid)) != NULL) {
102 br_forward(dst->dst, skb, false, true);
103 } else {
104 br_flood(br, skb, BR_PKT_UNICAST, false, true, vid);
105 }
106out:
107 rcu_read_unlock();
108 return NETDEV_TX_OK;
109}
110
111static struct lock_class_key bridge_netdev_addr_lock_key;
112
113static void br_set_lockdep_class(struct net_device *dev)
114{
115 lockdep_set_class(&dev->addr_list_lock, &bridge_netdev_addr_lock_key);
116}
117
118static int br_dev_init(struct net_device *dev)
119{
120 struct net_bridge *br = netdev_priv(dev);
121 int err;
122
123 dev->tstats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
124 if (!dev->tstats)
125 return -ENOMEM;
126
127 err = br_fdb_hash_init(br);
128 if (err) {
129 free_percpu(dev->tstats);
130 return err;
131 }
132
133 err = br_mdb_hash_init(br);
134 if (err) {
135 free_percpu(dev->tstats);
136 br_fdb_hash_fini(br);
137 return err;
138 }
139
140 err = br_vlan_init(br);
141 if (err) {
142 free_percpu(dev->tstats);
143 br_mdb_hash_fini(br);
144 br_fdb_hash_fini(br);
145 return err;
146 }
147
148 err = br_multicast_init_stats(br);
149 if (err) {
150 free_percpu(dev->tstats);
151 br_vlan_flush(br);
152 br_mdb_hash_fini(br);
153 br_fdb_hash_fini(br);
154 }
155
156 br_set_lockdep_class(dev);
157 return err;
158}
159
160static void br_dev_uninit(struct net_device *dev)
161{
162 struct net_bridge *br = netdev_priv(dev);
163
164 br_multicast_dev_del(br);
165 br_multicast_uninit_stats(br);
166 br_vlan_flush(br);
167 br_mdb_hash_fini(br);
168 br_fdb_hash_fini(br);
169 free_percpu(dev->tstats);
170}
171
172static int br_dev_open(struct net_device *dev)
173{
174 struct net_bridge *br = netdev_priv(dev);
175
176 netdev_update_features(dev);
177 netif_start_queue(dev);
178 br_stp_enable_bridge(br);
179 br_multicast_open(br);
180
181 if (br_opt_get(br, BROPT_MULTICAST_ENABLED))
182 br_multicast_join_snoopers(br);
183
184 return 0;
185}
186
187static void br_dev_set_multicast_list(struct net_device *dev)
188{
189}
190
191static void br_dev_change_rx_flags(struct net_device *dev, int change)
192{
193 if (change & IFF_PROMISC)
194 br_manage_promisc(netdev_priv(dev));
195}
196
197static int br_dev_stop(struct net_device *dev)
198{
199 struct net_bridge *br = netdev_priv(dev);
200
201 br_stp_disable_bridge(br);
202 br_multicast_stop(br);
203
204 if (br_opt_get(br, BROPT_MULTICAST_ENABLED))
205 br_multicast_leave_snoopers(br);
206
207 netif_stop_queue(dev);
208
209 return 0;
210}
211
212static int br_change_mtu(struct net_device *dev, int new_mtu)
213{
214 struct net_bridge *br = netdev_priv(dev);
215
216 dev->mtu = new_mtu;
217
218 /* this flag will be cleared if the MTU was automatically adjusted */
219 br_opt_toggle(br, BROPT_MTU_SET_BY_USER, true);
220#if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
221 /* remember the MTU in the rtable for PMTU */
222 dst_metric_set(&br->fake_rtable.dst, RTAX_MTU, new_mtu);
223#endif
224
225 return 0;
226}
227
228/* Allow setting mac address to any valid ethernet address. */
229static int br_set_mac_address(struct net_device *dev, void *p)
230{
231 struct net_bridge *br = netdev_priv(dev);
232 struct sockaddr *addr = p;
233
234 if (!is_valid_ether_addr(addr->sa_data))
235 return -EADDRNOTAVAIL;
236
237 /* dev_set_mac_addr() can be called by a master device on bridge's
238 * NETDEV_UNREGISTER, but since it's being destroyed do nothing
239 */
240 if (dev->reg_state != NETREG_REGISTERED)
241 return -EBUSY;
242
243 spin_lock_bh(&br->lock);
244 if (!ether_addr_equal(dev->dev_addr, addr->sa_data)) {
245 /* Mac address will be changed in br_stp_change_bridge_id(). */
246 br_stp_change_bridge_id(br, addr->sa_data);
247 }
248 spin_unlock_bh(&br->lock);
249
250 return 0;
251}
252
253static void br_getinfo(struct net_device *dev, struct ethtool_drvinfo *info)
254{
255 strscpy(info->driver, "bridge", sizeof(info->driver));
256 strscpy(info->version, BR_VERSION, sizeof(info->version));
257 strscpy(info->fw_version, "N/A", sizeof(info->fw_version));
258 strscpy(info->bus_info, "N/A", sizeof(info->bus_info));
259}
260
261static int br_get_link_ksettings(struct net_device *dev,
262 struct ethtool_link_ksettings *cmd)
263{
264 struct net_bridge *br = netdev_priv(dev);
265 struct net_bridge_port *p;
266
267 cmd->base.duplex = DUPLEX_UNKNOWN;
268 cmd->base.port = PORT_OTHER;
269 cmd->base.speed = SPEED_UNKNOWN;
270
271 list_for_each_entry(p, &br->port_list, list) {
272 struct ethtool_link_ksettings ecmd;
273 struct net_device *pdev = p->dev;
274
275 if (!netif_running(pdev) || !netif_oper_up(pdev))
276 continue;
277
278 if (__ethtool_get_link_ksettings(pdev, &ecmd))
279 continue;
280
281 if (ecmd.base.speed == (__u32)SPEED_UNKNOWN)
282 continue;
283
284 if (cmd->base.speed == (__u32)SPEED_UNKNOWN ||
285 cmd->base.speed < ecmd.base.speed)
286 cmd->base.speed = ecmd.base.speed;
287 }
288
289 return 0;
290}
291
292static netdev_features_t br_fix_features(struct net_device *dev,
293 netdev_features_t features)
294{
295 struct net_bridge *br = netdev_priv(dev);
296
297 return br_features_recompute(br, features);
298}
299
300#ifdef CONFIG_NET_POLL_CONTROLLER
301static void br_poll_controller(struct net_device *br_dev)
302{
303}
304
305static void br_netpoll_cleanup(struct net_device *dev)
306{
307 struct net_bridge *br = netdev_priv(dev);
308 struct net_bridge_port *p;
309
310 list_for_each_entry(p, &br->port_list, list)
311 br_netpoll_disable(p);
312}
313
314static int __br_netpoll_enable(struct net_bridge_port *p)
315{
316 struct netpoll *np;
317 int err;
318
319 np = kzalloc(sizeof(*p->np), GFP_KERNEL);
320 if (!np)
321 return -ENOMEM;
322
323 err = __netpoll_setup(np, p->dev);
324 if (err) {
325 kfree(np);
326 return err;
327 }
328
329 p->np = np;
330 return err;
331}
332
333int br_netpoll_enable(struct net_bridge_port *p)
334{
335 if (!p->br->dev->npinfo)
336 return 0;
337
338 return __br_netpoll_enable(p);
339}
340
341static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni)
342{
343 struct net_bridge *br = netdev_priv(dev);
344 struct net_bridge_port *p;
345 int err = 0;
346
347 list_for_each_entry(p, &br->port_list, list) {
348 if (!p->dev)
349 continue;
350 err = __br_netpoll_enable(p);
351 if (err)
352 goto fail;
353 }
354
355out:
356 return err;
357
358fail:
359 br_netpoll_cleanup(dev);
360 goto out;
361}
362
363void br_netpoll_disable(struct net_bridge_port *p)
364{
365 struct netpoll *np = p->np;
366
367 if (!np)
368 return;
369
370 p->np = NULL;
371
372 __netpoll_free(np);
373}
374
375#endif
376
377static int br_add_slave(struct net_device *dev, struct net_device *slave_dev,
378 struct netlink_ext_ack *extack)
379
380{
381 struct net_bridge *br = netdev_priv(dev);
382
383 return br_add_if(br, slave_dev, extack);
384}
385
386static int br_del_slave(struct net_device *dev, struct net_device *slave_dev)
387{
388 struct net_bridge *br = netdev_priv(dev);
389
390 return br_del_if(br, slave_dev);
391}
392
393static int br_fill_forward_path(struct net_device_path_ctx *ctx,
394 struct net_device_path *path)
395{
396 struct net_bridge_fdb_entry *f;
397 struct net_bridge_port *dst;
398 struct net_bridge *br;
399
400 if (netif_is_bridge_port(ctx->dev))
401 return -1;
402
403 br = netdev_priv(ctx->dev);
404
405 br_vlan_fill_forward_path_pvid(br, ctx, path);
406
407 f = br_fdb_find_rcu(br, ctx->daddr, path->bridge.vlan_id);
408 if (!f || !f->dst)
409 return -1;
410
411 dst = READ_ONCE(f->dst);
412 if (!dst)
413 return -1;
414
415 if (br_vlan_fill_forward_path_mode(br, dst, path))
416 return -1;
417
418 path->type = DEV_PATH_BRIDGE;
419 path->dev = dst->br->dev;
420 ctx->dev = dst->dev;
421
422 switch (path->bridge.vlan_mode) {
423 case DEV_PATH_BR_VLAN_TAG:
424 if (ctx->num_vlans >= ARRAY_SIZE(ctx->vlan))
425 return -ENOSPC;
426 ctx->vlan[ctx->num_vlans].id = path->bridge.vlan_id;
427 ctx->vlan[ctx->num_vlans].proto = path->bridge.vlan_proto;
428 ctx->num_vlans++;
429 break;
430 case DEV_PATH_BR_VLAN_UNTAG_HW:
431 case DEV_PATH_BR_VLAN_UNTAG:
432 ctx->num_vlans--;
433 break;
434 case DEV_PATH_BR_VLAN_KEEP:
435 break;
436 }
437
438 return 0;
439}
440
441static const struct ethtool_ops br_ethtool_ops = {
442 .get_drvinfo = br_getinfo,
443 .get_link = ethtool_op_get_link,
444 .get_link_ksettings = br_get_link_ksettings,
445};
446
447static const struct net_device_ops br_netdev_ops = {
448 .ndo_open = br_dev_open,
449 .ndo_stop = br_dev_stop,
450 .ndo_init = br_dev_init,
451 .ndo_uninit = br_dev_uninit,
452 .ndo_start_xmit = br_dev_xmit,
453 .ndo_get_stats64 = dev_get_tstats64,
454 .ndo_set_mac_address = br_set_mac_address,
455 .ndo_set_rx_mode = br_dev_set_multicast_list,
456 .ndo_change_rx_flags = br_dev_change_rx_flags,
457 .ndo_change_mtu = br_change_mtu,
458 .ndo_siocdevprivate = br_dev_siocdevprivate,
459#ifdef CONFIG_NET_POLL_CONTROLLER
460 .ndo_netpoll_setup = br_netpoll_setup,
461 .ndo_netpoll_cleanup = br_netpoll_cleanup,
462 .ndo_poll_controller = br_poll_controller,
463#endif
464 .ndo_add_slave = br_add_slave,
465 .ndo_del_slave = br_del_slave,
466 .ndo_fix_features = br_fix_features,
467 .ndo_fdb_add = br_fdb_add,
468 .ndo_fdb_del = br_fdb_delete,
469 .ndo_fdb_del_bulk = br_fdb_delete_bulk,
470 .ndo_fdb_dump = br_fdb_dump,
471 .ndo_fdb_get = br_fdb_get,
472 .ndo_mdb_add = br_mdb_add,
473 .ndo_mdb_del = br_mdb_del,
474 .ndo_mdb_del_bulk = br_mdb_del_bulk,
475 .ndo_mdb_dump = br_mdb_dump,
476 .ndo_mdb_get = br_mdb_get,
477 .ndo_bridge_getlink = br_getlink,
478 .ndo_bridge_setlink = br_setlink,
479 .ndo_bridge_dellink = br_dellink,
480 .ndo_features_check = passthru_features_check,
481 .ndo_fill_forward_path = br_fill_forward_path,
482};
483
484static struct device_type br_type = {
485 .name = "bridge",
486};
487
488void br_dev_setup(struct net_device *dev)
489{
490 struct net_bridge *br = netdev_priv(dev);
491
492 eth_hw_addr_random(dev);
493 ether_setup(dev);
494
495 dev->netdev_ops = &br_netdev_ops;
496 dev->needs_free_netdev = true;
497 dev->ethtool_ops = &br_ethtool_ops;
498 SET_NETDEV_DEVTYPE(dev, &br_type);
499 dev->priv_flags = IFF_EBRIDGE | IFF_NO_QUEUE;
500
501 dev->features = COMMON_FEATURES | NETIF_F_LLTX | NETIF_F_NETNS_LOCAL |
502 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_STAG_TX;
503 dev->hw_features = COMMON_FEATURES | NETIF_F_HW_VLAN_CTAG_TX |
504 NETIF_F_HW_VLAN_STAG_TX;
505 dev->vlan_features = COMMON_FEATURES;
506
507 br->dev = dev;
508 spin_lock_init(&br->lock);
509 INIT_LIST_HEAD(&br->port_list);
510 INIT_HLIST_HEAD(&br->fdb_list);
511 INIT_HLIST_HEAD(&br->frame_type_list);
512#if IS_ENABLED(CONFIG_BRIDGE_MRP)
513 INIT_HLIST_HEAD(&br->mrp_list);
514#endif
515#if IS_ENABLED(CONFIG_BRIDGE_CFM)
516 INIT_HLIST_HEAD(&br->mep_list);
517#endif
518 spin_lock_init(&br->hash_lock);
519
520 br->bridge_id.prio[0] = 0x80;
521 br->bridge_id.prio[1] = 0x00;
522
523 ether_addr_copy(br->group_addr, eth_stp_addr);
524
525 br->stp_enabled = BR_NO_STP;
526 br->group_fwd_mask = BR_GROUPFWD_DEFAULT;
527 br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
528
529 br->designated_root = br->bridge_id;
530 br->bridge_max_age = br->max_age = 20 * HZ;
531 br->bridge_hello_time = br->hello_time = 2 * HZ;
532 br->bridge_forward_delay = br->forward_delay = 15 * HZ;
533 br->bridge_ageing_time = br->ageing_time = BR_DEFAULT_AGEING_TIME;
534 dev->max_mtu = ETH_MAX_MTU;
535
536 br_netfilter_rtable_init(br);
537 br_stp_timer_init(br);
538 br_multicast_init(br);
539 INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup);
540}
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Device handling code
4 * Linux ethernet bridge
5 *
6 * Authors:
7 * Lennert Buytenhek <buytenh@gnu.org>
8 */
9
10#include <linux/kernel.h>
11#include <linux/netdevice.h>
12#include <linux/netpoll.h>
13#include <linux/etherdevice.h>
14#include <linux/ethtool.h>
15#include <linux/list.h>
16#include <linux/netfilter_bridge.h>
17
18#include <linux/uaccess.h>
19#include "br_private.h"
20
21#define COMMON_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HIGHDMA | \
22 NETIF_F_GSO_MASK | NETIF_F_HW_CSUM)
23
24const struct nf_br_ops __rcu *nf_br_ops __read_mostly;
25EXPORT_SYMBOL_GPL(nf_br_ops);
26
27/* net device transmit always called with BH disabled */
28netdev_tx_t br_dev_xmit(struct sk_buff *skb, struct net_device *dev)
29{
30 struct net_bridge *br = netdev_priv(dev);
31 struct net_bridge_fdb_entry *dst;
32 struct net_bridge_mdb_entry *mdst;
33 struct pcpu_sw_netstats *brstats = this_cpu_ptr(br->stats);
34 const struct nf_br_ops *nf_ops;
35 u8 state = BR_STATE_FORWARDING;
36 const unsigned char *dest;
37 u16 vid = 0;
38
39 memset(skb->cb, 0, sizeof(struct br_input_skb_cb));
40
41 rcu_read_lock();
42 nf_ops = rcu_dereference(nf_br_ops);
43 if (nf_ops && nf_ops->br_dev_xmit_hook(skb)) {
44 rcu_read_unlock();
45 return NETDEV_TX_OK;
46 }
47
48 u64_stats_update_begin(&brstats->syncp);
49 brstats->tx_packets++;
50 brstats->tx_bytes += skb->len;
51 u64_stats_update_end(&brstats->syncp);
52
53 br_switchdev_frame_unmark(skb);
54 BR_INPUT_SKB_CB(skb)->brdev = dev;
55 BR_INPUT_SKB_CB(skb)->frag_max_size = 0;
56
57 skb_reset_mac_header(skb);
58 skb_pull(skb, ETH_HLEN);
59
60 if (!br_allowed_ingress(br, br_vlan_group_rcu(br), skb, &vid, &state))
61 goto out;
62
63 if (IS_ENABLED(CONFIG_INET) &&
64 (eth_hdr(skb)->h_proto == htons(ETH_P_ARP) ||
65 eth_hdr(skb)->h_proto == htons(ETH_P_RARP)) &&
66 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED)) {
67 br_do_proxy_suppress_arp(skb, br, vid, NULL);
68 } else if (IS_ENABLED(CONFIG_IPV6) &&
69 skb->protocol == htons(ETH_P_IPV6) &&
70 br_opt_get(br, BROPT_NEIGH_SUPPRESS_ENABLED) &&
71 pskb_may_pull(skb, sizeof(struct ipv6hdr) +
72 sizeof(struct nd_msg)) &&
73 ipv6_hdr(skb)->nexthdr == IPPROTO_ICMPV6) {
74 struct nd_msg *msg, _msg;
75
76 msg = br_is_nd_neigh_msg(skb, &_msg);
77 if (msg)
78 br_do_suppress_nd(skb, br, vid, NULL, msg);
79 }
80
81 dest = eth_hdr(skb)->h_dest;
82 if (is_broadcast_ether_addr(dest)) {
83 br_flood(br, skb, BR_PKT_BROADCAST, false, true);
84 } else if (is_multicast_ether_addr(dest)) {
85 if (unlikely(netpoll_tx_running(dev))) {
86 br_flood(br, skb, BR_PKT_MULTICAST, false, true);
87 goto out;
88 }
89 if (br_multicast_rcv(br, NULL, skb, vid)) {
90 kfree_skb(skb);
91 goto out;
92 }
93
94 mdst = br_mdb_get(br, skb, vid);
95 if ((mdst || BR_INPUT_SKB_CB_MROUTERS_ONLY(skb)) &&
96 br_multicast_querier_exists(br, eth_hdr(skb)))
97 br_multicast_flood(mdst, skb, false, true);
98 else
99 br_flood(br, skb, BR_PKT_MULTICAST, false, true);
100 } else if ((dst = br_fdb_find_rcu(br, dest, vid)) != NULL) {
101 br_forward(dst->dst, skb, false, true);
102 } else {
103 br_flood(br, skb, BR_PKT_UNICAST, false, true);
104 }
105out:
106 rcu_read_unlock();
107 return NETDEV_TX_OK;
108}
109
110static struct lock_class_key bridge_netdev_addr_lock_key;
111
112static void br_set_lockdep_class(struct net_device *dev)
113{
114 lockdep_set_class(&dev->addr_list_lock, &bridge_netdev_addr_lock_key);
115}
116
117static int br_dev_init(struct net_device *dev)
118{
119 struct net_bridge *br = netdev_priv(dev);
120 int err;
121
122 br->stats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
123 if (!br->stats)
124 return -ENOMEM;
125
126 err = br_fdb_hash_init(br);
127 if (err) {
128 free_percpu(br->stats);
129 return err;
130 }
131
132 err = br_mdb_hash_init(br);
133 if (err) {
134 free_percpu(br->stats);
135 br_fdb_hash_fini(br);
136 return err;
137 }
138
139 err = br_vlan_init(br);
140 if (err) {
141 free_percpu(br->stats);
142 br_mdb_hash_fini(br);
143 br_fdb_hash_fini(br);
144 return err;
145 }
146
147 err = br_multicast_init_stats(br);
148 if (err) {
149 free_percpu(br->stats);
150 br_vlan_flush(br);
151 br_mdb_hash_fini(br);
152 br_fdb_hash_fini(br);
153 }
154
155 br_set_lockdep_class(dev);
156 return err;
157}
158
159static void br_dev_uninit(struct net_device *dev)
160{
161 struct net_bridge *br = netdev_priv(dev);
162
163 br_multicast_dev_del(br);
164 br_multicast_uninit_stats(br);
165 br_vlan_flush(br);
166 br_mdb_hash_fini(br);
167 br_fdb_hash_fini(br);
168 free_percpu(br->stats);
169}
170
171static int br_dev_open(struct net_device *dev)
172{
173 struct net_bridge *br = netdev_priv(dev);
174
175 netdev_update_features(dev);
176 netif_start_queue(dev);
177 br_stp_enable_bridge(br);
178 br_multicast_open(br);
179
180 return 0;
181}
182
183static void br_dev_set_multicast_list(struct net_device *dev)
184{
185}
186
187static void br_dev_change_rx_flags(struct net_device *dev, int change)
188{
189 if (change & IFF_PROMISC)
190 br_manage_promisc(netdev_priv(dev));
191}
192
193static int br_dev_stop(struct net_device *dev)
194{
195 struct net_bridge *br = netdev_priv(dev);
196
197 br_stp_disable_bridge(br);
198 br_multicast_stop(br);
199
200 netif_stop_queue(dev);
201
202 return 0;
203}
204
205static void br_get_stats64(struct net_device *dev,
206 struct rtnl_link_stats64 *stats)
207{
208 struct net_bridge *br = netdev_priv(dev);
209 struct pcpu_sw_netstats tmp, sum = { 0 };
210 unsigned int cpu;
211
212 for_each_possible_cpu(cpu) {
213 unsigned int start;
214 const struct pcpu_sw_netstats *bstats
215 = per_cpu_ptr(br->stats, cpu);
216 do {
217 start = u64_stats_fetch_begin_irq(&bstats->syncp);
218 memcpy(&tmp, bstats, sizeof(tmp));
219 } while (u64_stats_fetch_retry_irq(&bstats->syncp, start));
220 sum.tx_bytes += tmp.tx_bytes;
221 sum.tx_packets += tmp.tx_packets;
222 sum.rx_bytes += tmp.rx_bytes;
223 sum.rx_packets += tmp.rx_packets;
224 }
225
226 stats->tx_bytes = sum.tx_bytes;
227 stats->tx_packets = sum.tx_packets;
228 stats->rx_bytes = sum.rx_bytes;
229 stats->rx_packets = sum.rx_packets;
230}
231
232static int br_change_mtu(struct net_device *dev, int new_mtu)
233{
234 struct net_bridge *br = netdev_priv(dev);
235
236 dev->mtu = new_mtu;
237
238 /* this flag will be cleared if the MTU was automatically adjusted */
239 br_opt_toggle(br, BROPT_MTU_SET_BY_USER, true);
240#if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
241 /* remember the MTU in the rtable for PMTU */
242 dst_metric_set(&br->fake_rtable.dst, RTAX_MTU, new_mtu);
243#endif
244
245 return 0;
246}
247
248/* Allow setting mac address to any valid ethernet address. */
249static int br_set_mac_address(struct net_device *dev, void *p)
250{
251 struct net_bridge *br = netdev_priv(dev);
252 struct sockaddr *addr = p;
253
254 if (!is_valid_ether_addr(addr->sa_data))
255 return -EADDRNOTAVAIL;
256
257 /* dev_set_mac_addr() can be called by a master device on bridge's
258 * NETDEV_UNREGISTER, but since it's being destroyed do nothing
259 */
260 if (dev->reg_state != NETREG_REGISTERED)
261 return -EBUSY;
262
263 spin_lock_bh(&br->lock);
264 if (!ether_addr_equal(dev->dev_addr, addr->sa_data)) {
265 /* Mac address will be changed in br_stp_change_bridge_id(). */
266 br_stp_change_bridge_id(br, addr->sa_data);
267 }
268 spin_unlock_bh(&br->lock);
269
270 return 0;
271}
272
273static void br_getinfo(struct net_device *dev, struct ethtool_drvinfo *info)
274{
275 strlcpy(info->driver, "bridge", sizeof(info->driver));
276 strlcpy(info->version, BR_VERSION, sizeof(info->version));
277 strlcpy(info->fw_version, "N/A", sizeof(info->fw_version));
278 strlcpy(info->bus_info, "N/A", sizeof(info->bus_info));
279}
280
281static int br_get_link_ksettings(struct net_device *dev,
282 struct ethtool_link_ksettings *cmd)
283{
284 struct net_bridge *br = netdev_priv(dev);
285 struct net_bridge_port *p;
286
287 cmd->base.duplex = DUPLEX_UNKNOWN;
288 cmd->base.port = PORT_OTHER;
289 cmd->base.speed = SPEED_UNKNOWN;
290
291 list_for_each_entry(p, &br->port_list, list) {
292 struct ethtool_link_ksettings ecmd;
293 struct net_device *pdev = p->dev;
294
295 if (!netif_running(pdev) || !netif_oper_up(pdev))
296 continue;
297
298 if (__ethtool_get_link_ksettings(pdev, &ecmd))
299 continue;
300
301 if (ecmd.base.speed == (__u32)SPEED_UNKNOWN)
302 continue;
303
304 if (cmd->base.speed == (__u32)SPEED_UNKNOWN ||
305 cmd->base.speed < ecmd.base.speed)
306 cmd->base.speed = ecmd.base.speed;
307 }
308
309 return 0;
310}
311
312static netdev_features_t br_fix_features(struct net_device *dev,
313 netdev_features_t features)
314{
315 struct net_bridge *br = netdev_priv(dev);
316
317 return br_features_recompute(br, features);
318}
319
320#ifdef CONFIG_NET_POLL_CONTROLLER
321static void br_poll_controller(struct net_device *br_dev)
322{
323}
324
325static void br_netpoll_cleanup(struct net_device *dev)
326{
327 struct net_bridge *br = netdev_priv(dev);
328 struct net_bridge_port *p;
329
330 list_for_each_entry(p, &br->port_list, list)
331 br_netpoll_disable(p);
332}
333
334static int __br_netpoll_enable(struct net_bridge_port *p)
335{
336 struct netpoll *np;
337 int err;
338
339 np = kzalloc(sizeof(*p->np), GFP_KERNEL);
340 if (!np)
341 return -ENOMEM;
342
343 err = __netpoll_setup(np, p->dev);
344 if (err) {
345 kfree(np);
346 return err;
347 }
348
349 p->np = np;
350 return err;
351}
352
353int br_netpoll_enable(struct net_bridge_port *p)
354{
355 if (!p->br->dev->npinfo)
356 return 0;
357
358 return __br_netpoll_enable(p);
359}
360
361static int br_netpoll_setup(struct net_device *dev, struct netpoll_info *ni)
362{
363 struct net_bridge *br = netdev_priv(dev);
364 struct net_bridge_port *p;
365 int err = 0;
366
367 list_for_each_entry(p, &br->port_list, list) {
368 if (!p->dev)
369 continue;
370 err = __br_netpoll_enable(p);
371 if (err)
372 goto fail;
373 }
374
375out:
376 return err;
377
378fail:
379 br_netpoll_cleanup(dev);
380 goto out;
381}
382
383void br_netpoll_disable(struct net_bridge_port *p)
384{
385 struct netpoll *np = p->np;
386
387 if (!np)
388 return;
389
390 p->np = NULL;
391
392 __netpoll_free(np);
393}
394
395#endif
396
397static int br_add_slave(struct net_device *dev, struct net_device *slave_dev,
398 struct netlink_ext_ack *extack)
399
400{
401 struct net_bridge *br = netdev_priv(dev);
402
403 return br_add_if(br, slave_dev, extack);
404}
405
406static int br_del_slave(struct net_device *dev, struct net_device *slave_dev)
407{
408 struct net_bridge *br = netdev_priv(dev);
409
410 return br_del_if(br, slave_dev);
411}
412
413static const struct ethtool_ops br_ethtool_ops = {
414 .get_drvinfo = br_getinfo,
415 .get_link = ethtool_op_get_link,
416 .get_link_ksettings = br_get_link_ksettings,
417};
418
419static const struct net_device_ops br_netdev_ops = {
420 .ndo_open = br_dev_open,
421 .ndo_stop = br_dev_stop,
422 .ndo_init = br_dev_init,
423 .ndo_uninit = br_dev_uninit,
424 .ndo_start_xmit = br_dev_xmit,
425 .ndo_get_stats64 = br_get_stats64,
426 .ndo_set_mac_address = br_set_mac_address,
427 .ndo_set_rx_mode = br_dev_set_multicast_list,
428 .ndo_change_rx_flags = br_dev_change_rx_flags,
429 .ndo_change_mtu = br_change_mtu,
430 .ndo_do_ioctl = br_dev_ioctl,
431#ifdef CONFIG_NET_POLL_CONTROLLER
432 .ndo_netpoll_setup = br_netpoll_setup,
433 .ndo_netpoll_cleanup = br_netpoll_cleanup,
434 .ndo_poll_controller = br_poll_controller,
435#endif
436 .ndo_add_slave = br_add_slave,
437 .ndo_del_slave = br_del_slave,
438 .ndo_fix_features = br_fix_features,
439 .ndo_fdb_add = br_fdb_add,
440 .ndo_fdb_del = br_fdb_delete,
441 .ndo_fdb_dump = br_fdb_dump,
442 .ndo_fdb_get = br_fdb_get,
443 .ndo_bridge_getlink = br_getlink,
444 .ndo_bridge_setlink = br_setlink,
445 .ndo_bridge_dellink = br_dellink,
446 .ndo_features_check = passthru_features_check,
447};
448
449static struct device_type br_type = {
450 .name = "bridge",
451};
452
453void br_dev_setup(struct net_device *dev)
454{
455 struct net_bridge *br = netdev_priv(dev);
456
457 eth_hw_addr_random(dev);
458 ether_setup(dev);
459
460 dev->netdev_ops = &br_netdev_ops;
461 dev->needs_free_netdev = true;
462 dev->ethtool_ops = &br_ethtool_ops;
463 SET_NETDEV_DEVTYPE(dev, &br_type);
464 dev->priv_flags = IFF_EBRIDGE | IFF_NO_QUEUE;
465
466 dev->features = COMMON_FEATURES | NETIF_F_LLTX | NETIF_F_NETNS_LOCAL |
467 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_STAG_TX;
468 dev->hw_features = COMMON_FEATURES | NETIF_F_HW_VLAN_CTAG_TX |
469 NETIF_F_HW_VLAN_STAG_TX;
470 dev->vlan_features = COMMON_FEATURES;
471
472 br->dev = dev;
473 spin_lock_init(&br->lock);
474 INIT_LIST_HEAD(&br->port_list);
475 INIT_HLIST_HEAD(&br->fdb_list);
476#if IS_ENABLED(CONFIG_BRIDGE_MRP)
477 INIT_LIST_HEAD(&br->mrp_list);
478#endif
479 spin_lock_init(&br->hash_lock);
480
481 br->bridge_id.prio[0] = 0x80;
482 br->bridge_id.prio[1] = 0x00;
483
484 ether_addr_copy(br->group_addr, eth_stp_addr);
485
486 br->stp_enabled = BR_NO_STP;
487 br->group_fwd_mask = BR_GROUPFWD_DEFAULT;
488 br->group_fwd_mask_required = BR_GROUPFWD_DEFAULT;
489
490 br->designated_root = br->bridge_id;
491 br->bridge_max_age = br->max_age = 20 * HZ;
492 br->bridge_hello_time = br->hello_time = 2 * HZ;
493 br->bridge_forward_delay = br->forward_delay = 15 * HZ;
494 br->bridge_ageing_time = br->ageing_time = BR_DEFAULT_AGEING_TIME;
495 dev->max_mtu = ETH_MAX_MTU;
496
497 br_netfilter_rtable_init(br);
498 br_stp_timer_init(br);
499 br_multicast_init(br);
500 INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup);
501}