Linux Audio

Check our new training course

Loading...
v4.17
 
   1/*
   2 * net/dsa/slave.c - Slave device handling
   3 * Copyright (c) 2008-2009 Marvell Semiconductor
   4 *
   5 * This program is free software; you can redistribute it and/or modify
   6 * it under the terms of the GNU General Public License as published by
   7 * the Free Software Foundation; either version 2 of the License, or
   8 * (at your option) any later version.
   9 */
  10
  11#include <linux/list.h>
  12#include <linux/etherdevice.h>
  13#include <linux/netdevice.h>
  14#include <linux/phy.h>
  15#include <linux/phy_fixed.h>
 
  16#include <linux/of_net.h>
  17#include <linux/of_mdio.h>
  18#include <linux/mdio.h>
  19#include <net/rtnetlink.h>
  20#include <net/pkt_cls.h>
  21#include <net/tc_act/tc_mirred.h>
  22#include <linux/if_bridge.h>
  23#include <linux/netpoll.h>
  24#include <linux/ptp_classify.h>
  25
  26#include "dsa_priv.h"
  27
  28static bool dsa_slave_dev_check(struct net_device *dev);
  29
  30/* slave mii_bus handling ***************************************************/
  31static int dsa_slave_phy_read(struct mii_bus *bus, int addr, int reg)
  32{
  33	struct dsa_switch *ds = bus->priv;
  34
  35	if (ds->phys_mii_mask & (1 << addr))
  36		return ds->ops->phy_read(ds, addr, reg);
  37
  38	return 0xffff;
  39}
  40
  41static int dsa_slave_phy_write(struct mii_bus *bus, int addr, int reg, u16 val)
  42{
  43	struct dsa_switch *ds = bus->priv;
  44
  45	if (ds->phys_mii_mask & (1 << addr))
  46		return ds->ops->phy_write(ds, addr, reg, val);
  47
  48	return 0;
  49}
  50
  51void dsa_slave_mii_bus_init(struct dsa_switch *ds)
  52{
  53	ds->slave_mii_bus->priv = (void *)ds;
  54	ds->slave_mii_bus->name = "dsa slave smi";
  55	ds->slave_mii_bus->read = dsa_slave_phy_read;
  56	ds->slave_mii_bus->write = dsa_slave_phy_write;
  57	snprintf(ds->slave_mii_bus->id, MII_BUS_ID_SIZE, "dsa-%d.%d",
  58		 ds->dst->index, ds->index);
  59	ds->slave_mii_bus->parent = ds->dev;
  60	ds->slave_mii_bus->phy_mask = ~ds->phys_mii_mask;
  61}
  62
  63
  64/* slave device handling ****************************************************/
  65static int dsa_slave_get_iflink(const struct net_device *dev)
  66{
  67	return dsa_slave_to_master(dev)->ifindex;
  68}
  69
  70static int dsa_slave_open(struct net_device *dev)
  71{
  72	struct net_device *master = dsa_slave_to_master(dev);
  73	struct dsa_port *dp = dsa_slave_to_port(dev);
  74	int err;
  75
  76	if (!(master->flags & IFF_UP))
  77		return -ENETDOWN;
  78
  79	if (!ether_addr_equal(dev->dev_addr, master->dev_addr)) {
  80		err = dev_uc_add(master, dev->dev_addr);
  81		if (err < 0)
  82			goto out;
  83	}
  84
  85	if (dev->flags & IFF_ALLMULTI) {
  86		err = dev_set_allmulti(master, 1);
  87		if (err < 0)
  88			goto del_unicast;
  89	}
  90	if (dev->flags & IFF_PROMISC) {
  91		err = dev_set_promiscuity(master, 1);
  92		if (err < 0)
  93			goto clear_allmulti;
  94	}
  95
  96	err = dsa_port_enable(dp, dev->phydev);
  97	if (err)
  98		goto clear_promisc;
  99
 100	if (dev->phydev)
 101		phy_start(dev->phydev);
 102
 103	return 0;
 104
 105clear_promisc:
 106	if (dev->flags & IFF_PROMISC)
 107		dev_set_promiscuity(master, -1);
 108clear_allmulti:
 109	if (dev->flags & IFF_ALLMULTI)
 110		dev_set_allmulti(master, -1);
 111del_unicast:
 112	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
 113		dev_uc_del(master, dev->dev_addr);
 114out:
 115	return err;
 116}
 117
 118static int dsa_slave_close(struct net_device *dev)
 119{
 120	struct net_device *master = dsa_slave_to_master(dev);
 121	struct dsa_port *dp = dsa_slave_to_port(dev);
 122
 123	if (dev->phydev)
 124		phy_stop(dev->phydev);
 125
 126	dsa_port_disable(dp, dev->phydev);
 
 
 127
 128	dev_mc_unsync(master, dev);
 129	dev_uc_unsync(master, dev);
 130	if (dev->flags & IFF_ALLMULTI)
 131		dev_set_allmulti(master, -1);
 132	if (dev->flags & IFF_PROMISC)
 133		dev_set_promiscuity(master, -1);
 134
 135	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
 136		dev_uc_del(master, dev->dev_addr);
 137
 138	return 0;
 139}
 140
 141static void dsa_slave_change_rx_flags(struct net_device *dev, int change)
 142{
 143	struct net_device *master = dsa_slave_to_master(dev);
 144
 145	if (change & IFF_ALLMULTI)
 146		dev_set_allmulti(master, dev->flags & IFF_ALLMULTI ? 1 : -1);
 147	if (change & IFF_PROMISC)
 148		dev_set_promiscuity(master, dev->flags & IFF_PROMISC ? 1 : -1);
 
 
 
 149}
 150
 151static void dsa_slave_set_rx_mode(struct net_device *dev)
 152{
 153	struct net_device *master = dsa_slave_to_master(dev);
 154
 155	dev_mc_sync(master, dev);
 156	dev_uc_sync(master, dev);
 157}
 158
 159static int dsa_slave_set_mac_address(struct net_device *dev, void *a)
 160{
 161	struct net_device *master = dsa_slave_to_master(dev);
 162	struct sockaddr *addr = a;
 163	int err;
 164
 165	if (!is_valid_ether_addr(addr->sa_data))
 166		return -EADDRNOTAVAIL;
 167
 168	if (!(dev->flags & IFF_UP))
 169		goto out;
 170
 171	if (!ether_addr_equal(addr->sa_data, master->dev_addr)) {
 172		err = dev_uc_add(master, addr->sa_data);
 173		if (err < 0)
 174			return err;
 175	}
 176
 177	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
 178		dev_uc_del(master, dev->dev_addr);
 179
 180out:
 181	ether_addr_copy(dev->dev_addr, addr->sa_data);
 182
 183	return 0;
 184}
 185
 186struct dsa_slave_dump_ctx {
 187	struct net_device *dev;
 188	struct sk_buff *skb;
 189	struct netlink_callback *cb;
 190	int idx;
 191};
 192
 193static int
 194dsa_slave_port_fdb_do_dump(const unsigned char *addr, u16 vid,
 195			   bool is_static, void *data)
 196{
 197	struct dsa_slave_dump_ctx *dump = data;
 198	u32 portid = NETLINK_CB(dump->cb->skb).portid;
 199	u32 seq = dump->cb->nlh->nlmsg_seq;
 200	struct nlmsghdr *nlh;
 201	struct ndmsg *ndm;
 202
 203	if (dump->idx < dump->cb->args[2])
 204		goto skip;
 205
 206	nlh = nlmsg_put(dump->skb, portid, seq, RTM_NEWNEIGH,
 207			sizeof(*ndm), NLM_F_MULTI);
 208	if (!nlh)
 209		return -EMSGSIZE;
 210
 211	ndm = nlmsg_data(nlh);
 212	ndm->ndm_family  = AF_BRIDGE;
 213	ndm->ndm_pad1    = 0;
 214	ndm->ndm_pad2    = 0;
 215	ndm->ndm_flags   = NTF_SELF;
 216	ndm->ndm_type    = 0;
 217	ndm->ndm_ifindex = dump->dev->ifindex;
 218	ndm->ndm_state   = is_static ? NUD_NOARP : NUD_REACHABLE;
 219
 220	if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, addr))
 221		goto nla_put_failure;
 222
 223	if (vid && nla_put_u16(dump->skb, NDA_VLAN, vid))
 224		goto nla_put_failure;
 225
 226	nlmsg_end(dump->skb, nlh);
 227
 228skip:
 229	dump->idx++;
 230	return 0;
 231
 232nla_put_failure:
 233	nlmsg_cancel(dump->skb, nlh);
 234	return -EMSGSIZE;
 235}
 236
 237static int
 238dsa_slave_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
 239		   struct net_device *dev, struct net_device *filter_dev,
 240		   int *idx)
 241{
 242	struct dsa_port *dp = dsa_slave_to_port(dev);
 243	struct dsa_slave_dump_ctx dump = {
 244		.dev = dev,
 245		.skb = skb,
 246		.cb = cb,
 247		.idx = *idx,
 248	};
 249	int err;
 250
 251	err = dsa_port_fdb_dump(dp, dsa_slave_port_fdb_do_dump, &dump);
 252	*idx = dump.idx;
 253
 254	return err;
 255}
 256
 257static int dsa_slave_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
 258{
 259	struct dsa_slave_priv *p = netdev_priv(dev);
 260	struct dsa_switch *ds = p->dp->ds;
 261	int port = p->dp->index;
 262
 263	/* Pass through to switch driver if it supports timestamping */
 264	switch (cmd) {
 265	case SIOCGHWTSTAMP:
 266		if (ds->ops->port_hwtstamp_get)
 267			return ds->ops->port_hwtstamp_get(ds, port, ifr);
 268		break;
 269	case SIOCSHWTSTAMP:
 270		if (ds->ops->port_hwtstamp_set)
 271			return ds->ops->port_hwtstamp_set(ds, port, ifr);
 272		break;
 273	}
 274
 275	if (!dev->phydev)
 276		return -ENODEV;
 277
 278	return phy_mii_ioctl(dev->phydev, ifr, cmd);
 279}
 280
 281static int dsa_slave_port_attr_set(struct net_device *dev,
 282				   const struct switchdev_attr *attr,
 283				   struct switchdev_trans *trans)
 284{
 285	struct dsa_port *dp = dsa_slave_to_port(dev);
 286	int ret;
 287
 288	switch (attr->id) {
 289	case SWITCHDEV_ATTR_ID_PORT_STP_STATE:
 290		ret = dsa_port_set_state(dp, attr->u.stp_state, trans);
 291		break;
 292	case SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING:
 293		ret = dsa_port_vlan_filtering(dp, attr->u.vlan_filtering,
 294					      trans);
 295		break;
 296	case SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME:
 297		ret = dsa_port_ageing_time(dp, attr->u.ageing_time, trans);
 298		break;
 
 
 
 
 
 
 
 
 
 
 299	default:
 300		ret = -EOPNOTSUPP;
 301		break;
 302	}
 303
 304	return ret;
 305}
 306
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 307static int dsa_slave_port_obj_add(struct net_device *dev,
 308				  const struct switchdev_obj *obj,
 309				  struct switchdev_trans *trans)
 
 310{
 311	struct dsa_port *dp = dsa_slave_to_port(dev);
 312	int err;
 313
 314	/* For the prepare phase, ensure the full set of changes is feasable in
 315	 * one go in order to signal a failure properly. If an operation is not
 316	 * supported, return -EOPNOTSUPP.
 317	 */
 318
 319	switch (obj->id) {
 320	case SWITCHDEV_OBJ_ID_PORT_MDB:
 
 
 321		err = dsa_port_mdb_add(dp, SWITCHDEV_OBJ_PORT_MDB(obj), trans);
 322		break;
 323	case SWITCHDEV_OBJ_ID_HOST_MDB:
 324		/* DSA can directly translate this to a normal MDB add,
 325		 * but on the CPU port.
 326		 */
 327		err = dsa_port_mdb_add(dp->cpu_dp, SWITCHDEV_OBJ_PORT_MDB(obj),
 328				       trans);
 329		break;
 330	case SWITCHDEV_OBJ_ID_PORT_VLAN:
 331		err = dsa_port_vlan_add(dp, SWITCHDEV_OBJ_PORT_VLAN(obj),
 332					trans);
 333		break;
 334	default:
 335		err = -EOPNOTSUPP;
 336		break;
 337	}
 338
 339	return err;
 340}
 341
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 342static int dsa_slave_port_obj_del(struct net_device *dev,
 343				  const struct switchdev_obj *obj)
 344{
 345	struct dsa_port *dp = dsa_slave_to_port(dev);
 346	int err;
 347
 348	switch (obj->id) {
 349	case SWITCHDEV_OBJ_ID_PORT_MDB:
 
 
 350		err = dsa_port_mdb_del(dp, SWITCHDEV_OBJ_PORT_MDB(obj));
 351		break;
 352	case SWITCHDEV_OBJ_ID_HOST_MDB:
 353		/* DSA can directly translate this to a normal MDB add,
 354		 * but on the CPU port.
 355		 */
 356		err = dsa_port_mdb_del(dp->cpu_dp, SWITCHDEV_OBJ_PORT_MDB(obj));
 357		break;
 358	case SWITCHDEV_OBJ_ID_PORT_VLAN:
 359		err = dsa_port_vlan_del(dp, SWITCHDEV_OBJ_PORT_VLAN(obj));
 360		break;
 361	default:
 362		err = -EOPNOTSUPP;
 363		break;
 364	}
 365
 366	return err;
 367}
 368
 369static int dsa_slave_port_attr_get(struct net_device *dev,
 370				   struct switchdev_attr *attr)
 371{
 372	struct dsa_port *dp = dsa_slave_to_port(dev);
 373	struct dsa_switch *ds = dp->ds;
 374	struct dsa_switch_tree *dst = ds->dst;
 375
 376	switch (attr->id) {
 377	case SWITCHDEV_ATTR_ID_PORT_PARENT_ID:
 378		attr->u.ppid.id_len = sizeof(dst->index);
 379		memcpy(&attr->u.ppid.id, &dst->index, attr->u.ppid.id_len);
 380		break;
 381	case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS_SUPPORT:
 382		attr->u.brport_flags_support = 0;
 383		break;
 384	default:
 385		return -EOPNOTSUPP;
 386	}
 
 
 387
 388	return 0;
 389}
 390
 391static inline netdev_tx_t dsa_slave_netpoll_send_skb(struct net_device *dev,
 392						     struct sk_buff *skb)
 393{
 394#ifdef CONFIG_NET_POLL_CONTROLLER
 395	struct dsa_slave_priv *p = netdev_priv(dev);
 396
 397	if (p->netpoll)
 398		netpoll_send_skb(p->netpoll, skb);
 399#else
 400	BUG();
 401#endif
 402	return NETDEV_TX_OK;
 403}
 404
 405static void dsa_skb_tx_timestamp(struct dsa_slave_priv *p,
 406				 struct sk_buff *skb)
 407{
 408	struct dsa_switch *ds = p->dp->ds;
 409	struct sk_buff *clone;
 410	unsigned int type;
 411
 412	type = ptp_classify_raw(skb);
 413	if (type == PTP_CLASS_NONE)
 414		return;
 415
 416	if (!ds->ops->port_txtstamp)
 417		return;
 418
 419	clone = skb_clone_sk(skb);
 420	if (!clone)
 421		return;
 422
 
 
 423	if (ds->ops->port_txtstamp(ds, p->dp->index, clone, type))
 424		return;
 425
 426	kfree_skb(clone);
 427}
 428
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 429static netdev_tx_t dsa_slave_xmit(struct sk_buff *skb, struct net_device *dev)
 430{
 431	struct dsa_slave_priv *p = netdev_priv(dev);
 432	struct pcpu_sw_netstats *s;
 433	struct sk_buff *nskb;
 434
 435	s = this_cpu_ptr(p->stats64);
 436	u64_stats_update_begin(&s->syncp);
 437	s->tx_packets++;
 438	s->tx_bytes += skb->len;
 439	u64_stats_update_end(&s->syncp);
 440
 
 
 
 441	/* Identify PTP protocol packets, clone them, and pass them to the
 442	 * switch driver
 443	 */
 444	dsa_skb_tx_timestamp(p, skb);
 445
 446	/* Transmit function may have to reallocate the original SKB,
 447	 * in which case it must have freed it. Only free it here on error.
 448	 */
 449	nskb = p->xmit(skb, dev);
 450	if (!nskb) {
 451		kfree_skb(skb);
 
 452		return NETDEV_TX_OK;
 453	}
 454
 455	/* SKB for netpoll still need to be mangled with the protocol-specific
 456	 * tag to be successfully transmitted
 457	 */
 458	if (unlikely(netpoll_tx_running(dev)))
 459		return dsa_slave_netpoll_send_skb(dev, nskb);
 460
 461	/* Queue the SKB for transmission on the parent interface, but
 462	 * do not modify its EtherType
 463	 */
 464	nskb->dev = dsa_slave_to_master(dev);
 465	dev_queue_xmit(nskb);
 466
 467	return NETDEV_TX_OK;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 468}
 469
 470/* ethtool operations *******************************************************/
 471
 472static void dsa_slave_get_drvinfo(struct net_device *dev,
 473				  struct ethtool_drvinfo *drvinfo)
 474{
 475	strlcpy(drvinfo->driver, "dsa", sizeof(drvinfo->driver));
 476	strlcpy(drvinfo->fw_version, "N/A", sizeof(drvinfo->fw_version));
 477	strlcpy(drvinfo->bus_info, "platform", sizeof(drvinfo->bus_info));
 478}
 479
 480static int dsa_slave_get_regs_len(struct net_device *dev)
 481{
 482	struct dsa_port *dp = dsa_slave_to_port(dev);
 483	struct dsa_switch *ds = dp->ds;
 484
 485	if (ds->ops->get_regs_len)
 486		return ds->ops->get_regs_len(ds, dp->index);
 487
 488	return -EOPNOTSUPP;
 489}
 490
 491static void
 492dsa_slave_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *_p)
 493{
 494	struct dsa_port *dp = dsa_slave_to_port(dev);
 495	struct dsa_switch *ds = dp->ds;
 496
 497	if (ds->ops->get_regs)
 498		ds->ops->get_regs(ds, dp->index, regs, _p);
 499}
 500
 501static u32 dsa_slave_get_link(struct net_device *dev)
 502{
 503	if (!dev->phydev)
 504		return -ENODEV;
 505
 506	genphy_update_link(dev->phydev);
 507
 508	return dev->phydev->link;
 509}
 510
 511static int dsa_slave_get_eeprom_len(struct net_device *dev)
 512{
 513	struct dsa_port *dp = dsa_slave_to_port(dev);
 514	struct dsa_switch *ds = dp->ds;
 515
 516	if (ds->cd && ds->cd->eeprom_len)
 517		return ds->cd->eeprom_len;
 518
 519	if (ds->ops->get_eeprom_len)
 520		return ds->ops->get_eeprom_len(ds);
 521
 522	return 0;
 523}
 524
 525static int dsa_slave_get_eeprom(struct net_device *dev,
 526				struct ethtool_eeprom *eeprom, u8 *data)
 527{
 528	struct dsa_port *dp = dsa_slave_to_port(dev);
 529	struct dsa_switch *ds = dp->ds;
 530
 531	if (ds->ops->get_eeprom)
 532		return ds->ops->get_eeprom(ds, eeprom, data);
 533
 534	return -EOPNOTSUPP;
 535}
 536
 537static int dsa_slave_set_eeprom(struct net_device *dev,
 538				struct ethtool_eeprom *eeprom, u8 *data)
 539{
 540	struct dsa_port *dp = dsa_slave_to_port(dev);
 541	struct dsa_switch *ds = dp->ds;
 542
 543	if (ds->ops->set_eeprom)
 544		return ds->ops->set_eeprom(ds, eeprom, data);
 545
 546	return -EOPNOTSUPP;
 547}
 548
 549static void dsa_slave_get_strings(struct net_device *dev,
 550				  uint32_t stringset, uint8_t *data)
 551{
 552	struct dsa_port *dp = dsa_slave_to_port(dev);
 553	struct dsa_switch *ds = dp->ds;
 554
 555	if (stringset == ETH_SS_STATS) {
 556		int len = ETH_GSTRING_LEN;
 557
 558		strncpy(data, "tx_packets", len);
 559		strncpy(data + len, "tx_bytes", len);
 560		strncpy(data + 2 * len, "rx_packets", len);
 561		strncpy(data + 3 * len, "rx_bytes", len);
 562		if (ds->ops->get_strings)
 563			ds->ops->get_strings(ds, dp->index, data + 4 * len);
 
 564	}
 565}
 566
 567static void dsa_slave_get_ethtool_stats(struct net_device *dev,
 568					struct ethtool_stats *stats,
 569					uint64_t *data)
 570{
 571	struct dsa_port *dp = dsa_slave_to_port(dev);
 572	struct dsa_slave_priv *p = netdev_priv(dev);
 573	struct dsa_switch *ds = dp->ds;
 574	struct pcpu_sw_netstats *s;
 575	unsigned int start;
 576	int i;
 577
 578	for_each_possible_cpu(i) {
 579		u64 tx_packets, tx_bytes, rx_packets, rx_bytes;
 580
 581		s = per_cpu_ptr(p->stats64, i);
 582		do {
 583			start = u64_stats_fetch_begin_irq(&s->syncp);
 584			tx_packets = s->tx_packets;
 585			tx_bytes = s->tx_bytes;
 586			rx_packets = s->rx_packets;
 587			rx_bytes = s->rx_bytes;
 588		} while (u64_stats_fetch_retry_irq(&s->syncp, start));
 589		data[0] += tx_packets;
 590		data[1] += tx_bytes;
 591		data[2] += rx_packets;
 592		data[3] += rx_bytes;
 593	}
 594	if (ds->ops->get_ethtool_stats)
 595		ds->ops->get_ethtool_stats(ds, dp->index, data + 4);
 596}
 597
 598static int dsa_slave_get_sset_count(struct net_device *dev, int sset)
 599{
 600	struct dsa_port *dp = dsa_slave_to_port(dev);
 601	struct dsa_switch *ds = dp->ds;
 602
 603	if (sset == ETH_SS_STATS) {
 604		int count;
 605
 606		count = 4;
 607		if (ds->ops->get_sset_count)
 608			count += ds->ops->get_sset_count(ds, dp->index);
 609
 610		return count;
 611	}
 612
 613	return -EOPNOTSUPP;
 614}
 615
 616static void dsa_slave_get_wol(struct net_device *dev, struct ethtool_wolinfo *w)
 617{
 618	struct dsa_port *dp = dsa_slave_to_port(dev);
 619	struct dsa_switch *ds = dp->ds;
 620
 
 
 621	if (ds->ops->get_wol)
 622		ds->ops->get_wol(ds, dp->index, w);
 623}
 624
 625static int dsa_slave_set_wol(struct net_device *dev, struct ethtool_wolinfo *w)
 626{
 627	struct dsa_port *dp = dsa_slave_to_port(dev);
 628	struct dsa_switch *ds = dp->ds;
 629	int ret = -EOPNOTSUPP;
 630
 
 
 631	if (ds->ops->set_wol)
 632		ret = ds->ops->set_wol(ds, dp->index, w);
 633
 634	return ret;
 635}
 636
 637static int dsa_slave_set_eee(struct net_device *dev, struct ethtool_eee *e)
 638{
 639	struct dsa_port *dp = dsa_slave_to_port(dev);
 640	struct dsa_switch *ds = dp->ds;
 641	int ret;
 642
 643	/* Port's PHY and MAC both need to be EEE capable */
 644	if (!dev->phydev)
 645		return -ENODEV;
 646
 647	if (!ds->ops->set_mac_eee)
 648		return -EOPNOTSUPP;
 649
 650	ret = ds->ops->set_mac_eee(ds, dp->index, e);
 651	if (ret)
 652		return ret;
 653
 654	if (e->eee_enabled) {
 655		ret = phy_init_eee(dev->phydev, 0);
 656		if (ret)
 657			return ret;
 658	}
 659
 660	return phy_ethtool_set_eee(dev->phydev, e);
 661}
 662
 663static int dsa_slave_get_eee(struct net_device *dev, struct ethtool_eee *e)
 664{
 665	struct dsa_port *dp = dsa_slave_to_port(dev);
 666	struct dsa_switch *ds = dp->ds;
 667	int ret;
 668
 669	/* Port's PHY and MAC both need to be EEE capable */
 670	if (!dev->phydev)
 671		return -ENODEV;
 672
 673	if (!ds->ops->get_mac_eee)
 674		return -EOPNOTSUPP;
 675
 676	ret = ds->ops->get_mac_eee(ds, dp->index, e);
 677	if (ret)
 678		return ret;
 679
 680	return phy_ethtool_get_eee(dev->phydev, e);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 681}
 682
 683#ifdef CONFIG_NET_POLL_CONTROLLER
 684static int dsa_slave_netpoll_setup(struct net_device *dev,
 685				   struct netpoll_info *ni)
 686{
 687	struct net_device *master = dsa_slave_to_master(dev);
 688	struct dsa_slave_priv *p = netdev_priv(dev);
 689	struct netpoll *netpoll;
 690	int err = 0;
 691
 692	netpoll = kzalloc(sizeof(*netpoll), GFP_KERNEL);
 693	if (!netpoll)
 694		return -ENOMEM;
 695
 696	err = __netpoll_setup(netpoll, master);
 697	if (err) {
 698		kfree(netpoll);
 699		goto out;
 700	}
 701
 702	p->netpoll = netpoll;
 703out:
 704	return err;
 705}
 706
 707static void dsa_slave_netpoll_cleanup(struct net_device *dev)
 708{
 709	struct dsa_slave_priv *p = netdev_priv(dev);
 710	struct netpoll *netpoll = p->netpoll;
 711
 712	if (!netpoll)
 713		return;
 714
 715	p->netpoll = NULL;
 716
 717	__netpoll_free_async(netpoll);
 718}
 719
 720static void dsa_slave_poll_controller(struct net_device *dev)
 721{
 722}
 723#endif
 724
 725static int dsa_slave_get_phys_port_name(struct net_device *dev,
 726					char *name, size_t len)
 727{
 728	struct dsa_port *dp = dsa_slave_to_port(dev);
 729
 
 
 
 
 
 
 
 730	if (snprintf(name, len, "p%d", dp->index) >= len)
 731		return -EINVAL;
 732
 733	return 0;
 734}
 735
 736static struct dsa_mall_tc_entry *
 737dsa_slave_mall_tc_entry_find(struct net_device *dev, unsigned long cookie)
 738{
 739	struct dsa_slave_priv *p = netdev_priv(dev);
 740	struct dsa_mall_tc_entry *mall_tc_entry;
 741
 742	list_for_each_entry(mall_tc_entry, &p->mall_tc_list, list)
 743		if (mall_tc_entry->cookie == cookie)
 744			return mall_tc_entry;
 745
 746	return NULL;
 747}
 748
 749static int dsa_slave_add_cls_matchall(struct net_device *dev,
 750				      struct tc_cls_matchall_offload *cls,
 751				      bool ingress)
 752{
 753	struct dsa_port *dp = dsa_slave_to_port(dev);
 754	struct dsa_slave_priv *p = netdev_priv(dev);
 755	struct dsa_mall_tc_entry *mall_tc_entry;
 756	__be16 protocol = cls->common.protocol;
 757	struct dsa_switch *ds = dp->ds;
 758	struct net_device *to_dev;
 759	const struct tc_action *a;
 760	struct dsa_port *to_dp;
 761	int err = -EOPNOTSUPP;
 762	LIST_HEAD(actions);
 763
 764	if (!ds->ops->port_mirror_add)
 765		return err;
 766
 767	if (!tcf_exts_has_one_action(cls->exts))
 768		return err;
 769
 770	tcf_exts_to_list(cls->exts, &actions);
 771	a = list_first_entry(&actions, struct tc_action, list);
 772
 773	if (is_tcf_mirred_egress_mirror(a) && protocol == htons(ETH_P_ALL)) {
 774		struct dsa_mall_mirror_tc_entry *mirror;
 775
 776		to_dev = tcf_mirred_dev(a);
 777		if (!to_dev)
 778			return -EINVAL;
 779
 780		if (!dsa_slave_dev_check(to_dev))
 781			return -EOPNOTSUPP;
 782
 783		mall_tc_entry = kzalloc(sizeof(*mall_tc_entry), GFP_KERNEL);
 784		if (!mall_tc_entry)
 785			return -ENOMEM;
 786
 787		mall_tc_entry->cookie = cls->cookie;
 788		mall_tc_entry->type = DSA_PORT_MALL_MIRROR;
 789		mirror = &mall_tc_entry->mirror;
 790
 791		to_dp = dsa_slave_to_port(to_dev);
 792
 793		mirror->to_local_port = to_dp->index;
 794		mirror->ingress = ingress;
 795
 796		err = ds->ops->port_mirror_add(ds, dp->index, mirror, ingress);
 797		if (err) {
 798			kfree(mall_tc_entry);
 799			return err;
 800		}
 801
 802		list_add_tail(&mall_tc_entry->list, &p->mall_tc_list);
 803	}
 804
 805	return 0;
 806}
 807
 808static void dsa_slave_del_cls_matchall(struct net_device *dev,
 809				       struct tc_cls_matchall_offload *cls)
 810{
 811	struct dsa_port *dp = dsa_slave_to_port(dev);
 812	struct dsa_mall_tc_entry *mall_tc_entry;
 813	struct dsa_switch *ds = dp->ds;
 814
 815	if (!ds->ops->port_mirror_del)
 816		return;
 817
 818	mall_tc_entry = dsa_slave_mall_tc_entry_find(dev, cls->cookie);
 819	if (!mall_tc_entry)
 820		return;
 821
 822	list_del(&mall_tc_entry->list);
 823
 824	switch (mall_tc_entry->type) {
 825	case DSA_PORT_MALL_MIRROR:
 826		ds->ops->port_mirror_del(ds, dp->index, &mall_tc_entry->mirror);
 827		break;
 828	default:
 829		WARN_ON(1);
 830	}
 831
 832	kfree(mall_tc_entry);
 833}
 834
 835static int dsa_slave_setup_tc_cls_matchall(struct net_device *dev,
 836					   struct tc_cls_matchall_offload *cls,
 837					   bool ingress)
 838{
 839	if (cls->common.chain_index)
 840		return -EOPNOTSUPP;
 841
 842	switch (cls->command) {
 843	case TC_CLSMATCHALL_REPLACE:
 844		return dsa_slave_add_cls_matchall(dev, cls, ingress);
 845	case TC_CLSMATCHALL_DESTROY:
 846		dsa_slave_del_cls_matchall(dev, cls);
 847		return 0;
 848	default:
 849		return -EOPNOTSUPP;
 850	}
 851}
 852
 853static int dsa_slave_setup_tc_block_cb(enum tc_setup_type type, void *type_data,
 854				       void *cb_priv, bool ingress)
 855{
 856	struct net_device *dev = cb_priv;
 857
 858	if (!tc_can_offload(dev))
 859		return -EOPNOTSUPP;
 860
 861	switch (type) {
 862	case TC_SETUP_CLSMATCHALL:
 863		return dsa_slave_setup_tc_cls_matchall(dev, type_data, ingress);
 864	default:
 865		return -EOPNOTSUPP;
 866	}
 867}
 868
 869static int dsa_slave_setup_tc_block_cb_ig(enum tc_setup_type type,
 870					  void *type_data, void *cb_priv)
 871{
 872	return dsa_slave_setup_tc_block_cb(type, type_data, cb_priv, true);
 873}
 874
 875static int dsa_slave_setup_tc_block_cb_eg(enum tc_setup_type type,
 876					  void *type_data, void *cb_priv)
 877{
 878	return dsa_slave_setup_tc_block_cb(type, type_data, cb_priv, false);
 879}
 880
 
 
 881static int dsa_slave_setup_tc_block(struct net_device *dev,
 882				    struct tc_block_offload *f)
 883{
 884	tc_setup_cb_t *cb;
 
 885
 886	if (f->binder_type == TCF_BLOCK_BINDER_TYPE_CLSACT_INGRESS)
 887		cb = dsa_slave_setup_tc_block_cb_ig;
 888	else if (f->binder_type == TCF_BLOCK_BINDER_TYPE_CLSACT_EGRESS)
 889		cb = dsa_slave_setup_tc_block_cb_eg;
 890	else
 891		return -EOPNOTSUPP;
 892
 
 
 893	switch (f->command) {
 894	case TC_BLOCK_BIND:
 895		return tcf_block_cb_register(f->block, cb, dev, dev);
 896	case TC_BLOCK_UNBIND:
 897		tcf_block_cb_unregister(f->block, cb, dev);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 898		return 0;
 899	default:
 900		return -EOPNOTSUPP;
 901	}
 902}
 903
 904static int dsa_slave_setup_tc(struct net_device *dev, enum tc_setup_type type,
 905			      void *type_data)
 906{
 907	switch (type) {
 908	case TC_SETUP_BLOCK:
 
 
 909		return dsa_slave_setup_tc_block(dev, type_data);
 910	default:
 
 911		return -EOPNOTSUPP;
 912	}
 
 913}
 914
 915static void dsa_slave_get_stats64(struct net_device *dev,
 916				  struct rtnl_link_stats64 *stats)
 917{
 918	struct dsa_slave_priv *p = netdev_priv(dev);
 919	struct pcpu_sw_netstats *s;
 920	unsigned int start;
 921	int i;
 922
 923	netdev_stats_to_stats64(stats, &dev->stats);
 924	for_each_possible_cpu(i) {
 925		u64 tx_packets, tx_bytes, rx_packets, rx_bytes;
 926
 927		s = per_cpu_ptr(p->stats64, i);
 928		do {
 929			start = u64_stats_fetch_begin_irq(&s->syncp);
 930			tx_packets = s->tx_packets;
 931			tx_bytes = s->tx_bytes;
 932			rx_packets = s->rx_packets;
 933			rx_bytes = s->rx_bytes;
 934		} while (u64_stats_fetch_retry_irq(&s->syncp, start));
 935
 936		stats->tx_packets += tx_packets;
 937		stats->tx_bytes += tx_bytes;
 938		stats->rx_packets += rx_packets;
 939		stats->rx_bytes += rx_bytes;
 940	}
 941}
 942
 943static int dsa_slave_get_rxnfc(struct net_device *dev,
 944			       struct ethtool_rxnfc *nfc, u32 *rule_locs)
 945{
 946	struct dsa_port *dp = dsa_slave_to_port(dev);
 947	struct dsa_switch *ds = dp->ds;
 948
 949	if (!ds->ops->get_rxnfc)
 950		return -EOPNOTSUPP;
 951
 952	return ds->ops->get_rxnfc(ds, dp->index, nfc, rule_locs);
 953}
 954
 955static int dsa_slave_set_rxnfc(struct net_device *dev,
 956			       struct ethtool_rxnfc *nfc)
 957{
 958	struct dsa_port *dp = dsa_slave_to_port(dev);
 959	struct dsa_switch *ds = dp->ds;
 960
 961	if (!ds->ops->set_rxnfc)
 962		return -EOPNOTSUPP;
 963
 964	return ds->ops->set_rxnfc(ds, dp->index, nfc);
 965}
 966
 967static int dsa_slave_get_ts_info(struct net_device *dev,
 968				 struct ethtool_ts_info *ts)
 969{
 970	struct dsa_slave_priv *p = netdev_priv(dev);
 971	struct dsa_switch *ds = p->dp->ds;
 972
 973	if (!ds->ops->get_ts_info)
 974		return -EOPNOTSUPP;
 975
 976	return ds->ops->get_ts_info(ds, p->dp->index, ts);
 977}
 978
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 979static const struct ethtool_ops dsa_slave_ethtool_ops = {
 980	.get_drvinfo		= dsa_slave_get_drvinfo,
 981	.get_regs_len		= dsa_slave_get_regs_len,
 982	.get_regs		= dsa_slave_get_regs,
 983	.nway_reset		= phy_ethtool_nway_reset,
 984	.get_link		= dsa_slave_get_link,
 985	.get_eeprom_len		= dsa_slave_get_eeprom_len,
 986	.get_eeprom		= dsa_slave_get_eeprom,
 987	.set_eeprom		= dsa_slave_set_eeprom,
 988	.get_strings		= dsa_slave_get_strings,
 989	.get_ethtool_stats	= dsa_slave_get_ethtool_stats,
 990	.get_sset_count		= dsa_slave_get_sset_count,
 991	.set_wol		= dsa_slave_set_wol,
 992	.get_wol		= dsa_slave_get_wol,
 993	.set_eee		= dsa_slave_set_eee,
 994	.get_eee		= dsa_slave_get_eee,
 995	.get_link_ksettings	= phy_ethtool_get_link_ksettings,
 996	.set_link_ksettings	= phy_ethtool_set_link_ksettings,
 997	.get_rxnfc		= dsa_slave_get_rxnfc,
 998	.set_rxnfc		= dsa_slave_set_rxnfc,
 999	.get_ts_info		= dsa_slave_get_ts_info,
1000};
1001
1002/* legacy way, bypassing the bridge *****************************************/
1003int dsa_legacy_fdb_add(struct ndmsg *ndm, struct nlattr *tb[],
1004		       struct net_device *dev,
1005		       const unsigned char *addr, u16 vid,
1006		       u16 flags)
 
1007{
1008	struct dsa_port *dp = dsa_slave_to_port(dev);
1009
1010	return dsa_port_fdb_add(dp, addr, vid);
1011}
1012
1013int dsa_legacy_fdb_del(struct ndmsg *ndm, struct nlattr *tb[],
1014		       struct net_device *dev,
1015		       const unsigned char *addr, u16 vid)
1016{
1017	struct dsa_port *dp = dsa_slave_to_port(dev);
1018
1019	return dsa_port_fdb_del(dp, addr, vid);
1020}
1021
 
 
 
 
 
 
 
1022static const struct net_device_ops dsa_slave_netdev_ops = {
1023	.ndo_open	 	= dsa_slave_open,
1024	.ndo_stop		= dsa_slave_close,
1025	.ndo_start_xmit		= dsa_slave_xmit,
1026	.ndo_change_rx_flags	= dsa_slave_change_rx_flags,
1027	.ndo_set_rx_mode	= dsa_slave_set_rx_mode,
1028	.ndo_set_mac_address	= dsa_slave_set_mac_address,
1029	.ndo_fdb_add		= dsa_legacy_fdb_add,
1030	.ndo_fdb_del		= dsa_legacy_fdb_del,
1031	.ndo_fdb_dump		= dsa_slave_fdb_dump,
1032	.ndo_do_ioctl		= dsa_slave_ioctl,
1033	.ndo_get_iflink		= dsa_slave_get_iflink,
1034#ifdef CONFIG_NET_POLL_CONTROLLER
1035	.ndo_netpoll_setup	= dsa_slave_netpoll_setup,
1036	.ndo_netpoll_cleanup	= dsa_slave_netpoll_cleanup,
1037	.ndo_poll_controller	= dsa_slave_poll_controller,
1038#endif
1039	.ndo_get_phys_port_name	= dsa_slave_get_phys_port_name,
1040	.ndo_setup_tc		= dsa_slave_setup_tc,
1041	.ndo_get_stats64	= dsa_slave_get_stats64,
1042};
1043
1044static const struct switchdev_ops dsa_slave_switchdev_ops = {
1045	.switchdev_port_attr_get	= dsa_slave_port_attr_get,
1046	.switchdev_port_attr_set	= dsa_slave_port_attr_set,
1047	.switchdev_port_obj_add		= dsa_slave_port_obj_add,
1048	.switchdev_port_obj_del		= dsa_slave_port_obj_del,
1049};
1050
1051static struct device_type dsa_type = {
1052	.name	= "dsa",
1053};
1054
1055static void dsa_slave_adjust_link(struct net_device *dev)
1056{
1057	struct dsa_port *dp = dsa_slave_to_port(dev);
1058	struct dsa_slave_priv *p = netdev_priv(dev);
1059	struct dsa_switch *ds = dp->ds;
1060	unsigned int status_changed = 0;
1061
1062	if (p->old_link != dev->phydev->link) {
1063		status_changed = 1;
1064		p->old_link = dev->phydev->link;
1065	}
1066
1067	if (p->old_duplex != dev->phydev->duplex) {
1068		status_changed = 1;
1069		p->old_duplex = dev->phydev->duplex;
1070	}
1071
1072	if (p->old_pause != dev->phydev->pause) {
1073		status_changed = 1;
1074		p->old_pause = dev->phydev->pause;
1075	}
1076
1077	if (ds->ops->adjust_link && status_changed)
1078		ds->ops->adjust_link(ds, dp->index, dev->phydev);
1079
1080	if (status_changed)
1081		phy_print_status(dev->phydev);
1082}
 
1083
1084static int dsa_slave_fixed_link_update(struct net_device *dev,
1085				       struct fixed_phy_status *status)
1086{
1087	struct dsa_switch *ds;
1088	struct dsa_port *dp;
1089
1090	if (dev) {
1091		dp = dsa_slave_to_port(dev);
1092		ds = dp->ds;
1093		if (ds->ops->fixed_link_update)
1094			ds->ops->fixed_link_update(ds, dp->index, status);
1095	}
1096
1097	return 0;
 
 
 
1098}
1099
1100/* slave device setup *******************************************************/
1101static int dsa_slave_phy_connect(struct net_device *slave_dev, int addr)
1102{
1103	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1104	struct dsa_slave_priv *p = netdev_priv(slave_dev);
1105	struct dsa_switch *ds = dp->ds;
1106
1107	slave_dev->phydev = mdiobus_get_phy(ds->slave_mii_bus, addr);
1108	if (!slave_dev->phydev) {
1109		netdev_err(slave_dev, "no phy at %d\n", addr);
1110		return -ENODEV;
1111	}
1112
1113	/* Use already configured phy mode */
1114	if (p->phy_interface == PHY_INTERFACE_MODE_NA)
1115		p->phy_interface = slave_dev->phydev->interface;
1116
1117	return phy_connect_direct(slave_dev, slave_dev->phydev,
1118				  dsa_slave_adjust_link, p->phy_interface);
1119}
1120
1121static int dsa_slave_phy_setup(struct net_device *slave_dev)
1122{
1123	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1124	struct dsa_slave_priv *p = netdev_priv(slave_dev);
1125	struct device_node *port_dn = dp->dn;
1126	struct dsa_switch *ds = dp->ds;
1127	struct device_node *phy_dn;
1128	bool phy_is_fixed = false;
1129	u32 phy_flags = 0;
1130	int mode, ret;
1131
1132	mode = of_get_phy_mode(port_dn);
1133	if (mode < 0)
1134		mode = PHY_INTERFACE_MODE_NA;
1135	p->phy_interface = mode;
1136
1137	phy_dn = of_parse_phandle(port_dn, "phy-handle", 0);
1138	if (!phy_dn && of_phy_is_fixed_link(port_dn)) {
1139		/* In the case of a fixed PHY, the DT node associated
1140		 * to the fixed PHY is the Port DT node
1141		 */
1142		ret = of_phy_register_fixed_link(port_dn);
1143		if (ret) {
1144			netdev_err(slave_dev, "failed to register fixed PHY: %d\n", ret);
1145			return ret;
1146		}
1147		phy_is_fixed = true;
1148		phy_dn = of_node_get(port_dn);
1149	}
1150
 
 
 
 
 
 
 
1151	if (ds->ops->get_phy_flags)
1152		phy_flags = ds->ops->get_phy_flags(ds, dp->index);
1153
1154	if (phy_dn) {
1155		slave_dev->phydev = of_phy_connect(slave_dev, phy_dn,
1156						   dsa_slave_adjust_link,
1157						   phy_flags,
1158						   p->phy_interface);
1159		of_node_put(phy_dn);
1160	}
1161
1162	if (slave_dev->phydev && phy_is_fixed)
1163		fixed_phy_set_link_update(slave_dev->phydev,
1164					  dsa_slave_fixed_link_update);
1165
1166	/* We could not connect to a designated PHY, so use the switch internal
1167	 * MDIO bus instead
1168	 */
1169	if (!slave_dev->phydev) {
1170		ret = dsa_slave_phy_connect(slave_dev, dp->index);
1171		if (ret) {
1172			netdev_err(slave_dev, "failed to connect to port %d: %d\n",
 
1173				   dp->index, ret);
1174			if (phy_is_fixed)
1175				of_phy_deregister_fixed_link(port_dn);
1176			return ret;
1177		}
1178	}
1179
1180	phy_attached_info(slave_dev->phydev);
1181
1182	return 0;
1183}
1184
1185static struct lock_class_key dsa_slave_netdev_xmit_lock_key;
1186static void dsa_slave_set_lockdep_class_one(struct net_device *dev,
1187					    struct netdev_queue *txq,
1188					    void *_unused)
1189{
1190	lockdep_set_class(&txq->_xmit_lock,
1191			  &dsa_slave_netdev_xmit_lock_key);
1192}
1193
1194int dsa_slave_suspend(struct net_device *slave_dev)
1195{
1196	struct dsa_slave_priv *p = netdev_priv(slave_dev);
 
 
 
 
 
 
1197
1198	netif_device_detach(slave_dev);
1199
1200	if (slave_dev->phydev) {
1201		phy_stop(slave_dev->phydev);
1202		p->old_pause = -1;
1203		p->old_link = -1;
1204		p->old_duplex = -1;
1205		phy_suspend(slave_dev->phydev);
1206	}
1207
1208	return 0;
1209}
1210
1211int dsa_slave_resume(struct net_device *slave_dev)
1212{
 
 
 
 
 
1213	netif_device_attach(slave_dev);
1214
1215	if (slave_dev->phydev) {
1216		phy_resume(slave_dev->phydev);
1217		phy_start(slave_dev->phydev);
1218	}
1219
1220	return 0;
1221}
1222
1223static void dsa_slave_notify(struct net_device *dev, unsigned long val)
1224{
1225	struct net_device *master = dsa_slave_to_master(dev);
1226	struct dsa_port *dp = dsa_slave_to_port(dev);
1227	struct dsa_notifier_register_info rinfo = {
1228		.switch_number = dp->ds->index,
1229		.port_number = dp->index,
1230		.master = master,
1231		.info.dev = dev,
1232	};
1233
1234	call_dsa_notifiers(val, dev, &rinfo.info);
1235}
1236
1237int dsa_slave_create(struct dsa_port *port)
1238{
1239	const struct dsa_port *cpu_dp = port->cpu_dp;
1240	struct net_device *master = cpu_dp->master;
1241	struct dsa_switch *ds = port->ds;
1242	const char *name = port->name;
1243	struct net_device *slave_dev;
1244	struct dsa_slave_priv *p;
1245	int ret;
1246
1247	if (!ds->num_tx_queues)
1248		ds->num_tx_queues = 1;
1249
1250	slave_dev = alloc_netdev_mqs(sizeof(struct dsa_slave_priv), name,
1251				     NET_NAME_UNKNOWN, ether_setup,
1252				     ds->num_tx_queues, 1);
1253	if (slave_dev == NULL)
1254		return -ENOMEM;
1255
1256	slave_dev->features = master->vlan_features | NETIF_F_HW_TC;
 
 
1257	slave_dev->hw_features |= NETIF_F_HW_TC;
1258	slave_dev->ethtool_ops = &dsa_slave_ethtool_ops;
1259	eth_hw_addr_inherit(slave_dev, master);
 
 
 
1260	slave_dev->priv_flags |= IFF_NO_QUEUE;
1261	slave_dev->netdev_ops = &dsa_slave_netdev_ops;
1262	slave_dev->switchdev_ops = &dsa_slave_switchdev_ops;
1263	slave_dev->min_mtu = 0;
1264	slave_dev->max_mtu = ETH_MAX_MTU;
1265	SET_NETDEV_DEVTYPE(slave_dev, &dsa_type);
1266
1267	netdev_for_each_tx_queue(slave_dev, dsa_slave_set_lockdep_class_one,
1268				 NULL);
1269
1270	SET_NETDEV_DEV(slave_dev, port->ds->dev);
1271	slave_dev->dev.of_node = port->dn;
1272	slave_dev->vlan_features = master->vlan_features;
1273
1274	p = netdev_priv(slave_dev);
1275	p->stats64 = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
1276	if (!p->stats64) {
1277		free_netdev(slave_dev);
1278		return -ENOMEM;
1279	}
1280	p->dp = port;
1281	INIT_LIST_HEAD(&p->mall_tc_list);
 
 
1282	p->xmit = cpu_dp->tag_ops->xmit;
1283
1284	p->old_pause = -1;
1285	p->old_link = -1;
1286	p->old_duplex = -1;
1287
1288	port->slave = slave_dev;
1289
1290	netif_carrier_off(slave_dev);
1291
1292	ret = dsa_slave_phy_setup(slave_dev);
1293	if (ret) {
1294		netdev_err(master, "error %d setting up slave phy\n", ret);
1295		goto out_free;
1296	}
1297
1298	dsa_slave_notify(slave_dev, DSA_PORT_REGISTER);
1299
1300	ret = register_netdev(slave_dev);
1301	if (ret) {
1302		netdev_err(master, "error %d registering interface %s\n",
1303			   ret, slave_dev->name);
1304		goto out_phy;
1305	}
1306
1307	return 0;
1308
1309out_phy:
1310	phy_disconnect(slave_dev->phydev);
1311	if (of_phy_is_fixed_link(port->dn))
1312		of_phy_deregister_fixed_link(port->dn);
 
1313out_free:
1314	free_percpu(p->stats64);
1315	free_netdev(slave_dev);
1316	port->slave = NULL;
1317	return ret;
1318}
1319
1320void dsa_slave_destroy(struct net_device *slave_dev)
1321{
1322	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1323	struct dsa_slave_priv *p = netdev_priv(slave_dev);
1324	struct device_node *port_dn = dp->dn;
1325
1326	netif_carrier_off(slave_dev);
1327	if (slave_dev->phydev) {
1328		phy_disconnect(slave_dev->phydev);
 
1329
1330		if (of_phy_is_fixed_link(port_dn))
1331			of_phy_deregister_fixed_link(port_dn);
1332	}
1333	dsa_slave_notify(slave_dev, DSA_PORT_UNREGISTER);
1334	unregister_netdev(slave_dev);
 
1335	free_percpu(p->stats64);
1336	free_netdev(slave_dev);
1337}
1338
1339static bool dsa_slave_dev_check(struct net_device *dev)
1340{
1341	return dev->netdev_ops == &dsa_slave_netdev_ops;
1342}
1343
1344static int dsa_slave_changeupper(struct net_device *dev,
1345				 struct netdev_notifier_changeupper_info *info)
1346{
1347	struct dsa_port *dp = dsa_slave_to_port(dev);
1348	int err = NOTIFY_DONE;
1349
1350	if (netif_is_bridge_master(info->upper_dev)) {
1351		if (info->linking) {
1352			err = dsa_port_bridge_join(dp, info->upper_dev);
1353			err = notifier_from_errno(err);
1354		} else {
1355			dsa_port_bridge_leave(dp, info->upper_dev);
1356			err = NOTIFY_OK;
1357		}
1358	}
1359
1360	return err;
1361}
1362
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1363static int dsa_slave_netdevice_event(struct notifier_block *nb,
1364				     unsigned long event, void *ptr)
1365{
1366	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1367
1368	if (!dsa_slave_dev_check(dev))
1369		return NOTIFY_DONE;
 
1370
1371	if (event == NETDEV_CHANGEUPPER)
1372		return dsa_slave_changeupper(dev, ptr);
 
1373
1374	return NOTIFY_DONE;
1375}
1376
1377struct dsa_switchdev_event_work {
1378	struct work_struct work;
1379	struct switchdev_notifier_fdb_info fdb_info;
1380	struct net_device *dev;
1381	unsigned long event;
1382};
1383
1384static void dsa_slave_switchdev_event_work(struct work_struct *work)
1385{
1386	struct dsa_switchdev_event_work *switchdev_work =
1387		container_of(work, struct dsa_switchdev_event_work, work);
1388	struct net_device *dev = switchdev_work->dev;
1389	struct switchdev_notifier_fdb_info *fdb_info;
1390	struct dsa_port *dp = dsa_slave_to_port(dev);
1391	int err;
1392
1393	rtnl_lock();
1394	switch (switchdev_work->event) {
1395	case SWITCHDEV_FDB_ADD_TO_DEVICE:
1396		fdb_info = &switchdev_work->fdb_info;
 
 
 
1397		err = dsa_port_fdb_add(dp, fdb_info->addr, fdb_info->vid);
1398		if (err) {
1399			netdev_dbg(dev, "fdb add failed err=%d\n", err);
1400			break;
1401		}
 
1402		call_switchdev_notifiers(SWITCHDEV_FDB_OFFLOADED, dev,
1403					 &fdb_info->info);
1404		break;
1405
1406	case SWITCHDEV_FDB_DEL_TO_DEVICE:
1407		fdb_info = &switchdev_work->fdb_info;
 
 
 
1408		err = dsa_port_fdb_del(dp, fdb_info->addr, fdb_info->vid);
1409		if (err) {
1410			netdev_dbg(dev, "fdb del failed err=%d\n", err);
1411			dev_close(dev);
1412		}
1413		break;
1414	}
1415	rtnl_unlock();
1416
1417	kfree(switchdev_work->fdb_info.addr);
1418	kfree(switchdev_work);
1419	dev_put(dev);
1420}
1421
1422static int
1423dsa_slave_switchdev_fdb_work_init(struct dsa_switchdev_event_work *
1424				  switchdev_work,
1425				  const struct switchdev_notifier_fdb_info *
1426				  fdb_info)
1427{
1428	memcpy(&switchdev_work->fdb_info, fdb_info,
1429	       sizeof(switchdev_work->fdb_info));
1430	switchdev_work->fdb_info.addr = kzalloc(ETH_ALEN, GFP_ATOMIC);
1431	if (!switchdev_work->fdb_info.addr)
1432		return -ENOMEM;
1433	ether_addr_copy((u8 *)switchdev_work->fdb_info.addr,
1434			fdb_info->addr);
1435	return 0;
1436}
1437
1438/* Called under rcu_read_lock() */
1439static int dsa_slave_switchdev_event(struct notifier_block *unused,
1440				     unsigned long event, void *ptr)
1441{
1442	struct net_device *dev = switchdev_notifier_info_to_dev(ptr);
1443	struct dsa_switchdev_event_work *switchdev_work;
 
 
 
 
 
 
 
 
1444
1445	if (!dsa_slave_dev_check(dev))
1446		return NOTIFY_DONE;
1447
1448	switchdev_work = kzalloc(sizeof(*switchdev_work), GFP_ATOMIC);
1449	if (!switchdev_work)
1450		return NOTIFY_BAD;
1451
1452	INIT_WORK(&switchdev_work->work,
1453		  dsa_slave_switchdev_event_work);
1454	switchdev_work->dev = dev;
1455	switchdev_work->event = event;
1456
1457	switch (event) {
1458	case SWITCHDEV_FDB_ADD_TO_DEVICE: /* fall through */
1459	case SWITCHDEV_FDB_DEL_TO_DEVICE:
1460		if (dsa_slave_switchdev_fdb_work_init(switchdev_work,
1461						      ptr))
1462			goto err_fdb_work_init;
1463		dev_hold(dev);
1464		break;
1465	default:
1466		kfree(switchdev_work);
1467		return NOTIFY_DONE;
1468	}
1469
1470	dsa_schedule_work(&switchdev_work->work);
1471	return NOTIFY_OK;
1472
1473err_fdb_work_init:
1474	kfree(switchdev_work);
1475	return NOTIFY_BAD;
1476}
1477
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1478static struct notifier_block dsa_slave_nb __read_mostly = {
1479	.notifier_call  = dsa_slave_netdevice_event,
1480};
1481
1482static struct notifier_block dsa_slave_switchdev_notifier = {
1483	.notifier_call = dsa_slave_switchdev_event,
1484};
1485
 
 
 
 
1486int dsa_slave_register_notifier(void)
1487{
 
1488	int err;
1489
1490	err = register_netdevice_notifier(&dsa_slave_nb);
1491	if (err)
1492		return err;
1493
1494	err = register_switchdev_notifier(&dsa_slave_switchdev_notifier);
1495	if (err)
1496		goto err_switchdev_nb;
1497
 
 
 
 
 
1498	return 0;
1499
 
 
1500err_switchdev_nb:
1501	unregister_netdevice_notifier(&dsa_slave_nb);
1502	return err;
1503}
1504
1505void dsa_slave_unregister_notifier(void)
1506{
 
1507	int err;
 
 
 
 
 
1508
1509	err = unregister_switchdev_notifier(&dsa_slave_switchdev_notifier);
1510	if (err)
1511		pr_err("DSA: failed to unregister switchdev notifier (%d)\n", err);
1512
1513	err = unregister_netdevice_notifier(&dsa_slave_nb);
1514	if (err)
1515		pr_err("DSA: failed to unregister slave notifier (%d)\n", err);
1516}
v5.4
   1// SPDX-License-Identifier: GPL-2.0-or-later
   2/*
   3 * net/dsa/slave.c - Slave device handling
   4 * Copyright (c) 2008-2009 Marvell Semiconductor
 
 
 
 
 
   5 */
   6
   7#include <linux/list.h>
   8#include <linux/etherdevice.h>
   9#include <linux/netdevice.h>
  10#include <linux/phy.h>
  11#include <linux/phy_fixed.h>
  12#include <linux/phylink.h>
  13#include <linux/of_net.h>
  14#include <linux/of_mdio.h>
  15#include <linux/mdio.h>
  16#include <net/rtnetlink.h>
  17#include <net/pkt_cls.h>
  18#include <net/tc_act/tc_mirred.h>
  19#include <linux/if_bridge.h>
  20#include <linux/netpoll.h>
  21#include <linux/ptp_classify.h>
  22
  23#include "dsa_priv.h"
  24
  25static bool dsa_slave_dev_check(const struct net_device *dev);
  26
  27/* slave mii_bus handling ***************************************************/
  28static int dsa_slave_phy_read(struct mii_bus *bus, int addr, int reg)
  29{
  30	struct dsa_switch *ds = bus->priv;
  31
  32	if (ds->phys_mii_mask & (1 << addr))
  33		return ds->ops->phy_read(ds, addr, reg);
  34
  35	return 0xffff;
  36}
  37
  38static int dsa_slave_phy_write(struct mii_bus *bus, int addr, int reg, u16 val)
  39{
  40	struct dsa_switch *ds = bus->priv;
  41
  42	if (ds->phys_mii_mask & (1 << addr))
  43		return ds->ops->phy_write(ds, addr, reg, val);
  44
  45	return 0;
  46}
  47
  48void dsa_slave_mii_bus_init(struct dsa_switch *ds)
  49{
  50	ds->slave_mii_bus->priv = (void *)ds;
  51	ds->slave_mii_bus->name = "dsa slave smi";
  52	ds->slave_mii_bus->read = dsa_slave_phy_read;
  53	ds->slave_mii_bus->write = dsa_slave_phy_write;
  54	snprintf(ds->slave_mii_bus->id, MII_BUS_ID_SIZE, "dsa-%d.%d",
  55		 ds->dst->index, ds->index);
  56	ds->slave_mii_bus->parent = ds->dev;
  57	ds->slave_mii_bus->phy_mask = ~ds->phys_mii_mask;
  58}
  59
  60
  61/* slave device handling ****************************************************/
  62static int dsa_slave_get_iflink(const struct net_device *dev)
  63{
  64	return dsa_slave_to_master(dev)->ifindex;
  65}
  66
  67static int dsa_slave_open(struct net_device *dev)
  68{
  69	struct net_device *master = dsa_slave_to_master(dev);
  70	struct dsa_port *dp = dsa_slave_to_port(dev);
  71	int err;
  72
  73	if (!(master->flags & IFF_UP))
  74		return -ENETDOWN;
  75
  76	if (!ether_addr_equal(dev->dev_addr, master->dev_addr)) {
  77		err = dev_uc_add(master, dev->dev_addr);
  78		if (err < 0)
  79			goto out;
  80	}
  81
  82	if (dev->flags & IFF_ALLMULTI) {
  83		err = dev_set_allmulti(master, 1);
  84		if (err < 0)
  85			goto del_unicast;
  86	}
  87	if (dev->flags & IFF_PROMISC) {
  88		err = dev_set_promiscuity(master, 1);
  89		if (err < 0)
  90			goto clear_allmulti;
  91	}
  92
  93	err = dsa_port_enable(dp, dev->phydev);
  94	if (err)
  95		goto clear_promisc;
  96
  97	phylink_start(dp->pl);
 
  98
  99	return 0;
 100
 101clear_promisc:
 102	if (dev->flags & IFF_PROMISC)
 103		dev_set_promiscuity(master, -1);
 104clear_allmulti:
 105	if (dev->flags & IFF_ALLMULTI)
 106		dev_set_allmulti(master, -1);
 107del_unicast:
 108	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
 109		dev_uc_del(master, dev->dev_addr);
 110out:
 111	return err;
 112}
 113
 114static int dsa_slave_close(struct net_device *dev)
 115{
 116	struct net_device *master = dsa_slave_to_master(dev);
 117	struct dsa_port *dp = dsa_slave_to_port(dev);
 118
 119	cancel_work_sync(&dp->xmit_work);
 120	skb_queue_purge(&dp->xmit_queue);
 121
 122	phylink_stop(dp->pl);
 123
 124	dsa_port_disable(dp);
 125
 126	dev_mc_unsync(master, dev);
 127	dev_uc_unsync(master, dev);
 128	if (dev->flags & IFF_ALLMULTI)
 129		dev_set_allmulti(master, -1);
 130	if (dev->flags & IFF_PROMISC)
 131		dev_set_promiscuity(master, -1);
 132
 133	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
 134		dev_uc_del(master, dev->dev_addr);
 135
 136	return 0;
 137}
 138
 139static void dsa_slave_change_rx_flags(struct net_device *dev, int change)
 140{
 141	struct net_device *master = dsa_slave_to_master(dev);
 142	if (dev->flags & IFF_UP) {
 143		if (change & IFF_ALLMULTI)
 144			dev_set_allmulti(master,
 145					 dev->flags & IFF_ALLMULTI ? 1 : -1);
 146		if (change & IFF_PROMISC)
 147			dev_set_promiscuity(master,
 148					    dev->flags & IFF_PROMISC ? 1 : -1);
 149	}
 150}
 151
 152static void dsa_slave_set_rx_mode(struct net_device *dev)
 153{
 154	struct net_device *master = dsa_slave_to_master(dev);
 155
 156	dev_mc_sync(master, dev);
 157	dev_uc_sync(master, dev);
 158}
 159
 160static int dsa_slave_set_mac_address(struct net_device *dev, void *a)
 161{
 162	struct net_device *master = dsa_slave_to_master(dev);
 163	struct sockaddr *addr = a;
 164	int err;
 165
 166	if (!is_valid_ether_addr(addr->sa_data))
 167		return -EADDRNOTAVAIL;
 168
 169	if (!(dev->flags & IFF_UP))
 170		goto out;
 171
 172	if (!ether_addr_equal(addr->sa_data, master->dev_addr)) {
 173		err = dev_uc_add(master, addr->sa_data);
 174		if (err < 0)
 175			return err;
 176	}
 177
 178	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
 179		dev_uc_del(master, dev->dev_addr);
 180
 181out:
 182	ether_addr_copy(dev->dev_addr, addr->sa_data);
 183
 184	return 0;
 185}
 186
 187struct dsa_slave_dump_ctx {
 188	struct net_device *dev;
 189	struct sk_buff *skb;
 190	struct netlink_callback *cb;
 191	int idx;
 192};
 193
 194static int
 195dsa_slave_port_fdb_do_dump(const unsigned char *addr, u16 vid,
 196			   bool is_static, void *data)
 197{
 198	struct dsa_slave_dump_ctx *dump = data;
 199	u32 portid = NETLINK_CB(dump->cb->skb).portid;
 200	u32 seq = dump->cb->nlh->nlmsg_seq;
 201	struct nlmsghdr *nlh;
 202	struct ndmsg *ndm;
 203
 204	if (dump->idx < dump->cb->args[2])
 205		goto skip;
 206
 207	nlh = nlmsg_put(dump->skb, portid, seq, RTM_NEWNEIGH,
 208			sizeof(*ndm), NLM_F_MULTI);
 209	if (!nlh)
 210		return -EMSGSIZE;
 211
 212	ndm = nlmsg_data(nlh);
 213	ndm->ndm_family  = AF_BRIDGE;
 214	ndm->ndm_pad1    = 0;
 215	ndm->ndm_pad2    = 0;
 216	ndm->ndm_flags   = NTF_SELF;
 217	ndm->ndm_type    = 0;
 218	ndm->ndm_ifindex = dump->dev->ifindex;
 219	ndm->ndm_state   = is_static ? NUD_NOARP : NUD_REACHABLE;
 220
 221	if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, addr))
 222		goto nla_put_failure;
 223
 224	if (vid && nla_put_u16(dump->skb, NDA_VLAN, vid))
 225		goto nla_put_failure;
 226
 227	nlmsg_end(dump->skb, nlh);
 228
 229skip:
 230	dump->idx++;
 231	return 0;
 232
 233nla_put_failure:
 234	nlmsg_cancel(dump->skb, nlh);
 235	return -EMSGSIZE;
 236}
 237
 238static int
 239dsa_slave_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
 240		   struct net_device *dev, struct net_device *filter_dev,
 241		   int *idx)
 242{
 243	struct dsa_port *dp = dsa_slave_to_port(dev);
 244	struct dsa_slave_dump_ctx dump = {
 245		.dev = dev,
 246		.skb = skb,
 247		.cb = cb,
 248		.idx = *idx,
 249	};
 250	int err;
 251
 252	err = dsa_port_fdb_dump(dp, dsa_slave_port_fdb_do_dump, &dump);
 253	*idx = dump.idx;
 254
 255	return err;
 256}
 257
 258static int dsa_slave_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
 259{
 260	struct dsa_slave_priv *p = netdev_priv(dev);
 261	struct dsa_switch *ds = p->dp->ds;
 262	int port = p->dp->index;
 263
 264	/* Pass through to switch driver if it supports timestamping */
 265	switch (cmd) {
 266	case SIOCGHWTSTAMP:
 267		if (ds->ops->port_hwtstamp_get)
 268			return ds->ops->port_hwtstamp_get(ds, port, ifr);
 269		break;
 270	case SIOCSHWTSTAMP:
 271		if (ds->ops->port_hwtstamp_set)
 272			return ds->ops->port_hwtstamp_set(ds, port, ifr);
 273		break;
 274	}
 275
 276	return phylink_mii_ioctl(p->dp->pl, ifr, cmd);
 
 
 
 277}
 278
 279static int dsa_slave_port_attr_set(struct net_device *dev,
 280				   const struct switchdev_attr *attr,
 281				   struct switchdev_trans *trans)
 282{
 283	struct dsa_port *dp = dsa_slave_to_port(dev);
 284	int ret;
 285
 286	switch (attr->id) {
 287	case SWITCHDEV_ATTR_ID_PORT_STP_STATE:
 288		ret = dsa_port_set_state(dp, attr->u.stp_state, trans);
 289		break;
 290	case SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING:
 291		ret = dsa_port_vlan_filtering(dp, attr->u.vlan_filtering,
 292					      trans);
 293		break;
 294	case SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME:
 295		ret = dsa_port_ageing_time(dp, attr->u.ageing_time, trans);
 296		break;
 297	case SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS:
 298		ret = dsa_port_pre_bridge_flags(dp, attr->u.brport_flags,
 299						trans);
 300		break;
 301	case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS:
 302		ret = dsa_port_bridge_flags(dp, attr->u.brport_flags, trans);
 303		break;
 304	case SWITCHDEV_ATTR_ID_BRIDGE_MROUTER:
 305		ret = dsa_port_mrouter(dp->cpu_dp, attr->u.mrouter, trans);
 306		break;
 307	default:
 308		ret = -EOPNOTSUPP;
 309		break;
 310	}
 311
 312	return ret;
 313}
 314
 315static int dsa_slave_vlan_add(struct net_device *dev,
 316			      const struct switchdev_obj *obj,
 317			      struct switchdev_trans *trans)
 318{
 319	struct dsa_port *dp = dsa_slave_to_port(dev);
 320	struct switchdev_obj_port_vlan vlan;
 321	int err;
 322
 323	if (obj->orig_dev != dev)
 324		return -EOPNOTSUPP;
 325
 326	if (dp->bridge_dev && !br_vlan_enabled(dp->bridge_dev))
 327		return 0;
 328
 329	vlan = *SWITCHDEV_OBJ_PORT_VLAN(obj);
 330
 331	err = dsa_port_vlan_add(dp, &vlan, trans);
 332	if (err)
 333		return err;
 334
 335	/* We need the dedicated CPU port to be a member of the VLAN as well.
 336	 * Even though drivers often handle CPU membership in special ways,
 337	 * it doesn't make sense to program a PVID, so clear this flag.
 338	 */
 339	vlan.flags &= ~BRIDGE_VLAN_INFO_PVID;
 340
 341	err = dsa_port_vlan_add(dp->cpu_dp, &vlan, trans);
 342	if (err)
 343		return err;
 344
 345	return 0;
 346}
 347
 348static int dsa_slave_port_obj_add(struct net_device *dev,
 349				  const struct switchdev_obj *obj,
 350				  struct switchdev_trans *trans,
 351				  struct netlink_ext_ack *extack)
 352{
 353	struct dsa_port *dp = dsa_slave_to_port(dev);
 354	int err;
 355
 356	/* For the prepare phase, ensure the full set of changes is feasable in
 357	 * one go in order to signal a failure properly. If an operation is not
 358	 * supported, return -EOPNOTSUPP.
 359	 */
 360
 361	switch (obj->id) {
 362	case SWITCHDEV_OBJ_ID_PORT_MDB:
 363		if (obj->orig_dev != dev)
 364			return -EOPNOTSUPP;
 365		err = dsa_port_mdb_add(dp, SWITCHDEV_OBJ_PORT_MDB(obj), trans);
 366		break;
 367	case SWITCHDEV_OBJ_ID_HOST_MDB:
 368		/* DSA can directly translate this to a normal MDB add,
 369		 * but on the CPU port.
 370		 */
 371		err = dsa_port_mdb_add(dp->cpu_dp, SWITCHDEV_OBJ_PORT_MDB(obj),
 372				       trans);
 373		break;
 374	case SWITCHDEV_OBJ_ID_PORT_VLAN:
 375		err = dsa_slave_vlan_add(dev, obj, trans);
 
 376		break;
 377	default:
 378		err = -EOPNOTSUPP;
 379		break;
 380	}
 381
 382	return err;
 383}
 384
 385static int dsa_slave_vlan_del(struct net_device *dev,
 386			      const struct switchdev_obj *obj)
 387{
 388	struct dsa_port *dp = dsa_slave_to_port(dev);
 389
 390	if (obj->orig_dev != dev)
 391		return -EOPNOTSUPP;
 392
 393	if (dp->bridge_dev && !br_vlan_enabled(dp->bridge_dev))
 394		return 0;
 395
 396	/* Do not deprogram the CPU port as it may be shared with other user
 397	 * ports which can be members of this VLAN as well.
 398	 */
 399	return dsa_port_vlan_del(dp, SWITCHDEV_OBJ_PORT_VLAN(obj));
 400}
 401
 402static int dsa_slave_port_obj_del(struct net_device *dev,
 403				  const struct switchdev_obj *obj)
 404{
 405	struct dsa_port *dp = dsa_slave_to_port(dev);
 406	int err;
 407
 408	switch (obj->id) {
 409	case SWITCHDEV_OBJ_ID_PORT_MDB:
 410		if (obj->orig_dev != dev)
 411			return -EOPNOTSUPP;
 412		err = dsa_port_mdb_del(dp, SWITCHDEV_OBJ_PORT_MDB(obj));
 413		break;
 414	case SWITCHDEV_OBJ_ID_HOST_MDB:
 415		/* DSA can directly translate this to a normal MDB add,
 416		 * but on the CPU port.
 417		 */
 418		err = dsa_port_mdb_del(dp->cpu_dp, SWITCHDEV_OBJ_PORT_MDB(obj));
 419		break;
 420	case SWITCHDEV_OBJ_ID_PORT_VLAN:
 421		err = dsa_slave_vlan_del(dev, obj);
 422		break;
 423	default:
 424		err = -EOPNOTSUPP;
 425		break;
 426	}
 427
 428	return err;
 429}
 430
 431static int dsa_slave_get_port_parent_id(struct net_device *dev,
 432					struct netdev_phys_item_id *ppid)
 433{
 434	struct dsa_port *dp = dsa_slave_to_port(dev);
 435	struct dsa_switch *ds = dp->ds;
 436	struct dsa_switch_tree *dst = ds->dst;
 437
 438	/* For non-legacy ports, devlink is used and it takes
 439	 * care of the name generation. This ndo implementation
 440	 * should be removed with legacy support.
 441	 */
 442	if (dp->ds->devlink)
 
 
 
 
 443		return -EOPNOTSUPP;
 444
 445	ppid->id_len = sizeof(dst->index);
 446	memcpy(&ppid->id, &dst->index, ppid->id_len);
 447
 448	return 0;
 449}
 450
 451static inline netdev_tx_t dsa_slave_netpoll_send_skb(struct net_device *dev,
 452						     struct sk_buff *skb)
 453{
 454#ifdef CONFIG_NET_POLL_CONTROLLER
 455	struct dsa_slave_priv *p = netdev_priv(dev);
 456
 457	if (p->netpoll)
 458		netpoll_send_skb(p->netpoll, skb);
 459#else
 460	BUG();
 461#endif
 462	return NETDEV_TX_OK;
 463}
 464
 465static void dsa_skb_tx_timestamp(struct dsa_slave_priv *p,
 466				 struct sk_buff *skb)
 467{
 468	struct dsa_switch *ds = p->dp->ds;
 469	struct sk_buff *clone;
 470	unsigned int type;
 471
 472	type = ptp_classify_raw(skb);
 473	if (type == PTP_CLASS_NONE)
 474		return;
 475
 476	if (!ds->ops->port_txtstamp)
 477		return;
 478
 479	clone = skb_clone_sk(skb);
 480	if (!clone)
 481		return;
 482
 483	DSA_SKB_CB(skb)->clone = clone;
 484
 485	if (ds->ops->port_txtstamp(ds, p->dp->index, clone, type))
 486		return;
 487
 488	kfree_skb(clone);
 489}
 490
 491netdev_tx_t dsa_enqueue_skb(struct sk_buff *skb, struct net_device *dev)
 492{
 493	/* SKB for netpoll still need to be mangled with the protocol-specific
 494	 * tag to be successfully transmitted
 495	 */
 496	if (unlikely(netpoll_tx_running(dev)))
 497		return dsa_slave_netpoll_send_skb(dev, skb);
 498
 499	/* Queue the SKB for transmission on the parent interface, but
 500	 * do not modify its EtherType
 501	 */
 502	skb->dev = dsa_slave_to_master(dev);
 503	dev_queue_xmit(skb);
 504
 505	return NETDEV_TX_OK;
 506}
 507EXPORT_SYMBOL_GPL(dsa_enqueue_skb);
 508
 509static netdev_tx_t dsa_slave_xmit(struct sk_buff *skb, struct net_device *dev)
 510{
 511	struct dsa_slave_priv *p = netdev_priv(dev);
 512	struct pcpu_sw_netstats *s;
 513	struct sk_buff *nskb;
 514
 515	s = this_cpu_ptr(p->stats64);
 516	u64_stats_update_begin(&s->syncp);
 517	s->tx_packets++;
 518	s->tx_bytes += skb->len;
 519	u64_stats_update_end(&s->syncp);
 520
 521	DSA_SKB_CB(skb)->deferred_xmit = false;
 522	DSA_SKB_CB(skb)->clone = NULL;
 523
 524	/* Identify PTP protocol packets, clone them, and pass them to the
 525	 * switch driver
 526	 */
 527	dsa_skb_tx_timestamp(p, skb);
 528
 529	/* Transmit function may have to reallocate the original SKB,
 530	 * in which case it must have freed it. Only free it here on error.
 531	 */
 532	nskb = p->xmit(skb, dev);
 533	if (!nskb) {
 534		if (!DSA_SKB_CB(skb)->deferred_xmit)
 535			kfree_skb(skb);
 536		return NETDEV_TX_OK;
 537	}
 538
 539	return dsa_enqueue_skb(nskb, dev);
 540}
 
 
 
 541
 542void *dsa_defer_xmit(struct sk_buff *skb, struct net_device *dev)
 543{
 544	struct dsa_port *dp = dsa_slave_to_port(dev);
 
 
 545
 546	DSA_SKB_CB(skb)->deferred_xmit = true;
 547
 548	skb_queue_tail(&dp->xmit_queue, skb);
 549	schedule_work(&dp->xmit_work);
 550	return NULL;
 551}
 552EXPORT_SYMBOL_GPL(dsa_defer_xmit);
 553
 554static void dsa_port_xmit_work(struct work_struct *work)
 555{
 556	struct dsa_port *dp = container_of(work, struct dsa_port, xmit_work);
 557	struct dsa_switch *ds = dp->ds;
 558	struct sk_buff *skb;
 559
 560	if (unlikely(!ds->ops->port_deferred_xmit))
 561		return;
 562
 563	while ((skb = skb_dequeue(&dp->xmit_queue)) != NULL)
 564		ds->ops->port_deferred_xmit(ds, dp->index, skb);
 565}
 566
 567/* ethtool operations *******************************************************/
 568
 569static void dsa_slave_get_drvinfo(struct net_device *dev,
 570				  struct ethtool_drvinfo *drvinfo)
 571{
 572	strlcpy(drvinfo->driver, "dsa", sizeof(drvinfo->driver));
 573	strlcpy(drvinfo->fw_version, "N/A", sizeof(drvinfo->fw_version));
 574	strlcpy(drvinfo->bus_info, "platform", sizeof(drvinfo->bus_info));
 575}
 576
 577static int dsa_slave_get_regs_len(struct net_device *dev)
 578{
 579	struct dsa_port *dp = dsa_slave_to_port(dev);
 580	struct dsa_switch *ds = dp->ds;
 581
 582	if (ds->ops->get_regs_len)
 583		return ds->ops->get_regs_len(ds, dp->index);
 584
 585	return -EOPNOTSUPP;
 586}
 587
 588static void
 589dsa_slave_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *_p)
 590{
 591	struct dsa_port *dp = dsa_slave_to_port(dev);
 592	struct dsa_switch *ds = dp->ds;
 593
 594	if (ds->ops->get_regs)
 595		ds->ops->get_regs(ds, dp->index, regs, _p);
 596}
 597
 598static int dsa_slave_nway_reset(struct net_device *dev)
 599{
 600	struct dsa_port *dp = dsa_slave_to_port(dev);
 
 
 
 601
 602	return phylink_ethtool_nway_reset(dp->pl);
 603}
 604
 605static int dsa_slave_get_eeprom_len(struct net_device *dev)
 606{
 607	struct dsa_port *dp = dsa_slave_to_port(dev);
 608	struct dsa_switch *ds = dp->ds;
 609
 610	if (ds->cd && ds->cd->eeprom_len)
 611		return ds->cd->eeprom_len;
 612
 613	if (ds->ops->get_eeprom_len)
 614		return ds->ops->get_eeprom_len(ds);
 615
 616	return 0;
 617}
 618
 619static int dsa_slave_get_eeprom(struct net_device *dev,
 620				struct ethtool_eeprom *eeprom, u8 *data)
 621{
 622	struct dsa_port *dp = dsa_slave_to_port(dev);
 623	struct dsa_switch *ds = dp->ds;
 624
 625	if (ds->ops->get_eeprom)
 626		return ds->ops->get_eeprom(ds, eeprom, data);
 627
 628	return -EOPNOTSUPP;
 629}
 630
 631static int dsa_slave_set_eeprom(struct net_device *dev,
 632				struct ethtool_eeprom *eeprom, u8 *data)
 633{
 634	struct dsa_port *dp = dsa_slave_to_port(dev);
 635	struct dsa_switch *ds = dp->ds;
 636
 637	if (ds->ops->set_eeprom)
 638		return ds->ops->set_eeprom(ds, eeprom, data);
 639
 640	return -EOPNOTSUPP;
 641}
 642
 643static void dsa_slave_get_strings(struct net_device *dev,
 644				  uint32_t stringset, uint8_t *data)
 645{
 646	struct dsa_port *dp = dsa_slave_to_port(dev);
 647	struct dsa_switch *ds = dp->ds;
 648
 649	if (stringset == ETH_SS_STATS) {
 650		int len = ETH_GSTRING_LEN;
 651
 652		strncpy(data, "tx_packets", len);
 653		strncpy(data + len, "tx_bytes", len);
 654		strncpy(data + 2 * len, "rx_packets", len);
 655		strncpy(data + 3 * len, "rx_bytes", len);
 656		if (ds->ops->get_strings)
 657			ds->ops->get_strings(ds, dp->index, stringset,
 658					     data + 4 * len);
 659	}
 660}
 661
 662static void dsa_slave_get_ethtool_stats(struct net_device *dev,
 663					struct ethtool_stats *stats,
 664					uint64_t *data)
 665{
 666	struct dsa_port *dp = dsa_slave_to_port(dev);
 667	struct dsa_slave_priv *p = netdev_priv(dev);
 668	struct dsa_switch *ds = dp->ds;
 669	struct pcpu_sw_netstats *s;
 670	unsigned int start;
 671	int i;
 672
 673	for_each_possible_cpu(i) {
 674		u64 tx_packets, tx_bytes, rx_packets, rx_bytes;
 675
 676		s = per_cpu_ptr(p->stats64, i);
 677		do {
 678			start = u64_stats_fetch_begin_irq(&s->syncp);
 679			tx_packets = s->tx_packets;
 680			tx_bytes = s->tx_bytes;
 681			rx_packets = s->rx_packets;
 682			rx_bytes = s->rx_bytes;
 683		} while (u64_stats_fetch_retry_irq(&s->syncp, start));
 684		data[0] += tx_packets;
 685		data[1] += tx_bytes;
 686		data[2] += rx_packets;
 687		data[3] += rx_bytes;
 688	}
 689	if (ds->ops->get_ethtool_stats)
 690		ds->ops->get_ethtool_stats(ds, dp->index, data + 4);
 691}
 692
 693static int dsa_slave_get_sset_count(struct net_device *dev, int sset)
 694{
 695	struct dsa_port *dp = dsa_slave_to_port(dev);
 696	struct dsa_switch *ds = dp->ds;
 697
 698	if (sset == ETH_SS_STATS) {
 699		int count;
 700
 701		count = 4;
 702		if (ds->ops->get_sset_count)
 703			count += ds->ops->get_sset_count(ds, dp->index, sset);
 704
 705		return count;
 706	}
 707
 708	return -EOPNOTSUPP;
 709}
 710
 711static void dsa_slave_get_wol(struct net_device *dev, struct ethtool_wolinfo *w)
 712{
 713	struct dsa_port *dp = dsa_slave_to_port(dev);
 714	struct dsa_switch *ds = dp->ds;
 715
 716	phylink_ethtool_get_wol(dp->pl, w);
 717
 718	if (ds->ops->get_wol)
 719		ds->ops->get_wol(ds, dp->index, w);
 720}
 721
 722static int dsa_slave_set_wol(struct net_device *dev, struct ethtool_wolinfo *w)
 723{
 724	struct dsa_port *dp = dsa_slave_to_port(dev);
 725	struct dsa_switch *ds = dp->ds;
 726	int ret = -EOPNOTSUPP;
 727
 728	phylink_ethtool_set_wol(dp->pl, w);
 729
 730	if (ds->ops->set_wol)
 731		ret = ds->ops->set_wol(ds, dp->index, w);
 732
 733	return ret;
 734}
 735
 736static int dsa_slave_set_eee(struct net_device *dev, struct ethtool_eee *e)
 737{
 738	struct dsa_port *dp = dsa_slave_to_port(dev);
 739	struct dsa_switch *ds = dp->ds;
 740	int ret;
 741
 742	/* Port's PHY and MAC both need to be EEE capable */
 743	if (!dev->phydev || !dp->pl)
 744		return -ENODEV;
 745
 746	if (!ds->ops->set_mac_eee)
 747		return -EOPNOTSUPP;
 748
 749	ret = ds->ops->set_mac_eee(ds, dp->index, e);
 750	if (ret)
 751		return ret;
 752
 753	return phylink_ethtool_set_eee(dp->pl, e);
 
 
 
 
 
 
 754}
 755
 756static int dsa_slave_get_eee(struct net_device *dev, struct ethtool_eee *e)
 757{
 758	struct dsa_port *dp = dsa_slave_to_port(dev);
 759	struct dsa_switch *ds = dp->ds;
 760	int ret;
 761
 762	/* Port's PHY and MAC both need to be EEE capable */
 763	if (!dev->phydev || !dp->pl)
 764		return -ENODEV;
 765
 766	if (!ds->ops->get_mac_eee)
 767		return -EOPNOTSUPP;
 768
 769	ret = ds->ops->get_mac_eee(ds, dp->index, e);
 770	if (ret)
 771		return ret;
 772
 773	return phylink_ethtool_get_eee(dp->pl, e);
 774}
 775
 776static int dsa_slave_get_link_ksettings(struct net_device *dev,
 777					struct ethtool_link_ksettings *cmd)
 778{
 779	struct dsa_port *dp = dsa_slave_to_port(dev);
 780
 781	return phylink_ethtool_ksettings_get(dp->pl, cmd);
 782}
 783
 784static int dsa_slave_set_link_ksettings(struct net_device *dev,
 785					const struct ethtool_link_ksettings *cmd)
 786{
 787	struct dsa_port *dp = dsa_slave_to_port(dev);
 788
 789	return phylink_ethtool_ksettings_set(dp->pl, cmd);
 790}
 791
 792#ifdef CONFIG_NET_POLL_CONTROLLER
 793static int dsa_slave_netpoll_setup(struct net_device *dev,
 794				   struct netpoll_info *ni)
 795{
 796	struct net_device *master = dsa_slave_to_master(dev);
 797	struct dsa_slave_priv *p = netdev_priv(dev);
 798	struct netpoll *netpoll;
 799	int err = 0;
 800
 801	netpoll = kzalloc(sizeof(*netpoll), GFP_KERNEL);
 802	if (!netpoll)
 803		return -ENOMEM;
 804
 805	err = __netpoll_setup(netpoll, master);
 806	if (err) {
 807		kfree(netpoll);
 808		goto out;
 809	}
 810
 811	p->netpoll = netpoll;
 812out:
 813	return err;
 814}
 815
 816static void dsa_slave_netpoll_cleanup(struct net_device *dev)
 817{
 818	struct dsa_slave_priv *p = netdev_priv(dev);
 819	struct netpoll *netpoll = p->netpoll;
 820
 821	if (!netpoll)
 822		return;
 823
 824	p->netpoll = NULL;
 825
 826	__netpoll_free(netpoll);
 827}
 828
 829static void dsa_slave_poll_controller(struct net_device *dev)
 830{
 831}
 832#endif
 833
 834static int dsa_slave_get_phys_port_name(struct net_device *dev,
 835					char *name, size_t len)
 836{
 837	struct dsa_port *dp = dsa_slave_to_port(dev);
 838
 839	/* For non-legacy ports, devlink is used and it takes
 840	 * care of the name generation. This ndo implementation
 841	 * should be removed with legacy support.
 842	 */
 843	if (dp->ds->devlink)
 844		return -EOPNOTSUPP;
 845
 846	if (snprintf(name, len, "p%d", dp->index) >= len)
 847		return -EINVAL;
 848
 849	return 0;
 850}
 851
 852static struct dsa_mall_tc_entry *
 853dsa_slave_mall_tc_entry_find(struct net_device *dev, unsigned long cookie)
 854{
 855	struct dsa_slave_priv *p = netdev_priv(dev);
 856	struct dsa_mall_tc_entry *mall_tc_entry;
 857
 858	list_for_each_entry(mall_tc_entry, &p->mall_tc_list, list)
 859		if (mall_tc_entry->cookie == cookie)
 860			return mall_tc_entry;
 861
 862	return NULL;
 863}
 864
 865static int dsa_slave_add_cls_matchall(struct net_device *dev,
 866				      struct tc_cls_matchall_offload *cls,
 867				      bool ingress)
 868{
 869	struct dsa_port *dp = dsa_slave_to_port(dev);
 870	struct dsa_slave_priv *p = netdev_priv(dev);
 871	struct dsa_mall_tc_entry *mall_tc_entry;
 872	__be16 protocol = cls->common.protocol;
 873	struct dsa_switch *ds = dp->ds;
 874	struct flow_action_entry *act;
 
 875	struct dsa_port *to_dp;
 876	int err = -EOPNOTSUPP;
 
 877
 878	if (!ds->ops->port_mirror_add)
 879		return err;
 880
 881	if (!flow_offload_has_one_action(&cls->rule->action))
 882		return err;
 883
 884	act = &cls->rule->action.entries[0];
 
 885
 886	if (act->id == FLOW_ACTION_MIRRED && protocol == htons(ETH_P_ALL)) {
 887		struct dsa_mall_mirror_tc_entry *mirror;
 888
 889		if (!act->dev)
 
 890			return -EINVAL;
 891
 892		if (!dsa_slave_dev_check(act->dev))
 893			return -EOPNOTSUPP;
 894
 895		mall_tc_entry = kzalloc(sizeof(*mall_tc_entry), GFP_KERNEL);
 896		if (!mall_tc_entry)
 897			return -ENOMEM;
 898
 899		mall_tc_entry->cookie = cls->cookie;
 900		mall_tc_entry->type = DSA_PORT_MALL_MIRROR;
 901		mirror = &mall_tc_entry->mirror;
 902
 903		to_dp = dsa_slave_to_port(act->dev);
 904
 905		mirror->to_local_port = to_dp->index;
 906		mirror->ingress = ingress;
 907
 908		err = ds->ops->port_mirror_add(ds, dp->index, mirror, ingress);
 909		if (err) {
 910			kfree(mall_tc_entry);
 911			return err;
 912		}
 913
 914		list_add_tail(&mall_tc_entry->list, &p->mall_tc_list);
 915	}
 916
 917	return 0;
 918}
 919
 920static void dsa_slave_del_cls_matchall(struct net_device *dev,
 921				       struct tc_cls_matchall_offload *cls)
 922{
 923	struct dsa_port *dp = dsa_slave_to_port(dev);
 924	struct dsa_mall_tc_entry *mall_tc_entry;
 925	struct dsa_switch *ds = dp->ds;
 926
 927	if (!ds->ops->port_mirror_del)
 928		return;
 929
 930	mall_tc_entry = dsa_slave_mall_tc_entry_find(dev, cls->cookie);
 931	if (!mall_tc_entry)
 932		return;
 933
 934	list_del(&mall_tc_entry->list);
 935
 936	switch (mall_tc_entry->type) {
 937	case DSA_PORT_MALL_MIRROR:
 938		ds->ops->port_mirror_del(ds, dp->index, &mall_tc_entry->mirror);
 939		break;
 940	default:
 941		WARN_ON(1);
 942	}
 943
 944	kfree(mall_tc_entry);
 945}
 946
 947static int dsa_slave_setup_tc_cls_matchall(struct net_device *dev,
 948					   struct tc_cls_matchall_offload *cls,
 949					   bool ingress)
 950{
 951	if (cls->common.chain_index)
 952		return -EOPNOTSUPP;
 953
 954	switch (cls->command) {
 955	case TC_CLSMATCHALL_REPLACE:
 956		return dsa_slave_add_cls_matchall(dev, cls, ingress);
 957	case TC_CLSMATCHALL_DESTROY:
 958		dsa_slave_del_cls_matchall(dev, cls);
 959		return 0;
 960	default:
 961		return -EOPNOTSUPP;
 962	}
 963}
 964
 965static int dsa_slave_setup_tc_block_cb(enum tc_setup_type type, void *type_data,
 966				       void *cb_priv, bool ingress)
 967{
 968	struct net_device *dev = cb_priv;
 969
 970	if (!tc_can_offload(dev))
 971		return -EOPNOTSUPP;
 972
 973	switch (type) {
 974	case TC_SETUP_CLSMATCHALL:
 975		return dsa_slave_setup_tc_cls_matchall(dev, type_data, ingress);
 976	default:
 977		return -EOPNOTSUPP;
 978	}
 979}
 980
 981static int dsa_slave_setup_tc_block_cb_ig(enum tc_setup_type type,
 982					  void *type_data, void *cb_priv)
 983{
 984	return dsa_slave_setup_tc_block_cb(type, type_data, cb_priv, true);
 985}
 986
 987static int dsa_slave_setup_tc_block_cb_eg(enum tc_setup_type type,
 988					  void *type_data, void *cb_priv)
 989{
 990	return dsa_slave_setup_tc_block_cb(type, type_data, cb_priv, false);
 991}
 992
 993static LIST_HEAD(dsa_slave_block_cb_list);
 994
 995static int dsa_slave_setup_tc_block(struct net_device *dev,
 996				    struct flow_block_offload *f)
 997{
 998	struct flow_block_cb *block_cb;
 999	flow_setup_cb_t *cb;
1000
1001	if (f->binder_type == FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS)
1002		cb = dsa_slave_setup_tc_block_cb_ig;
1003	else if (f->binder_type == FLOW_BLOCK_BINDER_TYPE_CLSACT_EGRESS)
1004		cb = dsa_slave_setup_tc_block_cb_eg;
1005	else
1006		return -EOPNOTSUPP;
1007
1008	f->driver_block_list = &dsa_slave_block_cb_list;
1009
1010	switch (f->command) {
1011	case FLOW_BLOCK_BIND:
1012		if (flow_block_cb_is_busy(cb, dev, &dsa_slave_block_cb_list))
1013			return -EBUSY;
1014
1015		block_cb = flow_block_cb_alloc(cb, dev, dev, NULL);
1016		if (IS_ERR(block_cb))
1017			return PTR_ERR(block_cb);
1018
1019		flow_block_cb_add(block_cb, f);
1020		list_add_tail(&block_cb->driver_list, &dsa_slave_block_cb_list);
1021		return 0;
1022	case FLOW_BLOCK_UNBIND:
1023		block_cb = flow_block_cb_lookup(f->block, cb, dev);
1024		if (!block_cb)
1025			return -ENOENT;
1026
1027		flow_block_cb_remove(block_cb, f);
1028		list_del(&block_cb->driver_list);
1029		return 0;
1030	default:
1031		return -EOPNOTSUPP;
1032	}
1033}
1034
1035static int dsa_slave_setup_tc(struct net_device *dev, enum tc_setup_type type,
1036			      void *type_data)
1037{
1038	struct dsa_port *dp = dsa_slave_to_port(dev);
1039	struct dsa_switch *ds = dp->ds;
1040
1041	if (type == TC_SETUP_BLOCK)
1042		return dsa_slave_setup_tc_block(dev, type_data);
1043
1044	if (!ds->ops->port_setup_tc)
1045		return -EOPNOTSUPP;
1046
1047	return ds->ops->port_setup_tc(ds, dp->index, type, type_data);
1048}
1049
1050static void dsa_slave_get_stats64(struct net_device *dev,
1051				  struct rtnl_link_stats64 *stats)
1052{
1053	struct dsa_slave_priv *p = netdev_priv(dev);
1054	struct pcpu_sw_netstats *s;
1055	unsigned int start;
1056	int i;
1057
1058	netdev_stats_to_stats64(stats, &dev->stats);
1059	for_each_possible_cpu(i) {
1060		u64 tx_packets, tx_bytes, rx_packets, rx_bytes;
1061
1062		s = per_cpu_ptr(p->stats64, i);
1063		do {
1064			start = u64_stats_fetch_begin_irq(&s->syncp);
1065			tx_packets = s->tx_packets;
1066			tx_bytes = s->tx_bytes;
1067			rx_packets = s->rx_packets;
1068			rx_bytes = s->rx_bytes;
1069		} while (u64_stats_fetch_retry_irq(&s->syncp, start));
1070
1071		stats->tx_packets += tx_packets;
1072		stats->tx_bytes += tx_bytes;
1073		stats->rx_packets += rx_packets;
1074		stats->rx_bytes += rx_bytes;
1075	}
1076}
1077
1078static int dsa_slave_get_rxnfc(struct net_device *dev,
1079			       struct ethtool_rxnfc *nfc, u32 *rule_locs)
1080{
1081	struct dsa_port *dp = dsa_slave_to_port(dev);
1082	struct dsa_switch *ds = dp->ds;
1083
1084	if (!ds->ops->get_rxnfc)
1085		return -EOPNOTSUPP;
1086
1087	return ds->ops->get_rxnfc(ds, dp->index, nfc, rule_locs);
1088}
1089
1090static int dsa_slave_set_rxnfc(struct net_device *dev,
1091			       struct ethtool_rxnfc *nfc)
1092{
1093	struct dsa_port *dp = dsa_slave_to_port(dev);
1094	struct dsa_switch *ds = dp->ds;
1095
1096	if (!ds->ops->set_rxnfc)
1097		return -EOPNOTSUPP;
1098
1099	return ds->ops->set_rxnfc(ds, dp->index, nfc);
1100}
1101
1102static int dsa_slave_get_ts_info(struct net_device *dev,
1103				 struct ethtool_ts_info *ts)
1104{
1105	struct dsa_slave_priv *p = netdev_priv(dev);
1106	struct dsa_switch *ds = p->dp->ds;
1107
1108	if (!ds->ops->get_ts_info)
1109		return -EOPNOTSUPP;
1110
1111	return ds->ops->get_ts_info(ds, p->dp->index, ts);
1112}
1113
1114static int dsa_slave_vlan_rx_add_vid(struct net_device *dev, __be16 proto,
1115				     u16 vid)
1116{
1117	struct dsa_port *dp = dsa_slave_to_port(dev);
1118	struct bridge_vlan_info info;
1119	int ret;
1120
1121	/* Check for a possible bridge VLAN entry now since there is no
1122	 * need to emulate the switchdev prepare + commit phase.
1123	 */
1124	if (dp->bridge_dev) {
1125		if (!br_vlan_enabled(dp->bridge_dev))
1126			return 0;
1127
1128		/* br_vlan_get_info() returns -EINVAL or -ENOENT if the
1129		 * device, respectively the VID is not found, returning
1130		 * 0 means success, which is a failure for us here.
1131		 */
1132		ret = br_vlan_get_info(dp->bridge_dev, vid, &info);
1133		if (ret == 0)
1134			return -EBUSY;
1135	}
1136
1137	ret = dsa_port_vid_add(dp, vid, 0);
1138	if (ret)
1139		return ret;
1140
1141	ret = dsa_port_vid_add(dp->cpu_dp, vid, 0);
1142	if (ret)
1143		return ret;
1144
1145	return 0;
1146}
1147
1148static int dsa_slave_vlan_rx_kill_vid(struct net_device *dev, __be16 proto,
1149				      u16 vid)
1150{
1151	struct dsa_port *dp = dsa_slave_to_port(dev);
1152	struct bridge_vlan_info info;
1153	int ret;
1154
1155	/* Check for a possible bridge VLAN entry now since there is no
1156	 * need to emulate the switchdev prepare + commit phase.
1157	 */
1158	if (dp->bridge_dev) {
1159		if (!br_vlan_enabled(dp->bridge_dev))
1160			return 0;
1161
1162		/* br_vlan_get_info() returns -EINVAL or -ENOENT if the
1163		 * device, respectively the VID is not found, returning
1164		 * 0 means success, which is a failure for us here.
1165		 */
1166		ret = br_vlan_get_info(dp->bridge_dev, vid, &info);
1167		if (ret == 0)
1168			return -EBUSY;
1169	}
1170
1171	/* Do not deprogram the CPU port as it may be shared with other user
1172	 * ports which can be members of this VLAN as well.
1173	 */
1174	return dsa_port_vid_del(dp, vid);
1175}
1176
1177static const struct ethtool_ops dsa_slave_ethtool_ops = {
1178	.get_drvinfo		= dsa_slave_get_drvinfo,
1179	.get_regs_len		= dsa_slave_get_regs_len,
1180	.get_regs		= dsa_slave_get_regs,
1181	.nway_reset		= dsa_slave_nway_reset,
1182	.get_link		= ethtool_op_get_link,
1183	.get_eeprom_len		= dsa_slave_get_eeprom_len,
1184	.get_eeprom		= dsa_slave_get_eeprom,
1185	.set_eeprom		= dsa_slave_set_eeprom,
1186	.get_strings		= dsa_slave_get_strings,
1187	.get_ethtool_stats	= dsa_slave_get_ethtool_stats,
1188	.get_sset_count		= dsa_slave_get_sset_count,
1189	.set_wol		= dsa_slave_set_wol,
1190	.get_wol		= dsa_slave_get_wol,
1191	.set_eee		= dsa_slave_set_eee,
1192	.get_eee		= dsa_slave_get_eee,
1193	.get_link_ksettings	= dsa_slave_get_link_ksettings,
1194	.set_link_ksettings	= dsa_slave_set_link_ksettings,
1195	.get_rxnfc		= dsa_slave_get_rxnfc,
1196	.set_rxnfc		= dsa_slave_set_rxnfc,
1197	.get_ts_info		= dsa_slave_get_ts_info,
1198};
1199
1200/* legacy way, bypassing the bridge *****************************************/
1201int dsa_legacy_fdb_add(struct ndmsg *ndm, struct nlattr *tb[],
1202		       struct net_device *dev,
1203		       const unsigned char *addr, u16 vid,
1204		       u16 flags,
1205		       struct netlink_ext_ack *extack)
1206{
1207	struct dsa_port *dp = dsa_slave_to_port(dev);
1208
1209	return dsa_port_fdb_add(dp, addr, vid);
1210}
1211
1212int dsa_legacy_fdb_del(struct ndmsg *ndm, struct nlattr *tb[],
1213		       struct net_device *dev,
1214		       const unsigned char *addr, u16 vid)
1215{
1216	struct dsa_port *dp = dsa_slave_to_port(dev);
1217
1218	return dsa_port_fdb_del(dp, addr, vid);
1219}
1220
1221static struct devlink_port *dsa_slave_get_devlink_port(struct net_device *dev)
1222{
1223	struct dsa_port *dp = dsa_slave_to_port(dev);
1224
1225	return dp->ds->devlink ? &dp->devlink_port : NULL;
1226}
1227
1228static const struct net_device_ops dsa_slave_netdev_ops = {
1229	.ndo_open	 	= dsa_slave_open,
1230	.ndo_stop		= dsa_slave_close,
1231	.ndo_start_xmit		= dsa_slave_xmit,
1232	.ndo_change_rx_flags	= dsa_slave_change_rx_flags,
1233	.ndo_set_rx_mode	= dsa_slave_set_rx_mode,
1234	.ndo_set_mac_address	= dsa_slave_set_mac_address,
1235	.ndo_fdb_add		= dsa_legacy_fdb_add,
1236	.ndo_fdb_del		= dsa_legacy_fdb_del,
1237	.ndo_fdb_dump		= dsa_slave_fdb_dump,
1238	.ndo_do_ioctl		= dsa_slave_ioctl,
1239	.ndo_get_iflink		= dsa_slave_get_iflink,
1240#ifdef CONFIG_NET_POLL_CONTROLLER
1241	.ndo_netpoll_setup	= dsa_slave_netpoll_setup,
1242	.ndo_netpoll_cleanup	= dsa_slave_netpoll_cleanup,
1243	.ndo_poll_controller	= dsa_slave_poll_controller,
1244#endif
1245	.ndo_get_phys_port_name	= dsa_slave_get_phys_port_name,
1246	.ndo_setup_tc		= dsa_slave_setup_tc,
1247	.ndo_get_stats64	= dsa_slave_get_stats64,
1248	.ndo_get_port_parent_id	= dsa_slave_get_port_parent_id,
1249	.ndo_vlan_rx_add_vid	= dsa_slave_vlan_rx_add_vid,
1250	.ndo_vlan_rx_kill_vid	= dsa_slave_vlan_rx_kill_vid,
1251	.ndo_get_devlink_port	= dsa_slave_get_devlink_port,
 
 
 
1252};
1253
1254static struct device_type dsa_type = {
1255	.name	= "dsa",
1256};
1257
1258void dsa_port_phylink_mac_change(struct dsa_switch *ds, int port, bool up)
1259{
1260	const struct dsa_port *dp = dsa_to_port(ds, port);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1261
1262	phylink_mac_change(dp->pl, up);
 
 
 
 
1263}
1264EXPORT_SYMBOL_GPL(dsa_port_phylink_mac_change);
1265
1266static void dsa_slave_phylink_fixed_state(struct net_device *dev,
1267					  struct phylink_link_state *state)
1268{
1269	struct dsa_port *dp = dsa_slave_to_port(dev);
1270	struct dsa_switch *ds = dp->ds;
 
 
 
 
 
 
 
1271
1272	/* No need to check that this operation is valid, the callback would
1273	 * not be called if it was not.
1274	 */
1275	ds->ops->phylink_fixed_state(ds, dp->index, state);
1276}
1277
1278/* slave device setup *******************************************************/
1279static int dsa_slave_phy_connect(struct net_device *slave_dev, int addr)
1280{
1281	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
 
1282	struct dsa_switch *ds = dp->ds;
1283
1284	slave_dev->phydev = mdiobus_get_phy(ds->slave_mii_bus, addr);
1285	if (!slave_dev->phydev) {
1286		netdev_err(slave_dev, "no phy at %d\n", addr);
1287		return -ENODEV;
1288	}
1289
1290	return phylink_connect_phy(dp->pl, slave_dev->phydev);
 
 
 
 
 
1291}
1292
1293static int dsa_slave_phy_setup(struct net_device *slave_dev)
1294{
1295	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
 
1296	struct device_node *port_dn = dp->dn;
1297	struct dsa_switch *ds = dp->ds;
 
 
1298	u32 phy_flags = 0;
1299	int mode, ret;
1300
1301	mode = of_get_phy_mode(port_dn);
1302	if (mode < 0)
1303		mode = PHY_INTERFACE_MODE_NA;
 
1304
1305	dp->pl_config.dev = &slave_dev->dev;
1306	dp->pl_config.type = PHYLINK_NETDEV;
1307
1308	dp->pl = phylink_create(&dp->pl_config, of_fwnode_handle(port_dn), mode,
1309				&dsa_port_phylink_mac_ops);
1310	if (IS_ERR(dp->pl)) {
1311		netdev_err(slave_dev,
1312			   "error creating PHYLINK: %ld\n", PTR_ERR(dp->pl));
1313		return PTR_ERR(dp->pl);
 
 
 
1314	}
1315
1316	/* Register only if the switch provides such a callback, since this
1317	 * callback takes precedence over polling the link GPIO in PHYLINK
1318	 * (see phylink_get_fixed_state).
1319	 */
1320	if (ds->ops->phylink_fixed_state)
1321		phylink_fixed_state_cb(dp->pl, dsa_slave_phylink_fixed_state);
1322
1323	if (ds->ops->get_phy_flags)
1324		phy_flags = ds->ops->get_phy_flags(ds, dp->index);
1325
1326	ret = phylink_of_phy_connect(dp->pl, port_dn, phy_flags);
1327	if (ret == -ENODEV && ds->slave_mii_bus) {
1328		/* We could not connect to a designated PHY or SFP, so try to
1329		 * use the switch internal MDIO bus instead
1330		 */
 
 
 
 
 
 
 
 
 
 
 
1331		ret = dsa_slave_phy_connect(slave_dev, dp->index);
1332		if (ret) {
1333			netdev_err(slave_dev,
1334				   "failed to connect to port %d: %d\n",
1335				   dp->index, ret);
1336			phylink_destroy(dp->pl);
 
1337			return ret;
1338		}
1339	}
1340
1341	return ret;
 
 
 
 
 
 
 
 
 
 
 
1342}
1343
1344int dsa_slave_suspend(struct net_device *slave_dev)
1345{
1346	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1347
1348	if (!netif_running(slave_dev))
1349		return 0;
1350
1351	cancel_work_sync(&dp->xmit_work);
1352	skb_queue_purge(&dp->xmit_queue);
1353
1354	netif_device_detach(slave_dev);
1355
1356	rtnl_lock();
1357	phylink_stop(dp->pl);
1358	rtnl_unlock();
 
 
 
 
1359
1360	return 0;
1361}
1362
1363int dsa_slave_resume(struct net_device *slave_dev)
1364{
1365	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1366
1367	if (!netif_running(slave_dev))
1368		return 0;
1369
1370	netif_device_attach(slave_dev);
1371
1372	rtnl_lock();
1373	phylink_start(dp->pl);
1374	rtnl_unlock();
 
1375
1376	return 0;
1377}
1378
1379static void dsa_slave_notify(struct net_device *dev, unsigned long val)
1380{
1381	struct net_device *master = dsa_slave_to_master(dev);
1382	struct dsa_port *dp = dsa_slave_to_port(dev);
1383	struct dsa_notifier_register_info rinfo = {
1384		.switch_number = dp->ds->index,
1385		.port_number = dp->index,
1386		.master = master,
1387		.info.dev = dev,
1388	};
1389
1390	call_dsa_notifiers(val, dev, &rinfo.info);
1391}
1392
1393int dsa_slave_create(struct dsa_port *port)
1394{
1395	const struct dsa_port *cpu_dp = port->cpu_dp;
1396	struct net_device *master = cpu_dp->master;
1397	struct dsa_switch *ds = port->ds;
1398	const char *name = port->name;
1399	struct net_device *slave_dev;
1400	struct dsa_slave_priv *p;
1401	int ret;
1402
1403	if (!ds->num_tx_queues)
1404		ds->num_tx_queues = 1;
1405
1406	slave_dev = alloc_netdev_mqs(sizeof(struct dsa_slave_priv), name,
1407				     NET_NAME_UNKNOWN, ether_setup,
1408				     ds->num_tx_queues, 1);
1409	if (slave_dev == NULL)
1410		return -ENOMEM;
1411
1412	slave_dev->features = master->vlan_features | NETIF_F_HW_TC;
1413	if (ds->ops->port_vlan_add && ds->ops->port_vlan_del)
1414		slave_dev->features |= NETIF_F_HW_VLAN_CTAG_FILTER;
1415	slave_dev->hw_features |= NETIF_F_HW_TC;
1416	slave_dev->ethtool_ops = &dsa_slave_ethtool_ops;
1417	if (!IS_ERR_OR_NULL(port->mac))
1418		ether_addr_copy(slave_dev->dev_addr, port->mac);
1419	else
1420		eth_hw_addr_inherit(slave_dev, master);
1421	slave_dev->priv_flags |= IFF_NO_QUEUE;
1422	slave_dev->netdev_ops = &dsa_slave_netdev_ops;
 
1423	slave_dev->min_mtu = 0;
1424	slave_dev->max_mtu = ETH_MAX_MTU;
1425	SET_NETDEV_DEVTYPE(slave_dev, &dsa_type);
1426
 
 
 
1427	SET_NETDEV_DEV(slave_dev, port->ds->dev);
1428	slave_dev->dev.of_node = port->dn;
1429	slave_dev->vlan_features = master->vlan_features;
1430
1431	p = netdev_priv(slave_dev);
1432	p->stats64 = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
1433	if (!p->stats64) {
1434		free_netdev(slave_dev);
1435		return -ENOMEM;
1436	}
1437	p->dp = port;
1438	INIT_LIST_HEAD(&p->mall_tc_list);
1439	INIT_WORK(&port->xmit_work, dsa_port_xmit_work);
1440	skb_queue_head_init(&port->xmit_queue);
1441	p->xmit = cpu_dp->tag_ops->xmit;
 
 
 
 
 
1442	port->slave = slave_dev;
1443
1444	netif_carrier_off(slave_dev);
1445
1446	ret = dsa_slave_phy_setup(slave_dev);
1447	if (ret) {
1448		netdev_err(master, "error %d setting up slave phy\n", ret);
1449		goto out_free;
1450	}
1451
1452	dsa_slave_notify(slave_dev, DSA_PORT_REGISTER);
1453
1454	ret = register_netdev(slave_dev);
1455	if (ret) {
1456		netdev_err(master, "error %d registering interface %s\n",
1457			   ret, slave_dev->name);
1458		goto out_phy;
1459	}
1460
1461	return 0;
1462
1463out_phy:
1464	rtnl_lock();
1465	phylink_disconnect_phy(p->dp->pl);
1466	rtnl_unlock();
1467	phylink_destroy(p->dp->pl);
1468out_free:
1469	free_percpu(p->stats64);
1470	free_netdev(slave_dev);
1471	port->slave = NULL;
1472	return ret;
1473}
1474
1475void dsa_slave_destroy(struct net_device *slave_dev)
1476{
1477	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1478	struct dsa_slave_priv *p = netdev_priv(slave_dev);
 
1479
1480	netif_carrier_off(slave_dev);
1481	rtnl_lock();
1482	phylink_disconnect_phy(dp->pl);
1483	rtnl_unlock();
1484
 
 
 
1485	dsa_slave_notify(slave_dev, DSA_PORT_UNREGISTER);
1486	unregister_netdev(slave_dev);
1487	phylink_destroy(dp->pl);
1488	free_percpu(p->stats64);
1489	free_netdev(slave_dev);
1490}
1491
1492static bool dsa_slave_dev_check(const struct net_device *dev)
1493{
1494	return dev->netdev_ops == &dsa_slave_netdev_ops;
1495}
1496
1497static int dsa_slave_changeupper(struct net_device *dev,
1498				 struct netdev_notifier_changeupper_info *info)
1499{
1500	struct dsa_port *dp = dsa_slave_to_port(dev);
1501	int err = NOTIFY_DONE;
1502
1503	if (netif_is_bridge_master(info->upper_dev)) {
1504		if (info->linking) {
1505			err = dsa_port_bridge_join(dp, info->upper_dev);
1506			err = notifier_from_errno(err);
1507		} else {
1508			dsa_port_bridge_leave(dp, info->upper_dev);
1509			err = NOTIFY_OK;
1510		}
1511	}
1512
1513	return err;
1514}
1515
1516static int dsa_slave_upper_vlan_check(struct net_device *dev,
1517				      struct netdev_notifier_changeupper_info *
1518				      info)
1519{
1520	struct netlink_ext_ack *ext_ack;
1521	struct net_device *slave;
1522	struct dsa_port *dp;
1523
1524	ext_ack = netdev_notifier_info_to_extack(&info->info);
1525
1526	if (!is_vlan_dev(dev))
1527		return NOTIFY_DONE;
1528
1529	slave = vlan_dev_real_dev(dev);
1530	if (!dsa_slave_dev_check(slave))
1531		return NOTIFY_DONE;
1532
1533	dp = dsa_slave_to_port(slave);
1534	if (!dp->bridge_dev)
1535		return NOTIFY_DONE;
1536
1537	/* Deny enslaving a VLAN device into a VLAN-aware bridge */
1538	if (br_vlan_enabled(dp->bridge_dev) &&
1539	    netif_is_bridge_master(info->upper_dev) && info->linking) {
1540		NL_SET_ERR_MSG_MOD(ext_ack,
1541				   "Cannot enslave VLAN device into VLAN aware bridge");
1542		return notifier_from_errno(-EINVAL);
1543	}
1544
1545	return NOTIFY_DONE;
1546}
1547
1548static int dsa_slave_netdevice_event(struct notifier_block *nb,
1549				     unsigned long event, void *ptr)
1550{
1551	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1552
1553	if (event == NETDEV_CHANGEUPPER) {
1554		if (!dsa_slave_dev_check(dev))
1555			return dsa_slave_upper_vlan_check(dev, ptr);
1556
 
1557		return dsa_slave_changeupper(dev, ptr);
1558	}
1559
1560	return NOTIFY_DONE;
1561}
1562
1563struct dsa_switchdev_event_work {
1564	struct work_struct work;
1565	struct switchdev_notifier_fdb_info fdb_info;
1566	struct net_device *dev;
1567	unsigned long event;
1568};
1569
1570static void dsa_slave_switchdev_event_work(struct work_struct *work)
1571{
1572	struct dsa_switchdev_event_work *switchdev_work =
1573		container_of(work, struct dsa_switchdev_event_work, work);
1574	struct net_device *dev = switchdev_work->dev;
1575	struct switchdev_notifier_fdb_info *fdb_info;
1576	struct dsa_port *dp = dsa_slave_to_port(dev);
1577	int err;
1578
1579	rtnl_lock();
1580	switch (switchdev_work->event) {
1581	case SWITCHDEV_FDB_ADD_TO_DEVICE:
1582		fdb_info = &switchdev_work->fdb_info;
1583		if (!fdb_info->added_by_user)
1584			break;
1585
1586		err = dsa_port_fdb_add(dp, fdb_info->addr, fdb_info->vid);
1587		if (err) {
1588			netdev_dbg(dev, "fdb add failed err=%d\n", err);
1589			break;
1590		}
1591		fdb_info->offloaded = true;
1592		call_switchdev_notifiers(SWITCHDEV_FDB_OFFLOADED, dev,
1593					 &fdb_info->info, NULL);
1594		break;
1595
1596	case SWITCHDEV_FDB_DEL_TO_DEVICE:
1597		fdb_info = &switchdev_work->fdb_info;
1598		if (!fdb_info->added_by_user)
1599			break;
1600
1601		err = dsa_port_fdb_del(dp, fdb_info->addr, fdb_info->vid);
1602		if (err) {
1603			netdev_dbg(dev, "fdb del failed err=%d\n", err);
1604			dev_close(dev);
1605		}
1606		break;
1607	}
1608	rtnl_unlock();
1609
1610	kfree(switchdev_work->fdb_info.addr);
1611	kfree(switchdev_work);
1612	dev_put(dev);
1613}
1614
1615static int
1616dsa_slave_switchdev_fdb_work_init(struct dsa_switchdev_event_work *
1617				  switchdev_work,
1618				  const struct switchdev_notifier_fdb_info *
1619				  fdb_info)
1620{
1621	memcpy(&switchdev_work->fdb_info, fdb_info,
1622	       sizeof(switchdev_work->fdb_info));
1623	switchdev_work->fdb_info.addr = kzalloc(ETH_ALEN, GFP_ATOMIC);
1624	if (!switchdev_work->fdb_info.addr)
1625		return -ENOMEM;
1626	ether_addr_copy((u8 *)switchdev_work->fdb_info.addr,
1627			fdb_info->addr);
1628	return 0;
1629}
1630
1631/* Called under rcu_read_lock() */
1632static int dsa_slave_switchdev_event(struct notifier_block *unused,
1633				     unsigned long event, void *ptr)
1634{
1635	struct net_device *dev = switchdev_notifier_info_to_dev(ptr);
1636	struct dsa_switchdev_event_work *switchdev_work;
1637	int err;
1638
1639	if (event == SWITCHDEV_PORT_ATTR_SET) {
1640		err = switchdev_handle_port_attr_set(dev, ptr,
1641						     dsa_slave_dev_check,
1642						     dsa_slave_port_attr_set);
1643		return notifier_from_errno(err);
1644	}
1645
1646	if (!dsa_slave_dev_check(dev))
1647		return NOTIFY_DONE;
1648
1649	switchdev_work = kzalloc(sizeof(*switchdev_work), GFP_ATOMIC);
1650	if (!switchdev_work)
1651		return NOTIFY_BAD;
1652
1653	INIT_WORK(&switchdev_work->work,
1654		  dsa_slave_switchdev_event_work);
1655	switchdev_work->dev = dev;
1656	switchdev_work->event = event;
1657
1658	switch (event) {
1659	case SWITCHDEV_FDB_ADD_TO_DEVICE: /* fall through */
1660	case SWITCHDEV_FDB_DEL_TO_DEVICE:
1661		if (dsa_slave_switchdev_fdb_work_init(switchdev_work, ptr))
 
1662			goto err_fdb_work_init;
1663		dev_hold(dev);
1664		break;
1665	default:
1666		kfree(switchdev_work);
1667		return NOTIFY_DONE;
1668	}
1669
1670	dsa_schedule_work(&switchdev_work->work);
1671	return NOTIFY_OK;
1672
1673err_fdb_work_init:
1674	kfree(switchdev_work);
1675	return NOTIFY_BAD;
1676}
1677
1678static int dsa_slave_switchdev_blocking_event(struct notifier_block *unused,
1679					      unsigned long event, void *ptr)
1680{
1681	struct net_device *dev = switchdev_notifier_info_to_dev(ptr);
1682	int err;
1683
1684	switch (event) {
1685	case SWITCHDEV_PORT_OBJ_ADD:
1686		err = switchdev_handle_port_obj_add(dev, ptr,
1687						    dsa_slave_dev_check,
1688						    dsa_slave_port_obj_add);
1689		return notifier_from_errno(err);
1690	case SWITCHDEV_PORT_OBJ_DEL:
1691		err = switchdev_handle_port_obj_del(dev, ptr,
1692						    dsa_slave_dev_check,
1693						    dsa_slave_port_obj_del);
1694		return notifier_from_errno(err);
1695	case SWITCHDEV_PORT_ATTR_SET:
1696		err = switchdev_handle_port_attr_set(dev, ptr,
1697						     dsa_slave_dev_check,
1698						     dsa_slave_port_attr_set);
1699		return notifier_from_errno(err);
1700	}
1701
1702	return NOTIFY_DONE;
1703}
1704
1705static struct notifier_block dsa_slave_nb __read_mostly = {
1706	.notifier_call  = dsa_slave_netdevice_event,
1707};
1708
1709static struct notifier_block dsa_slave_switchdev_notifier = {
1710	.notifier_call = dsa_slave_switchdev_event,
1711};
1712
1713static struct notifier_block dsa_slave_switchdev_blocking_notifier = {
1714	.notifier_call = dsa_slave_switchdev_blocking_event,
1715};
1716
1717int dsa_slave_register_notifier(void)
1718{
1719	struct notifier_block *nb;
1720	int err;
1721
1722	err = register_netdevice_notifier(&dsa_slave_nb);
1723	if (err)
1724		return err;
1725
1726	err = register_switchdev_notifier(&dsa_slave_switchdev_notifier);
1727	if (err)
1728		goto err_switchdev_nb;
1729
1730	nb = &dsa_slave_switchdev_blocking_notifier;
1731	err = register_switchdev_blocking_notifier(nb);
1732	if (err)
1733		goto err_switchdev_blocking_nb;
1734
1735	return 0;
1736
1737err_switchdev_blocking_nb:
1738	unregister_switchdev_notifier(&dsa_slave_switchdev_notifier);
1739err_switchdev_nb:
1740	unregister_netdevice_notifier(&dsa_slave_nb);
1741	return err;
1742}
1743
1744void dsa_slave_unregister_notifier(void)
1745{
1746	struct notifier_block *nb;
1747	int err;
1748
1749	nb = &dsa_slave_switchdev_blocking_notifier;
1750	err = unregister_switchdev_blocking_notifier(nb);
1751	if (err)
1752		pr_err("DSA: failed to unregister switchdev blocking notifier (%d)\n", err);
1753
1754	err = unregister_switchdev_notifier(&dsa_slave_switchdev_notifier);
1755	if (err)
1756		pr_err("DSA: failed to unregister switchdev notifier (%d)\n", err);
1757
1758	err = unregister_netdevice_notifier(&dsa_slave_nb);
1759	if (err)
1760		pr_err("DSA: failed to unregister slave notifier (%d)\n", err);
1761}