Linux Audio

Check our new training course

Loading...
v5.14.15
   1// SPDX-License-Identifier: GPL-2.0+
   2/*
   3 * Copyright (C) 2015 Microchip Technology
   4 */
   5#include <linux/module.h>
   6#include <linux/netdevice.h>
   7#include <linux/etherdevice.h>
   8#include <linux/ethtool.h>
   9#include <linux/usb.h>
  10#include <linux/crc32.h>
  11#include <linux/signal.h>
  12#include <linux/slab.h>
  13#include <linux/if_vlan.h>
  14#include <linux/uaccess.h>
  15#include <linux/linkmode.h>
  16#include <linux/list.h>
  17#include <linux/ip.h>
  18#include <linux/ipv6.h>
  19#include <linux/mdio.h>
  20#include <linux/phy.h>
  21#include <net/ip6_checksum.h>
  22#include <net/vxlan.h>
  23#include <linux/interrupt.h>
  24#include <linux/irqdomain.h>
  25#include <linux/irq.h>
  26#include <linux/irqchip/chained_irq.h>
  27#include <linux/microchipphy.h>
  28#include <linux/phy_fixed.h>
  29#include <linux/of_mdio.h>
  30#include <linux/of_net.h>
  31#include "lan78xx.h"
  32
  33#define DRIVER_AUTHOR	"WOOJUNG HUH <woojung.huh@microchip.com>"
  34#define DRIVER_DESC	"LAN78XX USB 3.0 Gigabit Ethernet Devices"
  35#define DRIVER_NAME	"lan78xx"
  36
  37#define TX_TIMEOUT_JIFFIES		(5 * HZ)
  38#define THROTTLE_JIFFIES		(HZ / 8)
  39#define UNLINK_TIMEOUT_MS		3
  40
  41#define RX_MAX_QUEUE_MEMORY		(60 * 1518)
  42
  43#define SS_USB_PKT_SIZE			(1024)
  44#define HS_USB_PKT_SIZE			(512)
  45#define FS_USB_PKT_SIZE			(64)
  46
  47#define MAX_RX_FIFO_SIZE		(12 * 1024)
  48#define MAX_TX_FIFO_SIZE		(12 * 1024)
 
 
 
 
 
 
 
 
 
 
 
 
 
  49#define DEFAULT_BURST_CAP_SIZE		(MAX_TX_FIFO_SIZE)
  50#define DEFAULT_BULK_IN_DELAY		(0x0800)
  51#define MAX_SINGLE_PACKET_SIZE		(9000)
  52#define DEFAULT_TX_CSUM_ENABLE		(true)
  53#define DEFAULT_RX_CSUM_ENABLE		(true)
  54#define DEFAULT_TSO_CSUM_ENABLE		(true)
  55#define DEFAULT_VLAN_FILTER_ENABLE	(true)
  56#define DEFAULT_VLAN_RX_OFFLOAD		(true)
  57#define TX_OVERHEAD			(8)
  58#define RXW_PADDING			2
  59
  60#define LAN78XX_USB_VENDOR_ID		(0x0424)
  61#define LAN7800_USB_PRODUCT_ID		(0x7800)
  62#define LAN7850_USB_PRODUCT_ID		(0x7850)
  63#define LAN7801_USB_PRODUCT_ID		(0x7801)
  64#define LAN78XX_EEPROM_MAGIC		(0x78A5)
  65#define LAN78XX_OTP_MAGIC		(0x78F3)
 
 
  66
  67#define	MII_READ			1
  68#define	MII_WRITE			0
  69
  70#define EEPROM_INDICATOR		(0xA5)
  71#define EEPROM_MAC_OFFSET		(0x01)
  72#define MAX_EEPROM_SIZE			512
  73#define OTP_INDICATOR_1			(0xF3)
  74#define OTP_INDICATOR_2			(0xF7)
  75
  76#define WAKE_ALL			(WAKE_PHY | WAKE_UCAST | \
  77					 WAKE_MCAST | WAKE_BCAST | \
  78					 WAKE_ARP | WAKE_MAGIC)
  79
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  80/* USB related defines */
  81#define BULK_IN_PIPE			1
  82#define BULK_OUT_PIPE			2
  83
  84/* default autosuspend delay (mSec)*/
  85#define DEFAULT_AUTOSUSPEND_DELAY	(10 * 1000)
  86
  87/* statistic update interval (mSec) */
  88#define STAT_UPDATE_TIMER		(1 * 1000)
  89
 
 
 
 
 
 
  90/* defines interrupts from interrupt EP */
  91#define MAX_INT_EP			(32)
  92#define INT_EP_INTEP			(31)
  93#define INT_EP_OTP_WR_DONE		(28)
  94#define INT_EP_EEE_TX_LPI_START		(26)
  95#define INT_EP_EEE_TX_LPI_STOP		(25)
  96#define INT_EP_EEE_RX_LPI		(24)
  97#define INT_EP_MAC_RESET_TIMEOUT	(23)
  98#define INT_EP_RDFO			(22)
  99#define INT_EP_TXE			(21)
 100#define INT_EP_USB_STATUS		(20)
 101#define INT_EP_TX_DIS			(19)
 102#define INT_EP_RX_DIS			(18)
 103#define INT_EP_PHY			(17)
 104#define INT_EP_DP			(16)
 105#define INT_EP_MAC_ERR			(15)
 106#define INT_EP_TDFU			(14)
 107#define INT_EP_TDFO			(13)
 108#define INT_EP_UTX			(12)
 109#define INT_EP_GPIO_11			(11)
 110#define INT_EP_GPIO_10			(10)
 111#define INT_EP_GPIO_9			(9)
 112#define INT_EP_GPIO_8			(8)
 113#define INT_EP_GPIO_7			(7)
 114#define INT_EP_GPIO_6			(6)
 115#define INT_EP_GPIO_5			(5)
 116#define INT_EP_GPIO_4			(4)
 117#define INT_EP_GPIO_3			(3)
 118#define INT_EP_GPIO_2			(2)
 119#define INT_EP_GPIO_1			(1)
 120#define INT_EP_GPIO_0			(0)
 121
 122static const char lan78xx_gstrings[][ETH_GSTRING_LEN] = {
 123	"RX FCS Errors",
 124	"RX Alignment Errors",
 125	"Rx Fragment Errors",
 126	"RX Jabber Errors",
 127	"RX Undersize Frame Errors",
 128	"RX Oversize Frame Errors",
 129	"RX Dropped Frames",
 130	"RX Unicast Byte Count",
 131	"RX Broadcast Byte Count",
 132	"RX Multicast Byte Count",
 133	"RX Unicast Frames",
 134	"RX Broadcast Frames",
 135	"RX Multicast Frames",
 136	"RX Pause Frames",
 137	"RX 64 Byte Frames",
 138	"RX 65 - 127 Byte Frames",
 139	"RX 128 - 255 Byte Frames",
 140	"RX 256 - 511 Bytes Frames",
 141	"RX 512 - 1023 Byte Frames",
 142	"RX 1024 - 1518 Byte Frames",
 143	"RX Greater 1518 Byte Frames",
 144	"EEE RX LPI Transitions",
 145	"EEE RX LPI Time",
 146	"TX FCS Errors",
 147	"TX Excess Deferral Errors",
 148	"TX Carrier Errors",
 149	"TX Bad Byte Count",
 150	"TX Single Collisions",
 151	"TX Multiple Collisions",
 152	"TX Excessive Collision",
 153	"TX Late Collisions",
 154	"TX Unicast Byte Count",
 155	"TX Broadcast Byte Count",
 156	"TX Multicast Byte Count",
 157	"TX Unicast Frames",
 158	"TX Broadcast Frames",
 159	"TX Multicast Frames",
 160	"TX Pause Frames",
 161	"TX 64 Byte Frames",
 162	"TX 65 - 127 Byte Frames",
 163	"TX 128 - 255 Byte Frames",
 164	"TX 256 - 511 Bytes Frames",
 165	"TX 512 - 1023 Byte Frames",
 166	"TX 1024 - 1518 Byte Frames",
 167	"TX Greater 1518 Byte Frames",
 168	"EEE TX LPI Transitions",
 169	"EEE TX LPI Time",
 170};
 171
 172struct lan78xx_statstage {
 173	u32 rx_fcs_errors;
 174	u32 rx_alignment_errors;
 175	u32 rx_fragment_errors;
 176	u32 rx_jabber_errors;
 177	u32 rx_undersize_frame_errors;
 178	u32 rx_oversize_frame_errors;
 179	u32 rx_dropped_frames;
 180	u32 rx_unicast_byte_count;
 181	u32 rx_broadcast_byte_count;
 182	u32 rx_multicast_byte_count;
 183	u32 rx_unicast_frames;
 184	u32 rx_broadcast_frames;
 185	u32 rx_multicast_frames;
 186	u32 rx_pause_frames;
 187	u32 rx_64_byte_frames;
 188	u32 rx_65_127_byte_frames;
 189	u32 rx_128_255_byte_frames;
 190	u32 rx_256_511_bytes_frames;
 191	u32 rx_512_1023_byte_frames;
 192	u32 rx_1024_1518_byte_frames;
 193	u32 rx_greater_1518_byte_frames;
 194	u32 eee_rx_lpi_transitions;
 195	u32 eee_rx_lpi_time;
 196	u32 tx_fcs_errors;
 197	u32 tx_excess_deferral_errors;
 198	u32 tx_carrier_errors;
 199	u32 tx_bad_byte_count;
 200	u32 tx_single_collisions;
 201	u32 tx_multiple_collisions;
 202	u32 tx_excessive_collision;
 203	u32 tx_late_collisions;
 204	u32 tx_unicast_byte_count;
 205	u32 tx_broadcast_byte_count;
 206	u32 tx_multicast_byte_count;
 207	u32 tx_unicast_frames;
 208	u32 tx_broadcast_frames;
 209	u32 tx_multicast_frames;
 210	u32 tx_pause_frames;
 211	u32 tx_64_byte_frames;
 212	u32 tx_65_127_byte_frames;
 213	u32 tx_128_255_byte_frames;
 214	u32 tx_256_511_bytes_frames;
 215	u32 tx_512_1023_byte_frames;
 216	u32 tx_1024_1518_byte_frames;
 217	u32 tx_greater_1518_byte_frames;
 218	u32 eee_tx_lpi_transitions;
 219	u32 eee_tx_lpi_time;
 220};
 221
 222struct lan78xx_statstage64 {
 223	u64 rx_fcs_errors;
 224	u64 rx_alignment_errors;
 225	u64 rx_fragment_errors;
 226	u64 rx_jabber_errors;
 227	u64 rx_undersize_frame_errors;
 228	u64 rx_oversize_frame_errors;
 229	u64 rx_dropped_frames;
 230	u64 rx_unicast_byte_count;
 231	u64 rx_broadcast_byte_count;
 232	u64 rx_multicast_byte_count;
 233	u64 rx_unicast_frames;
 234	u64 rx_broadcast_frames;
 235	u64 rx_multicast_frames;
 236	u64 rx_pause_frames;
 237	u64 rx_64_byte_frames;
 238	u64 rx_65_127_byte_frames;
 239	u64 rx_128_255_byte_frames;
 240	u64 rx_256_511_bytes_frames;
 241	u64 rx_512_1023_byte_frames;
 242	u64 rx_1024_1518_byte_frames;
 243	u64 rx_greater_1518_byte_frames;
 244	u64 eee_rx_lpi_transitions;
 245	u64 eee_rx_lpi_time;
 246	u64 tx_fcs_errors;
 247	u64 tx_excess_deferral_errors;
 248	u64 tx_carrier_errors;
 249	u64 tx_bad_byte_count;
 250	u64 tx_single_collisions;
 251	u64 tx_multiple_collisions;
 252	u64 tx_excessive_collision;
 253	u64 tx_late_collisions;
 254	u64 tx_unicast_byte_count;
 255	u64 tx_broadcast_byte_count;
 256	u64 tx_multicast_byte_count;
 257	u64 tx_unicast_frames;
 258	u64 tx_broadcast_frames;
 259	u64 tx_multicast_frames;
 260	u64 tx_pause_frames;
 261	u64 tx_64_byte_frames;
 262	u64 tx_65_127_byte_frames;
 263	u64 tx_128_255_byte_frames;
 264	u64 tx_256_511_bytes_frames;
 265	u64 tx_512_1023_byte_frames;
 266	u64 tx_1024_1518_byte_frames;
 267	u64 tx_greater_1518_byte_frames;
 268	u64 eee_tx_lpi_transitions;
 269	u64 eee_tx_lpi_time;
 270};
 271
 272static u32 lan78xx_regs[] = {
 273	ID_REV,
 274	INT_STS,
 275	HW_CFG,
 276	PMT_CTL,
 277	E2P_CMD,
 278	E2P_DATA,
 279	USB_STATUS,
 280	VLAN_TYPE,
 281	MAC_CR,
 282	MAC_RX,
 283	MAC_TX,
 284	FLOW,
 285	ERR_STS,
 286	MII_ACC,
 287	MII_DATA,
 288	EEE_TX_LPI_REQ_DLY,
 289	EEE_TW_TX_SYS,
 290	EEE_TX_LPI_REM_DLY,
 291	WUCSR
 292};
 293
 294#define PHY_REG_SIZE (32 * sizeof(u32))
 295
 296struct lan78xx_net;
 297
 298struct lan78xx_priv {
 299	struct lan78xx_net *dev;
 300	u32 rfe_ctl;
 301	u32 mchash_table[DP_SEL_VHF_HASH_LEN]; /* multicast hash table */
 302	u32 pfilter_table[NUM_OF_MAF][2]; /* perfect filter table */
 303	u32 vlan_table[DP_SEL_VHF_VLAN_LEN];
 304	struct mutex dataport_mutex; /* for dataport access */
 305	spinlock_t rfe_ctl_lock; /* for rfe register access */
 306	struct work_struct set_multicast;
 307	struct work_struct set_vlan;
 308	u32 wol;
 309};
 310
 311enum skb_state {
 312	illegal = 0,
 313	tx_start,
 314	tx_done,
 315	rx_start,
 316	rx_done,
 317	rx_cleanup,
 318	unlink_start
 319};
 320
 321struct skb_data {		/* skb->cb is one of these */
 322	struct urb *urb;
 323	struct lan78xx_net *dev;
 324	enum skb_state state;
 325	size_t length;
 326	int num_of_packet;
 327};
 328
 329struct usb_context {
 330	struct usb_ctrlrequest req;
 331	struct lan78xx_net *dev;
 332};
 333
 334#define EVENT_TX_HALT			0
 335#define EVENT_RX_HALT			1
 336#define EVENT_RX_MEMORY			2
 337#define EVENT_STS_SPLIT			3
 338#define EVENT_LINK_RESET		4
 339#define EVENT_RX_PAUSED			5
 340#define EVENT_DEV_WAKING		6
 341#define EVENT_DEV_ASLEEP		7
 342#define EVENT_DEV_OPEN			8
 343#define EVENT_STAT_UPDATE		9
 
 344
 345struct statstage {
 346	struct mutex			access_lock;	/* for stats access */
 347	struct lan78xx_statstage	saved;
 348	struct lan78xx_statstage	rollover_count;
 349	struct lan78xx_statstage	rollover_max;
 350	struct lan78xx_statstage64	curr_stat;
 351};
 352
 353struct irq_domain_data {
 354	struct irq_domain	*irqdomain;
 355	unsigned int		phyirq;
 356	struct irq_chip		*irqchip;
 357	irq_flow_handler_t	irq_handler;
 358	u32			irqenable;
 359	struct mutex		irq_lock;		/* for irq bus access */
 360};
 361
 362struct lan78xx_net {
 363	struct net_device	*net;
 364	struct usb_device	*udev;
 365	struct usb_interface	*intf;
 366	void			*driver_priv;
 367
 368	int			rx_qlen;
 369	int			tx_qlen;
 
 
 
 
 
 370	struct sk_buff_head	rxq;
 
 
 
 371	struct sk_buff_head	txq;
 372	struct sk_buff_head	done;
 373	struct sk_buff_head	rxq_pause;
 374	struct sk_buff_head	txq_pend;
 375
 376	struct tasklet_struct	bh;
 
 377	struct delayed_work	wq;
 378
 379	int			msg_enable;
 380
 381	struct urb		*urb_intr;
 382	struct usb_anchor	deferred;
 383
 
 384	struct mutex		phy_mutex; /* for phy access */
 385	unsigned		pipe_in, pipe_out, pipe_intr;
 386
 387	u32			hard_mtu;	/* count any extra framing */
 388	size_t			rx_urb_size;	/* size for rx urbs */
 389
 390	unsigned long		flags;
 391
 392	wait_queue_head_t	*wait;
 393	unsigned char		suspend_count;
 394
 395	unsigned		maxpacket;
 396	struct timer_list	delay;
 397	struct timer_list	stat_monitor;
 398
 399	unsigned long		data[5];
 400
 401	int			link_on;
 402	u8			mdix_ctrl;
 403
 404	u32			chipid;
 405	u32			chiprev;
 406	struct mii_bus		*mdiobus;
 407	phy_interface_t		interface;
 408
 409	int			fc_autoneg;
 410	u8			fc_request_control;
 411
 412	int			delta;
 413	struct statstage	stats;
 414
 415	struct irq_domain_data	domain_data;
 416};
 417
 418/* define external phy id */
 419#define	PHY_LAN8835			(0x0007C130)
 420#define	PHY_KSZ9031RNX			(0x00221620)
 421
 422/* use ethtool to change the level for any given device */
 423static int msg_level = -1;
 424module_param(msg_level, int, 0);
 425MODULE_PARM_DESC(msg_level, "Override default message level");
 426
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 427static int lan78xx_read_reg(struct lan78xx_net *dev, u32 index, u32 *data)
 428{
 429	u32 *buf = kmalloc(sizeof(u32), GFP_KERNEL);
 430	int ret;
 431
 
 
 
 
 432	if (!buf)
 433		return -ENOMEM;
 434
 435	ret = usb_control_msg(dev->udev, usb_rcvctrlpipe(dev->udev, 0),
 436			      USB_VENDOR_REQUEST_READ_REGISTER,
 437			      USB_DIR_IN | USB_TYPE_VENDOR | USB_RECIP_DEVICE,
 438			      0, index, buf, 4, USB_CTRL_GET_TIMEOUT);
 439	if (likely(ret >= 0)) {
 440		le32_to_cpus(buf);
 441		*data = *buf;
 442	} else {
 443		netdev_warn(dev->net,
 444			    "Failed to read register index 0x%08x. ret = %d",
 445			    index, ret);
 446	}
 447
 448	kfree(buf);
 449
 450	return ret;
 451}
 452
 453static int lan78xx_write_reg(struct lan78xx_net *dev, u32 index, u32 data)
 454{
 455	u32 *buf = kmalloc(sizeof(u32), GFP_KERNEL);
 456	int ret;
 457
 
 
 
 
 458	if (!buf)
 459		return -ENOMEM;
 460
 461	*buf = data;
 462	cpu_to_le32s(buf);
 463
 464	ret = usb_control_msg(dev->udev, usb_sndctrlpipe(dev->udev, 0),
 465			      USB_VENDOR_REQUEST_WRITE_REGISTER,
 466			      USB_DIR_OUT | USB_TYPE_VENDOR | USB_RECIP_DEVICE,
 467			      0, index, buf, 4, USB_CTRL_SET_TIMEOUT);
 468	if (unlikely(ret < 0)) {
 
 469		netdev_warn(dev->net,
 470			    "Failed to write register index 0x%08x. ret = %d",
 471			    index, ret);
 472	}
 473
 474	kfree(buf);
 475
 476	return ret;
 477}
 478
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 479static int lan78xx_read_stats(struct lan78xx_net *dev,
 480			      struct lan78xx_statstage *data)
 481{
 482	int ret = 0;
 483	int i;
 484	struct lan78xx_statstage *stats;
 485	u32 *src;
 486	u32 *dst;
 487
 488	stats = kmalloc(sizeof(*stats), GFP_KERNEL);
 489	if (!stats)
 490		return -ENOMEM;
 491
 492	ret = usb_control_msg(dev->udev,
 493			      usb_rcvctrlpipe(dev->udev, 0),
 494			      USB_VENDOR_REQUEST_GET_STATS,
 495			      USB_DIR_IN | USB_TYPE_VENDOR | USB_RECIP_DEVICE,
 496			      0,
 497			      0,
 498			      (void *)stats,
 499			      sizeof(*stats),
 500			      USB_CTRL_SET_TIMEOUT);
 501	if (likely(ret >= 0)) {
 502		src = (u32 *)stats;
 503		dst = (u32 *)data;
 504		for (i = 0; i < sizeof(*stats)/sizeof(u32); i++) {
 505			le32_to_cpus(&src[i]);
 506			dst[i] = src[i];
 507		}
 508	} else {
 509		netdev_warn(dev->net,
 510			    "Failed to read stat ret = %d", ret);
 511	}
 512
 513	kfree(stats);
 514
 515	return ret;
 516}
 517
 518#define check_counter_rollover(struct1, dev_stats, member) {	\
 519	if (struct1->member < dev_stats.saved.member)		\
 520		dev_stats.rollover_count.member++;		\
 521	}
 
 522
 523static void lan78xx_check_stat_rollover(struct lan78xx_net *dev,
 524					struct lan78xx_statstage *stats)
 525{
 526	check_counter_rollover(stats, dev->stats, rx_fcs_errors);
 527	check_counter_rollover(stats, dev->stats, rx_alignment_errors);
 528	check_counter_rollover(stats, dev->stats, rx_fragment_errors);
 529	check_counter_rollover(stats, dev->stats, rx_jabber_errors);
 530	check_counter_rollover(stats, dev->stats, rx_undersize_frame_errors);
 531	check_counter_rollover(stats, dev->stats, rx_oversize_frame_errors);
 532	check_counter_rollover(stats, dev->stats, rx_dropped_frames);
 533	check_counter_rollover(stats, dev->stats, rx_unicast_byte_count);
 534	check_counter_rollover(stats, dev->stats, rx_broadcast_byte_count);
 535	check_counter_rollover(stats, dev->stats, rx_multicast_byte_count);
 536	check_counter_rollover(stats, dev->stats, rx_unicast_frames);
 537	check_counter_rollover(stats, dev->stats, rx_broadcast_frames);
 538	check_counter_rollover(stats, dev->stats, rx_multicast_frames);
 539	check_counter_rollover(stats, dev->stats, rx_pause_frames);
 540	check_counter_rollover(stats, dev->stats, rx_64_byte_frames);
 541	check_counter_rollover(stats, dev->stats, rx_65_127_byte_frames);
 542	check_counter_rollover(stats, dev->stats, rx_128_255_byte_frames);
 543	check_counter_rollover(stats, dev->stats, rx_256_511_bytes_frames);
 544	check_counter_rollover(stats, dev->stats, rx_512_1023_byte_frames);
 545	check_counter_rollover(stats, dev->stats, rx_1024_1518_byte_frames);
 546	check_counter_rollover(stats, dev->stats, rx_greater_1518_byte_frames);
 547	check_counter_rollover(stats, dev->stats, eee_rx_lpi_transitions);
 548	check_counter_rollover(stats, dev->stats, eee_rx_lpi_time);
 549	check_counter_rollover(stats, dev->stats, tx_fcs_errors);
 550	check_counter_rollover(stats, dev->stats, tx_excess_deferral_errors);
 551	check_counter_rollover(stats, dev->stats, tx_carrier_errors);
 552	check_counter_rollover(stats, dev->stats, tx_bad_byte_count);
 553	check_counter_rollover(stats, dev->stats, tx_single_collisions);
 554	check_counter_rollover(stats, dev->stats, tx_multiple_collisions);
 555	check_counter_rollover(stats, dev->stats, tx_excessive_collision);
 556	check_counter_rollover(stats, dev->stats, tx_late_collisions);
 557	check_counter_rollover(stats, dev->stats, tx_unicast_byte_count);
 558	check_counter_rollover(stats, dev->stats, tx_broadcast_byte_count);
 559	check_counter_rollover(stats, dev->stats, tx_multicast_byte_count);
 560	check_counter_rollover(stats, dev->stats, tx_unicast_frames);
 561	check_counter_rollover(stats, dev->stats, tx_broadcast_frames);
 562	check_counter_rollover(stats, dev->stats, tx_multicast_frames);
 563	check_counter_rollover(stats, dev->stats, tx_pause_frames);
 564	check_counter_rollover(stats, dev->stats, tx_64_byte_frames);
 565	check_counter_rollover(stats, dev->stats, tx_65_127_byte_frames);
 566	check_counter_rollover(stats, dev->stats, tx_128_255_byte_frames);
 567	check_counter_rollover(stats, dev->stats, tx_256_511_bytes_frames);
 568	check_counter_rollover(stats, dev->stats, tx_512_1023_byte_frames);
 569	check_counter_rollover(stats, dev->stats, tx_1024_1518_byte_frames);
 570	check_counter_rollover(stats, dev->stats, tx_greater_1518_byte_frames);
 571	check_counter_rollover(stats, dev->stats, eee_tx_lpi_transitions);
 572	check_counter_rollover(stats, dev->stats, eee_tx_lpi_time);
 573
 574	memcpy(&dev->stats.saved, stats, sizeof(struct lan78xx_statstage));
 575}
 576
 577static void lan78xx_update_stats(struct lan78xx_net *dev)
 578{
 579	u32 *p, *count, *max;
 580	u64 *data;
 581	int i;
 582	struct lan78xx_statstage lan78xx_stats;
 583
 584	if (usb_autopm_get_interface(dev->intf) < 0)
 585		return;
 586
 587	p = (u32 *)&lan78xx_stats;
 588	count = (u32 *)&dev->stats.rollover_count;
 589	max = (u32 *)&dev->stats.rollover_max;
 590	data = (u64 *)&dev->stats.curr_stat;
 591
 592	mutex_lock(&dev->stats.access_lock);
 593
 594	if (lan78xx_read_stats(dev, &lan78xx_stats) > 0)
 595		lan78xx_check_stat_rollover(dev, &lan78xx_stats);
 596
 597	for (i = 0; i < (sizeof(lan78xx_stats) / (sizeof(u32))); i++)
 598		data[i] = (u64)p[i] + ((u64)count[i] * ((u64)max[i] + 1));
 599
 600	mutex_unlock(&dev->stats.access_lock);
 601
 602	usb_autopm_put_interface(dev->intf);
 603}
 604
 605/* Loop until the read is completed with timeout called with phy_mutex held */
 606static int lan78xx_phy_wait_not_busy(struct lan78xx_net *dev)
 607{
 608	unsigned long start_time = jiffies;
 609	u32 val;
 610	int ret;
 611
 612	do {
 613		ret = lan78xx_read_reg(dev, MII_ACC, &val);
 614		if (unlikely(ret < 0))
 615			return -EIO;
 616
 617		if (!(val & MII_ACC_MII_BUSY_))
 618			return 0;
 619	} while (!time_after(jiffies, start_time + HZ));
 620
 621	return -EIO;
 622}
 623
 624static inline u32 mii_access(int id, int index, int read)
 625{
 626	u32 ret;
 627
 628	ret = ((u32)id << MII_ACC_PHY_ADDR_SHIFT_) & MII_ACC_PHY_ADDR_MASK_;
 629	ret |= ((u32)index << MII_ACC_MIIRINDA_SHIFT_) & MII_ACC_MIIRINDA_MASK_;
 630	if (read)
 631		ret |= MII_ACC_MII_READ_;
 632	else
 633		ret |= MII_ACC_MII_WRITE_;
 634	ret |= MII_ACC_MII_BUSY_;
 635
 636	return ret;
 637}
 638
 639static int lan78xx_wait_eeprom(struct lan78xx_net *dev)
 640{
 641	unsigned long start_time = jiffies;
 642	u32 val;
 643	int ret;
 644
 645	do {
 646		ret = lan78xx_read_reg(dev, E2P_CMD, &val);
 647		if (unlikely(ret < 0))
 648			return -EIO;
 649
 650		if (!(val & E2P_CMD_EPC_BUSY_) ||
 651		    (val & E2P_CMD_EPC_TIMEOUT_))
 652			break;
 653		usleep_range(40, 100);
 654	} while (!time_after(jiffies, start_time + HZ));
 655
 656	if (val & (E2P_CMD_EPC_TIMEOUT_ | E2P_CMD_EPC_BUSY_)) {
 657		netdev_warn(dev->net, "EEPROM read operation timeout");
 658		return -EIO;
 659	}
 660
 661	return 0;
 662}
 663
 664static int lan78xx_eeprom_confirm_not_busy(struct lan78xx_net *dev)
 665{
 666	unsigned long start_time = jiffies;
 667	u32 val;
 668	int ret;
 669
 670	do {
 671		ret = lan78xx_read_reg(dev, E2P_CMD, &val);
 672		if (unlikely(ret < 0))
 673			return -EIO;
 674
 675		if (!(val & E2P_CMD_EPC_BUSY_))
 676			return 0;
 677
 678		usleep_range(40, 100);
 679	} while (!time_after(jiffies, start_time + HZ));
 680
 681	netdev_warn(dev->net, "EEPROM is busy");
 682	return -EIO;
 683}
 684
 685static int lan78xx_read_raw_eeprom(struct lan78xx_net *dev, u32 offset,
 686				   u32 length, u8 *data)
 687{
 688	u32 val;
 689	u32 saved;
 690	int i, ret;
 691	int retval;
 692
 693	/* depends on chip, some EEPROM pins are muxed with LED function.
 694	 * disable & restore LED function to access EEPROM.
 695	 */
 696	ret = lan78xx_read_reg(dev, HW_CFG, &val);
 697	saved = val;
 698	if (dev->chipid == ID_REV_CHIP_ID_7800_) {
 699		val &= ~(HW_CFG_LED1_EN_ | HW_CFG_LED0_EN_);
 700		ret = lan78xx_write_reg(dev, HW_CFG, val);
 701	}
 702
 703	retval = lan78xx_eeprom_confirm_not_busy(dev);
 704	if (retval)
 705		return retval;
 706
 707	for (i = 0; i < length; i++) {
 708		val = E2P_CMD_EPC_BUSY_ | E2P_CMD_EPC_CMD_READ_;
 709		val |= (offset & E2P_CMD_EPC_ADDR_MASK_);
 710		ret = lan78xx_write_reg(dev, E2P_CMD, val);
 711		if (unlikely(ret < 0)) {
 712			retval = -EIO;
 713			goto exit;
 714		}
 715
 716		retval = lan78xx_wait_eeprom(dev);
 717		if (retval < 0)
 718			goto exit;
 719
 720		ret = lan78xx_read_reg(dev, E2P_DATA, &val);
 721		if (unlikely(ret < 0)) {
 722			retval = -EIO;
 723			goto exit;
 724		}
 725
 726		data[i] = val & 0xFF;
 727		offset++;
 728	}
 729
 730	retval = 0;
 731exit:
 732	if (dev->chipid == ID_REV_CHIP_ID_7800_)
 733		ret = lan78xx_write_reg(dev, HW_CFG, saved);
 734
 735	return retval;
 736}
 737
 738static int lan78xx_read_eeprom(struct lan78xx_net *dev, u32 offset,
 739			       u32 length, u8 *data)
 740{
 741	u8 sig;
 742	int ret;
 743
 744	ret = lan78xx_read_raw_eeprom(dev, 0, 1, &sig);
 745	if ((ret == 0) && (sig == EEPROM_INDICATOR))
 746		ret = lan78xx_read_raw_eeprom(dev, offset, length, data);
 747	else
 748		ret = -EINVAL;
 749
 750	return ret;
 751}
 752
 753static int lan78xx_write_raw_eeprom(struct lan78xx_net *dev, u32 offset,
 754				    u32 length, u8 *data)
 755{
 756	u32 val;
 757	u32 saved;
 758	int i, ret;
 759	int retval;
 760
 761	/* depends on chip, some EEPROM pins are muxed with LED function.
 762	 * disable & restore LED function to access EEPROM.
 763	 */
 764	ret = lan78xx_read_reg(dev, HW_CFG, &val);
 765	saved = val;
 766	if (dev->chipid == ID_REV_CHIP_ID_7800_) {
 767		val &= ~(HW_CFG_LED1_EN_ | HW_CFG_LED0_EN_);
 768		ret = lan78xx_write_reg(dev, HW_CFG, val);
 769	}
 770
 771	retval = lan78xx_eeprom_confirm_not_busy(dev);
 772	if (retval)
 773		goto exit;
 774
 775	/* Issue write/erase enable command */
 776	val = E2P_CMD_EPC_BUSY_ | E2P_CMD_EPC_CMD_EWEN_;
 777	ret = lan78xx_write_reg(dev, E2P_CMD, val);
 778	if (unlikely(ret < 0)) {
 779		retval = -EIO;
 780		goto exit;
 781	}
 782
 783	retval = lan78xx_wait_eeprom(dev);
 784	if (retval < 0)
 785		goto exit;
 786
 787	for (i = 0; i < length; i++) {
 788		/* Fill data register */
 789		val = data[i];
 790		ret = lan78xx_write_reg(dev, E2P_DATA, val);
 791		if (ret < 0) {
 792			retval = -EIO;
 793			goto exit;
 794		}
 795
 796		/* Send "write" command */
 797		val = E2P_CMD_EPC_BUSY_ | E2P_CMD_EPC_CMD_WRITE_;
 798		val |= (offset & E2P_CMD_EPC_ADDR_MASK_);
 799		ret = lan78xx_write_reg(dev, E2P_CMD, val);
 800		if (ret < 0) {
 801			retval = -EIO;
 802			goto exit;
 803		}
 804
 805		retval = lan78xx_wait_eeprom(dev);
 806		if (retval < 0)
 807			goto exit;
 808
 809		offset++;
 810	}
 811
 812	retval = 0;
 813exit:
 814	if (dev->chipid == ID_REV_CHIP_ID_7800_)
 815		ret = lan78xx_write_reg(dev, HW_CFG, saved);
 816
 817	return retval;
 818}
 819
 820static int lan78xx_read_raw_otp(struct lan78xx_net *dev, u32 offset,
 821				u32 length, u8 *data)
 822{
 823	int i;
 824	u32 buf;
 825	unsigned long timeout;
 826
 827	lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
 828
 829	if (buf & OTP_PWR_DN_PWRDN_N_) {
 830		/* clear it and wait to be cleared */
 831		lan78xx_write_reg(dev, OTP_PWR_DN, 0);
 832
 833		timeout = jiffies + HZ;
 834		do {
 835			usleep_range(1, 10);
 836			lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
 837			if (time_after(jiffies, timeout)) {
 838				netdev_warn(dev->net,
 839					    "timeout on OTP_PWR_DN");
 840				return -EIO;
 841			}
 842		} while (buf & OTP_PWR_DN_PWRDN_N_);
 843	}
 844
 845	for (i = 0; i < length; i++) {
 846		lan78xx_write_reg(dev, OTP_ADDR1,
 847					((offset + i) >> 8) & OTP_ADDR1_15_11);
 848		lan78xx_write_reg(dev, OTP_ADDR2,
 849					((offset + i) & OTP_ADDR2_10_3));
 850
 851		lan78xx_write_reg(dev, OTP_FUNC_CMD, OTP_FUNC_CMD_READ_);
 852		lan78xx_write_reg(dev, OTP_CMD_GO, OTP_CMD_GO_GO_);
 853
 854		timeout = jiffies + HZ;
 855		do {
 856			udelay(1);
 857			lan78xx_read_reg(dev, OTP_STATUS, &buf);
 858			if (time_after(jiffies, timeout)) {
 859				netdev_warn(dev->net,
 860					    "timeout on OTP_STATUS");
 861				return -EIO;
 862			}
 863		} while (buf & OTP_STATUS_BUSY_);
 864
 865		lan78xx_read_reg(dev, OTP_RD_DATA, &buf);
 866
 867		data[i] = (u8)(buf & 0xFF);
 868	}
 869
 870	return 0;
 871}
 872
 873static int lan78xx_write_raw_otp(struct lan78xx_net *dev, u32 offset,
 874				 u32 length, u8 *data)
 875{
 876	int i;
 877	u32 buf;
 878	unsigned long timeout;
 879
 880	lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
 881
 882	if (buf & OTP_PWR_DN_PWRDN_N_) {
 883		/* clear it and wait to be cleared */
 884		lan78xx_write_reg(dev, OTP_PWR_DN, 0);
 885
 886		timeout = jiffies + HZ;
 887		do {
 888			udelay(1);
 889			lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
 890			if (time_after(jiffies, timeout)) {
 891				netdev_warn(dev->net,
 892					    "timeout on OTP_PWR_DN completion");
 893				return -EIO;
 894			}
 895		} while (buf & OTP_PWR_DN_PWRDN_N_);
 896	}
 897
 898	/* set to BYTE program mode */
 899	lan78xx_write_reg(dev, OTP_PRGM_MODE, OTP_PRGM_MODE_BYTE_);
 900
 901	for (i = 0; i < length; i++) {
 902		lan78xx_write_reg(dev, OTP_ADDR1,
 903					((offset + i) >> 8) & OTP_ADDR1_15_11);
 904		lan78xx_write_reg(dev, OTP_ADDR2,
 905					((offset + i) & OTP_ADDR2_10_3));
 906		lan78xx_write_reg(dev, OTP_PRGM_DATA, data[i]);
 907		lan78xx_write_reg(dev, OTP_TST_CMD, OTP_TST_CMD_PRGVRFY_);
 908		lan78xx_write_reg(dev, OTP_CMD_GO, OTP_CMD_GO_GO_);
 909
 910		timeout = jiffies + HZ;
 911		do {
 912			udelay(1);
 913			lan78xx_read_reg(dev, OTP_STATUS, &buf);
 914			if (time_after(jiffies, timeout)) {
 915				netdev_warn(dev->net,
 916					    "Timeout on OTP_STATUS completion");
 917				return -EIO;
 918			}
 919		} while (buf & OTP_STATUS_BUSY_);
 920	}
 921
 922	return 0;
 923}
 924
 925static int lan78xx_read_otp(struct lan78xx_net *dev, u32 offset,
 926			    u32 length, u8 *data)
 927{
 928	u8 sig;
 929	int ret;
 930
 931	ret = lan78xx_read_raw_otp(dev, 0, 1, &sig);
 932
 933	if (ret == 0) {
 934		if (sig == OTP_INDICATOR_2)
 935			offset += 0x100;
 936		else if (sig != OTP_INDICATOR_1)
 937			ret = -EINVAL;
 938		if (!ret)
 939			ret = lan78xx_read_raw_otp(dev, offset, length, data);
 940	}
 941
 942	return ret;
 943}
 944
 945static int lan78xx_dataport_wait_not_busy(struct lan78xx_net *dev)
 946{
 947	int i, ret;
 948
 949	for (i = 0; i < 100; i++) {
 950		u32 dp_sel;
 951
 952		ret = lan78xx_read_reg(dev, DP_SEL, &dp_sel);
 953		if (unlikely(ret < 0))
 954			return -EIO;
 955
 956		if (dp_sel & DP_SEL_DPRDY_)
 957			return 0;
 958
 959		usleep_range(40, 100);
 960	}
 961
 962	netdev_warn(dev->net, "lan78xx_dataport_wait_not_busy timed out");
 963
 964	return -EIO;
 965}
 966
 967static int lan78xx_dataport_write(struct lan78xx_net *dev, u32 ram_select,
 968				  u32 addr, u32 length, u32 *buf)
 969{
 970	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
 971	u32 dp_sel;
 972	int i, ret;
 973
 974	if (usb_autopm_get_interface(dev->intf) < 0)
 975			return 0;
 976
 977	mutex_lock(&pdata->dataport_mutex);
 978
 979	ret = lan78xx_dataport_wait_not_busy(dev);
 980	if (ret < 0)
 981		goto done;
 982
 983	ret = lan78xx_read_reg(dev, DP_SEL, &dp_sel);
 984
 985	dp_sel &= ~DP_SEL_RSEL_MASK_;
 986	dp_sel |= ram_select;
 987	ret = lan78xx_write_reg(dev, DP_SEL, dp_sel);
 988
 989	for (i = 0; i < length; i++) {
 990		ret = lan78xx_write_reg(dev, DP_ADDR, addr + i);
 991
 992		ret = lan78xx_write_reg(dev, DP_DATA, buf[i]);
 993
 994		ret = lan78xx_write_reg(dev, DP_CMD, DP_CMD_WRITE_);
 995
 996		ret = lan78xx_dataport_wait_not_busy(dev);
 997		if (ret < 0)
 998			goto done;
 999	}
1000
1001done:
1002	mutex_unlock(&pdata->dataport_mutex);
1003	usb_autopm_put_interface(dev->intf);
1004
1005	return ret;
1006}
1007
1008static void lan78xx_set_addr_filter(struct lan78xx_priv *pdata,
1009				    int index, u8 addr[ETH_ALEN])
1010{
1011	u32 temp;
1012
1013	if ((pdata) && (index > 0) && (index < NUM_OF_MAF)) {
1014		temp = addr[3];
1015		temp = addr[2] | (temp << 8);
1016		temp = addr[1] | (temp << 8);
1017		temp = addr[0] | (temp << 8);
1018		pdata->pfilter_table[index][1] = temp;
1019		temp = addr[5];
1020		temp = addr[4] | (temp << 8);
1021		temp |= MAF_HI_VALID_ | MAF_HI_TYPE_DST_;
1022		pdata->pfilter_table[index][0] = temp;
1023	}
1024}
1025
1026/* returns hash bit number for given MAC address */
1027static inline u32 lan78xx_hash(char addr[ETH_ALEN])
1028{
1029	return (ether_crc(ETH_ALEN, addr) >> 23) & 0x1ff;
1030}
1031
1032static void lan78xx_deferred_multicast_write(struct work_struct *param)
1033{
1034	struct lan78xx_priv *pdata =
1035			container_of(param, struct lan78xx_priv, set_multicast);
1036	struct lan78xx_net *dev = pdata->dev;
1037	int i;
1038
1039	netif_dbg(dev, drv, dev->net, "deferred multicast write 0x%08x\n",
1040		  pdata->rfe_ctl);
1041
1042	lan78xx_dataport_write(dev, DP_SEL_RSEL_VLAN_DA_, DP_SEL_VHF_VLAN_LEN,
1043			       DP_SEL_VHF_HASH_LEN, pdata->mchash_table);
1044
1045	for (i = 1; i < NUM_OF_MAF; i++) {
1046		lan78xx_write_reg(dev, MAF_HI(i), 0);
1047		lan78xx_write_reg(dev, MAF_LO(i),
1048					pdata->pfilter_table[i][1]);
1049		lan78xx_write_reg(dev, MAF_HI(i),
1050					pdata->pfilter_table[i][0]);
1051	}
1052
1053	lan78xx_write_reg(dev, RFE_CTL, pdata->rfe_ctl);
1054}
1055
1056static void lan78xx_set_multicast(struct net_device *netdev)
1057{
1058	struct lan78xx_net *dev = netdev_priv(netdev);
1059	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1060	unsigned long flags;
1061	int i;
1062
1063	spin_lock_irqsave(&pdata->rfe_ctl_lock, flags);
1064
1065	pdata->rfe_ctl &= ~(RFE_CTL_UCAST_EN_ | RFE_CTL_MCAST_EN_ |
1066			    RFE_CTL_DA_PERFECT_ | RFE_CTL_MCAST_HASH_);
1067
1068	for (i = 0; i < DP_SEL_VHF_HASH_LEN; i++)
1069			pdata->mchash_table[i] = 0;
 
1070	/* pfilter_table[0] has own HW address */
1071	for (i = 1; i < NUM_OF_MAF; i++) {
1072			pdata->pfilter_table[i][0] =
1073			pdata->pfilter_table[i][1] = 0;
1074	}
1075
1076	pdata->rfe_ctl |= RFE_CTL_BCAST_EN_;
1077
1078	if (dev->net->flags & IFF_PROMISC) {
1079		netif_dbg(dev, drv, dev->net, "promiscuous mode enabled");
1080		pdata->rfe_ctl |= RFE_CTL_MCAST_EN_ | RFE_CTL_UCAST_EN_;
1081	} else {
1082		if (dev->net->flags & IFF_ALLMULTI) {
1083			netif_dbg(dev, drv, dev->net,
1084				  "receive all multicast enabled");
1085			pdata->rfe_ctl |= RFE_CTL_MCAST_EN_;
1086		}
1087	}
1088
1089	if (netdev_mc_count(dev->net)) {
1090		struct netdev_hw_addr *ha;
1091		int i;
1092
1093		netif_dbg(dev, drv, dev->net, "receive multicast hash filter");
1094
1095		pdata->rfe_ctl |= RFE_CTL_DA_PERFECT_;
1096
1097		i = 1;
1098		netdev_for_each_mc_addr(ha, netdev) {
1099			/* set first 32 into Perfect Filter */
1100			if (i < 33) {
1101				lan78xx_set_addr_filter(pdata, i, ha->addr);
1102			} else {
1103				u32 bitnum = lan78xx_hash(ha->addr);
1104
1105				pdata->mchash_table[bitnum / 32] |=
1106							(1 << (bitnum % 32));
1107				pdata->rfe_ctl |= RFE_CTL_MCAST_HASH_;
1108			}
1109			i++;
1110		}
1111	}
1112
1113	spin_unlock_irqrestore(&pdata->rfe_ctl_lock, flags);
1114
1115	/* defer register writes to a sleepable context */
1116	schedule_work(&pdata->set_multicast);
1117}
1118
1119static int lan78xx_update_flowcontrol(struct lan78xx_net *dev, u8 duplex,
1120				      u16 lcladv, u16 rmtadv)
1121{
1122	u32 flow = 0, fct_flow = 0;
1123	u8 cap;
1124
1125	if (dev->fc_autoneg)
1126		cap = mii_resolve_flowctrl_fdx(lcladv, rmtadv);
1127	else
1128		cap = dev->fc_request_control;
1129
1130	if (cap & FLOW_CTRL_TX)
1131		flow |= (FLOW_CR_TX_FCEN_ | 0xFFFF);
1132
1133	if (cap & FLOW_CTRL_RX)
1134		flow |= FLOW_CR_RX_FCEN_;
1135
1136	if (dev->udev->speed == USB_SPEED_SUPER)
1137		fct_flow = 0x817;
1138	else if (dev->udev->speed == USB_SPEED_HIGH)
1139		fct_flow = 0x211;
1140
1141	netif_dbg(dev, link, dev->net, "rx pause %s, tx pause %s",
1142		  (cap & FLOW_CTRL_RX ? "enabled" : "disabled"),
1143		  (cap & FLOW_CTRL_TX ? "enabled" : "disabled"));
1144
1145	lan78xx_write_reg(dev, FCT_FLOW, fct_flow);
1146
1147	/* threshold value should be set before enabling flow */
1148	lan78xx_write_reg(dev, FLOW, flow);
1149
1150	return 0;
1151}
1152
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1153static int lan78xx_link_reset(struct lan78xx_net *dev)
1154{
1155	struct phy_device *phydev = dev->net->phydev;
1156	struct ethtool_link_ksettings ecmd;
1157	int ladv, radv, ret, link;
1158	u32 buf;
1159
1160	/* clear LAN78xx interrupt status */
1161	ret = lan78xx_write_reg(dev, INT_STS, INT_STS_PHY_INT_);
1162	if (unlikely(ret < 0))
1163		return -EIO;
1164
1165	mutex_lock(&phydev->lock);
1166	phy_read_status(phydev);
1167	link = phydev->link;
1168	mutex_unlock(&phydev->lock);
1169
1170	if (!link && dev->link_on) {
1171		dev->link_on = false;
1172
1173		/* reset MAC */
1174		ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1175		if (unlikely(ret < 0))
1176			return -EIO;
1177		buf |= MAC_CR_RST_;
1178		ret = lan78xx_write_reg(dev, MAC_CR, buf);
1179		if (unlikely(ret < 0))
1180			return -EIO;
1181
1182		del_timer(&dev->stat_monitor);
1183	} else if (link && !dev->link_on) {
1184		dev->link_on = true;
1185
1186		phy_ethtool_ksettings_get(phydev, &ecmd);
1187
1188		if (dev->udev->speed == USB_SPEED_SUPER) {
1189			if (ecmd.base.speed == 1000) {
1190				/* disable U2 */
1191				ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
 
 
1192				buf &= ~USB_CFG1_DEV_U2_INIT_EN_;
1193				ret = lan78xx_write_reg(dev, USB_CFG1, buf);
 
 
1194				/* enable U1 */
1195				ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
 
 
1196				buf |= USB_CFG1_DEV_U1_INIT_EN_;
1197				ret = lan78xx_write_reg(dev, USB_CFG1, buf);
 
 
1198			} else {
1199				/* enable U1 & U2 */
1200				ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
 
 
1201				buf |= USB_CFG1_DEV_U2_INIT_EN_;
1202				buf |= USB_CFG1_DEV_U1_INIT_EN_;
1203				ret = lan78xx_write_reg(dev, USB_CFG1, buf);
 
 
1204			}
1205		}
1206
1207		ladv = phy_read(phydev, MII_ADVERTISE);
1208		if (ladv < 0)
1209			return ladv;
1210
1211		radv = phy_read(phydev, MII_LPA);
1212		if (radv < 0)
1213			return radv;
1214
1215		netif_dbg(dev, link, dev->net,
1216			  "speed: %u duplex: %d anadv: 0x%04x anlpa: 0x%04x",
1217			  ecmd.base.speed, ecmd.base.duplex, ladv, radv);
1218
1219		ret = lan78xx_update_flowcontrol(dev, ecmd.base.duplex, ladv,
1220						 radv);
 
 
1221
1222		if (!timer_pending(&dev->stat_monitor)) {
1223			dev->delta = 1;
1224			mod_timer(&dev->stat_monitor,
1225				  jiffies + STAT_UPDATE_TIMER);
1226		}
1227
1228		tasklet_schedule(&dev->bh);
 
 
1229	}
1230
1231	return ret;
1232}
1233
1234/* some work can't be done in tasklets, so we use keventd
1235 *
1236 * NOTE:  annoying asymmetry:  if it's active, schedule_work() fails,
1237 * but tasklet_schedule() doesn't.	hope the failure is rare.
1238 */
1239static void lan78xx_defer_kevent(struct lan78xx_net *dev, int work)
1240{
1241	set_bit(work, &dev->flags);
1242	if (!schedule_delayed_work(&dev->wq, 0))
1243		netdev_err(dev->net, "kevent %d may have been dropped\n", work);
1244}
1245
1246static void lan78xx_status(struct lan78xx_net *dev, struct urb *urb)
1247{
1248	u32 intdata;
1249
1250	if (urb->actual_length != 4) {
1251		netdev_warn(dev->net,
1252			    "unexpected urb length %d", urb->actual_length);
1253		return;
1254	}
1255
1256	intdata = get_unaligned_le32(urb->transfer_buffer);
1257
1258	if (intdata & INT_ENP_PHY_INT) {
1259		netif_dbg(dev, link, dev->net, "PHY INTR: 0x%08x\n", intdata);
1260		lan78xx_defer_kevent(dev, EVENT_LINK_RESET);
1261
1262		if (dev->domain_data.phyirq > 0) {
1263			local_irq_disable();
1264			generic_handle_irq(dev->domain_data.phyirq);
1265			local_irq_enable();
1266		}
1267	} else
1268		netdev_warn(dev->net,
1269			    "unexpected interrupt: 0x%08x\n", intdata);
 
1270}
1271
1272static int lan78xx_ethtool_get_eeprom_len(struct net_device *netdev)
1273{
1274	return MAX_EEPROM_SIZE;
1275}
1276
1277static int lan78xx_ethtool_get_eeprom(struct net_device *netdev,
1278				      struct ethtool_eeprom *ee, u8 *data)
1279{
1280	struct lan78xx_net *dev = netdev_priv(netdev);
1281	int ret;
1282
1283	ret = usb_autopm_get_interface(dev->intf);
1284	if (ret)
1285		return ret;
1286
1287	ee->magic = LAN78XX_EEPROM_MAGIC;
1288
1289	ret = lan78xx_read_raw_eeprom(dev, ee->offset, ee->len, data);
1290
1291	usb_autopm_put_interface(dev->intf);
1292
1293	return ret;
1294}
1295
1296static int lan78xx_ethtool_set_eeprom(struct net_device *netdev,
1297				      struct ethtool_eeprom *ee, u8 *data)
1298{
1299	struct lan78xx_net *dev = netdev_priv(netdev);
1300	int ret;
1301
1302	ret = usb_autopm_get_interface(dev->intf);
1303	if (ret)
1304		return ret;
1305
1306	/* Invalid EEPROM_INDICATOR at offset zero will result in a failure
1307	 * to load data from EEPROM
1308	 */
1309	if (ee->magic == LAN78XX_EEPROM_MAGIC)
1310		ret = lan78xx_write_raw_eeprom(dev, ee->offset, ee->len, data);
1311	else if ((ee->magic == LAN78XX_OTP_MAGIC) &&
1312		 (ee->offset == 0) &&
1313		 (ee->len == 512) &&
1314		 (data[0] == OTP_INDICATOR_1))
1315		ret = lan78xx_write_raw_otp(dev, ee->offset, ee->len, data);
1316
1317	usb_autopm_put_interface(dev->intf);
1318
1319	return ret;
1320}
1321
1322static void lan78xx_get_strings(struct net_device *netdev, u32 stringset,
1323				u8 *data)
1324{
1325	if (stringset == ETH_SS_STATS)
1326		memcpy(data, lan78xx_gstrings, sizeof(lan78xx_gstrings));
1327}
1328
1329static int lan78xx_get_sset_count(struct net_device *netdev, int sset)
1330{
1331	if (sset == ETH_SS_STATS)
1332		return ARRAY_SIZE(lan78xx_gstrings);
1333	else
1334		return -EOPNOTSUPP;
1335}
1336
1337static void lan78xx_get_stats(struct net_device *netdev,
1338			      struct ethtool_stats *stats, u64 *data)
1339{
1340	struct lan78xx_net *dev = netdev_priv(netdev);
1341
1342	lan78xx_update_stats(dev);
1343
1344	mutex_lock(&dev->stats.access_lock);
1345	memcpy(data, &dev->stats.curr_stat, sizeof(dev->stats.curr_stat));
1346	mutex_unlock(&dev->stats.access_lock);
1347}
1348
1349static void lan78xx_get_wol(struct net_device *netdev,
1350			    struct ethtool_wolinfo *wol)
1351{
1352	struct lan78xx_net *dev = netdev_priv(netdev);
1353	int ret;
1354	u32 buf;
1355	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1356
1357	if (usb_autopm_get_interface(dev->intf) < 0)
1358			return;
1359
1360	ret = lan78xx_read_reg(dev, USB_CFG0, &buf);
1361	if (unlikely(ret < 0)) {
1362		wol->supported = 0;
1363		wol->wolopts = 0;
1364	} else {
1365		if (buf & USB_CFG_RMT_WKP_) {
1366			wol->supported = WAKE_ALL;
1367			wol->wolopts = pdata->wol;
1368		} else {
1369			wol->supported = 0;
1370			wol->wolopts = 0;
1371		}
1372	}
1373
1374	usb_autopm_put_interface(dev->intf);
1375}
1376
1377static int lan78xx_set_wol(struct net_device *netdev,
1378			   struct ethtool_wolinfo *wol)
1379{
1380	struct lan78xx_net *dev = netdev_priv(netdev);
1381	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1382	int ret;
1383
1384	ret = usb_autopm_get_interface(dev->intf);
1385	if (ret < 0)
1386		return ret;
1387
1388	if (wol->wolopts & ~WAKE_ALL)
1389		return -EINVAL;
1390
1391	pdata->wol = wol->wolopts;
1392
1393	device_set_wakeup_enable(&dev->udev->dev, (bool)wol->wolopts);
1394
1395	phy_ethtool_set_wol(netdev->phydev, wol);
1396
1397	usb_autopm_put_interface(dev->intf);
1398
1399	return ret;
1400}
1401
1402static int lan78xx_get_eee(struct net_device *net, struct ethtool_eee *edata)
1403{
1404	struct lan78xx_net *dev = netdev_priv(net);
1405	struct phy_device *phydev = net->phydev;
1406	int ret;
1407	u32 buf;
1408
1409	ret = usb_autopm_get_interface(dev->intf);
1410	if (ret < 0)
1411		return ret;
1412
1413	ret = phy_ethtool_get_eee(phydev, edata);
1414	if (ret < 0)
1415		goto exit;
1416
1417	ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1418	if (buf & MAC_CR_EEE_EN_) {
1419		edata->eee_enabled = true;
1420		edata->eee_active = !!(edata->advertised &
1421				       edata->lp_advertised);
1422		edata->tx_lpi_enabled = true;
1423		/* EEE_TX_LPI_REQ_DLY & tx_lpi_timer are same uSec unit */
1424		ret = lan78xx_read_reg(dev, EEE_TX_LPI_REQ_DLY, &buf);
1425		edata->tx_lpi_timer = buf;
1426	} else {
1427		edata->eee_enabled = false;
1428		edata->eee_active = false;
1429		edata->tx_lpi_enabled = false;
1430		edata->tx_lpi_timer = 0;
1431	}
1432
1433	ret = 0;
1434exit:
1435	usb_autopm_put_interface(dev->intf);
1436
1437	return ret;
1438}
1439
1440static int lan78xx_set_eee(struct net_device *net, struct ethtool_eee *edata)
1441{
1442	struct lan78xx_net *dev = netdev_priv(net);
1443	int ret;
1444	u32 buf;
1445
1446	ret = usb_autopm_get_interface(dev->intf);
1447	if (ret < 0)
1448		return ret;
1449
1450	if (edata->eee_enabled) {
1451		ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1452		buf |= MAC_CR_EEE_EN_;
1453		ret = lan78xx_write_reg(dev, MAC_CR, buf);
1454
1455		phy_ethtool_set_eee(net->phydev, edata);
1456
1457		buf = (u32)edata->tx_lpi_timer;
1458		ret = lan78xx_write_reg(dev, EEE_TX_LPI_REQ_DLY, buf);
1459	} else {
1460		ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1461		buf &= ~MAC_CR_EEE_EN_;
1462		ret = lan78xx_write_reg(dev, MAC_CR, buf);
1463	}
1464
1465	usb_autopm_put_interface(dev->intf);
1466
1467	return 0;
1468}
1469
1470static u32 lan78xx_get_link(struct net_device *net)
1471{
1472	u32 link;
1473
1474	mutex_lock(&net->phydev->lock);
1475	phy_read_status(net->phydev);
1476	link = net->phydev->link;
1477	mutex_unlock(&net->phydev->lock);
1478
1479	return link;
1480}
1481
1482static void lan78xx_get_drvinfo(struct net_device *net,
1483				struct ethtool_drvinfo *info)
1484{
1485	struct lan78xx_net *dev = netdev_priv(net);
1486
1487	strncpy(info->driver, DRIVER_NAME, sizeof(info->driver));
1488	usb_make_path(dev->udev, info->bus_info, sizeof(info->bus_info));
1489}
1490
1491static u32 lan78xx_get_msglevel(struct net_device *net)
1492{
1493	struct lan78xx_net *dev = netdev_priv(net);
1494
1495	return dev->msg_enable;
1496}
1497
1498static void lan78xx_set_msglevel(struct net_device *net, u32 level)
1499{
1500	struct lan78xx_net *dev = netdev_priv(net);
1501
1502	dev->msg_enable = level;
1503}
1504
1505static int lan78xx_get_link_ksettings(struct net_device *net,
1506				      struct ethtool_link_ksettings *cmd)
1507{
1508	struct lan78xx_net *dev = netdev_priv(net);
1509	struct phy_device *phydev = net->phydev;
1510	int ret;
1511
1512	ret = usb_autopm_get_interface(dev->intf);
1513	if (ret < 0)
1514		return ret;
1515
1516	phy_ethtool_ksettings_get(phydev, cmd);
1517
1518	usb_autopm_put_interface(dev->intf);
1519
1520	return ret;
1521}
1522
1523static int lan78xx_set_link_ksettings(struct net_device *net,
1524				      const struct ethtool_link_ksettings *cmd)
1525{
1526	struct lan78xx_net *dev = netdev_priv(net);
1527	struct phy_device *phydev = net->phydev;
1528	int ret = 0;
1529	int temp;
1530
1531	ret = usb_autopm_get_interface(dev->intf);
1532	if (ret < 0)
1533		return ret;
1534
1535	/* change speed & duplex */
1536	ret = phy_ethtool_ksettings_set(phydev, cmd);
1537
1538	if (!cmd->base.autoneg) {
1539		/* force link down */
1540		temp = phy_read(phydev, MII_BMCR);
1541		phy_write(phydev, MII_BMCR, temp | BMCR_LOOPBACK);
1542		mdelay(1);
1543		phy_write(phydev, MII_BMCR, temp);
1544	}
1545
1546	usb_autopm_put_interface(dev->intf);
1547
1548	return ret;
1549}
1550
1551static void lan78xx_get_pause(struct net_device *net,
1552			      struct ethtool_pauseparam *pause)
1553{
1554	struct lan78xx_net *dev = netdev_priv(net);
1555	struct phy_device *phydev = net->phydev;
1556	struct ethtool_link_ksettings ecmd;
1557
1558	phy_ethtool_ksettings_get(phydev, &ecmd);
1559
1560	pause->autoneg = dev->fc_autoneg;
1561
1562	if (dev->fc_request_control & FLOW_CTRL_TX)
1563		pause->tx_pause = 1;
1564
1565	if (dev->fc_request_control & FLOW_CTRL_RX)
1566		pause->rx_pause = 1;
1567}
1568
1569static int lan78xx_set_pause(struct net_device *net,
1570			     struct ethtool_pauseparam *pause)
1571{
1572	struct lan78xx_net *dev = netdev_priv(net);
1573	struct phy_device *phydev = net->phydev;
1574	struct ethtool_link_ksettings ecmd;
1575	int ret;
1576
1577	phy_ethtool_ksettings_get(phydev, &ecmd);
1578
1579	if (pause->autoneg && !ecmd.base.autoneg) {
1580		ret = -EINVAL;
1581		goto exit;
1582	}
1583
1584	dev->fc_request_control = 0;
1585	if (pause->rx_pause)
1586		dev->fc_request_control |= FLOW_CTRL_RX;
1587
1588	if (pause->tx_pause)
1589		dev->fc_request_control |= FLOW_CTRL_TX;
1590
1591	if (ecmd.base.autoneg) {
1592		__ETHTOOL_DECLARE_LINK_MODE_MASK(fc) = { 0, };
1593		u32 mii_adv;
1594
1595		linkmode_clear_bit(ETHTOOL_LINK_MODE_Pause_BIT,
1596				   ecmd.link_modes.advertising);
1597		linkmode_clear_bit(ETHTOOL_LINK_MODE_Asym_Pause_BIT,
1598				   ecmd.link_modes.advertising);
1599		mii_adv = (u32)mii_advertise_flowctrl(dev->fc_request_control);
1600		mii_adv_to_linkmode_adv_t(fc, mii_adv);
1601		linkmode_or(ecmd.link_modes.advertising, fc,
1602			    ecmd.link_modes.advertising);
1603
1604		phy_ethtool_ksettings_set(phydev, &ecmd);
1605	}
1606
1607	dev->fc_autoneg = pause->autoneg;
1608
1609	ret = 0;
1610exit:
1611	return ret;
1612}
1613
1614static int lan78xx_get_regs_len(struct net_device *netdev)
1615{
1616	if (!netdev->phydev)
1617		return (sizeof(lan78xx_regs));
1618	else
1619		return (sizeof(lan78xx_regs) + PHY_REG_SIZE);
1620}
1621
1622static void
1623lan78xx_get_regs(struct net_device *netdev, struct ethtool_regs *regs,
1624		 void *buf)
1625{
1626	u32 *data = buf;
1627	int i, j;
1628	struct lan78xx_net *dev = netdev_priv(netdev);
1629
1630	/* Read Device/MAC registers */
1631	for (i = 0; i < ARRAY_SIZE(lan78xx_regs); i++)
1632		lan78xx_read_reg(dev, lan78xx_regs[i], &data[i]);
1633
1634	if (!netdev->phydev)
1635		return;
1636
1637	/* Read PHY registers */
1638	for (j = 0; j < 32; i++, j++)
1639		data[i] = phy_read(netdev->phydev, j);
1640}
1641
1642static const struct ethtool_ops lan78xx_ethtool_ops = {
1643	.get_link	= lan78xx_get_link,
1644	.nway_reset	= phy_ethtool_nway_reset,
1645	.get_drvinfo	= lan78xx_get_drvinfo,
1646	.get_msglevel	= lan78xx_get_msglevel,
1647	.set_msglevel	= lan78xx_set_msglevel,
1648	.get_eeprom_len = lan78xx_ethtool_get_eeprom_len,
1649	.get_eeprom	= lan78xx_ethtool_get_eeprom,
1650	.set_eeprom	= lan78xx_ethtool_set_eeprom,
1651	.get_ethtool_stats = lan78xx_get_stats,
1652	.get_sset_count = lan78xx_get_sset_count,
1653	.get_strings	= lan78xx_get_strings,
1654	.get_wol	= lan78xx_get_wol,
1655	.set_wol	= lan78xx_set_wol,
1656	.get_ts_info	= ethtool_op_get_ts_info,
1657	.get_eee	= lan78xx_get_eee,
1658	.set_eee	= lan78xx_set_eee,
1659	.get_pauseparam	= lan78xx_get_pause,
1660	.set_pauseparam	= lan78xx_set_pause,
1661	.get_link_ksettings = lan78xx_get_link_ksettings,
1662	.set_link_ksettings = lan78xx_set_link_ksettings,
1663	.get_regs_len	= lan78xx_get_regs_len,
1664	.get_regs	= lan78xx_get_regs,
1665};
1666
1667static void lan78xx_init_mac_address(struct lan78xx_net *dev)
1668{
1669	u32 addr_lo, addr_hi;
1670	u8 addr[6];
1671
1672	lan78xx_read_reg(dev, RX_ADDRL, &addr_lo);
1673	lan78xx_read_reg(dev, RX_ADDRH, &addr_hi);
1674
1675	addr[0] = addr_lo & 0xFF;
1676	addr[1] = (addr_lo >> 8) & 0xFF;
1677	addr[2] = (addr_lo >> 16) & 0xFF;
1678	addr[3] = (addr_lo >> 24) & 0xFF;
1679	addr[4] = addr_hi & 0xFF;
1680	addr[5] = (addr_hi >> 8) & 0xFF;
1681
1682	if (!is_valid_ether_addr(addr)) {
1683		if (!eth_platform_get_mac_address(&dev->udev->dev, addr)) {
1684			/* valid address present in Device Tree */
1685			netif_dbg(dev, ifup, dev->net,
1686				  "MAC address read from Device Tree");
1687		} else if (((lan78xx_read_eeprom(dev, EEPROM_MAC_OFFSET,
1688						 ETH_ALEN, addr) == 0) ||
1689			    (lan78xx_read_otp(dev, EEPROM_MAC_OFFSET,
1690					      ETH_ALEN, addr) == 0)) &&
1691			   is_valid_ether_addr(addr)) {
1692			/* eeprom values are valid so use them */
1693			netif_dbg(dev, ifup, dev->net,
1694				  "MAC address read from EEPROM");
1695		} else {
1696			/* generate random MAC */
1697			eth_random_addr(addr);
1698			netif_dbg(dev, ifup, dev->net,
1699				  "MAC address set to random addr");
1700		}
1701
1702		addr_lo = addr[0] | (addr[1] << 8) |
1703			  (addr[2] << 16) | (addr[3] << 24);
1704		addr_hi = addr[4] | (addr[5] << 8);
1705
1706		lan78xx_write_reg(dev, RX_ADDRL, addr_lo);
1707		lan78xx_write_reg(dev, RX_ADDRH, addr_hi);
1708	}
1709
1710	lan78xx_write_reg(dev, MAF_LO(0), addr_lo);
1711	lan78xx_write_reg(dev, MAF_HI(0), addr_hi | MAF_HI_VALID_);
1712
1713	ether_addr_copy(dev->net->dev_addr, addr);
1714}
1715
1716/* MDIO read and write wrappers for phylib */
1717static int lan78xx_mdiobus_read(struct mii_bus *bus, int phy_id, int idx)
1718{
1719	struct lan78xx_net *dev = bus->priv;
1720	u32 val, addr;
1721	int ret;
1722
1723	ret = usb_autopm_get_interface(dev->intf);
1724	if (ret < 0)
1725		return ret;
1726
1727	mutex_lock(&dev->phy_mutex);
1728
1729	/* confirm MII not busy */
1730	ret = lan78xx_phy_wait_not_busy(dev);
1731	if (ret < 0)
1732		goto done;
1733
1734	/* set the address, index & direction (read from PHY) */
1735	addr = mii_access(phy_id, idx, MII_READ);
1736	ret = lan78xx_write_reg(dev, MII_ACC, addr);
1737
1738	ret = lan78xx_phy_wait_not_busy(dev);
1739	if (ret < 0)
1740		goto done;
1741
1742	ret = lan78xx_read_reg(dev, MII_DATA, &val);
1743
1744	ret = (int)(val & 0xFFFF);
1745
1746done:
1747	mutex_unlock(&dev->phy_mutex);
1748	usb_autopm_put_interface(dev->intf);
1749
1750	return ret;
1751}
1752
1753static int lan78xx_mdiobus_write(struct mii_bus *bus, int phy_id, int idx,
1754				 u16 regval)
1755{
1756	struct lan78xx_net *dev = bus->priv;
1757	u32 val, addr;
1758	int ret;
1759
1760	ret = usb_autopm_get_interface(dev->intf);
1761	if (ret < 0)
1762		return ret;
1763
1764	mutex_lock(&dev->phy_mutex);
1765
1766	/* confirm MII not busy */
1767	ret = lan78xx_phy_wait_not_busy(dev);
1768	if (ret < 0)
1769		goto done;
1770
1771	val = (u32)regval;
1772	ret = lan78xx_write_reg(dev, MII_DATA, val);
1773
1774	/* set the address, index & direction (write to PHY) */
1775	addr = mii_access(phy_id, idx, MII_WRITE);
1776	ret = lan78xx_write_reg(dev, MII_ACC, addr);
1777
1778	ret = lan78xx_phy_wait_not_busy(dev);
1779	if (ret < 0)
1780		goto done;
1781
1782done:
1783	mutex_unlock(&dev->phy_mutex);
1784	usb_autopm_put_interface(dev->intf);
1785	return 0;
1786}
1787
1788static int lan78xx_mdio_init(struct lan78xx_net *dev)
1789{
1790	struct device_node *node;
1791	int ret;
1792
1793	dev->mdiobus = mdiobus_alloc();
1794	if (!dev->mdiobus) {
1795		netdev_err(dev->net, "can't allocate MDIO bus\n");
1796		return -ENOMEM;
1797	}
1798
1799	dev->mdiobus->priv = (void *)dev;
1800	dev->mdiobus->read = lan78xx_mdiobus_read;
1801	dev->mdiobus->write = lan78xx_mdiobus_write;
1802	dev->mdiobus->name = "lan78xx-mdiobus";
1803	dev->mdiobus->parent = &dev->udev->dev;
1804
1805	snprintf(dev->mdiobus->id, MII_BUS_ID_SIZE, "usb-%03d:%03d",
1806		 dev->udev->bus->busnum, dev->udev->devnum);
1807
1808	switch (dev->chipid) {
1809	case ID_REV_CHIP_ID_7800_:
1810	case ID_REV_CHIP_ID_7850_:
1811		/* set to internal PHY id */
1812		dev->mdiobus->phy_mask = ~(1 << 1);
1813		break;
1814	case ID_REV_CHIP_ID_7801_:
1815		/* scan thru PHYAD[2..0] */
1816		dev->mdiobus->phy_mask = ~(0xFF);
1817		break;
1818	}
1819
1820	node = of_get_child_by_name(dev->udev->dev.of_node, "mdio");
1821	ret = of_mdiobus_register(dev->mdiobus, node);
1822	of_node_put(node);
1823	if (ret) {
1824		netdev_err(dev->net, "can't register MDIO bus\n");
1825		goto exit1;
1826	}
1827
1828	netdev_dbg(dev->net, "registered mdiobus bus %s\n", dev->mdiobus->id);
1829	return 0;
1830exit1:
1831	mdiobus_free(dev->mdiobus);
1832	return ret;
1833}
1834
1835static void lan78xx_remove_mdio(struct lan78xx_net *dev)
1836{
1837	mdiobus_unregister(dev->mdiobus);
1838	mdiobus_free(dev->mdiobus);
1839}
1840
1841static void lan78xx_link_status_change(struct net_device *net)
1842{
1843	struct phy_device *phydev = net->phydev;
1844	int temp;
1845
1846	/* At forced 100 F/H mode, chip may fail to set mode correctly
1847	 * when cable is switched between long(~50+m) and short one.
1848	 * As workaround, set to 10 before setting to 100
1849	 * at forced 100 F/H mode.
1850	 */
1851	if (!phydev->autoneg && (phydev->speed == 100)) {
1852		/* disable phy interrupt */
1853		temp = phy_read(phydev, LAN88XX_INT_MASK);
1854		temp &= ~LAN88XX_INT_MASK_MDINTPIN_EN_;
1855		phy_write(phydev, LAN88XX_INT_MASK, temp);
1856
1857		temp = phy_read(phydev, MII_BMCR);
1858		temp &= ~(BMCR_SPEED100 | BMCR_SPEED1000);
1859		phy_write(phydev, MII_BMCR, temp); /* set to 10 first */
1860		temp |= BMCR_SPEED100;
1861		phy_write(phydev, MII_BMCR, temp); /* set to 100 later */
1862
1863		/* clear pending interrupt generated while workaround */
1864		temp = phy_read(phydev, LAN88XX_INT_STS);
1865
1866		/* enable phy interrupt back */
1867		temp = phy_read(phydev, LAN88XX_INT_MASK);
1868		temp |= LAN88XX_INT_MASK_MDINTPIN_EN_;
1869		phy_write(phydev, LAN88XX_INT_MASK, temp);
1870	}
1871}
1872
1873static int irq_map(struct irq_domain *d, unsigned int irq,
1874		   irq_hw_number_t hwirq)
1875{
1876	struct irq_domain_data *data = d->host_data;
1877
1878	irq_set_chip_data(irq, data);
1879	irq_set_chip_and_handler(irq, data->irqchip, data->irq_handler);
1880	irq_set_noprobe(irq);
1881
1882	return 0;
1883}
1884
1885static void irq_unmap(struct irq_domain *d, unsigned int irq)
1886{
1887	irq_set_chip_and_handler(irq, NULL, NULL);
1888	irq_set_chip_data(irq, NULL);
1889}
1890
1891static const struct irq_domain_ops chip_domain_ops = {
1892	.map	= irq_map,
1893	.unmap	= irq_unmap,
1894};
1895
1896static void lan78xx_irq_mask(struct irq_data *irqd)
1897{
1898	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
1899
1900	data->irqenable &= ~BIT(irqd_to_hwirq(irqd));
1901}
1902
1903static void lan78xx_irq_unmask(struct irq_data *irqd)
1904{
1905	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
1906
1907	data->irqenable |= BIT(irqd_to_hwirq(irqd));
1908}
1909
1910static void lan78xx_irq_bus_lock(struct irq_data *irqd)
1911{
1912	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
1913
1914	mutex_lock(&data->irq_lock);
1915}
1916
1917static void lan78xx_irq_bus_sync_unlock(struct irq_data *irqd)
1918{
1919	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
1920	struct lan78xx_net *dev =
1921			container_of(data, struct lan78xx_net, domain_data);
1922	u32 buf;
1923
1924	/* call register access here because irq_bus_lock & irq_bus_sync_unlock
1925	 * are only two callbacks executed in non-atomic contex.
1926	 */
1927	lan78xx_read_reg(dev, INT_EP_CTL, &buf);
1928	if (buf != data->irqenable)
1929		lan78xx_write_reg(dev, INT_EP_CTL, data->irqenable);
1930
1931	mutex_unlock(&data->irq_lock);
1932}
1933
1934static struct irq_chip lan78xx_irqchip = {
1935	.name			= "lan78xx-irqs",
1936	.irq_mask		= lan78xx_irq_mask,
1937	.irq_unmask		= lan78xx_irq_unmask,
1938	.irq_bus_lock		= lan78xx_irq_bus_lock,
1939	.irq_bus_sync_unlock	= lan78xx_irq_bus_sync_unlock,
1940};
1941
1942static int lan78xx_setup_irq_domain(struct lan78xx_net *dev)
1943{
1944	struct device_node *of_node;
1945	struct irq_domain *irqdomain;
1946	unsigned int irqmap = 0;
1947	u32 buf;
1948	int ret = 0;
1949
1950	of_node = dev->udev->dev.parent->of_node;
1951
1952	mutex_init(&dev->domain_data.irq_lock);
1953
1954	lan78xx_read_reg(dev, INT_EP_CTL, &buf);
1955	dev->domain_data.irqenable = buf;
1956
1957	dev->domain_data.irqchip = &lan78xx_irqchip;
1958	dev->domain_data.irq_handler = handle_simple_irq;
1959
1960	irqdomain = irq_domain_add_simple(of_node, MAX_INT_EP, 0,
1961					  &chip_domain_ops, &dev->domain_data);
1962	if (irqdomain) {
1963		/* create mapping for PHY interrupt */
1964		irqmap = irq_create_mapping(irqdomain, INT_EP_PHY);
1965		if (!irqmap) {
1966			irq_domain_remove(irqdomain);
1967
1968			irqdomain = NULL;
1969			ret = -EINVAL;
1970		}
1971	} else {
1972		ret = -EINVAL;
1973	}
1974
1975	dev->domain_data.irqdomain = irqdomain;
1976	dev->domain_data.phyirq = irqmap;
1977
1978	return ret;
1979}
1980
1981static void lan78xx_remove_irq_domain(struct lan78xx_net *dev)
1982{
1983	if (dev->domain_data.phyirq > 0) {
1984		irq_dispose_mapping(dev->domain_data.phyirq);
1985
1986		if (dev->domain_data.irqdomain)
1987			irq_domain_remove(dev->domain_data.irqdomain);
1988	}
1989	dev->domain_data.phyirq = 0;
1990	dev->domain_data.irqdomain = NULL;
1991}
1992
1993static int lan8835_fixup(struct phy_device *phydev)
1994{
1995	int buf;
1996	struct lan78xx_net *dev = netdev_priv(phydev->attached_dev);
1997
1998	/* LED2/PME_N/IRQ_N/RGMII_ID pin to IRQ_N mode */
1999	buf = phy_read_mmd(phydev, MDIO_MMD_PCS, 0x8010);
2000	buf &= ~0x1800;
2001	buf |= 0x0800;
2002	phy_write_mmd(phydev, MDIO_MMD_PCS, 0x8010, buf);
2003
2004	/* RGMII MAC TXC Delay Enable */
2005	lan78xx_write_reg(dev, MAC_RGMII_ID,
2006				MAC_RGMII_ID_TXC_DELAY_EN_);
2007
2008	/* RGMII TX DLL Tune Adjust */
2009	lan78xx_write_reg(dev, RGMII_TX_BYP_DLL, 0x3D00);
2010
2011	dev->interface = PHY_INTERFACE_MODE_RGMII_TXID;
2012
2013	return 1;
2014}
2015
2016static int ksz9031rnx_fixup(struct phy_device *phydev)
2017{
2018	struct lan78xx_net *dev = netdev_priv(phydev->attached_dev);
2019
2020	/* Micrel9301RNX PHY configuration */
2021	/* RGMII Control Signal Pad Skew */
2022	phy_write_mmd(phydev, MDIO_MMD_WIS, 4, 0x0077);
2023	/* RGMII RX Data Pad Skew */
2024	phy_write_mmd(phydev, MDIO_MMD_WIS, 5, 0x7777);
2025	/* RGMII RX Clock Pad Skew */
2026	phy_write_mmd(phydev, MDIO_MMD_WIS, 8, 0x1FF);
2027
2028	dev->interface = PHY_INTERFACE_MODE_RGMII_RXID;
2029
2030	return 1;
2031}
2032
2033static struct phy_device *lan7801_phy_init(struct lan78xx_net *dev)
2034{
2035	u32 buf;
2036	int ret;
2037	struct fixed_phy_status fphy_status = {
2038		.link = 1,
2039		.speed = SPEED_1000,
2040		.duplex = DUPLEX_FULL,
2041	};
2042	struct phy_device *phydev;
2043
2044	phydev = phy_find_first(dev->mdiobus);
2045	if (!phydev) {
2046		netdev_dbg(dev->net, "PHY Not Found!! Registering Fixed PHY\n");
2047		phydev = fixed_phy_register(PHY_POLL, &fphy_status, NULL);
2048		if (IS_ERR(phydev)) {
2049			netdev_err(dev->net, "No PHY/fixed_PHY found\n");
2050			return NULL;
2051		}
2052		netdev_dbg(dev->net, "Registered FIXED PHY\n");
2053		dev->interface = PHY_INTERFACE_MODE_RGMII;
2054		ret = lan78xx_write_reg(dev, MAC_RGMII_ID,
2055					MAC_RGMII_ID_TXC_DELAY_EN_);
2056		ret = lan78xx_write_reg(dev, RGMII_TX_BYP_DLL, 0x3D00);
2057		ret = lan78xx_read_reg(dev, HW_CFG, &buf);
2058		buf |= HW_CFG_CLK125_EN_;
2059		buf |= HW_CFG_REFCLK25_EN_;
2060		ret = lan78xx_write_reg(dev, HW_CFG, buf);
2061	} else {
2062		if (!phydev->drv) {
2063			netdev_err(dev->net, "no PHY driver found\n");
2064			return NULL;
2065		}
2066		dev->interface = PHY_INTERFACE_MODE_RGMII;
2067		/* external PHY fixup for KSZ9031RNX */
2068		ret = phy_register_fixup_for_uid(PHY_KSZ9031RNX, 0xfffffff0,
2069						 ksz9031rnx_fixup);
2070		if (ret < 0) {
2071			netdev_err(dev->net, "Failed to register fixup for PHY_KSZ9031RNX\n");
2072			return NULL;
2073		}
2074		/* external PHY fixup for LAN8835 */
2075		ret = phy_register_fixup_for_uid(PHY_LAN8835, 0xfffffff0,
2076						 lan8835_fixup);
2077		if (ret < 0) {
2078			netdev_err(dev->net, "Failed to register fixup for PHY_LAN8835\n");
2079			return NULL;
2080		}
2081		/* add more external PHY fixup here if needed */
2082
2083		phydev->is_internal = false;
2084	}
2085	return phydev;
2086}
2087
2088static int lan78xx_phy_init(struct lan78xx_net *dev)
2089{
2090	__ETHTOOL_DECLARE_LINK_MODE_MASK(fc) = { 0, };
2091	int ret;
2092	u32 mii_adv;
2093	struct phy_device *phydev;
2094
2095	switch (dev->chipid) {
2096	case ID_REV_CHIP_ID_7801_:
2097		phydev = lan7801_phy_init(dev);
2098		if (!phydev) {
2099			netdev_err(dev->net, "lan7801: PHY Init Failed");
2100			return -EIO;
2101		}
2102		break;
2103
2104	case ID_REV_CHIP_ID_7800_:
2105	case ID_REV_CHIP_ID_7850_:
2106		phydev = phy_find_first(dev->mdiobus);
2107		if (!phydev) {
2108			netdev_err(dev->net, "no PHY found\n");
2109			return -EIO;
2110		}
2111		phydev->is_internal = true;
2112		dev->interface = PHY_INTERFACE_MODE_GMII;
2113		break;
2114
2115	default:
2116		netdev_err(dev->net, "Unknown CHIP ID found\n");
2117		return -EIO;
2118	}
2119
2120	/* if phyirq is not set, use polling mode in phylib */
2121	if (dev->domain_data.phyirq > 0)
2122		phydev->irq = dev->domain_data.phyirq;
2123	else
2124		phydev->irq = 0;
2125	netdev_dbg(dev->net, "phydev->irq = %d\n", phydev->irq);
2126
2127	/* set to AUTOMDIX */
2128	phydev->mdix = ETH_TP_MDI_AUTO;
2129
2130	ret = phy_connect_direct(dev->net, phydev,
2131				 lan78xx_link_status_change,
2132				 dev->interface);
2133	if (ret) {
2134		netdev_err(dev->net, "can't attach PHY to %s\n",
2135			   dev->mdiobus->id);
2136		if (dev->chipid == ID_REV_CHIP_ID_7801_) {
2137			if (phy_is_pseudo_fixed_link(phydev)) {
2138				fixed_phy_unregister(phydev);
2139			} else {
2140				phy_unregister_fixup_for_uid(PHY_KSZ9031RNX,
2141							     0xfffffff0);
2142				phy_unregister_fixup_for_uid(PHY_LAN8835,
2143							     0xfffffff0);
2144			}
2145		}
2146		return -EIO;
2147	}
2148
2149	/* MAC doesn't support 1000T Half */
2150	phy_remove_link_mode(phydev, ETHTOOL_LINK_MODE_1000baseT_Half_BIT);
2151
2152	/* support both flow controls */
2153	dev->fc_request_control = (FLOW_CTRL_RX | FLOW_CTRL_TX);
2154	linkmode_clear_bit(ETHTOOL_LINK_MODE_Pause_BIT,
2155			   phydev->advertising);
2156	linkmode_clear_bit(ETHTOOL_LINK_MODE_Asym_Pause_BIT,
2157			   phydev->advertising);
2158	mii_adv = (u32)mii_advertise_flowctrl(dev->fc_request_control);
2159	mii_adv_to_linkmode_adv_t(fc, mii_adv);
2160	linkmode_or(phydev->advertising, fc, phydev->advertising);
2161
2162	if (phydev->mdio.dev.of_node) {
2163		u32 reg;
2164		int len;
2165
2166		len = of_property_count_elems_of_size(phydev->mdio.dev.of_node,
2167						      "microchip,led-modes",
2168						      sizeof(u32));
2169		if (len >= 0) {
2170			/* Ensure the appropriate LEDs are enabled */
2171			lan78xx_read_reg(dev, HW_CFG, &reg);
2172			reg &= ~(HW_CFG_LED0_EN_ |
2173				 HW_CFG_LED1_EN_ |
2174				 HW_CFG_LED2_EN_ |
2175				 HW_CFG_LED3_EN_);
2176			reg |= (len > 0) * HW_CFG_LED0_EN_ |
2177				(len > 1) * HW_CFG_LED1_EN_ |
2178				(len > 2) * HW_CFG_LED2_EN_ |
2179				(len > 3) * HW_CFG_LED3_EN_;
2180			lan78xx_write_reg(dev, HW_CFG, reg);
2181		}
2182	}
2183
2184	genphy_config_aneg(phydev);
2185
2186	dev->fc_autoneg = phydev->autoneg;
2187
2188	return 0;
2189}
2190
2191static int lan78xx_set_rx_max_frame_length(struct lan78xx_net *dev, int size)
2192{
2193	u32 buf;
2194	bool rxenabled;
2195
2196	lan78xx_read_reg(dev, MAC_RX, &buf);
2197
2198	rxenabled = ((buf & MAC_RX_RXEN_) != 0);
2199
2200	if (rxenabled) {
2201		buf &= ~MAC_RX_RXEN_;
2202		lan78xx_write_reg(dev, MAC_RX, buf);
2203	}
2204
2205	/* add 4 to size for FCS */
2206	buf &= ~MAC_RX_MAX_SIZE_MASK_;
2207	buf |= (((size + 4) << MAC_RX_MAX_SIZE_SHIFT_) & MAC_RX_MAX_SIZE_MASK_);
2208
2209	lan78xx_write_reg(dev, MAC_RX, buf);
2210
2211	if (rxenabled) {
2212		buf |= MAC_RX_RXEN_;
2213		lan78xx_write_reg(dev, MAC_RX, buf);
2214	}
2215
2216	return 0;
2217}
2218
2219static int unlink_urbs(struct lan78xx_net *dev, struct sk_buff_head *q)
2220{
2221	struct sk_buff *skb;
2222	unsigned long flags;
2223	int count = 0;
2224
2225	spin_lock_irqsave(&q->lock, flags);
2226	while (!skb_queue_empty(q)) {
2227		struct skb_data	*entry;
2228		struct urb *urb;
2229		int ret;
2230
2231		skb_queue_walk(q, skb) {
2232			entry = (struct skb_data *)skb->cb;
2233			if (entry->state != unlink_start)
2234				goto found;
2235		}
2236		break;
2237found:
2238		entry->state = unlink_start;
2239		urb = entry->urb;
2240
2241		/* Get reference count of the URB to avoid it to be
2242		 * freed during usb_unlink_urb, which may trigger
2243		 * use-after-free problem inside usb_unlink_urb since
2244		 * usb_unlink_urb is always racing with .complete
2245		 * handler(include defer_bh).
2246		 */
2247		usb_get_urb(urb);
2248		spin_unlock_irqrestore(&q->lock, flags);
2249		/* during some PM-driven resume scenarios,
2250		 * these (async) unlinks complete immediately
2251		 */
2252		ret = usb_unlink_urb(urb);
2253		if (ret != -EINPROGRESS && ret != 0)
2254			netdev_dbg(dev->net, "unlink urb err, %d\n", ret);
2255		else
2256			count++;
2257		usb_put_urb(urb);
2258		spin_lock_irqsave(&q->lock, flags);
2259	}
2260	spin_unlock_irqrestore(&q->lock, flags);
2261	return count;
2262}
2263
2264static int lan78xx_change_mtu(struct net_device *netdev, int new_mtu)
2265{
2266	struct lan78xx_net *dev = netdev_priv(netdev);
2267	int ll_mtu = new_mtu + netdev->hard_header_len;
2268	int old_hard_mtu = dev->hard_mtu;
2269	int old_rx_urb_size = dev->rx_urb_size;
2270
2271	/* no second zero-length packet read wanted after mtu-sized packets */
2272	if ((ll_mtu % dev->maxpacket) == 0)
2273		return -EDOM;
2274
2275	lan78xx_set_rx_max_frame_length(dev, new_mtu + VLAN_ETH_HLEN);
 
 
2276
2277	netdev->mtu = new_mtu;
 
 
2278
2279	dev->hard_mtu = netdev->mtu + netdev->hard_header_len;
2280	if (dev->rx_urb_size == old_hard_mtu) {
2281		dev->rx_urb_size = dev->hard_mtu;
2282		if (dev->rx_urb_size > old_rx_urb_size) {
2283			if (netif_running(dev->net)) {
2284				unlink_urbs(dev, &dev->rxq);
2285				tasklet_schedule(&dev->bh);
2286			}
2287		}
2288	}
2289
2290	return 0;
2291}
2292
2293static int lan78xx_set_mac_addr(struct net_device *netdev, void *p)
2294{
2295	struct lan78xx_net *dev = netdev_priv(netdev);
2296	struct sockaddr *addr = p;
2297	u32 addr_lo, addr_hi;
2298
2299	if (netif_running(netdev))
2300		return -EBUSY;
2301
2302	if (!is_valid_ether_addr(addr->sa_data))
2303		return -EADDRNOTAVAIL;
2304
2305	ether_addr_copy(netdev->dev_addr, addr->sa_data);
2306
2307	addr_lo = netdev->dev_addr[0] |
2308		  netdev->dev_addr[1] << 8 |
2309		  netdev->dev_addr[2] << 16 |
2310		  netdev->dev_addr[3] << 24;
2311	addr_hi = netdev->dev_addr[4] |
2312		  netdev->dev_addr[5] << 8;
2313
2314	lan78xx_write_reg(dev, RX_ADDRL, addr_lo);
2315	lan78xx_write_reg(dev, RX_ADDRH, addr_hi);
2316
2317	/* Added to support MAC address changes */
2318	lan78xx_write_reg(dev, MAF_LO(0), addr_lo);
2319	lan78xx_write_reg(dev, MAF_HI(0), addr_hi | MAF_HI_VALID_);
2320
2321	return 0;
2322}
2323
2324/* Enable or disable Rx checksum offload engine */
2325static int lan78xx_set_features(struct net_device *netdev,
2326				netdev_features_t features)
2327{
2328	struct lan78xx_net *dev = netdev_priv(netdev);
2329	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2330	unsigned long flags;
2331
2332	spin_lock_irqsave(&pdata->rfe_ctl_lock, flags);
2333
2334	if (features & NETIF_F_RXCSUM) {
2335		pdata->rfe_ctl |= RFE_CTL_TCPUDP_COE_ | RFE_CTL_IP_COE_;
2336		pdata->rfe_ctl |= RFE_CTL_ICMP_COE_ | RFE_CTL_IGMP_COE_;
2337	} else {
2338		pdata->rfe_ctl &= ~(RFE_CTL_TCPUDP_COE_ | RFE_CTL_IP_COE_);
2339		pdata->rfe_ctl &= ~(RFE_CTL_ICMP_COE_ | RFE_CTL_IGMP_COE_);
2340	}
2341
2342	if (features & NETIF_F_HW_VLAN_CTAG_RX)
2343		pdata->rfe_ctl |= RFE_CTL_VLAN_STRIP_;
2344	else
2345		pdata->rfe_ctl &= ~RFE_CTL_VLAN_STRIP_;
2346
2347	if (features & NETIF_F_HW_VLAN_CTAG_FILTER)
2348		pdata->rfe_ctl |= RFE_CTL_VLAN_FILTER_;
2349	else
2350		pdata->rfe_ctl &= ~RFE_CTL_VLAN_FILTER_;
2351
2352	spin_unlock_irqrestore(&pdata->rfe_ctl_lock, flags);
2353
2354	lan78xx_write_reg(dev, RFE_CTL, pdata->rfe_ctl);
2355
2356	return 0;
2357}
2358
2359static void lan78xx_deferred_vlan_write(struct work_struct *param)
2360{
2361	struct lan78xx_priv *pdata =
2362			container_of(param, struct lan78xx_priv, set_vlan);
2363	struct lan78xx_net *dev = pdata->dev;
2364
2365	lan78xx_dataport_write(dev, DP_SEL_RSEL_VLAN_DA_, 0,
2366			       DP_SEL_VHF_VLAN_LEN, pdata->vlan_table);
2367}
2368
2369static int lan78xx_vlan_rx_add_vid(struct net_device *netdev,
2370				   __be16 proto, u16 vid)
2371{
2372	struct lan78xx_net *dev = netdev_priv(netdev);
2373	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2374	u16 vid_bit_index;
2375	u16 vid_dword_index;
2376
2377	vid_dword_index = (vid >> 5) & 0x7F;
2378	vid_bit_index = vid & 0x1F;
2379
2380	pdata->vlan_table[vid_dword_index] |= (1 << vid_bit_index);
2381
2382	/* defer register writes to a sleepable context */
2383	schedule_work(&pdata->set_vlan);
2384
2385	return 0;
2386}
2387
2388static int lan78xx_vlan_rx_kill_vid(struct net_device *netdev,
2389				    __be16 proto, u16 vid)
2390{
2391	struct lan78xx_net *dev = netdev_priv(netdev);
2392	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2393	u16 vid_bit_index;
2394	u16 vid_dword_index;
2395
2396	vid_dword_index = (vid >> 5) & 0x7F;
2397	vid_bit_index = vid & 0x1F;
2398
2399	pdata->vlan_table[vid_dword_index] &= ~(1 << vid_bit_index);
2400
2401	/* defer register writes to a sleepable context */
2402	schedule_work(&pdata->set_vlan);
2403
2404	return 0;
2405}
2406
2407static void lan78xx_init_ltm(struct lan78xx_net *dev)
2408{
2409	int ret;
2410	u32 buf;
2411	u32 regs[6] = { 0 };
2412
2413	ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
2414	if (buf & USB_CFG1_LTM_ENABLE_) {
2415		u8 temp[2];
2416		/* Get values from EEPROM first */
2417		if (lan78xx_read_eeprom(dev, 0x3F, 2, temp) == 0) {
2418			if (temp[0] == 24) {
2419				ret = lan78xx_read_raw_eeprom(dev,
2420							      temp[1] * 2,
2421							      24,
2422							      (u8 *)regs);
2423				if (ret < 0)
2424					return;
2425			}
2426		} else if (lan78xx_read_otp(dev, 0x3F, 2, temp) == 0) {
2427			if (temp[0] == 24) {
2428				ret = lan78xx_read_raw_otp(dev,
2429							   temp[1] * 2,
2430							   24,
2431							   (u8 *)regs);
2432				if (ret < 0)
2433					return;
2434			}
2435		}
2436	}
2437
2438	lan78xx_write_reg(dev, LTM_BELT_IDLE0, regs[0]);
2439	lan78xx_write_reg(dev, LTM_BELT_IDLE1, regs[1]);
2440	lan78xx_write_reg(dev, LTM_BELT_ACT0, regs[2]);
2441	lan78xx_write_reg(dev, LTM_BELT_ACT1, regs[3]);
2442	lan78xx_write_reg(dev, LTM_INACTIVE0, regs[4]);
2443	lan78xx_write_reg(dev, LTM_INACTIVE1, regs[5]);
2444}
2445
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2446static int lan78xx_reset(struct lan78xx_net *dev)
2447{
2448	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2449	u32 buf;
2450	int ret = 0;
2451	unsigned long timeout;
 
 
2452	u8 sig;
2453
2454	ret = lan78xx_read_reg(dev, HW_CFG, &buf);
 
 
 
2455	buf |= HW_CFG_LRST_;
 
2456	ret = lan78xx_write_reg(dev, HW_CFG, buf);
 
 
2457
2458	timeout = jiffies + HZ;
2459	do {
2460		mdelay(1);
2461		ret = lan78xx_read_reg(dev, HW_CFG, &buf);
 
 
 
2462		if (time_after(jiffies, timeout)) {
2463			netdev_warn(dev->net,
2464				    "timeout on completion of LiteReset");
2465			return -EIO;
 
2466		}
2467	} while (buf & HW_CFG_LRST_);
2468
2469	lan78xx_init_mac_address(dev);
2470
2471	/* save DEVID for later usage */
2472	ret = lan78xx_read_reg(dev, ID_REV, &buf);
 
 
 
2473	dev->chipid = (buf & ID_REV_CHIP_ID_MASK_) >> 16;
2474	dev->chiprev = buf & ID_REV_CHIP_REV_MASK_;
2475
2476	/* Respond to the IN token with a NAK */
2477	ret = lan78xx_read_reg(dev, USB_CFG0, &buf);
 
 
 
2478	buf |= USB_CFG_BIR_;
 
2479	ret = lan78xx_write_reg(dev, USB_CFG0, buf);
 
 
2480
2481	/* Init LTM */
2482	lan78xx_init_ltm(dev);
2483
2484	if (dev->udev->speed == USB_SPEED_SUPER) {
2485		buf = DEFAULT_BURST_CAP_SIZE / SS_USB_PKT_SIZE;
2486		dev->rx_urb_size = DEFAULT_BURST_CAP_SIZE;
2487		dev->rx_qlen = 4;
2488		dev->tx_qlen = 4;
2489	} else if (dev->udev->speed == USB_SPEED_HIGH) {
2490		buf = DEFAULT_BURST_CAP_SIZE / HS_USB_PKT_SIZE;
2491		dev->rx_urb_size = DEFAULT_BURST_CAP_SIZE;
2492		dev->rx_qlen = RX_MAX_QUEUE_MEMORY / dev->rx_urb_size;
2493		dev->tx_qlen = RX_MAX_QUEUE_MEMORY / dev->hard_mtu;
2494	} else {
2495		buf = DEFAULT_BURST_CAP_SIZE / FS_USB_PKT_SIZE;
2496		dev->rx_urb_size = DEFAULT_BURST_CAP_SIZE;
2497		dev->rx_qlen = 4;
2498		dev->tx_qlen = 4;
2499	}
2500
2501	ret = lan78xx_write_reg(dev, BURST_CAP, buf);
2502	ret = lan78xx_write_reg(dev, BULK_IN_DLY, DEFAULT_BULK_IN_DELAY);
 
2503
2504	ret = lan78xx_read_reg(dev, HW_CFG, &buf);
 
 
 
2505	buf |= HW_CFG_MEF_;
 
2506	ret = lan78xx_write_reg(dev, HW_CFG, buf);
 
 
2507
2508	ret = lan78xx_read_reg(dev, USB_CFG0, &buf);
 
 
 
2509	buf |= USB_CFG_BCE_;
 
2510	ret = lan78xx_write_reg(dev, USB_CFG0, buf);
 
 
2511
2512	/* set FIFO sizes */
2513	buf = (MAX_RX_FIFO_SIZE - 512) / 512;
 
2514	ret = lan78xx_write_reg(dev, FCT_RX_FIFO_END, buf);
 
 
2515
2516	buf = (MAX_TX_FIFO_SIZE - 512) / 512;
 
2517	ret = lan78xx_write_reg(dev, FCT_TX_FIFO_END, buf);
 
 
2518
2519	ret = lan78xx_write_reg(dev, INT_STS, INT_STS_CLEAR_ALL_);
 
 
 
2520	ret = lan78xx_write_reg(dev, FLOW, 0);
 
 
 
2521	ret = lan78xx_write_reg(dev, FCT_FLOW, 0);
 
 
2522
2523	/* Don't need rfe_ctl_lock during initialisation */
2524	ret = lan78xx_read_reg(dev, RFE_CTL, &pdata->rfe_ctl);
 
 
 
2525	pdata->rfe_ctl |= RFE_CTL_BCAST_EN_ | RFE_CTL_DA_PERFECT_;
 
2526	ret = lan78xx_write_reg(dev, RFE_CTL, pdata->rfe_ctl);
 
 
2527
2528	/* Enable or disable checksum offload engines */
2529	lan78xx_set_features(dev->net, dev->net->features);
 
 
2530
2531	lan78xx_set_multicast(dev->net);
2532
2533	/* reset PHY */
2534	ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
 
 
 
2535	buf |= PMT_CTL_PHY_RST_;
 
2536	ret = lan78xx_write_reg(dev, PMT_CTL, buf);
 
 
2537
2538	timeout = jiffies + HZ;
2539	do {
2540		mdelay(1);
2541		ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
 
 
 
2542		if (time_after(jiffies, timeout)) {
2543			netdev_warn(dev->net, "timeout waiting for PHY Reset");
2544			return -EIO;
 
2545		}
2546	} while ((buf & PMT_CTL_PHY_RST_) || !(buf & PMT_CTL_READY_));
2547
2548	ret = lan78xx_read_reg(dev, MAC_CR, &buf);
 
 
 
2549	/* LAN7801 only has RGMII mode */
2550	if (dev->chipid == ID_REV_CHIP_ID_7801_)
2551		buf &= ~MAC_CR_GMII_EN_;
2552
2553	if (dev->chipid == ID_REV_CHIP_ID_7800_) {
2554		ret = lan78xx_read_raw_eeprom(dev, 0, 1, &sig);
2555		if (!ret && sig != EEPROM_INDICATOR) {
2556			/* Implies there is no external eeprom. Set mac speed */
2557			netdev_info(dev->net, "No External EEPROM. Setting MAC Speed\n");
2558			buf |= MAC_CR_AUTO_DUPLEX_ | MAC_CR_AUTO_SPEED_;
2559		}
2560	}
2561	ret = lan78xx_write_reg(dev, MAC_CR, buf);
2562
2563	ret = lan78xx_read_reg(dev, MAC_TX, &buf);
2564	buf |= MAC_TX_TXEN_;
2565	ret = lan78xx_write_reg(dev, MAC_TX, buf);
2566
2567	ret = lan78xx_read_reg(dev, FCT_TX_CTL, &buf);
2568	buf |= FCT_TX_CTL_EN_;
2569	ret = lan78xx_write_reg(dev, FCT_TX_CTL, buf);
2570
2571	ret = lan78xx_set_rx_max_frame_length(dev,
2572					      dev->net->mtu + VLAN_ETH_HLEN);
2573
2574	ret = lan78xx_read_reg(dev, MAC_RX, &buf);
2575	buf |= MAC_RX_RXEN_;
2576	ret = lan78xx_write_reg(dev, MAC_RX, buf);
2577
2578	ret = lan78xx_read_reg(dev, FCT_RX_CTL, &buf);
2579	buf |= FCT_RX_CTL_EN_;
2580	ret = lan78xx_write_reg(dev, FCT_RX_CTL, buf);
2581
2582	return 0;
2583}
2584
2585static void lan78xx_init_stats(struct lan78xx_net *dev)
2586{
2587	u32 *p;
2588	int i;
2589
2590	/* initialize for stats update
2591	 * some counters are 20bits and some are 32bits
2592	 */
2593	p = (u32 *)&dev->stats.rollover_max;
2594	for (i = 0; i < (sizeof(dev->stats.rollover_max) / (sizeof(u32))); i++)
2595		p[i] = 0xFFFFF;
2596
2597	dev->stats.rollover_max.rx_unicast_byte_count = 0xFFFFFFFF;
2598	dev->stats.rollover_max.rx_broadcast_byte_count = 0xFFFFFFFF;
2599	dev->stats.rollover_max.rx_multicast_byte_count = 0xFFFFFFFF;
2600	dev->stats.rollover_max.eee_rx_lpi_transitions = 0xFFFFFFFF;
2601	dev->stats.rollover_max.eee_rx_lpi_time = 0xFFFFFFFF;
2602	dev->stats.rollover_max.tx_unicast_byte_count = 0xFFFFFFFF;
2603	dev->stats.rollover_max.tx_broadcast_byte_count = 0xFFFFFFFF;
2604	dev->stats.rollover_max.tx_multicast_byte_count = 0xFFFFFFFF;
2605	dev->stats.rollover_max.eee_tx_lpi_transitions = 0xFFFFFFFF;
2606	dev->stats.rollover_max.eee_tx_lpi_time = 0xFFFFFFFF;
2607
2608	set_bit(EVENT_STAT_UPDATE, &dev->flags);
2609}
2610
2611static int lan78xx_open(struct net_device *net)
2612{
2613	struct lan78xx_net *dev = netdev_priv(net);
2614	int ret;
2615
 
 
2616	ret = usb_autopm_get_interface(dev->intf);
2617	if (ret < 0)
2618		goto out;
 
 
2619
2620	phy_start(net->phydev);
2621
2622	netif_dbg(dev, ifup, dev->net, "phy initialised successfully");
2623
2624	/* for Link Check */
2625	if (dev->urb_intr) {
2626		ret = usb_submit_urb(dev->urb_intr, GFP_KERNEL);
2627		if (ret < 0) {
2628			netif_err(dev, ifup, dev->net,
2629				  "intr submit %d\n", ret);
2630			goto done;
2631		}
2632	}
2633
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2634	lan78xx_init_stats(dev);
2635
2636	set_bit(EVENT_DEV_OPEN, &dev->flags);
2637
2638	netif_start_queue(net);
2639
2640	dev->link_on = false;
2641
 
 
2642	lan78xx_defer_kevent(dev, EVENT_LINK_RESET);
2643done:
 
 
2644	usb_autopm_put_interface(dev->intf);
2645
2646out:
2647	return ret;
2648}
2649
2650static void lan78xx_terminate_urbs(struct lan78xx_net *dev)
2651{
2652	DECLARE_WAIT_QUEUE_HEAD_ONSTACK(unlink_wakeup);
2653	DECLARE_WAITQUEUE(wait, current);
2654	int temp;
2655
2656	/* ensure there are no more active urbs */
2657	add_wait_queue(&unlink_wakeup, &wait);
2658	set_current_state(TASK_UNINTERRUPTIBLE);
2659	dev->wait = &unlink_wakeup;
2660	temp = unlink_urbs(dev, &dev->txq) + unlink_urbs(dev, &dev->rxq);
2661
2662	/* maybe wait for deletions to finish. */
2663	while (!skb_queue_empty(&dev->rxq) &&
2664	       !skb_queue_empty(&dev->txq) &&
2665	       !skb_queue_empty(&dev->done)) {
2666		schedule_timeout(msecs_to_jiffies(UNLINK_TIMEOUT_MS));
2667		set_current_state(TASK_UNINTERRUPTIBLE);
2668		netif_dbg(dev, ifdown, dev->net,
2669			  "waited for %d urb completions\n", temp);
2670	}
2671	set_current_state(TASK_RUNNING);
2672	dev->wait = NULL;
2673	remove_wait_queue(&unlink_wakeup, &wait);
 
 
 
 
 
 
 
 
 
 
 
2674}
2675
2676static int lan78xx_stop(struct net_device *net)
2677{
2678	struct lan78xx_net *dev = netdev_priv(net);
2679
 
 
 
 
2680	if (timer_pending(&dev->stat_monitor))
2681		del_timer_sync(&dev->stat_monitor);
2682
2683	if (net->phydev)
2684		phy_stop(net->phydev);
2685
2686	clear_bit(EVENT_DEV_OPEN, &dev->flags);
2687	netif_stop_queue(net);
 
 
 
2688
2689	netif_info(dev, ifdown, dev->net,
2690		   "stop stats: rx/tx %lu/%lu, errs %lu/%lu\n",
2691		   net->stats.rx_packets, net->stats.tx_packets,
2692		   net->stats.rx_errors, net->stats.tx_errors);
2693
2694	lan78xx_terminate_urbs(dev);
 
 
2695
2696	usb_kill_urb(dev->urb_intr);
 
2697
2698	skb_queue_purge(&dev->rxq_pause);
2699
2700	/* deferred work (task, timer, softirq) must also stop.
2701	 * can't flush_scheduled_work() until we drop rtnl (later),
2702	 * else workers could deadlock; so make workers a NOP.
2703	 */
2704	dev->flags = 0;
 
 
 
 
2705	cancel_delayed_work_sync(&dev->wq);
2706	tasklet_kill(&dev->bh);
2707
2708	usb_autopm_put_interface(dev->intf);
2709
2710	return 0;
2711}
2712
2713static struct sk_buff *lan78xx_tx_prep(struct lan78xx_net *dev,
2714				       struct sk_buff *skb, gfp_t flags)
2715{
2716	u32 tx_cmd_a, tx_cmd_b;
2717	void *ptr;
2718
2719	if (skb_cow_head(skb, TX_OVERHEAD)) {
2720		dev_kfree_skb_any(skb);
2721		return NULL;
2722	}
2723
2724	if (skb_linearize(skb)) {
2725		dev_kfree_skb_any(skb);
2726		return NULL;
2727	}
2728
2729	tx_cmd_a = (u32)(skb->len & TX_CMD_A_LEN_MASK_) | TX_CMD_A_FCS_;
2730
2731	if (skb->ip_summed == CHECKSUM_PARTIAL)
2732		tx_cmd_a |= TX_CMD_A_IPE_ | TX_CMD_A_TPE_;
2733
2734	tx_cmd_b = 0;
2735	if (skb_is_gso(skb)) {
2736		u16 mss = max(skb_shinfo(skb)->gso_size, TX_CMD_B_MSS_MIN_);
2737
2738		tx_cmd_b = (mss << TX_CMD_B_MSS_SHIFT_) & TX_CMD_B_MSS_MASK_;
2739
2740		tx_cmd_a |= TX_CMD_A_LSO_;
2741	}
2742
2743	if (skb_vlan_tag_present(skb)) {
2744		tx_cmd_a |= TX_CMD_A_IVTG_;
2745		tx_cmd_b |= skb_vlan_tag_get(skb) & TX_CMD_B_VTAG_MASK_;
2746	}
2747
2748	ptr = skb_push(skb, 8);
2749	put_unaligned_le32(tx_cmd_a, ptr);
2750	put_unaligned_le32(tx_cmd_b, ptr + 4);
2751
2752	return skb;
2753}
2754
2755static enum skb_state defer_bh(struct lan78xx_net *dev, struct sk_buff *skb,
2756			       struct sk_buff_head *list, enum skb_state state)
2757{
2758	unsigned long flags;
2759	enum skb_state old_state;
2760	struct skb_data *entry = (struct skb_data *)skb->cb;
2761
2762	spin_lock_irqsave(&list->lock, flags);
2763	old_state = entry->state;
2764	entry->state = state;
2765
2766	__skb_unlink(skb, list);
2767	spin_unlock(&list->lock);
2768	spin_lock(&dev->done.lock);
2769
2770	__skb_queue_tail(&dev->done, skb);
2771	if (skb_queue_len(&dev->done) == 1)
2772		tasklet_schedule(&dev->bh);
2773	spin_unlock_irqrestore(&dev->done.lock, flags);
 
2774
2775	return old_state;
2776}
2777
2778static void tx_complete(struct urb *urb)
2779{
2780	struct sk_buff *skb = (struct sk_buff *)urb->context;
2781	struct skb_data *entry = (struct skb_data *)skb->cb;
2782	struct lan78xx_net *dev = entry->dev;
2783
2784	if (urb->status == 0) {
2785		dev->net->stats.tx_packets += entry->num_of_packet;
2786		dev->net->stats.tx_bytes += entry->length;
2787	} else {
2788		dev->net->stats.tx_errors++;
2789
2790		switch (urb->status) {
2791		case -EPIPE:
2792			lan78xx_defer_kevent(dev, EVENT_TX_HALT);
2793			break;
2794
2795		/* software-driven interface shutdown */
2796		case -ECONNRESET:
2797		case -ESHUTDOWN:
 
 
 
2798			break;
2799
2800		case -EPROTO:
2801		case -ETIME:
2802		case -EILSEQ:
2803			netif_stop_queue(dev->net);
 
 
 
2804			break;
2805		default:
2806			netif_dbg(dev, tx_err, dev->net,
2807				  "tx err %d\n", entry->urb->status);
 
2808			break;
2809		}
2810	}
2811
2812	usb_autopm_put_interface_async(dev->intf);
2813
2814	defer_bh(dev, skb, &dev->txq, tx_done);
 
 
 
 
 
 
 
 
2815}
2816
2817static void lan78xx_queue_skb(struct sk_buff_head *list,
2818			      struct sk_buff *newsk, enum skb_state state)
2819{
2820	struct skb_data *entry = (struct skb_data *)newsk->cb;
2821
2822	__skb_queue_tail(list, newsk);
2823	entry->state = state;
2824}
2825
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2826static netdev_tx_t
2827lan78xx_start_xmit(struct sk_buff *skb, struct net_device *net)
2828{
2829	struct lan78xx_net *dev = netdev_priv(net);
2830	struct sk_buff *skb2 = NULL;
2831
2832	if (skb) {
2833		skb_tx_timestamp(skb);
2834		skb2 = lan78xx_tx_prep(dev, skb, GFP_ATOMIC);
2835	}
2836
2837	if (skb2) {
2838		skb_queue_tail(&dev->txq_pend, skb2);
2839
2840		/* throttle TX patch at slower than SUPER SPEED USB */
2841		if ((dev->udev->speed < USB_SPEED_SUPER) &&
2842		    (skb_queue_len(&dev->txq_pend) > 10))
2843			netif_stop_queue(net);
2844	} else {
2845		netif_dbg(dev, tx_err, dev->net,
2846			  "lan78xx_tx_prep return NULL\n");
2847		dev->net->stats.tx_errors++;
2848		dev->net->stats.tx_dropped++;
2849	}
 
 
 
 
 
2850
2851	tasklet_schedule(&dev->bh);
 
 
 
 
2852
2853	return NETDEV_TX_OK;
2854}
2855
2856static int lan78xx_bind(struct lan78xx_net *dev, struct usb_interface *intf)
2857{
2858	struct lan78xx_priv *pdata = NULL;
2859	int ret;
2860	int i;
2861
2862	dev->data[0] = (unsigned long)kzalloc(sizeof(*pdata), GFP_KERNEL);
2863
2864	pdata = (struct lan78xx_priv *)(dev->data[0]);
2865	if (!pdata) {
2866		netdev_warn(dev->net, "Unable to allocate lan78xx_priv");
2867		return -ENOMEM;
2868	}
2869
2870	pdata->dev = dev;
2871
2872	spin_lock_init(&pdata->rfe_ctl_lock);
2873	mutex_init(&pdata->dataport_mutex);
2874
2875	INIT_WORK(&pdata->set_multicast, lan78xx_deferred_multicast_write);
2876
2877	for (i = 0; i < DP_SEL_VHF_VLAN_LEN; i++)
2878		pdata->vlan_table[i] = 0;
2879
2880	INIT_WORK(&pdata->set_vlan, lan78xx_deferred_vlan_write);
2881
2882	dev->net->features = 0;
2883
2884	if (DEFAULT_TX_CSUM_ENABLE)
2885		dev->net->features |= NETIF_F_HW_CSUM;
2886
2887	if (DEFAULT_RX_CSUM_ENABLE)
2888		dev->net->features |= NETIF_F_RXCSUM;
2889
2890	if (DEFAULT_TSO_CSUM_ENABLE)
2891		dev->net->features |= NETIF_F_TSO | NETIF_F_TSO6 | NETIF_F_SG;
2892
2893	if (DEFAULT_VLAN_RX_OFFLOAD)
2894		dev->net->features |= NETIF_F_HW_VLAN_CTAG_RX;
2895
2896	if (DEFAULT_VLAN_FILTER_ENABLE)
2897		dev->net->features |= NETIF_F_HW_VLAN_CTAG_FILTER;
2898
2899	dev->net->hw_features = dev->net->features;
2900
2901	ret = lan78xx_setup_irq_domain(dev);
2902	if (ret < 0) {
2903		netdev_warn(dev->net,
2904			    "lan78xx_setup_irq_domain() failed : %d", ret);
2905		goto out1;
2906	}
2907
2908	dev->net->hard_header_len += TX_OVERHEAD;
2909	dev->hard_mtu = dev->net->mtu + dev->net->hard_header_len;
2910
2911	/* Init all registers */
2912	ret = lan78xx_reset(dev);
2913	if (ret) {
2914		netdev_warn(dev->net, "Registers INIT FAILED....");
2915		goto out2;
2916	}
2917
2918	ret = lan78xx_mdio_init(dev);
2919	if (ret) {
2920		netdev_warn(dev->net, "MDIO INIT FAILED.....");
2921		goto out2;
2922	}
2923
2924	dev->net->flags |= IFF_MULTICAST;
2925
2926	pdata->wol = WAKE_MAGIC;
2927
2928	return ret;
2929
2930out2:
2931	lan78xx_remove_irq_domain(dev);
2932
2933out1:
2934	netdev_warn(dev->net, "Bind routine FAILED");
2935	cancel_work_sync(&pdata->set_multicast);
2936	cancel_work_sync(&pdata->set_vlan);
2937	kfree(pdata);
2938	return ret;
2939}
2940
2941static void lan78xx_unbind(struct lan78xx_net *dev, struct usb_interface *intf)
2942{
2943	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2944
2945	lan78xx_remove_irq_domain(dev);
2946
2947	lan78xx_remove_mdio(dev);
2948
2949	if (pdata) {
2950		cancel_work_sync(&pdata->set_multicast);
2951		cancel_work_sync(&pdata->set_vlan);
2952		netif_dbg(dev, ifdown, dev->net, "free pdata");
2953		kfree(pdata);
2954		pdata = NULL;
2955		dev->data[0] = 0;
2956	}
2957}
2958
2959static void lan78xx_rx_csum_offload(struct lan78xx_net *dev,
2960				    struct sk_buff *skb,
2961				    u32 rx_cmd_a, u32 rx_cmd_b)
2962{
2963	/* HW Checksum offload appears to be flawed if used when not stripping
2964	 * VLAN headers. Drop back to S/W checksums under these conditions.
2965	 */
2966	if (!(dev->net->features & NETIF_F_RXCSUM) ||
2967	    unlikely(rx_cmd_a & RX_CMD_A_ICSM_) ||
2968	    ((rx_cmd_a & RX_CMD_A_FVTG_) &&
2969	     !(dev->net->features & NETIF_F_HW_VLAN_CTAG_RX))) {
2970		skb->ip_summed = CHECKSUM_NONE;
2971	} else {
2972		skb->csum = ntohs((u16)(rx_cmd_b >> RX_CMD_B_CSUM_SHIFT_));
2973		skb->ip_summed = CHECKSUM_COMPLETE;
2974	}
2975}
2976
2977static void lan78xx_rx_vlan_offload(struct lan78xx_net *dev,
2978				    struct sk_buff *skb,
2979				    u32 rx_cmd_a, u32 rx_cmd_b)
2980{
2981	if ((dev->net->features & NETIF_F_HW_VLAN_CTAG_RX) &&
2982	    (rx_cmd_a & RX_CMD_A_FVTG_))
2983		__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q),
2984				       (rx_cmd_b & 0xffff));
2985}
2986
2987static void lan78xx_skb_return(struct lan78xx_net *dev, struct sk_buff *skb)
2988{
2989	int status;
2990
2991	if (test_bit(EVENT_RX_PAUSED, &dev->flags)) {
2992		skb_queue_tail(&dev->rxq_pause, skb);
2993		return;
2994	}
2995
2996	dev->net->stats.rx_packets++;
2997	dev->net->stats.rx_bytes += skb->len;
2998
2999	skb->protocol = eth_type_trans(skb, dev->net);
3000
3001	netif_dbg(dev, rx_status, dev->net, "< rx, len %zu, type 0x%x\n",
3002		  skb->len + sizeof(struct ethhdr), skb->protocol);
3003	memset(skb->cb, 0, sizeof(struct skb_data));
3004
3005	if (skb_defer_rx_timestamp(skb))
3006		return;
3007
3008	status = netif_rx(skb);
3009	if (status != NET_RX_SUCCESS)
3010		netif_dbg(dev, rx_err, dev->net,
3011			  "netif_rx status %d\n", status);
3012}
3013
3014static int lan78xx_rx(struct lan78xx_net *dev, struct sk_buff *skb)
 
3015{
3016	if (skb->len < dev->net->hard_header_len)
3017		return 0;
3018
 
 
 
3019	while (skb->len > 0) {
3020		u32 rx_cmd_a, rx_cmd_b, align_count, size;
3021		u16 rx_cmd_c;
3022		struct sk_buff *skb2;
3023		unsigned char *packet;
3024
3025		rx_cmd_a = get_unaligned_le32(skb->data);
3026		skb_pull(skb, sizeof(rx_cmd_a));
3027
3028		rx_cmd_b = get_unaligned_le32(skb->data);
3029		skb_pull(skb, sizeof(rx_cmd_b));
3030
3031		rx_cmd_c = get_unaligned_le16(skb->data);
3032		skb_pull(skb, sizeof(rx_cmd_c));
3033
3034		packet = skb->data;
3035
3036		/* get the packet length */
3037		size = (rx_cmd_a & RX_CMD_A_LEN_MASK_);
3038		align_count = (4 - ((size + RXW_PADDING) % 4)) % 4;
3039
3040		if (unlikely(rx_cmd_a & RX_CMD_A_RED_)) {
3041			netif_dbg(dev, rx_err, dev->net,
3042				  "Error rx_cmd_a=0x%08x", rx_cmd_a);
3043		} else {
3044			/* last frame in this batch */
3045			if (skb->len == size) {
3046				lan78xx_rx_csum_offload(dev, skb,
3047							rx_cmd_a, rx_cmd_b);
3048				lan78xx_rx_vlan_offload(dev, skb,
3049							rx_cmd_a, rx_cmd_b);
3050
3051				skb_trim(skb, skb->len - 4); /* remove fcs */
3052				skb->truesize = size + sizeof(struct sk_buff);
3053
3054				return 1;
3055			}
3056
3057			skb2 = skb_clone(skb, GFP_ATOMIC);
3058			if (unlikely(!skb2)) {
3059				netdev_warn(dev->net, "Error allocating skb");
3060				return 0;
3061			}
3062
3063			skb2->len = size;
3064			skb2->data = packet;
3065			skb_set_tail_pointer(skb2, size);
3066
3067			lan78xx_rx_csum_offload(dev, skb2, rx_cmd_a, rx_cmd_b);
3068			lan78xx_rx_vlan_offload(dev, skb2, rx_cmd_a, rx_cmd_b);
3069
3070			skb_trim(skb2, skb2->len - 4); /* remove fcs */
3071			skb2->truesize = size + sizeof(struct sk_buff);
3072
3073			lan78xx_skb_return(dev, skb2);
 
 
 
 
 
 
 
3074		}
3075
3076		skb_pull(skb, size);
3077
3078		/* padding bytes before the next frame starts */
3079		if (skb->len)
3080			skb_pull(skb, align_count);
3081	}
3082
3083	return 1;
3084}
3085
3086static inline void rx_process(struct lan78xx_net *dev, struct sk_buff *skb)
 
3087{
3088	if (!lan78xx_rx(dev, skb)) {
 
3089		dev->net->stats.rx_errors++;
3090		goto done;
3091	}
3092
3093	if (skb->len) {
3094		lan78xx_skb_return(dev, skb);
3095		return;
3096	}
3097
3098	netif_dbg(dev, rx_err, dev->net, "drop\n");
3099	dev->net->stats.rx_errors++;
3100done:
3101	skb_queue_tail(&dev->done, skb);
3102}
3103
3104static void rx_complete(struct urb *urb);
3105
3106static int rx_submit(struct lan78xx_net *dev, struct urb *urb, gfp_t flags)
3107{
3108	struct sk_buff *skb;
3109	struct skb_data *entry;
3110	unsigned long lockflags;
3111	size_t size = dev->rx_urb_size;
3112	int ret = 0;
3113
3114	skb = netdev_alloc_skb_ip_align(dev->net, size);
3115	if (!skb) {
3116		usb_free_urb(urb);
3117		return -ENOMEM;
3118	}
3119
3120	entry = (struct skb_data *)skb->cb;
3121	entry->urb = urb;
3122	entry->dev = dev;
3123	entry->length = 0;
3124
3125	usb_fill_bulk_urb(urb, dev->udev, dev->pipe_in,
3126			  skb->data, size, rx_complete, skb);
3127
3128	spin_lock_irqsave(&dev->rxq.lock, lockflags);
3129
3130	if (netif_device_present(dev->net) &&
3131	    netif_running(dev->net) &&
3132	    !test_bit(EVENT_RX_HALT, &dev->flags) &&
3133	    !test_bit(EVENT_DEV_ASLEEP, &dev->flags)) {
3134		ret = usb_submit_urb(urb, GFP_ATOMIC);
3135		switch (ret) {
3136		case 0:
3137			lan78xx_queue_skb(&dev->rxq, skb, rx_start);
3138			break;
3139		case -EPIPE:
3140			lan78xx_defer_kevent(dev, EVENT_RX_HALT);
3141			break;
3142		case -ENODEV:
3143			netif_dbg(dev, ifdown, dev->net, "device gone\n");
3144			netif_device_detach(dev->net);
3145			break;
3146		case -EHOSTUNREACH:
3147			ret = -ENOLINK;
3148			break;
3149		default:
3150			netif_dbg(dev, rx_err, dev->net,
3151				  "rx submit, %d\n", ret);
3152			tasklet_schedule(&dev->bh);
3153		}
3154	} else {
3155		netif_dbg(dev, ifdown, dev->net, "rx: stopped\n");
3156		ret = -ENOLINK;
3157	}
3158	spin_unlock_irqrestore(&dev->rxq.lock, lockflags);
3159	if (ret) {
3160		dev_kfree_skb_any(skb);
3161		usb_free_urb(urb);
3162	}
3163	return ret;
3164}
3165
3166static void rx_complete(struct urb *urb)
3167{
3168	struct sk_buff	*skb = (struct sk_buff *)urb->context;
3169	struct skb_data	*entry = (struct skb_data *)skb->cb;
3170	struct lan78xx_net *dev = entry->dev;
3171	int urb_status = urb->status;
3172	enum skb_state state;
3173
 
 
 
3174	skb_put(skb, urb->actual_length);
3175	state = rx_done;
3176	entry->urb = NULL;
 
 
3177
3178	switch (urb_status) {
3179	case 0:
3180		if (skb->len < dev->net->hard_header_len) {
3181			state = rx_cleanup;
3182			dev->net->stats.rx_errors++;
3183			dev->net->stats.rx_length_errors++;
3184			netif_dbg(dev, rx_err, dev->net,
3185				  "rx length %d\n", skb->len);
3186		}
3187		usb_mark_last_busy(dev->udev);
3188		break;
3189	case -EPIPE:
3190		dev->net->stats.rx_errors++;
3191		lan78xx_defer_kevent(dev, EVENT_RX_HALT);
3192		fallthrough;
3193	case -ECONNRESET:				/* async unlink */
3194	case -ESHUTDOWN:				/* hardware gone */
3195		netif_dbg(dev, ifdown, dev->net,
3196			  "rx shutdown, code %d\n", urb_status);
3197		state = rx_cleanup;
3198		entry->urb = urb;
3199		urb = NULL;
3200		break;
3201	case -EPROTO:
3202	case -ETIME:
3203	case -EILSEQ:
3204		dev->net->stats.rx_errors++;
3205		state = rx_cleanup;
3206		entry->urb = urb;
3207		urb = NULL;
3208		break;
3209
3210	/* data overrun ... flush fifo? */
3211	case -EOVERFLOW:
3212		dev->net->stats.rx_over_errors++;
3213		fallthrough;
3214
3215	default:
3216		state = rx_cleanup;
3217		dev->net->stats.rx_errors++;
3218		netif_dbg(dev, rx_err, dev->net, "rx status %d\n", urb_status);
3219		break;
3220	}
3221
3222	state = defer_bh(dev, skb, &dev->rxq, state);
 
3223
3224	if (urb) {
3225		if (netif_running(dev->net) &&
3226		    !test_bit(EVENT_RX_HALT, &dev->flags) &&
3227		    state != unlink_start) {
3228			rx_submit(dev, urb, GFP_ATOMIC);
3229			return;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3230		}
3231		usb_free_urb(urb);
 
 
3232	}
3233	netif_dbg(dev, rx_err, dev->net, "no read resubmitted\n");
 
 
 
 
 
3234}
3235
3236static void lan78xx_tx_bh(struct lan78xx_net *dev)
3237{
3238	int length;
3239	struct urb *urb = NULL;
3240	struct skb_data *entry;
3241	unsigned long flags;
3242	struct sk_buff_head *tqp = &dev->txq_pend;
3243	struct sk_buff *skb, *skb2;
3244	int ret;
3245	int count, pos;
3246	int skb_totallen, pkt_cnt;
3247
3248	skb_totallen = 0;
3249	pkt_cnt = 0;
3250	count = 0;
3251	length = 0;
3252	spin_lock_irqsave(&tqp->lock, flags);
3253	skb_queue_walk(tqp, skb) {
3254		if (skb_is_gso(skb)) {
3255			if (!skb_queue_is_first(tqp, skb)) {
3256				/* handle previous packets first */
3257				break;
3258			}
3259			count = 1;
3260			length = skb->len - TX_OVERHEAD;
3261			__skb_unlink(skb, tqp);
3262			spin_unlock_irqrestore(&tqp->lock, flags);
3263			goto gso_skb;
3264		}
3265
3266		if ((skb_totallen + skb->len) > MAX_SINGLE_PACKET_SIZE)
3267			break;
3268		skb_totallen = skb->len + roundup(skb_totallen, sizeof(u32));
3269		pkt_cnt++;
3270	}
3271	spin_unlock_irqrestore(&tqp->lock, flags);
3272
3273	/* copy to a single skb */
3274	skb = alloc_skb(skb_totallen, GFP_ATOMIC);
3275	if (!skb)
3276		goto drop;
3277
3278	skb_put(skb, skb_totallen);
 
 
 
3279
3280	for (count = pos = 0; count < pkt_cnt; count++) {
3281		skb2 = skb_dequeue(tqp);
3282		if (skb2) {
3283			length += (skb2->len - TX_OVERHEAD);
3284			memcpy(skb->data + pos, skb2->data, skb2->len);
3285			pos += roundup(skb2->len, sizeof(u32));
3286			dev_kfree_skb(skb2);
3287		}
3288	}
3289
3290gso_skb:
3291	urb = usb_alloc_urb(0, GFP_ATOMIC);
3292	if (!urb)
3293		goto drop;
3294
3295	entry = (struct skb_data *)skb->cb;
3296	entry->urb = urb;
3297	entry->dev = dev;
3298	entry->length = length;
3299	entry->num_of_packet = count;
3300
3301	spin_lock_irqsave(&dev->txq.lock, flags);
3302	ret = usb_autopm_get_interface_async(dev->intf);
3303	if (ret < 0) {
3304		spin_unlock_irqrestore(&dev->txq.lock, flags);
3305		goto drop;
3306	}
3307
3308	usb_fill_bulk_urb(urb, dev->udev, dev->pipe_out,
3309			  skb->data, skb->len, tx_complete, skb);
 
3310
3311	if (length % dev->maxpacket == 0) {
3312		/* send USB_ZERO_PACKET */
3313		urb->transfer_flags |= URB_ZERO_PACKET;
3314	}
3315
3316#ifdef CONFIG_PM
3317	/* if this triggers the device is still a sleep */
3318	if (test_bit(EVENT_DEV_ASLEEP, &dev->flags)) {
3319		/* transmission will be done in resume */
3320		usb_anchor_urb(urb, &dev->deferred);
3321		/* no use to process more packets */
3322		netif_stop_queue(dev->net);
3323		usb_put_urb(urb);
3324		spin_unlock_irqrestore(&dev->txq.lock, flags);
3325		netdev_dbg(dev->net, "Delaying transmission for resumption\n");
3326		return;
3327	}
3328#endif
3329
3330	ret = usb_submit_urb(urb, GFP_ATOMIC);
3331	switch (ret) {
3332	case 0:
3333		netif_trans_update(dev->net);
3334		lan78xx_queue_skb(&dev->txq, skb, tx_start);
3335		if (skb_queue_len(&dev->txq) >= dev->tx_qlen)
3336			netif_stop_queue(dev->net);
3337		break;
3338	case -EPIPE:
3339		netif_stop_queue(dev->net);
3340		lan78xx_defer_kevent(dev, EVENT_TX_HALT);
3341		usb_autopm_put_interface_async(dev->intf);
3342		break;
3343	default:
3344		usb_autopm_put_interface_async(dev->intf);
3345		netif_dbg(dev, tx_err, dev->net,
3346			  "tx: submit urb err %d\n", ret);
3347		break;
3348	}
3349
3350	spin_unlock_irqrestore(&dev->txq.lock, flags);
 
 
3351
3352	if (ret) {
3353		netif_dbg(dev, tx_err, dev->net, "drop, code %d\n", ret);
3354drop:
3355		dev->net->stats.tx_dropped++;
3356		if (skb)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3357			dev_kfree_skb_any(skb);
3358		usb_free_urb(urb);
3359	} else
3360		netif_dbg(dev, tx_queued, dev->net,
3361			  "> tx, len %d, type 0x%x\n", length, skb->protocol);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3362}
3363
3364static void lan78xx_rx_bh(struct lan78xx_net *dev)
3365{
3366	struct urb *urb;
3367	int i;
3368
3369	if (skb_queue_len(&dev->rxq) < dev->rx_qlen) {
3370		for (i = 0; i < 10; i++) {
3371			if (skb_queue_len(&dev->rxq) >= dev->rx_qlen)
3372				break;
3373			urb = usb_alloc_urb(0, GFP_ATOMIC);
3374			if (urb)
3375				if (rx_submit(dev, urb, GFP_ATOMIC) == -ENOLINK)
3376					return;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3377		}
3378
3379		if (skb_queue_len(&dev->rxq) < dev->rx_qlen)
3380			tasklet_schedule(&dev->bh);
3381	}
3382	if (skb_queue_len(&dev->txq) < dev->tx_qlen)
3383		netif_wake_queue(dev->net);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3384}
3385
3386static void lan78xx_bh(struct tasklet_struct *t)
3387{
3388	struct lan78xx_net *dev = from_tasklet(dev, t, bh);
3389	struct sk_buff *skb;
3390	struct skb_data *entry;
 
 
 
 
 
 
 
 
 
 
3391
3392	while ((skb = skb_dequeue(&dev->done))) {
3393		entry = (struct skb_data *)(skb->cb);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3394		switch (entry->state) {
3395		case rx_done:
3396			entry->state = rx_cleanup;
3397			rx_process(dev, skb);
3398			continue;
3399		case tx_done:
3400			usb_free_urb(entry->urb);
3401			dev_kfree_skb(skb);
3402			continue;
3403		case rx_cleanup:
3404			usb_free_urb(entry->urb);
3405			dev_kfree_skb(skb);
3406			continue;
3407		default:
3408			netdev_dbg(dev->net, "skb state %d\n", entry->state);
3409			return;
 
3410		}
 
 
3411	}
3412
 
 
 
 
 
 
 
 
3413	if (netif_device_present(dev->net) && netif_running(dev->net)) {
3414		/* reset update timer delta */
3415		if (timer_pending(&dev->stat_monitor) && (dev->delta != 1)) {
3416			dev->delta = 1;
3417			mod_timer(&dev->stat_monitor,
3418				  jiffies + STAT_UPDATE_TIMER);
3419		}
3420
3421		if (!skb_queue_empty(&dev->txq_pend))
3422			lan78xx_tx_bh(dev);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3423
3424		if (!timer_pending(&dev->delay) &&
3425		    !test_bit(EVENT_RX_HALT, &dev->flags))
3426			lan78xx_rx_bh(dev);
3427	}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3428}
3429
3430static void lan78xx_delayedwork(struct work_struct *work)
3431{
3432	int status;
3433	struct lan78xx_net *dev;
3434
3435	dev = container_of(work, struct lan78xx_net, wq.work);
3436
 
 
 
 
 
 
3437	if (test_bit(EVENT_TX_HALT, &dev->flags)) {
3438		unlink_urbs(dev, &dev->txq);
3439		status = usb_autopm_get_interface(dev->intf);
3440		if (status < 0)
3441			goto fail_pipe;
3442		status = usb_clear_halt(dev->udev, dev->pipe_out);
3443		usb_autopm_put_interface(dev->intf);
3444		if (status < 0 &&
3445		    status != -EPIPE &&
3446		    status != -ESHUTDOWN) {
3447			if (netif_msg_tx_err(dev))
3448fail_pipe:
3449				netdev_err(dev->net,
3450					   "can't clear tx halt, status %d\n",
3451					   status);
3452		} else {
3453			clear_bit(EVENT_TX_HALT, &dev->flags);
3454			if (status != -ESHUTDOWN)
3455				netif_wake_queue(dev->net);
3456		}
3457	}
 
3458	if (test_bit(EVENT_RX_HALT, &dev->flags)) {
3459		unlink_urbs(dev, &dev->rxq);
3460		status = usb_autopm_get_interface(dev->intf);
3461		if (status < 0)
3462				goto fail_halt;
3463		status = usb_clear_halt(dev->udev, dev->pipe_in);
3464		usb_autopm_put_interface(dev->intf);
3465		if (status < 0 &&
3466		    status != -EPIPE &&
3467		    status != -ESHUTDOWN) {
3468			if (netif_msg_rx_err(dev))
3469fail_halt:
3470				netdev_err(dev->net,
3471					   "can't clear rx halt, status %d\n",
3472					   status);
3473		} else {
3474			clear_bit(EVENT_RX_HALT, &dev->flags);
3475			tasklet_schedule(&dev->bh);
3476		}
3477	}
3478
3479	if (test_bit(EVENT_LINK_RESET, &dev->flags)) {
3480		int ret = 0;
3481
3482		clear_bit(EVENT_LINK_RESET, &dev->flags);
3483		status = usb_autopm_get_interface(dev->intf);
3484		if (status < 0)
3485			goto skip_reset;
3486		if (lan78xx_link_reset(dev) < 0) {
3487			usb_autopm_put_interface(dev->intf);
3488skip_reset:
3489			netdev_info(dev->net, "link reset failed (%d)\n",
3490				    ret);
3491		} else {
3492			usb_autopm_put_interface(dev->intf);
3493		}
3494	}
3495
3496	if (test_bit(EVENT_STAT_UPDATE, &dev->flags)) {
3497		lan78xx_update_stats(dev);
3498
3499		clear_bit(EVENT_STAT_UPDATE, &dev->flags);
3500
3501		mod_timer(&dev->stat_monitor,
3502			  jiffies + (STAT_UPDATE_TIMER * dev->delta));
3503
3504		dev->delta = min((dev->delta * 2), 50);
3505	}
 
 
3506}
3507
3508static void intr_complete(struct urb *urb)
3509{
3510	struct lan78xx_net *dev = urb->context;
3511	int status = urb->status;
3512
3513	switch (status) {
3514	/* success */
3515	case 0:
3516		lan78xx_status(dev, urb);
3517		break;
3518
3519	/* software-driven interface shutdown */
3520	case -ENOENT:			/* urb killed */
 
3521	case -ESHUTDOWN:		/* hardware gone */
3522		netif_dbg(dev, ifdown, dev->net,
3523			  "intr shutdown, code %d\n", status);
3524		return;
3525
3526	/* NOTE:  not throttling like RX/TX, since this endpoint
3527	 * already polls infrequently
3528	 */
3529	default:
3530		netdev_dbg(dev->net, "intr status %d\n", status);
3531		break;
3532	}
3533
3534	if (!netif_running(dev->net))
 
 
3535		return;
 
3536
3537	memset(urb->transfer_buffer, 0, urb->transfer_buffer_length);
3538	status = usb_submit_urb(urb, GFP_ATOMIC);
3539	if (status != 0)
 
 
 
 
 
 
 
 
 
 
3540		netif_err(dev, timer, dev->net,
3541			  "intr resubmit --> %d\n", status);
 
 
3542}
3543
3544static void lan78xx_disconnect(struct usb_interface *intf)
3545{
3546	struct lan78xx_net *dev;
3547	struct usb_device *udev;
3548	struct net_device *net;
3549	struct phy_device *phydev;
3550
3551	dev = usb_get_intfdata(intf);
3552	usb_set_intfdata(intf, NULL);
3553	if (!dev)
3554		return;
3555
 
 
 
 
3556	udev = interface_to_usbdev(intf);
3557	net = dev->net;
 
 
 
 
 
3558	phydev = net->phydev;
3559
3560	phy_unregister_fixup_for_uid(PHY_KSZ9031RNX, 0xfffffff0);
3561	phy_unregister_fixup_for_uid(PHY_LAN8835, 0xfffffff0);
3562
3563	phy_disconnect(net->phydev);
3564
3565	if (phy_is_pseudo_fixed_link(phydev))
3566		fixed_phy_unregister(phydev);
3567
3568	unregister_netdev(net);
3569
3570	cancel_delayed_work_sync(&dev->wq);
3571
3572	usb_scuttle_anchored_urbs(&dev->deferred);
3573
 
 
 
3574	lan78xx_unbind(dev, intf);
3575
 
 
 
3576	usb_kill_urb(dev->urb_intr);
3577	usb_free_urb(dev->urb_intr);
3578
3579	free_netdev(net);
3580	usb_put_dev(udev);
3581}
3582
3583static void lan78xx_tx_timeout(struct net_device *net, unsigned int txqueue)
3584{
3585	struct lan78xx_net *dev = netdev_priv(net);
3586
3587	unlink_urbs(dev, &dev->txq);
3588	tasklet_schedule(&dev->bh);
3589}
3590
3591static netdev_features_t lan78xx_features_check(struct sk_buff *skb,
3592						struct net_device *netdev,
3593						netdev_features_t features)
3594{
3595	if (skb->len + TX_OVERHEAD > MAX_SINGLE_PACKET_SIZE)
 
 
3596		features &= ~NETIF_F_GSO_MASK;
3597
3598	features = vlan_features_check(skb, features);
3599	features = vxlan_features_check(skb, features);
3600
3601	return features;
3602}
3603
3604static const struct net_device_ops lan78xx_netdev_ops = {
3605	.ndo_open		= lan78xx_open,
3606	.ndo_stop		= lan78xx_stop,
3607	.ndo_start_xmit		= lan78xx_start_xmit,
3608	.ndo_tx_timeout		= lan78xx_tx_timeout,
3609	.ndo_change_mtu		= lan78xx_change_mtu,
3610	.ndo_set_mac_address	= lan78xx_set_mac_addr,
3611	.ndo_validate_addr	= eth_validate_addr,
3612	.ndo_do_ioctl		= phy_do_ioctl_running,
3613	.ndo_set_rx_mode	= lan78xx_set_multicast,
3614	.ndo_set_features	= lan78xx_set_features,
3615	.ndo_vlan_rx_add_vid	= lan78xx_vlan_rx_add_vid,
3616	.ndo_vlan_rx_kill_vid	= lan78xx_vlan_rx_kill_vid,
3617	.ndo_features_check	= lan78xx_features_check,
3618};
3619
3620static void lan78xx_stat_monitor(struct timer_list *t)
3621{
3622	struct lan78xx_net *dev = from_timer(dev, t, stat_monitor);
3623
3624	lan78xx_defer_kevent(dev, EVENT_STAT_UPDATE);
3625}
3626
3627static int lan78xx_probe(struct usb_interface *intf,
3628			 const struct usb_device_id *id)
3629{
3630	struct usb_host_endpoint *ep_blkin, *ep_blkout, *ep_intr;
3631	struct lan78xx_net *dev;
3632	struct net_device *netdev;
3633	struct usb_device *udev;
3634	int ret;
3635	unsigned maxp;
3636	unsigned period;
3637	u8 *buf = NULL;
3638
3639	udev = interface_to_usbdev(intf);
3640	udev = usb_get_dev(udev);
3641
3642	netdev = alloc_etherdev(sizeof(struct lan78xx_net));
3643	if (!netdev) {
3644		dev_err(&intf->dev, "Error: OOM\n");
3645		ret = -ENOMEM;
3646		goto out1;
3647	}
3648
3649	/* netdev_printk() needs this */
3650	SET_NETDEV_DEV(netdev, &intf->dev);
3651
3652	dev = netdev_priv(netdev);
3653	dev->udev = udev;
3654	dev->intf = intf;
3655	dev->net = netdev;
3656	dev->msg_enable = netif_msg_init(msg_level, NETIF_MSG_DRV
3657					| NETIF_MSG_PROBE | NETIF_MSG_LINK);
3658
3659	skb_queue_head_init(&dev->rxq);
3660	skb_queue_head_init(&dev->txq);
3661	skb_queue_head_init(&dev->done);
3662	skb_queue_head_init(&dev->rxq_pause);
3663	skb_queue_head_init(&dev->txq_pend);
 
3664	mutex_init(&dev->phy_mutex);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3665
3666	tasklet_setup(&dev->bh, lan78xx_bh);
3667	INIT_DELAYED_WORK(&dev->wq, lan78xx_delayedwork);
3668	init_usb_anchor(&dev->deferred);
3669
3670	netdev->netdev_ops = &lan78xx_netdev_ops;
3671	netdev->watchdog_timeo = TX_TIMEOUT_JIFFIES;
3672	netdev->ethtool_ops = &lan78xx_ethtool_ops;
3673
3674	dev->delta = 1;
3675	timer_setup(&dev->stat_monitor, lan78xx_stat_monitor, 0);
3676
3677	mutex_init(&dev->stats.access_lock);
3678
3679	if (intf->cur_altsetting->desc.bNumEndpoints < 3) {
3680		ret = -ENODEV;
3681		goto out2;
3682	}
3683
3684	dev->pipe_in = usb_rcvbulkpipe(udev, BULK_IN_PIPE);
3685	ep_blkin = usb_pipe_endpoint(udev, dev->pipe_in);
3686	if (!ep_blkin || !usb_endpoint_is_bulk_in(&ep_blkin->desc)) {
3687		ret = -ENODEV;
3688		goto out2;
3689	}
3690
3691	dev->pipe_out = usb_sndbulkpipe(udev, BULK_OUT_PIPE);
3692	ep_blkout = usb_pipe_endpoint(udev, dev->pipe_out);
3693	if (!ep_blkout || !usb_endpoint_is_bulk_out(&ep_blkout->desc)) {
3694		ret = -ENODEV;
3695		goto out2;
3696	}
3697
3698	ep_intr = &intf->cur_altsetting->endpoint[2];
3699	if (!usb_endpoint_is_int_in(&ep_intr->desc)) {
3700		ret = -ENODEV;
3701		goto out2;
3702	}
3703
3704	dev->pipe_intr = usb_rcvintpipe(dev->udev,
3705					usb_endpoint_num(&ep_intr->desc));
3706
3707	ret = lan78xx_bind(dev, intf);
3708	if (ret < 0)
3709		goto out2;
3710
3711	if (netdev->mtu > (dev->hard_mtu - netdev->hard_header_len))
3712		netdev->mtu = dev->hard_mtu - netdev->hard_header_len;
3713
3714	/* MTU range: 68 - 9000 */
3715	netdev->max_mtu = MAX_SINGLE_PACKET_SIZE;
3716	netif_set_gso_max_size(netdev, MAX_SINGLE_PACKET_SIZE - MAX_HEADER);
3717
3718	period = ep_intr->desc.bInterval;
3719	maxp = usb_maxpacket(dev->udev, dev->pipe_intr, 0);
3720	buf = kmalloc(maxp, GFP_KERNEL);
3721	if (buf) {
3722		dev->urb_intr = usb_alloc_urb(0, GFP_KERNEL);
3723		if (!dev->urb_intr) {
3724			ret = -ENOMEM;
3725			kfree(buf);
3726			goto out3;
3727		} else {
3728			usb_fill_int_urb(dev->urb_intr, dev->udev,
3729					 dev->pipe_intr, buf, maxp,
3730					 intr_complete, dev, period);
3731			dev->urb_intr->transfer_flags |= URB_FREE_BUFFER;
3732		}
 
 
3733	}
3734
3735	dev->maxpacket = usb_maxpacket(dev->udev, dev->pipe_out, 1);
 
 
 
 
 
 
3736
3737	/* driver requires remote-wakeup capability during autosuspend. */
3738	intf->needs_remote_wakeup = 1;
3739
3740	ret = lan78xx_phy_init(dev);
3741	if (ret < 0)
3742		goto out4;
3743
3744	ret = register_netdev(netdev);
3745	if (ret != 0) {
3746		netif_err(dev, probe, netdev, "couldn't register the device\n");
3747		goto out5;
3748	}
3749
3750	usb_set_intfdata(intf, dev);
3751
3752	ret = device_set_wakeup_enable(&udev->dev, true);
3753
3754	 /* Default delay of 2sec has more overhead than advantage.
3755	  * Set to 10sec as default.
3756	  */
3757	pm_runtime_set_autosuspend_delay(&udev->dev,
3758					 DEFAULT_AUTOSUSPEND_DELAY);
3759
3760	return 0;
3761
3762out5:
3763	phy_disconnect(netdev->phydev);
3764out4:
3765	usb_free_urb(dev->urb_intr);
3766out3:
 
 
3767	lan78xx_unbind(dev, intf);
 
 
 
 
 
3768out2:
3769	free_netdev(netdev);
3770out1:
3771	usb_put_dev(udev);
3772
3773	return ret;
3774}
3775
3776static u16 lan78xx_wakeframe_crc16(const u8 *buf, int len)
3777{
3778	const u16 crc16poly = 0x8005;
3779	int i;
3780	u16 bit, crc, msb;
3781	u8 data;
3782
3783	crc = 0xFFFF;
3784	for (i = 0; i < len; i++) {
3785		data = *buf++;
3786		for (bit = 0; bit < 8; bit++) {
3787			msb = crc >> 15;
3788			crc <<= 1;
3789
3790			if (msb ^ (u16)(data & 1)) {
3791				crc ^= crc16poly;
3792				crc |= (u16)0x0001U;
3793			}
3794			data >>= 1;
3795		}
3796	}
3797
3798	return crc;
3799}
3800
3801static int lan78xx_set_suspend(struct lan78xx_net *dev, u32 wol)
3802{
3803	u32 buf;
3804	int mask_index;
3805	u16 crc;
3806	u32 temp_wucsr;
3807	u32 temp_pmt_ctl;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3808	const u8 ipv4_multicast[3] = { 0x01, 0x00, 0x5E };
3809	const u8 ipv6_multicast[3] = { 0x33, 0x33 };
3810	const u8 arp_type[2] = { 0x08, 0x06 };
 
 
 
 
 
 
3811
3812	lan78xx_read_reg(dev, MAC_TX, &buf);
3813	buf &= ~MAC_TX_TXEN_;
3814	lan78xx_write_reg(dev, MAC_TX, buf);
3815	lan78xx_read_reg(dev, MAC_RX, &buf);
3816	buf &= ~MAC_RX_RXEN_;
3817	lan78xx_write_reg(dev, MAC_RX, buf);
3818
3819	lan78xx_write_reg(dev, WUCSR, 0);
3820	lan78xx_write_reg(dev, WUCSR2, 0);
3821	lan78xx_write_reg(dev, WK_SRC, 0xFFF1FF1FUL);
 
 
 
 
 
 
3822
3823	temp_wucsr = 0;
3824
3825	temp_pmt_ctl = 0;
3826	lan78xx_read_reg(dev, PMT_CTL, &temp_pmt_ctl);
 
 
 
 
3827	temp_pmt_ctl &= ~PMT_CTL_RES_CLR_WKP_EN_;
3828	temp_pmt_ctl |= PMT_CTL_RES_CLR_WKP_STS_;
3829
3830	for (mask_index = 0; mask_index < NUM_OF_WUF_CFG; mask_index++)
3831		lan78xx_write_reg(dev, WUF_CFG(mask_index), 0);
 
 
 
3832
3833	mask_index = 0;
3834	if (wol & WAKE_PHY) {
3835		temp_pmt_ctl |= PMT_CTL_PHY_WAKE_EN_;
3836
3837		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3838		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3839		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
3840	}
3841	if (wol & WAKE_MAGIC) {
3842		temp_wucsr |= WUCSR_MPEN_;
3843
3844		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3845		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3846		temp_pmt_ctl |= PMT_CTL_SUS_MODE_3_;
3847	}
3848	if (wol & WAKE_BCAST) {
3849		temp_wucsr |= WUCSR_BCST_EN_;
3850
3851		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3852		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3853		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
3854	}
3855	if (wol & WAKE_MCAST) {
3856		temp_wucsr |= WUCSR_WAKE_EN_;
3857
3858		/* set WUF_CFG & WUF_MASK for IPv4 Multicast */
3859		crc = lan78xx_wakeframe_crc16(ipv4_multicast, 3);
3860		lan78xx_write_reg(dev, WUF_CFG(mask_index),
3861					WUF_CFGX_EN_ |
3862					WUF_CFGX_TYPE_MCAST_ |
3863					(0 << WUF_CFGX_OFFSET_SHIFT_) |
3864					(crc & WUF_CFGX_CRC16_MASK_));
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3865
3866		lan78xx_write_reg(dev, WUF_MASK0(mask_index), 7);
3867		lan78xx_write_reg(dev, WUF_MASK1(mask_index), 0);
3868		lan78xx_write_reg(dev, WUF_MASK2(mask_index), 0);
3869		lan78xx_write_reg(dev, WUF_MASK3(mask_index), 0);
3870		mask_index++;
3871
3872		/* for IPv6 Multicast */
3873		crc = lan78xx_wakeframe_crc16(ipv6_multicast, 2);
3874		lan78xx_write_reg(dev, WUF_CFG(mask_index),
3875					WUF_CFGX_EN_ |
3876					WUF_CFGX_TYPE_MCAST_ |
3877					(0 << WUF_CFGX_OFFSET_SHIFT_) |
3878					(crc & WUF_CFGX_CRC16_MASK_));
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3879
3880		lan78xx_write_reg(dev, WUF_MASK0(mask_index), 3);
3881		lan78xx_write_reg(dev, WUF_MASK1(mask_index), 0);
3882		lan78xx_write_reg(dev, WUF_MASK2(mask_index), 0);
3883		lan78xx_write_reg(dev, WUF_MASK3(mask_index), 0);
3884		mask_index++;
3885
3886		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3887		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3888		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
3889	}
3890	if (wol & WAKE_UCAST) {
3891		temp_wucsr |= WUCSR_PFDA_EN_;
3892
3893		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3894		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3895		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
3896	}
3897	if (wol & WAKE_ARP) {
3898		temp_wucsr |= WUCSR_WAKE_EN_;
3899
3900		/* set WUF_CFG & WUF_MASK
3901		 * for packettype (offset 12,13) = ARP (0x0806)
3902		 */
3903		crc = lan78xx_wakeframe_crc16(arp_type, 2);
3904		lan78xx_write_reg(dev, WUF_CFG(mask_index),
3905					WUF_CFGX_EN_ |
3906					WUF_CFGX_TYPE_ALL_ |
3907					(0 << WUF_CFGX_OFFSET_SHIFT_) |
3908					(crc & WUF_CFGX_CRC16_MASK_));
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3909
3910		lan78xx_write_reg(dev, WUF_MASK0(mask_index), 0x3000);
3911		lan78xx_write_reg(dev, WUF_MASK1(mask_index), 0);
3912		lan78xx_write_reg(dev, WUF_MASK2(mask_index), 0);
3913		lan78xx_write_reg(dev, WUF_MASK3(mask_index), 0);
3914		mask_index++;
3915
3916		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3917		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3918		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
3919	}
3920
3921	lan78xx_write_reg(dev, WUCSR, temp_wucsr);
 
 
3922
3923	/* when multiple WOL bits are set */
3924	if (hweight_long((unsigned long)wol) > 1) {
3925		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
3926		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
3927		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
3928	}
3929	lan78xx_write_reg(dev, PMT_CTL, temp_pmt_ctl);
 
 
3930
3931	/* clear WUPS */
3932	lan78xx_read_reg(dev, PMT_CTL, &buf);
 
 
 
3933	buf |= PMT_CTL_WUPS_MASK_;
3934	lan78xx_write_reg(dev, PMT_CTL, buf);
3935
3936	lan78xx_read_reg(dev, MAC_RX, &buf);
3937	buf |= MAC_RX_RXEN_;
3938	lan78xx_write_reg(dev, MAC_RX, buf);
3939
3940	return 0;
 
 
3941}
3942
3943static int lan78xx_suspend(struct usb_interface *intf, pm_message_t message)
3944{
3945	struct lan78xx_net *dev = usb_get_intfdata(intf);
3946	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
3947	u32 buf;
3948	int ret;
3949
3950	if (!dev->suspend_count++) {
 
 
 
 
 
 
 
3951		spin_lock_irq(&dev->txq.lock);
3952		/* don't autosuspend while transmitting */
3953		if ((skb_queue_len(&dev->txq) ||
3954		     skb_queue_len(&dev->txq_pend)) &&
3955			PMSG_IS_AUTO(message)) {
3956			spin_unlock_irq(&dev->txq.lock);
3957			ret = -EBUSY;
3958			goto out;
3959		} else {
3960			set_bit(EVENT_DEV_ASLEEP, &dev->flags);
3961			spin_unlock_irq(&dev->txq.lock);
3962		}
3963
3964		/* stop TX & RX */
3965		ret = lan78xx_read_reg(dev, MAC_TX, &buf);
3966		buf &= ~MAC_TX_TXEN_;
3967		ret = lan78xx_write_reg(dev, MAC_TX, buf);
3968		ret = lan78xx_read_reg(dev, MAC_RX, &buf);
3969		buf &= ~MAC_RX_RXEN_;
3970		ret = lan78xx_write_reg(dev, MAC_RX, buf);
3971
3972		/* empty out the rx and queues */
 
 
 
 
 
 
 
 
 
3973		netif_device_detach(dev->net);
3974		lan78xx_terminate_urbs(dev);
3975		usb_kill_urb(dev->urb_intr);
3976
3977		/* reattach */
3978		netif_device_attach(dev->net);
3979	}
3980
3981	if (test_bit(EVENT_DEV_ASLEEP, &dev->flags)) {
3982		del_timer(&dev->stat_monitor);
3983
3984		if (PMSG_IS_AUTO(message)) {
3985			/* auto suspend (selective suspend) */
3986			ret = lan78xx_read_reg(dev, MAC_TX, &buf);
3987			buf &= ~MAC_TX_TXEN_;
3988			ret = lan78xx_write_reg(dev, MAC_TX, buf);
3989			ret = lan78xx_read_reg(dev, MAC_RX, &buf);
3990			buf &= ~MAC_RX_RXEN_;
3991			ret = lan78xx_write_reg(dev, MAC_RX, buf);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3992
3993			ret = lan78xx_write_reg(dev, WUCSR, 0);
3994			ret = lan78xx_write_reg(dev, WUCSR2, 0);
3995			ret = lan78xx_write_reg(dev, WK_SRC, 0xFFF1FF1FUL);
 
3996
3997			/* set goodframe wakeup */
3998			ret = lan78xx_read_reg(dev, WUCSR, &buf);
 
3999
4000			buf |= WUCSR_RFE_WAKE_EN_;
4001			buf |= WUCSR_STORE_WAKE_;
 
4002
4003			ret = lan78xx_write_reg(dev, WUCSR, buf);
4004
4005			ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
 
 
 
4006
4007			buf &= ~PMT_CTL_RES_CLR_WKP_EN_;
4008			buf |= PMT_CTL_RES_CLR_WKP_STS_;
 
4009
4010			buf |= PMT_CTL_PHY_WAKE_EN_;
4011			buf |= PMT_CTL_WOL_EN_;
4012			buf &= ~PMT_CTL_SUS_MODE_MASK_;
4013			buf |= PMT_CTL_SUS_MODE_3_;
4014
4015			ret = lan78xx_write_reg(dev, PMT_CTL, buf);
 
 
 
4016
4017			ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
 
 
4018
4019			buf |= PMT_CTL_WUPS_MASK_;
 
 
 
 
 
4020
4021			ret = lan78xx_write_reg(dev, PMT_CTL, buf);
4022
4023			ret = lan78xx_read_reg(dev, MAC_RX, &buf);
4024			buf |= MAC_RX_RXEN_;
4025			ret = lan78xx_write_reg(dev, MAC_RX, buf);
4026		} else {
4027			lan78xx_set_suspend(dev, pdata->wol);
 
 
 
 
 
 
 
4028		}
4029	}
4030
4031	ret = 0;
4032out:
4033	return ret;
4034}
4035
4036static int lan78xx_resume(struct usb_interface *intf)
4037{
4038	struct lan78xx_net *dev = usb_get_intfdata(intf);
4039	struct sk_buff *skb;
4040	struct urb *res;
4041	int ret;
4042	u32 buf;
4043
4044	if (!timer_pending(&dev->stat_monitor)) {
4045		dev->delta = 1;
4046		mod_timer(&dev->stat_monitor,
4047			  jiffies + STAT_UPDATE_TIMER);
4048	}
4049
4050	if (!--dev->suspend_count) {
4051		/* resume interrupt URBs */
4052		if (dev->urb_intr && test_bit(EVENT_DEV_OPEN, &dev->flags))
4053				usb_submit_urb(dev->urb_intr, GFP_NOIO);
 
 
 
 
 
 
 
 
 
4054
4055		spin_lock_irq(&dev->txq.lock);
4056		while ((res = usb_get_from_anchor(&dev->deferred))) {
4057			skb = (struct sk_buff *)res->context;
4058			ret = usb_submit_urb(res, GFP_ATOMIC);
4059			if (ret < 0) {
4060				dev_kfree_skb_any(skb);
4061				usb_free_urb(res);
4062				usb_autopm_put_interface_async(dev->intf);
4063			} else {
4064				netif_trans_update(dev->net);
4065				lan78xx_queue_skb(&dev->txq, skb, tx_start);
4066			}
4067		}
4068
 
 
 
 
 
 
 
 
 
4069		clear_bit(EVENT_DEV_ASLEEP, &dev->flags);
 
4070		spin_unlock_irq(&dev->txq.lock);
4071
4072		if (test_bit(EVENT_DEV_OPEN, &dev->flags)) {
4073			if (!(skb_queue_len(&dev->txq) >= dev->tx_qlen))
4074				netif_start_queue(dev->net);
4075			tasklet_schedule(&dev->bh);
 
 
 
 
 
 
 
 
 
 
 
4076		}
 
 
 
4077	}
4078
4079	ret = lan78xx_write_reg(dev, WUCSR2, 0);
 
 
4080	ret = lan78xx_write_reg(dev, WUCSR, 0);
 
 
4081	ret = lan78xx_write_reg(dev, WK_SRC, 0xFFF1FF1FUL);
 
 
4082
4083	ret = lan78xx_write_reg(dev, WUCSR2, WUCSR2_NS_RCD_ |
4084					     WUCSR2_ARP_RCD_ |
4085					     WUCSR2_IPV6_TCPSYN_RCD_ |
4086					     WUCSR2_IPV4_TCPSYN_RCD_);
 
 
4087
4088	ret = lan78xx_write_reg(dev, WUCSR, WUCSR_EEE_TX_WAKE_ |
4089					    WUCSR_EEE_RX_WAKE_ |
4090					    WUCSR_PFDA_FR_ |
4091					    WUCSR_RFE_WAKE_FR_ |
4092					    WUCSR_WUFR_ |
4093					    WUCSR_MPR_ |
4094					    WUCSR_BCST_FR_);
 
 
4095
4096	ret = lan78xx_read_reg(dev, MAC_TX, &buf);
4097	buf |= MAC_TX_TXEN_;
4098	ret = lan78xx_write_reg(dev, MAC_TX, buf);
4099
4100	return 0;
4101}
4102
4103static int lan78xx_reset_resume(struct usb_interface *intf)
4104{
4105	struct lan78xx_net *dev = usb_get_intfdata(intf);
 
4106
4107	lan78xx_reset(dev);
 
 
 
 
4108
4109	phy_start(dev->net->phydev);
4110
4111	return lan78xx_resume(intf);
 
 
4112}
4113
4114static const struct usb_device_id products[] = {
4115	{
4116	/* LAN7800 USB Gigabit Ethernet Device */
4117	USB_DEVICE(LAN78XX_USB_VENDOR_ID, LAN7800_USB_PRODUCT_ID),
4118	},
4119	{
4120	/* LAN7850 USB Gigabit Ethernet Device */
4121	USB_DEVICE(LAN78XX_USB_VENDOR_ID, LAN7850_USB_PRODUCT_ID),
4122	},
4123	{
4124	/* LAN7801 USB Gigabit Ethernet Device */
4125	USB_DEVICE(LAN78XX_USB_VENDOR_ID, LAN7801_USB_PRODUCT_ID),
 
 
 
 
4126	},
4127	{},
4128};
4129MODULE_DEVICE_TABLE(usb, products);
4130
4131static struct usb_driver lan78xx_driver = {
4132	.name			= DRIVER_NAME,
4133	.id_table		= products,
4134	.probe			= lan78xx_probe,
4135	.disconnect		= lan78xx_disconnect,
4136	.suspend		= lan78xx_suspend,
4137	.resume			= lan78xx_resume,
4138	.reset_resume		= lan78xx_reset_resume,
4139	.supports_autosuspend	= 1,
4140	.disable_hub_initiated_lpm = 1,
4141};
4142
4143module_usb_driver(lan78xx_driver);
4144
4145MODULE_AUTHOR(DRIVER_AUTHOR);
4146MODULE_DESCRIPTION(DRIVER_DESC);
4147MODULE_LICENSE("GPL");
v6.2
   1// SPDX-License-Identifier: GPL-2.0+
   2/*
   3 * Copyright (C) 2015 Microchip Technology
   4 */
   5#include <linux/module.h>
   6#include <linux/netdevice.h>
   7#include <linux/etherdevice.h>
   8#include <linux/ethtool.h>
   9#include <linux/usb.h>
  10#include <linux/crc32.h>
  11#include <linux/signal.h>
  12#include <linux/slab.h>
  13#include <linux/if_vlan.h>
  14#include <linux/uaccess.h>
  15#include <linux/linkmode.h>
  16#include <linux/list.h>
  17#include <linux/ip.h>
  18#include <linux/ipv6.h>
  19#include <linux/mdio.h>
  20#include <linux/phy.h>
  21#include <net/ip6_checksum.h>
  22#include <net/vxlan.h>
  23#include <linux/interrupt.h>
  24#include <linux/irqdomain.h>
  25#include <linux/irq.h>
  26#include <linux/irqchip/chained_irq.h>
  27#include <linux/microchipphy.h>
  28#include <linux/phy_fixed.h>
  29#include <linux/of_mdio.h>
  30#include <linux/of_net.h>
  31#include "lan78xx.h"
  32
  33#define DRIVER_AUTHOR	"WOOJUNG HUH <woojung.huh@microchip.com>"
  34#define DRIVER_DESC	"LAN78XX USB 3.0 Gigabit Ethernet Devices"
  35#define DRIVER_NAME	"lan78xx"
  36
  37#define TX_TIMEOUT_JIFFIES		(5 * HZ)
  38#define THROTTLE_JIFFIES		(HZ / 8)
  39#define UNLINK_TIMEOUT_MS		3
  40
  41#define RX_MAX_QUEUE_MEMORY		(60 * 1518)
  42
  43#define SS_USB_PKT_SIZE			(1024)
  44#define HS_USB_PKT_SIZE			(512)
  45#define FS_USB_PKT_SIZE			(64)
  46
  47#define MAX_RX_FIFO_SIZE		(12 * 1024)
  48#define MAX_TX_FIFO_SIZE		(12 * 1024)
  49
  50#define FLOW_THRESHOLD(n)		((((n) + 511) / 512) & 0x7F)
  51#define FLOW_CTRL_THRESHOLD(on, off)	((FLOW_THRESHOLD(on)  << 0) | \
  52					 (FLOW_THRESHOLD(off) << 8))
  53
  54/* Flow control turned on when Rx FIFO level rises above this level (bytes) */
  55#define FLOW_ON_SS			9216
  56#define FLOW_ON_HS			8704
  57
  58/* Flow control turned off when Rx FIFO level falls below this level (bytes) */
  59#define FLOW_OFF_SS			4096
  60#define FLOW_OFF_HS			1024
  61
  62#define DEFAULT_BURST_CAP_SIZE		(MAX_TX_FIFO_SIZE)
  63#define DEFAULT_BULK_IN_DELAY		(0x0800)
  64#define MAX_SINGLE_PACKET_SIZE		(9000)
  65#define DEFAULT_TX_CSUM_ENABLE		(true)
  66#define DEFAULT_RX_CSUM_ENABLE		(true)
  67#define DEFAULT_TSO_CSUM_ENABLE		(true)
  68#define DEFAULT_VLAN_FILTER_ENABLE	(true)
  69#define DEFAULT_VLAN_RX_OFFLOAD		(true)
  70#define TX_ALIGNMENT			(4)
  71#define RXW_PADDING			2
  72
  73#define LAN78XX_USB_VENDOR_ID		(0x0424)
  74#define LAN7800_USB_PRODUCT_ID		(0x7800)
  75#define LAN7850_USB_PRODUCT_ID		(0x7850)
  76#define LAN7801_USB_PRODUCT_ID		(0x7801)
  77#define LAN78XX_EEPROM_MAGIC		(0x78A5)
  78#define LAN78XX_OTP_MAGIC		(0x78F3)
  79#define AT29M2AF_USB_VENDOR_ID		(0x07C9)
  80#define AT29M2AF_USB_PRODUCT_ID	(0x0012)
  81
  82#define	MII_READ			1
  83#define	MII_WRITE			0
  84
  85#define EEPROM_INDICATOR		(0xA5)
  86#define EEPROM_MAC_OFFSET		(0x01)
  87#define MAX_EEPROM_SIZE			512
  88#define OTP_INDICATOR_1			(0xF3)
  89#define OTP_INDICATOR_2			(0xF7)
  90
  91#define WAKE_ALL			(WAKE_PHY | WAKE_UCAST | \
  92					 WAKE_MCAST | WAKE_BCAST | \
  93					 WAKE_ARP | WAKE_MAGIC)
  94
  95#define TX_URB_NUM			10
  96#define TX_SS_URB_NUM			TX_URB_NUM
  97#define TX_HS_URB_NUM			TX_URB_NUM
  98#define TX_FS_URB_NUM			TX_URB_NUM
  99
 100/* A single URB buffer must be large enough to hold a complete jumbo packet
 101 */
 102#define TX_SS_URB_SIZE			(32 * 1024)
 103#define TX_HS_URB_SIZE			(16 * 1024)
 104#define TX_FS_URB_SIZE			(10 * 1024)
 105
 106#define RX_SS_URB_NUM			30
 107#define RX_HS_URB_NUM			10
 108#define RX_FS_URB_NUM			10
 109#define RX_SS_URB_SIZE			TX_SS_URB_SIZE
 110#define RX_HS_URB_SIZE			TX_HS_URB_SIZE
 111#define RX_FS_URB_SIZE			TX_FS_URB_SIZE
 112
 113#define SS_BURST_CAP_SIZE		RX_SS_URB_SIZE
 114#define SS_BULK_IN_DELAY		0x2000
 115#define HS_BURST_CAP_SIZE		RX_HS_URB_SIZE
 116#define HS_BULK_IN_DELAY		0x2000
 117#define FS_BURST_CAP_SIZE		RX_FS_URB_SIZE
 118#define FS_BULK_IN_DELAY		0x2000
 119
 120#define TX_CMD_LEN			8
 121#define TX_SKB_MIN_LEN			(TX_CMD_LEN + ETH_HLEN)
 122#define LAN78XX_TSO_SIZE(dev)		((dev)->tx_urb_size - TX_SKB_MIN_LEN)
 123
 124#define RX_CMD_LEN			10
 125#define RX_SKB_MIN_LEN			(RX_CMD_LEN + ETH_HLEN)
 126#define RX_MAX_FRAME_LEN(mtu)		((mtu) + ETH_HLEN + VLAN_HLEN)
 127
 128/* USB related defines */
 129#define BULK_IN_PIPE			1
 130#define BULK_OUT_PIPE			2
 131
 132/* default autosuspend delay (mSec)*/
 133#define DEFAULT_AUTOSUSPEND_DELAY	(10 * 1000)
 134
 135/* statistic update interval (mSec) */
 136#define STAT_UPDATE_TIMER		(1 * 1000)
 137
 138/* time to wait for MAC or FCT to stop (jiffies) */
 139#define HW_DISABLE_TIMEOUT		(HZ / 10)
 140
 141/* time to wait between polling MAC or FCT state (ms) */
 142#define HW_DISABLE_DELAY_MS		1
 143
 144/* defines interrupts from interrupt EP */
 145#define MAX_INT_EP			(32)
 146#define INT_EP_INTEP			(31)
 147#define INT_EP_OTP_WR_DONE		(28)
 148#define INT_EP_EEE_TX_LPI_START		(26)
 149#define INT_EP_EEE_TX_LPI_STOP		(25)
 150#define INT_EP_EEE_RX_LPI		(24)
 151#define INT_EP_MAC_RESET_TIMEOUT	(23)
 152#define INT_EP_RDFO			(22)
 153#define INT_EP_TXE			(21)
 154#define INT_EP_USB_STATUS		(20)
 155#define INT_EP_TX_DIS			(19)
 156#define INT_EP_RX_DIS			(18)
 157#define INT_EP_PHY			(17)
 158#define INT_EP_DP			(16)
 159#define INT_EP_MAC_ERR			(15)
 160#define INT_EP_TDFU			(14)
 161#define INT_EP_TDFO			(13)
 162#define INT_EP_UTX			(12)
 163#define INT_EP_GPIO_11			(11)
 164#define INT_EP_GPIO_10			(10)
 165#define INT_EP_GPIO_9			(9)
 166#define INT_EP_GPIO_8			(8)
 167#define INT_EP_GPIO_7			(7)
 168#define INT_EP_GPIO_6			(6)
 169#define INT_EP_GPIO_5			(5)
 170#define INT_EP_GPIO_4			(4)
 171#define INT_EP_GPIO_3			(3)
 172#define INT_EP_GPIO_2			(2)
 173#define INT_EP_GPIO_1			(1)
 174#define INT_EP_GPIO_0			(0)
 175
 176static const char lan78xx_gstrings[][ETH_GSTRING_LEN] = {
 177	"RX FCS Errors",
 178	"RX Alignment Errors",
 179	"Rx Fragment Errors",
 180	"RX Jabber Errors",
 181	"RX Undersize Frame Errors",
 182	"RX Oversize Frame Errors",
 183	"RX Dropped Frames",
 184	"RX Unicast Byte Count",
 185	"RX Broadcast Byte Count",
 186	"RX Multicast Byte Count",
 187	"RX Unicast Frames",
 188	"RX Broadcast Frames",
 189	"RX Multicast Frames",
 190	"RX Pause Frames",
 191	"RX 64 Byte Frames",
 192	"RX 65 - 127 Byte Frames",
 193	"RX 128 - 255 Byte Frames",
 194	"RX 256 - 511 Bytes Frames",
 195	"RX 512 - 1023 Byte Frames",
 196	"RX 1024 - 1518 Byte Frames",
 197	"RX Greater 1518 Byte Frames",
 198	"EEE RX LPI Transitions",
 199	"EEE RX LPI Time",
 200	"TX FCS Errors",
 201	"TX Excess Deferral Errors",
 202	"TX Carrier Errors",
 203	"TX Bad Byte Count",
 204	"TX Single Collisions",
 205	"TX Multiple Collisions",
 206	"TX Excessive Collision",
 207	"TX Late Collisions",
 208	"TX Unicast Byte Count",
 209	"TX Broadcast Byte Count",
 210	"TX Multicast Byte Count",
 211	"TX Unicast Frames",
 212	"TX Broadcast Frames",
 213	"TX Multicast Frames",
 214	"TX Pause Frames",
 215	"TX 64 Byte Frames",
 216	"TX 65 - 127 Byte Frames",
 217	"TX 128 - 255 Byte Frames",
 218	"TX 256 - 511 Bytes Frames",
 219	"TX 512 - 1023 Byte Frames",
 220	"TX 1024 - 1518 Byte Frames",
 221	"TX Greater 1518 Byte Frames",
 222	"EEE TX LPI Transitions",
 223	"EEE TX LPI Time",
 224};
 225
 226struct lan78xx_statstage {
 227	u32 rx_fcs_errors;
 228	u32 rx_alignment_errors;
 229	u32 rx_fragment_errors;
 230	u32 rx_jabber_errors;
 231	u32 rx_undersize_frame_errors;
 232	u32 rx_oversize_frame_errors;
 233	u32 rx_dropped_frames;
 234	u32 rx_unicast_byte_count;
 235	u32 rx_broadcast_byte_count;
 236	u32 rx_multicast_byte_count;
 237	u32 rx_unicast_frames;
 238	u32 rx_broadcast_frames;
 239	u32 rx_multicast_frames;
 240	u32 rx_pause_frames;
 241	u32 rx_64_byte_frames;
 242	u32 rx_65_127_byte_frames;
 243	u32 rx_128_255_byte_frames;
 244	u32 rx_256_511_bytes_frames;
 245	u32 rx_512_1023_byte_frames;
 246	u32 rx_1024_1518_byte_frames;
 247	u32 rx_greater_1518_byte_frames;
 248	u32 eee_rx_lpi_transitions;
 249	u32 eee_rx_lpi_time;
 250	u32 tx_fcs_errors;
 251	u32 tx_excess_deferral_errors;
 252	u32 tx_carrier_errors;
 253	u32 tx_bad_byte_count;
 254	u32 tx_single_collisions;
 255	u32 tx_multiple_collisions;
 256	u32 tx_excessive_collision;
 257	u32 tx_late_collisions;
 258	u32 tx_unicast_byte_count;
 259	u32 tx_broadcast_byte_count;
 260	u32 tx_multicast_byte_count;
 261	u32 tx_unicast_frames;
 262	u32 tx_broadcast_frames;
 263	u32 tx_multicast_frames;
 264	u32 tx_pause_frames;
 265	u32 tx_64_byte_frames;
 266	u32 tx_65_127_byte_frames;
 267	u32 tx_128_255_byte_frames;
 268	u32 tx_256_511_bytes_frames;
 269	u32 tx_512_1023_byte_frames;
 270	u32 tx_1024_1518_byte_frames;
 271	u32 tx_greater_1518_byte_frames;
 272	u32 eee_tx_lpi_transitions;
 273	u32 eee_tx_lpi_time;
 274};
 275
 276struct lan78xx_statstage64 {
 277	u64 rx_fcs_errors;
 278	u64 rx_alignment_errors;
 279	u64 rx_fragment_errors;
 280	u64 rx_jabber_errors;
 281	u64 rx_undersize_frame_errors;
 282	u64 rx_oversize_frame_errors;
 283	u64 rx_dropped_frames;
 284	u64 rx_unicast_byte_count;
 285	u64 rx_broadcast_byte_count;
 286	u64 rx_multicast_byte_count;
 287	u64 rx_unicast_frames;
 288	u64 rx_broadcast_frames;
 289	u64 rx_multicast_frames;
 290	u64 rx_pause_frames;
 291	u64 rx_64_byte_frames;
 292	u64 rx_65_127_byte_frames;
 293	u64 rx_128_255_byte_frames;
 294	u64 rx_256_511_bytes_frames;
 295	u64 rx_512_1023_byte_frames;
 296	u64 rx_1024_1518_byte_frames;
 297	u64 rx_greater_1518_byte_frames;
 298	u64 eee_rx_lpi_transitions;
 299	u64 eee_rx_lpi_time;
 300	u64 tx_fcs_errors;
 301	u64 tx_excess_deferral_errors;
 302	u64 tx_carrier_errors;
 303	u64 tx_bad_byte_count;
 304	u64 tx_single_collisions;
 305	u64 tx_multiple_collisions;
 306	u64 tx_excessive_collision;
 307	u64 tx_late_collisions;
 308	u64 tx_unicast_byte_count;
 309	u64 tx_broadcast_byte_count;
 310	u64 tx_multicast_byte_count;
 311	u64 tx_unicast_frames;
 312	u64 tx_broadcast_frames;
 313	u64 tx_multicast_frames;
 314	u64 tx_pause_frames;
 315	u64 tx_64_byte_frames;
 316	u64 tx_65_127_byte_frames;
 317	u64 tx_128_255_byte_frames;
 318	u64 tx_256_511_bytes_frames;
 319	u64 tx_512_1023_byte_frames;
 320	u64 tx_1024_1518_byte_frames;
 321	u64 tx_greater_1518_byte_frames;
 322	u64 eee_tx_lpi_transitions;
 323	u64 eee_tx_lpi_time;
 324};
 325
 326static u32 lan78xx_regs[] = {
 327	ID_REV,
 328	INT_STS,
 329	HW_CFG,
 330	PMT_CTL,
 331	E2P_CMD,
 332	E2P_DATA,
 333	USB_STATUS,
 334	VLAN_TYPE,
 335	MAC_CR,
 336	MAC_RX,
 337	MAC_TX,
 338	FLOW,
 339	ERR_STS,
 340	MII_ACC,
 341	MII_DATA,
 342	EEE_TX_LPI_REQ_DLY,
 343	EEE_TW_TX_SYS,
 344	EEE_TX_LPI_REM_DLY,
 345	WUCSR
 346};
 347
 348#define PHY_REG_SIZE (32 * sizeof(u32))
 349
 350struct lan78xx_net;
 351
 352struct lan78xx_priv {
 353	struct lan78xx_net *dev;
 354	u32 rfe_ctl;
 355	u32 mchash_table[DP_SEL_VHF_HASH_LEN]; /* multicast hash table */
 356	u32 pfilter_table[NUM_OF_MAF][2]; /* perfect filter table */
 357	u32 vlan_table[DP_SEL_VHF_VLAN_LEN];
 358	struct mutex dataport_mutex; /* for dataport access */
 359	spinlock_t rfe_ctl_lock; /* for rfe register access */
 360	struct work_struct set_multicast;
 361	struct work_struct set_vlan;
 362	u32 wol;
 363};
 364
 365enum skb_state {
 366	illegal = 0,
 367	tx_start,
 368	tx_done,
 369	rx_start,
 370	rx_done,
 371	rx_cleanup,
 372	unlink_start
 373};
 374
 375struct skb_data {		/* skb->cb is one of these */
 376	struct urb *urb;
 377	struct lan78xx_net *dev;
 378	enum skb_state state;
 379	size_t length;
 380	int num_of_packet;
 381};
 382
 383struct usb_context {
 384	struct usb_ctrlrequest req;
 385	struct lan78xx_net *dev;
 386};
 387
 388#define EVENT_TX_HALT			0
 389#define EVENT_RX_HALT			1
 390#define EVENT_RX_MEMORY			2
 391#define EVENT_STS_SPLIT			3
 392#define EVENT_LINK_RESET		4
 393#define EVENT_RX_PAUSED			5
 394#define EVENT_DEV_WAKING		6
 395#define EVENT_DEV_ASLEEP		7
 396#define EVENT_DEV_OPEN			8
 397#define EVENT_STAT_UPDATE		9
 398#define EVENT_DEV_DISCONNECT		10
 399
 400struct statstage {
 401	struct mutex			access_lock;	/* for stats access */
 402	struct lan78xx_statstage	saved;
 403	struct lan78xx_statstage	rollover_count;
 404	struct lan78xx_statstage	rollover_max;
 405	struct lan78xx_statstage64	curr_stat;
 406};
 407
 408struct irq_domain_data {
 409	struct irq_domain	*irqdomain;
 410	unsigned int		phyirq;
 411	struct irq_chip		*irqchip;
 412	irq_flow_handler_t	irq_handler;
 413	u32			irqenable;
 414	struct mutex		irq_lock;		/* for irq bus access */
 415};
 416
 417struct lan78xx_net {
 418	struct net_device	*net;
 419	struct usb_device	*udev;
 420	struct usb_interface	*intf;
 421	void			*driver_priv;
 422
 423	unsigned int		tx_pend_data_len;
 424	size_t			n_tx_urbs;
 425	size_t			n_rx_urbs;
 426	size_t			tx_urb_size;
 427	size_t			rx_urb_size;
 428
 429	struct sk_buff_head	rxq_free;
 430	struct sk_buff_head	rxq;
 431	struct sk_buff_head	rxq_done;
 432	struct sk_buff_head	rxq_overflow;
 433	struct sk_buff_head	txq_free;
 434	struct sk_buff_head	txq;
 
 
 435	struct sk_buff_head	txq_pend;
 436
 437	struct napi_struct	napi;
 438
 439	struct delayed_work	wq;
 440
 441	int			msg_enable;
 442
 443	struct urb		*urb_intr;
 444	struct usb_anchor	deferred;
 445
 446	struct mutex		dev_mutex; /* serialise open/stop wrt suspend/resume */
 447	struct mutex		phy_mutex; /* for phy access */
 448	unsigned int		pipe_in, pipe_out, pipe_intr;
 449
 450	unsigned int		bulk_in_delay;
 451	unsigned int		burst_cap;
 452
 453	unsigned long		flags;
 454
 455	wait_queue_head_t	*wait;
 456	unsigned char		suspend_count;
 457
 458	unsigned int		maxpacket;
 
 459	struct timer_list	stat_monitor;
 460
 461	unsigned long		data[5];
 462
 463	int			link_on;
 464	u8			mdix_ctrl;
 465
 466	u32			chipid;
 467	u32			chiprev;
 468	struct mii_bus		*mdiobus;
 469	phy_interface_t		interface;
 470
 471	int			fc_autoneg;
 472	u8			fc_request_control;
 473
 474	int			delta;
 475	struct statstage	stats;
 476
 477	struct irq_domain_data	domain_data;
 478};
 479
 480/* define external phy id */
 481#define	PHY_LAN8835			(0x0007C130)
 482#define	PHY_KSZ9031RNX			(0x00221620)
 483
 484/* use ethtool to change the level for any given device */
 485static int msg_level = -1;
 486module_param(msg_level, int, 0);
 487MODULE_PARM_DESC(msg_level, "Override default message level");
 488
 489static struct sk_buff *lan78xx_get_buf(struct sk_buff_head *buf_pool)
 490{
 491	if (skb_queue_empty(buf_pool))
 492		return NULL;
 493
 494	return skb_dequeue(buf_pool);
 495}
 496
 497static void lan78xx_release_buf(struct sk_buff_head *buf_pool,
 498				struct sk_buff *buf)
 499{
 500	buf->data = buf->head;
 501	skb_reset_tail_pointer(buf);
 502
 503	buf->len = 0;
 504	buf->data_len = 0;
 505
 506	skb_queue_tail(buf_pool, buf);
 507}
 508
 509static void lan78xx_free_buf_pool(struct sk_buff_head *buf_pool)
 510{
 511	struct skb_data *entry;
 512	struct sk_buff *buf;
 513
 514	while (!skb_queue_empty(buf_pool)) {
 515		buf = skb_dequeue(buf_pool);
 516		if (buf) {
 517			entry = (struct skb_data *)buf->cb;
 518			usb_free_urb(entry->urb);
 519			dev_kfree_skb_any(buf);
 520		}
 521	}
 522}
 523
 524static int lan78xx_alloc_buf_pool(struct sk_buff_head *buf_pool,
 525				  size_t n_urbs, size_t urb_size,
 526				  struct lan78xx_net *dev)
 527{
 528	struct skb_data *entry;
 529	struct sk_buff *buf;
 530	struct urb *urb;
 531	int i;
 532
 533	skb_queue_head_init(buf_pool);
 534
 535	for (i = 0; i < n_urbs; i++) {
 536		buf = alloc_skb(urb_size, GFP_ATOMIC);
 537		if (!buf)
 538			goto error;
 539
 540		if (skb_linearize(buf) != 0) {
 541			dev_kfree_skb_any(buf);
 542			goto error;
 543		}
 544
 545		urb = usb_alloc_urb(0, GFP_ATOMIC);
 546		if (!urb) {
 547			dev_kfree_skb_any(buf);
 548			goto error;
 549		}
 550
 551		entry = (struct skb_data *)buf->cb;
 552		entry->urb = urb;
 553		entry->dev = dev;
 554		entry->length = 0;
 555		entry->num_of_packet = 0;
 556
 557		skb_queue_tail(buf_pool, buf);
 558	}
 559
 560	return 0;
 561
 562error:
 563	lan78xx_free_buf_pool(buf_pool);
 564
 565	return -ENOMEM;
 566}
 567
 568static struct sk_buff *lan78xx_get_rx_buf(struct lan78xx_net *dev)
 569{
 570	return lan78xx_get_buf(&dev->rxq_free);
 571}
 572
 573static void lan78xx_release_rx_buf(struct lan78xx_net *dev,
 574				   struct sk_buff *rx_buf)
 575{
 576	lan78xx_release_buf(&dev->rxq_free, rx_buf);
 577}
 578
 579static void lan78xx_free_rx_resources(struct lan78xx_net *dev)
 580{
 581	lan78xx_free_buf_pool(&dev->rxq_free);
 582}
 583
 584static int lan78xx_alloc_rx_resources(struct lan78xx_net *dev)
 585{
 586	return lan78xx_alloc_buf_pool(&dev->rxq_free,
 587				      dev->n_rx_urbs, dev->rx_urb_size, dev);
 588}
 589
 590static struct sk_buff *lan78xx_get_tx_buf(struct lan78xx_net *dev)
 591{
 592	return lan78xx_get_buf(&dev->txq_free);
 593}
 594
 595static void lan78xx_release_tx_buf(struct lan78xx_net *dev,
 596				   struct sk_buff *tx_buf)
 597{
 598	lan78xx_release_buf(&dev->txq_free, tx_buf);
 599}
 600
 601static void lan78xx_free_tx_resources(struct lan78xx_net *dev)
 602{
 603	lan78xx_free_buf_pool(&dev->txq_free);
 604}
 605
 606static int lan78xx_alloc_tx_resources(struct lan78xx_net *dev)
 607{
 608	return lan78xx_alloc_buf_pool(&dev->txq_free,
 609				      dev->n_tx_urbs, dev->tx_urb_size, dev);
 610}
 611
 612static int lan78xx_read_reg(struct lan78xx_net *dev, u32 index, u32 *data)
 613{
 614	u32 *buf;
 615	int ret;
 616
 617	if (test_bit(EVENT_DEV_DISCONNECT, &dev->flags))
 618		return -ENODEV;
 619
 620	buf = kmalloc(sizeof(u32), GFP_KERNEL);
 621	if (!buf)
 622		return -ENOMEM;
 623
 624	ret = usb_control_msg(dev->udev, usb_rcvctrlpipe(dev->udev, 0),
 625			      USB_VENDOR_REQUEST_READ_REGISTER,
 626			      USB_DIR_IN | USB_TYPE_VENDOR | USB_RECIP_DEVICE,
 627			      0, index, buf, 4, USB_CTRL_GET_TIMEOUT);
 628	if (likely(ret >= 0)) {
 629		le32_to_cpus(buf);
 630		*data = *buf;
 631	} else if (net_ratelimit()) {
 632		netdev_warn(dev->net,
 633			    "Failed to read register index 0x%08x. ret = %d",
 634			    index, ret);
 635	}
 636
 637	kfree(buf);
 638
 639	return ret;
 640}
 641
 642static int lan78xx_write_reg(struct lan78xx_net *dev, u32 index, u32 data)
 643{
 644	u32 *buf;
 645	int ret;
 646
 647	if (test_bit(EVENT_DEV_DISCONNECT, &dev->flags))
 648		return -ENODEV;
 649
 650	buf = kmalloc(sizeof(u32), GFP_KERNEL);
 651	if (!buf)
 652		return -ENOMEM;
 653
 654	*buf = data;
 655	cpu_to_le32s(buf);
 656
 657	ret = usb_control_msg(dev->udev, usb_sndctrlpipe(dev->udev, 0),
 658			      USB_VENDOR_REQUEST_WRITE_REGISTER,
 659			      USB_DIR_OUT | USB_TYPE_VENDOR | USB_RECIP_DEVICE,
 660			      0, index, buf, 4, USB_CTRL_SET_TIMEOUT);
 661	if (unlikely(ret < 0) &&
 662	    net_ratelimit()) {
 663		netdev_warn(dev->net,
 664			    "Failed to write register index 0x%08x. ret = %d",
 665			    index, ret);
 666	}
 667
 668	kfree(buf);
 669
 670	return ret;
 671}
 672
 673static int lan78xx_update_reg(struct lan78xx_net *dev, u32 reg, u32 mask,
 674			      u32 data)
 675{
 676	int ret;
 677	u32 buf;
 678
 679	ret = lan78xx_read_reg(dev, reg, &buf);
 680	if (ret < 0)
 681		return ret;
 682
 683	buf &= ~mask;
 684	buf |= (mask & data);
 685
 686	ret = lan78xx_write_reg(dev, reg, buf);
 687	if (ret < 0)
 688		return ret;
 689
 690	return 0;
 691}
 692
 693static int lan78xx_read_stats(struct lan78xx_net *dev,
 694			      struct lan78xx_statstage *data)
 695{
 696	int ret = 0;
 697	int i;
 698	struct lan78xx_statstage *stats;
 699	u32 *src;
 700	u32 *dst;
 701
 702	stats = kmalloc(sizeof(*stats), GFP_KERNEL);
 703	if (!stats)
 704		return -ENOMEM;
 705
 706	ret = usb_control_msg(dev->udev,
 707			      usb_rcvctrlpipe(dev->udev, 0),
 708			      USB_VENDOR_REQUEST_GET_STATS,
 709			      USB_DIR_IN | USB_TYPE_VENDOR | USB_RECIP_DEVICE,
 710			      0,
 711			      0,
 712			      (void *)stats,
 713			      sizeof(*stats),
 714			      USB_CTRL_SET_TIMEOUT);
 715	if (likely(ret >= 0)) {
 716		src = (u32 *)stats;
 717		dst = (u32 *)data;
 718		for (i = 0; i < sizeof(*stats) / sizeof(u32); i++) {
 719			le32_to_cpus(&src[i]);
 720			dst[i] = src[i];
 721		}
 722	} else {
 723		netdev_warn(dev->net,
 724			    "Failed to read stat ret = %d", ret);
 725	}
 726
 727	kfree(stats);
 728
 729	return ret;
 730}
 731
 732#define check_counter_rollover(struct1, dev_stats, member)		\
 733	do {								\
 734		if ((struct1)->member < (dev_stats).saved.member)	\
 735			(dev_stats).rollover_count.member++;		\
 736	} while (0)
 737
 738static void lan78xx_check_stat_rollover(struct lan78xx_net *dev,
 739					struct lan78xx_statstage *stats)
 740{
 741	check_counter_rollover(stats, dev->stats, rx_fcs_errors);
 742	check_counter_rollover(stats, dev->stats, rx_alignment_errors);
 743	check_counter_rollover(stats, dev->stats, rx_fragment_errors);
 744	check_counter_rollover(stats, dev->stats, rx_jabber_errors);
 745	check_counter_rollover(stats, dev->stats, rx_undersize_frame_errors);
 746	check_counter_rollover(stats, dev->stats, rx_oversize_frame_errors);
 747	check_counter_rollover(stats, dev->stats, rx_dropped_frames);
 748	check_counter_rollover(stats, dev->stats, rx_unicast_byte_count);
 749	check_counter_rollover(stats, dev->stats, rx_broadcast_byte_count);
 750	check_counter_rollover(stats, dev->stats, rx_multicast_byte_count);
 751	check_counter_rollover(stats, dev->stats, rx_unicast_frames);
 752	check_counter_rollover(stats, dev->stats, rx_broadcast_frames);
 753	check_counter_rollover(stats, dev->stats, rx_multicast_frames);
 754	check_counter_rollover(stats, dev->stats, rx_pause_frames);
 755	check_counter_rollover(stats, dev->stats, rx_64_byte_frames);
 756	check_counter_rollover(stats, dev->stats, rx_65_127_byte_frames);
 757	check_counter_rollover(stats, dev->stats, rx_128_255_byte_frames);
 758	check_counter_rollover(stats, dev->stats, rx_256_511_bytes_frames);
 759	check_counter_rollover(stats, dev->stats, rx_512_1023_byte_frames);
 760	check_counter_rollover(stats, dev->stats, rx_1024_1518_byte_frames);
 761	check_counter_rollover(stats, dev->stats, rx_greater_1518_byte_frames);
 762	check_counter_rollover(stats, dev->stats, eee_rx_lpi_transitions);
 763	check_counter_rollover(stats, dev->stats, eee_rx_lpi_time);
 764	check_counter_rollover(stats, dev->stats, tx_fcs_errors);
 765	check_counter_rollover(stats, dev->stats, tx_excess_deferral_errors);
 766	check_counter_rollover(stats, dev->stats, tx_carrier_errors);
 767	check_counter_rollover(stats, dev->stats, tx_bad_byte_count);
 768	check_counter_rollover(stats, dev->stats, tx_single_collisions);
 769	check_counter_rollover(stats, dev->stats, tx_multiple_collisions);
 770	check_counter_rollover(stats, dev->stats, tx_excessive_collision);
 771	check_counter_rollover(stats, dev->stats, tx_late_collisions);
 772	check_counter_rollover(stats, dev->stats, tx_unicast_byte_count);
 773	check_counter_rollover(stats, dev->stats, tx_broadcast_byte_count);
 774	check_counter_rollover(stats, dev->stats, tx_multicast_byte_count);
 775	check_counter_rollover(stats, dev->stats, tx_unicast_frames);
 776	check_counter_rollover(stats, dev->stats, tx_broadcast_frames);
 777	check_counter_rollover(stats, dev->stats, tx_multicast_frames);
 778	check_counter_rollover(stats, dev->stats, tx_pause_frames);
 779	check_counter_rollover(stats, dev->stats, tx_64_byte_frames);
 780	check_counter_rollover(stats, dev->stats, tx_65_127_byte_frames);
 781	check_counter_rollover(stats, dev->stats, tx_128_255_byte_frames);
 782	check_counter_rollover(stats, dev->stats, tx_256_511_bytes_frames);
 783	check_counter_rollover(stats, dev->stats, tx_512_1023_byte_frames);
 784	check_counter_rollover(stats, dev->stats, tx_1024_1518_byte_frames);
 785	check_counter_rollover(stats, dev->stats, tx_greater_1518_byte_frames);
 786	check_counter_rollover(stats, dev->stats, eee_tx_lpi_transitions);
 787	check_counter_rollover(stats, dev->stats, eee_tx_lpi_time);
 788
 789	memcpy(&dev->stats.saved, stats, sizeof(struct lan78xx_statstage));
 790}
 791
 792static void lan78xx_update_stats(struct lan78xx_net *dev)
 793{
 794	u32 *p, *count, *max;
 795	u64 *data;
 796	int i;
 797	struct lan78xx_statstage lan78xx_stats;
 798
 799	if (usb_autopm_get_interface(dev->intf) < 0)
 800		return;
 801
 802	p = (u32 *)&lan78xx_stats;
 803	count = (u32 *)&dev->stats.rollover_count;
 804	max = (u32 *)&dev->stats.rollover_max;
 805	data = (u64 *)&dev->stats.curr_stat;
 806
 807	mutex_lock(&dev->stats.access_lock);
 808
 809	if (lan78xx_read_stats(dev, &lan78xx_stats) > 0)
 810		lan78xx_check_stat_rollover(dev, &lan78xx_stats);
 811
 812	for (i = 0; i < (sizeof(lan78xx_stats) / (sizeof(u32))); i++)
 813		data[i] = (u64)p[i] + ((u64)count[i] * ((u64)max[i] + 1));
 814
 815	mutex_unlock(&dev->stats.access_lock);
 816
 817	usb_autopm_put_interface(dev->intf);
 818}
 819
 820/* Loop until the read is completed with timeout called with phy_mutex held */
 821static int lan78xx_phy_wait_not_busy(struct lan78xx_net *dev)
 822{
 823	unsigned long start_time = jiffies;
 824	u32 val;
 825	int ret;
 826
 827	do {
 828		ret = lan78xx_read_reg(dev, MII_ACC, &val);
 829		if (unlikely(ret < 0))
 830			return -EIO;
 831
 832		if (!(val & MII_ACC_MII_BUSY_))
 833			return 0;
 834	} while (!time_after(jiffies, start_time + HZ));
 835
 836	return -EIO;
 837}
 838
 839static inline u32 mii_access(int id, int index, int read)
 840{
 841	u32 ret;
 842
 843	ret = ((u32)id << MII_ACC_PHY_ADDR_SHIFT_) & MII_ACC_PHY_ADDR_MASK_;
 844	ret |= ((u32)index << MII_ACC_MIIRINDA_SHIFT_) & MII_ACC_MIIRINDA_MASK_;
 845	if (read)
 846		ret |= MII_ACC_MII_READ_;
 847	else
 848		ret |= MII_ACC_MII_WRITE_;
 849	ret |= MII_ACC_MII_BUSY_;
 850
 851	return ret;
 852}
 853
 854static int lan78xx_wait_eeprom(struct lan78xx_net *dev)
 855{
 856	unsigned long start_time = jiffies;
 857	u32 val;
 858	int ret;
 859
 860	do {
 861		ret = lan78xx_read_reg(dev, E2P_CMD, &val);
 862		if (unlikely(ret < 0))
 863			return -EIO;
 864
 865		if (!(val & E2P_CMD_EPC_BUSY_) ||
 866		    (val & E2P_CMD_EPC_TIMEOUT_))
 867			break;
 868		usleep_range(40, 100);
 869	} while (!time_after(jiffies, start_time + HZ));
 870
 871	if (val & (E2P_CMD_EPC_TIMEOUT_ | E2P_CMD_EPC_BUSY_)) {
 872		netdev_warn(dev->net, "EEPROM read operation timeout");
 873		return -EIO;
 874	}
 875
 876	return 0;
 877}
 878
 879static int lan78xx_eeprom_confirm_not_busy(struct lan78xx_net *dev)
 880{
 881	unsigned long start_time = jiffies;
 882	u32 val;
 883	int ret;
 884
 885	do {
 886		ret = lan78xx_read_reg(dev, E2P_CMD, &val);
 887		if (unlikely(ret < 0))
 888			return -EIO;
 889
 890		if (!(val & E2P_CMD_EPC_BUSY_))
 891			return 0;
 892
 893		usleep_range(40, 100);
 894	} while (!time_after(jiffies, start_time + HZ));
 895
 896	netdev_warn(dev->net, "EEPROM is busy");
 897	return -EIO;
 898}
 899
 900static int lan78xx_read_raw_eeprom(struct lan78xx_net *dev, u32 offset,
 901				   u32 length, u8 *data)
 902{
 903	u32 val;
 904	u32 saved;
 905	int i, ret;
 906	int retval;
 907
 908	/* depends on chip, some EEPROM pins are muxed with LED function.
 909	 * disable & restore LED function to access EEPROM.
 910	 */
 911	ret = lan78xx_read_reg(dev, HW_CFG, &val);
 912	saved = val;
 913	if (dev->chipid == ID_REV_CHIP_ID_7800_) {
 914		val &= ~(HW_CFG_LED1_EN_ | HW_CFG_LED0_EN_);
 915		ret = lan78xx_write_reg(dev, HW_CFG, val);
 916	}
 917
 918	retval = lan78xx_eeprom_confirm_not_busy(dev);
 919	if (retval)
 920		return retval;
 921
 922	for (i = 0; i < length; i++) {
 923		val = E2P_CMD_EPC_BUSY_ | E2P_CMD_EPC_CMD_READ_;
 924		val |= (offset & E2P_CMD_EPC_ADDR_MASK_);
 925		ret = lan78xx_write_reg(dev, E2P_CMD, val);
 926		if (unlikely(ret < 0)) {
 927			retval = -EIO;
 928			goto exit;
 929		}
 930
 931		retval = lan78xx_wait_eeprom(dev);
 932		if (retval < 0)
 933			goto exit;
 934
 935		ret = lan78xx_read_reg(dev, E2P_DATA, &val);
 936		if (unlikely(ret < 0)) {
 937			retval = -EIO;
 938			goto exit;
 939		}
 940
 941		data[i] = val & 0xFF;
 942		offset++;
 943	}
 944
 945	retval = 0;
 946exit:
 947	if (dev->chipid == ID_REV_CHIP_ID_7800_)
 948		ret = lan78xx_write_reg(dev, HW_CFG, saved);
 949
 950	return retval;
 951}
 952
 953static int lan78xx_read_eeprom(struct lan78xx_net *dev, u32 offset,
 954			       u32 length, u8 *data)
 955{
 956	u8 sig;
 957	int ret;
 958
 959	ret = lan78xx_read_raw_eeprom(dev, 0, 1, &sig);
 960	if ((ret == 0) && (sig == EEPROM_INDICATOR))
 961		ret = lan78xx_read_raw_eeprom(dev, offset, length, data);
 962	else
 963		ret = -EINVAL;
 964
 965	return ret;
 966}
 967
 968static int lan78xx_write_raw_eeprom(struct lan78xx_net *dev, u32 offset,
 969				    u32 length, u8 *data)
 970{
 971	u32 val;
 972	u32 saved;
 973	int i, ret;
 974	int retval;
 975
 976	/* depends on chip, some EEPROM pins are muxed with LED function.
 977	 * disable & restore LED function to access EEPROM.
 978	 */
 979	ret = lan78xx_read_reg(dev, HW_CFG, &val);
 980	saved = val;
 981	if (dev->chipid == ID_REV_CHIP_ID_7800_) {
 982		val &= ~(HW_CFG_LED1_EN_ | HW_CFG_LED0_EN_);
 983		ret = lan78xx_write_reg(dev, HW_CFG, val);
 984	}
 985
 986	retval = lan78xx_eeprom_confirm_not_busy(dev);
 987	if (retval)
 988		goto exit;
 989
 990	/* Issue write/erase enable command */
 991	val = E2P_CMD_EPC_BUSY_ | E2P_CMD_EPC_CMD_EWEN_;
 992	ret = lan78xx_write_reg(dev, E2P_CMD, val);
 993	if (unlikely(ret < 0)) {
 994		retval = -EIO;
 995		goto exit;
 996	}
 997
 998	retval = lan78xx_wait_eeprom(dev);
 999	if (retval < 0)
1000		goto exit;
1001
1002	for (i = 0; i < length; i++) {
1003		/* Fill data register */
1004		val = data[i];
1005		ret = lan78xx_write_reg(dev, E2P_DATA, val);
1006		if (ret < 0) {
1007			retval = -EIO;
1008			goto exit;
1009		}
1010
1011		/* Send "write" command */
1012		val = E2P_CMD_EPC_BUSY_ | E2P_CMD_EPC_CMD_WRITE_;
1013		val |= (offset & E2P_CMD_EPC_ADDR_MASK_);
1014		ret = lan78xx_write_reg(dev, E2P_CMD, val);
1015		if (ret < 0) {
1016			retval = -EIO;
1017			goto exit;
1018		}
1019
1020		retval = lan78xx_wait_eeprom(dev);
1021		if (retval < 0)
1022			goto exit;
1023
1024		offset++;
1025	}
1026
1027	retval = 0;
1028exit:
1029	if (dev->chipid == ID_REV_CHIP_ID_7800_)
1030		ret = lan78xx_write_reg(dev, HW_CFG, saved);
1031
1032	return retval;
1033}
1034
1035static int lan78xx_read_raw_otp(struct lan78xx_net *dev, u32 offset,
1036				u32 length, u8 *data)
1037{
1038	int i;
1039	u32 buf;
1040	unsigned long timeout;
1041
1042	lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
1043
1044	if (buf & OTP_PWR_DN_PWRDN_N_) {
1045		/* clear it and wait to be cleared */
1046		lan78xx_write_reg(dev, OTP_PWR_DN, 0);
1047
1048		timeout = jiffies + HZ;
1049		do {
1050			usleep_range(1, 10);
1051			lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
1052			if (time_after(jiffies, timeout)) {
1053				netdev_warn(dev->net,
1054					    "timeout on OTP_PWR_DN");
1055				return -EIO;
1056			}
1057		} while (buf & OTP_PWR_DN_PWRDN_N_);
1058	}
1059
1060	for (i = 0; i < length; i++) {
1061		lan78xx_write_reg(dev, OTP_ADDR1,
1062				  ((offset + i) >> 8) & OTP_ADDR1_15_11);
1063		lan78xx_write_reg(dev, OTP_ADDR2,
1064				  ((offset + i) & OTP_ADDR2_10_3));
1065
1066		lan78xx_write_reg(dev, OTP_FUNC_CMD, OTP_FUNC_CMD_READ_);
1067		lan78xx_write_reg(dev, OTP_CMD_GO, OTP_CMD_GO_GO_);
1068
1069		timeout = jiffies + HZ;
1070		do {
1071			udelay(1);
1072			lan78xx_read_reg(dev, OTP_STATUS, &buf);
1073			if (time_after(jiffies, timeout)) {
1074				netdev_warn(dev->net,
1075					    "timeout on OTP_STATUS");
1076				return -EIO;
1077			}
1078		} while (buf & OTP_STATUS_BUSY_);
1079
1080		lan78xx_read_reg(dev, OTP_RD_DATA, &buf);
1081
1082		data[i] = (u8)(buf & 0xFF);
1083	}
1084
1085	return 0;
1086}
1087
1088static int lan78xx_write_raw_otp(struct lan78xx_net *dev, u32 offset,
1089				 u32 length, u8 *data)
1090{
1091	int i;
1092	u32 buf;
1093	unsigned long timeout;
1094
1095	lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
1096
1097	if (buf & OTP_PWR_DN_PWRDN_N_) {
1098		/* clear it and wait to be cleared */
1099		lan78xx_write_reg(dev, OTP_PWR_DN, 0);
1100
1101		timeout = jiffies + HZ;
1102		do {
1103			udelay(1);
1104			lan78xx_read_reg(dev, OTP_PWR_DN, &buf);
1105			if (time_after(jiffies, timeout)) {
1106				netdev_warn(dev->net,
1107					    "timeout on OTP_PWR_DN completion");
1108				return -EIO;
1109			}
1110		} while (buf & OTP_PWR_DN_PWRDN_N_);
1111	}
1112
1113	/* set to BYTE program mode */
1114	lan78xx_write_reg(dev, OTP_PRGM_MODE, OTP_PRGM_MODE_BYTE_);
1115
1116	for (i = 0; i < length; i++) {
1117		lan78xx_write_reg(dev, OTP_ADDR1,
1118				  ((offset + i) >> 8) & OTP_ADDR1_15_11);
1119		lan78xx_write_reg(dev, OTP_ADDR2,
1120				  ((offset + i) & OTP_ADDR2_10_3));
1121		lan78xx_write_reg(dev, OTP_PRGM_DATA, data[i]);
1122		lan78xx_write_reg(dev, OTP_TST_CMD, OTP_TST_CMD_PRGVRFY_);
1123		lan78xx_write_reg(dev, OTP_CMD_GO, OTP_CMD_GO_GO_);
1124
1125		timeout = jiffies + HZ;
1126		do {
1127			udelay(1);
1128			lan78xx_read_reg(dev, OTP_STATUS, &buf);
1129			if (time_after(jiffies, timeout)) {
1130				netdev_warn(dev->net,
1131					    "Timeout on OTP_STATUS completion");
1132				return -EIO;
1133			}
1134		} while (buf & OTP_STATUS_BUSY_);
1135	}
1136
1137	return 0;
1138}
1139
1140static int lan78xx_read_otp(struct lan78xx_net *dev, u32 offset,
1141			    u32 length, u8 *data)
1142{
1143	u8 sig;
1144	int ret;
1145
1146	ret = lan78xx_read_raw_otp(dev, 0, 1, &sig);
1147
1148	if (ret == 0) {
1149		if (sig == OTP_INDICATOR_2)
1150			offset += 0x100;
1151		else if (sig != OTP_INDICATOR_1)
1152			ret = -EINVAL;
1153		if (!ret)
1154			ret = lan78xx_read_raw_otp(dev, offset, length, data);
1155	}
1156
1157	return ret;
1158}
1159
1160static int lan78xx_dataport_wait_not_busy(struct lan78xx_net *dev)
1161{
1162	int i, ret;
1163
1164	for (i = 0; i < 100; i++) {
1165		u32 dp_sel;
1166
1167		ret = lan78xx_read_reg(dev, DP_SEL, &dp_sel);
1168		if (unlikely(ret < 0))
1169			return -EIO;
1170
1171		if (dp_sel & DP_SEL_DPRDY_)
1172			return 0;
1173
1174		usleep_range(40, 100);
1175	}
1176
1177	netdev_warn(dev->net, "%s timed out", __func__);
1178
1179	return -EIO;
1180}
1181
1182static int lan78xx_dataport_write(struct lan78xx_net *dev, u32 ram_select,
1183				  u32 addr, u32 length, u32 *buf)
1184{
1185	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1186	u32 dp_sel;
1187	int i, ret;
1188
1189	if (usb_autopm_get_interface(dev->intf) < 0)
1190		return 0;
1191
1192	mutex_lock(&pdata->dataport_mutex);
1193
1194	ret = lan78xx_dataport_wait_not_busy(dev);
1195	if (ret < 0)
1196		goto done;
1197
1198	ret = lan78xx_read_reg(dev, DP_SEL, &dp_sel);
1199
1200	dp_sel &= ~DP_SEL_RSEL_MASK_;
1201	dp_sel |= ram_select;
1202	ret = lan78xx_write_reg(dev, DP_SEL, dp_sel);
1203
1204	for (i = 0; i < length; i++) {
1205		ret = lan78xx_write_reg(dev, DP_ADDR, addr + i);
1206
1207		ret = lan78xx_write_reg(dev, DP_DATA, buf[i]);
1208
1209		ret = lan78xx_write_reg(dev, DP_CMD, DP_CMD_WRITE_);
1210
1211		ret = lan78xx_dataport_wait_not_busy(dev);
1212		if (ret < 0)
1213			goto done;
1214	}
1215
1216done:
1217	mutex_unlock(&pdata->dataport_mutex);
1218	usb_autopm_put_interface(dev->intf);
1219
1220	return ret;
1221}
1222
1223static void lan78xx_set_addr_filter(struct lan78xx_priv *pdata,
1224				    int index, u8 addr[ETH_ALEN])
1225{
1226	u32 temp;
1227
1228	if ((pdata) && (index > 0) && (index < NUM_OF_MAF)) {
1229		temp = addr[3];
1230		temp = addr[2] | (temp << 8);
1231		temp = addr[1] | (temp << 8);
1232		temp = addr[0] | (temp << 8);
1233		pdata->pfilter_table[index][1] = temp;
1234		temp = addr[5];
1235		temp = addr[4] | (temp << 8);
1236		temp |= MAF_HI_VALID_ | MAF_HI_TYPE_DST_;
1237		pdata->pfilter_table[index][0] = temp;
1238	}
1239}
1240
1241/* returns hash bit number for given MAC address */
1242static inline u32 lan78xx_hash(char addr[ETH_ALEN])
1243{
1244	return (ether_crc(ETH_ALEN, addr) >> 23) & 0x1ff;
1245}
1246
1247static void lan78xx_deferred_multicast_write(struct work_struct *param)
1248{
1249	struct lan78xx_priv *pdata =
1250			container_of(param, struct lan78xx_priv, set_multicast);
1251	struct lan78xx_net *dev = pdata->dev;
1252	int i;
1253
1254	netif_dbg(dev, drv, dev->net, "deferred multicast write 0x%08x\n",
1255		  pdata->rfe_ctl);
1256
1257	lan78xx_dataport_write(dev, DP_SEL_RSEL_VLAN_DA_, DP_SEL_VHF_VLAN_LEN,
1258			       DP_SEL_VHF_HASH_LEN, pdata->mchash_table);
1259
1260	for (i = 1; i < NUM_OF_MAF; i++) {
1261		lan78xx_write_reg(dev, MAF_HI(i), 0);
1262		lan78xx_write_reg(dev, MAF_LO(i),
1263				  pdata->pfilter_table[i][1]);
1264		lan78xx_write_reg(dev, MAF_HI(i),
1265				  pdata->pfilter_table[i][0]);
1266	}
1267
1268	lan78xx_write_reg(dev, RFE_CTL, pdata->rfe_ctl);
1269}
1270
1271static void lan78xx_set_multicast(struct net_device *netdev)
1272{
1273	struct lan78xx_net *dev = netdev_priv(netdev);
1274	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1275	unsigned long flags;
1276	int i;
1277
1278	spin_lock_irqsave(&pdata->rfe_ctl_lock, flags);
1279
1280	pdata->rfe_ctl &= ~(RFE_CTL_UCAST_EN_ | RFE_CTL_MCAST_EN_ |
1281			    RFE_CTL_DA_PERFECT_ | RFE_CTL_MCAST_HASH_);
1282
1283	for (i = 0; i < DP_SEL_VHF_HASH_LEN; i++)
1284		pdata->mchash_table[i] = 0;
1285
1286	/* pfilter_table[0] has own HW address */
1287	for (i = 1; i < NUM_OF_MAF; i++) {
1288		pdata->pfilter_table[i][0] = 0;
1289		pdata->pfilter_table[i][1] = 0;
1290	}
1291
1292	pdata->rfe_ctl |= RFE_CTL_BCAST_EN_;
1293
1294	if (dev->net->flags & IFF_PROMISC) {
1295		netif_dbg(dev, drv, dev->net, "promiscuous mode enabled");
1296		pdata->rfe_ctl |= RFE_CTL_MCAST_EN_ | RFE_CTL_UCAST_EN_;
1297	} else {
1298		if (dev->net->flags & IFF_ALLMULTI) {
1299			netif_dbg(dev, drv, dev->net,
1300				  "receive all multicast enabled");
1301			pdata->rfe_ctl |= RFE_CTL_MCAST_EN_;
1302		}
1303	}
1304
1305	if (netdev_mc_count(dev->net)) {
1306		struct netdev_hw_addr *ha;
1307		int i;
1308
1309		netif_dbg(dev, drv, dev->net, "receive multicast hash filter");
1310
1311		pdata->rfe_ctl |= RFE_CTL_DA_PERFECT_;
1312
1313		i = 1;
1314		netdev_for_each_mc_addr(ha, netdev) {
1315			/* set first 32 into Perfect Filter */
1316			if (i < 33) {
1317				lan78xx_set_addr_filter(pdata, i, ha->addr);
1318			} else {
1319				u32 bitnum = lan78xx_hash(ha->addr);
1320
1321				pdata->mchash_table[bitnum / 32] |=
1322							(1 << (bitnum % 32));
1323				pdata->rfe_ctl |= RFE_CTL_MCAST_HASH_;
1324			}
1325			i++;
1326		}
1327	}
1328
1329	spin_unlock_irqrestore(&pdata->rfe_ctl_lock, flags);
1330
1331	/* defer register writes to a sleepable context */
1332	schedule_work(&pdata->set_multicast);
1333}
1334
1335static int lan78xx_update_flowcontrol(struct lan78xx_net *dev, u8 duplex,
1336				      u16 lcladv, u16 rmtadv)
1337{
1338	u32 flow = 0, fct_flow = 0;
1339	u8 cap;
1340
1341	if (dev->fc_autoneg)
1342		cap = mii_resolve_flowctrl_fdx(lcladv, rmtadv);
1343	else
1344		cap = dev->fc_request_control;
1345
1346	if (cap & FLOW_CTRL_TX)
1347		flow |= (FLOW_CR_TX_FCEN_ | 0xFFFF);
1348
1349	if (cap & FLOW_CTRL_RX)
1350		flow |= FLOW_CR_RX_FCEN_;
1351
1352	if (dev->udev->speed == USB_SPEED_SUPER)
1353		fct_flow = FLOW_CTRL_THRESHOLD(FLOW_ON_SS, FLOW_OFF_SS);
1354	else if (dev->udev->speed == USB_SPEED_HIGH)
1355		fct_flow = FLOW_CTRL_THRESHOLD(FLOW_ON_HS, FLOW_OFF_HS);
1356
1357	netif_dbg(dev, link, dev->net, "rx pause %s, tx pause %s",
1358		  (cap & FLOW_CTRL_RX ? "enabled" : "disabled"),
1359		  (cap & FLOW_CTRL_TX ? "enabled" : "disabled"));
1360
1361	lan78xx_write_reg(dev, FCT_FLOW, fct_flow);
1362
1363	/* threshold value should be set before enabling flow */
1364	lan78xx_write_reg(dev, FLOW, flow);
1365
1366	return 0;
1367}
1368
1369static void lan78xx_rx_urb_submit_all(struct lan78xx_net *dev);
1370
1371static int lan78xx_mac_reset(struct lan78xx_net *dev)
1372{
1373	unsigned long start_time = jiffies;
1374	u32 val;
1375	int ret;
1376
1377	mutex_lock(&dev->phy_mutex);
1378
1379	/* Resetting the device while there is activity on the MDIO
1380	 * bus can result in the MAC interface locking up and not
1381	 * completing register access transactions.
1382	 */
1383	ret = lan78xx_phy_wait_not_busy(dev);
1384	if (ret < 0)
1385		goto done;
1386
1387	ret = lan78xx_read_reg(dev, MAC_CR, &val);
1388	if (ret < 0)
1389		goto done;
1390
1391	val |= MAC_CR_RST_;
1392	ret = lan78xx_write_reg(dev, MAC_CR, val);
1393	if (ret < 0)
1394		goto done;
1395
1396	/* Wait for the reset to complete before allowing any further
1397	 * MAC register accesses otherwise the MAC may lock up.
1398	 */
1399	do {
1400		ret = lan78xx_read_reg(dev, MAC_CR, &val);
1401		if (ret < 0)
1402			goto done;
1403
1404		if (!(val & MAC_CR_RST_)) {
1405			ret = 0;
1406			goto done;
1407		}
1408	} while (!time_after(jiffies, start_time + HZ));
1409
1410	ret = -ETIMEDOUT;
1411done:
1412	mutex_unlock(&dev->phy_mutex);
1413
1414	return ret;
1415}
1416
1417static int lan78xx_link_reset(struct lan78xx_net *dev)
1418{
1419	struct phy_device *phydev = dev->net->phydev;
1420	struct ethtool_link_ksettings ecmd;
1421	int ladv, radv, ret, link;
1422	u32 buf;
1423
1424	/* clear LAN78xx interrupt status */
1425	ret = lan78xx_write_reg(dev, INT_STS, INT_STS_PHY_INT_);
1426	if (unlikely(ret < 0))
1427		return ret;
1428
1429	mutex_lock(&phydev->lock);
1430	phy_read_status(phydev);
1431	link = phydev->link;
1432	mutex_unlock(&phydev->lock);
1433
1434	if (!link && dev->link_on) {
1435		dev->link_on = false;
1436
1437		/* reset MAC */
1438		ret = lan78xx_mac_reset(dev);
1439		if (ret < 0)
1440			return ret;
 
 
 
 
1441
1442		del_timer(&dev->stat_monitor);
1443	} else if (link && !dev->link_on) {
1444		dev->link_on = true;
1445
1446		phy_ethtool_ksettings_get(phydev, &ecmd);
1447
1448		if (dev->udev->speed == USB_SPEED_SUPER) {
1449			if (ecmd.base.speed == 1000) {
1450				/* disable U2 */
1451				ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
1452				if (ret < 0)
1453					return ret;
1454				buf &= ~USB_CFG1_DEV_U2_INIT_EN_;
1455				ret = lan78xx_write_reg(dev, USB_CFG1, buf);
1456				if (ret < 0)
1457					return ret;
1458				/* enable U1 */
1459				ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
1460				if (ret < 0)
1461					return ret;
1462				buf |= USB_CFG1_DEV_U1_INIT_EN_;
1463				ret = lan78xx_write_reg(dev, USB_CFG1, buf);
1464				if (ret < 0)
1465					return ret;
1466			} else {
1467				/* enable U1 & U2 */
1468				ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
1469				if (ret < 0)
1470					return ret;
1471				buf |= USB_CFG1_DEV_U2_INIT_EN_;
1472				buf |= USB_CFG1_DEV_U1_INIT_EN_;
1473				ret = lan78xx_write_reg(dev, USB_CFG1, buf);
1474				if (ret < 0)
1475					return ret;
1476			}
1477		}
1478
1479		ladv = phy_read(phydev, MII_ADVERTISE);
1480		if (ladv < 0)
1481			return ladv;
1482
1483		radv = phy_read(phydev, MII_LPA);
1484		if (radv < 0)
1485			return radv;
1486
1487		netif_dbg(dev, link, dev->net,
1488			  "speed: %u duplex: %d anadv: 0x%04x anlpa: 0x%04x",
1489			  ecmd.base.speed, ecmd.base.duplex, ladv, radv);
1490
1491		ret = lan78xx_update_flowcontrol(dev, ecmd.base.duplex, ladv,
1492						 radv);
1493		if (ret < 0)
1494			return ret;
1495
1496		if (!timer_pending(&dev->stat_monitor)) {
1497			dev->delta = 1;
1498			mod_timer(&dev->stat_monitor,
1499				  jiffies + STAT_UPDATE_TIMER);
1500		}
1501
1502		lan78xx_rx_urb_submit_all(dev);
1503
1504		napi_schedule(&dev->napi);
1505	}
1506
1507	return 0;
1508}
1509
1510/* some work can't be done in tasklets, so we use keventd
1511 *
1512 * NOTE:  annoying asymmetry:  if it's active, schedule_work() fails,
1513 * but tasklet_schedule() doesn't.	hope the failure is rare.
1514 */
1515static void lan78xx_defer_kevent(struct lan78xx_net *dev, int work)
1516{
1517	set_bit(work, &dev->flags);
1518	if (!schedule_delayed_work(&dev->wq, 0))
1519		netdev_err(dev->net, "kevent %d may have been dropped\n", work);
1520}
1521
1522static void lan78xx_status(struct lan78xx_net *dev, struct urb *urb)
1523{
1524	u32 intdata;
1525
1526	if (urb->actual_length != 4) {
1527		netdev_warn(dev->net,
1528			    "unexpected urb length %d", urb->actual_length);
1529		return;
1530	}
1531
1532	intdata = get_unaligned_le32(urb->transfer_buffer);
1533
1534	if (intdata & INT_ENP_PHY_INT) {
1535		netif_dbg(dev, link, dev->net, "PHY INTR: 0x%08x\n", intdata);
1536		lan78xx_defer_kevent(dev, EVENT_LINK_RESET);
1537
1538		if (dev->domain_data.phyirq > 0)
1539			generic_handle_irq_safe(dev->domain_data.phyirq);
1540	} else {
 
 
 
1541		netdev_warn(dev->net,
1542			    "unexpected interrupt: 0x%08x\n", intdata);
1543	}
1544}
1545
1546static int lan78xx_ethtool_get_eeprom_len(struct net_device *netdev)
1547{
1548	return MAX_EEPROM_SIZE;
1549}
1550
1551static int lan78xx_ethtool_get_eeprom(struct net_device *netdev,
1552				      struct ethtool_eeprom *ee, u8 *data)
1553{
1554	struct lan78xx_net *dev = netdev_priv(netdev);
1555	int ret;
1556
1557	ret = usb_autopm_get_interface(dev->intf);
1558	if (ret)
1559		return ret;
1560
1561	ee->magic = LAN78XX_EEPROM_MAGIC;
1562
1563	ret = lan78xx_read_raw_eeprom(dev, ee->offset, ee->len, data);
1564
1565	usb_autopm_put_interface(dev->intf);
1566
1567	return ret;
1568}
1569
1570static int lan78xx_ethtool_set_eeprom(struct net_device *netdev,
1571				      struct ethtool_eeprom *ee, u8 *data)
1572{
1573	struct lan78xx_net *dev = netdev_priv(netdev);
1574	int ret;
1575
1576	ret = usb_autopm_get_interface(dev->intf);
1577	if (ret)
1578		return ret;
1579
1580	/* Invalid EEPROM_INDICATOR at offset zero will result in a failure
1581	 * to load data from EEPROM
1582	 */
1583	if (ee->magic == LAN78XX_EEPROM_MAGIC)
1584		ret = lan78xx_write_raw_eeprom(dev, ee->offset, ee->len, data);
1585	else if ((ee->magic == LAN78XX_OTP_MAGIC) &&
1586		 (ee->offset == 0) &&
1587		 (ee->len == 512) &&
1588		 (data[0] == OTP_INDICATOR_1))
1589		ret = lan78xx_write_raw_otp(dev, ee->offset, ee->len, data);
1590
1591	usb_autopm_put_interface(dev->intf);
1592
1593	return ret;
1594}
1595
1596static void lan78xx_get_strings(struct net_device *netdev, u32 stringset,
1597				u8 *data)
1598{
1599	if (stringset == ETH_SS_STATS)
1600		memcpy(data, lan78xx_gstrings, sizeof(lan78xx_gstrings));
1601}
1602
1603static int lan78xx_get_sset_count(struct net_device *netdev, int sset)
1604{
1605	if (sset == ETH_SS_STATS)
1606		return ARRAY_SIZE(lan78xx_gstrings);
1607	else
1608		return -EOPNOTSUPP;
1609}
1610
1611static void lan78xx_get_stats(struct net_device *netdev,
1612			      struct ethtool_stats *stats, u64 *data)
1613{
1614	struct lan78xx_net *dev = netdev_priv(netdev);
1615
1616	lan78xx_update_stats(dev);
1617
1618	mutex_lock(&dev->stats.access_lock);
1619	memcpy(data, &dev->stats.curr_stat, sizeof(dev->stats.curr_stat));
1620	mutex_unlock(&dev->stats.access_lock);
1621}
1622
1623static void lan78xx_get_wol(struct net_device *netdev,
1624			    struct ethtool_wolinfo *wol)
1625{
1626	struct lan78xx_net *dev = netdev_priv(netdev);
1627	int ret;
1628	u32 buf;
1629	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1630
1631	if (usb_autopm_get_interface(dev->intf) < 0)
1632		return;
1633
1634	ret = lan78xx_read_reg(dev, USB_CFG0, &buf);
1635	if (unlikely(ret < 0)) {
1636		wol->supported = 0;
1637		wol->wolopts = 0;
1638	} else {
1639		if (buf & USB_CFG_RMT_WKP_) {
1640			wol->supported = WAKE_ALL;
1641			wol->wolopts = pdata->wol;
1642		} else {
1643			wol->supported = 0;
1644			wol->wolopts = 0;
1645		}
1646	}
1647
1648	usb_autopm_put_interface(dev->intf);
1649}
1650
1651static int lan78xx_set_wol(struct net_device *netdev,
1652			   struct ethtool_wolinfo *wol)
1653{
1654	struct lan78xx_net *dev = netdev_priv(netdev);
1655	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
1656	int ret;
1657
1658	ret = usb_autopm_get_interface(dev->intf);
1659	if (ret < 0)
1660		return ret;
1661
1662	if (wol->wolopts & ~WAKE_ALL)
1663		return -EINVAL;
1664
1665	pdata->wol = wol->wolopts;
1666
1667	device_set_wakeup_enable(&dev->udev->dev, (bool)wol->wolopts);
1668
1669	phy_ethtool_set_wol(netdev->phydev, wol);
1670
1671	usb_autopm_put_interface(dev->intf);
1672
1673	return ret;
1674}
1675
1676static int lan78xx_get_eee(struct net_device *net, struct ethtool_eee *edata)
1677{
1678	struct lan78xx_net *dev = netdev_priv(net);
1679	struct phy_device *phydev = net->phydev;
1680	int ret;
1681	u32 buf;
1682
1683	ret = usb_autopm_get_interface(dev->intf);
1684	if (ret < 0)
1685		return ret;
1686
1687	ret = phy_ethtool_get_eee(phydev, edata);
1688	if (ret < 0)
1689		goto exit;
1690
1691	ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1692	if (buf & MAC_CR_EEE_EN_) {
1693		edata->eee_enabled = true;
1694		edata->eee_active = !!(edata->advertised &
1695				       edata->lp_advertised);
1696		edata->tx_lpi_enabled = true;
1697		/* EEE_TX_LPI_REQ_DLY & tx_lpi_timer are same uSec unit */
1698		ret = lan78xx_read_reg(dev, EEE_TX_LPI_REQ_DLY, &buf);
1699		edata->tx_lpi_timer = buf;
1700	} else {
1701		edata->eee_enabled = false;
1702		edata->eee_active = false;
1703		edata->tx_lpi_enabled = false;
1704		edata->tx_lpi_timer = 0;
1705	}
1706
1707	ret = 0;
1708exit:
1709	usb_autopm_put_interface(dev->intf);
1710
1711	return ret;
1712}
1713
1714static int lan78xx_set_eee(struct net_device *net, struct ethtool_eee *edata)
1715{
1716	struct lan78xx_net *dev = netdev_priv(net);
1717	int ret;
1718	u32 buf;
1719
1720	ret = usb_autopm_get_interface(dev->intf);
1721	if (ret < 0)
1722		return ret;
1723
1724	if (edata->eee_enabled) {
1725		ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1726		buf |= MAC_CR_EEE_EN_;
1727		ret = lan78xx_write_reg(dev, MAC_CR, buf);
1728
1729		phy_ethtool_set_eee(net->phydev, edata);
1730
1731		buf = (u32)edata->tx_lpi_timer;
1732		ret = lan78xx_write_reg(dev, EEE_TX_LPI_REQ_DLY, buf);
1733	} else {
1734		ret = lan78xx_read_reg(dev, MAC_CR, &buf);
1735		buf &= ~MAC_CR_EEE_EN_;
1736		ret = lan78xx_write_reg(dev, MAC_CR, buf);
1737	}
1738
1739	usb_autopm_put_interface(dev->intf);
1740
1741	return 0;
1742}
1743
1744static u32 lan78xx_get_link(struct net_device *net)
1745{
1746	u32 link;
1747
1748	mutex_lock(&net->phydev->lock);
1749	phy_read_status(net->phydev);
1750	link = net->phydev->link;
1751	mutex_unlock(&net->phydev->lock);
1752
1753	return link;
1754}
1755
1756static void lan78xx_get_drvinfo(struct net_device *net,
1757				struct ethtool_drvinfo *info)
1758{
1759	struct lan78xx_net *dev = netdev_priv(net);
1760
1761	strncpy(info->driver, DRIVER_NAME, sizeof(info->driver));
1762	usb_make_path(dev->udev, info->bus_info, sizeof(info->bus_info));
1763}
1764
1765static u32 lan78xx_get_msglevel(struct net_device *net)
1766{
1767	struct lan78xx_net *dev = netdev_priv(net);
1768
1769	return dev->msg_enable;
1770}
1771
1772static void lan78xx_set_msglevel(struct net_device *net, u32 level)
1773{
1774	struct lan78xx_net *dev = netdev_priv(net);
1775
1776	dev->msg_enable = level;
1777}
1778
1779static int lan78xx_get_link_ksettings(struct net_device *net,
1780				      struct ethtool_link_ksettings *cmd)
1781{
1782	struct lan78xx_net *dev = netdev_priv(net);
1783	struct phy_device *phydev = net->phydev;
1784	int ret;
1785
1786	ret = usb_autopm_get_interface(dev->intf);
1787	if (ret < 0)
1788		return ret;
1789
1790	phy_ethtool_ksettings_get(phydev, cmd);
1791
1792	usb_autopm_put_interface(dev->intf);
1793
1794	return ret;
1795}
1796
1797static int lan78xx_set_link_ksettings(struct net_device *net,
1798				      const struct ethtool_link_ksettings *cmd)
1799{
1800	struct lan78xx_net *dev = netdev_priv(net);
1801	struct phy_device *phydev = net->phydev;
1802	int ret = 0;
1803	int temp;
1804
1805	ret = usb_autopm_get_interface(dev->intf);
1806	if (ret < 0)
1807		return ret;
1808
1809	/* change speed & duplex */
1810	ret = phy_ethtool_ksettings_set(phydev, cmd);
1811
1812	if (!cmd->base.autoneg) {
1813		/* force link down */
1814		temp = phy_read(phydev, MII_BMCR);
1815		phy_write(phydev, MII_BMCR, temp | BMCR_LOOPBACK);
1816		mdelay(1);
1817		phy_write(phydev, MII_BMCR, temp);
1818	}
1819
1820	usb_autopm_put_interface(dev->intf);
1821
1822	return ret;
1823}
1824
1825static void lan78xx_get_pause(struct net_device *net,
1826			      struct ethtool_pauseparam *pause)
1827{
1828	struct lan78xx_net *dev = netdev_priv(net);
1829	struct phy_device *phydev = net->phydev;
1830	struct ethtool_link_ksettings ecmd;
1831
1832	phy_ethtool_ksettings_get(phydev, &ecmd);
1833
1834	pause->autoneg = dev->fc_autoneg;
1835
1836	if (dev->fc_request_control & FLOW_CTRL_TX)
1837		pause->tx_pause = 1;
1838
1839	if (dev->fc_request_control & FLOW_CTRL_RX)
1840		pause->rx_pause = 1;
1841}
1842
1843static int lan78xx_set_pause(struct net_device *net,
1844			     struct ethtool_pauseparam *pause)
1845{
1846	struct lan78xx_net *dev = netdev_priv(net);
1847	struct phy_device *phydev = net->phydev;
1848	struct ethtool_link_ksettings ecmd;
1849	int ret;
1850
1851	phy_ethtool_ksettings_get(phydev, &ecmd);
1852
1853	if (pause->autoneg && !ecmd.base.autoneg) {
1854		ret = -EINVAL;
1855		goto exit;
1856	}
1857
1858	dev->fc_request_control = 0;
1859	if (pause->rx_pause)
1860		dev->fc_request_control |= FLOW_CTRL_RX;
1861
1862	if (pause->tx_pause)
1863		dev->fc_request_control |= FLOW_CTRL_TX;
1864
1865	if (ecmd.base.autoneg) {
1866		__ETHTOOL_DECLARE_LINK_MODE_MASK(fc) = { 0, };
1867		u32 mii_adv;
1868
1869		linkmode_clear_bit(ETHTOOL_LINK_MODE_Pause_BIT,
1870				   ecmd.link_modes.advertising);
1871		linkmode_clear_bit(ETHTOOL_LINK_MODE_Asym_Pause_BIT,
1872				   ecmd.link_modes.advertising);
1873		mii_adv = (u32)mii_advertise_flowctrl(dev->fc_request_control);
1874		mii_adv_to_linkmode_adv_t(fc, mii_adv);
1875		linkmode_or(ecmd.link_modes.advertising, fc,
1876			    ecmd.link_modes.advertising);
1877
1878		phy_ethtool_ksettings_set(phydev, &ecmd);
1879	}
1880
1881	dev->fc_autoneg = pause->autoneg;
1882
1883	ret = 0;
1884exit:
1885	return ret;
1886}
1887
1888static int lan78xx_get_regs_len(struct net_device *netdev)
1889{
1890	if (!netdev->phydev)
1891		return (sizeof(lan78xx_regs));
1892	else
1893		return (sizeof(lan78xx_regs) + PHY_REG_SIZE);
1894}
1895
1896static void
1897lan78xx_get_regs(struct net_device *netdev, struct ethtool_regs *regs,
1898		 void *buf)
1899{
1900	u32 *data = buf;
1901	int i, j;
1902	struct lan78xx_net *dev = netdev_priv(netdev);
1903
1904	/* Read Device/MAC registers */
1905	for (i = 0; i < ARRAY_SIZE(lan78xx_regs); i++)
1906		lan78xx_read_reg(dev, lan78xx_regs[i], &data[i]);
1907
1908	if (!netdev->phydev)
1909		return;
1910
1911	/* Read PHY registers */
1912	for (j = 0; j < 32; i++, j++)
1913		data[i] = phy_read(netdev->phydev, j);
1914}
1915
1916static const struct ethtool_ops lan78xx_ethtool_ops = {
1917	.get_link	= lan78xx_get_link,
1918	.nway_reset	= phy_ethtool_nway_reset,
1919	.get_drvinfo	= lan78xx_get_drvinfo,
1920	.get_msglevel	= lan78xx_get_msglevel,
1921	.set_msglevel	= lan78xx_set_msglevel,
1922	.get_eeprom_len = lan78xx_ethtool_get_eeprom_len,
1923	.get_eeprom	= lan78xx_ethtool_get_eeprom,
1924	.set_eeprom	= lan78xx_ethtool_set_eeprom,
1925	.get_ethtool_stats = lan78xx_get_stats,
1926	.get_sset_count = lan78xx_get_sset_count,
1927	.get_strings	= lan78xx_get_strings,
1928	.get_wol	= lan78xx_get_wol,
1929	.set_wol	= lan78xx_set_wol,
1930	.get_ts_info	= ethtool_op_get_ts_info,
1931	.get_eee	= lan78xx_get_eee,
1932	.set_eee	= lan78xx_set_eee,
1933	.get_pauseparam	= lan78xx_get_pause,
1934	.set_pauseparam	= lan78xx_set_pause,
1935	.get_link_ksettings = lan78xx_get_link_ksettings,
1936	.set_link_ksettings = lan78xx_set_link_ksettings,
1937	.get_regs_len	= lan78xx_get_regs_len,
1938	.get_regs	= lan78xx_get_regs,
1939};
1940
1941static void lan78xx_init_mac_address(struct lan78xx_net *dev)
1942{
1943	u32 addr_lo, addr_hi;
1944	u8 addr[6];
1945
1946	lan78xx_read_reg(dev, RX_ADDRL, &addr_lo);
1947	lan78xx_read_reg(dev, RX_ADDRH, &addr_hi);
1948
1949	addr[0] = addr_lo & 0xFF;
1950	addr[1] = (addr_lo >> 8) & 0xFF;
1951	addr[2] = (addr_lo >> 16) & 0xFF;
1952	addr[3] = (addr_lo >> 24) & 0xFF;
1953	addr[4] = addr_hi & 0xFF;
1954	addr[5] = (addr_hi >> 8) & 0xFF;
1955
1956	if (!is_valid_ether_addr(addr)) {
1957		if (!eth_platform_get_mac_address(&dev->udev->dev, addr)) {
1958			/* valid address present in Device Tree */
1959			netif_dbg(dev, ifup, dev->net,
1960				  "MAC address read from Device Tree");
1961		} else if (((lan78xx_read_eeprom(dev, EEPROM_MAC_OFFSET,
1962						 ETH_ALEN, addr) == 0) ||
1963			    (lan78xx_read_otp(dev, EEPROM_MAC_OFFSET,
1964					      ETH_ALEN, addr) == 0)) &&
1965			   is_valid_ether_addr(addr)) {
1966			/* eeprom values are valid so use them */
1967			netif_dbg(dev, ifup, dev->net,
1968				  "MAC address read from EEPROM");
1969		} else {
1970			/* generate random MAC */
1971			eth_random_addr(addr);
1972			netif_dbg(dev, ifup, dev->net,
1973				  "MAC address set to random addr");
1974		}
1975
1976		addr_lo = addr[0] | (addr[1] << 8) |
1977			  (addr[2] << 16) | (addr[3] << 24);
1978		addr_hi = addr[4] | (addr[5] << 8);
1979
1980		lan78xx_write_reg(dev, RX_ADDRL, addr_lo);
1981		lan78xx_write_reg(dev, RX_ADDRH, addr_hi);
1982	}
1983
1984	lan78xx_write_reg(dev, MAF_LO(0), addr_lo);
1985	lan78xx_write_reg(dev, MAF_HI(0), addr_hi | MAF_HI_VALID_);
1986
1987	eth_hw_addr_set(dev->net, addr);
1988}
1989
1990/* MDIO read and write wrappers for phylib */
1991static int lan78xx_mdiobus_read(struct mii_bus *bus, int phy_id, int idx)
1992{
1993	struct lan78xx_net *dev = bus->priv;
1994	u32 val, addr;
1995	int ret;
1996
1997	ret = usb_autopm_get_interface(dev->intf);
1998	if (ret < 0)
1999		return ret;
2000
2001	mutex_lock(&dev->phy_mutex);
2002
2003	/* confirm MII not busy */
2004	ret = lan78xx_phy_wait_not_busy(dev);
2005	if (ret < 0)
2006		goto done;
2007
2008	/* set the address, index & direction (read from PHY) */
2009	addr = mii_access(phy_id, idx, MII_READ);
2010	ret = lan78xx_write_reg(dev, MII_ACC, addr);
2011
2012	ret = lan78xx_phy_wait_not_busy(dev);
2013	if (ret < 0)
2014		goto done;
2015
2016	ret = lan78xx_read_reg(dev, MII_DATA, &val);
2017
2018	ret = (int)(val & 0xFFFF);
2019
2020done:
2021	mutex_unlock(&dev->phy_mutex);
2022	usb_autopm_put_interface(dev->intf);
2023
2024	return ret;
2025}
2026
2027static int lan78xx_mdiobus_write(struct mii_bus *bus, int phy_id, int idx,
2028				 u16 regval)
2029{
2030	struct lan78xx_net *dev = bus->priv;
2031	u32 val, addr;
2032	int ret;
2033
2034	ret = usb_autopm_get_interface(dev->intf);
2035	if (ret < 0)
2036		return ret;
2037
2038	mutex_lock(&dev->phy_mutex);
2039
2040	/* confirm MII not busy */
2041	ret = lan78xx_phy_wait_not_busy(dev);
2042	if (ret < 0)
2043		goto done;
2044
2045	val = (u32)regval;
2046	ret = lan78xx_write_reg(dev, MII_DATA, val);
2047
2048	/* set the address, index & direction (write to PHY) */
2049	addr = mii_access(phy_id, idx, MII_WRITE);
2050	ret = lan78xx_write_reg(dev, MII_ACC, addr);
2051
2052	ret = lan78xx_phy_wait_not_busy(dev);
2053	if (ret < 0)
2054		goto done;
2055
2056done:
2057	mutex_unlock(&dev->phy_mutex);
2058	usb_autopm_put_interface(dev->intf);
2059	return 0;
2060}
2061
2062static int lan78xx_mdio_init(struct lan78xx_net *dev)
2063{
2064	struct device_node *node;
2065	int ret;
2066
2067	dev->mdiobus = mdiobus_alloc();
2068	if (!dev->mdiobus) {
2069		netdev_err(dev->net, "can't allocate MDIO bus\n");
2070		return -ENOMEM;
2071	}
2072
2073	dev->mdiobus->priv = (void *)dev;
2074	dev->mdiobus->read = lan78xx_mdiobus_read;
2075	dev->mdiobus->write = lan78xx_mdiobus_write;
2076	dev->mdiobus->name = "lan78xx-mdiobus";
2077	dev->mdiobus->parent = &dev->udev->dev;
2078
2079	snprintf(dev->mdiobus->id, MII_BUS_ID_SIZE, "usb-%03d:%03d",
2080		 dev->udev->bus->busnum, dev->udev->devnum);
2081
2082	switch (dev->chipid) {
2083	case ID_REV_CHIP_ID_7800_:
2084	case ID_REV_CHIP_ID_7850_:
2085		/* set to internal PHY id */
2086		dev->mdiobus->phy_mask = ~(1 << 1);
2087		break;
2088	case ID_REV_CHIP_ID_7801_:
2089		/* scan thru PHYAD[2..0] */
2090		dev->mdiobus->phy_mask = ~(0xFF);
2091		break;
2092	}
2093
2094	node = of_get_child_by_name(dev->udev->dev.of_node, "mdio");
2095	ret = of_mdiobus_register(dev->mdiobus, node);
2096	of_node_put(node);
2097	if (ret) {
2098		netdev_err(dev->net, "can't register MDIO bus\n");
2099		goto exit1;
2100	}
2101
2102	netdev_dbg(dev->net, "registered mdiobus bus %s\n", dev->mdiobus->id);
2103	return 0;
2104exit1:
2105	mdiobus_free(dev->mdiobus);
2106	return ret;
2107}
2108
2109static void lan78xx_remove_mdio(struct lan78xx_net *dev)
2110{
2111	mdiobus_unregister(dev->mdiobus);
2112	mdiobus_free(dev->mdiobus);
2113}
2114
2115static void lan78xx_link_status_change(struct net_device *net)
2116{
2117	struct phy_device *phydev = net->phydev;
2118	int temp;
2119
2120	/* At forced 100 F/H mode, chip may fail to set mode correctly
2121	 * when cable is switched between long(~50+m) and short one.
2122	 * As workaround, set to 10 before setting to 100
2123	 * at forced 100 F/H mode.
2124	 */
2125	if (!phydev->autoneg && (phydev->speed == 100)) {
2126		/* disable phy interrupt */
2127		temp = phy_read(phydev, LAN88XX_INT_MASK);
2128		temp &= ~LAN88XX_INT_MASK_MDINTPIN_EN_;
2129		phy_write(phydev, LAN88XX_INT_MASK, temp);
2130
2131		temp = phy_read(phydev, MII_BMCR);
2132		temp &= ~(BMCR_SPEED100 | BMCR_SPEED1000);
2133		phy_write(phydev, MII_BMCR, temp); /* set to 10 first */
2134		temp |= BMCR_SPEED100;
2135		phy_write(phydev, MII_BMCR, temp); /* set to 100 later */
2136
2137		/* clear pending interrupt generated while workaround */
2138		temp = phy_read(phydev, LAN88XX_INT_STS);
2139
2140		/* enable phy interrupt back */
2141		temp = phy_read(phydev, LAN88XX_INT_MASK);
2142		temp |= LAN88XX_INT_MASK_MDINTPIN_EN_;
2143		phy_write(phydev, LAN88XX_INT_MASK, temp);
2144	}
2145}
2146
2147static int irq_map(struct irq_domain *d, unsigned int irq,
2148		   irq_hw_number_t hwirq)
2149{
2150	struct irq_domain_data *data = d->host_data;
2151
2152	irq_set_chip_data(irq, data);
2153	irq_set_chip_and_handler(irq, data->irqchip, data->irq_handler);
2154	irq_set_noprobe(irq);
2155
2156	return 0;
2157}
2158
2159static void irq_unmap(struct irq_domain *d, unsigned int irq)
2160{
2161	irq_set_chip_and_handler(irq, NULL, NULL);
2162	irq_set_chip_data(irq, NULL);
2163}
2164
2165static const struct irq_domain_ops chip_domain_ops = {
2166	.map	= irq_map,
2167	.unmap	= irq_unmap,
2168};
2169
2170static void lan78xx_irq_mask(struct irq_data *irqd)
2171{
2172	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
2173
2174	data->irqenable &= ~BIT(irqd_to_hwirq(irqd));
2175}
2176
2177static void lan78xx_irq_unmask(struct irq_data *irqd)
2178{
2179	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
2180
2181	data->irqenable |= BIT(irqd_to_hwirq(irqd));
2182}
2183
2184static void lan78xx_irq_bus_lock(struct irq_data *irqd)
2185{
2186	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
2187
2188	mutex_lock(&data->irq_lock);
2189}
2190
2191static void lan78xx_irq_bus_sync_unlock(struct irq_data *irqd)
2192{
2193	struct irq_domain_data *data = irq_data_get_irq_chip_data(irqd);
2194	struct lan78xx_net *dev =
2195			container_of(data, struct lan78xx_net, domain_data);
2196	u32 buf;
2197
2198	/* call register access here because irq_bus_lock & irq_bus_sync_unlock
2199	 * are only two callbacks executed in non-atomic contex.
2200	 */
2201	lan78xx_read_reg(dev, INT_EP_CTL, &buf);
2202	if (buf != data->irqenable)
2203		lan78xx_write_reg(dev, INT_EP_CTL, data->irqenable);
2204
2205	mutex_unlock(&data->irq_lock);
2206}
2207
2208static struct irq_chip lan78xx_irqchip = {
2209	.name			= "lan78xx-irqs",
2210	.irq_mask		= lan78xx_irq_mask,
2211	.irq_unmask		= lan78xx_irq_unmask,
2212	.irq_bus_lock		= lan78xx_irq_bus_lock,
2213	.irq_bus_sync_unlock	= lan78xx_irq_bus_sync_unlock,
2214};
2215
2216static int lan78xx_setup_irq_domain(struct lan78xx_net *dev)
2217{
2218	struct device_node *of_node;
2219	struct irq_domain *irqdomain;
2220	unsigned int irqmap = 0;
2221	u32 buf;
2222	int ret = 0;
2223
2224	of_node = dev->udev->dev.parent->of_node;
2225
2226	mutex_init(&dev->domain_data.irq_lock);
2227
2228	lan78xx_read_reg(dev, INT_EP_CTL, &buf);
2229	dev->domain_data.irqenable = buf;
2230
2231	dev->domain_data.irqchip = &lan78xx_irqchip;
2232	dev->domain_data.irq_handler = handle_simple_irq;
2233
2234	irqdomain = irq_domain_add_simple(of_node, MAX_INT_EP, 0,
2235					  &chip_domain_ops, &dev->domain_data);
2236	if (irqdomain) {
2237		/* create mapping for PHY interrupt */
2238		irqmap = irq_create_mapping(irqdomain, INT_EP_PHY);
2239		if (!irqmap) {
2240			irq_domain_remove(irqdomain);
2241
2242			irqdomain = NULL;
2243			ret = -EINVAL;
2244		}
2245	} else {
2246		ret = -EINVAL;
2247	}
2248
2249	dev->domain_data.irqdomain = irqdomain;
2250	dev->domain_data.phyirq = irqmap;
2251
2252	return ret;
2253}
2254
2255static void lan78xx_remove_irq_domain(struct lan78xx_net *dev)
2256{
2257	if (dev->domain_data.phyirq > 0) {
2258		irq_dispose_mapping(dev->domain_data.phyirq);
2259
2260		if (dev->domain_data.irqdomain)
2261			irq_domain_remove(dev->domain_data.irqdomain);
2262	}
2263	dev->domain_data.phyirq = 0;
2264	dev->domain_data.irqdomain = NULL;
2265}
2266
2267static int lan8835_fixup(struct phy_device *phydev)
2268{
2269	int buf;
2270	struct lan78xx_net *dev = netdev_priv(phydev->attached_dev);
2271
2272	/* LED2/PME_N/IRQ_N/RGMII_ID pin to IRQ_N mode */
2273	buf = phy_read_mmd(phydev, MDIO_MMD_PCS, 0x8010);
2274	buf &= ~0x1800;
2275	buf |= 0x0800;
2276	phy_write_mmd(phydev, MDIO_MMD_PCS, 0x8010, buf);
2277
2278	/* RGMII MAC TXC Delay Enable */
2279	lan78xx_write_reg(dev, MAC_RGMII_ID,
2280			  MAC_RGMII_ID_TXC_DELAY_EN_);
2281
2282	/* RGMII TX DLL Tune Adjust */
2283	lan78xx_write_reg(dev, RGMII_TX_BYP_DLL, 0x3D00);
2284
2285	dev->interface = PHY_INTERFACE_MODE_RGMII_TXID;
2286
2287	return 1;
2288}
2289
2290static int ksz9031rnx_fixup(struct phy_device *phydev)
2291{
2292	struct lan78xx_net *dev = netdev_priv(phydev->attached_dev);
2293
2294	/* Micrel9301RNX PHY configuration */
2295	/* RGMII Control Signal Pad Skew */
2296	phy_write_mmd(phydev, MDIO_MMD_WIS, 4, 0x0077);
2297	/* RGMII RX Data Pad Skew */
2298	phy_write_mmd(phydev, MDIO_MMD_WIS, 5, 0x7777);
2299	/* RGMII RX Clock Pad Skew */
2300	phy_write_mmd(phydev, MDIO_MMD_WIS, 8, 0x1FF);
2301
2302	dev->interface = PHY_INTERFACE_MODE_RGMII_RXID;
2303
2304	return 1;
2305}
2306
2307static struct phy_device *lan7801_phy_init(struct lan78xx_net *dev)
2308{
2309	u32 buf;
2310	int ret;
2311	struct fixed_phy_status fphy_status = {
2312		.link = 1,
2313		.speed = SPEED_1000,
2314		.duplex = DUPLEX_FULL,
2315	};
2316	struct phy_device *phydev;
2317
2318	phydev = phy_find_first(dev->mdiobus);
2319	if (!phydev) {
2320		netdev_dbg(dev->net, "PHY Not Found!! Registering Fixed PHY\n");
2321		phydev = fixed_phy_register(PHY_POLL, &fphy_status, NULL);
2322		if (IS_ERR(phydev)) {
2323			netdev_err(dev->net, "No PHY/fixed_PHY found\n");
2324			return NULL;
2325		}
2326		netdev_dbg(dev->net, "Registered FIXED PHY\n");
2327		dev->interface = PHY_INTERFACE_MODE_RGMII;
2328		ret = lan78xx_write_reg(dev, MAC_RGMII_ID,
2329					MAC_RGMII_ID_TXC_DELAY_EN_);
2330		ret = lan78xx_write_reg(dev, RGMII_TX_BYP_DLL, 0x3D00);
2331		ret = lan78xx_read_reg(dev, HW_CFG, &buf);
2332		buf |= HW_CFG_CLK125_EN_;
2333		buf |= HW_CFG_REFCLK25_EN_;
2334		ret = lan78xx_write_reg(dev, HW_CFG, buf);
2335	} else {
2336		if (!phydev->drv) {
2337			netdev_err(dev->net, "no PHY driver found\n");
2338			return NULL;
2339		}
2340		dev->interface = PHY_INTERFACE_MODE_RGMII;
2341		/* external PHY fixup for KSZ9031RNX */
2342		ret = phy_register_fixup_for_uid(PHY_KSZ9031RNX, 0xfffffff0,
2343						 ksz9031rnx_fixup);
2344		if (ret < 0) {
2345			netdev_err(dev->net, "Failed to register fixup for PHY_KSZ9031RNX\n");
2346			return NULL;
2347		}
2348		/* external PHY fixup for LAN8835 */
2349		ret = phy_register_fixup_for_uid(PHY_LAN8835, 0xfffffff0,
2350						 lan8835_fixup);
2351		if (ret < 0) {
2352			netdev_err(dev->net, "Failed to register fixup for PHY_LAN8835\n");
2353			return NULL;
2354		}
2355		/* add more external PHY fixup here if needed */
2356
2357		phydev->is_internal = false;
2358	}
2359	return phydev;
2360}
2361
2362static int lan78xx_phy_init(struct lan78xx_net *dev)
2363{
2364	__ETHTOOL_DECLARE_LINK_MODE_MASK(fc) = { 0, };
2365	int ret;
2366	u32 mii_adv;
2367	struct phy_device *phydev;
2368
2369	switch (dev->chipid) {
2370	case ID_REV_CHIP_ID_7801_:
2371		phydev = lan7801_phy_init(dev);
2372		if (!phydev) {
2373			netdev_err(dev->net, "lan7801: PHY Init Failed");
2374			return -EIO;
2375		}
2376		break;
2377
2378	case ID_REV_CHIP_ID_7800_:
2379	case ID_REV_CHIP_ID_7850_:
2380		phydev = phy_find_first(dev->mdiobus);
2381		if (!phydev) {
2382			netdev_err(dev->net, "no PHY found\n");
2383			return -EIO;
2384		}
2385		phydev->is_internal = true;
2386		dev->interface = PHY_INTERFACE_MODE_GMII;
2387		break;
2388
2389	default:
2390		netdev_err(dev->net, "Unknown CHIP ID found\n");
2391		return -EIO;
2392	}
2393
2394	/* if phyirq is not set, use polling mode in phylib */
2395	if (dev->domain_data.phyirq > 0)
2396		phydev->irq = dev->domain_data.phyirq;
2397	else
2398		phydev->irq = PHY_POLL;
2399	netdev_dbg(dev->net, "phydev->irq = %d\n", phydev->irq);
2400
2401	/* set to AUTOMDIX */
2402	phydev->mdix = ETH_TP_MDI_AUTO;
2403
2404	ret = phy_connect_direct(dev->net, phydev,
2405				 lan78xx_link_status_change,
2406				 dev->interface);
2407	if (ret) {
2408		netdev_err(dev->net, "can't attach PHY to %s\n",
2409			   dev->mdiobus->id);
2410		if (dev->chipid == ID_REV_CHIP_ID_7801_) {
2411			if (phy_is_pseudo_fixed_link(phydev)) {
2412				fixed_phy_unregister(phydev);
2413			} else {
2414				phy_unregister_fixup_for_uid(PHY_KSZ9031RNX,
2415							     0xfffffff0);
2416				phy_unregister_fixup_for_uid(PHY_LAN8835,
2417							     0xfffffff0);
2418			}
2419		}
2420		return -EIO;
2421	}
2422
2423	/* MAC doesn't support 1000T Half */
2424	phy_remove_link_mode(phydev, ETHTOOL_LINK_MODE_1000baseT_Half_BIT);
2425
2426	/* support both flow controls */
2427	dev->fc_request_control = (FLOW_CTRL_RX | FLOW_CTRL_TX);
2428	linkmode_clear_bit(ETHTOOL_LINK_MODE_Pause_BIT,
2429			   phydev->advertising);
2430	linkmode_clear_bit(ETHTOOL_LINK_MODE_Asym_Pause_BIT,
2431			   phydev->advertising);
2432	mii_adv = (u32)mii_advertise_flowctrl(dev->fc_request_control);
2433	mii_adv_to_linkmode_adv_t(fc, mii_adv);
2434	linkmode_or(phydev->advertising, fc, phydev->advertising);
2435
2436	if (phydev->mdio.dev.of_node) {
2437		u32 reg;
2438		int len;
2439
2440		len = of_property_count_elems_of_size(phydev->mdio.dev.of_node,
2441						      "microchip,led-modes",
2442						      sizeof(u32));
2443		if (len >= 0) {
2444			/* Ensure the appropriate LEDs are enabled */
2445			lan78xx_read_reg(dev, HW_CFG, &reg);
2446			reg &= ~(HW_CFG_LED0_EN_ |
2447				 HW_CFG_LED1_EN_ |
2448				 HW_CFG_LED2_EN_ |
2449				 HW_CFG_LED3_EN_);
2450			reg |= (len > 0) * HW_CFG_LED0_EN_ |
2451				(len > 1) * HW_CFG_LED1_EN_ |
2452				(len > 2) * HW_CFG_LED2_EN_ |
2453				(len > 3) * HW_CFG_LED3_EN_;
2454			lan78xx_write_reg(dev, HW_CFG, reg);
2455		}
2456	}
2457
2458	genphy_config_aneg(phydev);
2459
2460	dev->fc_autoneg = phydev->autoneg;
2461
2462	return 0;
2463}
2464
2465static int lan78xx_set_rx_max_frame_length(struct lan78xx_net *dev, int size)
2466{
2467	u32 buf;
2468	bool rxenabled;
2469
2470	lan78xx_read_reg(dev, MAC_RX, &buf);
2471
2472	rxenabled = ((buf & MAC_RX_RXEN_) != 0);
2473
2474	if (rxenabled) {
2475		buf &= ~MAC_RX_RXEN_;
2476		lan78xx_write_reg(dev, MAC_RX, buf);
2477	}
2478
2479	/* add 4 to size for FCS */
2480	buf &= ~MAC_RX_MAX_SIZE_MASK_;
2481	buf |= (((size + 4) << MAC_RX_MAX_SIZE_SHIFT_) & MAC_RX_MAX_SIZE_MASK_);
2482
2483	lan78xx_write_reg(dev, MAC_RX, buf);
2484
2485	if (rxenabled) {
2486		buf |= MAC_RX_RXEN_;
2487		lan78xx_write_reg(dev, MAC_RX, buf);
2488	}
2489
2490	return 0;
2491}
2492
2493static int unlink_urbs(struct lan78xx_net *dev, struct sk_buff_head *q)
2494{
2495	struct sk_buff *skb;
2496	unsigned long flags;
2497	int count = 0;
2498
2499	spin_lock_irqsave(&q->lock, flags);
2500	while (!skb_queue_empty(q)) {
2501		struct skb_data	*entry;
2502		struct urb *urb;
2503		int ret;
2504
2505		skb_queue_walk(q, skb) {
2506			entry = (struct skb_data *)skb->cb;
2507			if (entry->state != unlink_start)
2508				goto found;
2509		}
2510		break;
2511found:
2512		entry->state = unlink_start;
2513		urb = entry->urb;
2514
2515		/* Get reference count of the URB to avoid it to be
2516		 * freed during usb_unlink_urb, which may trigger
2517		 * use-after-free problem inside usb_unlink_urb since
2518		 * usb_unlink_urb is always racing with .complete
2519		 * handler(include defer_bh).
2520		 */
2521		usb_get_urb(urb);
2522		spin_unlock_irqrestore(&q->lock, flags);
2523		/* during some PM-driven resume scenarios,
2524		 * these (async) unlinks complete immediately
2525		 */
2526		ret = usb_unlink_urb(urb);
2527		if (ret != -EINPROGRESS && ret != 0)
2528			netdev_dbg(dev->net, "unlink urb err, %d\n", ret);
2529		else
2530			count++;
2531		usb_put_urb(urb);
2532		spin_lock_irqsave(&q->lock, flags);
2533	}
2534	spin_unlock_irqrestore(&q->lock, flags);
2535	return count;
2536}
2537
2538static int lan78xx_change_mtu(struct net_device *netdev, int new_mtu)
2539{
2540	struct lan78xx_net *dev = netdev_priv(netdev);
2541	int max_frame_len = RX_MAX_FRAME_LEN(new_mtu);
2542	int ret;
 
2543
2544	/* no second zero-length packet read wanted after mtu-sized packets */
2545	if ((max_frame_len % dev->maxpacket) == 0)
2546		return -EDOM;
2547
2548	ret = usb_autopm_get_interface(dev->intf);
2549	if (ret < 0)
2550		return ret;
2551
2552	ret = lan78xx_set_rx_max_frame_length(dev, max_frame_len);
2553	if (!ret)
2554		netdev->mtu = new_mtu;
2555
2556	usb_autopm_put_interface(dev->intf);
 
 
 
 
 
 
 
 
 
2557
2558	return ret;
2559}
2560
2561static int lan78xx_set_mac_addr(struct net_device *netdev, void *p)
2562{
2563	struct lan78xx_net *dev = netdev_priv(netdev);
2564	struct sockaddr *addr = p;
2565	u32 addr_lo, addr_hi;
2566
2567	if (netif_running(netdev))
2568		return -EBUSY;
2569
2570	if (!is_valid_ether_addr(addr->sa_data))
2571		return -EADDRNOTAVAIL;
2572
2573	eth_hw_addr_set(netdev, addr->sa_data);
2574
2575	addr_lo = netdev->dev_addr[0] |
2576		  netdev->dev_addr[1] << 8 |
2577		  netdev->dev_addr[2] << 16 |
2578		  netdev->dev_addr[3] << 24;
2579	addr_hi = netdev->dev_addr[4] |
2580		  netdev->dev_addr[5] << 8;
2581
2582	lan78xx_write_reg(dev, RX_ADDRL, addr_lo);
2583	lan78xx_write_reg(dev, RX_ADDRH, addr_hi);
2584
2585	/* Added to support MAC address changes */
2586	lan78xx_write_reg(dev, MAF_LO(0), addr_lo);
2587	lan78xx_write_reg(dev, MAF_HI(0), addr_hi | MAF_HI_VALID_);
2588
2589	return 0;
2590}
2591
2592/* Enable or disable Rx checksum offload engine */
2593static int lan78xx_set_features(struct net_device *netdev,
2594				netdev_features_t features)
2595{
2596	struct lan78xx_net *dev = netdev_priv(netdev);
2597	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2598	unsigned long flags;
2599
2600	spin_lock_irqsave(&pdata->rfe_ctl_lock, flags);
2601
2602	if (features & NETIF_F_RXCSUM) {
2603		pdata->rfe_ctl |= RFE_CTL_TCPUDP_COE_ | RFE_CTL_IP_COE_;
2604		pdata->rfe_ctl |= RFE_CTL_ICMP_COE_ | RFE_CTL_IGMP_COE_;
2605	} else {
2606		pdata->rfe_ctl &= ~(RFE_CTL_TCPUDP_COE_ | RFE_CTL_IP_COE_);
2607		pdata->rfe_ctl &= ~(RFE_CTL_ICMP_COE_ | RFE_CTL_IGMP_COE_);
2608	}
2609
2610	if (features & NETIF_F_HW_VLAN_CTAG_RX)
2611		pdata->rfe_ctl |= RFE_CTL_VLAN_STRIP_;
2612	else
2613		pdata->rfe_ctl &= ~RFE_CTL_VLAN_STRIP_;
2614
2615	if (features & NETIF_F_HW_VLAN_CTAG_FILTER)
2616		pdata->rfe_ctl |= RFE_CTL_VLAN_FILTER_;
2617	else
2618		pdata->rfe_ctl &= ~RFE_CTL_VLAN_FILTER_;
2619
2620	spin_unlock_irqrestore(&pdata->rfe_ctl_lock, flags);
2621
2622	lan78xx_write_reg(dev, RFE_CTL, pdata->rfe_ctl);
2623
2624	return 0;
2625}
2626
2627static void lan78xx_deferred_vlan_write(struct work_struct *param)
2628{
2629	struct lan78xx_priv *pdata =
2630			container_of(param, struct lan78xx_priv, set_vlan);
2631	struct lan78xx_net *dev = pdata->dev;
2632
2633	lan78xx_dataport_write(dev, DP_SEL_RSEL_VLAN_DA_, 0,
2634			       DP_SEL_VHF_VLAN_LEN, pdata->vlan_table);
2635}
2636
2637static int lan78xx_vlan_rx_add_vid(struct net_device *netdev,
2638				   __be16 proto, u16 vid)
2639{
2640	struct lan78xx_net *dev = netdev_priv(netdev);
2641	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2642	u16 vid_bit_index;
2643	u16 vid_dword_index;
2644
2645	vid_dword_index = (vid >> 5) & 0x7F;
2646	vid_bit_index = vid & 0x1F;
2647
2648	pdata->vlan_table[vid_dword_index] |= (1 << vid_bit_index);
2649
2650	/* defer register writes to a sleepable context */
2651	schedule_work(&pdata->set_vlan);
2652
2653	return 0;
2654}
2655
2656static int lan78xx_vlan_rx_kill_vid(struct net_device *netdev,
2657				    __be16 proto, u16 vid)
2658{
2659	struct lan78xx_net *dev = netdev_priv(netdev);
2660	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
2661	u16 vid_bit_index;
2662	u16 vid_dword_index;
2663
2664	vid_dword_index = (vid >> 5) & 0x7F;
2665	vid_bit_index = vid & 0x1F;
2666
2667	pdata->vlan_table[vid_dword_index] &= ~(1 << vid_bit_index);
2668
2669	/* defer register writes to a sleepable context */
2670	schedule_work(&pdata->set_vlan);
2671
2672	return 0;
2673}
2674
2675static void lan78xx_init_ltm(struct lan78xx_net *dev)
2676{
2677	int ret;
2678	u32 buf;
2679	u32 regs[6] = { 0 };
2680
2681	ret = lan78xx_read_reg(dev, USB_CFG1, &buf);
2682	if (buf & USB_CFG1_LTM_ENABLE_) {
2683		u8 temp[2];
2684		/* Get values from EEPROM first */
2685		if (lan78xx_read_eeprom(dev, 0x3F, 2, temp) == 0) {
2686			if (temp[0] == 24) {
2687				ret = lan78xx_read_raw_eeprom(dev,
2688							      temp[1] * 2,
2689							      24,
2690							      (u8 *)regs);
2691				if (ret < 0)
2692					return;
2693			}
2694		} else if (lan78xx_read_otp(dev, 0x3F, 2, temp) == 0) {
2695			if (temp[0] == 24) {
2696				ret = lan78xx_read_raw_otp(dev,
2697							   temp[1] * 2,
2698							   24,
2699							   (u8 *)regs);
2700				if (ret < 0)
2701					return;
2702			}
2703		}
2704	}
2705
2706	lan78xx_write_reg(dev, LTM_BELT_IDLE0, regs[0]);
2707	lan78xx_write_reg(dev, LTM_BELT_IDLE1, regs[1]);
2708	lan78xx_write_reg(dev, LTM_BELT_ACT0, regs[2]);
2709	lan78xx_write_reg(dev, LTM_BELT_ACT1, regs[3]);
2710	lan78xx_write_reg(dev, LTM_INACTIVE0, regs[4]);
2711	lan78xx_write_reg(dev, LTM_INACTIVE1, regs[5]);
2712}
2713
2714static int lan78xx_urb_config_init(struct lan78xx_net *dev)
2715{
2716	int result = 0;
2717
2718	switch (dev->udev->speed) {
2719	case USB_SPEED_SUPER:
2720		dev->rx_urb_size = RX_SS_URB_SIZE;
2721		dev->tx_urb_size = TX_SS_URB_SIZE;
2722		dev->n_rx_urbs = RX_SS_URB_NUM;
2723		dev->n_tx_urbs = TX_SS_URB_NUM;
2724		dev->bulk_in_delay = SS_BULK_IN_DELAY;
2725		dev->burst_cap = SS_BURST_CAP_SIZE / SS_USB_PKT_SIZE;
2726		break;
2727	case USB_SPEED_HIGH:
2728		dev->rx_urb_size = RX_HS_URB_SIZE;
2729		dev->tx_urb_size = TX_HS_URB_SIZE;
2730		dev->n_rx_urbs = RX_HS_URB_NUM;
2731		dev->n_tx_urbs = TX_HS_URB_NUM;
2732		dev->bulk_in_delay = HS_BULK_IN_DELAY;
2733		dev->burst_cap = HS_BURST_CAP_SIZE / HS_USB_PKT_SIZE;
2734		break;
2735	case USB_SPEED_FULL:
2736		dev->rx_urb_size = RX_FS_URB_SIZE;
2737		dev->tx_urb_size = TX_FS_URB_SIZE;
2738		dev->n_rx_urbs = RX_FS_URB_NUM;
2739		dev->n_tx_urbs = TX_FS_URB_NUM;
2740		dev->bulk_in_delay = FS_BULK_IN_DELAY;
2741		dev->burst_cap = FS_BURST_CAP_SIZE / FS_USB_PKT_SIZE;
2742		break;
2743	default:
2744		netdev_warn(dev->net, "USB bus speed not supported\n");
2745		result = -EIO;
2746		break;
2747	}
2748
2749	return result;
2750}
2751
2752static int lan78xx_start_hw(struct lan78xx_net *dev, u32 reg, u32 hw_enable)
2753{
2754	return lan78xx_update_reg(dev, reg, hw_enable, hw_enable);
2755}
2756
2757static int lan78xx_stop_hw(struct lan78xx_net *dev, u32 reg, u32 hw_enabled,
2758			   u32 hw_disabled)
2759{
2760	unsigned long timeout;
2761	bool stopped = true;
2762	int ret;
2763	u32 buf;
2764
2765	/* Stop the h/w block (if not already stopped) */
2766
2767	ret = lan78xx_read_reg(dev, reg, &buf);
2768	if (ret < 0)
2769		return ret;
2770
2771	if (buf & hw_enabled) {
2772		buf &= ~hw_enabled;
2773
2774		ret = lan78xx_write_reg(dev, reg, buf);
2775		if (ret < 0)
2776			return ret;
2777
2778		stopped = false;
2779		timeout = jiffies + HW_DISABLE_TIMEOUT;
2780		do  {
2781			ret = lan78xx_read_reg(dev, reg, &buf);
2782			if (ret < 0)
2783				return ret;
2784
2785			if (buf & hw_disabled)
2786				stopped = true;
2787			else
2788				msleep(HW_DISABLE_DELAY_MS);
2789		} while (!stopped && !time_after(jiffies, timeout));
2790	}
2791
2792	ret = stopped ? 0 : -ETIME;
2793
2794	return ret;
2795}
2796
2797static int lan78xx_flush_fifo(struct lan78xx_net *dev, u32 reg, u32 fifo_flush)
2798{
2799	return lan78xx_update_reg(dev, reg, fifo_flush, fifo_flush);
2800}
2801
2802static int lan78xx_start_tx_path(struct lan78xx_net *dev)
2803{
2804	int ret;
2805
2806	netif_dbg(dev, drv, dev->net, "start tx path");
2807
2808	/* Start the MAC transmitter */
2809
2810	ret = lan78xx_start_hw(dev, MAC_TX, MAC_TX_TXEN_);
2811	if (ret < 0)
2812		return ret;
2813
2814	/* Start the Tx FIFO */
2815
2816	ret = lan78xx_start_hw(dev, FCT_TX_CTL, FCT_TX_CTL_EN_);
2817	if (ret < 0)
2818		return ret;
2819
2820	return 0;
2821}
2822
2823static int lan78xx_stop_tx_path(struct lan78xx_net *dev)
2824{
2825	int ret;
2826
2827	netif_dbg(dev, drv, dev->net, "stop tx path");
2828
2829	/* Stop the Tx FIFO */
2830
2831	ret = lan78xx_stop_hw(dev, FCT_TX_CTL, FCT_TX_CTL_EN_, FCT_TX_CTL_DIS_);
2832	if (ret < 0)
2833		return ret;
2834
2835	/* Stop the MAC transmitter */
2836
2837	ret = lan78xx_stop_hw(dev, MAC_TX, MAC_TX_TXEN_, MAC_TX_TXD_);
2838	if (ret < 0)
2839		return ret;
2840
2841	return 0;
2842}
2843
2844/* The caller must ensure the Tx path is stopped before calling
2845 * lan78xx_flush_tx_fifo().
2846 */
2847static int lan78xx_flush_tx_fifo(struct lan78xx_net *dev)
2848{
2849	return lan78xx_flush_fifo(dev, FCT_TX_CTL, FCT_TX_CTL_RST_);
2850}
2851
2852static int lan78xx_start_rx_path(struct lan78xx_net *dev)
2853{
2854	int ret;
2855
2856	netif_dbg(dev, drv, dev->net, "start rx path");
2857
2858	/* Start the Rx FIFO */
2859
2860	ret = lan78xx_start_hw(dev, FCT_RX_CTL, FCT_RX_CTL_EN_);
2861	if (ret < 0)
2862		return ret;
2863
2864	/* Start the MAC receiver*/
2865
2866	ret = lan78xx_start_hw(dev, MAC_RX, MAC_RX_RXEN_);
2867	if (ret < 0)
2868		return ret;
2869
2870	return 0;
2871}
2872
2873static int lan78xx_stop_rx_path(struct lan78xx_net *dev)
2874{
2875	int ret;
2876
2877	netif_dbg(dev, drv, dev->net, "stop rx path");
2878
2879	/* Stop the MAC receiver */
2880
2881	ret = lan78xx_stop_hw(dev, MAC_RX, MAC_RX_RXEN_, MAC_RX_RXD_);
2882	if (ret < 0)
2883		return ret;
2884
2885	/* Stop the Rx FIFO */
2886
2887	ret = lan78xx_stop_hw(dev, FCT_RX_CTL, FCT_RX_CTL_EN_, FCT_RX_CTL_DIS_);
2888	if (ret < 0)
2889		return ret;
2890
2891	return 0;
2892}
2893
2894/* The caller must ensure the Rx path is stopped before calling
2895 * lan78xx_flush_rx_fifo().
2896 */
2897static int lan78xx_flush_rx_fifo(struct lan78xx_net *dev)
2898{
2899	return lan78xx_flush_fifo(dev, FCT_RX_CTL, FCT_RX_CTL_RST_);
2900}
2901
2902static int lan78xx_reset(struct lan78xx_net *dev)
2903{
2904	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
 
 
2905	unsigned long timeout;
2906	int ret;
2907	u32 buf;
2908	u8 sig;
2909
2910	ret = lan78xx_read_reg(dev, HW_CFG, &buf);
2911	if (ret < 0)
2912		return ret;
2913
2914	buf |= HW_CFG_LRST_;
2915
2916	ret = lan78xx_write_reg(dev, HW_CFG, buf);
2917	if (ret < 0)
2918		return ret;
2919
2920	timeout = jiffies + HZ;
2921	do {
2922		mdelay(1);
2923		ret = lan78xx_read_reg(dev, HW_CFG, &buf);
2924		if (ret < 0)
2925			return ret;
2926
2927		if (time_after(jiffies, timeout)) {
2928			netdev_warn(dev->net,
2929				    "timeout on completion of LiteReset");
2930			ret = -ETIMEDOUT;
2931			return ret;
2932		}
2933	} while (buf & HW_CFG_LRST_);
2934
2935	lan78xx_init_mac_address(dev);
2936
2937	/* save DEVID for later usage */
2938	ret = lan78xx_read_reg(dev, ID_REV, &buf);
2939	if (ret < 0)
2940		return ret;
2941
2942	dev->chipid = (buf & ID_REV_CHIP_ID_MASK_) >> 16;
2943	dev->chiprev = buf & ID_REV_CHIP_REV_MASK_;
2944
2945	/* Respond to the IN token with a NAK */
2946	ret = lan78xx_read_reg(dev, USB_CFG0, &buf);
2947	if (ret < 0)
2948		return ret;
2949
2950	buf |= USB_CFG_BIR_;
2951
2952	ret = lan78xx_write_reg(dev, USB_CFG0, buf);
2953	if (ret < 0)
2954		return ret;
2955
2956	/* Init LTM */
2957	lan78xx_init_ltm(dev);
2958
2959	ret = lan78xx_write_reg(dev, BURST_CAP, dev->burst_cap);
2960	if (ret < 0)
2961		return ret;
 
 
 
 
 
 
 
 
 
 
 
 
 
2962
2963	ret = lan78xx_write_reg(dev, BULK_IN_DLY, dev->bulk_in_delay);
2964	if (ret < 0)
2965		return ret;
2966
2967	ret = lan78xx_read_reg(dev, HW_CFG, &buf);
2968	if (ret < 0)
2969		return ret;
2970
2971	buf |= HW_CFG_MEF_;
2972
2973	ret = lan78xx_write_reg(dev, HW_CFG, buf);
2974	if (ret < 0)
2975		return ret;
2976
2977	ret = lan78xx_read_reg(dev, USB_CFG0, &buf);
2978	if (ret < 0)
2979		return ret;
2980
2981	buf |= USB_CFG_BCE_;
2982
2983	ret = lan78xx_write_reg(dev, USB_CFG0, buf);
2984	if (ret < 0)
2985		return ret;
2986
2987	/* set FIFO sizes */
2988	buf = (MAX_RX_FIFO_SIZE - 512) / 512;
2989
2990	ret = lan78xx_write_reg(dev, FCT_RX_FIFO_END, buf);
2991	if (ret < 0)
2992		return ret;
2993
2994	buf = (MAX_TX_FIFO_SIZE - 512) / 512;
2995
2996	ret = lan78xx_write_reg(dev, FCT_TX_FIFO_END, buf);
2997	if (ret < 0)
2998		return ret;
2999
3000	ret = lan78xx_write_reg(dev, INT_STS, INT_STS_CLEAR_ALL_);
3001	if (ret < 0)
3002		return ret;
3003
3004	ret = lan78xx_write_reg(dev, FLOW, 0);
3005	if (ret < 0)
3006		return ret;
3007
3008	ret = lan78xx_write_reg(dev, FCT_FLOW, 0);
3009	if (ret < 0)
3010		return ret;
3011
3012	/* Don't need rfe_ctl_lock during initialisation */
3013	ret = lan78xx_read_reg(dev, RFE_CTL, &pdata->rfe_ctl);
3014	if (ret < 0)
3015		return ret;
3016
3017	pdata->rfe_ctl |= RFE_CTL_BCAST_EN_ | RFE_CTL_DA_PERFECT_;
3018
3019	ret = lan78xx_write_reg(dev, RFE_CTL, pdata->rfe_ctl);
3020	if (ret < 0)
3021		return ret;
3022
3023	/* Enable or disable checksum offload engines */
3024	ret = lan78xx_set_features(dev->net, dev->net->features);
3025	if (ret < 0)
3026		return ret;
3027
3028	lan78xx_set_multicast(dev->net);
3029
3030	/* reset PHY */
3031	ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
3032	if (ret < 0)
3033		return ret;
3034
3035	buf |= PMT_CTL_PHY_RST_;
3036
3037	ret = lan78xx_write_reg(dev, PMT_CTL, buf);
3038	if (ret < 0)
3039		return ret;
3040
3041	timeout = jiffies + HZ;
3042	do {
3043		mdelay(1);
3044		ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
3045		if (ret < 0)
3046			return ret;
3047
3048		if (time_after(jiffies, timeout)) {
3049			netdev_warn(dev->net, "timeout waiting for PHY Reset");
3050			ret = -ETIMEDOUT;
3051			return ret;
3052		}
3053	} while ((buf & PMT_CTL_PHY_RST_) || !(buf & PMT_CTL_READY_));
3054
3055	ret = lan78xx_read_reg(dev, MAC_CR, &buf);
3056	if (ret < 0)
3057		return ret;
3058
3059	/* LAN7801 only has RGMII mode */
3060	if (dev->chipid == ID_REV_CHIP_ID_7801_)
3061		buf &= ~MAC_CR_GMII_EN_;
3062
3063	if (dev->chipid == ID_REV_CHIP_ID_7800_) {
3064		ret = lan78xx_read_raw_eeprom(dev, 0, 1, &sig);
3065		if (!ret && sig != EEPROM_INDICATOR) {
3066			/* Implies there is no external eeprom. Set mac speed */
3067			netdev_info(dev->net, "No External EEPROM. Setting MAC Speed\n");
3068			buf |= MAC_CR_AUTO_DUPLEX_ | MAC_CR_AUTO_SPEED_;
3069		}
3070	}
3071	ret = lan78xx_write_reg(dev, MAC_CR, buf);
3072	if (ret < 0)
3073		return ret;
 
 
 
 
 
 
3074
3075	ret = lan78xx_set_rx_max_frame_length(dev,
3076					      RX_MAX_FRAME_LEN(dev->net->mtu));
3077
3078	return ret;
 
 
 
 
 
 
 
 
3079}
3080
3081static void lan78xx_init_stats(struct lan78xx_net *dev)
3082{
3083	u32 *p;
3084	int i;
3085
3086	/* initialize for stats update
3087	 * some counters are 20bits and some are 32bits
3088	 */
3089	p = (u32 *)&dev->stats.rollover_max;
3090	for (i = 0; i < (sizeof(dev->stats.rollover_max) / (sizeof(u32))); i++)
3091		p[i] = 0xFFFFF;
3092
3093	dev->stats.rollover_max.rx_unicast_byte_count = 0xFFFFFFFF;
3094	dev->stats.rollover_max.rx_broadcast_byte_count = 0xFFFFFFFF;
3095	dev->stats.rollover_max.rx_multicast_byte_count = 0xFFFFFFFF;
3096	dev->stats.rollover_max.eee_rx_lpi_transitions = 0xFFFFFFFF;
3097	dev->stats.rollover_max.eee_rx_lpi_time = 0xFFFFFFFF;
3098	dev->stats.rollover_max.tx_unicast_byte_count = 0xFFFFFFFF;
3099	dev->stats.rollover_max.tx_broadcast_byte_count = 0xFFFFFFFF;
3100	dev->stats.rollover_max.tx_multicast_byte_count = 0xFFFFFFFF;
3101	dev->stats.rollover_max.eee_tx_lpi_transitions = 0xFFFFFFFF;
3102	dev->stats.rollover_max.eee_tx_lpi_time = 0xFFFFFFFF;
3103
3104	set_bit(EVENT_STAT_UPDATE, &dev->flags);
3105}
3106
3107static int lan78xx_open(struct net_device *net)
3108{
3109	struct lan78xx_net *dev = netdev_priv(net);
3110	int ret;
3111
3112	netif_dbg(dev, ifup, dev->net, "open device");
3113
3114	ret = usb_autopm_get_interface(dev->intf);
3115	if (ret < 0)
3116		return ret;
3117
3118	mutex_lock(&dev->dev_mutex);
3119
3120	phy_start(net->phydev);
3121
3122	netif_dbg(dev, ifup, dev->net, "phy initialised successfully");
3123
3124	/* for Link Check */
3125	if (dev->urb_intr) {
3126		ret = usb_submit_urb(dev->urb_intr, GFP_KERNEL);
3127		if (ret < 0) {
3128			netif_err(dev, ifup, dev->net,
3129				  "intr submit %d\n", ret);
3130			goto done;
3131		}
3132	}
3133
3134	ret = lan78xx_flush_rx_fifo(dev);
3135	if (ret < 0)
3136		goto done;
3137	ret = lan78xx_flush_tx_fifo(dev);
3138	if (ret < 0)
3139		goto done;
3140
3141	ret = lan78xx_start_tx_path(dev);
3142	if (ret < 0)
3143		goto done;
3144	ret = lan78xx_start_rx_path(dev);
3145	if (ret < 0)
3146		goto done;
3147
3148	lan78xx_init_stats(dev);
3149
3150	set_bit(EVENT_DEV_OPEN, &dev->flags);
3151
3152	netif_start_queue(net);
3153
3154	dev->link_on = false;
3155
3156	napi_enable(&dev->napi);
3157
3158	lan78xx_defer_kevent(dev, EVENT_LINK_RESET);
3159done:
3160	mutex_unlock(&dev->dev_mutex);
3161
3162	usb_autopm_put_interface(dev->intf);
3163
 
3164	return ret;
3165}
3166
3167static void lan78xx_terminate_urbs(struct lan78xx_net *dev)
3168{
3169	DECLARE_WAIT_QUEUE_HEAD_ONSTACK(unlink_wakeup);
3170	DECLARE_WAITQUEUE(wait, current);
3171	int temp;
3172
3173	/* ensure there are no more active urbs */
3174	add_wait_queue(&unlink_wakeup, &wait);
3175	set_current_state(TASK_UNINTERRUPTIBLE);
3176	dev->wait = &unlink_wakeup;
3177	temp = unlink_urbs(dev, &dev->txq) + unlink_urbs(dev, &dev->rxq);
3178
3179	/* maybe wait for deletions to finish. */
3180	while (!skb_queue_empty(&dev->rxq) ||
3181	       !skb_queue_empty(&dev->txq)) {
 
3182		schedule_timeout(msecs_to_jiffies(UNLINK_TIMEOUT_MS));
3183		set_current_state(TASK_UNINTERRUPTIBLE);
3184		netif_dbg(dev, ifdown, dev->net,
3185			  "waited for %d urb completions", temp);
3186	}
3187	set_current_state(TASK_RUNNING);
3188	dev->wait = NULL;
3189	remove_wait_queue(&unlink_wakeup, &wait);
3190
3191	/* empty Rx done, Rx overflow and Tx pend queues
3192	 */
3193	while (!skb_queue_empty(&dev->rxq_done)) {
3194		struct sk_buff *skb = skb_dequeue(&dev->rxq_done);
3195
3196		lan78xx_release_rx_buf(dev, skb);
3197	}
3198
3199	skb_queue_purge(&dev->rxq_overflow);
3200	skb_queue_purge(&dev->txq_pend);
3201}
3202
3203static int lan78xx_stop(struct net_device *net)
3204{
3205	struct lan78xx_net *dev = netdev_priv(net);
3206
3207	netif_dbg(dev, ifup, dev->net, "stop device");
3208
3209	mutex_lock(&dev->dev_mutex);
3210
3211	if (timer_pending(&dev->stat_monitor))
3212		del_timer_sync(&dev->stat_monitor);
3213
 
 
 
3214	clear_bit(EVENT_DEV_OPEN, &dev->flags);
3215	netif_stop_queue(net);
3216	napi_disable(&dev->napi);
3217
3218	lan78xx_terminate_urbs(dev);
3219
3220	netif_info(dev, ifdown, dev->net,
3221		   "stop stats: rx/tx %lu/%lu, errs %lu/%lu\n",
3222		   net->stats.rx_packets, net->stats.tx_packets,
3223		   net->stats.rx_errors, net->stats.tx_errors);
3224
3225	/* ignore errors that occur stopping the Tx and Rx data paths */
3226	lan78xx_stop_tx_path(dev);
3227	lan78xx_stop_rx_path(dev);
3228
3229	if (net->phydev)
3230		phy_stop(net->phydev);
3231
3232	usb_kill_urb(dev->urb_intr);
3233
3234	/* deferred work (task, timer, softirq) must also stop.
3235	 * can't flush_scheduled_work() until we drop rtnl (later),
3236	 * else workers could deadlock; so make workers a NOP.
3237	 */
3238	clear_bit(EVENT_TX_HALT, &dev->flags);
3239	clear_bit(EVENT_RX_HALT, &dev->flags);
3240	clear_bit(EVENT_LINK_RESET, &dev->flags);
3241	clear_bit(EVENT_STAT_UPDATE, &dev->flags);
3242
3243	cancel_delayed_work_sync(&dev->wq);
 
3244
3245	usb_autopm_put_interface(dev->intf);
3246
3247	mutex_unlock(&dev->dev_mutex);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3248
3249	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3250}
3251
3252static enum skb_state defer_bh(struct lan78xx_net *dev, struct sk_buff *skb,
3253			       struct sk_buff_head *list, enum skb_state state)
3254{
3255	unsigned long flags;
3256	enum skb_state old_state;
3257	struct skb_data *entry = (struct skb_data *)skb->cb;
3258
3259	spin_lock_irqsave(&list->lock, flags);
3260	old_state = entry->state;
3261	entry->state = state;
3262
3263	__skb_unlink(skb, list);
3264	spin_unlock(&list->lock);
3265	spin_lock(&dev->rxq_done.lock);
3266
3267	__skb_queue_tail(&dev->rxq_done, skb);
3268	if (skb_queue_len(&dev->rxq_done) == 1)
3269		napi_schedule(&dev->napi);
3270
3271	spin_unlock_irqrestore(&dev->rxq_done.lock, flags);
3272
3273	return old_state;
3274}
3275
3276static void tx_complete(struct urb *urb)
3277{
3278	struct sk_buff *skb = (struct sk_buff *)urb->context;
3279	struct skb_data *entry = (struct skb_data *)skb->cb;
3280	struct lan78xx_net *dev = entry->dev;
3281
3282	if (urb->status == 0) {
3283		dev->net->stats.tx_packets += entry->num_of_packet;
3284		dev->net->stats.tx_bytes += entry->length;
3285	} else {
3286		dev->net->stats.tx_errors += entry->num_of_packet;
3287
3288		switch (urb->status) {
3289		case -EPIPE:
3290			lan78xx_defer_kevent(dev, EVENT_TX_HALT);
3291			break;
3292
3293		/* software-driven interface shutdown */
3294		case -ECONNRESET:
3295		case -ESHUTDOWN:
3296			netif_dbg(dev, tx_err, dev->net,
3297				  "tx err interface gone %d\n",
3298				  entry->urb->status);
3299			break;
3300
3301		case -EPROTO:
3302		case -ETIME:
3303		case -EILSEQ:
3304			netif_stop_queue(dev->net);
3305			netif_dbg(dev, tx_err, dev->net,
3306				  "tx err queue stopped %d\n",
3307				  entry->urb->status);
3308			break;
3309		default:
3310			netif_dbg(dev, tx_err, dev->net,
3311				  "unknown tx err %d\n",
3312				  entry->urb->status);
3313			break;
3314		}
3315	}
3316
3317	usb_autopm_put_interface_async(dev->intf);
3318
3319	skb_unlink(skb, &dev->txq);
3320
3321	lan78xx_release_tx_buf(dev, skb);
3322
3323	/* Re-schedule NAPI if Tx data pending but no URBs in progress.
3324	 */
3325	if (skb_queue_empty(&dev->txq) &&
3326	    !skb_queue_empty(&dev->txq_pend))
3327		napi_schedule(&dev->napi);
3328}
3329
3330static void lan78xx_queue_skb(struct sk_buff_head *list,
3331			      struct sk_buff *newsk, enum skb_state state)
3332{
3333	struct skb_data *entry = (struct skb_data *)newsk->cb;
3334
3335	__skb_queue_tail(list, newsk);
3336	entry->state = state;
3337}
3338
3339static unsigned int lan78xx_tx_urb_space(struct lan78xx_net *dev)
3340{
3341	return skb_queue_len(&dev->txq_free) * dev->tx_urb_size;
3342}
3343
3344static unsigned int lan78xx_tx_pend_data_len(struct lan78xx_net *dev)
3345{
3346	return dev->tx_pend_data_len;
3347}
3348
3349static void lan78xx_tx_pend_skb_add(struct lan78xx_net *dev,
3350				    struct sk_buff *skb,
3351				    unsigned int *tx_pend_data_len)
3352{
3353	unsigned long flags;
3354
3355	spin_lock_irqsave(&dev->txq_pend.lock, flags);
3356
3357	__skb_queue_tail(&dev->txq_pend, skb);
3358
3359	dev->tx_pend_data_len += skb->len;
3360	*tx_pend_data_len = dev->tx_pend_data_len;
3361
3362	spin_unlock_irqrestore(&dev->txq_pend.lock, flags);
3363}
3364
3365static void lan78xx_tx_pend_skb_head_add(struct lan78xx_net *dev,
3366					 struct sk_buff *skb,
3367					 unsigned int *tx_pend_data_len)
3368{
3369	unsigned long flags;
3370
3371	spin_lock_irqsave(&dev->txq_pend.lock, flags);
3372
3373	__skb_queue_head(&dev->txq_pend, skb);
3374
3375	dev->tx_pend_data_len += skb->len;
3376	*tx_pend_data_len = dev->tx_pend_data_len;
3377
3378	spin_unlock_irqrestore(&dev->txq_pend.lock, flags);
3379}
3380
3381static void lan78xx_tx_pend_skb_get(struct lan78xx_net *dev,
3382				    struct sk_buff **skb,
3383				    unsigned int *tx_pend_data_len)
3384{
3385	unsigned long flags;
3386
3387	spin_lock_irqsave(&dev->txq_pend.lock, flags);
3388
3389	*skb = __skb_dequeue(&dev->txq_pend);
3390	if (*skb)
3391		dev->tx_pend_data_len -= (*skb)->len;
3392	*tx_pend_data_len = dev->tx_pend_data_len;
3393
3394	spin_unlock_irqrestore(&dev->txq_pend.lock, flags);
3395}
3396
3397static netdev_tx_t
3398lan78xx_start_xmit(struct sk_buff *skb, struct net_device *net)
3399{
3400	struct lan78xx_net *dev = netdev_priv(net);
3401	unsigned int tx_pend_data_len;
3402
3403	if (test_bit(EVENT_DEV_ASLEEP, &dev->flags))
3404		schedule_delayed_work(&dev->wq, 0);
 
 
3405
3406	skb_tx_timestamp(skb);
 
3407
3408	lan78xx_tx_pend_skb_add(dev, skb, &tx_pend_data_len);
3409
3410	/* Set up a Tx URB if none is in progress */
3411
3412	if (skb_queue_empty(&dev->txq))
3413		napi_schedule(&dev->napi);
3414
3415	/* Stop stack Tx queue if we have enough data to fill
3416	 * all the free Tx URBs.
3417	 */
3418	if (tx_pend_data_len > lan78xx_tx_urb_space(dev)) {
3419		netif_stop_queue(net);
3420
3421		netif_dbg(dev, hw, dev->net, "tx data len: %u, urb space %u",
3422			  tx_pend_data_len, lan78xx_tx_urb_space(dev));
3423
3424		/* Kick off transmission of pending data */
3425
3426		if (!skb_queue_empty(&dev->txq_free))
3427			napi_schedule(&dev->napi);
3428	}
3429
3430	return NETDEV_TX_OK;
3431}
3432
3433static int lan78xx_bind(struct lan78xx_net *dev, struct usb_interface *intf)
3434{
3435	struct lan78xx_priv *pdata = NULL;
3436	int ret;
3437	int i;
3438
3439	dev->data[0] = (unsigned long)kzalloc(sizeof(*pdata), GFP_KERNEL);
3440
3441	pdata = (struct lan78xx_priv *)(dev->data[0]);
3442	if (!pdata) {
3443		netdev_warn(dev->net, "Unable to allocate lan78xx_priv");
3444		return -ENOMEM;
3445	}
3446
3447	pdata->dev = dev;
3448
3449	spin_lock_init(&pdata->rfe_ctl_lock);
3450	mutex_init(&pdata->dataport_mutex);
3451
3452	INIT_WORK(&pdata->set_multicast, lan78xx_deferred_multicast_write);
3453
3454	for (i = 0; i < DP_SEL_VHF_VLAN_LEN; i++)
3455		pdata->vlan_table[i] = 0;
3456
3457	INIT_WORK(&pdata->set_vlan, lan78xx_deferred_vlan_write);
3458
3459	dev->net->features = 0;
3460
3461	if (DEFAULT_TX_CSUM_ENABLE)
3462		dev->net->features |= NETIF_F_HW_CSUM;
3463
3464	if (DEFAULT_RX_CSUM_ENABLE)
3465		dev->net->features |= NETIF_F_RXCSUM;
3466
3467	if (DEFAULT_TSO_CSUM_ENABLE)
3468		dev->net->features |= NETIF_F_TSO | NETIF_F_TSO6 | NETIF_F_SG;
3469
3470	if (DEFAULT_VLAN_RX_OFFLOAD)
3471		dev->net->features |= NETIF_F_HW_VLAN_CTAG_RX;
3472
3473	if (DEFAULT_VLAN_FILTER_ENABLE)
3474		dev->net->features |= NETIF_F_HW_VLAN_CTAG_FILTER;
3475
3476	dev->net->hw_features = dev->net->features;
3477
3478	ret = lan78xx_setup_irq_domain(dev);
3479	if (ret < 0) {
3480		netdev_warn(dev->net,
3481			    "lan78xx_setup_irq_domain() failed : %d", ret);
3482		goto out1;
3483	}
3484
 
 
 
3485	/* Init all registers */
3486	ret = lan78xx_reset(dev);
3487	if (ret) {
3488		netdev_warn(dev->net, "Registers INIT FAILED....");
3489		goto out2;
3490	}
3491
3492	ret = lan78xx_mdio_init(dev);
3493	if (ret) {
3494		netdev_warn(dev->net, "MDIO INIT FAILED.....");
3495		goto out2;
3496	}
3497
3498	dev->net->flags |= IFF_MULTICAST;
3499
3500	pdata->wol = WAKE_MAGIC;
3501
3502	return ret;
3503
3504out2:
3505	lan78xx_remove_irq_domain(dev);
3506
3507out1:
3508	netdev_warn(dev->net, "Bind routine FAILED");
3509	cancel_work_sync(&pdata->set_multicast);
3510	cancel_work_sync(&pdata->set_vlan);
3511	kfree(pdata);
3512	return ret;
3513}
3514
3515static void lan78xx_unbind(struct lan78xx_net *dev, struct usb_interface *intf)
3516{
3517	struct lan78xx_priv *pdata = (struct lan78xx_priv *)(dev->data[0]);
3518
3519	lan78xx_remove_irq_domain(dev);
3520
3521	lan78xx_remove_mdio(dev);
3522
3523	if (pdata) {
3524		cancel_work_sync(&pdata->set_multicast);
3525		cancel_work_sync(&pdata->set_vlan);
3526		netif_dbg(dev, ifdown, dev->net, "free pdata");
3527		kfree(pdata);
3528		pdata = NULL;
3529		dev->data[0] = 0;
3530	}
3531}
3532
3533static void lan78xx_rx_csum_offload(struct lan78xx_net *dev,
3534				    struct sk_buff *skb,
3535				    u32 rx_cmd_a, u32 rx_cmd_b)
3536{
3537	/* HW Checksum offload appears to be flawed if used when not stripping
3538	 * VLAN headers. Drop back to S/W checksums under these conditions.
3539	 */
3540	if (!(dev->net->features & NETIF_F_RXCSUM) ||
3541	    unlikely(rx_cmd_a & RX_CMD_A_ICSM_) ||
3542	    ((rx_cmd_a & RX_CMD_A_FVTG_) &&
3543	     !(dev->net->features & NETIF_F_HW_VLAN_CTAG_RX))) {
3544		skb->ip_summed = CHECKSUM_NONE;
3545	} else {
3546		skb->csum = ntohs((u16)(rx_cmd_b >> RX_CMD_B_CSUM_SHIFT_));
3547		skb->ip_summed = CHECKSUM_COMPLETE;
3548	}
3549}
3550
3551static void lan78xx_rx_vlan_offload(struct lan78xx_net *dev,
3552				    struct sk_buff *skb,
3553				    u32 rx_cmd_a, u32 rx_cmd_b)
3554{
3555	if ((dev->net->features & NETIF_F_HW_VLAN_CTAG_RX) &&
3556	    (rx_cmd_a & RX_CMD_A_FVTG_))
3557		__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q),
3558				       (rx_cmd_b & 0xffff));
3559}
3560
3561static void lan78xx_skb_return(struct lan78xx_net *dev, struct sk_buff *skb)
3562{
 
 
 
 
 
 
 
3563	dev->net->stats.rx_packets++;
3564	dev->net->stats.rx_bytes += skb->len;
3565
3566	skb->protocol = eth_type_trans(skb, dev->net);
3567
3568	netif_dbg(dev, rx_status, dev->net, "< rx, len %zu, type 0x%x\n",
3569		  skb->len + sizeof(struct ethhdr), skb->protocol);
3570	memset(skb->cb, 0, sizeof(struct skb_data));
3571
3572	if (skb_defer_rx_timestamp(skb))
3573		return;
3574
3575	napi_gro_receive(&dev->napi, skb);
 
 
 
3576}
3577
3578static int lan78xx_rx(struct lan78xx_net *dev, struct sk_buff *skb,
3579		      int budget, int *work_done)
3580{
3581	if (skb->len < RX_SKB_MIN_LEN)
3582		return 0;
3583
3584	/* Extract frames from the URB buffer and pass each one to
3585	 * the stack in a new NAPI SKB.
3586	 */
3587	while (skb->len > 0) {
3588		u32 rx_cmd_a, rx_cmd_b, align_count, size;
3589		u16 rx_cmd_c;
 
3590		unsigned char *packet;
3591
3592		rx_cmd_a = get_unaligned_le32(skb->data);
3593		skb_pull(skb, sizeof(rx_cmd_a));
3594
3595		rx_cmd_b = get_unaligned_le32(skb->data);
3596		skb_pull(skb, sizeof(rx_cmd_b));
3597
3598		rx_cmd_c = get_unaligned_le16(skb->data);
3599		skb_pull(skb, sizeof(rx_cmd_c));
3600
3601		packet = skb->data;
3602
3603		/* get the packet length */
3604		size = (rx_cmd_a & RX_CMD_A_LEN_MASK_);
3605		align_count = (4 - ((size + RXW_PADDING) % 4)) % 4;
3606
3607		if (unlikely(rx_cmd_a & RX_CMD_A_RED_)) {
3608			netif_dbg(dev, rx_err, dev->net,
3609				  "Error rx_cmd_a=0x%08x", rx_cmd_a);
3610		} else {
3611			u32 frame_len = size - ETH_FCS_LEN;
3612			struct sk_buff *skb2;
 
 
 
 
3613
3614			skb2 = napi_alloc_skb(&dev->napi, frame_len);
3615			if (!skb2)
 
 
 
 
 
 
 
3616				return 0;
 
3617
3618			memcpy(skb2->data, packet, frame_len);
3619
3620			skb_put(skb2, frame_len);
3621
3622			lan78xx_rx_csum_offload(dev, skb2, rx_cmd_a, rx_cmd_b);
3623			lan78xx_rx_vlan_offload(dev, skb2, rx_cmd_a, rx_cmd_b);
3624
3625			/* Processing of the URB buffer must complete once
3626			 * it has started. If the NAPI work budget is exhausted
3627			 * while frames remain they are added to the overflow
3628			 * queue for delivery in the next NAPI polling cycle.
3629			 */
3630			if (*work_done < budget) {
3631				lan78xx_skb_return(dev, skb2);
3632				++(*work_done);
3633			} else {
3634				skb_queue_tail(&dev->rxq_overflow, skb2);
3635			}
3636		}
3637
3638		skb_pull(skb, size);
3639
3640		/* skip padding bytes before the next frame starts */
3641		if (skb->len)
3642			skb_pull(skb, align_count);
3643	}
3644
3645	return 1;
3646}
3647
3648static inline void rx_process(struct lan78xx_net *dev, struct sk_buff *skb,
3649			      int budget, int *work_done)
3650{
3651	if (!lan78xx_rx(dev, skb, budget, work_done)) {
3652		netif_dbg(dev, rx_err, dev->net, "drop\n");
3653		dev->net->stats.rx_errors++;
 
 
 
 
 
 
3654	}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3655}
3656
3657static void rx_complete(struct urb *urb)
3658{
3659	struct sk_buff	*skb = (struct sk_buff *)urb->context;
3660	struct skb_data	*entry = (struct skb_data *)skb->cb;
3661	struct lan78xx_net *dev = entry->dev;
3662	int urb_status = urb->status;
3663	enum skb_state state;
3664
3665	netif_dbg(dev, rx_status, dev->net,
3666		  "rx done: status %d", urb->status);
3667
3668	skb_put(skb, urb->actual_length);
3669	state = rx_done;
3670
3671	if (urb != entry->urb)
3672		netif_warn(dev, rx_err, dev->net, "URB pointer mismatch");
3673
3674	switch (urb_status) {
3675	case 0:
3676		if (skb->len < RX_SKB_MIN_LEN) {
3677			state = rx_cleanup;
3678			dev->net->stats.rx_errors++;
3679			dev->net->stats.rx_length_errors++;
3680			netif_dbg(dev, rx_err, dev->net,
3681				  "rx length %d\n", skb->len);
3682		}
3683		usb_mark_last_busy(dev->udev);
3684		break;
3685	case -EPIPE:
3686		dev->net->stats.rx_errors++;
3687		lan78xx_defer_kevent(dev, EVENT_RX_HALT);
3688		fallthrough;
3689	case -ECONNRESET:				/* async unlink */
3690	case -ESHUTDOWN:				/* hardware gone */
3691		netif_dbg(dev, ifdown, dev->net,
3692			  "rx shutdown, code %d\n", urb_status);
3693		state = rx_cleanup;
 
 
3694		break;
3695	case -EPROTO:
3696	case -ETIME:
3697	case -EILSEQ:
3698		dev->net->stats.rx_errors++;
3699		state = rx_cleanup;
 
 
3700		break;
3701
3702	/* data overrun ... flush fifo? */
3703	case -EOVERFLOW:
3704		dev->net->stats.rx_over_errors++;
3705		fallthrough;
3706
3707	default:
3708		state = rx_cleanup;
3709		dev->net->stats.rx_errors++;
3710		netif_dbg(dev, rx_err, dev->net, "rx status %d\n", urb_status);
3711		break;
3712	}
3713
3714	state = defer_bh(dev, skb, &dev->rxq, state);
3715}
3716
3717static int rx_submit(struct lan78xx_net *dev, struct sk_buff *skb, gfp_t flags)
3718{
3719	struct skb_data	*entry = (struct skb_data *)skb->cb;
3720	size_t size = dev->rx_urb_size;
3721	struct urb *urb = entry->urb;
3722	unsigned long lockflags;
3723	int ret = 0;
3724
3725	usb_fill_bulk_urb(urb, dev->udev, dev->pipe_in,
3726			  skb->data, size, rx_complete, skb);
3727
3728	spin_lock_irqsave(&dev->rxq.lock, lockflags);
3729
3730	if (netif_device_present(dev->net) &&
3731	    netif_running(dev->net) &&
3732	    !test_bit(EVENT_RX_HALT, &dev->flags) &&
3733	    !test_bit(EVENT_DEV_ASLEEP, &dev->flags)) {
3734		ret = usb_submit_urb(urb, flags);
3735		switch (ret) {
3736		case 0:
3737			lan78xx_queue_skb(&dev->rxq, skb, rx_start);
3738			break;
3739		case -EPIPE:
3740			lan78xx_defer_kevent(dev, EVENT_RX_HALT);
3741			break;
3742		case -ENODEV:
3743		case -ENOENT:
3744			netif_dbg(dev, ifdown, dev->net, "device gone\n");
3745			netif_device_detach(dev->net);
3746			break;
3747		case -EHOSTUNREACH:
3748			ret = -ENOLINK;
3749			napi_schedule(&dev->napi);
3750			break;
3751		default:
3752			netif_dbg(dev, rx_err, dev->net,
3753				  "rx submit, %d\n", ret);
3754			napi_schedule(&dev->napi);
3755			break;
3756		}
3757	} else {
3758		netif_dbg(dev, ifdown, dev->net, "rx: stopped\n");
3759		ret = -ENOLINK;
3760	}
3761	spin_unlock_irqrestore(&dev->rxq.lock, lockflags);
3762
3763	if (ret)
3764		lan78xx_release_rx_buf(dev, skb);
3765
3766	return ret;
3767}
3768
3769static void lan78xx_rx_urb_submit_all(struct lan78xx_net *dev)
3770{
3771	struct sk_buff *rx_buf;
 
 
 
 
 
 
 
 
3772
3773	/* Ensure the maximum number of Rx URBs is submitted
3774	 */
3775	while ((rx_buf = lan78xx_get_rx_buf(dev)) != NULL) {
3776		if (rx_submit(dev, rx_buf, GFP_ATOMIC) != 0)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3777			break;
 
 
3778	}
3779}
3780
3781static void lan78xx_rx_urb_resubmit(struct lan78xx_net *dev,
3782				    struct sk_buff *rx_buf)
3783{
3784	/* reset SKB data pointers */
3785
3786	rx_buf->data = rx_buf->head;
3787	skb_reset_tail_pointer(rx_buf);
3788	rx_buf->len = 0;
3789	rx_buf->data_len = 0;
3790
3791	rx_submit(dev, rx_buf, GFP_ATOMIC);
3792}
 
 
 
 
 
 
 
3793
3794static void lan78xx_fill_tx_cmd_words(struct sk_buff *skb, u8 *buffer)
3795{
3796	u32 tx_cmd_a;
3797	u32 tx_cmd_b;
3798
3799	tx_cmd_a = (u32)(skb->len & TX_CMD_A_LEN_MASK_) | TX_CMD_A_FCS_;
 
 
 
 
3800
3801	if (skb->ip_summed == CHECKSUM_PARTIAL)
3802		tx_cmd_a |= TX_CMD_A_IPE_ | TX_CMD_A_TPE_;
 
 
 
 
3803
3804	tx_cmd_b = 0;
3805	if (skb_is_gso(skb)) {
3806		u16 mss = max(skb_shinfo(skb)->gso_size, TX_CMD_B_MSS_MIN_);
3807
3808		tx_cmd_b = (mss << TX_CMD_B_MSS_SHIFT_) & TX_CMD_B_MSS_MASK_;
 
 
 
3809
3810		tx_cmd_a |= TX_CMD_A_LSO_;
 
 
 
 
 
 
 
 
 
 
3811	}
 
3812
3813	if (skb_vlan_tag_present(skb)) {
3814		tx_cmd_a |= TX_CMD_A_IVTG_;
3815		tx_cmd_b |= skb_vlan_tag_get(skb) & TX_CMD_B_VTAG_MASK_;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3816	}
3817
3818	put_unaligned_le32(tx_cmd_a, buffer);
3819	put_unaligned_le32(tx_cmd_b, buffer + 4);
3820}
3821
3822static struct skb_data *lan78xx_tx_buf_fill(struct lan78xx_net *dev,
3823					    struct sk_buff *tx_buf)
3824{
3825	struct skb_data *entry = (struct skb_data *)tx_buf->cb;
3826	int remain = dev->tx_urb_size;
3827	u8 *tx_data = tx_buf->data;
3828	u32 urb_len = 0;
3829
3830	entry->num_of_packet = 0;
3831	entry->length = 0;
3832
3833	/* Work through the pending SKBs and copy the data of each SKB into
3834	 * the URB buffer if there room for all the SKB data.
3835	 *
3836	 * There must be at least DST+SRC+TYPE in the SKB (with padding enabled)
3837	 */
3838	while (remain >= TX_SKB_MIN_LEN) {
3839		unsigned int pending_bytes;
3840		unsigned int align_bytes;
3841		struct sk_buff *skb;
3842		unsigned int len;
3843
3844		lan78xx_tx_pend_skb_get(dev, &skb, &pending_bytes);
3845
3846		if (!skb)
3847			break;
3848
3849		align_bytes = (TX_ALIGNMENT - (urb_len % TX_ALIGNMENT)) %
3850			      TX_ALIGNMENT;
3851		len = align_bytes + TX_CMD_LEN + skb->len;
3852		if (len > remain) {
3853			lan78xx_tx_pend_skb_head_add(dev, skb, &pending_bytes);
3854			break;
3855		}
3856
3857		tx_data += align_bytes;
3858
3859		lan78xx_fill_tx_cmd_words(skb, tx_data);
3860		tx_data += TX_CMD_LEN;
3861
3862		len = skb->len;
3863		if (skb_copy_bits(skb, 0, tx_data, len) < 0) {
3864			struct net_device_stats *stats = &dev->net->stats;
3865
3866			stats->tx_dropped++;
3867			dev_kfree_skb_any(skb);
3868			tx_data -= TX_CMD_LEN;
3869			continue;
3870		}
3871
3872		tx_data += len;
3873		entry->length += len;
3874		entry->num_of_packet += skb_shinfo(skb)->gso_segs ?: 1;
3875
3876		dev_kfree_skb_any(skb);
3877
3878		urb_len = (u32)(tx_data - (u8 *)tx_buf->data);
3879
3880		remain = dev->tx_urb_size - urb_len;
3881	}
3882
3883	skb_put(tx_buf, urb_len);
3884
3885	return entry;
3886}
3887
3888static void lan78xx_tx_bh(struct lan78xx_net *dev)
3889{
3890	int ret;
 
3891
3892	/* Start the stack Tx queue if it was stopped
3893	 */
3894	netif_tx_lock(dev->net);
3895	if (netif_queue_stopped(dev->net)) {
3896		if (lan78xx_tx_pend_data_len(dev) < lan78xx_tx_urb_space(dev))
3897			netif_wake_queue(dev->net);
3898	}
3899	netif_tx_unlock(dev->net);
3900
3901	/* Go through the Tx pending queue and set up URBs to transfer
3902	 * the data to the device. Stop if no more pending data or URBs,
3903	 * or if an error occurs when a URB is submitted.
3904	 */
3905	do {
3906		struct skb_data *entry;
3907		struct sk_buff *tx_buf;
3908		unsigned long flags;
3909
3910		if (skb_queue_empty(&dev->txq_pend))
3911			break;
3912
3913		tx_buf = lan78xx_get_tx_buf(dev);
3914		if (!tx_buf)
3915			break;
3916
3917		entry = lan78xx_tx_buf_fill(dev, tx_buf);
3918
3919		spin_lock_irqsave(&dev->txq.lock, flags);
3920		ret = usb_autopm_get_interface_async(dev->intf);
3921		if (ret < 0) {
3922			spin_unlock_irqrestore(&dev->txq.lock, flags);
3923			goto out;
3924		}
3925
3926		usb_fill_bulk_urb(entry->urb, dev->udev, dev->pipe_out,
3927				  tx_buf->data, tx_buf->len, tx_complete,
3928				  tx_buf);
3929
3930		if (tx_buf->len % dev->maxpacket == 0) {
3931			/* send USB_ZERO_PACKET */
3932			entry->urb->transfer_flags |= URB_ZERO_PACKET;
3933		}
3934
3935#ifdef CONFIG_PM
3936		/* if device is asleep stop outgoing packet processing */
3937		if (test_bit(EVENT_DEV_ASLEEP, &dev->flags)) {
3938			usb_anchor_urb(entry->urb, &dev->deferred);
3939			netif_stop_queue(dev->net);
3940			spin_unlock_irqrestore(&dev->txq.lock, flags);
3941			netdev_dbg(dev->net,
3942				   "Delaying transmission for resumption\n");
3943			return;
3944		}
3945#endif
3946		ret = usb_submit_urb(entry->urb, GFP_ATOMIC);
3947		switch (ret) {
3948		case 0:
3949			netif_trans_update(dev->net);
3950			lan78xx_queue_skb(&dev->txq, tx_buf, tx_start);
3951			break;
3952		case -EPIPE:
3953			netif_stop_queue(dev->net);
3954			lan78xx_defer_kevent(dev, EVENT_TX_HALT);
3955			usb_autopm_put_interface_async(dev->intf);
3956			break;
3957		case -ENODEV:
3958		case -ENOENT:
3959			netif_dbg(dev, tx_err, dev->net,
3960				  "tx submit urb err %d (disconnected?)", ret);
3961			netif_device_detach(dev->net);
3962			break;
3963		default:
3964			usb_autopm_put_interface_async(dev->intf);
3965			netif_dbg(dev, tx_err, dev->net,
3966				  "tx submit urb err %d\n", ret);
3967			break;
3968		}
3969
3970		spin_unlock_irqrestore(&dev->txq.lock, flags);
3971
3972		if (ret) {
3973			netdev_warn(dev->net, "failed to tx urb %d\n", ret);
3974out:
3975			dev->net->stats.tx_dropped += entry->num_of_packet;
3976			lan78xx_release_tx_buf(dev, tx_buf);
3977		}
3978	} while (ret == 0);
3979}
3980
3981static int lan78xx_bh(struct lan78xx_net *dev, int budget)
3982{
3983	struct sk_buff_head done;
3984	struct sk_buff *rx_buf;
3985	struct skb_data *entry;
3986	unsigned long flags;
3987	int work_done = 0;
3988
3989	/* Pass frames received in the last NAPI cycle before
3990	 * working on newly completed URBs.
3991	 */
3992	while (!skb_queue_empty(&dev->rxq_overflow)) {
3993		lan78xx_skb_return(dev, skb_dequeue(&dev->rxq_overflow));
3994		++work_done;
3995	}
3996
3997	/* Take a snapshot of the done queue and move items to a
3998	 * temporary queue. Rx URB completions will continue to add
3999	 * to the done queue.
4000	 */
4001	__skb_queue_head_init(&done);
4002
4003	spin_lock_irqsave(&dev->rxq_done.lock, flags);
4004	skb_queue_splice_init(&dev->rxq_done, &done);
4005	spin_unlock_irqrestore(&dev->rxq_done.lock, flags);
4006
4007	/* Extract receive frames from completed URBs and
4008	 * pass them to the stack. Re-submit each completed URB.
4009	 */
4010	while ((work_done < budget) &&
4011	       (rx_buf = __skb_dequeue(&done))) {
4012		entry = (struct skb_data *)(rx_buf->cb);
4013		switch (entry->state) {
4014		case rx_done:
4015			rx_process(dev, rx_buf, budget, &work_done);
4016			break;
 
 
 
 
 
4017		case rx_cleanup:
4018			break;
 
 
4019		default:
4020			netdev_dbg(dev->net, "rx buf state %d\n",
4021				   entry->state);
4022			break;
4023		}
4024
4025		lan78xx_rx_urb_resubmit(dev, rx_buf);
4026	}
4027
4028	/* If budget was consumed before processing all the URBs put them
4029	 * back on the front of the done queue. They will be first to be
4030	 * processed in the next NAPI cycle.
4031	 */
4032	spin_lock_irqsave(&dev->rxq_done.lock, flags);
4033	skb_queue_splice(&done, &dev->rxq_done);
4034	spin_unlock_irqrestore(&dev->rxq_done.lock, flags);
4035
4036	if (netif_device_present(dev->net) && netif_running(dev->net)) {
4037		/* reset update timer delta */
4038		if (timer_pending(&dev->stat_monitor) && (dev->delta != 1)) {
4039			dev->delta = 1;
4040			mod_timer(&dev->stat_monitor,
4041				  jiffies + STAT_UPDATE_TIMER);
4042		}
4043
4044		/* Submit all free Rx URBs */
4045
4046		if (!test_bit(EVENT_RX_HALT, &dev->flags))
4047			lan78xx_rx_urb_submit_all(dev);
4048
4049		/* Submit new Tx URBs */
4050
4051		lan78xx_tx_bh(dev);
4052	}
4053
4054	return work_done;
4055}
4056
4057static int lan78xx_poll(struct napi_struct *napi, int budget)
4058{
4059	struct lan78xx_net *dev = container_of(napi, struct lan78xx_net, napi);
4060	int result = budget;
4061	int work_done;
4062
4063	/* Don't do any work if the device is suspended */
4064
4065	if (test_bit(EVENT_DEV_ASLEEP, &dev->flags)) {
4066		napi_complete_done(napi, 0);
4067		return 0;
4068	}
4069
4070	/* Process completed URBs and submit new URBs */
4071
4072	work_done = lan78xx_bh(dev, budget);
4073
4074	if (work_done < budget) {
4075		napi_complete_done(napi, work_done);
4076
4077		/* Start a new polling cycle if data was received or
4078		 * data is waiting to be transmitted.
4079		 */
4080		if (!skb_queue_empty(&dev->rxq_done)) {
4081			napi_schedule(napi);
4082		} else if (netif_carrier_ok(dev->net)) {
4083			if (skb_queue_empty(&dev->txq) &&
4084			    !skb_queue_empty(&dev->txq_pend)) {
4085				napi_schedule(napi);
4086			} else {
4087				netif_tx_lock(dev->net);
4088				if (netif_queue_stopped(dev->net)) {
4089					netif_wake_queue(dev->net);
4090					napi_schedule(napi);
4091				}
4092				netif_tx_unlock(dev->net);
4093			}
4094		}
4095		result = work_done;
4096	}
4097
4098	return result;
4099}
4100
4101static void lan78xx_delayedwork(struct work_struct *work)
4102{
4103	int status;
4104	struct lan78xx_net *dev;
4105
4106	dev = container_of(work, struct lan78xx_net, wq.work);
4107
4108	if (test_bit(EVENT_DEV_DISCONNECT, &dev->flags))
4109		return;
4110
4111	if (usb_autopm_get_interface(dev->intf) < 0)
4112		return;
4113
4114	if (test_bit(EVENT_TX_HALT, &dev->flags)) {
4115		unlink_urbs(dev, &dev->txq);
4116
 
 
4117		status = usb_clear_halt(dev->udev, dev->pipe_out);
 
4118		if (status < 0 &&
4119		    status != -EPIPE &&
4120		    status != -ESHUTDOWN) {
4121			if (netif_msg_tx_err(dev))
 
4122				netdev_err(dev->net,
4123					   "can't clear tx halt, status %d\n",
4124					   status);
4125		} else {
4126			clear_bit(EVENT_TX_HALT, &dev->flags);
4127			if (status != -ESHUTDOWN)
4128				netif_wake_queue(dev->net);
4129		}
4130	}
4131
4132	if (test_bit(EVENT_RX_HALT, &dev->flags)) {
4133		unlink_urbs(dev, &dev->rxq);
 
 
 
4134		status = usb_clear_halt(dev->udev, dev->pipe_in);
 
4135		if (status < 0 &&
4136		    status != -EPIPE &&
4137		    status != -ESHUTDOWN) {
4138			if (netif_msg_rx_err(dev))
 
4139				netdev_err(dev->net,
4140					   "can't clear rx halt, status %d\n",
4141					   status);
4142		} else {
4143			clear_bit(EVENT_RX_HALT, &dev->flags);
4144			napi_schedule(&dev->napi);
4145		}
4146	}
4147
4148	if (test_bit(EVENT_LINK_RESET, &dev->flags)) {
4149		int ret = 0;
4150
4151		clear_bit(EVENT_LINK_RESET, &dev->flags);
 
 
 
4152		if (lan78xx_link_reset(dev) < 0) {
 
 
4153			netdev_info(dev->net, "link reset failed (%d)\n",
4154				    ret);
 
 
4155		}
4156	}
4157
4158	if (test_bit(EVENT_STAT_UPDATE, &dev->flags)) {
4159		lan78xx_update_stats(dev);
4160
4161		clear_bit(EVENT_STAT_UPDATE, &dev->flags);
4162
4163		mod_timer(&dev->stat_monitor,
4164			  jiffies + (STAT_UPDATE_TIMER * dev->delta));
4165
4166		dev->delta = min((dev->delta * 2), 50);
4167	}
4168
4169	usb_autopm_put_interface(dev->intf);
4170}
4171
4172static void intr_complete(struct urb *urb)
4173{
4174	struct lan78xx_net *dev = urb->context;
4175	int status = urb->status;
4176
4177	switch (status) {
4178	/* success */
4179	case 0:
4180		lan78xx_status(dev, urb);
4181		break;
4182
4183	/* software-driven interface shutdown */
4184	case -ENOENT:			/* urb killed */
4185	case -ENODEV:			/* hardware gone */
4186	case -ESHUTDOWN:		/* hardware gone */
4187		netif_dbg(dev, ifdown, dev->net,
4188			  "intr shutdown, code %d\n", status);
4189		return;
4190
4191	/* NOTE:  not throttling like RX/TX, since this endpoint
4192	 * already polls infrequently
4193	 */
4194	default:
4195		netdev_dbg(dev->net, "intr status %d\n", status);
4196		break;
4197	}
4198
4199	if (!netif_device_present(dev->net) ||
4200	    !netif_running(dev->net)) {
4201		netdev_warn(dev->net, "not submitting new status URB");
4202		return;
4203	}
4204
4205	memset(urb->transfer_buffer, 0, urb->transfer_buffer_length);
4206	status = usb_submit_urb(urb, GFP_ATOMIC);
4207
4208	switch (status) {
4209	case  0:
4210		break;
4211	case -ENODEV:
4212	case -ENOENT:
4213		netif_dbg(dev, timer, dev->net,
4214			  "intr resubmit %d (disconnect?)", status);
4215		netif_device_detach(dev->net);
4216		break;
4217	default:
4218		netif_err(dev, timer, dev->net,
4219			  "intr resubmit --> %d\n", status);
4220		break;
4221	}
4222}
4223
4224static void lan78xx_disconnect(struct usb_interface *intf)
4225{
4226	struct lan78xx_net *dev;
4227	struct usb_device *udev;
4228	struct net_device *net;
4229	struct phy_device *phydev;
4230
4231	dev = usb_get_intfdata(intf);
4232	usb_set_intfdata(intf, NULL);
4233	if (!dev)
4234		return;
4235
4236	set_bit(EVENT_DEV_DISCONNECT, &dev->flags);
4237
4238	netif_napi_del(&dev->napi);
4239
4240	udev = interface_to_usbdev(intf);
4241	net = dev->net;
4242
4243	unregister_netdev(net);
4244
4245	cancel_delayed_work_sync(&dev->wq);
4246
4247	phydev = net->phydev;
4248
4249	phy_unregister_fixup_for_uid(PHY_KSZ9031RNX, 0xfffffff0);
4250	phy_unregister_fixup_for_uid(PHY_LAN8835, 0xfffffff0);
4251
4252	phy_disconnect(net->phydev);
4253
4254	if (phy_is_pseudo_fixed_link(phydev))
4255		fixed_phy_unregister(phydev);
4256
 
 
 
 
4257	usb_scuttle_anchored_urbs(&dev->deferred);
4258
4259	if (timer_pending(&dev->stat_monitor))
4260		del_timer_sync(&dev->stat_monitor);
4261
4262	lan78xx_unbind(dev, intf);
4263
4264	lan78xx_free_tx_resources(dev);
4265	lan78xx_free_rx_resources(dev);
4266
4267	usb_kill_urb(dev->urb_intr);
4268	usb_free_urb(dev->urb_intr);
4269
4270	free_netdev(net);
4271	usb_put_dev(udev);
4272}
4273
4274static void lan78xx_tx_timeout(struct net_device *net, unsigned int txqueue)
4275{
4276	struct lan78xx_net *dev = netdev_priv(net);
4277
4278	unlink_urbs(dev, &dev->txq);
4279	napi_schedule(&dev->napi);
4280}
4281
4282static netdev_features_t lan78xx_features_check(struct sk_buff *skb,
4283						struct net_device *netdev,
4284						netdev_features_t features)
4285{
4286	struct lan78xx_net *dev = netdev_priv(netdev);
4287
4288	if (skb->len > LAN78XX_TSO_SIZE(dev))
4289		features &= ~NETIF_F_GSO_MASK;
4290
4291	features = vlan_features_check(skb, features);
4292	features = vxlan_features_check(skb, features);
4293
4294	return features;
4295}
4296
4297static const struct net_device_ops lan78xx_netdev_ops = {
4298	.ndo_open		= lan78xx_open,
4299	.ndo_stop		= lan78xx_stop,
4300	.ndo_start_xmit		= lan78xx_start_xmit,
4301	.ndo_tx_timeout		= lan78xx_tx_timeout,
4302	.ndo_change_mtu		= lan78xx_change_mtu,
4303	.ndo_set_mac_address	= lan78xx_set_mac_addr,
4304	.ndo_validate_addr	= eth_validate_addr,
4305	.ndo_eth_ioctl		= phy_do_ioctl_running,
4306	.ndo_set_rx_mode	= lan78xx_set_multicast,
4307	.ndo_set_features	= lan78xx_set_features,
4308	.ndo_vlan_rx_add_vid	= lan78xx_vlan_rx_add_vid,
4309	.ndo_vlan_rx_kill_vid	= lan78xx_vlan_rx_kill_vid,
4310	.ndo_features_check	= lan78xx_features_check,
4311};
4312
4313static void lan78xx_stat_monitor(struct timer_list *t)
4314{
4315	struct lan78xx_net *dev = from_timer(dev, t, stat_monitor);
4316
4317	lan78xx_defer_kevent(dev, EVENT_STAT_UPDATE);
4318}
4319
4320static int lan78xx_probe(struct usb_interface *intf,
4321			 const struct usb_device_id *id)
4322{
4323	struct usb_host_endpoint *ep_blkin, *ep_blkout, *ep_intr;
4324	struct lan78xx_net *dev;
4325	struct net_device *netdev;
4326	struct usb_device *udev;
4327	int ret;
4328	unsigned int maxp;
4329	unsigned int period;
4330	u8 *buf = NULL;
4331
4332	udev = interface_to_usbdev(intf);
4333	udev = usb_get_dev(udev);
4334
4335	netdev = alloc_etherdev(sizeof(struct lan78xx_net));
4336	if (!netdev) {
4337		dev_err(&intf->dev, "Error: OOM\n");
4338		ret = -ENOMEM;
4339		goto out1;
4340	}
4341
4342	/* netdev_printk() needs this */
4343	SET_NETDEV_DEV(netdev, &intf->dev);
4344
4345	dev = netdev_priv(netdev);
4346	dev->udev = udev;
4347	dev->intf = intf;
4348	dev->net = netdev;
4349	dev->msg_enable = netif_msg_init(msg_level, NETIF_MSG_DRV
4350					| NETIF_MSG_PROBE | NETIF_MSG_LINK);
4351
4352	skb_queue_head_init(&dev->rxq);
4353	skb_queue_head_init(&dev->txq);
4354	skb_queue_head_init(&dev->rxq_done);
 
4355	skb_queue_head_init(&dev->txq_pend);
4356	skb_queue_head_init(&dev->rxq_overflow);
4357	mutex_init(&dev->phy_mutex);
4358	mutex_init(&dev->dev_mutex);
4359
4360	ret = lan78xx_urb_config_init(dev);
4361	if (ret < 0)
4362		goto out2;
4363
4364	ret = lan78xx_alloc_tx_resources(dev);
4365	if (ret < 0)
4366		goto out2;
4367
4368	ret = lan78xx_alloc_rx_resources(dev);
4369	if (ret < 0)
4370		goto out3;
4371
4372	/* MTU range: 68 - 9000 */
4373	netdev->max_mtu = MAX_SINGLE_PACKET_SIZE;
4374
4375	netif_set_tso_max_size(netdev, LAN78XX_TSO_SIZE(dev));
4376
4377	netif_napi_add(netdev, &dev->napi, lan78xx_poll);
4378
 
4379	INIT_DELAYED_WORK(&dev->wq, lan78xx_delayedwork);
4380	init_usb_anchor(&dev->deferred);
4381
4382	netdev->netdev_ops = &lan78xx_netdev_ops;
4383	netdev->watchdog_timeo = TX_TIMEOUT_JIFFIES;
4384	netdev->ethtool_ops = &lan78xx_ethtool_ops;
4385
4386	dev->delta = 1;
4387	timer_setup(&dev->stat_monitor, lan78xx_stat_monitor, 0);
4388
4389	mutex_init(&dev->stats.access_lock);
4390
4391	if (intf->cur_altsetting->desc.bNumEndpoints < 3) {
4392		ret = -ENODEV;
4393		goto out4;
4394	}
4395
4396	dev->pipe_in = usb_rcvbulkpipe(udev, BULK_IN_PIPE);
4397	ep_blkin = usb_pipe_endpoint(udev, dev->pipe_in);
4398	if (!ep_blkin || !usb_endpoint_is_bulk_in(&ep_blkin->desc)) {
4399		ret = -ENODEV;
4400		goto out4;
4401	}
4402
4403	dev->pipe_out = usb_sndbulkpipe(udev, BULK_OUT_PIPE);
4404	ep_blkout = usb_pipe_endpoint(udev, dev->pipe_out);
4405	if (!ep_blkout || !usb_endpoint_is_bulk_out(&ep_blkout->desc)) {
4406		ret = -ENODEV;
4407		goto out4;
4408	}
4409
4410	ep_intr = &intf->cur_altsetting->endpoint[2];
4411	if (!usb_endpoint_is_int_in(&ep_intr->desc)) {
4412		ret = -ENODEV;
4413		goto out4;
4414	}
4415
4416	dev->pipe_intr = usb_rcvintpipe(dev->udev,
4417					usb_endpoint_num(&ep_intr->desc));
4418
4419	ret = lan78xx_bind(dev, intf);
4420	if (ret < 0)
4421		goto out4;
 
 
 
 
 
 
 
4422
4423	period = ep_intr->desc.bInterval;
4424	maxp = usb_maxpacket(dev->udev, dev->pipe_intr);
4425	buf = kmalloc(maxp, GFP_KERNEL);
4426	if (!buf) {
4427		ret = -ENOMEM;
4428		goto out5;
4429	}
4430
4431	dev->urb_intr = usb_alloc_urb(0, GFP_KERNEL);
4432	if (!dev->urb_intr) {
4433		ret = -ENOMEM;
4434		goto out6;
4435	} else {
4436		usb_fill_int_urb(dev->urb_intr, dev->udev,
4437				 dev->pipe_intr, buf, maxp,
4438				 intr_complete, dev, period);
4439		dev->urb_intr->transfer_flags |= URB_FREE_BUFFER;
4440	}
4441
4442	dev->maxpacket = usb_maxpacket(dev->udev, dev->pipe_out);
4443
4444	/* Reject broken descriptors. */
4445	if (dev->maxpacket == 0) {
4446		ret = -ENODEV;
4447		goto out6;
4448	}
4449
4450	/* driver requires remote-wakeup capability during autosuspend. */
4451	intf->needs_remote_wakeup = 1;
4452
4453	ret = lan78xx_phy_init(dev);
4454	if (ret < 0)
4455		goto out7;
4456
4457	ret = register_netdev(netdev);
4458	if (ret != 0) {
4459		netif_err(dev, probe, netdev, "couldn't register the device\n");
4460		goto out8;
4461	}
4462
4463	usb_set_intfdata(intf, dev);
4464
4465	ret = device_set_wakeup_enable(&udev->dev, true);
4466
4467	 /* Default delay of 2sec has more overhead than advantage.
4468	  * Set to 10sec as default.
4469	  */
4470	pm_runtime_set_autosuspend_delay(&udev->dev,
4471					 DEFAULT_AUTOSUSPEND_DELAY);
4472
4473	return 0;
4474
4475out8:
4476	phy_disconnect(netdev->phydev);
4477out7:
4478	usb_free_urb(dev->urb_intr);
4479out6:
4480	kfree(buf);
4481out5:
4482	lan78xx_unbind(dev, intf);
4483out4:
4484	netif_napi_del(&dev->napi);
4485	lan78xx_free_rx_resources(dev);
4486out3:
4487	lan78xx_free_tx_resources(dev);
4488out2:
4489	free_netdev(netdev);
4490out1:
4491	usb_put_dev(udev);
4492
4493	return ret;
4494}
4495
4496static u16 lan78xx_wakeframe_crc16(const u8 *buf, int len)
4497{
4498	const u16 crc16poly = 0x8005;
4499	int i;
4500	u16 bit, crc, msb;
4501	u8 data;
4502
4503	crc = 0xFFFF;
4504	for (i = 0; i < len; i++) {
4505		data = *buf++;
4506		for (bit = 0; bit < 8; bit++) {
4507			msb = crc >> 15;
4508			crc <<= 1;
4509
4510			if (msb ^ (u16)(data & 1)) {
4511				crc ^= crc16poly;
4512				crc |= (u16)0x0001U;
4513			}
4514			data >>= 1;
4515		}
4516	}
4517
4518	return crc;
4519}
4520
4521static int lan78xx_set_auto_suspend(struct lan78xx_net *dev)
4522{
4523	u32 buf;
4524	int ret;
4525
4526	ret = lan78xx_stop_tx_path(dev);
4527	if (ret < 0)
4528		return ret;
4529
4530	ret = lan78xx_stop_rx_path(dev);
4531	if (ret < 0)
4532		return ret;
4533
4534	/* auto suspend (selective suspend) */
4535
4536	ret = lan78xx_write_reg(dev, WUCSR, 0);
4537	if (ret < 0)
4538		return ret;
4539	ret = lan78xx_write_reg(dev, WUCSR2, 0);
4540	if (ret < 0)
4541		return ret;
4542	ret = lan78xx_write_reg(dev, WK_SRC, 0xFFF1FF1FUL);
4543	if (ret < 0)
4544		return ret;
4545
4546	/* set goodframe wakeup */
4547
4548	ret = lan78xx_read_reg(dev, WUCSR, &buf);
4549	if (ret < 0)
4550		return ret;
4551
4552	buf |= WUCSR_RFE_WAKE_EN_;
4553	buf |= WUCSR_STORE_WAKE_;
4554
4555	ret = lan78xx_write_reg(dev, WUCSR, buf);
4556	if (ret < 0)
4557		return ret;
4558
4559	ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
4560	if (ret < 0)
4561		return ret;
4562
4563	buf &= ~PMT_CTL_RES_CLR_WKP_EN_;
4564	buf |= PMT_CTL_RES_CLR_WKP_STS_;
4565	buf |= PMT_CTL_PHY_WAKE_EN_;
4566	buf |= PMT_CTL_WOL_EN_;
4567	buf &= ~PMT_CTL_SUS_MODE_MASK_;
4568	buf |= PMT_CTL_SUS_MODE_3_;
4569
4570	ret = lan78xx_write_reg(dev, PMT_CTL, buf);
4571	if (ret < 0)
4572		return ret;
4573
4574	ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
4575	if (ret < 0)
4576		return ret;
4577
4578	buf |= PMT_CTL_WUPS_MASK_;
4579
4580	ret = lan78xx_write_reg(dev, PMT_CTL, buf);
4581	if (ret < 0)
4582		return ret;
4583
4584	ret = lan78xx_start_rx_path(dev);
4585
4586	return ret;
4587}
4588
4589static int lan78xx_set_suspend(struct lan78xx_net *dev, u32 wol)
4590{
4591	const u8 ipv4_multicast[3] = { 0x01, 0x00, 0x5E };
4592	const u8 ipv6_multicast[3] = { 0x33, 0x33 };
4593	const u8 arp_type[2] = { 0x08, 0x06 };
4594	u32 temp_pmt_ctl;
4595	int mask_index;
4596	u32 temp_wucsr;
4597	u32 buf;
4598	u16 crc;
4599	int ret;
4600
4601	ret = lan78xx_stop_tx_path(dev);
4602	if (ret < 0)
4603		return ret;
4604	ret = lan78xx_stop_rx_path(dev);
4605	if (ret < 0)
4606		return ret;
4607
4608	ret = lan78xx_write_reg(dev, WUCSR, 0);
4609	if (ret < 0)
4610		return ret;
4611	ret = lan78xx_write_reg(dev, WUCSR2, 0);
4612	if (ret < 0)
4613		return ret;
4614	ret = lan78xx_write_reg(dev, WK_SRC, 0xFFF1FF1FUL);
4615	if (ret < 0)
4616		return ret;
4617
4618	temp_wucsr = 0;
4619
4620	temp_pmt_ctl = 0;
4621
4622	ret = lan78xx_read_reg(dev, PMT_CTL, &temp_pmt_ctl);
4623	if (ret < 0)
4624		return ret;
4625
4626	temp_pmt_ctl &= ~PMT_CTL_RES_CLR_WKP_EN_;
4627	temp_pmt_ctl |= PMT_CTL_RES_CLR_WKP_STS_;
4628
4629	for (mask_index = 0; mask_index < NUM_OF_WUF_CFG; mask_index++) {
4630		ret = lan78xx_write_reg(dev, WUF_CFG(mask_index), 0);
4631		if (ret < 0)
4632			return ret;
4633	}
4634
4635	mask_index = 0;
4636	if (wol & WAKE_PHY) {
4637		temp_pmt_ctl |= PMT_CTL_PHY_WAKE_EN_;
4638
4639		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4640		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4641		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
4642	}
4643	if (wol & WAKE_MAGIC) {
4644		temp_wucsr |= WUCSR_MPEN_;
4645
4646		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4647		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4648		temp_pmt_ctl |= PMT_CTL_SUS_MODE_3_;
4649	}
4650	if (wol & WAKE_BCAST) {
4651		temp_wucsr |= WUCSR_BCST_EN_;
4652
4653		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4654		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4655		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
4656	}
4657	if (wol & WAKE_MCAST) {
4658		temp_wucsr |= WUCSR_WAKE_EN_;
4659
4660		/* set WUF_CFG & WUF_MASK for IPv4 Multicast */
4661		crc = lan78xx_wakeframe_crc16(ipv4_multicast, 3);
4662		ret = lan78xx_write_reg(dev, WUF_CFG(mask_index),
4663					WUF_CFGX_EN_ |
4664					WUF_CFGX_TYPE_MCAST_ |
4665					(0 << WUF_CFGX_OFFSET_SHIFT_) |
4666					(crc & WUF_CFGX_CRC16_MASK_));
4667		if (ret < 0)
4668			return ret;
4669
4670		ret = lan78xx_write_reg(dev, WUF_MASK0(mask_index), 7);
4671		if (ret < 0)
4672			return ret;
4673		ret = lan78xx_write_reg(dev, WUF_MASK1(mask_index), 0);
4674		if (ret < 0)
4675			return ret;
4676		ret = lan78xx_write_reg(dev, WUF_MASK2(mask_index), 0);
4677		if (ret < 0)
4678			return ret;
4679		ret = lan78xx_write_reg(dev, WUF_MASK3(mask_index), 0);
4680		if (ret < 0)
4681			return ret;
4682
 
 
 
 
4683		mask_index++;
4684
4685		/* for IPv6 Multicast */
4686		crc = lan78xx_wakeframe_crc16(ipv6_multicast, 2);
4687		ret = lan78xx_write_reg(dev, WUF_CFG(mask_index),
4688					WUF_CFGX_EN_ |
4689					WUF_CFGX_TYPE_MCAST_ |
4690					(0 << WUF_CFGX_OFFSET_SHIFT_) |
4691					(crc & WUF_CFGX_CRC16_MASK_));
4692		if (ret < 0)
4693			return ret;
4694
4695		ret = lan78xx_write_reg(dev, WUF_MASK0(mask_index), 3);
4696		if (ret < 0)
4697			return ret;
4698		ret = lan78xx_write_reg(dev, WUF_MASK1(mask_index), 0);
4699		if (ret < 0)
4700			return ret;
4701		ret = lan78xx_write_reg(dev, WUF_MASK2(mask_index), 0);
4702		if (ret < 0)
4703			return ret;
4704		ret = lan78xx_write_reg(dev, WUF_MASK3(mask_index), 0);
4705		if (ret < 0)
4706			return ret;
4707
 
 
 
 
4708		mask_index++;
4709
4710		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4711		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4712		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
4713	}
4714	if (wol & WAKE_UCAST) {
4715		temp_wucsr |= WUCSR_PFDA_EN_;
4716
4717		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4718		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4719		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
4720	}
4721	if (wol & WAKE_ARP) {
4722		temp_wucsr |= WUCSR_WAKE_EN_;
4723
4724		/* set WUF_CFG & WUF_MASK
4725		 * for packettype (offset 12,13) = ARP (0x0806)
4726		 */
4727		crc = lan78xx_wakeframe_crc16(arp_type, 2);
4728		ret = lan78xx_write_reg(dev, WUF_CFG(mask_index),
4729					WUF_CFGX_EN_ |
4730					WUF_CFGX_TYPE_ALL_ |
4731					(0 << WUF_CFGX_OFFSET_SHIFT_) |
4732					(crc & WUF_CFGX_CRC16_MASK_));
4733		if (ret < 0)
4734			return ret;
4735
4736		ret = lan78xx_write_reg(dev, WUF_MASK0(mask_index), 0x3000);
4737		if (ret < 0)
4738			return ret;
4739		ret = lan78xx_write_reg(dev, WUF_MASK1(mask_index), 0);
4740		if (ret < 0)
4741			return ret;
4742		ret = lan78xx_write_reg(dev, WUF_MASK2(mask_index), 0);
4743		if (ret < 0)
4744			return ret;
4745		ret = lan78xx_write_reg(dev, WUF_MASK3(mask_index), 0);
4746		if (ret < 0)
4747			return ret;
4748
 
 
 
 
4749		mask_index++;
4750
4751		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4752		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4753		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
4754	}
4755
4756	ret = lan78xx_write_reg(dev, WUCSR, temp_wucsr);
4757	if (ret < 0)
4758		return ret;
4759
4760	/* when multiple WOL bits are set */
4761	if (hweight_long((unsigned long)wol) > 1) {
4762		temp_pmt_ctl |= PMT_CTL_WOL_EN_;
4763		temp_pmt_ctl &= ~PMT_CTL_SUS_MODE_MASK_;
4764		temp_pmt_ctl |= PMT_CTL_SUS_MODE_0_;
4765	}
4766	ret = lan78xx_write_reg(dev, PMT_CTL, temp_pmt_ctl);
4767	if (ret < 0)
4768		return ret;
4769
4770	/* clear WUPS */
4771	ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
4772	if (ret < 0)
4773		return ret;
4774
4775	buf |= PMT_CTL_WUPS_MASK_;
 
4776
4777	ret = lan78xx_write_reg(dev, PMT_CTL, buf);
4778	if (ret < 0)
4779		return ret;
4780
4781	ret = lan78xx_start_rx_path(dev);
4782
4783	return ret;
4784}
4785
4786static int lan78xx_suspend(struct usb_interface *intf, pm_message_t message)
4787{
4788	struct lan78xx_net *dev = usb_get_intfdata(intf);
4789	bool dev_open;
 
4790	int ret;
4791
4792	mutex_lock(&dev->dev_mutex);
4793
4794	netif_dbg(dev, ifdown, dev->net,
4795		  "suspending: pm event %#x", message.event);
4796
4797	dev_open = test_bit(EVENT_DEV_OPEN, &dev->flags);
4798
4799	if (dev_open) {
4800		spin_lock_irq(&dev->txq.lock);
4801		/* don't autosuspend while transmitting */
4802		if ((skb_queue_len(&dev->txq) ||
4803		     skb_queue_len(&dev->txq_pend)) &&
4804		    PMSG_IS_AUTO(message)) {
4805			spin_unlock_irq(&dev->txq.lock);
4806			ret = -EBUSY;
4807			goto out;
4808		} else {
4809			set_bit(EVENT_DEV_ASLEEP, &dev->flags);
4810			spin_unlock_irq(&dev->txq.lock);
4811		}
4812
4813		/* stop RX */
4814		ret = lan78xx_stop_rx_path(dev);
4815		if (ret < 0)
4816			goto out;
 
 
 
4817
4818		ret = lan78xx_flush_rx_fifo(dev);
4819		if (ret < 0)
4820			goto out;
4821
4822		/* stop Tx */
4823		ret = lan78xx_stop_tx_path(dev);
4824		if (ret < 0)
4825			goto out;
4826
4827		/* empty out the Rx and Tx queues */
4828		netif_device_detach(dev->net);
4829		lan78xx_terminate_urbs(dev);
4830		usb_kill_urb(dev->urb_intr);
4831
4832		/* reattach */
4833		netif_device_attach(dev->net);
 
4834
 
4835		del_timer(&dev->stat_monitor);
4836
4837		if (PMSG_IS_AUTO(message)) {
4838			ret = lan78xx_set_auto_suspend(dev);
4839			if (ret < 0)
4840				goto out;
4841		} else {
4842			struct lan78xx_priv *pdata;
4843
4844			pdata = (struct lan78xx_priv *)(dev->data[0]);
4845			netif_carrier_off(dev->net);
4846			ret = lan78xx_set_suspend(dev, pdata->wol);
4847			if (ret < 0)
4848				goto out;
4849		}
4850	} else {
4851		/* Interface is down; don't allow WOL and PHY
4852		 * events to wake up the host
4853		 */
4854		u32 buf;
4855
4856		set_bit(EVENT_DEV_ASLEEP, &dev->flags);
4857
4858		ret = lan78xx_write_reg(dev, WUCSR, 0);
4859		if (ret < 0)
4860			goto out;
4861		ret = lan78xx_write_reg(dev, WUCSR2, 0);
4862		if (ret < 0)
4863			goto out;
4864
4865		ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
4866		if (ret < 0)
4867			goto out;
4868
4869		buf &= ~PMT_CTL_RES_CLR_WKP_EN_;
4870		buf |= PMT_CTL_RES_CLR_WKP_STS_;
4871		buf &= ~PMT_CTL_SUS_MODE_MASK_;
4872		buf |= PMT_CTL_SUS_MODE_3_;
4873
4874		ret = lan78xx_write_reg(dev, PMT_CTL, buf);
4875		if (ret < 0)
4876			goto out;
4877
4878		ret = lan78xx_read_reg(dev, PMT_CTL, &buf);
4879		if (ret < 0)
4880			goto out;
4881
4882		buf |= PMT_CTL_WUPS_MASK_;
4883
4884		ret = lan78xx_write_reg(dev, PMT_CTL, buf);
4885		if (ret < 0)
4886			goto out;
4887	}
4888
4889	ret = 0;
4890out:
4891	mutex_unlock(&dev->dev_mutex);
4892
4893	return ret;
4894}
 
 
4895
4896static bool lan78xx_submit_deferred_urbs(struct lan78xx_net *dev)
4897{
4898	bool pipe_halted = false;
4899	struct urb *urb;
4900
4901	while ((urb = usb_get_from_anchor(&dev->deferred))) {
4902		struct sk_buff *skb = urb->context;
4903		int ret;
4904
4905		if (!netif_device_present(dev->net) ||
4906		    !netif_carrier_ok(dev->net) ||
4907		    pipe_halted) {
4908			lan78xx_release_tx_buf(dev, skb);
4909			continue;
4910		}
4911
4912		ret = usb_submit_urb(urb, GFP_ATOMIC);
4913
4914		if (ret == 0) {
4915			netif_trans_update(dev->net);
4916			lan78xx_queue_skb(&dev->txq, skb, tx_start);
4917		} else {
4918			if (ret == -EPIPE) {
4919				netif_stop_queue(dev->net);
4920				pipe_halted = true;
4921			} else if (ret == -ENODEV) {
4922				netif_device_detach(dev->net);
4923			}
4924
4925			lan78xx_release_tx_buf(dev, skb);
4926		}
4927	}
4928
4929	return pipe_halted;
 
 
4930}
4931
4932static int lan78xx_resume(struct usb_interface *intf)
4933{
4934	struct lan78xx_net *dev = usb_get_intfdata(intf);
4935	bool dev_open;
 
4936	int ret;
 
4937
4938	mutex_lock(&dev->dev_mutex);
 
 
 
 
4939
4940	netif_dbg(dev, ifup, dev->net, "resuming device");
4941
4942	dev_open = test_bit(EVENT_DEV_OPEN, &dev->flags);
4943
4944	if (dev_open) {
4945		bool pipe_halted = false;
4946
4947		ret = lan78xx_flush_tx_fifo(dev);
4948		if (ret < 0)
4949			goto out;
4950
4951		if (dev->urb_intr) {
4952			int ret = usb_submit_urb(dev->urb_intr, GFP_KERNEL);
4953
 
 
 
 
4954			if (ret < 0) {
4955				if (ret == -ENODEV)
4956					netif_device_detach(dev->net);
4957				netdev_warn(dev->net, "Failed to submit intr URB");
 
 
 
4958			}
4959		}
4960
4961		spin_lock_irq(&dev->txq.lock);
4962
4963		if (netif_device_present(dev->net)) {
4964			pipe_halted = lan78xx_submit_deferred_urbs(dev);
4965
4966			if (pipe_halted)
4967				lan78xx_defer_kevent(dev, EVENT_TX_HALT);
4968		}
4969
4970		clear_bit(EVENT_DEV_ASLEEP, &dev->flags);
4971
4972		spin_unlock_irq(&dev->txq.lock);
4973
4974		if (!pipe_halted &&
4975		    netif_device_present(dev->net) &&
4976		    (lan78xx_tx_pend_data_len(dev) < lan78xx_tx_urb_space(dev)))
4977			netif_start_queue(dev->net);
4978
4979		ret = lan78xx_start_tx_path(dev);
4980		if (ret < 0)
4981			goto out;
4982
4983		napi_schedule(&dev->napi);
4984
4985		if (!timer_pending(&dev->stat_monitor)) {
4986			dev->delta = 1;
4987			mod_timer(&dev->stat_monitor,
4988				  jiffies + STAT_UPDATE_TIMER);
4989		}
4990
4991	} else {
4992		clear_bit(EVENT_DEV_ASLEEP, &dev->flags);
4993	}
4994
4995	ret = lan78xx_write_reg(dev, WUCSR2, 0);
4996	if (ret < 0)
4997		goto out;
4998	ret = lan78xx_write_reg(dev, WUCSR, 0);
4999	if (ret < 0)
5000		goto out;
5001	ret = lan78xx_write_reg(dev, WK_SRC, 0xFFF1FF1FUL);
5002	if (ret < 0)
5003		goto out;
5004
5005	ret = lan78xx_write_reg(dev, WUCSR2, WUCSR2_NS_RCD_ |
5006					     WUCSR2_ARP_RCD_ |
5007					     WUCSR2_IPV6_TCPSYN_RCD_ |
5008					     WUCSR2_IPV4_TCPSYN_RCD_);
5009	if (ret < 0)
5010		goto out;
5011
5012	ret = lan78xx_write_reg(dev, WUCSR, WUCSR_EEE_TX_WAKE_ |
5013					    WUCSR_EEE_RX_WAKE_ |
5014					    WUCSR_PFDA_FR_ |
5015					    WUCSR_RFE_WAKE_FR_ |
5016					    WUCSR_WUFR_ |
5017					    WUCSR_MPR_ |
5018					    WUCSR_BCST_FR_);
5019	if (ret < 0)
5020		goto out;
5021
5022	ret = 0;
5023out:
5024	mutex_unlock(&dev->dev_mutex);
5025
5026	return ret;
5027}
5028
5029static int lan78xx_reset_resume(struct usb_interface *intf)
5030{
5031	struct lan78xx_net *dev = usb_get_intfdata(intf);
5032	int ret;
5033
5034	netif_dbg(dev, ifup, dev->net, "(reset) resuming device");
5035
5036	ret = lan78xx_reset(dev);
5037	if (ret < 0)
5038		return ret;
5039
5040	phy_start(dev->net->phydev);
5041
5042	ret = lan78xx_resume(intf);
5043
5044	return ret;
5045}
5046
5047static const struct usb_device_id products[] = {
5048	{
5049	/* LAN7800 USB Gigabit Ethernet Device */
5050	USB_DEVICE(LAN78XX_USB_VENDOR_ID, LAN7800_USB_PRODUCT_ID),
5051	},
5052	{
5053	/* LAN7850 USB Gigabit Ethernet Device */
5054	USB_DEVICE(LAN78XX_USB_VENDOR_ID, LAN7850_USB_PRODUCT_ID),
5055	},
5056	{
5057	/* LAN7801 USB Gigabit Ethernet Device */
5058	USB_DEVICE(LAN78XX_USB_VENDOR_ID, LAN7801_USB_PRODUCT_ID),
5059	},
5060	{
5061	/* ATM2-AF USB Gigabit Ethernet Device */
5062	USB_DEVICE(AT29M2AF_USB_VENDOR_ID, AT29M2AF_USB_PRODUCT_ID),
5063	},
5064	{},
5065};
5066MODULE_DEVICE_TABLE(usb, products);
5067
5068static struct usb_driver lan78xx_driver = {
5069	.name			= DRIVER_NAME,
5070	.id_table		= products,
5071	.probe			= lan78xx_probe,
5072	.disconnect		= lan78xx_disconnect,
5073	.suspend		= lan78xx_suspend,
5074	.resume			= lan78xx_resume,
5075	.reset_resume		= lan78xx_reset_resume,
5076	.supports_autosuspend	= 1,
5077	.disable_hub_initiated_lpm = 1,
5078};
5079
5080module_usb_driver(lan78xx_driver);
5081
5082MODULE_AUTHOR(DRIVER_AUTHOR);
5083MODULE_DESCRIPTION(DRIVER_DESC);
5084MODULE_LICENSE("GPL");