Linux Audio

Check our new training course

Loading...
   1/*
   2 * drivers/net/ethernet/nxp/lpc_eth.c
   3 *
   4 * Author: Kevin Wells <kevin.wells@nxp.com>
   5 *
   6 * Copyright (C) 2010 NXP Semiconductors
   7 * Copyright (C) 2012 Roland Stigge <stigge@antcom.de>
   8 *
   9 * This program is free software; you can redistribute it and/or modify
  10 * it under the terms of the GNU General Public License as published by
  11 * the Free Software Foundation; either version 2 of the License, or
  12 * (at your option) any later version.
  13 *
  14 * This program is distributed in the hope that it will be useful,
  15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
  16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
  17 * GNU General Public License for more details.
  18 */
  19
  20#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  21
  22#include <linux/init.h>
  23#include <linux/module.h>
  24#include <linux/kernel.h>
  25#include <linux/sched.h>
  26#include <linux/slab.h>
  27#include <linux/delay.h>
  28#include <linux/interrupt.h>
  29#include <linux/errno.h>
  30#include <linux/ioport.h>
  31#include <linux/crc32.h>
  32#include <linux/platform_device.h>
  33#include <linux/spinlock.h>
  34#include <linux/ethtool.h>
  35#include <linux/mii.h>
  36#include <linux/clk.h>
  37#include <linux/workqueue.h>
  38#include <linux/netdevice.h>
  39#include <linux/etherdevice.h>
  40#include <linux/skbuff.h>
  41#include <linux/phy.h>
  42#include <linux/dma-mapping.h>
  43#include <linux/of.h>
  44#include <linux/of_net.h>
  45#include <linux/types.h>
  46
  47#include <linux/delay.h>
  48#include <linux/io.h>
  49#include <mach/board.h>
  50#include <mach/platform.h>
  51#include <mach/hardware.h>
  52
  53#define MODNAME "lpc-eth"
  54#define DRV_VERSION "1.00"
  55#define PHYDEF_ADDR 0x00
  56
  57#define ENET_MAXF_SIZE 1536
  58#define ENET_RX_DESC 48
  59#define ENET_TX_DESC 16
  60
  61#define NAPI_WEIGHT 16
  62
  63/*
  64 * Ethernet MAC controller Register offsets
  65 */
  66#define LPC_ENET_MAC1(x)			(x + 0x000)
  67#define LPC_ENET_MAC2(x)			(x + 0x004)
  68#define LPC_ENET_IPGT(x)			(x + 0x008)
  69#define LPC_ENET_IPGR(x)			(x + 0x00C)
  70#define LPC_ENET_CLRT(x)			(x + 0x010)
  71#define LPC_ENET_MAXF(x)			(x + 0x014)
  72#define LPC_ENET_SUPP(x)			(x + 0x018)
  73#define LPC_ENET_TEST(x)			(x + 0x01C)
  74#define LPC_ENET_MCFG(x)			(x + 0x020)
  75#define LPC_ENET_MCMD(x)			(x + 0x024)
  76#define LPC_ENET_MADR(x)			(x + 0x028)
  77#define LPC_ENET_MWTD(x)			(x + 0x02C)
  78#define LPC_ENET_MRDD(x)			(x + 0x030)
  79#define LPC_ENET_MIND(x)			(x + 0x034)
  80#define LPC_ENET_SA0(x)				(x + 0x040)
  81#define LPC_ENET_SA1(x)				(x + 0x044)
  82#define LPC_ENET_SA2(x)				(x + 0x048)
  83#define LPC_ENET_COMMAND(x)			(x + 0x100)
  84#define LPC_ENET_STATUS(x)			(x + 0x104)
  85#define LPC_ENET_RXDESCRIPTOR(x)		(x + 0x108)
  86#define LPC_ENET_RXSTATUS(x)			(x + 0x10C)
  87#define LPC_ENET_RXDESCRIPTORNUMBER(x)		(x + 0x110)
  88#define LPC_ENET_RXPRODUCEINDEX(x)		(x + 0x114)
  89#define LPC_ENET_RXCONSUMEINDEX(x)		(x + 0x118)
  90#define LPC_ENET_TXDESCRIPTOR(x)		(x + 0x11C)
  91#define LPC_ENET_TXSTATUS(x)			(x + 0x120)
  92#define LPC_ENET_TXDESCRIPTORNUMBER(x)		(x + 0x124)
  93#define LPC_ENET_TXPRODUCEINDEX(x)		(x + 0x128)
  94#define LPC_ENET_TXCONSUMEINDEX(x)		(x + 0x12C)
  95#define LPC_ENET_TSV0(x)			(x + 0x158)
  96#define LPC_ENET_TSV1(x)			(x + 0x15C)
  97#define LPC_ENET_RSV(x)				(x + 0x160)
  98#define LPC_ENET_FLOWCONTROLCOUNTER(x)		(x + 0x170)
  99#define LPC_ENET_FLOWCONTROLSTATUS(x)		(x + 0x174)
 100#define LPC_ENET_RXFILTER_CTRL(x)		(x + 0x200)
 101#define LPC_ENET_RXFILTERWOLSTATUS(x)		(x + 0x204)
 102#define LPC_ENET_RXFILTERWOLCLEAR(x)		(x + 0x208)
 103#define LPC_ENET_HASHFILTERL(x)			(x + 0x210)
 104#define LPC_ENET_HASHFILTERH(x)			(x + 0x214)
 105#define LPC_ENET_INTSTATUS(x)			(x + 0xFE0)
 106#define LPC_ENET_INTENABLE(x)			(x + 0xFE4)
 107#define LPC_ENET_INTCLEAR(x)			(x + 0xFE8)
 108#define LPC_ENET_INTSET(x)			(x + 0xFEC)
 109#define LPC_ENET_POWERDOWN(x)			(x + 0xFF4)
 110
 111/*
 112 * mac1 register definitions
 113 */
 114#define LPC_MAC1_RECV_ENABLE			(1 << 0)
 115#define LPC_MAC1_PASS_ALL_RX_FRAMES		(1 << 1)
 116#define LPC_MAC1_RX_FLOW_CONTROL		(1 << 2)
 117#define LPC_MAC1_TX_FLOW_CONTROL		(1 << 3)
 118#define LPC_MAC1_LOOPBACK			(1 << 4)
 119#define LPC_MAC1_RESET_TX			(1 << 8)
 120#define LPC_MAC1_RESET_MCS_TX			(1 << 9)
 121#define LPC_MAC1_RESET_RX			(1 << 10)
 122#define LPC_MAC1_RESET_MCS_RX			(1 << 11)
 123#define LPC_MAC1_SIMULATION_RESET		(1 << 14)
 124#define LPC_MAC1_SOFT_RESET			(1 << 15)
 125
 126/*
 127 * mac2 register definitions
 128 */
 129#define LPC_MAC2_FULL_DUPLEX			(1 << 0)
 130#define LPC_MAC2_FRAME_LENGTH_CHECKING		(1 << 1)
 131#define LPC_MAC2_HUGH_LENGTH_CHECKING		(1 << 2)
 132#define LPC_MAC2_DELAYED_CRC			(1 << 3)
 133#define LPC_MAC2_CRC_ENABLE			(1 << 4)
 134#define LPC_MAC2_PAD_CRC_ENABLE			(1 << 5)
 135#define LPC_MAC2_VLAN_PAD_ENABLE		(1 << 6)
 136#define LPC_MAC2_AUTO_DETECT_PAD_ENABLE		(1 << 7)
 137#define LPC_MAC2_PURE_PREAMBLE_ENFORCEMENT	(1 << 8)
 138#define LPC_MAC2_LONG_PREAMBLE_ENFORCEMENT	(1 << 9)
 139#define LPC_MAC2_NO_BACKOFF			(1 << 12)
 140#define LPC_MAC2_BACK_PRESSURE			(1 << 13)
 141#define LPC_MAC2_EXCESS_DEFER			(1 << 14)
 142
 143/*
 144 * ipgt register definitions
 145 */
 146#define LPC_IPGT_LOAD(n)			((n) & 0x7F)
 147
 148/*
 149 * ipgr register definitions
 150 */
 151#define LPC_IPGR_LOAD_PART2(n)			((n) & 0x7F)
 152#define LPC_IPGR_LOAD_PART1(n)			(((n) & 0x7F) << 8)
 153
 154/*
 155 * clrt register definitions
 156 */
 157#define LPC_CLRT_LOAD_RETRY_MAX(n)		((n) & 0xF)
 158#define LPC_CLRT_LOAD_COLLISION_WINDOW(n)	(((n) & 0x3F) << 8)
 159
 160/*
 161 * maxf register definitions
 162 */
 163#define LPC_MAXF_LOAD_MAX_FRAME_LEN(n)		((n) & 0xFFFF)
 164
 165/*
 166 * supp register definitions
 167 */
 168#define LPC_SUPP_SPEED				(1 << 8)
 169#define LPC_SUPP_RESET_RMII			(1 << 11)
 170
 171/*
 172 * test register definitions
 173 */
 174#define LPC_TEST_SHORTCUT_PAUSE_QUANTA		(1 << 0)
 175#define LPC_TEST_PAUSE				(1 << 1)
 176#define LPC_TEST_BACKPRESSURE			(1 << 2)
 177
 178/*
 179 * mcfg register definitions
 180 */
 181#define LPC_MCFG_SCAN_INCREMENT			(1 << 0)
 182#define LPC_MCFG_SUPPRESS_PREAMBLE		(1 << 1)
 183#define LPC_MCFG_CLOCK_SELECT(n)		(((n) & 0x7) << 2)
 184#define LPC_MCFG_CLOCK_HOST_DIV_4		0
 185#define LPC_MCFG_CLOCK_HOST_DIV_6		2
 186#define LPC_MCFG_CLOCK_HOST_DIV_8		3
 187#define LPC_MCFG_CLOCK_HOST_DIV_10		4
 188#define LPC_MCFG_CLOCK_HOST_DIV_14		5
 189#define LPC_MCFG_CLOCK_HOST_DIV_20		6
 190#define LPC_MCFG_CLOCK_HOST_DIV_28		7
 191#define LPC_MCFG_RESET_MII_MGMT			(1 << 15)
 192
 193/*
 194 * mcmd register definitions
 195 */
 196#define LPC_MCMD_READ				(1 << 0)
 197#define LPC_MCMD_SCAN				(1 << 1)
 198
 199/*
 200 * madr register definitions
 201 */
 202#define LPC_MADR_REGISTER_ADDRESS(n)		((n) & 0x1F)
 203#define LPC_MADR_PHY_0ADDRESS(n)		(((n) & 0x1F) << 8)
 204
 205/*
 206 * mwtd register definitions
 207 */
 208#define LPC_MWDT_WRITE(n)			((n) & 0xFFFF)
 209
 210/*
 211 * mrdd register definitions
 212 */
 213#define LPC_MRDD_READ_MASK			0xFFFF
 214
 215/*
 216 * mind register definitions
 217 */
 218#define LPC_MIND_BUSY				(1 << 0)
 219#define LPC_MIND_SCANNING			(1 << 1)
 220#define LPC_MIND_NOT_VALID			(1 << 2)
 221#define LPC_MIND_MII_LINK_FAIL			(1 << 3)
 222
 223/*
 224 * command register definitions
 225 */
 226#define LPC_COMMAND_RXENABLE			(1 << 0)
 227#define LPC_COMMAND_TXENABLE			(1 << 1)
 228#define LPC_COMMAND_REG_RESET			(1 << 3)
 229#define LPC_COMMAND_TXRESET			(1 << 4)
 230#define LPC_COMMAND_RXRESET			(1 << 5)
 231#define LPC_COMMAND_PASSRUNTFRAME		(1 << 6)
 232#define LPC_COMMAND_PASSRXFILTER		(1 << 7)
 233#define LPC_COMMAND_TXFLOWCONTROL		(1 << 8)
 234#define LPC_COMMAND_RMII			(1 << 9)
 235#define LPC_COMMAND_FULLDUPLEX			(1 << 10)
 236
 237/*
 238 * status register definitions
 239 */
 240#define LPC_STATUS_RXACTIVE			(1 << 0)
 241#define LPC_STATUS_TXACTIVE			(1 << 1)
 242
 243/*
 244 * tsv0 register definitions
 245 */
 246#define LPC_TSV0_CRC_ERROR			(1 << 0)
 247#define LPC_TSV0_LENGTH_CHECK_ERROR		(1 << 1)
 248#define LPC_TSV0_LENGTH_OUT_OF_RANGE		(1 << 2)
 249#define LPC_TSV0_DONE				(1 << 3)
 250#define LPC_TSV0_MULTICAST			(1 << 4)
 251#define LPC_TSV0_BROADCAST			(1 << 5)
 252#define LPC_TSV0_PACKET_DEFER			(1 << 6)
 253#define LPC_TSV0_ESCESSIVE_DEFER		(1 << 7)
 254#define LPC_TSV0_ESCESSIVE_COLLISION		(1 << 8)
 255#define LPC_TSV0_LATE_COLLISION			(1 << 9)
 256#define LPC_TSV0_GIANT				(1 << 10)
 257#define LPC_TSV0_UNDERRUN			(1 << 11)
 258#define LPC_TSV0_TOTAL_BYTES(n)			(((n) >> 12) & 0xFFFF)
 259#define LPC_TSV0_CONTROL_FRAME			(1 << 28)
 260#define LPC_TSV0_PAUSE				(1 << 29)
 261#define LPC_TSV0_BACKPRESSURE			(1 << 30)
 262#define LPC_TSV0_VLAN				(1 << 31)
 263
 264/*
 265 * tsv1 register definitions
 266 */
 267#define LPC_TSV1_TRANSMIT_BYTE_COUNT(n)		((n) & 0xFFFF)
 268#define LPC_TSV1_COLLISION_COUNT(n)		(((n) >> 16) & 0xF)
 269
 270/*
 271 * rsv register definitions
 272 */
 273#define LPC_RSV_RECEIVED_BYTE_COUNT(n)		((n) & 0xFFFF)
 274#define LPC_RSV_RXDV_EVENT_IGNORED		(1 << 16)
 275#define LPC_RSV_RXDV_EVENT_PREVIOUSLY_SEEN	(1 << 17)
 276#define LPC_RSV_CARRIER_EVNT_PREVIOUS_SEEN	(1 << 18)
 277#define LPC_RSV_RECEIVE_CODE_VIOLATION		(1 << 19)
 278#define LPC_RSV_CRC_ERROR			(1 << 20)
 279#define LPC_RSV_LENGTH_CHECK_ERROR		(1 << 21)
 280#define LPC_RSV_LENGTH_OUT_OF_RANGE		(1 << 22)
 281#define LPC_RSV_RECEIVE_OK			(1 << 23)
 282#define LPC_RSV_MULTICAST			(1 << 24)
 283#define LPC_RSV_BROADCAST			(1 << 25)
 284#define LPC_RSV_DRIBBLE_NIBBLE			(1 << 26)
 285#define LPC_RSV_CONTROL_FRAME			(1 << 27)
 286#define LPC_RSV_PAUSE				(1 << 28)
 287#define LPC_RSV_UNSUPPORTED_OPCODE		(1 << 29)
 288#define LPC_RSV_VLAN				(1 << 30)
 289
 290/*
 291 * flowcontrolcounter register definitions
 292 */
 293#define LPC_FCCR_MIRRORCOUNTER(n)		((n) & 0xFFFF)
 294#define LPC_FCCR_PAUSETIMER(n)			(((n) >> 16) & 0xFFFF)
 295
 296/*
 297 * flowcontrolstatus register definitions
 298 */
 299#define LPC_FCCR_MIRRORCOUNTERCURRENT(n)	((n) & 0xFFFF)
 300
 301/*
 302 * rxfliterctrl, rxfilterwolstatus, and rxfilterwolclear shared
 303 * register definitions
 304 */
 305#define LPC_RXFLTRW_ACCEPTUNICAST		(1 << 0)
 306#define LPC_RXFLTRW_ACCEPTUBROADCAST		(1 << 1)
 307#define LPC_RXFLTRW_ACCEPTUMULTICAST		(1 << 2)
 308#define LPC_RXFLTRW_ACCEPTUNICASTHASH		(1 << 3)
 309#define LPC_RXFLTRW_ACCEPTUMULTICASTHASH	(1 << 4)
 310#define LPC_RXFLTRW_ACCEPTPERFECT		(1 << 5)
 311
 312/*
 313 * rxfliterctrl register definitions
 314 */
 315#define LPC_RXFLTRWSTS_MAGICPACKETENWOL		(1 << 12)
 316#define LPC_RXFLTRWSTS_RXFILTERENWOL		(1 << 13)
 317
 318/*
 319 * rxfilterwolstatus/rxfilterwolclear register definitions
 320 */
 321#define LPC_RXFLTRWSTS_RXFILTERWOL		(1 << 7)
 322#define LPC_RXFLTRWSTS_MAGICPACKETWOL		(1 << 8)
 323
 324/*
 325 * intstatus, intenable, intclear, and Intset shared register
 326 * definitions
 327 */
 328#define LPC_MACINT_RXOVERRUNINTEN		(1 << 0)
 329#define LPC_MACINT_RXERRORONINT			(1 << 1)
 330#define LPC_MACINT_RXFINISHEDINTEN		(1 << 2)
 331#define LPC_MACINT_RXDONEINTEN			(1 << 3)
 332#define LPC_MACINT_TXUNDERRUNINTEN		(1 << 4)
 333#define LPC_MACINT_TXERRORINTEN			(1 << 5)
 334#define LPC_MACINT_TXFINISHEDINTEN		(1 << 6)
 335#define LPC_MACINT_TXDONEINTEN			(1 << 7)
 336#define LPC_MACINT_SOFTINTEN			(1 << 12)
 337#define LPC_MACINT_WAKEUPINTEN			(1 << 13)
 338
 339/*
 340 * powerdown register definitions
 341 */
 342#define LPC_POWERDOWN_MACAHB			(1 << 31)
 343
 344static phy_interface_t lpc_phy_interface_mode(struct device *dev)
 345{
 346	if (dev && dev->of_node) {
 347		const char *mode = of_get_property(dev->of_node,
 348						   "phy-mode", NULL);
 349		if (mode && !strcmp(mode, "mii"))
 350			return PHY_INTERFACE_MODE_MII;
 351		return PHY_INTERFACE_MODE_RMII;
 352	}
 353
 354	/* non-DT */
 355#ifdef CONFIG_ARCH_LPC32XX_MII_SUPPORT
 356	return PHY_INTERFACE_MODE_MII;
 357#else
 358	return PHY_INTERFACE_MODE_RMII;
 359#endif
 360}
 361
 362static bool use_iram_for_net(struct device *dev)
 363{
 364	if (dev && dev->of_node)
 365		return of_property_read_bool(dev->of_node, "use-iram");
 366
 367	/* non-DT */
 368#ifdef CONFIG_ARCH_LPC32XX_IRAM_FOR_NET
 369	return true;
 370#else
 371	return false;
 372#endif
 373}
 374
 375/* Receive Status information word */
 376#define RXSTATUS_SIZE			0x000007FF
 377#define RXSTATUS_CONTROL		(1 << 18)
 378#define RXSTATUS_VLAN			(1 << 19)
 379#define RXSTATUS_FILTER			(1 << 20)
 380#define RXSTATUS_MULTICAST		(1 << 21)
 381#define RXSTATUS_BROADCAST		(1 << 22)
 382#define RXSTATUS_CRC			(1 << 23)
 383#define RXSTATUS_SYMBOL			(1 << 24)
 384#define RXSTATUS_LENGTH			(1 << 25)
 385#define RXSTATUS_RANGE			(1 << 26)
 386#define RXSTATUS_ALIGN			(1 << 27)
 387#define RXSTATUS_OVERRUN		(1 << 28)
 388#define RXSTATUS_NODESC			(1 << 29)
 389#define RXSTATUS_LAST			(1 << 30)
 390#define RXSTATUS_ERROR			(1 << 31)
 391
 392#define RXSTATUS_STATUS_ERROR \
 393	(RXSTATUS_NODESC | RXSTATUS_OVERRUN | RXSTATUS_ALIGN | \
 394	 RXSTATUS_RANGE | RXSTATUS_LENGTH | RXSTATUS_SYMBOL | RXSTATUS_CRC)
 395
 396/* Receive Descriptor control word */
 397#define RXDESC_CONTROL_SIZE		0x000007FF
 398#define RXDESC_CONTROL_INT		(1 << 31)
 399
 400/* Transmit Status information word */
 401#define TXSTATUS_COLLISIONS_GET(x)	(((x) >> 21) & 0xF)
 402#define TXSTATUS_DEFER			(1 << 25)
 403#define TXSTATUS_EXCESSDEFER		(1 << 26)
 404#define TXSTATUS_EXCESSCOLL		(1 << 27)
 405#define TXSTATUS_LATECOLL		(1 << 28)
 406#define TXSTATUS_UNDERRUN		(1 << 29)
 407#define TXSTATUS_NODESC			(1 << 30)
 408#define TXSTATUS_ERROR			(1 << 31)
 409
 410/* Transmit Descriptor control word */
 411#define TXDESC_CONTROL_SIZE		0x000007FF
 412#define TXDESC_CONTROL_OVERRIDE		(1 << 26)
 413#define TXDESC_CONTROL_HUGE		(1 << 27)
 414#define TXDESC_CONTROL_PAD		(1 << 28)
 415#define TXDESC_CONTROL_CRC		(1 << 29)
 416#define TXDESC_CONTROL_LAST		(1 << 30)
 417#define TXDESC_CONTROL_INT		(1 << 31)
 418
 419static int lpc_eth_hard_start_xmit(struct sk_buff *skb,
 420				   struct net_device *ndev);
 421
 422/*
 423 * Structure of a TX/RX descriptors and RX status
 424 */
 425struct txrx_desc_t {
 426	__le32 packet;
 427	__le32 control;
 428};
 429struct rx_status_t {
 430	__le32 statusinfo;
 431	__le32 statushashcrc;
 432};
 433
 434/*
 435 * Device driver data structure
 436 */
 437struct netdata_local {
 438	struct platform_device	*pdev;
 439	struct net_device	*ndev;
 440	spinlock_t		lock;
 441	void __iomem		*net_base;
 442	u32			msg_enable;
 443	struct sk_buff		*skb[ENET_TX_DESC];
 444	unsigned int		last_tx_idx;
 445	unsigned int		num_used_tx_buffs;
 446	struct mii_bus		*mii_bus;
 447	struct phy_device	*phy_dev;
 448	struct clk		*clk;
 449	dma_addr_t		dma_buff_base_p;
 450	void			*dma_buff_base_v;
 451	size_t			dma_buff_size;
 452	struct txrx_desc_t	*tx_desc_v;
 453	u32			*tx_stat_v;
 454	void			*tx_buff_v;
 455	struct txrx_desc_t	*rx_desc_v;
 456	struct rx_status_t	*rx_stat_v;
 457	void			*rx_buff_v;
 458	int			link;
 459	int			speed;
 460	int			duplex;
 461	struct napi_struct	napi;
 462};
 463
 464/*
 465 * MAC support functions
 466 */
 467static void __lpc_set_mac(struct netdata_local *pldat, u8 *mac)
 468{
 469	u32 tmp;
 470
 471	/* Set station address */
 472	tmp = mac[0] | ((u32)mac[1] << 8);
 473	writel(tmp, LPC_ENET_SA2(pldat->net_base));
 474	tmp = mac[2] | ((u32)mac[3] << 8);
 475	writel(tmp, LPC_ENET_SA1(pldat->net_base));
 476	tmp = mac[4] | ((u32)mac[5] << 8);
 477	writel(tmp, LPC_ENET_SA0(pldat->net_base));
 478
 479	netdev_dbg(pldat->ndev, "Ethernet MAC address %pM\n", mac);
 480}
 481
 482static void __lpc_get_mac(struct netdata_local *pldat, u8 *mac)
 483{
 484	u32 tmp;
 485
 486	/* Get station address */
 487	tmp = readl(LPC_ENET_SA2(pldat->net_base));
 488	mac[0] = tmp & 0xFF;
 489	mac[1] = tmp >> 8;
 490	tmp = readl(LPC_ENET_SA1(pldat->net_base));
 491	mac[2] = tmp & 0xFF;
 492	mac[3] = tmp >> 8;
 493	tmp = readl(LPC_ENET_SA0(pldat->net_base));
 494	mac[4] = tmp & 0xFF;
 495	mac[5] = tmp >> 8;
 496}
 497
 498static void __lpc_eth_clock_enable(struct netdata_local *pldat,
 499				   bool enable)
 500{
 501	if (enable)
 502		clk_enable(pldat->clk);
 503	else
 504		clk_disable(pldat->clk);
 505}
 506
 507static void __lpc_params_setup(struct netdata_local *pldat)
 508{
 509	u32 tmp;
 510
 511	if (pldat->duplex == DUPLEX_FULL) {
 512		tmp = readl(LPC_ENET_MAC2(pldat->net_base));
 513		tmp |= LPC_MAC2_FULL_DUPLEX;
 514		writel(tmp, LPC_ENET_MAC2(pldat->net_base));
 515		tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
 516		tmp |= LPC_COMMAND_FULLDUPLEX;
 517		writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
 518		writel(LPC_IPGT_LOAD(0x15), LPC_ENET_IPGT(pldat->net_base));
 519	} else {
 520		tmp = readl(LPC_ENET_MAC2(pldat->net_base));
 521		tmp &= ~LPC_MAC2_FULL_DUPLEX;
 522		writel(tmp, LPC_ENET_MAC2(pldat->net_base));
 523		tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
 524		tmp &= ~LPC_COMMAND_FULLDUPLEX;
 525		writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
 526		writel(LPC_IPGT_LOAD(0x12), LPC_ENET_IPGT(pldat->net_base));
 527	}
 528
 529	if (pldat->speed == SPEED_100)
 530		writel(LPC_SUPP_SPEED, LPC_ENET_SUPP(pldat->net_base));
 531	else
 532		writel(0, LPC_ENET_SUPP(pldat->net_base));
 533}
 534
 535static void __lpc_eth_reset(struct netdata_local *pldat)
 536{
 537	/* Reset all MAC logic */
 538	writel((LPC_MAC1_RESET_TX | LPC_MAC1_RESET_MCS_TX | LPC_MAC1_RESET_RX |
 539		LPC_MAC1_RESET_MCS_RX | LPC_MAC1_SIMULATION_RESET |
 540		LPC_MAC1_SOFT_RESET), LPC_ENET_MAC1(pldat->net_base));
 541	writel((LPC_COMMAND_REG_RESET | LPC_COMMAND_TXRESET |
 542		LPC_COMMAND_RXRESET), LPC_ENET_COMMAND(pldat->net_base));
 543}
 544
 545static int __lpc_mii_mngt_reset(struct netdata_local *pldat)
 546{
 547	/* Reset MII management hardware */
 548	writel(LPC_MCFG_RESET_MII_MGMT, LPC_ENET_MCFG(pldat->net_base));
 549
 550	/* Setup MII clock to slowest rate with a /28 divider */
 551	writel(LPC_MCFG_CLOCK_SELECT(LPC_MCFG_CLOCK_HOST_DIV_28),
 552	       LPC_ENET_MCFG(pldat->net_base));
 553
 554	return 0;
 555}
 556
 557static inline phys_addr_t __va_to_pa(void *addr, struct netdata_local *pldat)
 558{
 559	phys_addr_t phaddr;
 560
 561	phaddr = addr - pldat->dma_buff_base_v;
 562	phaddr += pldat->dma_buff_base_p;
 563
 564	return phaddr;
 565}
 566
 567static void lpc_eth_enable_int(void __iomem *regbase)
 568{
 569	writel((LPC_MACINT_RXDONEINTEN | LPC_MACINT_TXDONEINTEN),
 570	       LPC_ENET_INTENABLE(regbase));
 571}
 572
 573static void lpc_eth_disable_int(void __iomem *regbase)
 574{
 575	writel(0, LPC_ENET_INTENABLE(regbase));
 576}
 577
 578/* Setup TX/RX descriptors */
 579static void __lpc_txrx_desc_setup(struct netdata_local *pldat)
 580{
 581	u32 *ptxstat;
 582	void *tbuff;
 583	int i;
 584	struct txrx_desc_t *ptxrxdesc;
 585	struct rx_status_t *prxstat;
 586
 587	tbuff = PTR_ALIGN(pldat->dma_buff_base_v, 16);
 588
 589	/* Setup TX descriptors, status, and buffers */
 590	pldat->tx_desc_v = tbuff;
 591	tbuff += sizeof(struct txrx_desc_t) * ENET_TX_DESC;
 592
 593	pldat->tx_stat_v = tbuff;
 594	tbuff += sizeof(u32) * ENET_TX_DESC;
 595
 596	tbuff = PTR_ALIGN(tbuff, 16);
 597	pldat->tx_buff_v = tbuff;
 598	tbuff += ENET_MAXF_SIZE * ENET_TX_DESC;
 599
 600	/* Setup RX descriptors, status, and buffers */
 601	pldat->rx_desc_v = tbuff;
 602	tbuff += sizeof(struct txrx_desc_t) * ENET_RX_DESC;
 603
 604	tbuff = PTR_ALIGN(tbuff, 16);
 605	pldat->rx_stat_v = tbuff;
 606	tbuff += sizeof(struct rx_status_t) * ENET_RX_DESC;
 607
 608	tbuff = PTR_ALIGN(tbuff, 16);
 609	pldat->rx_buff_v = tbuff;
 610	tbuff += ENET_MAXF_SIZE * ENET_RX_DESC;
 611
 612	/* Map the TX descriptors to the TX buffers in hardware */
 613	for (i = 0; i < ENET_TX_DESC; i++) {
 614		ptxstat = &pldat->tx_stat_v[i];
 615		ptxrxdesc = &pldat->tx_desc_v[i];
 616
 617		ptxrxdesc->packet = __va_to_pa(
 618				pldat->tx_buff_v + i * ENET_MAXF_SIZE, pldat);
 619		ptxrxdesc->control = 0;
 620		*ptxstat = 0;
 621	}
 622
 623	/* Map the RX descriptors to the RX buffers in hardware */
 624	for (i = 0; i < ENET_RX_DESC; i++) {
 625		prxstat = &pldat->rx_stat_v[i];
 626		ptxrxdesc = &pldat->rx_desc_v[i];
 627
 628		ptxrxdesc->packet = __va_to_pa(
 629				pldat->rx_buff_v + i * ENET_MAXF_SIZE, pldat);
 630		ptxrxdesc->control = RXDESC_CONTROL_INT | (ENET_MAXF_SIZE - 1);
 631		prxstat->statusinfo = 0;
 632		prxstat->statushashcrc = 0;
 633	}
 634
 635	/* Setup base addresses in hardware to point to buffers and
 636	 * descriptors
 637	 */
 638	writel((ENET_TX_DESC - 1),
 639	       LPC_ENET_TXDESCRIPTORNUMBER(pldat->net_base));
 640	writel(__va_to_pa(pldat->tx_desc_v, pldat),
 641	       LPC_ENET_TXDESCRIPTOR(pldat->net_base));
 642	writel(__va_to_pa(pldat->tx_stat_v, pldat),
 643	       LPC_ENET_TXSTATUS(pldat->net_base));
 644	writel((ENET_RX_DESC - 1),
 645	       LPC_ENET_RXDESCRIPTORNUMBER(pldat->net_base));
 646	writel(__va_to_pa(pldat->rx_desc_v, pldat),
 647	       LPC_ENET_RXDESCRIPTOR(pldat->net_base));
 648	writel(__va_to_pa(pldat->rx_stat_v, pldat),
 649	       LPC_ENET_RXSTATUS(pldat->net_base));
 650}
 651
 652static void __lpc_eth_init(struct netdata_local *pldat)
 653{
 654	u32 tmp;
 655
 656	/* Disable controller and reset */
 657	tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
 658	tmp &= ~LPC_COMMAND_RXENABLE | LPC_COMMAND_TXENABLE;
 659	writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
 660	tmp = readl(LPC_ENET_MAC1(pldat->net_base));
 661	tmp &= ~LPC_MAC1_RECV_ENABLE;
 662	writel(tmp, LPC_ENET_MAC1(pldat->net_base));
 663
 664	/* Initial MAC setup */
 665	writel(LPC_MAC1_PASS_ALL_RX_FRAMES, LPC_ENET_MAC1(pldat->net_base));
 666	writel((LPC_MAC2_PAD_CRC_ENABLE | LPC_MAC2_CRC_ENABLE),
 667	       LPC_ENET_MAC2(pldat->net_base));
 668	writel(ENET_MAXF_SIZE, LPC_ENET_MAXF(pldat->net_base));
 669
 670	/* Collision window, gap */
 671	writel((LPC_CLRT_LOAD_RETRY_MAX(0xF) |
 672		LPC_CLRT_LOAD_COLLISION_WINDOW(0x37)),
 673	       LPC_ENET_CLRT(pldat->net_base));
 674	writel(LPC_IPGR_LOAD_PART2(0x12), LPC_ENET_IPGR(pldat->net_base));
 675
 676	if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII)
 677		writel(LPC_COMMAND_PASSRUNTFRAME,
 678		       LPC_ENET_COMMAND(pldat->net_base));
 679	else {
 680		writel((LPC_COMMAND_PASSRUNTFRAME | LPC_COMMAND_RMII),
 681		       LPC_ENET_COMMAND(pldat->net_base));
 682		writel(LPC_SUPP_RESET_RMII, LPC_ENET_SUPP(pldat->net_base));
 683	}
 684
 685	__lpc_params_setup(pldat);
 686
 687	/* Setup TX and RX descriptors */
 688	__lpc_txrx_desc_setup(pldat);
 689
 690	/* Setup packet filtering */
 691	writel((LPC_RXFLTRW_ACCEPTUBROADCAST | LPC_RXFLTRW_ACCEPTPERFECT),
 692	       LPC_ENET_RXFILTER_CTRL(pldat->net_base));
 693
 694	/* Get the next TX buffer output index */
 695	pldat->num_used_tx_buffs = 0;
 696	pldat->last_tx_idx =
 697		readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base));
 698
 699	/* Clear and enable interrupts */
 700	writel(0xFFFF, LPC_ENET_INTCLEAR(pldat->net_base));
 701	smp_wmb();
 702	lpc_eth_enable_int(pldat->net_base);
 703
 704	/* Enable controller */
 705	tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
 706	tmp |= LPC_COMMAND_RXENABLE | LPC_COMMAND_TXENABLE;
 707	writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
 708	tmp = readl(LPC_ENET_MAC1(pldat->net_base));
 709	tmp |= LPC_MAC1_RECV_ENABLE;
 710	writel(tmp, LPC_ENET_MAC1(pldat->net_base));
 711}
 712
 713static void __lpc_eth_shutdown(struct netdata_local *pldat)
 714{
 715	/* Reset ethernet and power down PHY */
 716	__lpc_eth_reset(pldat);
 717	writel(0, LPC_ENET_MAC1(pldat->net_base));
 718	writel(0, LPC_ENET_MAC2(pldat->net_base));
 719}
 720
 721/*
 722 * MAC<--->PHY support functions
 723 */
 724static int lpc_mdio_read(struct mii_bus *bus, int phy_id, int phyreg)
 725{
 726	struct netdata_local *pldat = bus->priv;
 727	unsigned long timeout = jiffies + msecs_to_jiffies(100);
 728	int lps;
 729
 730	writel(((phy_id << 8) | phyreg), LPC_ENET_MADR(pldat->net_base));
 731	writel(LPC_MCMD_READ, LPC_ENET_MCMD(pldat->net_base));
 732
 733	/* Wait for unbusy status */
 734	while (readl(LPC_ENET_MIND(pldat->net_base)) & LPC_MIND_BUSY) {
 735		if (time_after(jiffies, timeout))
 736			return -EIO;
 737		cpu_relax();
 738	}
 739
 740	lps = readl(LPC_ENET_MRDD(pldat->net_base));
 741	writel(0, LPC_ENET_MCMD(pldat->net_base));
 742
 743	return lps;
 744}
 745
 746static int lpc_mdio_write(struct mii_bus *bus, int phy_id, int phyreg,
 747			u16 phydata)
 748{
 749	struct netdata_local *pldat = bus->priv;
 750	unsigned long timeout = jiffies + msecs_to_jiffies(100);
 751
 752	writel(((phy_id << 8) | phyreg), LPC_ENET_MADR(pldat->net_base));
 753	writel(phydata, LPC_ENET_MWTD(pldat->net_base));
 754
 755	/* Wait for completion */
 756	while (readl(LPC_ENET_MIND(pldat->net_base)) & LPC_MIND_BUSY) {
 757		if (time_after(jiffies, timeout))
 758			return -EIO;
 759		cpu_relax();
 760	}
 761
 762	return 0;
 763}
 764
 765static int lpc_mdio_reset(struct mii_bus *bus)
 766{
 767	return __lpc_mii_mngt_reset((struct netdata_local *)bus->priv);
 768}
 769
 770static void lpc_handle_link_change(struct net_device *ndev)
 771{
 772	struct netdata_local *pldat = netdev_priv(ndev);
 773	struct phy_device *phydev = pldat->phy_dev;
 774	unsigned long flags;
 775
 776	bool status_change = false;
 777
 778	spin_lock_irqsave(&pldat->lock, flags);
 779
 780	if (phydev->link) {
 781		if ((pldat->speed != phydev->speed) ||
 782		    (pldat->duplex != phydev->duplex)) {
 783			pldat->speed = phydev->speed;
 784			pldat->duplex = phydev->duplex;
 785			status_change = true;
 786		}
 787	}
 788
 789	if (phydev->link != pldat->link) {
 790		if (!phydev->link) {
 791			pldat->speed = 0;
 792			pldat->duplex = -1;
 793		}
 794		pldat->link = phydev->link;
 795
 796		status_change = true;
 797	}
 798
 799	spin_unlock_irqrestore(&pldat->lock, flags);
 800
 801	if (status_change)
 802		__lpc_params_setup(pldat);
 803}
 804
 805static int lpc_mii_probe(struct net_device *ndev)
 806{
 807	struct netdata_local *pldat = netdev_priv(ndev);
 808	struct phy_device *phydev = phy_find_first(pldat->mii_bus);
 809
 810	if (!phydev) {
 811		netdev_err(ndev, "no PHY found\n");
 812		return -ENODEV;
 813	}
 814
 815	/* Attach to the PHY */
 816	if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII)
 817		netdev_info(ndev, "using MII interface\n");
 818	else
 819		netdev_info(ndev, "using RMII interface\n");
 820	phydev = phy_connect(ndev, dev_name(&phydev->dev),
 821			     &lpc_handle_link_change, 0,
 822			     lpc_phy_interface_mode(&pldat->pdev->dev));
 823
 824	if (IS_ERR(phydev)) {
 825		netdev_err(ndev, "Could not attach to PHY\n");
 826		return PTR_ERR(phydev);
 827	}
 828
 829	/* mask with MAC supported features */
 830	phydev->supported &= PHY_BASIC_FEATURES;
 831
 832	phydev->advertising = phydev->supported;
 833
 834	pldat->link = 0;
 835	pldat->speed = 0;
 836	pldat->duplex = -1;
 837	pldat->phy_dev = phydev;
 838
 839	netdev_info(ndev,
 840		"attached PHY driver [%s] (mii_bus:phy_addr=%s, irq=%d)\n",
 841		phydev->drv->name, dev_name(&phydev->dev), phydev->irq);
 842	return 0;
 843}
 844
 845static int lpc_mii_init(struct netdata_local *pldat)
 846{
 847	int err = -ENXIO, i;
 848
 849	pldat->mii_bus = mdiobus_alloc();
 850	if (!pldat->mii_bus) {
 851		err = -ENOMEM;
 852		goto err_out;
 853	}
 854
 855	/* Setup MII mode */
 856	if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII)
 857		writel(LPC_COMMAND_PASSRUNTFRAME,
 858		       LPC_ENET_COMMAND(pldat->net_base));
 859	else {
 860		writel((LPC_COMMAND_PASSRUNTFRAME | LPC_COMMAND_RMII),
 861		       LPC_ENET_COMMAND(pldat->net_base));
 862		writel(LPC_SUPP_RESET_RMII, LPC_ENET_SUPP(pldat->net_base));
 863	}
 864
 865	pldat->mii_bus->name = "lpc_mii_bus";
 866	pldat->mii_bus->read = &lpc_mdio_read;
 867	pldat->mii_bus->write = &lpc_mdio_write;
 868	pldat->mii_bus->reset = &lpc_mdio_reset;
 869	snprintf(pldat->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
 870		 pldat->pdev->name, pldat->pdev->id);
 871	pldat->mii_bus->priv = pldat;
 872	pldat->mii_bus->parent = &pldat->pdev->dev;
 873
 874	pldat->mii_bus->irq = kmalloc(sizeof(int) * PHY_MAX_ADDR, GFP_KERNEL);
 875	if (!pldat->mii_bus->irq) {
 876		err = -ENOMEM;
 877		goto err_out_1;
 878	}
 879
 880	for (i = 0; i < PHY_MAX_ADDR; i++)
 881		pldat->mii_bus->irq[i] = PHY_POLL;
 882
 883	platform_set_drvdata(pldat->pdev, pldat->mii_bus);
 884
 885	if (mdiobus_register(pldat->mii_bus))
 886		goto err_out_free_mdio_irq;
 887
 888	if (lpc_mii_probe(pldat->ndev) != 0)
 889		goto err_out_unregister_bus;
 890
 891	return 0;
 892
 893err_out_unregister_bus:
 894	mdiobus_unregister(pldat->mii_bus);
 895err_out_free_mdio_irq:
 896	kfree(pldat->mii_bus->irq);
 897err_out_1:
 898	mdiobus_free(pldat->mii_bus);
 899err_out:
 900	return err;
 901}
 902
 903static void __lpc_handle_xmit(struct net_device *ndev)
 904{
 905	struct netdata_local *pldat = netdev_priv(ndev);
 906	struct sk_buff *skb;
 907	u32 txcidx, *ptxstat, txstat;
 908
 909	txcidx = readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base));
 910	while (pldat->last_tx_idx != txcidx) {
 911		skb = pldat->skb[pldat->last_tx_idx];
 912
 913		/* A buffer is available, get buffer status */
 914		ptxstat = &pldat->tx_stat_v[pldat->last_tx_idx];
 915		txstat = *ptxstat;
 916
 917		/* Next buffer and decrement used buffer counter */
 918		pldat->num_used_tx_buffs--;
 919		pldat->last_tx_idx++;
 920		if (pldat->last_tx_idx >= ENET_TX_DESC)
 921			pldat->last_tx_idx = 0;
 922
 923		/* Update collision counter */
 924		ndev->stats.collisions += TXSTATUS_COLLISIONS_GET(txstat);
 925
 926		/* Any errors occurred? */
 927		if (txstat & TXSTATUS_ERROR) {
 928			if (txstat & TXSTATUS_UNDERRUN) {
 929				/* FIFO underrun */
 930				ndev->stats.tx_fifo_errors++;
 931			}
 932			if (txstat & TXSTATUS_LATECOLL) {
 933				/* Late collision */
 934				ndev->stats.tx_aborted_errors++;
 935			}
 936			if (txstat & TXSTATUS_EXCESSCOLL) {
 937				/* Excessive collision */
 938				ndev->stats.tx_aborted_errors++;
 939			}
 940			if (txstat & TXSTATUS_EXCESSDEFER) {
 941				/* Defer limit */
 942				ndev->stats.tx_aborted_errors++;
 943			}
 944			ndev->stats.tx_errors++;
 945		} else {
 946			/* Update stats */
 947			ndev->stats.tx_packets++;
 948			ndev->stats.tx_bytes += skb->len;
 949		}
 950		dev_kfree_skb_irq(skb);
 951
 952		txcidx = readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base));
 953	}
 954
 955	if (pldat->num_used_tx_buffs <= ENET_TX_DESC/2) {
 956		if (netif_queue_stopped(ndev))
 957			netif_wake_queue(ndev);
 958	}
 959}
 960
 961static int __lpc_handle_recv(struct net_device *ndev, int budget)
 962{
 963	struct netdata_local *pldat = netdev_priv(ndev);
 964	struct sk_buff *skb;
 965	u32 rxconsidx, len, ethst;
 966	struct rx_status_t *prxstat;
 967	u8 *prdbuf;
 968	int rx_done = 0;
 969
 970	/* Get the current RX buffer indexes */
 971	rxconsidx = readl(LPC_ENET_RXCONSUMEINDEX(pldat->net_base));
 972	while (rx_done < budget && rxconsidx !=
 973			readl(LPC_ENET_RXPRODUCEINDEX(pldat->net_base))) {
 974		/* Get pointer to receive status */
 975		prxstat = &pldat->rx_stat_v[rxconsidx];
 976		len = (prxstat->statusinfo & RXSTATUS_SIZE) + 1;
 977
 978		/* Status error? */
 979		ethst = prxstat->statusinfo;
 980		if ((ethst & (RXSTATUS_ERROR | RXSTATUS_STATUS_ERROR)) ==
 981		    (RXSTATUS_ERROR | RXSTATUS_RANGE))
 982			ethst &= ~RXSTATUS_ERROR;
 983
 984		if (ethst & RXSTATUS_ERROR) {
 985			int si = prxstat->statusinfo;
 986			/* Check statuses */
 987			if (si & RXSTATUS_OVERRUN) {
 988				/* Overrun error */
 989				ndev->stats.rx_fifo_errors++;
 990			} else if (si & RXSTATUS_CRC) {
 991				/* CRC error */
 992				ndev->stats.rx_crc_errors++;
 993			} else if (si & RXSTATUS_LENGTH) {
 994				/* Length error */
 995				ndev->stats.rx_length_errors++;
 996			} else if (si & RXSTATUS_ERROR) {
 997				/* Other error */
 998				ndev->stats.rx_length_errors++;
 999			}
1000			ndev->stats.rx_errors++;
1001		} else {
1002			/* Packet is good */
1003			skb = dev_alloc_skb(len);
1004			if (!skb) {
1005				ndev->stats.rx_dropped++;
1006			} else {
1007				prdbuf = skb_put(skb, len);
1008
1009				/* Copy packet from buffer */
1010				memcpy(prdbuf, pldat->rx_buff_v +
1011					rxconsidx * ENET_MAXF_SIZE, len);
1012
1013				/* Pass to upper layer */
1014				skb->protocol = eth_type_trans(skb, ndev);
1015				netif_receive_skb(skb);
1016				ndev->stats.rx_packets++;
1017				ndev->stats.rx_bytes += len;
1018			}
1019		}
1020
1021		/* Increment consume index */
1022		rxconsidx = rxconsidx + 1;
1023		if (rxconsidx >= ENET_RX_DESC)
1024			rxconsidx = 0;
1025		writel(rxconsidx,
1026		       LPC_ENET_RXCONSUMEINDEX(pldat->net_base));
1027		rx_done++;
1028	}
1029
1030	return rx_done;
1031}
1032
1033static int lpc_eth_poll(struct napi_struct *napi, int budget)
1034{
1035	struct netdata_local *pldat = container_of(napi,
1036			struct netdata_local, napi);
1037	struct net_device *ndev = pldat->ndev;
1038	int rx_done = 0;
1039	struct netdev_queue *txq = netdev_get_tx_queue(ndev, 0);
1040
1041	__netif_tx_lock(txq, smp_processor_id());
1042	__lpc_handle_xmit(ndev);
1043	__netif_tx_unlock(txq);
1044	rx_done = __lpc_handle_recv(ndev, budget);
1045
1046	if (rx_done < budget) {
1047		napi_complete(napi);
1048		lpc_eth_enable_int(pldat->net_base);
1049	}
1050
1051	return rx_done;
1052}
1053
1054static irqreturn_t __lpc_eth_interrupt(int irq, void *dev_id)
1055{
1056	struct net_device *ndev = dev_id;
1057	struct netdata_local *pldat = netdev_priv(ndev);
1058	u32 tmp;
1059
1060	spin_lock(&pldat->lock);
1061
1062	tmp = readl(LPC_ENET_INTSTATUS(pldat->net_base));
1063	/* Clear interrupts */
1064	writel(tmp, LPC_ENET_INTCLEAR(pldat->net_base));
1065
1066	lpc_eth_disable_int(pldat->net_base);
1067	if (likely(napi_schedule_prep(&pldat->napi)))
1068		__napi_schedule(&pldat->napi);
1069
1070	spin_unlock(&pldat->lock);
1071
1072	return IRQ_HANDLED;
1073}
1074
1075static int lpc_eth_close(struct net_device *ndev)
1076{
1077	unsigned long flags;
1078	struct netdata_local *pldat = netdev_priv(ndev);
1079
1080	if (netif_msg_ifdown(pldat))
1081		dev_dbg(&pldat->pdev->dev, "shutting down %s\n", ndev->name);
1082
1083	napi_disable(&pldat->napi);
1084	netif_stop_queue(ndev);
1085
1086	if (pldat->phy_dev)
1087		phy_stop(pldat->phy_dev);
1088
1089	spin_lock_irqsave(&pldat->lock, flags);
1090	__lpc_eth_reset(pldat);
1091	netif_carrier_off(ndev);
1092	writel(0, LPC_ENET_MAC1(pldat->net_base));
1093	writel(0, LPC_ENET_MAC2(pldat->net_base));
1094	spin_unlock_irqrestore(&pldat->lock, flags);
1095
1096	__lpc_eth_clock_enable(pldat, false);
1097
1098	return 0;
1099}
1100
1101static int lpc_eth_hard_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1102{
1103	struct netdata_local *pldat = netdev_priv(ndev);
1104	u32 len, txidx;
1105	u32 *ptxstat;
1106	struct txrx_desc_t *ptxrxdesc;
1107
1108	len = skb->len;
1109
1110	spin_lock_irq(&pldat->lock);
1111
1112	if (pldat->num_used_tx_buffs >= (ENET_TX_DESC - 1)) {
1113		/* This function should never be called when there are no
1114		   buffers */
1115		netif_stop_queue(ndev);
1116		spin_unlock_irq(&pldat->lock);
1117		WARN(1, "BUG! TX request when no free TX buffers!\n");
1118		return NETDEV_TX_BUSY;
1119	}
1120
1121	/* Get the next TX descriptor index */
1122	txidx = readl(LPC_ENET_TXPRODUCEINDEX(pldat->net_base));
1123
1124	/* Setup control for the transfer */
1125	ptxstat = &pldat->tx_stat_v[txidx];
1126	*ptxstat = 0;
1127	ptxrxdesc = &pldat->tx_desc_v[txidx];
1128	ptxrxdesc->control =
1129		(len - 1) | TXDESC_CONTROL_LAST | TXDESC_CONTROL_INT;
1130
1131	/* Copy data to the DMA buffer */
1132	memcpy(pldat->tx_buff_v + txidx * ENET_MAXF_SIZE, skb->data, len);
1133
1134	/* Save the buffer and increment the buffer counter */
1135	pldat->skb[txidx] = skb;
1136	pldat->num_used_tx_buffs++;
1137
1138	/* Start transmit */
1139	txidx++;
1140	if (txidx >= ENET_TX_DESC)
1141		txidx = 0;
1142	writel(txidx, LPC_ENET_TXPRODUCEINDEX(pldat->net_base));
1143
1144	/* Stop queue if no more TX buffers */
1145	if (pldat->num_used_tx_buffs >= (ENET_TX_DESC - 1))
1146		netif_stop_queue(ndev);
1147
1148	spin_unlock_irq(&pldat->lock);
1149
1150	return NETDEV_TX_OK;
1151}
1152
1153static int lpc_set_mac_address(struct net_device *ndev, void *p)
1154{
1155	struct sockaddr *addr = p;
1156	struct netdata_local *pldat = netdev_priv(ndev);
1157	unsigned long flags;
1158
1159	if (!is_valid_ether_addr(addr->sa_data))
1160		return -EADDRNOTAVAIL;
1161	memcpy(ndev->dev_addr, addr->sa_data, ETH_ALEN);
1162
1163	spin_lock_irqsave(&pldat->lock, flags);
1164
1165	/* Set station address */
1166	__lpc_set_mac(pldat, ndev->dev_addr);
1167
1168	spin_unlock_irqrestore(&pldat->lock, flags);
1169
1170	return 0;
1171}
1172
1173static void lpc_eth_set_multicast_list(struct net_device *ndev)
1174{
1175	struct netdata_local *pldat = netdev_priv(ndev);
1176	struct netdev_hw_addr_list *mcptr = &ndev->mc;
1177	struct netdev_hw_addr *ha;
1178	u32 tmp32, hash_val, hashlo, hashhi;
1179	unsigned long flags;
1180
1181	spin_lock_irqsave(&pldat->lock, flags);
1182
1183	/* Set station address */
1184	__lpc_set_mac(pldat, ndev->dev_addr);
1185
1186	tmp32 =  LPC_RXFLTRW_ACCEPTUBROADCAST | LPC_RXFLTRW_ACCEPTPERFECT;
1187
1188	if (ndev->flags & IFF_PROMISC)
1189		tmp32 |= LPC_RXFLTRW_ACCEPTUNICAST |
1190			LPC_RXFLTRW_ACCEPTUMULTICAST;
1191	if (ndev->flags & IFF_ALLMULTI)
1192		tmp32 |= LPC_RXFLTRW_ACCEPTUMULTICAST;
1193
1194	if (netdev_hw_addr_list_count(mcptr))
1195		tmp32 |= LPC_RXFLTRW_ACCEPTUMULTICASTHASH;
1196
1197	writel(tmp32, LPC_ENET_RXFILTER_CTRL(pldat->net_base));
1198
1199
1200	/* Set initial hash table */
1201	hashlo = 0x0;
1202	hashhi = 0x0;
1203
1204	/* 64 bits : multicast address in hash table */
1205	netdev_hw_addr_list_for_each(ha, mcptr) {
1206		hash_val = (ether_crc(6, ha->addr) >> 23) & 0x3F;
1207
1208		if (hash_val >= 32)
1209			hashhi |= 1 << (hash_val - 32);
1210		else
1211			hashlo |= 1 << hash_val;
1212	}
1213
1214	writel(hashlo, LPC_ENET_HASHFILTERL(pldat->net_base));
1215	writel(hashhi, LPC_ENET_HASHFILTERH(pldat->net_base));
1216
1217	spin_unlock_irqrestore(&pldat->lock, flags);
1218}
1219
1220static int lpc_eth_ioctl(struct net_device *ndev, struct ifreq *req, int cmd)
1221{
1222	struct netdata_local *pldat = netdev_priv(ndev);
1223	struct phy_device *phydev = pldat->phy_dev;
1224
1225	if (!netif_running(ndev))
1226		return -EINVAL;
1227
1228	if (!phydev)
1229		return -ENODEV;
1230
1231	return phy_mii_ioctl(phydev, req, cmd);
1232}
1233
1234static int lpc_eth_open(struct net_device *ndev)
1235{
1236	struct netdata_local *pldat = netdev_priv(ndev);
1237
1238	if (netif_msg_ifup(pldat))
1239		dev_dbg(&pldat->pdev->dev, "enabling %s\n", ndev->name);
1240
1241	if (!is_valid_ether_addr(ndev->dev_addr))
1242		return -EADDRNOTAVAIL;
1243
1244	__lpc_eth_clock_enable(pldat, true);
1245
1246	/* Reset and initialize */
1247	__lpc_eth_reset(pldat);
1248	__lpc_eth_init(pldat);
1249
1250	/* schedule a link state check */
1251	phy_start(pldat->phy_dev);
1252	netif_start_queue(ndev);
1253	napi_enable(&pldat->napi);
1254
1255	return 0;
1256}
1257
1258/*
1259 * Ethtool ops
1260 */
1261static void lpc_eth_ethtool_getdrvinfo(struct net_device *ndev,
1262	struct ethtool_drvinfo *info)
1263{
1264	strcpy(info->driver, MODNAME);
1265	strcpy(info->version, DRV_VERSION);
1266	strcpy(info->bus_info, dev_name(ndev->dev.parent));
1267}
1268
1269static u32 lpc_eth_ethtool_getmsglevel(struct net_device *ndev)
1270{
1271	struct netdata_local *pldat = netdev_priv(ndev);
1272
1273	return pldat->msg_enable;
1274}
1275
1276static void lpc_eth_ethtool_setmsglevel(struct net_device *ndev, u32 level)
1277{
1278	struct netdata_local *pldat = netdev_priv(ndev);
1279
1280	pldat->msg_enable = level;
1281}
1282
1283static int lpc_eth_ethtool_getsettings(struct net_device *ndev,
1284	struct ethtool_cmd *cmd)
1285{
1286	struct netdata_local *pldat = netdev_priv(ndev);
1287	struct phy_device *phydev = pldat->phy_dev;
1288
1289	if (!phydev)
1290		return -EOPNOTSUPP;
1291
1292	return phy_ethtool_gset(phydev, cmd);
1293}
1294
1295static int lpc_eth_ethtool_setsettings(struct net_device *ndev,
1296	struct ethtool_cmd *cmd)
1297{
1298	struct netdata_local *pldat = netdev_priv(ndev);
1299	struct phy_device *phydev = pldat->phy_dev;
1300
1301	if (!phydev)
1302		return -EOPNOTSUPP;
1303
1304	return phy_ethtool_sset(phydev, cmd);
1305}
1306
1307static const struct ethtool_ops lpc_eth_ethtool_ops = {
1308	.get_drvinfo	= lpc_eth_ethtool_getdrvinfo,
1309	.get_settings	= lpc_eth_ethtool_getsettings,
1310	.set_settings	= lpc_eth_ethtool_setsettings,
1311	.get_msglevel	= lpc_eth_ethtool_getmsglevel,
1312	.set_msglevel	= lpc_eth_ethtool_setmsglevel,
1313	.get_link	= ethtool_op_get_link,
1314};
1315
1316static const struct net_device_ops lpc_netdev_ops = {
1317	.ndo_open		= lpc_eth_open,
1318	.ndo_stop		= lpc_eth_close,
1319	.ndo_start_xmit		= lpc_eth_hard_start_xmit,
1320	.ndo_set_rx_mode	= lpc_eth_set_multicast_list,
1321	.ndo_do_ioctl		= lpc_eth_ioctl,
1322	.ndo_set_mac_address	= lpc_set_mac_address,
1323	.ndo_change_mtu		= eth_change_mtu,
1324};
1325
1326static int lpc_eth_drv_probe(struct platform_device *pdev)
1327{
1328	struct resource *res;
1329	struct net_device *ndev;
1330	struct netdata_local *pldat;
1331	struct phy_device *phydev;
1332	dma_addr_t dma_handle;
1333	int irq, ret;
1334	u32 tmp;
1335
1336	/* Setup network interface for RMII or MII mode */
1337	tmp = __raw_readl(LPC32XX_CLKPWR_MACCLK_CTRL);
1338	tmp &= ~LPC32XX_CLKPWR_MACCTRL_PINS_MSK;
1339	if (lpc_phy_interface_mode(&pdev->dev) == PHY_INTERFACE_MODE_MII)
1340		tmp |= LPC32XX_CLKPWR_MACCTRL_USE_MII_PINS;
1341	else
1342		tmp |= LPC32XX_CLKPWR_MACCTRL_USE_RMII_PINS;
1343	__raw_writel(tmp, LPC32XX_CLKPWR_MACCLK_CTRL);
1344
1345	/* Get platform resources */
1346	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1347	irq = platform_get_irq(pdev, 0);
1348	if ((!res) || (irq < 0) || (irq >= NR_IRQS)) {
1349		dev_err(&pdev->dev, "error getting resources.\n");
1350		ret = -ENXIO;
1351		goto err_exit;
1352	}
1353
1354	/* Allocate net driver data structure */
1355	ndev = alloc_etherdev(sizeof(struct netdata_local));
1356	if (!ndev) {
1357		dev_err(&pdev->dev, "could not allocate device.\n");
1358		ret = -ENOMEM;
1359		goto err_exit;
1360	}
1361
1362	SET_NETDEV_DEV(ndev, &pdev->dev);
1363
1364	pldat = netdev_priv(ndev);
1365	pldat->pdev = pdev;
1366	pldat->ndev = ndev;
1367
1368	spin_lock_init(&pldat->lock);
1369
1370	/* Save resources */
1371	ndev->irq = irq;
1372
1373	/* Get clock for the device */
1374	pldat->clk = clk_get(&pdev->dev, NULL);
1375	if (IS_ERR(pldat->clk)) {
1376		dev_err(&pdev->dev, "error getting clock.\n");
1377		ret = PTR_ERR(pldat->clk);
1378		goto err_out_free_dev;
1379	}
1380
1381	/* Enable network clock */
1382	__lpc_eth_clock_enable(pldat, true);
1383
1384	/* Map IO space */
1385	pldat->net_base = ioremap(res->start, res->end - res->start + 1);
1386	if (!pldat->net_base) {
1387		dev_err(&pdev->dev, "failed to map registers\n");
1388		ret = -ENOMEM;
1389		goto err_out_disable_clocks;
1390	}
1391	ret = request_irq(ndev->irq, __lpc_eth_interrupt, 0,
1392			  ndev->name, ndev);
1393	if (ret) {
1394		dev_err(&pdev->dev, "error requesting interrupt.\n");
1395		goto err_out_iounmap;
1396	}
1397
1398	/* Fill in the fields of the device structure with ethernet values. */
1399	ether_setup(ndev);
1400
1401	/* Setup driver functions */
1402	ndev->netdev_ops = &lpc_netdev_ops;
1403	ndev->ethtool_ops = &lpc_eth_ethtool_ops;
1404	ndev->watchdog_timeo = msecs_to_jiffies(2500);
1405
1406	/* Get size of DMA buffers/descriptors region */
1407	pldat->dma_buff_size = (ENET_TX_DESC + ENET_RX_DESC) * (ENET_MAXF_SIZE +
1408		sizeof(struct txrx_desc_t) + sizeof(struct rx_status_t));
1409	pldat->dma_buff_base_v = 0;
1410
1411	if (use_iram_for_net(&pldat->pdev->dev)) {
1412		dma_handle = LPC32XX_IRAM_BASE;
1413		if (pldat->dma_buff_size <= lpc32xx_return_iram_size())
1414			pldat->dma_buff_base_v =
1415				io_p2v(LPC32XX_IRAM_BASE);
1416		else
1417			netdev_err(ndev,
1418				"IRAM not big enough for net buffers, using SDRAM instead.\n");
1419	}
1420
1421	if (pldat->dma_buff_base_v == 0) {
1422		pldat->pdev->dev.coherent_dma_mask = 0xFFFFFFFF;
1423		pldat->pdev->dev.dma_mask = &pldat->pdev->dev.coherent_dma_mask;
1424		pldat->dma_buff_size = PAGE_ALIGN(pldat->dma_buff_size);
1425
1426		/* Allocate a chunk of memory for the DMA ethernet buffers
1427		   and descriptors */
1428		pldat->dma_buff_base_v =
1429			dma_alloc_coherent(&pldat->pdev->dev,
1430					   pldat->dma_buff_size, &dma_handle,
1431					   GFP_KERNEL);
1432
1433		if (pldat->dma_buff_base_v == NULL) {
1434			dev_err(&pdev->dev, "error getting DMA region.\n");
1435			ret = -ENOMEM;
1436			goto err_out_free_irq;
1437		}
1438	}
1439	pldat->dma_buff_base_p = dma_handle;
1440
1441	netdev_dbg(ndev, "IO address start     :0x%08x\n",
1442			res->start);
1443	netdev_dbg(ndev, "IO address size      :%d\n",
1444			res->end - res->start + 1);
1445	netdev_err(ndev, "IO address (mapped)  :0x%p\n",
1446			pldat->net_base);
1447	netdev_dbg(ndev, "IRQ number           :%d\n", ndev->irq);
1448	netdev_dbg(ndev, "DMA buffer size      :%d\n", pldat->dma_buff_size);
1449	netdev_dbg(ndev, "DMA buffer P address :0x%08x\n",
1450			pldat->dma_buff_base_p);
1451	netdev_dbg(ndev, "DMA buffer V address :0x%p\n",
1452			pldat->dma_buff_base_v);
1453
1454	/* Get MAC address from current HW setting (POR state is all zeros) */
1455	__lpc_get_mac(pldat, ndev->dev_addr);
1456
1457#ifdef CONFIG_OF_NET
1458	if (!is_valid_ether_addr(ndev->dev_addr)) {
1459		const char *macaddr = of_get_mac_address(pdev->dev.of_node);
1460		if (macaddr)
1461			memcpy(ndev->dev_addr, macaddr, ETH_ALEN);
1462	}
1463#endif
1464	if (!is_valid_ether_addr(ndev->dev_addr))
1465		eth_hw_addr_random(ndev);
1466
1467	/* Reset the ethernet controller */
1468	__lpc_eth_reset(pldat);
1469
1470	/* then shut everything down to save power */
1471	__lpc_eth_shutdown(pldat);
1472
1473	/* Set default parameters */
1474	pldat->msg_enable = NETIF_MSG_LINK;
1475
1476	/* Force an MII interface reset and clock setup */
1477	__lpc_mii_mngt_reset(pldat);
1478
1479	/* Force default PHY interface setup in chip, this will probably be
1480	   changed by the PHY driver */
1481	pldat->link = 0;
1482	pldat->speed = 100;
1483	pldat->duplex = DUPLEX_FULL;
1484	__lpc_params_setup(pldat);
1485
1486	netif_napi_add(ndev, &pldat->napi, lpc_eth_poll, NAPI_WEIGHT);
1487
1488	ret = register_netdev(ndev);
1489	if (ret) {
1490		dev_err(&pdev->dev, "Cannot register net device, aborting.\n");
1491		goto err_out_dma_unmap;
1492	}
1493	platform_set_drvdata(pdev, ndev);
1494
1495	if (lpc_mii_init(pldat) != 0)
1496		goto err_out_unregister_netdev;
1497
1498	netdev_info(ndev, "LPC mac at 0x%08x irq %d\n",
1499	       res->start, ndev->irq);
1500
1501	phydev = pldat->phy_dev;
1502
1503	device_init_wakeup(&pdev->dev, 1);
1504	device_set_wakeup_enable(&pdev->dev, 0);
1505
1506	return 0;
1507
1508err_out_unregister_netdev:
1509	platform_set_drvdata(pdev, NULL);
1510	unregister_netdev(ndev);
1511err_out_dma_unmap:
1512	if (!use_iram_for_net(&pldat->pdev->dev) ||
1513	    pldat->dma_buff_size > lpc32xx_return_iram_size())
1514		dma_free_coherent(&pldat->pdev->dev, pldat->dma_buff_size,
1515				  pldat->dma_buff_base_v,
1516				  pldat->dma_buff_base_p);
1517err_out_free_irq:
1518	free_irq(ndev->irq, ndev);
1519err_out_iounmap:
1520	iounmap(pldat->net_base);
1521err_out_disable_clocks:
1522	clk_disable(pldat->clk);
1523	clk_put(pldat->clk);
1524err_out_free_dev:
1525	free_netdev(ndev);
1526err_exit:
1527	pr_err("%s: not found (%d).\n", MODNAME, ret);
1528	return ret;
1529}
1530
1531static int lpc_eth_drv_remove(struct platform_device *pdev)
1532{
1533	struct net_device *ndev = platform_get_drvdata(pdev);
1534	struct netdata_local *pldat = netdev_priv(ndev);
1535
1536	unregister_netdev(ndev);
1537	platform_set_drvdata(pdev, NULL);
1538
1539	if (!use_iram_for_net(&pldat->pdev->dev) ||
1540	    pldat->dma_buff_size > lpc32xx_return_iram_size())
1541		dma_free_coherent(&pldat->pdev->dev, pldat->dma_buff_size,
1542				  pldat->dma_buff_base_v,
1543				  pldat->dma_buff_base_p);
1544	free_irq(ndev->irq, ndev);
1545	iounmap(pldat->net_base);
1546	mdiobus_free(pldat->mii_bus);
1547	clk_disable(pldat->clk);
1548	clk_put(pldat->clk);
1549	free_netdev(ndev);
1550
1551	return 0;
1552}
1553
1554#ifdef CONFIG_PM
1555static int lpc_eth_drv_suspend(struct platform_device *pdev,
1556	pm_message_t state)
1557{
1558	struct net_device *ndev = platform_get_drvdata(pdev);
1559	struct netdata_local *pldat = netdev_priv(ndev);
1560
1561	if (device_may_wakeup(&pdev->dev))
1562		enable_irq_wake(ndev->irq);
1563
1564	if (ndev) {
1565		if (netif_running(ndev)) {
1566			netif_device_detach(ndev);
1567			__lpc_eth_shutdown(pldat);
1568			clk_disable(pldat->clk);
1569
1570			/*
1571			 * Reset again now clock is disable to be sure
1572			 * EMC_MDC is down
1573			 */
1574			__lpc_eth_reset(pldat);
1575		}
1576	}
1577
1578	return 0;
1579}
1580
1581static int lpc_eth_drv_resume(struct platform_device *pdev)
1582{
1583	struct net_device *ndev = platform_get_drvdata(pdev);
1584	struct netdata_local *pldat;
1585
1586	if (device_may_wakeup(&pdev->dev))
1587		disable_irq_wake(ndev->irq);
1588
1589	if (ndev) {
1590		if (netif_running(ndev)) {
1591			pldat = netdev_priv(ndev);
1592
1593			/* Enable interface clock */
1594			clk_enable(pldat->clk);
1595
1596			/* Reset and initialize */
1597			__lpc_eth_reset(pldat);
1598			__lpc_eth_init(pldat);
1599
1600			netif_device_attach(ndev);
1601		}
1602	}
1603
1604	return 0;
1605}
1606#endif
1607
1608#ifdef CONFIG_OF
1609static const struct of_device_id lpc_eth_match[] = {
1610	{ .compatible = "nxp,lpc-eth" },
1611	{ }
1612};
1613MODULE_DEVICE_TABLE(of, lpc_eth_match);
1614#endif
1615
1616static struct platform_driver lpc_eth_driver = {
1617	.probe		= lpc_eth_drv_probe,
1618	.remove		= __devexit_p(lpc_eth_drv_remove),
1619#ifdef CONFIG_PM
1620	.suspend	= lpc_eth_drv_suspend,
1621	.resume		= lpc_eth_drv_resume,
1622#endif
1623	.driver		= {
1624		.name	= MODNAME,
1625		.of_match_table = of_match_ptr(lpc_eth_match),
1626	},
1627};
1628
1629module_platform_driver(lpc_eth_driver);
1630
1631MODULE_AUTHOR("Kevin Wells <kevin.wells@nxp.com>");
1632MODULE_AUTHOR("Roland Stigge <stigge@antcom.de>");
1633MODULE_DESCRIPTION("LPC Ethernet Driver");
1634MODULE_LICENSE("GPL");