Linux Audio

Check our new training course

Embedded Linux training

Mar 10-20, 2025, special US time zones
Register
Loading...
v6.8
   1// SPDX-License-Identifier: GPL-2.0-or-later
   2/*
   3 * RDC R6040 Fast Ethernet MAC support
   4 *
   5 * Copyright (C) 2004 Sten Wang <sten.wang@rdc.com.tw>
   6 * Copyright (C) 2007
   7 *	Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>
   8 * Copyright (C) 2007-2012 Florian Fainelli <f.fainelli@gmail.com>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
   9*/
  10
  11#include <linux/kernel.h>
  12#include <linux/module.h>
  13#include <linux/moduleparam.h>
  14#include <linux/string.h>
  15#include <linux/timer.h>
  16#include <linux/errno.h>
  17#include <linux/ioport.h>
  18#include <linux/interrupt.h>
  19#include <linux/pci.h>
  20#include <linux/netdevice.h>
  21#include <linux/etherdevice.h>
  22#include <linux/skbuff.h>
 
  23#include <linux/delay.h>
  24#include <linux/mii.h>
  25#include <linux/ethtool.h>
  26#include <linux/crc32.h>
  27#include <linux/spinlock.h>
  28#include <linux/bitops.h>
  29#include <linux/io.h>
  30#include <linux/irq.h>
  31#include <linux/uaccess.h>
  32#include <linux/phy.h>
  33
  34#include <asm/processor.h>
  35
  36#define DRV_NAME	"r6040"
  37#define DRV_VERSION	"0.29"
  38#define DRV_RELDATE	"04Jul2016"
  39
  40/* Time in jiffies before concluding the transmitter is hung. */
  41#define TX_TIMEOUT	(6000 * HZ / 1000)
  42
  43/* RDC MAC I/O Size */
  44#define R6040_IO_SIZE	256
  45
  46/* MAX RDC MAC */
  47#define MAX_MAC		2
  48
  49/* MAC registers */
  50#define MCR0		0x00	/* Control register 0 */
  51#define  MCR0_RCVEN	0x0002	/* Receive enable */
  52#define  MCR0_PROMISC	0x0020	/* Promiscuous mode */
  53#define  MCR0_HASH_EN	0x0100	/* Enable multicast hash table function */
  54#define  MCR0_XMTEN	0x1000	/* Transmission enable */
  55#define  MCR0_FD	0x8000	/* Full/Half duplex */
  56#define MCR1		0x04	/* Control register 1 */
  57#define  MAC_RST	0x0001	/* Reset the MAC */
  58#define MBCR		0x08	/* Bus control */
  59#define MT_ICR		0x0C	/* TX interrupt control */
  60#define MR_ICR		0x10	/* RX interrupt control */
  61#define MTPR		0x14	/* TX poll command register */
  62#define  TM2TX		0x0001	/* Trigger MAC to transmit */
  63#define MR_BSR		0x18	/* RX buffer size */
  64#define MR_DCR		0x1A	/* RX descriptor control */
  65#define MLSR		0x1C	/* Last status */
  66#define  TX_FIFO_UNDR	0x0200	/* TX FIFO under-run */
  67#define	 TX_EXCEEDC	0x2000	/* Transmit exceed collision */
  68#define  TX_LATEC	0x4000	/* Transmit late collision */
  69#define MMDIO		0x20	/* MDIO control register */
  70#define  MDIO_WRITE	0x4000	/* MDIO write */
  71#define  MDIO_READ	0x2000	/* MDIO read */
  72#define MMRD		0x24	/* MDIO read data register */
  73#define MMWD		0x28	/* MDIO write data register */
  74#define MTD_SA0		0x2C	/* TX descriptor start address 0 */
  75#define MTD_SA1		0x30	/* TX descriptor start address 1 */
  76#define MRD_SA0		0x34	/* RX descriptor start address 0 */
  77#define MRD_SA1		0x38	/* RX descriptor start address 1 */
  78#define MISR		0x3C	/* Status register */
  79#define MIER		0x40	/* INT enable register */
  80#define  MSK_INT	0x0000	/* Mask off interrupts */
  81#define  RX_FINISH	0x0001  /* RX finished */
  82#define  RX_NO_DESC	0x0002  /* No RX descriptor available */
  83#define  RX_FIFO_FULL	0x0004  /* RX FIFO full */
  84#define  RX_EARLY	0x0008  /* RX early */
  85#define  TX_FINISH	0x0010  /* TX finished */
  86#define  TX_EARLY	0x0080  /* TX early */
  87#define  EVENT_OVRFL	0x0100  /* Event counter overflow */
  88#define  LINK_CHANGED	0x0200  /* PHY link changed */
  89#define ME_CISR		0x44	/* Event counter INT status */
  90#define ME_CIER		0x48	/* Event counter INT enable  */
  91#define MR_CNT		0x50	/* Successfully received packet counter */
  92#define ME_CNT0		0x52	/* Event counter 0 */
  93#define ME_CNT1		0x54	/* Event counter 1 */
  94#define ME_CNT2		0x56	/* Event counter 2 */
  95#define ME_CNT3		0x58	/* Event counter 3 */
  96#define MT_CNT		0x5A	/* Successfully transmit packet counter */
  97#define ME_CNT4		0x5C	/* Event counter 4 */
  98#define MP_CNT		0x5E	/* Pause frame counter register */
  99#define MAR0		0x60	/* Hash table 0 */
 100#define MAR1		0x62	/* Hash table 1 */
 101#define MAR2		0x64	/* Hash table 2 */
 102#define MAR3		0x66	/* Hash table 3 */
 103#define MID_0L		0x68	/* Multicast address MID0 Low */
 104#define MID_0M		0x6A	/* Multicast address MID0 Medium */
 105#define MID_0H		0x6C	/* Multicast address MID0 High */
 106#define MID_1L		0x70	/* MID1 Low */
 107#define MID_1M		0x72	/* MID1 Medium */
 108#define MID_1H		0x74	/* MID1 High */
 109#define MID_2L		0x78	/* MID2 Low */
 110#define MID_2M		0x7A	/* MID2 Medium */
 111#define MID_2H		0x7C	/* MID2 High */
 112#define MID_3L		0x80	/* MID3 Low */
 113#define MID_3M		0x82	/* MID3 Medium */
 114#define MID_3H		0x84	/* MID3 High */
 115#define PHY_CC		0x88	/* PHY status change configuration register */
 116#define  SCEN		0x8000	/* PHY status change enable */
 117#define  PHYAD_SHIFT	8	/* PHY address shift */
 118#define  TMRDIV_SHIFT	0	/* Timer divider shift */
 119#define PHY_ST		0x8A	/* PHY status register */
 120#define MAC_SM		0xAC	/* MAC status machine */
 121#define  MAC_SM_RST	0x0002	/* MAC status machine reset */
 122#define MD_CSC		0xb6	/* MDC speed control register */
 123#define  MD_CSC_DEFAULT	0x0030
 124#define MAC_ID		0xBE	/* Identifier register */
 125
 126#define TX_DCNT		0x80	/* TX descriptor count */
 127#define RX_DCNT		0x80	/* RX descriptor count */
 128#define MAX_BUF_SIZE	0x600
 129#define RX_DESC_SIZE	(RX_DCNT * sizeof(struct r6040_descriptor))
 130#define TX_DESC_SIZE	(TX_DCNT * sizeof(struct r6040_descriptor))
 131#define MBCR_DEFAULT	0x012A	/* MAC Bus Control Register */
 132#define MCAST_MAX	3	/* Max number multicast addresses to filter */
 133
 134#define MAC_DEF_TIMEOUT	2048	/* Default MAC read/write operation timeout */
 135
 136/* Descriptor status */
 137#define DSC_OWNER_MAC	0x8000	/* MAC is the owner of this descriptor */
 138#define DSC_RX_OK	0x4000	/* RX was successful */
 139#define DSC_RX_ERR	0x0800	/* RX PHY error */
 140#define DSC_RX_ERR_DRI	0x0400	/* RX dribble packet */
 141#define DSC_RX_ERR_BUF	0x0200	/* RX length exceeds buffer size */
 142#define DSC_RX_ERR_LONG	0x0100	/* RX length > maximum packet length */
 143#define DSC_RX_ERR_RUNT	0x0080	/* RX packet length < 64 byte */
 144#define DSC_RX_ERR_CRC	0x0040	/* RX CRC error */
 145#define DSC_RX_BCAST	0x0020	/* RX broadcast (no error) */
 146#define DSC_RX_MCAST	0x0010	/* RX multicast (no error) */
 147#define DSC_RX_MCH_HIT	0x0008	/* RX multicast hit in hash table (no error) */
 148#define DSC_RX_MIDH_HIT	0x0004	/* RX MID table hit (no error) */
 149#define DSC_RX_IDX_MID_MASK 3	/* RX mask for the index of matched MIDx */
 150
 151MODULE_AUTHOR("Sten Wang <sten.wang@rdc.com.tw>,"
 152	"Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>,"
 153	"Florian Fainelli <f.fainelli@gmail.com>");
 154MODULE_LICENSE("GPL");
 155MODULE_DESCRIPTION("RDC R6040 NAPI PCI FastEthernet driver");
 156MODULE_VERSION(DRV_VERSION " " DRV_RELDATE);
 157
 158/* RX and TX interrupts that we handle */
 159#define RX_INTS			(RX_FIFO_FULL | RX_NO_DESC | RX_FINISH)
 160#define TX_INTS			(TX_FINISH)
 161#define INT_MASK		(RX_INTS | TX_INTS)
 162
 163struct r6040_descriptor {
 164	u16	status, len;		/* 0-3 */
 165	__le32	buf;			/* 4-7 */
 166	__le32	ndesc;			/* 8-B */
 167	u32	rev1;			/* C-F */
 168	char	*vbufp;			/* 10-13 */
 169	struct r6040_descriptor *vndescp;	/* 14-17 */
 170	struct sk_buff *skb_ptr;	/* 18-1B */
 171	u32	rev2;			/* 1C-1F */
 172} __aligned(32);
 173
 174struct r6040_private {
 175	spinlock_t lock;		/* driver lock */
 176	struct pci_dev *pdev;
 177	struct r6040_descriptor *rx_insert_ptr;
 178	struct r6040_descriptor *rx_remove_ptr;
 179	struct r6040_descriptor *tx_insert_ptr;
 180	struct r6040_descriptor *tx_remove_ptr;
 181	struct r6040_descriptor *rx_ring;
 182	struct r6040_descriptor *tx_ring;
 183	dma_addr_t rx_ring_dma;
 184	dma_addr_t tx_ring_dma;
 185	u16	tx_free_desc;
 186	u16	mcr0;
 187	struct net_device *dev;
 188	struct mii_bus *mii_bus;
 189	struct napi_struct napi;
 190	void __iomem *base;
 
 191	int old_link;
 192	int old_duplex;
 193};
 194
 195static char version[] = DRV_NAME
 196	": RDC R6040 NAPI net driver,"
 197	"version "DRV_VERSION " (" DRV_RELDATE ")";
 198
 199/* Read a word data from PHY Chip */
 200static int r6040_phy_read(void __iomem *ioaddr, int phy_addr, int reg)
 201{
 202	int limit = MAC_DEF_TIMEOUT;
 203	u16 cmd;
 204
 205	iowrite16(MDIO_READ | reg | (phy_addr << 8), ioaddr + MMDIO);
 206	/* Wait for the read bit to be cleared */
 207	while (limit--) {
 208		cmd = ioread16(ioaddr + MMDIO);
 209		if (!(cmd & MDIO_READ))
 210			break;
 211		udelay(1);
 212	}
 213
 214	if (limit < 0)
 215		return -ETIMEDOUT;
 216
 217	return ioread16(ioaddr + MMRD);
 218}
 219
 220/* Write a word data from PHY Chip */
 221static int r6040_phy_write(void __iomem *ioaddr,
 222					int phy_addr, int reg, u16 val)
 223{
 224	int limit = MAC_DEF_TIMEOUT;
 225	u16 cmd;
 226
 227	iowrite16(val, ioaddr + MMWD);
 228	/* Write the command to the MDIO bus */
 229	iowrite16(MDIO_WRITE | reg | (phy_addr << 8), ioaddr + MMDIO);
 230	/* Wait for the write bit to be cleared */
 231	while (limit--) {
 232		cmd = ioread16(ioaddr + MMDIO);
 233		if (!(cmd & MDIO_WRITE))
 234			break;
 235		udelay(1);
 236	}
 237
 238	return (limit < 0) ? -ETIMEDOUT : 0;
 239}
 240
 241static int r6040_mdiobus_read(struct mii_bus *bus, int phy_addr, int reg)
 242{
 243	struct net_device *dev = bus->priv;
 244	struct r6040_private *lp = netdev_priv(dev);
 245	void __iomem *ioaddr = lp->base;
 246
 247	return r6040_phy_read(ioaddr, phy_addr, reg);
 248}
 249
 250static int r6040_mdiobus_write(struct mii_bus *bus, int phy_addr,
 251						int reg, u16 value)
 252{
 253	struct net_device *dev = bus->priv;
 254	struct r6040_private *lp = netdev_priv(dev);
 255	void __iomem *ioaddr = lp->base;
 256
 257	return r6040_phy_write(ioaddr, phy_addr, reg, value);
 
 
 
 
 
 
 
 258}
 259
 260static void r6040_free_txbufs(struct net_device *dev)
 261{
 262	struct r6040_private *lp = netdev_priv(dev);
 263	int i;
 264
 265	for (i = 0; i < TX_DCNT; i++) {
 266		if (lp->tx_insert_ptr->skb_ptr) {
 267			dma_unmap_single(&lp->pdev->dev,
 268					 le32_to_cpu(lp->tx_insert_ptr->buf),
 269					 MAX_BUF_SIZE, DMA_TO_DEVICE);
 270			dev_kfree_skb(lp->tx_insert_ptr->skb_ptr);
 271			lp->tx_insert_ptr->skb_ptr = NULL;
 272		}
 273		lp->tx_insert_ptr = lp->tx_insert_ptr->vndescp;
 274	}
 275}
 276
 277static void r6040_free_rxbufs(struct net_device *dev)
 278{
 279	struct r6040_private *lp = netdev_priv(dev);
 280	int i;
 281
 282	for (i = 0; i < RX_DCNT; i++) {
 283		if (lp->rx_insert_ptr->skb_ptr) {
 284			dma_unmap_single(&lp->pdev->dev,
 285					 le32_to_cpu(lp->rx_insert_ptr->buf),
 286					 MAX_BUF_SIZE, DMA_FROM_DEVICE);
 287			dev_kfree_skb(lp->rx_insert_ptr->skb_ptr);
 288			lp->rx_insert_ptr->skb_ptr = NULL;
 289		}
 290		lp->rx_insert_ptr = lp->rx_insert_ptr->vndescp;
 291	}
 292}
 293
 294static void r6040_init_ring_desc(struct r6040_descriptor *desc_ring,
 295				 dma_addr_t desc_dma, int size)
 296{
 297	struct r6040_descriptor *desc = desc_ring;
 298	dma_addr_t mapping = desc_dma;
 299
 300	while (size-- > 0) {
 301		mapping += sizeof(*desc);
 302		desc->ndesc = cpu_to_le32(mapping);
 303		desc->vndescp = desc + 1;
 304		desc++;
 305	}
 306	desc--;
 307	desc->ndesc = cpu_to_le32(desc_dma);
 308	desc->vndescp = desc_ring;
 309}
 310
 311static void r6040_init_txbufs(struct net_device *dev)
 312{
 313	struct r6040_private *lp = netdev_priv(dev);
 314
 315	lp->tx_free_desc = TX_DCNT;
 316
 317	lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring;
 318	r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT);
 319}
 320
 321static int r6040_alloc_rxbufs(struct net_device *dev)
 322{
 323	struct r6040_private *lp = netdev_priv(dev);
 324	struct r6040_descriptor *desc;
 325	struct sk_buff *skb;
 326	int rc;
 327
 328	lp->rx_remove_ptr = lp->rx_insert_ptr = lp->rx_ring;
 329	r6040_init_ring_desc(lp->rx_ring, lp->rx_ring_dma, RX_DCNT);
 330
 331	/* Allocate skbs for the rx descriptors */
 332	desc = lp->rx_ring;
 333	do {
 334		skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
 335		if (!skb) {
 
 336			rc = -ENOMEM;
 337			goto err_exit;
 338		}
 339		desc->skb_ptr = skb;
 340		desc->buf = cpu_to_le32(dma_map_single(&lp->pdev->dev,
 341						       desc->skb_ptr->data,
 342						       MAX_BUF_SIZE,
 343						       DMA_FROM_DEVICE));
 344		desc->status = DSC_OWNER_MAC;
 345		desc = desc->vndescp;
 346	} while (desc != lp->rx_ring);
 347
 348	return 0;
 349
 350err_exit:
 351	/* Deallocate all previously allocated skbs */
 352	r6040_free_rxbufs(dev);
 353	return rc;
 354}
 355
 356static void r6040_reset_mac(struct r6040_private *lp)
 357{
 358	void __iomem *ioaddr = lp->base;
 359	int limit = MAC_DEF_TIMEOUT;
 360	u16 cmd, md_csc;
 361
 362	md_csc = ioread16(ioaddr + MD_CSC);
 363	iowrite16(MAC_RST, ioaddr + MCR1);
 364	while (limit--) {
 365		cmd = ioread16(ioaddr + MCR1);
 366		if (cmd & MAC_RST)
 367			break;
 368	}
 369
 370	/* Reset internal state machine */
 371	iowrite16(MAC_SM_RST, ioaddr + MAC_SM);
 372	iowrite16(0, ioaddr + MAC_SM);
 373	mdelay(5);
 374
 375	/* Restore MDIO clock frequency */
 376	if (md_csc != MD_CSC_DEFAULT)
 377		iowrite16(md_csc, ioaddr + MD_CSC);
 378}
 379
 380static void r6040_init_mac_regs(struct net_device *dev)
 381{
 382	struct r6040_private *lp = netdev_priv(dev);
 383	void __iomem *ioaddr = lp->base;
 384
 385	/* Mask Off Interrupt */
 386	iowrite16(MSK_INT, ioaddr + MIER);
 387
 388	/* Reset RDC MAC */
 389	r6040_reset_mac(lp);
 390
 391	/* MAC Bus Control Register */
 392	iowrite16(MBCR_DEFAULT, ioaddr + MBCR);
 393
 394	/* Buffer Size Register */
 395	iowrite16(MAX_BUF_SIZE, ioaddr + MR_BSR);
 396
 397	/* Write TX ring start address */
 398	iowrite16(lp->tx_ring_dma, ioaddr + MTD_SA0);
 399	iowrite16(lp->tx_ring_dma >> 16, ioaddr + MTD_SA1);
 400
 401	/* Write RX ring start address */
 402	iowrite16(lp->rx_ring_dma, ioaddr + MRD_SA0);
 403	iowrite16(lp->rx_ring_dma >> 16, ioaddr + MRD_SA1);
 404
 405	/* Set interrupt waiting time and packet numbers */
 406	iowrite16(0, ioaddr + MT_ICR);
 407	iowrite16(0, ioaddr + MR_ICR);
 408
 409	/* Enable interrupts */
 410	iowrite16(INT_MASK, ioaddr + MIER);
 411
 412	/* Enable TX and RX */
 413	iowrite16(lp->mcr0 | MCR0_RCVEN, ioaddr);
 414
 415	/* Let TX poll the descriptors
 416	 * we may got called by r6040_tx_timeout which has left
 417	 * some unsent tx buffers */
 418	iowrite16(TM2TX, ioaddr + MTPR);
 419}
 420
 421static void r6040_tx_timeout(struct net_device *dev, unsigned int txqueue)
 422{
 423	struct r6040_private *priv = netdev_priv(dev);
 424	void __iomem *ioaddr = priv->base;
 425
 426	netdev_warn(dev, "transmit timed out, int enable %4.4x "
 427		"status %4.4x\n",
 428		ioread16(ioaddr + MIER),
 429		ioread16(ioaddr + MISR));
 430
 431	dev->stats.tx_errors++;
 432
 433	/* Reset MAC and re-init all registers */
 434	r6040_init_mac_regs(dev);
 435}
 436
 437static struct net_device_stats *r6040_get_stats(struct net_device *dev)
 438{
 439	struct r6040_private *priv = netdev_priv(dev);
 440	void __iomem *ioaddr = priv->base;
 441	unsigned long flags;
 442
 443	spin_lock_irqsave(&priv->lock, flags);
 444	dev->stats.rx_crc_errors += ioread8(ioaddr + ME_CNT1);
 445	dev->stats.multicast += ioread8(ioaddr + ME_CNT0);
 446	spin_unlock_irqrestore(&priv->lock, flags);
 447
 448	return &dev->stats;
 449}
 450
 451/* Stop RDC MAC and Free the allocated resource */
 452static void r6040_down(struct net_device *dev)
 453{
 454	struct r6040_private *lp = netdev_priv(dev);
 455	void __iomem *ioaddr = lp->base;
 456	const u16 *adrp;
 457
 458	/* Stop MAC */
 459	iowrite16(MSK_INT, ioaddr + MIER);	/* Mask Off Interrupt */
 460
 461	/* Reset RDC MAC */
 462	r6040_reset_mac(lp);
 463
 464	/* Restore MAC Address to MIDx */
 465	adrp = (const u16 *) dev->dev_addr;
 466	iowrite16(adrp[0], ioaddr + MID_0L);
 467	iowrite16(adrp[1], ioaddr + MID_0M);
 468	iowrite16(adrp[2], ioaddr + MID_0H);
 
 
 469}
 470
 471static int r6040_close(struct net_device *dev)
 472{
 473	struct r6040_private *lp = netdev_priv(dev);
 474	struct pci_dev *pdev = lp->pdev;
 475
 476	phy_stop(dev->phydev);
 477	napi_disable(&lp->napi);
 478	netif_stop_queue(dev);
 479
 480	spin_lock_irq(&lp->lock);
 481	r6040_down(dev);
 482
 
 
 483	/* Free RX buffer */
 484	r6040_free_rxbufs(dev);
 485
 486	/* Free TX buffer */
 487	r6040_free_txbufs(dev);
 488
 489	spin_unlock_irq(&lp->lock);
 490
 491	free_irq(dev->irq, dev);
 492
 493	/* Free Descriptor memory */
 494	if (lp->rx_ring) {
 495		dma_free_coherent(&pdev->dev, RX_DESC_SIZE, lp->rx_ring,
 496				  lp->rx_ring_dma);
 497		lp->rx_ring = NULL;
 498	}
 499
 500	if (lp->tx_ring) {
 501		dma_free_coherent(&pdev->dev, TX_DESC_SIZE, lp->tx_ring,
 502				  lp->tx_ring_dma);
 503		lp->tx_ring = NULL;
 504	}
 505
 506	return 0;
 507}
 508
 
 
 
 
 
 
 
 
 
 
 509static int r6040_rx(struct net_device *dev, int limit)
 510{
 511	struct r6040_private *priv = netdev_priv(dev);
 512	struct r6040_descriptor *descptr = priv->rx_remove_ptr;
 513	struct sk_buff *skb_ptr, *new_skb;
 514	int count = 0;
 515	u16 err;
 516
 517	/* Limit not reached and the descriptor belongs to the CPU */
 518	while (count < limit && !(descptr->status & DSC_OWNER_MAC)) {
 519		/* Read the descriptor status */
 520		err = descptr->status;
 521		/* Global error status set */
 522		if (err & DSC_RX_ERR) {
 523			/* RX dribble */
 524			if (err & DSC_RX_ERR_DRI)
 525				dev->stats.rx_frame_errors++;
 526			/* Buffer length exceeded */
 527			if (err & DSC_RX_ERR_BUF)
 528				dev->stats.rx_length_errors++;
 529			/* Packet too long */
 530			if (err & DSC_RX_ERR_LONG)
 531				dev->stats.rx_length_errors++;
 532			/* Packet < 64 bytes */
 533			if (err & DSC_RX_ERR_RUNT)
 534				dev->stats.rx_length_errors++;
 535			/* CRC error */
 536			if (err & DSC_RX_ERR_CRC) {
 537				spin_lock(&priv->lock);
 538				dev->stats.rx_crc_errors++;
 539				spin_unlock(&priv->lock);
 540			}
 541			goto next_descr;
 542		}
 543
 544		/* Packet successfully received */
 545		new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
 546		if (!new_skb) {
 547			dev->stats.rx_dropped++;
 548			goto next_descr;
 549		}
 550		skb_ptr = descptr->skb_ptr;
 551		skb_ptr->dev = priv->dev;
 552
 553		/* Do not count the CRC */
 554		skb_put(skb_ptr, descptr->len - ETH_FCS_LEN);
 555		dma_unmap_single(&priv->pdev->dev, le32_to_cpu(descptr->buf),
 556				 MAX_BUF_SIZE, DMA_FROM_DEVICE);
 557		skb_ptr->protocol = eth_type_trans(skb_ptr, priv->dev);
 558
 559		/* Send to upper layer */
 560		netif_receive_skb(skb_ptr);
 561		dev->stats.rx_packets++;
 562		dev->stats.rx_bytes += descptr->len - ETH_FCS_LEN;
 563
 564		/* put new skb into descriptor */
 565		descptr->skb_ptr = new_skb;
 566		descptr->buf = cpu_to_le32(dma_map_single(&priv->pdev->dev,
 567							  descptr->skb_ptr->data,
 568							  MAX_BUF_SIZE,
 569							  DMA_FROM_DEVICE));
 570
 571next_descr:
 572		/* put the descriptor back to the MAC */
 573		descptr->status = DSC_OWNER_MAC;
 574		descptr = descptr->vndescp;
 575		count++;
 576	}
 577	priv->rx_remove_ptr = descptr;
 578
 579	return count;
 580}
 581
 582static void r6040_tx(struct net_device *dev)
 583{
 584	struct r6040_private *priv = netdev_priv(dev);
 585	struct r6040_descriptor *descptr;
 586	void __iomem *ioaddr = priv->base;
 587	struct sk_buff *skb_ptr;
 588	u16 err;
 589
 590	spin_lock(&priv->lock);
 591	descptr = priv->tx_remove_ptr;
 592	while (priv->tx_free_desc < TX_DCNT) {
 593		/* Check for errors */
 594		err = ioread16(ioaddr + MLSR);
 595
 596		if (err & TX_FIFO_UNDR)
 597			dev->stats.tx_fifo_errors++;
 598		if (err & (TX_EXCEEDC | TX_LATEC))
 599			dev->stats.tx_carrier_errors++;
 600
 601		if (descptr->status & DSC_OWNER_MAC)
 602			break; /* Not complete */
 603		skb_ptr = descptr->skb_ptr;
 604
 605		/* Statistic Counter */
 606		dev->stats.tx_packets++;
 607		dev->stats.tx_bytes += skb_ptr->len;
 608
 609		dma_unmap_single(&priv->pdev->dev, le32_to_cpu(descptr->buf),
 610				 skb_ptr->len, DMA_TO_DEVICE);
 611		/* Free buffer */
 612		dev_kfree_skb(skb_ptr);
 613		descptr->skb_ptr = NULL;
 614		/* To next descriptor */
 615		descptr = descptr->vndescp;
 616		priv->tx_free_desc++;
 617	}
 618	priv->tx_remove_ptr = descptr;
 619
 620	if (priv->tx_free_desc)
 621		netif_wake_queue(dev);
 622	spin_unlock(&priv->lock);
 623}
 624
 625static int r6040_poll(struct napi_struct *napi, int budget)
 626{
 627	struct r6040_private *priv =
 628		container_of(napi, struct r6040_private, napi);
 629	struct net_device *dev = priv->dev;
 630	void __iomem *ioaddr = priv->base;
 631	int work_done;
 632
 633	r6040_tx(dev);
 634
 635	work_done = r6040_rx(dev, budget);
 636
 637	if (work_done < budget) {
 638		napi_complete_done(napi, work_done);
 639		/* Enable RX/TX interrupt */
 640		iowrite16(ioread16(ioaddr + MIER) | RX_INTS | TX_INTS,
 641			  ioaddr + MIER);
 642	}
 643	return work_done;
 644}
 645
 646/* The RDC interrupt handler. */
 647static irqreturn_t r6040_interrupt(int irq, void *dev_id)
 648{
 649	struct net_device *dev = dev_id;
 650	struct r6040_private *lp = netdev_priv(dev);
 651	void __iomem *ioaddr = lp->base;
 652	u16 misr, status;
 653
 654	/* Save MIER */
 655	misr = ioread16(ioaddr + MIER);
 656	/* Mask off RDC MAC interrupt */
 657	iowrite16(MSK_INT, ioaddr + MIER);
 658	/* Read MISR status and clear */
 659	status = ioread16(ioaddr + MISR);
 660
 661	if (status == 0x0000 || status == 0xffff) {
 662		/* Restore RDC MAC interrupt */
 663		iowrite16(misr, ioaddr + MIER);
 664		return IRQ_NONE;
 665	}
 666
 667	/* RX interrupt request */
 668	if (status & (RX_INTS | TX_INTS)) {
 669		if (status & RX_NO_DESC) {
 670			/* RX descriptor unavailable */
 671			dev->stats.rx_dropped++;
 672			dev->stats.rx_missed_errors++;
 673		}
 674		if (status & RX_FIFO_FULL)
 675			dev->stats.rx_fifo_errors++;
 676
 677		if (likely(napi_schedule_prep(&lp->napi))) {
 678			/* Mask off RX interrupt */
 679			misr &= ~(RX_INTS | TX_INTS);
 680			__napi_schedule_irqoff(&lp->napi);
 681		}
 682	}
 683
 
 
 
 
 684	/* Restore RDC MAC interrupt */
 685	iowrite16(misr, ioaddr + MIER);
 686
 687	return IRQ_HANDLED;
 688}
 689
 690#ifdef CONFIG_NET_POLL_CONTROLLER
 691static void r6040_poll_controller(struct net_device *dev)
 692{
 693	disable_irq(dev->irq);
 694	r6040_interrupt(dev->irq, dev);
 695	enable_irq(dev->irq);
 696}
 697#endif
 698
 699/* Init RDC MAC */
 700static int r6040_up(struct net_device *dev)
 701{
 702	struct r6040_private *lp = netdev_priv(dev);
 703	void __iomem *ioaddr = lp->base;
 704	int ret;
 705
 706	/* Initialise and alloc RX/TX buffers */
 707	r6040_init_txbufs(dev);
 708	ret = r6040_alloc_rxbufs(dev);
 709	if (ret)
 710		return ret;
 711
 712	/* improve performance (by RDC guys) */
 713	r6040_phy_write(ioaddr, 30, 17,
 714			(r6040_phy_read(ioaddr, 30, 17) | 0x4000));
 715	r6040_phy_write(ioaddr, 30, 17,
 716			~((~r6040_phy_read(ioaddr, 30, 17)) | 0x2000));
 717	r6040_phy_write(ioaddr, 0, 19, 0x0000);
 718	r6040_phy_write(ioaddr, 0, 30, 0x01F0);
 719
 720	/* Initialize all MAC registers */
 721	r6040_init_mac_regs(dev);
 722
 723	phy_start(dev->phydev);
 724
 725	return 0;
 726}
 727
 728
 729/* Read/set MAC address routines */
 730static void r6040_mac_address(struct net_device *dev)
 731{
 732	struct r6040_private *lp = netdev_priv(dev);
 733	void __iomem *ioaddr = lp->base;
 734	const u16 *adrp;
 735
 736	/* Reset MAC */
 737	r6040_reset_mac(lp);
 738
 739	/* Restore MAC Address */
 740	adrp = (const u16 *) dev->dev_addr;
 741	iowrite16(adrp[0], ioaddr + MID_0L);
 742	iowrite16(adrp[1], ioaddr + MID_0M);
 743	iowrite16(adrp[2], ioaddr + MID_0H);
 
 
 
 744}
 745
 746static int r6040_open(struct net_device *dev)
 747{
 748	struct r6040_private *lp = netdev_priv(dev);
 749	int ret;
 750
 751	/* Request IRQ and Register interrupt handler */
 752	ret = request_irq(dev->irq, r6040_interrupt,
 753		IRQF_SHARED, dev->name, dev);
 754	if (ret)
 755		goto out;
 756
 757	/* Set MAC address */
 758	r6040_mac_address(dev);
 759
 760	/* Allocate Descriptor memory */
 761	lp->rx_ring =
 762		dma_alloc_coherent(&lp->pdev->dev, RX_DESC_SIZE,
 763				   &lp->rx_ring_dma, GFP_KERNEL);
 764	if (!lp->rx_ring) {
 765		ret = -ENOMEM;
 766		goto err_free_irq;
 767	}
 768
 769	lp->tx_ring =
 770		dma_alloc_coherent(&lp->pdev->dev, TX_DESC_SIZE,
 771				   &lp->tx_ring_dma, GFP_KERNEL);
 772	if (!lp->tx_ring) {
 773		ret = -ENOMEM;
 774		goto err_free_rx_ring;
 775	}
 776
 777	ret = r6040_up(dev);
 778	if (ret)
 779		goto err_free_tx_ring;
 780
 781	napi_enable(&lp->napi);
 782	netif_start_queue(dev);
 783
 784	return 0;
 785
 786err_free_tx_ring:
 787	dma_free_coherent(&lp->pdev->dev, TX_DESC_SIZE, lp->tx_ring,
 788			  lp->tx_ring_dma);
 789err_free_rx_ring:
 790	dma_free_coherent(&lp->pdev->dev, RX_DESC_SIZE, lp->rx_ring,
 791			  lp->rx_ring_dma);
 792err_free_irq:
 793	free_irq(dev->irq, dev);
 794out:
 795	return ret;
 796}
 797
 798static netdev_tx_t r6040_start_xmit(struct sk_buff *skb,
 799				    struct net_device *dev)
 800{
 801	struct r6040_private *lp = netdev_priv(dev);
 802	struct r6040_descriptor *descptr;
 803	void __iomem *ioaddr = lp->base;
 804	unsigned long flags;
 805
 806	if (skb_put_padto(skb, ETH_ZLEN) < 0)
 807		return NETDEV_TX_OK;
 808
 809	/* Critical Section */
 810	spin_lock_irqsave(&lp->lock, flags);
 811
 812	/* TX resource check */
 813	if (!lp->tx_free_desc) {
 814		spin_unlock_irqrestore(&lp->lock, flags);
 815		netif_stop_queue(dev);
 816		netdev_err(dev, ": no tx descriptor\n");
 817		return NETDEV_TX_BUSY;
 818	}
 819
 
 
 
 820	/* Set TX descriptor & Transmit it */
 821	lp->tx_free_desc--;
 822	descptr = lp->tx_insert_ptr;
 823	descptr->len = skb->len;
 
 
 
 
 824	descptr->skb_ptr = skb;
 825	descptr->buf = cpu_to_le32(dma_map_single(&lp->pdev->dev, skb->data,
 826						  skb->len, DMA_TO_DEVICE));
 827	descptr->status = DSC_OWNER_MAC;
 828
 829	skb_tx_timestamp(skb);
 830
 831	/* Trigger the MAC to check the TX descriptor */
 832	if (!netdev_xmit_more() || netif_queue_stopped(dev))
 833		iowrite16(TM2TX, ioaddr + MTPR);
 834	lp->tx_insert_ptr = descptr->vndescp;
 835
 836	/* If no tx resource, stop */
 837	if (!lp->tx_free_desc)
 838		netif_stop_queue(dev);
 839
 840	spin_unlock_irqrestore(&lp->lock, flags);
 841
 842	return NETDEV_TX_OK;
 843}
 844
 845static void r6040_multicast_list(struct net_device *dev)
 846{
 847	struct r6040_private *lp = netdev_priv(dev);
 848	void __iomem *ioaddr = lp->base;
 849	unsigned long flags;
 850	struct netdev_hw_addr *ha;
 851	int i;
 852	const u16 *adrp;
 853	u16 hash_table[4] = { 0 };
 854
 855	spin_lock_irqsave(&lp->lock, flags);
 856
 857	/* Keep our MAC Address */
 858	adrp = (const u16 *)dev->dev_addr;
 859	iowrite16(adrp[0], ioaddr + MID_0L);
 860	iowrite16(adrp[1], ioaddr + MID_0M);
 861	iowrite16(adrp[2], ioaddr + MID_0H);
 862
 863	/* Clear AMCP & PROM bits */
 864	lp->mcr0 = ioread16(ioaddr + MCR0) & ~(MCR0_PROMISC | MCR0_HASH_EN);
 865
 866	/* Promiscuous mode */
 867	if (dev->flags & IFF_PROMISC)
 868		lp->mcr0 |= MCR0_PROMISC;
 869
 870	/* Enable multicast hash table function to
 871	 * receive all multicast packets. */
 872	else if (dev->flags & IFF_ALLMULTI) {
 873		lp->mcr0 |= MCR0_HASH_EN;
 874
 875		for (i = 0; i < MCAST_MAX ; i++) {
 876			iowrite16(0, ioaddr + MID_1L + 8 * i);
 877			iowrite16(0, ioaddr + MID_1M + 8 * i);
 878			iowrite16(0, ioaddr + MID_1H + 8 * i);
 879		}
 880
 881		for (i = 0; i < 4; i++)
 882			hash_table[i] = 0xffff;
 883	}
 884	/* Use internal multicast address registers if the number of
 885	 * multicast addresses is not greater than MCAST_MAX. */
 886	else if (netdev_mc_count(dev) <= MCAST_MAX) {
 887		i = 0;
 888		netdev_for_each_mc_addr(ha, dev) {
 889			u16 *adrp = (u16 *) ha->addr;
 890			iowrite16(adrp[0], ioaddr + MID_1L + 8 * i);
 891			iowrite16(adrp[1], ioaddr + MID_1M + 8 * i);
 892			iowrite16(adrp[2], ioaddr + MID_1H + 8 * i);
 893			i++;
 894		}
 895		while (i < MCAST_MAX) {
 896			iowrite16(0, ioaddr + MID_1L + 8 * i);
 897			iowrite16(0, ioaddr + MID_1M + 8 * i);
 898			iowrite16(0, ioaddr + MID_1H + 8 * i);
 899			i++;
 900		}
 901	}
 902	/* Otherwise, Enable multicast hash table function. */
 903	else {
 904		u32 crc;
 905
 906		lp->mcr0 |= MCR0_HASH_EN;
 907
 908		for (i = 0; i < MCAST_MAX ; i++) {
 909			iowrite16(0, ioaddr + MID_1L + 8 * i);
 910			iowrite16(0, ioaddr + MID_1M + 8 * i);
 911			iowrite16(0, ioaddr + MID_1H + 8 * i);
 912		}
 913
 914		/* Build multicast hash table */
 915		netdev_for_each_mc_addr(ha, dev) {
 916			u8 *addrs = ha->addr;
 917
 918			crc = ether_crc(ETH_ALEN, addrs);
 919			crc >>= 26;
 920			hash_table[crc >> 4] |= 1 << (crc & 0xf);
 921		}
 922	}
 923
 924	iowrite16(lp->mcr0, ioaddr + MCR0);
 925
 926	/* Fill the MAC hash tables with their values */
 927	if (lp->mcr0 & MCR0_HASH_EN) {
 928		iowrite16(hash_table[0], ioaddr + MAR0);
 929		iowrite16(hash_table[1], ioaddr + MAR1);
 930		iowrite16(hash_table[2], ioaddr + MAR2);
 931		iowrite16(hash_table[3], ioaddr + MAR3);
 932	}
 933
 934	spin_unlock_irqrestore(&lp->lock, flags);
 935}
 936
 937static void netdev_get_drvinfo(struct net_device *dev,
 938			struct ethtool_drvinfo *info)
 939{
 940	struct r6040_private *rp = netdev_priv(dev);
 941
 942	strscpy(info->driver, DRV_NAME, sizeof(info->driver));
 943	strscpy(info->version, DRV_VERSION, sizeof(info->version));
 944	strscpy(info->bus_info, pci_name(rp->pdev), sizeof(info->bus_info));
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 945}
 946
 947static const struct ethtool_ops netdev_ethtool_ops = {
 948	.get_drvinfo		= netdev_get_drvinfo,
 
 
 949	.get_link		= ethtool_op_get_link,
 950	.get_ts_info		= ethtool_op_get_ts_info,
 951	.get_link_ksettings     = phy_ethtool_get_link_ksettings,
 952	.set_link_ksettings     = phy_ethtool_set_link_ksettings,
 953	.nway_reset		= phy_ethtool_nway_reset,
 954};
 955
 956static const struct net_device_ops r6040_netdev_ops = {
 957	.ndo_open		= r6040_open,
 958	.ndo_stop		= r6040_close,
 959	.ndo_start_xmit		= r6040_start_xmit,
 960	.ndo_get_stats		= r6040_get_stats,
 961	.ndo_set_rx_mode	= r6040_multicast_list,
 
 962	.ndo_validate_addr	= eth_validate_addr,
 963	.ndo_set_mac_address	= eth_mac_addr,
 964	.ndo_eth_ioctl		= phy_do_ioctl,
 965	.ndo_tx_timeout		= r6040_tx_timeout,
 966#ifdef CONFIG_NET_POLL_CONTROLLER
 967	.ndo_poll_controller	= r6040_poll_controller,
 968#endif
 969};
 970
 971static void r6040_adjust_link(struct net_device *dev)
 972{
 973	struct r6040_private *lp = netdev_priv(dev);
 974	struct phy_device *phydev = dev->phydev;
 975	int status_changed = 0;
 976	void __iomem *ioaddr = lp->base;
 977
 978	BUG_ON(!phydev);
 979
 980	if (lp->old_link != phydev->link) {
 981		status_changed = 1;
 982		lp->old_link = phydev->link;
 983	}
 984
 985	/* reflect duplex change */
 986	if (phydev->link && (lp->old_duplex != phydev->duplex)) {
 987		lp->mcr0 |= (phydev->duplex == DUPLEX_FULL ? MCR0_FD : 0);
 988		iowrite16(lp->mcr0, ioaddr);
 989
 990		status_changed = 1;
 991		lp->old_duplex = phydev->duplex;
 992	}
 993
 994	if (status_changed)
 995		phy_print_status(phydev);
 
 
 
 
 
 
 996}
 997
 998static int r6040_mii_probe(struct net_device *dev)
 999{
1000	struct r6040_private *lp = netdev_priv(dev);
1001	struct phy_device *phydev = NULL;
1002
1003	phydev = phy_find_first(lp->mii_bus);
1004	if (!phydev) {
1005		dev_err(&lp->pdev->dev, "no PHY found\n");
1006		return -ENODEV;
1007	}
1008
1009	phydev = phy_connect(dev, phydev_name(phydev), &r6040_adjust_link,
1010			     PHY_INTERFACE_MODE_MII);
1011
1012	if (IS_ERR(phydev)) {
1013		dev_err(&lp->pdev->dev, "could not attach to PHY\n");
1014		return PTR_ERR(phydev);
1015	}
1016
1017	phy_set_max_speed(phydev, SPEED_100);
 
 
 
 
 
 
 
1018
 
 
1019	lp->old_link = 0;
1020	lp->old_duplex = -1;
1021
1022	phy_attached_info(phydev);
 
 
1023
1024	return 0;
1025}
1026
1027static int r6040_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
 
1028{
1029	struct net_device *dev;
1030	struct r6040_private *lp;
1031	void __iomem *ioaddr;
1032	int err, io_size = R6040_IO_SIZE;
1033	static int card_idx = -1;
1034	u16 addr[ETH_ALEN / 2];
1035	int bar = 0;
 
 
1036
1037	pr_info("%s\n", version);
1038
1039	err = pci_enable_device(pdev);
1040	if (err)
1041		goto err_out;
1042
1043	/* this should always be supported */
1044	err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
1045	if (err) {
1046		dev_err(&pdev->dev, "32-bit PCI DMA addresses not supported by the card\n");
 
1047		goto err_out_disable_dev;
1048	}
1049	err = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
1050	if (err) {
1051		dev_err(&pdev->dev, "32-bit PCI DMA addresses not supported by the card\n");
 
1052		goto err_out_disable_dev;
1053	}
1054
1055	/* IO Size check */
1056	if (pci_resource_len(pdev, bar) < io_size) {
1057		dev_err(&pdev->dev, "Insufficient PCI resources, aborting\n");
1058		err = -EIO;
1059		goto err_out_disable_dev;
1060	}
1061
1062	pci_set_master(pdev);
1063
1064	dev = alloc_etherdev(sizeof(struct r6040_private));
1065	if (!dev) {
1066		err = -ENOMEM;
1067		goto err_out_disable_dev;
1068	}
1069	SET_NETDEV_DEV(dev, &pdev->dev);
1070	lp = netdev_priv(dev);
1071
1072	err = pci_request_regions(pdev, DRV_NAME);
1073
1074	if (err) {
1075		dev_err(&pdev->dev, "Failed to request PCI regions\n");
1076		goto err_out_free_dev;
1077	}
1078
1079	ioaddr = pci_iomap(pdev, bar, io_size);
1080	if (!ioaddr) {
1081		dev_err(&pdev->dev, "ioremap failed for device\n");
1082		err = -EIO;
1083		goto err_out_free_res;
1084	}
1085
1086	/* If PHY status change register is still set to zero it means the
1087	 * bootloader didn't initialize it, so we set it to:
1088	 * - enable phy status change
1089	 * - enable all phy addresses
1090	 * - set to lowest timer divider */
1091	if (ioread16(ioaddr + PHY_CC) == 0)
1092		iowrite16(SCEN | PHY_MAX_ADDR << PHYAD_SHIFT |
1093				7 << TMRDIV_SHIFT, ioaddr + PHY_CC);
1094
1095	/* Init system & device */
1096	lp->base = ioaddr;
1097	dev->irq = pdev->irq;
1098
1099	spin_lock_init(&lp->lock);
1100	pci_set_drvdata(pdev, dev);
1101
1102	/* Set MAC address */
1103	card_idx++;
1104
1105	addr[0] = ioread16(ioaddr + MID_0L);
1106	addr[1] = ioread16(ioaddr + MID_0M);
1107	addr[2] = ioread16(ioaddr + MID_0H);
1108	eth_hw_addr_set(dev, (u8 *)addr);
1109
1110	/* Some bootloader/BIOSes do not initialize
1111	 * MAC address, warn about that */
1112	if (!(addr[0] || addr[1] || addr[2])) {
1113		netdev_warn(dev, "MAC address not initialized, "
1114					"generating random\n");
1115		eth_hw_addr_random(dev);
1116	}
1117
1118	/* Link new device into r6040_root_dev */
1119	lp->pdev = pdev;
1120	lp->dev = dev;
1121
1122	/* Init RDC private data */
1123	lp->mcr0 = MCR0_XMTEN | MCR0_RCVEN;
1124
1125	/* The RDC-specific entries in the device structure. */
1126	dev->netdev_ops = &r6040_netdev_ops;
1127	dev->ethtool_ops = &netdev_ethtool_ops;
1128	dev->watchdog_timeo = TX_TIMEOUT;
1129
1130	netif_napi_add(dev, &lp->napi, r6040_poll);
1131
1132	lp->mii_bus = mdiobus_alloc();
1133	if (!lp->mii_bus) {
1134		dev_err(&pdev->dev, "mdiobus_alloc() failed\n");
1135		err = -ENOMEM;
1136		goto err_out_unmap;
1137	}
1138
1139	lp->mii_bus->priv = dev;
1140	lp->mii_bus->read = r6040_mdiobus_read;
1141	lp->mii_bus->write = r6040_mdiobus_write;
 
1142	lp->mii_bus->name = "r6040_eth_mii";
1143	snprintf(lp->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
1144		dev_name(&pdev->dev), card_idx);
 
 
 
 
 
 
 
 
 
1145
1146	err = mdiobus_register(lp->mii_bus);
1147	if (err) {
1148		dev_err(&pdev->dev, "failed to register MII bus\n");
1149		goto err_out_mdio;
1150	}
1151
1152	err = r6040_mii_probe(dev);
1153	if (err) {
1154		dev_err(&pdev->dev, "failed to probe MII bus\n");
1155		goto err_out_mdio_unregister;
1156	}
1157
1158	/* Register net device. After this dev->name assign */
1159	err = register_netdev(dev);
1160	if (err) {
1161		dev_err(&pdev->dev, "Failed to register net device\n");
1162		goto err_out_phy_disconnect;
1163	}
1164	return 0;
1165
1166err_out_phy_disconnect:
1167	phy_disconnect(dev->phydev);
1168err_out_mdio_unregister:
1169	mdiobus_unregister(lp->mii_bus);
 
 
1170err_out_mdio:
1171	mdiobus_free(lp->mii_bus);
1172err_out_unmap:
1173	netif_napi_del(&lp->napi);
 
1174	pci_iounmap(pdev, ioaddr);
1175err_out_free_res:
1176	pci_release_regions(pdev);
1177err_out_free_dev:
1178	free_netdev(dev);
1179err_out_disable_dev:
1180	pci_disable_device(pdev);
1181err_out:
1182	return err;
1183}
1184
1185static void r6040_remove_one(struct pci_dev *pdev)
1186{
1187	struct net_device *dev = pci_get_drvdata(pdev);
1188	struct r6040_private *lp = netdev_priv(dev);
1189
1190	unregister_netdev(dev);
1191	phy_disconnect(dev->phydev);
1192	mdiobus_unregister(lp->mii_bus);
 
1193	mdiobus_free(lp->mii_bus);
1194	netif_napi_del(&lp->napi);
 
1195	pci_iounmap(pdev, lp->base);
1196	pci_release_regions(pdev);
1197	free_netdev(dev);
1198	pci_disable_device(pdev);
 
1199}
1200
1201
1202static const struct pci_device_id r6040_pci_tbl[] = {
1203	{ PCI_DEVICE(PCI_VENDOR_ID_RDC, 0x6040) },
1204	{ 0 }
1205};
1206MODULE_DEVICE_TABLE(pci, r6040_pci_tbl);
1207
1208static struct pci_driver r6040_driver = {
1209	.name		= DRV_NAME,
1210	.id_table	= r6040_pci_tbl,
1211	.probe		= r6040_init_one,
1212	.remove		= r6040_remove_one,
1213};
1214
1215module_pci_driver(r6040_driver);
 
 
 
 
 
 
 
 
 
 
 
 
 
v3.5.6
 
   1/*
   2 * RDC R6040 Fast Ethernet MAC support
   3 *
   4 * Copyright (C) 2004 Sten Wang <sten.wang@rdc.com.tw>
   5 * Copyright (C) 2007
   6 *	Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>
   7 * Copyright (C) 2007-2012 Florian Fainelli <florian@openwrt.org>
   8 *
   9 * This program is free software; you can redistribute it and/or
  10 * modify it under the terms of the GNU General Public License
  11 * as published by the Free Software Foundation; either version 2
  12 * of the License, or (at your option) any later version.
  13 *
  14 * This program is distributed in the hope that it will be useful,
  15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
  16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
  17 * GNU General Public License for more details.
  18 *
  19 * You should have received a copy of the GNU General Public License
  20 * along with this program; if not, write to the
  21 * Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
  22 * Boston, MA  02110-1301, USA.
  23*/
  24
  25#include <linux/kernel.h>
  26#include <linux/module.h>
  27#include <linux/moduleparam.h>
  28#include <linux/string.h>
  29#include <linux/timer.h>
  30#include <linux/errno.h>
  31#include <linux/ioport.h>
  32#include <linux/interrupt.h>
  33#include <linux/pci.h>
  34#include <linux/netdevice.h>
  35#include <linux/etherdevice.h>
  36#include <linux/skbuff.h>
  37#include <linux/init.h>
  38#include <linux/delay.h>
  39#include <linux/mii.h>
  40#include <linux/ethtool.h>
  41#include <linux/crc32.h>
  42#include <linux/spinlock.h>
  43#include <linux/bitops.h>
  44#include <linux/io.h>
  45#include <linux/irq.h>
  46#include <linux/uaccess.h>
  47#include <linux/phy.h>
  48
  49#include <asm/processor.h>
  50
  51#define DRV_NAME	"r6040"
  52#define DRV_VERSION	"0.28"
  53#define DRV_RELDATE	"07Oct2011"
  54
  55/* Time in jiffies before concluding the transmitter is hung. */
  56#define TX_TIMEOUT	(6000 * HZ / 1000)
  57
  58/* RDC MAC I/O Size */
  59#define R6040_IO_SIZE	256
  60
  61/* MAX RDC MAC */
  62#define MAX_MAC		2
  63
  64/* MAC registers */
  65#define MCR0		0x00	/* Control register 0 */
  66#define  MCR0_RCVEN	0x0002	/* Receive enable */
  67#define  MCR0_PROMISC	0x0020	/* Promiscuous mode */
  68#define  MCR0_HASH_EN	0x0100	/* Enable multicast hash table function */
  69#define  MCR0_XMTEN	0x1000	/* Transmission enable */
  70#define  MCR0_FD	0x8000	/* Full/Half duplex */
  71#define MCR1		0x04	/* Control register 1 */
  72#define  MAC_RST	0x0001	/* Reset the MAC */
  73#define MBCR		0x08	/* Bus control */
  74#define MT_ICR		0x0C	/* TX interrupt control */
  75#define MR_ICR		0x10	/* RX interrupt control */
  76#define MTPR		0x14	/* TX poll command register */
  77#define  TM2TX		0x0001	/* Trigger MAC to transmit */
  78#define MR_BSR		0x18	/* RX buffer size */
  79#define MR_DCR		0x1A	/* RX descriptor control */
  80#define MLSR		0x1C	/* Last status */
  81#define  TX_FIFO_UNDR	0x0200	/* TX FIFO under-run */
  82#define	 TX_EXCEEDC	0x2000	/* Transmit exceed collision */
  83#define  TX_LATEC	0x4000	/* Transmit late collision */
  84#define MMDIO		0x20	/* MDIO control register */
  85#define  MDIO_WRITE	0x4000	/* MDIO write */
  86#define  MDIO_READ	0x2000	/* MDIO read */
  87#define MMRD		0x24	/* MDIO read data register */
  88#define MMWD		0x28	/* MDIO write data register */
  89#define MTD_SA0		0x2C	/* TX descriptor start address 0 */
  90#define MTD_SA1		0x30	/* TX descriptor start address 1 */
  91#define MRD_SA0		0x34	/* RX descriptor start address 0 */
  92#define MRD_SA1		0x38	/* RX descriptor start address 1 */
  93#define MISR		0x3C	/* Status register */
  94#define MIER		0x40	/* INT enable register */
  95#define  MSK_INT	0x0000	/* Mask off interrupts */
  96#define  RX_FINISH	0x0001  /* RX finished */
  97#define  RX_NO_DESC	0x0002  /* No RX descriptor available */
  98#define  RX_FIFO_FULL	0x0004  /* RX FIFO full */
  99#define  RX_EARLY	0x0008  /* RX early */
 100#define  TX_FINISH	0x0010  /* TX finished */
 101#define  TX_EARLY	0x0080  /* TX early */
 102#define  EVENT_OVRFL	0x0100  /* Event counter overflow */
 103#define  LINK_CHANGED	0x0200  /* PHY link changed */
 104#define ME_CISR		0x44	/* Event counter INT status */
 105#define ME_CIER		0x48	/* Event counter INT enable  */
 106#define MR_CNT		0x50	/* Successfully received packet counter */
 107#define ME_CNT0		0x52	/* Event counter 0 */
 108#define ME_CNT1		0x54	/* Event counter 1 */
 109#define ME_CNT2		0x56	/* Event counter 2 */
 110#define ME_CNT3		0x58	/* Event counter 3 */
 111#define MT_CNT		0x5A	/* Successfully transmit packet counter */
 112#define ME_CNT4		0x5C	/* Event counter 4 */
 113#define MP_CNT		0x5E	/* Pause frame counter register */
 114#define MAR0		0x60	/* Hash table 0 */
 115#define MAR1		0x62	/* Hash table 1 */
 116#define MAR2		0x64	/* Hash table 2 */
 117#define MAR3		0x66	/* Hash table 3 */
 118#define MID_0L		0x68	/* Multicast address MID0 Low */
 119#define MID_0M		0x6A	/* Multicast address MID0 Medium */
 120#define MID_0H		0x6C	/* Multicast address MID0 High */
 121#define MID_1L		0x70	/* MID1 Low */
 122#define MID_1M		0x72	/* MID1 Medium */
 123#define MID_1H		0x74	/* MID1 High */
 124#define MID_2L		0x78	/* MID2 Low */
 125#define MID_2M		0x7A	/* MID2 Medium */
 126#define MID_2H		0x7C	/* MID2 High */
 127#define MID_3L		0x80	/* MID3 Low */
 128#define MID_3M		0x82	/* MID3 Medium */
 129#define MID_3H		0x84	/* MID3 High */
 130#define PHY_CC		0x88	/* PHY status change configuration register */
 131#define  SCEN		0x8000	/* PHY status change enable */
 132#define  PHYAD_SHIFT	8	/* PHY address shift */
 133#define  TMRDIV_SHIFT	0	/* Timer divider shift */
 134#define PHY_ST		0x8A	/* PHY status register */
 135#define MAC_SM		0xAC	/* MAC status machine */
 136#define  MAC_SM_RST	0x0002	/* MAC status machine reset */
 
 
 137#define MAC_ID		0xBE	/* Identifier register */
 138
 139#define TX_DCNT		0x80	/* TX descriptor count */
 140#define RX_DCNT		0x80	/* RX descriptor count */
 141#define MAX_BUF_SIZE	0x600
 142#define RX_DESC_SIZE	(RX_DCNT * sizeof(struct r6040_descriptor))
 143#define TX_DESC_SIZE	(TX_DCNT * sizeof(struct r6040_descriptor))
 144#define MBCR_DEFAULT	0x012A	/* MAC Bus Control Register */
 145#define MCAST_MAX	3	/* Max number multicast addresses to filter */
 146
 147#define MAC_DEF_TIMEOUT	2048	/* Default MAC read/write operation timeout */
 148
 149/* Descriptor status */
 150#define DSC_OWNER_MAC	0x8000	/* MAC is the owner of this descriptor */
 151#define DSC_RX_OK	0x4000	/* RX was successful */
 152#define DSC_RX_ERR	0x0800	/* RX PHY error */
 153#define DSC_RX_ERR_DRI	0x0400	/* RX dribble packet */
 154#define DSC_RX_ERR_BUF	0x0200	/* RX length exceeds buffer size */
 155#define DSC_RX_ERR_LONG	0x0100	/* RX length > maximum packet length */
 156#define DSC_RX_ERR_RUNT	0x0080	/* RX packet length < 64 byte */
 157#define DSC_RX_ERR_CRC	0x0040	/* RX CRC error */
 158#define DSC_RX_BCAST	0x0020	/* RX broadcast (no error) */
 159#define DSC_RX_MCAST	0x0010	/* RX multicast (no error) */
 160#define DSC_RX_MCH_HIT	0x0008	/* RX multicast hit in hash table (no error) */
 161#define DSC_RX_MIDH_HIT	0x0004	/* RX MID table hit (no error) */
 162#define DSC_RX_IDX_MID_MASK 3	/* RX mask for the index of matched MIDx */
 163
 164MODULE_AUTHOR("Sten Wang <sten.wang@rdc.com.tw>,"
 165	"Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>,"
 166	"Florian Fainelli <florian@openwrt.org>");
 167MODULE_LICENSE("GPL");
 168MODULE_DESCRIPTION("RDC R6040 NAPI PCI FastEthernet driver");
 169MODULE_VERSION(DRV_VERSION " " DRV_RELDATE);
 170
 171/* RX and TX interrupts that we handle */
 172#define RX_INTS			(RX_FIFO_FULL | RX_NO_DESC | RX_FINISH)
 173#define TX_INTS			(TX_FINISH)
 174#define INT_MASK		(RX_INTS | TX_INTS)
 175
 176struct r6040_descriptor {
 177	u16	status, len;		/* 0-3 */
 178	__le32	buf;			/* 4-7 */
 179	__le32	ndesc;			/* 8-B */
 180	u32	rev1;			/* C-F */
 181	char	*vbufp;			/* 10-13 */
 182	struct r6040_descriptor *vndescp;	/* 14-17 */
 183	struct sk_buff *skb_ptr;	/* 18-1B */
 184	u32	rev2;			/* 1C-1F */
 185} __aligned(32);
 186
 187struct r6040_private {
 188	spinlock_t lock;		/* driver lock */
 189	struct pci_dev *pdev;
 190	struct r6040_descriptor *rx_insert_ptr;
 191	struct r6040_descriptor *rx_remove_ptr;
 192	struct r6040_descriptor *tx_insert_ptr;
 193	struct r6040_descriptor *tx_remove_ptr;
 194	struct r6040_descriptor *rx_ring;
 195	struct r6040_descriptor *tx_ring;
 196	dma_addr_t rx_ring_dma;
 197	dma_addr_t tx_ring_dma;
 198	u16	tx_free_desc;
 199	u16	mcr0;
 200	struct net_device *dev;
 201	struct mii_bus *mii_bus;
 202	struct napi_struct napi;
 203	void __iomem *base;
 204	struct phy_device *phydev;
 205	int old_link;
 206	int old_duplex;
 207};
 208
 209static char version[] __devinitdata = DRV_NAME
 210	": RDC R6040 NAPI net driver,"
 211	"version "DRV_VERSION " (" DRV_RELDATE ")";
 212
 213/* Read a word data from PHY Chip */
 214static int r6040_phy_read(void __iomem *ioaddr, int phy_addr, int reg)
 215{
 216	int limit = MAC_DEF_TIMEOUT;
 217	u16 cmd;
 218
 219	iowrite16(MDIO_READ + reg + (phy_addr << 8), ioaddr + MMDIO);
 220	/* Wait for the read bit to be cleared */
 221	while (limit--) {
 222		cmd = ioread16(ioaddr + MMDIO);
 223		if (!(cmd & MDIO_READ))
 224			break;
 
 225	}
 226
 
 
 
 227	return ioread16(ioaddr + MMRD);
 228}
 229
 230/* Write a word data from PHY Chip */
 231static void r6040_phy_write(void __iomem *ioaddr,
 232					int phy_addr, int reg, u16 val)
 233{
 234	int limit = MAC_DEF_TIMEOUT;
 235	u16 cmd;
 236
 237	iowrite16(val, ioaddr + MMWD);
 238	/* Write the command to the MDIO bus */
 239	iowrite16(MDIO_WRITE + reg + (phy_addr << 8), ioaddr + MMDIO);
 240	/* Wait for the write bit to be cleared */
 241	while (limit--) {
 242		cmd = ioread16(ioaddr + MMDIO);
 243		if (!(cmd & MDIO_WRITE))
 244			break;
 
 245	}
 
 
 246}
 247
 248static int r6040_mdiobus_read(struct mii_bus *bus, int phy_addr, int reg)
 249{
 250	struct net_device *dev = bus->priv;
 251	struct r6040_private *lp = netdev_priv(dev);
 252	void __iomem *ioaddr = lp->base;
 253
 254	return r6040_phy_read(ioaddr, phy_addr, reg);
 255}
 256
 257static int r6040_mdiobus_write(struct mii_bus *bus, int phy_addr,
 258						int reg, u16 value)
 259{
 260	struct net_device *dev = bus->priv;
 261	struct r6040_private *lp = netdev_priv(dev);
 262	void __iomem *ioaddr = lp->base;
 263
 264	r6040_phy_write(ioaddr, phy_addr, reg, value);
 265
 266	return 0;
 267}
 268
 269static int r6040_mdiobus_reset(struct mii_bus *bus)
 270{
 271	return 0;
 272}
 273
 274static void r6040_free_txbufs(struct net_device *dev)
 275{
 276	struct r6040_private *lp = netdev_priv(dev);
 277	int i;
 278
 279	for (i = 0; i < TX_DCNT; i++) {
 280		if (lp->tx_insert_ptr->skb_ptr) {
 281			pci_unmap_single(lp->pdev,
 282				le32_to_cpu(lp->tx_insert_ptr->buf),
 283				MAX_BUF_SIZE, PCI_DMA_TODEVICE);
 284			dev_kfree_skb(lp->tx_insert_ptr->skb_ptr);
 285			lp->tx_insert_ptr->skb_ptr = NULL;
 286		}
 287		lp->tx_insert_ptr = lp->tx_insert_ptr->vndescp;
 288	}
 289}
 290
 291static void r6040_free_rxbufs(struct net_device *dev)
 292{
 293	struct r6040_private *lp = netdev_priv(dev);
 294	int i;
 295
 296	for (i = 0; i < RX_DCNT; i++) {
 297		if (lp->rx_insert_ptr->skb_ptr) {
 298			pci_unmap_single(lp->pdev,
 299				le32_to_cpu(lp->rx_insert_ptr->buf),
 300				MAX_BUF_SIZE, PCI_DMA_FROMDEVICE);
 301			dev_kfree_skb(lp->rx_insert_ptr->skb_ptr);
 302			lp->rx_insert_ptr->skb_ptr = NULL;
 303		}
 304		lp->rx_insert_ptr = lp->rx_insert_ptr->vndescp;
 305	}
 306}
 307
 308static void r6040_init_ring_desc(struct r6040_descriptor *desc_ring,
 309				 dma_addr_t desc_dma, int size)
 310{
 311	struct r6040_descriptor *desc = desc_ring;
 312	dma_addr_t mapping = desc_dma;
 313
 314	while (size-- > 0) {
 315		mapping += sizeof(*desc);
 316		desc->ndesc = cpu_to_le32(mapping);
 317		desc->vndescp = desc + 1;
 318		desc++;
 319	}
 320	desc--;
 321	desc->ndesc = cpu_to_le32(desc_dma);
 322	desc->vndescp = desc_ring;
 323}
 324
 325static void r6040_init_txbufs(struct net_device *dev)
 326{
 327	struct r6040_private *lp = netdev_priv(dev);
 328
 329	lp->tx_free_desc = TX_DCNT;
 330
 331	lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring;
 332	r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT);
 333}
 334
 335static int r6040_alloc_rxbufs(struct net_device *dev)
 336{
 337	struct r6040_private *lp = netdev_priv(dev);
 338	struct r6040_descriptor *desc;
 339	struct sk_buff *skb;
 340	int rc;
 341
 342	lp->rx_remove_ptr = lp->rx_insert_ptr = lp->rx_ring;
 343	r6040_init_ring_desc(lp->rx_ring, lp->rx_ring_dma, RX_DCNT);
 344
 345	/* Allocate skbs for the rx descriptors */
 346	desc = lp->rx_ring;
 347	do {
 348		skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
 349		if (!skb) {
 350			netdev_err(dev, "failed to alloc skb for rx\n");
 351			rc = -ENOMEM;
 352			goto err_exit;
 353		}
 354		desc->skb_ptr = skb;
 355		desc->buf = cpu_to_le32(pci_map_single(lp->pdev,
 356					desc->skb_ptr->data,
 357					MAX_BUF_SIZE, PCI_DMA_FROMDEVICE));
 
 358		desc->status = DSC_OWNER_MAC;
 359		desc = desc->vndescp;
 360	} while (desc != lp->rx_ring);
 361
 362	return 0;
 363
 364err_exit:
 365	/* Deallocate all previously allocated skbs */
 366	r6040_free_rxbufs(dev);
 367	return rc;
 368}
 369
 370static void r6040_reset_mac(struct r6040_private *lp)
 371{
 372	void __iomem *ioaddr = lp->base;
 373	int limit = MAC_DEF_TIMEOUT;
 374	u16 cmd;
 375
 
 376	iowrite16(MAC_RST, ioaddr + MCR1);
 377	while (limit--) {
 378		cmd = ioread16(ioaddr + MCR1);
 379		if (cmd & MAC_RST)
 380			break;
 381	}
 382
 383	/* Reset internal state machine */
 384	iowrite16(MAC_SM_RST, ioaddr + MAC_SM);
 385	iowrite16(0, ioaddr + MAC_SM);
 386	mdelay(5);
 
 
 
 
 387}
 388
 389static void r6040_init_mac_regs(struct net_device *dev)
 390{
 391	struct r6040_private *lp = netdev_priv(dev);
 392	void __iomem *ioaddr = lp->base;
 393
 394	/* Mask Off Interrupt */
 395	iowrite16(MSK_INT, ioaddr + MIER);
 396
 397	/* Reset RDC MAC */
 398	r6040_reset_mac(lp);
 399
 400	/* MAC Bus Control Register */
 401	iowrite16(MBCR_DEFAULT, ioaddr + MBCR);
 402
 403	/* Buffer Size Register */
 404	iowrite16(MAX_BUF_SIZE, ioaddr + MR_BSR);
 405
 406	/* Write TX ring start address */
 407	iowrite16(lp->tx_ring_dma, ioaddr + MTD_SA0);
 408	iowrite16(lp->tx_ring_dma >> 16, ioaddr + MTD_SA1);
 409
 410	/* Write RX ring start address */
 411	iowrite16(lp->rx_ring_dma, ioaddr + MRD_SA0);
 412	iowrite16(lp->rx_ring_dma >> 16, ioaddr + MRD_SA1);
 413
 414	/* Set interrupt waiting time and packet numbers */
 415	iowrite16(0, ioaddr + MT_ICR);
 416	iowrite16(0, ioaddr + MR_ICR);
 417
 418	/* Enable interrupts */
 419	iowrite16(INT_MASK, ioaddr + MIER);
 420
 421	/* Enable TX and RX */
 422	iowrite16(lp->mcr0 | MCR0_RCVEN, ioaddr);
 423
 424	/* Let TX poll the descriptors
 425	 * we may got called by r6040_tx_timeout which has left
 426	 * some unsent tx buffers */
 427	iowrite16(TM2TX, ioaddr + MTPR);
 428}
 429
 430static void r6040_tx_timeout(struct net_device *dev)
 431{
 432	struct r6040_private *priv = netdev_priv(dev);
 433	void __iomem *ioaddr = priv->base;
 434
 435	netdev_warn(dev, "transmit timed out, int enable %4.4x "
 436		"status %4.4x\n",
 437		ioread16(ioaddr + MIER),
 438		ioread16(ioaddr + MISR));
 439
 440	dev->stats.tx_errors++;
 441
 442	/* Reset MAC and re-init all registers */
 443	r6040_init_mac_regs(dev);
 444}
 445
 446static struct net_device_stats *r6040_get_stats(struct net_device *dev)
 447{
 448	struct r6040_private *priv = netdev_priv(dev);
 449	void __iomem *ioaddr = priv->base;
 450	unsigned long flags;
 451
 452	spin_lock_irqsave(&priv->lock, flags);
 453	dev->stats.rx_crc_errors += ioread8(ioaddr + ME_CNT1);
 454	dev->stats.multicast += ioread8(ioaddr + ME_CNT0);
 455	spin_unlock_irqrestore(&priv->lock, flags);
 456
 457	return &dev->stats;
 458}
 459
 460/* Stop RDC MAC and Free the allocated resource */
 461static void r6040_down(struct net_device *dev)
 462{
 463	struct r6040_private *lp = netdev_priv(dev);
 464	void __iomem *ioaddr = lp->base;
 465	u16 *adrp;
 466
 467	/* Stop MAC */
 468	iowrite16(MSK_INT, ioaddr + MIER);	/* Mask Off Interrupt */
 469
 470	/* Reset RDC MAC */
 471	r6040_reset_mac(lp);
 472
 473	/* Restore MAC Address to MIDx */
 474	adrp = (u16 *) dev->dev_addr;
 475	iowrite16(adrp[0], ioaddr + MID_0L);
 476	iowrite16(adrp[1], ioaddr + MID_0M);
 477	iowrite16(adrp[2], ioaddr + MID_0H);
 478
 479	phy_stop(lp->phydev);
 480}
 481
 482static int r6040_close(struct net_device *dev)
 483{
 484	struct r6040_private *lp = netdev_priv(dev);
 485	struct pci_dev *pdev = lp->pdev;
 486
 487	spin_lock_irq(&lp->lock);
 488	napi_disable(&lp->napi);
 489	netif_stop_queue(dev);
 
 
 490	r6040_down(dev);
 491
 492	free_irq(dev->irq, dev);
 493
 494	/* Free RX buffer */
 495	r6040_free_rxbufs(dev);
 496
 497	/* Free TX buffer */
 498	r6040_free_txbufs(dev);
 499
 500	spin_unlock_irq(&lp->lock);
 501
 
 
 502	/* Free Descriptor memory */
 503	if (lp->rx_ring) {
 504		pci_free_consistent(pdev,
 505				RX_DESC_SIZE, lp->rx_ring, lp->rx_ring_dma);
 506		lp->rx_ring = NULL;
 507	}
 508
 509	if (lp->tx_ring) {
 510		pci_free_consistent(pdev,
 511				TX_DESC_SIZE, lp->tx_ring, lp->tx_ring_dma);
 512		lp->tx_ring = NULL;
 513	}
 514
 515	return 0;
 516}
 517
 518static int r6040_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
 519{
 520	struct r6040_private *lp = netdev_priv(dev);
 521
 522	if (!lp->phydev)
 523		return -EINVAL;
 524
 525	return phy_mii_ioctl(lp->phydev, rq, cmd);
 526}
 527
 528static int r6040_rx(struct net_device *dev, int limit)
 529{
 530	struct r6040_private *priv = netdev_priv(dev);
 531	struct r6040_descriptor *descptr = priv->rx_remove_ptr;
 532	struct sk_buff *skb_ptr, *new_skb;
 533	int count = 0;
 534	u16 err;
 535
 536	/* Limit not reached and the descriptor belongs to the CPU */
 537	while (count < limit && !(descptr->status & DSC_OWNER_MAC)) {
 538		/* Read the descriptor status */
 539		err = descptr->status;
 540		/* Global error status set */
 541		if (err & DSC_RX_ERR) {
 542			/* RX dribble */
 543			if (err & DSC_RX_ERR_DRI)
 544				dev->stats.rx_frame_errors++;
 545			/* Buffer length exceeded */
 546			if (err & DSC_RX_ERR_BUF)
 547				dev->stats.rx_length_errors++;
 548			/* Packet too long */
 549			if (err & DSC_RX_ERR_LONG)
 550				dev->stats.rx_length_errors++;
 551			/* Packet < 64 bytes */
 552			if (err & DSC_RX_ERR_RUNT)
 553				dev->stats.rx_length_errors++;
 554			/* CRC error */
 555			if (err & DSC_RX_ERR_CRC) {
 556				spin_lock(&priv->lock);
 557				dev->stats.rx_crc_errors++;
 558				spin_unlock(&priv->lock);
 559			}
 560			goto next_descr;
 561		}
 562
 563		/* Packet successfully received */
 564		new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
 565		if (!new_skb) {
 566			dev->stats.rx_dropped++;
 567			goto next_descr;
 568		}
 569		skb_ptr = descptr->skb_ptr;
 570		skb_ptr->dev = priv->dev;
 571
 572		/* Do not count the CRC */
 573		skb_put(skb_ptr, descptr->len - 4);
 574		pci_unmap_single(priv->pdev, le32_to_cpu(descptr->buf),
 575					MAX_BUF_SIZE, PCI_DMA_FROMDEVICE);
 576		skb_ptr->protocol = eth_type_trans(skb_ptr, priv->dev);
 577
 578		/* Send to upper layer */
 579		netif_receive_skb(skb_ptr);
 580		dev->stats.rx_packets++;
 581		dev->stats.rx_bytes += descptr->len - 4;
 582
 583		/* put new skb into descriptor */
 584		descptr->skb_ptr = new_skb;
 585		descptr->buf = cpu_to_le32(pci_map_single(priv->pdev,
 586						descptr->skb_ptr->data,
 587					MAX_BUF_SIZE, PCI_DMA_FROMDEVICE));
 
 588
 589next_descr:
 590		/* put the descriptor back to the MAC */
 591		descptr->status = DSC_OWNER_MAC;
 592		descptr = descptr->vndescp;
 593		count++;
 594	}
 595	priv->rx_remove_ptr = descptr;
 596
 597	return count;
 598}
 599
 600static void r6040_tx(struct net_device *dev)
 601{
 602	struct r6040_private *priv = netdev_priv(dev);
 603	struct r6040_descriptor *descptr;
 604	void __iomem *ioaddr = priv->base;
 605	struct sk_buff *skb_ptr;
 606	u16 err;
 607
 608	spin_lock(&priv->lock);
 609	descptr = priv->tx_remove_ptr;
 610	while (priv->tx_free_desc < TX_DCNT) {
 611		/* Check for errors */
 612		err = ioread16(ioaddr + MLSR);
 613
 614		if (err & TX_FIFO_UNDR)
 615			dev->stats.tx_fifo_errors++;
 616		if (err & (TX_EXCEEDC | TX_LATEC))
 617			dev->stats.tx_carrier_errors++;
 618
 619		if (descptr->status & DSC_OWNER_MAC)
 620			break; /* Not complete */
 621		skb_ptr = descptr->skb_ptr;
 622		pci_unmap_single(priv->pdev, le32_to_cpu(descptr->buf),
 623			skb_ptr->len, PCI_DMA_TODEVICE);
 
 
 
 
 
 624		/* Free buffer */
 625		dev_kfree_skb_irq(skb_ptr);
 626		descptr->skb_ptr = NULL;
 627		/* To next descriptor */
 628		descptr = descptr->vndescp;
 629		priv->tx_free_desc++;
 630	}
 631	priv->tx_remove_ptr = descptr;
 632
 633	if (priv->tx_free_desc)
 634		netif_wake_queue(dev);
 635	spin_unlock(&priv->lock);
 636}
 637
 638static int r6040_poll(struct napi_struct *napi, int budget)
 639{
 640	struct r6040_private *priv =
 641		container_of(napi, struct r6040_private, napi);
 642	struct net_device *dev = priv->dev;
 643	void __iomem *ioaddr = priv->base;
 644	int work_done;
 645
 
 
 646	work_done = r6040_rx(dev, budget);
 647
 648	if (work_done < budget) {
 649		napi_complete(napi);
 650		/* Enable RX interrupt */
 651		iowrite16(ioread16(ioaddr + MIER) | RX_INTS, ioaddr + MIER);
 
 652	}
 653	return work_done;
 654}
 655
 656/* The RDC interrupt handler. */
 657static irqreturn_t r6040_interrupt(int irq, void *dev_id)
 658{
 659	struct net_device *dev = dev_id;
 660	struct r6040_private *lp = netdev_priv(dev);
 661	void __iomem *ioaddr = lp->base;
 662	u16 misr, status;
 663
 664	/* Save MIER */
 665	misr = ioread16(ioaddr + MIER);
 666	/* Mask off RDC MAC interrupt */
 667	iowrite16(MSK_INT, ioaddr + MIER);
 668	/* Read MISR status and clear */
 669	status = ioread16(ioaddr + MISR);
 670
 671	if (status == 0x0000 || status == 0xffff) {
 672		/* Restore RDC MAC interrupt */
 673		iowrite16(misr, ioaddr + MIER);
 674		return IRQ_NONE;
 675	}
 676
 677	/* RX interrupt request */
 678	if (status & RX_INTS) {
 679		if (status & RX_NO_DESC) {
 680			/* RX descriptor unavailable */
 681			dev->stats.rx_dropped++;
 682			dev->stats.rx_missed_errors++;
 683		}
 684		if (status & RX_FIFO_FULL)
 685			dev->stats.rx_fifo_errors++;
 686
 687		if (likely(napi_schedule_prep(&lp->napi))) {
 688			/* Mask off RX interrupt */
 689			misr &= ~RX_INTS;
 690			__napi_schedule(&lp->napi);
 691		}
 692	}
 693
 694	/* TX interrupt request */
 695	if (status & TX_INTS)
 696		r6040_tx(dev);
 697
 698	/* Restore RDC MAC interrupt */
 699	iowrite16(misr, ioaddr + MIER);
 700
 701	return IRQ_HANDLED;
 702}
 703
 704#ifdef CONFIG_NET_POLL_CONTROLLER
 705static void r6040_poll_controller(struct net_device *dev)
 706{
 707	disable_irq(dev->irq);
 708	r6040_interrupt(dev->irq, dev);
 709	enable_irq(dev->irq);
 710}
 711#endif
 712
 713/* Init RDC MAC */
 714static int r6040_up(struct net_device *dev)
 715{
 716	struct r6040_private *lp = netdev_priv(dev);
 717	void __iomem *ioaddr = lp->base;
 718	int ret;
 719
 720	/* Initialise and alloc RX/TX buffers */
 721	r6040_init_txbufs(dev);
 722	ret = r6040_alloc_rxbufs(dev);
 723	if (ret)
 724		return ret;
 725
 726	/* improve performance (by RDC guys) */
 727	r6040_phy_write(ioaddr, 30, 17,
 728			(r6040_phy_read(ioaddr, 30, 17) | 0x4000));
 729	r6040_phy_write(ioaddr, 30, 17,
 730			~((~r6040_phy_read(ioaddr, 30, 17)) | 0x2000));
 731	r6040_phy_write(ioaddr, 0, 19, 0x0000);
 732	r6040_phy_write(ioaddr, 0, 30, 0x01F0);
 733
 734	/* Initialize all MAC registers */
 735	r6040_init_mac_regs(dev);
 736
 737	phy_start(lp->phydev);
 738
 739	return 0;
 740}
 741
 742
 743/* Read/set MAC address routines */
 744static void r6040_mac_address(struct net_device *dev)
 745{
 746	struct r6040_private *lp = netdev_priv(dev);
 747	void __iomem *ioaddr = lp->base;
 748	u16 *adrp;
 749
 750	/* Reset MAC */
 751	r6040_reset_mac(lp);
 752
 753	/* Restore MAC Address */
 754	adrp = (u16 *) dev->dev_addr;
 755	iowrite16(adrp[0], ioaddr + MID_0L);
 756	iowrite16(adrp[1], ioaddr + MID_0M);
 757	iowrite16(adrp[2], ioaddr + MID_0H);
 758
 759	/* Store MAC Address in perm_addr */
 760	memcpy(dev->perm_addr, dev->dev_addr, ETH_ALEN);
 761}
 762
 763static int r6040_open(struct net_device *dev)
 764{
 765	struct r6040_private *lp = netdev_priv(dev);
 766	int ret;
 767
 768	/* Request IRQ and Register interrupt handler */
 769	ret = request_irq(dev->irq, r6040_interrupt,
 770		IRQF_SHARED, dev->name, dev);
 771	if (ret)
 772		goto out;
 773
 774	/* Set MAC address */
 775	r6040_mac_address(dev);
 776
 777	/* Allocate Descriptor memory */
 778	lp->rx_ring =
 779		pci_alloc_consistent(lp->pdev, RX_DESC_SIZE, &lp->rx_ring_dma);
 
 780	if (!lp->rx_ring) {
 781		ret = -ENOMEM;
 782		goto err_free_irq;
 783	}
 784
 785	lp->tx_ring =
 786		pci_alloc_consistent(lp->pdev, TX_DESC_SIZE, &lp->tx_ring_dma);
 
 787	if (!lp->tx_ring) {
 788		ret = -ENOMEM;
 789		goto err_free_rx_ring;
 790	}
 791
 792	ret = r6040_up(dev);
 793	if (ret)
 794		goto err_free_tx_ring;
 795
 796	napi_enable(&lp->napi);
 797	netif_start_queue(dev);
 798
 799	return 0;
 800
 801err_free_tx_ring:
 802	pci_free_consistent(lp->pdev, TX_DESC_SIZE, lp->tx_ring,
 803			lp->tx_ring_dma);
 804err_free_rx_ring:
 805	pci_free_consistent(lp->pdev, RX_DESC_SIZE, lp->rx_ring,
 806			lp->rx_ring_dma);
 807err_free_irq:
 808	free_irq(dev->irq, dev);
 809out:
 810	return ret;
 811}
 812
 813static netdev_tx_t r6040_start_xmit(struct sk_buff *skb,
 814				    struct net_device *dev)
 815{
 816	struct r6040_private *lp = netdev_priv(dev);
 817	struct r6040_descriptor *descptr;
 818	void __iomem *ioaddr = lp->base;
 819	unsigned long flags;
 820
 
 
 
 821	/* Critical Section */
 822	spin_lock_irqsave(&lp->lock, flags);
 823
 824	/* TX resource check */
 825	if (!lp->tx_free_desc) {
 826		spin_unlock_irqrestore(&lp->lock, flags);
 827		netif_stop_queue(dev);
 828		netdev_err(dev, ": no tx descriptor\n");
 829		return NETDEV_TX_BUSY;
 830	}
 831
 832	/* Statistic Counter */
 833	dev->stats.tx_packets++;
 834	dev->stats.tx_bytes += skb->len;
 835	/* Set TX descriptor & Transmit it */
 836	lp->tx_free_desc--;
 837	descptr = lp->tx_insert_ptr;
 838	if (skb->len < MISR)
 839		descptr->len = MISR;
 840	else
 841		descptr->len = skb->len;
 842
 843	descptr->skb_ptr = skb;
 844	descptr->buf = cpu_to_le32(pci_map_single(lp->pdev,
 845		skb->data, skb->len, PCI_DMA_TODEVICE));
 846	descptr->status = DSC_OWNER_MAC;
 847
 848	skb_tx_timestamp(skb);
 849
 850	/* Trigger the MAC to check the TX descriptor */
 851	iowrite16(TM2TX, ioaddr + MTPR);
 
 852	lp->tx_insert_ptr = descptr->vndescp;
 853
 854	/* If no tx resource, stop */
 855	if (!lp->tx_free_desc)
 856		netif_stop_queue(dev);
 857
 858	spin_unlock_irqrestore(&lp->lock, flags);
 859
 860	return NETDEV_TX_OK;
 861}
 862
 863static void r6040_multicast_list(struct net_device *dev)
 864{
 865	struct r6040_private *lp = netdev_priv(dev);
 866	void __iomem *ioaddr = lp->base;
 867	unsigned long flags;
 868	struct netdev_hw_addr *ha;
 869	int i;
 870	u16 *adrp;
 871	u16 hash_table[4] = { 0 };
 872
 873	spin_lock_irqsave(&lp->lock, flags);
 874
 875	/* Keep our MAC Address */
 876	adrp = (u16 *)dev->dev_addr;
 877	iowrite16(adrp[0], ioaddr + MID_0L);
 878	iowrite16(adrp[1], ioaddr + MID_0M);
 879	iowrite16(adrp[2], ioaddr + MID_0H);
 880
 881	/* Clear AMCP & PROM bits */
 882	lp->mcr0 = ioread16(ioaddr + MCR0) & ~(MCR0_PROMISC | MCR0_HASH_EN);
 883
 884	/* Promiscuous mode */
 885	if (dev->flags & IFF_PROMISC)
 886		lp->mcr0 |= MCR0_PROMISC;
 887
 888	/* Enable multicast hash table function to
 889	 * receive all multicast packets. */
 890	else if (dev->flags & IFF_ALLMULTI) {
 891		lp->mcr0 |= MCR0_HASH_EN;
 892
 893		for (i = 0; i < MCAST_MAX ; i++) {
 894			iowrite16(0, ioaddr + MID_1L + 8 * i);
 895			iowrite16(0, ioaddr + MID_1M + 8 * i);
 896			iowrite16(0, ioaddr + MID_1H + 8 * i);
 897		}
 898
 899		for (i = 0; i < 4; i++)
 900			hash_table[i] = 0xffff;
 901	}
 902	/* Use internal multicast address registers if the number of
 903	 * multicast addresses is not greater than MCAST_MAX. */
 904	else if (netdev_mc_count(dev) <= MCAST_MAX) {
 905		i = 0;
 906		netdev_for_each_mc_addr(ha, dev) {
 907			u16 *adrp = (u16 *) ha->addr;
 908			iowrite16(adrp[0], ioaddr + MID_1L + 8 * i);
 909			iowrite16(adrp[1], ioaddr + MID_1M + 8 * i);
 910			iowrite16(adrp[2], ioaddr + MID_1H + 8 * i);
 911			i++;
 912		}
 913		while (i < MCAST_MAX) {
 914			iowrite16(0, ioaddr + MID_1L + 8 * i);
 915			iowrite16(0, ioaddr + MID_1M + 8 * i);
 916			iowrite16(0, ioaddr + MID_1H + 8 * i);
 917			i++;
 918		}
 919	}
 920	/* Otherwise, Enable multicast hash table function. */
 921	else {
 922		u32 crc;
 923
 924		lp->mcr0 |= MCR0_HASH_EN;
 925
 926		for (i = 0; i < MCAST_MAX ; i++) {
 927			iowrite16(0, ioaddr + MID_1L + 8 * i);
 928			iowrite16(0, ioaddr + MID_1M + 8 * i);
 929			iowrite16(0, ioaddr + MID_1H + 8 * i);
 930		}
 931
 932		/* Build multicast hash table */
 933		netdev_for_each_mc_addr(ha, dev) {
 934			u8 *addrs = ha->addr;
 935
 936			crc = ether_crc(ETH_ALEN, addrs);
 937			crc >>= 26;
 938			hash_table[crc >> 4] |= 1 << (crc & 0xf);
 939		}
 940	}
 941
 942	iowrite16(lp->mcr0, ioaddr + MCR0);
 943
 944	/* Fill the MAC hash tables with their values */
 945	if (lp->mcr0 & MCR0_HASH_EN) {
 946		iowrite16(hash_table[0], ioaddr + MAR0);
 947		iowrite16(hash_table[1], ioaddr + MAR1);
 948		iowrite16(hash_table[2], ioaddr + MAR2);
 949		iowrite16(hash_table[3], ioaddr + MAR3);
 950	}
 951
 952	spin_unlock_irqrestore(&lp->lock, flags);
 953}
 954
 955static void netdev_get_drvinfo(struct net_device *dev,
 956			struct ethtool_drvinfo *info)
 957{
 958	struct r6040_private *rp = netdev_priv(dev);
 959
 960	strcpy(info->driver, DRV_NAME);
 961	strcpy(info->version, DRV_VERSION);
 962	strcpy(info->bus_info, pci_name(rp->pdev));
 963}
 964
 965static int netdev_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
 966{
 967	struct r6040_private *rp = netdev_priv(dev);
 968
 969	return  phy_ethtool_gset(rp->phydev, cmd);
 970}
 971
 972static int netdev_set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
 973{
 974	struct r6040_private *rp = netdev_priv(dev);
 975
 976	return phy_ethtool_sset(rp->phydev, cmd);
 977}
 978
 979static const struct ethtool_ops netdev_ethtool_ops = {
 980	.get_drvinfo		= netdev_get_drvinfo,
 981	.get_settings		= netdev_get_settings,
 982	.set_settings		= netdev_set_settings,
 983	.get_link		= ethtool_op_get_link,
 984	.get_ts_info		= ethtool_op_get_ts_info,
 
 
 
 985};
 986
 987static const struct net_device_ops r6040_netdev_ops = {
 988	.ndo_open		= r6040_open,
 989	.ndo_stop		= r6040_close,
 990	.ndo_start_xmit		= r6040_start_xmit,
 991	.ndo_get_stats		= r6040_get_stats,
 992	.ndo_set_rx_mode	= r6040_multicast_list,
 993	.ndo_change_mtu		= eth_change_mtu,
 994	.ndo_validate_addr	= eth_validate_addr,
 995	.ndo_set_mac_address	= eth_mac_addr,
 996	.ndo_do_ioctl		= r6040_ioctl,
 997	.ndo_tx_timeout		= r6040_tx_timeout,
 998#ifdef CONFIG_NET_POLL_CONTROLLER
 999	.ndo_poll_controller	= r6040_poll_controller,
1000#endif
1001};
1002
1003static void r6040_adjust_link(struct net_device *dev)
1004{
1005	struct r6040_private *lp = netdev_priv(dev);
1006	struct phy_device *phydev = lp->phydev;
1007	int status_changed = 0;
1008	void __iomem *ioaddr = lp->base;
1009
1010	BUG_ON(!phydev);
1011
1012	if (lp->old_link != phydev->link) {
1013		status_changed = 1;
1014		lp->old_link = phydev->link;
1015	}
1016
1017	/* reflect duplex change */
1018	if (phydev->link && (lp->old_duplex != phydev->duplex)) {
1019		lp->mcr0 |= (phydev->duplex == DUPLEX_FULL ? MCR0_FD : 0);
1020		iowrite16(lp->mcr0, ioaddr);
1021
1022		status_changed = 1;
1023		lp->old_duplex = phydev->duplex;
1024	}
1025
1026	if (status_changed) {
1027		pr_info("%s: link %s", dev->name, phydev->link ?
1028			"UP" : "DOWN");
1029		if (phydev->link)
1030			pr_cont(" - %d/%s", phydev->speed,
1031			DUPLEX_FULL == phydev->duplex ? "full" : "half");
1032		pr_cont("\n");
1033	}
1034}
1035
1036static int r6040_mii_probe(struct net_device *dev)
1037{
1038	struct r6040_private *lp = netdev_priv(dev);
1039	struct phy_device *phydev = NULL;
1040
1041	phydev = phy_find_first(lp->mii_bus);
1042	if (!phydev) {
1043		dev_err(&lp->pdev->dev, "no PHY found\n");
1044		return -ENODEV;
1045	}
1046
1047	phydev = phy_connect(dev, dev_name(&phydev->dev), &r6040_adjust_link,
1048				0, PHY_INTERFACE_MODE_MII);
1049
1050	if (IS_ERR(phydev)) {
1051		dev_err(&lp->pdev->dev, "could not attach to PHY\n");
1052		return PTR_ERR(phydev);
1053	}
1054
1055	/* mask with MAC supported features */
1056	phydev->supported &= (SUPPORTED_10baseT_Half
1057				| SUPPORTED_10baseT_Full
1058				| SUPPORTED_100baseT_Half
1059				| SUPPORTED_100baseT_Full
1060				| SUPPORTED_Autoneg
1061				| SUPPORTED_MII
1062				| SUPPORTED_TP);
1063
1064	phydev->advertising = phydev->supported;
1065	lp->phydev = phydev;
1066	lp->old_link = 0;
1067	lp->old_duplex = -1;
1068
1069	dev_info(&lp->pdev->dev, "attached PHY driver [%s] "
1070		"(mii_bus:phy_addr=%s)\n",
1071		phydev->drv->name, dev_name(&phydev->dev));
1072
1073	return 0;
1074}
1075
1076static int __devinit r6040_init_one(struct pci_dev *pdev,
1077					 const struct pci_device_id *ent)
1078{
1079	struct net_device *dev;
1080	struct r6040_private *lp;
1081	void __iomem *ioaddr;
1082	int err, io_size = R6040_IO_SIZE;
1083	static int card_idx = -1;
 
1084	int bar = 0;
1085	u16 *adrp;
1086	int i;
1087
1088	pr_info("%s\n", version);
1089
1090	err = pci_enable_device(pdev);
1091	if (err)
1092		goto err_out;
1093
1094	/* this should always be supported */
1095	err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
1096	if (err) {
1097		dev_err(&pdev->dev, "32-bit PCI DMA addresses"
1098				"not supported by the card\n");
1099		goto err_out_disable_dev;
1100	}
1101	err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
1102	if (err) {
1103		dev_err(&pdev->dev, "32-bit PCI DMA addresses"
1104				"not supported by the card\n");
1105		goto err_out_disable_dev;
1106	}
1107
1108	/* IO Size check */
1109	if (pci_resource_len(pdev, bar) < io_size) {
1110		dev_err(&pdev->dev, "Insufficient PCI resources, aborting\n");
1111		err = -EIO;
1112		goto err_out_disable_dev;
1113	}
1114
1115	pci_set_master(pdev);
1116
1117	dev = alloc_etherdev(sizeof(struct r6040_private));
1118	if (!dev) {
1119		err = -ENOMEM;
1120		goto err_out_disable_dev;
1121	}
1122	SET_NETDEV_DEV(dev, &pdev->dev);
1123	lp = netdev_priv(dev);
1124
1125	err = pci_request_regions(pdev, DRV_NAME);
1126
1127	if (err) {
1128		dev_err(&pdev->dev, "Failed to request PCI regions\n");
1129		goto err_out_free_dev;
1130	}
1131
1132	ioaddr = pci_iomap(pdev, bar, io_size);
1133	if (!ioaddr) {
1134		dev_err(&pdev->dev, "ioremap failed for device\n");
1135		err = -EIO;
1136		goto err_out_free_res;
1137	}
1138
1139	/* If PHY status change register is still set to zero it means the
1140	 * bootloader didn't initialize it, so we set it to:
1141	 * - enable phy status change
1142	 * - enable all phy addresses
1143	 * - set to lowest timer divider */
1144	if (ioread16(ioaddr + PHY_CC) == 0)
1145		iowrite16(SCEN | PHY_MAX_ADDR << PHYAD_SHIFT |
1146				7 << TMRDIV_SHIFT, ioaddr + PHY_CC);
1147
1148	/* Init system & device */
1149	lp->base = ioaddr;
1150	dev->irq = pdev->irq;
1151
1152	spin_lock_init(&lp->lock);
1153	pci_set_drvdata(pdev, dev);
1154
1155	/* Set MAC address */
1156	card_idx++;
1157
1158	adrp = (u16 *)dev->dev_addr;
1159	adrp[0] = ioread16(ioaddr + MID_0L);
1160	adrp[1] = ioread16(ioaddr + MID_0M);
1161	adrp[2] = ioread16(ioaddr + MID_0H);
1162
1163	/* Some bootloader/BIOSes do not initialize
1164	 * MAC address, warn about that */
1165	if (!(adrp[0] || adrp[1] || adrp[2])) {
1166		netdev_warn(dev, "MAC address not initialized, "
1167					"generating random\n");
1168		eth_hw_addr_random(dev);
1169	}
1170
1171	/* Link new device into r6040_root_dev */
1172	lp->pdev = pdev;
1173	lp->dev = dev;
1174
1175	/* Init RDC private data */
1176	lp->mcr0 = MCR0_XMTEN | MCR0_RCVEN;
1177
1178	/* The RDC-specific entries in the device structure. */
1179	dev->netdev_ops = &r6040_netdev_ops;
1180	dev->ethtool_ops = &netdev_ethtool_ops;
1181	dev->watchdog_timeo = TX_TIMEOUT;
1182
1183	netif_napi_add(dev, &lp->napi, r6040_poll, 64);
1184
1185	lp->mii_bus = mdiobus_alloc();
1186	if (!lp->mii_bus) {
1187		dev_err(&pdev->dev, "mdiobus_alloc() failed\n");
1188		err = -ENOMEM;
1189		goto err_out_unmap;
1190	}
1191
1192	lp->mii_bus->priv = dev;
1193	lp->mii_bus->read = r6040_mdiobus_read;
1194	lp->mii_bus->write = r6040_mdiobus_write;
1195	lp->mii_bus->reset = r6040_mdiobus_reset;
1196	lp->mii_bus->name = "r6040_eth_mii";
1197	snprintf(lp->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
1198		dev_name(&pdev->dev), card_idx);
1199	lp->mii_bus->irq = kmalloc(sizeof(int)*PHY_MAX_ADDR, GFP_KERNEL);
1200	if (!lp->mii_bus->irq) {
1201		dev_err(&pdev->dev, "mii_bus irq allocation failed\n");
1202		err = -ENOMEM;
1203		goto err_out_mdio;
1204	}
1205
1206	for (i = 0; i < PHY_MAX_ADDR; i++)
1207		lp->mii_bus->irq[i] = PHY_POLL;
1208
1209	err = mdiobus_register(lp->mii_bus);
1210	if (err) {
1211		dev_err(&pdev->dev, "failed to register MII bus\n");
1212		goto err_out_mdio_irq;
1213	}
1214
1215	err = r6040_mii_probe(dev);
1216	if (err) {
1217		dev_err(&pdev->dev, "failed to probe MII bus\n");
1218		goto err_out_mdio_unregister;
1219	}
1220
1221	/* Register net device. After this dev->name assign */
1222	err = register_netdev(dev);
1223	if (err) {
1224		dev_err(&pdev->dev, "Failed to register net device\n");
1225		goto err_out_mdio_unregister;
1226	}
1227	return 0;
1228
 
 
1229err_out_mdio_unregister:
1230	mdiobus_unregister(lp->mii_bus);
1231err_out_mdio_irq:
1232	kfree(lp->mii_bus->irq);
1233err_out_mdio:
1234	mdiobus_free(lp->mii_bus);
1235err_out_unmap:
1236	netif_napi_del(&lp->napi);
1237	pci_set_drvdata(pdev, NULL);
1238	pci_iounmap(pdev, ioaddr);
1239err_out_free_res:
1240	pci_release_regions(pdev);
1241err_out_free_dev:
1242	free_netdev(dev);
1243err_out_disable_dev:
1244	pci_disable_device(pdev);
1245err_out:
1246	return err;
1247}
1248
1249static void __devexit r6040_remove_one(struct pci_dev *pdev)
1250{
1251	struct net_device *dev = pci_get_drvdata(pdev);
1252	struct r6040_private *lp = netdev_priv(dev);
1253
1254	unregister_netdev(dev);
 
1255	mdiobus_unregister(lp->mii_bus);
1256	kfree(lp->mii_bus->irq);
1257	mdiobus_free(lp->mii_bus);
1258	netif_napi_del(&lp->napi);
1259	pci_set_drvdata(pdev, NULL);
1260	pci_iounmap(pdev, lp->base);
1261	pci_release_regions(pdev);
1262	free_netdev(dev);
1263	pci_disable_device(pdev);
1264	pci_set_drvdata(pdev, NULL);
1265}
1266
1267
1268static DEFINE_PCI_DEVICE_TABLE(r6040_pci_tbl) = {
1269	{ PCI_DEVICE(PCI_VENDOR_ID_RDC, 0x6040) },
1270	{ 0 }
1271};
1272MODULE_DEVICE_TABLE(pci, r6040_pci_tbl);
1273
1274static struct pci_driver r6040_driver = {
1275	.name		= DRV_NAME,
1276	.id_table	= r6040_pci_tbl,
1277	.probe		= r6040_init_one,
1278	.remove		= __devexit_p(r6040_remove_one),
1279};
1280
1281
1282static int __init r6040_init(void)
1283{
1284	return pci_register_driver(&r6040_driver);
1285}
1286
1287
1288static void __exit r6040_cleanup(void)
1289{
1290	pci_unregister_driver(&r6040_driver);
1291}
1292
1293module_init(r6040_init);
1294module_exit(r6040_cleanup);