Linux Audio

Check our new training course

Yocto / OpenEmbedded training

Mar 24-27, 2025, special US time zones
Register
Loading...
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * wanXL serial card driver for Linux
  4 * host part
  5 *
  6 * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
  7 *
 
 
 
 
  8 * Status:
  9 *   - Only DTE (external clock) support with NRZ and NRZI encodings
 10 *   - wanXL100 will require minor driver modifications, no access to hw
 11 */
 12
 13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
 14
 15#include <linux/module.h>
 16#include <linux/kernel.h>
 17#include <linux/slab.h>
 18#include <linux/sched.h>
 19#include <linux/types.h>
 20#include <linux/fcntl.h>
 21#include <linux/string.h>
 22#include <linux/errno.h>
 23#include <linux/init.h>
 24#include <linux/interrupt.h>
 25#include <linux/ioport.h>
 26#include <linux/netdevice.h>
 27#include <linux/hdlc.h>
 28#include <linux/pci.h>
 29#include <linux/dma-mapping.h>
 30#include <linux/delay.h>
 31#include <asm/io.h>
 32
 33#include "wanxl.h"
 34
 35static const char *version = "wanXL serial card driver version: 0.48";
 36
 37#define PLX_CTL_RESET   0x40000000 /* adapter reset */
 38
 39#undef DEBUG_PKT
 40#undef DEBUG_PCI
 41
 42/* MAILBOX #1 - PUTS COMMANDS */
 43#define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
 44#ifdef __LITTLE_ENDIAN
 45#define MBX1_CMD_BSWAP  0x8C000001 /* little-endian Byte Swap Mode */
 46#else
 47#define MBX1_CMD_BSWAP  0x8C000000 /* big-endian Byte Swap Mode */
 48#endif
 49
 50/* MAILBOX #2 - DRAM SIZE */
 51#define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
 52
 
 53struct port {
 54	struct net_device *dev;
 55	struct card *card;
 56	spinlock_t lock;	/* for wanxl_xmit */
 57	int node;		/* physical port #0 - 3 */
 58	unsigned int clock_type;
 59	int tx_in, tx_out;
 60	struct sk_buff *tx_skbs[TX_BUFFERS];
 61};
 62
 
 63struct card_status {
 64	desc_t rx_descs[RX_QUEUE_LENGTH];
 65	port_status_t port_status[4];
 66};
 67
 
 68struct card {
 69	int n_ports;		/* 1, 2 or 4 ports */
 70	u8 irq;
 71
 72	u8 __iomem *plx;	/* PLX PCI9060 virtual base address */
 73	struct pci_dev *pdev;	/* for pci_name(pdev) */
 74	int rx_in;
 75	struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
 76	struct card_status *status;	/* shared between host and card */
 77	dma_addr_t status_address;
 78	struct port ports[];	/* 1 - 4 port structures follow */
 79};
 80
 
 
 81static inline struct port *dev_to_port(struct net_device *dev)
 82{
 83	return (struct port *)dev_to_hdlc(dev)->priv;
 84}
 85
 
 86static inline port_status_t *get_status(struct port *port)
 87{
 88	return &port->card->status->port_status[port->node];
 89}
 90
 
 91#ifdef DEBUG_PCI
 92static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
 93					      size_t size, int direction)
 94{
 95	dma_addr_t addr = dma_map_single(&pdev->dev, ptr, size, direction);
 96
 97	if (addr + size > 0x100000000LL)
 98		pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
 99			pci_name(pdev), (unsigned long long)addr);
100	return addr;
101}
102
103#undef pci_map_single
104#define pci_map_single pci_map_single_debug
105#endif
106
 
107/* Cable and/or personality module change interrupt service */
108static inline void wanxl_cable_intr(struct port *port)
109{
110	u32 value = get_status(port)->cable;
111	int valid = 1;
112	const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
113
114	switch (value & 0x7) {
115	case STATUS_CABLE_V35:
116		cable = "V.35";
117		break;
118	case STATUS_CABLE_X21:
119		cable = "X.21";
120		break;
121	case STATUS_CABLE_V24:
122		cable = "V.24";
123		break;
124	case STATUS_CABLE_EIA530:
125		cable = "EIA530";
126		break;
127	case STATUS_CABLE_NONE:
128		cable = "no";
129		break;
130	default:
131		cable = "invalid";
132	}
133
134	switch ((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
135	case STATUS_CABLE_V35:
136		pm = "V.35";
137		break;
138	case STATUS_CABLE_X21:
139		pm = "X.21";
140		break;
141	case STATUS_CABLE_V24:
142		pm = "V.24";
143		break;
144	case STATUS_CABLE_EIA530:
145		pm = "EIA530";
146		break;
147	case STATUS_CABLE_NONE:
148		pm = "no personality";
149		valid = 0;
150		break;
151	default:
152		pm = "invalid personality";
153		valid = 0;
154	}
155
156	if (valid) {
157		if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
158			dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
159				", DSR off";
160			dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
161				", carrier off";
162		}
163		dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
164	}
165	netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
166		    pm, dte, cable, dsr, dcd);
167
168	if (value & STATUS_CABLE_DCD)
169		netif_carrier_on(port->dev);
170	else
171		netif_carrier_off(port->dev);
172}
173
 
 
174/* Transmit complete interrupt service */
175static inline void wanxl_tx_intr(struct port *port)
176{
177	struct net_device *dev = port->dev;
178
179	while (1) {
180		desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
181		struct sk_buff *skb = port->tx_skbs[port->tx_in];
182
183		switch (desc->stat) {
184		case PACKET_FULL:
185		case PACKET_EMPTY:
186			netif_wake_queue(dev);
187			return;
188
189		case PACKET_UNDERRUN:
190			dev->stats.tx_errors++;
191			dev->stats.tx_fifo_errors++;
192			break;
193
194		default:
195			dev->stats.tx_packets++;
196			dev->stats.tx_bytes += skb->len;
197		}
198		desc->stat = PACKET_EMPTY; /* Free descriptor */
199		dma_unmap_single(&port->card->pdev->dev, desc->address,
200				 skb->len, DMA_TO_DEVICE);
201		dev_consume_skb_irq(skb);
202		port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
203	}
204}
205
 
 
206/* Receive complete interrupt service */
207static inline void wanxl_rx_intr(struct card *card)
208{
209	desc_t *desc;
210
211	while (desc = &card->status->rx_descs[card->rx_in],
212	       desc->stat != PACKET_EMPTY) {
213		if ((desc->stat & PACKET_PORT_MASK) > card->n_ports) {
214			pr_crit("%s: received packet for nonexistent port\n",
215				pci_name(card->pdev));
216		} else {
217			struct sk_buff *skb = card->rx_skbs[card->rx_in];
218			struct port *port = &card->ports[desc->stat &
219						    PACKET_PORT_MASK];
220			struct net_device *dev = port->dev;
221
222			if (!skb) {
223				dev->stats.rx_dropped++;
224			} else {
225				dma_unmap_single(&card->pdev->dev,
226						 desc->address, BUFFER_LENGTH,
227						 DMA_FROM_DEVICE);
228				skb_put(skb, desc->length);
229
230#ifdef DEBUG_PKT
231				printk(KERN_DEBUG "%s RX(%i):", dev->name,
232				       skb->len);
233				debug_frame(skb);
234#endif
235				dev->stats.rx_packets++;
236				dev->stats.rx_bytes += skb->len;
237				skb->protocol = hdlc_type_trans(skb, dev);
238				netif_rx(skb);
239				skb = NULL;
240			}
241
242			if (!skb) {
243				skb = dev_alloc_skb(BUFFER_LENGTH);
244				desc->address = skb ?
245					dma_map_single(&card->pdev->dev,
246						       skb->data,
247						       BUFFER_LENGTH,
248						       DMA_FROM_DEVICE) : 0;
249				card->rx_skbs[card->rx_in] = skb;
250			}
251		}
252		desc->stat = PACKET_EMPTY; /* Free descriptor */
253		card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
254	}
255}
256
257static irqreturn_t wanxl_intr(int irq, void *dev_id)
 
 
258{
259	struct card *card = dev_id;
260	int i;
261	u32 stat;
262	int handled = 0;
263
264	while ((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
265		handled = 1;
 
266		writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
267
268		for (i = 0; i < card->n_ports; i++) {
269			if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
270				wanxl_tx_intr(&card->ports[i]);
271			if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
272				wanxl_cable_intr(&card->ports[i]);
273		}
274		if (stat & (1 << DOORBELL_FROM_CARD_RX))
275			wanxl_rx_intr(card);
276	}
277
278	return IRQ_RETVAL(handled);
279}
280
 
 
281static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
282{
283	struct port *port = dev_to_port(dev);
284	desc_t *desc;
285
286	spin_lock(&port->lock);
287
288	desc = &get_status(port)->tx_descs[port->tx_out];
289	if (desc->stat != PACKET_EMPTY) {
290		/* should never happen - previous xmit should stop queue */
291#ifdef DEBUG_PKT
292                printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
293#endif
294		netif_stop_queue(dev);
295		spin_unlock(&port->lock);
296		return NETDEV_TX_BUSY;       /* request packet to be queued */
297	}
298
299#ifdef DEBUG_PKT
300	printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
301	debug_frame(skb);
302#endif
303
304	port->tx_skbs[port->tx_out] = skb;
305	desc->address = dma_map_single(&port->card->pdev->dev, skb->data,
306				       skb->len, DMA_TO_DEVICE);
307	desc->length = skb->len;
308	desc->stat = PACKET_FULL;
309	writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
310	       port->card->plx + PLX_DOORBELL_TO_CARD);
311
312	port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
313
314	if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
315		netif_stop_queue(dev);
316#ifdef DEBUG_PKT
317		printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
318#endif
319	}
320
321	spin_unlock(&port->lock);
322	return NETDEV_TX_OK;
323}
324
 
 
325static int wanxl_attach(struct net_device *dev, unsigned short encoding,
326			unsigned short parity)
327{
328	struct port *port = dev_to_port(dev);
329
330	if (encoding != ENCODING_NRZ &&
331	    encoding != ENCODING_NRZI)
332		return -EINVAL;
333
334	if (parity != PARITY_NONE &&
335	    parity != PARITY_CRC32_PR1_CCITT &&
336	    parity != PARITY_CRC16_PR1_CCITT &&
337	    parity != PARITY_CRC32_PR0_CCITT &&
338	    parity != PARITY_CRC16_PR0_CCITT)
339		return -EINVAL;
340
341	get_status(port)->encoding = encoding;
342	get_status(port)->parity = parity;
343	return 0;
344}
345
346static int wanxl_ioctl(struct net_device *dev, struct if_settings *ifs)
 
 
347{
348	const size_t size = sizeof(sync_serial_settings);
349	sync_serial_settings line;
350	struct port *port = dev_to_port(dev);
351
352	switch (ifs->type) {
 
 
 
353	case IF_GET_IFACE:
354		ifs->type = IF_IFACE_SYNC_SERIAL;
355		if (ifs->size < size) {
356			ifs->size = size; /* data size wanted */
357			return -ENOBUFS;
358		}
359		memset(&line, 0, sizeof(line));
360		line.clock_type = get_status(port)->clocking;
361		line.clock_rate = 0;
362		line.loopback = 0;
363
364		if (copy_to_user(ifs->ifs_ifsu.sync, &line, size))
365			return -EFAULT;
366		return 0;
367
368	case IF_IFACE_SYNC_SERIAL:
369		if (!capable(CAP_NET_ADMIN))
370			return -EPERM;
371		if (dev->flags & IFF_UP)
372			return -EBUSY;
373
374		if (copy_from_user(&line, ifs->ifs_ifsu.sync,
375				   size))
376			return -EFAULT;
377
378		if (line.clock_type != CLOCK_EXT &&
379		    line.clock_type != CLOCK_TXFROMRX)
380			return -EINVAL; /* No such clock setting */
381
382		if (line.loopback != 0)
383			return -EINVAL;
384
385		get_status(port)->clocking = line.clock_type;
386		return 0;
387
388	default:
389		return hdlc_ioctl(dev, ifs);
390	}
391}
392
 
 
393static int wanxl_open(struct net_device *dev)
394{
395	struct port *port = dev_to_port(dev);
396	u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
397	unsigned long timeout;
398	int i;
399
400	if (get_status(port)->open) {
401		netdev_err(dev, "port already open\n");
402		return -EIO;
403	}
404
405	i = hdlc_open(dev);
406	if (i)
407		return i;
408
409	port->tx_in = port->tx_out = 0;
410	for (i = 0; i < TX_BUFFERS; i++)
411		get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
412	/* signal the card */
413	writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
414
415	timeout = jiffies + HZ;
416	do {
417		if (get_status(port)->open) {
418			netif_start_queue(dev);
419			return 0;
420		}
421	} while (time_after(timeout, jiffies));
422
423	netdev_err(dev, "unable to open port\n");
424	/* ask the card to close the port, should it be still alive */
425	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
426	return -EFAULT;
427}
428
 
 
429static int wanxl_close(struct net_device *dev)
430{
431	struct port *port = dev_to_port(dev);
432	unsigned long timeout;
433	int i;
434
435	hdlc_close(dev);
436	/* signal the card */
437	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
438	       port->card->plx + PLX_DOORBELL_TO_CARD);
439
440	timeout = jiffies + HZ;
441	do {
442		if (!get_status(port)->open)
443			break;
444	} while (time_after(timeout, jiffies));
445
446	if (get_status(port)->open)
447		netdev_err(dev, "unable to close port\n");
448
449	netif_stop_queue(dev);
450
451	for (i = 0; i < TX_BUFFERS; i++) {
452		desc_t *desc = &get_status(port)->tx_descs[i];
453
454		if (desc->stat != PACKET_EMPTY) {
455			desc->stat = PACKET_EMPTY;
456			dma_unmap_single(&port->card->pdev->dev,
457					 desc->address, port->tx_skbs[i]->len,
458					 DMA_TO_DEVICE);
459			dev_kfree_skb(port->tx_skbs[i]);
460		}
461	}
462	return 0;
463}
464
 
 
465static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
466{
467	struct port *port = dev_to_port(dev);
468
469	dev->stats.rx_over_errors = get_status(port)->rx_overruns;
470	dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
471	dev->stats.rx_errors = dev->stats.rx_over_errors +
472		dev->stats.rx_frame_errors;
473	return &dev->stats;
474}
475
 
 
476static int wanxl_puts_command(struct card *card, u32 cmd)
477{
478	unsigned long timeout = jiffies + 5 * HZ;
479
480	writel(cmd, card->plx + PLX_MAILBOX_1);
481	do {
482		if (readl(card->plx + PLX_MAILBOX_1) == 0)
483			return 0;
484
485		schedule();
486	} while (time_after(timeout, jiffies));
487
488	return -1;
489}
490
 
 
491static void wanxl_reset(struct card *card)
492{
493	u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
494
495	writel(0x80, card->plx + PLX_MAILBOX_0);
496	writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
497	readl(card->plx + PLX_CONTROL); /* wait for posted write */
498	udelay(1);
499	writel(old_value, card->plx + PLX_CONTROL);
500	readl(card->plx + PLX_CONTROL); /* wait for posted write */
501}
502
 
 
503static void wanxl_pci_remove_one(struct pci_dev *pdev)
504{
505	struct card *card = pci_get_drvdata(pdev);
506	int i;
507
508	for (i = 0; i < card->n_ports; i++) {
509		unregister_hdlc_device(card->ports[i].dev);
510		free_netdev(card->ports[i].dev);
511	}
512
513	/* unregister and free all host resources */
514	if (card->irq)
515		free_irq(card->irq, card);
516
517	wanxl_reset(card);
518
519	for (i = 0; i < RX_QUEUE_LENGTH; i++)
520		if (card->rx_skbs[i]) {
521			dma_unmap_single(&card->pdev->dev,
522					 card->status->rx_descs[i].address,
523					 BUFFER_LENGTH, DMA_FROM_DEVICE);
524			dev_kfree_skb(card->rx_skbs[i]);
525		}
526
527	if (card->plx)
528		iounmap(card->plx);
529
530	if (card->status)
531		dma_free_coherent(&pdev->dev, sizeof(struct card_status),
532				  card->status, card->status_address);
533
534	pci_release_regions(pdev);
535	pci_disable_device(pdev);
536	kfree(card);
537}
538
 
539#include "wanxlfw.inc"
540
541static const struct net_device_ops wanxl_ops = {
542	.ndo_open       = wanxl_open,
543	.ndo_stop       = wanxl_close,
 
544	.ndo_start_xmit = hdlc_start_xmit,
545	.ndo_siocwandev = wanxl_ioctl,
546	.ndo_get_stats  = wanxl_get_stats,
547};
548
549static int wanxl_pci_init_one(struct pci_dev *pdev,
550			      const struct pci_device_id *ent)
551{
552	struct card *card;
553	u32 ramsize, stat;
554	unsigned long timeout;
555	u32 plx_phy;		/* PLX PCI base address */
556	u32 mem_phy;		/* memory PCI base addr */
557	u8 __iomem *mem;	/* memory virtual base addr */
558	int i, ports;
559
560#ifndef MODULE
561	pr_info_once("%s\n", version);
562#endif
563
564	i = pci_enable_device(pdev);
565	if (i)
566		return i;
567
568	/* QUICC can only access first 256 MB of host RAM directly,
569	 * but PLX9060 DMA does 32-bits for actual packet data transfers
570	 */
571
572	/* FIXME when PCI/DMA subsystems are fixed.
573	 * We set both dma_mask and consistent_dma_mask to 28 bits
574	 * and pray pci_alloc_consistent() will use this info. It should
575	 * work on most platforms
576	 */
577	if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(28)) ||
578	    dma_set_mask(&pdev->dev, DMA_BIT_MASK(28))) {
579		pr_err("No usable DMA configuration\n");
580		pci_disable_device(pdev);
581		return -EIO;
582	}
583
584	i = pci_request_regions(pdev, "wanXL");
585	if (i) {
586		pci_disable_device(pdev);
587		return i;
588	}
589
590	switch (pdev->device) {
591	case PCI_DEVICE_ID_SBE_WANXL100:
592		ports = 1;
593		break;
594	case PCI_DEVICE_ID_SBE_WANXL200:
595		ports = 2;
596		break;
597	default:
598		ports = 4;
599	}
600
601	card = kzalloc(struct_size(card, ports, ports), GFP_KERNEL);
602	if (!card) {
 
603		pci_release_regions(pdev);
604		pci_disable_device(pdev);
605		return -ENOBUFS;
606	}
607
608	pci_set_drvdata(pdev, card);
609	card->pdev = pdev;
610
611	card->status = dma_alloc_coherent(&pdev->dev,
612					  sizeof(struct card_status),
613					  &card->status_address, GFP_KERNEL);
614	if (!card->status) {
615		wanxl_pci_remove_one(pdev);
616		return -ENOBUFS;
617	}
618
619#ifdef DEBUG_PCI
620	printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
621	       " at 0x%LX\n", pci_name(pdev),
622	       (unsigned long long)card->status_address);
623#endif
624
625	/* FIXME when PCI/DMA subsystems are fixed.
626	 * We set both dma_mask and consistent_dma_mask back to 32 bits
627	 * to indicate the card can do 32-bit DMA addressing
628	 */
629	if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)) ||
630	    dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
631		pr_err("No usable DMA configuration\n");
632		wanxl_pci_remove_one(pdev);
633		return -EIO;
634	}
635
636	/* set up PLX mapping */
637	plx_phy = pci_resource_start(pdev, 0);
638
639	card->plx = ioremap(plx_phy, 0x70);
640	if (!card->plx) {
641		pr_err("ioremap() failed\n");
642		wanxl_pci_remove_one(pdev);
643		return -EFAULT;
644	}
645
646#if RESET_WHILE_LOADING
647	wanxl_reset(card);
648#endif
649
650	timeout = jiffies + 20 * HZ;
651	while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
652		if (time_before(timeout, jiffies)) {
653			pr_warn("%s: timeout waiting for PUTS to complete\n",
654				pci_name(pdev));
655			wanxl_pci_remove_one(pdev);
656			return -ENODEV;
657		}
658
659		switch (stat & 0xC0) {
660		case 0x00:	/* hmm - PUTS completed with non-zero code? */
661		case 0x80:	/* PUTS still testing the hardware */
662			break;
663
664		default:
665			pr_warn("%s: PUTS test 0x%X failed\n",
666				pci_name(pdev), stat & 0x30);
667			wanxl_pci_remove_one(pdev);
668			return -ENODEV;
669		}
670
671		schedule();
672	}
673
674	/* get on-board memory size (PUTS detects no more than 4 MB) */
675	ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
676
677	/* set up on-board RAM mapping */
678	mem_phy = pci_resource_start(pdev, 2);
679
 
680	/* sanity check the board's reported memory size */
681	if (ramsize < BUFFERS_ADDR +
682	    (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
683		pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
684			pci_name(pdev), ramsize,
685			BUFFERS_ADDR +
686			(TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
687		wanxl_pci_remove_one(pdev);
688		return -ENODEV;
689	}
690
691	if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
692		pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
693		wanxl_pci_remove_one(pdev);
694		return -ENODEV;
695	}
696
697	for (i = 0; i < RX_QUEUE_LENGTH; i++) {
698		struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
699
700		card->rx_skbs[i] = skb;
701		if (skb)
702			card->status->rx_descs[i].address =
703				dma_map_single(&card->pdev->dev, skb->data,
704					       BUFFER_LENGTH, DMA_FROM_DEVICE);
 
705	}
706
707	mem = ioremap(mem_phy, PDM_OFFSET + sizeof(firmware));
708	if (!mem) {
709		pr_err("ioremap() failed\n");
710		wanxl_pci_remove_one(pdev);
711		return -EFAULT;
712	}
713
714	for (i = 0; i < sizeof(firmware); i += 4)
715		writel(ntohl(*(__be32 *)(firmware + i)), mem + PDM_OFFSET + i);
716
717	for (i = 0; i < ports; i++)
718		writel(card->status_address +
719		       (void *)&card->status->port_status[i] -
720		       (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
721	writel(card->status_address, mem + PDM_OFFSET + 20);
722	writel(PDM_OFFSET, mem);
723	iounmap(mem);
724
725	writel(0, card->plx + PLX_MAILBOX_5);
726
727	if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
728		pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
729		wanxl_pci_remove_one(pdev);
730		return -ENODEV;
731	}
732
 
733	timeout = jiffies + 5 * HZ;
734	do {
735		stat = readl(card->plx + PLX_MAILBOX_5);
736		if (stat)
737			break;
738		schedule();
739	} while (time_after(timeout, jiffies));
740
741	if (!stat) {
742		pr_warn("%s: timeout while initializing card firmware\n",
743			pci_name(pdev));
744		wanxl_pci_remove_one(pdev);
745		return -ENODEV;
746	}
747
748#if DETECT_RAM
749	ramsize = stat;
750#endif
751
752	pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
753		pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
754
755	/* Allocate IRQ */
756	if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
757		pr_warn("%s: could not allocate IRQ%i\n",
758			pci_name(pdev), pdev->irq);
759		wanxl_pci_remove_one(pdev);
760		return -EBUSY;
761	}
762	card->irq = pdev->irq;
763
764	for (i = 0; i < ports; i++) {
765		hdlc_device *hdlc;
766		struct port *port = &card->ports[i];
767		struct net_device *dev = alloc_hdlcdev(port);
768
769		if (!dev) {
770			pr_err("%s: unable to allocate memory\n",
771			       pci_name(pdev));
772			wanxl_pci_remove_one(pdev);
773			return -ENOMEM;
774		}
775
776		port->dev = dev;
777		hdlc = dev_to_hdlc(dev);
778		spin_lock_init(&port->lock);
779		dev->tx_queue_len = 50;
780		dev->netdev_ops = &wanxl_ops;
781		hdlc->attach = wanxl_attach;
782		hdlc->xmit = wanxl_xmit;
783		port->card = card;
784		port->node = i;
785		get_status(port)->clocking = CLOCK_EXT;
786		if (register_hdlc_device(dev)) {
787			pr_err("%s: unable to register hdlc device\n",
788			       pci_name(pdev));
789			free_netdev(dev);
790			wanxl_pci_remove_one(pdev);
791			return -ENOBUFS;
792		}
793		card->n_ports++;
794	}
795
796	pr_info("%s: port", pci_name(pdev));
797	for (i = 0; i < ports; i++)
798		pr_cont("%s #%i: %s",
799			i ? "," : "", i, card->ports[i].dev->name);
800	pr_cont("\n");
801
802	for (i = 0; i < ports; i++)
803		wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
804
805	return 0;
806}
807
808static const struct pci_device_id wanxl_pci_tbl[] = {
809	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
810	  PCI_ANY_ID, 0, 0, 0 },
811	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
812	  PCI_ANY_ID, 0, 0, 0 },
813	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
814	  PCI_ANY_ID, 0, 0, 0 },
815	{ 0, }
816};
817
 
818static struct pci_driver wanxl_pci_driver = {
819	.name		= "wanXL",
820	.id_table	= wanxl_pci_tbl,
821	.probe		= wanxl_pci_init_one,
822	.remove		= wanxl_pci_remove_one,
823};
824
 
825static int __init wanxl_init_module(void)
826{
827#ifdef MODULE
828	pr_info("%s\n", version);
829#endif
830	return pci_register_driver(&wanxl_pci_driver);
831}
832
833static void __exit wanxl_cleanup_module(void)
834{
835	pci_unregister_driver(&wanxl_pci_driver);
836}
 
837
838MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
839MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
840MODULE_LICENSE("GPL v2");
841MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
842
843module_init(wanxl_init_module);
844module_exit(wanxl_cleanup_module);
v4.6
 
  1/*
  2 * wanXL serial card driver for Linux
  3 * host part
  4 *
  5 * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
  6 *
  7 * This program is free software; you can redistribute it and/or modify it
  8 * under the terms of version 2 of the GNU General Public License
  9 * as published by the Free Software Foundation.
 10 *
 11 * Status:
 12 *   - Only DTE (external clock) support with NRZ and NRZI encodings
 13 *   - wanXL100 will require minor driver modifications, no access to hw
 14 */
 15
 16#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
 17
 18#include <linux/module.h>
 19#include <linux/kernel.h>
 20#include <linux/slab.h>
 21#include <linux/sched.h>
 22#include <linux/types.h>
 23#include <linux/fcntl.h>
 24#include <linux/string.h>
 25#include <linux/errno.h>
 26#include <linux/init.h>
 27#include <linux/interrupt.h>
 28#include <linux/ioport.h>
 29#include <linux/netdevice.h>
 30#include <linux/hdlc.h>
 31#include <linux/pci.h>
 32#include <linux/dma-mapping.h>
 33#include <linux/delay.h>
 34#include <asm/io.h>
 35
 36#include "wanxl.h"
 37
 38static const char* version = "wanXL serial card driver version: 0.48";
 39
 40#define PLX_CTL_RESET   0x40000000 /* adapter reset */
 41
 42#undef DEBUG_PKT
 43#undef DEBUG_PCI
 44
 45/* MAILBOX #1 - PUTS COMMANDS */
 46#define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
 47#ifdef __LITTLE_ENDIAN
 48#define MBX1_CMD_BSWAP  0x8C000001 /* little-endian Byte Swap Mode */
 49#else
 50#define MBX1_CMD_BSWAP  0x8C000000 /* big-endian Byte Swap Mode */
 51#endif
 52
 53/* MAILBOX #2 - DRAM SIZE */
 54#define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
 55
 56
 57struct port {
 58	struct net_device *dev;
 59	struct card *card;
 60	spinlock_t lock;	/* for wanxl_xmit */
 61        int node;		/* physical port #0 - 3 */
 62	unsigned int clock_type;
 63	int tx_in, tx_out;
 64	struct sk_buff *tx_skbs[TX_BUFFERS];
 65};
 66
 67
 68struct card_status {
 69	desc_t rx_descs[RX_QUEUE_LENGTH];
 70	port_status_t port_status[4];
 71};
 72
 73
 74struct card {
 75	int n_ports;		/* 1, 2 or 4 ports */
 76	u8 irq;
 77
 78	u8 __iomem *plx;	/* PLX PCI9060 virtual base address */
 79	struct pci_dev *pdev;	/* for pci_name(pdev) */
 80	int rx_in;
 81	struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
 82	struct card_status *status;	/* shared between host and card */
 83	dma_addr_t status_address;
 84	struct port ports[0];	/* 1 - 4 port structures follow */
 85};
 86
 87
 88
 89static inline struct port *dev_to_port(struct net_device *dev)
 90{
 91	return (struct port *)dev_to_hdlc(dev)->priv;
 92}
 93
 94
 95static inline port_status_t *get_status(struct port *port)
 96{
 97	return &port->card->status->port_status[port->node];
 98}
 99
100
101#ifdef DEBUG_PCI
102static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
103					      size_t size, int direction)
104{
105	dma_addr_t addr = pci_map_single(pdev, ptr, size, direction);
 
106	if (addr + size > 0x100000000LL)
107		pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
108			pci_name(pdev), (unsigned long long)addr);
109	return addr;
110}
111
112#undef pci_map_single
113#define pci_map_single pci_map_single_debug
114#endif
115
116
117/* Cable and/or personality module change interrupt service */
118static inline void wanxl_cable_intr(struct port *port)
119{
120	u32 value = get_status(port)->cable;
121	int valid = 1;
122	const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
123
124	switch(value & 0x7) {
125	case STATUS_CABLE_V35: cable = "V.35"; break;
126	case STATUS_CABLE_X21: cable = "X.21"; break;
127	case STATUS_CABLE_V24: cable = "V.24"; break;
128	case STATUS_CABLE_EIA530: cable = "EIA530"; break;
129	case STATUS_CABLE_NONE: cable = "no"; break;
130	default: cable = "invalid";
131	}
132
133	switch((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
134	case STATUS_CABLE_V35: pm = "V.35"; break;
135	case STATUS_CABLE_X21: pm = "X.21"; break;
136	case STATUS_CABLE_V24: pm = "V.24"; break;
137	case STATUS_CABLE_EIA530: pm = "EIA530"; break;
138	case STATUS_CABLE_NONE: pm = "no personality"; valid = 0; break;
139	default: pm = "invalid personality"; valid = 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
140	}
141
142	if (valid) {
143		if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
144			dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
145				", DSR off";
146			dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
147				", carrier off";
148		}
149		dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
150	}
151	netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
152		    pm, dte, cable, dsr, dcd);
153
154	if (value & STATUS_CABLE_DCD)
155		netif_carrier_on(port->dev);
156	else
157		netif_carrier_off(port->dev);
158}
159
160
161
162/* Transmit complete interrupt service */
163static inline void wanxl_tx_intr(struct port *port)
164{
165	struct net_device *dev = port->dev;
 
166	while (1) {
167                desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
168		struct sk_buff *skb = port->tx_skbs[port->tx_in];
169
170		switch (desc->stat) {
171		case PACKET_FULL:
172		case PACKET_EMPTY:
173			netif_wake_queue(dev);
174			return;
175
176		case PACKET_UNDERRUN:
177			dev->stats.tx_errors++;
178			dev->stats.tx_fifo_errors++;
179			break;
180
181		default:
182			dev->stats.tx_packets++;
183			dev->stats.tx_bytes += skb->len;
184		}
185                desc->stat = PACKET_EMPTY; /* Free descriptor */
186		pci_unmap_single(port->card->pdev, desc->address, skb->len,
187				 PCI_DMA_TODEVICE);
188		dev_kfree_skb_irq(skb);
189                port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
190        }
191}
192
193
194
195/* Receive complete interrupt service */
196static inline void wanxl_rx_intr(struct card *card)
197{
198	desc_t *desc;
 
199	while (desc = &card->status->rx_descs[card->rx_in],
200	       desc->stat != PACKET_EMPTY) {
201		if ((desc->stat & PACKET_PORT_MASK) > card->n_ports)
202			pr_crit("%s: received packet for nonexistent port\n",
203				pci_name(card->pdev));
204		else {
205			struct sk_buff *skb = card->rx_skbs[card->rx_in];
206			struct port *port = &card->ports[desc->stat &
207						    PACKET_PORT_MASK];
208			struct net_device *dev = port->dev;
209
210			if (!skb)
211				dev->stats.rx_dropped++;
212			else {
213				pci_unmap_single(card->pdev, desc->address,
214						 BUFFER_LENGTH,
215						 PCI_DMA_FROMDEVICE);
216				skb_put(skb, desc->length);
217
218#ifdef DEBUG_PKT
219				printk(KERN_DEBUG "%s RX(%i):", dev->name,
220				       skb->len);
221				debug_frame(skb);
222#endif
223				dev->stats.rx_packets++;
224				dev->stats.rx_bytes += skb->len;
225				skb->protocol = hdlc_type_trans(skb, dev);
226				netif_rx(skb);
227				skb = NULL;
228			}
229
230			if (!skb) {
231				skb = dev_alloc_skb(BUFFER_LENGTH);
232				desc->address = skb ?
233					pci_map_single(card->pdev, skb->data,
 
234						       BUFFER_LENGTH,
235						       PCI_DMA_FROMDEVICE) : 0;
236				card->rx_skbs[card->rx_in] = skb;
237			}
238		}
239		desc->stat = PACKET_EMPTY; /* Free descriptor */
240		card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
241	}
242}
243
244
245
246static irqreturn_t wanxl_intr(int irq, void* dev_id)
247{
248	struct card *card = dev_id;
249        int i;
250        u32 stat;
251        int handled = 0;
252
253
254        while((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
255                handled = 1;
256		writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
257
258                for (i = 0; i < card->n_ports; i++) {
259			if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
260				wanxl_tx_intr(&card->ports[i]);
261			if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
262				wanxl_cable_intr(&card->ports[i]);
263		}
264		if (stat & (1 << DOORBELL_FROM_CARD_RX))
265			wanxl_rx_intr(card);
266        }
267
268        return IRQ_RETVAL(handled);
269}
270
271
272
273static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
274{
275	struct port *port = dev_to_port(dev);
276	desc_t *desc;
277
278        spin_lock(&port->lock);
279
280	desc = &get_status(port)->tx_descs[port->tx_out];
281        if (desc->stat != PACKET_EMPTY) {
282                /* should never happen - previous xmit should stop queue */
283#ifdef DEBUG_PKT
284                printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
285#endif
286		netif_stop_queue(dev);
287		spin_unlock(&port->lock);
288		return NETDEV_TX_BUSY;       /* request packet to be queued */
289	}
290
291#ifdef DEBUG_PKT
292	printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
293	debug_frame(skb);
294#endif
295
296	port->tx_skbs[port->tx_out] = skb;
297	desc->address = pci_map_single(port->card->pdev, skb->data, skb->len,
298				       PCI_DMA_TODEVICE);
299	desc->length = skb->len;
300	desc->stat = PACKET_FULL;
301	writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
302	       port->card->plx + PLX_DOORBELL_TO_CARD);
303
304	port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
305
306	if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
307		netif_stop_queue(dev);
308#ifdef DEBUG_PKT
309		printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
310#endif
311	}
312
313	spin_unlock(&port->lock);
314	return NETDEV_TX_OK;
315}
316
317
318
319static int wanxl_attach(struct net_device *dev, unsigned short encoding,
320			unsigned short parity)
321{
322	struct port *port = dev_to_port(dev);
323
324	if (encoding != ENCODING_NRZ &&
325	    encoding != ENCODING_NRZI)
326		return -EINVAL;
327
328	if (parity != PARITY_NONE &&
329	    parity != PARITY_CRC32_PR1_CCITT &&
330	    parity != PARITY_CRC16_PR1_CCITT &&
331	    parity != PARITY_CRC32_PR0_CCITT &&
332	    parity != PARITY_CRC16_PR0_CCITT)
333		return -EINVAL;
334
335	get_status(port)->encoding = encoding;
336	get_status(port)->parity = parity;
337	return 0;
338}
339
340
341
342static int wanxl_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
343{
344	const size_t size = sizeof(sync_serial_settings);
345	sync_serial_settings line;
346	struct port *port = dev_to_port(dev);
347
348	if (cmd != SIOCWANDEV)
349		return hdlc_ioctl(dev, ifr, cmd);
350
351	switch (ifr->ifr_settings.type) {
352	case IF_GET_IFACE:
353		ifr->ifr_settings.type = IF_IFACE_SYNC_SERIAL;
354		if (ifr->ifr_settings.size < size) {
355			ifr->ifr_settings.size = size; /* data size wanted */
356			return -ENOBUFS;
357		}
358		memset(&line, 0, sizeof(line));
359		line.clock_type = get_status(port)->clocking;
360		line.clock_rate = 0;
361		line.loopback = 0;
362
363		if (copy_to_user(ifr->ifr_settings.ifs_ifsu.sync, &line, size))
364			return -EFAULT;
365		return 0;
366
367	case IF_IFACE_SYNC_SERIAL:
368		if (!capable(CAP_NET_ADMIN))
369			return -EPERM;
370		if (dev->flags & IFF_UP)
371			return -EBUSY;
372
373		if (copy_from_user(&line, ifr->ifr_settings.ifs_ifsu.sync,
374				   size))
375			return -EFAULT;
376
377		if (line.clock_type != CLOCK_EXT &&
378		    line.clock_type != CLOCK_TXFROMRX)
379			return -EINVAL; /* No such clock setting */
380
381		if (line.loopback != 0)
382			return -EINVAL;
383
384		get_status(port)->clocking = line.clock_type;
385		return 0;
386
387	default:
388		return hdlc_ioctl(dev, ifr, cmd);
389        }
390}
391
392
393
394static int wanxl_open(struct net_device *dev)
395{
396	struct port *port = dev_to_port(dev);
397	u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
398	unsigned long timeout;
399	int i;
400
401	if (get_status(port)->open) {
402		netdev_err(dev, "port already open\n");
403		return -EIO;
404	}
405	if ((i = hdlc_open(dev)) != 0)
 
 
406		return i;
407
408	port->tx_in = port->tx_out = 0;
409	for (i = 0; i < TX_BUFFERS; i++)
410		get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
411	/* signal the card */
412	writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
413
414	timeout = jiffies + HZ;
415	do {
416		if (get_status(port)->open) {
417			netif_start_queue(dev);
418			return 0;
419		}
420	} while (time_after(timeout, jiffies));
421
422	netdev_err(dev, "unable to open port\n");
423	/* ask the card to close the port, should it be still alive */
424	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
425	return -EFAULT;
426}
427
428
429
430static int wanxl_close(struct net_device *dev)
431{
432	struct port *port = dev_to_port(dev);
433	unsigned long timeout;
434	int i;
435
436	hdlc_close(dev);
437	/* signal the card */
438	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
439	       port->card->plx + PLX_DOORBELL_TO_CARD);
440
441	timeout = jiffies + HZ;
442	do {
443		if (!get_status(port)->open)
444			break;
445	} while (time_after(timeout, jiffies));
446
447	if (get_status(port)->open)
448		netdev_err(dev, "unable to close port\n");
449
450	netif_stop_queue(dev);
451
452	for (i = 0; i < TX_BUFFERS; i++) {
453		desc_t *desc = &get_status(port)->tx_descs[i];
454
455		if (desc->stat != PACKET_EMPTY) {
456			desc->stat = PACKET_EMPTY;
457			pci_unmap_single(port->card->pdev, desc->address,
458					 port->tx_skbs[i]->len,
459					 PCI_DMA_TODEVICE);
460			dev_kfree_skb(port->tx_skbs[i]);
461		}
462	}
463	return 0;
464}
465
466
467
468static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
469{
470	struct port *port = dev_to_port(dev);
471
472	dev->stats.rx_over_errors = get_status(port)->rx_overruns;
473	dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
474	dev->stats.rx_errors = dev->stats.rx_over_errors +
475		dev->stats.rx_frame_errors;
476	return &dev->stats;
477}
478
479
480
481static int wanxl_puts_command(struct card *card, u32 cmd)
482{
483	unsigned long timeout = jiffies + 5 * HZ;
484
485	writel(cmd, card->plx + PLX_MAILBOX_1);
486	do {
487		if (readl(card->plx + PLX_MAILBOX_1) == 0)
488			return 0;
489
490		schedule();
491	}while (time_after(timeout, jiffies));
492
493	return -1;
494}
495
496
497
498static void wanxl_reset(struct card *card)
499{
500	u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
501
502	writel(0x80, card->plx + PLX_MAILBOX_0);
503	writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
504	readl(card->plx + PLX_CONTROL); /* wait for posted write */
505	udelay(1);
506	writel(old_value, card->plx + PLX_CONTROL);
507	readl(card->plx + PLX_CONTROL); /* wait for posted write */
508}
509
510
511
512static void wanxl_pci_remove_one(struct pci_dev *pdev)
513{
514	struct card *card = pci_get_drvdata(pdev);
515	int i;
516
517	for (i = 0; i < card->n_ports; i++) {
518		unregister_hdlc_device(card->ports[i].dev);
519		free_netdev(card->ports[i].dev);
520	}
521
522	/* unregister and free all host resources */
523	if (card->irq)
524		free_irq(card->irq, card);
525
526	wanxl_reset(card);
527
528	for (i = 0; i < RX_QUEUE_LENGTH; i++)
529		if (card->rx_skbs[i]) {
530			pci_unmap_single(card->pdev,
531					 card->status->rx_descs[i].address,
532					 BUFFER_LENGTH, PCI_DMA_FROMDEVICE);
533			dev_kfree_skb(card->rx_skbs[i]);
534		}
535
536	if (card->plx)
537		iounmap(card->plx);
538
539	if (card->status)
540		pci_free_consistent(pdev, sizeof(struct card_status),
541				    card->status, card->status_address);
542
543	pci_release_regions(pdev);
544	pci_disable_device(pdev);
545	kfree(card);
546}
547
548
549#include "wanxlfw.inc"
550
551static const struct net_device_ops wanxl_ops = {
552	.ndo_open       = wanxl_open,
553	.ndo_stop       = wanxl_close,
554	.ndo_change_mtu = hdlc_change_mtu,
555	.ndo_start_xmit = hdlc_start_xmit,
556	.ndo_do_ioctl   = wanxl_ioctl,
557	.ndo_get_stats  = wanxl_get_stats,
558};
559
560static int wanxl_pci_init_one(struct pci_dev *pdev,
561			      const struct pci_device_id *ent)
562{
563	struct card *card;
564	u32 ramsize, stat;
565	unsigned long timeout;
566	u32 plx_phy;		/* PLX PCI base address */
567	u32 mem_phy;		/* memory PCI base addr */
568	u8 __iomem *mem;	/* memory virtual base addr */
569	int i, ports, alloc_size;
570
571#ifndef MODULE
572	pr_info_once("%s\n", version);
573#endif
574
575	i = pci_enable_device(pdev);
576	if (i)
577		return i;
578
579	/* QUICC can only access first 256 MB of host RAM directly,
580	   but PLX9060 DMA does 32-bits for actual packet data transfers */
 
581
582	/* FIXME when PCI/DMA subsystems are fixed.
583	   We set both dma_mask and consistent_dma_mask to 28 bits
584	   and pray pci_alloc_consistent() will use this info. It should
585	   work on most platforms */
586	if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(28)) ||
587	    pci_set_dma_mask(pdev, DMA_BIT_MASK(28))) {
 
588		pr_err("No usable DMA configuration\n");
589		pci_disable_device(pdev);
590		return -EIO;
591	}
592
593	i = pci_request_regions(pdev, "wanXL");
594	if (i) {
595		pci_disable_device(pdev);
596		return i;
597	}
598
599	switch (pdev->device) {
600	case PCI_DEVICE_ID_SBE_WANXL100: ports = 1; break;
601	case PCI_DEVICE_ID_SBE_WANXL200: ports = 2; break;
602	default: ports = 4;
 
 
 
 
 
603	}
604
605	alloc_size = sizeof(struct card) + ports * sizeof(struct port);
606	card = kzalloc(alloc_size, GFP_KERNEL);
607	if (card == NULL) {
608		pci_release_regions(pdev);
609		pci_disable_device(pdev);
610		return -ENOBUFS;
611	}
612
613	pci_set_drvdata(pdev, card);
614	card->pdev = pdev;
615
616	card->status = pci_alloc_consistent(pdev,
617					    sizeof(struct card_status),
618					    &card->status_address);
619	if (card->status == NULL) {
620		wanxl_pci_remove_one(pdev);
621		return -ENOBUFS;
622	}
623
624#ifdef DEBUG_PCI
625	printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
626	       " at 0x%LX\n", pci_name(pdev),
627	       (unsigned long long)card->status_address);
628#endif
629
630	/* FIXME when PCI/DMA subsystems are fixed.
631	   We set both dma_mask and consistent_dma_mask back to 32 bits
632	   to indicate the card can do 32-bit DMA addressing */
633	if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) ||
634	    pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
 
635		pr_err("No usable DMA configuration\n");
636		wanxl_pci_remove_one(pdev);
637		return -EIO;
638	}
639
640	/* set up PLX mapping */
641	plx_phy = pci_resource_start(pdev, 0);
642
643	card->plx = ioremap_nocache(plx_phy, 0x70);
644	if (!card->plx) {
645		pr_err("ioremap() failed\n");
646 		wanxl_pci_remove_one(pdev);
647		return -EFAULT;
648	}
649
650#if RESET_WHILE_LOADING
651	wanxl_reset(card);
652#endif
653
654	timeout = jiffies + 20 * HZ;
655	while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
656		if (time_before(timeout, jiffies)) {
657			pr_warn("%s: timeout waiting for PUTS to complete\n",
658				pci_name(pdev));
659			wanxl_pci_remove_one(pdev);
660			return -ENODEV;
661		}
662
663		switch(stat & 0xC0) {
664		case 0x00:	/* hmm - PUTS completed with non-zero code? */
665		case 0x80:	/* PUTS still testing the hardware */
666			break;
667
668		default:
669			pr_warn("%s: PUTS test 0x%X failed\n",
670				pci_name(pdev), stat & 0x30);
671			wanxl_pci_remove_one(pdev);
672			return -ENODEV;
673		}
674
675		schedule();
676	}
677
678	/* get on-board memory size (PUTS detects no more than 4 MB) */
679	ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
680
681	/* set up on-board RAM mapping */
682	mem_phy = pci_resource_start(pdev, 2);
683
684
685	/* sanity check the board's reported memory size */
686	if (ramsize < BUFFERS_ADDR +
687	    (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
688		pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
689			pci_name(pdev), ramsize,
690			BUFFERS_ADDR +
691			(TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
692		wanxl_pci_remove_one(pdev);
693		return -ENODEV;
694	}
695
696	if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
697		pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
698		wanxl_pci_remove_one(pdev);
699		return -ENODEV;
700	}
701
702	for (i = 0; i < RX_QUEUE_LENGTH; i++) {
703		struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
 
704		card->rx_skbs[i] = skb;
705		if (skb)
706			card->status->rx_descs[i].address =
707				pci_map_single(card->pdev, skb->data,
708					       BUFFER_LENGTH,
709					       PCI_DMA_FROMDEVICE);
710	}
711
712	mem = ioremap_nocache(mem_phy, PDM_OFFSET + sizeof(firmware));
713	if (!mem) {
714		pr_err("ioremap() failed\n");
715 		wanxl_pci_remove_one(pdev);
716		return -EFAULT;
717	}
718
719	for (i = 0; i < sizeof(firmware); i += 4)
720		writel(ntohl(*(__be32*)(firmware + i)), mem + PDM_OFFSET + i);
721
722	for (i = 0; i < ports; i++)
723		writel(card->status_address +
724		       (void *)&card->status->port_status[i] -
725		       (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
726	writel(card->status_address, mem + PDM_OFFSET + 20);
727	writel(PDM_OFFSET, mem);
728	iounmap(mem);
729
730	writel(0, card->plx + PLX_MAILBOX_5);
731
732	if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
733		pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
734		wanxl_pci_remove_one(pdev);
735		return -ENODEV;
736	}
737
738	stat = 0;
739	timeout = jiffies + 5 * HZ;
740	do {
741		if ((stat = readl(card->plx + PLX_MAILBOX_5)) != 0)
 
742			break;
743		schedule();
744	}while (time_after(timeout, jiffies));
745
746	if (!stat) {
747		pr_warn("%s: timeout while initializing card firmware\n",
748			pci_name(pdev));
749		wanxl_pci_remove_one(pdev);
750		return -ENODEV;
751	}
752
753#if DETECT_RAM
754	ramsize = stat;
755#endif
756
757	pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
758		pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
759
760	/* Allocate IRQ */
761	if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
762		pr_warn("%s: could not allocate IRQ%i\n",
763			pci_name(pdev), pdev->irq);
764		wanxl_pci_remove_one(pdev);
765		return -EBUSY;
766	}
767	card->irq = pdev->irq;
768
769	for (i = 0; i < ports; i++) {
770		hdlc_device *hdlc;
771		struct port *port = &card->ports[i];
772		struct net_device *dev = alloc_hdlcdev(port);
 
773		if (!dev) {
774			pr_err("%s: unable to allocate memory\n",
775			       pci_name(pdev));
776			wanxl_pci_remove_one(pdev);
777			return -ENOMEM;
778		}
779
780		port->dev = dev;
781		hdlc = dev_to_hdlc(dev);
782		spin_lock_init(&port->lock);
783		dev->tx_queue_len = 50;
784		dev->netdev_ops = &wanxl_ops;
785		hdlc->attach = wanxl_attach;
786		hdlc->xmit = wanxl_xmit;
787		port->card = card;
788		port->node = i;
789		get_status(port)->clocking = CLOCK_EXT;
790		if (register_hdlc_device(dev)) {
791			pr_err("%s: unable to register hdlc device\n",
792			       pci_name(pdev));
793			free_netdev(dev);
794			wanxl_pci_remove_one(pdev);
795			return -ENOBUFS;
796		}
797		card->n_ports++;
798	}
799
800	pr_info("%s: port", pci_name(pdev));
801	for (i = 0; i < ports; i++)
802		pr_cont("%s #%i: %s",
803			i ? "," : "", i, card->ports[i].dev->name);
804	pr_cont("\n");
805
806	for (i = 0; i < ports; i++)
807		wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
808
809	return 0;
810}
811
812static const struct pci_device_id wanxl_pci_tbl[] = {
813	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
814	  PCI_ANY_ID, 0, 0, 0 },
815	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
816	  PCI_ANY_ID, 0, 0, 0 },
817	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
818	  PCI_ANY_ID, 0, 0, 0 },
819	{ 0, }
820};
821
822
823static struct pci_driver wanxl_pci_driver = {
824	.name		= "wanXL",
825	.id_table	= wanxl_pci_tbl,
826	.probe		= wanxl_pci_init_one,
827	.remove		= wanxl_pci_remove_one,
828};
829
830
831static int __init wanxl_init_module(void)
832{
833#ifdef MODULE
834	pr_info("%s\n", version);
835#endif
836	return pci_register_driver(&wanxl_pci_driver);
837}
838
839static void __exit wanxl_cleanup_module(void)
840{
841	pci_unregister_driver(&wanxl_pci_driver);
842}
843
844
845MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
846MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
847MODULE_LICENSE("GPL v2");
848MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
849
850module_init(wanxl_init_module);
851module_exit(wanxl_cleanup_module);