Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * wanXL serial card driver for Linux
4 * host part
5 *
6 * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
7 *
8 * Status:
9 * - Only DTE (external clock) support with NRZ and NRZI encodings
10 * - wanXL100 will require minor driver modifications, no access to hw
11 */
12
13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15#include <linux/module.h>
16#include <linux/kernel.h>
17#include <linux/slab.h>
18#include <linux/sched.h>
19#include <linux/types.h>
20#include <linux/fcntl.h>
21#include <linux/string.h>
22#include <linux/errno.h>
23#include <linux/init.h>
24#include <linux/interrupt.h>
25#include <linux/ioport.h>
26#include <linux/netdevice.h>
27#include <linux/hdlc.h>
28#include <linux/pci.h>
29#include <linux/dma-mapping.h>
30#include <linux/delay.h>
31#include <asm/io.h>
32
33#include "wanxl.h"
34
35static const char *version = "wanXL serial card driver version: 0.48";
36
37#define PLX_CTL_RESET 0x40000000 /* adapter reset */
38
39#undef DEBUG_PKT
40#undef DEBUG_PCI
41
42/* MAILBOX #1 - PUTS COMMANDS */
43#define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
44#ifdef __LITTLE_ENDIAN
45#define MBX1_CMD_BSWAP 0x8C000001 /* little-endian Byte Swap Mode */
46#else
47#define MBX1_CMD_BSWAP 0x8C000000 /* big-endian Byte Swap Mode */
48#endif
49
50/* MAILBOX #2 - DRAM SIZE */
51#define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
52
53struct port {
54 struct net_device *dev;
55 struct card *card;
56 spinlock_t lock; /* for wanxl_xmit */
57 int node; /* physical port #0 - 3 */
58 unsigned int clock_type;
59 int tx_in, tx_out;
60 struct sk_buff *tx_skbs[TX_BUFFERS];
61};
62
63struct card_status {
64 desc_t rx_descs[RX_QUEUE_LENGTH];
65 port_status_t port_status[4];
66};
67
68struct card {
69 int n_ports; /* 1, 2 or 4 ports */
70 u8 irq;
71
72 u8 __iomem *plx; /* PLX PCI9060 virtual base address */
73 struct pci_dev *pdev; /* for pci_name(pdev) */
74 int rx_in;
75 struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
76 struct card_status *status; /* shared between host and card */
77 dma_addr_t status_address;
78 struct port ports[]; /* 1 - 4 port structures follow */
79};
80
81static inline struct port *dev_to_port(struct net_device *dev)
82{
83 return (struct port *)dev_to_hdlc(dev)->priv;
84}
85
86static inline port_status_t *get_status(struct port *port)
87{
88 return &port->card->status->port_status[port->node];
89}
90
91#ifdef DEBUG_PCI
92static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
93 size_t size, int direction)
94{
95 dma_addr_t addr = dma_map_single(&pdev->dev, ptr, size, direction);
96
97 if (addr + size > 0x100000000LL)
98 pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
99 pci_name(pdev), (unsigned long long)addr);
100 return addr;
101}
102
103#undef pci_map_single
104#define pci_map_single pci_map_single_debug
105#endif
106
107/* Cable and/or personality module change interrupt service */
108static inline void wanxl_cable_intr(struct port *port)
109{
110 u32 value = get_status(port)->cable;
111 int valid = 1;
112 const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
113
114 switch (value & 0x7) {
115 case STATUS_CABLE_V35:
116 cable = "V.35";
117 break;
118 case STATUS_CABLE_X21:
119 cable = "X.21";
120 break;
121 case STATUS_CABLE_V24:
122 cable = "V.24";
123 break;
124 case STATUS_CABLE_EIA530:
125 cable = "EIA530";
126 break;
127 case STATUS_CABLE_NONE:
128 cable = "no";
129 break;
130 default:
131 cable = "invalid";
132 }
133
134 switch ((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
135 case STATUS_CABLE_V35:
136 pm = "V.35";
137 break;
138 case STATUS_CABLE_X21:
139 pm = "X.21";
140 break;
141 case STATUS_CABLE_V24:
142 pm = "V.24";
143 break;
144 case STATUS_CABLE_EIA530:
145 pm = "EIA530";
146 break;
147 case STATUS_CABLE_NONE:
148 pm = "no personality";
149 valid = 0;
150 break;
151 default:
152 pm = "invalid personality";
153 valid = 0;
154 }
155
156 if (valid) {
157 if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
158 dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
159 ", DSR off";
160 dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
161 ", carrier off";
162 }
163 dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
164 }
165 netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
166 pm, dte, cable, dsr, dcd);
167
168 if (value & STATUS_CABLE_DCD)
169 netif_carrier_on(port->dev);
170 else
171 netif_carrier_off(port->dev);
172}
173
174/* Transmit complete interrupt service */
175static inline void wanxl_tx_intr(struct port *port)
176{
177 struct net_device *dev = port->dev;
178
179 while (1) {
180 desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
181 struct sk_buff *skb = port->tx_skbs[port->tx_in];
182
183 switch (desc->stat) {
184 case PACKET_FULL:
185 case PACKET_EMPTY:
186 netif_wake_queue(dev);
187 return;
188
189 case PACKET_UNDERRUN:
190 dev->stats.tx_errors++;
191 dev->stats.tx_fifo_errors++;
192 break;
193
194 default:
195 dev->stats.tx_packets++;
196 dev->stats.tx_bytes += skb->len;
197 }
198 desc->stat = PACKET_EMPTY; /* Free descriptor */
199 dma_unmap_single(&port->card->pdev->dev, desc->address,
200 skb->len, DMA_TO_DEVICE);
201 dev_consume_skb_irq(skb);
202 port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
203 }
204}
205
206/* Receive complete interrupt service */
207static inline void wanxl_rx_intr(struct card *card)
208{
209 desc_t *desc;
210
211 while (desc = &card->status->rx_descs[card->rx_in],
212 desc->stat != PACKET_EMPTY) {
213 if ((desc->stat & PACKET_PORT_MASK) > card->n_ports) {
214 pr_crit("%s: received packet for nonexistent port\n",
215 pci_name(card->pdev));
216 } else {
217 struct sk_buff *skb = card->rx_skbs[card->rx_in];
218 struct port *port = &card->ports[desc->stat &
219 PACKET_PORT_MASK];
220 struct net_device *dev = port->dev;
221
222 if (!skb) {
223 dev->stats.rx_dropped++;
224 } else {
225 dma_unmap_single(&card->pdev->dev,
226 desc->address, BUFFER_LENGTH,
227 DMA_FROM_DEVICE);
228 skb_put(skb, desc->length);
229
230#ifdef DEBUG_PKT
231 printk(KERN_DEBUG "%s RX(%i):", dev->name,
232 skb->len);
233 debug_frame(skb);
234#endif
235 dev->stats.rx_packets++;
236 dev->stats.rx_bytes += skb->len;
237 skb->protocol = hdlc_type_trans(skb, dev);
238 netif_rx(skb);
239 skb = NULL;
240 }
241
242 if (!skb) {
243 skb = dev_alloc_skb(BUFFER_LENGTH);
244 desc->address = skb ?
245 dma_map_single(&card->pdev->dev,
246 skb->data,
247 BUFFER_LENGTH,
248 DMA_FROM_DEVICE) : 0;
249 card->rx_skbs[card->rx_in] = skb;
250 }
251 }
252 desc->stat = PACKET_EMPTY; /* Free descriptor */
253 card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
254 }
255}
256
257static irqreturn_t wanxl_intr(int irq, void *dev_id)
258{
259 struct card *card = dev_id;
260 int i;
261 u32 stat;
262 int handled = 0;
263
264 while ((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
265 handled = 1;
266 writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
267
268 for (i = 0; i < card->n_ports; i++) {
269 if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
270 wanxl_tx_intr(&card->ports[i]);
271 if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
272 wanxl_cable_intr(&card->ports[i]);
273 }
274 if (stat & (1 << DOORBELL_FROM_CARD_RX))
275 wanxl_rx_intr(card);
276 }
277
278 return IRQ_RETVAL(handled);
279}
280
281static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
282{
283 struct port *port = dev_to_port(dev);
284 desc_t *desc;
285
286 spin_lock(&port->lock);
287
288 desc = &get_status(port)->tx_descs[port->tx_out];
289 if (desc->stat != PACKET_EMPTY) {
290 /* should never happen - previous xmit should stop queue */
291#ifdef DEBUG_PKT
292 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
293#endif
294 netif_stop_queue(dev);
295 spin_unlock(&port->lock);
296 return NETDEV_TX_BUSY; /* request packet to be queued */
297 }
298
299#ifdef DEBUG_PKT
300 printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
301 debug_frame(skb);
302#endif
303
304 port->tx_skbs[port->tx_out] = skb;
305 desc->address = dma_map_single(&port->card->pdev->dev, skb->data,
306 skb->len, DMA_TO_DEVICE);
307 desc->length = skb->len;
308 desc->stat = PACKET_FULL;
309 writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
310 port->card->plx + PLX_DOORBELL_TO_CARD);
311
312 port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
313
314 if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
315 netif_stop_queue(dev);
316#ifdef DEBUG_PKT
317 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
318#endif
319 }
320
321 spin_unlock(&port->lock);
322 return NETDEV_TX_OK;
323}
324
325static int wanxl_attach(struct net_device *dev, unsigned short encoding,
326 unsigned short parity)
327{
328 struct port *port = dev_to_port(dev);
329
330 if (encoding != ENCODING_NRZ &&
331 encoding != ENCODING_NRZI)
332 return -EINVAL;
333
334 if (parity != PARITY_NONE &&
335 parity != PARITY_CRC32_PR1_CCITT &&
336 parity != PARITY_CRC16_PR1_CCITT &&
337 parity != PARITY_CRC32_PR0_CCITT &&
338 parity != PARITY_CRC16_PR0_CCITT)
339 return -EINVAL;
340
341 get_status(port)->encoding = encoding;
342 get_status(port)->parity = parity;
343 return 0;
344}
345
346static int wanxl_ioctl(struct net_device *dev, struct if_settings *ifs)
347{
348 const size_t size = sizeof(sync_serial_settings);
349 sync_serial_settings line;
350 struct port *port = dev_to_port(dev);
351
352 switch (ifs->type) {
353 case IF_GET_IFACE:
354 ifs->type = IF_IFACE_SYNC_SERIAL;
355 if (ifs->size < size) {
356 ifs->size = size; /* data size wanted */
357 return -ENOBUFS;
358 }
359 memset(&line, 0, sizeof(line));
360 line.clock_type = get_status(port)->clocking;
361 line.clock_rate = 0;
362 line.loopback = 0;
363
364 if (copy_to_user(ifs->ifs_ifsu.sync, &line, size))
365 return -EFAULT;
366 return 0;
367
368 case IF_IFACE_SYNC_SERIAL:
369 if (!capable(CAP_NET_ADMIN))
370 return -EPERM;
371 if (dev->flags & IFF_UP)
372 return -EBUSY;
373
374 if (copy_from_user(&line, ifs->ifs_ifsu.sync,
375 size))
376 return -EFAULT;
377
378 if (line.clock_type != CLOCK_EXT &&
379 line.clock_type != CLOCK_TXFROMRX)
380 return -EINVAL; /* No such clock setting */
381
382 if (line.loopback != 0)
383 return -EINVAL;
384
385 get_status(port)->clocking = line.clock_type;
386 return 0;
387
388 default:
389 return hdlc_ioctl(dev, ifs);
390 }
391}
392
393static int wanxl_open(struct net_device *dev)
394{
395 struct port *port = dev_to_port(dev);
396 u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
397 unsigned long timeout;
398 int i;
399
400 if (get_status(port)->open) {
401 netdev_err(dev, "port already open\n");
402 return -EIO;
403 }
404
405 i = hdlc_open(dev);
406 if (i)
407 return i;
408
409 port->tx_in = port->tx_out = 0;
410 for (i = 0; i < TX_BUFFERS; i++)
411 get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
412 /* signal the card */
413 writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
414
415 timeout = jiffies + HZ;
416 do {
417 if (get_status(port)->open) {
418 netif_start_queue(dev);
419 return 0;
420 }
421 } while (time_after(timeout, jiffies));
422
423 netdev_err(dev, "unable to open port\n");
424 /* ask the card to close the port, should it be still alive */
425 writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
426 return -EFAULT;
427}
428
429static int wanxl_close(struct net_device *dev)
430{
431 struct port *port = dev_to_port(dev);
432 unsigned long timeout;
433 int i;
434
435 hdlc_close(dev);
436 /* signal the card */
437 writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
438 port->card->plx + PLX_DOORBELL_TO_CARD);
439
440 timeout = jiffies + HZ;
441 do {
442 if (!get_status(port)->open)
443 break;
444 } while (time_after(timeout, jiffies));
445
446 if (get_status(port)->open)
447 netdev_err(dev, "unable to close port\n");
448
449 netif_stop_queue(dev);
450
451 for (i = 0; i < TX_BUFFERS; i++) {
452 desc_t *desc = &get_status(port)->tx_descs[i];
453
454 if (desc->stat != PACKET_EMPTY) {
455 desc->stat = PACKET_EMPTY;
456 dma_unmap_single(&port->card->pdev->dev,
457 desc->address, port->tx_skbs[i]->len,
458 DMA_TO_DEVICE);
459 dev_kfree_skb(port->tx_skbs[i]);
460 }
461 }
462 return 0;
463}
464
465static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
466{
467 struct port *port = dev_to_port(dev);
468
469 dev->stats.rx_over_errors = get_status(port)->rx_overruns;
470 dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
471 dev->stats.rx_errors = dev->stats.rx_over_errors +
472 dev->stats.rx_frame_errors;
473 return &dev->stats;
474}
475
476static int wanxl_puts_command(struct card *card, u32 cmd)
477{
478 unsigned long timeout = jiffies + 5 * HZ;
479
480 writel(cmd, card->plx + PLX_MAILBOX_1);
481 do {
482 if (readl(card->plx + PLX_MAILBOX_1) == 0)
483 return 0;
484
485 schedule();
486 } while (time_after(timeout, jiffies));
487
488 return -1;
489}
490
491static void wanxl_reset(struct card *card)
492{
493 u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
494
495 writel(0x80, card->plx + PLX_MAILBOX_0);
496 writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
497 readl(card->plx + PLX_CONTROL); /* wait for posted write */
498 udelay(1);
499 writel(old_value, card->plx + PLX_CONTROL);
500 readl(card->plx + PLX_CONTROL); /* wait for posted write */
501}
502
503static void wanxl_pci_remove_one(struct pci_dev *pdev)
504{
505 struct card *card = pci_get_drvdata(pdev);
506 int i;
507
508 for (i = 0; i < card->n_ports; i++) {
509 unregister_hdlc_device(card->ports[i].dev);
510 free_netdev(card->ports[i].dev);
511 }
512
513 /* unregister and free all host resources */
514 if (card->irq)
515 free_irq(card->irq, card);
516
517 wanxl_reset(card);
518
519 for (i = 0; i < RX_QUEUE_LENGTH; i++)
520 if (card->rx_skbs[i]) {
521 dma_unmap_single(&card->pdev->dev,
522 card->status->rx_descs[i].address,
523 BUFFER_LENGTH, DMA_FROM_DEVICE);
524 dev_kfree_skb(card->rx_skbs[i]);
525 }
526
527 if (card->plx)
528 iounmap(card->plx);
529
530 if (card->status)
531 dma_free_coherent(&pdev->dev, sizeof(struct card_status),
532 card->status, card->status_address);
533
534 pci_release_regions(pdev);
535 pci_disable_device(pdev);
536 kfree(card);
537}
538
539#include "wanxlfw.inc"
540
541static const struct net_device_ops wanxl_ops = {
542 .ndo_open = wanxl_open,
543 .ndo_stop = wanxl_close,
544 .ndo_start_xmit = hdlc_start_xmit,
545 .ndo_siocwandev = wanxl_ioctl,
546 .ndo_get_stats = wanxl_get_stats,
547};
548
549static int wanxl_pci_init_one(struct pci_dev *pdev,
550 const struct pci_device_id *ent)
551{
552 struct card *card;
553 u32 ramsize, stat;
554 unsigned long timeout;
555 u32 plx_phy; /* PLX PCI base address */
556 u32 mem_phy; /* memory PCI base addr */
557 u8 __iomem *mem; /* memory virtual base addr */
558 int i, ports;
559
560#ifndef MODULE
561 pr_info_once("%s\n", version);
562#endif
563
564 i = pci_enable_device(pdev);
565 if (i)
566 return i;
567
568 /* QUICC can only access first 256 MB of host RAM directly,
569 * but PLX9060 DMA does 32-bits for actual packet data transfers
570 */
571
572 /* FIXME when PCI/DMA subsystems are fixed.
573 * We set both dma_mask and consistent_dma_mask to 28 bits
574 * and pray pci_alloc_consistent() will use this info. It should
575 * work on most platforms
576 */
577 if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(28)) ||
578 dma_set_mask(&pdev->dev, DMA_BIT_MASK(28))) {
579 pr_err("No usable DMA configuration\n");
580 pci_disable_device(pdev);
581 return -EIO;
582 }
583
584 i = pci_request_regions(pdev, "wanXL");
585 if (i) {
586 pci_disable_device(pdev);
587 return i;
588 }
589
590 switch (pdev->device) {
591 case PCI_DEVICE_ID_SBE_WANXL100:
592 ports = 1;
593 break;
594 case PCI_DEVICE_ID_SBE_WANXL200:
595 ports = 2;
596 break;
597 default:
598 ports = 4;
599 }
600
601 card = kzalloc(struct_size(card, ports, ports), GFP_KERNEL);
602 if (!card) {
603 pci_release_regions(pdev);
604 pci_disable_device(pdev);
605 return -ENOBUFS;
606 }
607
608 pci_set_drvdata(pdev, card);
609 card->pdev = pdev;
610
611 card->status = dma_alloc_coherent(&pdev->dev,
612 sizeof(struct card_status),
613 &card->status_address, GFP_KERNEL);
614 if (!card->status) {
615 wanxl_pci_remove_one(pdev);
616 return -ENOBUFS;
617 }
618
619#ifdef DEBUG_PCI
620 printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
621 " at 0x%LX\n", pci_name(pdev),
622 (unsigned long long)card->status_address);
623#endif
624
625 /* FIXME when PCI/DMA subsystems are fixed.
626 * We set both dma_mask and consistent_dma_mask back to 32 bits
627 * to indicate the card can do 32-bit DMA addressing
628 */
629 if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)) ||
630 dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
631 pr_err("No usable DMA configuration\n");
632 wanxl_pci_remove_one(pdev);
633 return -EIO;
634 }
635
636 /* set up PLX mapping */
637 plx_phy = pci_resource_start(pdev, 0);
638
639 card->plx = ioremap(plx_phy, 0x70);
640 if (!card->plx) {
641 pr_err("ioremap() failed\n");
642 wanxl_pci_remove_one(pdev);
643 return -EFAULT;
644 }
645
646#if RESET_WHILE_LOADING
647 wanxl_reset(card);
648#endif
649
650 timeout = jiffies + 20 * HZ;
651 while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
652 if (time_before(timeout, jiffies)) {
653 pr_warn("%s: timeout waiting for PUTS to complete\n",
654 pci_name(pdev));
655 wanxl_pci_remove_one(pdev);
656 return -ENODEV;
657 }
658
659 switch (stat & 0xC0) {
660 case 0x00: /* hmm - PUTS completed with non-zero code? */
661 case 0x80: /* PUTS still testing the hardware */
662 break;
663
664 default:
665 pr_warn("%s: PUTS test 0x%X failed\n",
666 pci_name(pdev), stat & 0x30);
667 wanxl_pci_remove_one(pdev);
668 return -ENODEV;
669 }
670
671 schedule();
672 }
673
674 /* get on-board memory size (PUTS detects no more than 4 MB) */
675 ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
676
677 /* set up on-board RAM mapping */
678 mem_phy = pci_resource_start(pdev, 2);
679
680 /* sanity check the board's reported memory size */
681 if (ramsize < BUFFERS_ADDR +
682 (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
683 pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
684 pci_name(pdev), ramsize,
685 BUFFERS_ADDR +
686 (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
687 wanxl_pci_remove_one(pdev);
688 return -ENODEV;
689 }
690
691 if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
692 pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
693 wanxl_pci_remove_one(pdev);
694 return -ENODEV;
695 }
696
697 for (i = 0; i < RX_QUEUE_LENGTH; i++) {
698 struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
699
700 card->rx_skbs[i] = skb;
701 if (skb)
702 card->status->rx_descs[i].address =
703 dma_map_single(&card->pdev->dev, skb->data,
704 BUFFER_LENGTH, DMA_FROM_DEVICE);
705 }
706
707 mem = ioremap(mem_phy, PDM_OFFSET + sizeof(firmware));
708 if (!mem) {
709 pr_err("ioremap() failed\n");
710 wanxl_pci_remove_one(pdev);
711 return -EFAULT;
712 }
713
714 for (i = 0; i < sizeof(firmware); i += 4)
715 writel(ntohl(*(__be32 *)(firmware + i)), mem + PDM_OFFSET + i);
716
717 for (i = 0; i < ports; i++)
718 writel(card->status_address +
719 (void *)&card->status->port_status[i] -
720 (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
721 writel(card->status_address, mem + PDM_OFFSET + 20);
722 writel(PDM_OFFSET, mem);
723 iounmap(mem);
724
725 writel(0, card->plx + PLX_MAILBOX_5);
726
727 if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
728 pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
729 wanxl_pci_remove_one(pdev);
730 return -ENODEV;
731 }
732
733 timeout = jiffies + 5 * HZ;
734 do {
735 stat = readl(card->plx + PLX_MAILBOX_5);
736 if (stat)
737 break;
738 schedule();
739 } while (time_after(timeout, jiffies));
740
741 if (!stat) {
742 pr_warn("%s: timeout while initializing card firmware\n",
743 pci_name(pdev));
744 wanxl_pci_remove_one(pdev);
745 return -ENODEV;
746 }
747
748#if DETECT_RAM
749 ramsize = stat;
750#endif
751
752 pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
753 pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
754
755 /* Allocate IRQ */
756 if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
757 pr_warn("%s: could not allocate IRQ%i\n",
758 pci_name(pdev), pdev->irq);
759 wanxl_pci_remove_one(pdev);
760 return -EBUSY;
761 }
762 card->irq = pdev->irq;
763
764 for (i = 0; i < ports; i++) {
765 hdlc_device *hdlc;
766 struct port *port = &card->ports[i];
767 struct net_device *dev = alloc_hdlcdev(port);
768
769 if (!dev) {
770 pr_err("%s: unable to allocate memory\n",
771 pci_name(pdev));
772 wanxl_pci_remove_one(pdev);
773 return -ENOMEM;
774 }
775
776 port->dev = dev;
777 hdlc = dev_to_hdlc(dev);
778 spin_lock_init(&port->lock);
779 dev->tx_queue_len = 50;
780 dev->netdev_ops = &wanxl_ops;
781 hdlc->attach = wanxl_attach;
782 hdlc->xmit = wanxl_xmit;
783 port->card = card;
784 port->node = i;
785 get_status(port)->clocking = CLOCK_EXT;
786 if (register_hdlc_device(dev)) {
787 pr_err("%s: unable to register hdlc device\n",
788 pci_name(pdev));
789 free_netdev(dev);
790 wanxl_pci_remove_one(pdev);
791 return -ENOBUFS;
792 }
793 card->n_ports++;
794 }
795
796 pr_info("%s: port", pci_name(pdev));
797 for (i = 0; i < ports; i++)
798 pr_cont("%s #%i: %s",
799 i ? "," : "", i, card->ports[i].dev->name);
800 pr_cont("\n");
801
802 for (i = 0; i < ports; i++)
803 wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
804
805 return 0;
806}
807
808static const struct pci_device_id wanxl_pci_tbl[] = {
809 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
810 PCI_ANY_ID, 0, 0, 0 },
811 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
812 PCI_ANY_ID, 0, 0, 0 },
813 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
814 PCI_ANY_ID, 0, 0, 0 },
815 { 0, }
816};
817
818static struct pci_driver wanxl_pci_driver = {
819 .name = "wanXL",
820 .id_table = wanxl_pci_tbl,
821 .probe = wanxl_pci_init_one,
822 .remove = wanxl_pci_remove_one,
823};
824
825static int __init wanxl_init_module(void)
826{
827#ifdef MODULE
828 pr_info("%s\n", version);
829#endif
830 return pci_register_driver(&wanxl_pci_driver);
831}
832
833static void __exit wanxl_cleanup_module(void)
834{
835 pci_unregister_driver(&wanxl_pci_driver);
836}
837
838MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
839MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
840MODULE_LICENSE("GPL v2");
841MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
842
843module_init(wanxl_init_module);
844module_exit(wanxl_cleanup_module);
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * wanXL serial card driver for Linux
4 * host part
5 *
6 * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
7 *
8 * Status:
9 * - Only DTE (external clock) support with NRZ and NRZI encodings
10 * - wanXL100 will require minor driver modifications, no access to hw
11 */
12
13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15#include <linux/module.h>
16#include <linux/kernel.h>
17#include <linux/slab.h>
18#include <linux/sched.h>
19#include <linux/types.h>
20#include <linux/fcntl.h>
21#include <linux/string.h>
22#include <linux/errno.h>
23#include <linux/init.h>
24#include <linux/interrupt.h>
25#include <linux/ioport.h>
26#include <linux/netdevice.h>
27#include <linux/hdlc.h>
28#include <linux/pci.h>
29#include <linux/dma-mapping.h>
30#include <linux/delay.h>
31#include <asm/io.h>
32
33#include "wanxl.h"
34
35static const char* version = "wanXL serial card driver version: 0.48";
36
37#define PLX_CTL_RESET 0x40000000 /* adapter reset */
38
39#undef DEBUG_PKT
40#undef DEBUG_PCI
41
42/* MAILBOX #1 - PUTS COMMANDS */
43#define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
44#ifdef __LITTLE_ENDIAN
45#define MBX1_CMD_BSWAP 0x8C000001 /* little-endian Byte Swap Mode */
46#else
47#define MBX1_CMD_BSWAP 0x8C000000 /* big-endian Byte Swap Mode */
48#endif
49
50/* MAILBOX #2 - DRAM SIZE */
51#define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
52
53
54struct port {
55 struct net_device *dev;
56 struct card *card;
57 spinlock_t lock; /* for wanxl_xmit */
58 int node; /* physical port #0 - 3 */
59 unsigned int clock_type;
60 int tx_in, tx_out;
61 struct sk_buff *tx_skbs[TX_BUFFERS];
62};
63
64
65struct card_status {
66 desc_t rx_descs[RX_QUEUE_LENGTH];
67 port_status_t port_status[4];
68};
69
70
71struct card {
72 int n_ports; /* 1, 2 or 4 ports */
73 u8 irq;
74
75 u8 __iomem *plx; /* PLX PCI9060 virtual base address */
76 struct pci_dev *pdev; /* for pci_name(pdev) */
77 int rx_in;
78 struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
79 struct card_status *status; /* shared between host and card */
80 dma_addr_t status_address;
81 struct port ports[0]; /* 1 - 4 port structures follow */
82};
83
84
85
86static inline struct port *dev_to_port(struct net_device *dev)
87{
88 return (struct port *)dev_to_hdlc(dev)->priv;
89}
90
91
92static inline port_status_t *get_status(struct port *port)
93{
94 return &port->card->status->port_status[port->node];
95}
96
97
98#ifdef DEBUG_PCI
99static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
100 size_t size, int direction)
101{
102 dma_addr_t addr = pci_map_single(pdev, ptr, size, direction);
103 if (addr + size > 0x100000000LL)
104 pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
105 pci_name(pdev), (unsigned long long)addr);
106 return addr;
107}
108
109#undef pci_map_single
110#define pci_map_single pci_map_single_debug
111#endif
112
113
114/* Cable and/or personality module change interrupt service */
115static inline void wanxl_cable_intr(struct port *port)
116{
117 u32 value = get_status(port)->cable;
118 int valid = 1;
119 const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
120
121 switch(value & 0x7) {
122 case STATUS_CABLE_V35: cable = "V.35"; break;
123 case STATUS_CABLE_X21: cable = "X.21"; break;
124 case STATUS_CABLE_V24: cable = "V.24"; break;
125 case STATUS_CABLE_EIA530: cable = "EIA530"; break;
126 case STATUS_CABLE_NONE: cable = "no"; break;
127 default: cable = "invalid";
128 }
129
130 switch((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
131 case STATUS_CABLE_V35: pm = "V.35"; break;
132 case STATUS_CABLE_X21: pm = "X.21"; break;
133 case STATUS_CABLE_V24: pm = "V.24"; break;
134 case STATUS_CABLE_EIA530: pm = "EIA530"; break;
135 case STATUS_CABLE_NONE: pm = "no personality"; valid = 0; break;
136 default: pm = "invalid personality"; valid = 0;
137 }
138
139 if (valid) {
140 if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
141 dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
142 ", DSR off";
143 dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
144 ", carrier off";
145 }
146 dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
147 }
148 netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
149 pm, dte, cable, dsr, dcd);
150
151 if (value & STATUS_CABLE_DCD)
152 netif_carrier_on(port->dev);
153 else
154 netif_carrier_off(port->dev);
155}
156
157
158
159/* Transmit complete interrupt service */
160static inline void wanxl_tx_intr(struct port *port)
161{
162 struct net_device *dev = port->dev;
163 while (1) {
164 desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
165 struct sk_buff *skb = port->tx_skbs[port->tx_in];
166
167 switch (desc->stat) {
168 case PACKET_FULL:
169 case PACKET_EMPTY:
170 netif_wake_queue(dev);
171 return;
172
173 case PACKET_UNDERRUN:
174 dev->stats.tx_errors++;
175 dev->stats.tx_fifo_errors++;
176 break;
177
178 default:
179 dev->stats.tx_packets++;
180 dev->stats.tx_bytes += skb->len;
181 }
182 desc->stat = PACKET_EMPTY; /* Free descriptor */
183 pci_unmap_single(port->card->pdev, desc->address, skb->len,
184 PCI_DMA_TODEVICE);
185 dev_consume_skb_irq(skb);
186 port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
187 }
188}
189
190
191
192/* Receive complete interrupt service */
193static inline void wanxl_rx_intr(struct card *card)
194{
195 desc_t *desc;
196 while (desc = &card->status->rx_descs[card->rx_in],
197 desc->stat != PACKET_EMPTY) {
198 if ((desc->stat & PACKET_PORT_MASK) > card->n_ports)
199 pr_crit("%s: received packet for nonexistent port\n",
200 pci_name(card->pdev));
201 else {
202 struct sk_buff *skb = card->rx_skbs[card->rx_in];
203 struct port *port = &card->ports[desc->stat &
204 PACKET_PORT_MASK];
205 struct net_device *dev = port->dev;
206
207 if (!skb)
208 dev->stats.rx_dropped++;
209 else {
210 pci_unmap_single(card->pdev, desc->address,
211 BUFFER_LENGTH,
212 PCI_DMA_FROMDEVICE);
213 skb_put(skb, desc->length);
214
215#ifdef DEBUG_PKT
216 printk(KERN_DEBUG "%s RX(%i):", dev->name,
217 skb->len);
218 debug_frame(skb);
219#endif
220 dev->stats.rx_packets++;
221 dev->stats.rx_bytes += skb->len;
222 skb->protocol = hdlc_type_trans(skb, dev);
223 netif_rx(skb);
224 skb = NULL;
225 }
226
227 if (!skb) {
228 skb = dev_alloc_skb(BUFFER_LENGTH);
229 desc->address = skb ?
230 pci_map_single(card->pdev, skb->data,
231 BUFFER_LENGTH,
232 PCI_DMA_FROMDEVICE) : 0;
233 card->rx_skbs[card->rx_in] = skb;
234 }
235 }
236 desc->stat = PACKET_EMPTY; /* Free descriptor */
237 card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
238 }
239}
240
241
242
243static irqreturn_t wanxl_intr(int irq, void* dev_id)
244{
245 struct card *card = dev_id;
246 int i;
247 u32 stat;
248 int handled = 0;
249
250
251 while((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
252 handled = 1;
253 writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
254
255 for (i = 0; i < card->n_ports; i++) {
256 if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
257 wanxl_tx_intr(&card->ports[i]);
258 if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
259 wanxl_cable_intr(&card->ports[i]);
260 }
261 if (stat & (1 << DOORBELL_FROM_CARD_RX))
262 wanxl_rx_intr(card);
263 }
264
265 return IRQ_RETVAL(handled);
266}
267
268
269
270static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
271{
272 struct port *port = dev_to_port(dev);
273 desc_t *desc;
274
275 spin_lock(&port->lock);
276
277 desc = &get_status(port)->tx_descs[port->tx_out];
278 if (desc->stat != PACKET_EMPTY) {
279 /* should never happen - previous xmit should stop queue */
280#ifdef DEBUG_PKT
281 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
282#endif
283 netif_stop_queue(dev);
284 spin_unlock(&port->lock);
285 return NETDEV_TX_BUSY; /* request packet to be queued */
286 }
287
288#ifdef DEBUG_PKT
289 printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
290 debug_frame(skb);
291#endif
292
293 port->tx_skbs[port->tx_out] = skb;
294 desc->address = pci_map_single(port->card->pdev, skb->data, skb->len,
295 PCI_DMA_TODEVICE);
296 desc->length = skb->len;
297 desc->stat = PACKET_FULL;
298 writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
299 port->card->plx + PLX_DOORBELL_TO_CARD);
300
301 port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
302
303 if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
304 netif_stop_queue(dev);
305#ifdef DEBUG_PKT
306 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
307#endif
308 }
309
310 spin_unlock(&port->lock);
311 return NETDEV_TX_OK;
312}
313
314
315
316static int wanxl_attach(struct net_device *dev, unsigned short encoding,
317 unsigned short parity)
318{
319 struct port *port = dev_to_port(dev);
320
321 if (encoding != ENCODING_NRZ &&
322 encoding != ENCODING_NRZI)
323 return -EINVAL;
324
325 if (parity != PARITY_NONE &&
326 parity != PARITY_CRC32_PR1_CCITT &&
327 parity != PARITY_CRC16_PR1_CCITT &&
328 parity != PARITY_CRC32_PR0_CCITT &&
329 parity != PARITY_CRC16_PR0_CCITT)
330 return -EINVAL;
331
332 get_status(port)->encoding = encoding;
333 get_status(port)->parity = parity;
334 return 0;
335}
336
337
338
339static int wanxl_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
340{
341 const size_t size = sizeof(sync_serial_settings);
342 sync_serial_settings line;
343 struct port *port = dev_to_port(dev);
344
345 if (cmd != SIOCWANDEV)
346 return hdlc_ioctl(dev, ifr, cmd);
347
348 switch (ifr->ifr_settings.type) {
349 case IF_GET_IFACE:
350 ifr->ifr_settings.type = IF_IFACE_SYNC_SERIAL;
351 if (ifr->ifr_settings.size < size) {
352 ifr->ifr_settings.size = size; /* data size wanted */
353 return -ENOBUFS;
354 }
355 memset(&line, 0, sizeof(line));
356 line.clock_type = get_status(port)->clocking;
357 line.clock_rate = 0;
358 line.loopback = 0;
359
360 if (copy_to_user(ifr->ifr_settings.ifs_ifsu.sync, &line, size))
361 return -EFAULT;
362 return 0;
363
364 case IF_IFACE_SYNC_SERIAL:
365 if (!capable(CAP_NET_ADMIN))
366 return -EPERM;
367 if (dev->flags & IFF_UP)
368 return -EBUSY;
369
370 if (copy_from_user(&line, ifr->ifr_settings.ifs_ifsu.sync,
371 size))
372 return -EFAULT;
373
374 if (line.clock_type != CLOCK_EXT &&
375 line.clock_type != CLOCK_TXFROMRX)
376 return -EINVAL; /* No such clock setting */
377
378 if (line.loopback != 0)
379 return -EINVAL;
380
381 get_status(port)->clocking = line.clock_type;
382 return 0;
383
384 default:
385 return hdlc_ioctl(dev, ifr, cmd);
386 }
387}
388
389
390
391static int wanxl_open(struct net_device *dev)
392{
393 struct port *port = dev_to_port(dev);
394 u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
395 unsigned long timeout;
396 int i;
397
398 if (get_status(port)->open) {
399 netdev_err(dev, "port already open\n");
400 return -EIO;
401 }
402 if ((i = hdlc_open(dev)) != 0)
403 return i;
404
405 port->tx_in = port->tx_out = 0;
406 for (i = 0; i < TX_BUFFERS; i++)
407 get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
408 /* signal the card */
409 writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
410
411 timeout = jiffies + HZ;
412 do {
413 if (get_status(port)->open) {
414 netif_start_queue(dev);
415 return 0;
416 }
417 } while (time_after(timeout, jiffies));
418
419 netdev_err(dev, "unable to open port\n");
420 /* ask the card to close the port, should it be still alive */
421 writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
422 return -EFAULT;
423}
424
425
426
427static int wanxl_close(struct net_device *dev)
428{
429 struct port *port = dev_to_port(dev);
430 unsigned long timeout;
431 int i;
432
433 hdlc_close(dev);
434 /* signal the card */
435 writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
436 port->card->plx + PLX_DOORBELL_TO_CARD);
437
438 timeout = jiffies + HZ;
439 do {
440 if (!get_status(port)->open)
441 break;
442 } while (time_after(timeout, jiffies));
443
444 if (get_status(port)->open)
445 netdev_err(dev, "unable to close port\n");
446
447 netif_stop_queue(dev);
448
449 for (i = 0; i < TX_BUFFERS; i++) {
450 desc_t *desc = &get_status(port)->tx_descs[i];
451
452 if (desc->stat != PACKET_EMPTY) {
453 desc->stat = PACKET_EMPTY;
454 pci_unmap_single(port->card->pdev, desc->address,
455 port->tx_skbs[i]->len,
456 PCI_DMA_TODEVICE);
457 dev_kfree_skb(port->tx_skbs[i]);
458 }
459 }
460 return 0;
461}
462
463
464
465static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
466{
467 struct port *port = dev_to_port(dev);
468
469 dev->stats.rx_over_errors = get_status(port)->rx_overruns;
470 dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
471 dev->stats.rx_errors = dev->stats.rx_over_errors +
472 dev->stats.rx_frame_errors;
473 return &dev->stats;
474}
475
476
477
478static int wanxl_puts_command(struct card *card, u32 cmd)
479{
480 unsigned long timeout = jiffies + 5 * HZ;
481
482 writel(cmd, card->plx + PLX_MAILBOX_1);
483 do {
484 if (readl(card->plx + PLX_MAILBOX_1) == 0)
485 return 0;
486
487 schedule();
488 }while (time_after(timeout, jiffies));
489
490 return -1;
491}
492
493
494
495static void wanxl_reset(struct card *card)
496{
497 u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
498
499 writel(0x80, card->plx + PLX_MAILBOX_0);
500 writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
501 readl(card->plx + PLX_CONTROL); /* wait for posted write */
502 udelay(1);
503 writel(old_value, card->plx + PLX_CONTROL);
504 readl(card->plx + PLX_CONTROL); /* wait for posted write */
505}
506
507
508
509static void wanxl_pci_remove_one(struct pci_dev *pdev)
510{
511 struct card *card = pci_get_drvdata(pdev);
512 int i;
513
514 for (i = 0; i < card->n_ports; i++) {
515 unregister_hdlc_device(card->ports[i].dev);
516 free_netdev(card->ports[i].dev);
517 }
518
519 /* unregister and free all host resources */
520 if (card->irq)
521 free_irq(card->irq, card);
522
523 wanxl_reset(card);
524
525 for (i = 0; i < RX_QUEUE_LENGTH; i++)
526 if (card->rx_skbs[i]) {
527 pci_unmap_single(card->pdev,
528 card->status->rx_descs[i].address,
529 BUFFER_LENGTH, PCI_DMA_FROMDEVICE);
530 dev_kfree_skb(card->rx_skbs[i]);
531 }
532
533 if (card->plx)
534 iounmap(card->plx);
535
536 if (card->status)
537 pci_free_consistent(pdev, sizeof(struct card_status),
538 card->status, card->status_address);
539
540 pci_release_regions(pdev);
541 pci_disable_device(pdev);
542 kfree(card);
543}
544
545
546#include "wanxlfw.inc"
547
548static const struct net_device_ops wanxl_ops = {
549 .ndo_open = wanxl_open,
550 .ndo_stop = wanxl_close,
551 .ndo_start_xmit = hdlc_start_xmit,
552 .ndo_do_ioctl = wanxl_ioctl,
553 .ndo_get_stats = wanxl_get_stats,
554};
555
556static int wanxl_pci_init_one(struct pci_dev *pdev,
557 const struct pci_device_id *ent)
558{
559 struct card *card;
560 u32 ramsize, stat;
561 unsigned long timeout;
562 u32 plx_phy; /* PLX PCI base address */
563 u32 mem_phy; /* memory PCI base addr */
564 u8 __iomem *mem; /* memory virtual base addr */
565 int i, ports;
566
567#ifndef MODULE
568 pr_info_once("%s\n", version);
569#endif
570
571 i = pci_enable_device(pdev);
572 if (i)
573 return i;
574
575 /* QUICC can only access first 256 MB of host RAM directly,
576 but PLX9060 DMA does 32-bits for actual packet data transfers */
577
578 /* FIXME when PCI/DMA subsystems are fixed.
579 We set both dma_mask and consistent_dma_mask to 28 bits
580 and pray pci_alloc_consistent() will use this info. It should
581 work on most platforms */
582 if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(28)) ||
583 pci_set_dma_mask(pdev, DMA_BIT_MASK(28))) {
584 pr_err("No usable DMA configuration\n");
585 pci_disable_device(pdev);
586 return -EIO;
587 }
588
589 i = pci_request_regions(pdev, "wanXL");
590 if (i) {
591 pci_disable_device(pdev);
592 return i;
593 }
594
595 switch (pdev->device) {
596 case PCI_DEVICE_ID_SBE_WANXL100: ports = 1; break;
597 case PCI_DEVICE_ID_SBE_WANXL200: ports = 2; break;
598 default: ports = 4;
599 }
600
601 card = kzalloc(struct_size(card, ports, ports), GFP_KERNEL);
602 if (card == NULL) {
603 pci_release_regions(pdev);
604 pci_disable_device(pdev);
605 return -ENOBUFS;
606 }
607
608 pci_set_drvdata(pdev, card);
609 card->pdev = pdev;
610
611 card->status = pci_alloc_consistent(pdev,
612 sizeof(struct card_status),
613 &card->status_address);
614 if (card->status == NULL) {
615 wanxl_pci_remove_one(pdev);
616 return -ENOBUFS;
617 }
618
619#ifdef DEBUG_PCI
620 printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
621 " at 0x%LX\n", pci_name(pdev),
622 (unsigned long long)card->status_address);
623#endif
624
625 /* FIXME when PCI/DMA subsystems are fixed.
626 We set both dma_mask and consistent_dma_mask back to 32 bits
627 to indicate the card can do 32-bit DMA addressing */
628 if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) ||
629 pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
630 pr_err("No usable DMA configuration\n");
631 wanxl_pci_remove_one(pdev);
632 return -EIO;
633 }
634
635 /* set up PLX mapping */
636 plx_phy = pci_resource_start(pdev, 0);
637
638 card->plx = ioremap_nocache(plx_phy, 0x70);
639 if (!card->plx) {
640 pr_err("ioremap() failed\n");
641 wanxl_pci_remove_one(pdev);
642 return -EFAULT;
643 }
644
645#if RESET_WHILE_LOADING
646 wanxl_reset(card);
647#endif
648
649 timeout = jiffies + 20 * HZ;
650 while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
651 if (time_before(timeout, jiffies)) {
652 pr_warn("%s: timeout waiting for PUTS to complete\n",
653 pci_name(pdev));
654 wanxl_pci_remove_one(pdev);
655 return -ENODEV;
656 }
657
658 switch(stat & 0xC0) {
659 case 0x00: /* hmm - PUTS completed with non-zero code? */
660 case 0x80: /* PUTS still testing the hardware */
661 break;
662
663 default:
664 pr_warn("%s: PUTS test 0x%X failed\n",
665 pci_name(pdev), stat & 0x30);
666 wanxl_pci_remove_one(pdev);
667 return -ENODEV;
668 }
669
670 schedule();
671 }
672
673 /* get on-board memory size (PUTS detects no more than 4 MB) */
674 ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
675
676 /* set up on-board RAM mapping */
677 mem_phy = pci_resource_start(pdev, 2);
678
679
680 /* sanity check the board's reported memory size */
681 if (ramsize < BUFFERS_ADDR +
682 (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
683 pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
684 pci_name(pdev), ramsize,
685 BUFFERS_ADDR +
686 (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
687 wanxl_pci_remove_one(pdev);
688 return -ENODEV;
689 }
690
691 if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
692 pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
693 wanxl_pci_remove_one(pdev);
694 return -ENODEV;
695 }
696
697 for (i = 0; i < RX_QUEUE_LENGTH; i++) {
698 struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
699 card->rx_skbs[i] = skb;
700 if (skb)
701 card->status->rx_descs[i].address =
702 pci_map_single(card->pdev, skb->data,
703 BUFFER_LENGTH,
704 PCI_DMA_FROMDEVICE);
705 }
706
707 mem = ioremap_nocache(mem_phy, PDM_OFFSET + sizeof(firmware));
708 if (!mem) {
709 pr_err("ioremap() failed\n");
710 wanxl_pci_remove_one(pdev);
711 return -EFAULT;
712 }
713
714 for (i = 0; i < sizeof(firmware); i += 4)
715 writel(ntohl(*(__be32*)(firmware + i)), mem + PDM_OFFSET + i);
716
717 for (i = 0; i < ports; i++)
718 writel(card->status_address +
719 (void *)&card->status->port_status[i] -
720 (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
721 writel(card->status_address, mem + PDM_OFFSET + 20);
722 writel(PDM_OFFSET, mem);
723 iounmap(mem);
724
725 writel(0, card->plx + PLX_MAILBOX_5);
726
727 if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
728 pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
729 wanxl_pci_remove_one(pdev);
730 return -ENODEV;
731 }
732
733 timeout = jiffies + 5 * HZ;
734 do {
735 if ((stat = readl(card->plx + PLX_MAILBOX_5)) != 0)
736 break;
737 schedule();
738 }while (time_after(timeout, jiffies));
739
740 if (!stat) {
741 pr_warn("%s: timeout while initializing card firmware\n",
742 pci_name(pdev));
743 wanxl_pci_remove_one(pdev);
744 return -ENODEV;
745 }
746
747#if DETECT_RAM
748 ramsize = stat;
749#endif
750
751 pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
752 pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
753
754 /* Allocate IRQ */
755 if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
756 pr_warn("%s: could not allocate IRQ%i\n",
757 pci_name(pdev), pdev->irq);
758 wanxl_pci_remove_one(pdev);
759 return -EBUSY;
760 }
761 card->irq = pdev->irq;
762
763 for (i = 0; i < ports; i++) {
764 hdlc_device *hdlc;
765 struct port *port = &card->ports[i];
766 struct net_device *dev = alloc_hdlcdev(port);
767 if (!dev) {
768 pr_err("%s: unable to allocate memory\n",
769 pci_name(pdev));
770 wanxl_pci_remove_one(pdev);
771 return -ENOMEM;
772 }
773
774 port->dev = dev;
775 hdlc = dev_to_hdlc(dev);
776 spin_lock_init(&port->lock);
777 dev->tx_queue_len = 50;
778 dev->netdev_ops = &wanxl_ops;
779 hdlc->attach = wanxl_attach;
780 hdlc->xmit = wanxl_xmit;
781 port->card = card;
782 port->node = i;
783 get_status(port)->clocking = CLOCK_EXT;
784 if (register_hdlc_device(dev)) {
785 pr_err("%s: unable to register hdlc device\n",
786 pci_name(pdev));
787 free_netdev(dev);
788 wanxl_pci_remove_one(pdev);
789 return -ENOBUFS;
790 }
791 card->n_ports++;
792 }
793
794 pr_info("%s: port", pci_name(pdev));
795 for (i = 0; i < ports; i++)
796 pr_cont("%s #%i: %s",
797 i ? "," : "", i, card->ports[i].dev->name);
798 pr_cont("\n");
799
800 for (i = 0; i < ports; i++)
801 wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
802
803 return 0;
804}
805
806static const struct pci_device_id wanxl_pci_tbl[] = {
807 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
808 PCI_ANY_ID, 0, 0, 0 },
809 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
810 PCI_ANY_ID, 0, 0, 0 },
811 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
812 PCI_ANY_ID, 0, 0, 0 },
813 { 0, }
814};
815
816
817static struct pci_driver wanxl_pci_driver = {
818 .name = "wanXL",
819 .id_table = wanxl_pci_tbl,
820 .probe = wanxl_pci_init_one,
821 .remove = wanxl_pci_remove_one,
822};
823
824
825static int __init wanxl_init_module(void)
826{
827#ifdef MODULE
828 pr_info("%s\n", version);
829#endif
830 return pci_register_driver(&wanxl_pci_driver);
831}
832
833static void __exit wanxl_cleanup_module(void)
834{
835 pci_unregister_driver(&wanxl_pci_driver);
836}
837
838
839MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
840MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
841MODULE_LICENSE("GPL v2");
842MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
843
844module_init(wanxl_init_module);
845module_exit(wanxl_cleanup_module);