Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * slip.c This module implements the SLIP protocol for kernel-based
4 * devices like TTY. It interfaces between a raw TTY, and the
5 * kernel's INET protocol layers.
6 *
7 * Version: @(#)slip.c 0.8.3 12/24/94
8 *
9 * Authors: Laurence Culhane, <loz@holmes.demon.co.uk>
10 * Fred N. van Kempen, <waltje@uwalt.nl.mugnet.org>
11 *
12 * Fixes:
13 * Alan Cox : Sanity checks and avoid tx overruns.
14 * Has a new sl->mtu field.
15 * Alan Cox : Found cause of overrun. ifconfig sl0
16 * mtu upwards. Driver now spots this
17 * and grows/shrinks its buffers(hack!).
18 * Memory leak if you run out of memory
19 * setting up a slip driver fixed.
20 * Matt Dillon : Printable slip (borrowed from NET2E)
21 * Pauline Middelink : Slip driver fixes.
22 * Alan Cox : Honours the old SL_COMPRESSED flag
23 * Alan Cox : KISS AX.25 and AXUI IP support
24 * Michael Riepe : Automatic CSLIP recognition added
25 * Charles Hedrick : CSLIP header length problem fix.
26 * Alan Cox : Corrected non-IP cases of the above.
27 * Alan Cox : Now uses hardware type as per FvK.
28 * Alan Cox : Default to 192.168.0.0 (RFC 1597)
29 * A.N.Kuznetsov : dev_tint() recursion fix.
30 * Dmitry Gorodchanin : SLIP memory leaks
31 * Dmitry Gorodchanin : Code cleanup. Reduce tty driver
32 * buffering from 4096 to 256 bytes.
33 * Improving SLIP response time.
34 * CONFIG_SLIP_MODE_SLIP6.
35 * ifconfig sl? up & down now works
36 * correctly.
37 * Modularization.
38 * Alan Cox : Oops - fix AX.25 buffer lengths
39 * Dmitry Gorodchanin : Even more cleanups. Preserve CSLIP
40 * statistics. Include CSLIP code only
41 * if it really needed.
42 * Alan Cox : Free slhc buffers in the right place.
43 * Alan Cox : Allow for digipeated IP over AX.25
44 * Matti Aarnio : Dynamic SLIP devices, with ideas taken
45 * from Jim Freeman's <jfree@caldera.com>
46 * dynamic PPP devices. We do NOT kfree()
47 * device entries, just reg./unreg. them
48 * as they are needed. We kfree() them
49 * at module cleanup.
50 * With MODULE-loading ``insmod'', user
51 * can issue parameter: slip_maxdev=1024
52 * (Or how much he/she wants.. Default
53 * is 256)
54 * Stanislav Voronyi : Slip line checking, with ideas taken
55 * from multislip BSDI driver which was
56 * written by Igor Chechik, RELCOM Corp.
57 * Only algorithms have been ported to
58 * Linux SLIP driver.
59 * Vitaly E. Lavrov : Sane behaviour on tty hangup.
60 * Alexey Kuznetsov : Cleanup interfaces to tty & netdevice
61 * modules.
62 */
63
64#define SL_CHECK_TRANSMIT
65#include <linux/module.h>
66#include <linux/moduleparam.h>
67
68#include <linux/uaccess.h>
69#include <linux/bitops.h>
70#include <linux/sched/signal.h>
71#include <linux/string.h>
72#include <linux/mm.h>
73#include <linux/interrupt.h>
74#include <linux/in.h>
75#include <linux/tty.h>
76#include <linux/errno.h>
77#include <linux/netdevice.h>
78#include <linux/etherdevice.h>
79#include <linux/skbuff.h>
80#include <linux/rtnetlink.h>
81#include <linux/if_arp.h>
82#include <linux/if_slip.h>
83#include <linux/delay.h>
84#include <linux/init.h>
85#include <linux/slab.h>
86#include <linux/workqueue.h>
87#include "slip.h"
88#ifdef CONFIG_INET
89#include <linux/ip.h>
90#include <linux/tcp.h>
91#include <net/slhc_vj.h>
92#endif
93
94#define SLIP_VERSION "0.8.4-NET3.019-NEWTTY"
95
96static struct net_device **slip_devs;
97
98static int slip_maxdev = SL_NRUNIT;
99module_param(slip_maxdev, int, 0);
100MODULE_PARM_DESC(slip_maxdev, "Maximum number of slip devices");
101
102static int slip_esc(unsigned char *p, unsigned char *d, int len);
103static void slip_unesc(struct slip *sl, unsigned char c);
104#ifdef CONFIG_SLIP_MODE_SLIP6
105static int slip_esc6(unsigned char *p, unsigned char *d, int len);
106static void slip_unesc6(struct slip *sl, unsigned char c);
107#endif
108#ifdef CONFIG_SLIP_SMART
109static void sl_keepalive(struct timer_list *t);
110static void sl_outfill(struct timer_list *t);
111static int sl_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
112#endif
113
114/********************************
115* Buffer administration routines:
116* sl_alloc_bufs()
117* sl_free_bufs()
118* sl_realloc_bufs()
119*
120* NOTE: sl_realloc_bufs != sl_free_bufs + sl_alloc_bufs, because
121* sl_realloc_bufs provides strong atomicity and reallocation
122* on actively running device.
123*********************************/
124
125/*
126 Allocate channel buffers.
127 */
128
129static int sl_alloc_bufs(struct slip *sl, int mtu)
130{
131 int err = -ENOBUFS;
132 unsigned long len;
133 char *rbuff = NULL;
134 char *xbuff = NULL;
135#ifdef SL_INCLUDE_CSLIP
136 char *cbuff = NULL;
137 struct slcompress *slcomp = NULL;
138#endif
139
140 /*
141 * Allocate the SLIP frame buffers:
142 *
143 * rbuff Receive buffer.
144 * xbuff Transmit buffer.
145 * cbuff Temporary compression buffer.
146 */
147 len = mtu * 2;
148
149 /*
150 * allow for arrival of larger UDP packets, even if we say not to
151 * also fixes a bug in which SunOS sends 512-byte packets even with
152 * an MSS of 128
153 */
154 if (len < 576 * 2)
155 len = 576 * 2;
156 rbuff = kmalloc(len + 4, GFP_KERNEL);
157 if (rbuff == NULL)
158 goto err_exit;
159 xbuff = kmalloc(len + 4, GFP_KERNEL);
160 if (xbuff == NULL)
161 goto err_exit;
162#ifdef SL_INCLUDE_CSLIP
163 cbuff = kmalloc(len + 4, GFP_KERNEL);
164 if (cbuff == NULL)
165 goto err_exit;
166 slcomp = slhc_init(16, 16);
167 if (IS_ERR(slcomp))
168 goto err_exit;
169#endif
170 spin_lock_bh(&sl->lock);
171 if (sl->tty == NULL) {
172 spin_unlock_bh(&sl->lock);
173 err = -ENODEV;
174 goto err_exit;
175 }
176 sl->mtu = mtu;
177 sl->buffsize = len;
178 sl->rcount = 0;
179 sl->xleft = 0;
180 rbuff = xchg(&sl->rbuff, rbuff);
181 xbuff = xchg(&sl->xbuff, xbuff);
182#ifdef SL_INCLUDE_CSLIP
183 cbuff = xchg(&sl->cbuff, cbuff);
184 slcomp = xchg(&sl->slcomp, slcomp);
185#endif
186#ifdef CONFIG_SLIP_MODE_SLIP6
187 sl->xdata = 0;
188 sl->xbits = 0;
189#endif
190 spin_unlock_bh(&sl->lock);
191 err = 0;
192
193 /* Cleanup */
194err_exit:
195#ifdef SL_INCLUDE_CSLIP
196 kfree(cbuff);
197 slhc_free(slcomp);
198#endif
199 kfree(xbuff);
200 kfree(rbuff);
201 return err;
202}
203
204/* Free a SLIP channel buffers. */
205static void sl_free_bufs(struct slip *sl)
206{
207 /* Free all SLIP frame buffers. */
208 kfree(xchg(&sl->rbuff, NULL));
209 kfree(xchg(&sl->xbuff, NULL));
210#ifdef SL_INCLUDE_CSLIP
211 kfree(xchg(&sl->cbuff, NULL));
212 slhc_free(xchg(&sl->slcomp, NULL));
213#endif
214}
215
216/*
217 Reallocate slip channel buffers.
218 */
219
220static int sl_realloc_bufs(struct slip *sl, int mtu)
221{
222 int err = 0;
223 struct net_device *dev = sl->dev;
224 unsigned char *xbuff, *rbuff;
225#ifdef SL_INCLUDE_CSLIP
226 unsigned char *cbuff;
227#endif
228 int len = mtu * 2;
229
230/*
231 * allow for arrival of larger UDP packets, even if we say not to
232 * also fixes a bug in which SunOS sends 512-byte packets even with
233 * an MSS of 128
234 */
235 if (len < 576 * 2)
236 len = 576 * 2;
237
238 xbuff = kmalloc(len + 4, GFP_ATOMIC);
239 rbuff = kmalloc(len + 4, GFP_ATOMIC);
240#ifdef SL_INCLUDE_CSLIP
241 cbuff = kmalloc(len + 4, GFP_ATOMIC);
242#endif
243
244
245#ifdef SL_INCLUDE_CSLIP
246 if (xbuff == NULL || rbuff == NULL || cbuff == NULL) {
247#else
248 if (xbuff == NULL || rbuff == NULL) {
249#endif
250 if (mtu > sl->mtu) {
251 printk(KERN_WARNING "%s: unable to grow slip buffers, MTU change cancelled.\n",
252 dev->name);
253 err = -ENOBUFS;
254 }
255 goto done;
256 }
257 spin_lock_bh(&sl->lock);
258
259 err = -ENODEV;
260 if (sl->tty == NULL)
261 goto done_on_bh;
262
263 xbuff = xchg(&sl->xbuff, xbuff);
264 rbuff = xchg(&sl->rbuff, rbuff);
265#ifdef SL_INCLUDE_CSLIP
266 cbuff = xchg(&sl->cbuff, cbuff);
267#endif
268 if (sl->xleft) {
269 if (sl->xleft <= len) {
270 memcpy(sl->xbuff, sl->xhead, sl->xleft);
271 } else {
272 sl->xleft = 0;
273 dev->stats.tx_dropped++;
274 }
275 }
276 sl->xhead = sl->xbuff;
277
278 if (sl->rcount) {
279 if (sl->rcount <= len) {
280 memcpy(sl->rbuff, rbuff, sl->rcount);
281 } else {
282 sl->rcount = 0;
283 dev->stats.rx_over_errors++;
284 set_bit(SLF_ERROR, &sl->flags);
285 }
286 }
287 sl->mtu = mtu;
288 dev->mtu = mtu;
289 sl->buffsize = len;
290 err = 0;
291
292done_on_bh:
293 spin_unlock_bh(&sl->lock);
294
295done:
296 kfree(xbuff);
297 kfree(rbuff);
298#ifdef SL_INCLUDE_CSLIP
299 kfree(cbuff);
300#endif
301 return err;
302}
303
304
305/* Set the "sending" flag. This must be atomic hence the set_bit. */
306static inline void sl_lock(struct slip *sl)
307{
308 netif_stop_queue(sl->dev);
309}
310
311
312/* Clear the "sending" flag. This must be atomic, hence the ASM. */
313static inline void sl_unlock(struct slip *sl)
314{
315 netif_wake_queue(sl->dev);
316}
317
318/* Send one completely decapsulated IP datagram to the IP layer. */
319static void sl_bump(struct slip *sl)
320{
321 struct net_device *dev = sl->dev;
322 struct sk_buff *skb;
323 int count;
324
325 count = sl->rcount;
326#ifdef SL_INCLUDE_CSLIP
327 if (sl->mode & (SL_MODE_ADAPTIVE | SL_MODE_CSLIP)) {
328 unsigned char c = sl->rbuff[0];
329 if (c & SL_TYPE_COMPRESSED_TCP) {
330 /* ignore compressed packets when CSLIP is off */
331 if (!(sl->mode & SL_MODE_CSLIP)) {
332 printk(KERN_WARNING "%s: compressed packet ignored\n", dev->name);
333 return;
334 }
335 /* make sure we've reserved enough space for uncompress
336 to use */
337 if (count + 80 > sl->buffsize) {
338 dev->stats.rx_over_errors++;
339 return;
340 }
341 count = slhc_uncompress(sl->slcomp, sl->rbuff, count);
342 if (count <= 0)
343 return;
344 } else if (c >= SL_TYPE_UNCOMPRESSED_TCP) {
345 if (!(sl->mode & SL_MODE_CSLIP)) {
346 /* turn on header compression */
347 sl->mode |= SL_MODE_CSLIP;
348 sl->mode &= ~SL_MODE_ADAPTIVE;
349 printk(KERN_INFO "%s: header compression turned on\n", dev->name);
350 }
351 sl->rbuff[0] &= 0x4f;
352 if (slhc_remember(sl->slcomp, sl->rbuff, count) <= 0)
353 return;
354 }
355 }
356#endif /* SL_INCLUDE_CSLIP */
357
358 dev->stats.rx_bytes += count;
359
360 skb = dev_alloc_skb(count);
361 if (skb == NULL) {
362 printk(KERN_WARNING "%s: memory squeeze, dropping packet.\n", dev->name);
363 dev->stats.rx_dropped++;
364 return;
365 }
366 skb->dev = dev;
367 skb_put_data(skb, sl->rbuff, count);
368 skb_reset_mac_header(skb);
369 skb->protocol = htons(ETH_P_IP);
370 netif_rx_ni(skb);
371 dev->stats.rx_packets++;
372}
373
374/* Encapsulate one IP datagram and stuff into a TTY queue. */
375static void sl_encaps(struct slip *sl, unsigned char *icp, int len)
376{
377 unsigned char *p;
378 int actual, count;
379
380 if (len > sl->mtu) { /* Sigh, shouldn't occur BUT ... */
381 printk(KERN_WARNING "%s: truncating oversized transmit packet!\n", sl->dev->name);
382 sl->dev->stats.tx_dropped++;
383 sl_unlock(sl);
384 return;
385 }
386
387 p = icp;
388#ifdef SL_INCLUDE_CSLIP
389 if (sl->mode & SL_MODE_CSLIP)
390 len = slhc_compress(sl->slcomp, p, len, sl->cbuff, &p, 1);
391#endif
392#ifdef CONFIG_SLIP_MODE_SLIP6
393 if (sl->mode & SL_MODE_SLIP6)
394 count = slip_esc6(p, sl->xbuff, len);
395 else
396#endif
397 count = slip_esc(p, sl->xbuff, len);
398
399 /* Order of next two lines is *very* important.
400 * When we are sending a little amount of data,
401 * the transfer may be completed inside the ops->write()
402 * routine, because it's running with interrupts enabled.
403 * In this case we *never* got WRITE_WAKEUP event,
404 * if we did not request it before write operation.
405 * 14 Oct 1994 Dmitry Gorodchanin.
406 */
407 set_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
408 actual = sl->tty->ops->write(sl->tty, sl->xbuff, count);
409#ifdef SL_CHECK_TRANSMIT
410 netif_trans_update(sl->dev);
411#endif
412 sl->xleft = count - actual;
413 sl->xhead = sl->xbuff + actual;
414#ifdef CONFIG_SLIP_SMART
415 /* VSV */
416 clear_bit(SLF_OUTWAIT, &sl->flags); /* reset outfill flag */
417#endif
418}
419
420/* Write out any remaining transmit buffer. Scheduled when tty is writable */
421static void slip_transmit(struct work_struct *work)
422{
423 struct slip *sl = container_of(work, struct slip, tx_work);
424 int actual;
425
426 spin_lock_bh(&sl->lock);
427 /* First make sure we're connected. */
428 if (!sl->tty || sl->magic != SLIP_MAGIC || !netif_running(sl->dev)) {
429 spin_unlock_bh(&sl->lock);
430 return;
431 }
432
433 if (sl->xleft <= 0) {
434 /* Now serial buffer is almost free & we can start
435 * transmission of another packet */
436 sl->dev->stats.tx_packets++;
437 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
438 spin_unlock_bh(&sl->lock);
439 sl_unlock(sl);
440 return;
441 }
442
443 actual = sl->tty->ops->write(sl->tty, sl->xhead, sl->xleft);
444 sl->xleft -= actual;
445 sl->xhead += actual;
446 spin_unlock_bh(&sl->lock);
447}
448
449/*
450 * Called by the driver when there's room for more data.
451 * Schedule the transmit.
452 */
453static void slip_write_wakeup(struct tty_struct *tty)
454{
455 struct slip *sl;
456
457 rcu_read_lock();
458 sl = rcu_dereference(tty->disc_data);
459 if (sl)
460 schedule_work(&sl->tx_work);
461 rcu_read_unlock();
462}
463
464static void sl_tx_timeout(struct net_device *dev, unsigned int txqueue)
465{
466 struct slip *sl = netdev_priv(dev);
467
468 spin_lock(&sl->lock);
469
470 if (netif_queue_stopped(dev)) {
471 if (!netif_running(dev))
472 goto out;
473
474 /* May be we must check transmitter timeout here ?
475 * 14 Oct 1994 Dmitry Gorodchanin.
476 */
477#ifdef SL_CHECK_TRANSMIT
478 if (time_before(jiffies, dev_trans_start(dev) + 20 * HZ)) {
479 /* 20 sec timeout not reached */
480 goto out;
481 }
482 printk(KERN_WARNING "%s: transmit timed out, %s?\n",
483 dev->name,
484 (tty_chars_in_buffer(sl->tty) || sl->xleft) ?
485 "bad line quality" : "driver error");
486 sl->xleft = 0;
487 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
488 sl_unlock(sl);
489#endif
490 }
491out:
492 spin_unlock(&sl->lock);
493}
494
495
496/* Encapsulate an IP datagram and kick it into a TTY queue. */
497static netdev_tx_t
498sl_xmit(struct sk_buff *skb, struct net_device *dev)
499{
500 struct slip *sl = netdev_priv(dev);
501
502 spin_lock(&sl->lock);
503 if (!netif_running(dev)) {
504 spin_unlock(&sl->lock);
505 printk(KERN_WARNING "%s: xmit call when iface is down\n", dev->name);
506 dev_kfree_skb(skb);
507 return NETDEV_TX_OK;
508 }
509 if (sl->tty == NULL) {
510 spin_unlock(&sl->lock);
511 dev_kfree_skb(skb);
512 return NETDEV_TX_OK;
513 }
514
515 sl_lock(sl);
516 dev->stats.tx_bytes += skb->len;
517 sl_encaps(sl, skb->data, skb->len);
518 spin_unlock(&sl->lock);
519
520 dev_kfree_skb(skb);
521 return NETDEV_TX_OK;
522}
523
524
525/******************************************
526 * Routines looking at netdevice side.
527 ******************************************/
528
529/* Netdevice UP -> DOWN routine */
530
531static int
532sl_close(struct net_device *dev)
533{
534 struct slip *sl = netdev_priv(dev);
535
536 spin_lock_bh(&sl->lock);
537 if (sl->tty)
538 /* TTY discipline is running. */
539 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
540 netif_stop_queue(dev);
541 sl->rcount = 0;
542 sl->xleft = 0;
543 spin_unlock_bh(&sl->lock);
544
545 return 0;
546}
547
548/* Netdevice DOWN -> UP routine */
549
550static int sl_open(struct net_device *dev)
551{
552 struct slip *sl = netdev_priv(dev);
553
554 if (sl->tty == NULL)
555 return -ENODEV;
556
557 sl->flags &= (1 << SLF_INUSE);
558 netif_start_queue(dev);
559 return 0;
560}
561
562/* Netdevice change MTU request */
563
564static int sl_change_mtu(struct net_device *dev, int new_mtu)
565{
566 struct slip *sl = netdev_priv(dev);
567
568 return sl_realloc_bufs(sl, new_mtu);
569}
570
571/* Netdevice get statistics request */
572
573static void
574sl_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
575{
576 struct net_device_stats *devstats = &dev->stats;
577#ifdef SL_INCLUDE_CSLIP
578 struct slip *sl = netdev_priv(dev);
579 struct slcompress *comp = sl->slcomp;
580#endif
581 stats->rx_packets = devstats->rx_packets;
582 stats->tx_packets = devstats->tx_packets;
583 stats->rx_bytes = devstats->rx_bytes;
584 stats->tx_bytes = devstats->tx_bytes;
585 stats->rx_dropped = devstats->rx_dropped;
586 stats->tx_dropped = devstats->tx_dropped;
587 stats->tx_errors = devstats->tx_errors;
588 stats->rx_errors = devstats->rx_errors;
589 stats->rx_over_errors = devstats->rx_over_errors;
590
591#ifdef SL_INCLUDE_CSLIP
592 if (comp) {
593 /* Generic compressed statistics */
594 stats->rx_compressed = comp->sls_i_compressed;
595 stats->tx_compressed = comp->sls_o_compressed;
596
597 /* Are we really still needs this? */
598 stats->rx_fifo_errors += comp->sls_i_compressed;
599 stats->rx_dropped += comp->sls_i_tossed;
600 stats->tx_fifo_errors += comp->sls_o_compressed;
601 stats->collisions += comp->sls_o_misses;
602 }
603#endif
604}
605
606/* Netdevice register callback */
607
608static int sl_init(struct net_device *dev)
609{
610 struct slip *sl = netdev_priv(dev);
611
612 /*
613 * Finish setting up the DEVICE info.
614 */
615
616 dev->mtu = sl->mtu;
617 dev->type = ARPHRD_SLIP + sl->mode;
618#ifdef SL_CHECK_TRANSMIT
619 dev->watchdog_timeo = 20*HZ;
620#endif
621 return 0;
622}
623
624
625static void sl_uninit(struct net_device *dev)
626{
627 struct slip *sl = netdev_priv(dev);
628
629 sl_free_bufs(sl);
630}
631
632/* Hook the destructor so we can free slip devices at the right point in time */
633static void sl_free_netdev(struct net_device *dev)
634{
635 int i = dev->base_addr;
636
637 slip_devs[i] = NULL;
638}
639
640static const struct net_device_ops sl_netdev_ops = {
641 .ndo_init = sl_init,
642 .ndo_uninit = sl_uninit,
643 .ndo_open = sl_open,
644 .ndo_stop = sl_close,
645 .ndo_start_xmit = sl_xmit,
646 .ndo_get_stats64 = sl_get_stats64,
647 .ndo_change_mtu = sl_change_mtu,
648 .ndo_tx_timeout = sl_tx_timeout,
649#ifdef CONFIG_SLIP_SMART
650 .ndo_do_ioctl = sl_ioctl,
651#endif
652};
653
654
655static void sl_setup(struct net_device *dev)
656{
657 dev->netdev_ops = &sl_netdev_ops;
658 dev->needs_free_netdev = true;
659 dev->priv_destructor = sl_free_netdev;
660
661 dev->hard_header_len = 0;
662 dev->addr_len = 0;
663 dev->tx_queue_len = 10;
664
665 /* MTU range: 68 - 65534 */
666 dev->min_mtu = 68;
667 dev->max_mtu = 65534;
668
669 /* New-style flags. */
670 dev->flags = IFF_NOARP|IFF_POINTOPOINT|IFF_MULTICAST;
671}
672
673/******************************************
674 Routines looking at TTY side.
675 ******************************************/
676
677
678/*
679 * Handle the 'receiver data ready' interrupt.
680 * This function is called by the 'tty_io' module in the kernel when
681 * a block of SLIP data has been received, which can now be decapsulated
682 * and sent on to some IP layer for further processing. This will not
683 * be re-entered while running but other ldisc functions may be called
684 * in parallel
685 */
686
687static void slip_receive_buf(struct tty_struct *tty, const unsigned char *cp,
688 char *fp, int count)
689{
690 struct slip *sl = tty->disc_data;
691
692 if (!sl || sl->magic != SLIP_MAGIC || !netif_running(sl->dev))
693 return;
694
695 /* Read the characters out of the buffer */
696 while (count--) {
697 if (fp && *fp++) {
698 if (!test_and_set_bit(SLF_ERROR, &sl->flags))
699 sl->dev->stats.rx_errors++;
700 cp++;
701 continue;
702 }
703#ifdef CONFIG_SLIP_MODE_SLIP6
704 if (sl->mode & SL_MODE_SLIP6)
705 slip_unesc6(sl, *cp++);
706 else
707#endif
708 slip_unesc(sl, *cp++);
709 }
710}
711
712/************************************
713 * slip_open helper routines.
714 ************************************/
715
716/* Collect hanged up channels */
717static void sl_sync(void)
718{
719 int i;
720 struct net_device *dev;
721 struct slip *sl;
722
723 for (i = 0; i < slip_maxdev; i++) {
724 dev = slip_devs[i];
725 if (dev == NULL)
726 break;
727
728 sl = netdev_priv(dev);
729 if (sl->tty || sl->leased)
730 continue;
731 if (dev->flags & IFF_UP)
732 dev_close(dev);
733 }
734}
735
736
737/* Find a free SLIP channel, and link in this `tty' line. */
738static struct slip *sl_alloc(void)
739{
740 int i;
741 char name[IFNAMSIZ];
742 struct net_device *dev = NULL;
743 struct slip *sl;
744
745 for (i = 0; i < slip_maxdev; i++) {
746 dev = slip_devs[i];
747 if (dev == NULL)
748 break;
749 }
750 /* Sorry, too many, all slots in use */
751 if (i >= slip_maxdev)
752 return NULL;
753
754 sprintf(name, "sl%d", i);
755 dev = alloc_netdev(sizeof(*sl), name, NET_NAME_UNKNOWN, sl_setup);
756 if (!dev)
757 return NULL;
758
759 dev->base_addr = i;
760 sl = netdev_priv(dev);
761
762 /* Initialize channel control data */
763 sl->magic = SLIP_MAGIC;
764 sl->dev = dev;
765 spin_lock_init(&sl->lock);
766 INIT_WORK(&sl->tx_work, slip_transmit);
767 sl->mode = SL_MODE_DEFAULT;
768#ifdef CONFIG_SLIP_SMART
769 /* initialize timer_list struct */
770 timer_setup(&sl->keepalive_timer, sl_keepalive, 0);
771 timer_setup(&sl->outfill_timer, sl_outfill, 0);
772#endif
773 slip_devs[i] = dev;
774 return sl;
775}
776
777/*
778 * Open the high-level part of the SLIP channel.
779 * This function is called by the TTY module when the
780 * SLIP line discipline is called for. Because we are
781 * sure the tty line exists, we only have to link it to
782 * a free SLIP channel...
783 *
784 * Called in process context serialized from other ldisc calls.
785 */
786
787static int slip_open(struct tty_struct *tty)
788{
789 struct slip *sl;
790 int err;
791
792 if (!capable(CAP_NET_ADMIN))
793 return -EPERM;
794
795 if (tty->ops->write == NULL)
796 return -EOPNOTSUPP;
797
798 /* RTnetlink lock is misused here to serialize concurrent
799 opens of slip channels. There are better ways, but it is
800 the simplest one.
801 */
802 rtnl_lock();
803
804 /* Collect hanged up channels. */
805 sl_sync();
806
807 sl = tty->disc_data;
808
809 err = -EEXIST;
810 /* First make sure we're not already connected. */
811 if (sl && sl->magic == SLIP_MAGIC)
812 goto err_exit;
813
814 /* OK. Find a free SLIP channel to use. */
815 err = -ENFILE;
816 sl = sl_alloc();
817 if (sl == NULL)
818 goto err_exit;
819
820 sl->tty = tty;
821 tty->disc_data = sl;
822 sl->pid = current->pid;
823
824 if (!test_bit(SLF_INUSE, &sl->flags)) {
825 /* Perform the low-level SLIP initialization. */
826 err = sl_alloc_bufs(sl, SL_MTU);
827 if (err)
828 goto err_free_chan;
829
830 set_bit(SLF_INUSE, &sl->flags);
831
832 err = register_netdevice(sl->dev);
833 if (err)
834 goto err_free_bufs;
835 }
836
837#ifdef CONFIG_SLIP_SMART
838 if (sl->keepalive) {
839 sl->keepalive_timer.expires = jiffies + sl->keepalive * HZ;
840 add_timer(&sl->keepalive_timer);
841 }
842 if (sl->outfill) {
843 sl->outfill_timer.expires = jiffies + sl->outfill * HZ;
844 add_timer(&sl->outfill_timer);
845 }
846#endif
847
848 /* Done. We have linked the TTY line to a channel. */
849 rtnl_unlock();
850 tty->receive_room = 65536; /* We don't flow control */
851
852 /* TTY layer expects 0 on success */
853 return 0;
854
855err_free_bufs:
856 sl_free_bufs(sl);
857
858err_free_chan:
859 sl->tty = NULL;
860 tty->disc_data = NULL;
861 clear_bit(SLF_INUSE, &sl->flags);
862 sl_free_netdev(sl->dev);
863 /* do not call free_netdev before rtnl_unlock */
864 rtnl_unlock();
865 free_netdev(sl->dev);
866 return err;
867
868err_exit:
869 rtnl_unlock();
870
871 /* Count references from TTY module */
872 return err;
873}
874
875/*
876 * Close down a SLIP channel.
877 * This means flushing out any pending queues, and then returning. This
878 * call is serialized against other ldisc functions.
879 *
880 * We also use this method fo a hangup event
881 */
882
883static void slip_close(struct tty_struct *tty)
884{
885 struct slip *sl = tty->disc_data;
886
887 /* First make sure we're connected. */
888 if (!sl || sl->magic != SLIP_MAGIC || sl->tty != tty)
889 return;
890
891 spin_lock_bh(&sl->lock);
892 rcu_assign_pointer(tty->disc_data, NULL);
893 sl->tty = NULL;
894 spin_unlock_bh(&sl->lock);
895
896 synchronize_rcu();
897 flush_work(&sl->tx_work);
898
899 /* VSV = very important to remove timers */
900#ifdef CONFIG_SLIP_SMART
901 del_timer_sync(&sl->keepalive_timer);
902 del_timer_sync(&sl->outfill_timer);
903#endif
904 /* Flush network side */
905 unregister_netdev(sl->dev);
906 /* This will complete via sl_free_netdev */
907}
908
909static int slip_hangup(struct tty_struct *tty)
910{
911 slip_close(tty);
912 return 0;
913}
914 /************************************************************************
915 * STANDARD SLIP ENCAPSULATION *
916 ************************************************************************/
917
918static int slip_esc(unsigned char *s, unsigned char *d, int len)
919{
920 unsigned char *ptr = d;
921 unsigned char c;
922
923 /*
924 * Send an initial END character to flush out any
925 * data that may have accumulated in the receiver
926 * due to line noise.
927 */
928
929 *ptr++ = END;
930
931 /*
932 * For each byte in the packet, send the appropriate
933 * character sequence, according to the SLIP protocol.
934 */
935
936 while (len-- > 0) {
937 switch (c = *s++) {
938 case END:
939 *ptr++ = ESC;
940 *ptr++ = ESC_END;
941 break;
942 case ESC:
943 *ptr++ = ESC;
944 *ptr++ = ESC_ESC;
945 break;
946 default:
947 *ptr++ = c;
948 break;
949 }
950 }
951 *ptr++ = END;
952 return ptr - d;
953}
954
955static void slip_unesc(struct slip *sl, unsigned char s)
956{
957
958 switch (s) {
959 case END:
960#ifdef CONFIG_SLIP_SMART
961 /* drop keeptest bit = VSV */
962 if (test_bit(SLF_KEEPTEST, &sl->flags))
963 clear_bit(SLF_KEEPTEST, &sl->flags);
964#endif
965
966 if (!test_and_clear_bit(SLF_ERROR, &sl->flags) &&
967 (sl->rcount > 2))
968 sl_bump(sl);
969 clear_bit(SLF_ESCAPE, &sl->flags);
970 sl->rcount = 0;
971 return;
972
973 case ESC:
974 set_bit(SLF_ESCAPE, &sl->flags);
975 return;
976 case ESC_ESC:
977 if (test_and_clear_bit(SLF_ESCAPE, &sl->flags))
978 s = ESC;
979 break;
980 case ESC_END:
981 if (test_and_clear_bit(SLF_ESCAPE, &sl->flags))
982 s = END;
983 break;
984 }
985 if (!test_bit(SLF_ERROR, &sl->flags)) {
986 if (sl->rcount < sl->buffsize) {
987 sl->rbuff[sl->rcount++] = s;
988 return;
989 }
990 sl->dev->stats.rx_over_errors++;
991 set_bit(SLF_ERROR, &sl->flags);
992 }
993}
994
995
996#ifdef CONFIG_SLIP_MODE_SLIP6
997/************************************************************************
998 * 6 BIT SLIP ENCAPSULATION *
999 ************************************************************************/
1000
1001static int slip_esc6(unsigned char *s, unsigned char *d, int len)
1002{
1003 unsigned char *ptr = d;
1004 unsigned char c;
1005 int i;
1006 unsigned short v = 0;
1007 short bits = 0;
1008
1009 /*
1010 * Send an initial END character to flush out any
1011 * data that may have accumulated in the receiver
1012 * due to line noise.
1013 */
1014
1015 *ptr++ = 0x70;
1016
1017 /*
1018 * Encode the packet into printable ascii characters
1019 */
1020
1021 for (i = 0; i < len; ++i) {
1022 v = (v << 8) | s[i];
1023 bits += 8;
1024 while (bits >= 6) {
1025 bits -= 6;
1026 c = 0x30 + ((v >> bits) & 0x3F);
1027 *ptr++ = c;
1028 }
1029 }
1030 if (bits) {
1031 c = 0x30 + ((v << (6 - bits)) & 0x3F);
1032 *ptr++ = c;
1033 }
1034 *ptr++ = 0x70;
1035 return ptr - d;
1036}
1037
1038static void slip_unesc6(struct slip *sl, unsigned char s)
1039{
1040 unsigned char c;
1041
1042 if (s == 0x70) {
1043#ifdef CONFIG_SLIP_SMART
1044 /* drop keeptest bit = VSV */
1045 if (test_bit(SLF_KEEPTEST, &sl->flags))
1046 clear_bit(SLF_KEEPTEST, &sl->flags);
1047#endif
1048
1049 if (!test_and_clear_bit(SLF_ERROR, &sl->flags) &&
1050 (sl->rcount > 2))
1051 sl_bump(sl);
1052 sl->rcount = 0;
1053 sl->xbits = 0;
1054 sl->xdata = 0;
1055 } else if (s >= 0x30 && s < 0x70) {
1056 sl->xdata = (sl->xdata << 6) | ((s - 0x30) & 0x3F);
1057 sl->xbits += 6;
1058 if (sl->xbits >= 8) {
1059 sl->xbits -= 8;
1060 c = (unsigned char)(sl->xdata >> sl->xbits);
1061 if (!test_bit(SLF_ERROR, &sl->flags)) {
1062 if (sl->rcount < sl->buffsize) {
1063 sl->rbuff[sl->rcount++] = c;
1064 return;
1065 }
1066 sl->dev->stats.rx_over_errors++;
1067 set_bit(SLF_ERROR, &sl->flags);
1068 }
1069 }
1070 }
1071}
1072#endif /* CONFIG_SLIP_MODE_SLIP6 */
1073
1074/* Perform I/O control on an active SLIP channel. */
1075static int slip_ioctl(struct tty_struct *tty, struct file *file,
1076 unsigned int cmd, unsigned long arg)
1077{
1078 struct slip *sl = tty->disc_data;
1079 unsigned int tmp;
1080 int __user *p = (int __user *)arg;
1081
1082 /* First make sure we're connected. */
1083 if (!sl || sl->magic != SLIP_MAGIC)
1084 return -EINVAL;
1085
1086 switch (cmd) {
1087 case SIOCGIFNAME:
1088 tmp = strlen(sl->dev->name) + 1;
1089 if (copy_to_user((void __user *)arg, sl->dev->name, tmp))
1090 return -EFAULT;
1091 return 0;
1092
1093 case SIOCGIFENCAP:
1094 if (put_user(sl->mode, p))
1095 return -EFAULT;
1096 return 0;
1097
1098 case SIOCSIFENCAP:
1099 if (get_user(tmp, p))
1100 return -EFAULT;
1101#ifndef SL_INCLUDE_CSLIP
1102 if (tmp & (SL_MODE_CSLIP|SL_MODE_ADAPTIVE))
1103 return -EINVAL;
1104#else
1105 if ((tmp & (SL_MODE_ADAPTIVE | SL_MODE_CSLIP)) ==
1106 (SL_MODE_ADAPTIVE | SL_MODE_CSLIP))
1107 /* return -EINVAL; */
1108 tmp &= ~SL_MODE_ADAPTIVE;
1109#endif
1110#ifndef CONFIG_SLIP_MODE_SLIP6
1111 if (tmp & SL_MODE_SLIP6)
1112 return -EINVAL;
1113#endif
1114 sl->mode = tmp;
1115 sl->dev->type = ARPHRD_SLIP + sl->mode;
1116 return 0;
1117
1118 case SIOCSIFHWADDR:
1119 return -EINVAL;
1120
1121#ifdef CONFIG_SLIP_SMART
1122 /* VSV changes start here */
1123 case SIOCSKEEPALIVE:
1124 if (get_user(tmp, p))
1125 return -EFAULT;
1126 if (tmp > 255) /* max for unchar */
1127 return -EINVAL;
1128
1129 spin_lock_bh(&sl->lock);
1130 if (!sl->tty) {
1131 spin_unlock_bh(&sl->lock);
1132 return -ENODEV;
1133 }
1134 sl->keepalive = (u8)tmp;
1135 if (sl->keepalive != 0) {
1136 mod_timer(&sl->keepalive_timer,
1137 jiffies + sl->keepalive * HZ);
1138 set_bit(SLF_KEEPTEST, &sl->flags);
1139 } else
1140 del_timer(&sl->keepalive_timer);
1141 spin_unlock_bh(&sl->lock);
1142 return 0;
1143
1144 case SIOCGKEEPALIVE:
1145 if (put_user(sl->keepalive, p))
1146 return -EFAULT;
1147 return 0;
1148
1149 case SIOCSOUTFILL:
1150 if (get_user(tmp, p))
1151 return -EFAULT;
1152 if (tmp > 255) /* max for unchar */
1153 return -EINVAL;
1154 spin_lock_bh(&sl->lock);
1155 if (!sl->tty) {
1156 spin_unlock_bh(&sl->lock);
1157 return -ENODEV;
1158 }
1159 sl->outfill = (u8)tmp;
1160 if (sl->outfill != 0) {
1161 mod_timer(&sl->outfill_timer,
1162 jiffies + sl->outfill * HZ);
1163 set_bit(SLF_OUTWAIT, &sl->flags);
1164 } else
1165 del_timer(&sl->outfill_timer);
1166 spin_unlock_bh(&sl->lock);
1167 return 0;
1168
1169 case SIOCGOUTFILL:
1170 if (put_user(sl->outfill, p))
1171 return -EFAULT;
1172 return 0;
1173 /* VSV changes end */
1174#endif
1175 default:
1176 return tty_mode_ioctl(tty, file, cmd, arg);
1177 }
1178}
1179
1180/* VSV changes start here */
1181#ifdef CONFIG_SLIP_SMART
1182/* function do_ioctl called from net/core/dev.c
1183 to allow get/set outfill/keepalive parameter
1184 by ifconfig */
1185
1186static int sl_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
1187{
1188 struct slip *sl = netdev_priv(dev);
1189 unsigned long *p = (unsigned long *)&rq->ifr_ifru;
1190
1191 if (sl == NULL) /* Allocation failed ?? */
1192 return -ENODEV;
1193
1194 spin_lock_bh(&sl->lock);
1195
1196 if (!sl->tty) {
1197 spin_unlock_bh(&sl->lock);
1198 return -ENODEV;
1199 }
1200
1201 switch (cmd) {
1202 case SIOCSKEEPALIVE:
1203 /* max for unchar */
1204 if ((unsigned)*p > 255) {
1205 spin_unlock_bh(&sl->lock);
1206 return -EINVAL;
1207 }
1208 sl->keepalive = (u8)*p;
1209 if (sl->keepalive != 0) {
1210 sl->keepalive_timer.expires =
1211 jiffies + sl->keepalive * HZ;
1212 mod_timer(&sl->keepalive_timer,
1213 jiffies + sl->keepalive * HZ);
1214 set_bit(SLF_KEEPTEST, &sl->flags);
1215 } else
1216 del_timer(&sl->keepalive_timer);
1217 break;
1218
1219 case SIOCGKEEPALIVE:
1220 *p = sl->keepalive;
1221 break;
1222
1223 case SIOCSOUTFILL:
1224 if ((unsigned)*p > 255) { /* max for unchar */
1225 spin_unlock_bh(&sl->lock);
1226 return -EINVAL;
1227 }
1228 sl->outfill = (u8)*p;
1229 if (sl->outfill != 0) {
1230 mod_timer(&sl->outfill_timer,
1231 jiffies + sl->outfill * HZ);
1232 set_bit(SLF_OUTWAIT, &sl->flags);
1233 } else
1234 del_timer(&sl->outfill_timer);
1235 break;
1236
1237 case SIOCGOUTFILL:
1238 *p = sl->outfill;
1239 break;
1240
1241 case SIOCSLEASE:
1242 /* Resolve race condition, when ioctl'ing hanged up
1243 and opened by another process device.
1244 */
1245 if (sl->tty != current->signal->tty &&
1246 sl->pid != current->pid) {
1247 spin_unlock_bh(&sl->lock);
1248 return -EPERM;
1249 }
1250 sl->leased = 0;
1251 if (*p)
1252 sl->leased = 1;
1253 break;
1254
1255 case SIOCGLEASE:
1256 *p = sl->leased;
1257 }
1258 spin_unlock_bh(&sl->lock);
1259 return 0;
1260}
1261#endif
1262/* VSV changes end */
1263
1264static struct tty_ldisc_ops sl_ldisc = {
1265 .owner = THIS_MODULE,
1266 .magic = TTY_LDISC_MAGIC,
1267 .name = "slip",
1268 .open = slip_open,
1269 .close = slip_close,
1270 .hangup = slip_hangup,
1271 .ioctl = slip_ioctl,
1272 .receive_buf = slip_receive_buf,
1273 .write_wakeup = slip_write_wakeup,
1274};
1275
1276static int __init slip_init(void)
1277{
1278 int status;
1279
1280 if (slip_maxdev < 4)
1281 slip_maxdev = 4; /* Sanity */
1282
1283 printk(KERN_INFO "SLIP: version %s (dynamic channels, max=%d)"
1284#ifdef CONFIG_SLIP_MODE_SLIP6
1285 " (6 bit encapsulation enabled)"
1286#endif
1287 ".\n",
1288 SLIP_VERSION, slip_maxdev);
1289#if defined(SL_INCLUDE_CSLIP)
1290 printk(KERN_INFO "CSLIP: code copyright 1989 Regents of the University of California.\n");
1291#endif
1292#ifdef CONFIG_SLIP_SMART
1293 printk(KERN_INFO "SLIP linefill/keepalive option.\n");
1294#endif
1295
1296 slip_devs = kcalloc(slip_maxdev, sizeof(struct net_device *),
1297 GFP_KERNEL);
1298 if (!slip_devs)
1299 return -ENOMEM;
1300
1301 /* Fill in our line protocol discipline, and register it */
1302 status = tty_register_ldisc(N_SLIP, &sl_ldisc);
1303 if (status != 0) {
1304 printk(KERN_ERR "SLIP: can't register line discipline (err = %d)\n", status);
1305 kfree(slip_devs);
1306 }
1307 return status;
1308}
1309
1310static void __exit slip_exit(void)
1311{
1312 int i;
1313 struct net_device *dev;
1314 struct slip *sl;
1315 unsigned long timeout = jiffies + HZ;
1316 int busy = 0;
1317
1318 if (slip_devs == NULL)
1319 return;
1320
1321 /* First of all: check for active disciplines and hangup them.
1322 */
1323 do {
1324 if (busy)
1325 msleep_interruptible(100);
1326
1327 busy = 0;
1328 for (i = 0; i < slip_maxdev; i++) {
1329 dev = slip_devs[i];
1330 if (!dev)
1331 continue;
1332 sl = netdev_priv(dev);
1333 spin_lock_bh(&sl->lock);
1334 if (sl->tty) {
1335 busy++;
1336 tty_hangup(sl->tty);
1337 }
1338 spin_unlock_bh(&sl->lock);
1339 }
1340 } while (busy && time_before(jiffies, timeout));
1341
1342 /* FIXME: hangup is async so we should wait when doing this second
1343 phase */
1344
1345 for (i = 0; i < slip_maxdev; i++) {
1346 dev = slip_devs[i];
1347 if (!dev)
1348 continue;
1349 slip_devs[i] = NULL;
1350
1351 sl = netdev_priv(dev);
1352 if (sl->tty) {
1353 printk(KERN_ERR "%s: tty discipline still running\n",
1354 dev->name);
1355 }
1356
1357 unregister_netdev(dev);
1358 }
1359
1360 kfree(slip_devs);
1361 slip_devs = NULL;
1362
1363 i = tty_unregister_ldisc(N_SLIP);
1364 if (i != 0)
1365 printk(KERN_ERR "SLIP: can't unregister line discipline (err = %d)\n", i);
1366}
1367
1368module_init(slip_init);
1369module_exit(slip_exit);
1370
1371#ifdef CONFIG_SLIP_SMART
1372/*
1373 * This is start of the code for multislip style line checking
1374 * added by Stanislav Voronyi. All changes before marked VSV
1375 */
1376
1377static void sl_outfill(struct timer_list *t)
1378{
1379 struct slip *sl = from_timer(sl, t, outfill_timer);
1380
1381 spin_lock(&sl->lock);
1382
1383 if (sl->tty == NULL)
1384 goto out;
1385
1386 if (sl->outfill) {
1387 if (test_bit(SLF_OUTWAIT, &sl->flags)) {
1388 /* no packets were transmitted, do outfill */
1389#ifdef CONFIG_SLIP_MODE_SLIP6
1390 unsigned char s = (sl->mode & SL_MODE_SLIP6)?0x70:END;
1391#else
1392 unsigned char s = END;
1393#endif
1394 /* put END into tty queue. Is it right ??? */
1395 if (!netif_queue_stopped(sl->dev)) {
1396 /* if device busy no outfill */
1397 sl->tty->ops->write(sl->tty, &s, 1);
1398 }
1399 } else
1400 set_bit(SLF_OUTWAIT, &sl->flags);
1401
1402 mod_timer(&sl->outfill_timer, jiffies+sl->outfill*HZ);
1403 }
1404out:
1405 spin_unlock(&sl->lock);
1406}
1407
1408static void sl_keepalive(struct timer_list *t)
1409{
1410 struct slip *sl = from_timer(sl, t, keepalive_timer);
1411
1412 spin_lock(&sl->lock);
1413
1414 if (sl->tty == NULL)
1415 goto out;
1416
1417 if (sl->keepalive) {
1418 if (test_bit(SLF_KEEPTEST, &sl->flags)) {
1419 /* keepalive still high :(, we must hangup */
1420 if (sl->outfill)
1421 /* outfill timer must be deleted too */
1422 (void)del_timer(&sl->outfill_timer);
1423 printk(KERN_DEBUG "%s: no packets received during keepalive timeout, hangup.\n", sl->dev->name);
1424 /* this must hangup tty & close slip */
1425 tty_hangup(sl->tty);
1426 /* I think we need not something else */
1427 goto out;
1428 } else
1429 set_bit(SLF_KEEPTEST, &sl->flags);
1430
1431 mod_timer(&sl->keepalive_timer, jiffies+sl->keepalive*HZ);
1432 }
1433out:
1434 spin_unlock(&sl->lock);
1435}
1436
1437#endif
1438MODULE_LICENSE("GPL");
1439MODULE_ALIAS_LDISC(N_SLIP);
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * slip.c This module implements the SLIP protocol for kernel-based
4 * devices like TTY. It interfaces between a raw TTY, and the
5 * kernel's INET protocol layers.
6 *
7 * Version: @(#)slip.c 0.8.3 12/24/94
8 *
9 * Authors: Laurence Culhane, <loz@holmes.demon.co.uk>
10 * Fred N. van Kempen, <waltje@uwalt.nl.mugnet.org>
11 *
12 * Fixes:
13 * Alan Cox : Sanity checks and avoid tx overruns.
14 * Has a new sl->mtu field.
15 * Alan Cox : Found cause of overrun. ifconfig sl0
16 * mtu upwards. Driver now spots this
17 * and grows/shrinks its buffers(hack!).
18 * Memory leak if you run out of memory
19 * setting up a slip driver fixed.
20 * Matt Dillon : Printable slip (borrowed from NET2E)
21 * Pauline Middelink : Slip driver fixes.
22 * Alan Cox : Honours the old SL_COMPRESSED flag
23 * Alan Cox : KISS AX.25 and AXUI IP support
24 * Michael Riepe : Automatic CSLIP recognition added
25 * Charles Hedrick : CSLIP header length problem fix.
26 * Alan Cox : Corrected non-IP cases of the above.
27 * Alan Cox : Now uses hardware type as per FvK.
28 * Alan Cox : Default to 192.168.0.0 (RFC 1597)
29 * A.N.Kuznetsov : dev_tint() recursion fix.
30 * Dmitry Gorodchanin : SLIP memory leaks
31 * Dmitry Gorodchanin : Code cleanup. Reduce tty driver
32 * buffering from 4096 to 256 bytes.
33 * Improving SLIP response time.
34 * CONFIG_SLIP_MODE_SLIP6.
35 * ifconfig sl? up & down now works
36 * correctly.
37 * Modularization.
38 * Alan Cox : Oops - fix AX.25 buffer lengths
39 * Dmitry Gorodchanin : Even more cleanups. Preserve CSLIP
40 * statistics. Include CSLIP code only
41 * if it really needed.
42 * Alan Cox : Free slhc buffers in the right place.
43 * Alan Cox : Allow for digipeated IP over AX.25
44 * Matti Aarnio : Dynamic SLIP devices, with ideas taken
45 * from Jim Freeman's <jfree@caldera.com>
46 * dynamic PPP devices. We do NOT kfree()
47 * device entries, just reg./unreg. them
48 * as they are needed. We kfree() them
49 * at module cleanup.
50 * With MODULE-loading ``insmod'', user
51 * can issue parameter: slip_maxdev=1024
52 * (Or how much he/she wants.. Default
53 * is 256)
54 * Stanislav Voronyi : Slip line checking, with ideas taken
55 * from multislip BSDI driver which was
56 * written by Igor Chechik, RELCOM Corp.
57 * Only algorithms have been ported to
58 * Linux SLIP driver.
59 * Vitaly E. Lavrov : Sane behaviour on tty hangup.
60 * Alexey Kuznetsov : Cleanup interfaces to tty & netdevice
61 * modules.
62 */
63
64#define SL_CHECK_TRANSMIT
65#include <linux/module.h>
66#include <linux/moduleparam.h>
67
68#include <linux/uaccess.h>
69#include <linux/bitops.h>
70#include <linux/sched/signal.h>
71#include <linux/string.h>
72#include <linux/mm.h>
73#include <linux/interrupt.h>
74#include <linux/in.h>
75#include <linux/tty.h>
76#include <linux/errno.h>
77#include <linux/netdevice.h>
78#include <linux/etherdevice.h>
79#include <linux/skbuff.h>
80#include <linux/rtnetlink.h>
81#include <linux/if_arp.h>
82#include <linux/if_slip.h>
83#include <linux/delay.h>
84#include <linux/init.h>
85#include <linux/slab.h>
86#include <linux/workqueue.h>
87#include "slip.h"
88#ifdef CONFIG_INET
89#include <linux/ip.h>
90#include <linux/tcp.h>
91#include <net/slhc_vj.h>
92#endif
93
94#define SLIP_VERSION "0.8.4-NET3.019-NEWTTY"
95
96static struct net_device **slip_devs;
97
98static int slip_maxdev = SL_NRUNIT;
99module_param(slip_maxdev, int, 0);
100MODULE_PARM_DESC(slip_maxdev, "Maximum number of slip devices");
101
102static int slip_esc(unsigned char *p, unsigned char *d, int len);
103static void slip_unesc(struct slip *sl, unsigned char c);
104#ifdef CONFIG_SLIP_MODE_SLIP6
105static int slip_esc6(unsigned char *p, unsigned char *d, int len);
106static void slip_unesc6(struct slip *sl, unsigned char c);
107#endif
108#ifdef CONFIG_SLIP_SMART
109static void sl_keepalive(struct timer_list *t);
110static void sl_outfill(struct timer_list *t);
111static int sl_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
112#endif
113
114/********************************
115* Buffer administration routines:
116* sl_alloc_bufs()
117* sl_free_bufs()
118* sl_realloc_bufs()
119*
120* NOTE: sl_realloc_bufs != sl_free_bufs + sl_alloc_bufs, because
121* sl_realloc_bufs provides strong atomicity and reallocation
122* on actively running device.
123*********************************/
124
125/*
126 Allocate channel buffers.
127 */
128
129static int sl_alloc_bufs(struct slip *sl, int mtu)
130{
131 int err = -ENOBUFS;
132 unsigned long len;
133 char *rbuff = NULL;
134 char *xbuff = NULL;
135#ifdef SL_INCLUDE_CSLIP
136 char *cbuff = NULL;
137 struct slcompress *slcomp = NULL;
138#endif
139
140 /*
141 * Allocate the SLIP frame buffers:
142 *
143 * rbuff Receive buffer.
144 * xbuff Transmit buffer.
145 * cbuff Temporary compression buffer.
146 */
147 len = mtu * 2;
148
149 /*
150 * allow for arrival of larger UDP packets, even if we say not to
151 * also fixes a bug in which SunOS sends 512-byte packets even with
152 * an MSS of 128
153 */
154 if (len < 576 * 2)
155 len = 576 * 2;
156 rbuff = kmalloc(len + 4, GFP_KERNEL);
157 if (rbuff == NULL)
158 goto err_exit;
159 xbuff = kmalloc(len + 4, GFP_KERNEL);
160 if (xbuff == NULL)
161 goto err_exit;
162#ifdef SL_INCLUDE_CSLIP
163 cbuff = kmalloc(len + 4, GFP_KERNEL);
164 if (cbuff == NULL)
165 goto err_exit;
166 slcomp = slhc_init(16, 16);
167 if (IS_ERR(slcomp))
168 goto err_exit;
169#endif
170 spin_lock_bh(&sl->lock);
171 if (sl->tty == NULL) {
172 spin_unlock_bh(&sl->lock);
173 err = -ENODEV;
174 goto err_exit;
175 }
176 sl->mtu = mtu;
177 sl->buffsize = len;
178 sl->rcount = 0;
179 sl->xleft = 0;
180 rbuff = xchg(&sl->rbuff, rbuff);
181 xbuff = xchg(&sl->xbuff, xbuff);
182#ifdef SL_INCLUDE_CSLIP
183 cbuff = xchg(&sl->cbuff, cbuff);
184 slcomp = xchg(&sl->slcomp, slcomp);
185#endif
186#ifdef CONFIG_SLIP_MODE_SLIP6
187 sl->xdata = 0;
188 sl->xbits = 0;
189#endif
190 spin_unlock_bh(&sl->lock);
191 err = 0;
192
193 /* Cleanup */
194err_exit:
195#ifdef SL_INCLUDE_CSLIP
196 kfree(cbuff);
197 slhc_free(slcomp);
198#endif
199 kfree(xbuff);
200 kfree(rbuff);
201 return err;
202}
203
204/* Free a SLIP channel buffers. */
205static void sl_free_bufs(struct slip *sl)
206{
207 /* Free all SLIP frame buffers. */
208 kfree(xchg(&sl->rbuff, NULL));
209 kfree(xchg(&sl->xbuff, NULL));
210#ifdef SL_INCLUDE_CSLIP
211 kfree(xchg(&sl->cbuff, NULL));
212 slhc_free(xchg(&sl->slcomp, NULL));
213#endif
214}
215
216/*
217 Reallocate slip channel buffers.
218 */
219
220static int sl_realloc_bufs(struct slip *sl, int mtu)
221{
222 int err = 0;
223 struct net_device *dev = sl->dev;
224 unsigned char *xbuff, *rbuff;
225#ifdef SL_INCLUDE_CSLIP
226 unsigned char *cbuff;
227#endif
228 int len = mtu * 2;
229
230/*
231 * allow for arrival of larger UDP packets, even if we say not to
232 * also fixes a bug in which SunOS sends 512-byte packets even with
233 * an MSS of 128
234 */
235 if (len < 576 * 2)
236 len = 576 * 2;
237
238 xbuff = kmalloc(len + 4, GFP_ATOMIC);
239 rbuff = kmalloc(len + 4, GFP_ATOMIC);
240#ifdef SL_INCLUDE_CSLIP
241 cbuff = kmalloc(len + 4, GFP_ATOMIC);
242#endif
243
244
245#ifdef SL_INCLUDE_CSLIP
246 if (xbuff == NULL || rbuff == NULL || cbuff == NULL) {
247#else
248 if (xbuff == NULL || rbuff == NULL) {
249#endif
250 if (mtu > sl->mtu) {
251 printk(KERN_WARNING "%s: unable to grow slip buffers, MTU change cancelled.\n",
252 dev->name);
253 err = -ENOBUFS;
254 }
255 goto done;
256 }
257 spin_lock_bh(&sl->lock);
258
259 err = -ENODEV;
260 if (sl->tty == NULL)
261 goto done_on_bh;
262
263 xbuff = xchg(&sl->xbuff, xbuff);
264 rbuff = xchg(&sl->rbuff, rbuff);
265#ifdef SL_INCLUDE_CSLIP
266 cbuff = xchg(&sl->cbuff, cbuff);
267#endif
268 if (sl->xleft) {
269 if (sl->xleft <= len) {
270 memcpy(sl->xbuff, sl->xhead, sl->xleft);
271 } else {
272 sl->xleft = 0;
273 dev->stats.tx_dropped++;
274 }
275 }
276 sl->xhead = sl->xbuff;
277
278 if (sl->rcount) {
279 if (sl->rcount <= len) {
280 memcpy(sl->rbuff, rbuff, sl->rcount);
281 } else {
282 sl->rcount = 0;
283 dev->stats.rx_over_errors++;
284 set_bit(SLF_ERROR, &sl->flags);
285 }
286 }
287 sl->mtu = mtu;
288 dev->mtu = mtu;
289 sl->buffsize = len;
290 err = 0;
291
292done_on_bh:
293 spin_unlock_bh(&sl->lock);
294
295done:
296 kfree(xbuff);
297 kfree(rbuff);
298#ifdef SL_INCLUDE_CSLIP
299 kfree(cbuff);
300#endif
301 return err;
302}
303
304
305/* Set the "sending" flag. This must be atomic hence the set_bit. */
306static inline void sl_lock(struct slip *sl)
307{
308 netif_stop_queue(sl->dev);
309}
310
311
312/* Clear the "sending" flag. This must be atomic, hence the ASM. */
313static inline void sl_unlock(struct slip *sl)
314{
315 netif_wake_queue(sl->dev);
316}
317
318/* Send one completely decapsulated IP datagram to the IP layer. */
319static void sl_bump(struct slip *sl)
320{
321 struct net_device *dev = sl->dev;
322 struct sk_buff *skb;
323 int count;
324
325 count = sl->rcount;
326#ifdef SL_INCLUDE_CSLIP
327 if (sl->mode & (SL_MODE_ADAPTIVE | SL_MODE_CSLIP)) {
328 unsigned char c = sl->rbuff[0];
329 if (c & SL_TYPE_COMPRESSED_TCP) {
330 /* ignore compressed packets when CSLIP is off */
331 if (!(sl->mode & SL_MODE_CSLIP)) {
332 printk(KERN_WARNING "%s: compressed packet ignored\n", dev->name);
333 return;
334 }
335 /* make sure we've reserved enough space for uncompress
336 to use */
337 if (count + 80 > sl->buffsize) {
338 dev->stats.rx_over_errors++;
339 return;
340 }
341 count = slhc_uncompress(sl->slcomp, sl->rbuff, count);
342 if (count <= 0)
343 return;
344 } else if (c >= SL_TYPE_UNCOMPRESSED_TCP) {
345 if (!(sl->mode & SL_MODE_CSLIP)) {
346 /* turn on header compression */
347 sl->mode |= SL_MODE_CSLIP;
348 sl->mode &= ~SL_MODE_ADAPTIVE;
349 printk(KERN_INFO "%s: header compression turned on\n", dev->name);
350 }
351 sl->rbuff[0] &= 0x4f;
352 if (slhc_remember(sl->slcomp, sl->rbuff, count) <= 0)
353 return;
354 }
355 }
356#endif /* SL_INCLUDE_CSLIP */
357
358 dev->stats.rx_bytes += count;
359
360 skb = dev_alloc_skb(count);
361 if (skb == NULL) {
362 printk(KERN_WARNING "%s: memory squeeze, dropping packet.\n", dev->name);
363 dev->stats.rx_dropped++;
364 return;
365 }
366 skb->dev = dev;
367 skb_put_data(skb, sl->rbuff, count);
368 skb_reset_mac_header(skb);
369 skb->protocol = htons(ETH_P_IP);
370 netif_rx_ni(skb);
371 dev->stats.rx_packets++;
372}
373
374/* Encapsulate one IP datagram and stuff into a TTY queue. */
375static void sl_encaps(struct slip *sl, unsigned char *icp, int len)
376{
377 unsigned char *p;
378 int actual, count;
379
380 if (len > sl->mtu) { /* Sigh, shouldn't occur BUT ... */
381 printk(KERN_WARNING "%s: truncating oversized transmit packet!\n", sl->dev->name);
382 sl->dev->stats.tx_dropped++;
383 sl_unlock(sl);
384 return;
385 }
386
387 p = icp;
388#ifdef SL_INCLUDE_CSLIP
389 if (sl->mode & SL_MODE_CSLIP)
390 len = slhc_compress(sl->slcomp, p, len, sl->cbuff, &p, 1);
391#endif
392#ifdef CONFIG_SLIP_MODE_SLIP6
393 if (sl->mode & SL_MODE_SLIP6)
394 count = slip_esc6(p, sl->xbuff, len);
395 else
396#endif
397 count = slip_esc(p, sl->xbuff, len);
398
399 /* Order of next two lines is *very* important.
400 * When we are sending a little amount of data,
401 * the transfer may be completed inside the ops->write()
402 * routine, because it's running with interrupts enabled.
403 * In this case we *never* got WRITE_WAKEUP event,
404 * if we did not request it before write operation.
405 * 14 Oct 1994 Dmitry Gorodchanin.
406 */
407 set_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
408 actual = sl->tty->ops->write(sl->tty, sl->xbuff, count);
409#ifdef SL_CHECK_TRANSMIT
410 netif_trans_update(sl->dev);
411#endif
412 sl->xleft = count - actual;
413 sl->xhead = sl->xbuff + actual;
414#ifdef CONFIG_SLIP_SMART
415 /* VSV */
416 clear_bit(SLF_OUTWAIT, &sl->flags); /* reset outfill flag */
417#endif
418}
419
420/* Write out any remaining transmit buffer. Scheduled when tty is writable */
421static void slip_transmit(struct work_struct *work)
422{
423 struct slip *sl = container_of(work, struct slip, tx_work);
424 int actual;
425
426 spin_lock_bh(&sl->lock);
427 /* First make sure we're connected. */
428 if (!sl->tty || sl->magic != SLIP_MAGIC || !netif_running(sl->dev)) {
429 spin_unlock_bh(&sl->lock);
430 return;
431 }
432
433 if (sl->xleft <= 0) {
434 /* Now serial buffer is almost free & we can start
435 * transmission of another packet */
436 sl->dev->stats.tx_packets++;
437 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
438 spin_unlock_bh(&sl->lock);
439 sl_unlock(sl);
440 return;
441 }
442
443 actual = sl->tty->ops->write(sl->tty, sl->xhead, sl->xleft);
444 sl->xleft -= actual;
445 sl->xhead += actual;
446 spin_unlock_bh(&sl->lock);
447}
448
449/*
450 * Called by the driver when there's room for more data.
451 * Schedule the transmit.
452 */
453static void slip_write_wakeup(struct tty_struct *tty)
454{
455 struct slip *sl = tty->disc_data;
456
457 schedule_work(&sl->tx_work);
458}
459
460static void sl_tx_timeout(struct net_device *dev)
461{
462 struct slip *sl = netdev_priv(dev);
463
464 spin_lock(&sl->lock);
465
466 if (netif_queue_stopped(dev)) {
467 if (!netif_running(dev))
468 goto out;
469
470 /* May be we must check transmitter timeout here ?
471 * 14 Oct 1994 Dmitry Gorodchanin.
472 */
473#ifdef SL_CHECK_TRANSMIT
474 if (time_before(jiffies, dev_trans_start(dev) + 20 * HZ)) {
475 /* 20 sec timeout not reached */
476 goto out;
477 }
478 printk(KERN_WARNING "%s: transmit timed out, %s?\n",
479 dev->name,
480 (tty_chars_in_buffer(sl->tty) || sl->xleft) ?
481 "bad line quality" : "driver error");
482 sl->xleft = 0;
483 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
484 sl_unlock(sl);
485#endif
486 }
487out:
488 spin_unlock(&sl->lock);
489}
490
491
492/* Encapsulate an IP datagram and kick it into a TTY queue. */
493static netdev_tx_t
494sl_xmit(struct sk_buff *skb, struct net_device *dev)
495{
496 struct slip *sl = netdev_priv(dev);
497
498 spin_lock(&sl->lock);
499 if (!netif_running(dev)) {
500 spin_unlock(&sl->lock);
501 printk(KERN_WARNING "%s: xmit call when iface is down\n", dev->name);
502 dev_kfree_skb(skb);
503 return NETDEV_TX_OK;
504 }
505 if (sl->tty == NULL) {
506 spin_unlock(&sl->lock);
507 dev_kfree_skb(skb);
508 return NETDEV_TX_OK;
509 }
510
511 sl_lock(sl);
512 dev->stats.tx_bytes += skb->len;
513 sl_encaps(sl, skb->data, skb->len);
514 spin_unlock(&sl->lock);
515
516 dev_kfree_skb(skb);
517 return NETDEV_TX_OK;
518}
519
520
521/******************************************
522 * Routines looking at netdevice side.
523 ******************************************/
524
525/* Netdevice UP -> DOWN routine */
526
527static int
528sl_close(struct net_device *dev)
529{
530 struct slip *sl = netdev_priv(dev);
531
532 spin_lock_bh(&sl->lock);
533 if (sl->tty)
534 /* TTY discipline is running. */
535 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
536 netif_stop_queue(dev);
537 sl->rcount = 0;
538 sl->xleft = 0;
539 spin_unlock_bh(&sl->lock);
540
541 return 0;
542}
543
544/* Netdevice DOWN -> UP routine */
545
546static int sl_open(struct net_device *dev)
547{
548 struct slip *sl = netdev_priv(dev);
549
550 if (sl->tty == NULL)
551 return -ENODEV;
552
553 sl->flags &= (1 << SLF_INUSE);
554 netif_start_queue(dev);
555 return 0;
556}
557
558/* Netdevice change MTU request */
559
560static int sl_change_mtu(struct net_device *dev, int new_mtu)
561{
562 struct slip *sl = netdev_priv(dev);
563
564 return sl_realloc_bufs(sl, new_mtu);
565}
566
567/* Netdevice get statistics request */
568
569static void
570sl_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
571{
572 struct net_device_stats *devstats = &dev->stats;
573#ifdef SL_INCLUDE_CSLIP
574 struct slip *sl = netdev_priv(dev);
575 struct slcompress *comp = sl->slcomp;
576#endif
577 stats->rx_packets = devstats->rx_packets;
578 stats->tx_packets = devstats->tx_packets;
579 stats->rx_bytes = devstats->rx_bytes;
580 stats->tx_bytes = devstats->tx_bytes;
581 stats->rx_dropped = devstats->rx_dropped;
582 stats->tx_dropped = devstats->tx_dropped;
583 stats->tx_errors = devstats->tx_errors;
584 stats->rx_errors = devstats->rx_errors;
585 stats->rx_over_errors = devstats->rx_over_errors;
586
587#ifdef SL_INCLUDE_CSLIP
588 if (comp) {
589 /* Generic compressed statistics */
590 stats->rx_compressed = comp->sls_i_compressed;
591 stats->tx_compressed = comp->sls_o_compressed;
592
593 /* Are we really still needs this? */
594 stats->rx_fifo_errors += comp->sls_i_compressed;
595 stats->rx_dropped += comp->sls_i_tossed;
596 stats->tx_fifo_errors += comp->sls_o_compressed;
597 stats->collisions += comp->sls_o_misses;
598 }
599#endif
600}
601
602/* Netdevice register callback */
603
604static int sl_init(struct net_device *dev)
605{
606 struct slip *sl = netdev_priv(dev);
607
608 /*
609 * Finish setting up the DEVICE info.
610 */
611
612 dev->mtu = sl->mtu;
613 dev->type = ARPHRD_SLIP + sl->mode;
614#ifdef SL_CHECK_TRANSMIT
615 dev->watchdog_timeo = 20*HZ;
616#endif
617 return 0;
618}
619
620
621static void sl_uninit(struct net_device *dev)
622{
623 struct slip *sl = netdev_priv(dev);
624
625 sl_free_bufs(sl);
626}
627
628/* Hook the destructor so we can free slip devices at the right point in time */
629static void sl_free_netdev(struct net_device *dev)
630{
631 int i = dev->base_addr;
632
633 slip_devs[i] = NULL;
634}
635
636static const struct net_device_ops sl_netdev_ops = {
637 .ndo_init = sl_init,
638 .ndo_uninit = sl_uninit,
639 .ndo_open = sl_open,
640 .ndo_stop = sl_close,
641 .ndo_start_xmit = sl_xmit,
642 .ndo_get_stats64 = sl_get_stats64,
643 .ndo_change_mtu = sl_change_mtu,
644 .ndo_tx_timeout = sl_tx_timeout,
645#ifdef CONFIG_SLIP_SMART
646 .ndo_do_ioctl = sl_ioctl,
647#endif
648};
649
650
651static void sl_setup(struct net_device *dev)
652{
653 dev->netdev_ops = &sl_netdev_ops;
654 dev->needs_free_netdev = true;
655 dev->priv_destructor = sl_free_netdev;
656
657 dev->hard_header_len = 0;
658 dev->addr_len = 0;
659 dev->tx_queue_len = 10;
660
661 /* MTU range: 68 - 65534 */
662 dev->min_mtu = 68;
663 dev->max_mtu = 65534;
664
665 /* New-style flags. */
666 dev->flags = IFF_NOARP|IFF_POINTOPOINT|IFF_MULTICAST;
667}
668
669/******************************************
670 Routines looking at TTY side.
671 ******************************************/
672
673
674/*
675 * Handle the 'receiver data ready' interrupt.
676 * This function is called by the 'tty_io' module in the kernel when
677 * a block of SLIP data has been received, which can now be decapsulated
678 * and sent on to some IP layer for further processing. This will not
679 * be re-entered while running but other ldisc functions may be called
680 * in parallel
681 */
682
683static void slip_receive_buf(struct tty_struct *tty, const unsigned char *cp,
684 char *fp, int count)
685{
686 struct slip *sl = tty->disc_data;
687
688 if (!sl || sl->magic != SLIP_MAGIC || !netif_running(sl->dev))
689 return;
690
691 /* Read the characters out of the buffer */
692 while (count--) {
693 if (fp && *fp++) {
694 if (!test_and_set_bit(SLF_ERROR, &sl->flags))
695 sl->dev->stats.rx_errors++;
696 cp++;
697 continue;
698 }
699#ifdef CONFIG_SLIP_MODE_SLIP6
700 if (sl->mode & SL_MODE_SLIP6)
701 slip_unesc6(sl, *cp++);
702 else
703#endif
704 slip_unesc(sl, *cp++);
705 }
706}
707
708/************************************
709 * slip_open helper routines.
710 ************************************/
711
712/* Collect hanged up channels */
713static void sl_sync(void)
714{
715 int i;
716 struct net_device *dev;
717 struct slip *sl;
718
719 for (i = 0; i < slip_maxdev; i++) {
720 dev = slip_devs[i];
721 if (dev == NULL)
722 break;
723
724 sl = netdev_priv(dev);
725 if (sl->tty || sl->leased)
726 continue;
727 if (dev->flags & IFF_UP)
728 dev_close(dev);
729 }
730}
731
732
733/* Find a free SLIP channel, and link in this `tty' line. */
734static struct slip *sl_alloc(void)
735{
736 int i;
737 char name[IFNAMSIZ];
738 struct net_device *dev = NULL;
739 struct slip *sl;
740
741 for (i = 0; i < slip_maxdev; i++) {
742 dev = slip_devs[i];
743 if (dev == NULL)
744 break;
745 }
746 /* Sorry, too many, all slots in use */
747 if (i >= slip_maxdev)
748 return NULL;
749
750 sprintf(name, "sl%d", i);
751 dev = alloc_netdev(sizeof(*sl), name, NET_NAME_UNKNOWN, sl_setup);
752 if (!dev)
753 return NULL;
754
755 dev->base_addr = i;
756 sl = netdev_priv(dev);
757
758 /* Initialize channel control data */
759 sl->magic = SLIP_MAGIC;
760 sl->dev = dev;
761 spin_lock_init(&sl->lock);
762 INIT_WORK(&sl->tx_work, slip_transmit);
763 sl->mode = SL_MODE_DEFAULT;
764#ifdef CONFIG_SLIP_SMART
765 /* initialize timer_list struct */
766 timer_setup(&sl->keepalive_timer, sl_keepalive, 0);
767 timer_setup(&sl->outfill_timer, sl_outfill, 0);
768#endif
769 slip_devs[i] = dev;
770 return sl;
771}
772
773/*
774 * Open the high-level part of the SLIP channel.
775 * This function is called by the TTY module when the
776 * SLIP line discipline is called for. Because we are
777 * sure the tty line exists, we only have to link it to
778 * a free SLIP channel...
779 *
780 * Called in process context serialized from other ldisc calls.
781 */
782
783static int slip_open(struct tty_struct *tty)
784{
785 struct slip *sl;
786 int err;
787
788 if (!capable(CAP_NET_ADMIN))
789 return -EPERM;
790
791 if (tty->ops->write == NULL)
792 return -EOPNOTSUPP;
793
794 /* RTnetlink lock is misused here to serialize concurrent
795 opens of slip channels. There are better ways, but it is
796 the simplest one.
797 */
798 rtnl_lock();
799
800 /* Collect hanged up channels. */
801 sl_sync();
802
803 sl = tty->disc_data;
804
805 err = -EEXIST;
806 /* First make sure we're not already connected. */
807 if (sl && sl->magic == SLIP_MAGIC)
808 goto err_exit;
809
810 /* OK. Find a free SLIP channel to use. */
811 err = -ENFILE;
812 sl = sl_alloc();
813 if (sl == NULL)
814 goto err_exit;
815
816 sl->tty = tty;
817 tty->disc_data = sl;
818 sl->pid = current->pid;
819
820 if (!test_bit(SLF_INUSE, &sl->flags)) {
821 /* Perform the low-level SLIP initialization. */
822 err = sl_alloc_bufs(sl, SL_MTU);
823 if (err)
824 goto err_free_chan;
825
826 set_bit(SLF_INUSE, &sl->flags);
827
828 err = register_netdevice(sl->dev);
829 if (err)
830 goto err_free_bufs;
831 }
832
833#ifdef CONFIG_SLIP_SMART
834 if (sl->keepalive) {
835 sl->keepalive_timer.expires = jiffies + sl->keepalive * HZ;
836 add_timer(&sl->keepalive_timer);
837 }
838 if (sl->outfill) {
839 sl->outfill_timer.expires = jiffies + sl->outfill * HZ;
840 add_timer(&sl->outfill_timer);
841 }
842#endif
843
844 /* Done. We have linked the TTY line to a channel. */
845 rtnl_unlock();
846 tty->receive_room = 65536; /* We don't flow control */
847
848 /* TTY layer expects 0 on success */
849 return 0;
850
851err_free_bufs:
852 sl_free_bufs(sl);
853
854err_free_chan:
855 sl->tty = NULL;
856 tty->disc_data = NULL;
857 clear_bit(SLF_INUSE, &sl->flags);
858 free_netdev(sl->dev);
859
860err_exit:
861 rtnl_unlock();
862
863 /* Count references from TTY module */
864 return err;
865}
866
867/*
868 * Close down a SLIP channel.
869 * This means flushing out any pending queues, and then returning. This
870 * call is serialized against other ldisc functions.
871 *
872 * We also use this method fo a hangup event
873 */
874
875static void slip_close(struct tty_struct *tty)
876{
877 struct slip *sl = tty->disc_data;
878
879 /* First make sure we're connected. */
880 if (!sl || sl->magic != SLIP_MAGIC || sl->tty != tty)
881 return;
882
883 spin_lock_bh(&sl->lock);
884 tty->disc_data = NULL;
885 sl->tty = NULL;
886 spin_unlock_bh(&sl->lock);
887
888 flush_work(&sl->tx_work);
889
890 /* VSV = very important to remove timers */
891#ifdef CONFIG_SLIP_SMART
892 del_timer_sync(&sl->keepalive_timer);
893 del_timer_sync(&sl->outfill_timer);
894#endif
895 /* Flush network side */
896 unregister_netdev(sl->dev);
897 /* This will complete via sl_free_netdev */
898}
899
900static int slip_hangup(struct tty_struct *tty)
901{
902 slip_close(tty);
903 return 0;
904}
905 /************************************************************************
906 * STANDARD SLIP ENCAPSULATION *
907 ************************************************************************/
908
909static int slip_esc(unsigned char *s, unsigned char *d, int len)
910{
911 unsigned char *ptr = d;
912 unsigned char c;
913
914 /*
915 * Send an initial END character to flush out any
916 * data that may have accumulated in the receiver
917 * due to line noise.
918 */
919
920 *ptr++ = END;
921
922 /*
923 * For each byte in the packet, send the appropriate
924 * character sequence, according to the SLIP protocol.
925 */
926
927 while (len-- > 0) {
928 switch (c = *s++) {
929 case END:
930 *ptr++ = ESC;
931 *ptr++ = ESC_END;
932 break;
933 case ESC:
934 *ptr++ = ESC;
935 *ptr++ = ESC_ESC;
936 break;
937 default:
938 *ptr++ = c;
939 break;
940 }
941 }
942 *ptr++ = END;
943 return ptr - d;
944}
945
946static void slip_unesc(struct slip *sl, unsigned char s)
947{
948
949 switch (s) {
950 case END:
951#ifdef CONFIG_SLIP_SMART
952 /* drop keeptest bit = VSV */
953 if (test_bit(SLF_KEEPTEST, &sl->flags))
954 clear_bit(SLF_KEEPTEST, &sl->flags);
955#endif
956
957 if (!test_and_clear_bit(SLF_ERROR, &sl->flags) &&
958 (sl->rcount > 2))
959 sl_bump(sl);
960 clear_bit(SLF_ESCAPE, &sl->flags);
961 sl->rcount = 0;
962 return;
963
964 case ESC:
965 set_bit(SLF_ESCAPE, &sl->flags);
966 return;
967 case ESC_ESC:
968 if (test_and_clear_bit(SLF_ESCAPE, &sl->flags))
969 s = ESC;
970 break;
971 case ESC_END:
972 if (test_and_clear_bit(SLF_ESCAPE, &sl->flags))
973 s = END;
974 break;
975 }
976 if (!test_bit(SLF_ERROR, &sl->flags)) {
977 if (sl->rcount < sl->buffsize) {
978 sl->rbuff[sl->rcount++] = s;
979 return;
980 }
981 sl->dev->stats.rx_over_errors++;
982 set_bit(SLF_ERROR, &sl->flags);
983 }
984}
985
986
987#ifdef CONFIG_SLIP_MODE_SLIP6
988/************************************************************************
989 * 6 BIT SLIP ENCAPSULATION *
990 ************************************************************************/
991
992static int slip_esc6(unsigned char *s, unsigned char *d, int len)
993{
994 unsigned char *ptr = d;
995 unsigned char c;
996 int i;
997 unsigned short v = 0;
998 short bits = 0;
999
1000 /*
1001 * Send an initial END character to flush out any
1002 * data that may have accumulated in the receiver
1003 * due to line noise.
1004 */
1005
1006 *ptr++ = 0x70;
1007
1008 /*
1009 * Encode the packet into printable ascii characters
1010 */
1011
1012 for (i = 0; i < len; ++i) {
1013 v = (v << 8) | s[i];
1014 bits += 8;
1015 while (bits >= 6) {
1016 bits -= 6;
1017 c = 0x30 + ((v >> bits) & 0x3F);
1018 *ptr++ = c;
1019 }
1020 }
1021 if (bits) {
1022 c = 0x30 + ((v << (6 - bits)) & 0x3F);
1023 *ptr++ = c;
1024 }
1025 *ptr++ = 0x70;
1026 return ptr - d;
1027}
1028
1029static void slip_unesc6(struct slip *sl, unsigned char s)
1030{
1031 unsigned char c;
1032
1033 if (s == 0x70) {
1034#ifdef CONFIG_SLIP_SMART
1035 /* drop keeptest bit = VSV */
1036 if (test_bit(SLF_KEEPTEST, &sl->flags))
1037 clear_bit(SLF_KEEPTEST, &sl->flags);
1038#endif
1039
1040 if (!test_and_clear_bit(SLF_ERROR, &sl->flags) &&
1041 (sl->rcount > 2))
1042 sl_bump(sl);
1043 sl->rcount = 0;
1044 sl->xbits = 0;
1045 sl->xdata = 0;
1046 } else if (s >= 0x30 && s < 0x70) {
1047 sl->xdata = (sl->xdata << 6) | ((s - 0x30) & 0x3F);
1048 sl->xbits += 6;
1049 if (sl->xbits >= 8) {
1050 sl->xbits -= 8;
1051 c = (unsigned char)(sl->xdata >> sl->xbits);
1052 if (!test_bit(SLF_ERROR, &sl->flags)) {
1053 if (sl->rcount < sl->buffsize) {
1054 sl->rbuff[sl->rcount++] = c;
1055 return;
1056 }
1057 sl->dev->stats.rx_over_errors++;
1058 set_bit(SLF_ERROR, &sl->flags);
1059 }
1060 }
1061 }
1062}
1063#endif /* CONFIG_SLIP_MODE_SLIP6 */
1064
1065/* Perform I/O control on an active SLIP channel. */
1066static int slip_ioctl(struct tty_struct *tty, struct file *file,
1067 unsigned int cmd, unsigned long arg)
1068{
1069 struct slip *sl = tty->disc_data;
1070 unsigned int tmp;
1071 int __user *p = (int __user *)arg;
1072
1073 /* First make sure we're connected. */
1074 if (!sl || sl->magic != SLIP_MAGIC)
1075 return -EINVAL;
1076
1077 switch (cmd) {
1078 case SIOCGIFNAME:
1079 tmp = strlen(sl->dev->name) + 1;
1080 if (copy_to_user((void __user *)arg, sl->dev->name, tmp))
1081 return -EFAULT;
1082 return 0;
1083
1084 case SIOCGIFENCAP:
1085 if (put_user(sl->mode, p))
1086 return -EFAULT;
1087 return 0;
1088
1089 case SIOCSIFENCAP:
1090 if (get_user(tmp, p))
1091 return -EFAULT;
1092#ifndef SL_INCLUDE_CSLIP
1093 if (tmp & (SL_MODE_CSLIP|SL_MODE_ADAPTIVE))
1094 return -EINVAL;
1095#else
1096 if ((tmp & (SL_MODE_ADAPTIVE | SL_MODE_CSLIP)) ==
1097 (SL_MODE_ADAPTIVE | SL_MODE_CSLIP))
1098 /* return -EINVAL; */
1099 tmp &= ~SL_MODE_ADAPTIVE;
1100#endif
1101#ifndef CONFIG_SLIP_MODE_SLIP6
1102 if (tmp & SL_MODE_SLIP6)
1103 return -EINVAL;
1104#endif
1105 sl->mode = tmp;
1106 sl->dev->type = ARPHRD_SLIP + sl->mode;
1107 return 0;
1108
1109 case SIOCSIFHWADDR:
1110 return -EINVAL;
1111
1112#ifdef CONFIG_SLIP_SMART
1113 /* VSV changes start here */
1114 case SIOCSKEEPALIVE:
1115 if (get_user(tmp, p))
1116 return -EFAULT;
1117 if (tmp > 255) /* max for unchar */
1118 return -EINVAL;
1119
1120 spin_lock_bh(&sl->lock);
1121 if (!sl->tty) {
1122 spin_unlock_bh(&sl->lock);
1123 return -ENODEV;
1124 }
1125 sl->keepalive = (u8)tmp;
1126 if (sl->keepalive != 0) {
1127 mod_timer(&sl->keepalive_timer,
1128 jiffies + sl->keepalive * HZ);
1129 set_bit(SLF_KEEPTEST, &sl->flags);
1130 } else
1131 del_timer(&sl->keepalive_timer);
1132 spin_unlock_bh(&sl->lock);
1133 return 0;
1134
1135 case SIOCGKEEPALIVE:
1136 if (put_user(sl->keepalive, p))
1137 return -EFAULT;
1138 return 0;
1139
1140 case SIOCSOUTFILL:
1141 if (get_user(tmp, p))
1142 return -EFAULT;
1143 if (tmp > 255) /* max for unchar */
1144 return -EINVAL;
1145 spin_lock_bh(&sl->lock);
1146 if (!sl->tty) {
1147 spin_unlock_bh(&sl->lock);
1148 return -ENODEV;
1149 }
1150 sl->outfill = (u8)tmp;
1151 if (sl->outfill != 0) {
1152 mod_timer(&sl->outfill_timer,
1153 jiffies + sl->outfill * HZ);
1154 set_bit(SLF_OUTWAIT, &sl->flags);
1155 } else
1156 del_timer(&sl->outfill_timer);
1157 spin_unlock_bh(&sl->lock);
1158 return 0;
1159
1160 case SIOCGOUTFILL:
1161 if (put_user(sl->outfill, p))
1162 return -EFAULT;
1163 return 0;
1164 /* VSV changes end */
1165#endif
1166 default:
1167 return tty_mode_ioctl(tty, file, cmd, arg);
1168 }
1169}
1170
1171/* VSV changes start here */
1172#ifdef CONFIG_SLIP_SMART
1173/* function do_ioctl called from net/core/dev.c
1174 to allow get/set outfill/keepalive parameter
1175 by ifconfig */
1176
1177static int sl_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
1178{
1179 struct slip *sl = netdev_priv(dev);
1180 unsigned long *p = (unsigned long *)&rq->ifr_ifru;
1181
1182 if (sl == NULL) /* Allocation failed ?? */
1183 return -ENODEV;
1184
1185 spin_lock_bh(&sl->lock);
1186
1187 if (!sl->tty) {
1188 spin_unlock_bh(&sl->lock);
1189 return -ENODEV;
1190 }
1191
1192 switch (cmd) {
1193 case SIOCSKEEPALIVE:
1194 /* max for unchar */
1195 if ((unsigned)*p > 255) {
1196 spin_unlock_bh(&sl->lock);
1197 return -EINVAL;
1198 }
1199 sl->keepalive = (u8)*p;
1200 if (sl->keepalive != 0) {
1201 sl->keepalive_timer.expires =
1202 jiffies + sl->keepalive * HZ;
1203 mod_timer(&sl->keepalive_timer,
1204 jiffies + sl->keepalive * HZ);
1205 set_bit(SLF_KEEPTEST, &sl->flags);
1206 } else
1207 del_timer(&sl->keepalive_timer);
1208 break;
1209
1210 case SIOCGKEEPALIVE:
1211 *p = sl->keepalive;
1212 break;
1213
1214 case SIOCSOUTFILL:
1215 if ((unsigned)*p > 255) { /* max for unchar */
1216 spin_unlock_bh(&sl->lock);
1217 return -EINVAL;
1218 }
1219 sl->outfill = (u8)*p;
1220 if (sl->outfill != 0) {
1221 mod_timer(&sl->outfill_timer,
1222 jiffies + sl->outfill * HZ);
1223 set_bit(SLF_OUTWAIT, &sl->flags);
1224 } else
1225 del_timer(&sl->outfill_timer);
1226 break;
1227
1228 case SIOCGOUTFILL:
1229 *p = sl->outfill;
1230 break;
1231
1232 case SIOCSLEASE:
1233 /* Resolve race condition, when ioctl'ing hanged up
1234 and opened by another process device.
1235 */
1236 if (sl->tty != current->signal->tty &&
1237 sl->pid != current->pid) {
1238 spin_unlock_bh(&sl->lock);
1239 return -EPERM;
1240 }
1241 sl->leased = 0;
1242 if (*p)
1243 sl->leased = 1;
1244 break;
1245
1246 case SIOCGLEASE:
1247 *p = sl->leased;
1248 }
1249 spin_unlock_bh(&sl->lock);
1250 return 0;
1251}
1252#endif
1253/* VSV changes end */
1254
1255static struct tty_ldisc_ops sl_ldisc = {
1256 .owner = THIS_MODULE,
1257 .magic = TTY_LDISC_MAGIC,
1258 .name = "slip",
1259 .open = slip_open,
1260 .close = slip_close,
1261 .hangup = slip_hangup,
1262 .ioctl = slip_ioctl,
1263 .receive_buf = slip_receive_buf,
1264 .write_wakeup = slip_write_wakeup,
1265};
1266
1267static int __init slip_init(void)
1268{
1269 int status;
1270
1271 if (slip_maxdev < 4)
1272 slip_maxdev = 4; /* Sanity */
1273
1274 printk(KERN_INFO "SLIP: version %s (dynamic channels, max=%d)"
1275#ifdef CONFIG_SLIP_MODE_SLIP6
1276 " (6 bit encapsulation enabled)"
1277#endif
1278 ".\n",
1279 SLIP_VERSION, slip_maxdev);
1280#if defined(SL_INCLUDE_CSLIP)
1281 printk(KERN_INFO "CSLIP: code copyright 1989 Regents of the University of California.\n");
1282#endif
1283#ifdef CONFIG_SLIP_SMART
1284 printk(KERN_INFO "SLIP linefill/keepalive option.\n");
1285#endif
1286
1287 slip_devs = kcalloc(slip_maxdev, sizeof(struct net_device *),
1288 GFP_KERNEL);
1289 if (!slip_devs)
1290 return -ENOMEM;
1291
1292 /* Fill in our line protocol discipline, and register it */
1293 status = tty_register_ldisc(N_SLIP, &sl_ldisc);
1294 if (status != 0) {
1295 printk(KERN_ERR "SLIP: can't register line discipline (err = %d)\n", status);
1296 kfree(slip_devs);
1297 }
1298 return status;
1299}
1300
1301static void __exit slip_exit(void)
1302{
1303 int i;
1304 struct net_device *dev;
1305 struct slip *sl;
1306 unsigned long timeout = jiffies + HZ;
1307 int busy = 0;
1308
1309 if (slip_devs == NULL)
1310 return;
1311
1312 /* First of all: check for active disciplines and hangup them.
1313 */
1314 do {
1315 if (busy)
1316 msleep_interruptible(100);
1317
1318 busy = 0;
1319 for (i = 0; i < slip_maxdev; i++) {
1320 dev = slip_devs[i];
1321 if (!dev)
1322 continue;
1323 sl = netdev_priv(dev);
1324 spin_lock_bh(&sl->lock);
1325 if (sl->tty) {
1326 busy++;
1327 tty_hangup(sl->tty);
1328 }
1329 spin_unlock_bh(&sl->lock);
1330 }
1331 } while (busy && time_before(jiffies, timeout));
1332
1333 /* FIXME: hangup is async so we should wait when doing this second
1334 phase */
1335
1336 for (i = 0; i < slip_maxdev; i++) {
1337 dev = slip_devs[i];
1338 if (!dev)
1339 continue;
1340 slip_devs[i] = NULL;
1341
1342 sl = netdev_priv(dev);
1343 if (sl->tty) {
1344 printk(KERN_ERR "%s: tty discipline still running\n",
1345 dev->name);
1346 }
1347
1348 unregister_netdev(dev);
1349 }
1350
1351 kfree(slip_devs);
1352 slip_devs = NULL;
1353
1354 i = tty_unregister_ldisc(N_SLIP);
1355 if (i != 0)
1356 printk(KERN_ERR "SLIP: can't unregister line discipline (err = %d)\n", i);
1357}
1358
1359module_init(slip_init);
1360module_exit(slip_exit);
1361
1362#ifdef CONFIG_SLIP_SMART
1363/*
1364 * This is start of the code for multislip style line checking
1365 * added by Stanislav Voronyi. All changes before marked VSV
1366 */
1367
1368static void sl_outfill(struct timer_list *t)
1369{
1370 struct slip *sl = from_timer(sl, t, outfill_timer);
1371
1372 spin_lock(&sl->lock);
1373
1374 if (sl->tty == NULL)
1375 goto out;
1376
1377 if (sl->outfill) {
1378 if (test_bit(SLF_OUTWAIT, &sl->flags)) {
1379 /* no packets were transmitted, do outfill */
1380#ifdef CONFIG_SLIP_MODE_SLIP6
1381 unsigned char s = (sl->mode & SL_MODE_SLIP6)?0x70:END;
1382#else
1383 unsigned char s = END;
1384#endif
1385 /* put END into tty queue. Is it right ??? */
1386 if (!netif_queue_stopped(sl->dev)) {
1387 /* if device busy no outfill */
1388 sl->tty->ops->write(sl->tty, &s, 1);
1389 }
1390 } else
1391 set_bit(SLF_OUTWAIT, &sl->flags);
1392
1393 mod_timer(&sl->outfill_timer, jiffies+sl->outfill*HZ);
1394 }
1395out:
1396 spin_unlock(&sl->lock);
1397}
1398
1399static void sl_keepalive(struct timer_list *t)
1400{
1401 struct slip *sl = from_timer(sl, t, keepalive_timer);
1402
1403 spin_lock(&sl->lock);
1404
1405 if (sl->tty == NULL)
1406 goto out;
1407
1408 if (sl->keepalive) {
1409 if (test_bit(SLF_KEEPTEST, &sl->flags)) {
1410 /* keepalive still high :(, we must hangup */
1411 if (sl->outfill)
1412 /* outfill timer must be deleted too */
1413 (void)del_timer(&sl->outfill_timer);
1414 printk(KERN_DEBUG "%s: no packets received during keepalive timeout, hangup.\n", sl->dev->name);
1415 /* this must hangup tty & close slip */
1416 tty_hangup(sl->tty);
1417 /* I think we need not something else */
1418 goto out;
1419 } else
1420 set_bit(SLF_KEEPTEST, &sl->flags);
1421
1422 mod_timer(&sl->keepalive_timer, jiffies+sl->keepalive*HZ);
1423 }
1424out:
1425 spin_unlock(&sl->lock);
1426}
1427
1428#endif
1429MODULE_LICENSE("GPL");
1430MODULE_ALIAS_LDISC(N_SLIP);