1 // SPDX-License-Identifier: GPL-2.0-only
3 * slip.c This module implements the SLIP protocol for kernel-based
4 * devices like TTY. It interfaces between a raw TTY, and the
5 * kernel's INET protocol layers.
7 * Version: @(#)slip.c 0.8.3 12/24/94
9 * Authors: Laurence Culhane, <loz@holmes.demon.co.uk>
10 * Fred N. van Kempen, <waltje@uwalt.nl.mugnet.org>
13 * Alan Cox : Sanity checks and avoid tx overruns.
14 * Has a new sl->mtu field.
15 * Alan Cox : Found cause of overrun. ifconfig sl0
16 * mtu upwards. Driver now spots this
17 * and grows/shrinks its buffers(hack!).
18 * Memory leak if you run out of memory
19 * setting up a slip driver fixed.
20 * Matt Dillon : Printable slip (borrowed from NET2E)
21 * Pauline Middelink : Slip driver fixes.
22 * Alan Cox : Honours the old SL_COMPRESSED flag
23 * Alan Cox : KISS AX.25 and AXUI IP support
24 * Michael Riepe : Automatic CSLIP recognition added
25 * Charles Hedrick : CSLIP header length problem fix.
26 * Alan Cox : Corrected non-IP cases of the above.
27 * Alan Cox : Now uses hardware type as per FvK.
28 * Alan Cox : Default to 192.168.0.0 (RFC 1597)
29 * A.N.Kuznetsov : dev_tint() recursion fix.
30 * Dmitry Gorodchanin : SLIP memory leaks
31 * Dmitry Gorodchanin : Code cleanup. Reduce tty driver
32 * buffering from 4096 to 256 bytes.
33 * Improving SLIP response time.
34 * CONFIG_SLIP_MODE_SLIP6.
35 * ifconfig sl? up & down now works
38 * Alan Cox : Oops - fix AX.25 buffer lengths
39 * Dmitry Gorodchanin : Even more cleanups. Preserve CSLIP
40 * statistics. Include CSLIP code only
41 * if it really needed.
42 * Alan Cox : Free slhc buffers in the right place.
43 * Alan Cox : Allow for digipeated IP over AX.25
44 * Matti Aarnio : Dynamic SLIP devices, with ideas taken
45 * from Jim Freeman's <jfree@caldera.com>
46 * dynamic PPP devices. We do NOT kfree()
47 * device entries, just reg./unreg. them
48 * as they are needed. We kfree() them
50 * With MODULE-loading ``insmod'', user
51 * can issue parameter: slip_maxdev=1024
52 * (Or how much he/she wants.. Default
54 * Stanislav Voronyi : Slip line checking, with ideas taken
55 * from multislip BSDI driver which was
56 * written by Igor Chechik, RELCOM Corp.
57 * Only algorithms have been ported to
59 * Vitaly E. Lavrov : Sane behaviour on tty hangup.
60 * Alexey Kuznetsov : Cleanup interfaces to tty & netdevice
64 #define SL_CHECK_TRANSMIT
65 #include <linux/compat.h>
66 #include <linux/module.h>
67 #include <linux/moduleparam.h>
69 #include <linux/uaccess.h>
70 #include <linux/bitops.h>
71 #include <linux/sched/signal.h>
72 #include <linux/string.h>
74 #include <linux/interrupt.h>
76 #include <linux/tty.h>
77 #include <linux/errno.h>
78 #include <linux/netdevice.h>
79 #include <linux/etherdevice.h>
80 #include <linux/skbuff.h>
81 #include <linux/rtnetlink.h>
82 #include <linux/if_arp.h>
83 #include <linux/if_slip.h>
84 #include <linux/delay.h>
85 #include <linux/init.h>
86 #include <linux/slab.h>
87 #include <linux/workqueue.h>
91 #include <linux/tcp.h>
92 #include <net/slhc_vj.h>
95 #define SLIP_VERSION "0.8.4-NET3.019-NEWTTY"
97 static struct net_device **slip_devs;
99 static int slip_maxdev = SL_NRUNIT;
100 module_param(slip_maxdev, int, 0);
101 MODULE_PARM_DESC(slip_maxdev, "Maximum number of slip devices");
103 static int slip_esc(unsigned char *p, unsigned char *d, int len);
104 static void slip_unesc(struct slip *sl, unsigned char c);
105 #ifdef CONFIG_SLIP_MODE_SLIP6
106 static int slip_esc6(unsigned char *p, unsigned char *d, int len);
107 static void slip_unesc6(struct slip *sl, unsigned char c);
109 #ifdef CONFIG_SLIP_SMART
110 static void sl_keepalive(struct timer_list *t);
111 static void sl_outfill(struct timer_list *t);
112 static int sl_siocdevprivate(struct net_device *dev, struct ifreq *rq, void __user *data, int cmd);
115 /********************************
116 * Buffer administration routines:
121 * NOTE: sl_realloc_bufs != sl_free_bufs + sl_alloc_bufs, because
122 * sl_realloc_bufs provides strong atomicity and reallocation
123 * on actively running device.
124 *********************************/
127 Allocate channel buffers.
130 static int sl_alloc_bufs(struct slip *sl, int mtu)
136 #ifdef SL_INCLUDE_CSLIP
138 struct slcompress *slcomp = NULL;
142 * Allocate the SLIP frame buffers:
144 * rbuff Receive buffer.
145 * xbuff Transmit buffer.
146 * cbuff Temporary compression buffer.
151 * allow for arrival of larger UDP packets, even if we say not to
152 * also fixes a bug in which SunOS sends 512-byte packets even with
157 rbuff = kmalloc(len + 4, GFP_KERNEL);
160 xbuff = kmalloc(len + 4, GFP_KERNEL);
163 #ifdef SL_INCLUDE_CSLIP
164 cbuff = kmalloc(len + 4, GFP_KERNEL);
167 slcomp = slhc_init(16, 16);
171 spin_lock_bh(&sl->lock);
172 if (sl->tty == NULL) {
173 spin_unlock_bh(&sl->lock);
181 rbuff = xchg(&sl->rbuff, rbuff);
182 xbuff = xchg(&sl->xbuff, xbuff);
183 #ifdef SL_INCLUDE_CSLIP
184 cbuff = xchg(&sl->cbuff, cbuff);
185 slcomp = xchg(&sl->slcomp, slcomp);
187 #ifdef CONFIG_SLIP_MODE_SLIP6
191 spin_unlock_bh(&sl->lock);
196 #ifdef SL_INCLUDE_CSLIP
205 /* Free a SLIP channel buffers. */
206 static void sl_free_bufs(struct slip *sl)
208 /* Free all SLIP frame buffers. */
209 kfree(xchg(&sl->rbuff, NULL));
210 kfree(xchg(&sl->xbuff, NULL));
211 #ifdef SL_INCLUDE_CSLIP
212 kfree(xchg(&sl->cbuff, NULL));
213 slhc_free(xchg(&sl->slcomp, NULL));
218 Reallocate slip channel buffers.
221 static int sl_realloc_bufs(struct slip *sl, int mtu)
224 struct net_device *dev = sl->dev;
225 unsigned char *xbuff, *rbuff;
226 #ifdef SL_INCLUDE_CSLIP
227 unsigned char *cbuff;
232 * allow for arrival of larger UDP packets, even if we say not to
233 * also fixes a bug in which SunOS sends 512-byte packets even with
239 xbuff = kmalloc(len + 4, GFP_ATOMIC);
240 rbuff = kmalloc(len + 4, GFP_ATOMIC);
241 #ifdef SL_INCLUDE_CSLIP
242 cbuff = kmalloc(len + 4, GFP_ATOMIC);
246 #ifdef SL_INCLUDE_CSLIP
247 if (xbuff == NULL || rbuff == NULL || cbuff == NULL) {
249 if (xbuff == NULL || rbuff == NULL) {
252 printk(KERN_WARNING "%s: unable to grow slip buffers, MTU change cancelled.\n",
258 spin_lock_bh(&sl->lock);
264 xbuff = xchg(&sl->xbuff, xbuff);
265 rbuff = xchg(&sl->rbuff, rbuff);
266 #ifdef SL_INCLUDE_CSLIP
267 cbuff = xchg(&sl->cbuff, cbuff);
270 if (sl->xleft <= len) {
271 memcpy(sl->xbuff, sl->xhead, sl->xleft);
274 dev->stats.tx_dropped++;
277 sl->xhead = sl->xbuff;
280 if (sl->rcount <= len) {
281 memcpy(sl->rbuff, rbuff, sl->rcount);
284 dev->stats.rx_over_errors++;
285 set_bit(SLF_ERROR, &sl->flags);
294 spin_unlock_bh(&sl->lock);
299 #ifdef SL_INCLUDE_CSLIP
306 /* Set the "sending" flag. This must be atomic hence the set_bit. */
307 static inline void sl_lock(struct slip *sl)
309 netif_stop_queue(sl->dev);
313 /* Clear the "sending" flag. This must be atomic, hence the ASM. */
314 static inline void sl_unlock(struct slip *sl)
316 netif_wake_queue(sl->dev);
319 /* Send one completely decapsulated IP datagram to the IP layer. */
320 static void sl_bump(struct slip *sl)
322 struct net_device *dev = sl->dev;
327 #ifdef SL_INCLUDE_CSLIP
328 if (sl->mode & (SL_MODE_ADAPTIVE | SL_MODE_CSLIP)) {
329 unsigned char c = sl->rbuff[0];
330 if (c & SL_TYPE_COMPRESSED_TCP) {
331 /* ignore compressed packets when CSLIP is off */
332 if (!(sl->mode & SL_MODE_CSLIP)) {
333 printk(KERN_WARNING "%s: compressed packet ignored\n", dev->name);
336 /* make sure we've reserved enough space for uncompress
338 if (count + 80 > sl->buffsize) {
339 dev->stats.rx_over_errors++;
342 count = slhc_uncompress(sl->slcomp, sl->rbuff, count);
345 } else if (c >= SL_TYPE_UNCOMPRESSED_TCP) {
346 if (!(sl->mode & SL_MODE_CSLIP)) {
347 /* turn on header compression */
348 sl->mode |= SL_MODE_CSLIP;
349 sl->mode &= ~SL_MODE_ADAPTIVE;
350 printk(KERN_INFO "%s: header compression turned on\n", dev->name);
352 sl->rbuff[0] &= 0x4f;
353 if (slhc_remember(sl->slcomp, sl->rbuff, count) <= 0)
357 #endif /* SL_INCLUDE_CSLIP */
359 dev->stats.rx_bytes += count;
361 skb = dev_alloc_skb(count);
363 printk(KERN_WARNING "%s: memory squeeze, dropping packet.\n", dev->name);
364 dev->stats.rx_dropped++;
368 skb_put_data(skb, sl->rbuff, count);
369 skb_reset_mac_header(skb);
370 skb->protocol = htons(ETH_P_IP);
372 dev->stats.rx_packets++;
375 /* Encapsulate one IP datagram and stuff into a TTY queue. */
376 static void sl_encaps(struct slip *sl, unsigned char *icp, int len)
381 if (len > sl->mtu) { /* Sigh, shouldn't occur BUT ... */
382 printk(KERN_WARNING "%s: truncating oversized transmit packet!\n", sl->dev->name);
383 sl->dev->stats.tx_dropped++;
389 #ifdef SL_INCLUDE_CSLIP
390 if (sl->mode & SL_MODE_CSLIP)
391 len = slhc_compress(sl->slcomp, p, len, sl->cbuff, &p, 1);
393 #ifdef CONFIG_SLIP_MODE_SLIP6
394 if (sl->mode & SL_MODE_SLIP6)
395 count = slip_esc6(p, sl->xbuff, len);
398 count = slip_esc(p, sl->xbuff, len);
400 /* Order of next two lines is *very* important.
401 * When we are sending a little amount of data,
402 * the transfer may be completed inside the ops->write()
403 * routine, because it's running with interrupts enabled.
404 * In this case we *never* got WRITE_WAKEUP event,
405 * if we did not request it before write operation.
406 * 14 Oct 1994 Dmitry Gorodchanin.
408 set_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
409 actual = sl->tty->ops->write(sl->tty, sl->xbuff, count);
410 #ifdef SL_CHECK_TRANSMIT
411 netif_trans_update(sl->dev);
413 sl->xleft = count - actual;
414 sl->xhead = sl->xbuff + actual;
415 #ifdef CONFIG_SLIP_SMART
417 clear_bit(SLF_OUTWAIT, &sl->flags); /* reset outfill flag */
421 /* Write out any remaining transmit buffer. Scheduled when tty is writable */
422 static void slip_transmit(struct work_struct *work)
424 struct slip *sl = container_of(work, struct slip, tx_work);
427 spin_lock_bh(&sl->lock);
428 /* First make sure we're connected. */
429 if (!sl->tty || sl->magic != SLIP_MAGIC || !netif_running(sl->dev)) {
430 spin_unlock_bh(&sl->lock);
434 if (sl->xleft <= 0) {
435 /* Now serial buffer is almost free & we can start
436 * transmission of another packet */
437 sl->dev->stats.tx_packets++;
438 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
439 spin_unlock_bh(&sl->lock);
444 actual = sl->tty->ops->write(sl->tty, sl->xhead, sl->xleft);
447 spin_unlock_bh(&sl->lock);
451 * Called by the driver when there's room for more data.
452 * Schedule the transmit.
454 static void slip_write_wakeup(struct tty_struct *tty)
459 sl = rcu_dereference(tty->disc_data);
461 schedule_work(&sl->tx_work);
465 static void sl_tx_timeout(struct net_device *dev, unsigned int txqueue)
467 struct slip *sl = netdev_priv(dev);
469 spin_lock(&sl->lock);
471 if (netif_queue_stopped(dev)) {
472 if (!netif_running(dev) || !sl->tty)
475 /* May be we must check transmitter timeout here ?
476 * 14 Oct 1994 Dmitry Gorodchanin.
478 #ifdef SL_CHECK_TRANSMIT
479 if (time_before(jiffies, dev_trans_start(dev) + 20 * HZ)) {
480 /* 20 sec timeout not reached */
483 printk(KERN_WARNING "%s: transmit timed out, %s?\n",
485 (tty_chars_in_buffer(sl->tty) || sl->xleft) ?
486 "bad line quality" : "driver error");
488 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
493 spin_unlock(&sl->lock);
497 /* Encapsulate an IP datagram and kick it into a TTY queue. */
499 sl_xmit(struct sk_buff *skb, struct net_device *dev)
501 struct slip *sl = netdev_priv(dev);
503 spin_lock(&sl->lock);
504 if (!netif_running(dev)) {
505 spin_unlock(&sl->lock);
506 printk(KERN_WARNING "%s: xmit call when iface is down\n", dev->name);
510 if (sl->tty == NULL) {
511 spin_unlock(&sl->lock);
517 dev->stats.tx_bytes += skb->len;
518 sl_encaps(sl, skb->data, skb->len);
519 spin_unlock(&sl->lock);
526 /******************************************
527 * Routines looking at netdevice side.
528 ******************************************/
530 /* Netdevice UP -> DOWN routine */
533 sl_close(struct net_device *dev)
535 struct slip *sl = netdev_priv(dev);
537 spin_lock_bh(&sl->lock);
539 /* TTY discipline is running. */
540 clear_bit(TTY_DO_WRITE_WAKEUP, &sl->tty->flags);
541 netif_stop_queue(dev);
544 spin_unlock_bh(&sl->lock);
549 /* Netdevice DOWN -> UP routine */
551 static int sl_open(struct net_device *dev)
553 struct slip *sl = netdev_priv(dev);
558 sl->flags &= (1 << SLF_INUSE);
559 netif_start_queue(dev);
563 /* Netdevice change MTU request */
565 static int sl_change_mtu(struct net_device *dev, int new_mtu)
567 struct slip *sl = netdev_priv(dev);
569 return sl_realloc_bufs(sl, new_mtu);
572 /* Netdevice get statistics request */
575 sl_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
577 struct net_device_stats *devstats = &dev->stats;
578 #ifdef SL_INCLUDE_CSLIP
579 struct slip *sl = netdev_priv(dev);
580 struct slcompress *comp = sl->slcomp;
582 stats->rx_packets = devstats->rx_packets;
583 stats->tx_packets = devstats->tx_packets;
584 stats->rx_bytes = devstats->rx_bytes;
585 stats->tx_bytes = devstats->tx_bytes;
586 stats->rx_dropped = devstats->rx_dropped;
587 stats->tx_dropped = devstats->tx_dropped;
588 stats->tx_errors = devstats->tx_errors;
589 stats->rx_errors = devstats->rx_errors;
590 stats->rx_over_errors = devstats->rx_over_errors;
592 #ifdef SL_INCLUDE_CSLIP
594 /* Generic compressed statistics */
595 stats->rx_compressed = comp->sls_i_compressed;
596 stats->tx_compressed = comp->sls_o_compressed;
598 /* Are we really still needs this? */
599 stats->rx_fifo_errors += comp->sls_i_compressed;
600 stats->rx_dropped += comp->sls_i_tossed;
601 stats->tx_fifo_errors += comp->sls_o_compressed;
602 stats->collisions += comp->sls_o_misses;
607 /* Netdevice register callback */
609 static int sl_init(struct net_device *dev)
611 struct slip *sl = netdev_priv(dev);
614 * Finish setting up the DEVICE info.
618 dev->type = ARPHRD_SLIP + sl->mode;
619 #ifdef SL_CHECK_TRANSMIT
620 dev->watchdog_timeo = 20*HZ;
626 static void sl_uninit(struct net_device *dev)
628 struct slip *sl = netdev_priv(dev);
633 /* Hook the destructor so we can free slip devices at the right point in time */
634 static void sl_free_netdev(struct net_device *dev)
636 int i = dev->base_addr;
641 static const struct net_device_ops sl_netdev_ops = {
643 .ndo_uninit = sl_uninit,
645 .ndo_stop = sl_close,
646 .ndo_start_xmit = sl_xmit,
647 .ndo_get_stats64 = sl_get_stats64,
648 .ndo_change_mtu = sl_change_mtu,
649 .ndo_tx_timeout = sl_tx_timeout,
650 #ifdef CONFIG_SLIP_SMART
651 .ndo_siocdevprivate = sl_siocdevprivate,
656 static void sl_setup(struct net_device *dev)
658 dev->netdev_ops = &sl_netdev_ops;
659 dev->needs_free_netdev = true;
660 dev->priv_destructor = sl_free_netdev;
662 dev->hard_header_len = 0;
664 dev->tx_queue_len = 10;
666 /* MTU range: 68 - 65534 */
668 dev->max_mtu = 65534;
670 /* New-style flags. */
671 dev->flags = IFF_NOARP|IFF_POINTOPOINT|IFF_MULTICAST;
674 /******************************************
675 Routines looking at TTY side.
676 ******************************************/
680 * Handle the 'receiver data ready' interrupt.
681 * This function is called by the 'tty_io' module in the kernel when
682 * a block of SLIP data has been received, which can now be decapsulated
683 * and sent on to some IP layer for further processing. This will not
684 * be re-entered while running but other ldisc functions may be called
688 static void slip_receive_buf(struct tty_struct *tty, const unsigned char *cp,
689 const char *fp, int count)
691 struct slip *sl = tty->disc_data;
693 if (!sl || sl->magic != SLIP_MAGIC || !netif_running(sl->dev))
696 /* Read the characters out of the buffer */
699 if (!test_and_set_bit(SLF_ERROR, &sl->flags))
700 sl->dev->stats.rx_errors++;
704 #ifdef CONFIG_SLIP_MODE_SLIP6
705 if (sl->mode & SL_MODE_SLIP6)
706 slip_unesc6(sl, *cp++);
709 slip_unesc(sl, *cp++);
713 /************************************
714 * slip_open helper routines.
715 ************************************/
717 /* Collect hanged up channels */
718 static void sl_sync(void)
721 struct net_device *dev;
724 for (i = 0; i < slip_maxdev; i++) {
729 sl = netdev_priv(dev);
730 if (sl->tty || sl->leased)
732 if (dev->flags & IFF_UP)
738 /* Find a free SLIP channel, and link in this `tty' line. */
739 static struct slip *sl_alloc(void)
743 struct net_device *dev = NULL;
746 for (i = 0; i < slip_maxdev; i++) {
751 /* Sorry, too many, all slots in use */
752 if (i >= slip_maxdev)
755 sprintf(name, "sl%d", i);
756 dev = alloc_netdev(sizeof(*sl), name, NET_NAME_UNKNOWN, sl_setup);
761 sl = netdev_priv(dev);
763 /* Initialize channel control data */
764 sl->magic = SLIP_MAGIC;
766 spin_lock_init(&sl->lock);
767 INIT_WORK(&sl->tx_work, slip_transmit);
768 sl->mode = SL_MODE_DEFAULT;
769 #ifdef CONFIG_SLIP_SMART
770 /* initialize timer_list struct */
771 timer_setup(&sl->keepalive_timer, sl_keepalive, 0);
772 timer_setup(&sl->outfill_timer, sl_outfill, 0);
779 * Open the high-level part of the SLIP channel.
780 * This function is called by the TTY module when the
781 * SLIP line discipline is called for. Because we are
782 * sure the tty line exists, we only have to link it to
783 * a free SLIP channel...
785 * Called in process context serialized from other ldisc calls.
788 static int slip_open(struct tty_struct *tty)
793 if (!capable(CAP_NET_ADMIN))
796 if (tty->ops->write == NULL)
799 /* RTnetlink lock is misused here to serialize concurrent
800 opens of slip channels. There are better ways, but it is
805 /* Collect hanged up channels. */
811 /* First make sure we're not already connected. */
812 if (sl && sl->magic == SLIP_MAGIC)
815 /* OK. Find a free SLIP channel to use. */
823 sl->pid = current->pid;
825 if (!test_bit(SLF_INUSE, &sl->flags)) {
826 /* Perform the low-level SLIP initialization. */
827 err = sl_alloc_bufs(sl, SL_MTU);
831 set_bit(SLF_INUSE, &sl->flags);
833 err = register_netdevice(sl->dev);
838 #ifdef CONFIG_SLIP_SMART
840 sl->keepalive_timer.expires = jiffies + sl->keepalive * HZ;
841 add_timer(&sl->keepalive_timer);
844 sl->outfill_timer.expires = jiffies + sl->outfill * HZ;
845 add_timer(&sl->outfill_timer);
849 /* Done. We have linked the TTY line to a channel. */
851 tty->receive_room = 65536; /* We don't flow control */
853 /* TTY layer expects 0 on success */
861 tty->disc_data = NULL;
862 clear_bit(SLF_INUSE, &sl->flags);
863 sl_free_netdev(sl->dev);
864 /* do not call free_netdev before rtnl_unlock */
866 free_netdev(sl->dev);
872 /* Count references from TTY module */
877 * Close down a SLIP channel.
878 * This means flushing out any pending queues, and then returning. This
879 * call is serialized against other ldisc functions.
881 * We also use this method fo a hangup event
884 static void slip_close(struct tty_struct *tty)
886 struct slip *sl = tty->disc_data;
888 /* First make sure we're connected. */
889 if (!sl || sl->magic != SLIP_MAGIC || sl->tty != tty)
892 spin_lock_bh(&sl->lock);
893 rcu_assign_pointer(tty->disc_data, NULL);
895 spin_unlock_bh(&sl->lock);
898 flush_work(&sl->tx_work);
900 /* VSV = very important to remove timers */
901 #ifdef CONFIG_SLIP_SMART
902 del_timer_sync(&sl->keepalive_timer);
903 del_timer_sync(&sl->outfill_timer);
905 /* Flush network side */
906 unregister_netdev(sl->dev);
907 /* This will complete via sl_free_netdev */
910 static void slip_hangup(struct tty_struct *tty)
914 /************************************************************************
915 * STANDARD SLIP ENCAPSULATION *
916 ************************************************************************/
918 static int slip_esc(unsigned char *s, unsigned char *d, int len)
920 unsigned char *ptr = d;
924 * Send an initial END character to flush out any
925 * data that may have accumulated in the receiver
932 * For each byte in the packet, send the appropriate
933 * character sequence, according to the SLIP protocol.
955 static void slip_unesc(struct slip *sl, unsigned char s)
960 #ifdef CONFIG_SLIP_SMART
961 /* drop keeptest bit = VSV */
962 if (test_bit(SLF_KEEPTEST, &sl->flags))
963 clear_bit(SLF_KEEPTEST, &sl->flags);
966 if (!test_and_clear_bit(SLF_ERROR, &sl->flags) &&
969 clear_bit(SLF_ESCAPE, &sl->flags);
974 set_bit(SLF_ESCAPE, &sl->flags);
977 if (test_and_clear_bit(SLF_ESCAPE, &sl->flags))
981 if (test_and_clear_bit(SLF_ESCAPE, &sl->flags))
985 if (!test_bit(SLF_ERROR, &sl->flags)) {
986 if (sl->rcount < sl->buffsize) {
987 sl->rbuff[sl->rcount++] = s;
990 sl->dev->stats.rx_over_errors++;
991 set_bit(SLF_ERROR, &sl->flags);
996 #ifdef CONFIG_SLIP_MODE_SLIP6
997 /************************************************************************
998 * 6 BIT SLIP ENCAPSULATION *
999 ************************************************************************/
1001 static int slip_esc6(unsigned char *s, unsigned char *d, int len)
1003 unsigned char *ptr = d;
1006 unsigned short v = 0;
1010 * Send an initial END character to flush out any
1011 * data that may have accumulated in the receiver
1012 * due to line noise.
1018 * Encode the packet into printable ascii characters
1021 for (i = 0; i < len; ++i) {
1022 v = (v << 8) | s[i];
1026 c = 0x30 + ((v >> bits) & 0x3F);
1031 c = 0x30 + ((v << (6 - bits)) & 0x3F);
1038 static void slip_unesc6(struct slip *sl, unsigned char s)
1043 #ifdef CONFIG_SLIP_SMART
1044 /* drop keeptest bit = VSV */
1045 if (test_bit(SLF_KEEPTEST, &sl->flags))
1046 clear_bit(SLF_KEEPTEST, &sl->flags);
1049 if (!test_and_clear_bit(SLF_ERROR, &sl->flags) &&
1055 } else if (s >= 0x30 && s < 0x70) {
1056 sl->xdata = (sl->xdata << 6) | ((s - 0x30) & 0x3F);
1058 if (sl->xbits >= 8) {
1060 c = (unsigned char)(sl->xdata >> sl->xbits);
1061 if (!test_bit(SLF_ERROR, &sl->flags)) {
1062 if (sl->rcount < sl->buffsize) {
1063 sl->rbuff[sl->rcount++] = c;
1066 sl->dev->stats.rx_over_errors++;
1067 set_bit(SLF_ERROR, &sl->flags);
1072 #endif /* CONFIG_SLIP_MODE_SLIP6 */
1074 /* Perform I/O control on an active SLIP channel. */
1075 static int slip_ioctl(struct tty_struct *tty, unsigned int cmd,
1078 struct slip *sl = tty->disc_data;
1080 int __user *p = (int __user *)arg;
1082 /* First make sure we're connected. */
1083 if (!sl || sl->magic != SLIP_MAGIC)
1088 tmp = strlen(sl->dev->name) + 1;
1089 if (copy_to_user((void __user *)arg, sl->dev->name, tmp))
1094 if (put_user(sl->mode, p))
1099 if (get_user(tmp, p))
1101 #ifndef SL_INCLUDE_CSLIP
1102 if (tmp & (SL_MODE_CSLIP|SL_MODE_ADAPTIVE))
1105 if ((tmp & (SL_MODE_ADAPTIVE | SL_MODE_CSLIP)) ==
1106 (SL_MODE_ADAPTIVE | SL_MODE_CSLIP))
1107 /* return -EINVAL; */
1108 tmp &= ~SL_MODE_ADAPTIVE;
1110 #ifndef CONFIG_SLIP_MODE_SLIP6
1111 if (tmp & SL_MODE_SLIP6)
1115 sl->dev->type = ARPHRD_SLIP + sl->mode;
1121 #ifdef CONFIG_SLIP_SMART
1122 /* VSV changes start here */
1123 case SIOCSKEEPALIVE:
1124 if (get_user(tmp, p))
1126 if (tmp > 255) /* max for unchar */
1129 spin_lock_bh(&sl->lock);
1131 spin_unlock_bh(&sl->lock);
1134 sl->keepalive = (u8)tmp;
1135 if (sl->keepalive != 0) {
1136 mod_timer(&sl->keepalive_timer,
1137 jiffies + sl->keepalive * HZ);
1138 set_bit(SLF_KEEPTEST, &sl->flags);
1140 del_timer(&sl->keepalive_timer);
1141 spin_unlock_bh(&sl->lock);
1144 case SIOCGKEEPALIVE:
1145 if (put_user(sl->keepalive, p))
1150 if (get_user(tmp, p))
1152 if (tmp > 255) /* max for unchar */
1154 spin_lock_bh(&sl->lock);
1156 spin_unlock_bh(&sl->lock);
1159 sl->outfill = (u8)tmp;
1160 if (sl->outfill != 0) {
1161 mod_timer(&sl->outfill_timer,
1162 jiffies + sl->outfill * HZ);
1163 set_bit(SLF_OUTWAIT, &sl->flags);
1165 del_timer(&sl->outfill_timer);
1166 spin_unlock_bh(&sl->lock);
1170 if (put_user(sl->outfill, p))
1173 /* VSV changes end */
1176 return tty_mode_ioctl(tty, cmd, arg);
1180 /* VSV changes start here */
1181 #ifdef CONFIG_SLIP_SMART
1182 /* function sl_siocdevprivate called from net/core/dev.c
1183 to allow get/set outfill/keepalive parameter
1186 static int sl_siocdevprivate(struct net_device *dev, struct ifreq *rq,
1187 void __user *data, int cmd)
1189 struct slip *sl = netdev_priv(dev);
1190 unsigned long *p = (unsigned long *)&rq->ifr_ifru;
1192 if (sl == NULL) /* Allocation failed ?? */
1195 if (in_compat_syscall())
1198 spin_lock_bh(&sl->lock);
1201 spin_unlock_bh(&sl->lock);
1206 case SIOCSKEEPALIVE:
1207 /* max for unchar */
1208 if ((unsigned)*p > 255) {
1209 spin_unlock_bh(&sl->lock);
1212 sl->keepalive = (u8)*p;
1213 if (sl->keepalive != 0) {
1214 sl->keepalive_timer.expires =
1215 jiffies + sl->keepalive * HZ;
1216 mod_timer(&sl->keepalive_timer,
1217 jiffies + sl->keepalive * HZ);
1218 set_bit(SLF_KEEPTEST, &sl->flags);
1220 del_timer(&sl->keepalive_timer);
1223 case SIOCGKEEPALIVE:
1228 if ((unsigned)*p > 255) { /* max for unchar */
1229 spin_unlock_bh(&sl->lock);
1232 sl->outfill = (u8)*p;
1233 if (sl->outfill != 0) {
1234 mod_timer(&sl->outfill_timer,
1235 jiffies + sl->outfill * HZ);
1236 set_bit(SLF_OUTWAIT, &sl->flags);
1238 del_timer(&sl->outfill_timer);
1246 /* Resolve race condition, when ioctl'ing hanged up
1247 and opened by another process device.
1249 if (sl->tty != current->signal->tty &&
1250 sl->pid != current->pid) {
1251 spin_unlock_bh(&sl->lock);
1262 spin_unlock_bh(&sl->lock);
1266 /* VSV changes end */
1268 static struct tty_ldisc_ops sl_ldisc = {
1269 .owner = THIS_MODULE,
1273 .close = slip_close,
1274 .hangup = slip_hangup,
1275 .ioctl = slip_ioctl,
1276 .receive_buf = slip_receive_buf,
1277 .write_wakeup = slip_write_wakeup,
1280 static int __init slip_init(void)
1284 if (slip_maxdev < 4)
1285 slip_maxdev = 4; /* Sanity */
1287 printk(KERN_INFO "SLIP: version %s (dynamic channels, max=%d)"
1288 #ifdef CONFIG_SLIP_MODE_SLIP6
1289 " (6 bit encapsulation enabled)"
1292 SLIP_VERSION, slip_maxdev);
1293 #if defined(SL_INCLUDE_CSLIP)
1294 printk(KERN_INFO "CSLIP: code copyright 1989 Regents of the University of California.\n");
1296 #ifdef CONFIG_SLIP_SMART
1297 printk(KERN_INFO "SLIP linefill/keepalive option.\n");
1300 slip_devs = kcalloc(slip_maxdev, sizeof(struct net_device *),
1305 /* Fill in our line protocol discipline, and register it */
1306 status = tty_register_ldisc(&sl_ldisc);
1308 printk(KERN_ERR "SLIP: can't register line discipline (err = %d)\n", status);
1314 static void __exit slip_exit(void)
1317 struct net_device *dev;
1319 unsigned long timeout = jiffies + HZ;
1322 if (slip_devs == NULL)
1325 /* First of all: check for active disciplines and hangup them.
1329 msleep_interruptible(100);
1332 for (i = 0; i < slip_maxdev; i++) {
1336 sl = netdev_priv(dev);
1337 spin_lock_bh(&sl->lock);
1340 tty_hangup(sl->tty);
1342 spin_unlock_bh(&sl->lock);
1344 } while (busy && time_before(jiffies, timeout));
1346 /* FIXME: hangup is async so we should wait when doing this second
1349 for (i = 0; i < slip_maxdev; i++) {
1353 slip_devs[i] = NULL;
1355 sl = netdev_priv(dev);
1357 printk(KERN_ERR "%s: tty discipline still running\n",
1361 unregister_netdev(dev);
1367 tty_unregister_ldisc(&sl_ldisc);
1370 module_init(slip_init);
1371 module_exit(slip_exit);
1373 #ifdef CONFIG_SLIP_SMART
1375 * This is start of the code for multislip style line checking
1376 * added by Stanislav Voronyi. All changes before marked VSV
1379 static void sl_outfill(struct timer_list *t)
1381 struct slip *sl = from_timer(sl, t, outfill_timer);
1383 spin_lock(&sl->lock);
1385 if (sl->tty == NULL)
1389 if (test_bit(SLF_OUTWAIT, &sl->flags)) {
1390 /* no packets were transmitted, do outfill */
1391 #ifdef CONFIG_SLIP_MODE_SLIP6
1392 unsigned char s = (sl->mode & SL_MODE_SLIP6)?0x70:END;
1394 unsigned char s = END;
1396 /* put END into tty queue. Is it right ??? */
1397 if (!netif_queue_stopped(sl->dev)) {
1398 /* if device busy no outfill */
1399 sl->tty->ops->write(sl->tty, &s, 1);
1402 set_bit(SLF_OUTWAIT, &sl->flags);
1404 mod_timer(&sl->outfill_timer, jiffies+sl->outfill*HZ);
1407 spin_unlock(&sl->lock);
1410 static void sl_keepalive(struct timer_list *t)
1412 struct slip *sl = from_timer(sl, t, keepalive_timer);
1414 spin_lock(&sl->lock);
1416 if (sl->tty == NULL)
1419 if (sl->keepalive) {
1420 if (test_bit(SLF_KEEPTEST, &sl->flags)) {
1421 /* keepalive still high :(, we must hangup */
1423 /* outfill timer must be deleted too */
1424 (void)del_timer(&sl->outfill_timer);
1425 printk(KERN_DEBUG "%s: no packets received during keepalive timeout, hangup.\n", sl->dev->name);
1426 /* this must hangup tty & close slip */
1427 tty_hangup(sl->tty);
1428 /* I think we need not something else */
1431 set_bit(SLF_KEEPTEST, &sl->flags);
1433 mod_timer(&sl->keepalive_timer, jiffies+sl->keepalive*HZ);
1436 spin_unlock(&sl->lock);
1440 MODULE_LICENSE("GPL");
1441 MODULE_ALIAS_LDISC(N_SLIP);