xref: /linux/net/rose/af_rose.c (revision 8f7aa3d3c7323f4ca2768a9e74ebbe359c4f8f88)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *
4  * Copyright (C) Jonathan Naylor G4KLX (g4klx@g4klx.demon.co.uk)
5  * Copyright (C) Alan Cox GW4PTS (alan@lxorguk.ukuu.org.uk)
6  * Copyright (C) Terry Dawson VK2KTJ (terry@animats.net)
7  * Copyright (C) Tomi Manninen OH2BNS (oh2bns@sral.fi)
8  */
9 
10 #include <linux/capability.h>
11 #include <linux/module.h>
12 #include <linux/moduleparam.h>
13 #include <linux/init.h>
14 #include <linux/errno.h>
15 #include <linux/types.h>
16 #include <linux/socket.h>
17 #include <linux/in.h>
18 #include <linux/slab.h>
19 #include <linux/kernel.h>
20 #include <linux/sched/signal.h>
21 #include <linux/spinlock.h>
22 #include <linux/timer.h>
23 #include <linux/string.h>
24 #include <linux/sockios.h>
25 #include <linux/net.h>
26 #include <linux/stat.h>
27 #include <net/net_namespace.h>
28 #include <net/ax25.h>
29 #include <linux/inet.h>
30 #include <linux/netdevice.h>
31 #include <linux/if_arp.h>
32 #include <linux/skbuff.h>
33 #include <net/sock.h>
34 #include <linux/uaccess.h>
35 #include <linux/fcntl.h>
36 #include <linux/termios.h>
37 #include <linux/mm.h>
38 #include <linux/interrupt.h>
39 #include <linux/notifier.h>
40 #include <net/rose.h>
41 #include <linux/proc_fs.h>
42 #include <linux/seq_file.h>
43 #include <net/tcp_states.h>
44 #include <net/ip.h>
45 #include <net/arp.h>
46 
47 static int rose_ndevs = 10;
48 
49 int sysctl_rose_restart_request_timeout = ROSE_DEFAULT_T0;
50 int sysctl_rose_call_request_timeout    = ROSE_DEFAULT_T1;
51 int sysctl_rose_reset_request_timeout   = ROSE_DEFAULT_T2;
52 int sysctl_rose_clear_request_timeout   = ROSE_DEFAULT_T3;
53 int sysctl_rose_no_activity_timeout     = ROSE_DEFAULT_IDLE;
54 int sysctl_rose_ack_hold_back_timeout   = ROSE_DEFAULT_HB;
55 int sysctl_rose_routing_control         = ROSE_DEFAULT_ROUTING;
56 int sysctl_rose_link_fail_timeout       = ROSE_DEFAULT_FAIL_TIMEOUT;
57 int sysctl_rose_maximum_vcs             = ROSE_DEFAULT_MAXVC;
58 int sysctl_rose_window_size             = ROSE_DEFAULT_WINDOW_SIZE;
59 
60 static HLIST_HEAD(rose_list);
61 static DEFINE_SPINLOCK(rose_list_lock);
62 
63 static const struct proto_ops rose_proto_ops;
64 
65 ax25_address rose_callsign;
66 
67 /*
68  * ROSE network devices are virtual network devices encapsulating ROSE
69  * frames into AX.25 which will be sent through an AX.25 device, so form a
70  * special "super class" of normal net devices; split their locks off into a
71  * separate class since they always nest.
72  */
73 static struct lock_class_key rose_netdev_xmit_lock_key;
74 static struct lock_class_key rose_netdev_addr_lock_key;
75 
76 static void rose_set_lockdep_one(struct net_device *dev,
77 				 struct netdev_queue *txq,
78 				 void *_unused)
79 {
80 	lockdep_set_class(&txq->_xmit_lock, &rose_netdev_xmit_lock_key);
81 }
82 
83 static void rose_set_lockdep_key(struct net_device *dev)
84 {
85 	lockdep_set_class(&dev->addr_list_lock, &rose_netdev_addr_lock_key);
86 	netdev_for_each_tx_queue(dev, rose_set_lockdep_one, NULL);
87 }
88 
89 /*
90  *	Convert a ROSE address into text.
91  */
92 char *rose2asc(char *buf, const rose_address *addr)
93 {
94 	if (addr->rose_addr[0] == 0x00 && addr->rose_addr[1] == 0x00 &&
95 	    addr->rose_addr[2] == 0x00 && addr->rose_addr[3] == 0x00 &&
96 	    addr->rose_addr[4] == 0x00) {
97 		strcpy(buf, "*");
98 	} else {
99 		sprintf(buf, "%02X%02X%02X%02X%02X", addr->rose_addr[0] & 0xFF,
100 						addr->rose_addr[1] & 0xFF,
101 						addr->rose_addr[2] & 0xFF,
102 						addr->rose_addr[3] & 0xFF,
103 						addr->rose_addr[4] & 0xFF);
104 	}
105 
106 	return buf;
107 }
108 
109 /*
110  *	Compare two ROSE addresses, 0 == equal.
111  */
112 int rosecmp(const rose_address *addr1, const rose_address *addr2)
113 {
114 	int i;
115 
116 	for (i = 0; i < 5; i++)
117 		if (addr1->rose_addr[i] != addr2->rose_addr[i])
118 			return 1;
119 
120 	return 0;
121 }
122 
123 /*
124  *	Compare two ROSE addresses for only mask digits, 0 == equal.
125  */
126 int rosecmpm(const rose_address *addr1, const rose_address *addr2,
127 	     unsigned short mask)
128 {
129 	unsigned int i, j;
130 
131 	if (mask > 10)
132 		return 1;
133 
134 	for (i = 0; i < mask; i++) {
135 		j = i / 2;
136 
137 		if ((i % 2) != 0) {
138 			if ((addr1->rose_addr[j] & 0x0F) != (addr2->rose_addr[j] & 0x0F))
139 				return 1;
140 		} else {
141 			if ((addr1->rose_addr[j] & 0xF0) != (addr2->rose_addr[j] & 0xF0))
142 				return 1;
143 		}
144 	}
145 
146 	return 0;
147 }
148 
149 /*
150  *	Socket removal during an interrupt is now safe.
151  */
152 static void rose_remove_socket(struct sock *sk)
153 {
154 	spin_lock_bh(&rose_list_lock);
155 	sk_del_node_init(sk);
156 	spin_unlock_bh(&rose_list_lock);
157 }
158 
159 /*
160  *	Kill all bound sockets on a broken link layer connection to a
161  *	particular neighbour.
162  */
163 void rose_kill_by_neigh(struct rose_neigh *neigh)
164 {
165 	struct sock *s;
166 
167 	spin_lock_bh(&rose_list_lock);
168 	sk_for_each(s, &rose_list) {
169 		struct rose_sock *rose = rose_sk(s);
170 
171 		if (rose->neighbour == neigh) {
172 			rose_disconnect(s, ENETUNREACH, ROSE_OUT_OF_ORDER, 0);
173 			rose_neigh_put(rose->neighbour);
174 			rose->neighbour = NULL;
175 		}
176 	}
177 	spin_unlock_bh(&rose_list_lock);
178 }
179 
180 /*
181  *	Kill all bound sockets on a dropped device.
182  */
183 static void rose_kill_by_device(struct net_device *dev)
184 {
185 	struct sock *sk, *array[16];
186 	struct rose_sock *rose;
187 	bool rescan;
188 	int i, cnt;
189 
190 start:
191 	rescan = false;
192 	cnt = 0;
193 	spin_lock_bh(&rose_list_lock);
194 	sk_for_each(sk, &rose_list) {
195 		rose = rose_sk(sk);
196 		if (rose->device == dev) {
197 			if (cnt == ARRAY_SIZE(array)) {
198 				rescan = true;
199 				break;
200 			}
201 			sock_hold(sk);
202 			array[cnt++] = sk;
203 		}
204 	}
205 	spin_unlock_bh(&rose_list_lock);
206 
207 	for (i = 0; i < cnt; i++) {
208 		sk = array[cnt];
209 		rose = rose_sk(sk);
210 		lock_sock(sk);
211 		spin_lock_bh(&rose_list_lock);
212 		if (rose->device == dev) {
213 			rose_disconnect(sk, ENETUNREACH, ROSE_OUT_OF_ORDER, 0);
214 			if (rose->neighbour)
215 				rose_neigh_put(rose->neighbour);
216 			netdev_put(rose->device, &rose->dev_tracker);
217 			rose->device = NULL;
218 		}
219 		spin_unlock_bh(&rose_list_lock);
220 		release_sock(sk);
221 		sock_put(sk);
222 		cond_resched();
223 	}
224 	if (rescan)
225 		goto start;
226 }
227 
228 /*
229  *	Handle device status changes.
230  */
231 static int rose_device_event(struct notifier_block *this,
232 			     unsigned long event, void *ptr)
233 {
234 	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
235 
236 	if (!net_eq(dev_net(dev), &init_net))
237 		return NOTIFY_DONE;
238 
239 	if (event != NETDEV_DOWN)
240 		return NOTIFY_DONE;
241 
242 	switch (dev->type) {
243 	case ARPHRD_ROSE:
244 		rose_kill_by_device(dev);
245 		break;
246 	case ARPHRD_AX25:
247 		rose_link_device_down(dev);
248 		rose_rt_device_down(dev);
249 		break;
250 	}
251 
252 	return NOTIFY_DONE;
253 }
254 
255 /*
256  *	Add a socket to the bound sockets list.
257  */
258 static void rose_insert_socket(struct sock *sk)
259 {
260 
261 	spin_lock_bh(&rose_list_lock);
262 	sk_add_node(sk, &rose_list);
263 	spin_unlock_bh(&rose_list_lock);
264 }
265 
266 /*
267  *	Find a socket that wants to accept the Call Request we just
268  *	received.
269  */
270 static struct sock *rose_find_listener(rose_address *addr, ax25_address *call)
271 {
272 	struct sock *s;
273 
274 	spin_lock_bh(&rose_list_lock);
275 	sk_for_each(s, &rose_list) {
276 		struct rose_sock *rose = rose_sk(s);
277 
278 		if (!rosecmp(&rose->source_addr, addr) &&
279 		    !ax25cmp(&rose->source_call, call) &&
280 		    !rose->source_ndigis && s->sk_state == TCP_LISTEN)
281 			goto found;
282 	}
283 
284 	sk_for_each(s, &rose_list) {
285 		struct rose_sock *rose = rose_sk(s);
286 
287 		if (!rosecmp(&rose->source_addr, addr) &&
288 		    !ax25cmp(&rose->source_call, &null_ax25_address) &&
289 		    s->sk_state == TCP_LISTEN)
290 			goto found;
291 	}
292 	s = NULL;
293 found:
294 	spin_unlock_bh(&rose_list_lock);
295 	return s;
296 }
297 
298 /*
299  *	Find a connected ROSE socket given my LCI and device.
300  */
301 struct sock *rose_find_socket(unsigned int lci, struct rose_neigh *neigh)
302 {
303 	struct sock *s;
304 
305 	spin_lock_bh(&rose_list_lock);
306 	sk_for_each(s, &rose_list) {
307 		struct rose_sock *rose = rose_sk(s);
308 
309 		if (rose->lci == lci && rose->neighbour == neigh)
310 			goto found;
311 	}
312 	s = NULL;
313 found:
314 	spin_unlock_bh(&rose_list_lock);
315 	return s;
316 }
317 
318 /*
319  *	Find a unique LCI for a given device.
320  */
321 unsigned int rose_new_lci(struct rose_neigh *neigh)
322 {
323 	int lci;
324 
325 	if (neigh->dce_mode) {
326 		for (lci = 1; lci <= sysctl_rose_maximum_vcs; lci++)
327 			if (rose_find_socket(lci, neigh) == NULL && rose_route_free_lci(lci, neigh) == NULL)
328 				return lci;
329 	} else {
330 		for (lci = sysctl_rose_maximum_vcs; lci > 0; lci--)
331 			if (rose_find_socket(lci, neigh) == NULL && rose_route_free_lci(lci, neigh) == NULL)
332 				return lci;
333 	}
334 
335 	return 0;
336 }
337 
338 /*
339  *	Deferred destroy.
340  */
341 void rose_destroy_socket(struct sock *);
342 
343 /*
344  *	Handler for deferred kills.
345  */
346 static void rose_destroy_timer(struct timer_list *t)
347 {
348 	struct sock *sk = timer_container_of(sk, t, sk_timer);
349 
350 	rose_destroy_socket(sk);
351 }
352 
353 /*
354  *	This is called from user mode and the timers. Thus it protects itself
355  *	against interrupt users but doesn't worry about being called during
356  *	work.  Once it is removed from the queue no interrupt or bottom half
357  *	will touch it and we are (fairly 8-) ) safe.
358  */
359 void rose_destroy_socket(struct sock *sk)
360 {
361 	struct sk_buff *skb;
362 
363 	rose_remove_socket(sk);
364 	rose_stop_heartbeat(sk);
365 	rose_stop_idletimer(sk);
366 	rose_stop_timer(sk);
367 
368 	rose_clear_queues(sk);		/* Flush the queues */
369 
370 	while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) {
371 		if (skb->sk != sk) {	/* A pending connection */
372 			/* Queue the unaccepted socket for death */
373 			sock_set_flag(skb->sk, SOCK_DEAD);
374 			rose_start_heartbeat(skb->sk);
375 			rose_sk(skb->sk)->state = ROSE_STATE_0;
376 		}
377 
378 		kfree_skb(skb);
379 	}
380 
381 	if (sk_has_allocations(sk)) {
382 		/* Defer: outstanding buffers */
383 		timer_setup(&sk->sk_timer, rose_destroy_timer, 0);
384 		sk->sk_timer.expires  = jiffies + 10 * HZ;
385 		add_timer(&sk->sk_timer);
386 	} else
387 		sock_put(sk);
388 }
389 
390 /*
391  *	Handling for system calls applied via the various interfaces to a
392  *	ROSE socket object.
393  */
394 
395 static int rose_setsockopt(struct socket *sock, int level, int optname,
396 		sockptr_t optval, unsigned int optlen)
397 {
398 	struct sock *sk = sock->sk;
399 	struct rose_sock *rose = rose_sk(sk);
400 	unsigned int opt;
401 
402 	if (level != SOL_ROSE)
403 		return -ENOPROTOOPT;
404 
405 	if (optlen < sizeof(unsigned int))
406 		return -EINVAL;
407 
408 	if (copy_from_sockptr(&opt, optval, sizeof(unsigned int)))
409 		return -EFAULT;
410 
411 	switch (optname) {
412 	case ROSE_DEFER:
413 		rose->defer = opt ? 1 : 0;
414 		return 0;
415 
416 	case ROSE_T1:
417 		if (opt < 1 || opt > UINT_MAX / HZ)
418 			return -EINVAL;
419 		rose->t1 = opt * HZ;
420 		return 0;
421 
422 	case ROSE_T2:
423 		if (opt < 1 || opt > UINT_MAX / HZ)
424 			return -EINVAL;
425 		rose->t2 = opt * HZ;
426 		return 0;
427 
428 	case ROSE_T3:
429 		if (opt < 1 || opt > UINT_MAX / HZ)
430 			return -EINVAL;
431 		rose->t3 = opt * HZ;
432 		return 0;
433 
434 	case ROSE_HOLDBACK:
435 		if (opt < 1 || opt > UINT_MAX / HZ)
436 			return -EINVAL;
437 		rose->hb = opt * HZ;
438 		return 0;
439 
440 	case ROSE_IDLE:
441 		if (opt > UINT_MAX / (60 * HZ))
442 			return -EINVAL;
443 		rose->idle = opt * 60 * HZ;
444 		return 0;
445 
446 	case ROSE_QBITINCL:
447 		rose->qbitincl = opt ? 1 : 0;
448 		return 0;
449 
450 	default:
451 		return -ENOPROTOOPT;
452 	}
453 }
454 
455 static int rose_getsockopt(struct socket *sock, int level, int optname,
456 	char __user *optval, int __user *optlen)
457 {
458 	struct sock *sk = sock->sk;
459 	struct rose_sock *rose = rose_sk(sk);
460 	int val = 0;
461 	int len;
462 
463 	if (level != SOL_ROSE)
464 		return -ENOPROTOOPT;
465 
466 	if (get_user(len, optlen))
467 		return -EFAULT;
468 
469 	if (len < 0)
470 		return -EINVAL;
471 
472 	switch (optname) {
473 	case ROSE_DEFER:
474 		val = rose->defer;
475 		break;
476 
477 	case ROSE_T1:
478 		val = rose->t1 / HZ;
479 		break;
480 
481 	case ROSE_T2:
482 		val = rose->t2 / HZ;
483 		break;
484 
485 	case ROSE_T3:
486 		val = rose->t3 / HZ;
487 		break;
488 
489 	case ROSE_HOLDBACK:
490 		val = rose->hb / HZ;
491 		break;
492 
493 	case ROSE_IDLE:
494 		val = rose->idle / (60 * HZ);
495 		break;
496 
497 	case ROSE_QBITINCL:
498 		val = rose->qbitincl;
499 		break;
500 
501 	default:
502 		return -ENOPROTOOPT;
503 	}
504 
505 	len = min_t(unsigned int, len, sizeof(int));
506 
507 	if (put_user(len, optlen))
508 		return -EFAULT;
509 
510 	return copy_to_user(optval, &val, len) ? -EFAULT : 0;
511 }
512 
513 static int rose_listen(struct socket *sock, int backlog)
514 {
515 	struct sock *sk = sock->sk;
516 
517 	lock_sock(sk);
518 	if (sock->state != SS_UNCONNECTED) {
519 		release_sock(sk);
520 		return -EINVAL;
521 	}
522 
523 	if (sk->sk_state != TCP_LISTEN) {
524 		struct rose_sock *rose = rose_sk(sk);
525 
526 		rose->dest_ndigis = 0;
527 		memset(&rose->dest_addr, 0, ROSE_ADDR_LEN);
528 		memset(&rose->dest_call, 0, AX25_ADDR_LEN);
529 		memset(rose->dest_digis, 0, AX25_ADDR_LEN * ROSE_MAX_DIGIS);
530 		sk->sk_max_ack_backlog = backlog;
531 		sk->sk_state           = TCP_LISTEN;
532 		release_sock(sk);
533 		return 0;
534 	}
535 	release_sock(sk);
536 
537 	return -EOPNOTSUPP;
538 }
539 
540 static struct proto rose_proto = {
541 	.name	  = "ROSE",
542 	.owner	  = THIS_MODULE,
543 	.obj_size = sizeof(struct rose_sock),
544 };
545 
546 static int rose_create(struct net *net, struct socket *sock, int protocol,
547 		       int kern)
548 {
549 	struct sock *sk;
550 	struct rose_sock *rose;
551 
552 	if (!net_eq(net, &init_net))
553 		return -EAFNOSUPPORT;
554 
555 	if (sock->type != SOCK_SEQPACKET || protocol != 0)
556 		return -ESOCKTNOSUPPORT;
557 
558 	sk = sk_alloc(net, PF_ROSE, GFP_ATOMIC, &rose_proto, kern);
559 	if (sk == NULL)
560 		return -ENOMEM;
561 
562 	rose = rose_sk(sk);
563 
564 	sock_init_data(sock, sk);
565 
566 	skb_queue_head_init(&rose->ack_queue);
567 #ifdef M_BIT
568 	skb_queue_head_init(&rose->frag_queue);
569 	rose->fraglen    = 0;
570 #endif
571 
572 	sock->ops    = &rose_proto_ops;
573 	sk->sk_protocol = protocol;
574 
575 	timer_setup(&rose->timer, NULL, 0);
576 	timer_setup(&rose->idletimer, NULL, 0);
577 
578 	rose->t1   = msecs_to_jiffies(sysctl_rose_call_request_timeout);
579 	rose->t2   = msecs_to_jiffies(sysctl_rose_reset_request_timeout);
580 	rose->t3   = msecs_to_jiffies(sysctl_rose_clear_request_timeout);
581 	rose->hb   = msecs_to_jiffies(sysctl_rose_ack_hold_back_timeout);
582 	rose->idle = msecs_to_jiffies(sysctl_rose_no_activity_timeout);
583 
584 	rose->state = ROSE_STATE_0;
585 
586 	return 0;
587 }
588 
589 static struct sock *rose_make_new(struct sock *osk)
590 {
591 	struct sock *sk;
592 	struct rose_sock *rose, *orose;
593 
594 	if (osk->sk_type != SOCK_SEQPACKET)
595 		return NULL;
596 
597 	sk = sk_alloc(sock_net(osk), PF_ROSE, GFP_ATOMIC, &rose_proto, 0);
598 	if (sk == NULL)
599 		return NULL;
600 
601 	rose = rose_sk(sk);
602 
603 	sock_init_data(NULL, sk);
604 
605 	skb_queue_head_init(&rose->ack_queue);
606 #ifdef M_BIT
607 	skb_queue_head_init(&rose->frag_queue);
608 	rose->fraglen  = 0;
609 #endif
610 
611 	sk->sk_type     = osk->sk_type;
612 	sk->sk_priority = READ_ONCE(osk->sk_priority);
613 	sk->sk_protocol = osk->sk_protocol;
614 	sk->sk_rcvbuf   = osk->sk_rcvbuf;
615 	sk->sk_sndbuf   = osk->sk_sndbuf;
616 	sk->sk_state    = TCP_ESTABLISHED;
617 	sock_copy_flags(sk, osk);
618 
619 	timer_setup(&rose->timer, NULL, 0);
620 	timer_setup(&rose->idletimer, NULL, 0);
621 
622 	orose		= rose_sk(osk);
623 	rose->t1	= orose->t1;
624 	rose->t2	= orose->t2;
625 	rose->t3	= orose->t3;
626 	rose->hb	= orose->hb;
627 	rose->idle	= orose->idle;
628 	rose->defer	= orose->defer;
629 	rose->device	= orose->device;
630 	if (rose->device)
631 		netdev_hold(rose->device, &rose->dev_tracker, GFP_ATOMIC);
632 	rose->qbitincl	= orose->qbitincl;
633 
634 	return sk;
635 }
636 
637 static int rose_release(struct socket *sock)
638 {
639 	struct sock *sk = sock->sk;
640 	struct rose_sock *rose;
641 
642 	if (sk == NULL) return 0;
643 
644 	sock_hold(sk);
645 	sock_orphan(sk);
646 	lock_sock(sk);
647 	rose = rose_sk(sk);
648 
649 	switch (rose->state) {
650 	case ROSE_STATE_0:
651 		release_sock(sk);
652 		rose_disconnect(sk, 0, -1, -1);
653 		lock_sock(sk);
654 		rose_destroy_socket(sk);
655 		break;
656 
657 	case ROSE_STATE_2:
658 		rose_neigh_put(rose->neighbour);
659 		release_sock(sk);
660 		rose_disconnect(sk, 0, -1, -1);
661 		lock_sock(sk);
662 		rose_destroy_socket(sk);
663 		break;
664 
665 	case ROSE_STATE_1:
666 	case ROSE_STATE_3:
667 	case ROSE_STATE_4:
668 	case ROSE_STATE_5:
669 		rose_clear_queues(sk);
670 		rose_stop_idletimer(sk);
671 		rose_write_internal(sk, ROSE_CLEAR_REQUEST);
672 		rose_start_t3timer(sk);
673 		rose->state  = ROSE_STATE_2;
674 		sk->sk_state    = TCP_CLOSE;
675 		sk->sk_shutdown |= SEND_SHUTDOWN;
676 		sk->sk_state_change(sk);
677 		sock_set_flag(sk, SOCK_DEAD);
678 		sock_set_flag(sk, SOCK_DESTROY);
679 		break;
680 
681 	default:
682 		break;
683 	}
684 
685 	spin_lock_bh(&rose_list_lock);
686 	netdev_put(rose->device, &rose->dev_tracker);
687 	rose->device = NULL;
688 	spin_unlock_bh(&rose_list_lock);
689 	sock->sk = NULL;
690 	release_sock(sk);
691 	sock_put(sk);
692 
693 	return 0;
694 }
695 
696 static int rose_bind(struct socket *sock, struct sockaddr_unsized *uaddr, int addr_len)
697 {
698 	struct sock *sk = sock->sk;
699 	struct rose_sock *rose = rose_sk(sk);
700 	struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr;
701 	struct net_device *dev;
702 	ax25_address *source;
703 	ax25_uid_assoc *user;
704 	int err = -EINVAL;
705 	int n;
706 
707 	if (addr_len != sizeof(struct sockaddr_rose) && addr_len != sizeof(struct full_sockaddr_rose))
708 		return -EINVAL;
709 
710 	if (addr->srose_family != AF_ROSE)
711 		return -EINVAL;
712 
713 	if (addr_len == sizeof(struct sockaddr_rose) && addr->srose_ndigis > 1)
714 		return -EINVAL;
715 
716 	if ((unsigned int) addr->srose_ndigis > ROSE_MAX_DIGIS)
717 		return -EINVAL;
718 
719 	lock_sock(sk);
720 
721 	if (!sock_flag(sk, SOCK_ZAPPED))
722 		goto out_release;
723 
724 	err = -EADDRNOTAVAIL;
725 	dev = rose_dev_get(&addr->srose_addr);
726 	if (!dev)
727 		goto out_release;
728 
729 	source = &addr->srose_call;
730 
731 	user = ax25_findbyuid(current_euid());
732 	if (user) {
733 		rose->source_call = user->call;
734 		ax25_uid_put(user);
735 	} else {
736 		if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) {
737 			dev_put(dev);
738 			err = -EACCES;
739 			goto out_release;
740 		}
741 		rose->source_call   = *source;
742 	}
743 
744 	rose->source_addr   = addr->srose_addr;
745 	rose->device        = dev;
746 	netdev_tracker_alloc(rose->device, &rose->dev_tracker, GFP_KERNEL);
747 	rose->source_ndigis = addr->srose_ndigis;
748 
749 	if (addr_len == sizeof(struct full_sockaddr_rose)) {
750 		struct full_sockaddr_rose *full_addr = (struct full_sockaddr_rose *)uaddr;
751 		for (n = 0 ; n < addr->srose_ndigis ; n++)
752 			rose->source_digis[n] = full_addr->srose_digis[n];
753 	} else {
754 		if (rose->source_ndigis == 1) {
755 			rose->source_digis[0] = addr->srose_digi;
756 		}
757 	}
758 
759 	rose_insert_socket(sk);
760 
761 	sock_reset_flag(sk, SOCK_ZAPPED);
762 	err = 0;
763 out_release:
764 	release_sock(sk);
765 	return err;
766 }
767 
768 static int rose_connect(struct socket *sock, struct sockaddr_unsized *uaddr, int addr_len,
769 			int flags)
770 {
771 	struct sock *sk = sock->sk;
772 	struct rose_sock *rose = rose_sk(sk);
773 	struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr;
774 	unsigned char cause, diagnostic;
775 	ax25_uid_assoc *user;
776 	int n, err = 0;
777 
778 	if (addr_len != sizeof(struct sockaddr_rose) && addr_len != sizeof(struct full_sockaddr_rose))
779 		return -EINVAL;
780 
781 	if (addr->srose_family != AF_ROSE)
782 		return -EINVAL;
783 
784 	if (addr_len == sizeof(struct sockaddr_rose) && addr->srose_ndigis > 1)
785 		return -EINVAL;
786 
787 	if ((unsigned int) addr->srose_ndigis > ROSE_MAX_DIGIS)
788 		return -EINVAL;
789 
790 	/* Source + Destination digis should not exceed ROSE_MAX_DIGIS */
791 	if ((rose->source_ndigis + addr->srose_ndigis) > ROSE_MAX_DIGIS)
792 		return -EINVAL;
793 
794 	lock_sock(sk);
795 
796 	if (sk->sk_state == TCP_ESTABLISHED && sock->state == SS_CONNECTING) {
797 		/* Connect completed during a ERESTARTSYS event */
798 		sock->state = SS_CONNECTED;
799 		goto out_release;
800 	}
801 
802 	if (sk->sk_state == TCP_CLOSE && sock->state == SS_CONNECTING) {
803 		sock->state = SS_UNCONNECTED;
804 		err = -ECONNREFUSED;
805 		goto out_release;
806 	}
807 
808 	if (sk->sk_state == TCP_ESTABLISHED) {
809 		/* No reconnect on a seqpacket socket */
810 		err = -EISCONN;
811 		goto out_release;
812 	}
813 
814 	sk->sk_state   = TCP_CLOSE;
815 	sock->state = SS_UNCONNECTED;
816 
817 	rose->neighbour = rose_get_neigh(&addr->srose_addr, &cause,
818 					 &diagnostic, 0);
819 	if (!rose->neighbour) {
820 		err = -ENETUNREACH;
821 		goto out_release;
822 	}
823 
824 	rose->lci = rose_new_lci(rose->neighbour);
825 	if (!rose->lci) {
826 		err = -ENETUNREACH;
827 		rose_neigh_put(rose->neighbour);
828 		goto out_release;
829 	}
830 
831 	if (sock_flag(sk, SOCK_ZAPPED)) {	/* Must bind first - autobinding in this may or may not work */
832 		struct net_device *dev;
833 
834 		sock_reset_flag(sk, SOCK_ZAPPED);
835 
836 		dev = rose_dev_first();
837 		if (!dev) {
838 			err = -ENETUNREACH;
839 			rose_neigh_put(rose->neighbour);
840 			goto out_release;
841 		}
842 
843 		user = ax25_findbyuid(current_euid());
844 		if (!user) {
845 			err = -EINVAL;
846 			rose_neigh_put(rose->neighbour);
847 			dev_put(dev);
848 			goto out_release;
849 		}
850 
851 		memcpy(&rose->source_addr, dev->dev_addr, ROSE_ADDR_LEN);
852 		rose->source_call = user->call;
853 		rose->device      = dev;
854 		netdev_tracker_alloc(rose->device, &rose->dev_tracker,
855 				     GFP_KERNEL);
856 		ax25_uid_put(user);
857 
858 		rose_insert_socket(sk);		/* Finish the bind */
859 	}
860 	rose->dest_addr   = addr->srose_addr;
861 	rose->dest_call   = addr->srose_call;
862 	rose->rand        = ((long)rose & 0xFFFF) + rose->lci;
863 	rose->dest_ndigis = addr->srose_ndigis;
864 
865 	if (addr_len == sizeof(struct full_sockaddr_rose)) {
866 		struct full_sockaddr_rose *full_addr = (struct full_sockaddr_rose *)uaddr;
867 		for (n = 0 ; n < addr->srose_ndigis ; n++)
868 			rose->dest_digis[n] = full_addr->srose_digis[n];
869 	} else {
870 		if (rose->dest_ndigis == 1) {
871 			rose->dest_digis[0] = addr->srose_digi;
872 		}
873 	}
874 
875 	/* Move to connecting socket, start sending Connect Requests */
876 	sock->state   = SS_CONNECTING;
877 	sk->sk_state     = TCP_SYN_SENT;
878 
879 	rose->state = ROSE_STATE_1;
880 
881 	rose_write_internal(sk, ROSE_CALL_REQUEST);
882 	rose_start_heartbeat(sk);
883 	rose_start_t1timer(sk);
884 
885 	/* Now the loop */
886 	if (sk->sk_state != TCP_ESTABLISHED && (flags & O_NONBLOCK)) {
887 		err = -EINPROGRESS;
888 		goto out_release;
889 	}
890 
891 	/*
892 	 * A Connect Ack with Choke or timeout or failed routing will go to
893 	 * closed.
894 	 */
895 	if (sk->sk_state == TCP_SYN_SENT) {
896 		DEFINE_WAIT(wait);
897 
898 		for (;;) {
899 			prepare_to_wait(sk_sleep(sk), &wait,
900 					TASK_INTERRUPTIBLE);
901 			if (sk->sk_state != TCP_SYN_SENT)
902 				break;
903 			if (!signal_pending(current)) {
904 				release_sock(sk);
905 				schedule();
906 				lock_sock(sk);
907 				continue;
908 			}
909 			err = -ERESTARTSYS;
910 			break;
911 		}
912 		finish_wait(sk_sleep(sk), &wait);
913 
914 		if (err)
915 			goto out_release;
916 	}
917 
918 	if (sk->sk_state != TCP_ESTABLISHED) {
919 		sock->state = SS_UNCONNECTED;
920 		err = sock_error(sk);	/* Always set at this point */
921 		goto out_release;
922 	}
923 
924 	sock->state = SS_CONNECTED;
925 
926 out_release:
927 	release_sock(sk);
928 
929 	return err;
930 }
931 
932 static int rose_accept(struct socket *sock, struct socket *newsock,
933 		       struct proto_accept_arg *arg)
934 {
935 	struct sk_buff *skb;
936 	struct sock *newsk;
937 	DEFINE_WAIT(wait);
938 	struct sock *sk;
939 	int err = 0;
940 
941 	if ((sk = sock->sk) == NULL)
942 		return -EINVAL;
943 
944 	lock_sock(sk);
945 	if (sk->sk_type != SOCK_SEQPACKET) {
946 		err = -EOPNOTSUPP;
947 		goto out_release;
948 	}
949 
950 	if (sk->sk_state != TCP_LISTEN) {
951 		err = -EINVAL;
952 		goto out_release;
953 	}
954 
955 	/*
956 	 *	The write queue this time is holding sockets ready to use
957 	 *	hooked into the SABM we saved
958 	 */
959 	for (;;) {
960 		prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
961 
962 		skb = skb_dequeue(&sk->sk_receive_queue);
963 		if (skb)
964 			break;
965 
966 		if (arg->flags & O_NONBLOCK) {
967 			err = -EWOULDBLOCK;
968 			break;
969 		}
970 		if (!signal_pending(current)) {
971 			release_sock(sk);
972 			schedule();
973 			lock_sock(sk);
974 			continue;
975 		}
976 		err = -ERESTARTSYS;
977 		break;
978 	}
979 	finish_wait(sk_sleep(sk), &wait);
980 	if (err)
981 		goto out_release;
982 
983 	newsk = skb->sk;
984 	sock_graft(newsk, newsock);
985 
986 	/* Now attach up the new socket */
987 	skb->sk = NULL;
988 	kfree_skb(skb);
989 	sk_acceptq_removed(sk);
990 
991 out_release:
992 	release_sock(sk);
993 
994 	return err;
995 }
996 
997 static int rose_getname(struct socket *sock, struct sockaddr *uaddr,
998 	int peer)
999 {
1000 	struct full_sockaddr_rose *srose = (struct full_sockaddr_rose *)uaddr;
1001 	struct sock *sk = sock->sk;
1002 	struct rose_sock *rose = rose_sk(sk);
1003 	int n;
1004 
1005 	memset(srose, 0, sizeof(*srose));
1006 	if (peer != 0) {
1007 		if (sk->sk_state != TCP_ESTABLISHED)
1008 			return -ENOTCONN;
1009 		srose->srose_family = AF_ROSE;
1010 		srose->srose_addr   = rose->dest_addr;
1011 		srose->srose_call   = rose->dest_call;
1012 		srose->srose_ndigis = rose->dest_ndigis;
1013 		for (n = 0; n < rose->dest_ndigis; n++)
1014 			srose->srose_digis[n] = rose->dest_digis[n];
1015 	} else {
1016 		srose->srose_family = AF_ROSE;
1017 		srose->srose_addr   = rose->source_addr;
1018 		srose->srose_call   = rose->source_call;
1019 		srose->srose_ndigis = rose->source_ndigis;
1020 		for (n = 0; n < rose->source_ndigis; n++)
1021 			srose->srose_digis[n] = rose->source_digis[n];
1022 	}
1023 
1024 	return sizeof(struct full_sockaddr_rose);
1025 }
1026 
1027 int rose_rx_call_request(struct sk_buff *skb, struct net_device *dev, struct rose_neigh *neigh, unsigned int lci)
1028 {
1029 	struct sock *sk;
1030 	struct sock *make;
1031 	struct rose_sock *make_rose;
1032 	struct rose_facilities_struct facilities;
1033 	int n;
1034 
1035 	skb->sk = NULL;		/* Initially we don't know who it's for */
1036 
1037 	/*
1038 	 *	skb->data points to the rose frame start
1039 	 */
1040 	memset(&facilities, 0x00, sizeof(struct rose_facilities_struct));
1041 
1042 	if (!rose_parse_facilities(skb->data + ROSE_CALL_REQ_FACILITIES_OFF,
1043 				   skb->len - ROSE_CALL_REQ_FACILITIES_OFF,
1044 				   &facilities)) {
1045 		rose_transmit_clear_request(neigh, lci, ROSE_INVALID_FACILITY, 76);
1046 		return 0;
1047 	}
1048 
1049 	sk = rose_find_listener(&facilities.source_addr, &facilities.source_call);
1050 
1051 	/*
1052 	 * We can't accept the Call Request.
1053 	 */
1054 	if (sk == NULL || sk_acceptq_is_full(sk) ||
1055 	    (make = rose_make_new(sk)) == NULL) {
1056 		rose_transmit_clear_request(neigh, lci, ROSE_NETWORK_CONGESTION, 120);
1057 		return 0;
1058 	}
1059 
1060 	skb->sk     = make;
1061 	make->sk_state = TCP_ESTABLISHED;
1062 	make_rose = rose_sk(make);
1063 
1064 	make_rose->lci           = lci;
1065 	make_rose->dest_addr     = facilities.dest_addr;
1066 	make_rose->dest_call     = facilities.dest_call;
1067 	make_rose->dest_ndigis   = facilities.dest_ndigis;
1068 	for (n = 0 ; n < facilities.dest_ndigis ; n++)
1069 		make_rose->dest_digis[n] = facilities.dest_digis[n];
1070 	make_rose->source_addr   = facilities.source_addr;
1071 	make_rose->source_call   = facilities.source_call;
1072 	make_rose->source_ndigis = facilities.source_ndigis;
1073 	for (n = 0 ; n < facilities.source_ndigis ; n++)
1074 		make_rose->source_digis[n] = facilities.source_digis[n];
1075 	make_rose->neighbour     = neigh;
1076 	make_rose->device        = dev;
1077 	/* Caller got a reference for us. */
1078 	netdev_tracker_alloc(make_rose->device, &make_rose->dev_tracker,
1079 			     GFP_ATOMIC);
1080 	make_rose->facilities    = facilities;
1081 
1082 	rose_neigh_hold(make_rose->neighbour);
1083 
1084 	if (rose_sk(sk)->defer) {
1085 		make_rose->state = ROSE_STATE_5;
1086 	} else {
1087 		rose_write_internal(make, ROSE_CALL_ACCEPTED);
1088 		make_rose->state = ROSE_STATE_3;
1089 		rose_start_idletimer(make);
1090 	}
1091 
1092 	make_rose->condition = 0x00;
1093 	make_rose->vs        = 0;
1094 	make_rose->va        = 0;
1095 	make_rose->vr        = 0;
1096 	make_rose->vl        = 0;
1097 	sk_acceptq_added(sk);
1098 
1099 	rose_insert_socket(make);
1100 
1101 	skb_queue_head(&sk->sk_receive_queue, skb);
1102 
1103 	rose_start_heartbeat(make);
1104 
1105 	if (!sock_flag(sk, SOCK_DEAD))
1106 		sk->sk_data_ready(sk);
1107 
1108 	return 1;
1109 }
1110 
1111 static int rose_sendmsg(struct socket *sock, struct msghdr *msg, size_t len)
1112 {
1113 	struct sock *sk = sock->sk;
1114 	struct rose_sock *rose = rose_sk(sk);
1115 	DECLARE_SOCKADDR(struct sockaddr_rose *, usrose, msg->msg_name);
1116 	int err;
1117 	struct full_sockaddr_rose srose;
1118 	struct sk_buff *skb;
1119 	unsigned char *asmptr;
1120 	int n, size, qbit = 0;
1121 
1122 	if (msg->msg_flags & ~(MSG_DONTWAIT|MSG_EOR|MSG_CMSG_COMPAT))
1123 		return -EINVAL;
1124 
1125 	if (sock_flag(sk, SOCK_ZAPPED))
1126 		return -EADDRNOTAVAIL;
1127 
1128 	if (sk->sk_shutdown & SEND_SHUTDOWN) {
1129 		send_sig(SIGPIPE, current, 0);
1130 		return -EPIPE;
1131 	}
1132 
1133 	if (rose->neighbour == NULL || rose->device == NULL)
1134 		return -ENETUNREACH;
1135 
1136 	if (usrose != NULL) {
1137 		if (msg->msg_namelen != sizeof(struct sockaddr_rose) && msg->msg_namelen != sizeof(struct full_sockaddr_rose))
1138 			return -EINVAL;
1139 		memset(&srose, 0, sizeof(struct full_sockaddr_rose));
1140 		memcpy(&srose, usrose, msg->msg_namelen);
1141 		if (rosecmp(&rose->dest_addr, &srose.srose_addr) != 0 ||
1142 		    ax25cmp(&rose->dest_call, &srose.srose_call) != 0)
1143 			return -EISCONN;
1144 		if (srose.srose_ndigis != rose->dest_ndigis)
1145 			return -EISCONN;
1146 		if (srose.srose_ndigis == rose->dest_ndigis) {
1147 			for (n = 0 ; n < srose.srose_ndigis ; n++)
1148 				if (ax25cmp(&rose->dest_digis[n],
1149 					    &srose.srose_digis[n]))
1150 					return -EISCONN;
1151 		}
1152 		if (srose.srose_family != AF_ROSE)
1153 			return -EINVAL;
1154 	} else {
1155 		if (sk->sk_state != TCP_ESTABLISHED)
1156 			return -ENOTCONN;
1157 
1158 		srose.srose_family = AF_ROSE;
1159 		srose.srose_addr   = rose->dest_addr;
1160 		srose.srose_call   = rose->dest_call;
1161 		srose.srose_ndigis = rose->dest_ndigis;
1162 		for (n = 0 ; n < rose->dest_ndigis ; n++)
1163 			srose.srose_digis[n] = rose->dest_digis[n];
1164 	}
1165 
1166 	/* Build a packet */
1167 	/* Sanity check the packet size */
1168 	if (len > 65535)
1169 		return -EMSGSIZE;
1170 
1171 	size = len + AX25_BPQ_HEADER_LEN + AX25_MAX_HEADER_LEN + ROSE_MIN_LEN;
1172 
1173 	if ((skb = sock_alloc_send_skb(sk, size, msg->msg_flags & MSG_DONTWAIT, &err)) == NULL)
1174 		return err;
1175 
1176 	skb_reserve(skb, AX25_BPQ_HEADER_LEN + AX25_MAX_HEADER_LEN + ROSE_MIN_LEN);
1177 
1178 	/*
1179 	 *	Put the data on the end
1180 	 */
1181 
1182 	skb_reset_transport_header(skb);
1183 	skb_put(skb, len);
1184 
1185 	err = memcpy_from_msg(skb_transport_header(skb), msg, len);
1186 	if (err) {
1187 		kfree_skb(skb);
1188 		return err;
1189 	}
1190 
1191 	/*
1192 	 *	If the Q BIT Include socket option is in force, the first
1193 	 *	byte of the user data is the logical value of the Q Bit.
1194 	 */
1195 	if (rose->qbitincl) {
1196 		qbit = skb->data[0];
1197 		skb_pull(skb, 1);
1198 	}
1199 
1200 	/*
1201 	 *	Push down the ROSE header
1202 	 */
1203 	asmptr = skb_push(skb, ROSE_MIN_LEN);
1204 
1205 	/* Build a ROSE Network header */
1206 	asmptr[0] = ((rose->lci >> 8) & 0x0F) | ROSE_GFI;
1207 	asmptr[1] = (rose->lci >> 0) & 0xFF;
1208 	asmptr[2] = ROSE_DATA;
1209 
1210 	if (qbit)
1211 		asmptr[0] |= ROSE_Q_BIT;
1212 
1213 	if (sk->sk_state != TCP_ESTABLISHED) {
1214 		kfree_skb(skb);
1215 		return -ENOTCONN;
1216 	}
1217 
1218 #ifdef M_BIT
1219 #define ROSE_PACLEN (256-ROSE_MIN_LEN)
1220 	if (skb->len - ROSE_MIN_LEN > ROSE_PACLEN) {
1221 		unsigned char header[ROSE_MIN_LEN];
1222 		struct sk_buff *skbn;
1223 		int frontlen;
1224 		int lg;
1225 
1226 		/* Save a copy of the Header */
1227 		skb_copy_from_linear_data(skb, header, ROSE_MIN_LEN);
1228 		skb_pull(skb, ROSE_MIN_LEN);
1229 
1230 		frontlen = skb_headroom(skb);
1231 
1232 		while (skb->len > 0) {
1233 			if ((skbn = sock_alloc_send_skb(sk, frontlen + ROSE_PACLEN, 0, &err)) == NULL) {
1234 				kfree_skb(skb);
1235 				return err;
1236 			}
1237 
1238 			skbn->sk   = sk;
1239 			skbn->free = 1;
1240 			skbn->arp  = 1;
1241 
1242 			skb_reserve(skbn, frontlen);
1243 
1244 			lg = (ROSE_PACLEN > skb->len) ? skb->len : ROSE_PACLEN;
1245 
1246 			/* Copy the user data */
1247 			skb_copy_from_linear_data(skb, skb_put(skbn, lg), lg);
1248 			skb_pull(skb, lg);
1249 
1250 			/* Duplicate the Header */
1251 			skb_push(skbn, ROSE_MIN_LEN);
1252 			skb_copy_to_linear_data(skbn, header, ROSE_MIN_LEN);
1253 
1254 			if (skb->len > 0)
1255 				skbn->data[2] |= M_BIT;
1256 
1257 			skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */
1258 		}
1259 
1260 		skb->free = 1;
1261 		kfree_skb(skb);
1262 	} else {
1263 		skb_queue_tail(&sk->sk_write_queue, skb);		/* Throw it on the queue */
1264 	}
1265 #else
1266 	skb_queue_tail(&sk->sk_write_queue, skb);	/* Shove it onto the queue */
1267 #endif
1268 
1269 	rose_kick(sk);
1270 
1271 	return len;
1272 }
1273 
1274 
1275 static int rose_recvmsg(struct socket *sock, struct msghdr *msg, size_t size,
1276 			int flags)
1277 {
1278 	struct sock *sk = sock->sk;
1279 	struct rose_sock *rose = rose_sk(sk);
1280 	size_t copied;
1281 	unsigned char *asmptr;
1282 	struct sk_buff *skb;
1283 	int n, er, qbit;
1284 
1285 	/*
1286 	 * This works for seqpacket too. The receiver has ordered the queue for
1287 	 * us! We do one quick check first though
1288 	 */
1289 	if (sk->sk_state != TCP_ESTABLISHED)
1290 		return -ENOTCONN;
1291 
1292 	/* Now we can treat all alike */
1293 	skb = skb_recv_datagram(sk, flags, &er);
1294 	if (!skb)
1295 		return er;
1296 
1297 	qbit = (skb->data[0] & ROSE_Q_BIT) == ROSE_Q_BIT;
1298 
1299 	skb_pull(skb, ROSE_MIN_LEN);
1300 
1301 	if (rose->qbitincl) {
1302 		asmptr  = skb_push(skb, 1);
1303 		*asmptr = qbit;
1304 	}
1305 
1306 	skb_reset_transport_header(skb);
1307 	copied     = skb->len;
1308 
1309 	if (copied > size) {
1310 		copied = size;
1311 		msg->msg_flags |= MSG_TRUNC;
1312 	}
1313 
1314 	skb_copy_datagram_msg(skb, 0, msg, copied);
1315 
1316 	if (msg->msg_name) {
1317 		struct sockaddr_rose *srose;
1318 		DECLARE_SOCKADDR(struct full_sockaddr_rose *, full_srose,
1319 				 msg->msg_name);
1320 
1321 		memset(msg->msg_name, 0, sizeof(struct full_sockaddr_rose));
1322 		srose = msg->msg_name;
1323 		srose->srose_family = AF_ROSE;
1324 		srose->srose_addr   = rose->dest_addr;
1325 		srose->srose_call   = rose->dest_call;
1326 		srose->srose_ndigis = rose->dest_ndigis;
1327 		for (n = 0 ; n < rose->dest_ndigis ; n++)
1328 			full_srose->srose_digis[n] = rose->dest_digis[n];
1329 		msg->msg_namelen = sizeof(struct full_sockaddr_rose);
1330 	}
1331 
1332 	skb_free_datagram(sk, skb);
1333 
1334 	return copied;
1335 }
1336 
1337 
1338 static int rose_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg)
1339 {
1340 	struct sock *sk = sock->sk;
1341 	struct rose_sock *rose = rose_sk(sk);
1342 	void __user *argp = (void __user *)arg;
1343 
1344 	switch (cmd) {
1345 	case TIOCOUTQ: {
1346 		long amount;
1347 
1348 		amount = sk->sk_sndbuf - sk_wmem_alloc_get(sk);
1349 		if (amount < 0)
1350 			amount = 0;
1351 		return put_user(amount, (unsigned int __user *) argp);
1352 	}
1353 
1354 	case TIOCINQ: {
1355 		struct sk_buff *skb;
1356 		long amount = 0L;
1357 
1358 		spin_lock_irq(&sk->sk_receive_queue.lock);
1359 		if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL)
1360 			amount = skb->len;
1361 		spin_unlock_irq(&sk->sk_receive_queue.lock);
1362 		return put_user(amount, (unsigned int __user *) argp);
1363 	}
1364 
1365 	case SIOCGIFADDR:
1366 	case SIOCSIFADDR:
1367 	case SIOCGIFDSTADDR:
1368 	case SIOCSIFDSTADDR:
1369 	case SIOCGIFBRDADDR:
1370 	case SIOCSIFBRDADDR:
1371 	case SIOCGIFNETMASK:
1372 	case SIOCSIFNETMASK:
1373 	case SIOCGIFMETRIC:
1374 	case SIOCSIFMETRIC:
1375 		return -EINVAL;
1376 
1377 	case SIOCADDRT:
1378 	case SIOCDELRT:
1379 	case SIOCRSCLRRT:
1380 		if (!capable(CAP_NET_ADMIN))
1381 			return -EPERM;
1382 		return rose_rt_ioctl(cmd, argp);
1383 
1384 	case SIOCRSGCAUSE: {
1385 		struct rose_cause_struct rose_cause;
1386 		rose_cause.cause      = rose->cause;
1387 		rose_cause.diagnostic = rose->diagnostic;
1388 		return copy_to_user(argp, &rose_cause, sizeof(struct rose_cause_struct)) ? -EFAULT : 0;
1389 	}
1390 
1391 	case SIOCRSSCAUSE: {
1392 		struct rose_cause_struct rose_cause;
1393 		if (copy_from_user(&rose_cause, argp, sizeof(struct rose_cause_struct)))
1394 			return -EFAULT;
1395 		rose->cause      = rose_cause.cause;
1396 		rose->diagnostic = rose_cause.diagnostic;
1397 		return 0;
1398 	}
1399 
1400 	case SIOCRSSL2CALL:
1401 		if (!capable(CAP_NET_ADMIN)) return -EPERM;
1402 		if (ax25cmp(&rose_callsign, &null_ax25_address) != 0)
1403 			ax25_listen_release(&rose_callsign, NULL);
1404 		if (copy_from_user(&rose_callsign, argp, sizeof(ax25_address)))
1405 			return -EFAULT;
1406 		if (ax25cmp(&rose_callsign, &null_ax25_address) != 0)
1407 			return ax25_listen_register(&rose_callsign, NULL);
1408 
1409 		return 0;
1410 
1411 	case SIOCRSGL2CALL:
1412 		return copy_to_user(argp, &rose_callsign, sizeof(ax25_address)) ? -EFAULT : 0;
1413 
1414 	case SIOCRSACCEPT:
1415 		if (rose->state == ROSE_STATE_5) {
1416 			rose_write_internal(sk, ROSE_CALL_ACCEPTED);
1417 			rose_start_idletimer(sk);
1418 			rose->condition = 0x00;
1419 			rose->vs        = 0;
1420 			rose->va        = 0;
1421 			rose->vr        = 0;
1422 			rose->vl        = 0;
1423 			rose->state     = ROSE_STATE_3;
1424 		}
1425 		return 0;
1426 
1427 	default:
1428 		return -ENOIOCTLCMD;
1429 	}
1430 
1431 	return 0;
1432 }
1433 
1434 #ifdef CONFIG_PROC_FS
1435 static void *rose_info_start(struct seq_file *seq, loff_t *pos)
1436 	__acquires(rose_list_lock)
1437 {
1438 	spin_lock_bh(&rose_list_lock);
1439 	return seq_hlist_start_head(&rose_list, *pos);
1440 }
1441 
1442 static void *rose_info_next(struct seq_file *seq, void *v, loff_t *pos)
1443 {
1444 	return seq_hlist_next(v, &rose_list, pos);
1445 }
1446 
1447 static void rose_info_stop(struct seq_file *seq, void *v)
1448 	__releases(rose_list_lock)
1449 {
1450 	spin_unlock_bh(&rose_list_lock);
1451 }
1452 
1453 static int rose_info_show(struct seq_file *seq, void *v)
1454 {
1455 	char buf[11], rsbuf[11];
1456 
1457 	if (v == SEQ_START_TOKEN)
1458 		seq_puts(seq,
1459 			 "dest_addr  dest_call src_addr   src_call  dev   lci neigh st vs vr va   t  t1  t2  t3  hb    idle Snd-Q Rcv-Q inode\n");
1460 
1461 	else {
1462 		struct sock *s = sk_entry(v);
1463 		struct rose_sock *rose = rose_sk(s);
1464 		const char *devname, *callsign;
1465 		const struct net_device *dev = rose->device;
1466 
1467 		if (!dev)
1468 			devname = "???";
1469 		else
1470 			devname = dev->name;
1471 
1472 		seq_printf(seq, "%-10s %-9s ",
1473 			   rose2asc(rsbuf, &rose->dest_addr),
1474 			   ax2asc(buf, &rose->dest_call));
1475 
1476 		if (ax25cmp(&rose->source_call, &null_ax25_address) == 0)
1477 			callsign = "??????-?";
1478 		else
1479 			callsign = ax2asc(buf, &rose->source_call);
1480 
1481 		seq_printf(seq,
1482 			   "%-10s %-9s %-5s %3.3X %05d  %d  %d  %d  %d %3lu %3lu %3lu %3lu %3lu %3lu/%03lu %5d %5d %ld\n",
1483 			rose2asc(rsbuf, &rose->source_addr),
1484 			callsign,
1485 			devname,
1486 			rose->lci & 0x0FFF,
1487 			(rose->neighbour) ? rose->neighbour->number : 0,
1488 			rose->state,
1489 			rose->vs,
1490 			rose->vr,
1491 			rose->va,
1492 			ax25_display_timer(&rose->timer) / HZ,
1493 			rose->t1 / HZ,
1494 			rose->t2 / HZ,
1495 			rose->t3 / HZ,
1496 			rose->hb / HZ,
1497 			ax25_display_timer(&rose->idletimer) / (60 * HZ),
1498 			rose->idle / (60 * HZ),
1499 			sk_wmem_alloc_get(s),
1500 			sk_rmem_alloc_get(s),
1501 			s->sk_socket ? SOCK_INODE(s->sk_socket)->i_ino : 0L);
1502 	}
1503 
1504 	return 0;
1505 }
1506 
1507 static const struct seq_operations rose_info_seqops = {
1508 	.start = rose_info_start,
1509 	.next = rose_info_next,
1510 	.stop = rose_info_stop,
1511 	.show = rose_info_show,
1512 };
1513 #endif	/* CONFIG_PROC_FS */
1514 
1515 static const struct net_proto_family rose_family_ops = {
1516 	.family		=	PF_ROSE,
1517 	.create		=	rose_create,
1518 	.owner		=	THIS_MODULE,
1519 };
1520 
1521 static const struct proto_ops rose_proto_ops = {
1522 	.family		=	PF_ROSE,
1523 	.owner		=	THIS_MODULE,
1524 	.release	=	rose_release,
1525 	.bind		=	rose_bind,
1526 	.connect	=	rose_connect,
1527 	.socketpair	=	sock_no_socketpair,
1528 	.accept		=	rose_accept,
1529 	.getname	=	rose_getname,
1530 	.poll		=	datagram_poll,
1531 	.ioctl		=	rose_ioctl,
1532 	.gettstamp	=	sock_gettstamp,
1533 	.listen		=	rose_listen,
1534 	.shutdown	=	sock_no_shutdown,
1535 	.setsockopt	=	rose_setsockopt,
1536 	.getsockopt	=	rose_getsockopt,
1537 	.sendmsg	=	rose_sendmsg,
1538 	.recvmsg	=	rose_recvmsg,
1539 	.mmap		=	sock_no_mmap,
1540 };
1541 
1542 static struct notifier_block rose_dev_notifier = {
1543 	.notifier_call	=	rose_device_event,
1544 };
1545 
1546 static struct net_device **dev_rose;
1547 
1548 static struct ax25_protocol rose_pid = {
1549 	.pid	= AX25_P_ROSE,
1550 	.func	= rose_route_frame
1551 };
1552 
1553 static struct ax25_linkfail rose_linkfail_notifier = {
1554 	.func	= rose_link_failed
1555 };
1556 
1557 static int __init rose_proto_init(void)
1558 {
1559 	int i;
1560 	int rc;
1561 
1562 	if (rose_ndevs > 0x7FFFFFFF/sizeof(struct net_device *)) {
1563 		printk(KERN_ERR "ROSE: rose_proto_init - rose_ndevs parameter too large\n");
1564 		rc = -EINVAL;
1565 		goto out;
1566 	}
1567 
1568 	rc = proto_register(&rose_proto, 0);
1569 	if (rc != 0)
1570 		goto out;
1571 
1572 	rose_callsign = null_ax25_address;
1573 
1574 	dev_rose = kcalloc(rose_ndevs, sizeof(struct net_device *),
1575 			   GFP_KERNEL);
1576 	if (dev_rose == NULL) {
1577 		printk(KERN_ERR "ROSE: rose_proto_init - unable to allocate device structure\n");
1578 		rc = -ENOMEM;
1579 		goto out_proto_unregister;
1580 	}
1581 
1582 	for (i = 0; i < rose_ndevs; i++) {
1583 		struct net_device *dev;
1584 		char name[IFNAMSIZ];
1585 
1586 		sprintf(name, "rose%d", i);
1587 		dev = alloc_netdev(0, name, NET_NAME_UNKNOWN, rose_setup);
1588 		if (!dev) {
1589 			printk(KERN_ERR "ROSE: rose_proto_init - unable to allocate memory\n");
1590 			rc = -ENOMEM;
1591 			goto fail;
1592 		}
1593 		rc = register_netdev(dev);
1594 		if (rc) {
1595 			printk(KERN_ERR "ROSE: netdevice registration failed\n");
1596 			free_netdev(dev);
1597 			goto fail;
1598 		}
1599 		rose_set_lockdep_key(dev);
1600 		dev_rose[i] = dev;
1601 	}
1602 
1603 	sock_register(&rose_family_ops);
1604 	register_netdevice_notifier(&rose_dev_notifier);
1605 
1606 	ax25_register_pid(&rose_pid);
1607 	ax25_linkfail_register(&rose_linkfail_notifier);
1608 
1609 #ifdef CONFIG_SYSCTL
1610 	rose_register_sysctl();
1611 #endif
1612 	rose_loopback_init();
1613 
1614 	rose_add_loopback_neigh();
1615 
1616 	proc_create_seq("rose", 0444, init_net.proc_net, &rose_info_seqops);
1617 	proc_create_seq("rose_neigh", 0444, init_net.proc_net,
1618 		    &rose_neigh_seqops);
1619 	proc_create_seq("rose_nodes", 0444, init_net.proc_net,
1620 		    &rose_node_seqops);
1621 	proc_create_seq("rose_routes", 0444, init_net.proc_net,
1622 		    &rose_route_seqops);
1623 out:
1624 	return rc;
1625 fail:
1626 	while (--i >= 0) {
1627 		unregister_netdev(dev_rose[i]);
1628 		free_netdev(dev_rose[i]);
1629 	}
1630 	kfree(dev_rose);
1631 out_proto_unregister:
1632 	proto_unregister(&rose_proto);
1633 	goto out;
1634 }
1635 module_init(rose_proto_init);
1636 
1637 module_param(rose_ndevs, int, 0);
1638 MODULE_PARM_DESC(rose_ndevs, "number of ROSE devices");
1639 
1640 MODULE_AUTHOR("Jonathan Naylor G4KLX <g4klx@g4klx.demon.co.uk>");
1641 MODULE_DESCRIPTION("The amateur radio ROSE network layer protocol");
1642 MODULE_LICENSE("GPL");
1643 MODULE_ALIAS_NETPROTO(PF_ROSE);
1644 
1645 static void __exit rose_exit(void)
1646 {
1647 	int i;
1648 
1649 	remove_proc_entry("rose", init_net.proc_net);
1650 	remove_proc_entry("rose_neigh", init_net.proc_net);
1651 	remove_proc_entry("rose_nodes", init_net.proc_net);
1652 	remove_proc_entry("rose_routes", init_net.proc_net);
1653 	rose_loopback_clear();
1654 
1655 	rose_rt_free();
1656 
1657 	ax25_protocol_release(AX25_P_ROSE);
1658 	ax25_linkfail_release(&rose_linkfail_notifier);
1659 
1660 	if (ax25cmp(&rose_callsign, &null_ax25_address) != 0)
1661 		ax25_listen_release(&rose_callsign, NULL);
1662 
1663 #ifdef CONFIG_SYSCTL
1664 	rose_unregister_sysctl();
1665 #endif
1666 	unregister_netdevice_notifier(&rose_dev_notifier);
1667 
1668 	sock_unregister(PF_ROSE);
1669 
1670 	for (i = 0; i < rose_ndevs; i++) {
1671 		struct net_device *dev = dev_rose[i];
1672 
1673 		if (dev) {
1674 			unregister_netdev(dev);
1675 			free_netdev(dev);
1676 		}
1677 	}
1678 
1679 	kfree(dev_rose);
1680 	proto_unregister(&rose_proto);
1681 }
1682 
1683 module_exit(rose_exit);
1684