Linux Audio

Check our new training course

Loading...
v3.5.6
   1/*
   2 * NET4:	Implementation of BSD Unix domain sockets.
   3 *
   4 * Authors:	Alan Cox, <alan@lxorguk.ukuu.org.uk>
   5 *
   6 *		This program is free software; you can redistribute it and/or
   7 *		modify it under the terms of the GNU General Public License
   8 *		as published by the Free Software Foundation; either version
   9 *		2 of the License, or (at your option) any later version.
  10 *
  11 * Fixes:
  12 *		Linus Torvalds	:	Assorted bug cures.
  13 *		Niibe Yutaka	:	async I/O support.
  14 *		Carsten Paeth	:	PF_UNIX check, address fixes.
  15 *		Alan Cox	:	Limit size of allocated blocks.
  16 *		Alan Cox	:	Fixed the stupid socketpair bug.
  17 *		Alan Cox	:	BSD compatibility fine tuning.
  18 *		Alan Cox	:	Fixed a bug in connect when interrupted.
  19 *		Alan Cox	:	Sorted out a proper draft version of
  20 *					file descriptor passing hacked up from
  21 *					Mike Shaver's work.
  22 *		Marty Leisner	:	Fixes to fd passing
  23 *		Nick Nevin	:	recvmsg bugfix.
  24 *		Alan Cox	:	Started proper garbage collector
  25 *		Heiko EiBfeldt	:	Missing verify_area check
  26 *		Alan Cox	:	Started POSIXisms
  27 *		Andreas Schwab	:	Replace inode by dentry for proper
  28 *					reference counting
  29 *		Kirk Petersen	:	Made this a module
  30 *	    Christoph Rohland	:	Elegant non-blocking accept/connect algorithm.
  31 *					Lots of bug fixes.
  32 *	     Alexey Kuznetosv	:	Repaired (I hope) bugs introduces
  33 *					by above two patches.
  34 *	     Andrea Arcangeli	:	If possible we block in connect(2)
  35 *					if the max backlog of the listen socket
  36 *					is been reached. This won't break
  37 *					old apps and it will avoid huge amount
  38 *					of socks hashed (this for unix_gc()
  39 *					performances reasons).
  40 *					Security fix that limits the max
  41 *					number of socks to 2*max_files and
  42 *					the number of skb queueable in the
  43 *					dgram receiver.
  44 *		Artur Skawina   :	Hash function optimizations
  45 *	     Alexey Kuznetsov   :	Full scale SMP. Lot of bugs are introduced 8)
  46 *	      Malcolm Beattie   :	Set peercred for socketpair
  47 *	     Michal Ostrowski   :       Module initialization cleanup.
  48 *	     Arnaldo C. Melo	:	Remove MOD_{INC,DEC}_USE_COUNT,
  49 *	     				the core infrastructure is doing that
  50 *	     				for all net proto families now (2.5.69+)
  51 *
  52 *
  53 * Known differences from reference BSD that was tested:
  54 *
  55 *	[TO FIX]
  56 *	ECONNREFUSED is not returned from one end of a connected() socket to the
  57 *		other the moment one end closes.
  58 *	fstat() doesn't return st_dev=0, and give the blksize as high water mark
  59 *		and a fake inode identifier (nor the BSD first socket fstat twice bug).
  60 *	[NOT TO FIX]
  61 *	accept() returns a path name even if the connecting socket has closed
  62 *		in the meantime (BSD loses the path and gives up).
  63 *	accept() returns 0 length path for an unbound connector. BSD returns 16
  64 *		and a null first byte in the path (but not for gethost/peername - BSD bug ??)
  65 *	socketpair(...SOCK_RAW..) doesn't panic the kernel.
  66 *	BSD af_unix apparently has connect forgetting to block properly.
  67 *		(need to check this with the POSIX spec in detail)
  68 *
  69 * Differences from 2.0.0-11-... (ANK)
  70 *	Bug fixes and improvements.
  71 *		- client shutdown killed server socket.
  72 *		- removed all useless cli/sti pairs.
  73 *
  74 *	Semantic changes/extensions.
  75 *		- generic control message passing.
  76 *		- SCM_CREDENTIALS control message.
  77 *		- "Abstract" (not FS based) socket bindings.
  78 *		  Abstract names are sequences of bytes (not zero terminated)
  79 *		  started by 0, so that this name space does not intersect
  80 *		  with BSD names.
  81 */
  82
 
 
  83#include <linux/module.h>
  84#include <linux/kernel.h>
  85#include <linux/signal.h>
  86#include <linux/sched.h>
  87#include <linux/errno.h>
  88#include <linux/string.h>
  89#include <linux/stat.h>
  90#include <linux/dcache.h>
  91#include <linux/namei.h>
  92#include <linux/socket.h>
  93#include <linux/un.h>
  94#include <linux/fcntl.h>
  95#include <linux/termios.h>
  96#include <linux/sockios.h>
  97#include <linux/net.h>
  98#include <linux/in.h>
  99#include <linux/fs.h>
 100#include <linux/slab.h>
 101#include <asm/uaccess.h>
 102#include <linux/skbuff.h>
 103#include <linux/netdevice.h>
 104#include <net/net_namespace.h>
 105#include <net/sock.h>
 106#include <net/tcp_states.h>
 107#include <net/af_unix.h>
 108#include <linux/proc_fs.h>
 109#include <linux/seq_file.h>
 110#include <net/scm.h>
 111#include <linux/init.h>
 112#include <linux/poll.h>
 113#include <linux/rtnetlink.h>
 114#include <linux/mount.h>
 115#include <net/checksum.h>
 116#include <linux/security.h>
 
 117
 118struct hlist_head unix_socket_table[UNIX_HASH_SIZE + 1];
 119EXPORT_SYMBOL_GPL(unix_socket_table);
 120DEFINE_SPINLOCK(unix_table_lock);
 121EXPORT_SYMBOL_GPL(unix_table_lock);
 122static atomic_long_t unix_nr_socks;
 123
 124#define unix_sockets_unbound	(&unix_socket_table[UNIX_HASH_SIZE])
 125
 126#define UNIX_ABSTRACT(sk)	(unix_sk(sk)->addr->hash != UNIX_HASH_SIZE)
 
 
 
 
 
 
 
 
 
 
 127
 128#ifdef CONFIG_SECURITY_NETWORK
 129static void unix_get_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 130{
 131	memcpy(UNIXSID(skb), &scm->secid, sizeof(u32));
 132}
 133
 134static inline void unix_set_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 135{
 136	scm->secid = *UNIXSID(skb);
 137}
 138#else
 139static inline void unix_get_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 140{ }
 141
 142static inline void unix_set_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 143{ }
 144#endif /* CONFIG_SECURITY_NETWORK */
 145
 146/*
 147 *  SMP locking strategy:
 148 *    hash table is protected with spinlock unix_table_lock
 149 *    each socket state is protected by separate spin lock.
 150 */
 151
 152static inline unsigned int unix_hash_fold(__wsum n)
 153{
 154	unsigned int hash = (__force unsigned int)n;
 155
 156	hash ^= hash>>16;
 157	hash ^= hash>>8;
 158	return hash&(UNIX_HASH_SIZE-1);
 159}
 160
 161#define unix_peer(sk) (unix_sk(sk)->peer)
 162
 163static inline int unix_our_peer(struct sock *sk, struct sock *osk)
 164{
 165	return unix_peer(osk) == sk;
 166}
 167
 168static inline int unix_may_send(struct sock *sk, struct sock *osk)
 169{
 170	return unix_peer(osk) == NULL || unix_our_peer(sk, osk);
 171}
 172
 173static inline int unix_recvq_full(struct sock const *sk)
 174{
 175	return skb_queue_len(&sk->sk_receive_queue) > sk->sk_max_ack_backlog;
 176}
 177
 178struct sock *unix_peer_get(struct sock *s)
 179{
 180	struct sock *peer;
 181
 182	unix_state_lock(s);
 183	peer = unix_peer(s);
 184	if (peer)
 185		sock_hold(peer);
 186	unix_state_unlock(s);
 187	return peer;
 188}
 189EXPORT_SYMBOL_GPL(unix_peer_get);
 190
 191static inline void unix_release_addr(struct unix_address *addr)
 192{
 193	if (atomic_dec_and_test(&addr->refcnt))
 194		kfree(addr);
 195}
 196
 197/*
 198 *	Check unix socket name:
 199 *		- should be not zero length.
 200 *	        - if started by not zero, should be NULL terminated (FS object)
 201 *		- if started by zero, it is abstract name.
 202 */
 203
 204static int unix_mkname(struct sockaddr_un *sunaddr, int len, unsigned int *hashp)
 205{
 206	if (len <= sizeof(short) || len > sizeof(*sunaddr))
 207		return -EINVAL;
 208	if (!sunaddr || sunaddr->sun_family != AF_UNIX)
 209		return -EINVAL;
 210	if (sunaddr->sun_path[0]) {
 211		/*
 212		 * This may look like an off by one error but it is a bit more
 213		 * subtle. 108 is the longest valid AF_UNIX path for a binding.
 214		 * sun_path[108] doesn't as such exist.  However in kernel space
 215		 * we are guaranteed that it is a valid memory location in our
 216		 * kernel address buffer.
 217		 */
 218		((char *)sunaddr)[len] = 0;
 219		len = strlen(sunaddr->sun_path)+1+sizeof(short);
 220		return len;
 221	}
 222
 223	*hashp = unix_hash_fold(csum_partial(sunaddr, len, 0));
 224	return len;
 225}
 226
 227static void __unix_remove_socket(struct sock *sk)
 228{
 229	sk_del_node_init(sk);
 230}
 231
 232static void __unix_insert_socket(struct hlist_head *list, struct sock *sk)
 233{
 234	WARN_ON(!sk_unhashed(sk));
 235	sk_add_node(sk, list);
 236}
 237
 238static inline void unix_remove_socket(struct sock *sk)
 239{
 240	spin_lock(&unix_table_lock);
 241	__unix_remove_socket(sk);
 242	spin_unlock(&unix_table_lock);
 243}
 244
 245static inline void unix_insert_socket(struct hlist_head *list, struct sock *sk)
 246{
 247	spin_lock(&unix_table_lock);
 248	__unix_insert_socket(list, sk);
 249	spin_unlock(&unix_table_lock);
 250}
 251
 252static struct sock *__unix_find_socket_byname(struct net *net,
 253					      struct sockaddr_un *sunname,
 254					      int len, int type, unsigned int hash)
 255{
 256	struct sock *s;
 257	struct hlist_node *node;
 258
 259	sk_for_each(s, node, &unix_socket_table[hash ^ type]) {
 260		struct unix_sock *u = unix_sk(s);
 261
 262		if (!net_eq(sock_net(s), net))
 263			continue;
 264
 265		if (u->addr->len == len &&
 266		    !memcmp(u->addr->name, sunname, len))
 267			goto found;
 268	}
 269	s = NULL;
 270found:
 271	return s;
 272}
 273
 274static inline struct sock *unix_find_socket_byname(struct net *net,
 275						   struct sockaddr_un *sunname,
 276						   int len, int type,
 277						   unsigned int hash)
 278{
 279	struct sock *s;
 280
 281	spin_lock(&unix_table_lock);
 282	s = __unix_find_socket_byname(net, sunname, len, type, hash);
 283	if (s)
 284		sock_hold(s);
 285	spin_unlock(&unix_table_lock);
 286	return s;
 287}
 288
 289static struct sock *unix_find_socket_byinode(struct inode *i)
 290{
 291	struct sock *s;
 292	struct hlist_node *node;
 293
 294	spin_lock(&unix_table_lock);
 295	sk_for_each(s, node,
 296		    &unix_socket_table[i->i_ino & (UNIX_HASH_SIZE - 1)]) {
 297		struct dentry *dentry = unix_sk(s)->path.dentry;
 298
 299		if (dentry && dentry->d_inode == i) {
 300			sock_hold(s);
 301			goto found;
 302		}
 303	}
 304	s = NULL;
 305found:
 306	spin_unlock(&unix_table_lock);
 307	return s;
 308}
 309
 310static inline int unix_writable(struct sock *sk)
 311{
 312	return (atomic_read(&sk->sk_wmem_alloc) << 2) <= sk->sk_sndbuf;
 313}
 314
 315static void unix_write_space(struct sock *sk)
 316{
 317	struct socket_wq *wq;
 318
 319	rcu_read_lock();
 320	if (unix_writable(sk)) {
 321		wq = rcu_dereference(sk->sk_wq);
 322		if (wq_has_sleeper(wq))
 323			wake_up_interruptible_sync_poll(&wq->wait,
 324				POLLOUT | POLLWRNORM | POLLWRBAND);
 325		sk_wake_async(sk, SOCK_WAKE_SPACE, POLL_OUT);
 326	}
 327	rcu_read_unlock();
 328}
 329
 330/* When dgram socket disconnects (or changes its peer), we clear its receive
 331 * queue of packets arrived from previous peer. First, it allows to do
 332 * flow control based only on wmem_alloc; second, sk connected to peer
 333 * may receive messages only from that peer. */
 334static void unix_dgram_disconnected(struct sock *sk, struct sock *other)
 335{
 336	if (!skb_queue_empty(&sk->sk_receive_queue)) {
 337		skb_queue_purge(&sk->sk_receive_queue);
 338		wake_up_interruptible_all(&unix_sk(sk)->peer_wait);
 339
 340		/* If one link of bidirectional dgram pipe is disconnected,
 341		 * we signal error. Messages are lost. Do not make this,
 342		 * when peer was not connected to us.
 343		 */
 344		if (!sock_flag(other, SOCK_DEAD) && unix_peer(other) == sk) {
 345			other->sk_err = ECONNRESET;
 346			other->sk_error_report(other);
 347		}
 348	}
 349}
 350
 351static void unix_sock_destructor(struct sock *sk)
 352{
 353	struct unix_sock *u = unix_sk(sk);
 354
 355	skb_queue_purge(&sk->sk_receive_queue);
 356
 357	WARN_ON(atomic_read(&sk->sk_wmem_alloc));
 358	WARN_ON(!sk_unhashed(sk));
 359	WARN_ON(sk->sk_socket);
 360	if (!sock_flag(sk, SOCK_DEAD)) {
 361		printk(KERN_INFO "Attempt to release alive unix socket: %p\n", sk);
 362		return;
 363	}
 364
 365	if (u->addr)
 366		unix_release_addr(u->addr);
 367
 368	atomic_long_dec(&unix_nr_socks);
 369	local_bh_disable();
 370	sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1);
 371	local_bh_enable();
 372#ifdef UNIX_REFCNT_DEBUG
 373	printk(KERN_DEBUG "UNIX %p is destroyed, %ld are still alive.\n", sk,
 374		atomic_long_read(&unix_nr_socks));
 375#endif
 376}
 377
 378static int unix_release_sock(struct sock *sk, int embrion)
 379{
 380	struct unix_sock *u = unix_sk(sk);
 381	struct path path;
 382	struct sock *skpair;
 383	struct sk_buff *skb;
 384	int state;
 385
 386	unix_remove_socket(sk);
 387
 388	/* Clear state */
 389	unix_state_lock(sk);
 390	sock_orphan(sk);
 391	sk->sk_shutdown = SHUTDOWN_MASK;
 392	path	     = u->path;
 393	u->path.dentry = NULL;
 394	u->path.mnt = NULL;
 395	state = sk->sk_state;
 396	sk->sk_state = TCP_CLOSE;
 397	unix_state_unlock(sk);
 398
 399	wake_up_interruptible_all(&u->peer_wait);
 400
 401	skpair = unix_peer(sk);
 402
 403	if (skpair != NULL) {
 404		if (sk->sk_type == SOCK_STREAM || sk->sk_type == SOCK_SEQPACKET) {
 405			unix_state_lock(skpair);
 406			/* No more writes */
 407			skpair->sk_shutdown = SHUTDOWN_MASK;
 408			if (!skb_queue_empty(&sk->sk_receive_queue) || embrion)
 409				skpair->sk_err = ECONNRESET;
 410			unix_state_unlock(skpair);
 411			skpair->sk_state_change(skpair);
 412			sk_wake_async(skpair, SOCK_WAKE_WAITD, POLL_HUP);
 413		}
 414		sock_put(skpair); /* It may now die */
 415		unix_peer(sk) = NULL;
 416	}
 417
 418	/* Try to flush out this socket. Throw out buffers at least */
 419
 420	while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) {
 421		if (state == TCP_LISTEN)
 422			unix_release_sock(skb->sk, 1);
 423		/* passed fds are erased in the kfree_skb hook	      */
 424		kfree_skb(skb);
 425	}
 426
 427	if (path.dentry)
 428		path_put(&path);
 429
 430	sock_put(sk);
 431
 432	/* ---- Socket is dead now and most probably destroyed ---- */
 433
 434	/*
 435	 * Fixme: BSD difference: In BSD all sockets connected to use get
 436	 *	  ECONNRESET and we die on the spot. In Linux we behave
 437	 *	  like files and pipes do and wait for the last
 438	 *	  dereference.
 439	 *
 440	 * Can't we simply set sock->err?
 441	 *
 442	 *	  What the above comment does talk about? --ANK(980817)
 443	 */
 444
 445	if (unix_tot_inflight)
 446		unix_gc();		/* Garbage collect fds */
 447
 448	return 0;
 449}
 450
 451static void init_peercred(struct sock *sk)
 452{
 453	put_pid(sk->sk_peer_pid);
 454	if (sk->sk_peer_cred)
 455		put_cred(sk->sk_peer_cred);
 456	sk->sk_peer_pid  = get_pid(task_tgid(current));
 457	sk->sk_peer_cred = get_current_cred();
 458}
 459
 460static void copy_peercred(struct sock *sk, struct sock *peersk)
 461{
 462	put_pid(sk->sk_peer_pid);
 463	if (sk->sk_peer_cred)
 464		put_cred(sk->sk_peer_cred);
 465	sk->sk_peer_pid  = get_pid(peersk->sk_peer_pid);
 466	sk->sk_peer_cred = get_cred(peersk->sk_peer_cred);
 467}
 468
 469static int unix_listen(struct socket *sock, int backlog)
 470{
 471	int err;
 472	struct sock *sk = sock->sk;
 473	struct unix_sock *u = unix_sk(sk);
 474	struct pid *old_pid = NULL;
 475	const struct cred *old_cred = NULL;
 476
 477	err = -EOPNOTSUPP;
 478	if (sock->type != SOCK_STREAM && sock->type != SOCK_SEQPACKET)
 479		goto out;	/* Only stream/seqpacket sockets accept */
 480	err = -EINVAL;
 481	if (!u->addr)
 482		goto out;	/* No listens on an unbound socket */
 483	unix_state_lock(sk);
 484	if (sk->sk_state != TCP_CLOSE && sk->sk_state != TCP_LISTEN)
 485		goto out_unlock;
 486	if (backlog > sk->sk_max_ack_backlog)
 487		wake_up_interruptible_all(&u->peer_wait);
 488	sk->sk_max_ack_backlog	= backlog;
 489	sk->sk_state		= TCP_LISTEN;
 490	/* set credentials so connect can copy them */
 491	init_peercred(sk);
 492	err = 0;
 493
 494out_unlock:
 495	unix_state_unlock(sk);
 496	put_pid(old_pid);
 497	if (old_cred)
 498		put_cred(old_cred);
 499out:
 500	return err;
 501}
 502
 503static int unix_release(struct socket *);
 504static int unix_bind(struct socket *, struct sockaddr *, int);
 505static int unix_stream_connect(struct socket *, struct sockaddr *,
 506			       int addr_len, int flags);
 507static int unix_socketpair(struct socket *, struct socket *);
 508static int unix_accept(struct socket *, struct socket *, int);
 509static int unix_getname(struct socket *, struct sockaddr *, int *, int);
 510static unsigned int unix_poll(struct file *, struct socket *, poll_table *);
 511static unsigned int unix_dgram_poll(struct file *, struct socket *,
 512				    poll_table *);
 513static int unix_ioctl(struct socket *, unsigned int, unsigned long);
 514static int unix_shutdown(struct socket *, int);
 515static int unix_stream_sendmsg(struct kiocb *, struct socket *,
 516			       struct msghdr *, size_t);
 517static int unix_stream_recvmsg(struct kiocb *, struct socket *,
 518			       struct msghdr *, size_t, int);
 519static int unix_dgram_sendmsg(struct kiocb *, struct socket *,
 520			      struct msghdr *, size_t);
 521static int unix_dgram_recvmsg(struct kiocb *, struct socket *,
 522			      struct msghdr *, size_t, int);
 523static int unix_dgram_connect(struct socket *, struct sockaddr *,
 524			      int, int);
 525static int unix_seqpacket_sendmsg(struct kiocb *, struct socket *,
 526				  struct msghdr *, size_t);
 527static int unix_seqpacket_recvmsg(struct kiocb *, struct socket *,
 528				  struct msghdr *, size_t, int);
 529
 530static void unix_set_peek_off(struct sock *sk, int val)
 531{
 532	struct unix_sock *u = unix_sk(sk);
 533
 534	mutex_lock(&u->readlock);
 
 
 535	sk->sk_peek_off = val;
 536	mutex_unlock(&u->readlock);
 
 
 537}
 538
 539
 540static const struct proto_ops unix_stream_ops = {
 541	.family =	PF_UNIX,
 542	.owner =	THIS_MODULE,
 543	.release =	unix_release,
 544	.bind =		unix_bind,
 545	.connect =	unix_stream_connect,
 546	.socketpair =	unix_socketpair,
 547	.accept =	unix_accept,
 548	.getname =	unix_getname,
 549	.poll =		unix_poll,
 550	.ioctl =	unix_ioctl,
 551	.listen =	unix_listen,
 552	.shutdown =	unix_shutdown,
 553	.setsockopt =	sock_no_setsockopt,
 554	.getsockopt =	sock_no_getsockopt,
 555	.sendmsg =	unix_stream_sendmsg,
 556	.recvmsg =	unix_stream_recvmsg,
 557	.mmap =		sock_no_mmap,
 558	.sendpage =	sock_no_sendpage,
 559	.set_peek_off =	unix_set_peek_off,
 560};
 561
 562static const struct proto_ops unix_dgram_ops = {
 563	.family =	PF_UNIX,
 564	.owner =	THIS_MODULE,
 565	.release =	unix_release,
 566	.bind =		unix_bind,
 567	.connect =	unix_dgram_connect,
 568	.socketpair =	unix_socketpair,
 569	.accept =	sock_no_accept,
 570	.getname =	unix_getname,
 571	.poll =		unix_dgram_poll,
 572	.ioctl =	unix_ioctl,
 573	.listen =	sock_no_listen,
 574	.shutdown =	unix_shutdown,
 575	.setsockopt =	sock_no_setsockopt,
 576	.getsockopt =	sock_no_getsockopt,
 577	.sendmsg =	unix_dgram_sendmsg,
 578	.recvmsg =	unix_dgram_recvmsg,
 579	.mmap =		sock_no_mmap,
 580	.sendpage =	sock_no_sendpage,
 581	.set_peek_off =	unix_set_peek_off,
 582};
 583
 584static const struct proto_ops unix_seqpacket_ops = {
 585	.family =	PF_UNIX,
 586	.owner =	THIS_MODULE,
 587	.release =	unix_release,
 588	.bind =		unix_bind,
 589	.connect =	unix_stream_connect,
 590	.socketpair =	unix_socketpair,
 591	.accept =	unix_accept,
 592	.getname =	unix_getname,
 593	.poll =		unix_dgram_poll,
 594	.ioctl =	unix_ioctl,
 595	.listen =	unix_listen,
 596	.shutdown =	unix_shutdown,
 597	.setsockopt =	sock_no_setsockopt,
 598	.getsockopt =	sock_no_getsockopt,
 599	.sendmsg =	unix_seqpacket_sendmsg,
 600	.recvmsg =	unix_seqpacket_recvmsg,
 601	.mmap =		sock_no_mmap,
 602	.sendpage =	sock_no_sendpage,
 603	.set_peek_off =	unix_set_peek_off,
 604};
 605
 606static struct proto unix_proto = {
 607	.name			= "UNIX",
 608	.owner			= THIS_MODULE,
 609	.obj_size		= sizeof(struct unix_sock),
 610};
 611
 612/*
 613 * AF_UNIX sockets do not interact with hardware, hence they
 614 * dont trigger interrupts - so it's safe for them to have
 615 * bh-unsafe locking for their sk_receive_queue.lock. Split off
 616 * this special lock-class by reinitializing the spinlock key:
 617 */
 618static struct lock_class_key af_unix_sk_receive_queue_lock_key;
 619
 620static struct sock *unix_create1(struct net *net, struct socket *sock)
 621{
 622	struct sock *sk = NULL;
 623	struct unix_sock *u;
 624
 625	atomic_long_inc(&unix_nr_socks);
 626	if (atomic_long_read(&unix_nr_socks) > 2 * get_max_files())
 627		goto out;
 628
 629	sk = sk_alloc(net, PF_UNIX, GFP_KERNEL, &unix_proto);
 630	if (!sk)
 631		goto out;
 632
 633	sock_init_data(sock, sk);
 634	lockdep_set_class(&sk->sk_receive_queue.lock,
 635				&af_unix_sk_receive_queue_lock_key);
 636
 637	sk->sk_write_space	= unix_write_space;
 638	sk->sk_max_ack_backlog	= net->unx.sysctl_max_dgram_qlen;
 639	sk->sk_destruct		= unix_sock_destructor;
 640	u	  = unix_sk(sk);
 641	u->path.dentry = NULL;
 642	u->path.mnt = NULL;
 643	spin_lock_init(&u->lock);
 644	atomic_long_set(&u->inflight, 0);
 645	INIT_LIST_HEAD(&u->link);
 646	mutex_init(&u->readlock); /* single task reading lock */
 647	init_waitqueue_head(&u->peer_wait);
 648	unix_insert_socket(unix_sockets_unbound, sk);
 649out:
 650	if (sk == NULL)
 651		atomic_long_dec(&unix_nr_socks);
 652	else {
 653		local_bh_disable();
 654		sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1);
 655		local_bh_enable();
 656	}
 657	return sk;
 658}
 659
 660static int unix_create(struct net *net, struct socket *sock, int protocol,
 661		       int kern)
 662{
 663	if (protocol && protocol != PF_UNIX)
 664		return -EPROTONOSUPPORT;
 665
 666	sock->state = SS_UNCONNECTED;
 667
 668	switch (sock->type) {
 669	case SOCK_STREAM:
 670		sock->ops = &unix_stream_ops;
 671		break;
 672		/*
 673		 *	Believe it or not BSD has AF_UNIX, SOCK_RAW though
 674		 *	nothing uses it.
 675		 */
 676	case SOCK_RAW:
 677		sock->type = SOCK_DGRAM;
 678	case SOCK_DGRAM:
 679		sock->ops = &unix_dgram_ops;
 680		break;
 681	case SOCK_SEQPACKET:
 682		sock->ops = &unix_seqpacket_ops;
 683		break;
 684	default:
 685		return -ESOCKTNOSUPPORT;
 686	}
 687
 688	return unix_create1(net, sock) ? 0 : -ENOMEM;
 689}
 690
 691static int unix_release(struct socket *sock)
 692{
 693	struct sock *sk = sock->sk;
 694
 695	if (!sk)
 696		return 0;
 697
 
 698	sock->sk = NULL;
 699
 700	return unix_release_sock(sk, 0);
 701}
 702
 703static int unix_autobind(struct socket *sock)
 704{
 705	struct sock *sk = sock->sk;
 706	struct net *net = sock_net(sk);
 707	struct unix_sock *u = unix_sk(sk);
 708	static u32 ordernum = 1;
 709	struct unix_address *addr;
 710	int err;
 711	unsigned int retries = 0;
 712
 713	mutex_lock(&u->readlock);
 
 
 714
 715	err = 0;
 716	if (u->addr)
 717		goto out;
 718
 719	err = -ENOMEM;
 720	addr = kzalloc(sizeof(*addr) + sizeof(short) + 16, GFP_KERNEL);
 721	if (!addr)
 722		goto out;
 723
 724	addr->name->sun_family = AF_UNIX;
 725	atomic_set(&addr->refcnt, 1);
 726
 727retry:
 728	addr->len = sprintf(addr->name->sun_path+1, "%05x", ordernum) + 1 + sizeof(short);
 729	addr->hash = unix_hash_fold(csum_partial(addr->name, addr->len, 0));
 730
 731	spin_lock(&unix_table_lock);
 732	ordernum = (ordernum+1)&0xFFFFF;
 733
 734	if (__unix_find_socket_byname(net, addr->name, addr->len, sock->type,
 735				      addr->hash)) {
 736		spin_unlock(&unix_table_lock);
 737		/*
 738		 * __unix_find_socket_byname() may take long time if many names
 739		 * are already in use.
 740		 */
 741		cond_resched();
 742		/* Give up if all names seems to be in use. */
 743		if (retries++ == 0xFFFFF) {
 744			err = -ENOSPC;
 745			kfree(addr);
 746			goto out;
 747		}
 748		goto retry;
 749	}
 750	addr->hash ^= sk->sk_type;
 751
 752	__unix_remove_socket(sk);
 753	u->addr = addr;
 754	__unix_insert_socket(&unix_socket_table[addr->hash], sk);
 755	spin_unlock(&unix_table_lock);
 756	err = 0;
 757
 758out:	mutex_unlock(&u->readlock);
 759	return err;
 760}
 761
 762static struct sock *unix_find_other(struct net *net,
 763				    struct sockaddr_un *sunname, int len,
 764				    int type, unsigned int hash, int *error)
 765{
 766	struct sock *u;
 767	struct path path;
 768	int err = 0;
 769
 770	if (sunname->sun_path[0]) {
 771		struct inode *inode;
 772		err = kern_path(sunname->sun_path, LOOKUP_FOLLOW, &path);
 773		if (err)
 774			goto fail;
 775		inode = path.dentry->d_inode;
 776		err = inode_permission(inode, MAY_WRITE);
 777		if (err)
 778			goto put_fail;
 779
 780		err = -ECONNREFUSED;
 781		if (!S_ISSOCK(inode->i_mode))
 782			goto put_fail;
 783		u = unix_find_socket_byinode(inode);
 784		if (!u)
 785			goto put_fail;
 786
 787		if (u->sk_type == type)
 788			touch_atime(&path);
 789
 790		path_put(&path);
 791
 792		err = -EPROTOTYPE;
 793		if (u->sk_type != type) {
 794			sock_put(u);
 795			goto fail;
 796		}
 797	} else {
 798		err = -ECONNREFUSED;
 799		u = unix_find_socket_byname(net, sunname, len, type, hash);
 800		if (u) {
 801			struct dentry *dentry;
 802			dentry = unix_sk(u)->path.dentry;
 803			if (dentry)
 804				touch_atime(&unix_sk(u)->path);
 805		} else
 806			goto fail;
 807	}
 808	return u;
 809
 810put_fail:
 811	path_put(&path);
 812fail:
 813	*error = err;
 814	return NULL;
 815}
 816
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 817
 818static int unix_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len)
 819{
 820	struct sock *sk = sock->sk;
 821	struct net *net = sock_net(sk);
 822	struct unix_sock *u = unix_sk(sk);
 823	struct sockaddr_un *sunaddr = (struct sockaddr_un *)uaddr;
 824	char *sun_path = sunaddr->sun_path;
 825	struct dentry *dentry = NULL;
 826	struct path path;
 827	int err;
 828	unsigned int hash;
 829	struct unix_address *addr;
 830	struct hlist_head *list;
 831
 832	err = -EINVAL;
 833	if (sunaddr->sun_family != AF_UNIX)
 834		goto out;
 835
 836	if (addr_len == sizeof(short)) {
 837		err = unix_autobind(sock);
 838		goto out;
 839	}
 840
 841	err = unix_mkname(sunaddr, addr_len, &hash);
 842	if (err < 0)
 843		goto out;
 844	addr_len = err;
 845
 846	mutex_lock(&u->readlock);
 
 
 847
 848	err = -EINVAL;
 849	if (u->addr)
 850		goto out_up;
 851
 852	err = -ENOMEM;
 853	addr = kmalloc(sizeof(*addr)+addr_len, GFP_KERNEL);
 854	if (!addr)
 855		goto out_up;
 856
 857	memcpy(addr->name, sunaddr, addr_len);
 858	addr->len = addr_len;
 859	addr->hash = hash ^ sk->sk_type;
 860	atomic_set(&addr->refcnt, 1);
 861
 862	if (sun_path[0]) {
 863		umode_t mode;
 864		err = 0;
 865		/*
 866		 * Get the parent directory, calculate the hash for last
 867		 * component.
 868		 */
 869		dentry = kern_path_create(AT_FDCWD, sun_path, &path, 0);
 870		err = PTR_ERR(dentry);
 871		if (IS_ERR(dentry))
 872			goto out_mknod_parent;
 873
 874		/*
 875		 * All right, let's create it.
 876		 */
 877		mode = S_IFSOCK |
 878		       (SOCK_INODE(sock)->i_mode & ~current_umask());
 879		err = mnt_want_write(path.mnt);
 880		if (err)
 881			goto out_mknod_dput;
 882		err = security_path_mknod(&path, dentry, mode, 0);
 883		if (err)
 884			goto out_mknod_drop_write;
 885		err = vfs_mknod(path.dentry->d_inode, dentry, mode, 0);
 886out_mknod_drop_write:
 887		mnt_drop_write(path.mnt);
 888		if (err)
 889			goto out_mknod_dput;
 890		mutex_unlock(&path.dentry->d_inode->i_mutex);
 891		dput(path.dentry);
 892		path.dentry = dentry;
 893
 894		addr->hash = UNIX_HASH_SIZE;
 895	}
 896
 897	spin_lock(&unix_table_lock);
 898
 899	if (!sun_path[0]) {
 
 900		err = -EADDRINUSE;
 901		if (__unix_find_socket_byname(net, sunaddr, addr_len,
 902					      sk->sk_type, hash)) {
 903			unix_release_addr(addr);
 904			goto out_unlock;
 905		}
 906
 907		list = &unix_socket_table[addr->hash];
 908	} else {
 909		list = &unix_socket_table[dentry->d_inode->i_ino & (UNIX_HASH_SIZE-1)];
 910		u->path = path;
 911	}
 912
 913	err = 0;
 914	__unix_remove_socket(sk);
 915	u->addr = addr;
 916	__unix_insert_socket(list, sk);
 917
 918out_unlock:
 919	spin_unlock(&unix_table_lock);
 920out_up:
 921	mutex_unlock(&u->readlock);
 922out:
 923	return err;
 924
 925out_mknod_dput:
 926	dput(dentry);
 927	mutex_unlock(&path.dentry->d_inode->i_mutex);
 928	path_put(&path);
 929out_mknod_parent:
 930	if (err == -EEXIST)
 931		err = -EADDRINUSE;
 932	unix_release_addr(addr);
 933	goto out_up;
 934}
 935
 936static void unix_state_double_lock(struct sock *sk1, struct sock *sk2)
 937{
 938	if (unlikely(sk1 == sk2) || !sk2) {
 939		unix_state_lock(sk1);
 940		return;
 941	}
 942	if (sk1 < sk2) {
 943		unix_state_lock(sk1);
 944		unix_state_lock_nested(sk2);
 945	} else {
 946		unix_state_lock(sk2);
 947		unix_state_lock_nested(sk1);
 948	}
 949}
 950
 951static void unix_state_double_unlock(struct sock *sk1, struct sock *sk2)
 952{
 953	if (unlikely(sk1 == sk2) || !sk2) {
 954		unix_state_unlock(sk1);
 955		return;
 956	}
 957	unix_state_unlock(sk1);
 958	unix_state_unlock(sk2);
 959}
 960
 961static int unix_dgram_connect(struct socket *sock, struct sockaddr *addr,
 962			      int alen, int flags)
 963{
 964	struct sock *sk = sock->sk;
 965	struct net *net = sock_net(sk);
 966	struct sockaddr_un *sunaddr = (struct sockaddr_un *)addr;
 967	struct sock *other;
 968	unsigned int hash;
 969	int err;
 970
 971	if (addr->sa_family != AF_UNSPEC) {
 972		err = unix_mkname(sunaddr, alen, &hash);
 973		if (err < 0)
 974			goto out;
 975		alen = err;
 976
 977		if (test_bit(SOCK_PASSCRED, &sock->flags) &&
 978		    !unix_sk(sk)->addr && (err = unix_autobind(sock)) != 0)
 979			goto out;
 980
 981restart:
 982		other = unix_find_other(net, sunaddr, alen, sock->type, hash, &err);
 983		if (!other)
 984			goto out;
 985
 986		unix_state_double_lock(sk, other);
 987
 988		/* Apparently VFS overslept socket death. Retry. */
 989		if (sock_flag(other, SOCK_DEAD)) {
 990			unix_state_double_unlock(sk, other);
 991			sock_put(other);
 992			goto restart;
 993		}
 994
 995		err = -EPERM;
 996		if (!unix_may_send(sk, other))
 997			goto out_unlock;
 998
 999		err = security_unix_may_send(sk->sk_socket, other->sk_socket);
1000		if (err)
1001			goto out_unlock;
1002
1003	} else {
1004		/*
1005		 *	1003.1g breaking connected state with AF_UNSPEC
1006		 */
1007		other = NULL;
1008		unix_state_double_lock(sk, other);
1009	}
1010
1011	/*
1012	 * If it was connected, reconnect.
1013	 */
1014	if (unix_peer(sk)) {
1015		struct sock *old_peer = unix_peer(sk);
1016		unix_peer(sk) = other;
1017		unix_state_double_unlock(sk, other);
1018
1019		if (other != old_peer)
1020			unix_dgram_disconnected(sk, old_peer);
1021		sock_put(old_peer);
1022	} else {
1023		unix_peer(sk) = other;
1024		unix_state_double_unlock(sk, other);
1025	}
1026	return 0;
1027
1028out_unlock:
1029	unix_state_double_unlock(sk, other);
1030	sock_put(other);
1031out:
1032	return err;
1033}
1034
1035static long unix_wait_for_peer(struct sock *other, long timeo)
1036{
1037	struct unix_sock *u = unix_sk(other);
1038	int sched;
1039	DEFINE_WAIT(wait);
1040
1041	prepare_to_wait_exclusive(&u->peer_wait, &wait, TASK_INTERRUPTIBLE);
1042
1043	sched = !sock_flag(other, SOCK_DEAD) &&
1044		!(other->sk_shutdown & RCV_SHUTDOWN) &&
1045		unix_recvq_full(other);
1046
1047	unix_state_unlock(other);
1048
1049	if (sched)
1050		timeo = schedule_timeout(timeo);
1051
1052	finish_wait(&u->peer_wait, &wait);
1053	return timeo;
1054}
1055
1056static int unix_stream_connect(struct socket *sock, struct sockaddr *uaddr,
1057			       int addr_len, int flags)
1058{
1059	struct sockaddr_un *sunaddr = (struct sockaddr_un *)uaddr;
1060	struct sock *sk = sock->sk;
1061	struct net *net = sock_net(sk);
1062	struct unix_sock *u = unix_sk(sk), *newu, *otheru;
1063	struct sock *newsk = NULL;
1064	struct sock *other = NULL;
1065	struct sk_buff *skb = NULL;
1066	unsigned int hash;
1067	int st;
1068	int err;
1069	long timeo;
1070
1071	err = unix_mkname(sunaddr, addr_len, &hash);
1072	if (err < 0)
1073		goto out;
1074	addr_len = err;
1075
1076	if (test_bit(SOCK_PASSCRED, &sock->flags) && !u->addr &&
1077	    (err = unix_autobind(sock)) != 0)
1078		goto out;
1079
1080	timeo = sock_sndtimeo(sk, flags & O_NONBLOCK);
1081
1082	/* First of all allocate resources.
1083	   If we will make it after state is locked,
1084	   we will have to recheck all again in any case.
1085	 */
1086
1087	err = -ENOMEM;
1088
1089	/* create new sock for complete connection */
1090	newsk = unix_create1(sock_net(sk), NULL);
1091	if (newsk == NULL)
1092		goto out;
1093
1094	/* Allocate skb for sending to listening sock */
1095	skb = sock_wmalloc(newsk, 1, 0, GFP_KERNEL);
1096	if (skb == NULL)
1097		goto out;
1098
1099restart:
1100	/*  Find listening sock. */
1101	other = unix_find_other(net, sunaddr, addr_len, sk->sk_type, hash, &err);
1102	if (!other)
1103		goto out;
1104
1105	/* Latch state of peer */
1106	unix_state_lock(other);
1107
1108	/* Apparently VFS overslept socket death. Retry. */
1109	if (sock_flag(other, SOCK_DEAD)) {
1110		unix_state_unlock(other);
1111		sock_put(other);
1112		goto restart;
1113	}
1114
1115	err = -ECONNREFUSED;
1116	if (other->sk_state != TCP_LISTEN)
1117		goto out_unlock;
1118	if (other->sk_shutdown & RCV_SHUTDOWN)
1119		goto out_unlock;
1120
1121	if (unix_recvq_full(other)) {
1122		err = -EAGAIN;
1123		if (!timeo)
1124			goto out_unlock;
1125
1126		timeo = unix_wait_for_peer(other, timeo);
1127
1128		err = sock_intr_errno(timeo);
1129		if (signal_pending(current))
1130			goto out;
1131		sock_put(other);
1132		goto restart;
1133	}
1134
1135	/* Latch our state.
1136
1137	   It is tricky place. We need to grab our state lock and cannot
1138	   drop lock on peer. It is dangerous because deadlock is
1139	   possible. Connect to self case and simultaneous
1140	   attempt to connect are eliminated by checking socket
1141	   state. other is TCP_LISTEN, if sk is TCP_LISTEN we
1142	   check this before attempt to grab lock.
1143
1144	   Well, and we have to recheck the state after socket locked.
1145	 */
1146	st = sk->sk_state;
1147
1148	switch (st) {
1149	case TCP_CLOSE:
1150		/* This is ok... continue with connect */
1151		break;
1152	case TCP_ESTABLISHED:
1153		/* Socket is already connected */
1154		err = -EISCONN;
1155		goto out_unlock;
1156	default:
1157		err = -EINVAL;
1158		goto out_unlock;
1159	}
1160
1161	unix_state_lock_nested(sk);
1162
1163	if (sk->sk_state != st) {
1164		unix_state_unlock(sk);
1165		unix_state_unlock(other);
1166		sock_put(other);
1167		goto restart;
1168	}
1169
1170	err = security_unix_stream_connect(sk, other, newsk);
1171	if (err) {
1172		unix_state_unlock(sk);
1173		goto out_unlock;
1174	}
1175
1176	/* The way is open! Fastly set all the necessary fields... */
1177
1178	sock_hold(sk);
1179	unix_peer(newsk)	= sk;
1180	newsk->sk_state		= TCP_ESTABLISHED;
1181	newsk->sk_type		= sk->sk_type;
1182	init_peercred(newsk);
1183	newu = unix_sk(newsk);
1184	RCU_INIT_POINTER(newsk->sk_wq, &newu->peer_wq);
1185	otheru = unix_sk(other);
1186
1187	/* copy address information from listening to new sock*/
1188	if (otheru->addr) {
1189		atomic_inc(&otheru->addr->refcnt);
1190		newu->addr = otheru->addr;
1191	}
1192	if (otheru->path.dentry) {
1193		path_get(&otheru->path);
1194		newu->path = otheru->path;
1195	}
1196
1197	/* Set credentials */
1198	copy_peercred(sk, other);
1199
1200	sock->state	= SS_CONNECTED;
1201	sk->sk_state	= TCP_ESTABLISHED;
1202	sock_hold(newsk);
1203
1204	smp_mb__after_atomic_inc();	/* sock_hold() does an atomic_inc() */
1205	unix_peer(sk)	= newsk;
1206
1207	unix_state_unlock(sk);
1208
1209	/* take ten and and send info to listening sock */
1210	spin_lock(&other->sk_receive_queue.lock);
1211	__skb_queue_tail(&other->sk_receive_queue, skb);
1212	spin_unlock(&other->sk_receive_queue.lock);
1213	unix_state_unlock(other);
1214	other->sk_data_ready(other, 0);
1215	sock_put(other);
1216	return 0;
1217
1218out_unlock:
1219	if (other)
1220		unix_state_unlock(other);
1221
1222out:
1223	kfree_skb(skb);
1224	if (newsk)
1225		unix_release_sock(newsk, 0);
1226	if (other)
1227		sock_put(other);
1228	return err;
1229}
1230
1231static int unix_socketpair(struct socket *socka, struct socket *sockb)
1232{
1233	struct sock *ska = socka->sk, *skb = sockb->sk;
1234
1235	/* Join our sockets back to back */
1236	sock_hold(ska);
1237	sock_hold(skb);
1238	unix_peer(ska) = skb;
1239	unix_peer(skb) = ska;
1240	init_peercred(ska);
1241	init_peercred(skb);
1242
1243	if (ska->sk_type != SOCK_DGRAM) {
1244		ska->sk_state = TCP_ESTABLISHED;
1245		skb->sk_state = TCP_ESTABLISHED;
1246		socka->state  = SS_CONNECTED;
1247		sockb->state  = SS_CONNECTED;
1248	}
1249	return 0;
1250}
1251
 
 
 
 
 
 
 
 
 
1252static int unix_accept(struct socket *sock, struct socket *newsock, int flags)
1253{
1254	struct sock *sk = sock->sk;
1255	struct sock *tsk;
1256	struct sk_buff *skb;
1257	int err;
1258
1259	err = -EOPNOTSUPP;
1260	if (sock->type != SOCK_STREAM && sock->type != SOCK_SEQPACKET)
1261		goto out;
1262
1263	err = -EINVAL;
1264	if (sk->sk_state != TCP_LISTEN)
1265		goto out;
1266
1267	/* If socket state is TCP_LISTEN it cannot change (for now...),
1268	 * so that no locks are necessary.
1269	 */
1270
1271	skb = skb_recv_datagram(sk, 0, flags&O_NONBLOCK, &err);
1272	if (!skb) {
1273		/* This means receive shutdown. */
1274		if (err == 0)
1275			err = -EINVAL;
1276		goto out;
1277	}
1278
1279	tsk = skb->sk;
1280	skb_free_datagram(sk, skb);
1281	wake_up_interruptible(&unix_sk(sk)->peer_wait);
1282
1283	/* attach accepted sock to socket */
1284	unix_state_lock(tsk);
1285	newsock->state = SS_CONNECTED;
 
1286	sock_graft(tsk, newsock);
1287	unix_state_unlock(tsk);
1288	return 0;
1289
1290out:
1291	return err;
1292}
1293
1294
1295static int unix_getname(struct socket *sock, struct sockaddr *uaddr, int *uaddr_len, int peer)
1296{
1297	struct sock *sk = sock->sk;
1298	struct unix_sock *u;
1299	DECLARE_SOCKADDR(struct sockaddr_un *, sunaddr, uaddr);
1300	int err = 0;
1301
1302	if (peer) {
1303		sk = unix_peer_get(sk);
1304
1305		err = -ENOTCONN;
1306		if (!sk)
1307			goto out;
1308		err = 0;
1309	} else {
1310		sock_hold(sk);
1311	}
1312
1313	u = unix_sk(sk);
1314	unix_state_lock(sk);
1315	if (!u->addr) {
1316		sunaddr->sun_family = AF_UNIX;
1317		sunaddr->sun_path[0] = 0;
1318		*uaddr_len = sizeof(short);
1319	} else {
1320		struct unix_address *addr = u->addr;
1321
1322		*uaddr_len = addr->len;
1323		memcpy(sunaddr, addr->name, *uaddr_len);
1324	}
1325	unix_state_unlock(sk);
1326	sock_put(sk);
1327out:
1328	return err;
1329}
1330
1331static void unix_detach_fds(struct scm_cookie *scm, struct sk_buff *skb)
1332{
1333	int i;
1334
1335	scm->fp = UNIXCB(skb).fp;
1336	UNIXCB(skb).fp = NULL;
1337
1338	for (i = scm->fp->count-1; i >= 0; i--)
1339		unix_notinflight(scm->fp->fp[i]);
1340}
1341
1342static void unix_destruct_scm(struct sk_buff *skb)
1343{
1344	struct scm_cookie scm;
1345	memset(&scm, 0, sizeof(scm));
1346	scm.pid  = UNIXCB(skb).pid;
1347	scm.cred = UNIXCB(skb).cred;
1348	if (UNIXCB(skb).fp)
1349		unix_detach_fds(&scm, skb);
1350
1351	/* Alas, it calls VFS */
1352	/* So fscking what? fput() had been SMP-safe since the last Summer */
1353	scm_destroy(&scm);
1354	sock_wfree(skb);
1355}
1356
1357#define MAX_RECURSION_LEVEL 4
1358
1359static int unix_attach_fds(struct scm_cookie *scm, struct sk_buff *skb)
1360{
1361	int i;
1362	unsigned char max_level = 0;
1363	int unix_sock_count = 0;
1364
1365	for (i = scm->fp->count - 1; i >= 0; i--) {
1366		struct sock *sk = unix_get_socket(scm->fp->fp[i]);
1367
1368		if (sk) {
1369			unix_sock_count++;
1370			max_level = max(max_level,
1371					unix_sk(sk)->recursion_level);
1372		}
1373	}
1374	if (unlikely(max_level > MAX_RECURSION_LEVEL))
1375		return -ETOOMANYREFS;
1376
1377	/*
1378	 * Need to duplicate file references for the sake of garbage
1379	 * collection.  Otherwise a socket in the fps might become a
1380	 * candidate for GC while the skb is not yet queued.
1381	 */
1382	UNIXCB(skb).fp = scm_fp_dup(scm->fp);
1383	if (!UNIXCB(skb).fp)
1384		return -ENOMEM;
1385
1386	if (unix_sock_count) {
1387		for (i = scm->fp->count - 1; i >= 0; i--)
1388			unix_inflight(scm->fp->fp[i]);
1389	}
1390	return max_level;
1391}
1392
1393static int unix_scm_to_skb(struct scm_cookie *scm, struct sk_buff *skb, bool send_fds)
1394{
1395	int err = 0;
1396
1397	UNIXCB(skb).pid  = get_pid(scm->pid);
1398	if (scm->cred)
1399		UNIXCB(skb).cred = get_cred(scm->cred);
1400	UNIXCB(skb).fp = NULL;
1401	if (scm->fp && send_fds)
1402		err = unix_attach_fds(scm, skb);
1403
1404	skb->destructor = unix_destruct_scm;
1405	return err;
1406}
1407
1408/*
1409 * Some apps rely on write() giving SCM_CREDENTIALS
1410 * We include credentials if source or destination socket
1411 * asserted SOCK_PASSCRED.
1412 */
1413static void maybe_add_creds(struct sk_buff *skb, const struct socket *sock,
1414			    const struct sock *other)
1415{
1416	if (UNIXCB(skb).cred)
1417		return;
1418	if (test_bit(SOCK_PASSCRED, &sock->flags) ||
1419	    !other->sk_socket ||
1420	    test_bit(SOCK_PASSCRED, &other->sk_socket->flags)) {
1421		UNIXCB(skb).pid  = get_pid(task_tgid(current));
1422		UNIXCB(skb).cred = get_current_cred();
1423	}
1424}
1425
1426/*
1427 *	Send AF_UNIX data.
1428 */
1429
1430static int unix_dgram_sendmsg(struct kiocb *kiocb, struct socket *sock,
1431			      struct msghdr *msg, size_t len)
1432{
1433	struct sock_iocb *siocb = kiocb_to_siocb(kiocb);
1434	struct sock *sk = sock->sk;
1435	struct net *net = sock_net(sk);
1436	struct unix_sock *u = unix_sk(sk);
1437	struct sockaddr_un *sunaddr = msg->msg_name;
1438	struct sock *other = NULL;
1439	int namelen = 0; /* fake GCC */
1440	int err;
1441	unsigned int hash;
1442	struct sk_buff *skb;
1443	long timeo;
1444	struct scm_cookie tmp_scm;
1445	int max_level;
1446	int data_len = 0;
1447
1448	if (NULL == siocb->scm)
1449		siocb->scm = &tmp_scm;
1450	wait_for_unix_gc();
1451	err = scm_send(sock, msg, siocb->scm, false);
1452	if (err < 0)
1453		return err;
1454
1455	err = -EOPNOTSUPP;
1456	if (msg->msg_flags&MSG_OOB)
1457		goto out;
1458
1459	if (msg->msg_namelen) {
1460		err = unix_mkname(sunaddr, msg->msg_namelen, &hash);
1461		if (err < 0)
1462			goto out;
1463		namelen = err;
1464	} else {
1465		sunaddr = NULL;
1466		err = -ENOTCONN;
1467		other = unix_peer_get(sk);
1468		if (!other)
1469			goto out;
1470	}
1471
1472	if (test_bit(SOCK_PASSCRED, &sock->flags) && !u->addr
1473	    && (err = unix_autobind(sock)) != 0)
1474		goto out;
1475
1476	err = -EMSGSIZE;
1477	if (len > sk->sk_sndbuf - 32)
1478		goto out;
1479
1480	if (len > SKB_MAX_ALLOC)
1481		data_len = min_t(size_t,
1482				 len - SKB_MAX_ALLOC,
1483				 MAX_SKB_FRAGS * PAGE_SIZE);
1484
1485	skb = sock_alloc_send_pskb(sk, len - data_len, data_len,
1486				   msg->msg_flags & MSG_DONTWAIT, &err);
 
1487	if (skb == NULL)
1488		goto out;
1489
1490	err = unix_scm_to_skb(siocb->scm, skb, true);
1491	if (err < 0)
1492		goto out_free;
1493	max_level = err + 1;
1494	unix_get_secdata(siocb->scm, skb);
1495
1496	skb_put(skb, len - data_len);
1497	skb->data_len = data_len;
1498	skb->len = len;
1499	err = skb_copy_datagram_from_iovec(skb, 0, msg->msg_iov, 0, len);
1500	if (err)
1501		goto out_free;
1502
1503	timeo = sock_sndtimeo(sk, msg->msg_flags & MSG_DONTWAIT);
1504
1505restart:
1506	if (!other) {
1507		err = -ECONNRESET;
1508		if (sunaddr == NULL)
1509			goto out_free;
1510
1511		other = unix_find_other(net, sunaddr, namelen, sk->sk_type,
1512					hash, &err);
1513		if (other == NULL)
1514			goto out_free;
1515	}
1516
1517	if (sk_filter(other, skb) < 0) {
1518		/* Toss the packet but do not return any error to the sender */
1519		err = len;
1520		goto out_free;
1521	}
1522
1523	unix_state_lock(other);
1524	err = -EPERM;
1525	if (!unix_may_send(sk, other))
1526		goto out_unlock;
1527
1528	if (sock_flag(other, SOCK_DEAD)) {
1529		/*
1530		 *	Check with 1003.1g - what should
1531		 *	datagram error
1532		 */
1533		unix_state_unlock(other);
1534		sock_put(other);
1535
1536		err = 0;
1537		unix_state_lock(sk);
1538		if (unix_peer(sk) == other) {
1539			unix_peer(sk) = NULL;
1540			unix_state_unlock(sk);
1541
1542			unix_dgram_disconnected(sk, other);
1543			sock_put(other);
1544			err = -ECONNREFUSED;
1545		} else {
1546			unix_state_unlock(sk);
1547		}
1548
1549		other = NULL;
1550		if (err)
1551			goto out_free;
1552		goto restart;
1553	}
1554
1555	err = -EPIPE;
1556	if (other->sk_shutdown & RCV_SHUTDOWN)
1557		goto out_unlock;
1558
1559	if (sk->sk_type != SOCK_SEQPACKET) {
1560		err = security_unix_may_send(sk->sk_socket, other->sk_socket);
1561		if (err)
1562			goto out_unlock;
1563	}
1564
1565	if (unix_peer(other) != sk && unix_recvq_full(other)) {
1566		if (!timeo) {
1567			err = -EAGAIN;
1568			goto out_unlock;
1569		}
1570
1571		timeo = unix_wait_for_peer(other, timeo);
1572
1573		err = sock_intr_errno(timeo);
1574		if (signal_pending(current))
1575			goto out_free;
1576
1577		goto restart;
1578	}
1579
1580	if (sock_flag(other, SOCK_RCVTSTAMP))
1581		__net_timestamp(skb);
1582	maybe_add_creds(skb, sock, other);
1583	skb_queue_tail(&other->sk_receive_queue, skb);
1584	if (max_level > unix_sk(other)->recursion_level)
1585		unix_sk(other)->recursion_level = max_level;
1586	unix_state_unlock(other);
1587	other->sk_data_ready(other, len);
1588	sock_put(other);
1589	scm_destroy(siocb->scm);
1590	return len;
1591
1592out_unlock:
1593	unix_state_unlock(other);
1594out_free:
1595	kfree_skb(skb);
1596out:
1597	if (other)
1598		sock_put(other);
1599	scm_destroy(siocb->scm);
1600	return err;
1601}
1602
 
 
 
 
1603
1604static int unix_stream_sendmsg(struct kiocb *kiocb, struct socket *sock,
1605			       struct msghdr *msg, size_t len)
1606{
1607	struct sock_iocb *siocb = kiocb_to_siocb(kiocb);
1608	struct sock *sk = sock->sk;
1609	struct sock *other = NULL;
1610	int err, size;
1611	struct sk_buff *skb;
1612	int sent = 0;
1613	struct scm_cookie tmp_scm;
1614	bool fds_sent = false;
1615	int max_level;
 
1616
1617	if (NULL == siocb->scm)
1618		siocb->scm = &tmp_scm;
1619	wait_for_unix_gc();
1620	err = scm_send(sock, msg, siocb->scm, false);
1621	if (err < 0)
1622		return err;
1623
1624	err = -EOPNOTSUPP;
1625	if (msg->msg_flags&MSG_OOB)
1626		goto out_err;
1627
1628	if (msg->msg_namelen) {
1629		err = sk->sk_state == TCP_ESTABLISHED ? -EISCONN : -EOPNOTSUPP;
1630		goto out_err;
1631	} else {
1632		err = -ENOTCONN;
1633		other = unix_peer(sk);
1634		if (!other)
1635			goto out_err;
1636	}
1637
1638	if (sk->sk_shutdown & SEND_SHUTDOWN)
1639		goto pipe_err;
1640
1641	while (sent < len) {
1642		/*
1643		 *	Optimisation for the fact that under 0.01% of X
1644		 *	messages typically need breaking up.
1645		 */
1646
1647		size = len-sent;
1648
1649		/* Keep two messages in the pipe so it schedules better */
1650		if (size > ((sk->sk_sndbuf >> 1) - 64))
1651			size = (sk->sk_sndbuf >> 1) - 64;
1652
1653		if (size > SKB_MAX_ALLOC)
1654			size = SKB_MAX_ALLOC;
1655
1656		/*
1657		 *	Grab a buffer
1658		 */
1659
1660		skb = sock_alloc_send_skb(sk, size, msg->msg_flags&MSG_DONTWAIT,
1661					  &err);
1662
1663		if (skb == NULL)
 
 
 
1664			goto out_err;
1665
1666		/*
1667		 *	If you pass two values to the sock_alloc_send_skb
1668		 *	it tries to grab the large buffer with GFP_NOFS
1669		 *	(which can fail easily), and if it fails grab the
1670		 *	fallback size buffer which is under a page and will
1671		 *	succeed. [Alan]
1672		 */
1673		size = min_t(int, size, skb_tailroom(skb));
1674
1675
1676		/* Only send the fds in the first buffer */
1677		err = unix_scm_to_skb(siocb->scm, skb, !fds_sent);
1678		if (err < 0) {
1679			kfree_skb(skb);
1680			goto out_err;
1681		}
1682		max_level = err + 1;
1683		fds_sent = true;
1684
1685		err = memcpy_fromiovec(skb_put(skb, size), msg->msg_iov, size);
 
 
 
 
1686		if (err) {
1687			kfree_skb(skb);
1688			goto out_err;
1689		}
1690
1691		unix_state_lock(other);
1692
1693		if (sock_flag(other, SOCK_DEAD) ||
1694		    (other->sk_shutdown & RCV_SHUTDOWN))
1695			goto pipe_err_free;
1696
1697		maybe_add_creds(skb, sock, other);
1698		skb_queue_tail(&other->sk_receive_queue, skb);
1699		if (max_level > unix_sk(other)->recursion_level)
1700			unix_sk(other)->recursion_level = max_level;
1701		unix_state_unlock(other);
1702		other->sk_data_ready(other, size);
1703		sent += size;
1704	}
1705
1706	scm_destroy(siocb->scm);
1707	siocb->scm = NULL;
1708
1709	return sent;
1710
1711pipe_err_free:
1712	unix_state_unlock(other);
1713	kfree_skb(skb);
1714pipe_err:
1715	if (sent == 0 && !(msg->msg_flags&MSG_NOSIGNAL))
1716		send_sig(SIGPIPE, current, 0);
1717	err = -EPIPE;
1718out_err:
1719	scm_destroy(siocb->scm);
1720	siocb->scm = NULL;
1721	return sent ? : err;
1722}
1723
1724static int unix_seqpacket_sendmsg(struct kiocb *kiocb, struct socket *sock,
1725				  struct msghdr *msg, size_t len)
1726{
1727	int err;
1728	struct sock *sk = sock->sk;
1729
1730	err = sock_error(sk);
1731	if (err)
1732		return err;
1733
1734	if (sk->sk_state != TCP_ESTABLISHED)
1735		return -ENOTCONN;
1736
1737	if (msg->msg_namelen)
1738		msg->msg_namelen = 0;
1739
1740	return unix_dgram_sendmsg(kiocb, sock, msg, len);
1741}
1742
1743static int unix_seqpacket_recvmsg(struct kiocb *iocb, struct socket *sock,
1744			      struct msghdr *msg, size_t size,
1745			      int flags)
1746{
1747	struct sock *sk = sock->sk;
1748
1749	if (sk->sk_state != TCP_ESTABLISHED)
1750		return -ENOTCONN;
1751
1752	return unix_dgram_recvmsg(iocb, sock, msg, size, flags);
1753}
1754
1755static void unix_copy_addr(struct msghdr *msg, struct sock *sk)
1756{
1757	struct unix_sock *u = unix_sk(sk);
1758
1759	msg->msg_namelen = 0;
1760	if (u->addr) {
1761		msg->msg_namelen = u->addr->len;
1762		memcpy(msg->msg_name, u->addr->name, u->addr->len);
1763	}
1764}
1765
1766static int unix_dgram_recvmsg(struct kiocb *iocb, struct socket *sock,
1767			      struct msghdr *msg, size_t size,
1768			      int flags)
1769{
1770	struct sock_iocb *siocb = kiocb_to_siocb(iocb);
1771	struct scm_cookie tmp_scm;
1772	struct sock *sk = sock->sk;
1773	struct unix_sock *u = unix_sk(sk);
1774	int noblock = flags & MSG_DONTWAIT;
1775	struct sk_buff *skb;
1776	int err;
1777	int peeked, skip;
1778
1779	err = -EOPNOTSUPP;
1780	if (flags&MSG_OOB)
1781		goto out;
1782
1783	msg->msg_namelen = 0;
1784
1785	err = mutex_lock_interruptible(&u->readlock);
1786	if (err) {
1787		err = sock_intr_errno(sock_rcvtimeo(sk, noblock));
 
 
 
1788		goto out;
1789	}
1790
1791	skip = sk_peek_offset(sk, flags);
1792
1793	skb = __skb_recv_datagram(sk, flags, &peeked, &skip, &err);
1794	if (!skb) {
1795		unix_state_lock(sk);
1796		/* Signal EOF on disconnected non-blocking SEQPACKET socket. */
1797		if (sk->sk_type == SOCK_SEQPACKET && err == -EAGAIN &&
1798		    (sk->sk_shutdown & RCV_SHUTDOWN))
1799			err = 0;
1800		unix_state_unlock(sk);
1801		goto out_unlock;
1802	}
1803
1804	wake_up_interruptible_sync_poll(&u->peer_wait,
1805					POLLOUT | POLLWRNORM | POLLWRBAND);
1806
1807	if (msg->msg_name)
1808		unix_copy_addr(msg, skb->sk);
1809
1810	if (size > skb->len - skip)
1811		size = skb->len - skip;
1812	else if (size < skb->len - skip)
1813		msg->msg_flags |= MSG_TRUNC;
1814
1815	err = skb_copy_datagram_iovec(skb, skip, msg->msg_iov, size);
1816	if (err)
1817		goto out_free;
1818
1819	if (sock_flag(sk, SOCK_RCVTSTAMP))
1820		__sock_recv_timestamp(msg, sk, skb);
1821
1822	if (!siocb->scm) {
1823		siocb->scm = &tmp_scm;
1824		memset(&tmp_scm, 0, sizeof(tmp_scm));
1825	}
1826	scm_set_cred(siocb->scm, UNIXCB(skb).pid, UNIXCB(skb).cred);
1827	unix_set_secdata(siocb->scm, skb);
1828
1829	if (!(flags & MSG_PEEK)) {
1830		if (UNIXCB(skb).fp)
1831			unix_detach_fds(siocb->scm, skb);
1832
1833		sk_peek_offset_bwd(sk, skb->len);
1834	} else {
1835		/* It is questionable: on PEEK we could:
1836		   - do not return fds - good, but too simple 8)
1837		   - return fds, and do not return them on read (old strategy,
1838		     apparently wrong)
1839		   - clone fds (I chose it for now, it is the most universal
1840		     solution)
1841
1842		   POSIX 1003.1g does not actually define this clearly
1843		   at all. POSIX 1003.1g doesn't define a lot of things
1844		   clearly however!
1845
1846		*/
1847
1848		sk_peek_offset_fwd(sk, size);
1849
1850		if (UNIXCB(skb).fp)
1851			siocb->scm->fp = scm_fp_dup(UNIXCB(skb).fp);
1852	}
1853	err = (flags & MSG_TRUNC) ? skb->len - skip : size;
1854
1855	scm_recv(sock, msg, siocb->scm, flags);
1856
1857out_free:
1858	skb_free_datagram(sk, skb);
1859out_unlock:
1860	mutex_unlock(&u->readlock);
1861out:
1862	return err;
1863}
1864
1865/*
1866 *	Sleep until data has arrive. But check for races..
1867 */
1868
1869static long unix_stream_data_wait(struct sock *sk, long timeo)
1870{
1871	DEFINE_WAIT(wait);
1872
1873	unix_state_lock(sk);
1874
1875	for (;;) {
1876		prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
1877
1878		if (!skb_queue_empty(&sk->sk_receive_queue) ||
1879		    sk->sk_err ||
1880		    (sk->sk_shutdown & RCV_SHUTDOWN) ||
1881		    signal_pending(current) ||
1882		    !timeo)
1883			break;
1884
1885		set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1886		unix_state_unlock(sk);
1887		timeo = schedule_timeout(timeo);
1888		unix_state_lock(sk);
1889		clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1890	}
1891
1892	finish_wait(sk_sleep(sk), &wait);
1893	unix_state_unlock(sk);
1894	return timeo;
1895}
1896
1897
 
 
 
1898
1899static int unix_stream_recvmsg(struct kiocb *iocb, struct socket *sock,
1900			       struct msghdr *msg, size_t size,
1901			       int flags)
1902{
1903	struct sock_iocb *siocb = kiocb_to_siocb(iocb);
1904	struct scm_cookie tmp_scm;
1905	struct sock *sk = sock->sk;
1906	struct unix_sock *u = unix_sk(sk);
1907	struct sockaddr_un *sunaddr = msg->msg_name;
1908	int copied = 0;
 
1909	int check_creds = 0;
1910	int target;
1911	int err = 0;
1912	long timeo;
1913	int skip;
1914
1915	err = -EINVAL;
1916	if (sk->sk_state != TCP_ESTABLISHED)
1917		goto out;
1918
1919	err = -EOPNOTSUPP;
1920	if (flags&MSG_OOB)
1921		goto out;
1922
1923	target = sock_rcvlowat(sk, flags&MSG_WAITALL, size);
1924	timeo = sock_rcvtimeo(sk, flags&MSG_DONTWAIT);
1925
1926	msg->msg_namelen = 0;
1927
1928	/* Lock the socket to prevent queue disordering
1929	 * while sleeps in memcpy_tomsg
1930	 */
1931
1932	if (!siocb->scm) {
1933		siocb->scm = &tmp_scm;
1934		memset(&tmp_scm, 0, sizeof(tmp_scm));
1935	}
1936
1937	err = mutex_lock_interruptible(&u->readlock);
1938	if (err) {
1939		err = sock_intr_errno(timeo);
 
 
 
1940		goto out;
1941	}
1942
1943	skip = sk_peek_offset(sk, flags);
1944
1945	do {
1946		int chunk;
1947		struct sk_buff *skb;
1948
1949		unix_state_lock(sk);
1950		skb = skb_peek(&sk->sk_receive_queue);
1951again:
1952		if (skb == NULL) {
1953			unix_sk(sk)->recursion_level = 0;
1954			if (copied >= target)
1955				goto unlock;
1956
1957			/*
1958			 *	POSIX 1003.1g mandates this order.
1959			 */
1960
1961			err = sock_error(sk);
1962			if (err)
1963				goto unlock;
1964			if (sk->sk_shutdown & RCV_SHUTDOWN)
1965				goto unlock;
1966
1967			unix_state_unlock(sk);
1968			err = -EAGAIN;
1969			if (!timeo)
1970				break;
1971			mutex_unlock(&u->readlock);
1972
1973			timeo = unix_stream_data_wait(sk, timeo);
1974
1975			if (signal_pending(current)
1976			    ||  mutex_lock_interruptible(&u->readlock)) {
1977				err = sock_intr_errno(timeo);
1978				goto out;
1979			}
1980
1981			continue;
1982 unlock:
1983			unix_state_unlock(sk);
1984			break;
1985		}
1986
1987		if (skip >= skb->len) {
1988			skip -= skb->len;
 
 
1989			skb = skb_peek_next(skb, &sk->sk_receive_queue);
1990			goto again;
 
1991		}
1992
1993		unix_state_unlock(sk);
1994
1995		if (check_creds) {
1996			/* Never glue messages from different writers */
1997			if ((UNIXCB(skb).pid  != siocb->scm->pid) ||
1998			    (UNIXCB(skb).cred != siocb->scm->cred))
 
1999				break;
2000		} else {
2001			/* Copy credentials */
2002			scm_set_cred(siocb->scm, UNIXCB(skb).pid, UNIXCB(skb).cred);
2003			check_creds = 1;
2004		}
2005
2006		/* Copy address just once */
2007		if (sunaddr) {
2008			unix_copy_addr(msg, skb->sk);
2009			sunaddr = NULL;
2010		}
2011
2012		chunk = min_t(unsigned int, skb->len - skip, size);
2013		if (memcpy_toiovec(msg->msg_iov, skb->data + skip, chunk)) {
 
2014			if (copied == 0)
2015				copied = -EFAULT;
2016			break;
2017		}
2018		copied += chunk;
2019		size -= chunk;
2020
2021		/* Mark read part of skb as used */
2022		if (!(flags & MSG_PEEK)) {
2023			skb_pull(skb, chunk);
2024
2025			sk_peek_offset_bwd(sk, chunk);
2026
2027			if (UNIXCB(skb).fp)
2028				unix_detach_fds(siocb->scm, skb);
2029
2030			if (skb->len)
2031				break;
2032
2033			skb_unlink(skb, &sk->sk_receive_queue);
2034			consume_skb(skb);
2035
2036			if (siocb->scm->fp)
2037				break;
2038		} else {
2039			/* It is questionable, see note in unix_dgram_recvmsg.
2040			 */
2041			if (UNIXCB(skb).fp)
2042				siocb->scm->fp = scm_fp_dup(UNIXCB(skb).fp);
2043
2044			sk_peek_offset_fwd(sk, chunk);
2045
2046			break;
2047		}
2048	} while (size);
2049
2050	mutex_unlock(&u->readlock);
2051	scm_recv(sock, msg, siocb->scm, flags);
2052out:
2053	return copied ? : err;
2054}
2055
2056static int unix_shutdown(struct socket *sock, int mode)
2057{
2058	struct sock *sk = sock->sk;
2059	struct sock *other;
2060
2061	mode = (mode+1)&(RCV_SHUTDOWN|SEND_SHUTDOWN);
2062
2063	if (!mode)
2064		return 0;
 
 
 
 
2065
2066	unix_state_lock(sk);
2067	sk->sk_shutdown |= mode;
2068	other = unix_peer(sk);
2069	if (other)
2070		sock_hold(other);
2071	unix_state_unlock(sk);
2072	sk->sk_state_change(sk);
2073
2074	if (other &&
2075		(sk->sk_type == SOCK_STREAM || sk->sk_type == SOCK_SEQPACKET)) {
2076
2077		int peer_mode = 0;
2078
2079		if (mode&RCV_SHUTDOWN)
2080			peer_mode |= SEND_SHUTDOWN;
2081		if (mode&SEND_SHUTDOWN)
2082			peer_mode |= RCV_SHUTDOWN;
2083		unix_state_lock(other);
2084		other->sk_shutdown |= peer_mode;
2085		unix_state_unlock(other);
2086		other->sk_state_change(other);
2087		if (peer_mode == SHUTDOWN_MASK)
2088			sk_wake_async(other, SOCK_WAKE_WAITD, POLL_HUP);
2089		else if (peer_mode & RCV_SHUTDOWN)
2090			sk_wake_async(other, SOCK_WAKE_WAITD, POLL_IN);
2091	}
2092	if (other)
2093		sock_put(other);
2094
2095	return 0;
2096}
2097
2098long unix_inq_len(struct sock *sk)
2099{
2100	struct sk_buff *skb;
2101	long amount = 0;
2102
2103	if (sk->sk_state == TCP_LISTEN)
2104		return -EINVAL;
2105
2106	spin_lock(&sk->sk_receive_queue.lock);
2107	if (sk->sk_type == SOCK_STREAM ||
2108	    sk->sk_type == SOCK_SEQPACKET) {
2109		skb_queue_walk(&sk->sk_receive_queue, skb)
2110			amount += skb->len;
2111	} else {
2112		skb = skb_peek(&sk->sk_receive_queue);
2113		if (skb)
2114			amount = skb->len;
2115	}
2116	spin_unlock(&sk->sk_receive_queue.lock);
2117
2118	return amount;
2119}
2120EXPORT_SYMBOL_GPL(unix_inq_len);
2121
2122long unix_outq_len(struct sock *sk)
2123{
2124	return sk_wmem_alloc_get(sk);
2125}
2126EXPORT_SYMBOL_GPL(unix_outq_len);
2127
2128static int unix_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg)
2129{
2130	struct sock *sk = sock->sk;
2131	long amount = 0;
2132	int err;
2133
2134	switch (cmd) {
2135	case SIOCOUTQ:
2136		amount = unix_outq_len(sk);
2137		err = put_user(amount, (int __user *)arg);
2138		break;
2139	case SIOCINQ:
2140		amount = unix_inq_len(sk);
2141		if (amount < 0)
2142			err = amount;
2143		else
2144			err = put_user(amount, (int __user *)arg);
2145		break;
2146	default:
2147		err = -ENOIOCTLCMD;
2148		break;
2149	}
2150	return err;
2151}
2152
2153static unsigned int unix_poll(struct file *file, struct socket *sock, poll_table *wait)
2154{
2155	struct sock *sk = sock->sk;
2156	unsigned int mask;
2157
2158	sock_poll_wait(file, sk_sleep(sk), wait);
2159	mask = 0;
2160
2161	/* exceptional events? */
2162	if (sk->sk_err)
2163		mask |= POLLERR;
2164	if (sk->sk_shutdown == SHUTDOWN_MASK)
2165		mask |= POLLHUP;
2166	if (sk->sk_shutdown & RCV_SHUTDOWN)
2167		mask |= POLLRDHUP | POLLIN | POLLRDNORM;
2168
2169	/* readable? */
2170	if (!skb_queue_empty(&sk->sk_receive_queue))
2171		mask |= POLLIN | POLLRDNORM;
2172
2173	/* Connection-based need to check for termination and startup */
2174	if ((sk->sk_type == SOCK_STREAM || sk->sk_type == SOCK_SEQPACKET) &&
2175	    sk->sk_state == TCP_CLOSE)
2176		mask |= POLLHUP;
2177
2178	/*
2179	 * we set writable also when the other side has shut down the
2180	 * connection. This prevents stuck sockets.
2181	 */
2182	if (unix_writable(sk))
2183		mask |= POLLOUT | POLLWRNORM | POLLWRBAND;
2184
2185	return mask;
2186}
2187
2188static unsigned int unix_dgram_poll(struct file *file, struct socket *sock,
2189				    poll_table *wait)
2190{
2191	struct sock *sk = sock->sk, *other;
2192	unsigned int mask, writable;
2193
2194	sock_poll_wait(file, sk_sleep(sk), wait);
2195	mask = 0;
2196
2197	/* exceptional events? */
2198	if (sk->sk_err || !skb_queue_empty(&sk->sk_error_queue))
2199		mask |= POLLERR;
 
 
2200	if (sk->sk_shutdown & RCV_SHUTDOWN)
2201		mask |= POLLRDHUP | POLLIN | POLLRDNORM;
2202	if (sk->sk_shutdown == SHUTDOWN_MASK)
2203		mask |= POLLHUP;
2204
2205	/* readable? */
2206	if (!skb_queue_empty(&sk->sk_receive_queue))
2207		mask |= POLLIN | POLLRDNORM;
2208
2209	/* Connection-based need to check for termination and startup */
2210	if (sk->sk_type == SOCK_SEQPACKET) {
2211		if (sk->sk_state == TCP_CLOSE)
2212			mask |= POLLHUP;
2213		/* connection hasn't started yet? */
2214		if (sk->sk_state == TCP_SYN_SENT)
2215			return mask;
2216	}
2217
2218	/* No write status requested, avoid expensive OUT tests. */
2219	if (!(poll_requested_events(wait) & (POLLWRBAND|POLLWRNORM|POLLOUT)))
2220		return mask;
2221
2222	writable = unix_writable(sk);
2223	other = unix_peer_get(sk);
2224	if (other) {
2225		if (unix_peer(other) != sk) {
2226			sock_poll_wait(file, &unix_sk(other)->peer_wait, wait);
2227			if (unix_recvq_full(other))
2228				writable = 0;
2229		}
2230		sock_put(other);
2231	}
2232
2233	if (writable)
2234		mask |= POLLOUT | POLLWRNORM | POLLWRBAND;
2235	else
2236		set_bit(SOCK_ASYNC_NOSPACE, &sk->sk_socket->flags);
2237
2238	return mask;
2239}
2240
2241#ifdef CONFIG_PROC_FS
2242static struct sock *first_unix_socket(int *i)
2243{
2244	for (*i = 0; *i <= UNIX_HASH_SIZE; (*i)++) {
2245		if (!hlist_empty(&unix_socket_table[*i]))
2246			return __sk_head(&unix_socket_table[*i]);
2247	}
2248	return NULL;
2249}
2250
2251static struct sock *next_unix_socket(int *i, struct sock *s)
 
 
 
 
 
 
2252{
2253	struct sock *next = sk_next(s);
2254	/* More in this chain? */
2255	if (next)
2256		return next;
2257	/* Look for next non-empty chain. */
2258	for ((*i)++; *i <= UNIX_HASH_SIZE; (*i)++) {
2259		if (!hlist_empty(&unix_socket_table[*i]))
2260			return __sk_head(&unix_socket_table[*i]);
 
 
2261	}
2262	return NULL;
2263}
2264
2265struct unix_iter_state {
2266	struct seq_net_private p;
2267	int i;
2268};
2269
2270static struct sock *unix_seq_idx(struct seq_file *seq, loff_t pos)
 
 
2271{
2272	struct unix_iter_state *iter = seq->private;
2273	loff_t off = 0;
2274	struct sock *s;
2275
2276	for (s = first_unix_socket(&iter->i); s; s = next_unix_socket(&iter->i, s)) {
2277		if (sock_net(s) != seq_file_net(seq))
2278			continue;
2279		if (off == pos)
2280			return s;
2281		++off;
2282	}
 
 
 
 
 
 
 
 
 
 
 
2283	return NULL;
2284}
2285
2286static void *unix_seq_start(struct seq_file *seq, loff_t *pos)
2287	__acquires(unix_table_lock)
2288{
2289	spin_lock(&unix_table_lock);
2290	return *pos ? unix_seq_idx(seq, *pos - 1) : SEQ_START_TOKEN;
 
 
 
 
 
 
 
2291}
2292
2293static void *unix_seq_next(struct seq_file *seq, void *v, loff_t *pos)
2294{
2295	struct unix_iter_state *iter = seq->private;
2296	struct sock *sk = v;
2297	++*pos;
2298
2299	if (v == SEQ_START_TOKEN)
2300		sk = first_unix_socket(&iter->i);
2301	else
2302		sk = next_unix_socket(&iter->i, sk);
2303	while (sk && (sock_net(sk) != seq_file_net(seq)))
2304		sk = next_unix_socket(&iter->i, sk);
2305	return sk;
2306}
2307
2308static void unix_seq_stop(struct seq_file *seq, void *v)
2309	__releases(unix_table_lock)
2310{
2311	spin_unlock(&unix_table_lock);
2312}
2313
2314static int unix_seq_show(struct seq_file *seq, void *v)
2315{
2316
2317	if (v == SEQ_START_TOKEN)
2318		seq_puts(seq, "Num       RefCount Protocol Flags    Type St "
2319			 "Inode Path\n");
2320	else {
2321		struct sock *s = v;
2322		struct unix_sock *u = unix_sk(s);
2323		unix_state_lock(s);
2324
2325		seq_printf(seq, "%pK: %08X %08X %08X %04X %02X %5lu",
2326			s,
2327			atomic_read(&s->sk_refcnt),
2328			0,
2329			s->sk_state == TCP_LISTEN ? __SO_ACCEPTCON : 0,
2330			s->sk_type,
2331			s->sk_socket ?
2332			(s->sk_state == TCP_ESTABLISHED ? SS_CONNECTED : SS_UNCONNECTED) :
2333			(s->sk_state == TCP_ESTABLISHED ? SS_CONNECTING : SS_DISCONNECTING),
2334			sock_i_ino(s));
2335
2336		if (u->addr) {
2337			int i, len;
2338			seq_putc(seq, ' ');
2339
2340			i = 0;
2341			len = u->addr->len - sizeof(short);
2342			if (!UNIX_ABSTRACT(s))
2343				len--;
2344			else {
2345				seq_putc(seq, '@');
2346				i++;
2347			}
2348			for ( ; i < len; i++)
2349				seq_putc(seq, u->addr->name->sun_path[i]);
2350		}
2351		unix_state_unlock(s);
2352		seq_putc(seq, '\n');
2353	}
2354
2355	return 0;
2356}
2357
2358static const struct seq_operations unix_seq_ops = {
2359	.start  = unix_seq_start,
2360	.next   = unix_seq_next,
2361	.stop   = unix_seq_stop,
2362	.show   = unix_seq_show,
2363};
2364
2365static int unix_seq_open(struct inode *inode, struct file *file)
2366{
2367	return seq_open_net(inode, file, &unix_seq_ops,
2368			    sizeof(struct unix_iter_state));
2369}
2370
2371static const struct file_operations unix_seq_fops = {
2372	.owner		= THIS_MODULE,
2373	.open		= unix_seq_open,
2374	.read		= seq_read,
2375	.llseek		= seq_lseek,
2376	.release	= seq_release_net,
2377};
2378
2379#endif
2380
2381static const struct net_proto_family unix_family_ops = {
2382	.family = PF_UNIX,
2383	.create = unix_create,
2384	.owner	= THIS_MODULE,
2385};
2386
2387
2388static int __net_init unix_net_init(struct net *net)
2389{
2390	int error = -ENOMEM;
2391
2392	net->unx.sysctl_max_dgram_qlen = 10;
2393	if (unix_sysctl_register(net))
2394		goto out;
2395
2396#ifdef CONFIG_PROC_FS
2397	if (!proc_net_fops_create(net, "unix", 0, &unix_seq_fops)) {
2398		unix_sysctl_unregister(net);
2399		goto out;
2400	}
2401#endif
2402	error = 0;
2403out:
2404	return error;
2405}
2406
2407static void __net_exit unix_net_exit(struct net *net)
2408{
2409	unix_sysctl_unregister(net);
2410	proc_net_remove(net, "unix");
2411}
2412
2413static struct pernet_operations unix_net_ops = {
2414	.init = unix_net_init,
2415	.exit = unix_net_exit,
2416};
2417
2418static int __init af_unix_init(void)
2419{
2420	int rc = -1;
2421	struct sk_buff *dummy_skb;
2422
2423	BUILD_BUG_ON(sizeof(struct unix_skb_parms) > sizeof(dummy_skb->cb));
2424
2425	rc = proto_register(&unix_proto, 1);
2426	if (rc != 0) {
2427		printk(KERN_CRIT "%s: Cannot create unix_sock SLAB cache!\n",
2428		       __func__);
2429		goto out;
2430	}
2431
2432	sock_register(&unix_family_ops);
2433	register_pernet_subsys(&unix_net_ops);
2434out:
2435	return rc;
2436}
2437
2438static void __exit af_unix_exit(void)
2439{
2440	sock_unregister(PF_UNIX);
2441	proto_unregister(&unix_proto);
2442	unregister_pernet_subsys(&unix_net_ops);
2443}
2444
2445/* Earlier than device_initcall() so that other drivers invoking
2446   request_module() don't end up in a loop when modprobe tries
2447   to use a UNIX socket. But later than subsys_initcall() because
2448   we depend on stuff initialised there */
2449fs_initcall(af_unix_init);
2450module_exit(af_unix_exit);
2451
2452MODULE_LICENSE("GPL");
2453MODULE_ALIAS_NETPROTO(PF_UNIX);
v3.15
   1/*
   2 * NET4:	Implementation of BSD Unix domain sockets.
   3 *
   4 * Authors:	Alan Cox, <alan@lxorguk.ukuu.org.uk>
   5 *
   6 *		This program is free software; you can redistribute it and/or
   7 *		modify it under the terms of the GNU General Public License
   8 *		as published by the Free Software Foundation; either version
   9 *		2 of the License, or (at your option) any later version.
  10 *
  11 * Fixes:
  12 *		Linus Torvalds	:	Assorted bug cures.
  13 *		Niibe Yutaka	:	async I/O support.
  14 *		Carsten Paeth	:	PF_UNIX check, address fixes.
  15 *		Alan Cox	:	Limit size of allocated blocks.
  16 *		Alan Cox	:	Fixed the stupid socketpair bug.
  17 *		Alan Cox	:	BSD compatibility fine tuning.
  18 *		Alan Cox	:	Fixed a bug in connect when interrupted.
  19 *		Alan Cox	:	Sorted out a proper draft version of
  20 *					file descriptor passing hacked up from
  21 *					Mike Shaver's work.
  22 *		Marty Leisner	:	Fixes to fd passing
  23 *		Nick Nevin	:	recvmsg bugfix.
  24 *		Alan Cox	:	Started proper garbage collector
  25 *		Heiko EiBfeldt	:	Missing verify_area check
  26 *		Alan Cox	:	Started POSIXisms
  27 *		Andreas Schwab	:	Replace inode by dentry for proper
  28 *					reference counting
  29 *		Kirk Petersen	:	Made this a module
  30 *	    Christoph Rohland	:	Elegant non-blocking accept/connect algorithm.
  31 *					Lots of bug fixes.
  32 *	     Alexey Kuznetosv	:	Repaired (I hope) bugs introduces
  33 *					by above two patches.
  34 *	     Andrea Arcangeli	:	If possible we block in connect(2)
  35 *					if the max backlog of the listen socket
  36 *					is been reached. This won't break
  37 *					old apps and it will avoid huge amount
  38 *					of socks hashed (this for unix_gc()
  39 *					performances reasons).
  40 *					Security fix that limits the max
  41 *					number of socks to 2*max_files and
  42 *					the number of skb queueable in the
  43 *					dgram receiver.
  44 *		Artur Skawina   :	Hash function optimizations
  45 *	     Alexey Kuznetsov   :	Full scale SMP. Lot of bugs are introduced 8)
  46 *	      Malcolm Beattie   :	Set peercred for socketpair
  47 *	     Michal Ostrowski   :       Module initialization cleanup.
  48 *	     Arnaldo C. Melo	:	Remove MOD_{INC,DEC}_USE_COUNT,
  49 *	     				the core infrastructure is doing that
  50 *	     				for all net proto families now (2.5.69+)
  51 *
  52 *
  53 * Known differences from reference BSD that was tested:
  54 *
  55 *	[TO FIX]
  56 *	ECONNREFUSED is not returned from one end of a connected() socket to the
  57 *		other the moment one end closes.
  58 *	fstat() doesn't return st_dev=0, and give the blksize as high water mark
  59 *		and a fake inode identifier (nor the BSD first socket fstat twice bug).
  60 *	[NOT TO FIX]
  61 *	accept() returns a path name even if the connecting socket has closed
  62 *		in the meantime (BSD loses the path and gives up).
  63 *	accept() returns 0 length path for an unbound connector. BSD returns 16
  64 *		and a null first byte in the path (but not for gethost/peername - BSD bug ??)
  65 *	socketpair(...SOCK_RAW..) doesn't panic the kernel.
  66 *	BSD af_unix apparently has connect forgetting to block properly.
  67 *		(need to check this with the POSIX spec in detail)
  68 *
  69 * Differences from 2.0.0-11-... (ANK)
  70 *	Bug fixes and improvements.
  71 *		- client shutdown killed server socket.
  72 *		- removed all useless cli/sti pairs.
  73 *
  74 *	Semantic changes/extensions.
  75 *		- generic control message passing.
  76 *		- SCM_CREDENTIALS control message.
  77 *		- "Abstract" (not FS based) socket bindings.
  78 *		  Abstract names are sequences of bytes (not zero terminated)
  79 *		  started by 0, so that this name space does not intersect
  80 *		  with BSD names.
  81 */
  82
  83#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  84
  85#include <linux/module.h>
  86#include <linux/kernel.h>
  87#include <linux/signal.h>
  88#include <linux/sched.h>
  89#include <linux/errno.h>
  90#include <linux/string.h>
  91#include <linux/stat.h>
  92#include <linux/dcache.h>
  93#include <linux/namei.h>
  94#include <linux/socket.h>
  95#include <linux/un.h>
  96#include <linux/fcntl.h>
  97#include <linux/termios.h>
  98#include <linux/sockios.h>
  99#include <linux/net.h>
 100#include <linux/in.h>
 101#include <linux/fs.h>
 102#include <linux/slab.h>
 103#include <asm/uaccess.h>
 104#include <linux/skbuff.h>
 105#include <linux/netdevice.h>
 106#include <net/net_namespace.h>
 107#include <net/sock.h>
 108#include <net/tcp_states.h>
 109#include <net/af_unix.h>
 110#include <linux/proc_fs.h>
 111#include <linux/seq_file.h>
 112#include <net/scm.h>
 113#include <linux/init.h>
 114#include <linux/poll.h>
 115#include <linux/rtnetlink.h>
 116#include <linux/mount.h>
 117#include <net/checksum.h>
 118#include <linux/security.h>
 119#include <linux/freezer.h>
 120
 121struct hlist_head unix_socket_table[2 * UNIX_HASH_SIZE];
 122EXPORT_SYMBOL_GPL(unix_socket_table);
 123DEFINE_SPINLOCK(unix_table_lock);
 124EXPORT_SYMBOL_GPL(unix_table_lock);
 125static atomic_long_t unix_nr_socks;
 126
 
 127
 128static struct hlist_head *unix_sockets_unbound(void *addr)
 129{
 130	unsigned long hash = (unsigned long)addr;
 131
 132	hash ^= hash >> 16;
 133	hash ^= hash >> 8;
 134	hash %= UNIX_HASH_SIZE;
 135	return &unix_socket_table[UNIX_HASH_SIZE + hash];
 136}
 137
 138#define UNIX_ABSTRACT(sk)	(unix_sk(sk)->addr->hash < UNIX_HASH_SIZE)
 139
 140#ifdef CONFIG_SECURITY_NETWORK
 141static void unix_get_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 142{
 143	memcpy(UNIXSID(skb), &scm->secid, sizeof(u32));
 144}
 145
 146static inline void unix_set_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 147{
 148	scm->secid = *UNIXSID(skb);
 149}
 150#else
 151static inline void unix_get_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 152{ }
 153
 154static inline void unix_set_secdata(struct scm_cookie *scm, struct sk_buff *skb)
 155{ }
 156#endif /* CONFIG_SECURITY_NETWORK */
 157
 158/*
 159 *  SMP locking strategy:
 160 *    hash table is protected with spinlock unix_table_lock
 161 *    each socket state is protected by separate spin lock.
 162 */
 163
 164static inline unsigned int unix_hash_fold(__wsum n)
 165{
 166	unsigned int hash = (__force unsigned int)csum_fold(n);
 167
 
 168	hash ^= hash>>8;
 169	return hash&(UNIX_HASH_SIZE-1);
 170}
 171
 172#define unix_peer(sk) (unix_sk(sk)->peer)
 173
 174static inline int unix_our_peer(struct sock *sk, struct sock *osk)
 175{
 176	return unix_peer(osk) == sk;
 177}
 178
 179static inline int unix_may_send(struct sock *sk, struct sock *osk)
 180{
 181	return unix_peer(osk) == NULL || unix_our_peer(sk, osk);
 182}
 183
 184static inline int unix_recvq_full(struct sock const *sk)
 185{
 186	return skb_queue_len(&sk->sk_receive_queue) > sk->sk_max_ack_backlog;
 187}
 188
 189struct sock *unix_peer_get(struct sock *s)
 190{
 191	struct sock *peer;
 192
 193	unix_state_lock(s);
 194	peer = unix_peer(s);
 195	if (peer)
 196		sock_hold(peer);
 197	unix_state_unlock(s);
 198	return peer;
 199}
 200EXPORT_SYMBOL_GPL(unix_peer_get);
 201
 202static inline void unix_release_addr(struct unix_address *addr)
 203{
 204	if (atomic_dec_and_test(&addr->refcnt))
 205		kfree(addr);
 206}
 207
 208/*
 209 *	Check unix socket name:
 210 *		- should be not zero length.
 211 *	        - if started by not zero, should be NULL terminated (FS object)
 212 *		- if started by zero, it is abstract name.
 213 */
 214
 215static int unix_mkname(struct sockaddr_un *sunaddr, int len, unsigned int *hashp)
 216{
 217	if (len <= sizeof(short) || len > sizeof(*sunaddr))
 218		return -EINVAL;
 219	if (!sunaddr || sunaddr->sun_family != AF_UNIX)
 220		return -EINVAL;
 221	if (sunaddr->sun_path[0]) {
 222		/*
 223		 * This may look like an off by one error but it is a bit more
 224		 * subtle. 108 is the longest valid AF_UNIX path for a binding.
 225		 * sun_path[108] doesn't as such exist.  However in kernel space
 226		 * we are guaranteed that it is a valid memory location in our
 227		 * kernel address buffer.
 228		 */
 229		((char *)sunaddr)[len] = 0;
 230		len = strlen(sunaddr->sun_path)+1+sizeof(short);
 231		return len;
 232	}
 233
 234	*hashp = unix_hash_fold(csum_partial(sunaddr, len, 0));
 235	return len;
 236}
 237
 238static void __unix_remove_socket(struct sock *sk)
 239{
 240	sk_del_node_init(sk);
 241}
 242
 243static void __unix_insert_socket(struct hlist_head *list, struct sock *sk)
 244{
 245	WARN_ON(!sk_unhashed(sk));
 246	sk_add_node(sk, list);
 247}
 248
 249static inline void unix_remove_socket(struct sock *sk)
 250{
 251	spin_lock(&unix_table_lock);
 252	__unix_remove_socket(sk);
 253	spin_unlock(&unix_table_lock);
 254}
 255
 256static inline void unix_insert_socket(struct hlist_head *list, struct sock *sk)
 257{
 258	spin_lock(&unix_table_lock);
 259	__unix_insert_socket(list, sk);
 260	spin_unlock(&unix_table_lock);
 261}
 262
 263static struct sock *__unix_find_socket_byname(struct net *net,
 264					      struct sockaddr_un *sunname,
 265					      int len, int type, unsigned int hash)
 266{
 267	struct sock *s;
 
 268
 269	sk_for_each(s, &unix_socket_table[hash ^ type]) {
 270		struct unix_sock *u = unix_sk(s);
 271
 272		if (!net_eq(sock_net(s), net))
 273			continue;
 274
 275		if (u->addr->len == len &&
 276		    !memcmp(u->addr->name, sunname, len))
 277			goto found;
 278	}
 279	s = NULL;
 280found:
 281	return s;
 282}
 283
 284static inline struct sock *unix_find_socket_byname(struct net *net,
 285						   struct sockaddr_un *sunname,
 286						   int len, int type,
 287						   unsigned int hash)
 288{
 289	struct sock *s;
 290
 291	spin_lock(&unix_table_lock);
 292	s = __unix_find_socket_byname(net, sunname, len, type, hash);
 293	if (s)
 294		sock_hold(s);
 295	spin_unlock(&unix_table_lock);
 296	return s;
 297}
 298
 299static struct sock *unix_find_socket_byinode(struct inode *i)
 300{
 301	struct sock *s;
 
 302
 303	spin_lock(&unix_table_lock);
 304	sk_for_each(s,
 305		    &unix_socket_table[i->i_ino & (UNIX_HASH_SIZE - 1)]) {
 306		struct dentry *dentry = unix_sk(s)->path.dentry;
 307
 308		if (dentry && dentry->d_inode == i) {
 309			sock_hold(s);
 310			goto found;
 311		}
 312	}
 313	s = NULL;
 314found:
 315	spin_unlock(&unix_table_lock);
 316	return s;
 317}
 318
 319static inline int unix_writable(struct sock *sk)
 320{
 321	return (atomic_read(&sk->sk_wmem_alloc) << 2) <= sk->sk_sndbuf;
 322}
 323
 324static void unix_write_space(struct sock *sk)
 325{
 326	struct socket_wq *wq;
 327
 328	rcu_read_lock();
 329	if (unix_writable(sk)) {
 330		wq = rcu_dereference(sk->sk_wq);
 331		if (wq_has_sleeper(wq))
 332			wake_up_interruptible_sync_poll(&wq->wait,
 333				POLLOUT | POLLWRNORM | POLLWRBAND);
 334		sk_wake_async(sk, SOCK_WAKE_SPACE, POLL_OUT);
 335	}
 336	rcu_read_unlock();
 337}
 338
 339/* When dgram socket disconnects (or changes its peer), we clear its receive
 340 * queue of packets arrived from previous peer. First, it allows to do
 341 * flow control based only on wmem_alloc; second, sk connected to peer
 342 * may receive messages only from that peer. */
 343static void unix_dgram_disconnected(struct sock *sk, struct sock *other)
 344{
 345	if (!skb_queue_empty(&sk->sk_receive_queue)) {
 346		skb_queue_purge(&sk->sk_receive_queue);
 347		wake_up_interruptible_all(&unix_sk(sk)->peer_wait);
 348
 349		/* If one link of bidirectional dgram pipe is disconnected,
 350		 * we signal error. Messages are lost. Do not make this,
 351		 * when peer was not connected to us.
 352		 */
 353		if (!sock_flag(other, SOCK_DEAD) && unix_peer(other) == sk) {
 354			other->sk_err = ECONNRESET;
 355			other->sk_error_report(other);
 356		}
 357	}
 358}
 359
 360static void unix_sock_destructor(struct sock *sk)
 361{
 362	struct unix_sock *u = unix_sk(sk);
 363
 364	skb_queue_purge(&sk->sk_receive_queue);
 365
 366	WARN_ON(atomic_read(&sk->sk_wmem_alloc));
 367	WARN_ON(!sk_unhashed(sk));
 368	WARN_ON(sk->sk_socket);
 369	if (!sock_flag(sk, SOCK_DEAD)) {
 370		pr_info("Attempt to release alive unix socket: %p\n", sk);
 371		return;
 372	}
 373
 374	if (u->addr)
 375		unix_release_addr(u->addr);
 376
 377	atomic_long_dec(&unix_nr_socks);
 378	local_bh_disable();
 379	sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1);
 380	local_bh_enable();
 381#ifdef UNIX_REFCNT_DEBUG
 382	pr_debug("UNIX %p is destroyed, %ld are still alive.\n", sk,
 383		atomic_long_read(&unix_nr_socks));
 384#endif
 385}
 386
 387static void unix_release_sock(struct sock *sk, int embrion)
 388{
 389	struct unix_sock *u = unix_sk(sk);
 390	struct path path;
 391	struct sock *skpair;
 392	struct sk_buff *skb;
 393	int state;
 394
 395	unix_remove_socket(sk);
 396
 397	/* Clear state */
 398	unix_state_lock(sk);
 399	sock_orphan(sk);
 400	sk->sk_shutdown = SHUTDOWN_MASK;
 401	path	     = u->path;
 402	u->path.dentry = NULL;
 403	u->path.mnt = NULL;
 404	state = sk->sk_state;
 405	sk->sk_state = TCP_CLOSE;
 406	unix_state_unlock(sk);
 407
 408	wake_up_interruptible_all(&u->peer_wait);
 409
 410	skpair = unix_peer(sk);
 411
 412	if (skpair != NULL) {
 413		if (sk->sk_type == SOCK_STREAM || sk->sk_type == SOCK_SEQPACKET) {
 414			unix_state_lock(skpair);
 415			/* No more writes */
 416			skpair->sk_shutdown = SHUTDOWN_MASK;
 417			if (!skb_queue_empty(&sk->sk_receive_queue) || embrion)
 418				skpair->sk_err = ECONNRESET;
 419			unix_state_unlock(skpair);
 420			skpair->sk_state_change(skpair);
 421			sk_wake_async(skpair, SOCK_WAKE_WAITD, POLL_HUP);
 422		}
 423		sock_put(skpair); /* It may now die */
 424		unix_peer(sk) = NULL;
 425	}
 426
 427	/* Try to flush out this socket. Throw out buffers at least */
 428
 429	while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) {
 430		if (state == TCP_LISTEN)
 431			unix_release_sock(skb->sk, 1);
 432		/* passed fds are erased in the kfree_skb hook	      */
 433		kfree_skb(skb);
 434	}
 435
 436	if (path.dentry)
 437		path_put(&path);
 438
 439	sock_put(sk);
 440
 441	/* ---- Socket is dead now and most probably destroyed ---- */
 442
 443	/*
 444	 * Fixme: BSD difference: In BSD all sockets connected to us get
 445	 *	  ECONNRESET and we die on the spot. In Linux we behave
 446	 *	  like files and pipes do and wait for the last
 447	 *	  dereference.
 448	 *
 449	 * Can't we simply set sock->err?
 450	 *
 451	 *	  What the above comment does talk about? --ANK(980817)
 452	 */
 453
 454	if (unix_tot_inflight)
 455		unix_gc();		/* Garbage collect fds */
 
 
 456}
 457
 458static void init_peercred(struct sock *sk)
 459{
 460	put_pid(sk->sk_peer_pid);
 461	if (sk->sk_peer_cred)
 462		put_cred(sk->sk_peer_cred);
 463	sk->sk_peer_pid  = get_pid(task_tgid(current));
 464	sk->sk_peer_cred = get_current_cred();
 465}
 466
 467static void copy_peercred(struct sock *sk, struct sock *peersk)
 468{
 469	put_pid(sk->sk_peer_pid);
 470	if (sk->sk_peer_cred)
 471		put_cred(sk->sk_peer_cred);
 472	sk->sk_peer_pid  = get_pid(peersk->sk_peer_pid);
 473	sk->sk_peer_cred = get_cred(peersk->sk_peer_cred);
 474}
 475
 476static int unix_listen(struct socket *sock, int backlog)
 477{
 478	int err;
 479	struct sock *sk = sock->sk;
 480	struct unix_sock *u = unix_sk(sk);
 481	struct pid *old_pid = NULL;
 
 482
 483	err = -EOPNOTSUPP;
 484	if (sock->type != SOCK_STREAM && sock->type != SOCK_SEQPACKET)
 485		goto out;	/* Only stream/seqpacket sockets accept */
 486	err = -EINVAL;
 487	if (!u->addr)
 488		goto out;	/* No listens on an unbound socket */
 489	unix_state_lock(sk);
 490	if (sk->sk_state != TCP_CLOSE && sk->sk_state != TCP_LISTEN)
 491		goto out_unlock;
 492	if (backlog > sk->sk_max_ack_backlog)
 493		wake_up_interruptible_all(&u->peer_wait);
 494	sk->sk_max_ack_backlog	= backlog;
 495	sk->sk_state		= TCP_LISTEN;
 496	/* set credentials so connect can copy them */
 497	init_peercred(sk);
 498	err = 0;
 499
 500out_unlock:
 501	unix_state_unlock(sk);
 502	put_pid(old_pid);
 
 
 503out:
 504	return err;
 505}
 506
 507static int unix_release(struct socket *);
 508static int unix_bind(struct socket *, struct sockaddr *, int);
 509static int unix_stream_connect(struct socket *, struct sockaddr *,
 510			       int addr_len, int flags);
 511static int unix_socketpair(struct socket *, struct socket *);
 512static int unix_accept(struct socket *, struct socket *, int);
 513static int unix_getname(struct socket *, struct sockaddr *, int *, int);
 514static unsigned int unix_poll(struct file *, struct socket *, poll_table *);
 515static unsigned int unix_dgram_poll(struct file *, struct socket *,
 516				    poll_table *);
 517static int unix_ioctl(struct socket *, unsigned int, unsigned long);
 518static int unix_shutdown(struct socket *, int);
 519static int unix_stream_sendmsg(struct kiocb *, struct socket *,
 520			       struct msghdr *, size_t);
 521static int unix_stream_recvmsg(struct kiocb *, struct socket *,
 522			       struct msghdr *, size_t, int);
 523static int unix_dgram_sendmsg(struct kiocb *, struct socket *,
 524			      struct msghdr *, size_t);
 525static int unix_dgram_recvmsg(struct kiocb *, struct socket *,
 526			      struct msghdr *, size_t, int);
 527static int unix_dgram_connect(struct socket *, struct sockaddr *,
 528			      int, int);
 529static int unix_seqpacket_sendmsg(struct kiocb *, struct socket *,
 530				  struct msghdr *, size_t);
 531static int unix_seqpacket_recvmsg(struct kiocb *, struct socket *,
 532				  struct msghdr *, size_t, int);
 533
 534static int unix_set_peek_off(struct sock *sk, int val)
 535{
 536	struct unix_sock *u = unix_sk(sk);
 537
 538	if (mutex_lock_interruptible(&u->readlock))
 539		return -EINTR;
 540
 541	sk->sk_peek_off = val;
 542	mutex_unlock(&u->readlock);
 543
 544	return 0;
 545}
 546
 547
 548static const struct proto_ops unix_stream_ops = {
 549	.family =	PF_UNIX,
 550	.owner =	THIS_MODULE,
 551	.release =	unix_release,
 552	.bind =		unix_bind,
 553	.connect =	unix_stream_connect,
 554	.socketpair =	unix_socketpair,
 555	.accept =	unix_accept,
 556	.getname =	unix_getname,
 557	.poll =		unix_poll,
 558	.ioctl =	unix_ioctl,
 559	.listen =	unix_listen,
 560	.shutdown =	unix_shutdown,
 561	.setsockopt =	sock_no_setsockopt,
 562	.getsockopt =	sock_no_getsockopt,
 563	.sendmsg =	unix_stream_sendmsg,
 564	.recvmsg =	unix_stream_recvmsg,
 565	.mmap =		sock_no_mmap,
 566	.sendpage =	sock_no_sendpage,
 567	.set_peek_off =	unix_set_peek_off,
 568};
 569
 570static const struct proto_ops unix_dgram_ops = {
 571	.family =	PF_UNIX,
 572	.owner =	THIS_MODULE,
 573	.release =	unix_release,
 574	.bind =		unix_bind,
 575	.connect =	unix_dgram_connect,
 576	.socketpair =	unix_socketpair,
 577	.accept =	sock_no_accept,
 578	.getname =	unix_getname,
 579	.poll =		unix_dgram_poll,
 580	.ioctl =	unix_ioctl,
 581	.listen =	sock_no_listen,
 582	.shutdown =	unix_shutdown,
 583	.setsockopt =	sock_no_setsockopt,
 584	.getsockopt =	sock_no_getsockopt,
 585	.sendmsg =	unix_dgram_sendmsg,
 586	.recvmsg =	unix_dgram_recvmsg,
 587	.mmap =		sock_no_mmap,
 588	.sendpage =	sock_no_sendpage,
 589	.set_peek_off =	unix_set_peek_off,
 590};
 591
 592static const struct proto_ops unix_seqpacket_ops = {
 593	.family =	PF_UNIX,
 594	.owner =	THIS_MODULE,
 595	.release =	unix_release,
 596	.bind =		unix_bind,
 597	.connect =	unix_stream_connect,
 598	.socketpair =	unix_socketpair,
 599	.accept =	unix_accept,
 600	.getname =	unix_getname,
 601	.poll =		unix_dgram_poll,
 602	.ioctl =	unix_ioctl,
 603	.listen =	unix_listen,
 604	.shutdown =	unix_shutdown,
 605	.setsockopt =	sock_no_setsockopt,
 606	.getsockopt =	sock_no_getsockopt,
 607	.sendmsg =	unix_seqpacket_sendmsg,
 608	.recvmsg =	unix_seqpacket_recvmsg,
 609	.mmap =		sock_no_mmap,
 610	.sendpage =	sock_no_sendpage,
 611	.set_peek_off =	unix_set_peek_off,
 612};
 613
 614static struct proto unix_proto = {
 615	.name			= "UNIX",
 616	.owner			= THIS_MODULE,
 617	.obj_size		= sizeof(struct unix_sock),
 618};
 619
 620/*
 621 * AF_UNIX sockets do not interact with hardware, hence they
 622 * dont trigger interrupts - so it's safe for them to have
 623 * bh-unsafe locking for their sk_receive_queue.lock. Split off
 624 * this special lock-class by reinitializing the spinlock key:
 625 */
 626static struct lock_class_key af_unix_sk_receive_queue_lock_key;
 627
 628static struct sock *unix_create1(struct net *net, struct socket *sock)
 629{
 630	struct sock *sk = NULL;
 631	struct unix_sock *u;
 632
 633	atomic_long_inc(&unix_nr_socks);
 634	if (atomic_long_read(&unix_nr_socks) > 2 * get_max_files())
 635		goto out;
 636
 637	sk = sk_alloc(net, PF_UNIX, GFP_KERNEL, &unix_proto);
 638	if (!sk)
 639		goto out;
 640
 641	sock_init_data(sock, sk);
 642	lockdep_set_class(&sk->sk_receive_queue.lock,
 643				&af_unix_sk_receive_queue_lock_key);
 644
 645	sk->sk_write_space	= unix_write_space;
 646	sk->sk_max_ack_backlog	= net->unx.sysctl_max_dgram_qlen;
 647	sk->sk_destruct		= unix_sock_destructor;
 648	u	  = unix_sk(sk);
 649	u->path.dentry = NULL;
 650	u->path.mnt = NULL;
 651	spin_lock_init(&u->lock);
 652	atomic_long_set(&u->inflight, 0);
 653	INIT_LIST_HEAD(&u->link);
 654	mutex_init(&u->readlock); /* single task reading lock */
 655	init_waitqueue_head(&u->peer_wait);
 656	unix_insert_socket(unix_sockets_unbound(sk), sk);
 657out:
 658	if (sk == NULL)
 659		atomic_long_dec(&unix_nr_socks);
 660	else {
 661		local_bh_disable();
 662		sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1);
 663		local_bh_enable();
 664	}
 665	return sk;
 666}
 667
 668static int unix_create(struct net *net, struct socket *sock, int protocol,
 669		       int kern)
 670{
 671	if (protocol && protocol != PF_UNIX)
 672		return -EPROTONOSUPPORT;
 673
 674	sock->state = SS_UNCONNECTED;
 675
 676	switch (sock->type) {
 677	case SOCK_STREAM:
 678		sock->ops = &unix_stream_ops;
 679		break;
 680		/*
 681		 *	Believe it or not BSD has AF_UNIX, SOCK_RAW though
 682		 *	nothing uses it.
 683		 */
 684	case SOCK_RAW:
 685		sock->type = SOCK_DGRAM;
 686	case SOCK_DGRAM:
 687		sock->ops = &unix_dgram_ops;
 688		break;
 689	case SOCK_SEQPACKET:
 690		sock->ops = &unix_seqpacket_ops;
 691		break;
 692	default:
 693		return -ESOCKTNOSUPPORT;
 694	}
 695
 696	return unix_create1(net, sock) ? 0 : -ENOMEM;
 697}
 698
 699static int unix_release(struct socket *sock)
 700{
 701	struct sock *sk = sock->sk;
 702
 703	if (!sk)
 704		return 0;
 705
 706	unix_release_sock(sk, 0);
 707	sock->sk = NULL;
 708
 709	return 0;
 710}
 711
 712static int unix_autobind(struct socket *sock)
 713{
 714	struct sock *sk = sock->sk;
 715	struct net *net = sock_net(sk);
 716	struct unix_sock *u = unix_sk(sk);
 717	static u32 ordernum = 1;
 718	struct unix_address *addr;
 719	int err;
 720	unsigned int retries = 0;
 721
 722	err = mutex_lock_interruptible(&u->readlock);
 723	if (err)
 724		return err;
 725
 726	err = 0;
 727	if (u->addr)
 728		goto out;
 729
 730	err = -ENOMEM;
 731	addr = kzalloc(sizeof(*addr) + sizeof(short) + 16, GFP_KERNEL);
 732	if (!addr)
 733		goto out;
 734
 735	addr->name->sun_family = AF_UNIX;
 736	atomic_set(&addr->refcnt, 1);
 737
 738retry:
 739	addr->len = sprintf(addr->name->sun_path+1, "%05x", ordernum) + 1 + sizeof(short);
 740	addr->hash = unix_hash_fold(csum_partial(addr->name, addr->len, 0));
 741
 742	spin_lock(&unix_table_lock);
 743	ordernum = (ordernum+1)&0xFFFFF;
 744
 745	if (__unix_find_socket_byname(net, addr->name, addr->len, sock->type,
 746				      addr->hash)) {
 747		spin_unlock(&unix_table_lock);
 748		/*
 749		 * __unix_find_socket_byname() may take long time if many names
 750		 * are already in use.
 751		 */
 752		cond_resched();
 753		/* Give up if all names seems to be in use. */
 754		if (retries++ == 0xFFFFF) {
 755			err = -ENOSPC;
 756			kfree(addr);
 757			goto out;
 758		}
 759		goto retry;
 760	}
 761	addr->hash ^= sk->sk_type;
 762
 763	__unix_remove_socket(sk);
 764	u->addr = addr;
 765	__unix_insert_socket(&unix_socket_table[addr->hash], sk);
 766	spin_unlock(&unix_table_lock);
 767	err = 0;
 768
 769out:	mutex_unlock(&u->readlock);
 770	return err;
 771}
 772
 773static struct sock *unix_find_other(struct net *net,
 774				    struct sockaddr_un *sunname, int len,
 775				    int type, unsigned int hash, int *error)
 776{
 777	struct sock *u;
 778	struct path path;
 779	int err = 0;
 780
 781	if (sunname->sun_path[0]) {
 782		struct inode *inode;
 783		err = kern_path(sunname->sun_path, LOOKUP_FOLLOW, &path);
 784		if (err)
 785			goto fail;
 786		inode = path.dentry->d_inode;
 787		err = inode_permission(inode, MAY_WRITE);
 788		if (err)
 789			goto put_fail;
 790
 791		err = -ECONNREFUSED;
 792		if (!S_ISSOCK(inode->i_mode))
 793			goto put_fail;
 794		u = unix_find_socket_byinode(inode);
 795		if (!u)
 796			goto put_fail;
 797
 798		if (u->sk_type == type)
 799			touch_atime(&path);
 800
 801		path_put(&path);
 802
 803		err = -EPROTOTYPE;
 804		if (u->sk_type != type) {
 805			sock_put(u);
 806			goto fail;
 807		}
 808	} else {
 809		err = -ECONNREFUSED;
 810		u = unix_find_socket_byname(net, sunname, len, type, hash);
 811		if (u) {
 812			struct dentry *dentry;
 813			dentry = unix_sk(u)->path.dentry;
 814			if (dentry)
 815				touch_atime(&unix_sk(u)->path);
 816		} else
 817			goto fail;
 818	}
 819	return u;
 820
 821put_fail:
 822	path_put(&path);
 823fail:
 824	*error = err;
 825	return NULL;
 826}
 827
 828static int unix_mknod(const char *sun_path, umode_t mode, struct path *res)
 829{
 830	struct dentry *dentry;
 831	struct path path;
 832	int err = 0;
 833	/*
 834	 * Get the parent directory, calculate the hash for last
 835	 * component.
 836	 */
 837	dentry = kern_path_create(AT_FDCWD, sun_path, &path, 0);
 838	err = PTR_ERR(dentry);
 839	if (IS_ERR(dentry))
 840		return err;
 841
 842	/*
 843	 * All right, let's create it.
 844	 */
 845	err = security_path_mknod(&path, dentry, mode, 0);
 846	if (!err) {
 847		err = vfs_mknod(path.dentry->d_inode, dentry, mode, 0);
 848		if (!err) {
 849			res->mnt = mntget(path.mnt);
 850			res->dentry = dget(dentry);
 851		}
 852	}
 853	done_path_create(&path, dentry);
 854	return err;
 855}
 856
 857static int unix_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len)
 858{
 859	struct sock *sk = sock->sk;
 860	struct net *net = sock_net(sk);
 861	struct unix_sock *u = unix_sk(sk);
 862	struct sockaddr_un *sunaddr = (struct sockaddr_un *)uaddr;
 863	char *sun_path = sunaddr->sun_path;
 
 
 864	int err;
 865	unsigned int hash;
 866	struct unix_address *addr;
 867	struct hlist_head *list;
 868
 869	err = -EINVAL;
 870	if (sunaddr->sun_family != AF_UNIX)
 871		goto out;
 872
 873	if (addr_len == sizeof(short)) {
 874		err = unix_autobind(sock);
 875		goto out;
 876	}
 877
 878	err = unix_mkname(sunaddr, addr_len, &hash);
 879	if (err < 0)
 880		goto out;
 881	addr_len = err;
 882
 883	err = mutex_lock_interruptible(&u->readlock);
 884	if (err)
 885		goto out;
 886
 887	err = -EINVAL;
 888	if (u->addr)
 889		goto out_up;
 890
 891	err = -ENOMEM;
 892	addr = kmalloc(sizeof(*addr)+addr_len, GFP_KERNEL);
 893	if (!addr)
 894		goto out_up;
 895
 896	memcpy(addr->name, sunaddr, addr_len);
 897	addr->len = addr_len;
 898	addr->hash = hash ^ sk->sk_type;
 899	atomic_set(&addr->refcnt, 1);
 900
 901	if (sun_path[0]) {
 902		struct path path;
 903		umode_t mode = S_IFSOCK |
 
 
 
 
 
 
 
 
 
 
 
 
 
 904		       (SOCK_INODE(sock)->i_mode & ~current_umask());
 905		err = unix_mknod(sun_path, mode, &path);
 906		if (err) {
 907			if (err == -EEXIST)
 908				err = -EADDRINUSE;
 909			unix_release_addr(addr);
 910			goto out_up;
 911		}
 
 
 
 
 
 
 
 
 912		addr->hash = UNIX_HASH_SIZE;
 913		hash = path.dentry->d_inode->i_ino & (UNIX_HASH_SIZE-1);
 914		spin_lock(&unix_table_lock);
 915		u->path = path;
 916		list = &unix_socket_table[hash];
 917	} else {
 918		spin_lock(&unix_table_lock);
 919		err = -EADDRINUSE;
 920		if (__unix_find_socket_byname(net, sunaddr, addr_len,
 921					      sk->sk_type, hash)) {
 922			unix_release_addr(addr);
 923			goto out_unlock;
 924		}
 925
 926		list = &unix_socket_table[addr->hash];
 
 
 
 927	}
 928
 929	err = 0;
 930	__unix_remove_socket(sk);
 931	u->addr = addr;
 932	__unix_insert_socket(list, sk);
 933
 934out_unlock:
 935	spin_unlock(&unix_table_lock);
 936out_up:
 937	mutex_unlock(&u->readlock);
 938out:
 939	return err;
 
 
 
 
 
 
 
 
 
 
 940}
 941
 942static void unix_state_double_lock(struct sock *sk1, struct sock *sk2)
 943{
 944	if (unlikely(sk1 == sk2) || !sk2) {
 945		unix_state_lock(sk1);
 946		return;
 947	}
 948	if (sk1 < sk2) {
 949		unix_state_lock(sk1);
 950		unix_state_lock_nested(sk2);
 951	} else {
 952		unix_state_lock(sk2);
 953		unix_state_lock_nested(sk1);
 954	}
 955}
 956
 957static void unix_state_double_unlock(struct sock *sk1, struct sock *sk2)
 958{
 959	if (unlikely(sk1 == sk2) || !sk2) {
 960		unix_state_unlock(sk1);
 961		return;
 962	}
 963	unix_state_unlock(sk1);
 964	unix_state_unlock(sk2);
 965}
 966
 967static int unix_dgram_connect(struct socket *sock, struct sockaddr *addr,
 968			      int alen, int flags)
 969{
 970	struct sock *sk = sock->sk;
 971	struct net *net = sock_net(sk);
 972	struct sockaddr_un *sunaddr = (struct sockaddr_un *)addr;
 973	struct sock *other;
 974	unsigned int hash;
 975	int err;
 976
 977	if (addr->sa_family != AF_UNSPEC) {
 978		err = unix_mkname(sunaddr, alen, &hash);
 979		if (err < 0)
 980			goto out;
 981		alen = err;
 982
 983		if (test_bit(SOCK_PASSCRED, &sock->flags) &&
 984		    !unix_sk(sk)->addr && (err = unix_autobind(sock)) != 0)
 985			goto out;
 986
 987restart:
 988		other = unix_find_other(net, sunaddr, alen, sock->type, hash, &err);
 989		if (!other)
 990			goto out;
 991
 992		unix_state_double_lock(sk, other);
 993
 994		/* Apparently VFS overslept socket death. Retry. */
 995		if (sock_flag(other, SOCK_DEAD)) {
 996			unix_state_double_unlock(sk, other);
 997			sock_put(other);
 998			goto restart;
 999		}
1000
1001		err = -EPERM;
1002		if (!unix_may_send(sk, other))
1003			goto out_unlock;
1004
1005		err = security_unix_may_send(sk->sk_socket, other->sk_socket);
1006		if (err)
1007			goto out_unlock;
1008
1009	} else {
1010		/*
1011		 *	1003.1g breaking connected state with AF_UNSPEC
1012		 */
1013		other = NULL;
1014		unix_state_double_lock(sk, other);
1015	}
1016
1017	/*
1018	 * If it was connected, reconnect.
1019	 */
1020	if (unix_peer(sk)) {
1021		struct sock *old_peer = unix_peer(sk);
1022		unix_peer(sk) = other;
1023		unix_state_double_unlock(sk, other);
1024
1025		if (other != old_peer)
1026			unix_dgram_disconnected(sk, old_peer);
1027		sock_put(old_peer);
1028	} else {
1029		unix_peer(sk) = other;
1030		unix_state_double_unlock(sk, other);
1031	}
1032	return 0;
1033
1034out_unlock:
1035	unix_state_double_unlock(sk, other);
1036	sock_put(other);
1037out:
1038	return err;
1039}
1040
1041static long unix_wait_for_peer(struct sock *other, long timeo)
1042{
1043	struct unix_sock *u = unix_sk(other);
1044	int sched;
1045	DEFINE_WAIT(wait);
1046
1047	prepare_to_wait_exclusive(&u->peer_wait, &wait, TASK_INTERRUPTIBLE);
1048
1049	sched = !sock_flag(other, SOCK_DEAD) &&
1050		!(other->sk_shutdown & RCV_SHUTDOWN) &&
1051		unix_recvq_full(other);
1052
1053	unix_state_unlock(other);
1054
1055	if (sched)
1056		timeo = schedule_timeout(timeo);
1057
1058	finish_wait(&u->peer_wait, &wait);
1059	return timeo;
1060}
1061
1062static int unix_stream_connect(struct socket *sock, struct sockaddr *uaddr,
1063			       int addr_len, int flags)
1064{
1065	struct sockaddr_un *sunaddr = (struct sockaddr_un *)uaddr;
1066	struct sock *sk = sock->sk;
1067	struct net *net = sock_net(sk);
1068	struct unix_sock *u = unix_sk(sk), *newu, *otheru;
1069	struct sock *newsk = NULL;
1070	struct sock *other = NULL;
1071	struct sk_buff *skb = NULL;
1072	unsigned int hash;
1073	int st;
1074	int err;
1075	long timeo;
1076
1077	err = unix_mkname(sunaddr, addr_len, &hash);
1078	if (err < 0)
1079		goto out;
1080	addr_len = err;
1081
1082	if (test_bit(SOCK_PASSCRED, &sock->flags) && !u->addr &&
1083	    (err = unix_autobind(sock)) != 0)
1084		goto out;
1085
1086	timeo = sock_sndtimeo(sk, flags & O_NONBLOCK);
1087
1088	/* First of all allocate resources.
1089	   If we will make it after state is locked,
1090	   we will have to recheck all again in any case.
1091	 */
1092
1093	err = -ENOMEM;
1094
1095	/* create new sock for complete connection */
1096	newsk = unix_create1(sock_net(sk), NULL);
1097	if (newsk == NULL)
1098		goto out;
1099
1100	/* Allocate skb for sending to listening sock */
1101	skb = sock_wmalloc(newsk, 1, 0, GFP_KERNEL);
1102	if (skb == NULL)
1103		goto out;
1104
1105restart:
1106	/*  Find listening sock. */
1107	other = unix_find_other(net, sunaddr, addr_len, sk->sk_type, hash, &err);
1108	if (!other)
1109		goto out;
1110
1111	/* Latch state of peer */
1112	unix_state_lock(other);
1113
1114	/* Apparently VFS overslept socket death. Retry. */
1115	if (sock_flag(other, SOCK_DEAD)) {
1116		unix_state_unlock(other);
1117		sock_put(other);
1118		goto restart;
1119	}
1120
1121	err = -ECONNREFUSED;
1122	if (other->sk_state != TCP_LISTEN)
1123		goto out_unlock;
1124	if (other->sk_shutdown & RCV_SHUTDOWN)
1125		goto out_unlock;
1126
1127	if (unix_recvq_full(other)) {
1128		err = -EAGAIN;
1129		if (!timeo)
1130			goto out_unlock;
1131
1132		timeo = unix_wait_for_peer(other, timeo);
1133
1134		err = sock_intr_errno(timeo);
1135		if (signal_pending(current))
1136			goto out;
1137		sock_put(other);
1138		goto restart;
1139	}
1140
1141	/* Latch our state.
1142
1143	   It is tricky place. We need to grab our state lock and cannot
1144	   drop lock on peer. It is dangerous because deadlock is
1145	   possible. Connect to self case and simultaneous
1146	   attempt to connect are eliminated by checking socket
1147	   state. other is TCP_LISTEN, if sk is TCP_LISTEN we
1148	   check this before attempt to grab lock.
1149
1150	   Well, and we have to recheck the state after socket locked.
1151	 */
1152	st = sk->sk_state;
1153
1154	switch (st) {
1155	case TCP_CLOSE:
1156		/* This is ok... continue with connect */
1157		break;
1158	case TCP_ESTABLISHED:
1159		/* Socket is already connected */
1160		err = -EISCONN;
1161		goto out_unlock;
1162	default:
1163		err = -EINVAL;
1164		goto out_unlock;
1165	}
1166
1167	unix_state_lock_nested(sk);
1168
1169	if (sk->sk_state != st) {
1170		unix_state_unlock(sk);
1171		unix_state_unlock(other);
1172		sock_put(other);
1173		goto restart;
1174	}
1175
1176	err = security_unix_stream_connect(sk, other, newsk);
1177	if (err) {
1178		unix_state_unlock(sk);
1179		goto out_unlock;
1180	}
1181
1182	/* The way is open! Fastly set all the necessary fields... */
1183
1184	sock_hold(sk);
1185	unix_peer(newsk)	= sk;
1186	newsk->sk_state		= TCP_ESTABLISHED;
1187	newsk->sk_type		= sk->sk_type;
1188	init_peercred(newsk);
1189	newu = unix_sk(newsk);
1190	RCU_INIT_POINTER(newsk->sk_wq, &newu->peer_wq);
1191	otheru = unix_sk(other);
1192
1193	/* copy address information from listening to new sock*/
1194	if (otheru->addr) {
1195		atomic_inc(&otheru->addr->refcnt);
1196		newu->addr = otheru->addr;
1197	}
1198	if (otheru->path.dentry) {
1199		path_get(&otheru->path);
1200		newu->path = otheru->path;
1201	}
1202
1203	/* Set credentials */
1204	copy_peercred(sk, other);
1205
1206	sock->state	= SS_CONNECTED;
1207	sk->sk_state	= TCP_ESTABLISHED;
1208	sock_hold(newsk);
1209
1210	smp_mb__after_atomic_inc();	/* sock_hold() does an atomic_inc() */
1211	unix_peer(sk)	= newsk;
1212
1213	unix_state_unlock(sk);
1214
1215	/* take ten and and send info to listening sock */
1216	spin_lock(&other->sk_receive_queue.lock);
1217	__skb_queue_tail(&other->sk_receive_queue, skb);
1218	spin_unlock(&other->sk_receive_queue.lock);
1219	unix_state_unlock(other);
1220	other->sk_data_ready(other);
1221	sock_put(other);
1222	return 0;
1223
1224out_unlock:
1225	if (other)
1226		unix_state_unlock(other);
1227
1228out:
1229	kfree_skb(skb);
1230	if (newsk)
1231		unix_release_sock(newsk, 0);
1232	if (other)
1233		sock_put(other);
1234	return err;
1235}
1236
1237static int unix_socketpair(struct socket *socka, struct socket *sockb)
1238{
1239	struct sock *ska = socka->sk, *skb = sockb->sk;
1240
1241	/* Join our sockets back to back */
1242	sock_hold(ska);
1243	sock_hold(skb);
1244	unix_peer(ska) = skb;
1245	unix_peer(skb) = ska;
1246	init_peercred(ska);
1247	init_peercred(skb);
1248
1249	if (ska->sk_type != SOCK_DGRAM) {
1250		ska->sk_state = TCP_ESTABLISHED;
1251		skb->sk_state = TCP_ESTABLISHED;
1252		socka->state  = SS_CONNECTED;
1253		sockb->state  = SS_CONNECTED;
1254	}
1255	return 0;
1256}
1257
1258static void unix_sock_inherit_flags(const struct socket *old,
1259				    struct socket *new)
1260{
1261	if (test_bit(SOCK_PASSCRED, &old->flags))
1262		set_bit(SOCK_PASSCRED, &new->flags);
1263	if (test_bit(SOCK_PASSSEC, &old->flags))
1264		set_bit(SOCK_PASSSEC, &new->flags);
1265}
1266
1267static int unix_accept(struct socket *sock, struct socket *newsock, int flags)
1268{
1269	struct sock *sk = sock->sk;
1270	struct sock *tsk;
1271	struct sk_buff *skb;
1272	int err;
1273
1274	err = -EOPNOTSUPP;
1275	if (sock->type != SOCK_STREAM && sock->type != SOCK_SEQPACKET)
1276		goto out;
1277
1278	err = -EINVAL;
1279	if (sk->sk_state != TCP_LISTEN)
1280		goto out;
1281
1282	/* If socket state is TCP_LISTEN it cannot change (for now...),
1283	 * so that no locks are necessary.
1284	 */
1285
1286	skb = skb_recv_datagram(sk, 0, flags&O_NONBLOCK, &err);
1287	if (!skb) {
1288		/* This means receive shutdown. */
1289		if (err == 0)
1290			err = -EINVAL;
1291		goto out;
1292	}
1293
1294	tsk = skb->sk;
1295	skb_free_datagram(sk, skb);
1296	wake_up_interruptible(&unix_sk(sk)->peer_wait);
1297
1298	/* attach accepted sock to socket */
1299	unix_state_lock(tsk);
1300	newsock->state = SS_CONNECTED;
1301	unix_sock_inherit_flags(sock, newsock);
1302	sock_graft(tsk, newsock);
1303	unix_state_unlock(tsk);
1304	return 0;
1305
1306out:
1307	return err;
1308}
1309
1310
1311static int unix_getname(struct socket *sock, struct sockaddr *uaddr, int *uaddr_len, int peer)
1312{
1313	struct sock *sk = sock->sk;
1314	struct unix_sock *u;
1315	DECLARE_SOCKADDR(struct sockaddr_un *, sunaddr, uaddr);
1316	int err = 0;
1317
1318	if (peer) {
1319		sk = unix_peer_get(sk);
1320
1321		err = -ENOTCONN;
1322		if (!sk)
1323			goto out;
1324		err = 0;
1325	} else {
1326		sock_hold(sk);
1327	}
1328
1329	u = unix_sk(sk);
1330	unix_state_lock(sk);
1331	if (!u->addr) {
1332		sunaddr->sun_family = AF_UNIX;
1333		sunaddr->sun_path[0] = 0;
1334		*uaddr_len = sizeof(short);
1335	} else {
1336		struct unix_address *addr = u->addr;
1337
1338		*uaddr_len = addr->len;
1339		memcpy(sunaddr, addr->name, *uaddr_len);
1340	}
1341	unix_state_unlock(sk);
1342	sock_put(sk);
1343out:
1344	return err;
1345}
1346
1347static void unix_detach_fds(struct scm_cookie *scm, struct sk_buff *skb)
1348{
1349	int i;
1350
1351	scm->fp = UNIXCB(skb).fp;
1352	UNIXCB(skb).fp = NULL;
1353
1354	for (i = scm->fp->count-1; i >= 0; i--)
1355		unix_notinflight(scm->fp->fp[i]);
1356}
1357
1358static void unix_destruct_scm(struct sk_buff *skb)
1359{
1360	struct scm_cookie scm;
1361	memset(&scm, 0, sizeof(scm));
1362	scm.pid  = UNIXCB(skb).pid;
 
1363	if (UNIXCB(skb).fp)
1364		unix_detach_fds(&scm, skb);
1365
1366	/* Alas, it calls VFS */
1367	/* So fscking what? fput() had been SMP-safe since the last Summer */
1368	scm_destroy(&scm);
1369	sock_wfree(skb);
1370}
1371
1372#define MAX_RECURSION_LEVEL 4
1373
1374static int unix_attach_fds(struct scm_cookie *scm, struct sk_buff *skb)
1375{
1376	int i;
1377	unsigned char max_level = 0;
1378	int unix_sock_count = 0;
1379
1380	for (i = scm->fp->count - 1; i >= 0; i--) {
1381		struct sock *sk = unix_get_socket(scm->fp->fp[i]);
1382
1383		if (sk) {
1384			unix_sock_count++;
1385			max_level = max(max_level,
1386					unix_sk(sk)->recursion_level);
1387		}
1388	}
1389	if (unlikely(max_level > MAX_RECURSION_LEVEL))
1390		return -ETOOMANYREFS;
1391
1392	/*
1393	 * Need to duplicate file references for the sake of garbage
1394	 * collection.  Otherwise a socket in the fps might become a
1395	 * candidate for GC while the skb is not yet queued.
1396	 */
1397	UNIXCB(skb).fp = scm_fp_dup(scm->fp);
1398	if (!UNIXCB(skb).fp)
1399		return -ENOMEM;
1400
1401	if (unix_sock_count) {
1402		for (i = scm->fp->count - 1; i >= 0; i--)
1403			unix_inflight(scm->fp->fp[i]);
1404	}
1405	return max_level;
1406}
1407
1408static int unix_scm_to_skb(struct scm_cookie *scm, struct sk_buff *skb, bool send_fds)
1409{
1410	int err = 0;
1411
1412	UNIXCB(skb).pid  = get_pid(scm->pid);
1413	UNIXCB(skb).uid = scm->creds.uid;
1414	UNIXCB(skb).gid = scm->creds.gid;
1415	UNIXCB(skb).fp = NULL;
1416	if (scm->fp && send_fds)
1417		err = unix_attach_fds(scm, skb);
1418
1419	skb->destructor = unix_destruct_scm;
1420	return err;
1421}
1422
1423/*
1424 * Some apps rely on write() giving SCM_CREDENTIALS
1425 * We include credentials if source or destination socket
1426 * asserted SOCK_PASSCRED.
1427 */
1428static void maybe_add_creds(struct sk_buff *skb, const struct socket *sock,
1429			    const struct sock *other)
1430{
1431	if (UNIXCB(skb).pid)
1432		return;
1433	if (test_bit(SOCK_PASSCRED, &sock->flags) ||
1434	    !other->sk_socket ||
1435	    test_bit(SOCK_PASSCRED, &other->sk_socket->flags)) {
1436		UNIXCB(skb).pid  = get_pid(task_tgid(current));
1437		current_uid_gid(&UNIXCB(skb).uid, &UNIXCB(skb).gid);
1438	}
1439}
1440
1441/*
1442 *	Send AF_UNIX data.
1443 */
1444
1445static int unix_dgram_sendmsg(struct kiocb *kiocb, struct socket *sock,
1446			      struct msghdr *msg, size_t len)
1447{
1448	struct sock_iocb *siocb = kiocb_to_siocb(kiocb);
1449	struct sock *sk = sock->sk;
1450	struct net *net = sock_net(sk);
1451	struct unix_sock *u = unix_sk(sk);
1452	DECLARE_SOCKADDR(struct sockaddr_un *, sunaddr, msg->msg_name);
1453	struct sock *other = NULL;
1454	int namelen = 0; /* fake GCC */
1455	int err;
1456	unsigned int hash;
1457	struct sk_buff *skb;
1458	long timeo;
1459	struct scm_cookie tmp_scm;
1460	int max_level;
1461	int data_len = 0;
1462
1463	if (NULL == siocb->scm)
1464		siocb->scm = &tmp_scm;
1465	wait_for_unix_gc();
1466	err = scm_send(sock, msg, siocb->scm, false);
1467	if (err < 0)
1468		return err;
1469
1470	err = -EOPNOTSUPP;
1471	if (msg->msg_flags&MSG_OOB)
1472		goto out;
1473
1474	if (msg->msg_namelen) {
1475		err = unix_mkname(sunaddr, msg->msg_namelen, &hash);
1476		if (err < 0)
1477			goto out;
1478		namelen = err;
1479	} else {
1480		sunaddr = NULL;
1481		err = -ENOTCONN;
1482		other = unix_peer_get(sk);
1483		if (!other)
1484			goto out;
1485	}
1486
1487	if (test_bit(SOCK_PASSCRED, &sock->flags) && !u->addr
1488	    && (err = unix_autobind(sock)) != 0)
1489		goto out;
1490
1491	err = -EMSGSIZE;
1492	if (len > sk->sk_sndbuf - 32)
1493		goto out;
1494
1495	if (len > SKB_MAX_ALLOC)
1496		data_len = min_t(size_t,
1497				 len - SKB_MAX_ALLOC,
1498				 MAX_SKB_FRAGS * PAGE_SIZE);
1499
1500	skb = sock_alloc_send_pskb(sk, len - data_len, data_len,
1501				   msg->msg_flags & MSG_DONTWAIT, &err,
1502				   PAGE_ALLOC_COSTLY_ORDER);
1503	if (skb == NULL)
1504		goto out;
1505
1506	err = unix_scm_to_skb(siocb->scm, skb, true);
1507	if (err < 0)
1508		goto out_free;
1509	max_level = err + 1;
1510	unix_get_secdata(siocb->scm, skb);
1511
1512	skb_put(skb, len - data_len);
1513	skb->data_len = data_len;
1514	skb->len = len;
1515	err = skb_copy_datagram_from_iovec(skb, 0, msg->msg_iov, 0, len);
1516	if (err)
1517		goto out_free;
1518
1519	timeo = sock_sndtimeo(sk, msg->msg_flags & MSG_DONTWAIT);
1520
1521restart:
1522	if (!other) {
1523		err = -ECONNRESET;
1524		if (sunaddr == NULL)
1525			goto out_free;
1526
1527		other = unix_find_other(net, sunaddr, namelen, sk->sk_type,
1528					hash, &err);
1529		if (other == NULL)
1530			goto out_free;
1531	}
1532
1533	if (sk_filter(other, skb) < 0) {
1534		/* Toss the packet but do not return any error to the sender */
1535		err = len;
1536		goto out_free;
1537	}
1538
1539	unix_state_lock(other);
1540	err = -EPERM;
1541	if (!unix_may_send(sk, other))
1542		goto out_unlock;
1543
1544	if (sock_flag(other, SOCK_DEAD)) {
1545		/*
1546		 *	Check with 1003.1g - what should
1547		 *	datagram error
1548		 */
1549		unix_state_unlock(other);
1550		sock_put(other);
1551
1552		err = 0;
1553		unix_state_lock(sk);
1554		if (unix_peer(sk) == other) {
1555			unix_peer(sk) = NULL;
1556			unix_state_unlock(sk);
1557
1558			unix_dgram_disconnected(sk, other);
1559			sock_put(other);
1560			err = -ECONNREFUSED;
1561		} else {
1562			unix_state_unlock(sk);
1563		}
1564
1565		other = NULL;
1566		if (err)
1567			goto out_free;
1568		goto restart;
1569	}
1570
1571	err = -EPIPE;
1572	if (other->sk_shutdown & RCV_SHUTDOWN)
1573		goto out_unlock;
1574
1575	if (sk->sk_type != SOCK_SEQPACKET) {
1576		err = security_unix_may_send(sk->sk_socket, other->sk_socket);
1577		if (err)
1578			goto out_unlock;
1579	}
1580
1581	if (unix_peer(other) != sk && unix_recvq_full(other)) {
1582		if (!timeo) {
1583			err = -EAGAIN;
1584			goto out_unlock;
1585		}
1586
1587		timeo = unix_wait_for_peer(other, timeo);
1588
1589		err = sock_intr_errno(timeo);
1590		if (signal_pending(current))
1591			goto out_free;
1592
1593		goto restart;
1594	}
1595
1596	if (sock_flag(other, SOCK_RCVTSTAMP))
1597		__net_timestamp(skb);
1598	maybe_add_creds(skb, sock, other);
1599	skb_queue_tail(&other->sk_receive_queue, skb);
1600	if (max_level > unix_sk(other)->recursion_level)
1601		unix_sk(other)->recursion_level = max_level;
1602	unix_state_unlock(other);
1603	other->sk_data_ready(other);
1604	sock_put(other);
1605	scm_destroy(siocb->scm);
1606	return len;
1607
1608out_unlock:
1609	unix_state_unlock(other);
1610out_free:
1611	kfree_skb(skb);
1612out:
1613	if (other)
1614		sock_put(other);
1615	scm_destroy(siocb->scm);
1616	return err;
1617}
1618
1619/* We use paged skbs for stream sockets, and limit occupancy to 32768
1620 * bytes, and a minimun of a full page.
1621 */
1622#define UNIX_SKB_FRAGS_SZ (PAGE_SIZE << get_order(32768))
1623
1624static int unix_stream_sendmsg(struct kiocb *kiocb, struct socket *sock,
1625			       struct msghdr *msg, size_t len)
1626{
1627	struct sock_iocb *siocb = kiocb_to_siocb(kiocb);
1628	struct sock *sk = sock->sk;
1629	struct sock *other = NULL;
1630	int err, size;
1631	struct sk_buff *skb;
1632	int sent = 0;
1633	struct scm_cookie tmp_scm;
1634	bool fds_sent = false;
1635	int max_level;
1636	int data_len;
1637
1638	if (NULL == siocb->scm)
1639		siocb->scm = &tmp_scm;
1640	wait_for_unix_gc();
1641	err = scm_send(sock, msg, siocb->scm, false);
1642	if (err < 0)
1643		return err;
1644
1645	err = -EOPNOTSUPP;
1646	if (msg->msg_flags&MSG_OOB)
1647		goto out_err;
1648
1649	if (msg->msg_namelen) {
1650		err = sk->sk_state == TCP_ESTABLISHED ? -EISCONN : -EOPNOTSUPP;
1651		goto out_err;
1652	} else {
1653		err = -ENOTCONN;
1654		other = unix_peer(sk);
1655		if (!other)
1656			goto out_err;
1657	}
1658
1659	if (sk->sk_shutdown & SEND_SHUTDOWN)
1660		goto pipe_err;
1661
1662	while (sent < len) {
1663		size = len - sent;
 
 
 
 
 
1664
1665		/* Keep two messages in the pipe so it schedules better */
1666		size = min_t(int, size, (sk->sk_sndbuf >> 1) - 64);
 
 
 
 
1667
1668		/* allow fallback to order-0 allocations */
1669		size = min_t(int, size, SKB_MAX_HEAD(0) + UNIX_SKB_FRAGS_SZ);
 
1670
1671		data_len = max_t(int, 0, size - SKB_MAX_HEAD(0));
 
1672
1673		skb = sock_alloc_send_pskb(sk, size - data_len, data_len,
1674					   msg->msg_flags & MSG_DONTWAIT, &err,
1675					   get_order(UNIX_SKB_FRAGS_SZ));
1676		if (!skb)
1677			goto out_err;
1678
 
 
 
 
 
 
 
 
 
 
1679		/* Only send the fds in the first buffer */
1680		err = unix_scm_to_skb(siocb->scm, skb, !fds_sent);
1681		if (err < 0) {
1682			kfree_skb(skb);
1683			goto out_err;
1684		}
1685		max_level = err + 1;
1686		fds_sent = true;
1687
1688		skb_put(skb, size - data_len);
1689		skb->data_len = data_len;
1690		skb->len = size;
1691		err = skb_copy_datagram_from_iovec(skb, 0, msg->msg_iov,
1692						   sent, size);
1693		if (err) {
1694			kfree_skb(skb);
1695			goto out_err;
1696		}
1697
1698		unix_state_lock(other);
1699
1700		if (sock_flag(other, SOCK_DEAD) ||
1701		    (other->sk_shutdown & RCV_SHUTDOWN))
1702			goto pipe_err_free;
1703
1704		maybe_add_creds(skb, sock, other);
1705		skb_queue_tail(&other->sk_receive_queue, skb);
1706		if (max_level > unix_sk(other)->recursion_level)
1707			unix_sk(other)->recursion_level = max_level;
1708		unix_state_unlock(other);
1709		other->sk_data_ready(other);
1710		sent += size;
1711	}
1712
1713	scm_destroy(siocb->scm);
1714	siocb->scm = NULL;
1715
1716	return sent;
1717
1718pipe_err_free:
1719	unix_state_unlock(other);
1720	kfree_skb(skb);
1721pipe_err:
1722	if (sent == 0 && !(msg->msg_flags&MSG_NOSIGNAL))
1723		send_sig(SIGPIPE, current, 0);
1724	err = -EPIPE;
1725out_err:
1726	scm_destroy(siocb->scm);
1727	siocb->scm = NULL;
1728	return sent ? : err;
1729}
1730
1731static int unix_seqpacket_sendmsg(struct kiocb *kiocb, struct socket *sock,
1732				  struct msghdr *msg, size_t len)
1733{
1734	int err;
1735	struct sock *sk = sock->sk;
1736
1737	err = sock_error(sk);
1738	if (err)
1739		return err;
1740
1741	if (sk->sk_state != TCP_ESTABLISHED)
1742		return -ENOTCONN;
1743
1744	if (msg->msg_namelen)
1745		msg->msg_namelen = 0;
1746
1747	return unix_dgram_sendmsg(kiocb, sock, msg, len);
1748}
1749
1750static int unix_seqpacket_recvmsg(struct kiocb *iocb, struct socket *sock,
1751			      struct msghdr *msg, size_t size,
1752			      int flags)
1753{
1754	struct sock *sk = sock->sk;
1755
1756	if (sk->sk_state != TCP_ESTABLISHED)
1757		return -ENOTCONN;
1758
1759	return unix_dgram_recvmsg(iocb, sock, msg, size, flags);
1760}
1761
1762static void unix_copy_addr(struct msghdr *msg, struct sock *sk)
1763{
1764	struct unix_sock *u = unix_sk(sk);
1765
 
1766	if (u->addr) {
1767		msg->msg_namelen = u->addr->len;
1768		memcpy(msg->msg_name, u->addr->name, u->addr->len);
1769	}
1770}
1771
1772static int unix_dgram_recvmsg(struct kiocb *iocb, struct socket *sock,
1773			      struct msghdr *msg, size_t size,
1774			      int flags)
1775{
1776	struct sock_iocb *siocb = kiocb_to_siocb(iocb);
1777	struct scm_cookie tmp_scm;
1778	struct sock *sk = sock->sk;
1779	struct unix_sock *u = unix_sk(sk);
1780	int noblock = flags & MSG_DONTWAIT;
1781	struct sk_buff *skb;
1782	int err;
1783	int peeked, skip;
1784
1785	err = -EOPNOTSUPP;
1786	if (flags&MSG_OOB)
1787		goto out;
1788
 
 
1789	err = mutex_lock_interruptible(&u->readlock);
1790	if (unlikely(err)) {
1791		/* recvmsg() in non blocking mode is supposed to return -EAGAIN
1792		 * sk_rcvtimeo is not honored by mutex_lock_interruptible()
1793		 */
1794		err = noblock ? -EAGAIN : -ERESTARTSYS;
1795		goto out;
1796	}
1797
1798	skip = sk_peek_offset(sk, flags);
1799
1800	skb = __skb_recv_datagram(sk, flags, &peeked, &skip, &err);
1801	if (!skb) {
1802		unix_state_lock(sk);
1803		/* Signal EOF on disconnected non-blocking SEQPACKET socket. */
1804		if (sk->sk_type == SOCK_SEQPACKET && err == -EAGAIN &&
1805		    (sk->sk_shutdown & RCV_SHUTDOWN))
1806			err = 0;
1807		unix_state_unlock(sk);
1808		goto out_unlock;
1809	}
1810
1811	wake_up_interruptible_sync_poll(&u->peer_wait,
1812					POLLOUT | POLLWRNORM | POLLWRBAND);
1813
1814	if (msg->msg_name)
1815		unix_copy_addr(msg, skb->sk);
1816
1817	if (size > skb->len - skip)
1818		size = skb->len - skip;
1819	else if (size < skb->len - skip)
1820		msg->msg_flags |= MSG_TRUNC;
1821
1822	err = skb_copy_datagram_iovec(skb, skip, msg->msg_iov, size);
1823	if (err)
1824		goto out_free;
1825
1826	if (sock_flag(sk, SOCK_RCVTSTAMP))
1827		__sock_recv_timestamp(msg, sk, skb);
1828
1829	if (!siocb->scm) {
1830		siocb->scm = &tmp_scm;
1831		memset(&tmp_scm, 0, sizeof(tmp_scm));
1832	}
1833	scm_set_cred(siocb->scm, UNIXCB(skb).pid, UNIXCB(skb).uid, UNIXCB(skb).gid);
1834	unix_set_secdata(siocb->scm, skb);
1835
1836	if (!(flags & MSG_PEEK)) {
1837		if (UNIXCB(skb).fp)
1838			unix_detach_fds(siocb->scm, skb);
1839
1840		sk_peek_offset_bwd(sk, skb->len);
1841	} else {
1842		/* It is questionable: on PEEK we could:
1843		   - do not return fds - good, but too simple 8)
1844		   - return fds, and do not return them on read (old strategy,
1845		     apparently wrong)
1846		   - clone fds (I chose it for now, it is the most universal
1847		     solution)
1848
1849		   POSIX 1003.1g does not actually define this clearly
1850		   at all. POSIX 1003.1g doesn't define a lot of things
1851		   clearly however!
1852
1853		*/
1854
1855		sk_peek_offset_fwd(sk, size);
1856
1857		if (UNIXCB(skb).fp)
1858			siocb->scm->fp = scm_fp_dup(UNIXCB(skb).fp);
1859	}
1860	err = (flags & MSG_TRUNC) ? skb->len - skip : size;
1861
1862	scm_recv(sock, msg, siocb->scm, flags);
1863
1864out_free:
1865	skb_free_datagram(sk, skb);
1866out_unlock:
1867	mutex_unlock(&u->readlock);
1868out:
1869	return err;
1870}
1871
1872/*
1873 *	Sleep until more data has arrived. But check for races..
1874 */
1875static long unix_stream_data_wait(struct sock *sk, long timeo,
1876				  struct sk_buff *last)
1877{
1878	DEFINE_WAIT(wait);
1879
1880	unix_state_lock(sk);
1881
1882	for (;;) {
1883		prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
1884
1885		if (skb_peek_tail(&sk->sk_receive_queue) != last ||
1886		    sk->sk_err ||
1887		    (sk->sk_shutdown & RCV_SHUTDOWN) ||
1888		    signal_pending(current) ||
1889		    !timeo)
1890			break;
1891
1892		set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1893		unix_state_unlock(sk);
1894		timeo = freezable_schedule_timeout(timeo);
1895		unix_state_lock(sk);
1896		clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1897	}
1898
1899	finish_wait(sk_sleep(sk), &wait);
1900	unix_state_unlock(sk);
1901	return timeo;
1902}
1903
1904static unsigned int unix_skb_len(const struct sk_buff *skb)
1905{
1906	return skb->len - UNIXCB(skb).consumed;
1907}
1908
1909static int unix_stream_recvmsg(struct kiocb *iocb, struct socket *sock,
1910			       struct msghdr *msg, size_t size,
1911			       int flags)
1912{
1913	struct sock_iocb *siocb = kiocb_to_siocb(iocb);
1914	struct scm_cookie tmp_scm;
1915	struct sock *sk = sock->sk;
1916	struct unix_sock *u = unix_sk(sk);
1917	DECLARE_SOCKADDR(struct sockaddr_un *, sunaddr, msg->msg_name);
1918	int copied = 0;
1919	int noblock = flags & MSG_DONTWAIT;
1920	int check_creds = 0;
1921	int target;
1922	int err = 0;
1923	long timeo;
1924	int skip;
1925
1926	err = -EINVAL;
1927	if (sk->sk_state != TCP_ESTABLISHED)
1928		goto out;
1929
1930	err = -EOPNOTSUPP;
1931	if (flags&MSG_OOB)
1932		goto out;
1933
1934	target = sock_rcvlowat(sk, flags&MSG_WAITALL, size);
1935	timeo = sock_rcvtimeo(sk, noblock);
 
 
1936
1937	/* Lock the socket to prevent queue disordering
1938	 * while sleeps in memcpy_tomsg
1939	 */
1940
1941	if (!siocb->scm) {
1942		siocb->scm = &tmp_scm;
1943		memset(&tmp_scm, 0, sizeof(tmp_scm));
1944	}
1945
1946	err = mutex_lock_interruptible(&u->readlock);
1947	if (unlikely(err)) {
1948		/* recvmsg() in non blocking mode is supposed to return -EAGAIN
1949		 * sk_rcvtimeo is not honored by mutex_lock_interruptible()
1950		 */
1951		err = noblock ? -EAGAIN : -ERESTARTSYS;
1952		goto out;
1953	}
1954
 
 
1955	do {
1956		int chunk;
1957		struct sk_buff *skb, *last;
1958
1959		unix_state_lock(sk);
1960		last = skb = skb_peek(&sk->sk_receive_queue);
1961again:
1962		if (skb == NULL) {
1963			unix_sk(sk)->recursion_level = 0;
1964			if (copied >= target)
1965				goto unlock;
1966
1967			/*
1968			 *	POSIX 1003.1g mandates this order.
1969			 */
1970
1971			err = sock_error(sk);
1972			if (err)
1973				goto unlock;
1974			if (sk->sk_shutdown & RCV_SHUTDOWN)
1975				goto unlock;
1976
1977			unix_state_unlock(sk);
1978			err = -EAGAIN;
1979			if (!timeo)
1980				break;
1981			mutex_unlock(&u->readlock);
1982
1983			timeo = unix_stream_data_wait(sk, timeo, last);
1984
1985			if (signal_pending(current)
1986			    ||  mutex_lock_interruptible(&u->readlock)) {
1987				err = sock_intr_errno(timeo);
1988				goto out;
1989			}
1990
1991			continue;
1992 unlock:
1993			unix_state_unlock(sk);
1994			break;
1995		}
1996
1997		skip = sk_peek_offset(sk, flags);
1998		while (skip >= unix_skb_len(skb)) {
1999			skip -= unix_skb_len(skb);
2000			last = skb;
2001			skb = skb_peek_next(skb, &sk->sk_receive_queue);
2002			if (!skb)
2003				goto again;
2004		}
2005
2006		unix_state_unlock(sk);
2007
2008		if (check_creds) {
2009			/* Never glue messages from different writers */
2010			if ((UNIXCB(skb).pid  != siocb->scm->pid) ||
2011			    !uid_eq(UNIXCB(skb).uid, siocb->scm->creds.uid) ||
2012			    !gid_eq(UNIXCB(skb).gid, siocb->scm->creds.gid))
2013				break;
2014		} else if (test_bit(SOCK_PASSCRED, &sock->flags)) {
2015			/* Copy credentials */
2016			scm_set_cred(siocb->scm, UNIXCB(skb).pid, UNIXCB(skb).uid, UNIXCB(skb).gid);
2017			check_creds = 1;
2018		}
2019
2020		/* Copy address just once */
2021		if (sunaddr) {
2022			unix_copy_addr(msg, skb->sk);
2023			sunaddr = NULL;
2024		}
2025
2026		chunk = min_t(unsigned int, unix_skb_len(skb) - skip, size);
2027		if (skb_copy_datagram_iovec(skb, UNIXCB(skb).consumed + skip,
2028					    msg->msg_iov, chunk)) {
2029			if (copied == 0)
2030				copied = -EFAULT;
2031			break;
2032		}
2033		copied += chunk;
2034		size -= chunk;
2035
2036		/* Mark read part of skb as used */
2037		if (!(flags & MSG_PEEK)) {
2038			UNIXCB(skb).consumed += chunk;
2039
2040			sk_peek_offset_bwd(sk, chunk);
2041
2042			if (UNIXCB(skb).fp)
2043				unix_detach_fds(siocb->scm, skb);
2044
2045			if (unix_skb_len(skb))
2046				break;
2047
2048			skb_unlink(skb, &sk->sk_receive_queue);
2049			consume_skb(skb);
2050
2051			if (siocb->scm->fp)
2052				break;
2053		} else {
2054			/* It is questionable, see note in unix_dgram_recvmsg.
2055			 */
2056			if (UNIXCB(skb).fp)
2057				siocb->scm->fp = scm_fp_dup(UNIXCB(skb).fp);
2058
2059			sk_peek_offset_fwd(sk, chunk);
2060
2061			break;
2062		}
2063	} while (size);
2064
2065	mutex_unlock(&u->readlock);
2066	scm_recv(sock, msg, siocb->scm, flags);
2067out:
2068	return copied ? : err;
2069}
2070
2071static int unix_shutdown(struct socket *sock, int mode)
2072{
2073	struct sock *sk = sock->sk;
2074	struct sock *other;
2075
2076	if (mode < SHUT_RD || mode > SHUT_RDWR)
2077		return -EINVAL;
2078	/* This maps:
2079	 * SHUT_RD   (0) -> RCV_SHUTDOWN  (1)
2080	 * SHUT_WR   (1) -> SEND_SHUTDOWN (2)
2081	 * SHUT_RDWR (2) -> SHUTDOWN_MASK (3)
2082	 */
2083	++mode;
2084
2085	unix_state_lock(sk);
2086	sk->sk_shutdown |= mode;
2087	other = unix_peer(sk);
2088	if (other)
2089		sock_hold(other);
2090	unix_state_unlock(sk);
2091	sk->sk_state_change(sk);
2092
2093	if (other &&
2094		(sk->sk_type == SOCK_STREAM || sk->sk_type == SOCK_SEQPACKET)) {
2095
2096		int peer_mode = 0;
2097
2098		if (mode&RCV_SHUTDOWN)
2099			peer_mode |= SEND_SHUTDOWN;
2100		if (mode&SEND_SHUTDOWN)
2101			peer_mode |= RCV_SHUTDOWN;
2102		unix_state_lock(other);
2103		other->sk_shutdown |= peer_mode;
2104		unix_state_unlock(other);
2105		other->sk_state_change(other);
2106		if (peer_mode == SHUTDOWN_MASK)
2107			sk_wake_async(other, SOCK_WAKE_WAITD, POLL_HUP);
2108		else if (peer_mode & RCV_SHUTDOWN)
2109			sk_wake_async(other, SOCK_WAKE_WAITD, POLL_IN);
2110	}
2111	if (other)
2112		sock_put(other);
2113
2114	return 0;
2115}
2116
2117long unix_inq_len(struct sock *sk)
2118{
2119	struct sk_buff *skb;
2120	long amount = 0;
2121
2122	if (sk->sk_state == TCP_LISTEN)
2123		return -EINVAL;
2124
2125	spin_lock(&sk->sk_receive_queue.lock);
2126	if (sk->sk_type == SOCK_STREAM ||
2127	    sk->sk_type == SOCK_SEQPACKET) {
2128		skb_queue_walk(&sk->sk_receive_queue, skb)
2129			amount += unix_skb_len(skb);
2130	} else {
2131		skb = skb_peek(&sk->sk_receive_queue);
2132		if (skb)
2133			amount = skb->len;
2134	}
2135	spin_unlock(&sk->sk_receive_queue.lock);
2136
2137	return amount;
2138}
2139EXPORT_SYMBOL_GPL(unix_inq_len);
2140
2141long unix_outq_len(struct sock *sk)
2142{
2143	return sk_wmem_alloc_get(sk);
2144}
2145EXPORT_SYMBOL_GPL(unix_outq_len);
2146
2147static int unix_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg)
2148{
2149	struct sock *sk = sock->sk;
2150	long amount = 0;
2151	int err;
2152
2153	switch (cmd) {
2154	case SIOCOUTQ:
2155		amount = unix_outq_len(sk);
2156		err = put_user(amount, (int __user *)arg);
2157		break;
2158	case SIOCINQ:
2159		amount = unix_inq_len(sk);
2160		if (amount < 0)
2161			err = amount;
2162		else
2163			err = put_user(amount, (int __user *)arg);
2164		break;
2165	default:
2166		err = -ENOIOCTLCMD;
2167		break;
2168	}
2169	return err;
2170}
2171
2172static unsigned int unix_poll(struct file *file, struct socket *sock, poll_table *wait)
2173{
2174	struct sock *sk = sock->sk;
2175	unsigned int mask;
2176
2177	sock_poll_wait(file, sk_sleep(sk), wait);
2178	mask = 0;
2179
2180	/* exceptional events? */
2181	if (sk->sk_err)
2182		mask |= POLLERR;
2183	if (sk->sk_shutdown == SHUTDOWN_MASK)
2184		mask |= POLLHUP;
2185	if (sk->sk_shutdown & RCV_SHUTDOWN)
2186		mask |= POLLRDHUP | POLLIN | POLLRDNORM;
2187
2188	/* readable? */
2189	if (!skb_queue_empty(&sk->sk_receive_queue))
2190		mask |= POLLIN | POLLRDNORM;
2191
2192	/* Connection-based need to check for termination and startup */
2193	if ((sk->sk_type == SOCK_STREAM || sk->sk_type == SOCK_SEQPACKET) &&
2194	    sk->sk_state == TCP_CLOSE)
2195		mask |= POLLHUP;
2196
2197	/*
2198	 * we set writable also when the other side has shut down the
2199	 * connection. This prevents stuck sockets.
2200	 */
2201	if (unix_writable(sk))
2202		mask |= POLLOUT | POLLWRNORM | POLLWRBAND;
2203
2204	return mask;
2205}
2206
2207static unsigned int unix_dgram_poll(struct file *file, struct socket *sock,
2208				    poll_table *wait)
2209{
2210	struct sock *sk = sock->sk, *other;
2211	unsigned int mask, writable;
2212
2213	sock_poll_wait(file, sk_sleep(sk), wait);
2214	mask = 0;
2215
2216	/* exceptional events? */
2217	if (sk->sk_err || !skb_queue_empty(&sk->sk_error_queue))
2218		mask |= POLLERR |
2219			(sock_flag(sk, SOCK_SELECT_ERR_QUEUE) ? POLLPRI : 0);
2220
2221	if (sk->sk_shutdown & RCV_SHUTDOWN)
2222		mask |= POLLRDHUP | POLLIN | POLLRDNORM;
2223	if (sk->sk_shutdown == SHUTDOWN_MASK)
2224		mask |= POLLHUP;
2225
2226	/* readable? */
2227	if (!skb_queue_empty(&sk->sk_receive_queue))
2228		mask |= POLLIN | POLLRDNORM;
2229
2230	/* Connection-based need to check for termination and startup */
2231	if (sk->sk_type == SOCK_SEQPACKET) {
2232		if (sk->sk_state == TCP_CLOSE)
2233			mask |= POLLHUP;
2234		/* connection hasn't started yet? */
2235		if (sk->sk_state == TCP_SYN_SENT)
2236			return mask;
2237	}
2238
2239	/* No write status requested, avoid expensive OUT tests. */
2240	if (!(poll_requested_events(wait) & (POLLWRBAND|POLLWRNORM|POLLOUT)))
2241		return mask;
2242
2243	writable = unix_writable(sk);
2244	other = unix_peer_get(sk);
2245	if (other) {
2246		if (unix_peer(other) != sk) {
2247			sock_poll_wait(file, &unix_sk(other)->peer_wait, wait);
2248			if (unix_recvq_full(other))
2249				writable = 0;
2250		}
2251		sock_put(other);
2252	}
2253
2254	if (writable)
2255		mask |= POLLOUT | POLLWRNORM | POLLWRBAND;
2256	else
2257		set_bit(SOCK_ASYNC_NOSPACE, &sk->sk_socket->flags);
2258
2259	return mask;
2260}
2261
2262#ifdef CONFIG_PROC_FS
 
 
 
 
 
 
 
 
2263
2264#define BUCKET_SPACE (BITS_PER_LONG - (UNIX_HASH_BITS + 1) - 1)
2265
2266#define get_bucket(x) ((x) >> BUCKET_SPACE)
2267#define get_offset(x) ((x) & ((1L << BUCKET_SPACE) - 1))
2268#define set_bucket_offset(b, o) ((b) << BUCKET_SPACE | (o))
2269
2270static struct sock *unix_from_bucket(struct seq_file *seq, loff_t *pos)
2271{
2272	unsigned long offset = get_offset(*pos);
2273	unsigned long bucket = get_bucket(*pos);
2274	struct sock *sk;
2275	unsigned long count = 0;
2276
2277	for (sk = sk_head(&unix_socket_table[bucket]); sk; sk = sk_next(sk)) {
2278		if (sock_net(sk) != seq_file_net(seq))
2279			continue;
2280		if (++count == offset)
2281			break;
2282	}
 
 
2283
2284	return sk;
2285}
 
 
2286
2287static struct sock *unix_next_socket(struct seq_file *seq,
2288				     struct sock *sk,
2289				     loff_t *pos)
2290{
2291	unsigned long bucket;
 
 
2292
2293	while (sk > (struct sock *)SEQ_START_TOKEN) {
2294		sk = sk_next(sk);
2295		if (!sk)
2296			goto next_bucket;
2297		if (sock_net(sk) == seq_file_net(seq))
2298			return sk;
2299	}
2300
2301	do {
2302		sk = unix_from_bucket(seq, pos);
2303		if (sk)
2304			return sk;
2305
2306next_bucket:
2307		bucket = get_bucket(*pos) + 1;
2308		*pos = set_bucket_offset(bucket, 1);
2309	} while (bucket < ARRAY_SIZE(unix_socket_table));
2310
2311	return NULL;
2312}
2313
2314static void *unix_seq_start(struct seq_file *seq, loff_t *pos)
2315	__acquires(unix_table_lock)
2316{
2317	spin_lock(&unix_table_lock);
2318
2319	if (!*pos)
2320		return SEQ_START_TOKEN;
2321
2322	if (get_bucket(*pos) >= ARRAY_SIZE(unix_socket_table))
2323		return NULL;
2324
2325	return unix_next_socket(seq, NULL, pos);
2326}
2327
2328static void *unix_seq_next(struct seq_file *seq, void *v, loff_t *pos)
2329{
 
 
2330	++*pos;
2331	return unix_next_socket(seq, v, pos);
 
 
 
 
 
 
 
2332}
2333
2334static void unix_seq_stop(struct seq_file *seq, void *v)
2335	__releases(unix_table_lock)
2336{
2337	spin_unlock(&unix_table_lock);
2338}
2339
2340static int unix_seq_show(struct seq_file *seq, void *v)
2341{
2342
2343	if (v == SEQ_START_TOKEN)
2344		seq_puts(seq, "Num       RefCount Protocol Flags    Type St "
2345			 "Inode Path\n");
2346	else {
2347		struct sock *s = v;
2348		struct unix_sock *u = unix_sk(s);
2349		unix_state_lock(s);
2350
2351		seq_printf(seq, "%pK: %08X %08X %08X %04X %02X %5lu",
2352			s,
2353			atomic_read(&s->sk_refcnt),
2354			0,
2355			s->sk_state == TCP_LISTEN ? __SO_ACCEPTCON : 0,
2356			s->sk_type,
2357			s->sk_socket ?
2358			(s->sk_state == TCP_ESTABLISHED ? SS_CONNECTED : SS_UNCONNECTED) :
2359			(s->sk_state == TCP_ESTABLISHED ? SS_CONNECTING : SS_DISCONNECTING),
2360			sock_i_ino(s));
2361
2362		if (u->addr) {
2363			int i, len;
2364			seq_putc(seq, ' ');
2365
2366			i = 0;
2367			len = u->addr->len - sizeof(short);
2368			if (!UNIX_ABSTRACT(s))
2369				len--;
2370			else {
2371				seq_putc(seq, '@');
2372				i++;
2373			}
2374			for ( ; i < len; i++)
2375				seq_putc(seq, u->addr->name->sun_path[i]);
2376		}
2377		unix_state_unlock(s);
2378		seq_putc(seq, '\n');
2379	}
2380
2381	return 0;
2382}
2383
2384static const struct seq_operations unix_seq_ops = {
2385	.start  = unix_seq_start,
2386	.next   = unix_seq_next,
2387	.stop   = unix_seq_stop,
2388	.show   = unix_seq_show,
2389};
2390
2391static int unix_seq_open(struct inode *inode, struct file *file)
2392{
2393	return seq_open_net(inode, file, &unix_seq_ops,
2394			    sizeof(struct seq_net_private));
2395}
2396
2397static const struct file_operations unix_seq_fops = {
2398	.owner		= THIS_MODULE,
2399	.open		= unix_seq_open,
2400	.read		= seq_read,
2401	.llseek		= seq_lseek,
2402	.release	= seq_release_net,
2403};
2404
2405#endif
2406
2407static const struct net_proto_family unix_family_ops = {
2408	.family = PF_UNIX,
2409	.create = unix_create,
2410	.owner	= THIS_MODULE,
2411};
2412
2413
2414static int __net_init unix_net_init(struct net *net)
2415{
2416	int error = -ENOMEM;
2417
2418	net->unx.sysctl_max_dgram_qlen = 10;
2419	if (unix_sysctl_register(net))
2420		goto out;
2421
2422#ifdef CONFIG_PROC_FS
2423	if (!proc_create("unix", 0, net->proc_net, &unix_seq_fops)) {
2424		unix_sysctl_unregister(net);
2425		goto out;
2426	}
2427#endif
2428	error = 0;
2429out:
2430	return error;
2431}
2432
2433static void __net_exit unix_net_exit(struct net *net)
2434{
2435	unix_sysctl_unregister(net);
2436	remove_proc_entry("unix", net->proc_net);
2437}
2438
2439static struct pernet_operations unix_net_ops = {
2440	.init = unix_net_init,
2441	.exit = unix_net_exit,
2442};
2443
2444static int __init af_unix_init(void)
2445{
2446	int rc = -1;
 
2447
2448	BUILD_BUG_ON(sizeof(struct unix_skb_parms) > FIELD_SIZEOF(struct sk_buff, cb));
2449
2450	rc = proto_register(&unix_proto, 1);
2451	if (rc != 0) {
2452		pr_crit("%s: Cannot create unix_sock SLAB cache!\n", __func__);
 
2453		goto out;
2454	}
2455
2456	sock_register(&unix_family_ops);
2457	register_pernet_subsys(&unix_net_ops);
2458out:
2459	return rc;
2460}
2461
2462static void __exit af_unix_exit(void)
2463{
2464	sock_unregister(PF_UNIX);
2465	proto_unregister(&unix_proto);
2466	unregister_pernet_subsys(&unix_net_ops);
2467}
2468
2469/* Earlier than device_initcall() so that other drivers invoking
2470   request_module() don't end up in a loop when modprobe tries
2471   to use a UNIX socket. But later than subsys_initcall() because
2472   we depend on stuff initialised there */
2473fs_initcall(af_unix_init);
2474module_exit(af_unix_exit);
2475
2476MODULE_LICENSE("GPL");
2477MODULE_ALIAS_NETPROTO(PF_UNIX);