sock.h 30.9 KB
Newer Older
Linus Torvalds's avatar
Linus Torvalds committed
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
/*
 * INET		An implementation of the TCP/IP protocol suite for the LINUX
 *		operating system.  INET is implemented using the  BSD Socket
 *		interface as the means of communication with the user level.
 *
 *		Definitions for the AF_INET socket handler.
 *
 * Version:	@(#)sock.h	1.0.4	05/13/93
 *
 * Authors:	Ross Biro, <bir7@leland.Stanford.Edu>
 *		Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
 *		Corey Minyard <wf-rch!minyard@relay.EU.net>
 *		Florian La Roche <flla@stud.uni-sb.de>
 *
 * Fixes:
 *		Alan Cox	:	Volatiles in skbuff pointers. See
 *					skbuff comments. May be overdone,
 *					better to prove they can be removed
 *					than the reverse.
 *		Alan Cox	:	Added a zapped field for tcp to note
 *					a socket is reset and must stay shut up
 *		Alan Cox	:	New fields for options
 *	Pauline Middelink	:	identd support
 *		Alan Cox	:	Eliminate low level recv/recvfrom
 *		David S. Miller	:	New socket lookup architecture.
 *              Steve Whitehouse:       Default routines for sock_ops
27 28 29 30 31
 *              Arnaldo C. Melo :	removed net_pinfo, tp_pinfo and made
 *              			protinfo be just a void pointer, as the
 *              			protocol specific parts were moved to
 *              			respective headers and ipv4/v6, etc now
 *              			use private slabcaches for its socks
32 33
 *              Pedro Hortas	:	New flags field for socket options
 *
Linus Torvalds's avatar
Linus Torvalds committed
34 35 36 37 38 39 40 41 42 43
 *
 *		This program is free software; you can redistribute it and/or
 *		modify it under the terms of the GNU General Public License
 *		as published by the Free Software Foundation; either version
 *		2 of the License, or (at your option) any later version.
 */
#ifndef _SOCK_H
#define _SOCK_H

#include <linux/config.h>
44
#include <linux/list.h>
Linus Torvalds's avatar
Linus Torvalds committed
45 46
#include <linux/timer.h>
#include <linux/cache.h>
47
#include <linux/module.h>
Linus Torvalds's avatar
Linus Torvalds committed
48 49
#include <linux/netdevice.h>
#include <linux/skbuff.h>	/* struct sk_buff */
50
#include <linux/security.h>
Linus Torvalds's avatar
Linus Torvalds committed
51 52 53 54 55 56 57 58 59 60 61 62

#include <linux/filter.h>

#include <asm/atomic.h>
#include <net/dst.h>

/*
 * This structure really needs to be cleaned up.
 * Most of it is for TCP, and not used by any of
 * the other protocols.
 */

63
/* Define this to get the sk->sk_debug debugging facility. */
Linus Torvalds's avatar
Linus Torvalds committed
64 65
#define SOCK_DEBUGGING
#ifdef SOCK_DEBUGGING
66 67
#define SOCK_DEBUG(sk, msg...) do { if ((sk) && ((sk)->sk_debug)) \
					printk(KERN_DEBUG msg); } while (0)
Linus Torvalds's avatar
Linus Torvalds committed
68 69 70 71 72 73 74 75
#else
#define SOCK_DEBUG(sk, msg...) do { } while (0)
#endif

/* This is the per-socket lock.  The spinlock provides a synchronization
 * between user contexts and software interrupt processing, whereas the
 * mini-semaphore synchronizes multiple users amongst themselves.
 */
76
struct sock_iocb;
Linus Torvalds's avatar
Linus Torvalds committed
77 78
typedef struct {
	spinlock_t		slock;
79
	struct sock_iocb	*owner;
Linus Torvalds's avatar
Linus Torvalds committed
80 81 82 83
	wait_queue_head_t	wq;
} socket_lock_t;

#define sock_lock_init(__sk) \
84 85 86
do {	spin_lock_init(&((__sk)->sk_lock.slock)); \
	(__sk)->sk_lock.owner = NULL; \
	init_waitqueue_head(&((__sk)->sk_lock.wq)); \
87
} while(0)
Linus Torvalds's avatar
Linus Torvalds committed
88

89 90 91 92 93 94 95 96
struct sock;

/**
  *	struct sock_common - minimal network layer representation of sockets
  *	@skc_family - network address family
  *	@skc_state - Connection state
  *	@skc_reuse - %SO_REUSEADDR setting
  *	@skc_bound_dev_if - bound device index if != 0
97 98
  *	@skc_node - main hash linkage for various protocol lookup tables
  *	@skc_bind_node - bind hash linkage for various protocol lookup tables
99 100 101 102 103 104 105 106 107 108
  *	@skc_refcnt - reference count
  *
  *	This is the minimal network layer representation of sockets, the header
  *	for struct sock and struct tcp_tw_bucket.
  */
struct sock_common {
	unsigned short		skc_family;
	volatile unsigned char	skc_state;
	unsigned char		skc_reuse;
	int			skc_bound_dev_if;
109 110
	struct hlist_node	skc_node;
	struct hlist_node	skc_bind_node;
111 112 113
	atomic_t		skc_refcnt;
};

114 115
/**
  *	struct sock - network layer representation of sockets
116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174
  *	@__sk_common - shared layout with tcp_tw_bucket
  *	@sk_zapped - ax25 & ipx means !linked
  *	@sk_shutdown - mask of %SEND_SHUTDOWN and/or %RCV_SHUTDOWN
  *	@sk_use_write_queue - wheter to call sk->sk_write_space in sock_wfree
  *	@sk_userlocks - %SO_SNDBUF and %SO_RCVBUF settings
  *	@sk_lock -	synchronizer
  *	@sk_rcvbuf - size of receive buffer in bytes
  *	@sk_sleep - sock wait queue
  *	@sk_dst_cache - destination cache
  *	@sk_dst_lock - destination cache lock
  *	@sk_policy - flow policy
  *	@sk_rmem_alloc - receive queue bytes committed
  *	@sk_receive_queue - incoming packets
  *	@sk_wmem_alloc - transmit queue bytes committed
  *	@sk_write_queue - Packet sending queue
  *	@sk_omem_alloc - "o" is "option" or "other"
  *	@sk_wmem_queued - persistent queue size
  *	@sk_forward_alloc - space allocated forward
  *	@sk_allocation - allocation mode
  *	@sk_sndbuf - size of send buffer in bytes
  *	@sk_flags - %SO_LINGER (l_onoff), %SO_BROADCAST, %SO_KEEPALIVE, %SO_OOBINLINE settings
  *	@sk_no_check - %SO_NO_CHECK setting, wether or not checkup packets
  *	@sk_debug - %SO_DEBUG setting
  *	@sk_rcvtstamp - %SO_TIMESTAMP setting
  *	@sk_no_largesend - whether to sent large segments or not
  *	@sk_route_caps - route capabilities (e.g. %NETIF_F_TSO)
  *	@sk_lingertime - %SO_LINGER l_linger setting
  *	@sk_hashent - hash entry in several tables (e.g. tcp_ehash)
  *	@sk_pair - socket pair (e.g. AF_UNIX/unix_peer)
  *	@sk_backlog - always used with the per-socket spinlock held
  *	@sk_callback_lock - used with the callbacks in the end of this struct
  *	@sk_error_queue - rarely used
  *	@sk_prot - protocol handlers inside a network family
  *	@sk_err - last error
  *	@sk_err_soft - errors that don't cause failure but are the cause of a persistent failure not just 'timed out'
  *	@sk_ack_backlog - current listen backlog
  *	@sk_max_ack_backlog - listen backlog set in listen()
  *	@sk_priority - %SO_PRIORITY setting
  *	@sk_type - socket type (%SOCK_STREAM, etc)
  *	@sk_localroute - route locally only, %SO_DONTROUTE setting
  *	@sk_protocol - which protocol this socket belongs in this network family
  *	@sk_peercred - %SO_PEERCRED setting
  *	@sk_rcvlowat - %SO_RCVLOWAT setting
  *	@sk_rcvtimeo - %SO_RCVTIMEO setting
  *	@sk_sndtimeo - %SO_SNDTIMEO setting
  *	@sk_filter - socket filtering instructions
  *	@sk_protinfo - private area, net family specific, when not using slab
  *	@sk_slab - the slabcache this instance was allocated from
  *	@sk_timer - sock cleanup timer
  *	@sk_stamp - time stamp of last packet received
  *	@sk_socket - Identd and reporting IO signals
  *	@sk_user_data - RPC layer private data
  *	@sk_owner - module that owns this socket
  *	@sk_state_change - callback to indicate change in the state of the sock
  *	@sk_data_ready - callback to indicate there is data to be processed
  *	@sk_write_space - callback to indicate there is bf sending space available
  *	@sk_error_report - callback to indicate errors (e.g. %MSG_ERRQUEUE)
  *	@sk_backlog_rcv - callback to process the backlog
  *	@sk_destruct - called at sock freeing time, i.e. when all refcnt == 0
175
 */
Linus Torvalds's avatar
Linus Torvalds committed
176
struct sock {
177 178 179 180 181 182 183 184 185
	/*
	 * Now struct tcp_tw_bucket also uses sock_common, so please just
	 * don't add nothing before this first member (__sk_common) --acme
	 */
	struct sock_common	__sk_common;
#define sk_family		__sk_common.skc_family
#define sk_state		__sk_common.skc_state
#define sk_reuse		__sk_common.skc_reuse
#define sk_bound_dev_if		__sk_common.skc_bound_dev_if
186 187
#define sk_node			__sk_common.skc_node
#define sk_bind_node		__sk_common.skc_bind_node
188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216
#define sk_refcnt		__sk_common.skc_refcnt
	volatile unsigned char	sk_zapped;
	unsigned char		sk_shutdown;
	unsigned char		sk_use_write_queue;
	unsigned char		sk_userlocks;
	socket_lock_t		sk_lock;
	int			sk_rcvbuf;
	wait_queue_head_t	*sk_sleep;
	struct dst_entry	*sk_dst_cache;
	rwlock_t		sk_dst_lock;
	struct xfrm_policy	*sk_policy[2];
	atomic_t		sk_rmem_alloc;
	struct sk_buff_head	sk_receive_queue;
	atomic_t		sk_wmem_alloc;
	struct sk_buff_head	sk_write_queue;
	atomic_t		sk_omem_alloc;
	int			sk_wmem_queued;
	int			sk_forward_alloc;
	unsigned int		sk_allocation;
	int			sk_sndbuf;
	unsigned long 		sk_flags;
	char		 	sk_no_check;
	unsigned char		sk_debug;
	unsigned char		sk_rcvtstamp;
	unsigned char		sk_no_largesend;
	int			sk_route_caps;
	unsigned long	        sk_lingertime;
	int			sk_hashent;
	struct sock		*sk_pair;
217 218
	/*
	 * The backlog queue is special, it is always used with
Linus Torvalds's avatar
Linus Torvalds committed
219
	 * the per-socket spinlock held and requires low latency
220
	 * access. Therefore we special case it's implementation.
Linus Torvalds's avatar
Linus Torvalds committed
221 222 223 224
	 */
	struct {
		struct sk_buff *head;
		struct sk_buff *tail;
225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255
	} sk_backlog;
	rwlock_t		sk_callback_lock;
	struct sk_buff_head	sk_error_queue;
	struct proto		*sk_prot;
	int			sk_err,
				sk_err_soft;
	unsigned short		sk_ack_backlog;
	unsigned short		sk_max_ack_backlog;
	__u32			sk_priority;
	unsigned short		sk_type;
	unsigned char		sk_localroute;
	unsigned char		sk_protocol;
	struct ucred		sk_peercred;
	int			sk_rcvlowat;
	long			sk_rcvtimeo;
	long			sk_sndtimeo;
	struct sk_filter      	*sk_filter;
	void			*sk_protinfo;
	kmem_cache_t		*sk_slab;
	struct timer_list	sk_timer;
	struct timeval		sk_stamp;
	struct socket		*sk_socket;
	void			*sk_user_data;
	struct module		*sk_owner;
	void			(*sk_state_change)(struct sock *sk);
	void			(*sk_data_ready)(struct sock *sk, int bytes);
	void			(*sk_write_space)(struct sock *sk);
	void			(*sk_error_report)(struct sock *sk);
  	int			(*sk_backlog_rcv)(struct sock *sk,
						  struct sk_buff *skb);  
	void                    (*sk_destruct)(struct sock *sk);
Linus Torvalds's avatar
Linus Torvalds committed
256 257
};

258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291
/*
 * Hashed lists helper routines
 */
static inline struct sock *__sk_head(struct hlist_head *head)
{
	return hlist_entry(head->first, struct sock, sk_node);
}

static inline struct sock *sk_head(struct hlist_head *head)
{
	return hlist_empty(head) ? NULL : __sk_head(head);
}

static inline struct sock *sk_next(struct sock *sk)
{
	return sk->sk_node.next ?
		hlist_entry(sk->sk_node.next, struct sock, sk_node) : NULL;
}

static inline int sk_unhashed(struct sock *sk)
{
	return hlist_unhashed(&sk->sk_node);
}

static inline int sk_hashed(struct sock *sk)
{
	return sk->sk_node.pprev != NULL;
}

static __inline__ void sk_node_init(struct hlist_node *node)
{
	node->pprev = NULL;
}

292 293 294 295 296 297
static __inline__ void __sk_del_node(struct sock *sk)
{
	__hlist_del(&sk->sk_node);
}

static __inline__ int __sk_del_node_init(struct sock *sk)
298 299
{
	if (sk_hashed(sk)) {
300
		__sk_del_node(sk);
301 302 303 304 305 306
		sk_node_init(&sk->sk_node);
		return 1;
	}
	return 0;
}

307 308 309 310 311 312 313 314 315 316 317 318 319 320 321
static inline void __sock_put(struct sock *sk);

static __inline__ int sk_del_node_init(struct sock *sk)
{
	int rc = __sk_del_node_init(sk);

	if (rc) {
		/* paranoid for a while -acme */
		WARN_ON(atomic_read(&sk->sk_refcnt) == 1);
		__sock_put(sk);
	}
	return rc;
}

static __inline__ void __sk_add_node(struct sock *sk, struct hlist_head *list)
322 323 324 325
{
	hlist_add_head(&sk->sk_node, list);
}

326 327 328 329 330 331 332 333 334 335 336 337 338
static inline void sock_hold(struct sock *sk);

static __inline__ void sk_add_node(struct sock *sk, struct hlist_head *list)
{
	sock_hold(sk);
	__sk_add_node(sk, list);
}

static __inline__ void __sk_del_bind_node(struct sock *sk)
{
	__hlist_del(&sk->sk_bind_node);
}

339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357
static __inline__ void sk_add_bind_node(struct sock *sk,
					struct hlist_head *list)
{
	hlist_add_head(&sk->sk_bind_node, list);
}

#define sk_for_each(__sk, node, list) \
	hlist_for_each_entry(__sk, node, list, sk_node)
#define sk_for_each_from(__sk, node) \
	if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
		hlist_for_each_entry_from(__sk, node, sk_node)
#define sk_for_each_continue(__sk, node) \
	if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
		hlist_for_each_entry_continue(__sk, node, sk_node)
#define sk_for_each_safe(__sk, node, tmp, list) \
	hlist_for_each_entry_safe(__sk, node, tmp, list, sk_node)
#define sk_for_each_bound(__sk, node, list) \
	hlist_for_each_entry(__sk, node, list, sk_bind_node)

358 359 360 361 362 363 364 365 366 367 368 369 370
/* Sock flags */
enum sock_flags {
	SOCK_DEAD,
	SOCK_DONE,
	SOCK_URGINLINE,
	SOCK_KEEPOPEN,
	SOCK_LINGER,
	SOCK_DESTROY,
	SOCK_BROADCAST,
};

static inline void sock_set_flag(struct sock *sk, enum sock_flags flag)
{
371
	__set_bit(flag, &sk->sk_flags);
372 373 374 375
}

static inline void sock_reset_flag(struct sock *sk, enum sock_flags flag)
{
376
	__clear_bit(flag, &sk->sk_flags);
377 378 379 380
}

static inline int sock_flag(struct sock *sk, enum sock_flags flag)
{
381
	return test_bit(flag, &sk->sk_flags);
382 383
}

Linus Torvalds's avatar
Linus Torvalds committed
384
/* The per-socket spinlock must be held here. */
385 386 387 388 389 390 391 392 393
#define sk_add_backlog(__sk, __skb)				\
do {	if (!(__sk)->sk_backlog.tail) {				\
		(__sk)->sk_backlog.head =			\
		     (__sk)->sk_backlog.tail = (__skb);		\
	} else {						\
		((__sk)->sk_backlog.tail)->next = (__skb);	\
		(__sk)->sk_backlog.tail = (__skb);		\
	}							\
	(__skb)->next = NULL;					\
Linus Torvalds's avatar
Linus Torvalds committed
394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419
} while(0)

/* IP protocol blocks we attach to sockets.
 * socket layer -> transport layer interface
 * transport -> network interface is defined by struct inet_proto
 */
struct proto {
	void			(*close)(struct sock *sk, 
					long timeout);
	int			(*connect)(struct sock *sk,
				        struct sockaddr *uaddr, 
					int addr_len);
	int			(*disconnect)(struct sock *sk, int flags);

	struct sock *		(*accept) (struct sock *sk, int flags, int *err);

	int			(*ioctl)(struct sock *sk, int cmd,
					 unsigned long arg);
	int			(*init)(struct sock *sk);
	int			(*destroy)(struct sock *sk);
	void			(*shutdown)(struct sock *sk, int how);
	int			(*setsockopt)(struct sock *sk, int level, 
					int optname, char *optval, int optlen);
	int			(*getsockopt)(struct sock *sk, int level, 
					int optname, char *optval, 
					int *option);  	 
Benjamin LaHaise's avatar
Benjamin LaHaise committed
420 421 422 423
	int			(*sendmsg)(struct kiocb *iocb, struct sock *sk,
					   struct msghdr *msg, int len);
	int			(*recvmsg)(struct kiocb *iocb, struct sock *sk,
					   struct msghdr *msg,
Linus Torvalds's avatar
Linus Torvalds committed
424 425
					int len, int noblock, int flags, 
					int *addr_len);
426 427
	int			(*sendpage)(struct sock *sk, struct page *page,
					int offset, size_t size, int flags);
Linus Torvalds's avatar
Linus Torvalds committed
428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446
	int			(*bind)(struct sock *sk, 
					struct sockaddr *uaddr, int addr_len);

	int			(*backlog_rcv) (struct sock *sk, 
						struct sk_buff *skb);

	/* Keeping track of sk's, looking them up, and port selection methods. */
	void			(*hash)(struct sock *sk);
	void			(*unhash)(struct sock *sk);
	int			(*get_port)(struct sock *sk, unsigned short snum);

	char			name[32];

	struct {
		int inuse;
		u8  __pad[SMP_CACHE_BYTES - sizeof(int)];
	} stats[NR_CPUS];
};

447 448 449 450 451 452 453 454 455 456 457
static __inline__ void sk_set_owner(struct sock *sk, struct module *owner)
{
	/*
	 * One should use sk_set_owner just once, after struct sock creation,
	 * be it shortly after sk_alloc or after a function that returns a new
	 * struct sock (and that down the call chain called sk_alloc), e.g. the
	 * IPv4 and IPv6 modules share tcp_create_openreq_child, so if
	 * tcp_create_openreq_child called sk_set_owner IPv6 would have to
	 * change the ownership of this struct sock, with one not needed
	 * transient sk_set_owner call.
	 */
458
	if (unlikely(sk->sk_owner != NULL))
459
		BUG();
460
	sk->sk_owner = owner;
461 462 463
	__module_get(owner);
}

Linus Torvalds's avatar
Linus Torvalds committed
464
/* Called with local bh disabled */
Linus Torvalds's avatar
Linus Torvalds committed
465
static __inline__ void sock_prot_inc_use(struct proto *prot)
Linus Torvalds's avatar
Linus Torvalds committed
466 467 468 469
{
	prot->stats[smp_processor_id()].inuse++;
}

Linus Torvalds's avatar
Linus Torvalds committed
470
static __inline__ void sock_prot_dec_use(struct proto *prot)
Linus Torvalds's avatar
Linus Torvalds committed
471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489
{
	prot->stats[smp_processor_id()].inuse--;
}

/* About 10 seconds */
#define SOCK_DESTROY_TIME (10*HZ)

/* Sockets 0-1023 can't be bound to unless you are superuser */
#define PROT_SOCK	1024

#define SHUTDOWN_MASK	3
#define RCV_SHUTDOWN	1
#define SEND_SHUTDOWN	2

#define SOCK_SNDBUF_LOCK	1
#define SOCK_RCVBUF_LOCK	2
#define SOCK_BINDADDR_LOCK	4
#define SOCK_BINDPORT_LOCK	8

Benjamin LaHaise's avatar
Benjamin LaHaise committed
490 491 492 493 494 495 496 497
/* sock_iocb: used to kick off async processing of socket ios */
struct sock_iocb {
	struct list_head	list;

	int			flags;
	int			size;
	struct socket		*sock;
	struct sock		*sk;
498
	struct scm_cookie	*scm;
Benjamin LaHaise's avatar
Benjamin LaHaise committed
499 500 501 502 503 504 505 506 507 508 509 510 511 512
	struct msghdr		*msg, async_msg;
	struct iovec		async_iov;
};

static inline struct sock_iocb *kiocb_to_siocb(struct kiocb *iocb)
{
	BUG_ON(sizeof(struct sock_iocb) > KIOCB_PRIVATE_SIZE);
	return (struct sock_iocb *)iocb->private;
}

static inline struct kiocb *siocb_to_kiocb(struct sock_iocb *si)
{
	return container_of((void *)si, struct kiocb, private);
}
Linus Torvalds's avatar
Linus Torvalds committed
513

514 515 516 517 518 519 520 521 522 523 524 525 526 527 528
struct socket_alloc {
	struct socket socket;
	struct inode vfs_inode;
};

static inline struct socket *SOCKET_I(struct inode *inode)
{
	return &container_of(inode, struct socket_alloc, vfs_inode)->socket;
}

static inline struct inode *SOCK_INODE(struct socket *socket)
{
	return &container_of(socket, struct socket_alloc, socket)->vfs_inode;
}

Linus Torvalds's avatar
Linus Torvalds committed
529 530 531 532 533 534 535 536 537 538 539 540 541 542 543
/* Used by processes to "lock" a socket state, so that
 * interrupts and bottom half handlers won't change it
 * from under us. It essentially blocks any incoming
 * packets, so that we won't get any new data or any
 * packets that change the state of the socket.
 *
 * While locked, BH processing will add new packets to
 * the backlog queue.  This queue is processed by the
 * owner of the socket lock right before it is released.
 *
 * Since ~2.3.5 it is also exclusive sleep lock serializing
 * accesses from user process context.
 */
extern void __lock_sock(struct sock *sk);
extern void __release_sock(struct sock *sk);
544
#define sock_owned_by_user(sk)	((sk)->sk_lock.owner)
Linus Torvalds's avatar
Linus Torvalds committed
545
#define lock_sock(__sk) \
546
do {	might_sleep(); \
547 548
	spin_lock_bh(&((__sk)->sk_lock.slock)); \
	if ((__sk)->sk_lock.owner) \
Linus Torvalds's avatar
Linus Torvalds committed
549
		__lock_sock(__sk); \
550 551
	(__sk)->sk_lock.owner = (void *)1; \
	spin_unlock_bh(&((__sk)->sk_lock.slock)); \
Linus Torvalds's avatar
Linus Torvalds committed
552 553 554
} while(0)

#define release_sock(__sk) \
555 556
do {	spin_lock_bh(&((__sk)->sk_lock.slock)); \
	if ((__sk)->sk_backlog.tail) \
Linus Torvalds's avatar
Linus Torvalds committed
557
		__release_sock(__sk); \
558 559 560 561
	(__sk)->sk_lock.owner = NULL; \
        if (waitqueue_active(&((__sk)->sk_lock.wq))) \
		wake_up(&((__sk)->sk_lock.wq)); \
	spin_unlock_bh(&((__sk)->sk_lock.slock)); \
Linus Torvalds's avatar
Linus Torvalds committed
562 563 564
} while(0)

/* BH context may only use the following locking interface. */
565 566
#define bh_lock_sock(__sk)	spin_lock(&((__sk)->sk_lock.slock))
#define bh_unlock_sock(__sk)	spin_unlock(&((__sk)->sk_lock.slock))
Linus Torvalds's avatar
Linus Torvalds committed
567

568 569
extern struct sock *		sk_alloc(int family, int priority, int zero_it,
					 kmem_cache_t *slab);
Linus Torvalds's avatar
Linus Torvalds committed
570 571 572 573 574 575 576 577 578 579 580 581
extern void			sk_free(struct sock *sk);

extern struct sk_buff		*sock_wmalloc(struct sock *sk,
					      unsigned long size, int force,
					      int priority);
extern struct sk_buff		*sock_rmalloc(struct sock *sk,
					      unsigned long size, int force,
					      int priority);
extern void			sock_wfree(struct sk_buff *skb);
extern void			sock_rfree(struct sk_buff *skb);

extern int			sock_setsockopt(struct socket *sock, int level,
582
						int op, char __user *optval,
Linus Torvalds's avatar
Linus Torvalds committed
583 584 585
						int optlen);

extern int			sock_getsockopt(struct socket *sock, int level,
586 587
						int op, char __user *optval, 
						int __user *optlen);
Linus Torvalds's avatar
Linus Torvalds committed
588 589 590 591
extern struct sk_buff 		*sock_alloc_send_skb(struct sock *sk,
						     unsigned long size,
						     int noblock,
						     int *errcode);
Linus Torvalds's avatar
Linus Torvalds committed
592 593 594 595 596
extern struct sk_buff 		*sock_alloc_send_pskb(struct sock *sk,
						      unsigned long header_len,
						      unsigned long data_len,
						      int noblock,
						      int *errcode);
Linus Torvalds's avatar
Linus Torvalds committed
597 598
extern void *sock_kmalloc(struct sock *sk, int size, int priority);
extern void sock_kfree_s(struct sock *sk, void *mem, int size);
599
extern void sk_send_sigurg(struct sock *sk);
Linus Torvalds's avatar
Linus Torvalds committed
600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625

/*
 * Functions to fill in entries in struct proto_ops when a protocol
 * does not implement a particular function.
 */
extern int                      sock_no_release(struct socket *);
extern int                      sock_no_bind(struct socket *, 
					     struct sockaddr *, int);
extern int                      sock_no_connect(struct socket *,
						struct sockaddr *, int, int);
extern int                      sock_no_socketpair(struct socket *,
						   struct socket *);
extern int                      sock_no_accept(struct socket *,
					       struct socket *, int);
extern int                      sock_no_getname(struct socket *,
						struct sockaddr *, int *, int);
extern unsigned int             sock_no_poll(struct file *, struct socket *,
					     struct poll_table_struct *);
extern int                      sock_no_ioctl(struct socket *, unsigned int,
					      unsigned long);
extern int			sock_no_listen(struct socket *, int);
extern int                      sock_no_shutdown(struct socket *, int);
extern int			sock_no_getsockopt(struct socket *, int , int,
						   char *, int *);
extern int			sock_no_setsockopt(struct socket *, int, int,
						   char *, int);
Benjamin LaHaise's avatar
Benjamin LaHaise committed
626
extern int                      sock_no_sendmsg(struct kiocb *, struct socket *,
627
						struct msghdr *, int);
Benjamin LaHaise's avatar
Benjamin LaHaise committed
628
extern int                      sock_no_recvmsg(struct kiocb *, struct socket *,
629
						struct msghdr *, int, int);
Linus Torvalds's avatar
Linus Torvalds committed
630 631 632
extern int			sock_no_mmap(struct file *file,
					     struct socket *sock,
					     struct vm_area_struct *vma);
Linus Torvalds's avatar
Linus Torvalds committed
633 634 635 636
extern ssize_t			sock_no_sendpage(struct socket *sock,
						struct page *page,
						int offset, size_t size, 
						int flags);
Linus Torvalds's avatar
Linus Torvalds committed
637 638 639 640 641 642 643 644 645 646 647

/*
 *	Default socket callbacks and setup code
 */
 
extern void sock_def_destruct(struct sock *);

/* Initialise core socket variables */
extern void sock_init_data(struct socket *sock, struct sock *sk);

/**
Chris Wright's avatar
Chris Wright committed
648
 *	sk_filter - run a packet through a socket filter
649
 *	@sk: sock associated with &sk_buff
Linus Torvalds's avatar
Linus Torvalds committed
650
 *	@skb: buffer to filter
651
 *	@needlock: set to 1 if the sock is not locked by caller.
Linus Torvalds's avatar
Linus Torvalds committed
652 653 654 655 656
 *
 * Run the filter code and then cut skb->data to correct size returned by
 * sk_run_filter. If pkt_len is 0 we toss packet. If skb->len is smaller
 * than pkt_len we keep whole skb->data. This is the socket level
 * wrapper to sk_run_filter. It returns 0 if the packet should
657 658
 * be accepted or -EPERM if the packet should be tossed.
 *
Linus Torvalds's avatar
Linus Torvalds committed
659 660
 */

Chris Wright's avatar
Chris Wright committed
661
static inline int sk_filter(struct sock *sk, struct sk_buff *skb, int needlock)
662
{
Chris Wright's avatar
Chris Wright committed
663 664 665 666 667 668
	int err;
	
	err = security_sock_rcv_skb(sk, skb);
	if (err)
		return err;
	
669
	if (sk->sk_filter) {
670 671 672 673 674
		struct sk_filter *filter;
		
		if (needlock)
			bh_lock_sock(sk);
		
675
		filter = sk->sk_filter;
676 677 678 679 680 681 682 683 684 685 686 687 688
		if (filter) {
			int pkt_len = sk_run_filter(skb, filter->insns,
						    filter->len);
			if (!pkt_len)
				err = -EPERM;
			else
				skb_trim(skb, pkt_len);
		}

		if (needlock)
			bh_unlock_sock(sk);
	}
	return err;
Linus Torvalds's avatar
Linus Torvalds committed
689 690 691 692 693 694 695 696 697 698 699 700 701 702
}

/**
 *	sk_filter_release: Release a socket filter
 *	@sk: socket
 *	@fp: filter to remove
 *
 *	Remove a filter from a socket and release its resources.
 */
 
static inline void sk_filter_release(struct sock *sk, struct sk_filter *fp)
{
	unsigned int size = sk_filter_len(fp);

703
	atomic_sub(size, &sk->sk_omem_alloc);
Linus Torvalds's avatar
Linus Torvalds committed
704 705 706 707 708 709 710 711

	if (atomic_dec_and_test(&fp->refcnt))
		kfree(fp);
}

static inline void sk_filter_charge(struct sock *sk, struct sk_filter *fp)
{
	atomic_inc(&fp->refcnt);
712
	atomic_add(sk_filter_len(fp), &sk->sk_omem_alloc);
Linus Torvalds's avatar
Linus Torvalds committed
713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747
}

/*
 * Socket reference counting postulates.
 *
 * * Each user of socket SHOULD hold a reference count.
 * * Each access point to socket (an hash table bucket, reference from a list,
 *   running timer, skb in flight MUST hold a reference count.
 * * When reference count hits 0, it means it will never increase back.
 * * When reference count hits 0, it means that no references from
 *   outside exist to this socket and current process on current CPU
 *   is last user and may/should destroy this socket.
 * * sk_free is called from any context: process, BH, IRQ. When
 *   it is called, socket has no references from outside -> sk_free
 *   may release descendant resources allocated by the socket, but
 *   to the time when it is called, socket is NOT referenced by any
 *   hash tables, lists etc.
 * * Packets, delivered from outside (from network or from another process)
 *   and enqueued on receive/error queues SHOULD NOT grab reference count,
 *   when they sit in queue. Otherwise, packets will leak to hole, when
 *   socket is looked up by one cpu and unhasing is made by another CPU.
 *   It is true for udp/raw, netlink (leak to receive and error queues), tcp
 *   (leak to backlog). Packet socket does all the processing inside
 *   BR_NETPROTO_LOCK, so that it has not this race condition. UNIX sockets
 *   use separate SMP lock, so that they are prone too.
 */

/* Grab socket reference count. This operation is valid only
   when sk is ALREADY grabbed f.e. it is found in hash table
   or a list and the lookup is made under lock preventing hash table
   modifications.
 */

static inline void sock_hold(struct sock *sk)
{
748
	atomic_inc(&sk->sk_refcnt);
Linus Torvalds's avatar
Linus Torvalds committed
749 750 751 752 753 754 755
}

/* Ungrab socket in the context, which assumes that socket refcnt
   cannot hit zero, f.e. it is true in context of any socketcall.
 */
static inline void __sock_put(struct sock *sk)
{
756
	atomic_dec(&sk->sk_refcnt);
Linus Torvalds's avatar
Linus Torvalds committed
757 758 759 760 761
}

/* Ungrab socket and destroy it, if it was the last reference. */
static inline void sock_put(struct sock *sk)
{
762
	if (atomic_dec_and_test(&sk->sk_refcnt))
Linus Torvalds's avatar
Linus Torvalds committed
763 764 765 766 767 768 769 770 771 772 773 774
		sk_free(sk);
}

/* Detach socket from process context.
 * Announce socket dead, detach it from wait queue and inode.
 * Note that parent inode held reference count on this struct sock,
 * we do not release it in this function, because protocol
 * probably wants some additional cleanups or even continuing
 * to work with this socket (TCP).
 */
static inline void sock_orphan(struct sock *sk)
{
775
	write_lock_bh(&sk->sk_callback_lock);
776
	sock_set_flag(sk, SOCK_DEAD);
777 778 779
	sk->sk_socket = NULL;
	sk->sk_sleep  = NULL;
	write_unlock_bh(&sk->sk_callback_lock);
Linus Torvalds's avatar
Linus Torvalds committed
780 781 782 783
}

static inline void sock_graft(struct sock *sk, struct socket *parent)
{
784 785
	write_lock_bh(&sk->sk_callback_lock);
	sk->sk_sleep = &parent->wait;
Linus Torvalds's avatar
Linus Torvalds committed
786
	parent->sk = sk;
787 788
	sk->sk_socket = parent;
	write_unlock_bh(&sk->sk_callback_lock);
Linus Torvalds's avatar
Linus Torvalds committed
789 790 791 792 793 794
}

static inline int sock_i_uid(struct sock *sk)
{
	int uid;

795 796 797
	read_lock(&sk->sk_callback_lock);
	uid = sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_uid : 0;
	read_unlock(&sk->sk_callback_lock);
Linus Torvalds's avatar
Linus Torvalds committed
798 799 800 801 802 803 804
	return uid;
}

static inline unsigned long sock_i_ino(struct sock *sk)
{
	unsigned long ino;

805 806 807
	read_lock(&sk->sk_callback_lock);
	ino = sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_ino : 0;
	read_unlock(&sk->sk_callback_lock);
Linus Torvalds's avatar
Linus Torvalds committed
808 809 810 811 812 813
	return ino;
}

static inline struct dst_entry *
__sk_dst_get(struct sock *sk)
{
814
	return sk->sk_dst_cache;
Linus Torvalds's avatar
Linus Torvalds committed
815 816 817 818 819 820 821
}

static inline struct dst_entry *
sk_dst_get(struct sock *sk)
{
	struct dst_entry *dst;

822 823
	read_lock(&sk->sk_dst_lock);
	dst = sk->sk_dst_cache;
Linus Torvalds's avatar
Linus Torvalds committed
824 825
	if (dst)
		dst_hold(dst);
826
	read_unlock(&sk->sk_dst_lock);
Linus Torvalds's avatar
Linus Torvalds committed
827 828 829 830 831 832 833 834
	return dst;
}

static inline void
__sk_dst_set(struct sock *sk, struct dst_entry *dst)
{
	struct dst_entry *old_dst;

835 836
	old_dst = sk->sk_dst_cache;
	sk->sk_dst_cache = dst;
Linus Torvalds's avatar
Linus Torvalds committed
837 838 839 840 841 842
	dst_release(old_dst);
}

static inline void
sk_dst_set(struct sock *sk, struct dst_entry *dst)
{
843
	write_lock(&sk->sk_dst_lock);
Linus Torvalds's avatar
Linus Torvalds committed
844
	__sk_dst_set(sk, dst);
845
	write_unlock(&sk->sk_dst_lock);
Linus Torvalds's avatar
Linus Torvalds committed
846 847 848 849 850 851 852
}

static inline void
__sk_dst_reset(struct sock *sk)
{
	struct dst_entry *old_dst;

853 854
	old_dst = sk->sk_dst_cache;
	sk->sk_dst_cache = NULL;
Linus Torvalds's avatar
Linus Torvalds committed
855 856 857 858 859 860
	dst_release(old_dst);
}

static inline void
sk_dst_reset(struct sock *sk)
{
861
	write_lock(&sk->sk_dst_lock);
Linus Torvalds's avatar
Linus Torvalds committed
862
	__sk_dst_reset(sk);
863
	write_unlock(&sk->sk_dst_lock);
Linus Torvalds's avatar
Linus Torvalds committed
864 865 866 867 868
}

static inline struct dst_entry *
__sk_dst_check(struct sock *sk, u32 cookie)
{
869
	struct dst_entry *dst = sk->sk_dst_cache;
Linus Torvalds's avatar
Linus Torvalds committed
870 871

	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
872
		sk->sk_dst_cache = NULL;
Linus Torvalds's avatar
Linus Torvalds committed
873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906
		return NULL;
	}

	return dst;
}

static inline struct dst_entry *
sk_dst_check(struct sock *sk, u32 cookie)
{
	struct dst_entry *dst = sk_dst_get(sk);

	if (dst && dst->obsolete && dst->ops->check(dst, cookie) == NULL) {
		sk_dst_reset(sk);
		return NULL;
	}

	return dst;
}


/*
 * 	Queue a received datagram if it will fit. Stream and sequenced
 *	protocols can't normally use this as they need to fit buffers in
 *	and play with them.
 *
 * 	Inlined as it's very short and called for pretty much every
 *	packet ever received.
 */

static inline void skb_set_owner_w(struct sk_buff *skb, struct sock *sk)
{
	sock_hold(sk);
	skb->sk = sk;
	skb->destructor = sock_wfree;
907
	atomic_add(skb->truesize, &sk->sk_wmem_alloc);
Linus Torvalds's avatar
Linus Torvalds committed
908 909 910 911 912 913
}

static inline void skb_set_owner_r(struct sk_buff *skb, struct sock *sk)
{
	skb->sk = sk;
	skb->destructor = sock_rfree;
914
	atomic_add(skb->truesize, &sk->sk_rmem_alloc);
Linus Torvalds's avatar
Linus Torvalds committed
915 916 917 918
}

static inline int sock_queue_rcv_skb(struct sock *sk, struct sk_buff *skb)
{
919 920
	int err = 0;

Linus Torvalds's avatar
Linus Torvalds committed
921 922 923
	/* Cast skb->rcvbuf to unsigned... It's pointless, but reduces
	   number of warnings when compiling with -W --ANK
	 */
924 925
	if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >=
	    (unsigned)sk->sk_rcvbuf) {
926 927
		err = -ENOMEM;
		goto out;
Linus Torvalds's avatar
Linus Torvalds committed
928
	}
929 930 931 932 933 934 935 936

	/* It would be deadlock, if sock_queue_rcv_skb is used
	   with socket lock! We assume that users of this
	   function are lock free.
	*/
	err = sk_filter(sk, skb, 1);
	if (err)
		goto out;
Linus Torvalds's avatar
Linus Torvalds committed
937 938 939

	skb->dev = NULL;
	skb_set_owner_r(skb, sk);
940
	skb_queue_tail(&sk->sk_receive_queue, skb);
941
	if (!sock_flag(sk, SOCK_DEAD))
942
		sk->sk_data_ready(sk, skb->len);
943 944
out:
	return err;
Linus Torvalds's avatar
Linus Torvalds committed
945 946 947 948 949 950 951
}

static inline int sock_queue_err_skb(struct sock *sk, struct sk_buff *skb)
{
	/* Cast skb->rcvbuf to unsigned... It's pointless, but reduces
	   number of warnings when compiling with -W --ANK
	 */
952 953
	if (atomic_read(&sk->sk_rmem_alloc) + skb->truesize >=
	    (unsigned)sk->sk_rcvbuf)
Linus Torvalds's avatar
Linus Torvalds committed
954 955
		return -ENOMEM;
	skb_set_owner_r(skb, sk);
956
	skb_queue_tail(&sk->sk_error_queue, skb);
957
	if (!sock_flag(sk, SOCK_DEAD))
958
		sk->sk_data_ready(sk, skb->len);
Linus Torvalds's avatar
Linus Torvalds committed
959 960 961 962 963 964 965 966 967
	return 0;
}

/*
 *	Recover an error report and clear atomically
 */
 
static inline int sock_error(struct sock *sk)
{
968
	int err = xchg(&sk->sk_err, 0);
Linus Torvalds's avatar
Linus Torvalds committed
969 970 971 972 973 974 975
	return -err;
}

static inline unsigned long sock_wspace(struct sock *sk)
{
	int amt = 0;

976 977
	if (!(sk->sk_shutdown & SEND_SHUTDOWN)) {
		amt = sk->sk_sndbuf - atomic_read(&sk->sk_wmem_alloc);
Linus Torvalds's avatar
Linus Torvalds committed
978 979 980 981 982 983 984 985
		if (amt < 0) 
			amt = 0;
	}
	return amt;
}

static inline void sk_wake_async(struct sock *sk, int how, int band)
{
986 987
	if (sk->sk_socket && sk->sk_socket->fasync_list)
		sock_wake_async(sk->sk_socket, how, band);
Linus Torvalds's avatar
Linus Torvalds committed
988 989 990 991 992 993 994 995 996 997
}

#define SOCK_MIN_SNDBUF 2048
#define SOCK_MIN_RCVBUF 256

/*
 *	Default write policy as shown to user space via poll/select/SIGIO
 */
static inline int sock_writeable(struct sock *sk) 
{
998
	return atomic_read(&sk->sk_wmem_alloc) < (sk->sk_sndbuf / 2);
Linus Torvalds's avatar
Linus Torvalds committed
999 1000 1001 1002 1003 1004 1005 1006 1007
}

static inline int gfp_any(void)
{
	return in_softirq() ? GFP_ATOMIC : GFP_KERNEL;
}

static inline long sock_rcvtimeo(struct sock *sk, int noblock)
{
1008
	return noblock ? 0 : sk->sk_rcvtimeo;
Linus Torvalds's avatar
Linus Torvalds committed
1009 1010 1011 1012
}

static inline long sock_sndtimeo(struct sock *sk, int noblock)
{
1013
	return noblock ? 0 : sk->sk_sndtimeo;
Linus Torvalds's avatar
Linus Torvalds committed
1014 1015 1016 1017
}

static inline int sock_rcvlowat(struct sock *sk, int waitall, int len)
{
1018
	return (waitall ? len : min_t(int, sk->sk_rcvlowat, len)) ? : 1;
Linus Torvalds's avatar
Linus Torvalds committed
1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031
}

/* Alas, with timeout socket operations are not restartable.
 * Compare this to poll().
 */
static inline int sock_intr_errno(long timeo)
{
	return timeo == MAX_SCHEDULE_TIMEOUT ? -ERESTARTSYS : -EINTR;
}

static __inline__ void
sock_recv_timestamp(struct msghdr *msg, struct sock *sk, struct sk_buff *skb)
{
1032
	if (sk->sk_rcvtstamp)
Linus Torvalds's avatar
Linus Torvalds committed
1033 1034
		put_cmsg(msg, SOL_SOCKET, SO_TIMESTAMP, sizeof(skb->stamp), &skb->stamp);
	else
1035
		sk->sk_stamp = skb->stamp;
Linus Torvalds's avatar
Linus Torvalds committed
1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055
}

/* 
 *	Enable debug/info messages 
 */

#if 0
#define NETDEBUG(x)	do { } while (0)
#else
#define NETDEBUG(x)	do { x; } while (0)
#endif

/*
 * Macros for sleeping on a socket. Use them like this:
 *
 * SOCK_SLEEP_PRE(sk)
 * if (condition)
 * 	schedule();
 * SOCK_SLEEP_POST(sk)
 *
1056 1057 1058
 * N.B. These are now obsolete and were, afaik, only ever used in DECnet
 * and when the last use of them in DECnet has gone, I'm intending to
 * remove them.
Linus Torvalds's avatar
Linus Torvalds committed
1059 1060 1061 1062 1063
 */

#define SOCK_SLEEP_PRE(sk) 	{ struct task_struct *tsk = current; \
				DECLARE_WAITQUEUE(wait, tsk); \
				tsk->state = TASK_INTERRUPTIBLE; \
1064
				add_wait_queue((sk)->sk_sleep, &wait); \
Linus Torvalds's avatar
Linus Torvalds committed
1065 1066 1067
				release_sock(sk);

#define SOCK_SLEEP_POST(sk)	tsk->state = TASK_RUNNING; \
1068
				remove_wait_queue((sk)->sk_sleep, &wait); \
Linus Torvalds's avatar
Linus Torvalds committed
1069 1070 1071
				lock_sock(sk); \
				}

1072 1073 1074
static inline void sock_valbool_flag(struct sock *sk, int bit, int valbool)
{
	if (valbool)
1075
		sock_set_flag(sk, bit);
1076
	else
1077
		sock_reset_flag(sk, bit);
1078 1079
}

Linus Torvalds's avatar
Linus Torvalds committed
1080 1081 1082
extern __u32 sysctl_wmem_max;
extern __u32 sysctl_rmem_max;

1083 1084
int siocdevprivate_ioctl(unsigned int fd, unsigned int cmd, unsigned long arg);

Linus Torvalds's avatar
Linus Torvalds committed
1085
#endif	/* _SOCK_H */