[NET]: Ensure device name passed to SO_BINDTODEVICE is NULL terminated.
[linux-2.6/suspend2-2.6.18.git] / net / core / sock.c
blobe110b9004147b4418391467a4b31715a5dc2902a
1 /*
2 * INET An implementation of the TCP/IP protocol suite for the LINUX
3 * operating system. INET is implemented using the BSD Socket
4 * interface as the means of communication with the user level.
6 * Generic socket support routines. Memory allocators, socket lock/release
7 * handler for protocols to use and generic option handler.
10 * Version: $Id: sock.c,v 1.117 2002/02/01 22:01:03 davem Exp $
12 * Authors: Ross Biro
13 * Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
14 * Florian La Roche, <flla@stud.uni-sb.de>
15 * Alan Cox, <A.Cox@swansea.ac.uk>
17 * Fixes:
18 * Alan Cox : Numerous verify_area() problems
19 * Alan Cox : Connecting on a connecting socket
20 * now returns an error for tcp.
21 * Alan Cox : sock->protocol is set correctly.
22 * and is not sometimes left as 0.
23 * Alan Cox : connect handles icmp errors on a
24 * connect properly. Unfortunately there
25 * is a restart syscall nasty there. I
26 * can't match BSD without hacking the C
27 * library. Ideas urgently sought!
28 * Alan Cox : Disallow bind() to addresses that are
29 * not ours - especially broadcast ones!!
30 * Alan Cox : Socket 1024 _IS_ ok for users. (fencepost)
31 * Alan Cox : sock_wfree/sock_rfree don't destroy sockets,
32 * instead they leave that for the DESTROY timer.
33 * Alan Cox : Clean up error flag in accept
34 * Alan Cox : TCP ack handling is buggy, the DESTROY timer
35 * was buggy. Put a remove_sock() in the handler
36 * for memory when we hit 0. Also altered the timer
37 * code. The ACK stuff can wait and needs major
38 * TCP layer surgery.
39 * Alan Cox : Fixed TCP ack bug, removed remove sock
40 * and fixed timer/inet_bh race.
41 * Alan Cox : Added zapped flag for TCP
42 * Alan Cox : Move kfree_skb into skbuff.c and tidied up surplus code
43 * Alan Cox : for new sk_buff allocations wmalloc/rmalloc now call alloc_skb
44 * Alan Cox : kfree_s calls now are kfree_skbmem so we can track skb resources
45 * Alan Cox : Supports socket option broadcast now as does udp. Packet and raw need fixing.
46 * Alan Cox : Added RCVBUF,SNDBUF size setting. It suddenly occurred to me how easy it was so...
47 * Rick Sladkey : Relaxed UDP rules for matching packets.
48 * C.E.Hawkins : IFF_PROMISC/SIOCGHWADDR support
49 * Pauline Middelink : identd support
50 * Alan Cox : Fixed connect() taking signals I think.
51 * Alan Cox : SO_LINGER supported
52 * Alan Cox : Error reporting fixes
53 * Anonymous : inet_create tidied up (sk->reuse setting)
54 * Alan Cox : inet sockets don't set sk->type!
55 * Alan Cox : Split socket option code
56 * Alan Cox : Callbacks
57 * Alan Cox : Nagle flag for Charles & Johannes stuff
58 * Alex : Removed restriction on inet fioctl
59 * Alan Cox : Splitting INET from NET core
60 * Alan Cox : Fixed bogus SO_TYPE handling in getsockopt()
61 * Adam Caldwell : Missing return in SO_DONTROUTE/SO_DEBUG code
62 * Alan Cox : Split IP from generic code
63 * Alan Cox : New kfree_skbmem()
64 * Alan Cox : Make SO_DEBUG superuser only.
65 * Alan Cox : Allow anyone to clear SO_DEBUG
66 * (compatibility fix)
67 * Alan Cox : Added optimistic memory grabbing for AF_UNIX throughput.
68 * Alan Cox : Allocator for a socket is settable.
69 * Alan Cox : SO_ERROR includes soft errors.
70 * Alan Cox : Allow NULL arguments on some SO_ opts
71 * Alan Cox : Generic socket allocation to make hooks
72 * easier (suggested by Craig Metz).
73 * Michael Pall : SO_ERROR returns positive errno again
74 * Steve Whitehouse: Added default destructor to free
75 * protocol private data.
76 * Steve Whitehouse: Added various other default routines
77 * common to several socket families.
78 * Chris Evans : Call suser() check last on F_SETOWN
79 * Jay Schulist : Added SO_ATTACH_FILTER and SO_DETACH_FILTER.
80 * Andi Kleen : Add sock_kmalloc()/sock_kfree_s()
81 * Andi Kleen : Fix write_space callback
82 * Chris Evans : Security fixes - signedness again
83 * Arnaldo C. Melo : cleanups, use skb_queue_purge
85 * To Fix:
88 * This program is free software; you can redistribute it and/or
89 * modify it under the terms of the GNU General Public License
90 * as published by the Free Software Foundation; either version
91 * 2 of the License, or (at your option) any later version.
94 #include <linux/capability.h>
95 #include <linux/config.h>
96 #include <linux/errno.h>
97 #include <linux/types.h>
98 #include <linux/socket.h>
99 #include <linux/in.h>
100 #include <linux/kernel.h>
101 #include <linux/module.h>
102 #include <linux/proc_fs.h>
103 #include <linux/seq_file.h>
104 #include <linux/sched.h>
105 #include <linux/timer.h>
106 #include <linux/string.h>
107 #include <linux/sockios.h>
108 #include <linux/net.h>
109 #include <linux/mm.h>
110 #include <linux/slab.h>
111 #include <linux/interrupt.h>
112 #include <linux/poll.h>
113 #include <linux/tcp.h>
114 #include <linux/init.h>
116 #include <asm/uaccess.h>
117 #include <asm/system.h>
119 #include <linux/netdevice.h>
120 #include <net/protocol.h>
121 #include <linux/skbuff.h>
122 #include <net/request_sock.h>
123 #include <net/sock.h>
124 #include <net/xfrm.h>
125 #include <linux/ipsec.h>
127 #include <linux/filter.h>
129 #ifdef CONFIG_INET
130 #include <net/tcp.h>
131 #endif
133 /* Take into consideration the size of the struct sk_buff overhead in the
134 * determination of these values, since that is non-constant across
135 * platforms. This makes socket queueing behavior and performance
136 * not depend upon such differences.
138 #define _SK_MEM_PACKETS 256
139 #define _SK_MEM_OVERHEAD (sizeof(struct sk_buff) + 256)
140 #define SK_WMEM_MAX (_SK_MEM_OVERHEAD * _SK_MEM_PACKETS)
141 #define SK_RMEM_MAX (_SK_MEM_OVERHEAD * _SK_MEM_PACKETS)
143 /* Run time adjustable parameters. */
144 __u32 sysctl_wmem_max = SK_WMEM_MAX;
145 __u32 sysctl_rmem_max = SK_RMEM_MAX;
146 __u32 sysctl_wmem_default = SK_WMEM_MAX;
147 __u32 sysctl_rmem_default = SK_RMEM_MAX;
149 /* Maximal space eaten by iovec or ancilliary data plus some space */
150 int sysctl_optmem_max = sizeof(unsigned long)*(2*UIO_MAXIOV + 512);
152 static int sock_set_timeout(long *timeo_p, char __user *optval, int optlen)
154 struct timeval tv;
156 if (optlen < sizeof(tv))
157 return -EINVAL;
158 if (copy_from_user(&tv, optval, sizeof(tv)))
159 return -EFAULT;
161 *timeo_p = MAX_SCHEDULE_TIMEOUT;
162 if (tv.tv_sec == 0 && tv.tv_usec == 0)
163 return 0;
164 if (tv.tv_sec < (MAX_SCHEDULE_TIMEOUT/HZ - 1))
165 *timeo_p = tv.tv_sec*HZ + (tv.tv_usec+(1000000/HZ-1))/(1000000/HZ);
166 return 0;
169 static void sock_warn_obsolete_bsdism(const char *name)
171 static int warned;
172 static char warncomm[TASK_COMM_LEN];
173 if (strcmp(warncomm, current->comm) && warned < 5) {
174 strcpy(warncomm, current->comm);
175 printk(KERN_WARNING "process `%s' is using obsolete "
176 "%s SO_BSDCOMPAT\n", warncomm, name);
177 warned++;
181 static void sock_disable_timestamp(struct sock *sk)
183 if (sock_flag(sk, SOCK_TIMESTAMP)) {
184 sock_reset_flag(sk, SOCK_TIMESTAMP);
185 net_disable_timestamp();
191 * This is meant for all protocols to use and covers goings on
192 * at the socket level. Everything here is generic.
195 int sock_setsockopt(struct socket *sock, int level, int optname,
196 char __user *optval, int optlen)
198 struct sock *sk=sock->sk;
199 struct sk_filter *filter;
200 int val;
201 int valbool;
202 struct linger ling;
203 int ret = 0;
206 * Options without arguments
209 #ifdef SO_DONTLINGER /* Compatibility item... */
210 if (optname == SO_DONTLINGER) {
211 lock_sock(sk);
212 sock_reset_flag(sk, SOCK_LINGER);
213 release_sock(sk);
214 return 0;
216 #endif
218 if(optlen<sizeof(int))
219 return(-EINVAL);
221 if (get_user(val, (int __user *)optval))
222 return -EFAULT;
224 valbool = val?1:0;
226 lock_sock(sk);
228 switch(optname)
230 case SO_DEBUG:
231 if(val && !capable(CAP_NET_ADMIN))
233 ret = -EACCES;
235 else if (valbool)
236 sock_set_flag(sk, SOCK_DBG);
237 else
238 sock_reset_flag(sk, SOCK_DBG);
239 break;
240 case SO_REUSEADDR:
241 sk->sk_reuse = valbool;
242 break;
243 case SO_TYPE:
244 case SO_ERROR:
245 ret = -ENOPROTOOPT;
246 break;
247 case SO_DONTROUTE:
248 if (valbool)
249 sock_set_flag(sk, SOCK_LOCALROUTE);
250 else
251 sock_reset_flag(sk, SOCK_LOCALROUTE);
252 break;
253 case SO_BROADCAST:
254 sock_valbool_flag(sk, SOCK_BROADCAST, valbool);
255 break;
256 case SO_SNDBUF:
257 /* Don't error on this BSD doesn't and if you think
258 about it this is right. Otherwise apps have to
259 play 'guess the biggest size' games. RCVBUF/SNDBUF
260 are treated in BSD as hints */
262 if (val > sysctl_wmem_max)
263 val = sysctl_wmem_max;
264 set_sndbuf:
265 sk->sk_userlocks |= SOCK_SNDBUF_LOCK;
266 if ((val * 2) < SOCK_MIN_SNDBUF)
267 sk->sk_sndbuf = SOCK_MIN_SNDBUF;
268 else
269 sk->sk_sndbuf = val * 2;
272 * Wake up sending tasks if we
273 * upped the value.
275 sk->sk_write_space(sk);
276 break;
278 case SO_SNDBUFFORCE:
279 if (!capable(CAP_NET_ADMIN)) {
280 ret = -EPERM;
281 break;
283 goto set_sndbuf;
285 case SO_RCVBUF:
286 /* Don't error on this BSD doesn't and if you think
287 about it this is right. Otherwise apps have to
288 play 'guess the biggest size' games. RCVBUF/SNDBUF
289 are treated in BSD as hints */
291 if (val > sysctl_rmem_max)
292 val = sysctl_rmem_max;
293 set_rcvbuf:
294 sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
295 /* FIXME: is this lower bound the right one? */
296 if ((val * 2) < SOCK_MIN_RCVBUF)
297 sk->sk_rcvbuf = SOCK_MIN_RCVBUF;
298 else
299 sk->sk_rcvbuf = val * 2;
300 break;
302 case SO_RCVBUFFORCE:
303 if (!capable(CAP_NET_ADMIN)) {
304 ret = -EPERM;
305 break;
307 goto set_rcvbuf;
309 case SO_KEEPALIVE:
310 #ifdef CONFIG_INET
311 if (sk->sk_protocol == IPPROTO_TCP)
312 tcp_set_keepalive(sk, valbool);
313 #endif
314 sock_valbool_flag(sk, SOCK_KEEPOPEN, valbool);
315 break;
317 case SO_OOBINLINE:
318 sock_valbool_flag(sk, SOCK_URGINLINE, valbool);
319 break;
321 case SO_NO_CHECK:
322 sk->sk_no_check = valbool;
323 break;
325 case SO_PRIORITY:
326 if ((val >= 0 && val <= 6) || capable(CAP_NET_ADMIN))
327 sk->sk_priority = val;
328 else
329 ret = -EPERM;
330 break;
332 case SO_LINGER:
333 if(optlen<sizeof(ling)) {
334 ret = -EINVAL; /* 1003.1g */
335 break;
337 if (copy_from_user(&ling,optval,sizeof(ling))) {
338 ret = -EFAULT;
339 break;
341 if (!ling.l_onoff)
342 sock_reset_flag(sk, SOCK_LINGER);
343 else {
344 #if (BITS_PER_LONG == 32)
345 if ((unsigned int)ling.l_linger >= MAX_SCHEDULE_TIMEOUT/HZ)
346 sk->sk_lingertime = MAX_SCHEDULE_TIMEOUT;
347 else
348 #endif
349 sk->sk_lingertime = (unsigned int)ling.l_linger * HZ;
350 sock_set_flag(sk, SOCK_LINGER);
352 break;
354 case SO_BSDCOMPAT:
355 sock_warn_obsolete_bsdism("setsockopt");
356 break;
358 case SO_PASSCRED:
359 if (valbool)
360 set_bit(SOCK_PASSCRED, &sock->flags);
361 else
362 clear_bit(SOCK_PASSCRED, &sock->flags);
363 break;
365 case SO_TIMESTAMP:
366 if (valbool) {
367 sock_set_flag(sk, SOCK_RCVTSTAMP);
368 sock_enable_timestamp(sk);
369 } else
370 sock_reset_flag(sk, SOCK_RCVTSTAMP);
371 break;
373 case SO_RCVLOWAT:
374 if (val < 0)
375 val = INT_MAX;
376 sk->sk_rcvlowat = val ? : 1;
377 break;
379 case SO_RCVTIMEO:
380 ret = sock_set_timeout(&sk->sk_rcvtimeo, optval, optlen);
381 break;
383 case SO_SNDTIMEO:
384 ret = sock_set_timeout(&sk->sk_sndtimeo, optval, optlen);
385 break;
387 #ifdef CONFIG_NETDEVICES
388 case SO_BINDTODEVICE:
390 char devname[IFNAMSIZ];
392 /* Sorry... */
393 if (!capable(CAP_NET_RAW)) {
394 ret = -EPERM;
395 break;
398 /* Bind this socket to a particular device like "eth0",
399 * as specified in the passed interface name. If the
400 * name is "" or the option length is zero the socket
401 * is not bound.
404 if (!valbool) {
405 sk->sk_bound_dev_if = 0;
406 } else {
407 if (optlen > IFNAMSIZ - 1)
408 optlen = IFNAMSIZ - 1;
409 memset(devname, 0, sizeof(devname));
410 if (copy_from_user(devname, optval, optlen)) {
411 ret = -EFAULT;
412 break;
415 /* Remove any cached route for this socket. */
416 sk_dst_reset(sk);
418 if (devname[0] == '\0') {
419 sk->sk_bound_dev_if = 0;
420 } else {
421 struct net_device *dev = dev_get_by_name(devname);
422 if (!dev) {
423 ret = -ENODEV;
424 break;
426 sk->sk_bound_dev_if = dev->ifindex;
427 dev_put(dev);
430 break;
432 #endif
435 case SO_ATTACH_FILTER:
436 ret = -EINVAL;
437 if (optlen == sizeof(struct sock_fprog)) {
438 struct sock_fprog fprog;
440 ret = -EFAULT;
441 if (copy_from_user(&fprog, optval, sizeof(fprog)))
442 break;
444 ret = sk_attach_filter(&fprog, sk);
446 break;
448 case SO_DETACH_FILTER:
449 spin_lock_bh(&sk->sk_lock.slock);
450 filter = sk->sk_filter;
451 if (filter) {
452 sk->sk_filter = NULL;
453 spin_unlock_bh(&sk->sk_lock.slock);
454 sk_filter_release(sk, filter);
455 break;
457 spin_unlock_bh(&sk->sk_lock.slock);
458 ret = -ENONET;
459 break;
461 /* We implement the SO_SNDLOWAT etc to
462 not be settable (1003.1g 5.3) */
463 default:
464 ret = -ENOPROTOOPT;
465 break;
467 release_sock(sk);
468 return ret;
472 int sock_getsockopt(struct socket *sock, int level, int optname,
473 char __user *optval, int __user *optlen)
475 struct sock *sk = sock->sk;
477 union
479 int val;
480 struct linger ling;
481 struct timeval tm;
482 } v;
484 unsigned int lv = sizeof(int);
485 int len;
487 if(get_user(len,optlen))
488 return -EFAULT;
489 if(len < 0)
490 return -EINVAL;
492 switch(optname)
494 case SO_DEBUG:
495 v.val = sock_flag(sk, SOCK_DBG);
496 break;
498 case SO_DONTROUTE:
499 v.val = sock_flag(sk, SOCK_LOCALROUTE);
500 break;
502 case SO_BROADCAST:
503 v.val = !!sock_flag(sk, SOCK_BROADCAST);
504 break;
506 case SO_SNDBUF:
507 v.val = sk->sk_sndbuf;
508 break;
510 case SO_RCVBUF:
511 v.val = sk->sk_rcvbuf;
512 break;
514 case SO_REUSEADDR:
515 v.val = sk->sk_reuse;
516 break;
518 case SO_KEEPALIVE:
519 v.val = !!sock_flag(sk, SOCK_KEEPOPEN);
520 break;
522 case SO_TYPE:
523 v.val = sk->sk_type;
524 break;
526 case SO_ERROR:
527 v.val = -sock_error(sk);
528 if(v.val==0)
529 v.val = xchg(&sk->sk_err_soft, 0);
530 break;
532 case SO_OOBINLINE:
533 v.val = !!sock_flag(sk, SOCK_URGINLINE);
534 break;
536 case SO_NO_CHECK:
537 v.val = sk->sk_no_check;
538 break;
540 case SO_PRIORITY:
541 v.val = sk->sk_priority;
542 break;
544 case SO_LINGER:
545 lv = sizeof(v.ling);
546 v.ling.l_onoff = !!sock_flag(sk, SOCK_LINGER);
547 v.ling.l_linger = sk->sk_lingertime / HZ;
548 break;
550 case SO_BSDCOMPAT:
551 sock_warn_obsolete_bsdism("getsockopt");
552 break;
554 case SO_TIMESTAMP:
555 v.val = sock_flag(sk, SOCK_RCVTSTAMP);
556 break;
558 case SO_RCVTIMEO:
559 lv=sizeof(struct timeval);
560 if (sk->sk_rcvtimeo == MAX_SCHEDULE_TIMEOUT) {
561 v.tm.tv_sec = 0;
562 v.tm.tv_usec = 0;
563 } else {
564 v.tm.tv_sec = sk->sk_rcvtimeo / HZ;
565 v.tm.tv_usec = ((sk->sk_rcvtimeo % HZ) * 1000000) / HZ;
567 break;
569 case SO_SNDTIMEO:
570 lv=sizeof(struct timeval);
571 if (sk->sk_sndtimeo == MAX_SCHEDULE_TIMEOUT) {
572 v.tm.tv_sec = 0;
573 v.tm.tv_usec = 0;
574 } else {
575 v.tm.tv_sec = sk->sk_sndtimeo / HZ;
576 v.tm.tv_usec = ((sk->sk_sndtimeo % HZ) * 1000000) / HZ;
578 break;
580 case SO_RCVLOWAT:
581 v.val = sk->sk_rcvlowat;
582 break;
584 case SO_SNDLOWAT:
585 v.val=1;
586 break;
588 case SO_PASSCRED:
589 v.val = test_bit(SOCK_PASSCRED, &sock->flags) ? 1 : 0;
590 break;
592 case SO_PEERCRED:
593 if (len > sizeof(sk->sk_peercred))
594 len = sizeof(sk->sk_peercred);
595 if (copy_to_user(optval, &sk->sk_peercred, len))
596 return -EFAULT;
597 goto lenout;
599 case SO_PEERNAME:
601 char address[128];
603 if (sock->ops->getname(sock, (struct sockaddr *)address, &lv, 2))
604 return -ENOTCONN;
605 if (lv < len)
606 return -EINVAL;
607 if (copy_to_user(optval, address, len))
608 return -EFAULT;
609 goto lenout;
612 /* Dubious BSD thing... Probably nobody even uses it, but
613 * the UNIX standard wants it for whatever reason... -DaveM
615 case SO_ACCEPTCONN:
616 v.val = sk->sk_state == TCP_LISTEN;
617 break;
619 case SO_PEERSEC:
620 return security_socket_getpeersec_stream(sock, optval, optlen, len);
622 default:
623 return(-ENOPROTOOPT);
625 if (len > lv)
626 len = lv;
627 if (copy_to_user(optval, &v, len))
628 return -EFAULT;
629 lenout:
630 if (put_user(len, optlen))
631 return -EFAULT;
632 return 0;
636 * sk_alloc - All socket objects are allocated here
637 * @family: protocol family
638 * @priority: for allocation (%GFP_KERNEL, %GFP_ATOMIC, etc)
639 * @prot: struct proto associated with this new sock instance
640 * @zero_it: if we should zero the newly allocated sock
642 struct sock *sk_alloc(int family, gfp_t priority,
643 struct proto *prot, int zero_it)
645 struct sock *sk = NULL;
646 kmem_cache_t *slab = prot->slab;
648 if (slab != NULL)
649 sk = kmem_cache_alloc(slab, priority);
650 else
651 sk = kmalloc(prot->obj_size, priority);
653 if (sk) {
654 if (zero_it) {
655 memset(sk, 0, prot->obj_size);
656 sk->sk_family = family;
658 * See comment in struct sock definition to understand
659 * why we need sk_prot_creator -acme
661 sk->sk_prot = sk->sk_prot_creator = prot;
662 sock_lock_init(sk);
665 if (security_sk_alloc(sk, family, priority))
666 goto out_free;
668 if (!try_module_get(prot->owner))
669 goto out_free;
671 return sk;
673 out_free:
674 if (slab != NULL)
675 kmem_cache_free(slab, sk);
676 else
677 kfree(sk);
678 return NULL;
681 void sk_free(struct sock *sk)
683 struct sk_filter *filter;
684 struct module *owner = sk->sk_prot_creator->owner;
686 if (sk->sk_destruct)
687 sk->sk_destruct(sk);
689 filter = sk->sk_filter;
690 if (filter) {
691 sk_filter_release(sk, filter);
692 sk->sk_filter = NULL;
695 sock_disable_timestamp(sk);
697 if (atomic_read(&sk->sk_omem_alloc))
698 printk(KERN_DEBUG "%s: optmem leakage (%d bytes) detected.\n",
699 __FUNCTION__, atomic_read(&sk->sk_omem_alloc));
701 security_sk_free(sk);
702 if (sk->sk_prot_creator->slab != NULL)
703 kmem_cache_free(sk->sk_prot_creator->slab, sk);
704 else
705 kfree(sk);
706 module_put(owner);
709 struct sock *sk_clone(const struct sock *sk, const gfp_t priority)
711 struct sock *newsk = sk_alloc(sk->sk_family, priority, sk->sk_prot, 0);
713 if (newsk != NULL) {
714 struct sk_filter *filter;
716 memcpy(newsk, sk, sk->sk_prot->obj_size);
718 /* SANITY */
719 sk_node_init(&newsk->sk_node);
720 sock_lock_init(newsk);
721 bh_lock_sock(newsk);
723 atomic_set(&newsk->sk_rmem_alloc, 0);
724 atomic_set(&newsk->sk_wmem_alloc, 0);
725 atomic_set(&newsk->sk_omem_alloc, 0);
726 skb_queue_head_init(&newsk->sk_receive_queue);
727 skb_queue_head_init(&newsk->sk_write_queue);
729 rwlock_init(&newsk->sk_dst_lock);
730 rwlock_init(&newsk->sk_callback_lock);
732 newsk->sk_dst_cache = NULL;
733 newsk->sk_wmem_queued = 0;
734 newsk->sk_forward_alloc = 0;
735 newsk->sk_send_head = NULL;
736 newsk->sk_backlog.head = newsk->sk_backlog.tail = NULL;
737 newsk->sk_userlocks = sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
739 sock_reset_flag(newsk, SOCK_DONE);
740 skb_queue_head_init(&newsk->sk_error_queue);
742 filter = newsk->sk_filter;
743 if (filter != NULL)
744 sk_filter_charge(newsk, filter);
746 if (unlikely(xfrm_sk_clone_policy(newsk))) {
747 /* It is still raw copy of parent, so invalidate
748 * destructor and make plain sk_free() */
749 newsk->sk_destruct = NULL;
750 sk_free(newsk);
751 newsk = NULL;
752 goto out;
755 newsk->sk_err = 0;
756 newsk->sk_priority = 0;
757 atomic_set(&newsk->sk_refcnt, 2);
760 * Increment the counter in the same struct proto as the master
761 * sock (sk_refcnt_debug_inc uses newsk->sk_prot->socks, that
762 * is the same as sk->sk_prot->socks, as this field was copied
763 * with memcpy).
765 * This _changes_ the previous behaviour, where
766 * tcp_create_openreq_child always was incrementing the
767 * equivalent to tcp_prot->socks (inet_sock_nr), so this have
768 * to be taken into account in all callers. -acme
770 sk_refcnt_debug_inc(newsk);
771 newsk->sk_socket = NULL;
772 newsk->sk_sleep = NULL;
774 if (newsk->sk_prot->sockets_allocated)
775 atomic_inc(newsk->sk_prot->sockets_allocated);
777 out:
778 return newsk;
781 EXPORT_SYMBOL_GPL(sk_clone);
783 void __init sk_init(void)
785 if (num_physpages <= 4096) {
786 sysctl_wmem_max = 32767;
787 sysctl_rmem_max = 32767;
788 sysctl_wmem_default = 32767;
789 sysctl_rmem_default = 32767;
790 } else if (num_physpages >= 131072) {
791 sysctl_wmem_max = 131071;
792 sysctl_rmem_max = 131071;
797 * Simple resource managers for sockets.
802 * Write buffer destructor automatically called from kfree_skb.
804 void sock_wfree(struct sk_buff *skb)
806 struct sock *sk = skb->sk;
808 /* In case it might be waiting for more memory. */
809 atomic_sub(skb->truesize, &sk->sk_wmem_alloc);
810 if (!sock_flag(sk, SOCK_USE_WRITE_QUEUE))
811 sk->sk_write_space(sk);
812 sock_put(sk);
816 * Read buffer destructor automatically called from kfree_skb.
818 void sock_rfree(struct sk_buff *skb)
820 struct sock *sk = skb->sk;
822 atomic_sub(skb->truesize, &sk->sk_rmem_alloc);
826 int sock_i_uid(struct sock *sk)
828 int uid;
830 read_lock(&sk->sk_callback_lock);
831 uid = sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_uid : 0;
832 read_unlock(&sk->sk_callback_lock);
833 return uid;
836 unsigned long sock_i_ino(struct sock *sk)
838 unsigned long ino;
840 read_lock(&sk->sk_callback_lock);
841 ino = sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_ino : 0;
842 read_unlock(&sk->sk_callback_lock);
843 return ino;
847 * Allocate a skb from the socket's send buffer.
849 struct sk_buff *sock_wmalloc(struct sock *sk, unsigned long size, int force,
850 gfp_t priority)
852 if (force || atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf) {
853 struct sk_buff * skb = alloc_skb(size, priority);
854 if (skb) {
855 skb_set_owner_w(skb, sk);
856 return skb;
859 return NULL;
863 * Allocate a skb from the socket's receive buffer.
865 struct sk_buff *sock_rmalloc(struct sock *sk, unsigned long size, int force,
866 gfp_t priority)
868 if (force || atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf) {
869 struct sk_buff *skb = alloc_skb(size, priority);
870 if (skb) {
871 skb_set_owner_r(skb, sk);
872 return skb;
875 return NULL;
879 * Allocate a memory block from the socket's option memory buffer.
881 void *sock_kmalloc(struct sock *sk, int size, gfp_t priority)
883 if ((unsigned)size <= sysctl_optmem_max &&
884 atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) {
885 void *mem;
886 /* First do the add, to avoid the race if kmalloc
887 * might sleep.
889 atomic_add(size, &sk->sk_omem_alloc);
890 mem = kmalloc(size, priority);
891 if (mem)
892 return mem;
893 atomic_sub(size, &sk->sk_omem_alloc);
895 return NULL;
899 * Free an option memory block.
901 void sock_kfree_s(struct sock *sk, void *mem, int size)
903 kfree(mem);
904 atomic_sub(size, &sk->sk_omem_alloc);
907 /* It is almost wait_for_tcp_memory minus release_sock/lock_sock.
908 I think, these locks should be removed for datagram sockets.
910 static long sock_wait_for_wmem(struct sock * sk, long timeo)
912 DEFINE_WAIT(wait);
914 clear_bit(SOCK_ASYNC_NOSPACE, &sk->sk_socket->flags);
915 for (;;) {
916 if (!timeo)
917 break;
918 if (signal_pending(current))
919 break;
920 set_bit(SOCK_NOSPACE, &sk->sk_socket->flags);
921 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
922 if (atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf)
923 break;
924 if (sk->sk_shutdown & SEND_SHUTDOWN)
925 break;
926 if (sk->sk_err)
927 break;
928 timeo = schedule_timeout(timeo);
930 finish_wait(sk->sk_sleep, &wait);
931 return timeo;
936 * Generic send/receive buffer handlers
939 static struct sk_buff *sock_alloc_send_pskb(struct sock *sk,
940 unsigned long header_len,
941 unsigned long data_len,
942 int noblock, int *errcode)
944 struct sk_buff *skb;
945 gfp_t gfp_mask;
946 long timeo;
947 int err;
949 gfp_mask = sk->sk_allocation;
950 if (gfp_mask & __GFP_WAIT)
951 gfp_mask |= __GFP_REPEAT;
953 timeo = sock_sndtimeo(sk, noblock);
954 while (1) {
955 err = sock_error(sk);
956 if (err != 0)
957 goto failure;
959 err = -EPIPE;
960 if (sk->sk_shutdown & SEND_SHUTDOWN)
961 goto failure;
963 if (atomic_read(&sk->sk_wmem_alloc) < sk->sk_sndbuf) {
964 skb = alloc_skb(header_len, sk->sk_allocation);
965 if (skb) {
966 int npages;
967 int i;
969 /* No pages, we're done... */
970 if (!data_len)
971 break;
973 npages = (data_len + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
974 skb->truesize += data_len;
975 skb_shinfo(skb)->nr_frags = npages;
976 for (i = 0; i < npages; i++) {
977 struct page *page;
978 skb_frag_t *frag;
980 page = alloc_pages(sk->sk_allocation, 0);
981 if (!page) {
982 err = -ENOBUFS;
983 skb_shinfo(skb)->nr_frags = i;
984 kfree_skb(skb);
985 goto failure;
988 frag = &skb_shinfo(skb)->frags[i];
989 frag->page = page;
990 frag->page_offset = 0;
991 frag->size = (data_len >= PAGE_SIZE ?
992 PAGE_SIZE :
993 data_len);
994 data_len -= PAGE_SIZE;
997 /* Full success... */
998 break;
1000 err = -ENOBUFS;
1001 goto failure;
1003 set_bit(SOCK_ASYNC_NOSPACE, &sk->sk_socket->flags);
1004 set_bit(SOCK_NOSPACE, &sk->sk_socket->flags);
1005 err = -EAGAIN;
1006 if (!timeo)
1007 goto failure;
1008 if (signal_pending(current))
1009 goto interrupted;
1010 timeo = sock_wait_for_wmem(sk, timeo);
1013 skb_set_owner_w(skb, sk);
1014 return skb;
1016 interrupted:
1017 err = sock_intr_errno(timeo);
1018 failure:
1019 *errcode = err;
1020 return NULL;
1023 struct sk_buff *sock_alloc_send_skb(struct sock *sk, unsigned long size,
1024 int noblock, int *errcode)
1026 return sock_alloc_send_pskb(sk, size, 0, noblock, errcode);
1029 static void __lock_sock(struct sock *sk)
1031 DEFINE_WAIT(wait);
1033 for(;;) {
1034 prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait,
1035 TASK_UNINTERRUPTIBLE);
1036 spin_unlock_bh(&sk->sk_lock.slock);
1037 schedule();
1038 spin_lock_bh(&sk->sk_lock.slock);
1039 if(!sock_owned_by_user(sk))
1040 break;
1042 finish_wait(&sk->sk_lock.wq, &wait);
1045 static void __release_sock(struct sock *sk)
1047 struct sk_buff *skb = sk->sk_backlog.head;
1049 do {
1050 sk->sk_backlog.head = sk->sk_backlog.tail = NULL;
1051 bh_unlock_sock(sk);
1053 do {
1054 struct sk_buff *next = skb->next;
1056 skb->next = NULL;
1057 sk->sk_backlog_rcv(sk, skb);
1060 * We are in process context here with softirqs
1061 * disabled, use cond_resched_softirq() to preempt.
1062 * This is safe to do because we've taken the backlog
1063 * queue private:
1065 cond_resched_softirq();
1067 skb = next;
1068 } while (skb != NULL);
1070 bh_lock_sock(sk);
1071 } while((skb = sk->sk_backlog.head) != NULL);
1075 * sk_wait_data - wait for data to arrive at sk_receive_queue
1076 * @sk: sock to wait on
1077 * @timeo: for how long
1079 * Now socket state including sk->sk_err is changed only under lock,
1080 * hence we may omit checks after joining wait queue.
1081 * We check receive queue before schedule() only as optimization;
1082 * it is very likely that release_sock() added new data.
1084 int sk_wait_data(struct sock *sk, long *timeo)
1086 int rc;
1087 DEFINE_WAIT(wait);
1089 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
1090 set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1091 rc = sk_wait_event(sk, timeo, !skb_queue_empty(&sk->sk_receive_queue));
1092 clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1093 finish_wait(sk->sk_sleep, &wait);
1094 return rc;
1097 EXPORT_SYMBOL(sk_wait_data);
1100 * Set of default routines for initialising struct proto_ops when
1101 * the protocol does not support a particular function. In certain
1102 * cases where it makes no sense for a protocol to have a "do nothing"
1103 * function, some default processing is provided.
1106 int sock_no_bind(struct socket *sock, struct sockaddr *saddr, int len)
1108 return -EOPNOTSUPP;
1111 int sock_no_connect(struct socket *sock, struct sockaddr *saddr,
1112 int len, int flags)
1114 return -EOPNOTSUPP;
1117 int sock_no_socketpair(struct socket *sock1, struct socket *sock2)
1119 return -EOPNOTSUPP;
1122 int sock_no_accept(struct socket *sock, struct socket *newsock, int flags)
1124 return -EOPNOTSUPP;
1127 int sock_no_getname(struct socket *sock, struct sockaddr *saddr,
1128 int *len, int peer)
1130 return -EOPNOTSUPP;
1133 unsigned int sock_no_poll(struct file * file, struct socket *sock, poll_table *pt)
1135 return 0;
1138 int sock_no_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg)
1140 return -EOPNOTSUPP;
1143 int sock_no_listen(struct socket *sock, int backlog)
1145 return -EOPNOTSUPP;
1148 int sock_no_shutdown(struct socket *sock, int how)
1150 return -EOPNOTSUPP;
1153 int sock_no_setsockopt(struct socket *sock, int level, int optname,
1154 char __user *optval, int optlen)
1156 return -EOPNOTSUPP;
1159 int sock_no_getsockopt(struct socket *sock, int level, int optname,
1160 char __user *optval, int __user *optlen)
1162 return -EOPNOTSUPP;
1165 int sock_no_sendmsg(struct kiocb *iocb, struct socket *sock, struct msghdr *m,
1166 size_t len)
1168 return -EOPNOTSUPP;
1171 int sock_no_recvmsg(struct kiocb *iocb, struct socket *sock, struct msghdr *m,
1172 size_t len, int flags)
1174 return -EOPNOTSUPP;
1177 int sock_no_mmap(struct file *file, struct socket *sock, struct vm_area_struct *vma)
1179 /* Mirror missing mmap method error code */
1180 return -ENODEV;
1183 ssize_t sock_no_sendpage(struct socket *sock, struct page *page, int offset, size_t size, int flags)
1185 ssize_t res;
1186 struct msghdr msg = {.msg_flags = flags};
1187 struct kvec iov;
1188 char *kaddr = kmap(page);
1189 iov.iov_base = kaddr + offset;
1190 iov.iov_len = size;
1191 res = kernel_sendmsg(sock, &msg, &iov, 1, size);
1192 kunmap(page);
1193 return res;
1197 * Default Socket Callbacks
1200 static void sock_def_wakeup(struct sock *sk)
1202 read_lock(&sk->sk_callback_lock);
1203 if (sk->sk_sleep && waitqueue_active(sk->sk_sleep))
1204 wake_up_interruptible_all(sk->sk_sleep);
1205 read_unlock(&sk->sk_callback_lock);
1208 static void sock_def_error_report(struct sock *sk)
1210 read_lock(&sk->sk_callback_lock);
1211 if (sk->sk_sleep && waitqueue_active(sk->sk_sleep))
1212 wake_up_interruptible(sk->sk_sleep);
1213 sk_wake_async(sk,0,POLL_ERR);
1214 read_unlock(&sk->sk_callback_lock);
1217 static void sock_def_readable(struct sock *sk, int len)
1219 read_lock(&sk->sk_callback_lock);
1220 if (sk->sk_sleep && waitqueue_active(sk->sk_sleep))
1221 wake_up_interruptible(sk->sk_sleep);
1222 sk_wake_async(sk,1,POLL_IN);
1223 read_unlock(&sk->sk_callback_lock);
1226 static void sock_def_write_space(struct sock *sk)
1228 read_lock(&sk->sk_callback_lock);
1230 /* Do not wake up a writer until he can make "significant"
1231 * progress. --DaveM
1233 if((atomic_read(&sk->sk_wmem_alloc) << 1) <= sk->sk_sndbuf) {
1234 if (sk->sk_sleep && waitqueue_active(sk->sk_sleep))
1235 wake_up_interruptible(sk->sk_sleep);
1237 /* Should agree with poll, otherwise some programs break */
1238 if (sock_writeable(sk))
1239 sk_wake_async(sk, 2, POLL_OUT);
1242 read_unlock(&sk->sk_callback_lock);
1245 static void sock_def_destruct(struct sock *sk)
1247 kfree(sk->sk_protinfo);
1250 void sk_send_sigurg(struct sock *sk)
1252 if (sk->sk_socket && sk->sk_socket->file)
1253 if (send_sigurg(&sk->sk_socket->file->f_owner))
1254 sk_wake_async(sk, 3, POLL_PRI);
1257 void sk_reset_timer(struct sock *sk, struct timer_list* timer,
1258 unsigned long expires)
1260 if (!mod_timer(timer, expires))
1261 sock_hold(sk);
1264 EXPORT_SYMBOL(sk_reset_timer);
1266 void sk_stop_timer(struct sock *sk, struct timer_list* timer)
1268 if (timer_pending(timer) && del_timer(timer))
1269 __sock_put(sk);
1272 EXPORT_SYMBOL(sk_stop_timer);
1274 void sock_init_data(struct socket *sock, struct sock *sk)
1276 skb_queue_head_init(&sk->sk_receive_queue);
1277 skb_queue_head_init(&sk->sk_write_queue);
1278 skb_queue_head_init(&sk->sk_error_queue);
1280 sk->sk_send_head = NULL;
1282 init_timer(&sk->sk_timer);
1284 sk->sk_allocation = GFP_KERNEL;
1285 sk->sk_rcvbuf = sysctl_rmem_default;
1286 sk->sk_sndbuf = sysctl_wmem_default;
1287 sk->sk_state = TCP_CLOSE;
1288 sk->sk_socket = sock;
1290 sock_set_flag(sk, SOCK_ZAPPED);
1292 if(sock)
1294 sk->sk_type = sock->type;
1295 sk->sk_sleep = &sock->wait;
1296 sock->sk = sk;
1297 } else
1298 sk->sk_sleep = NULL;
1300 rwlock_init(&sk->sk_dst_lock);
1301 rwlock_init(&sk->sk_callback_lock);
1303 sk->sk_state_change = sock_def_wakeup;
1304 sk->sk_data_ready = sock_def_readable;
1305 sk->sk_write_space = sock_def_write_space;
1306 sk->sk_error_report = sock_def_error_report;
1307 sk->sk_destruct = sock_def_destruct;
1309 sk->sk_sndmsg_page = NULL;
1310 sk->sk_sndmsg_off = 0;
1312 sk->sk_peercred.pid = 0;
1313 sk->sk_peercred.uid = -1;
1314 sk->sk_peercred.gid = -1;
1315 sk->sk_write_pending = 0;
1316 sk->sk_rcvlowat = 1;
1317 sk->sk_rcvtimeo = MAX_SCHEDULE_TIMEOUT;
1318 sk->sk_sndtimeo = MAX_SCHEDULE_TIMEOUT;
1320 sk->sk_stamp.tv_sec = -1L;
1321 sk->sk_stamp.tv_usec = -1L;
1323 atomic_set(&sk->sk_refcnt, 1);
1326 void fastcall lock_sock(struct sock *sk)
1328 might_sleep();
1329 spin_lock_bh(&(sk->sk_lock.slock));
1330 if (sk->sk_lock.owner)
1331 __lock_sock(sk);
1332 sk->sk_lock.owner = (void *)1;
1333 spin_unlock_bh(&(sk->sk_lock.slock));
1336 EXPORT_SYMBOL(lock_sock);
1338 void fastcall release_sock(struct sock *sk)
1340 spin_lock_bh(&(sk->sk_lock.slock));
1341 if (sk->sk_backlog.tail)
1342 __release_sock(sk);
1343 sk->sk_lock.owner = NULL;
1344 if (waitqueue_active(&(sk->sk_lock.wq)))
1345 wake_up(&(sk->sk_lock.wq));
1346 spin_unlock_bh(&(sk->sk_lock.slock));
1348 EXPORT_SYMBOL(release_sock);
1350 int sock_get_timestamp(struct sock *sk, struct timeval __user *userstamp)
1352 if (!sock_flag(sk, SOCK_TIMESTAMP))
1353 sock_enable_timestamp(sk);
1354 if (sk->sk_stamp.tv_sec == -1)
1355 return -ENOENT;
1356 if (sk->sk_stamp.tv_sec == 0)
1357 do_gettimeofday(&sk->sk_stamp);
1358 return copy_to_user(userstamp, &sk->sk_stamp, sizeof(struct timeval)) ?
1359 -EFAULT : 0;
1361 EXPORT_SYMBOL(sock_get_timestamp);
1363 void sock_enable_timestamp(struct sock *sk)
1365 if (!sock_flag(sk, SOCK_TIMESTAMP)) {
1366 sock_set_flag(sk, SOCK_TIMESTAMP);
1367 net_enable_timestamp();
1370 EXPORT_SYMBOL(sock_enable_timestamp);
1373 * Get a socket option on an socket.
1375 * FIX: POSIX 1003.1g is very ambiguous here. It states that
1376 * asynchronous errors should be reported by getsockopt. We assume
1377 * this means if you specify SO_ERROR (otherwise whats the point of it).
1379 int sock_common_getsockopt(struct socket *sock, int level, int optname,
1380 char __user *optval, int __user *optlen)
1382 struct sock *sk = sock->sk;
1384 return sk->sk_prot->getsockopt(sk, level, optname, optval, optlen);
1387 EXPORT_SYMBOL(sock_common_getsockopt);
1389 #ifdef CONFIG_COMPAT
1390 int compat_sock_common_getsockopt(struct socket *sock, int level, int optname,
1391 char __user *optval, int __user *optlen)
1393 struct sock *sk = sock->sk;
1395 if (sk->sk_prot->compat_setsockopt != NULL)
1396 return sk->sk_prot->compat_getsockopt(sk, level, optname,
1397 optval, optlen);
1398 return sk->sk_prot->getsockopt(sk, level, optname, optval, optlen);
1400 EXPORT_SYMBOL(compat_sock_common_getsockopt);
1401 #endif
1403 int sock_common_recvmsg(struct kiocb *iocb, struct socket *sock,
1404 struct msghdr *msg, size_t size, int flags)
1406 struct sock *sk = sock->sk;
1407 int addr_len = 0;
1408 int err;
1410 err = sk->sk_prot->recvmsg(iocb, sk, msg, size, flags & MSG_DONTWAIT,
1411 flags & ~MSG_DONTWAIT, &addr_len);
1412 if (err >= 0)
1413 msg->msg_namelen = addr_len;
1414 return err;
1417 EXPORT_SYMBOL(sock_common_recvmsg);
1420 * Set socket options on an inet socket.
1422 int sock_common_setsockopt(struct socket *sock, int level, int optname,
1423 char __user *optval, int optlen)
1425 struct sock *sk = sock->sk;
1427 return sk->sk_prot->setsockopt(sk, level, optname, optval, optlen);
1430 EXPORT_SYMBOL(sock_common_setsockopt);
1432 #ifdef CONFIG_COMPAT
1433 int compat_sock_common_setsockopt(struct socket *sock, int level, int optname,
1434 char __user *optval, int optlen)
1436 struct sock *sk = sock->sk;
1438 if (sk->sk_prot->compat_setsockopt != NULL)
1439 return sk->sk_prot->compat_setsockopt(sk, level, optname,
1440 optval, optlen);
1441 return sk->sk_prot->setsockopt(sk, level, optname, optval, optlen);
1443 EXPORT_SYMBOL(compat_sock_common_setsockopt);
1444 #endif
1446 void sk_common_release(struct sock *sk)
1448 if (sk->sk_prot->destroy)
1449 sk->sk_prot->destroy(sk);
1452 * Observation: when sock_common_release is called, processes have
1453 * no access to socket. But net still has.
1454 * Step one, detach it from networking:
1456 * A. Remove from hash tables.
1459 sk->sk_prot->unhash(sk);
1462 * In this point socket cannot receive new packets, but it is possible
1463 * that some packets are in flight because some CPU runs receiver and
1464 * did hash table lookup before we unhashed socket. They will achieve
1465 * receive queue and will be purged by socket destructor.
1467 * Also we still have packets pending on receive queue and probably,
1468 * our own packets waiting in device queues. sock_destroy will drain
1469 * receive queue, but transmitted packets will delay socket destruction
1470 * until the last reference will be released.
1473 sock_orphan(sk);
1475 xfrm_sk_free_policy(sk);
1477 sk_refcnt_debug_release(sk);
1478 sock_put(sk);
1481 EXPORT_SYMBOL(sk_common_release);
1483 static DEFINE_RWLOCK(proto_list_lock);
1484 static LIST_HEAD(proto_list);
1486 int proto_register(struct proto *prot, int alloc_slab)
1488 char *request_sock_slab_name = NULL;
1489 char *timewait_sock_slab_name;
1490 int rc = -ENOBUFS;
1492 if (alloc_slab) {
1493 prot->slab = kmem_cache_create(prot->name, prot->obj_size, 0,
1494 SLAB_HWCACHE_ALIGN, NULL, NULL);
1496 if (prot->slab == NULL) {
1497 printk(KERN_CRIT "%s: Can't create sock SLAB cache!\n",
1498 prot->name);
1499 goto out;
1502 if (prot->rsk_prot != NULL) {
1503 static const char mask[] = "request_sock_%s";
1505 request_sock_slab_name = kmalloc(strlen(prot->name) + sizeof(mask) - 1, GFP_KERNEL);
1506 if (request_sock_slab_name == NULL)
1507 goto out_free_sock_slab;
1509 sprintf(request_sock_slab_name, mask, prot->name);
1510 prot->rsk_prot->slab = kmem_cache_create(request_sock_slab_name,
1511 prot->rsk_prot->obj_size, 0,
1512 SLAB_HWCACHE_ALIGN, NULL, NULL);
1514 if (prot->rsk_prot->slab == NULL) {
1515 printk(KERN_CRIT "%s: Can't create request sock SLAB cache!\n",
1516 prot->name);
1517 goto out_free_request_sock_slab_name;
1521 if (prot->twsk_prot != NULL) {
1522 static const char mask[] = "tw_sock_%s";
1524 timewait_sock_slab_name = kmalloc(strlen(prot->name) + sizeof(mask) - 1, GFP_KERNEL);
1526 if (timewait_sock_slab_name == NULL)
1527 goto out_free_request_sock_slab;
1529 sprintf(timewait_sock_slab_name, mask, prot->name);
1530 prot->twsk_prot->twsk_slab =
1531 kmem_cache_create(timewait_sock_slab_name,
1532 prot->twsk_prot->twsk_obj_size,
1533 0, SLAB_HWCACHE_ALIGN,
1534 NULL, NULL);
1535 if (prot->twsk_prot->twsk_slab == NULL)
1536 goto out_free_timewait_sock_slab_name;
1540 write_lock(&proto_list_lock);
1541 list_add(&prot->node, &proto_list);
1542 write_unlock(&proto_list_lock);
1543 rc = 0;
1544 out:
1545 return rc;
1546 out_free_timewait_sock_slab_name:
1547 kfree(timewait_sock_slab_name);
1548 out_free_request_sock_slab:
1549 if (prot->rsk_prot && prot->rsk_prot->slab) {
1550 kmem_cache_destroy(prot->rsk_prot->slab);
1551 prot->rsk_prot->slab = NULL;
1553 out_free_request_sock_slab_name:
1554 kfree(request_sock_slab_name);
1555 out_free_sock_slab:
1556 kmem_cache_destroy(prot->slab);
1557 prot->slab = NULL;
1558 goto out;
1561 EXPORT_SYMBOL(proto_register);
1563 void proto_unregister(struct proto *prot)
1565 write_lock(&proto_list_lock);
1566 list_del(&prot->node);
1567 write_unlock(&proto_list_lock);
1569 if (prot->slab != NULL) {
1570 kmem_cache_destroy(prot->slab);
1571 prot->slab = NULL;
1574 if (prot->rsk_prot != NULL && prot->rsk_prot->slab != NULL) {
1575 const char *name = kmem_cache_name(prot->rsk_prot->slab);
1577 kmem_cache_destroy(prot->rsk_prot->slab);
1578 kfree(name);
1579 prot->rsk_prot->slab = NULL;
1582 if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) {
1583 const char *name = kmem_cache_name(prot->twsk_prot->twsk_slab);
1585 kmem_cache_destroy(prot->twsk_prot->twsk_slab);
1586 kfree(name);
1587 prot->twsk_prot->twsk_slab = NULL;
1591 EXPORT_SYMBOL(proto_unregister);
1593 #ifdef CONFIG_PROC_FS
1594 static inline struct proto *__proto_head(void)
1596 return list_entry(proto_list.next, struct proto, node);
1599 static inline struct proto *proto_head(void)
1601 return list_empty(&proto_list) ? NULL : __proto_head();
1604 static inline struct proto *proto_next(struct proto *proto)
1606 return proto->node.next == &proto_list ? NULL :
1607 list_entry(proto->node.next, struct proto, node);
1610 static inline struct proto *proto_get_idx(loff_t pos)
1612 struct proto *proto;
1613 loff_t i = 0;
1615 list_for_each_entry(proto, &proto_list, node)
1616 if (i++ == pos)
1617 goto out;
1619 proto = NULL;
1620 out:
1621 return proto;
1624 static void *proto_seq_start(struct seq_file *seq, loff_t *pos)
1626 read_lock(&proto_list_lock);
1627 return *pos ? proto_get_idx(*pos - 1) : SEQ_START_TOKEN;
1630 static void *proto_seq_next(struct seq_file *seq, void *v, loff_t *pos)
1632 ++*pos;
1633 return v == SEQ_START_TOKEN ? proto_head() : proto_next(v);
1636 static void proto_seq_stop(struct seq_file *seq, void *v)
1638 read_unlock(&proto_list_lock);
1641 static char proto_method_implemented(const void *method)
1643 return method == NULL ? 'n' : 'y';
1646 static void proto_seq_printf(struct seq_file *seq, struct proto *proto)
1648 seq_printf(seq, "%-9s %4u %6d %6d %-3s %6u %-3s %-10s "
1649 "%2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c %2c\n",
1650 proto->name,
1651 proto->obj_size,
1652 proto->sockets_allocated != NULL ? atomic_read(proto->sockets_allocated) : -1,
1653 proto->memory_allocated != NULL ? atomic_read(proto->memory_allocated) : -1,
1654 proto->memory_pressure != NULL ? *proto->memory_pressure ? "yes" : "no" : "NI",
1655 proto->max_header,
1656 proto->slab == NULL ? "no" : "yes",
1657 module_name(proto->owner),
1658 proto_method_implemented(proto->close),
1659 proto_method_implemented(proto->connect),
1660 proto_method_implemented(proto->disconnect),
1661 proto_method_implemented(proto->accept),
1662 proto_method_implemented(proto->ioctl),
1663 proto_method_implemented(proto->init),
1664 proto_method_implemented(proto->destroy),
1665 proto_method_implemented(proto->shutdown),
1666 proto_method_implemented(proto->setsockopt),
1667 proto_method_implemented(proto->getsockopt),
1668 proto_method_implemented(proto->sendmsg),
1669 proto_method_implemented(proto->recvmsg),
1670 proto_method_implemented(proto->sendpage),
1671 proto_method_implemented(proto->bind),
1672 proto_method_implemented(proto->backlog_rcv),
1673 proto_method_implemented(proto->hash),
1674 proto_method_implemented(proto->unhash),
1675 proto_method_implemented(proto->get_port),
1676 proto_method_implemented(proto->enter_memory_pressure));
1679 static int proto_seq_show(struct seq_file *seq, void *v)
1681 if (v == SEQ_START_TOKEN)
1682 seq_printf(seq, "%-9s %-4s %-8s %-6s %-5s %-7s %-4s %-10s %s",
1683 "protocol",
1684 "size",
1685 "sockets",
1686 "memory",
1687 "press",
1688 "maxhdr",
1689 "slab",
1690 "module",
1691 "cl co di ac io in de sh ss gs se re sp bi br ha uh gp em\n");
1692 else
1693 proto_seq_printf(seq, v);
1694 return 0;
1697 static struct seq_operations proto_seq_ops = {
1698 .start = proto_seq_start,
1699 .next = proto_seq_next,
1700 .stop = proto_seq_stop,
1701 .show = proto_seq_show,
1704 static int proto_seq_open(struct inode *inode, struct file *file)
1706 return seq_open(file, &proto_seq_ops);
1709 static struct file_operations proto_seq_fops = {
1710 .owner = THIS_MODULE,
1711 .open = proto_seq_open,
1712 .read = seq_read,
1713 .llseek = seq_lseek,
1714 .release = seq_release,
1717 static int __init proto_init(void)
1719 /* register /proc/net/protocols */
1720 return proc_net_fops_create("protocols", S_IRUGO, &proto_seq_fops) == NULL ? -ENOBUFS : 0;
1723 subsys_initcall(proto_init);
1725 #endif /* PROC_FS */
1727 EXPORT_SYMBOL(sk_alloc);
1728 EXPORT_SYMBOL(sk_free);
1729 EXPORT_SYMBOL(sk_send_sigurg);
1730 EXPORT_SYMBOL(sock_alloc_send_skb);
1731 EXPORT_SYMBOL(sock_init_data);
1732 EXPORT_SYMBOL(sock_kfree_s);
1733 EXPORT_SYMBOL(sock_kmalloc);
1734 EXPORT_SYMBOL(sock_no_accept);
1735 EXPORT_SYMBOL(sock_no_bind);
1736 EXPORT_SYMBOL(sock_no_connect);
1737 EXPORT_SYMBOL(sock_no_getname);
1738 EXPORT_SYMBOL(sock_no_getsockopt);
1739 EXPORT_SYMBOL(sock_no_ioctl);
1740 EXPORT_SYMBOL(sock_no_listen);
1741 EXPORT_SYMBOL(sock_no_mmap);
1742 EXPORT_SYMBOL(sock_no_poll);
1743 EXPORT_SYMBOL(sock_no_recvmsg);
1744 EXPORT_SYMBOL(sock_no_sendmsg);
1745 EXPORT_SYMBOL(sock_no_sendpage);
1746 EXPORT_SYMBOL(sock_no_setsockopt);
1747 EXPORT_SYMBOL(sock_no_shutdown);
1748 EXPORT_SYMBOL(sock_no_socketpair);
1749 EXPORT_SYMBOL(sock_rfree);
1750 EXPORT_SYMBOL(sock_setsockopt);
1751 EXPORT_SYMBOL(sock_wfree);
1752 EXPORT_SYMBOL(sock_wmalloc);
1753 EXPORT_SYMBOL(sock_i_uid);
1754 EXPORT_SYMBOL(sock_i_ino);
1755 EXPORT_SYMBOL(sysctl_optmem_max);
1756 #ifdef CONFIG_SYSCTL
1757 EXPORT_SYMBOL(sysctl_rmem_max);
1758 EXPORT_SYMBOL(sysctl_wmem_max);
1759 #endif