ACPI: thinkpad-acpi: add development version tag
[linux-2.6/linux-acpi-2.6/ibm-acpi-2.6.git] / net / sunrpc / svcauth_unix.c
blob97cc3de7432e83702ae7913cdf4eb5d93bd4594f
1 #include <linux/types.h>
2 #include <linux/sched.h>
3 #include <linux/module.h>
4 #include <linux/sunrpc/types.h>
5 #include <linux/sunrpc/xdr.h>
6 #include <linux/sunrpc/svcsock.h>
7 #include <linux/sunrpc/svcauth.h>
8 #include <linux/sunrpc/gss_api.h>
9 #include <linux/err.h>
10 #include <linux/seq_file.h>
11 #include <linux/hash.h>
12 #include <linux/string.h>
13 #include <net/sock.h>
14 #include <net/ipv6.h>
15 #include <linux/kernel.h>
16 #define RPCDBG_FACILITY RPCDBG_AUTH
20 * AUTHUNIX and AUTHNULL credentials are both handled here.
21 * AUTHNULL is treated just like AUTHUNIX except that the uid/gid
22 * are always nobody (-2). i.e. we do the same IP address checks for
23 * AUTHNULL as for AUTHUNIX, and that is done here.
27 struct unix_domain {
28 struct auth_domain h;
29 int addr_changes;
30 /* other stuff later */
33 extern struct auth_ops svcauth_unix;
35 struct auth_domain *unix_domain_find(char *name)
37 struct auth_domain *rv;
38 struct unix_domain *new = NULL;
40 rv = auth_domain_lookup(name, NULL);
41 while(1) {
42 if (rv) {
43 if (new && rv != &new->h)
44 auth_domain_put(&new->h);
46 if (rv->flavour != &svcauth_unix) {
47 auth_domain_put(rv);
48 return NULL;
50 return rv;
53 new = kmalloc(sizeof(*new), GFP_KERNEL);
54 if (new == NULL)
55 return NULL;
56 kref_init(&new->h.ref);
57 new->h.name = kstrdup(name, GFP_KERNEL);
58 if (new->h.name == NULL) {
59 kfree(new);
60 return NULL;
62 new->h.flavour = &svcauth_unix;
63 new->addr_changes = 0;
64 rv = auth_domain_lookup(name, &new->h);
67 EXPORT_SYMBOL_GPL(unix_domain_find);
69 static void svcauth_unix_domain_release(struct auth_domain *dom)
71 struct unix_domain *ud = container_of(dom, struct unix_domain, h);
73 kfree(dom->name);
74 kfree(ud);
78 /**************************************************
79 * cache for IP address to unix_domain
80 * as needed by AUTH_UNIX
82 #define IP_HASHBITS 8
83 #define IP_HASHMAX (1<<IP_HASHBITS)
84 #define IP_HASHMASK (IP_HASHMAX-1)
86 struct ip_map {
87 struct cache_head h;
88 char m_class[8]; /* e.g. "nfsd" */
89 struct in6_addr m_addr;
90 struct unix_domain *m_client;
91 int m_add_change;
93 static struct cache_head *ip_table[IP_HASHMAX];
95 static void ip_map_put(struct kref *kref)
97 struct cache_head *item = container_of(kref, struct cache_head, ref);
98 struct ip_map *im = container_of(item, struct ip_map,h);
100 if (test_bit(CACHE_VALID, &item->flags) &&
101 !test_bit(CACHE_NEGATIVE, &item->flags))
102 auth_domain_put(&im->m_client->h);
103 kfree(im);
106 #if IP_HASHBITS == 8
107 /* hash_long on a 64 bit machine is currently REALLY BAD for
108 * IP addresses in reverse-endian (i.e. on a little-endian machine).
109 * So use a trivial but reliable hash instead
111 static inline int hash_ip(__be32 ip)
113 int hash = (__force u32)ip ^ ((__force u32)ip>>16);
114 return (hash ^ (hash>>8)) & 0xff;
116 #endif
117 static inline int hash_ip6(struct in6_addr ip)
119 return (hash_ip(ip.s6_addr32[0]) ^
120 hash_ip(ip.s6_addr32[1]) ^
121 hash_ip(ip.s6_addr32[2]) ^
122 hash_ip(ip.s6_addr32[3]));
124 static int ip_map_match(struct cache_head *corig, struct cache_head *cnew)
126 struct ip_map *orig = container_of(corig, struct ip_map, h);
127 struct ip_map *new = container_of(cnew, struct ip_map, h);
128 return strcmp(orig->m_class, new->m_class) == 0
129 && ipv6_addr_equal(&orig->m_addr, &new->m_addr);
131 static void ip_map_init(struct cache_head *cnew, struct cache_head *citem)
133 struct ip_map *new = container_of(cnew, struct ip_map, h);
134 struct ip_map *item = container_of(citem, struct ip_map, h);
136 strcpy(new->m_class, item->m_class);
137 ipv6_addr_copy(&new->m_addr, &item->m_addr);
139 static void update(struct cache_head *cnew, struct cache_head *citem)
141 struct ip_map *new = container_of(cnew, struct ip_map, h);
142 struct ip_map *item = container_of(citem, struct ip_map, h);
144 kref_get(&item->m_client->h.ref);
145 new->m_client = item->m_client;
146 new->m_add_change = item->m_add_change;
148 static struct cache_head *ip_map_alloc(void)
150 struct ip_map *i = kmalloc(sizeof(*i), GFP_KERNEL);
151 if (i)
152 return &i->h;
153 else
154 return NULL;
157 static void ip_map_request(struct cache_detail *cd,
158 struct cache_head *h,
159 char **bpp, int *blen)
161 char text_addr[40];
162 struct ip_map *im = container_of(h, struct ip_map, h);
164 if (ipv6_addr_v4mapped(&(im->m_addr))) {
165 snprintf(text_addr, 20, "%pI4", &im->m_addr.s6_addr32[3]);
166 } else {
167 snprintf(text_addr, 40, "%pI6", &im->m_addr);
169 qword_add(bpp, blen, im->m_class);
170 qword_add(bpp, blen, text_addr);
171 (*bpp)[-1] = '\n';
174 static int ip_map_upcall(struct cache_detail *cd, struct cache_head *h)
176 return sunrpc_cache_pipe_upcall(cd, h, ip_map_request);
179 static struct ip_map *ip_map_lookup(char *class, struct in6_addr *addr);
180 static int ip_map_update(struct ip_map *ipm, struct unix_domain *udom, time_t expiry);
182 static int ip_map_parse(struct cache_detail *cd,
183 char *mesg, int mlen)
185 /* class ipaddress [domainname] */
186 /* should be safe just to use the start of the input buffer
187 * for scratch: */
188 char *buf = mesg;
189 int len;
190 int b1, b2, b3, b4, b5, b6, b7, b8;
191 char c;
192 char class[8];
193 struct in6_addr addr;
194 int err;
196 struct ip_map *ipmp;
197 struct auth_domain *dom;
198 time_t expiry;
200 if (mesg[mlen-1] != '\n')
201 return -EINVAL;
202 mesg[mlen-1] = 0;
204 /* class */
205 len = qword_get(&mesg, class, sizeof(class));
206 if (len <= 0) return -EINVAL;
208 /* ip address */
209 len = qword_get(&mesg, buf, mlen);
210 if (len <= 0) return -EINVAL;
212 if (sscanf(buf, "%u.%u.%u.%u%c", &b1, &b2, &b3, &b4, &c) == 4) {
213 addr.s6_addr32[0] = 0;
214 addr.s6_addr32[1] = 0;
215 addr.s6_addr32[2] = htonl(0xffff);
216 addr.s6_addr32[3] =
217 htonl((((((b1<<8)|b2)<<8)|b3)<<8)|b4);
218 } else if (sscanf(buf, "%04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x%c",
219 &b1, &b2, &b3, &b4, &b5, &b6, &b7, &b8, &c) == 8) {
220 addr.s6_addr16[0] = htons(b1);
221 addr.s6_addr16[1] = htons(b2);
222 addr.s6_addr16[2] = htons(b3);
223 addr.s6_addr16[3] = htons(b4);
224 addr.s6_addr16[4] = htons(b5);
225 addr.s6_addr16[5] = htons(b6);
226 addr.s6_addr16[6] = htons(b7);
227 addr.s6_addr16[7] = htons(b8);
228 } else
229 return -EINVAL;
231 expiry = get_expiry(&mesg);
232 if (expiry ==0)
233 return -EINVAL;
235 /* domainname, or empty for NEGATIVE */
236 len = qword_get(&mesg, buf, mlen);
237 if (len < 0) return -EINVAL;
239 if (len) {
240 dom = unix_domain_find(buf);
241 if (dom == NULL)
242 return -ENOENT;
243 } else
244 dom = NULL;
246 ipmp = ip_map_lookup(class, &addr);
247 if (ipmp) {
248 err = ip_map_update(ipmp,
249 container_of(dom, struct unix_domain, h),
250 expiry);
251 } else
252 err = -ENOMEM;
254 if (dom)
255 auth_domain_put(dom);
257 cache_flush();
258 return err;
261 static int ip_map_show(struct seq_file *m,
262 struct cache_detail *cd,
263 struct cache_head *h)
265 struct ip_map *im;
266 struct in6_addr addr;
267 char *dom = "-no-domain-";
269 if (h == NULL) {
270 seq_puts(m, "#class IP domain\n");
271 return 0;
273 im = container_of(h, struct ip_map, h);
274 /* class addr domain */
275 ipv6_addr_copy(&addr, &im->m_addr);
277 if (test_bit(CACHE_VALID, &h->flags) &&
278 !test_bit(CACHE_NEGATIVE, &h->flags))
279 dom = im->m_client->h.name;
281 if (ipv6_addr_v4mapped(&addr)) {
282 seq_printf(m, "%s %pI4 %s\n",
283 im->m_class, &addr.s6_addr32[3], dom);
284 } else {
285 seq_printf(m, "%s %pI6 %s\n", im->m_class, &addr, dom);
287 return 0;
291 struct cache_detail ip_map_cache = {
292 .owner = THIS_MODULE,
293 .hash_size = IP_HASHMAX,
294 .hash_table = ip_table,
295 .name = "auth.unix.ip",
296 .cache_put = ip_map_put,
297 .cache_upcall = ip_map_upcall,
298 .cache_parse = ip_map_parse,
299 .cache_show = ip_map_show,
300 .match = ip_map_match,
301 .init = ip_map_init,
302 .update = update,
303 .alloc = ip_map_alloc,
306 static struct ip_map *ip_map_lookup(char *class, struct in6_addr *addr)
308 struct ip_map ip;
309 struct cache_head *ch;
311 strcpy(ip.m_class, class);
312 ipv6_addr_copy(&ip.m_addr, addr);
313 ch = sunrpc_cache_lookup(&ip_map_cache, &ip.h,
314 hash_str(class, IP_HASHBITS) ^
315 hash_ip6(*addr));
317 if (ch)
318 return container_of(ch, struct ip_map, h);
319 else
320 return NULL;
323 static int ip_map_update(struct ip_map *ipm, struct unix_domain *udom, time_t expiry)
325 struct ip_map ip;
326 struct cache_head *ch;
328 ip.m_client = udom;
329 ip.h.flags = 0;
330 if (!udom)
331 set_bit(CACHE_NEGATIVE, &ip.h.flags);
332 else {
333 ip.m_add_change = udom->addr_changes;
334 /* if this is from the legacy set_client system call,
335 * we need m_add_change to be one higher
337 if (expiry == NEVER)
338 ip.m_add_change++;
340 ip.h.expiry_time = expiry;
341 ch = sunrpc_cache_update(&ip_map_cache,
342 &ip.h, &ipm->h,
343 hash_str(ipm->m_class, IP_HASHBITS) ^
344 hash_ip6(ipm->m_addr));
345 if (!ch)
346 return -ENOMEM;
347 cache_put(ch, &ip_map_cache);
348 return 0;
351 int auth_unix_add_addr(struct in6_addr *addr, struct auth_domain *dom)
353 struct unix_domain *udom;
354 struct ip_map *ipmp;
356 if (dom->flavour != &svcauth_unix)
357 return -EINVAL;
358 udom = container_of(dom, struct unix_domain, h);
359 ipmp = ip_map_lookup("nfsd", addr);
361 if (ipmp)
362 return ip_map_update(ipmp, udom, NEVER);
363 else
364 return -ENOMEM;
366 EXPORT_SYMBOL_GPL(auth_unix_add_addr);
368 int auth_unix_forget_old(struct auth_domain *dom)
370 struct unix_domain *udom;
372 if (dom->flavour != &svcauth_unix)
373 return -EINVAL;
374 udom = container_of(dom, struct unix_domain, h);
375 udom->addr_changes++;
376 return 0;
378 EXPORT_SYMBOL_GPL(auth_unix_forget_old);
380 struct auth_domain *auth_unix_lookup(struct in6_addr *addr)
382 struct ip_map *ipm;
383 struct auth_domain *rv;
385 ipm = ip_map_lookup("nfsd", addr);
387 if (!ipm)
388 return NULL;
389 if (cache_check(&ip_map_cache, &ipm->h, NULL))
390 return NULL;
392 if ((ipm->m_client->addr_changes - ipm->m_add_change) >0) {
393 if (test_and_set_bit(CACHE_NEGATIVE, &ipm->h.flags) == 0)
394 auth_domain_put(&ipm->m_client->h);
395 rv = NULL;
396 } else {
397 rv = &ipm->m_client->h;
398 kref_get(&rv->ref);
400 cache_put(&ipm->h, &ip_map_cache);
401 return rv;
403 EXPORT_SYMBOL_GPL(auth_unix_lookup);
405 void svcauth_unix_purge(void)
407 cache_purge(&ip_map_cache);
409 EXPORT_SYMBOL_GPL(svcauth_unix_purge);
411 static inline struct ip_map *
412 ip_map_cached_get(struct svc_rqst *rqstp)
414 struct ip_map *ipm = NULL;
415 struct svc_xprt *xprt = rqstp->rq_xprt;
417 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) {
418 spin_lock(&xprt->xpt_lock);
419 ipm = xprt->xpt_auth_cache;
420 if (ipm != NULL) {
421 if (!cache_valid(&ipm->h)) {
423 * The entry has been invalidated since it was
424 * remembered, e.g. by a second mount from the
425 * same IP address.
427 xprt->xpt_auth_cache = NULL;
428 spin_unlock(&xprt->xpt_lock);
429 cache_put(&ipm->h, &ip_map_cache);
430 return NULL;
432 cache_get(&ipm->h);
434 spin_unlock(&xprt->xpt_lock);
436 return ipm;
439 static inline void
440 ip_map_cached_put(struct svc_rqst *rqstp, struct ip_map *ipm)
442 struct svc_xprt *xprt = rqstp->rq_xprt;
444 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) {
445 spin_lock(&xprt->xpt_lock);
446 if (xprt->xpt_auth_cache == NULL) {
447 /* newly cached, keep the reference */
448 xprt->xpt_auth_cache = ipm;
449 ipm = NULL;
451 spin_unlock(&xprt->xpt_lock);
453 if (ipm)
454 cache_put(&ipm->h, &ip_map_cache);
457 void
458 svcauth_unix_info_release(void *info)
460 struct ip_map *ipm = info;
461 cache_put(&ipm->h, &ip_map_cache);
464 /****************************************************************************
465 * auth.unix.gid cache
466 * simple cache to map a UID to a list of GIDs
467 * because AUTH_UNIX aka AUTH_SYS has a max of 16
469 #define GID_HASHBITS 8
470 #define GID_HASHMAX (1<<GID_HASHBITS)
471 #define GID_HASHMASK (GID_HASHMAX - 1)
473 struct unix_gid {
474 struct cache_head h;
475 uid_t uid;
476 struct group_info *gi;
478 static struct cache_head *gid_table[GID_HASHMAX];
480 static void unix_gid_put(struct kref *kref)
482 struct cache_head *item = container_of(kref, struct cache_head, ref);
483 struct unix_gid *ug = container_of(item, struct unix_gid, h);
484 if (test_bit(CACHE_VALID, &item->flags) &&
485 !test_bit(CACHE_NEGATIVE, &item->flags))
486 put_group_info(ug->gi);
487 kfree(ug);
490 static int unix_gid_match(struct cache_head *corig, struct cache_head *cnew)
492 struct unix_gid *orig = container_of(corig, struct unix_gid, h);
493 struct unix_gid *new = container_of(cnew, struct unix_gid, h);
494 return orig->uid == new->uid;
496 static void unix_gid_init(struct cache_head *cnew, struct cache_head *citem)
498 struct unix_gid *new = container_of(cnew, struct unix_gid, h);
499 struct unix_gid *item = container_of(citem, struct unix_gid, h);
500 new->uid = item->uid;
502 static void unix_gid_update(struct cache_head *cnew, struct cache_head *citem)
504 struct unix_gid *new = container_of(cnew, struct unix_gid, h);
505 struct unix_gid *item = container_of(citem, struct unix_gid, h);
507 get_group_info(item->gi);
508 new->gi = item->gi;
510 static struct cache_head *unix_gid_alloc(void)
512 struct unix_gid *g = kmalloc(sizeof(*g), GFP_KERNEL);
513 if (g)
514 return &g->h;
515 else
516 return NULL;
519 static void unix_gid_request(struct cache_detail *cd,
520 struct cache_head *h,
521 char **bpp, int *blen)
523 char tuid[20];
524 struct unix_gid *ug = container_of(h, struct unix_gid, h);
526 snprintf(tuid, 20, "%u", ug->uid);
527 qword_add(bpp, blen, tuid);
528 (*bpp)[-1] = '\n';
531 static int unix_gid_upcall(struct cache_detail *cd, struct cache_head *h)
533 return sunrpc_cache_pipe_upcall(cd, h, unix_gid_request);
536 static struct unix_gid *unix_gid_lookup(uid_t uid);
537 extern struct cache_detail unix_gid_cache;
539 static int unix_gid_parse(struct cache_detail *cd,
540 char *mesg, int mlen)
542 /* uid expiry Ngid gid0 gid1 ... gidN-1 */
543 int uid;
544 int gids;
545 int rv;
546 int i;
547 int err;
548 time_t expiry;
549 struct unix_gid ug, *ugp;
551 if (mlen <= 0 || mesg[mlen-1] != '\n')
552 return -EINVAL;
553 mesg[mlen-1] = 0;
555 rv = get_int(&mesg, &uid);
556 if (rv)
557 return -EINVAL;
558 ug.uid = uid;
560 expiry = get_expiry(&mesg);
561 if (expiry == 0)
562 return -EINVAL;
564 rv = get_int(&mesg, &gids);
565 if (rv || gids < 0 || gids > 8192)
566 return -EINVAL;
568 ug.gi = groups_alloc(gids);
569 if (!ug.gi)
570 return -ENOMEM;
572 for (i = 0 ; i < gids ; i++) {
573 int gid;
574 rv = get_int(&mesg, &gid);
575 err = -EINVAL;
576 if (rv)
577 goto out;
578 GROUP_AT(ug.gi, i) = gid;
581 ugp = unix_gid_lookup(uid);
582 if (ugp) {
583 struct cache_head *ch;
584 ug.h.flags = 0;
585 ug.h.expiry_time = expiry;
586 ch = sunrpc_cache_update(&unix_gid_cache,
587 &ug.h, &ugp->h,
588 hash_long(uid, GID_HASHBITS));
589 if (!ch)
590 err = -ENOMEM;
591 else {
592 err = 0;
593 cache_put(ch, &unix_gid_cache);
595 } else
596 err = -ENOMEM;
597 out:
598 if (ug.gi)
599 put_group_info(ug.gi);
600 return err;
603 static int unix_gid_show(struct seq_file *m,
604 struct cache_detail *cd,
605 struct cache_head *h)
607 struct unix_gid *ug;
608 int i;
609 int glen;
611 if (h == NULL) {
612 seq_puts(m, "#uid cnt: gids...\n");
613 return 0;
615 ug = container_of(h, struct unix_gid, h);
616 if (test_bit(CACHE_VALID, &h->flags) &&
617 !test_bit(CACHE_NEGATIVE, &h->flags))
618 glen = ug->gi->ngroups;
619 else
620 glen = 0;
622 seq_printf(m, "%d %d:", ug->uid, glen);
623 for (i = 0; i < glen; i++)
624 seq_printf(m, " %d", GROUP_AT(ug->gi, i));
625 seq_printf(m, "\n");
626 return 0;
629 struct cache_detail unix_gid_cache = {
630 .owner = THIS_MODULE,
631 .hash_size = GID_HASHMAX,
632 .hash_table = gid_table,
633 .name = "auth.unix.gid",
634 .cache_put = unix_gid_put,
635 .cache_upcall = unix_gid_upcall,
636 .cache_parse = unix_gid_parse,
637 .cache_show = unix_gid_show,
638 .match = unix_gid_match,
639 .init = unix_gid_init,
640 .update = unix_gid_update,
641 .alloc = unix_gid_alloc,
644 static struct unix_gid *unix_gid_lookup(uid_t uid)
646 struct unix_gid ug;
647 struct cache_head *ch;
649 ug.uid = uid;
650 ch = sunrpc_cache_lookup(&unix_gid_cache, &ug.h,
651 hash_long(uid, GID_HASHBITS));
652 if (ch)
653 return container_of(ch, struct unix_gid, h);
654 else
655 return NULL;
658 static struct group_info *unix_gid_find(uid_t uid, struct svc_rqst *rqstp)
660 struct unix_gid *ug;
661 struct group_info *gi;
662 int ret;
664 ug = unix_gid_lookup(uid);
665 if (!ug)
666 return ERR_PTR(-EAGAIN);
667 ret = cache_check(&unix_gid_cache, &ug->h, &rqstp->rq_chandle);
668 switch (ret) {
669 case -ENOENT:
670 return ERR_PTR(-ENOENT);
671 case 0:
672 gi = get_group_info(ug->gi);
673 cache_put(&ug->h, &unix_gid_cache);
674 return gi;
675 default:
676 return ERR_PTR(-EAGAIN);
681 svcauth_unix_set_client(struct svc_rqst *rqstp)
683 struct sockaddr_in *sin;
684 struct sockaddr_in6 *sin6, sin6_storage;
685 struct ip_map *ipm;
686 struct group_info *gi;
687 struct svc_cred *cred = &rqstp->rq_cred;
689 switch (rqstp->rq_addr.ss_family) {
690 case AF_INET:
691 sin = svc_addr_in(rqstp);
692 sin6 = &sin6_storage;
693 ipv6_addr_set(&sin6->sin6_addr, 0, 0,
694 htonl(0x0000FFFF), sin->sin_addr.s_addr);
695 break;
696 case AF_INET6:
697 sin6 = svc_addr_in6(rqstp);
698 break;
699 default:
700 BUG();
703 rqstp->rq_client = NULL;
704 if (rqstp->rq_proc == 0)
705 return SVC_OK;
707 ipm = ip_map_cached_get(rqstp);
708 if (ipm == NULL)
709 ipm = ip_map_lookup(rqstp->rq_server->sv_program->pg_class,
710 &sin6->sin6_addr);
712 if (ipm == NULL)
713 return SVC_DENIED;
715 switch (cache_check(&ip_map_cache, &ipm->h, &rqstp->rq_chandle)) {
716 default:
717 BUG();
718 case -EAGAIN:
719 case -ETIMEDOUT:
720 return SVC_DROP;
721 case -ENOENT:
722 return SVC_DENIED;
723 case 0:
724 rqstp->rq_client = &ipm->m_client->h;
725 kref_get(&rqstp->rq_client->ref);
726 ip_map_cached_put(rqstp, ipm);
727 break;
730 gi = unix_gid_find(cred->cr_uid, rqstp);
731 switch (PTR_ERR(gi)) {
732 case -EAGAIN:
733 return SVC_DROP;
734 case -ENOENT:
735 break;
736 default:
737 put_group_info(cred->cr_group_info);
738 cred->cr_group_info = gi;
740 return SVC_OK;
743 EXPORT_SYMBOL_GPL(svcauth_unix_set_client);
745 static int
746 svcauth_null_accept(struct svc_rqst *rqstp, __be32 *authp)
748 struct kvec *argv = &rqstp->rq_arg.head[0];
749 struct kvec *resv = &rqstp->rq_res.head[0];
750 struct svc_cred *cred = &rqstp->rq_cred;
752 cred->cr_group_info = NULL;
753 rqstp->rq_client = NULL;
755 if (argv->iov_len < 3*4)
756 return SVC_GARBAGE;
758 if (svc_getu32(argv) != 0) {
759 dprintk("svc: bad null cred\n");
760 *authp = rpc_autherr_badcred;
761 return SVC_DENIED;
763 if (svc_getu32(argv) != htonl(RPC_AUTH_NULL) || svc_getu32(argv) != 0) {
764 dprintk("svc: bad null verf\n");
765 *authp = rpc_autherr_badverf;
766 return SVC_DENIED;
769 /* Signal that mapping to nobody uid/gid is required */
770 cred->cr_uid = (uid_t) -1;
771 cred->cr_gid = (gid_t) -1;
772 cred->cr_group_info = groups_alloc(0);
773 if (cred->cr_group_info == NULL)
774 return SVC_DROP; /* kmalloc failure - client must retry */
776 /* Put NULL verifier */
777 svc_putnl(resv, RPC_AUTH_NULL);
778 svc_putnl(resv, 0);
780 rqstp->rq_flavor = RPC_AUTH_NULL;
781 return SVC_OK;
784 static int
785 svcauth_null_release(struct svc_rqst *rqstp)
787 if (rqstp->rq_client)
788 auth_domain_put(rqstp->rq_client);
789 rqstp->rq_client = NULL;
790 if (rqstp->rq_cred.cr_group_info)
791 put_group_info(rqstp->rq_cred.cr_group_info);
792 rqstp->rq_cred.cr_group_info = NULL;
794 return 0; /* don't drop */
798 struct auth_ops svcauth_null = {
799 .name = "null",
800 .owner = THIS_MODULE,
801 .flavour = RPC_AUTH_NULL,
802 .accept = svcauth_null_accept,
803 .release = svcauth_null_release,
804 .set_client = svcauth_unix_set_client,
808 static int
809 svcauth_unix_accept(struct svc_rqst *rqstp, __be32 *authp)
811 struct kvec *argv = &rqstp->rq_arg.head[0];
812 struct kvec *resv = &rqstp->rq_res.head[0];
813 struct svc_cred *cred = &rqstp->rq_cred;
814 u32 slen, i;
815 int len = argv->iov_len;
817 cred->cr_group_info = NULL;
818 rqstp->rq_client = NULL;
820 if ((len -= 3*4) < 0)
821 return SVC_GARBAGE;
823 svc_getu32(argv); /* length */
824 svc_getu32(argv); /* time stamp */
825 slen = XDR_QUADLEN(svc_getnl(argv)); /* machname length */
826 if (slen > 64 || (len -= (slen + 3)*4) < 0)
827 goto badcred;
828 argv->iov_base = (void*)((__be32*)argv->iov_base + slen); /* skip machname */
829 argv->iov_len -= slen*4;
831 cred->cr_uid = svc_getnl(argv); /* uid */
832 cred->cr_gid = svc_getnl(argv); /* gid */
833 slen = svc_getnl(argv); /* gids length */
834 if (slen > 16 || (len -= (slen + 2)*4) < 0)
835 goto badcred;
836 cred->cr_group_info = groups_alloc(slen);
837 if (cred->cr_group_info == NULL)
838 return SVC_DROP;
839 for (i = 0; i < slen; i++)
840 GROUP_AT(cred->cr_group_info, i) = svc_getnl(argv);
841 if (svc_getu32(argv) != htonl(RPC_AUTH_NULL) || svc_getu32(argv) != 0) {
842 *authp = rpc_autherr_badverf;
843 return SVC_DENIED;
846 /* Put NULL verifier */
847 svc_putnl(resv, RPC_AUTH_NULL);
848 svc_putnl(resv, 0);
850 rqstp->rq_flavor = RPC_AUTH_UNIX;
851 return SVC_OK;
853 badcred:
854 *authp = rpc_autherr_badcred;
855 return SVC_DENIED;
858 static int
859 svcauth_unix_release(struct svc_rqst *rqstp)
861 /* Verifier (such as it is) is already in place.
863 if (rqstp->rq_client)
864 auth_domain_put(rqstp->rq_client);
865 rqstp->rq_client = NULL;
866 if (rqstp->rq_cred.cr_group_info)
867 put_group_info(rqstp->rq_cred.cr_group_info);
868 rqstp->rq_cred.cr_group_info = NULL;
870 return 0;
874 struct auth_ops svcauth_unix = {
875 .name = "unix",
876 .owner = THIS_MODULE,
877 .flavour = RPC_AUTH_UNIX,
878 .accept = svcauth_unix_accept,
879 .release = svcauth_unix_release,
880 .domain_release = svcauth_unix_domain_release,
881 .set_client = svcauth_unix_set_client,