1 // SPDX-License-Identifier: GPL-2.0-only
2 #include <linux/types.h>
3 #include <linux/sched.h>
4 #include <linux/module.h>
5 #include <linux/sunrpc/types.h>
6 #include <linux/sunrpc/xdr.h>
7 #include <linux/sunrpc/svcsock.h>
8 #include <linux/sunrpc/svcauth.h>
9 #include <linux/sunrpc/gss_api.h>
10 #include <linux/sunrpc/addr.h>
11 #include <linux/err.h>
12 #include <linux/seq_file.h>
13 #include <linux/hash.h>
14 #include <linux/string.h>
15 #include <linux/slab.h>
18 #include <linux/kernel.h>
19 #include <linux/user_namespace.h>
20 #define RPCDBG_FACILITY RPCDBG_AUTH
26 * AUTHUNIX and AUTHNULL credentials are both handled here.
27 * AUTHNULL is treated just like AUTHUNIX except that the uid/gid
28 * are always nobody (-2). i.e. we do the same IP address checks for
29 * AUTHNULL as for AUTHUNIX, and that is done here.
35 /* other stuff later */
38 extern struct auth_ops svcauth_null;
39 extern struct auth_ops svcauth_unix;
40 extern struct auth_ops svcauth_tls;
42 static void svcauth_unix_domain_release_rcu(struct rcu_head *head)
44 struct auth_domain *dom = container_of(head, struct auth_domain, rcu_head);
45 struct unix_domain *ud = container_of(dom, struct unix_domain, h);
51 static void svcauth_unix_domain_release(struct auth_domain *dom)
53 call_rcu(&dom->rcu_head, svcauth_unix_domain_release_rcu);
56 struct auth_domain *unix_domain_find(char *name)
58 struct auth_domain *rv;
59 struct unix_domain *new = NULL;
61 rv = auth_domain_find(name);
64 if (new && rv != &new->h)
65 svcauth_unix_domain_release(&new->h);
67 if (rv->flavour != &svcauth_unix) {
74 new = kmalloc(sizeof(*new), GFP_KERNEL);
77 kref_init(&new->h.ref);
78 new->h.name = kstrdup(name, GFP_KERNEL);
79 if (new->h.name == NULL) {
83 new->h.flavour = &svcauth_unix;
84 rv = auth_domain_lookup(name, &new->h);
87 EXPORT_SYMBOL_GPL(unix_domain_find);
90 /**************************************************
91 * cache for IP address to unix_domain
92 * as needed by AUTH_UNIX
95 #define IP_HASHMAX (1<<IP_HASHBITS)
99 char m_class[8]; /* e.g. "nfsd" */
100 struct in6_addr m_addr;
101 struct unix_domain *m_client;
102 struct rcu_head m_rcu;
105 static void ip_map_put(struct kref *kref)
107 struct cache_head *item = container_of(kref, struct cache_head, ref);
108 struct ip_map *im = container_of(item, struct ip_map,h);
110 if (test_bit(CACHE_VALID, &item->flags) &&
111 !test_bit(CACHE_NEGATIVE, &item->flags))
112 auth_domain_put(&im->m_client->h);
113 kfree_rcu(im, m_rcu);
116 static inline int hash_ip6(const struct in6_addr *ip)
118 return hash_32(ipv6_addr_hash(ip), IP_HASHBITS);
120 static int ip_map_match(struct cache_head *corig, struct cache_head *cnew)
122 struct ip_map *orig = container_of(corig, struct ip_map, h);
123 struct ip_map *new = container_of(cnew, struct ip_map, h);
124 return strcmp(orig->m_class, new->m_class) == 0 &&
125 ipv6_addr_equal(&orig->m_addr, &new->m_addr);
127 static void ip_map_init(struct cache_head *cnew, struct cache_head *citem)
129 struct ip_map *new = container_of(cnew, struct ip_map, h);
130 struct ip_map *item = container_of(citem, struct ip_map, h);
132 strcpy(new->m_class, item->m_class);
133 new->m_addr = item->m_addr;
135 static void update(struct cache_head *cnew, struct cache_head *citem)
137 struct ip_map *new = container_of(cnew, struct ip_map, h);
138 struct ip_map *item = container_of(citem, struct ip_map, h);
140 kref_get(&item->m_client->h.ref);
141 new->m_client = item->m_client;
143 static struct cache_head *ip_map_alloc(void)
145 struct ip_map *i = kmalloc(sizeof(*i), GFP_KERNEL);
152 static int ip_map_upcall(struct cache_detail *cd, struct cache_head *h)
154 return sunrpc_cache_pipe_upcall(cd, h);
157 static void ip_map_request(struct cache_detail *cd,
158 struct cache_head *h,
159 char **bpp, int *blen)
162 struct ip_map *im = container_of(h, struct ip_map, h);
164 if (ipv6_addr_v4mapped(&(im->m_addr))) {
165 snprintf(text_addr, 20, "%pI4", &im->m_addr.s6_addr32[3]);
167 snprintf(text_addr, 40, "%pI6", &im->m_addr);
169 qword_add(bpp, blen, im->m_class);
170 qword_add(bpp, blen, text_addr);
174 static struct ip_map *__ip_map_lookup(struct cache_detail *cd, char *class, struct in6_addr *addr);
175 static int __ip_map_update(struct cache_detail *cd, struct ip_map *ipm, struct unix_domain *udom, time64_t expiry);
177 static int ip_map_parse(struct cache_detail *cd,
178 char *mesg, int mlen)
180 /* class ipaddress [domainname] */
181 /* should be safe just to use the start of the input buffer
188 struct sockaddr_in s4;
189 struct sockaddr_in6 s6;
191 struct sockaddr_in6 sin6;
195 struct auth_domain *dom;
198 if (mesg[mlen-1] != '\n')
203 len = qword_get(&mesg, class, sizeof(class));
204 if (len <= 0) return -EINVAL;
207 len = qword_get(&mesg, buf, mlen);
208 if (len <= 0) return -EINVAL;
210 if (rpc_pton(cd->net, buf, len, &address.sa, sizeof(address)) == 0)
212 switch (address.sa.sa_family) {
214 /* Form a mapped IPv4 address in sin6 */
215 sin6.sin6_family = AF_INET6;
216 ipv6_addr_set_v4mapped(address.s4.sin_addr.s_addr,
219 #if IS_ENABLED(CONFIG_IPV6)
221 memcpy(&sin6, &address.s6, sizeof(sin6));
228 expiry = get_expiry(&mesg);
232 /* domainname, or empty for NEGATIVE */
233 len = qword_get(&mesg, buf, mlen);
234 if (len < 0) return -EINVAL;
237 dom = unix_domain_find(buf);
243 /* IPv6 scope IDs are ignored for now */
244 ipmp = __ip_map_lookup(cd, class, &sin6.sin6_addr);
246 err = __ip_map_update(cd, ipmp,
247 container_of(dom, struct unix_domain, h),
253 auth_domain_put(dom);
259 static int ip_map_show(struct seq_file *m,
260 struct cache_detail *cd,
261 struct cache_head *h)
264 struct in6_addr addr;
265 char *dom = "-no-domain-";
268 seq_puts(m, "#class IP domain\n");
271 im = container_of(h, struct ip_map, h);
272 /* class addr domain */
275 if (test_bit(CACHE_VALID, &h->flags) &&
276 !test_bit(CACHE_NEGATIVE, &h->flags))
277 dom = im->m_client->h.name;
279 if (ipv6_addr_v4mapped(&addr)) {
280 seq_printf(m, "%s %pI4 %s\n",
281 im->m_class, &addr.s6_addr32[3], dom);
283 seq_printf(m, "%s %pI6 %s\n", im->m_class, &addr, dom);
289 static struct ip_map *__ip_map_lookup(struct cache_detail *cd, char *class,
290 struct in6_addr *addr)
293 struct cache_head *ch;
295 strcpy(ip.m_class, class);
297 ch = sunrpc_cache_lookup_rcu(cd, &ip.h,
298 hash_str(class, IP_HASHBITS) ^
302 return container_of(ch, struct ip_map, h);
307 static int __ip_map_update(struct cache_detail *cd, struct ip_map *ipm,
308 struct unix_domain *udom, time64_t expiry)
311 struct cache_head *ch;
316 set_bit(CACHE_NEGATIVE, &ip.h.flags);
317 ip.h.expiry_time = expiry;
318 ch = sunrpc_cache_update(cd, &ip.h, &ipm->h,
319 hash_str(ipm->m_class, IP_HASHBITS) ^
320 hash_ip6(&ipm->m_addr));
327 void svcauth_unix_purge(struct net *net)
329 struct sunrpc_net *sn;
331 sn = net_generic(net, sunrpc_net_id);
332 cache_purge(sn->ip_map_cache);
334 EXPORT_SYMBOL_GPL(svcauth_unix_purge);
336 static inline struct ip_map *
337 ip_map_cached_get(struct svc_xprt *xprt)
339 struct ip_map *ipm = NULL;
340 struct sunrpc_net *sn;
342 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) {
343 spin_lock(&xprt->xpt_lock);
344 ipm = xprt->xpt_auth_cache;
346 sn = net_generic(xprt->xpt_net, sunrpc_net_id);
347 if (cache_is_expired(sn->ip_map_cache, &ipm->h)) {
349 * The entry has been invalidated since it was
350 * remembered, e.g. by a second mount from the
353 xprt->xpt_auth_cache = NULL;
354 spin_unlock(&xprt->xpt_lock);
355 cache_put(&ipm->h, sn->ip_map_cache);
360 spin_unlock(&xprt->xpt_lock);
366 ip_map_cached_put(struct svc_xprt *xprt, struct ip_map *ipm)
368 if (test_bit(XPT_CACHE_AUTH, &xprt->xpt_flags)) {
369 spin_lock(&xprt->xpt_lock);
370 if (xprt->xpt_auth_cache == NULL) {
371 /* newly cached, keep the reference */
372 xprt->xpt_auth_cache = ipm;
375 spin_unlock(&xprt->xpt_lock);
378 struct sunrpc_net *sn;
380 sn = net_generic(xprt->xpt_net, sunrpc_net_id);
381 cache_put(&ipm->h, sn->ip_map_cache);
386 svcauth_unix_info_release(struct svc_xprt *xpt)
390 ipm = xpt->xpt_auth_cache;
392 struct sunrpc_net *sn;
394 sn = net_generic(xpt->xpt_net, sunrpc_net_id);
395 cache_put(&ipm->h, sn->ip_map_cache);
399 /****************************************************************************
400 * auth.unix.gid cache
401 * simple cache to map a UID to a list of GIDs
402 * because AUTH_UNIX aka AUTH_SYS has a max of UNX_NGROUPS
404 #define GID_HASHBITS 8
405 #define GID_HASHMAX (1<<GID_HASHBITS)
410 struct group_info *gi;
414 static int unix_gid_hash(kuid_t uid)
416 return hash_long(from_kuid(&init_user_ns, uid), GID_HASHBITS);
419 static void unix_gid_put(struct kref *kref)
421 struct cache_head *item = container_of(kref, struct cache_head, ref);
422 struct unix_gid *ug = container_of(item, struct unix_gid, h);
423 if (test_bit(CACHE_VALID, &item->flags) &&
424 !test_bit(CACHE_NEGATIVE, &item->flags))
425 put_group_info(ug->gi);
429 static int unix_gid_match(struct cache_head *corig, struct cache_head *cnew)
431 struct unix_gid *orig = container_of(corig, struct unix_gid, h);
432 struct unix_gid *new = container_of(cnew, struct unix_gid, h);
433 return uid_eq(orig->uid, new->uid);
435 static void unix_gid_init(struct cache_head *cnew, struct cache_head *citem)
437 struct unix_gid *new = container_of(cnew, struct unix_gid, h);
438 struct unix_gid *item = container_of(citem, struct unix_gid, h);
439 new->uid = item->uid;
441 static void unix_gid_update(struct cache_head *cnew, struct cache_head *citem)
443 struct unix_gid *new = container_of(cnew, struct unix_gid, h);
444 struct unix_gid *item = container_of(citem, struct unix_gid, h);
446 get_group_info(item->gi);
449 static struct cache_head *unix_gid_alloc(void)
451 struct unix_gid *g = kmalloc(sizeof(*g), GFP_KERNEL);
458 static int unix_gid_upcall(struct cache_detail *cd, struct cache_head *h)
460 return sunrpc_cache_pipe_upcall_timeout(cd, h);
463 static void unix_gid_request(struct cache_detail *cd,
464 struct cache_head *h,
465 char **bpp, int *blen)
468 struct unix_gid *ug = container_of(h, struct unix_gid, h);
470 snprintf(tuid, 20, "%u", from_kuid(&init_user_ns, ug->uid));
471 qword_add(bpp, blen, tuid);
475 static struct unix_gid *unix_gid_lookup(struct cache_detail *cd, kuid_t uid);
477 static int unix_gid_parse(struct cache_detail *cd,
478 char *mesg, int mlen)
480 /* uid expiry Ngid gid0 gid1 ... gidN-1 */
488 struct unix_gid ug, *ugp;
490 if (mesg[mlen - 1] != '\n')
494 rv = get_int(&mesg, &id);
497 uid = make_kuid(current_user_ns(), id);
500 expiry = get_expiry(&mesg);
504 rv = get_int(&mesg, &gids);
505 if (rv || gids < 0 || gids > 8192)
508 ug.gi = groups_alloc(gids);
512 for (i = 0 ; i < gids ; i++) {
515 rv = get_int(&mesg, &gid);
519 kgid = make_kgid(current_user_ns(), gid);
520 if (!gid_valid(kgid))
522 ug.gi->gid[i] = kgid;
526 ugp = unix_gid_lookup(cd, uid);
528 struct cache_head *ch;
530 ug.h.expiry_time = expiry;
531 ch = sunrpc_cache_update(cd,
544 put_group_info(ug.gi);
548 static int unix_gid_show(struct seq_file *m,
549 struct cache_detail *cd,
550 struct cache_head *h)
552 struct user_namespace *user_ns = m->file->f_cred->user_ns;
558 seq_puts(m, "#uid cnt: gids...\n");
561 ug = container_of(h, struct unix_gid, h);
562 if (test_bit(CACHE_VALID, &h->flags) &&
563 !test_bit(CACHE_NEGATIVE, &h->flags))
564 glen = ug->gi->ngroups;
568 seq_printf(m, "%u %d:", from_kuid_munged(user_ns, ug->uid), glen);
569 for (i = 0; i < glen; i++)
570 seq_printf(m, " %d", from_kgid_munged(user_ns, ug->gi->gid[i]));
575 static const struct cache_detail unix_gid_cache_template = {
576 .owner = THIS_MODULE,
577 .hash_size = GID_HASHMAX,
578 .name = "auth.unix.gid",
579 .cache_put = unix_gid_put,
580 .cache_upcall = unix_gid_upcall,
581 .cache_request = unix_gid_request,
582 .cache_parse = unix_gid_parse,
583 .cache_show = unix_gid_show,
584 .match = unix_gid_match,
585 .init = unix_gid_init,
586 .update = unix_gid_update,
587 .alloc = unix_gid_alloc,
590 int unix_gid_cache_create(struct net *net)
592 struct sunrpc_net *sn = net_generic(net, sunrpc_net_id);
593 struct cache_detail *cd;
596 cd = cache_create_net(&unix_gid_cache_template, net);
599 err = cache_register_net(cd, net);
601 cache_destroy_net(cd, net);
604 sn->unix_gid_cache = cd;
608 void unix_gid_cache_destroy(struct net *net)
610 struct sunrpc_net *sn = net_generic(net, sunrpc_net_id);
611 struct cache_detail *cd = sn->unix_gid_cache;
613 sn->unix_gid_cache = NULL;
615 cache_unregister_net(cd, net);
616 cache_destroy_net(cd, net);
619 static struct unix_gid *unix_gid_lookup(struct cache_detail *cd, kuid_t uid)
622 struct cache_head *ch;
625 ch = sunrpc_cache_lookup_rcu(cd, &ug.h, unix_gid_hash(uid));
627 return container_of(ch, struct unix_gid, h);
632 static struct group_info *unix_gid_find(kuid_t uid, struct svc_rqst *rqstp)
635 struct group_info *gi;
637 struct sunrpc_net *sn = net_generic(rqstp->rq_xprt->xpt_net,
640 ug = unix_gid_lookup(sn->unix_gid_cache, uid);
642 return ERR_PTR(-EAGAIN);
643 ret = cache_check(sn->unix_gid_cache, &ug->h, &rqstp->rq_chandle);
646 return ERR_PTR(-ENOENT);
648 return ERR_PTR(-ESHUTDOWN);
650 gi = get_group_info(ug->gi);
651 cache_put(&ug->h, sn->unix_gid_cache);
654 return ERR_PTR(-EAGAIN);
659 svcauth_unix_set_client(struct svc_rqst *rqstp)
661 struct sockaddr_in *sin;
662 struct sockaddr_in6 *sin6, sin6_storage;
664 struct group_info *gi;
665 struct svc_cred *cred = &rqstp->rq_cred;
666 struct svc_xprt *xprt = rqstp->rq_xprt;
667 struct net *net = xprt->xpt_net;
668 struct sunrpc_net *sn = net_generic(net, sunrpc_net_id);
670 switch (rqstp->rq_addr.ss_family) {
672 sin = svc_addr_in(rqstp);
673 sin6 = &sin6_storage;
674 ipv6_addr_set_v4mapped(sin->sin_addr.s_addr, &sin6->sin6_addr);
677 sin6 = svc_addr_in6(rqstp);
683 rqstp->rq_client = NULL;
684 if (rqstp->rq_proc == 0)
687 rqstp->rq_auth_stat = rpc_autherr_badcred;
688 ipm = ip_map_cached_get(xprt);
690 ipm = __ip_map_lookup(sn->ip_map_cache, rqstp->rq_server->sv_program->pg_class,
696 switch (cache_check(sn->ip_map_cache, &ipm->h, &rqstp->rq_chandle)) {
706 rqstp->rq_client = &ipm->m_client->h;
707 kref_get(&rqstp->rq_client->ref);
708 ip_map_cached_put(xprt, ipm);
712 gi = unix_gid_find(cred->cr_uid, rqstp);
713 switch (PTR_ERR(gi)) {
721 put_group_info(cred->cr_group_info);
722 cred->cr_group_info = gi;
726 rqstp->rq_auth_stat = rpc_auth_ok;
730 EXPORT_SYMBOL_GPL(svcauth_unix_set_client);
733 svcauth_null_accept(struct svc_rqst *rqstp)
735 struct kvec *argv = &rqstp->rq_arg.head[0];
736 struct kvec *resv = &rqstp->rq_res.head[0];
737 struct svc_cred *cred = &rqstp->rq_cred;
739 if (argv->iov_len < 3*4)
742 if (svc_getu32(argv) != 0) {
743 dprintk("svc: bad null cred\n");
744 rqstp->rq_auth_stat = rpc_autherr_badcred;
747 if (svc_getu32(argv) != htonl(RPC_AUTH_NULL) || svc_getu32(argv) != 0) {
748 dprintk("svc: bad null verf\n");
749 rqstp->rq_auth_stat = rpc_autherr_badverf;
753 /* Signal that mapping to nobody uid/gid is required */
754 cred->cr_uid = INVALID_UID;
755 cred->cr_gid = INVALID_GID;
756 cred->cr_group_info = groups_alloc(0);
757 if (cred->cr_group_info == NULL)
758 return SVC_CLOSE; /* kmalloc failure - client must retry */
760 /* Put NULL verifier */
761 svc_putnl(resv, RPC_AUTH_NULL);
764 rqstp->rq_cred.cr_flavor = RPC_AUTH_NULL;
769 svcauth_null_release(struct svc_rqst *rqstp)
771 if (rqstp->rq_client)
772 auth_domain_put(rqstp->rq_client);
773 rqstp->rq_client = NULL;
774 if (rqstp->rq_cred.cr_group_info)
775 put_group_info(rqstp->rq_cred.cr_group_info);
776 rqstp->rq_cred.cr_group_info = NULL;
778 return 0; /* don't drop */
782 struct auth_ops svcauth_null = {
784 .owner = THIS_MODULE,
785 .flavour = RPC_AUTH_NULL,
786 .accept = svcauth_null_accept,
787 .release = svcauth_null_release,
788 .set_client = svcauth_unix_set_client,
793 svcauth_tls_accept(struct svc_rqst *rqstp)
795 struct svc_cred *cred = &rqstp->rq_cred;
796 struct kvec *argv = rqstp->rq_arg.head;
797 struct kvec *resv = rqstp->rq_res.head;
799 if (argv->iov_len < XDR_UNIT * 3)
802 /* Call's cred length */
803 if (svc_getu32(argv) != xdr_zero) {
804 rqstp->rq_auth_stat = rpc_autherr_badcred;
808 /* Call's verifier flavor and its length */
809 if (svc_getu32(argv) != rpc_auth_null ||
810 svc_getu32(argv) != xdr_zero) {
811 rqstp->rq_auth_stat = rpc_autherr_badverf;
815 /* AUTH_TLS is not valid on non-NULL procedures */
816 if (rqstp->rq_proc != 0) {
817 rqstp->rq_auth_stat = rpc_autherr_badcred;
821 /* Mapping to nobody uid/gid is required */
822 cred->cr_uid = INVALID_UID;
823 cred->cr_gid = INVALID_GID;
824 cred->cr_group_info = groups_alloc(0);
825 if (cred->cr_group_info == NULL)
826 return SVC_CLOSE; /* kmalloc failure - client must retry */
828 /* Reply's verifier */
829 svc_putnl(resv, RPC_AUTH_NULL);
830 if (rqstp->rq_xprt->xpt_ops->xpo_start_tls) {
832 memcpy(resv->iov_base + resv->iov_len, "STARTTLS", 8);
837 rqstp->rq_cred.cr_flavor = RPC_AUTH_TLS;
841 struct auth_ops svcauth_tls = {
843 .owner = THIS_MODULE,
844 .flavour = RPC_AUTH_TLS,
845 .accept = svcauth_tls_accept,
846 .release = svcauth_null_release,
847 .set_client = svcauth_unix_set_client,
852 svcauth_unix_accept(struct svc_rqst *rqstp)
854 struct kvec *argv = &rqstp->rq_arg.head[0];
855 struct kvec *resv = &rqstp->rq_res.head[0];
856 struct svc_cred *cred = &rqstp->rq_cred;
857 struct user_namespace *userns;
859 int len = argv->iov_len;
861 if ((len -= 3*4) < 0)
864 svc_getu32(argv); /* length */
865 svc_getu32(argv); /* time stamp */
866 slen = XDR_QUADLEN(svc_getnl(argv)); /* machname length */
867 if (slen > 64 || (len -= (slen + 3)*4) < 0)
869 argv->iov_base = (void*)((__be32*)argv->iov_base + slen); /* skip machname */
870 argv->iov_len -= slen*4;
872 * Note: we skip uid_valid()/gid_valid() checks here for
873 * backwards compatibility with clients that use -1 id's.
874 * Instead, -1 uid or gid is later mapped to the
875 * (export-specific) anonymous id by nfsd_setuser.
876 * Supplementary gid's will be left alone.
878 userns = (rqstp->rq_xprt && rqstp->rq_xprt->xpt_cred) ?
879 rqstp->rq_xprt->xpt_cred->user_ns : &init_user_ns;
880 cred->cr_uid = make_kuid(userns, svc_getnl(argv)); /* uid */
881 cred->cr_gid = make_kgid(userns, svc_getnl(argv)); /* gid */
882 slen = svc_getnl(argv); /* gids length */
883 if (slen > UNX_NGROUPS || (len -= (slen + 2)*4) < 0)
885 cred->cr_group_info = groups_alloc(slen);
886 if (cred->cr_group_info == NULL)
888 for (i = 0; i < slen; i++) {
889 kgid_t kgid = make_kgid(userns, svc_getnl(argv));
890 cred->cr_group_info->gid[i] = kgid;
892 groups_sort(cred->cr_group_info);
893 if (svc_getu32(argv) != htonl(RPC_AUTH_NULL) || svc_getu32(argv) != 0) {
894 rqstp->rq_auth_stat = rpc_autherr_badverf;
898 /* Put NULL verifier */
899 svc_putnl(resv, RPC_AUTH_NULL);
902 rqstp->rq_cred.cr_flavor = RPC_AUTH_UNIX;
906 rqstp->rq_auth_stat = rpc_autherr_badcred;
911 svcauth_unix_release(struct svc_rqst *rqstp)
913 /* Verifier (such as it is) is already in place.
915 if (rqstp->rq_client)
916 auth_domain_put(rqstp->rq_client);
917 rqstp->rq_client = NULL;
918 if (rqstp->rq_cred.cr_group_info)
919 put_group_info(rqstp->rq_cred.cr_group_info);
920 rqstp->rq_cred.cr_group_info = NULL;
926 struct auth_ops svcauth_unix = {
928 .owner = THIS_MODULE,
929 .flavour = RPC_AUTH_UNIX,
930 .accept = svcauth_unix_accept,
931 .release = svcauth_unix_release,
932 .domain_release = svcauth_unix_domain_release,
933 .set_client = svcauth_unix_set_client,
936 static const struct cache_detail ip_map_cache_template = {
937 .owner = THIS_MODULE,
938 .hash_size = IP_HASHMAX,
939 .name = "auth.unix.ip",
940 .cache_put = ip_map_put,
941 .cache_upcall = ip_map_upcall,
942 .cache_request = ip_map_request,
943 .cache_parse = ip_map_parse,
944 .cache_show = ip_map_show,
945 .match = ip_map_match,
948 .alloc = ip_map_alloc,
951 int ip_map_cache_create(struct net *net)
953 struct sunrpc_net *sn = net_generic(net, sunrpc_net_id);
954 struct cache_detail *cd;
957 cd = cache_create_net(&ip_map_cache_template, net);
960 err = cache_register_net(cd, net);
962 cache_destroy_net(cd, net);
965 sn->ip_map_cache = cd;
969 void ip_map_cache_destroy(struct net *net)
971 struct sunrpc_net *sn = net_generic(net, sunrpc_net_id);
972 struct cache_detail *cd = sn->ip_map_cache;
974 sn->ip_map_cache = NULL;
976 cache_unregister_net(cd, net);
977 cache_destroy_net(cd, net);