mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-26 19:36:41 +00:00
a217813f90
Update the write handler for the portlist file to allow creating new listening endpoints on a transport. The general form of the string is: <transport_name><space><port number> For example: echo "tcp 2049" > /proc/fs/nfsd/portlist This is intended to support the creation of a listening endpoint for RDMA transports without adding #ifdef code to the nfssvc.c file. Transports can also be removed as follows: '-'<transport_name><space><port number> For example: echo "-tcp 2049" > /proc/fs/nfsd/portlist Attempting to add a listener with an invalid transport string results in EPROTONOSUPPORT and a perror string of "Protocol not supported". Attempting to remove an non-existent listener (.e.g. bad proto or port) results in ENOTCONN and a perror string of "Transport endpoint is not connected" Signed-off-by: Tom Tucker <tom@opengridcomputing.com> Acked-by: Neil Brown <neilb@suse.de> Reviewed-by: Chuck Lever <chuck.lever@oracle.com> Reviewed-by: Greg Banks <gnb@sgi.com> Signed-off-by: J. Bruce Fields <bfields@citi.umich.edu>
576 lines
14 KiB
C
576 lines
14 KiB
C
/*
|
|
* linux/fs/lockd/svc.c
|
|
*
|
|
* This is the central lockd service.
|
|
*
|
|
* FIXME: Separate the lockd NFS server functionality from the lockd NFS
|
|
* client functionality. Oh why didn't Sun create two separate
|
|
* services in the first place?
|
|
*
|
|
* Authors: Olaf Kirch (okir@monad.swb.de)
|
|
*
|
|
* Copyright (C) 1995, 1996 Olaf Kirch <okir@monad.swb.de>
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/sysctl.h>
|
|
#include <linux/moduleparam.h>
|
|
|
|
#include <linux/sched.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/in.h>
|
|
#include <linux/uio.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/smp_lock.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/freezer.h>
|
|
|
|
#include <linux/sunrpc/types.h>
|
|
#include <linux/sunrpc/stats.h>
|
|
#include <linux/sunrpc/clnt.h>
|
|
#include <linux/sunrpc/svc.h>
|
|
#include <linux/sunrpc/svcsock.h>
|
|
#include <net/ip.h>
|
|
#include <linux/lockd/lockd.h>
|
|
#include <linux/lockd/sm_inter.h>
|
|
#include <linux/nfs.h>
|
|
|
|
#define NLMDBG_FACILITY NLMDBG_SVC
|
|
#define LOCKD_BUFSIZE (1024 + NLMSVC_XDRSIZE)
|
|
#define ALLOWED_SIGS (sigmask(SIGKILL))
|
|
|
|
static struct svc_program nlmsvc_program;
|
|
|
|
struct nlmsvc_binding * nlmsvc_ops;
|
|
EXPORT_SYMBOL(nlmsvc_ops);
|
|
|
|
static DEFINE_MUTEX(nlmsvc_mutex);
|
|
static unsigned int nlmsvc_users;
|
|
static pid_t nlmsvc_pid;
|
|
static struct svc_serv *nlmsvc_serv;
|
|
int nlmsvc_grace_period;
|
|
unsigned long nlmsvc_timeout;
|
|
|
|
static DECLARE_COMPLETION(lockd_start_done);
|
|
static DECLARE_WAIT_QUEUE_HEAD(lockd_exit);
|
|
|
|
/*
|
|
* These can be set at insmod time (useful for NFS as root filesystem),
|
|
* and also changed through the sysctl interface. -- Jamie Lokier, Aug 2003
|
|
*/
|
|
static unsigned long nlm_grace_period;
|
|
static unsigned long nlm_timeout = LOCKD_DFLT_TIMEO;
|
|
static int nlm_udpport, nlm_tcpport;
|
|
int nsm_use_hostnames = 0;
|
|
|
|
/*
|
|
* Constants needed for the sysctl interface.
|
|
*/
|
|
static const unsigned long nlm_grace_period_min = 0;
|
|
static const unsigned long nlm_grace_period_max = 240;
|
|
static const unsigned long nlm_timeout_min = 3;
|
|
static const unsigned long nlm_timeout_max = 20;
|
|
static const int nlm_port_min = 0, nlm_port_max = 65535;
|
|
|
|
static struct ctl_table_header * nlm_sysctl_table;
|
|
|
|
static unsigned long get_lockd_grace_period(void)
|
|
{
|
|
/* Note: nlm_timeout should always be nonzero */
|
|
if (nlm_grace_period)
|
|
return roundup(nlm_grace_period, nlm_timeout) * HZ;
|
|
else
|
|
return nlm_timeout * 5 * HZ;
|
|
}
|
|
|
|
unsigned long get_nfs_grace_period(void)
|
|
{
|
|
unsigned long lockdgrace = get_lockd_grace_period();
|
|
unsigned long nfsdgrace = 0;
|
|
|
|
if (nlmsvc_ops)
|
|
nfsdgrace = nlmsvc_ops->get_grace_period();
|
|
|
|
return max(lockdgrace, nfsdgrace);
|
|
}
|
|
EXPORT_SYMBOL(get_nfs_grace_period);
|
|
|
|
static unsigned long set_grace_period(void)
|
|
{
|
|
nlmsvc_grace_period = 1;
|
|
return get_nfs_grace_period() + jiffies;
|
|
}
|
|
|
|
static inline void clear_grace_period(void)
|
|
{
|
|
nlmsvc_grace_period = 0;
|
|
}
|
|
|
|
/*
|
|
* This is the lockd kernel thread
|
|
*/
|
|
static void
|
|
lockd(struct svc_rqst *rqstp)
|
|
{
|
|
int err = 0;
|
|
unsigned long grace_period_expire;
|
|
|
|
/* Lock module and set up kernel thread */
|
|
/* lockd_up is waiting for us to startup, so will
|
|
* be holding a reference to this module, so it
|
|
* is safe to just claim another reference
|
|
*/
|
|
__module_get(THIS_MODULE);
|
|
lock_kernel();
|
|
|
|
/*
|
|
* Let our maker know we're running.
|
|
*/
|
|
nlmsvc_pid = current->pid;
|
|
nlmsvc_serv = rqstp->rq_server;
|
|
complete(&lockd_start_done);
|
|
|
|
daemonize("lockd");
|
|
set_freezable();
|
|
|
|
/* Process request with signals blocked, but allow SIGKILL. */
|
|
allow_signal(SIGKILL);
|
|
|
|
dprintk("NFS locking service started (ver " LOCKD_VERSION ").\n");
|
|
|
|
if (!nlm_timeout)
|
|
nlm_timeout = LOCKD_DFLT_TIMEO;
|
|
nlmsvc_timeout = nlm_timeout * HZ;
|
|
|
|
grace_period_expire = set_grace_period();
|
|
|
|
/*
|
|
* The main request loop. We don't terminate until the last
|
|
* NFS mount or NFS daemon has gone away, and we've been sent a
|
|
* signal, or else another process has taken over our job.
|
|
*/
|
|
while ((nlmsvc_users || !signalled()) && nlmsvc_pid == current->pid) {
|
|
long timeout = MAX_SCHEDULE_TIMEOUT;
|
|
char buf[RPC_MAX_ADDRBUFLEN];
|
|
|
|
if (signalled()) {
|
|
flush_signals(current);
|
|
if (nlmsvc_ops) {
|
|
nlmsvc_invalidate_all();
|
|
grace_period_expire = set_grace_period();
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Retry any blocked locks that have been notified by
|
|
* the VFS. Don't do this during grace period.
|
|
* (Theoretically, there shouldn't even be blocked locks
|
|
* during grace period).
|
|
*/
|
|
if (!nlmsvc_grace_period) {
|
|
timeout = nlmsvc_retry_blocked();
|
|
} else if (time_before(grace_period_expire, jiffies))
|
|
clear_grace_period();
|
|
|
|
/*
|
|
* Find a socket with data available and call its
|
|
* recvfrom routine.
|
|
*/
|
|
err = svc_recv(rqstp, timeout);
|
|
if (err == -EAGAIN || err == -EINTR)
|
|
continue;
|
|
if (err < 0) {
|
|
printk(KERN_WARNING
|
|
"lockd: terminating on error %d\n",
|
|
-err);
|
|
break;
|
|
}
|
|
|
|
dprintk("lockd: request from %s\n",
|
|
svc_print_addr(rqstp, buf, sizeof(buf)));
|
|
|
|
svc_process(rqstp);
|
|
}
|
|
|
|
flush_signals(current);
|
|
|
|
/*
|
|
* Check whether there's a new lockd process before
|
|
* shutting down the hosts and clearing the slot.
|
|
*/
|
|
if (!nlmsvc_pid || current->pid == nlmsvc_pid) {
|
|
if (nlmsvc_ops)
|
|
nlmsvc_invalidate_all();
|
|
nlm_shutdown_hosts();
|
|
nlmsvc_pid = 0;
|
|
nlmsvc_serv = NULL;
|
|
} else
|
|
printk(KERN_DEBUG
|
|
"lockd: new process, skipping host shutdown\n");
|
|
wake_up(&lockd_exit);
|
|
|
|
/* Exit the RPC thread */
|
|
svc_exit_thread(rqstp);
|
|
|
|
/* Release module */
|
|
unlock_kernel();
|
|
module_put_and_exit(0);
|
|
}
|
|
|
|
/*
|
|
* Make any sockets that are needed but not present.
|
|
* If nlm_udpport or nlm_tcpport were set as module
|
|
* options, make those sockets unconditionally
|
|
*/
|
|
static int make_socks(struct svc_serv *serv, int proto)
|
|
{
|
|
static int warned;
|
|
struct svc_xprt *xprt;
|
|
int err = 0;
|
|
|
|
if (proto == IPPROTO_UDP || nlm_udpport) {
|
|
xprt = svc_find_xprt(serv, "udp", 0, 0);
|
|
if (!xprt)
|
|
err = svc_create_xprt(serv, "udp", nlm_udpport,
|
|
SVC_SOCK_DEFAULTS);
|
|
else
|
|
svc_xprt_put(xprt);
|
|
}
|
|
if (err >= 0 && (proto == IPPROTO_TCP || nlm_tcpport)) {
|
|
xprt = svc_find_xprt(serv, "tcp", 0, 0);
|
|
if (!xprt)
|
|
err = svc_create_xprt(serv, "tcp", nlm_tcpport,
|
|
SVC_SOCK_DEFAULTS);
|
|
else
|
|
svc_xprt_put(xprt);
|
|
}
|
|
if (err >= 0) {
|
|
warned = 0;
|
|
err = 0;
|
|
} else if (warned++ == 0)
|
|
printk(KERN_WARNING
|
|
"lockd_up: makesock failed, error=%d\n", err);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* Bring up the lockd process if it's not already up.
|
|
*/
|
|
int
|
|
lockd_up(int proto) /* Maybe add a 'family' option when IPv6 is supported ?? */
|
|
{
|
|
struct svc_serv * serv;
|
|
int error = 0;
|
|
|
|
mutex_lock(&nlmsvc_mutex);
|
|
/*
|
|
* Check whether we're already up and running.
|
|
*/
|
|
if (nlmsvc_pid) {
|
|
if (proto)
|
|
error = make_socks(nlmsvc_serv, proto);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Sanity check: if there's no pid,
|
|
* we should be the first user ...
|
|
*/
|
|
if (nlmsvc_users)
|
|
printk(KERN_WARNING
|
|
"lockd_up: no pid, %d users??\n", nlmsvc_users);
|
|
|
|
error = -ENOMEM;
|
|
serv = svc_create(&nlmsvc_program, LOCKD_BUFSIZE, NULL);
|
|
if (!serv) {
|
|
printk(KERN_WARNING "lockd_up: create service failed\n");
|
|
goto out;
|
|
}
|
|
|
|
if ((error = make_socks(serv, proto)) < 0)
|
|
goto destroy_and_out;
|
|
|
|
/*
|
|
* Create the kernel thread and wait for it to start.
|
|
*/
|
|
error = svc_create_thread(lockd, serv);
|
|
if (error) {
|
|
printk(KERN_WARNING
|
|
"lockd_up: create thread failed, error=%d\n", error);
|
|
goto destroy_and_out;
|
|
}
|
|
wait_for_completion(&lockd_start_done);
|
|
|
|
/*
|
|
* Note: svc_serv structures have an initial use count of 1,
|
|
* so we exit through here on both success and failure.
|
|
*/
|
|
destroy_and_out:
|
|
svc_destroy(serv);
|
|
out:
|
|
if (!error)
|
|
nlmsvc_users++;
|
|
mutex_unlock(&nlmsvc_mutex);
|
|
return error;
|
|
}
|
|
EXPORT_SYMBOL(lockd_up);
|
|
|
|
/*
|
|
* Decrement the user count and bring down lockd if we're the last.
|
|
*/
|
|
void
|
|
lockd_down(void)
|
|
{
|
|
static int warned;
|
|
|
|
mutex_lock(&nlmsvc_mutex);
|
|
if (nlmsvc_users) {
|
|
if (--nlmsvc_users)
|
|
goto out;
|
|
} else
|
|
printk(KERN_WARNING "lockd_down: no users! pid=%d\n", nlmsvc_pid);
|
|
|
|
if (!nlmsvc_pid) {
|
|
if (warned++ == 0)
|
|
printk(KERN_WARNING "lockd_down: no lockd running.\n");
|
|
goto out;
|
|
}
|
|
warned = 0;
|
|
|
|
kill_proc(nlmsvc_pid, SIGKILL, 1);
|
|
/*
|
|
* Wait for the lockd process to exit, but since we're holding
|
|
* the lockd semaphore, we can't wait around forever ...
|
|
*/
|
|
clear_thread_flag(TIF_SIGPENDING);
|
|
interruptible_sleep_on_timeout(&lockd_exit, HZ);
|
|
if (nlmsvc_pid) {
|
|
printk(KERN_WARNING
|
|
"lockd_down: lockd failed to exit, clearing pid\n");
|
|
nlmsvc_pid = 0;
|
|
}
|
|
spin_lock_irq(¤t->sighand->siglock);
|
|
recalc_sigpending();
|
|
spin_unlock_irq(¤t->sighand->siglock);
|
|
out:
|
|
mutex_unlock(&nlmsvc_mutex);
|
|
}
|
|
EXPORT_SYMBOL(lockd_down);
|
|
|
|
/*
|
|
* Sysctl parameters (same as module parameters, different interface).
|
|
*/
|
|
|
|
static ctl_table nlm_sysctls[] = {
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nlm_grace_period",
|
|
.data = &nlm_grace_period,
|
|
.maxlen = sizeof(unsigned long),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_doulongvec_minmax,
|
|
.extra1 = (unsigned long *) &nlm_grace_period_min,
|
|
.extra2 = (unsigned long *) &nlm_grace_period_max,
|
|
},
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nlm_timeout",
|
|
.data = &nlm_timeout,
|
|
.maxlen = sizeof(unsigned long),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_doulongvec_minmax,
|
|
.extra1 = (unsigned long *) &nlm_timeout_min,
|
|
.extra2 = (unsigned long *) &nlm_timeout_max,
|
|
},
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nlm_udpport",
|
|
.data = &nlm_udpport,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_dointvec_minmax,
|
|
.extra1 = (int *) &nlm_port_min,
|
|
.extra2 = (int *) &nlm_port_max,
|
|
},
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nlm_tcpport",
|
|
.data = &nlm_tcpport,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_dointvec_minmax,
|
|
.extra1 = (int *) &nlm_port_min,
|
|
.extra2 = (int *) &nlm_port_max,
|
|
},
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nsm_use_hostnames",
|
|
.data = &nsm_use_hostnames,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_dointvec,
|
|
},
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nsm_local_state",
|
|
.data = &nsm_local_state,
|
|
.maxlen = sizeof(int),
|
|
.mode = 0644,
|
|
.proc_handler = &proc_dointvec,
|
|
},
|
|
{ .ctl_name = 0 }
|
|
};
|
|
|
|
static ctl_table nlm_sysctl_dir[] = {
|
|
{
|
|
.ctl_name = CTL_UNNUMBERED,
|
|
.procname = "nfs",
|
|
.mode = 0555,
|
|
.child = nlm_sysctls,
|
|
},
|
|
{ .ctl_name = 0 }
|
|
};
|
|
|
|
static ctl_table nlm_sysctl_root[] = {
|
|
{
|
|
.ctl_name = CTL_FS,
|
|
.procname = "fs",
|
|
.mode = 0555,
|
|
.child = nlm_sysctl_dir,
|
|
},
|
|
{ .ctl_name = 0 }
|
|
};
|
|
|
|
/*
|
|
* Module (and sysfs) parameters.
|
|
*/
|
|
|
|
#define param_set_min_max(name, type, which_strtol, min, max) \
|
|
static int param_set_##name(const char *val, struct kernel_param *kp) \
|
|
{ \
|
|
char *endp; \
|
|
__typeof__(type) num = which_strtol(val, &endp, 0); \
|
|
if (endp == val || *endp || num < (min) || num > (max)) \
|
|
return -EINVAL; \
|
|
*((int *) kp->arg) = num; \
|
|
return 0; \
|
|
}
|
|
|
|
static inline int is_callback(u32 proc)
|
|
{
|
|
return proc == NLMPROC_GRANTED
|
|
|| proc == NLMPROC_GRANTED_MSG
|
|
|| proc == NLMPROC_TEST_RES
|
|
|| proc == NLMPROC_LOCK_RES
|
|
|| proc == NLMPROC_CANCEL_RES
|
|
|| proc == NLMPROC_UNLOCK_RES
|
|
|| proc == NLMPROC_NSM_NOTIFY;
|
|
}
|
|
|
|
|
|
static int lockd_authenticate(struct svc_rqst *rqstp)
|
|
{
|
|
rqstp->rq_client = NULL;
|
|
switch (rqstp->rq_authop->flavour) {
|
|
case RPC_AUTH_NULL:
|
|
case RPC_AUTH_UNIX:
|
|
if (rqstp->rq_proc == 0)
|
|
return SVC_OK;
|
|
if (is_callback(rqstp->rq_proc)) {
|
|
/* Leave it to individual procedures to
|
|
* call nlmsvc_lookup_host(rqstp)
|
|
*/
|
|
return SVC_OK;
|
|
}
|
|
return svc_set_client(rqstp);
|
|
}
|
|
return SVC_DENIED;
|
|
}
|
|
|
|
|
|
param_set_min_max(port, int, simple_strtol, 0, 65535)
|
|
param_set_min_max(grace_period, unsigned long, simple_strtoul,
|
|
nlm_grace_period_min, nlm_grace_period_max)
|
|
param_set_min_max(timeout, unsigned long, simple_strtoul,
|
|
nlm_timeout_min, nlm_timeout_max)
|
|
|
|
MODULE_AUTHOR("Olaf Kirch <okir@monad.swb.de>");
|
|
MODULE_DESCRIPTION("NFS file locking service version " LOCKD_VERSION ".");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
module_param_call(nlm_grace_period, param_set_grace_period, param_get_ulong,
|
|
&nlm_grace_period, 0644);
|
|
module_param_call(nlm_timeout, param_set_timeout, param_get_ulong,
|
|
&nlm_timeout, 0644);
|
|
module_param_call(nlm_udpport, param_set_port, param_get_int,
|
|
&nlm_udpport, 0644);
|
|
module_param_call(nlm_tcpport, param_set_port, param_get_int,
|
|
&nlm_tcpport, 0644);
|
|
module_param(nsm_use_hostnames, bool, 0644);
|
|
|
|
/*
|
|
* Initialising and terminating the module.
|
|
*/
|
|
|
|
static int __init init_nlm(void)
|
|
{
|
|
nlm_sysctl_table = register_sysctl_table(nlm_sysctl_root);
|
|
return nlm_sysctl_table ? 0 : -ENOMEM;
|
|
}
|
|
|
|
static void __exit exit_nlm(void)
|
|
{
|
|
/* FIXME: delete all NLM clients */
|
|
nlm_shutdown_hosts();
|
|
unregister_sysctl_table(nlm_sysctl_table);
|
|
}
|
|
|
|
module_init(init_nlm);
|
|
module_exit(exit_nlm);
|
|
|
|
/*
|
|
* Define NLM program and procedures
|
|
*/
|
|
static struct svc_version nlmsvc_version1 = {
|
|
.vs_vers = 1,
|
|
.vs_nproc = 17,
|
|
.vs_proc = nlmsvc_procedures,
|
|
.vs_xdrsize = NLMSVC_XDRSIZE,
|
|
};
|
|
static struct svc_version nlmsvc_version3 = {
|
|
.vs_vers = 3,
|
|
.vs_nproc = 24,
|
|
.vs_proc = nlmsvc_procedures,
|
|
.vs_xdrsize = NLMSVC_XDRSIZE,
|
|
};
|
|
#ifdef CONFIG_LOCKD_V4
|
|
static struct svc_version nlmsvc_version4 = {
|
|
.vs_vers = 4,
|
|
.vs_nproc = 24,
|
|
.vs_proc = nlmsvc_procedures4,
|
|
.vs_xdrsize = NLMSVC_XDRSIZE,
|
|
};
|
|
#endif
|
|
static struct svc_version * nlmsvc_version[] = {
|
|
[1] = &nlmsvc_version1,
|
|
[3] = &nlmsvc_version3,
|
|
#ifdef CONFIG_LOCKD_V4
|
|
[4] = &nlmsvc_version4,
|
|
#endif
|
|
};
|
|
|
|
static struct svc_stat nlmsvc_stats;
|
|
|
|
#define NLM_NRVERS ARRAY_SIZE(nlmsvc_version)
|
|
static struct svc_program nlmsvc_program = {
|
|
.pg_prog = NLM_PROGRAM, /* program number */
|
|
.pg_nvers = NLM_NRVERS, /* number of entries in nlmsvc_version */
|
|
.pg_vers = nlmsvc_version, /* version table */
|
|
.pg_name = "lockd", /* service name */
|
|
.pg_class = "nfsd", /* share authentication with nfsd */
|
|
.pg_stats = &nlmsvc_stats, /* stats table */
|
|
.pg_authenticate = &lockd_authenticate /* export authentication */
|
|
};
|