Commit 150030b7 authored by Matthew Wilcox's avatar Matthew Wilcox Committed by Matthew Wilcox

NFS: Switch from intr mount option to TASK_KILLABLE

By using the TASK_KILLABLE infrastructure, we can get rid of the 'intr'
mount option.  We have to use _killable everywhere instead of _interruptible
as we get rid of rpc_clnt_sigmask/sigunmask.
Signed-off-by: default avatarLiam R. Howlett <howlett@gmail.com>
Signed-off-by: default avatarMatthew Wilcox <willy@linux.intel.com>
parent 009e577e
...@@ -302,7 +302,7 @@ static struct nfs_client *nfs_get_client(const char *hostname, ...@@ -302,7 +302,7 @@ static struct nfs_client *nfs_get_client(const char *hostname,
if (new) if (new)
nfs_free_client(new); nfs_free_client(new);
error = wait_event_interruptible(nfs_client_active_wq, error = wait_event_killable(nfs_client_active_wq,
clp->cl_cons_state != NFS_CS_INITING); clp->cl_cons_state != NFS_CS_INITING);
if (error < 0) { if (error < 0) {
nfs_put_client(clp); nfs_put_client(clp);
...@@ -494,10 +494,6 @@ static int nfs_init_server_rpcclient(struct nfs_server *server, rpc_authflavor_t ...@@ -494,10 +494,6 @@ static int nfs_init_server_rpcclient(struct nfs_server *server, rpc_authflavor_t
if (server->flags & NFS_MOUNT_SOFT) if (server->flags & NFS_MOUNT_SOFT)
server->client->cl_softrtry = 1; server->client->cl_softrtry = 1;
server->client->cl_intr = 0;
if (server->flags & NFS4_MOUNT_INTR)
server->client->cl_intr = 1;
return 0; return 0;
} }
......
...@@ -193,7 +193,7 @@ static ssize_t nfs_direct_wait(struct nfs_direct_req *dreq) ...@@ -193,7 +193,7 @@ static ssize_t nfs_direct_wait(struct nfs_direct_req *dreq)
if (dreq->iocb) if (dreq->iocb)
goto out; goto out;
result = wait_for_completion_interruptible(&dreq->completion); result = wait_for_completion_killable(&dreq->completion);
if (!result) if (!result)
result = dreq->error; result = dreq->error;
...@@ -391,9 +391,7 @@ static ssize_t nfs_direct_read(struct kiocb *iocb, const struct iovec *iov, ...@@ -391,9 +391,7 @@ static ssize_t nfs_direct_read(struct kiocb *iocb, const struct iovec *iov,
unsigned long nr_segs, loff_t pos) unsigned long nr_segs, loff_t pos)
{ {
ssize_t result = 0; ssize_t result = 0;
sigset_t oldset;
struct inode *inode = iocb->ki_filp->f_mapping->host; struct inode *inode = iocb->ki_filp->f_mapping->host;
struct rpc_clnt *clnt = NFS_CLIENT(inode);
struct nfs_direct_req *dreq; struct nfs_direct_req *dreq;
dreq = nfs_direct_req_alloc(); dreq = nfs_direct_req_alloc();
...@@ -405,11 +403,9 @@ static ssize_t nfs_direct_read(struct kiocb *iocb, const struct iovec *iov, ...@@ -405,11 +403,9 @@ static ssize_t nfs_direct_read(struct kiocb *iocb, const struct iovec *iov,
if (!is_sync_kiocb(iocb)) if (!is_sync_kiocb(iocb))
dreq->iocb = iocb; dreq->iocb = iocb;
rpc_clnt_sigmask(clnt, &oldset);
result = nfs_direct_read_schedule_iovec(dreq, iov, nr_segs, pos); result = nfs_direct_read_schedule_iovec(dreq, iov, nr_segs, pos);
if (!result) if (!result)
result = nfs_direct_wait(dreq); result = nfs_direct_wait(dreq);
rpc_clnt_sigunmask(clnt, &oldset);
nfs_direct_req_release(dreq); nfs_direct_req_release(dreq);
return result; return result;
...@@ -767,9 +763,7 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, const struct iovec *iov, ...@@ -767,9 +763,7 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, const struct iovec *iov,
size_t count) size_t count)
{ {
ssize_t result = 0; ssize_t result = 0;
sigset_t oldset;
struct inode *inode = iocb->ki_filp->f_mapping->host; struct inode *inode = iocb->ki_filp->f_mapping->host;
struct rpc_clnt *clnt = NFS_CLIENT(inode);
struct nfs_direct_req *dreq; struct nfs_direct_req *dreq;
size_t wsize = NFS_SERVER(inode)->wsize; size_t wsize = NFS_SERVER(inode)->wsize;
int sync = 0; int sync = 0;
...@@ -787,11 +781,9 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, const struct iovec *iov, ...@@ -787,11 +781,9 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, const struct iovec *iov,
if (!is_sync_kiocb(iocb)) if (!is_sync_kiocb(iocb))
dreq->iocb = iocb; dreq->iocb = iocb;
rpc_clnt_sigmask(clnt, &oldset);
result = nfs_direct_write_schedule_iovec(dreq, iov, nr_segs, pos, sync); result = nfs_direct_write_schedule_iovec(dreq, iov, nr_segs, pos, sync);
if (!result) if (!result)
result = nfs_direct_wait(dreq); result = nfs_direct_wait(dreq);
rpc_clnt_sigunmask(clnt, &oldset);
nfs_direct_req_release(dreq); nfs_direct_req_release(dreq);
return result; return result;
......
...@@ -433,15 +433,11 @@ static int nfs_wait_schedule(void *word) ...@@ -433,15 +433,11 @@ static int nfs_wait_schedule(void *word)
*/ */
static int nfs_wait_on_inode(struct inode *inode) static int nfs_wait_on_inode(struct inode *inode)
{ {
struct rpc_clnt *clnt = NFS_CLIENT(inode);
struct nfs_inode *nfsi = NFS_I(inode); struct nfs_inode *nfsi = NFS_I(inode);
sigset_t oldmask;
int error; int error;
rpc_clnt_sigmask(clnt, &oldmask);
error = wait_on_bit_lock(&nfsi->flags, NFS_INO_REVALIDATING, error = wait_on_bit_lock(&nfsi->flags, NFS_INO_REVALIDATING,
nfs_wait_schedule, TASK_INTERRUPTIBLE); nfs_wait_schedule, TASK_KILLABLE);
rpc_clnt_sigunmask(clnt, &oldmask);
return error; return error;
} }
......
...@@ -56,7 +56,7 @@ int nfs_mount(struct sockaddr *addr, size_t len, char *hostname, char *path, ...@@ -56,7 +56,7 @@ int nfs_mount(struct sockaddr *addr, size_t len, char *hostname, char *path,
.program = &mnt_program, .program = &mnt_program,
.version = version, .version = version,
.authflavor = RPC_AUTH_UNIX, .authflavor = RPC_AUTH_UNIX,
.flags = RPC_CLNT_CREATE_INTR, .flags = 0,
}; };
struct rpc_clnt *mnt_clnt; struct rpc_clnt *mnt_clnt;
int status; int status;
......
...@@ -27,17 +27,14 @@ ...@@ -27,17 +27,14 @@
static int static int
nfs3_rpc_wrapper(struct rpc_clnt *clnt, struct rpc_message *msg, int flags) nfs3_rpc_wrapper(struct rpc_clnt *clnt, struct rpc_message *msg, int flags)
{ {
sigset_t oldset;
int res; int res;
rpc_clnt_sigmask(clnt, &oldset);
do { do {
res = rpc_call_sync(clnt, msg, flags); res = rpc_call_sync(clnt, msg, flags);
if (res != -EJUKEBOX) if (res != -EJUKEBOX)
break; break;
schedule_timeout_interruptible(NFS_JUKEBOX_RETRY_TIME); schedule_timeout_killable(NFS_JUKEBOX_RETRY_TIME);
res = -ERESTARTSYS; res = -ERESTARTSYS;
} while (!signalled()); } while (!fatal_signal_pending(current));
rpc_clnt_sigunmask(clnt, &oldset);
return res; return res;
} }
......
...@@ -316,12 +316,9 @@ static void nfs4_opendata_put(struct nfs4_opendata *p) ...@@ -316,12 +316,9 @@ static void nfs4_opendata_put(struct nfs4_opendata *p)
static int nfs4_wait_for_completion_rpc_task(struct rpc_task *task) static int nfs4_wait_for_completion_rpc_task(struct rpc_task *task)
{ {
sigset_t oldset;
int ret; int ret;
rpc_clnt_sigmask(task->tk_client, &oldset);
ret = rpc_wait_for_completion_task(task); ret = rpc_wait_for_completion_task(task);
rpc_clnt_sigunmask(task->tk_client, &oldset);
return ret; return ret;
} }
...@@ -2806,9 +2803,9 @@ nfs4_async_handle_error(struct rpc_task *task, const struct nfs_server *server) ...@@ -2806,9 +2803,9 @@ nfs4_async_handle_error(struct rpc_task *task, const struct nfs_server *server)
return 0; return 0;
} }
static int nfs4_wait_bit_interruptible(void *word) static int nfs4_wait_bit_killable(void *word)
{ {
if (signal_pending(current)) if (fatal_signal_pending(current))
return -ERESTARTSYS; return -ERESTARTSYS;
schedule(); schedule();
return 0; return 0;
...@@ -2816,18 +2813,14 @@ static int nfs4_wait_bit_interruptible(void *word) ...@@ -2816,18 +2813,14 @@ static int nfs4_wait_bit_interruptible(void *word)
static int nfs4_wait_clnt_recover(struct rpc_clnt *clnt, struct nfs_client *clp) static int nfs4_wait_clnt_recover(struct rpc_clnt *clnt, struct nfs_client *clp)
{ {
sigset_t oldset;
int res; int res;
might_sleep(); might_sleep();
rwsem_acquire(&clp->cl_sem.dep_map, 0, 0, _RET_IP_); rwsem_acquire(&clp->cl_sem.dep_map, 0, 0, _RET_IP_);
rpc_clnt_sigmask(clnt, &oldset);
res = wait_on_bit(&clp->cl_state, NFS4CLNT_STATE_RECOVER, res = wait_on_bit(&clp->cl_state, NFS4CLNT_STATE_RECOVER,
nfs4_wait_bit_interruptible, nfs4_wait_bit_killable, TASK_KILLABLE);
TASK_INTERRUPTIBLE);
rpc_clnt_sigunmask(clnt, &oldset);
rwsem_release(&clp->cl_sem.dep_map, 1, _RET_IP_); rwsem_release(&clp->cl_sem.dep_map, 1, _RET_IP_);
return res; return res;
...@@ -2835,7 +2828,6 @@ static int nfs4_wait_clnt_recover(struct rpc_clnt *clnt, struct nfs_client *clp) ...@@ -2835,7 +2828,6 @@ static int nfs4_wait_clnt_recover(struct rpc_clnt *clnt, struct nfs_client *clp)
static int nfs4_delay(struct rpc_clnt *clnt, long *timeout) static int nfs4_delay(struct rpc_clnt *clnt, long *timeout)
{ {
sigset_t oldset;
int res = 0; int res = 0;
might_sleep(); might_sleep();
...@@ -2844,14 +2836,9 @@ static int nfs4_delay(struct rpc_clnt *clnt, long *timeout) ...@@ -2844,14 +2836,9 @@ static int nfs4_delay(struct rpc_clnt *clnt, long *timeout)
*timeout = NFS4_POLL_RETRY_MIN; *timeout = NFS4_POLL_RETRY_MIN;
if (*timeout > NFS4_POLL_RETRY_MAX) if (*timeout > NFS4_POLL_RETRY_MAX)
*timeout = NFS4_POLL_RETRY_MAX; *timeout = NFS4_POLL_RETRY_MAX;
rpc_clnt_sigmask(clnt, &oldset); schedule_timeout_killable(*timeout);
if (clnt->cl_intr) { if (fatal_signal_pending(current))
schedule_timeout_interruptible(*timeout); res = -ERESTARTSYS;
if (signalled())
res = -ERESTARTSYS;
} else
schedule_timeout_uninterruptible(*timeout);
rpc_clnt_sigunmask(clnt, &oldset);
*timeout <<= 1; *timeout <<= 1;
return res; return res;
} }
...@@ -3085,7 +3072,7 @@ int nfs4_proc_delegreturn(struct inode *inode, struct rpc_cred *cred, const nfs4 ...@@ -3085,7 +3072,7 @@ int nfs4_proc_delegreturn(struct inode *inode, struct rpc_cred *cred, const nfs4
static unsigned long static unsigned long
nfs4_set_lock_task_retry(unsigned long timeout) nfs4_set_lock_task_retry(unsigned long timeout)
{ {
schedule_timeout_interruptible(timeout); schedule_timeout_killable(timeout);
timeout <<= 1; timeout <<= 1;
if (timeout > NFS4_LOCK_MAXTIMEOUT) if (timeout > NFS4_LOCK_MAXTIMEOUT)
return NFS4_LOCK_MAXTIMEOUT; return NFS4_LOCK_MAXTIMEOUT;
......
...@@ -228,10 +228,7 @@ static int __init root_nfs_parse(char *name, char *buf) ...@@ -228,10 +228,7 @@ static int __init root_nfs_parse(char *name, char *buf)
nfs_data.flags &= ~NFS_MOUNT_SOFT; nfs_data.flags &= ~NFS_MOUNT_SOFT;
break; break;
case Opt_intr: case Opt_intr:
nfs_data.flags |= NFS_MOUNT_INTR;
break;
case Opt_nointr: case Opt_nointr:
nfs_data.flags &= ~NFS_MOUNT_INTR;
break; break;
case Opt_posix: case Opt_posix:
nfs_data.flags |= NFS_MOUNT_POSIX; nfs_data.flags |= NFS_MOUNT_POSIX;
......
...@@ -67,7 +67,7 @@ nfs_create_request(struct nfs_open_context *ctx, struct inode *inode, ...@@ -67,7 +67,7 @@ nfs_create_request(struct nfs_open_context *ctx, struct inode *inode,
if (req != NULL) if (req != NULL)
break; break;
if (signalled() && (server->flags & NFS_MOUNT_INTR)) if (fatal_signal_pending(current))
return ERR_PTR(-ERESTARTSYS); return ERR_PTR(-ERESTARTSYS);
yield(); yield();
} }
...@@ -175,11 +175,11 @@ void nfs_release_request(struct nfs_page *req) ...@@ -175,11 +175,11 @@ void nfs_release_request(struct nfs_page *req)
kref_put(&req->wb_kref, nfs_free_request); kref_put(&req->wb_kref, nfs_free_request);
} }
static int nfs_wait_bit_interruptible(void *word) static int nfs_wait_bit_killable(void *word)
{ {
int ret = 0; int ret = 0;
if (signal_pending(current)) if (fatal_signal_pending(current))
ret = -ERESTARTSYS; ret = -ERESTARTSYS;
else else
schedule(); schedule();
...@@ -190,26 +190,18 @@ static int nfs_wait_bit_interruptible(void *word) ...@@ -190,26 +190,18 @@ static int nfs_wait_bit_interruptible(void *word)
* nfs_wait_on_request - Wait for a request to complete. * nfs_wait_on_request - Wait for a request to complete.
* @req: request to wait upon. * @req: request to wait upon.
* *
* Interruptible by signals only if mounted with intr flag. * Interruptible by fatal signals only.
* The user is responsible for holding a count on the request. * The user is responsible for holding a count on the request.
*/ */
int int
nfs_wait_on_request(struct nfs_page *req) nfs_wait_on_request(struct nfs_page *req)
{ {
struct rpc_clnt *clnt = NFS_CLIENT(req->wb_context->path.dentry->d_inode);
sigset_t oldmask;
int ret = 0; int ret = 0;
if (!test_bit(PG_BUSY, &req->wb_flags)) if (!test_bit(PG_BUSY, &req->wb_flags))
goto out; goto out;
/*
* Note: the call to rpc_clnt_sigmask() suffices to ensure that we
* are not interrupted if intr flag is not set
*/
rpc_clnt_sigmask(clnt, &oldmask);
ret = out_of_line_wait_on_bit(&req->wb_flags, PG_BUSY, ret = out_of_line_wait_on_bit(&req->wb_flags, PG_BUSY,
nfs_wait_bit_interruptible, TASK_INTERRUPTIBLE); nfs_wait_bit_killable, TASK_KILLABLE);
rpc_clnt_sigunmask(clnt, &oldmask);
out: out:
return ret; return ret;
} }
......
...@@ -212,12 +212,7 @@ nfs_async_read_error(struct list_head *head) ...@@ -212,12 +212,7 @@ nfs_async_read_error(struct list_head *head)
*/ */
static void nfs_execute_read(struct nfs_read_data *data) static void nfs_execute_read(struct nfs_read_data *data)
{ {
struct rpc_clnt *clnt = NFS_CLIENT(data->inode);
sigset_t oldset;
rpc_clnt_sigmask(clnt, &oldset);
rpc_execute(&data->task); rpc_execute(&data->task);
rpc_clnt_sigunmask(clnt, &oldset);
} }
/* /*
......
...@@ -424,7 +424,6 @@ static void nfs_show_mount_options(struct seq_file *m, struct nfs_server *nfss, ...@@ -424,7 +424,6 @@ static void nfs_show_mount_options(struct seq_file *m, struct nfs_server *nfss,
const char *nostr; const char *nostr;
} nfs_info[] = { } nfs_info[] = {
{ NFS_MOUNT_SOFT, ",soft", ",hard" }, { NFS_MOUNT_SOFT, ",soft", ",hard" },
{ NFS_MOUNT_INTR, ",intr", ",nointr" },
{ NFS_MOUNT_NOCTO, ",nocto", "" }, { NFS_MOUNT_NOCTO, ",nocto", "" },
{ NFS_MOUNT_NOAC, ",noac", "" }, { NFS_MOUNT_NOAC, ",noac", "" },
{ NFS_MOUNT_NONLM, ",nolock", "" }, { NFS_MOUNT_NONLM, ",nolock", "" },
...@@ -624,10 +623,7 @@ static int nfs_parse_mount_options(char *raw, ...@@ -624,10 +623,7 @@ static int nfs_parse_mount_options(char *raw,
mnt->flags &= ~NFS_MOUNT_SOFT; mnt->flags &= ~NFS_MOUNT_SOFT;
break; break;
case Opt_intr: case Opt_intr:
mnt->flags |= NFS_MOUNT_INTR;
break;
case Opt_nointr: case Opt_nointr:
mnt->flags &= ~NFS_MOUNT_INTR;
break; break;
case Opt_posix: case Opt_posix:
mnt->flags |= NFS_MOUNT_POSIX; mnt->flags |= NFS_MOUNT_POSIX;
......
...@@ -490,7 +490,7 @@ int nfs_reschedule_unstable_write(struct nfs_page *req) ...@@ -490,7 +490,7 @@ int nfs_reschedule_unstable_write(struct nfs_page *req)
/* /*
* Wait for a request to complete. * Wait for a request to complete.
* *
* Interruptible by signals only if mounted with intr flag. * Interruptible by fatal signals only.
*/ */
static int nfs_wait_on_requests_locked(struct inode *inode, pgoff_t idx_start, unsigned int npages) static int nfs_wait_on_requests_locked(struct inode *inode, pgoff_t idx_start, unsigned int npages)
{ {
...@@ -816,12 +816,7 @@ static void nfs_write_rpcsetup(struct nfs_page *req, ...@@ -816,12 +816,7 @@ static void nfs_write_rpcsetup(struct nfs_page *req,
static void nfs_execute_write(struct nfs_write_data *data) static void nfs_execute_write(struct nfs_write_data *data)
{ {
struct rpc_clnt *clnt = NFS_CLIENT(data->inode);
sigset_t oldset;
rpc_clnt_sigmask(clnt, &oldset);
rpc_execute(&data->task); rpc_execute(&data->task);
rpc_clnt_sigunmask(clnt, &oldset);
} }
/* /*
......
...@@ -516,14 +516,7 @@ extern void * nfs_root_data(void); ...@@ -516,14 +516,7 @@ extern void * nfs_root_data(void);
#define nfs_wait_event(clnt, wq, condition) \ #define nfs_wait_event(clnt, wq, condition) \
({ \ ({ \
int __retval = 0; \ int __retval = wait_event_killable(wq, condition); \
if (clnt->cl_intr) { \
sigset_t oldmask; \
rpc_clnt_sigmask(clnt, &oldmask); \
__retval = wait_event_interruptible(wq, condition); \
rpc_clnt_sigunmask(clnt, &oldmask); \
} else \
wait_event(wq, condition); \
__retval; \ __retval; \
}) })
......
...@@ -48,7 +48,7 @@ struct nfs_mount_data { ...@@ -48,7 +48,7 @@ struct nfs_mount_data {
/* bits in the flags field */ /* bits in the flags field */
#define NFS_MOUNT_SOFT 0x0001 /* 1 */ #define NFS_MOUNT_SOFT 0x0001 /* 1 */
#define NFS_MOUNT_INTR 0x0002 /* 1 */ #define NFS_MOUNT_INTR 0x0002 /* 1 */ /* now unused, but ABI */
#define NFS_MOUNT_SECURE 0x0004 /* 1 */ #define NFS_MOUNT_SECURE 0x0004 /* 1 */
#define NFS_MOUNT_POSIX 0x0008 /* 1 */ #define NFS_MOUNT_POSIX 0x0008 /* 1 */
#define NFS_MOUNT_NOCTO 0x0010 /* 1 */ #define NFS_MOUNT_NOCTO 0x0010 /* 1 */
......
...@@ -41,7 +41,6 @@ struct rpc_clnt { ...@@ -41,7 +41,6 @@ struct rpc_clnt {
struct rpc_iostats * cl_metrics; /* per-client statistics */ struct rpc_iostats * cl_metrics; /* per-client statistics */
unsigned int cl_softrtry : 1,/* soft timeouts */ unsigned int cl_softrtry : 1,/* soft timeouts */
cl_intr : 1,/* interruptible */
cl_discrtry : 1,/* disconnect before retry */ cl_discrtry : 1,/* disconnect before retry */
cl_autobind : 1;/* use getport() */ cl_autobind : 1;/* use getport() */
...@@ -109,7 +108,6 @@ struct rpc_create_args { ...@@ -109,7 +108,6 @@ struct rpc_create_args {
/* Values for "flags" field */ /* Values for "flags" field */
#define RPC_CLNT_CREATE_HARDRTRY (1UL << 0) #define RPC_CLNT_CREATE_HARDRTRY (1UL << 0)
#define RPC_CLNT_CREATE_INTR (1UL << 1)
#define RPC_CLNT_CREATE_AUTOBIND (1UL << 2) #define RPC_CLNT_CREATE_AUTOBIND (1UL << 2)
#define RPC_CLNT_CREATE_NONPRIVPORT (1UL << 3) #define RPC_CLNT_CREATE_NONPRIVPORT (1UL << 3)
#define RPC_CLNT_CREATE_NOPING (1UL << 4) #define RPC_CLNT_CREATE_NOPING (1UL << 4)
......
...@@ -128,7 +128,6 @@ struct rpc_call_ops { ...@@ -128,7 +128,6 @@ struct rpc_call_ops {
#define RPC_TASK_DYNAMIC 0x0080 /* task was kmalloc'ed */ #define RPC_TASK_DYNAMIC 0x0080 /* task was kmalloc'ed */
#define RPC_TASK_KILLED 0x0100 /* task was killed */ #define RPC_TASK_KILLED 0x0100 /* task was killed */
#define RPC_TASK_SOFT 0x0200 /* Use soft timeouts */ #define RPC_TASK_SOFT 0x0200 /* Use soft timeouts */
#define RPC_TASK_NOINTR 0x0400 /* uninterruptible task */
#define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC) #define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC)
#define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER) #define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER)
...@@ -136,7 +135,6 @@ struct rpc_call_ops { ...@@ -136,7 +135,6 @@ struct rpc_call_ops {
#define RPC_ASSASSINATED(t) ((t)->tk_flags & RPC_TASK_KILLED) #define RPC_ASSASSINATED(t) ((t)->tk_flags & RPC_TASK_KILLED)
#define RPC_DO_CALLBACK(t) ((t)->tk_callback != NULL) #define RPC_DO_CALLBACK(t) ((t)->tk_callback != NULL)
#define RPC_IS_SOFT(t) ((t)->tk_flags & RPC_TASK_SOFT) #define RPC_IS_SOFT(t) ((t)->tk_flags & RPC_TASK_SOFT)
#define RPC_TASK_UNINTERRUPTIBLE(t) ((t)->tk_flags & RPC_TASK_NOINTR)
#define RPC_TASK_RUNNING 0 #define RPC_TASK_RUNNING 0
#define RPC_TASK_QUEUED 1 #define RPC_TASK_QUEUED 1
......
...@@ -281,7 +281,7 @@ struct rpc_clnt *rpc_create(struct rpc_create_args *args) ...@@ -281,7 +281,7 @@ struct rpc_clnt *rpc_create(struct rpc_create_args *args)
return clnt; return clnt;
if (!(args->flags & RPC_CLNT_CREATE_NOPING)) { if (!(args->flags & RPC_CLNT_CREATE_NOPING)) {
int err = rpc_ping(clnt, RPC_TASK_SOFT|RPC_TASK_NOINTR); int err = rpc_ping(clnt, RPC_TASK_SOFT);
if (err != 0) { if (err != 0) {
rpc_shutdown_client(clnt); rpc_shutdown_client(clnt);
return ERR_PTR(err); return ERR_PTR(err);
...@@ -292,8 +292,6 @@ struct rpc_clnt *rpc_create(struct rpc_create_args *args) ...@@ -292,8 +292,6 @@ struct rpc_clnt *rpc_create(struct rpc_create_args *args)
if (args->flags & RPC_CLNT_CREATE_HARDRTRY) if (args->flags & RPC_CLNT_CREATE_HARDRTRY)
clnt->cl_softrtry = 0; clnt->cl_softrtry = 0;
if (args->flags & RPC_CLNT_CREATE_INTR)
clnt->cl_intr = 1;
if (args->flags & RPC_CLNT_CREATE_AUTOBIND) if (args->flags & RPC_CLNT_CREATE_AUTOBIND)
clnt->cl_autobind = 1; clnt->cl_autobind = 1;
if (args->flags & RPC_CLNT_CREATE_DISCRTRY) if (args->flags & RPC_CLNT_CREATE_DISCRTRY)
...@@ -459,7 +457,7 @@ struct rpc_clnt *rpc_bind_new_program(struct rpc_clnt *old, ...@@ -459,7 +457,7 @@ struct rpc_clnt *rpc_bind_new_program(struct rpc_clnt *old,
clnt->cl_prog = program->number; clnt->cl_prog = program->number;
clnt->cl_vers = version->number; clnt->cl_vers = version->number;
clnt->cl_stats = program->stats; clnt->cl_stats = program->stats;
err = rpc_ping(clnt, RPC_TASK_SOFT|RPC_TASK_NOINTR); err = rpc_ping(clnt, RPC_TASK_SOFT);
if (err != 0) { if (err != 0) {
rpc_shutdown_client(clnt); rpc_shutdown_client(clnt);
clnt = ERR_PTR(err); clnt = ERR_PTR(err);
...@@ -480,44 +478,6 @@ static const struct rpc_call_ops rpc_default_ops = { ...@@ -480,44 +478,6 @@ static const struct rpc_call_ops rpc_default_ops = {
.rpc_call_done = rpc_default_callback, .rpc_call_done = rpc_default_callback,
}; };
/*
* Export the signal mask handling for synchronous code that
* sleeps on RPC calls
*/
#define RPC_INTR_SIGNALS (sigmask(SIGHUP) | sigmask(SIGINT) | sigmask(SIGQUIT) | sigmask(SIGTERM))
static void rpc_save_sigmask(sigset_t *oldset, int intr)
{
unsigned long sigallow = sigmask(SIGKILL);
sigset_t sigmask;
/* Block all signals except those listed in sigallow */
if (intr)
sigallow |= RPC_INTR_SIGNALS;
siginitsetinv(&sigmask, sigallow);
sigprocmask(SIG_BLOCK, &sigmask, oldset);
}
static inline void rpc_task_sigmask(struct rpc_task *task, sigset_t *oldset)
{
rpc_save_sigmask(oldset, !RPC_TASK_UNINTERRUPTIBLE(task));
}
static inline void rpc_restore_sigmask(sigset_t *oldset)
{
sigprocmask(SIG_SETMASK, oldset, NULL);
}
void rpc_clnt_sigmask(struct rpc_clnt *clnt, sigset_t *oldset)
{
rpc_save_sigmask(oldset, clnt->cl_intr);
}
void rpc_clnt_sigunmask(struct rpc_clnt *clnt, sigset_t *oldset)
{
rpc_restore_sigmask(oldset);
}
static static
struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt, struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt,
struct rpc_message *msg, struct rpc_message *msg,
...@@ -526,7 +486,6 @@ struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt, ...@@ -526,7 +486,6 @@ struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt,
void *data) void *data)
{ {
struct rpc_task *task, *ret; struct rpc_task *task, *ret;
sigset_t oldset;
task = rpc_new_task(clnt, flags, ops, data); task = rpc_new_task(clnt, flags, ops, data);
if (task == NULL) { if (task == NULL) {
...@@ -535,7 +494,6 @@ struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt, ...@@ -535,7 +494,6 @@ struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt,
} }
/* Mask signals on synchronous RPC calls and RPCSEC_GSS upcalls */ /* Mask signals on synchronous RPC calls and RPCSEC_GSS upcalls */
rpc_task_sigmask(task, &oldset);
if (msg != NULL) { if (msg != NULL) {
rpc_call_setup(task, msg, 0); rpc_call_setup(task, msg, 0);
if (task->tk_status != 0) { if (task->tk_status != 0) {
...@@ -548,7 +506,6 @@ struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt, ...@@ -548,7 +506,6 @@ struct rpc_task *rpc_do_run_task(struct rpc_clnt *clnt,
rpc_execute(task); rpc_execute(task);
ret = task; ret = task;
out: out:
rpc_restore_sigmask(&oldset);
return ret; return ret;
} }
......
...@@ -172,8 +172,7 @@ static struct rpc_clnt *rpcb_create(char *hostname, struct sockaddr *srvaddr, ...@@ -172,8 +172,7 @@ static struct rpc_clnt *rpcb_create(char *hostname, struct sockaddr *srvaddr,
.program = &rpcb_program, .program = &rpcb_program,
.version = version, .version = version,
.authflavor = RPC_AUTH_UNIX, .authflavor = RPC_AUTH_UNIX,
.flags = (RPC_CLNT_CREATE_NOPING | .flags = RPC_CLNT_CREATE_NOPING,
RPC_CLNT_CREATE_INTR),
}; };
switch (srvaddr->sa_family) { switch (srvaddr->sa_family) {
......
...@@ -245,9 +245,9 @@ void rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname) ...@@ -245,9 +245,9 @@ void rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname)
} }
EXPORT_SYMBOL(rpc_init_wait_queue); EXPORT_SYMBOL(rpc_init_wait_queue);
static int rpc_wait_bit_interruptible(void *word) static int rpc_wait_bit_killable(void *word)
{ {
if (signal_pending(current)) if (fatal_signal_pending(current))
return -ERESTARTSYS; return -ERESTARTSYS;
schedule(); schedule();
return 0; return 0;
...@@ -299,9 +299,9 @@ static void rpc_mark_complete_task(struct rpc_task *task) ...@@ -299,9 +299,9 @@ static void rpc_mark_complete_task(struct rpc_task *task)
int __rpc_wait_for_completion_task(struct rpc_task *task, int (*action)(void *)) int __rpc_wait_for_completion_task(struct rpc_task *task, int (*action)(void *))
{ {
if (action == NULL) if (action == NULL)
action = rpc_wait_bit_interruptible; action = rpc_wait_bit_killable;
return wait_on_bit(&task->tk_runstate, RPC_TASK_ACTIVE, return wait_on_bit(&task->tk_runstate, RPC_TASK_ACTIVE,
action, TASK_INTERRUPTIBLE); action, TASK_KILLABLE);
} }
EXPORT_SYMBOL(__rpc_wait_for_completion_task); EXPORT_SYMBOL(__rpc_wait_for_completion_task);
...@@ -690,10 +690,9 @@ static void __rpc_execute(struct rpc_task *task) ...@@ -690,10 +690,9 @@ static void __rpc_execute(struct rpc_task *task)
/* sync task: sleep here */ /* sync task: sleep here */
dprintk("RPC: %5u sync task going to sleep\n", task->tk_pid); dprintk("RPC: %5u sync task going to sleep\n", task->tk_pid);
/* Note: Caller should be using rpc_clnt_sigmask() */
status = out_of_line_wait_on_bit(&task->tk_runstate, status = out_of_line_wait_on_bit(&task->tk_runstate,
RPC_TASK_QUEUED, rpc_wait_bit_interruptible, RPC_TASK_QUEUED, rpc_wait_bit_killable,
TASK_INTERRUPTIBLE); TASK_KILLABLE);
if (status == -ERESTARTSYS) { if (status == -ERESTARTSYS) {
/* /*
* When a sync task receives a signal, it exits with * When a sync task receives a signal, it exits with
...@@ -837,8 +836,6 @@ void rpc_init_task(struct rpc_task *task, struct rpc_clnt *clnt, int flags, cons ...@@ -837,8 +836,6 @@ void rpc_init_task(struct rpc_task *task, struct rpc_clnt *clnt, int flags, cons
kref_get(&clnt->cl_kref); kref_get(&clnt->cl_kref);
if (clnt->cl_softrtry) if (clnt->cl_softrtry)
task->tk_flags |= RPC_TASK_SOFT; task->tk_flags |= RPC_TASK_SOFT;
if (!clnt->cl_intr)
task->tk_flags |= RPC_TASK_NOINTR;
} }
BUG_ON(task->tk_ops == NULL); BUG_ON(task->tk_ops == NULL);
......
...@@ -38,8 +38,6 @@ EXPORT_SYMBOL(rpc_killall_tasks); ...@@ -38,8 +38,6 @@ EXPORT_SYMBOL(rpc_killall_tasks);
EXPORT_SYMBOL(rpc_call_sync); EXPORT_SYMBOL(rpc_call_sync);
EXPORT_SYMBOL(rpc_call_async); EXPORT_SYMBOL(rpc_call_async);
EXPORT_SYMBOL(rpc_call_setup); EXPORT_SYMBOL(rpc_call_setup);
EXPORT_SYMBOL(rpc_clnt_sigmask);
EXPORT_SYMBOL(rpc_clnt_sigunmask);
EXPORT_SYMBOL(rpc_delay); EXPORT_SYMBOL(rpc_delay);
EXPORT_SYMBOL(rpc_restart_call); EXPORT_SYMBOL(rpc_restart_call);
EXPORT_SYMBOL(rpc_setbufsize); EXPORT_SYMBOL(rpc_setbufsize);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment