Commit b62bf732 authored by Ingo Molnar's avatar Ingo Molnar Committed by Jens Axboe

[PATCH] sys_exit_group(), threading, 2.5.34

This is another step to have better threading support under Linux, it
implements the sys_exit_group() system call.

It's a straightforward extension of the generic 'thread group' concept,
which extension also comes handy to solve a number of problems when
implementing POSIX threads.

POSIX exit() [the C library function] has the following semantics: all
thread have to exit and the waiting parent has to get the exit code that
was specified for the exit() function.  It also has to be ensured that
every thread has truly finished its work by the time the parent gets the
notification.  The exit code has to be propagated properly to the parent
thread even if not the thread group leader calls the exit() function.

Normal single-thread exit is done via the pthread_exit() function, which
calls sys_exit().

Previous incarnations of Linux POSIX threads implementations chose the
following solution: send a 'thread management' signal to the thread
group leader via tkill(), which thread goes around and kills every
thread in the group (except itself), then calls sys_exit() with the
proper exit code.  Both old libpthreads and NGPT use this solution.

This works to a certain degree, unless a userspace threading library
uses the initial thread for normal thread work [like the new
libpthreads], which 'work' can cause the initial thread to exit
prematurely.

At this point the threading library has to catch the group leader in
pthread_exit() and has to keep the management thread 'hanging around'
artificially, waiting for the management signal. Besides being slightly
confusing to users ('why is this thread still around?') even this variant
is unrobust: if the initial thread is killed by the kernel (SIGSEGV or any
other thread-specific event that triggers do_exit()) then the thread goes
away without the thread library having a chance to intervene.

the sys_exit_group() syscall implements the mechanism within the kernel,
which, besides robustness, is also *much* faster. Instead of the threading
library having to tkill() every thread available, the kernel can use the
already existing 'broadcast signal' capability. (the threading library
cannot use broadcast signals because that would kill the initial thread as
well.)

as a side-effect of the completion mechanism used by sys_exit_group() it
was also possible to make the initial thread hang around as a zombie until
every other thread in the group has exited. A 'Z' state thread is much
easier to understand by users - it's around because it has to wait for all
other threads to exit first.

and as a side-effect of the initial thread hanging around in a guaranteed
way, there are three advantages:

 - signals sent to the thread group via sys_kill() work again. Previously
   if the initial thread exited then all subsequent sys_kill() calls to
   the group PID failed with a -ESRCH.

 - the get_pid() function got faster: it does not have to check for tgid
   collision anymore.

 - procps has an easier job displaying threaded applications - since the
   thread group leader is always around, no thread group can 'hide' from
   procps just because the thread group leader has exited.

 [ - NOTE: the same mechanism can/will also be used by the upcoming
     threaded-coredumps patch. ]

there's also another (small) advantage for threading libraries: eg. the
new libpthreads does not even have any notion of 'group of threads'
anymore - it does not maintain any global list of threads. Via this
syscall it can purely rely on the kernel to manage thread groups.

the patch itself does some internal changes to the way a thread exits: now
the unhashing of the PID and the signal-freeing is done atomically. This
is needed to make sure the thread group leader unhashes itself precisely
when the last thread group member has exited.

(the sys_exit_group() syscall has been used by glibc's new libpthreads
code for the past couple of weeks and the concept is working just fine.)
parent 4c21fddc
...@@ -759,6 +759,7 @@ ENTRY(sys_call_table) ...@@ -759,6 +759,7 @@ ENTRY(sys_call_table)
.long sys_io_getevents .long sys_io_getevents
.long sys_io_submit .long sys_io_submit
.long sys_io_cancel .long sys_io_cancel
.long sys_exit_group /* 250 */
.rept NR_syscalls-(.-sys_call_table)/4 .rept NR_syscalls-(.-sys_call_table)/4
.long sys_ni_syscall .long sys_ni_syscall
......
...@@ -513,6 +513,9 @@ static inline int make_private_signals(void) ...@@ -513,6 +513,9 @@ static inline int make_private_signals(void)
return -ENOMEM; return -ENOMEM;
spin_lock_init(&newsig->siglock); spin_lock_init(&newsig->siglock);
atomic_set(&newsig->count, 1); atomic_set(&newsig->count, 1);
newsig->group_exit = 0;
newsig->group_exit_code = 0;
init_completion(&newsig->group_exit_done);
memcpy(newsig->action, current->sig->action, sizeof(newsig->action)); memcpy(newsig->action, current->sig->action, sizeof(newsig->action));
init_sigpending(&newsig->shared_pending); init_sigpending(&newsig->shared_pending);
......
...@@ -27,6 +27,7 @@ extern unsigned long event; ...@@ -27,6 +27,7 @@ extern unsigned long event;
#include <linux/securebits.h> #include <linux/securebits.h>
#include <linux/fs_struct.h> #include <linux/fs_struct.h>
#include <linux/compiler.h> #include <linux/compiler.h>
#include <linux/completion.h>
struct exec_domain; struct exec_domain;
...@@ -128,8 +129,6 @@ struct sched_param { ...@@ -128,8 +129,6 @@ struct sched_param {
int sched_priority; int sched_priority;
}; };
struct completion;
#ifdef __KERNEL__ #ifdef __KERNEL__
#include <linux/spinlock.h> #include <linux/spinlock.h>
...@@ -216,6 +215,12 @@ struct signal_struct { ...@@ -216,6 +215,12 @@ struct signal_struct {
task_t *curr_target; task_t *curr_target;
struct sigpending shared_pending; struct sigpending shared_pending;
/* thread group exit support */
int group_exit;
int group_exit_code;
struct completion group_exit_done;
}; };
/* /*
...@@ -555,6 +560,7 @@ extern void notify_parent(struct task_struct *, int); ...@@ -555,6 +560,7 @@ extern void notify_parent(struct task_struct *, int);
extern void do_notify_parent(struct task_struct *, int); extern void do_notify_parent(struct task_struct *, int);
extern void force_sig(int, struct task_struct *); extern void force_sig(int, struct task_struct *);
extern int send_sig(int, struct task_struct *, int); extern int send_sig(int, struct task_struct *, int);
extern int __broadcast_thread_group(struct task_struct *p, int sig);
extern int kill_pg(pid_t, int, int); extern int kill_pg(pid_t, int, int);
extern int kill_sl(pid_t, int, int); extern int kill_sl(pid_t, int, int);
extern int kill_proc(pid_t, int, int); extern int kill_proc(pid_t, int, int);
...@@ -661,6 +667,7 @@ extern void exit_thread(void); ...@@ -661,6 +667,7 @@ extern void exit_thread(void);
extern void exit_mm(struct task_struct *); extern void exit_mm(struct task_struct *);
extern void exit_files(struct task_struct *); extern void exit_files(struct task_struct *);
extern void exit_sighand(struct task_struct *); extern void exit_sighand(struct task_struct *);
extern void __exit_sighand(struct task_struct *);
extern void remove_thread_group(struct task_struct *tsk, struct signal_struct *sig); extern void remove_thread_group(struct task_struct *tsk, struct signal_struct *sig);
extern void reparent_to_init(void); extern void reparent_to_init(void);
......
...@@ -29,10 +29,9 @@ extern struct task_struct *child_reaper; ...@@ -29,10 +29,9 @@ extern struct task_struct *child_reaper;
int getrusage(struct task_struct *, int, struct rusage *); int getrusage(struct task_struct *, int, struct rusage *);
static inline void __unhash_process(struct task_struct *p) static struct dentry * __unhash_process(struct task_struct *p)
{ {
struct dentry *proc_dentry; struct dentry *proc_dentry;
write_lock_irq(&tasklist_lock);
nr_threads--; nr_threads--;
unhash_pid(p); unhash_pid(p);
REMOVE_LINKS(p); REMOVE_LINKS(p);
...@@ -47,15 +46,13 @@ static inline void __unhash_process(struct task_struct *p) ...@@ -47,15 +46,13 @@ static inline void __unhash_process(struct task_struct *p)
proc_dentry = NULL; proc_dentry = NULL;
spin_unlock(&dcache_lock); spin_unlock(&dcache_lock);
} }
write_unlock_irq(&tasklist_lock); return proc_dentry;
if (unlikely(proc_dentry != NULL)) {
shrink_dcache_parent(proc_dentry);
dput(proc_dentry);
}
} }
static void release_task(struct task_struct * p) static void release_task(struct task_struct * p)
{ {
struct dentry *proc_dentry;
if (p->state != TASK_ZOMBIE) if (p->state != TASK_ZOMBIE)
BUG(); BUG();
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
...@@ -71,8 +68,14 @@ static void release_task(struct task_struct * p) ...@@ -71,8 +68,14 @@ static void release_task(struct task_struct * p)
write_unlock_irq(&tasklist_lock); write_unlock_irq(&tasklist_lock);
} }
BUG_ON(!list_empty(&p->ptrace_list) || !list_empty(&p->ptrace_children)); BUG_ON(!list_empty(&p->ptrace_list) || !list_empty(&p->ptrace_children));
unhash_process(p); write_lock_irq(&tasklist_lock);
exit_sighand(p); __exit_sighand(p);
proc_dentry = __unhash_process(p);
write_unlock_irq(&tasklist_lock);
if (unlikely(proc_dentry != NULL)) {
shrink_dcache_parent(proc_dentry);
dput(proc_dentry);
}
release_thread(p); release_thread(p);
if (p != current) { if (p != current) {
...@@ -88,7 +91,16 @@ static void release_task(struct task_struct * p) ...@@ -88,7 +91,16 @@ static void release_task(struct task_struct * p)
void unhash_process(struct task_struct *p) void unhash_process(struct task_struct *p)
{ {
return __unhash_process(p); struct dentry *proc_dentry;
write_lock_irq(&tasklist_lock);
proc_dentry = __unhash_process(p);
write_unlock_irq(&tasklist_lock);
if (unlikely(proc_dentry != NULL)) {
shrink_dcache_parent(proc_dentry);
dput(proc_dentry);
}
} }
/* /*
...@@ -653,6 +665,30 @@ asmlinkage long sys_exit(int error_code) ...@@ -653,6 +665,30 @@ asmlinkage long sys_exit(int error_code)
do_exit((error_code&0xff)<<8); do_exit((error_code&0xff)<<8);
} }
/*
* this kills every thread in the thread group. Note that any externally
* wait4()-ing process will get the correct exit code - even if this
* thread is not the thread group leader.
*/
asmlinkage long sys_exit_group(int error_code)
{
struct signal_struct *sig = current->sig;
spin_lock_irq(&sig->siglock);
if (sig->group_exit) {
spin_unlock_irq(&sig->siglock);
/* another thread was faster: */
do_exit(sig->group_exit_code);
}
sig->group_exit = 1;
sig->group_exit_code = (error_code & 0xff) << 8;
__broadcast_thread_group(current, SIGKILL);
spin_unlock_irq(&sig->siglock);
do_exit(sig->group_exit_code);
}
static inline int eligible_child(pid_t pid, int options, task_t *p) static inline int eligible_child(pid_t pid, int options, task_t *p)
{ {
if (pid>0) { if (pid>0) {
......
...@@ -181,7 +181,6 @@ static int get_pid(unsigned long flags) ...@@ -181,7 +181,6 @@ static int get_pid(unsigned long flags)
for_each_task(p) { for_each_task(p) {
if (p->pid == last_pid || if (p->pid == last_pid ||
p->pgrp == last_pid || p->pgrp == last_pid ||
p->tgid == last_pid ||
p->session == last_pid) { p->session == last_pid) {
if (++last_pid >= next_safe) { if (++last_pid >= next_safe) {
if (last_pid >= pid_max) if (last_pid >= pid_max)
...@@ -194,8 +193,6 @@ static int get_pid(unsigned long flags) ...@@ -194,8 +193,6 @@ static int get_pid(unsigned long flags)
next_safe = p->pid; next_safe = p->pid;
if (p->pgrp > last_pid && next_safe > p->pgrp) if (p->pgrp > last_pid && next_safe > p->pgrp)
next_safe = p->pgrp; next_safe = p->pgrp;
if (p->tgid > last_pid && next_safe > p->tgid)
next_safe = p->tgid;
if (p->session > last_pid && next_safe > p->session) if (p->session > last_pid && next_safe > p->session)
next_safe = p->session; next_safe = p->session;
} }
...@@ -629,7 +626,10 @@ static inline int copy_sighand(unsigned long clone_flags, struct task_struct * t ...@@ -629,7 +626,10 @@ static inline int copy_sighand(unsigned long clone_flags, struct task_struct * t
return -1; return -1;
spin_lock_init(&sig->siglock); spin_lock_init(&sig->siglock);
atomic_set(&sig->count, 1); atomic_set(&sig->count, 1);
memcpy(tsk->sig->action, current->sig->action, sizeof(tsk->sig->action)); sig->group_exit = 0;
sig->group_exit_code = 0;
init_completion(&sig->group_exit_done);
memcpy(sig->action, current->sig->action, sizeof(sig->action));
sig->curr_target = NULL; sig->curr_target = NULL;
init_sigpending(&sig->shared_pending); init_sigpending(&sig->shared_pending);
...@@ -853,6 +853,16 @@ static struct task_struct *copy_process(unsigned long clone_flags, ...@@ -853,6 +853,16 @@ static struct task_struct *copy_process(unsigned long clone_flags,
if (clone_flags & CLONE_THREAD) { if (clone_flags & CLONE_THREAD) {
spin_lock(&current->sig->siglock); spin_lock(&current->sig->siglock);
/*
* Important: if an exit-all has been started then
* do not create this new thread - the whole thread
* group is supposed to exit anyway.
*/
if (current->sig->group_exit) {
spin_unlock(&current->sig->siglock);
write_unlock_irq(&tasklist_lock);
goto bad_fork_cleanup_namespace;
}
p->tgid = current->tgid; p->tgid = current->tgid;
list_add(&p->thread_group, &current->thread_group); list_add(&p->thread_group, &current->thread_group);
spin_unlock(&current->sig->siglock); spin_unlock(&current->sig->siglock);
......
...@@ -221,20 +221,28 @@ flush_signals(struct task_struct *t) ...@@ -221,20 +221,28 @@ flush_signals(struct task_struct *t)
flush_sigqueue(&t->pending); flush_sigqueue(&t->pending);
} }
static inline void __remove_thread_group(struct task_struct *tsk, struct signal_struct *sig)
{
if (tsk == sig->curr_target)
sig->curr_target = next_thread(tsk);
list_del_init(&tsk->thread_group);
}
void remove_thread_group(struct task_struct *tsk, struct signal_struct *sig) void remove_thread_group(struct task_struct *tsk, struct signal_struct *sig)
{ {
write_lock_irq(&tasklist_lock); write_lock_irq(&tasklist_lock);
spin_lock(&tsk->sig->siglock); spin_lock(&tsk->sig->siglock);
if (tsk == sig->curr_target) __remove_thread_group(tsk, sig);
sig->curr_target = next_thread(tsk);
list_del_init(&tsk->thread_group);
spin_unlock(&tsk->sig->siglock); spin_unlock(&tsk->sig->siglock);
write_unlock_irq(&tasklist_lock); write_unlock_irq(&tasklist_lock);
} }
void exit_sighand(struct task_struct *tsk) /*
* This function expects the tasklist_lock write-locked.
*/
void __exit_sighand(struct task_struct *tsk)
{ {
struct signal_struct * sig = tsk->sig; struct signal_struct * sig = tsk->sig;
...@@ -242,19 +250,45 @@ void exit_sighand(struct task_struct *tsk) ...@@ -242,19 +250,45 @@ void exit_sighand(struct task_struct *tsk)
BUG(); BUG();
if (!atomic_read(&sig->count)) if (!atomic_read(&sig->count))
BUG(); BUG();
remove_thread_group(tsk, sig); spin_lock(&sig->siglock);
/*
* Do not let the thread group leader exit until all other
* threads are done:
*/
while (current->tgid == current->pid && atomic_read(&sig->count) > 1) {
spin_unlock(&sig->siglock);
write_unlock_irq(&tasklist_lock);
spin_lock_irq(&tsk->sigmask_lock); wait_for_completion(&sig->group_exit_done);
if (sig) {
tsk->sig = NULL; write_lock_irq(&tasklist_lock);
if (atomic_dec_and_test(&sig->count)) { spin_lock(&sig->siglock);
flush_sigqueue(&sig->shared_pending); }
kmem_cache_free(sigact_cachep, sig);
} __remove_thread_group(tsk, sig);
spin_lock(&tsk->sigmask_lock);
tsk->sig = NULL;
if (atomic_dec_and_test(&sig->count)) {
spin_unlock(&sig->siglock);
flush_sigqueue(&sig->shared_pending);
kmem_cache_free(sigact_cachep, sig);
} else {
if (atomic_read(&sig->count) == 1)
complete(&sig->group_exit_done);
spin_unlock(&sig->siglock);
} }
clear_tsk_thread_flag(tsk,TIF_SIGPENDING); clear_tsk_thread_flag(tsk,TIF_SIGPENDING);
flush_sigqueue(&tsk->pending); flush_sigqueue(&tsk->pending);
spin_unlock_irq(&tsk->sigmask_lock);
spin_unlock(&tsk->sigmask_lock);
}
void exit_sighand(struct task_struct *tsk)
{
write_lock_irq(&tasklist_lock);
__exit_sighand(tsk);
write_unlock_irq(&tasklist_lock);
} }
/* /*
...@@ -286,6 +320,9 @@ sig_exit(int sig, int exit_code, struct siginfo *info) ...@@ -286,6 +320,9 @@ sig_exit(int sig, int exit_code, struct siginfo *info)
recalc_sigpending(); recalc_sigpending();
current->flags |= PF_SIGNALED; current->flags |= PF_SIGNALED;
if (current->sig->group_exit)
exit_code = current->sig->group_exit_code;
do_exit(exit_code); do_exit(exit_code);
/* NOTREACHED */ /* NOTREACHED */
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment