From 035406972c5e9ae212deea49cd2c4d7e369210a8 Mon Sep 17 00:00:00 2001 From: Rusty Russell <rusty@rustcorp.com.au> Date: Fri, 13 Jun 2003 21:57:53 -0700 Subject: [PATCH] [PATCH] sched.c neatening and fixes. 1) Fix the comments for the migration_thread. A while back Ingo agreed they were exactly wrong, IIRC. 8). 2) Changed spin_lock_irqsave to spin_lock_irq, since it's in a kernel thread. 3) Don't repeat if the task has moved off the original CPU, just finish. This is because we are simply trying to push the task off this CPU: if it's already moved, great. Currently we might theoretically move a task which is actually running on another CPU, which is v. bad. 4) Replace the __ffs(p->cpus_allowed) with any_online_cpu(), since that's what it's for, and __ffs() can give the wrong answer, eg. if there's no CPU 0. 5) Move the core functionality of migrate_task into a separate function, move_task_away, which I want for the hotplug CPU patch. --- kernel/sched.c | 73 +++++++++++++++++++++++++------------------------- 1 file changed, 36 insertions(+), 37 deletions(-) diff --git a/kernel/sched.c b/kernel/sched.c index 79c8794c1769..dead0e9dcad5 100644 --- a/kernel/sched.c +++ b/kernel/sched.c @@ -2296,7 +2296,7 @@ void set_cpus_allowed(task_t *p, unsigned long new_mask) * it is sufficient to simply update the task's cpu field. */ if (!p->array && !task_running(rq, p)) { - set_task_cpu(p, __ffs(p->cpus_allowed)); + set_task_cpu(p, any_online_cpu(p->cpus_allowed)); task_rq_unlock(rq, &flags); return; } @@ -2310,9 +2310,35 @@ void set_cpus_allowed(task_t *p, unsigned long new_mask) wait_for_completion(&req.done); } +/* Move (not current) task off this cpu, onto dest cpu. */ +static void move_task_away(struct task_struct *p, int dest_cpu) +{ + runqueue_t *rq_dest; + unsigned long flags; + + rq_dest = cpu_rq(dest_cpu); + + local_irq_save(flags); + double_rq_lock(this_rq(), rq_dest); + if (task_cpu(p) != smp_processor_id()) + goto out; /* Already moved */ + + set_task_cpu(p, dest_cpu); + if (p->array) { + deactivate_task(p, this_rq()); + activate_task(p, rq_dest); + if (p->prio < rq_dest->curr->prio) + resched_task(rq_dest->curr); + } + out: + double_rq_unlock(this_rq(), rq_dest); + local_irq_restore(flags); +} + /* * migration_thread - this is a highprio system thread that performs - * thread migration by 'pulling' threads into the target runqueue. + * thread migration by bumping thread off CPU then 'pushing' onto + * another runqueue. */ static int migration_thread(void * data) { @@ -2326,8 +2352,9 @@ static int migration_thread(void * data) set_fs(KERNEL_DS); /* - * Either we are running on the right CPU, or there's a - * a migration thread on the target CPU, guaranteed. + * Either we are running on the right CPU, or there's a a + * migration thread on this CPU, guaranteed (we're started + * serially). */ set_cpus_allowed(current, 1UL << cpu); @@ -2337,51 +2364,23 @@ static int migration_thread(void * data) rq->migration_thread = current; for (;;) { - runqueue_t *rq_src, *rq_dest; struct list_head *head; - int cpu_src, cpu_dest; migration_req_t *req; - unsigned long flags; - task_t *p; - spin_lock_irqsave(&rq->lock, flags); + spin_lock_irq(&rq->lock); head = &rq->migration_queue; current->state = TASK_INTERRUPTIBLE; if (list_empty(head)) { - spin_unlock_irqrestore(&rq->lock, flags); + spin_unlock_irq(&rq->lock); schedule(); continue; } req = list_entry(head->next, migration_req_t, list); list_del_init(head->next); - spin_unlock_irqrestore(&rq->lock, flags); - - p = req->task; - cpu_dest = __ffs(p->cpus_allowed & cpu_online_map); - rq_dest = cpu_rq(cpu_dest); -repeat: - cpu_src = task_cpu(p); - rq_src = cpu_rq(cpu_src); - - local_irq_save(flags); - double_rq_lock(rq_src, rq_dest); - if (task_cpu(p) != cpu_src) { - double_rq_unlock(rq_src, rq_dest); - local_irq_restore(flags); - goto repeat; - } - if (rq_src == rq) { - set_task_cpu(p, cpu_dest); - if (p->array) { - deactivate_task(p, rq_src); - __activate_task(p, rq_dest); - if (p->prio < rq_dest->curr->prio) - resched_task(rq_dest->curr); - } - } - double_rq_unlock(rq_src, rq_dest); - local_irq_restore(flags); + spin_unlock_irq(&rq->lock); + move_task_away(req->task, + any_online_cpu(req->task->cpus_allowed)); complete(&req->done); } } -- 2.30.9