Commit f2f1b44c authored by Zwane Mwaikambo's avatar Zwane Mwaikambo Committed by Linus Torvalds

[PATCH] Remove RCU abuse in cpu_idle()

Introduce cpu_idle_wait() on architectures requiring modification of
pm_idle from modules, this will ensure that all processors have updated
their cached values of pm_idle upon exit.  This patch is to address the bug
report at http://bugme.osdl.org/show_bug.cgi?id=1716 and replaces the
current code fix which is in violation of normal RCU usage as pointed out
by Stephen, Dipankar and Paul.
Signed-off-by: default avatarZwane Mwaikambo <zwane@arm.linux.org.uk>
Signed-off-by: default avatarAndrew Morton <akpm@osdl.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent a61e7286
...@@ -2369,7 +2369,7 @@ static void __exit apm_exit(void) ...@@ -2369,7 +2369,7 @@ static void __exit apm_exit(void)
* (pm_idle), Wait for all processors to update cached/local * (pm_idle), Wait for all processors to update cached/local
* copies of pm_idle before proceeding. * copies of pm_idle before proceeding.
*/ */
synchronize_kernel(); cpu_idle_wait();
} }
if (((apm_info.bios.flags & APM_BIOS_DISENGAGED) == 0) if (((apm_info.bios.flags & APM_BIOS_DISENGAGED) == 0)
&& (apm_info.connection_version > 0x0100)) { && (apm_info.connection_version > 0x0100)) {
......
...@@ -72,6 +72,7 @@ unsigned long thread_saved_pc(struct task_struct *tsk) ...@@ -72,6 +72,7 @@ unsigned long thread_saved_pc(struct task_struct *tsk)
* Powermanagement idle function, if any.. * Powermanagement idle function, if any..
*/ */
void (*pm_idle)(void); void (*pm_idle)(void);
static cpumask_t cpu_idle_map;
void disable_hlt(void) void disable_hlt(void)
{ {
...@@ -93,7 +94,7 @@ EXPORT_SYMBOL(enable_hlt); ...@@ -93,7 +94,7 @@ EXPORT_SYMBOL(enable_hlt);
*/ */
void default_idle(void) void default_idle(void)
{ {
if (!hlt_counter && current_cpu_data.hlt_works_ok) { if (!hlt_counter && boot_cpu_data.hlt_works_ok) {
local_irq_disable(); local_irq_disable();
if (!need_resched()) if (!need_resched())
safe_halt(); safe_halt();
...@@ -144,29 +145,44 @@ static void poll_idle (void) ...@@ -144,29 +145,44 @@ static void poll_idle (void)
*/ */
void cpu_idle (void) void cpu_idle (void)
{ {
int cpu = smp_processor_id();
/* endless idle loop with no priority at all */ /* endless idle loop with no priority at all */
while (1) { while (1) {
while (!need_resched()) { while (!need_resched()) {
void (*idle)(void); void (*idle)(void);
/*
* Mark this as an RCU critical section so that if (cpu_isset(cpu, cpu_idle_map))
* synchronize_kernel() in the unload path waits cpu_clear(cpu, cpu_idle_map);
* for our completion. rmb();
*/
rcu_read_lock();
idle = pm_idle; idle = pm_idle;
if (!idle) if (!idle)
idle = default_idle; idle = default_idle;
irq_stat[smp_processor_id()].idle_timestamp = jiffies; irq_stat[cpu].idle_timestamp = jiffies;
idle(); idle();
rcu_read_unlock();
} }
schedule(); schedule();
} }
} }
void cpu_idle_wait(void)
{
int cpu;
cpumask_t map;
for_each_online_cpu(cpu)
cpu_set(cpu, cpu_idle_map);
wmb();
do {
ssleep(1);
cpus_and(map, cpu_idle_map, cpu_online_map);
} while (!cpus_empty(map));
}
EXPORT_SYMBOL_GPL(cpu_idle_wait);
/* /*
* This uses new MONITOR/MWAIT instructions on P4 processors with PNI, * This uses new MONITOR/MWAIT instructions on P4 processors with PNI,
* which can obviate IPI to trigger checking of need_resched. * which can obviate IPI to trigger checking of need_resched.
......
...@@ -25,6 +25,7 @@ ...@@ -25,6 +25,7 @@
#include <linux/unistd.h> #include <linux/unistd.h>
#include <linux/efi.h> #include <linux/efi.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/delay.h>
#include <asm/cpu.h> #include <asm/cpu.h>
#include <asm/delay.h> #include <asm/delay.h>
...@@ -46,6 +47,7 @@ ...@@ -46,6 +47,7 @@
#include "sigframe.h" #include "sigframe.h"
void (*ia64_mark_idle)(int); void (*ia64_mark_idle)(int);
static cpumask_t cpu_idle_map;
unsigned long boot_option_idle_override = 0; unsigned long boot_option_idle_override = 0;
EXPORT_SYMBOL(boot_option_idle_override); EXPORT_SYMBOL(boot_option_idle_override);
...@@ -225,10 +227,28 @@ static inline void play_dead(void) ...@@ -225,10 +227,28 @@ static inline void play_dead(void)
} }
#endif /* CONFIG_HOTPLUG_CPU */ #endif /* CONFIG_HOTPLUG_CPU */
void cpu_idle_wait(void)
{
int cpu;
cpumask_t map;
for_each_online_cpu(cpu)
cpu_set(cpu, cpu_idle_map);
wmb();
do {
ssleep(1);
cpus_and(map, cpu_idle_map, cpu_online_map);
} while (!cpus_empty(map));
}
EXPORT_SYMBOL_GPL(cpu_idle_wait);
void __attribute__((noreturn)) void __attribute__((noreturn))
cpu_idle (void *unused) cpu_idle (void *unused)
{ {
void (*mark_idle)(int) = ia64_mark_idle; void (*mark_idle)(int) = ia64_mark_idle;
int cpu = smp_processor_id();
/* endless idle loop with no priority at all */ /* endless idle loop with no priority at all */
while (1) { while (1) {
...@@ -241,17 +261,14 @@ cpu_idle (void *unused) ...@@ -241,17 +261,14 @@ cpu_idle (void *unused)
if (mark_idle) if (mark_idle)
(*mark_idle)(1); (*mark_idle)(1);
/*
* Mark this as an RCU critical section so that if (cpu_isset(cpu, cpu_idle_map))
* synchronize_kernel() in the unload path waits cpu_clear(cpu, cpu_idle_map);
* for our completion. rmb();
*/
rcu_read_lock();
idle = pm_idle; idle = pm_idle;
if (!idle) if (!idle)
idle = default_idle; idle = default_idle;
(*idle)(); (*idle)();
rcu_read_unlock();
} }
if (mark_idle) if (mark_idle)
......
...@@ -61,6 +61,7 @@ EXPORT_SYMBOL(boot_option_idle_override); ...@@ -61,6 +61,7 @@ EXPORT_SYMBOL(boot_option_idle_override);
* Powermanagement idle function, if any.. * Powermanagement idle function, if any..
*/ */
void (*pm_idle)(void); void (*pm_idle)(void);
static cpumask_t cpu_idle_map;
void disable_hlt(void) void disable_hlt(void)
{ {
...@@ -123,6 +124,23 @@ static void poll_idle (void) ...@@ -123,6 +124,23 @@ static void poll_idle (void)
} }
} }
void cpu_idle_wait(void)
{
int cpu;
cpumask_t map;
for_each_online_cpu(cpu)
cpu_set(cpu, cpu_idle_map);
wmb();
do {
ssleep(1);
cpus_and(map, cpu_idle_map, cpu_online_map);
} while (!cpus_empty(map));
}
EXPORT_SYMBOL_GPL(cpu_idle_wait);
/* /*
* The idle thread. There's no useful work to be * The idle thread. There's no useful work to be
* done, so just try to conserve power and have a * done, so just try to conserve power and have a
...@@ -131,21 +149,20 @@ static void poll_idle (void) ...@@ -131,21 +149,20 @@ static void poll_idle (void)
*/ */
void cpu_idle (void) void cpu_idle (void)
{ {
int cpu = smp_processor_id();
/* endless idle loop with no priority at all */ /* endless idle loop with no priority at all */
while (1) { while (1) {
while (!need_resched()) { while (!need_resched()) {
void (*idle)(void); void (*idle)(void);
/*
* Mark this as an RCU critical section so that if (cpu_isset(cpu, cpu_idle_map))
* synchronize_kernel() in the unload path waits cpu_clear(cpu, cpu_idle_map);
* for our completion. rmb();
*/
rcu_read_lock();
idle = pm_idle; idle = pm_idle;
if (!idle) if (!idle)
idle = default_idle; idle = default_idle;
idle(); idle();
rcu_read_unlock();
} }
schedule(); schedule();
} }
......
...@@ -988,7 +988,7 @@ int acpi_processor_power_exit(struct acpi_processor *pr, struct acpi_device *dev ...@@ -988,7 +988,7 @@ int acpi_processor_power_exit(struct acpi_processor *pr, struct acpi_device *dev
* (pm_idle), Wait for all processors to update cached/local * (pm_idle), Wait for all processors to update cached/local
* copies of pm_idle before proceeding. * copies of pm_idle before proceeding.
*/ */
synchronize_kernel(); cpu_idle_wait();
} }
return_VALUE(0); return_VALUE(0);
......
...@@ -466,5 +466,6 @@ void disable_hlt(void); ...@@ -466,5 +466,6 @@ void disable_hlt(void);
void enable_hlt(void); void enable_hlt(void);
extern int es7000_plat; extern int es7000_plat;
void cpu_idle_wait(void);
#endif #endif
...@@ -284,6 +284,7 @@ do { \ ...@@ -284,6 +284,7 @@ do { \
#define ia64_platform_is(x) (strcmp(x, platform_name) == 0) #define ia64_platform_is(x) (strcmp(x, platform_name) == 0)
void cpu_idle_wait(void);
#endif /* __KERNEL__ */ #endif /* __KERNEL__ */
#endif /* __ASSEMBLY__ */ #endif /* __ASSEMBLY__ */
......
...@@ -326,6 +326,8 @@ static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, ...@@ -326,6 +326,8 @@ static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old,
/* For spinlocks etc */ /* For spinlocks etc */
#define local_irq_save(x) do { warn_if_not_ulong(x); __asm__ __volatile__("# local_irq_save \n\t pushfq ; popq %0 ; cli":"=g" (x): /* no input */ :"memory"); } while (0) #define local_irq_save(x) do { warn_if_not_ulong(x); __asm__ __volatile__("# local_irq_save \n\t pushfq ; popq %0 ; cli":"=g" (x): /* no input */ :"memory"); } while (0)
void cpu_idle_wait(void);
/* /*
* disable hlt during certain critical i/o operations * disable hlt during certain critical i/o operations
*/ */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment