Commit be833f33 authored by Thomas Gleixner's avatar Thomas Gleixner Committed by Benjamin Herrenschmidt

powerpc: Convert context_lock to raw_spinlock

context_lock needs to be a real spinlock in RT. Convert it to
raw_spinlock.
Signed-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
Signed-off-by: default avatarBenjamin Herrenschmidt <benh@kernel.crashing.org>
parent 87d31345
...@@ -56,7 +56,7 @@ static unsigned int next_context, nr_free_contexts; ...@@ -56,7 +56,7 @@ static unsigned int next_context, nr_free_contexts;
static unsigned long *context_map; static unsigned long *context_map;
static unsigned long *stale_map[NR_CPUS]; static unsigned long *stale_map[NR_CPUS];
static struct mm_struct **context_mm; static struct mm_struct **context_mm;
static DEFINE_SPINLOCK(context_lock); static DEFINE_RAW_SPINLOCK(context_lock);
#define CTX_MAP_SIZE \ #define CTX_MAP_SIZE \
(sizeof(unsigned long) * (last_context / BITS_PER_LONG + 1)) (sizeof(unsigned long) * (last_context / BITS_PER_LONG + 1))
...@@ -121,9 +121,9 @@ static unsigned int steal_context_smp(unsigned int id) ...@@ -121,9 +121,9 @@ static unsigned int steal_context_smp(unsigned int id)
/* This will happen if you have more CPUs than available contexts, /* This will happen if you have more CPUs than available contexts,
* all we can do here is wait a bit and try again * all we can do here is wait a bit and try again
*/ */
spin_unlock(&context_lock); raw_spin_unlock(&context_lock);
cpu_relax(); cpu_relax();
spin_lock(&context_lock); raw_spin_lock(&context_lock);
/* This will cause the caller to try again */ /* This will cause the caller to try again */
return MMU_NO_CONTEXT; return MMU_NO_CONTEXT;
...@@ -194,7 +194,7 @@ void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) ...@@ -194,7 +194,7 @@ void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next)
unsigned long *map; unsigned long *map;
/* No lockless fast path .. yet */ /* No lockless fast path .. yet */
spin_lock(&context_lock); raw_spin_lock(&context_lock);
pr_hard("[%d] activating context for mm @%p, active=%d, id=%d", pr_hard("[%d] activating context for mm @%p, active=%d, id=%d",
cpu, next, next->context.active, next->context.id); cpu, next, next->context.active, next->context.id);
...@@ -278,7 +278,7 @@ void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) ...@@ -278,7 +278,7 @@ void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next)
/* Flick the MMU and release lock */ /* Flick the MMU and release lock */
pr_hardcont(" -> %d\n", id); pr_hardcont(" -> %d\n", id);
set_context(id, next->pgd); set_context(id, next->pgd);
spin_unlock(&context_lock); raw_spin_unlock(&context_lock);
} }
/* /*
...@@ -307,7 +307,7 @@ void destroy_context(struct mm_struct *mm) ...@@ -307,7 +307,7 @@ void destroy_context(struct mm_struct *mm)
WARN_ON(mm->context.active != 0); WARN_ON(mm->context.active != 0);
spin_lock_irqsave(&context_lock, flags); raw_spin_lock_irqsave(&context_lock, flags);
id = mm->context.id; id = mm->context.id;
if (id != MMU_NO_CONTEXT) { if (id != MMU_NO_CONTEXT) {
__clear_bit(id, context_map); __clear_bit(id, context_map);
...@@ -318,7 +318,7 @@ void destroy_context(struct mm_struct *mm) ...@@ -318,7 +318,7 @@ void destroy_context(struct mm_struct *mm)
context_mm[id] = NULL; context_mm[id] = NULL;
nr_free_contexts++; nr_free_contexts++;
} }
spin_unlock_irqrestore(&context_lock, flags); raw_spin_unlock_irqrestore(&context_lock, flags);
} }
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment