Commit 6a3ee3d5 authored by Jeremy Fitzhardinge's avatar Jeremy Fitzhardinge Committed by Linus Torvalds

i386: fix voyager build

This adds an smp_ops for voyager, and hooks things up appropriately.  This is
the first baby-step to making subarch runtime switchable.
Signed-off-by: default avatarJeremy Fitzhardinge <jeremy@xensource.com>
Cc: James Bottomley <James.Bottomley@HansenPartnership.com>
Cc: Eric W. Biederman <ebiederm@xmission.com>
Cc: Andi Kleen <ak@suse.de>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 297d9c03
...@@ -27,7 +27,6 @@ ...@@ -27,7 +27,6 @@
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
#include <asm/tlbflush.h> #include <asm/tlbflush.h>
#include <asm/arch_hooks.h> #include <asm/arch_hooks.h>
#include <asm/pda.h>
/* TLB state -- visible externally, indexed physically */ /* TLB state -- visible externally, indexed physically */
DEFINE_PER_CPU(struct tlb_state, cpu_tlbstate) ____cacheline_aligned = { &init_mm, 0 }; DEFINE_PER_CPU(struct tlb_state, cpu_tlbstate) ____cacheline_aligned = { &init_mm, 0 };
...@@ -422,7 +421,7 @@ find_smp_config(void) ...@@ -422,7 +421,7 @@ find_smp_config(void)
VOYAGER_SUS_IN_CONTROL_PORT); VOYAGER_SUS_IN_CONTROL_PORT);
current_thread_info()->cpu = boot_cpu_id; current_thread_info()->cpu = boot_cpu_id;
write_pda(cpu_number, boot_cpu_id); x86_write_percpu(cpu_number, boot_cpu_id);
} }
/* /*
...@@ -435,7 +434,7 @@ smp_store_cpu_info(int id) ...@@ -435,7 +434,7 @@ smp_store_cpu_info(int id)
*c = boot_cpu_data; *c = boot_cpu_data;
identify_cpu(c); identify_secondary_cpu(c);
} }
/* set up the trampoline and return the physical address of the code */ /* set up the trampoline and return the physical address of the code */
...@@ -459,7 +458,7 @@ start_secondary(void *unused) ...@@ -459,7 +458,7 @@ start_secondary(void *unused)
/* external functions not defined in the headers */ /* external functions not defined in the headers */
extern void calibrate_delay(void); extern void calibrate_delay(void);
secondary_cpu_init(); cpu_init();
/* OK, we're in the routine */ /* OK, we're in the routine */
ack_CPI(VIC_CPU_BOOT_CPI); ack_CPI(VIC_CPU_BOOT_CPI);
...@@ -572,7 +571,9 @@ do_boot_cpu(__u8 cpu) ...@@ -572,7 +571,9 @@ do_boot_cpu(__u8 cpu)
/* init_tasks (in sched.c) is indexed logically */ /* init_tasks (in sched.c) is indexed logically */
stack_start.esp = (void *) idle->thread.esp; stack_start.esp = (void *) idle->thread.esp;
init_gdt(cpu, idle); init_gdt(cpu);
per_cpu(current_task, cpu) = idle;
early_gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu);
irq_ctx_init(cpu); irq_ctx_init(cpu);
/* Note: Don't modify initial ss override */ /* Note: Don't modify initial ss override */
...@@ -859,8 +860,8 @@ smp_invalidate_interrupt(void) ...@@ -859,8 +860,8 @@ smp_invalidate_interrupt(void)
/* This routine is called with a physical cpu mask */ /* This routine is called with a physical cpu mask */
static void static void
flush_tlb_others (unsigned long cpumask, struct mm_struct *mm, voyager_flush_tlb_others (unsigned long cpumask, struct mm_struct *mm,
unsigned long va) unsigned long va)
{ {
int stuck = 50000; int stuck = 50000;
...@@ -912,7 +913,7 @@ flush_tlb_current_task(void) ...@@ -912,7 +913,7 @@ flush_tlb_current_task(void)
cpu_mask = cpus_addr(mm->cpu_vm_mask)[0] & ~(1 << smp_processor_id()); cpu_mask = cpus_addr(mm->cpu_vm_mask)[0] & ~(1 << smp_processor_id());
local_flush_tlb(); local_flush_tlb();
if (cpu_mask) if (cpu_mask)
flush_tlb_others(cpu_mask, mm, FLUSH_ALL); voyager_flush_tlb_others(cpu_mask, mm, FLUSH_ALL);
preempt_enable(); preempt_enable();
} }
...@@ -934,7 +935,7 @@ flush_tlb_mm (struct mm_struct * mm) ...@@ -934,7 +935,7 @@ flush_tlb_mm (struct mm_struct * mm)
leave_mm(smp_processor_id()); leave_mm(smp_processor_id());
} }
if (cpu_mask) if (cpu_mask)
flush_tlb_others(cpu_mask, mm, FLUSH_ALL); voyager_flush_tlb_others(cpu_mask, mm, FLUSH_ALL);
preempt_enable(); preempt_enable();
} }
...@@ -955,7 +956,7 @@ void flush_tlb_page(struct vm_area_struct * vma, unsigned long va) ...@@ -955,7 +956,7 @@ void flush_tlb_page(struct vm_area_struct * vma, unsigned long va)
} }
if (cpu_mask) if (cpu_mask)
flush_tlb_others(cpu_mask, mm, va); voyager_flush_tlb_others(cpu_mask, mm, va);
preempt_enable(); preempt_enable();
} }
...@@ -1044,10 +1045,12 @@ smp_call_function_interrupt(void) ...@@ -1044,10 +1045,12 @@ smp_call_function_interrupt(void)
} }
static int static int
__smp_call_function_mask (void (*func) (void *info), void *info, int retry, voyager_smp_call_function_mask (cpumask_t cpumask,
int wait, __u32 mask) void (*func) (void *info), void *info,
int wait)
{ {
struct call_data_struct data; struct call_data_struct data;
u32 mask = cpus_addr(cpumask)[0];
mask &= ~(1<<smp_processor_id()); mask &= ~(1<<smp_processor_id());
...@@ -1083,47 +1086,6 @@ __smp_call_function_mask (void (*func) (void *info), void *info, int retry, ...@@ -1083,47 +1086,6 @@ __smp_call_function_mask (void (*func) (void *info), void *info, int retry,
return 0; return 0;
} }
/* Call this function on all CPUs using the function_interrupt above
<func> The function to run. This must be fast and non-blocking.
<info> An arbitrary pointer to pass to the function.
<retry> If true, keep retrying until ready.
<wait> If true, wait until function has completed on other CPUs.
[RETURNS] 0 on success, else a negative status code. Does not return until
remote CPUs are nearly ready to execute <<func>> or are or have executed.
*/
int
smp_call_function(void (*func) (void *info), void *info, int retry,
int wait)
{
__u32 mask = cpus_addr(cpu_online_map)[0];
return __smp_call_function_mask(func, info, retry, wait, mask);
}
EXPORT_SYMBOL(smp_call_function);
/*
* smp_call_function_single - Run a function on another CPU
* @func: The function to run. This must be fast and non-blocking.
* @info: An arbitrary pointer to pass to the function.
* @nonatomic: Currently unused.
* @wait: If true, wait until function has completed on other CPUs.
*
* Retrurns 0 on success, else a negative status code.
*
* Does not return until the remote CPU is nearly ready to execute <func>
* or is or has executed.
*/
int
smp_call_function_single(int cpu, void (*func) (void *info), void *info,
int nonatomic, int wait)
{
__u32 mask = 1 << cpu;
return __smp_call_function_mask(func, info, nonatomic, wait, mask);
}
EXPORT_SYMBOL(smp_call_function_single);
/* Sorry about the name. In an APIC based system, the APICs /* Sorry about the name. In an APIC based system, the APICs
* themselves are programmed to send a timer interrupt. This is used * themselves are programmed to send a timer interrupt. This is used
* by linux to reschedule the processor. Voyager doesn't have this, * by linux to reschedule the processor. Voyager doesn't have this,
...@@ -1237,8 +1199,8 @@ smp_alloc_memory(void) ...@@ -1237,8 +1199,8 @@ smp_alloc_memory(void)
} }
/* send a reschedule CPI to one CPU by physical CPU number*/ /* send a reschedule CPI to one CPU by physical CPU number*/
void static void
smp_send_reschedule(int cpu) voyager_smp_send_reschedule(int cpu)
{ {
send_one_CPI(cpu, VIC_RESCHEDULE_CPI); send_one_CPI(cpu, VIC_RESCHEDULE_CPI);
} }
...@@ -1267,8 +1229,8 @@ safe_smp_processor_id(void) ...@@ -1267,8 +1229,8 @@ safe_smp_processor_id(void)
} }
/* broadcast a halt to all other CPUs */ /* broadcast a halt to all other CPUs */
void static void
smp_send_stop(void) voyager_smp_send_stop(void)
{ {
smp_call_function(smp_stop_cpu_function, NULL, 1, 1); smp_call_function(smp_stop_cpu_function, NULL, 1, 1);
} }
...@@ -1930,23 +1892,26 @@ smp_voyager_power_off(void *dummy) ...@@ -1930,23 +1892,26 @@ smp_voyager_power_off(void *dummy)
smp_stop_cpu_function(NULL); smp_stop_cpu_function(NULL);
} }
void __init static void __init
smp_prepare_cpus(unsigned int max_cpus) voyager_smp_prepare_cpus(unsigned int max_cpus)
{ {
/* FIXME: ignore max_cpus for now */ /* FIXME: ignore max_cpus for now */
smp_boot_cpus(); smp_boot_cpus();
} }
void __devinit smp_prepare_boot_cpu(void) static void __devinit voyager_smp_prepare_boot_cpu(void)
{ {
init_gdt(smp_processor_id());
switch_to_new_gdt();
cpu_set(smp_processor_id(), cpu_online_map); cpu_set(smp_processor_id(), cpu_online_map);
cpu_set(smp_processor_id(), cpu_callout_map); cpu_set(smp_processor_id(), cpu_callout_map);
cpu_set(smp_processor_id(), cpu_possible_map); cpu_set(smp_processor_id(), cpu_possible_map);
cpu_set(smp_processor_id(), cpu_present_map); cpu_set(smp_processor_id(), cpu_present_map);
} }
int __devinit static int __devinit
__cpu_up(unsigned int cpu) voyager_cpu_up(unsigned int cpu)
{ {
/* This only works at boot for x86. See "rewrite" above. */ /* This only works at boot for x86. See "rewrite" above. */
if (cpu_isset(cpu, smp_commenced_mask)) if (cpu_isset(cpu, smp_commenced_mask))
...@@ -1962,8 +1927,8 @@ __cpu_up(unsigned int cpu) ...@@ -1962,8 +1927,8 @@ __cpu_up(unsigned int cpu)
return 0; return 0;
} }
void __init static void __init
smp_cpus_done(unsigned int max_cpus) voyager_smp_cpus_done(unsigned int max_cpus)
{ {
zap_low_mappings(); zap_low_mappings();
} }
...@@ -1972,5 +1937,16 @@ void __init ...@@ -1972,5 +1937,16 @@ void __init
smp_setup_processor_id(void) smp_setup_processor_id(void)
{ {
current_thread_info()->cpu = hard_smp_processor_id(); current_thread_info()->cpu = hard_smp_processor_id();
write_pda(cpu_number, hard_smp_processor_id()); x86_write_percpu(cpu_number, hard_smp_processor_id());
} }
struct smp_ops smp_ops = {
.smp_prepare_boot_cpu = voyager_smp_prepare_boot_cpu,
.smp_prepare_cpus = voyager_smp_prepare_cpus,
.cpu_up = voyager_cpu_up,
.smp_cpus_done = voyager_smp_cpus_done,
.smp_send_stop = voyager_smp_send_stop,
.smp_send_reschedule = voyager_smp_send_reschedule,
.smp_call_function_mask = voyager_smp_call_function_mask,
};
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment