Commit f63c2f24 authored by Tej's avatar Tej Committed by Ingo Molnar

xen: whitespace/checkpatch cleanup

Impact: cleanup
Signed-off-by: default avatarTej <bewith.tej@gmail.com>
Signed-off-by: default avatarJeremy Fitzhardinge <jeremy.fitzhardinge@citrix.com>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent 1bda7128
...@@ -793,7 +793,7 @@ static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high) ...@@ -793,7 +793,7 @@ static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high)
ret = 0; ret = 0;
switch(msr) { switch (msr) {
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
unsigned which; unsigned which;
u64 base; u64 base;
...@@ -1453,7 +1453,7 @@ static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn) ...@@ -1453,7 +1453,7 @@ static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn)
ident_pte = 0; ident_pte = 0;
pfn = 0; pfn = 0;
for(pmdidx = 0; pmdidx < PTRS_PER_PMD && pfn < max_pfn; pmdidx++) { for (pmdidx = 0; pmdidx < PTRS_PER_PMD && pfn < max_pfn; pmdidx++) {
pte_t *pte_page; pte_t *pte_page;
/* Reuse or allocate a page of ptes */ /* Reuse or allocate a page of ptes */
...@@ -1471,7 +1471,7 @@ static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn) ...@@ -1471,7 +1471,7 @@ static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn)
} }
/* Install mappings */ /* Install mappings */
for(pteidx = 0; pteidx < PTRS_PER_PTE; pteidx++, pfn++) { for (pteidx = 0; pteidx < PTRS_PER_PTE; pteidx++, pfn++) {
pte_t pte; pte_t pte;
if (pfn > max_pfn_mapped) if (pfn > max_pfn_mapped)
...@@ -1485,7 +1485,7 @@ static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn) ...@@ -1485,7 +1485,7 @@ static __init void xen_map_identity_early(pmd_t *pmd, unsigned long max_pfn)
} }
} }
for(pteidx = 0; pteidx < ident_pte; pteidx += PTRS_PER_PTE) for (pteidx = 0; pteidx < ident_pte; pteidx += PTRS_PER_PTE)
set_page_prot(&level1_ident_pgt[pteidx], PAGE_KERNEL_RO); set_page_prot(&level1_ident_pgt[pteidx], PAGE_KERNEL_RO);
set_page_prot(pmd, PAGE_KERNEL_RO); set_page_prot(pmd, PAGE_KERNEL_RO);
...@@ -1499,7 +1499,7 @@ static void convert_pfn_mfn(void *v) ...@@ -1499,7 +1499,7 @@ static void convert_pfn_mfn(void *v)
/* All levels are converted the same way, so just treat them /* All levels are converted the same way, so just treat them
as ptes. */ as ptes. */
for(i = 0; i < PTRS_PER_PTE; i++) for (i = 0; i < PTRS_PER_PTE; i++)
pte[i] = xen_make_pte(pte[i].pte); pte[i] = xen_make_pte(pte[i].pte);
} }
...@@ -1514,7 +1514,8 @@ static void convert_pfn_mfn(void *v) ...@@ -1514,7 +1514,8 @@ static void convert_pfn_mfn(void *v)
* of the physical mapping once some sort of allocator has been set * of the physical mapping once some sort of allocator has been set
* up. * up.
*/ */
static __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd, unsigned long max_pfn) static __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd,
unsigned long max_pfn)
{ {
pud_t *l3; pud_t *l3;
pmd_t *l2; pmd_t *l2;
...@@ -1577,7 +1578,8 @@ static __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd, unsigned long max_pf ...@@ -1577,7 +1578,8 @@ static __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd, unsigned long max_pf
#else /* !CONFIG_X86_64 */ #else /* !CONFIG_X86_64 */
static pmd_t level2_kernel_pgt[PTRS_PER_PMD] __page_aligned_bss; static pmd_t level2_kernel_pgt[PTRS_PER_PMD] __page_aligned_bss;
static __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd, unsigned long max_pfn) static __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd,
unsigned long max_pfn)
{ {
pmd_t *kernel_pmd; pmd_t *kernel_pmd;
......
...@@ -154,13 +154,13 @@ void xen_setup_mfn_list_list(void) ...@@ -154,13 +154,13 @@ void xen_setup_mfn_list_list(void)
{ {
unsigned pfn, idx; unsigned pfn, idx;
for(pfn = 0; pfn < MAX_DOMAIN_PAGES; pfn += P2M_ENTRIES_PER_PAGE) { for (pfn = 0; pfn < MAX_DOMAIN_PAGES; pfn += P2M_ENTRIES_PER_PAGE) {
unsigned topidx = p2m_top_index(pfn); unsigned topidx = p2m_top_index(pfn);
p2m_top_mfn[topidx] = virt_to_mfn(p2m_top[topidx]); p2m_top_mfn[topidx] = virt_to_mfn(p2m_top[topidx]);
} }
for(idx = 0; idx < ARRAY_SIZE(p2m_top_mfn_list); idx++) { for (idx = 0; idx < ARRAY_SIZE(p2m_top_mfn_list); idx++) {
unsigned topidx = idx * P2M_ENTRIES_PER_PAGE; unsigned topidx = idx * P2M_ENTRIES_PER_PAGE;
p2m_top_mfn_list[idx] = virt_to_mfn(&p2m_top_mfn[topidx]); p2m_top_mfn_list[idx] = virt_to_mfn(&p2m_top_mfn[topidx]);
} }
...@@ -179,7 +179,7 @@ void __init xen_build_dynamic_phys_to_machine(void) ...@@ -179,7 +179,7 @@ void __init xen_build_dynamic_phys_to_machine(void)
unsigned long max_pfn = min(MAX_DOMAIN_PAGES, xen_start_info->nr_pages); unsigned long max_pfn = min(MAX_DOMAIN_PAGES, xen_start_info->nr_pages);
unsigned pfn; unsigned pfn;
for(pfn = 0; pfn < max_pfn; pfn += P2M_ENTRIES_PER_PAGE) { for (pfn = 0; pfn < max_pfn; pfn += P2M_ENTRIES_PER_PAGE) {
unsigned topidx = p2m_top_index(pfn); unsigned topidx = p2m_top_index(pfn);
p2m_top[topidx] = &mfn_list[pfn]; p2m_top[topidx] = &mfn_list[pfn];
...@@ -207,7 +207,7 @@ static void alloc_p2m(unsigned long **pp, unsigned long *mfnp) ...@@ -207,7 +207,7 @@ static void alloc_p2m(unsigned long **pp, unsigned long *mfnp)
p = (void *)__get_free_page(GFP_KERNEL | __GFP_NOFAIL); p = (void *)__get_free_page(GFP_KERNEL | __GFP_NOFAIL);
BUG_ON(p == NULL); BUG_ON(p == NULL);
for(i = 0; i < P2M_ENTRIES_PER_PAGE; i++) for (i = 0; i < P2M_ENTRIES_PER_PAGE; i++)
p[i] = INVALID_P2M_ENTRY; p[i] = INVALID_P2M_ENTRY;
if (cmpxchg(pp, p2m_missing, p) != p2m_missing) if (cmpxchg(pp, p2m_missing, p) != p2m_missing)
...@@ -407,7 +407,8 @@ void xen_set_pte_at(struct mm_struct *mm, unsigned long addr, ...@@ -407,7 +407,8 @@ void xen_set_pte_at(struct mm_struct *mm, unsigned long addr,
preempt_enable(); preempt_enable();
} }
pte_t xen_ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, pte_t *ptep) pte_t xen_ptep_modify_prot_start(struct mm_struct *mm,
unsigned long addr, pte_t *ptep)
{ {
/* Just return the pte as-is. We preserve the bits on commit */ /* Just return the pte as-is. We preserve the bits on commit */
return *ptep; return *ptep;
...@@ -878,7 +879,8 @@ static void __xen_pgd_pin(struct mm_struct *mm, pgd_t *pgd) ...@@ -878,7 +879,8 @@ static void __xen_pgd_pin(struct mm_struct *mm, pgd_t *pgd)
if (user_pgd) { if (user_pgd) {
xen_pin_page(mm, virt_to_page(user_pgd), PT_PGD); xen_pin_page(mm, virt_to_page(user_pgd), PT_PGD);
xen_do_pin(MMUEXT_PIN_L4_TABLE, PFN_DOWN(__pa(user_pgd))); xen_do_pin(MMUEXT_PIN_L4_TABLE,
PFN_DOWN(__pa(user_pgd)));
} }
} }
#else /* CONFIG_X86_32 */ #else /* CONFIG_X86_32 */
...@@ -993,7 +995,8 @@ static void __xen_pgd_unpin(struct mm_struct *mm, pgd_t *pgd) ...@@ -993,7 +995,8 @@ static void __xen_pgd_unpin(struct mm_struct *mm, pgd_t *pgd)
pgd_t *user_pgd = xen_get_user_pgd(pgd); pgd_t *user_pgd = xen_get_user_pgd(pgd);
if (user_pgd) { if (user_pgd) {
xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(user_pgd))); xen_do_pin(MMUEXT_UNPIN_TABLE,
PFN_DOWN(__pa(user_pgd)));
xen_unpin_page(mm, virt_to_page(user_pgd), PT_PGD); xen_unpin_page(mm, virt_to_page(user_pgd), PT_PGD);
} }
} }
......
...@@ -154,7 +154,7 @@ void xen_mc_flush(void) ...@@ -154,7 +154,7 @@ void xen_mc_flush(void)
ret, smp_processor_id()); ret, smp_processor_id());
dump_stack(); dump_stack();
for (i = 0; i < b->mcidx; i++) { for (i = 0; i < b->mcidx; i++) {
printk(" call %2d/%d: op=%lu arg=[%lx] result=%ld\n", printk(KERN_DEBUG " call %2d/%d: op=%lu arg=[%lx] result=%ld\n",
i+1, b->mcidx, i+1, b->mcidx,
b->debug[i].op, b->debug[i].op,
b->debug[i].args[0], b->debug[i].args[0],
......
...@@ -28,6 +28,9 @@ ...@@ -28,6 +28,9 @@
/* These are code, but not functions. Defined in entry.S */ /* These are code, but not functions. Defined in entry.S */
extern const char xen_hypervisor_callback[]; extern const char xen_hypervisor_callback[];
extern const char xen_failsafe_callback[]; extern const char xen_failsafe_callback[];
extern void xen_sysenter_target(void);
extern void xen_syscall_target(void);
extern void xen_syscall32_target(void);
/** /**
...@@ -110,7 +113,6 @@ static __cpuinit int register_callback(unsigned type, const void *func) ...@@ -110,7 +113,6 @@ static __cpuinit int register_callback(unsigned type, const void *func)
void __cpuinit xen_enable_sysenter(void) void __cpuinit xen_enable_sysenter(void)
{ {
extern void xen_sysenter_target(void);
int ret; int ret;
unsigned sysenter_feature; unsigned sysenter_feature;
...@@ -132,8 +134,6 @@ void __cpuinit xen_enable_syscall(void) ...@@ -132,8 +134,6 @@ void __cpuinit xen_enable_syscall(void)
{ {
#ifdef CONFIG_X86_64 #ifdef CONFIG_X86_64
int ret; int ret;
extern void xen_syscall_target(void);
extern void xen_syscall32_target(void);
ret = register_callback(CALLBACKTYPE_syscall, xen_syscall_target); ret = register_callback(CALLBACKTYPE_syscall, xen_syscall_target);
if (ret != 0) { if (ret != 0) {
...@@ -160,7 +160,8 @@ void __init xen_arch_setup(void) ...@@ -160,7 +160,8 @@ void __init xen_arch_setup(void)
HYPERVISOR_vm_assist(VMASST_CMD_enable, VMASST_TYPE_writable_pagetables); HYPERVISOR_vm_assist(VMASST_CMD_enable, VMASST_TYPE_writable_pagetables);
if (!xen_feature(XENFEAT_auto_translated_physmap)) if (!xen_feature(XENFEAT_auto_translated_physmap))
HYPERVISOR_vm_assist(VMASST_CMD_enable, VMASST_TYPE_pae_extended_cr3); HYPERVISOR_vm_assist(VMASST_CMD_enable,
VMASST_TYPE_pae_extended_cr3);
if (register_callback(CALLBACKTYPE_event, xen_hypervisor_callback) || if (register_callback(CALLBACKTYPE_event, xen_hypervisor_callback) ||
register_callback(CALLBACKTYPE_failsafe, xen_failsafe_callback)) register_callback(CALLBACKTYPE_failsafe, xen_failsafe_callback))
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment