Commit 93a2cdff authored by Rusty Russell's avatar Rusty Russell

lguest: assume Switcher text is a single page.

ie. SHARED_SWITCHER_PAGES == 1.  It is well under a page, and it's a
minor simplification: it's nice to have *one* simplification in a
patch series!
Signed-off-by: default avatarRusty Russell <rusty@rustcorp.com.au>
parent 856c6088
...@@ -11,11 +11,8 @@ ...@@ -11,11 +11,8 @@
#define GUEST_PL 1 #define GUEST_PL 1
/* Every guest maps the core switcher code. */ /* Page for Switcher text itself, then two pages per cpu */
#define SHARED_SWITCHER_PAGES \ #define TOTAL_SWITCHER_PAGES (1 + 2 * nr_cpu_ids)
DIV_ROUND_UP(end_switcher_text - start_switcher_text, PAGE_SIZE)
/* Pages for switcher itself, then two pages per cpu */
#define TOTAL_SWITCHER_PAGES (SHARED_SWITCHER_PAGES + 2 * nr_cpu_ids)
/* We map at -4M (-2M for PAE) for ease of mapping (one PTE page). */ /* We map at -4M (-2M for PAE) for ease of mapping (one PTE page). */
#ifdef CONFIG_X86_PAE #ifdef CONFIG_X86_PAE
......
...@@ -52,6 +52,13 @@ static __init int map_switcher(void) ...@@ -52,6 +52,13 @@ static __init int map_switcher(void)
* easy. * easy.
*/ */
/* We assume Switcher text fits into a single page. */
if (end_switcher_text - start_switcher_text > PAGE_SIZE) {
printk(KERN_ERR "lguest: switcher text too large (%zu)\n",
end_switcher_text - start_switcher_text);
return -EINVAL;
}
/* /*
* We allocate an array of struct page pointers. map_vm_area() wants * We allocate an array of struct page pointers. map_vm_area() wants
* this, rather than just an array of pages. * this, rather than just an array of pages.
...@@ -326,7 +333,7 @@ static int __init init(void) ...@@ -326,7 +333,7 @@ static int __init init(void)
goto out; goto out;
/* Now we set up the pagetable implementation for the Guests. */ /* Now we set up the pagetable implementation for the Guests. */
err = init_pagetables(switcher_pages, SHARED_SWITCHER_PAGES); err = init_pagetables(switcher_pages);
if (err) if (err)
goto unmap; goto unmap;
......
...@@ -15,7 +15,7 @@ ...@@ -15,7 +15,7 @@
#include <asm/lguest.h> #include <asm/lguest.h>
void free_pagetables(void); void free_pagetables(void);
int init_pagetables(struct page **switcher_pages, unsigned int pages); int init_pagetables(struct page **switcher_pages);
struct pgdir { struct pgdir {
unsigned long gpgdir; unsigned long gpgdir;
......
...@@ -1079,25 +1079,20 @@ static void free_switcher_pte_pages(void) ...@@ -1079,25 +1079,20 @@ static void free_switcher_pte_pages(void)
/*H:520 /*H:520
* Setting up the Switcher PTE page for given CPU is fairly easy, given * Setting up the Switcher PTE page for given CPU is fairly easy, given
* the CPU number and the "struct page"s for the Switcher code itself. * the CPU number and the "struct page"s for the Switcher and per-cpu pages.
*
* Currently the Switcher is less than a page long, so "pages" is always 1.
*/ */
static __init void populate_switcher_pte_page(unsigned int cpu, static __init void populate_switcher_pte_page(unsigned int cpu,
struct page *switcher_pages[], struct page *switcher_pages[])
unsigned int pages)
{ {
unsigned int i;
pte_t *pte = switcher_pte_page(cpu); pte_t *pte = switcher_pte_page(cpu);
int i;
/* The first entries are easy: they map the Switcher code. */ /* The first entries maps the Switcher code. */
for (i = 0; i < pages; i++) { set_pte(&pte[0], mk_pte(switcher_pages[0],
set_pte(&pte[i], mk_pte(switcher_pages[i],
__pgprot(_PAGE_PRESENT|_PAGE_ACCESSED))); __pgprot(_PAGE_PRESENT|_PAGE_ACCESSED)));
}
/* The only other thing we map is this CPU's pair of pages. */ /* The only other thing we map is this CPU's pair of pages. */
i = pages + cpu*2; i = 1 + cpu*2;
/* First page (Guest registers) is writable from the Guest */ /* First page (Guest registers) is writable from the Guest */
set_pte(&pte[i], pfn_pte(page_to_pfn(switcher_pages[i]), set_pte(&pte[i], pfn_pte(page_to_pfn(switcher_pages[i]),
...@@ -1128,7 +1123,7 @@ static __init void populate_switcher_pte_page(unsigned int cpu, ...@@ -1128,7 +1123,7 @@ static __init void populate_switcher_pte_page(unsigned int cpu,
* At boot or module load time, init_pagetables() allocates and populates * At boot or module load time, init_pagetables() allocates and populates
* the Switcher PTE page for each CPU. * the Switcher PTE page for each CPU.
*/ */
__init int init_pagetables(struct page **switcher_pages, unsigned int pages) __init int init_pagetables(struct page **switcher_pages)
{ {
unsigned int i; unsigned int i;
...@@ -1138,7 +1133,7 @@ __init int init_pagetables(struct page **switcher_pages, unsigned int pages) ...@@ -1138,7 +1133,7 @@ __init int init_pagetables(struct page **switcher_pages, unsigned int pages)
free_switcher_pte_pages(); free_switcher_pte_pages();
return -ENOMEM; return -ENOMEM;
} }
populate_switcher_pte_page(i, switcher_pages, pages); populate_switcher_pte_page(i, switcher_pages);
} }
return 0; return 0;
} }
......
...@@ -62,11 +62,10 @@ static unsigned long switcher_offset(void) ...@@ -62,11 +62,10 @@ static unsigned long switcher_offset(void)
return switcher_addr - (unsigned long)start_switcher_text; return switcher_addr - (unsigned long)start_switcher_text;
} }
/* This cpu's struct lguest_pages. */ /* This cpu's struct lguest_pages (after the Switcher text page) */
static struct lguest_pages *lguest_pages(unsigned int cpu) static struct lguest_pages *lguest_pages(unsigned int cpu)
{ {
return &(((struct lguest_pages *) return &(((struct lguest_pages *)(switcher_addr + PAGE_SIZE))[cpu]);
(switcher_addr + SHARED_SWITCHER_PAGES*PAGE_SIZE))[cpu]);
} }
static DEFINE_PER_CPU(struct lg_cpu *, lg_last_cpu); static DEFINE_PER_CPU(struct lg_cpu *, lg_last_cpu);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment