Commit 9b4006dc authored by David S. Miller's avatar David S. Miller

[SPARC64]: Use SLAB caches for TSB tables.

Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent b52439c2
...@@ -192,6 +192,9 @@ config ARCH_SPARSEMEM_ENABLE ...@@ -192,6 +192,9 @@ config ARCH_SPARSEMEM_ENABLE
config ARCH_SPARSEMEM_DEFAULT config ARCH_SPARSEMEM_DEFAULT
def_bool y def_bool y
config LARGE_ALLOCS
def_bool y
source "mm/Kconfig" source "mm/Kconfig"
config GENERIC_ISA_DMA config GENERIC_ISA_DMA
......
...@@ -165,6 +165,8 @@ static void zero_ctor(void *addr, kmem_cache_t *cache, unsigned long flags) ...@@ -165,6 +165,8 @@ static void zero_ctor(void *addr, kmem_cache_t *cache, unsigned long flags)
clear_page(addr); clear_page(addr);
} }
extern void tsb_cache_init(void);
void pgtable_cache_init(void) void pgtable_cache_init(void)
{ {
pgtable_cache = kmem_cache_create("pgtable_cache", pgtable_cache = kmem_cache_create("pgtable_cache",
...@@ -174,9 +176,10 @@ void pgtable_cache_init(void) ...@@ -174,9 +176,10 @@ void pgtable_cache_init(void)
zero_ctor, zero_ctor,
NULL); NULL);
if (!pgtable_cache) { if (!pgtable_cache) {
prom_printf("pgtable_cache_init(): Could not create!\n"); prom_printf("Could not create pgtable_cache\n");
prom_halt(); prom_halt();
} }
tsb_cache_init();
} }
#ifdef CONFIG_DEBUG_DCFLUSH #ifdef CONFIG_DEBUG_DCFLUSH
......
...@@ -11,6 +11,7 @@ ...@@ -11,6 +11,7 @@
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
#include <asm/tsb.h> #include <asm/tsb.h>
#include <asm/oplib.h>
extern struct tsb swapper_tsb[KERNEL_TSB_NENTRIES]; extern struct tsb swapper_tsb[KERNEL_TSB_NENTRIES];
...@@ -207,6 +208,39 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes) ...@@ -207,6 +208,39 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
} }
} }
static kmem_cache_t *tsb_caches[8] __read_mostly;
static const char *tsb_cache_names[8] = {
"tsb_8KB",
"tsb_16KB",
"tsb_32KB",
"tsb_64KB",
"tsb_128KB",
"tsb_256KB",
"tsb_512KB",
"tsb_1MB",
};
void __init tsb_cache_init(void)
{
unsigned long i;
for (i = 0; i < 8; i++) {
unsigned long size = 8192 << i;
const char *name = tsb_cache_names[i];
tsb_caches[i] = kmem_cache_create(name,
size, size,
SLAB_HWCACHE_ALIGN |
SLAB_MUST_HWCACHE_ALIGN,
NULL, NULL);
if (!tsb_caches[i]) {
prom_printf("Could not create %s cache\n", name);
prom_halt();
}
}
}
/* When the RSS of an address space exceeds mm->context.tsb_rss_limit, /* When the RSS of an address space exceeds mm->context.tsb_rss_limit,
* do_sparc64_fault() invokes this routine to try and grow the TSB. * do_sparc64_fault() invokes this routine to try and grow the TSB.
* *
...@@ -226,45 +260,48 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes) ...@@ -226,45 +260,48 @@ static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
void tsb_grow(struct mm_struct *mm, unsigned long rss) void tsb_grow(struct mm_struct *mm, unsigned long rss)
{ {
unsigned long max_tsb_size = 1 * 1024 * 1024; unsigned long max_tsb_size = 1 * 1024 * 1024;
unsigned long size, old_size, flags; unsigned long new_size, old_size, flags;
struct page *page;
struct tsb *old_tsb, *new_tsb; struct tsb *old_tsb, *new_tsb;
unsigned long order, new_rss_limit; unsigned long new_cache_index, old_cache_index;
unsigned long new_rss_limit;
gfp_t gfp_flags; gfp_t gfp_flags;
if (max_tsb_size > (PAGE_SIZE << MAX_ORDER)) if (max_tsb_size > (PAGE_SIZE << MAX_ORDER))
max_tsb_size = (PAGE_SIZE << MAX_ORDER); max_tsb_size = (PAGE_SIZE << MAX_ORDER);
for (size = PAGE_SIZE; size < max_tsb_size; size <<= 1UL) { new_cache_index = 0;
unsigned long n_entries = size / sizeof(struct tsb); for (new_size = 8192; new_size < max_tsb_size; new_size <<= 1UL) {
unsigned long n_entries = new_size / sizeof(struct tsb);
n_entries = (n_entries * 3) / 4; n_entries = (n_entries * 3) / 4;
if (n_entries > rss) if (n_entries > rss)
break; break;
new_cache_index++;
} }
if (size == max_tsb_size) if (new_size == max_tsb_size)
new_rss_limit = ~0UL; new_rss_limit = ~0UL;
else else
new_rss_limit = ((size / sizeof(struct tsb)) * 3) / 4; new_rss_limit = ((new_size / sizeof(struct tsb)) * 3) / 4;
retry_page_alloc: retry_tsb_alloc:
order = get_order(size);
gfp_flags = GFP_KERNEL; gfp_flags = GFP_KERNEL;
if (order > 1) if (new_size > (PAGE_SIZE * 2))
gfp_flags = __GFP_NOWARN | __GFP_NORETRY; gfp_flags = __GFP_NOWARN | __GFP_NORETRY;
page = alloc_pages(gfp_flags, order); new_tsb = kmem_cache_alloc(tsb_caches[new_cache_index], gfp_flags);
if (unlikely(!page)) { if (unlikely(!new_tsb)) {
/* Not being able to fork due to a high-order TSB /* Not being able to fork due to a high-order TSB
* allocation failure is very bad behavior. Just back * allocation failure is very bad behavior. Just back
* down to a 0-order allocation and force no TSB * down to a 0-order allocation and force no TSB
* growing for this address space. * growing for this address space.
*/ */
if (mm->context.tsb == NULL && order > 0) { if (mm->context.tsb == NULL && new_cache_index > 0) {
size = PAGE_SIZE; new_cache_index = 0;
new_size = 8192;
new_rss_limit = ~0UL; new_rss_limit = ~0UL;
goto retry_page_alloc; goto retry_tsb_alloc;
} }
/* If we failed on a TSB grow, we are under serious /* If we failed on a TSB grow, we are under serious
...@@ -276,8 +313,7 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss) ...@@ -276,8 +313,7 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss)
} }
/* Mark all tags as invalid. */ /* Mark all tags as invalid. */
new_tsb = page_address(page); memset(new_tsb, 0x40, new_size);
memset(new_tsb, 0x40, size);
/* Ok, we are about to commit the changes. If we are /* Ok, we are about to commit the changes. If we are
* growing an existing TSB the locking is very tricky, * growing an existing TSB the locking is very tricky,
...@@ -304,8 +340,10 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss) ...@@ -304,8 +340,10 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss)
spin_lock_irqsave(&mm->context.lock, flags); spin_lock_irqsave(&mm->context.lock, flags);
old_tsb = mm->context.tsb; old_tsb = mm->context.tsb;
old_cache_index = (mm->context.tsb_reg_val & 0x7UL);
old_size = mm->context.tsb_nentries * sizeof(struct tsb); old_size = mm->context.tsb_nentries * sizeof(struct tsb);
/* Handle multiple threads trying to grow the TSB at the same time. /* Handle multiple threads trying to grow the TSB at the same time.
* One will get in here first, and bump the size and the RSS limit. * One will get in here first, and bump the size and the RSS limit.
* The others will get in here next and hit this check. * The others will get in here next and hit this check.
...@@ -313,7 +351,7 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss) ...@@ -313,7 +351,7 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss)
if (unlikely(old_tsb && (rss < mm->context.tsb_rss_limit))) { if (unlikely(old_tsb && (rss < mm->context.tsb_rss_limit))) {
spin_unlock_irqrestore(&mm->context.lock, flags); spin_unlock_irqrestore(&mm->context.lock, flags);
free_pages((unsigned long) new_tsb, get_order(size)); kmem_cache_free(tsb_caches[new_cache_index], new_tsb);
return; return;
} }
...@@ -331,11 +369,11 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss) ...@@ -331,11 +369,11 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss)
old_tsb_base = __pa(old_tsb_base); old_tsb_base = __pa(old_tsb_base);
new_tsb_base = __pa(new_tsb_base); new_tsb_base = __pa(new_tsb_base);
} }
copy_tsb(old_tsb_base, old_size, new_tsb_base, size); copy_tsb(old_tsb_base, old_size, new_tsb_base, new_size);
} }
mm->context.tsb = new_tsb; mm->context.tsb = new_tsb;
setup_tsb_params(mm, size); setup_tsb_params(mm, new_size);
spin_unlock_irqrestore(&mm->context.lock, flags); spin_unlock_irqrestore(&mm->context.lock, flags);
...@@ -350,7 +388,7 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss) ...@@ -350,7 +388,7 @@ void tsb_grow(struct mm_struct *mm, unsigned long rss)
smp_tsb_sync(mm); smp_tsb_sync(mm);
/* Now it is safe to free the old tsb. */ /* Now it is safe to free the old tsb. */
free_pages((unsigned long) old_tsb, get_order(old_size)); kmem_cache_free(tsb_caches[old_cache_index], old_tsb);
} }
} }
...@@ -379,10 +417,10 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm) ...@@ -379,10 +417,10 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
void destroy_context(struct mm_struct *mm) void destroy_context(struct mm_struct *mm)
{ {
unsigned long size = mm->context.tsb_nentries * sizeof(struct tsb); unsigned long flags, cache_index;
unsigned long flags;
free_pages((unsigned long) mm->context.tsb, get_order(size)); cache_index = (mm->context.tsb_reg_val & 0x7UL);
kmem_cache_free(tsb_caches[cache_index], mm->context.tsb);
/* We can remove these later, but for now it's useful /* We can remove these later, but for now it's useful
* to catch any bogus post-destroy_context() references * to catch any bogus post-destroy_context() references
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment