Commit 8d408b4b authored by Tejun Heo's avatar Tejun Heo

percpu: give more latitude to arch specific first chunk initialization

Impact: more latitude for first percpu chunk allocation

The first percpu chunk serves the kernel static percpu area and may or
may not contain extra room for further dynamic allocation.
Initialization of the first chunk needs to be done before normal
memory allocation service is up, so it has its own init path -
pcpu_setup_static().

It seems archs need more latitude while initializing the first chunk
for example to take advantage of large page mapping.  This patch makes
the following changes to allow this.

* Define PERCPU_DYNAMIC_RESERVE to give arch hint about how much space
  to reserve in the first chunk for further dynamic allocation.

* Rename pcpu_setup_static() to pcpu_setup_first_chunk().

* Make pcpu_setup_first_chunk() much more flexible by fetching page
  pointer by callback and adding optional @unit_size, @free_size and
  @base_addr arguments which allow archs to selectively part of chunk
  initialization to their likings.
Signed-off-by: default avatarTejun Heo <tj@kernel.org>
parent d9b55eeb
...@@ -41,6 +41,16 @@ unsigned long __per_cpu_offset[NR_CPUS] __read_mostly = { ...@@ -41,6 +41,16 @@ unsigned long __per_cpu_offset[NR_CPUS] __read_mostly = {
}; };
EXPORT_SYMBOL(__per_cpu_offset); EXPORT_SYMBOL(__per_cpu_offset);
static struct page **pcpu4k_pages __initdata;
static int pcpu4k_nr_static_pages __initdata;
static struct page * __init pcpu4k_get_page(unsigned int cpu, int pageno)
{
if (pageno < pcpu4k_nr_static_pages)
return pcpu4k_pages[cpu * pcpu4k_nr_static_pages + pageno];
return NULL;
}
static void __init pcpu4k_populate_pte(unsigned long addr) static void __init pcpu4k_populate_pte(unsigned long addr)
{ {
populate_extra_pte(addr); populate_extra_pte(addr);
...@@ -109,7 +119,10 @@ void __init setup_per_cpu_areas(void) ...@@ -109,7 +119,10 @@ void __init setup_per_cpu_areas(void)
} }
} }
pcpu_unit_size = pcpu_setup_static(pcpu4k_populate_pte, pages, size); pcpu4k_pages = pages;
pcpu4k_nr_static_pages = nr_cpu_pages;
pcpu_unit_size = pcpu_setup_first_chunk(pcpu4k_get_page, size, 0, 0,
NULL, pcpu4k_populate_pte);
free_bootmem(__pa(pages), pages_size); free_bootmem(__pa(pages), pages_size);
......
...@@ -78,12 +78,47 @@ ...@@ -78,12 +78,47 @@
#ifdef CONFIG_HAVE_DYNAMIC_PER_CPU_AREA #ifdef CONFIG_HAVE_DYNAMIC_PER_CPU_AREA
/* minimum unit size, also is the maximum supported allocation size */
#define PCPU_MIN_UNIT_SIZE (16UL << PAGE_SHIFT)
/*
* PERCPU_DYNAMIC_RESERVE indicates the amount of free area to piggy
* back on the first chunk if arch is manually allocating and mapping
* it for faster access (as a part of large page mapping for example).
* Note that dynamic percpu allocator covers both static and dynamic
* areas, so these values are bigger than PERCPU_MODULE_RESERVE.
*
* On typical configuration with modules, the following values leave
* about 8k of free space on the first chunk after boot on both x86_32
* and 64 when module support is enabled. When module support is
* disabled, it's much tighter.
*/
#ifndef PERCPU_DYNAMIC_RESERVE
# if BITS_PER_LONG > 32
# ifdef CONFIG_MODULES
# define PERCPU_DYNAMIC_RESERVE (6 << PAGE_SHIFT)
# else
# define PERCPU_DYNAMIC_RESERVE (4 << PAGE_SHIFT)
# endif
# else
# ifdef CONFIG_MODULES
# define PERCPU_DYNAMIC_RESERVE (4 << PAGE_SHIFT)
# else
# define PERCPU_DYNAMIC_RESERVE (2 << PAGE_SHIFT)
# endif
# endif
#endif /* PERCPU_DYNAMIC_RESERVE */
extern void *pcpu_base_addr; extern void *pcpu_base_addr;
typedef struct page * (*pcpu_get_page_fn_t)(unsigned int cpu, int pageno);
typedef void (*pcpu_populate_pte_fn_t)(unsigned long addr); typedef void (*pcpu_populate_pte_fn_t)(unsigned long addr);
extern size_t __init pcpu_setup_static(pcpu_populate_pte_fn_t populate_pte_fn, extern size_t __init pcpu_setup_first_chunk(pcpu_get_page_fn_t get_page_fn,
struct page **pages, size_t cpu_size); size_t static_size, size_t unit_size,
size_t free_size, void *base_addr,
pcpu_populate_pte_fn_t populate_pte_fn);
/* /*
* Use this to get to a cpu's version of the per-cpu object * Use this to get to a cpu's version of the per-cpu object
* dynamically allocated. Non-atomic access to the current CPU's * dynamically allocated. Non-atomic access to the current CPU's
......
...@@ -48,8 +48,8 @@ ...@@ -48,8 +48,8 @@
* - define __addr_to_pcpu_ptr() and __pcpu_ptr_to_addr() to translate * - define __addr_to_pcpu_ptr() and __pcpu_ptr_to_addr() to translate
* regular address to percpu pointer and back * regular address to percpu pointer and back
* *
* - use pcpu_setup_static() during percpu area initialization to * - use pcpu_setup_first_chunk() during percpu area initialization to
* setup kernel static percpu area * setup the first chunk containing the kernel static percpu area
*/ */
#include <linux/bitmap.h> #include <linux/bitmap.h>
...@@ -67,7 +67,6 @@ ...@@ -67,7 +67,6 @@
#include <asm/cacheflush.h> #include <asm/cacheflush.h>
#include <asm/tlbflush.h> #include <asm/tlbflush.h>
#define PCPU_MIN_UNIT_PAGES 16 /* max alloc size in pages */
#define PCPU_SLOT_BASE_SHIFT 5 /* 1-31 shares the same slot */ #define PCPU_SLOT_BASE_SHIFT 5 /* 1-31 shares the same slot */
#define PCPU_DFL_MAP_ALLOC 16 /* start a map with 16 ents */ #define PCPU_DFL_MAP_ALLOC 16 /* start a map with 16 ents */
...@@ -80,6 +79,7 @@ struct pcpu_chunk { ...@@ -80,6 +79,7 @@ struct pcpu_chunk {
int map_used; /* # of map entries used */ int map_used; /* # of map entries used */
int map_alloc; /* # of map entries allocated */ int map_alloc; /* # of map entries allocated */
int *map; /* allocation map */ int *map; /* allocation map */
bool immutable; /* no [de]population allowed */
struct page *page[]; /* #cpus * UNIT_PAGES */ struct page *page[]; /* #cpus * UNIT_PAGES */
}; };
...@@ -521,6 +521,9 @@ static void pcpu_unmap(struct pcpu_chunk *chunk, int page_start, int page_end, ...@@ -521,6 +521,9 @@ static void pcpu_unmap(struct pcpu_chunk *chunk, int page_start, int page_end,
unsigned int last = num_possible_cpus() - 1; unsigned int last = num_possible_cpus() - 1;
unsigned int cpu; unsigned int cpu;
/* unmap must not be done on immutable chunk */
WARN_ON(chunk->immutable);
/* /*
* Each flushing trial can be very expensive, issue flush on * Each flushing trial can be very expensive, issue flush on
* the whole region at once rather than doing it for each cpu. * the whole region at once rather than doing it for each cpu.
...@@ -602,6 +605,9 @@ static int pcpu_map(struct pcpu_chunk *chunk, int page_start, int page_end) ...@@ -602,6 +605,9 @@ static int pcpu_map(struct pcpu_chunk *chunk, int page_start, int page_end)
unsigned int cpu; unsigned int cpu;
int err; int err;
/* map must not be done on immutable chunk */
WARN_ON(chunk->immutable);
for_each_possible_cpu(cpu) { for_each_possible_cpu(cpu) {
err = map_kernel_range_noflush( err = map_kernel_range_noflush(
pcpu_chunk_addr(chunk, cpu, page_start), pcpu_chunk_addr(chunk, cpu, page_start),
...@@ -727,8 +733,7 @@ void *__alloc_percpu(size_t size, size_t align) ...@@ -727,8 +733,7 @@ void *__alloc_percpu(size_t size, size_t align)
struct pcpu_chunk *chunk; struct pcpu_chunk *chunk;
int slot, off; int slot, off;
if (unlikely(!size || size > PCPU_MIN_UNIT_PAGES * PAGE_SIZE || if (unlikely(!size || size > PCPU_MIN_UNIT_SIZE || align > PAGE_SIZE)) {
align > PAGE_SIZE)) {
WARN(true, "illegal size (%zu) or align (%zu) for " WARN(true, "illegal size (%zu) or align (%zu) for "
"percpu allocation\n", size, align); "percpu allocation\n", size, align);
return NULL; return NULL;
...@@ -776,6 +781,7 @@ EXPORT_SYMBOL_GPL(__alloc_percpu); ...@@ -776,6 +781,7 @@ EXPORT_SYMBOL_GPL(__alloc_percpu);
static void pcpu_kill_chunk(struct pcpu_chunk *chunk) static void pcpu_kill_chunk(struct pcpu_chunk *chunk)
{ {
WARN_ON(chunk->immutable);
pcpu_depopulate_chunk(chunk, 0, pcpu_unit_size, false); pcpu_depopulate_chunk(chunk, 0, pcpu_unit_size, false);
list_del(&chunk->list); list_del(&chunk->list);
rb_erase(&chunk->rb_node, &pcpu_addr_root); rb_erase(&chunk->rb_node, &pcpu_addr_root);
...@@ -821,33 +827,73 @@ void free_percpu(void *ptr) ...@@ -821,33 +827,73 @@ void free_percpu(void *ptr)
EXPORT_SYMBOL_GPL(free_percpu); EXPORT_SYMBOL_GPL(free_percpu);
/** /**
* pcpu_setup_static - initialize kernel static percpu area * pcpu_setup_first_chunk - initialize the first percpu chunk
* @populate_pte_fn: callback to allocate pagetable * @get_page_fn: callback to fetch page pointer
* @pages: num_possible_cpus() * PFN_UP(cpu_size) pages * @static_size: the size of static percpu area in bytes
* @cpu_size: the size of static percpu area in bytes * @unit_size: unit size in bytes, must be multiple of PAGE_SIZE, 0 for auto
* * @free_size: free size in bytes, 0 for auto
* Initialize kernel static percpu area. The caller should allocate * @base_addr: mapped address, NULL for auto
* all the necessary pages and pass them in @pages. * @populate_pte_fn: callback to allocate pagetable, NULL if unnecessary
* @populate_pte_fn() is called on each page to be used for percpu *
* mapping and is responsible for making sure all the necessary page * Initialize the first percpu chunk which contains the kernel static
* tables for the page is allocated. * perpcu area. This function is to be called from arch percpu area
* setup path. The first two parameters are mandatory. The rest are
* optional.
*
* @get_page_fn() should return pointer to percpu page given cpu
* number and page number. It should at least return enough pages to
* cover the static area. The returned pages for static area should
* have been initialized with valid data. If @unit_size is specified,
* it can also return pages after the static area. NULL return
* indicates end of pages for the cpu. Note that @get_page_fn() must
* return the same number of pages for all cpus.
*
* @unit_size, if non-zero, determines unit size and must be aligned
* to PAGE_SIZE and equal to or larger than @static_size + @free_size.
*
* @free_size determines the number of free bytes after the static
* area in the first chunk. If zero, whatever left is available.
* Specifying non-zero value make percpu leave the area after
* @static_size + @free_size alone.
*
* Non-null @base_addr means that the caller already allocated virtual
* region for the first chunk and mapped it. percpu must not mess
* with the chunk. Note that @base_addr with 0 @unit_size or non-NULL
* @populate_pte_fn doesn't make any sense.
*
* @populate_pte_fn is used to populate the pagetable. NULL means the
* caller already populated the pagetable.
* *
* RETURNS: * RETURNS:
* The determined pcpu_unit_size which can be used to initialize * The determined pcpu_unit_size which can be used to initialize
* percpu access. * percpu access.
*/ */
size_t __init pcpu_setup_static(pcpu_populate_pte_fn_t populate_pte_fn, size_t __init pcpu_setup_first_chunk(pcpu_get_page_fn_t get_page_fn,
struct page **pages, size_t cpu_size) size_t static_size, size_t unit_size,
size_t free_size, void *base_addr,
pcpu_populate_pte_fn_t populate_pte_fn)
{ {
static struct vm_struct static_vm; static struct vm_struct static_vm;
struct pcpu_chunk *static_chunk; struct pcpu_chunk *static_chunk;
int nr_cpu_pages = DIV_ROUND_UP(cpu_size, PAGE_SIZE);
unsigned int cpu; unsigned int cpu;
int nr_pages;
int err, i; int err, i;
pcpu_unit_pages = max_t(int, PCPU_MIN_UNIT_PAGES, PFN_UP(cpu_size)); /* santiy checks */
BUG_ON(!static_size);
BUG_ON(!unit_size && free_size);
BUG_ON(unit_size && unit_size < static_size + free_size);
BUG_ON(unit_size & ~PAGE_MASK);
BUG_ON(base_addr && !unit_size);
BUG_ON(base_addr && populate_pte_fn);
pcpu_static_size = cpu_size; if (unit_size)
pcpu_unit_pages = unit_size >> PAGE_SHIFT;
else
pcpu_unit_pages = max_t(int, PCPU_MIN_UNIT_SIZE >> PAGE_SHIFT,
PFN_UP(static_size));
pcpu_static_size = static_size;
pcpu_unit_size = pcpu_unit_pages << PAGE_SHIFT; pcpu_unit_size = pcpu_unit_pages << PAGE_SHIFT;
pcpu_chunk_size = num_possible_cpus() * pcpu_unit_size; pcpu_chunk_size = num_possible_cpus() * pcpu_unit_size;
pcpu_chunk_struct_size = sizeof(struct pcpu_chunk) pcpu_chunk_struct_size = sizeof(struct pcpu_chunk)
...@@ -862,29 +908,66 @@ size_t __init pcpu_setup_static(pcpu_populate_pte_fn_t populate_pte_fn, ...@@ -862,29 +908,66 @@ size_t __init pcpu_setup_static(pcpu_populate_pte_fn_t populate_pte_fn,
for (i = 0; i < pcpu_nr_slots; i++) for (i = 0; i < pcpu_nr_slots; i++)
INIT_LIST_HEAD(&pcpu_slot[i]); INIT_LIST_HEAD(&pcpu_slot[i]);
/* init and register vm area */
static_vm.flags = VM_ALLOC;
static_vm.size = pcpu_chunk_size;
vm_area_register_early(&static_vm, PAGE_SIZE);
/* init static_chunk */ /* init static_chunk */
static_chunk = alloc_bootmem(pcpu_chunk_struct_size); static_chunk = alloc_bootmem(pcpu_chunk_struct_size);
INIT_LIST_HEAD(&static_chunk->list); INIT_LIST_HEAD(&static_chunk->list);
static_chunk->vm = &static_vm; static_chunk->vm = &static_vm;
static_chunk->free_size = pcpu_unit_size - pcpu_static_size;
if (free_size)
static_chunk->free_size = free_size;
else
static_chunk->free_size = pcpu_unit_size - pcpu_static_size;
static_chunk->contig_hint = static_chunk->free_size; static_chunk->contig_hint = static_chunk->free_size;
/* assign pages and map them */ /* allocate vm address */
static_vm.flags = VM_ALLOC;
static_vm.size = pcpu_chunk_size;
if (!base_addr)
vm_area_register_early(&static_vm, PAGE_SIZE);
else {
/*
* Pages already mapped. No need to remap into
* vmalloc area. In this case the static chunk can't
* be mapped or unmapped by percpu and is marked
* immutable.
*/
static_vm.addr = base_addr;
static_chunk->immutable = true;
}
/* assign pages */
nr_pages = -1;
for_each_possible_cpu(cpu) { for_each_possible_cpu(cpu) {
for (i = 0; i < nr_cpu_pages; i++) { for (i = 0; i < pcpu_unit_pages; i++) {
*pcpu_chunk_pagep(static_chunk, cpu, i) = *pages++; struct page *page = get_page_fn(cpu, i);
populate_pte_fn(pcpu_chunk_addr(static_chunk, cpu, i));
if (!page)
break;
*pcpu_chunk_pagep(static_chunk, cpu, i) = page;
} }
BUG_ON(i < PFN_UP(pcpu_static_size));
if (nr_pages < 0)
nr_pages = i;
else
BUG_ON(nr_pages != i);
} }
err = pcpu_map(static_chunk, 0, nr_cpu_pages); /* map them */
if (err) if (populate_pte_fn) {
panic("failed to setup static percpu area, err=%d\n", err); for_each_possible_cpu(cpu)
for (i = 0; i < nr_pages; i++)
populate_pte_fn(pcpu_chunk_addr(static_chunk,
cpu, i));
err = pcpu_map(static_chunk, 0, nr_pages);
if (err)
panic("failed to setup static percpu area, err=%d\n",
err);
}
/* link static_chunk in */ /* link static_chunk in */
pcpu_chunk_relocate(static_chunk, -1); pcpu_chunk_relocate(static_chunk, -1);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment