Commit 4f42047b authored by Minchan Kim's avatar Minchan Kim Committed by Linus Torvalds

zsmalloc: use accessor

Upcoming patch will change how to encode zspage meta so for easy review,
this patch wraps code to access metadata as accessor.

Link: http://lkml.kernel.org/r/1464736881-24886-7-git-send-email-minchan@kernel.orgSigned-off-by: default avatarMinchan Kim <minchan@kernel.org>
Reviewed-by: default avatarSergey Senozhatsky <sergey.senozhatsky@gmail.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 1b8320b6
...@@ -268,10 +268,14 @@ struct zs_pool { ...@@ -268,10 +268,14 @@ struct zs_pool {
* A zspage's class index and fullness group * A zspage's class index and fullness group
* are encoded in its (first)page->mapping * are encoded in its (first)page->mapping
*/ */
#define CLASS_IDX_BITS 28
#define FULLNESS_BITS 4 #define FULLNESS_BITS 4
#define CLASS_IDX_MASK ((1 << CLASS_IDX_BITS) - 1) #define CLASS_BITS 28
#define FULLNESS_MASK ((1 << FULLNESS_BITS) - 1)
#define FULLNESS_SHIFT 0
#define CLASS_SHIFT (FULLNESS_SHIFT + FULLNESS_BITS)
#define FULLNESS_MASK ((1UL << FULLNESS_BITS) - 1)
#define CLASS_MASK ((1UL << CLASS_BITS) - 1)
struct mapping_area { struct mapping_area {
#ifdef CONFIG_PGTABLE_MAPPING #ifdef CONFIG_PGTABLE_MAPPING
...@@ -418,6 +422,41 @@ static int is_last_page(struct page *page) ...@@ -418,6 +422,41 @@ static int is_last_page(struct page *page)
return PagePrivate2(page); return PagePrivate2(page);
} }
static inline int get_zspage_inuse(struct page *first_page)
{
return first_page->inuse;
}
static inline void set_zspage_inuse(struct page *first_page, int val)
{
first_page->inuse = val;
}
static inline void mod_zspage_inuse(struct page *first_page, int val)
{
first_page->inuse += val;
}
static inline int get_first_obj_offset(struct page *page)
{
return page->index;
}
static inline void set_first_obj_offset(struct page *page, int offset)
{
page->index = offset;
}
static inline unsigned long get_freeobj(struct page *first_page)
{
return (unsigned long)first_page->freelist;
}
static inline void set_freeobj(struct page *first_page, unsigned long obj)
{
first_page->freelist = (void *)obj;
}
static void get_zspage_mapping(struct page *first_page, static void get_zspage_mapping(struct page *first_page,
unsigned int *class_idx, unsigned int *class_idx,
enum fullness_group *fullness) enum fullness_group *fullness)
...@@ -426,8 +465,8 @@ static void get_zspage_mapping(struct page *first_page, ...@@ -426,8 +465,8 @@ static void get_zspage_mapping(struct page *first_page,
VM_BUG_ON_PAGE(!is_first_page(first_page), first_page); VM_BUG_ON_PAGE(!is_first_page(first_page), first_page);
m = (unsigned long)first_page->mapping; m = (unsigned long)first_page->mapping;
*fullness = m & FULLNESS_MASK; *fullness = (m >> FULLNESS_SHIFT) & FULLNESS_MASK;
*class_idx = (m >> FULLNESS_BITS) & CLASS_IDX_MASK; *class_idx = (m >> CLASS_SHIFT) & CLASS_MASK;
} }
static void set_zspage_mapping(struct page *first_page, static void set_zspage_mapping(struct page *first_page,
...@@ -437,8 +476,7 @@ static void set_zspage_mapping(struct page *first_page, ...@@ -437,8 +476,7 @@ static void set_zspage_mapping(struct page *first_page,
unsigned long m; unsigned long m;
VM_BUG_ON_PAGE(!is_first_page(first_page), first_page); VM_BUG_ON_PAGE(!is_first_page(first_page), first_page);
m = ((class_idx & CLASS_IDX_MASK) << FULLNESS_BITS) | m = (class_idx << CLASS_SHIFT) | (fullness << FULLNESS_SHIFT);
(fullness & FULLNESS_MASK);
first_page->mapping = (struct address_space *)m; first_page->mapping = (struct address_space *)m;
} }
...@@ -638,7 +676,7 @@ static enum fullness_group get_fullness_group(struct size_class *class, ...@@ -638,7 +676,7 @@ static enum fullness_group get_fullness_group(struct size_class *class,
VM_BUG_ON_PAGE(!is_first_page(first_page), first_page); VM_BUG_ON_PAGE(!is_first_page(first_page), first_page);
inuse = first_page->inuse; inuse = get_zspage_inuse(first_page);
objs_per_zspage = class->objs_per_zspage; objs_per_zspage = class->objs_per_zspage;
if (inuse == 0) if (inuse == 0)
...@@ -684,7 +722,7 @@ static void insert_zspage(struct size_class *class, ...@@ -684,7 +722,7 @@ static void insert_zspage(struct size_class *class,
* empty/full. Put pages with higher ->inuse first. * empty/full. Put pages with higher ->inuse first.
*/ */
list_add_tail(&first_page->lru, &(*head)->lru); list_add_tail(&first_page->lru, &(*head)->lru);
if (first_page->inuse >= (*head)->inuse) if (get_zspage_inuse(first_page) >= get_zspage_inuse(*head))
*head = first_page; *head = first_page;
} }
...@@ -861,7 +899,7 @@ static unsigned long obj_idx_to_offset(struct page *page, ...@@ -861,7 +899,7 @@ static unsigned long obj_idx_to_offset(struct page *page,
unsigned long off = 0; unsigned long off = 0;
if (!is_first_page(page)) if (!is_first_page(page))
off = page->index; off = get_first_obj_offset(page);
return off + obj_idx * class_size; return off + obj_idx * class_size;
} }
...@@ -896,7 +934,7 @@ static void free_zspage(struct page *first_page) ...@@ -896,7 +934,7 @@ static void free_zspage(struct page *first_page)
struct page *nextp, *tmp, *head_extra; struct page *nextp, *tmp, *head_extra;
VM_BUG_ON_PAGE(!is_first_page(first_page), first_page); VM_BUG_ON_PAGE(!is_first_page(first_page), first_page);
VM_BUG_ON_PAGE(first_page->inuse, first_page); VM_BUG_ON_PAGE(get_zspage_inuse(first_page), first_page);
head_extra = (struct page *)page_private(first_page); head_extra = (struct page *)page_private(first_page);
...@@ -937,7 +975,7 @@ static void init_zspage(struct size_class *class, struct page *first_page) ...@@ -937,7 +975,7 @@ static void init_zspage(struct size_class *class, struct page *first_page)
* head of corresponding zspage's freelist. * head of corresponding zspage's freelist.
*/ */
if (page != first_page) if (page != first_page)
page->index = off; set_first_obj_offset(page, off);
vaddr = kmap_atomic(page); vaddr = kmap_atomic(page);
link = (struct link_free *)vaddr + off / sizeof(*link); link = (struct link_free *)vaddr + off / sizeof(*link);
...@@ -992,7 +1030,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags) ...@@ -992,7 +1030,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags)
SetPagePrivate(page); SetPagePrivate(page);
set_page_private(page, 0); set_page_private(page, 0);
first_page = page; first_page = page;
first_page->inuse = 0; set_zspage_inuse(first_page, 0);
} }
if (i == 1) if (i == 1)
set_page_private(first_page, (unsigned long)page); set_page_private(first_page, (unsigned long)page);
...@@ -1007,7 +1045,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags) ...@@ -1007,7 +1045,7 @@ static struct page *alloc_zspage(struct size_class *class, gfp_t flags)
init_zspage(class, first_page); init_zspage(class, first_page);
first_page->freelist = location_to_obj(first_page, 0); set_freeobj(first_page, (unsigned long)location_to_obj(first_page, 0));
error = 0; /* Success */ error = 0; /* Success */
cleanup: cleanup:
...@@ -1239,7 +1277,7 @@ static bool zspage_full(struct size_class *class, struct page *first_page) ...@@ -1239,7 +1277,7 @@ static bool zspage_full(struct size_class *class, struct page *first_page)
{ {
VM_BUG_ON_PAGE(!is_first_page(first_page), first_page); VM_BUG_ON_PAGE(!is_first_page(first_page), first_page);
return first_page->inuse == class->objs_per_zspage; return get_zspage_inuse(first_page) == class->objs_per_zspage;
} }
unsigned long zs_get_total_pages(struct zs_pool *pool) unsigned long zs_get_total_pages(struct zs_pool *pool)
...@@ -1358,13 +1396,13 @@ static unsigned long obj_malloc(struct size_class *class, ...@@ -1358,13 +1396,13 @@ static unsigned long obj_malloc(struct size_class *class,
void *vaddr; void *vaddr;
handle |= OBJ_ALLOCATED_TAG; handle |= OBJ_ALLOCATED_TAG;
obj = (unsigned long)first_page->freelist; obj = get_freeobj(first_page);
obj_to_location(obj, &m_page, &m_objidx); obj_to_location(obj, &m_page, &m_objidx);
m_offset = obj_idx_to_offset(m_page, m_objidx, class->size); m_offset = obj_idx_to_offset(m_page, m_objidx, class->size);
vaddr = kmap_atomic(m_page); vaddr = kmap_atomic(m_page);
link = (struct link_free *)vaddr + m_offset / sizeof(*link); link = (struct link_free *)vaddr + m_offset / sizeof(*link);
first_page->freelist = link->next; set_freeobj(first_page, (unsigned long)link->next);
if (!class->huge) if (!class->huge)
/* record handle in the header of allocated chunk */ /* record handle in the header of allocated chunk */
link->handle = handle; link->handle = handle;
...@@ -1372,7 +1410,7 @@ static unsigned long obj_malloc(struct size_class *class, ...@@ -1372,7 +1410,7 @@ static unsigned long obj_malloc(struct size_class *class,
/* record handle in first_page->private */ /* record handle in first_page->private */
set_page_private(first_page, handle); set_page_private(first_page, handle);
kunmap_atomic(vaddr); kunmap_atomic(vaddr);
first_page->inuse++; mod_zspage_inuse(first_page, 1);
zs_stat_inc(class, OBJ_USED, 1); zs_stat_inc(class, OBJ_USED, 1);
return obj; return obj;
...@@ -1452,12 +1490,12 @@ static void obj_free(struct size_class *class, unsigned long obj) ...@@ -1452,12 +1490,12 @@ static void obj_free(struct size_class *class, unsigned long obj)
/* Insert this object in containing zspage's freelist */ /* Insert this object in containing zspage's freelist */
link = (struct link_free *)(vaddr + f_offset); link = (struct link_free *)(vaddr + f_offset);
link->next = first_page->freelist; link->next = (void *)get_freeobj(first_page);
if (class->huge) if (class->huge)
set_page_private(first_page, 0); set_page_private(first_page, 0);
kunmap_atomic(vaddr); kunmap_atomic(vaddr);
first_page->freelist = (void *)obj; set_freeobj(first_page, obj);
first_page->inuse--; mod_zspage_inuse(first_page, -1);
zs_stat_dec(class, OBJ_USED, 1); zs_stat_dec(class, OBJ_USED, 1);
} }
...@@ -1573,7 +1611,7 @@ static unsigned long find_alloced_obj(struct size_class *class, ...@@ -1573,7 +1611,7 @@ static unsigned long find_alloced_obj(struct size_class *class,
void *addr = kmap_atomic(page); void *addr = kmap_atomic(page);
if (!is_first_page(page)) if (!is_first_page(page))
offset = page->index; offset = get_first_obj_offset(page);
offset += class->size * index; offset += class->size * index;
while (offset < PAGE_SIZE) { while (offset < PAGE_SIZE) {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment