Commit 64f768c6 authored by Nhat Pham's avatar Nhat Pham Committed by Andrew Morton

zsmalloc: add a LRU to zs_pool to keep track of zspages in LRU order

This helps determines the coldest zspages as candidates for writeback.

Link: https://lkml.kernel.org/r/20221128191616.1261026-5-nphamcs@gmail.comSigned-off-by: default avatarNhat Pham <nphamcs@gmail.com>
Acked-by: default avatarJohannes Weiner <hannes@cmpxchg.org>
Reviewed-by: default avatarSergey Senozhatsky <senozhatsky@chromium.org>
Cc: Dan Streetman <ddstreet@ieee.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Nitin Gupta <ngupta@vflare.org>
Cc: Seth Jennings <sjenning@redhat.com>
Cc: Vitaly Wool <vitaly.wool@konsulko.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent c0547d0b
...@@ -239,6 +239,11 @@ struct zs_pool { ...@@ -239,6 +239,11 @@ struct zs_pool {
/* Compact classes */ /* Compact classes */
struct shrinker shrinker; struct shrinker shrinker;
#ifdef CONFIG_ZPOOL
/* List tracking the zspages in LRU order by most recently added object */
struct list_head lru;
#endif
#ifdef CONFIG_ZSMALLOC_STAT #ifdef CONFIG_ZSMALLOC_STAT
struct dentry *stat_dentry; struct dentry *stat_dentry;
#endif #endif
...@@ -260,6 +265,12 @@ struct zspage { ...@@ -260,6 +265,12 @@ struct zspage {
unsigned int freeobj; unsigned int freeobj;
struct page *first_page; struct page *first_page;
struct list_head list; /* fullness list */ struct list_head list; /* fullness list */
#ifdef CONFIG_ZPOOL
/* links the zspage to the lru list in the pool */
struct list_head lru;
#endif
struct zs_pool *pool; struct zs_pool *pool;
#ifdef CONFIG_COMPACTION #ifdef CONFIG_COMPACTION
rwlock_t lock; rwlock_t lock;
...@@ -953,6 +964,9 @@ static void free_zspage(struct zs_pool *pool, struct size_class *class, ...@@ -953,6 +964,9 @@ static void free_zspage(struct zs_pool *pool, struct size_class *class,
} }
remove_zspage(class, zspage, ZS_EMPTY); remove_zspage(class, zspage, ZS_EMPTY);
#ifdef CONFIG_ZPOOL
list_del(&zspage->lru);
#endif
__free_zspage(pool, class, zspage); __free_zspage(pool, class, zspage);
} }
...@@ -998,6 +1012,10 @@ static void init_zspage(struct size_class *class, struct zspage *zspage) ...@@ -998,6 +1012,10 @@ static void init_zspage(struct size_class *class, struct zspage *zspage)
off %= PAGE_SIZE; off %= PAGE_SIZE;
} }
#ifdef CONFIG_ZPOOL
INIT_LIST_HEAD(&zspage->lru);
#endif
set_freeobj(zspage, 0); set_freeobj(zspage, 0);
} }
...@@ -1270,6 +1288,31 @@ void *zs_map_object(struct zs_pool *pool, unsigned long handle, ...@@ -1270,6 +1288,31 @@ void *zs_map_object(struct zs_pool *pool, unsigned long handle,
obj_to_location(obj, &page, &obj_idx); obj_to_location(obj, &page, &obj_idx);
zspage = get_zspage(page); zspage = get_zspage(page);
#ifdef CONFIG_ZPOOL
/*
* Move the zspage to front of pool's LRU.
*
* Note that this is swap-specific, so by definition there are no ongoing
* accesses to the memory while the page is swapped out that would make
* it "hot". A new entry is hot, then ages to the tail until it gets either
* written back or swaps back in.
*
* Furthermore, map is also called during writeback. We must not put an
* isolated page on the LRU mid-reclaim.
*
* As a result, only update the LRU when the page is mapped for write
* when it's first instantiated.
*
* This is a deviation from the other backends, which perform this update
* in the allocation function (zbud_alloc, z3fold_alloc).
*/
if (mm == ZS_MM_WO) {
if (!list_empty(&zspage->lru))
list_del(&zspage->lru);
list_add(&zspage->lru, &pool->lru);
}
#endif
/* /*
* migration cannot move any zpages in this zspage. Here, pool->lock * migration cannot move any zpages in this zspage. Here, pool->lock
* is too heavy since callers would take some time until they calls * is too heavy since callers would take some time until they calls
...@@ -1988,6 +2031,9 @@ static void async_free_zspage(struct work_struct *work) ...@@ -1988,6 +2031,9 @@ static void async_free_zspage(struct work_struct *work)
VM_BUG_ON(fullness != ZS_EMPTY); VM_BUG_ON(fullness != ZS_EMPTY);
class = pool->size_class[class_idx]; class = pool->size_class[class_idx];
spin_lock(&pool->lock); spin_lock(&pool->lock);
#ifdef CONFIG_ZPOOL
list_del(&zspage->lru);
#endif
__free_zspage(pool, class, zspage); __free_zspage(pool, class, zspage);
spin_unlock(&pool->lock); spin_unlock(&pool->lock);
} }
...@@ -2299,6 +2345,10 @@ struct zs_pool *zs_create_pool(const char *name) ...@@ -2299,6 +2345,10 @@ struct zs_pool *zs_create_pool(const char *name)
*/ */
zs_register_shrinker(pool); zs_register_shrinker(pool);
#ifdef CONFIG_ZPOOL
INIT_LIST_HEAD(&pool->lru);
#endif
return pool; return pool;
err: err:
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment