Commit 59807685 authored by Huang Ying's avatar Huang Ying Committed by Linus Torvalds

mm, THP, swap: support splitting THP for THP swap out

After adding swapping out support for THP (Transparent Huge Page), it is
possible that a THP in swap cache (partly swapped out) need to be split.
To split such a THP, the swap cluster backing the THP need to be split
too, that is, the CLUSTER_FLAG_HUGE flag need to be cleared for the swap
cluster.  The patch implemented this.

And because the THP swap writing needs the THP keeps as huge page during
writing.  The PageWriteback flag is checked before splitting.

Link: http://lkml.kernel.org/r/20170724051840.2309-8-ying.huang@intel.comSigned-off-by: default avatar"Huang, Ying" <ying.huang@intel.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Hugh Dickins <hughd@google.com>
Cc: Shaohua Li <shli@kernel.org>
Cc: Rik van Riel <riel@redhat.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: "Kirill A . Shutemov" <kirill.shutemov@linux.intel.com>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Ross Zwisler <ross.zwisler@intel.com> [for brd.c, zram_drv.c, pmem.c]
Cc: Vishal L Verma <vishal.l.verma@intel.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 225311a4
...@@ -527,6 +527,15 @@ static inline swp_entry_t get_swap_page(struct page *page) ...@@ -527,6 +527,15 @@ static inline swp_entry_t get_swap_page(struct page *page)
#endif /* CONFIG_SWAP */ #endif /* CONFIG_SWAP */
#ifdef CONFIG_THP_SWAP
extern int split_swap_cluster(swp_entry_t entry);
#else
static inline int split_swap_cluster(swp_entry_t entry)
{
return 0;
}
#endif
#ifdef CONFIG_MEMCG #ifdef CONFIG_MEMCG
static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg)
{ {
......
...@@ -2481,6 +2481,9 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) ...@@ -2481,6 +2481,9 @@ int split_huge_page_to_list(struct page *page, struct list_head *list)
VM_BUG_ON_PAGE(!PageLocked(page), page); VM_BUG_ON_PAGE(!PageLocked(page), page);
VM_BUG_ON_PAGE(!PageCompound(page), page); VM_BUG_ON_PAGE(!PageCompound(page), page);
if (PageWriteback(page))
return -EBUSY;
if (PageAnon(head)) { if (PageAnon(head)) {
/* /*
* The caller does not necessarily hold an mmap_sem that would * The caller does not necessarily hold an mmap_sem that would
...@@ -2558,6 +2561,11 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) ...@@ -2558,6 +2561,11 @@ int split_huge_page_to_list(struct page *page, struct list_head *list)
__dec_node_page_state(page, NR_SHMEM_THPS); __dec_node_page_state(page, NR_SHMEM_THPS);
spin_unlock(&pgdata->split_queue_lock); spin_unlock(&pgdata->split_queue_lock);
__split_huge_page(page, list, flags); __split_huge_page(page, list, flags);
if (PageSwapCache(head)) {
swp_entry_t entry = { .val = page_private(head) };
ret = split_swap_cluster(entry);
} else
ret = 0; ret = 0;
} else { } else {
if (IS_ENABLED(CONFIG_DEBUG_VM) && mapcount) { if (IS_ENABLED(CONFIG_DEBUG_VM) && mapcount) {
......
...@@ -1216,6 +1216,21 @@ static void swapcache_free_cluster(swp_entry_t entry) ...@@ -1216,6 +1216,21 @@ static void swapcache_free_cluster(swp_entry_t entry)
} }
} }
} }
int split_swap_cluster(swp_entry_t entry)
{
struct swap_info_struct *si;
struct swap_cluster_info *ci;
unsigned long offset = swp_offset(entry);
si = _swap_info_get(entry);
if (!si)
return -EBUSY;
ci = lock_cluster(si, offset);
cluster_clear_huge(ci);
unlock_cluster(ci);
return 0;
}
#else #else
static inline void swapcache_free_cluster(swp_entry_t entry) static inline void swapcache_free_cluster(swp_entry_t entry)
{ {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment