Commit cd39d4e9 authored by Muchun Song's avatar Muchun Song Committed by Linus Torvalds

mm: hugetlb: gather discrete indexes of tail page

For HugeTLB page, there are more metadata to save in the struct page.  But
the head struct page cannot meet our needs, so we have to abuse other tail
struct page to store the metadata.  In order to avoid conflicts caused by
subsequent use of more tail struct pages, we can gather these discrete
indexes of tail struct page.  In this case, it will be easier to add a new
tail page index later.

Link: https://lkml.kernel.org/r/20210510030027.56044-4-songmuchun@bytedance.comSigned-off-by: default avatarMuchun Song <songmuchun@bytedance.com>
Reviewed-by: default avatarOscar Salvador <osalvador@suse.de>
Reviewed-by: default avatarMiaohe Lin <linmiaohe@huawei.com>
Tested-by: default avatarChen Huang <chenhuang5@huawei.com>
Tested-by: default avatarBodeddula Balasubramaniam <bodeddub@amazon.com>
Acked-by: default avatarMichal Hocko <mhocko@suse.com>
Reviewed-by: default avatarMike Kravetz <mike.kravetz@oracle.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Andy Lutomirski <luto@kernel.org>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Balbir Singh <bsingharora@gmail.com>
Cc: Barry Song <song.bao.hua@hisilicon.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: David Rientjes <rientjes@google.com>
Cc: HORIGUCHI NAOYA <naoya.horiguchi@nec.com>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Joao Martins <joao.m.martins@oracle.com>
Cc: Joerg Roedel <jroedel@suse.de>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mina Almasry <almasrymina@google.com>
Cc: Oliver Neukum <oneukum@suse.com>
Cc: Paul E. McKenney <paulmck@kernel.org>
Cc: Pawan Gupta <pawan.kumar.gupta@linux.intel.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Randy Dunlap <rdunlap@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 6be24bed
...@@ -29,6 +29,23 @@ typedef struct { unsigned long pd; } hugepd_t; ...@@ -29,6 +29,23 @@ typedef struct { unsigned long pd; } hugepd_t;
#include <linux/shm.h> #include <linux/shm.h>
#include <asm/tlbflush.h> #include <asm/tlbflush.h>
/*
* For HugeTLB page, there are more metadata to save in the struct page. But
* the head struct page cannot meet our needs, so we have to abuse other tail
* struct page to store the metadata. In order to avoid conflicts caused by
* subsequent use of more tail struct pages, we gather these discrete indexes
* of tail struct page here.
*/
enum {
SUBPAGE_INDEX_SUBPOOL = 1, /* reuse page->private */
#ifdef CONFIG_CGROUP_HUGETLB
SUBPAGE_INDEX_CGROUP, /* reuse page->private */
SUBPAGE_INDEX_CGROUP_RSVD, /* reuse page->private */
__MAX_CGROUP_SUBPAGE_INDEX = SUBPAGE_INDEX_CGROUP_RSVD,
#endif
__NR_USED_SUBPAGE,
};
struct hugepage_subpool { struct hugepage_subpool {
spinlock_t lock; spinlock_t lock;
long count; long count;
...@@ -635,13 +652,13 @@ extern unsigned int default_hstate_idx; ...@@ -635,13 +652,13 @@ extern unsigned int default_hstate_idx;
*/ */
static inline struct hugepage_subpool *hugetlb_page_subpool(struct page *hpage) static inline struct hugepage_subpool *hugetlb_page_subpool(struct page *hpage)
{ {
return (struct hugepage_subpool *)(hpage+1)->private; return (void *)page_private(hpage + SUBPAGE_INDEX_SUBPOOL);
} }
static inline void hugetlb_set_page_subpool(struct page *hpage, static inline void hugetlb_set_page_subpool(struct page *hpage,
struct hugepage_subpool *subpool) struct hugepage_subpool *subpool)
{ {
set_page_private(hpage+1, (unsigned long)subpool); set_page_private(hpage + SUBPAGE_INDEX_SUBPOOL, (unsigned long)subpool);
} }
static inline struct hstate *hstate_file(struct file *f) static inline struct hstate *hstate_file(struct file *f)
......
...@@ -21,15 +21,16 @@ struct hugetlb_cgroup; ...@@ -21,15 +21,16 @@ struct hugetlb_cgroup;
struct resv_map; struct resv_map;
struct file_region; struct file_region;
#ifdef CONFIG_CGROUP_HUGETLB
/* /*
* Minimum page order trackable by hugetlb cgroup. * Minimum page order trackable by hugetlb cgroup.
* At least 4 pages are necessary for all the tracking information. * At least 4 pages are necessary for all the tracking information.
* The second tail page (hpage[2]) is the fault usage cgroup. * The second tail page (hpage[SUBPAGE_INDEX_CGROUP]) is the fault
* The third tail page (hpage[3]) is the reservation usage cgroup. * usage cgroup. The third tail page (hpage[SUBPAGE_INDEX_CGROUP_RSVD])
* is the reservation usage cgroup.
*/ */
#define HUGETLB_CGROUP_MIN_ORDER 2 #define HUGETLB_CGROUP_MIN_ORDER order_base_2(__MAX_CGROUP_SUBPAGE_INDEX + 1)
#ifdef CONFIG_CGROUP_HUGETLB
enum hugetlb_memory_event { enum hugetlb_memory_event {
HUGETLB_MAX, HUGETLB_MAX,
HUGETLB_NR_MEMORY_EVENTS, HUGETLB_NR_MEMORY_EVENTS,
...@@ -66,9 +67,9 @@ __hugetlb_cgroup_from_page(struct page *page, bool rsvd) ...@@ -66,9 +67,9 @@ __hugetlb_cgroup_from_page(struct page *page, bool rsvd)
if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER)
return NULL; return NULL;
if (rsvd) if (rsvd)
return (struct hugetlb_cgroup *)page[3].private; return (void *)page_private(page + SUBPAGE_INDEX_CGROUP_RSVD);
else else
return (struct hugetlb_cgroup *)page[2].private; return (void *)page_private(page + SUBPAGE_INDEX_CGROUP);
} }
static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page)
...@@ -90,9 +91,11 @@ static inline int __set_hugetlb_cgroup(struct page *page, ...@@ -90,9 +91,11 @@ static inline int __set_hugetlb_cgroup(struct page *page,
if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER)
return -1; return -1;
if (rsvd) if (rsvd)
page[3].private = (unsigned long)h_cg; set_page_private(page + SUBPAGE_INDEX_CGROUP_RSVD,
(unsigned long)h_cg);
else else
page[2].private = (unsigned long)h_cg; set_page_private(page + SUBPAGE_INDEX_CGROUP,
(unsigned long)h_cg);
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment