Commit c62d2555 authored by Michal Hocko's avatar Michal Hocko Committed by Linus Torvalds

mm, fs: introduce mapping_gfp_constraint()

There are many places which use mapping_gfp_mask to restrict a more
generic gfp mask which would be used for allocations which are not
directly related to the page cache but they are performed in the same
context.

Let's introduce a helper function which makes the restriction explicit and
easier to track.  This patch doesn't introduce any functional changes.

[akpm@linux-foundation.org: coding-style fixes]
Signed-off-by: default avatarMichal Hocko <mhocko@suse.com>
Suggested-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 89903327
...@@ -491,7 +491,7 @@ struct page **drm_gem_get_pages(struct drm_gem_object *obj) ...@@ -491,7 +491,7 @@ struct page **drm_gem_get_pages(struct drm_gem_object *obj)
* __GFP_DMA32 to be set in mapping_gfp_mask(inode->i_mapping) * __GFP_DMA32 to be set in mapping_gfp_mask(inode->i_mapping)
* so shmem can relocate pages during swapin if required. * so shmem can relocate pages during swapin if required.
*/ */
BUG_ON((mapping_gfp_mask(mapping) & __GFP_DMA32) && BUG_ON(mapping_gfp_constraint(mapping, __GFP_DMA32) &&
(page_to_pfn(p) >= 0x00100000UL)); (page_to_pfn(p) >= 0x00100000UL));
} }
......
...@@ -2214,9 +2214,8 @@ i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj) ...@@ -2214,9 +2214,8 @@ i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj)
* Fail silently without starting the shrinker * Fail silently without starting the shrinker
*/ */
mapping = file_inode(obj->base.filp)->i_mapping; mapping = file_inode(obj->base.filp)->i_mapping;
gfp = mapping_gfp_mask(mapping); gfp = mapping_gfp_constraint(mapping, ~(__GFP_IO | __GFP_RECLAIM));
gfp |= __GFP_NORETRY | __GFP_NOWARN; gfp |= __GFP_NORETRY | __GFP_NOWARN;
gfp &= ~(__GFP_IO | __GFP_RECLAIM);
sg = st->sgl; sg = st->sgl;
st->nents = 0; st->nents = 0;
for (i = 0; i < page_count; i++) { for (i = 0; i < page_count; i++) {
......
...@@ -482,13 +482,12 @@ static noinline int add_ra_bio_pages(struct inode *inode, ...@@ -482,13 +482,12 @@ static noinline int add_ra_bio_pages(struct inode *inode,
goto next; goto next;
} }
page = __page_cache_alloc(mapping_gfp_mask(mapping) & page = __page_cache_alloc(mapping_gfp_constraint(mapping,
~__GFP_FS); ~__GFP_FS));
if (!page) if (!page)
break; break;
if (add_to_page_cache_lru(page, mapping, pg_index, if (add_to_page_cache_lru(page, mapping, pg_index, GFP_NOFS)) {
GFP_NOFS)) {
page_cache_release(page); page_cache_release(page);
goto next; goto next;
} }
......
...@@ -3316,7 +3316,7 @@ static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info) ...@@ -3316,7 +3316,7 @@ static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info)
static inline gfp_t btrfs_alloc_write_mask(struct address_space *mapping) static inline gfp_t btrfs_alloc_write_mask(struct address_space *mapping)
{ {
return mapping_gfp_mask(mapping) & ~__GFP_FS; return mapping_gfp_constraint(mapping, ~__GFP_FS);
} }
/* extent-tree.c */ /* extent-tree.c */
......
...@@ -85,8 +85,8 @@ static struct inode *__lookup_free_space_inode(struct btrfs_root *root, ...@@ -85,8 +85,8 @@ static struct inode *__lookup_free_space_inode(struct btrfs_root *root,
} }
mapping_set_gfp_mask(inode->i_mapping, mapping_set_gfp_mask(inode->i_mapping,
mapping_gfp_mask(inode->i_mapping) & mapping_gfp_constraint(inode->i_mapping,
~(__GFP_FS | __GFP_HIGHMEM)); ~(__GFP_FS | __GFP_HIGHMEM)));
return inode; return inode;
} }
......
...@@ -999,7 +999,7 @@ grow_dev_page(struct block_device *bdev, sector_t block, ...@@ -999,7 +999,7 @@ grow_dev_page(struct block_device *bdev, sector_t block,
int ret = 0; /* Will call free_more_memory() */ int ret = 0; /* Will call free_more_memory() */
gfp_t gfp_mask; gfp_t gfp_mask;
gfp_mask = (mapping_gfp_mask(inode->i_mapping) & ~__GFP_FS) | gfp; gfp_mask = mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS) | gfp;
/* /*
* XXX: __getblk_slow() can not really deal with failure and * XXX: __getblk_slow() can not really deal with failure and
......
...@@ -1283,8 +1283,8 @@ static int ceph_filemap_fault(struct vm_area_struct *vma, struct vm_fault *vmf) ...@@ -1283,8 +1283,8 @@ static int ceph_filemap_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
int ret1; int ret1;
struct address_space *mapping = inode->i_mapping; struct address_space *mapping = inode->i_mapping;
struct page *page = find_or_create_page(mapping, 0, struct page *page = find_or_create_page(mapping, 0,
mapping_gfp_mask(mapping) & mapping_gfp_constraint(mapping,
~__GFP_FS); ~__GFP_FS));
if (!page) { if (!page) {
ret = VM_FAULT_OOM; ret = VM_FAULT_OOM;
goto out; goto out;
...@@ -1428,7 +1428,8 @@ void ceph_fill_inline_data(struct inode *inode, struct page *locked_page, ...@@ -1428,7 +1428,8 @@ void ceph_fill_inline_data(struct inode *inode, struct page *locked_page,
if (i_size_read(inode) == 0) if (i_size_read(inode) == 0)
return; return;
page = find_or_create_page(mapping, 0, page = find_or_create_page(mapping, 0,
mapping_gfp_mask(mapping) & ~__GFP_FS); mapping_gfp_constraint(mapping,
~__GFP_FS));
if (!page) if (!page)
return; return;
if (PageUptodate(page)) { if (PageUptodate(page)) {
......
...@@ -3380,7 +3380,7 @@ readpages_get_pages(struct address_space *mapping, struct list_head *page_list, ...@@ -3380,7 +3380,7 @@ readpages_get_pages(struct address_space *mapping, struct list_head *page_list,
struct page *page, *tpage; struct page *page, *tpage;
unsigned int expected_index; unsigned int expected_index;
int rc; int rc;
gfp_t gfp = GFP_KERNEL & mapping_gfp_mask(mapping); gfp_t gfp = mapping_gfp_constraint(mapping, GFP_KERNEL);
INIT_LIST_HEAD(tmplist); INIT_LIST_HEAD(tmplist);
......
...@@ -3344,7 +3344,7 @@ static int __ext4_block_zero_page_range(handle_t *handle, ...@@ -3344,7 +3344,7 @@ static int __ext4_block_zero_page_range(handle_t *handle,
int err = 0; int err = 0;
page = find_or_create_page(mapping, from >> PAGE_CACHE_SHIFT, page = find_or_create_page(mapping, from >> PAGE_CACHE_SHIFT,
mapping_gfp_mask(mapping) & ~__GFP_FS); mapping_gfp_constraint(mapping, ~__GFP_FS));
if (!page) if (!page)
return -ENOMEM; return -ENOMEM;
......
...@@ -166,7 +166,7 @@ int ext4_mpage_readpages(struct address_space *mapping, ...@@ -166,7 +166,7 @@ int ext4_mpage_readpages(struct address_space *mapping,
page = list_entry(pages->prev, struct page, lru); page = list_entry(pages->prev, struct page, lru);
list_del(&page->lru); list_del(&page->lru);
if (add_to_page_cache_lru(page, mapping, page->index, if (add_to_page_cache_lru(page, mapping, page->index,
GFP_KERNEL & mapping_gfp_mask(mapping))) mapping_gfp_constraint(mapping, GFP_KERNEL)))
goto next_page; goto next_page;
} }
......
...@@ -57,7 +57,7 @@ static struct page *get_mapping_page(struct super_block *sb, pgoff_t index, ...@@ -57,7 +57,7 @@ static struct page *get_mapping_page(struct super_block *sb, pgoff_t index,
filler_t *filler = super->s_devops->readpage; filler_t *filler = super->s_devops->readpage;
struct page *page; struct page *page;
BUG_ON(mapping_gfp_mask(mapping) & __GFP_FS); BUG_ON(mapping_gfp_constraint(mapping, __GFP_FS));
if (use_filler) if (use_filler)
page = read_cache_page(mapping, index, filler, sb); page = read_cache_page(mapping, index, filler, sb);
else { else {
......
...@@ -361,7 +361,7 @@ mpage_readpages(struct address_space *mapping, struct list_head *pages, ...@@ -361,7 +361,7 @@ mpage_readpages(struct address_space *mapping, struct list_head *pages,
sector_t last_block_in_bio = 0; sector_t last_block_in_bio = 0;
struct buffer_head map_bh; struct buffer_head map_bh;
unsigned long first_logical_block = 0; unsigned long first_logical_block = 0;
gfp_t gfp = GFP_KERNEL & mapping_gfp_mask(mapping); gfp_t gfp = mapping_gfp_constraint(mapping, GFP_KERNEL);
map_bh.b_state = 0; map_bh.b_state = 0;
map_bh.b_size = 0; map_bh.b_size = 0;
...@@ -397,7 +397,7 @@ int mpage_readpage(struct page *page, get_block_t get_block) ...@@ -397,7 +397,7 @@ int mpage_readpage(struct page *page, get_block_t get_block)
sector_t last_block_in_bio = 0; sector_t last_block_in_bio = 0;
struct buffer_head map_bh; struct buffer_head map_bh;
unsigned long first_logical_block = 0; unsigned long first_logical_block = 0;
gfp_t gfp = GFP_KERNEL & mapping_gfp_mask(page->mapping); gfp_t gfp = mapping_gfp_constraint(page->mapping, GFP_KERNEL);
map_bh.b_state = 0; map_bh.b_state = 0;
map_bh.b_size = 0; map_bh.b_size = 0;
......
...@@ -4604,7 +4604,7 @@ EXPORT_SYMBOL(__page_symlink); ...@@ -4604,7 +4604,7 @@ EXPORT_SYMBOL(__page_symlink);
int page_symlink(struct inode *inode, const char *symname, int len) int page_symlink(struct inode *inode, const char *symname, int len)
{ {
return __page_symlink(inode, symname, len, return __page_symlink(inode, symname, len,
!(mapping_gfp_mask(inode->i_mapping) & __GFP_FS)); !mapping_gfp_constraint(inode->i_mapping, __GFP_FS));
} }
EXPORT_SYMBOL(page_symlink); EXPORT_SYMBOL(page_symlink);
......
...@@ -356,7 +356,7 @@ struct inode *nilfs_new_inode(struct inode *dir, umode_t mode) ...@@ -356,7 +356,7 @@ struct inode *nilfs_new_inode(struct inode *dir, umode_t mode)
goto failed; goto failed;
mapping_set_gfp_mask(inode->i_mapping, mapping_set_gfp_mask(inode->i_mapping,
mapping_gfp_mask(inode->i_mapping) & ~__GFP_FS); mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS));
root = NILFS_I(dir)->i_root; root = NILFS_I(dir)->i_root;
ii = NILFS_I(inode); ii = NILFS_I(inode);
...@@ -522,7 +522,7 @@ static int __nilfs_read_inode(struct super_block *sb, ...@@ -522,7 +522,7 @@ static int __nilfs_read_inode(struct super_block *sb,
up_read(&NILFS_MDT(nilfs->ns_dat)->mi_sem); up_read(&NILFS_MDT(nilfs->ns_dat)->mi_sem);
nilfs_set_inode_flags(inode); nilfs_set_inode_flags(inode);
mapping_set_gfp_mask(inode->i_mapping, mapping_set_gfp_mask(inode->i_mapping,
mapping_gfp_mask(inode->i_mapping) & ~__GFP_FS); mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS));
return 0; return 0;
failed_unmap: failed_unmap:
......
...@@ -525,8 +525,8 @@ static inline int __ntfs_grab_cache_pages(struct address_space *mapping, ...@@ -525,8 +525,8 @@ static inline int __ntfs_grab_cache_pages(struct address_space *mapping,
} }
} }
err = add_to_page_cache_lru(*cached_page, mapping, err = add_to_page_cache_lru(*cached_page, mapping,
index, index,
GFP_KERNEL & mapping_gfp_mask(mapping)); mapping_gfp_constraint(mapping, GFP_KERNEL));
if (unlikely(err)) { if (unlikely(err)) {
if (err == -EEXIST) if (err == -EEXIST)
continue; continue;
......
...@@ -360,7 +360,7 @@ __generic_file_splice_read(struct file *in, loff_t *ppos, ...@@ -360,7 +360,7 @@ __generic_file_splice_read(struct file *in, loff_t *ppos,
break; break;
error = add_to_page_cache_lru(page, mapping, index, error = add_to_page_cache_lru(page, mapping, index,
GFP_KERNEL & mapping_gfp_mask(mapping)); mapping_gfp_constraint(mapping, GFP_KERNEL));
if (unlikely(error)) { if (unlikely(error)) {
page_cache_release(page); page_cache_release(page);
if (error == -EEXIST) if (error == -EEXIST)
......
...@@ -69,6 +69,13 @@ static inline gfp_t mapping_gfp_mask(struct address_space * mapping) ...@@ -69,6 +69,13 @@ static inline gfp_t mapping_gfp_mask(struct address_space * mapping)
return (__force gfp_t)mapping->flags & __GFP_BITS_MASK; return (__force gfp_t)mapping->flags & __GFP_BITS_MASK;
} }
/* Restricts the given gfp_mask to what the mapping allows. */
static inline gfp_t mapping_gfp_constraint(struct address_space *mapping,
gfp_t gfp_mask)
{
return mapping_gfp_mask(mapping) & gfp_mask;
}
/* /*
* This is non-atomic. Only to be used before the mapping is activated. * This is non-atomic. Only to be used before the mapping is activated.
* Probably needs a barrier... * Probably needs a barrier...
......
...@@ -1722,7 +1722,7 @@ static ssize_t do_generic_file_read(struct file *filp, loff_t *ppos, ...@@ -1722,7 +1722,7 @@ static ssize_t do_generic_file_read(struct file *filp, loff_t *ppos,
goto out; goto out;
} }
error = add_to_page_cache_lru(page, mapping, index, error = add_to_page_cache_lru(page, mapping, index,
GFP_KERNEL & mapping_gfp_mask(mapping)); mapping_gfp_constraint(mapping, GFP_KERNEL));
if (error) { if (error) {
page_cache_release(page); page_cache_release(page);
if (error == -EEXIST) { if (error == -EEXIST) {
...@@ -1824,7 +1824,7 @@ static int page_cache_read(struct file *file, pgoff_t offset) ...@@ -1824,7 +1824,7 @@ static int page_cache_read(struct file *file, pgoff_t offset)
return -ENOMEM; return -ENOMEM;
ret = add_to_page_cache_lru(page, mapping, offset, ret = add_to_page_cache_lru(page, mapping, offset,
GFP_KERNEL & mapping_gfp_mask(mapping)); mapping_gfp_constraint(mapping, GFP_KERNEL));
if (ret == 0) if (ret == 0)
ret = mapping->a_ops->readpage(file, page); ret = mapping->a_ops->readpage(file, page);
else if (ret == -EEXIST) else if (ret == -EEXIST)
......
...@@ -90,7 +90,7 @@ int read_cache_pages(struct address_space *mapping, struct list_head *pages, ...@@ -90,7 +90,7 @@ int read_cache_pages(struct address_space *mapping, struct list_head *pages,
page = list_to_page(pages); page = list_to_page(pages);
list_del(&page->lru); list_del(&page->lru);
if (add_to_page_cache_lru(page, mapping, page->index, if (add_to_page_cache_lru(page, mapping, page->index,
GFP_KERNEL & mapping_gfp_mask(mapping))) { mapping_gfp_constraint(mapping, GFP_KERNEL))) {
read_cache_pages_invalidate_page(mapping, page); read_cache_pages_invalidate_page(mapping, page);
continue; continue;
} }
...@@ -128,7 +128,7 @@ static int read_pages(struct address_space *mapping, struct file *filp, ...@@ -128,7 +128,7 @@ static int read_pages(struct address_space *mapping, struct file *filp,
struct page *page = list_to_page(pages); struct page *page = list_to_page(pages);
list_del(&page->lru); list_del(&page->lru);
if (!add_to_page_cache_lru(page, mapping, page->index, if (!add_to_page_cache_lru(page, mapping, page->index,
GFP_KERNEL & mapping_gfp_mask(mapping))) { mapping_gfp_constraint(mapping, GFP_KERNEL))) {
mapping->a_ops->readpage(filp, page); mapping->a_ops->readpage(filp, page);
} }
page_cache_release(page); page_cache_release(page);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment