Commit c202baf0 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'akpm' (patches from Andrew)

Merge misc fixes from Andrew Morton:
 "13 fixes"

* emailed patches from Andrew Morton <akpm@linux-foundation.org>:
  memcg: disable hierarchy support if bound to the legacy cgroup hierarchy
  mm: reorder can_do_mlock to fix audit denial
  kasan, module: move MODULE_ALIGN macro into <linux/moduleloader.h>
  kasan, module, vmalloc: rework shadow allocation for modules
  fanotify: fix event filtering with FAN_ONDIR set
  mm/nommu.c: export symbol max_mapnr
  arch/c6x/include/asm/pgtable.h: define dummy pgprot_writecombine for !MMU
  nilfs2: fix deadlock of segment constructor during recovery
  mm: cma: fix CMA aligned offset calculation
  mm, hugetlb: close race when setting PageTail for gigantic pages
  mm, oom: do not fail __GFP_NOFAIL allocation if oom killer is disabled
  drivers/rtc/rtc-s3c.c: add .needs_src_clk to s3c6410 RTC data
  ocfs2: make append_dio an incompat feature
parents ba68bc01 7feee590
...@@ -67,6 +67,11 @@ extern unsigned long empty_zero_page; ...@@ -67,6 +67,11 @@ extern unsigned long empty_zero_page;
*/ */
#define pgtable_cache_init() do { } while (0) #define pgtable_cache_init() do { } while (0)
/*
* c6x is !MMU, so define the simpliest implementation
*/
#define pgprot_writecombine pgprot_noncached
#include <asm-generic/pgtable.h> #include <asm-generic/pgtable.h>
#endif /* _ASM_C6X_PGTABLE_H */ #endif /* _ASM_C6X_PGTABLE_H */
...@@ -849,6 +849,7 @@ static struct s3c_rtc_data const s3c2443_rtc_data = { ...@@ -849,6 +849,7 @@ static struct s3c_rtc_data const s3c2443_rtc_data = {
static struct s3c_rtc_data const s3c6410_rtc_data = { static struct s3c_rtc_data const s3c6410_rtc_data = {
.max_user_freq = 32768, .max_user_freq = 32768,
.needs_src_clk = true,
.irq_handler = s3c6410_rtc_irq, .irq_handler = s3c6410_rtc_irq,
.set_freq = s3c6410_rtc_setfreq, .set_freq = s3c6410_rtc_setfreq,
.enable_tick = s3c6410_rtc_enable_tick, .enable_tick = s3c6410_rtc_enable_tick,
......
...@@ -1907,6 +1907,7 @@ static void nilfs_segctor_drop_written_files(struct nilfs_sc_info *sci, ...@@ -1907,6 +1907,7 @@ static void nilfs_segctor_drop_written_files(struct nilfs_sc_info *sci,
struct the_nilfs *nilfs) struct the_nilfs *nilfs)
{ {
struct nilfs_inode_info *ii, *n; struct nilfs_inode_info *ii, *n;
int during_mount = !(sci->sc_super->s_flags & MS_ACTIVE);
int defer_iput = false; int defer_iput = false;
spin_lock(&nilfs->ns_inode_lock); spin_lock(&nilfs->ns_inode_lock);
...@@ -1919,10 +1920,10 @@ static void nilfs_segctor_drop_written_files(struct nilfs_sc_info *sci, ...@@ -1919,10 +1920,10 @@ static void nilfs_segctor_drop_written_files(struct nilfs_sc_info *sci,
brelse(ii->i_bh); brelse(ii->i_bh);
ii->i_bh = NULL; ii->i_bh = NULL;
list_del_init(&ii->i_dirty); list_del_init(&ii->i_dirty);
if (!ii->vfs_inode.i_nlink) { if (!ii->vfs_inode.i_nlink || during_mount) {
/* /*
* Defer calling iput() to avoid a deadlock * Defer calling iput() to avoid deadlocks if
* over I_SYNC flag for inodes with i_nlink == 0 * i_nlink == 0 or mount is not yet finished.
*/ */
list_add_tail(&ii->i_dirty, &sci->sc_iput_queue); list_add_tail(&ii->i_dirty, &sci->sc_iput_queue);
defer_iput = true; defer_iput = true;
......
...@@ -143,7 +143,8 @@ static bool fanotify_should_send_event(struct fsnotify_mark *inode_mark, ...@@ -143,7 +143,8 @@ static bool fanotify_should_send_event(struct fsnotify_mark *inode_mark,
!(marks_mask & FS_ISDIR & ~marks_ignored_mask)) !(marks_mask & FS_ISDIR & ~marks_ignored_mask))
return false; return false;
if (event_mask & marks_mask & ~marks_ignored_mask) if (event_mask & FAN_ALL_OUTGOING_EVENTS & marks_mask &
~marks_ignored_mask)
return true; return true;
return false; return false;
......
...@@ -502,7 +502,7 @@ static inline int ocfs2_writes_unwritten_extents(struct ocfs2_super *osb) ...@@ -502,7 +502,7 @@ static inline int ocfs2_writes_unwritten_extents(struct ocfs2_super *osb)
static inline int ocfs2_supports_append_dio(struct ocfs2_super *osb) static inline int ocfs2_supports_append_dio(struct ocfs2_super *osb)
{ {
if (osb->s_feature_ro_compat & OCFS2_FEATURE_RO_COMPAT_APPEND_DIO) if (osb->s_feature_incompat & OCFS2_FEATURE_INCOMPAT_APPEND_DIO)
return 1; return 1;
return 0; return 0;
} }
......
...@@ -102,11 +102,11 @@ ...@@ -102,11 +102,11 @@
| OCFS2_FEATURE_INCOMPAT_INDEXED_DIRS \ | OCFS2_FEATURE_INCOMPAT_INDEXED_DIRS \
| OCFS2_FEATURE_INCOMPAT_REFCOUNT_TREE \ | OCFS2_FEATURE_INCOMPAT_REFCOUNT_TREE \
| OCFS2_FEATURE_INCOMPAT_DISCONTIG_BG \ | OCFS2_FEATURE_INCOMPAT_DISCONTIG_BG \
| OCFS2_FEATURE_INCOMPAT_CLUSTERINFO) | OCFS2_FEATURE_INCOMPAT_CLUSTERINFO \
| OCFS2_FEATURE_INCOMPAT_APPEND_DIO)
#define OCFS2_FEATURE_RO_COMPAT_SUPP (OCFS2_FEATURE_RO_COMPAT_UNWRITTEN \ #define OCFS2_FEATURE_RO_COMPAT_SUPP (OCFS2_FEATURE_RO_COMPAT_UNWRITTEN \
| OCFS2_FEATURE_RO_COMPAT_USRQUOTA \ | OCFS2_FEATURE_RO_COMPAT_USRQUOTA \
| OCFS2_FEATURE_RO_COMPAT_GRPQUOTA \ | OCFS2_FEATURE_RO_COMPAT_GRPQUOTA)
| OCFS2_FEATURE_RO_COMPAT_APPEND_DIO)
/* /*
* Heartbeat-only devices are missing journals and other files. The * Heartbeat-only devices are missing journals and other files. The
...@@ -178,6 +178,11 @@ ...@@ -178,6 +178,11 @@
*/ */
#define OCFS2_FEATURE_INCOMPAT_CLUSTERINFO 0x4000 #define OCFS2_FEATURE_INCOMPAT_CLUSTERINFO 0x4000
/*
* Append Direct IO support
*/
#define OCFS2_FEATURE_INCOMPAT_APPEND_DIO 0x8000
/* /*
* backup superblock flag is used to indicate that this volume * backup superblock flag is used to indicate that this volume
* has backup superblocks. * has backup superblocks.
...@@ -200,10 +205,6 @@ ...@@ -200,10 +205,6 @@
#define OCFS2_FEATURE_RO_COMPAT_USRQUOTA 0x0002 #define OCFS2_FEATURE_RO_COMPAT_USRQUOTA 0x0002
#define OCFS2_FEATURE_RO_COMPAT_GRPQUOTA 0x0004 #define OCFS2_FEATURE_RO_COMPAT_GRPQUOTA 0x0004
/*
* Append Direct IO support
*/
#define OCFS2_FEATURE_RO_COMPAT_APPEND_DIO 0x0008
/* The byte offset of the first backup block will be 1G. /* The byte offset of the first backup block will be 1G.
* The following will be 4G, 16G, 64G, 256G and 1T. * The following will be 4G, 16G, 64G, 256G and 1T.
......
...@@ -5,6 +5,7 @@ ...@@ -5,6 +5,7 @@
struct kmem_cache; struct kmem_cache;
struct page; struct page;
struct vm_struct;
#ifdef CONFIG_KASAN #ifdef CONFIG_KASAN
...@@ -49,15 +50,11 @@ void kasan_krealloc(const void *object, size_t new_size); ...@@ -49,15 +50,11 @@ void kasan_krealloc(const void *object, size_t new_size);
void kasan_slab_alloc(struct kmem_cache *s, void *object); void kasan_slab_alloc(struct kmem_cache *s, void *object);
void kasan_slab_free(struct kmem_cache *s, void *object); void kasan_slab_free(struct kmem_cache *s, void *object);
#define MODULE_ALIGN (PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT)
int kasan_module_alloc(void *addr, size_t size); int kasan_module_alloc(void *addr, size_t size);
void kasan_module_free(void *addr); void kasan_free_shadow(const struct vm_struct *vm);
#else /* CONFIG_KASAN */ #else /* CONFIG_KASAN */
#define MODULE_ALIGN 1
static inline void kasan_unpoison_shadow(const void *address, size_t size) {} static inline void kasan_unpoison_shadow(const void *address, size_t size) {}
static inline void kasan_enable_current(void) {} static inline void kasan_enable_current(void) {}
...@@ -82,7 +79,7 @@ static inline void kasan_slab_alloc(struct kmem_cache *s, void *object) {} ...@@ -82,7 +79,7 @@ static inline void kasan_slab_alloc(struct kmem_cache *s, void *object) {}
static inline void kasan_slab_free(struct kmem_cache *s, void *object) {} static inline void kasan_slab_free(struct kmem_cache *s, void *object) {}
static inline int kasan_module_alloc(void *addr, size_t size) { return 0; } static inline int kasan_module_alloc(void *addr, size_t size) { return 0; }
static inline void kasan_module_free(void *addr) {} static inline void kasan_free_shadow(const struct vm_struct *vm) {}
#endif /* CONFIG_KASAN */ #endif /* CONFIG_KASAN */
......
...@@ -84,4 +84,12 @@ void module_arch_cleanup(struct module *mod); ...@@ -84,4 +84,12 @@ void module_arch_cleanup(struct module *mod);
/* Any cleanup before freeing mod->module_init */ /* Any cleanup before freeing mod->module_init */
void module_arch_freeing_init(struct module *mod); void module_arch_freeing_init(struct module *mod);
#ifdef CONFIG_KASAN
#include <linux/kasan.h>
#define MODULE_ALIGN (PAGE_SIZE << KASAN_SHADOW_SCALE_SHIFT)
#else
#define MODULE_ALIGN PAGE_SIZE
#endif
#endif #endif
...@@ -17,6 +17,7 @@ struct vm_area_struct; /* vma defining user mapping in mm_types.h */ ...@@ -17,6 +17,7 @@ struct vm_area_struct; /* vma defining user mapping in mm_types.h */
#define VM_VPAGES 0x00000010 /* buffer for pages was vmalloc'ed */ #define VM_VPAGES 0x00000010 /* buffer for pages was vmalloc'ed */
#define VM_UNINITIALIZED 0x00000020 /* vm_struct is not fully initialized */ #define VM_UNINITIALIZED 0x00000020 /* vm_struct is not fully initialized */
#define VM_NO_GUARD 0x00000040 /* don't add guard page */ #define VM_NO_GUARD 0x00000040 /* don't add guard page */
#define VM_KASAN 0x00000080 /* has allocated kasan shadow memory */
/* bits [20..32] reserved for arch specific ioremap internals */ /* bits [20..32] reserved for arch specific ioremap internals */
/* /*
......
...@@ -56,7 +56,6 @@ ...@@ -56,7 +56,6 @@
#include <linux/async.h> #include <linux/async.h>
#include <linux/percpu.h> #include <linux/percpu.h>
#include <linux/kmemleak.h> #include <linux/kmemleak.h>
#include <linux/kasan.h>
#include <linux/jump_label.h> #include <linux/jump_label.h>
#include <linux/pfn.h> #include <linux/pfn.h>
#include <linux/bsearch.h> #include <linux/bsearch.h>
...@@ -1814,7 +1813,6 @@ static void unset_module_init_ro_nx(struct module *mod) { } ...@@ -1814,7 +1813,6 @@ static void unset_module_init_ro_nx(struct module *mod) { }
void __weak module_memfree(void *module_region) void __weak module_memfree(void *module_region)
{ {
vfree(module_region); vfree(module_region);
kasan_module_free(module_region);
} }
void __weak module_arch_cleanup(struct module *mod) void __weak module_arch_cleanup(struct module *mod)
......
...@@ -64,15 +64,17 @@ static unsigned long cma_bitmap_aligned_mask(struct cma *cma, int align_order) ...@@ -64,15 +64,17 @@ static unsigned long cma_bitmap_aligned_mask(struct cma *cma, int align_order)
return (1UL << (align_order - cma->order_per_bit)) - 1; return (1UL << (align_order - cma->order_per_bit)) - 1;
} }
/*
* Find a PFN aligned to the specified order and return an offset represented in
* order_per_bits.
*/
static unsigned long cma_bitmap_aligned_offset(struct cma *cma, int align_order) static unsigned long cma_bitmap_aligned_offset(struct cma *cma, int align_order)
{ {
unsigned int alignment;
if (align_order <= cma->order_per_bit) if (align_order <= cma->order_per_bit)
return 0; return 0;
alignment = 1UL << (align_order - cma->order_per_bit);
return ALIGN(cma->base_pfn, alignment) - return (ALIGN(cma->base_pfn, (1UL << align_order))
(cma->base_pfn >> cma->order_per_bit); - cma->base_pfn) >> cma->order_per_bit;
} }
static unsigned long cma_bitmap_maxno(struct cma *cma) static unsigned long cma_bitmap_maxno(struct cma *cma)
......
...@@ -917,7 +917,6 @@ static void prep_compound_gigantic_page(struct page *page, unsigned long order) ...@@ -917,7 +917,6 @@ static void prep_compound_gigantic_page(struct page *page, unsigned long order)
__SetPageHead(page); __SetPageHead(page);
__ClearPageReserved(page); __ClearPageReserved(page);
for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) { for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) {
__SetPageTail(p);
/* /*
* For gigantic hugepages allocated through bootmem at * For gigantic hugepages allocated through bootmem at
* boot, it's safer to be consistent with the not-gigantic * boot, it's safer to be consistent with the not-gigantic
...@@ -933,6 +932,9 @@ static void prep_compound_gigantic_page(struct page *page, unsigned long order) ...@@ -933,6 +932,9 @@ static void prep_compound_gigantic_page(struct page *page, unsigned long order)
__ClearPageReserved(p); __ClearPageReserved(p);
set_page_count(p, 0); set_page_count(p, 0);
p->first_page = page; p->first_page = page;
/* Make sure p->first_page is always valid for PageTail() */
smp_wmb();
__SetPageTail(p);
} }
} }
......
...@@ -29,6 +29,7 @@ ...@@ -29,6 +29,7 @@
#include <linux/stacktrace.h> #include <linux/stacktrace.h>
#include <linux/string.h> #include <linux/string.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/vmalloc.h>
#include <linux/kasan.h> #include <linux/kasan.h>
#include "kasan.h" #include "kasan.h"
...@@ -414,12 +415,19 @@ int kasan_module_alloc(void *addr, size_t size) ...@@ -414,12 +415,19 @@ int kasan_module_alloc(void *addr, size_t size)
GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO, GFP_KERNEL | __GFP_HIGHMEM | __GFP_ZERO,
PAGE_KERNEL, VM_NO_GUARD, NUMA_NO_NODE, PAGE_KERNEL, VM_NO_GUARD, NUMA_NO_NODE,
__builtin_return_address(0)); __builtin_return_address(0));
return ret ? 0 : -ENOMEM;
if (ret) {
find_vm_area(addr)->flags |= VM_KASAN;
return 0;
}
return -ENOMEM;
} }
void kasan_module_free(void *addr) void kasan_free_shadow(const struct vm_struct *vm)
{ {
vfree(kasan_mem_to_shadow(addr)); if (vm->flags & VM_KASAN)
vfree(kasan_mem_to_shadow(vm->addr));
} }
static void register_global(struct kasan_global *global) static void register_global(struct kasan_global *global)
......
...@@ -5232,7 +5232,9 @@ static void mem_cgroup_bind(struct cgroup_subsys_state *root_css) ...@@ -5232,7 +5232,9 @@ static void mem_cgroup_bind(struct cgroup_subsys_state *root_css)
* on for the root memcg is enough. * on for the root memcg is enough.
*/ */
if (cgroup_on_dfl(root_css->cgroup)) if (cgroup_on_dfl(root_css->cgroup))
mem_cgroup_from_css(root_css)->use_hierarchy = true; root_mem_cgroup->use_hierarchy = true;
else
root_mem_cgroup->use_hierarchy = false;
} }
static u64 memory_current_read(struct cgroup_subsys_state *css, static u64 memory_current_read(struct cgroup_subsys_state *css,
......
...@@ -26,10 +26,10 @@ ...@@ -26,10 +26,10 @@
int can_do_mlock(void) int can_do_mlock(void)
{ {
if (capable(CAP_IPC_LOCK))
return 1;
if (rlimit(RLIMIT_MEMLOCK) != 0) if (rlimit(RLIMIT_MEMLOCK) != 0)
return 1; return 1;
if (capable(CAP_IPC_LOCK))
return 1;
return 0; return 0;
} }
EXPORT_SYMBOL(can_do_mlock); EXPORT_SYMBOL(can_do_mlock);
......
...@@ -62,6 +62,7 @@ void *high_memory; ...@@ -62,6 +62,7 @@ void *high_memory;
EXPORT_SYMBOL(high_memory); EXPORT_SYMBOL(high_memory);
struct page *mem_map; struct page *mem_map;
unsigned long max_mapnr; unsigned long max_mapnr;
EXPORT_SYMBOL(max_mapnr);
unsigned long highest_memmap_pfn; unsigned long highest_memmap_pfn;
struct percpu_counter vm_committed_as; struct percpu_counter vm_committed_as;
int sysctl_overcommit_memory = OVERCOMMIT_GUESS; /* heuristic overcommit */ int sysctl_overcommit_memory = OVERCOMMIT_GUESS; /* heuristic overcommit */
......
...@@ -2373,7 +2373,8 @@ __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, ...@@ -2373,7 +2373,8 @@ __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order,
goto out; goto out;
} }
/* Exhausted what can be done so it's blamo time */ /* Exhausted what can be done so it's blamo time */
if (out_of_memory(ac->zonelist, gfp_mask, order, ac->nodemask, false)) if (out_of_memory(ac->zonelist, gfp_mask, order, ac->nodemask, false)
|| WARN_ON_ONCE(gfp_mask & __GFP_NOFAIL))
*did_some_progress = 1; *did_some_progress = 1;
out: out:
oom_zonelist_unlock(ac->zonelist, gfp_mask); oom_zonelist_unlock(ac->zonelist, gfp_mask);
......
...@@ -1418,6 +1418,7 @@ struct vm_struct *remove_vm_area(const void *addr) ...@@ -1418,6 +1418,7 @@ struct vm_struct *remove_vm_area(const void *addr)
spin_unlock(&vmap_area_lock); spin_unlock(&vmap_area_lock);
vmap_debug_free_range(va->va_start, va->va_end); vmap_debug_free_range(va->va_start, va->va_end);
kasan_free_shadow(vm);
free_unmap_vmap_area(va); free_unmap_vmap_area(va);
vm->size -= PAGE_SIZE; vm->size -= PAGE_SIZE;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment