Commit cecee739 authored by Andrew Morton's avatar Andrew Morton Committed by Linus Torvalds

[PATCH] mm/mmap.c whitespace cleanups

- Don't require a 160-col xterm

- Coding style consistency
parent 6b2ca90b
...@@ -491,7 +491,9 @@ extern int do_munmap(struct mm_struct *, unsigned long, size_t); ...@@ -491,7 +491,9 @@ extern int do_munmap(struct mm_struct *, unsigned long, size_t);
extern unsigned long do_brk(unsigned long, unsigned long); extern unsigned long do_brk(unsigned long, unsigned long);
static inline void __vma_unlink(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev) static inline void
__vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma,
struct vm_area_struct *prev)
{ {
prev->vm_next = vma->vm_next; prev->vm_next = vma->vm_next;
rb_erase(&vma->vm_rb, &mm->mm_rb); rb_erase(&vma->vm_rb, &mm->mm_rb);
...@@ -499,7 +501,8 @@ static inline void __vma_unlink(struct mm_struct * mm, struct vm_area_struct * v ...@@ -499,7 +501,8 @@ static inline void __vma_unlink(struct mm_struct * mm, struct vm_area_struct * v
mm->mmap_cache = prev; mm->mmap_cache = prev;
} }
static inline int can_vma_merge(struct vm_area_struct * vma, unsigned long vm_flags) static inline int
can_vma_merge(struct vm_area_struct *vma, unsigned long vm_flags)
{ {
if (!vma->vm_file && vma->vm_flags == vm_flags) if (!vma->vm_file && vma->vm_flags == vm_flags)
return 1; return 1;
......
...@@ -205,7 +205,8 @@ asmlinkage unsigned long sys_brk(unsigned long brk) ...@@ -205,7 +205,8 @@ asmlinkage unsigned long sys_brk(unsigned long brk)
* internally. Essentially, translate the "PROT_xxx" and "MAP_xxx" bits * internally. Essentially, translate the "PROT_xxx" and "MAP_xxx" bits
* into "VM_xxx". * into "VM_xxx".
*/ */
static inline unsigned long calc_vm_flags(unsigned long prot, unsigned long flags) static inline unsigned long
calc_vm_flags(unsigned long prot, unsigned long flags)
{ {
#define _trans(x,bit1,bit2) \ #define _trans(x,bit1,bit2) \
((bit1==bit2)?(x&bit1):(x&bit1)?bit2:0) ((bit1==bit2)?(x&bit1):(x&bit1)?bit2:0)
...@@ -254,10 +255,10 @@ static void validate_mm(struct mm_struct * mm) { ...@@ -254,10 +255,10 @@ static void validate_mm(struct mm_struct * mm) {
#define validate_mm(mm) do { } while (0) #define validate_mm(mm) do { } while (0)
#endif #endif
static struct vm_area_struct * find_vma_prepare(struct mm_struct * mm, unsigned long addr, static struct vm_area_struct *
struct vm_area_struct ** pprev, find_vma_prepare(struct mm_struct *mm, unsigned long addr,
struct rb_node *** rb_link, struct vm_area_struct **pprev, struct rb_node ***rb_link,
struct rb_node ** rb_parent) struct rb_node ** rb_parent)
{ {
struct vm_area_struct * vma; struct vm_area_struct * vma;
struct rb_node ** __rb_link, * __rb_parent, * rb_prev; struct rb_node ** __rb_link, * __rb_parent, * rb_prev;
...@@ -291,8 +292,9 @@ static struct vm_area_struct * find_vma_prepare(struct mm_struct * mm, unsigned ...@@ -291,8 +292,9 @@ static struct vm_area_struct * find_vma_prepare(struct mm_struct * mm, unsigned
return vma; return vma;
} }
static inline void __vma_link_list(struct mm_struct * mm, struct vm_area_struct * vma, static inline void
struct vm_area_struct * prev, struct rb_node * rb_parent) __vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma,
struct vm_area_struct *prev, struct rb_node *rb_parent)
{ {
if (prev) { if (prev) {
vma->vm_next = prev->vm_next; vma->vm_next = prev->vm_next;
...@@ -300,20 +302,21 @@ static inline void __vma_link_list(struct mm_struct * mm, struct vm_area_struct ...@@ -300,20 +302,21 @@ static inline void __vma_link_list(struct mm_struct * mm, struct vm_area_struct
} else { } else {
mm->mmap = vma; mm->mmap = vma;
if (rb_parent) if (rb_parent)
vma->vm_next = rb_entry(rb_parent, struct vm_area_struct, vm_rb); vma->vm_next = rb_entry(rb_parent,
struct vm_area_struct, vm_rb);
else else
vma->vm_next = NULL; vma->vm_next = NULL;
} }
} }
static void __vma_link_rb(struct mm_struct * mm, struct vm_area_struct * vma, static void __vma_link_rb(struct mm_struct *mm, struct vm_area_struct *vma,
struct rb_node ** rb_link, struct rb_node * rb_parent) struct rb_node **rb_link, struct rb_node *rb_parent)
{ {
rb_link_node(&vma->vm_rb, rb_parent, rb_link); rb_link_node(&vma->vm_rb, rb_parent, rb_link);
rb_insert_color(&vma->vm_rb, &mm->mm_rb); rb_insert_color(&vma->vm_rb, &mm->mm_rb);
} }
static inline void __vma_link_file(struct vm_area_struct * vma) static inline void __vma_link_file(struct vm_area_struct *vma)
{ {
struct file * file; struct file * file;
...@@ -332,8 +335,10 @@ static inline void __vma_link_file(struct vm_area_struct * vma) ...@@ -332,8 +335,10 @@ static inline void __vma_link_file(struct vm_area_struct * vma)
} }
} }
static void __vma_link(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, static void
struct rb_node ** rb_link, struct rb_node * rb_parent) __vma_link(struct mm_struct *mm, struct vm_area_struct *vma,
struct vm_area_struct *prev, struct rb_node **rb_link,
struct rb_node *rb_parent)
{ {
__vma_link_list(mm, vma, prev, rb_parent); __vma_link_list(mm, vma, prev, rb_parent);
__vma_link_rb(mm, vma, rb_link, rb_parent); __vma_link_rb(mm, vma, rb_link, rb_parent);
...@@ -529,7 +534,8 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr, ...@@ -529,7 +534,8 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
* to. we assume access permissions have been handled by the open * to. we assume access permissions have been handled by the open
* of the memory object, so we don't do any here. * of the memory object, so we don't do any here.
*/ */
vm_flags = calc_vm_flags(prot,flags) | mm->def_flags | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC; vm_flags = calc_vm_flags(prot,flags) | mm->def_flags |
VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
if (flags & MAP_LOCKED) { if (flags & MAP_LOCKED) {
if (!capable(CAP_IPC_LOCK)) if (!capable(CAP_IPC_LOCK))
...@@ -549,14 +555,19 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr, ...@@ -549,14 +555,19 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
if (file) { if (file) {
switch (flags & MAP_TYPE) { switch (flags & MAP_TYPE) {
case MAP_SHARED: case MAP_SHARED:
if ((prot & PROT_WRITE) && !(file->f_mode & FMODE_WRITE)) if ((prot&PROT_WRITE) && !(file->f_mode&FMODE_WRITE))
return -EACCES; return -EACCES;
/* Make sure we don't allow writing to an append-only file.. */ /*
* Make sure we don't allow writing to an append-only
* file..
*/
if (IS_APPEND(inode) && (file->f_mode & FMODE_WRITE)) if (IS_APPEND(inode) && (file->f_mode & FMODE_WRITE))
return -EACCES; return -EACCES;
/* make sure there are no mandatory locks on the file. */ /*
* Make sure there are no mandatory locks on the file.
*/
if (locks_verify_locked(inode)) if (locks_verify_locked(inode))
return -EAGAIN; return -EAGAIN;
...@@ -610,7 +621,9 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr, ...@@ -610,7 +621,9 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
/* Check memory availability in shmem_file_setup? */ /* Check memory availability in shmem_file_setup? */
vm_flags |= VM_ACCOUNT; vm_flags |= VM_ACCOUNT;
} else if (vm_flags & VM_WRITE) { } else if (vm_flags & VM_WRITE) {
/* Private writable mapping: check memory availability */ /*
* Private writable mapping: check memory availability
*/
charged = len >> PAGE_SHIFT; charged = len >> PAGE_SHIFT;
if (!vm_enough_memory(charged)) if (!vm_enough_memory(charged))
return -ENOMEM; return -ENOMEM;
...@@ -620,10 +633,12 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr, ...@@ -620,10 +633,12 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
/* Can we just expand an old anonymous mapping? */ /* Can we just expand an old anonymous mapping? */
if (!file && !(vm_flags & VM_SHARED) && rb_parent) if (!file && !(vm_flags & VM_SHARED) && rb_parent)
if (vma_merge(mm, prev, rb_parent, addr, addr + len, vm_flags, NULL, 0)) if (vma_merge(mm, prev, rb_parent, addr, addr + len,
vm_flags, NULL, 0))
goto out; goto out;
/* Determine the object being mapped and call the appropriate /*
* Determine the object being mapped and call the appropriate
* specific mapper. the address has already been validated, but * specific mapper. the address has already been validated, but
* not unmapped, but the maps are removed from the list. * not unmapped, but the maps are removed from the list.
*/ */
...@@ -734,7 +749,9 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr, ...@@ -734,7 +749,9 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
* This function "knows" that -ENOMEM has the bits set. * This function "knows" that -ENOMEM has the bits set.
*/ */
#ifndef HAVE_ARCH_UNMAPPED_AREA #ifndef HAVE_ARCH_UNMAPPED_AREA
static inline unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) static inline unsigned long
arch_get_unmapped_area(struct file *filp, unsigned long addr,
unsigned long len, unsigned long pgoff, unsigned long flags)
{ {
struct mm_struct *mm = current->mm; struct mm_struct *mm = current->mm;
struct vm_area_struct *vma; struct vm_area_struct *vma;
...@@ -769,10 +786,14 @@ static inline unsigned long arch_get_unmapped_area(struct file *filp, unsigned l ...@@ -769,10 +786,14 @@ static inline unsigned long arch_get_unmapped_area(struct file *filp, unsigned l
} }
} }
#else #else
extern unsigned long arch_get_unmapped_area(struct file *, unsigned long, unsigned long, unsigned long, unsigned long); extern unsigned long
arch_get_unmapped_area(struct file *, unsigned long, unsigned long,
unsigned long, unsigned long);
#endif #endif
unsigned long get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) unsigned long
get_unmapped_area(struct file *file, unsigned long addr, unsigned long len,
unsigned long pgoff, unsigned long flags)
{ {
if (flags & MAP_FIXED) { if (flags & MAP_FIXED) {
if (addr > TASK_SIZE - len) if (addr > TASK_SIZE - len)
...@@ -783,7 +804,8 @@ unsigned long get_unmapped_area(struct file *file, unsigned long addr, unsigned ...@@ -783,7 +804,8 @@ unsigned long get_unmapped_area(struct file *file, unsigned long addr, unsigned
} }
if (file && file->f_op && file->f_op->get_unmapped_area) if (file && file->f_op && file->f_op->get_unmapped_area)
return file->f_op->get_unmapped_area(file, addr, len, pgoff, flags); return file->f_op->get_unmapped_area(file, addr, len,
pgoff, flags);
return arch_get_unmapped_area(file, addr, len, pgoff, flags); return arch_get_unmapped_area(file, addr, len, pgoff, flags);
} }
...@@ -806,7 +828,8 @@ struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr) ...@@ -806,7 +828,8 @@ struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr)
while (rb_node) { while (rb_node) {
struct vm_area_struct * vma_tmp; struct vm_area_struct * vma_tmp;
vma_tmp = rb_entry(rb_node, struct vm_area_struct, vm_rb); vma_tmp = rb_entry(rb_node,
struct vm_area_struct, vm_rb);
if (vma_tmp->vm_end > addr) { if (vma_tmp->vm_end > addr) {
vma = vma_tmp; vma = vma_tmp;
...@@ -824,8 +847,9 @@ struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr) ...@@ -824,8 +847,9 @@ struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr)
} }
/* Same as find_vma, but also return a pointer to the previous VMA in *pprev. */ /* Same as find_vma, but also return a pointer to the previous VMA in *pprev. */
struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr, struct vm_area_struct *
struct vm_area_struct **pprev) find_vma_prev(struct mm_struct *mm, unsigned long addr,
struct vm_area_struct **pprev)
{ {
struct vm_area_struct *vma = NULL, *prev = NULL; struct vm_area_struct *vma = NULL, *prev = NULL;
struct rb_node * rb_node; struct rb_node * rb_node;
...@@ -852,7 +876,7 @@ struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr, ...@@ -852,7 +876,7 @@ struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
} }
} }
out: out:
*pprev = prev; *pprev = prev;
return prev ? prev->vm_next : vma; return prev ? prev->vm_next : vma;
} }
...@@ -899,7 +923,8 @@ int expand_stack(struct vm_area_struct * vma, unsigned long address) ...@@ -899,7 +923,8 @@ int expand_stack(struct vm_area_struct * vma, unsigned long address)
return 0; return 0;
} }
struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long addr) struct vm_area_struct *
find_extend_vma(struct mm_struct *mm, unsigned long addr)
{ {
struct vm_area_struct *vma, *prev; struct vm_area_struct *vma, *prev;
...@@ -918,7 +943,7 @@ struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long add ...@@ -918,7 +943,7 @@ struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long add
/* /*
* vma is the first one with address < vma->vm_start. Have to extend vma. * vma is the first one with address < vma->vm_start. Have to extend vma.
*/ */
int expand_stack(struct vm_area_struct * vma, unsigned long address) int expand_stack(struct vm_area_struct *vma, unsigned long address)
{ {
unsigned long grow; unsigned long grow;
...@@ -953,7 +978,8 @@ int expand_stack(struct vm_area_struct * vma, unsigned long address) ...@@ -953,7 +978,8 @@ int expand_stack(struct vm_area_struct * vma, unsigned long address)
return 0; return 0;
} }
struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long addr) struct vm_area_struct *
find_extend_vma(struct mm_struct * mm, unsigned long addr)
{ {
struct vm_area_struct * vma; struct vm_area_struct * vma;
unsigned long start; unsigned long start;
...@@ -1023,7 +1049,7 @@ static void free_pgtables(struct mmu_gather *tlb, struct vm_area_struct *prev, ...@@ -1023,7 +1049,7 @@ static void free_pgtables(struct mmu_gather *tlb, struct vm_area_struct *prev,
break; break;
} }
no_mmaps: no_mmaps:
if (last < first) /* needed for arches with discontiguous pgd indices */ if (last < first) /* for arches with discontiguous pgd indices */
return; return;
/* /*
* If the PGD bits are not consecutive in the virtual address, the * If the PGD bits are not consecutive in the virtual address, the
...@@ -1298,7 +1324,8 @@ unsigned long do_brk(unsigned long addr, unsigned long len) ...@@ -1298,7 +1324,8 @@ unsigned long do_brk(unsigned long addr, unsigned long len)
flags = VM_DATA_DEFAULT_FLAGS | VM_ACCOUNT | mm->def_flags; flags = VM_DATA_DEFAULT_FLAGS | VM_ACCOUNT | mm->def_flags;
/* Can we just expand an old anonymous mapping? */ /* Can we just expand an old anonymous mapping? */
if (rb_parent && vma_merge(mm, prev, rb_parent, addr, addr + len, flags, NULL, 0)) if (rb_parent && vma_merge(mm, prev, rb_parent, addr, addr + len,
flags, NULL, 0))
goto out; goto out;
/* /*
...@@ -1408,7 +1435,7 @@ void insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) ...@@ -1408,7 +1435,7 @@ void insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma)
struct vm_area_struct * __vma, * prev; struct vm_area_struct * __vma, * prev;
struct rb_node ** rb_link, * rb_parent; struct rb_node ** rb_link, * rb_parent;
__vma = find_vma_prepare(mm, vma->vm_start, &prev, &rb_link, &rb_parent); __vma = find_vma_prepare(mm,vma->vm_start,&prev,&rb_link,&rb_parent);
if (__vma && __vma->vm_start < vma->vm_end) if (__vma && __vma->vm_start < vma->vm_end)
BUG(); BUG();
vma_link(mm, vma, prev, rb_link, rb_parent); vma_link(mm, vma, prev, rb_link, rb_parent);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment