Commit b3ae5acb authored by Matt Mackall's avatar Matt Mackall Committed by Linus Torvalds

maps4: use pagewalker in clear_refs and smaps

Use the generic pagewalker for smaps and clear_refs
Signed-off-by: default avatarMatt Mackall <mpm@selenic.com>
Cc: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent e6473092
...@@ -135,6 +135,7 @@ static void pad_len_spaces(struct seq_file *m, int len) ...@@ -135,6 +135,7 @@ static void pad_len_spaces(struct seq_file *m, int len)
struct mem_size_stats struct mem_size_stats
{ {
struct vm_area_struct *vma;
unsigned long resident; unsigned long resident;
unsigned long shared_clean; unsigned long shared_clean;
unsigned long shared_dirty; unsigned long shared_dirty;
...@@ -144,13 +145,6 @@ struct mem_size_stats ...@@ -144,13 +145,6 @@ struct mem_size_stats
u64 pss; u64 pss;
}; };
struct pmd_walker {
struct vm_area_struct *vma;
void *private;
void (*action)(struct vm_area_struct *, pmd_t *, unsigned long,
unsigned long, void *);
};
static int show_map_internal(struct seq_file *m, void *v, struct mem_size_stats *mss) static int show_map_internal(struct seq_file *m, void *v, struct mem_size_stats *mss)
{ {
struct proc_maps_private *priv = m->private; struct proc_maps_private *priv = m->private;
...@@ -240,11 +234,11 @@ static int show_map(struct seq_file *m, void *v) ...@@ -240,11 +234,11 @@ static int show_map(struct seq_file *m, void *v)
return show_map_internal(m, v, NULL); return show_map_internal(m, v, NULL);
} }
static void smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd, static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
unsigned long addr, unsigned long end, void *private)
void *private)
{ {
struct mem_size_stats *mss = private; struct mem_size_stats *mss = private;
struct vm_area_struct *vma = mss->vma;
pte_t *pte, ptent; pte_t *pte, ptent;
spinlock_t *ptl; spinlock_t *ptl;
struct page *page; struct page *page;
...@@ -282,12 +276,13 @@ static void smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd, ...@@ -282,12 +276,13 @@ static void smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
} }
pte_unmap_unlock(pte - 1, ptl); pte_unmap_unlock(pte - 1, ptl);
cond_resched(); cond_resched();
return 0;
} }
static void clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd, static int clear_refs_pte_range(pmd_t *pmd, unsigned long addr,
unsigned long addr, unsigned long end, unsigned long end, void *private)
void *private)
{ {
struct vm_area_struct *vma = private;
pte_t *pte, ptent; pte_t *pte, ptent;
spinlock_t *ptl; spinlock_t *ptl;
struct page *page; struct page *page;
...@@ -308,71 +303,10 @@ static void clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd, ...@@ -308,71 +303,10 @@ static void clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
} }
pte_unmap_unlock(pte - 1, ptl); pte_unmap_unlock(pte - 1, ptl);
cond_resched(); cond_resched();
return 0;
} }
static inline void walk_pmd_range(struct pmd_walker *walker, pud_t *pud, static struct mm_walk smaps_walk = { .pmd_entry = smaps_pte_range };
unsigned long addr, unsigned long end)
{
pmd_t *pmd;
unsigned long next;
for (pmd = pmd_offset(pud, addr); addr != end;
pmd++, addr = next) {
next = pmd_addr_end(addr, end);
if (pmd_none_or_clear_bad(pmd))
continue;
walker->action(walker->vma, pmd, addr, next, walker->private);
}
}
static inline void walk_pud_range(struct pmd_walker *walker, pgd_t *pgd,
unsigned long addr, unsigned long end)
{
pud_t *pud;
unsigned long next;
for (pud = pud_offset(pgd, addr); addr != end;
pud++, addr = next) {
next = pud_addr_end(addr, end);
if (pud_none_or_clear_bad(pud))
continue;
walk_pmd_range(walker, pud, addr, next);
}
}
/*
* walk_page_range - walk the page tables of a VMA with a callback
* @vma - VMA to walk
* @action - callback invoked for every bottom-level (PTE) page table
* @private - private data passed to the callback function
*
* Recursively walk the page table for the memory area in a VMA, calling
* a callback for every bottom-level (PTE) page table.
*/
static inline void walk_page_range(struct vm_area_struct *vma,
void (*action)(struct vm_area_struct *,
pmd_t *, unsigned long,
unsigned long, void *),
void *private)
{
unsigned long addr = vma->vm_start;
unsigned long end = vma->vm_end;
struct pmd_walker walker = {
.vma = vma,
.private = private,
.action = action,
};
pgd_t *pgd;
unsigned long next;
for (pgd = pgd_offset(vma->vm_mm, addr); addr != end;
pgd++, addr = next) {
next = pgd_addr_end(addr, end);
if (pgd_none_or_clear_bad(pgd))
continue;
walk_pud_range(&walker, pgd, addr, next);
}
}
static int show_smap(struct seq_file *m, void *v) static int show_smap(struct seq_file *m, void *v)
{ {
...@@ -380,11 +314,15 @@ static int show_smap(struct seq_file *m, void *v) ...@@ -380,11 +314,15 @@ static int show_smap(struct seq_file *m, void *v)
struct mem_size_stats mss; struct mem_size_stats mss;
memset(&mss, 0, sizeof mss); memset(&mss, 0, sizeof mss);
mss.vma = vma;
if (vma->vm_mm && !is_vm_hugetlb_page(vma)) if (vma->vm_mm && !is_vm_hugetlb_page(vma))
walk_page_range(vma, smaps_pte_range, &mss); walk_page_range(vma->vm_mm, vma->vm_start, vma->vm_end,
&smaps_walk, &mss);
return show_map_internal(m, v, &mss); return show_map_internal(m, v, &mss);
} }
static struct mm_walk clear_refs_walk = { .pmd_entry = clear_refs_pte_range };
void clear_refs_smap(struct mm_struct *mm) void clear_refs_smap(struct mm_struct *mm)
{ {
struct vm_area_struct *vma; struct vm_area_struct *vma;
...@@ -392,7 +330,8 @@ void clear_refs_smap(struct mm_struct *mm) ...@@ -392,7 +330,8 @@ void clear_refs_smap(struct mm_struct *mm)
down_read(&mm->mmap_sem); down_read(&mm->mmap_sem);
for (vma = mm->mmap; vma; vma = vma->vm_next) for (vma = mm->mmap; vma; vma = vma->vm_next)
if (vma->vm_mm && !is_vm_hugetlb_page(vma)) if (vma->vm_mm && !is_vm_hugetlb_page(vma))
walk_page_range(vma, clear_refs_pte_range, NULL); walk_page_range(vma->vm_mm, vma->vm_start, vma->vm_end,
&clear_refs_walk, vma);
flush_tlb_mm(mm); flush_tlb_mm(mm);
up_read(&mm->mmap_sem); up_read(&mm->mmap_sem);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment