maps4: use pagewalker in clear_refs and smaps
Use the generic pagewalker for smaps and clear_refs Signed-off-by: Matt Mackall <mpm@selenic.com> Cc: Dave Hansen <haveblue@us.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
e6473092bd
commit
b3ae5acbbb
1 changed files with 17 additions and 78 deletions
|
@ -135,6 +135,7 @@ static void pad_len_spaces(struct seq_file *m, int len)
|
||||||
|
|
||||||
struct mem_size_stats
|
struct mem_size_stats
|
||||||
{
|
{
|
||||||
|
struct vm_area_struct *vma;
|
||||||
unsigned long resident;
|
unsigned long resident;
|
||||||
unsigned long shared_clean;
|
unsigned long shared_clean;
|
||||||
unsigned long shared_dirty;
|
unsigned long shared_dirty;
|
||||||
|
@ -144,13 +145,6 @@ struct mem_size_stats
|
||||||
u64 pss;
|
u64 pss;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct pmd_walker {
|
|
||||||
struct vm_area_struct *vma;
|
|
||||||
void *private;
|
|
||||||
void (*action)(struct vm_area_struct *, pmd_t *, unsigned long,
|
|
||||||
unsigned long, void *);
|
|
||||||
};
|
|
||||||
|
|
||||||
static int show_map_internal(struct seq_file *m, void *v, struct mem_size_stats *mss)
|
static int show_map_internal(struct seq_file *m, void *v, struct mem_size_stats *mss)
|
||||||
{
|
{
|
||||||
struct proc_maps_private *priv = m->private;
|
struct proc_maps_private *priv = m->private;
|
||||||
|
@ -240,11 +234,11 @@ static int show_map(struct seq_file *m, void *v)
|
||||||
return show_map_internal(m, v, NULL);
|
return show_map_internal(m, v, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
|
static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
|
||||||
unsigned long addr, unsigned long end,
|
void *private)
|
||||||
void *private)
|
|
||||||
{
|
{
|
||||||
struct mem_size_stats *mss = private;
|
struct mem_size_stats *mss = private;
|
||||||
|
struct vm_area_struct *vma = mss->vma;
|
||||||
pte_t *pte, ptent;
|
pte_t *pte, ptent;
|
||||||
spinlock_t *ptl;
|
spinlock_t *ptl;
|
||||||
struct page *page;
|
struct page *page;
|
||||||
|
@ -282,12 +276,13 @@ static void smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
|
||||||
}
|
}
|
||||||
pte_unmap_unlock(pte - 1, ptl);
|
pte_unmap_unlock(pte - 1, ptl);
|
||||||
cond_resched();
|
cond_resched();
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
|
static int clear_refs_pte_range(pmd_t *pmd, unsigned long addr,
|
||||||
unsigned long addr, unsigned long end,
|
unsigned long end, void *private)
|
||||||
void *private)
|
|
||||||
{
|
{
|
||||||
|
struct vm_area_struct *vma = private;
|
||||||
pte_t *pte, ptent;
|
pte_t *pte, ptent;
|
||||||
spinlock_t *ptl;
|
spinlock_t *ptl;
|
||||||
struct page *page;
|
struct page *page;
|
||||||
|
@ -308,71 +303,10 @@ static void clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
|
||||||
}
|
}
|
||||||
pte_unmap_unlock(pte - 1, ptl);
|
pte_unmap_unlock(pte - 1, ptl);
|
||||||
cond_resched();
|
cond_resched();
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void walk_pmd_range(struct pmd_walker *walker, pud_t *pud,
|
static struct mm_walk smaps_walk = { .pmd_entry = smaps_pte_range };
|
||||||
unsigned long addr, unsigned long end)
|
|
||||||
{
|
|
||||||
pmd_t *pmd;
|
|
||||||
unsigned long next;
|
|
||||||
|
|
||||||
for (pmd = pmd_offset(pud, addr); addr != end;
|
|
||||||
pmd++, addr = next) {
|
|
||||||
next = pmd_addr_end(addr, end);
|
|
||||||
if (pmd_none_or_clear_bad(pmd))
|
|
||||||
continue;
|
|
||||||
walker->action(walker->vma, pmd, addr, next, walker->private);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline void walk_pud_range(struct pmd_walker *walker, pgd_t *pgd,
|
|
||||||
unsigned long addr, unsigned long end)
|
|
||||||
{
|
|
||||||
pud_t *pud;
|
|
||||||
unsigned long next;
|
|
||||||
|
|
||||||
for (pud = pud_offset(pgd, addr); addr != end;
|
|
||||||
pud++, addr = next) {
|
|
||||||
next = pud_addr_end(addr, end);
|
|
||||||
if (pud_none_or_clear_bad(pud))
|
|
||||||
continue;
|
|
||||||
walk_pmd_range(walker, pud, addr, next);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* walk_page_range - walk the page tables of a VMA with a callback
|
|
||||||
* @vma - VMA to walk
|
|
||||||
* @action - callback invoked for every bottom-level (PTE) page table
|
|
||||||
* @private - private data passed to the callback function
|
|
||||||
*
|
|
||||||
* Recursively walk the page table for the memory area in a VMA, calling
|
|
||||||
* a callback for every bottom-level (PTE) page table.
|
|
||||||
*/
|
|
||||||
static inline void walk_page_range(struct vm_area_struct *vma,
|
|
||||||
void (*action)(struct vm_area_struct *,
|
|
||||||
pmd_t *, unsigned long,
|
|
||||||
unsigned long, void *),
|
|
||||||
void *private)
|
|
||||||
{
|
|
||||||
unsigned long addr = vma->vm_start;
|
|
||||||
unsigned long end = vma->vm_end;
|
|
||||||
struct pmd_walker walker = {
|
|
||||||
.vma = vma,
|
|
||||||
.private = private,
|
|
||||||
.action = action,
|
|
||||||
};
|
|
||||||
pgd_t *pgd;
|
|
||||||
unsigned long next;
|
|
||||||
|
|
||||||
for (pgd = pgd_offset(vma->vm_mm, addr); addr != end;
|
|
||||||
pgd++, addr = next) {
|
|
||||||
next = pgd_addr_end(addr, end);
|
|
||||||
if (pgd_none_or_clear_bad(pgd))
|
|
||||||
continue;
|
|
||||||
walk_pud_range(&walker, pgd, addr, next);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static int show_smap(struct seq_file *m, void *v)
|
static int show_smap(struct seq_file *m, void *v)
|
||||||
{
|
{
|
||||||
|
@ -380,11 +314,15 @@ static int show_smap(struct seq_file *m, void *v)
|
||||||
struct mem_size_stats mss;
|
struct mem_size_stats mss;
|
||||||
|
|
||||||
memset(&mss, 0, sizeof mss);
|
memset(&mss, 0, sizeof mss);
|
||||||
|
mss.vma = vma;
|
||||||
if (vma->vm_mm && !is_vm_hugetlb_page(vma))
|
if (vma->vm_mm && !is_vm_hugetlb_page(vma))
|
||||||
walk_page_range(vma, smaps_pte_range, &mss);
|
walk_page_range(vma->vm_mm, vma->vm_start, vma->vm_end,
|
||||||
|
&smaps_walk, &mss);
|
||||||
return show_map_internal(m, v, &mss);
|
return show_map_internal(m, v, &mss);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static struct mm_walk clear_refs_walk = { .pmd_entry = clear_refs_pte_range };
|
||||||
|
|
||||||
void clear_refs_smap(struct mm_struct *mm)
|
void clear_refs_smap(struct mm_struct *mm)
|
||||||
{
|
{
|
||||||
struct vm_area_struct *vma;
|
struct vm_area_struct *vma;
|
||||||
|
@ -392,7 +330,8 @@ void clear_refs_smap(struct mm_struct *mm)
|
||||||
down_read(&mm->mmap_sem);
|
down_read(&mm->mmap_sem);
|
||||||
for (vma = mm->mmap; vma; vma = vma->vm_next)
|
for (vma = mm->mmap; vma; vma = vma->vm_next)
|
||||||
if (vma->vm_mm && !is_vm_hugetlb_page(vma))
|
if (vma->vm_mm && !is_vm_hugetlb_page(vma))
|
||||||
walk_page_range(vma, clear_refs_pte_range, NULL);
|
walk_page_range(vma->vm_mm, vma->vm_start, vma->vm_end,
|
||||||
|
&clear_refs_walk, vma);
|
||||||
flush_tlb_mm(mm);
|
flush_tlb_mm(mm);
|
||||||
up_read(&mm->mmap_sem);
|
up_read(&mm->mmap_sem);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Reference in a new issue