]> git.karo-electronics.de Git - karo-tx-linux.git/commitdiff
pagemap: redefine callback functions for page table walker
authorNaoya Horiguchi <n-horiguchi@ah.jp.nec.com>
Thu, 22 May 2014 00:42:41 +0000 (10:42 +1000)
committerStephen Rothwell <sfr@canb.auug.org.au>
Thu, 22 May 2014 00:42:41 +0000 (10:42 +1000)
pagemap_pte_range() connected to pmd_entry() does both of pmd loop and pte
loop.  So this patch moves pte part into pagemap_pte() on pte_entry().

We remove VM_SOFTDIRTY check in pagemap_pte_range(), because in the new
page table walker we call __walk_page_range() for each vma separately, so
we never experience multiple vmas in single pgd/pud/pmd/pte loop.

Signed-off-by: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>
Cc: Matt Mackall <mpm@selenic.com>
Cc: Cliff Wickman <cpw@sgi.com>
Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Cc: Michal Hocko <mhocko@suse.cz>
Cc: "Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com>
Cc: Pavel Emelyanov <xemul@parallels.com>
Cc: Rik van Riel <riel@redhat.com>
Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
fs/proc/task_mmu.c

index 3e4a9f736faf817693b44929bcfc365cee1b6aa4..2ab77b86fdfe2494a53226c008e10d31394d5944 100644 (file)
@@ -958,19 +958,33 @@ static inline void thp_pmd_to_pagemap_entry(pagemap_entry_t *pme, struct pagemap
 }
 #endif
 
-static int pagemap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
+static int pagemap_pte(pte_t *pte, unsigned long addr, unsigned long end,
                             struct mm_walk *walk)
 {
-       struct vm_area_struct *vma;
+       struct vm_area_struct *vma = walk->vma;
        struct pagemapread *pm = walk->private;
-       spinlock_t *ptl;
-       pte_t *pte;
+       pagemap_entry_t pme = make_pme(PM_NOT_PRESENT(pm->v2));
+
+       if (vma && vma->vm_start <= addr && end <= vma->vm_end) {
+               pte_to_pagemap_entry(&pme, pm, vma, addr, *pte);
+               /* unmap before userspace copy */
+               pte_unmap(pte);
+       }
+       return add_to_pagemap(addr, &pme, pm);
+}
+
+static int pagemap_pmd(pmd_t *pmd, unsigned long addr, unsigned long end,
+                            struct mm_walk *walk)
+{
        int err = 0;
+       struct vm_area_struct *vma = walk->vma;
+       struct pagemapread *pm = walk->private;
        pagemap_entry_t pme = make_pme(PM_NOT_PRESENT(pm->v2));
+       spinlock_t *ptl;
 
-       /* find the first VMA at or above 'addr' */
-       vma = find_vma(walk->mm, addr);
-       if (vma && pmd_trans_huge_lock(pmd, vma, &ptl) == 1) {
+       if (!vma)
+               return err;
+       if (pmd_trans_huge_lock(pmd, vma, &ptl) == 1) {
                int pmd_flags2;
 
                if ((vma->vm_flags & VM_SOFTDIRTY) || pmd_soft_dirty(*pmd))
@@ -989,41 +1003,9 @@ static int pagemap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
                                break;
                }
                spin_unlock(ptl);
-               return err;
-       }
-
-       if (pmd_trans_unstable(pmd))
-               return 0;
-       for (; addr != end; addr += PAGE_SIZE) {
-               int flags2;
-
-               /* check to see if we've left 'vma' behind
-                * and need a new, higher one */
-               if (vma && (addr >= vma->vm_end)) {
-                       vma = find_vma(walk->mm, addr);
-                       if (vma && (vma->vm_flags & VM_SOFTDIRTY))
-                               flags2 = __PM_SOFT_DIRTY;
-                       else
-                               flags2 = 0;
-                       pme = make_pme(PM_NOT_PRESENT(pm->v2) | PM_STATUS2(pm->v2, flags2));
-               }
-
-               /* check that 'vma' actually covers this address,
-                * and that it isn't a huge page vma */
-               if (vma && (vma->vm_start <= addr) &&
-                   !is_vm_hugetlb_page(vma)) {
-                       pte = pte_offset_map(pmd, addr);
-                       pte_to_pagemap_entry(&pme, pm, vma, addr, *pte);
-                       /* unmap before userspace copy */
-                       pte_unmap(pte);
-               }
-               err = add_to_pagemap(addr, &pme, pm);
-               if (err)
-                       return err;
+               /* don't call pagemap_pte() */
+               walk->skip = 1;
        }
-
-       cond_resched();
-
        return err;
 }
 
@@ -1046,12 +1028,11 @@ static int pagemap_hugetlb_range(pte_t *pte, unsigned long hmask,
                                 struct mm_walk *walk)
 {
        struct pagemapread *pm = walk->private;
-       struct vm_area_struct *vma;
+       struct vm_area_struct *vma = walk->vma;
        int err = 0;
        int flags2;
        pagemap_entry_t pme;
 
-       vma = find_vma(walk->mm, addr);
        WARN_ON_ONCE(!vma);
 
        if (vma && (vma->vm_flags & VM_SOFTDIRTY))
@@ -1059,6 +1040,7 @@ static int pagemap_hugetlb_range(pte_t *pte, unsigned long hmask,
        else
                flags2 = 0;
 
+       hmask = huge_page_mask(hstate_vma(vma));
        for (; addr != end; addr += PAGE_SIZE) {
                int offset = (addr & ~hmask) >> PAGE_SHIFT;
                huge_pte_to_pagemap_entry(&pme, pm, *pte, offset, flags2);
@@ -1066,9 +1048,6 @@ static int pagemap_hugetlb_range(pte_t *pte, unsigned long hmask,
                if (err)
                        return err;
        }
-
-       cond_resched();
-
        return err;
 }
 #endif /* HUGETLB_PAGE */
@@ -1135,10 +1114,11 @@ static ssize_t pagemap_read(struct file *file, char __user *buf,
        if (!mm || IS_ERR(mm))
                goto out_free;
 
-       pagemap_walk.pmd_entry = pagemap_pte_range;
+       pagemap_walk.pte_entry = pagemap_pte;
+       pagemap_walk.pmd_entry = pagemap_pmd;
        pagemap_walk.pte_hole = pagemap_pte_hole;
 #ifdef CONFIG_HUGETLB_PAGE
-       pagemap_walk.hugetlb_entry = pagemap_hugetlb_range;
+       pagemap_walk.hugetlb_entry = pagemap_hugetlb;
 #endif
        pagemap_walk.mm = mm;
        pagemap_walk.private = &pm;