#include <asm/tlb.h>
-struct madvise_free_private {
- struct vm_area_struct *vma;
- struct mmu_gather *tlb;
-};
-
/*
* Any behaviour which results in changes to the vma->vm_flags needs to
* take mmap_sem for writing. Others, which simply traverse vmas, need
unsigned long end, struct mm_walk *walk)
{
- struct madvise_free_private *fp = walk->private;
- struct mmu_gather *tlb = fp->tlb;
+ struct mmu_gather *tlb = walk->private;
struct mm_struct *mm = tlb->mm;
- struct vm_area_struct *vma = fp->vma;
+ struct vm_area_struct *vma = walk->vma;
spinlock_t *ptl;
pte_t *pte, ptent;
struct page *page;
struct vm_area_struct *vma,
unsigned long addr, unsigned long end)
{
- struct madvise_free_private fp = {
- .vma = vma,
- .tlb = tlb,
- };
-
struct mm_walk free_walk = {
.pmd_entry = madvise_free_pte_range,
.mm = vma->vm_mm,
- .private = &fp,
+ .private = tlb,
};
BUG_ON(addr >= end);
spinlock_t *ptl;
int ret = SWAP_AGAIN;
enum ttu_flags flags = (enum ttu_flags)arg;
- int dirty = 0;
/* munlock has nothing to gain from examining un-locked vmas */
if ((flags & TTU_MUNLOCK) && !(vma->vm_flags & VM_LOCKED))
}
/* Move the dirty bit to the physical page now the pte is gone. */
- dirty = pte_dirty(pteval);
- if (dirty)
+ if (pte_dirty(pteval))
set_page_dirty(page);
/* Update high watermark before we lower rss */
if (flags & TTU_FREE) {
VM_BUG_ON_PAGE(PageSwapCache(page), page);
- if (!dirty && !PageDirty(page)) {
+ if (!PageDirty(page)) {
/* It's a freeable page by MADV_FREE */
dec_mm_counter(mm, MM_ANONPAGES);
goto discard;