return pte_offset_kernel(pmd, address);
}
-static void set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte)
+static void __set_pmd_pte(pte_t *kpte, unsigned long address, pte_t pte)
{
- unsigned long flags;
struct page *page;
/* change init_mm */
if (SHARED_KERNEL_PMD)
return;
- spin_lock_irqsave(&pgd_lock, flags);
for (page = pgd_list; page; page = (struct page *)page->index) {
pgd_t *pgd;
pud_t *pud;
pmd = pmd_offset(pud, address);
set_pte_atomic((pte_t *)pmd, pte);
}
- spin_unlock_irqrestore(&pgd_lock, flags);
}
static int split_large_page(pte_t *kpte, unsigned long address)
{
pgprot_t ref_prot = pte_pgprot(pte_clrhuge(*kpte));
+ unsigned long flags;
unsigned long addr;
pte_t *pbase, *tmp;
struct page *base;
if (!base)
return -ENOMEM;
- down_write(&init_mm.mmap_sem);
+ spin_lock_irqsave(&pgd_lock, flags);
/*
* Check for races, another CPU might have split this page
* up for us already:
/*
* Install the new, split up pagetable:
*/
- set_pmd_pte(kpte, address, mk_pte(base, ref_prot));
+ __set_pmd_pte(kpte, address, mk_pte(base, ref_prot));
base = NULL;
out_unlock:
- up_write(&init_mm.mmap_sem);
+ spin_unlock_irqrestore(&pgd_lock, flags);
if (base)
__free_pages(base, 0);