|
@@ -3351,9 +3351,6 @@ static int do_numa_page(struct fault_env *fe, pte_t pte)
|
|
|
bool was_writable = pte_write(pte);
|
|
|
int flags = 0;
|
|
|
|
|
|
- /* A PROT_NONE fault should not end up here */
|
|
|
- BUG_ON(!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE)));
|
|
|
-
|
|
|
/*
|
|
|
* The "pte" at this point cannot be used safely without
|
|
|
* validation through pte_unmap_same(). It's of NUMA type but
|
|
@@ -3458,6 +3455,11 @@ static int wp_huge_pmd(struct fault_env *fe, pmd_t orig_pmd)
|
|
|
return VM_FAULT_FALLBACK;
|
|
|
}
|
|
|
|
|
|
+static inline bool vma_is_accessible(struct vm_area_struct *vma)
|
|
|
+{
|
|
|
+ return vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE);
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* These routines also need to handle stuff like marking pages dirty
|
|
|
* and/or accessed for architectures that don't do it in hardware (most
|
|
@@ -3524,7 +3526,7 @@ static int handle_pte_fault(struct fault_env *fe)
|
|
|
if (!pte_present(entry))
|
|
|
return do_swap_page(fe, entry);
|
|
|
|
|
|
- if (pte_protnone(entry))
|
|
|
+ if (pte_protnone(entry) && vma_is_accessible(fe->vma))
|
|
|
return do_numa_page(fe, entry);
|
|
|
|
|
|
fe->ptl = pte_lockptr(fe->vma->vm_mm, fe->pmd);
|
|
@@ -3590,7 +3592,7 @@ static int __handle_mm_fault(struct vm_area_struct *vma, unsigned long address,
|
|
|
|
|
|
barrier();
|
|
|
if (pmd_trans_huge(orig_pmd) || pmd_devmap(orig_pmd)) {
|
|
|
- if (pmd_protnone(orig_pmd))
|
|
|
+ if (pmd_protnone(orig_pmd) && vma_is_accessible(vma))
|
|
|
return do_huge_pmd_numa_page(&fe, orig_pmd);
|
|
|
|
|
|
if ((fe.flags & FAULT_FLAG_WRITE) &&
|