|
@@ -351,6 +351,10 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *
|
|
|
|
|
|
mm = vma->vm_mm;
|
|
|
|
|
|
+ /* Don't insert a non-valid PTE into the TSB, we'll deadlock. */
|
|
|
+ if (!pte_accessible(mm, pte))
|
|
|
+ return;
|
|
|
+
|
|
|
spin_lock_irqsave(&mm->context.lock, flags);
|
|
|
|
|
|
#if defined(CONFIG_HUGETLB_PAGE) || defined(CONFIG_TRANSPARENT_HUGEPAGE)
|
|
@@ -2619,6 +2623,10 @@ void update_mmu_cache_pmd(struct vm_area_struct *vma, unsigned long addr,
|
|
|
|
|
|
pte = pmd_val(entry);
|
|
|
|
|
|
+ /* Don't insert a non-valid PMD into the TSB, we'll deadlock. */
|
|
|
+ if (!(pte & _PAGE_VALID))
|
|
|
+ return;
|
|
|
+
|
|
|
/* We are fabricating 8MB pages using 4MB real hw pages. */
|
|
|
pte |= (addr & (1UL << REAL_HPAGE_SHIFT));
|
|
|
|