|
@@ -1164,7 +1164,7 @@ struct task_struct *__switch_to(struct task_struct *prev,
|
|
|
}
|
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
|
|
-#ifdef CONFIG_PPC_STD_MMU_64
|
|
|
+#ifdef CONFIG_PPC_BOOK3S_64
|
|
|
batch = this_cpu_ptr(&ppc64_tlb_batch);
|
|
|
if (batch->active) {
|
|
|
current_thread_info()->local_flags |= _TLF_LAZY_MMU;
|
|
@@ -1172,7 +1172,7 @@ struct task_struct *__switch_to(struct task_struct *prev,
|
|
|
__flush_tlb_pending(batch);
|
|
|
batch->active = 0;
|
|
|
}
|
|
|
-#endif /* CONFIG_PPC_STD_MMU_64 */
|
|
|
+#endif /* CONFIG_PPC_BOOK3S_64 */
|
|
|
|
|
|
#ifdef CONFIG_PPC_ADV_DEBUG_REGS
|
|
|
switch_booke_debug_regs(&new->thread.debug);
|
|
@@ -1218,7 +1218,7 @@ struct task_struct *__switch_to(struct task_struct *prev,
|
|
|
|
|
|
last = _switch(old_thread, new_thread);
|
|
|
|
|
|
-#ifdef CONFIG_PPC_STD_MMU_64
|
|
|
+#ifdef CONFIG_PPC_BOOK3S_64
|
|
|
if (current_thread_info()->local_flags & _TLF_LAZY_MMU) {
|
|
|
current_thread_info()->local_flags &= ~_TLF_LAZY_MMU;
|
|
|
batch = this_cpu_ptr(&ppc64_tlb_batch);
|
|
@@ -1247,7 +1247,7 @@ struct task_struct *__switch_to(struct task_struct *prev,
|
|
|
: : "r"(dummy_copy_buffer), "r"(0));
|
|
|
}
|
|
|
}
|
|
|
-#endif /* CONFIG_PPC_STD_MMU_64 */
|
|
|
+#endif /* CONFIG_PPC_BOOK3S_64 */
|
|
|
|
|
|
return last;
|
|
|
}
|
|
@@ -1476,7 +1476,7 @@ int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src)
|
|
|
|
|
|
static void setup_ksp_vsid(struct task_struct *p, unsigned long sp)
|
|
|
{
|
|
|
-#ifdef CONFIG_PPC_STD_MMU_64
|
|
|
+#ifdef CONFIG_PPC_BOOK3S_64
|
|
|
unsigned long sp_vsid;
|
|
|
unsigned long llp = mmu_psize_defs[mmu_linear_psize].sllp;
|
|
|
|
|
@@ -2056,7 +2056,7 @@ unsigned long arch_randomize_brk(struct mm_struct *mm)
|
|
|
unsigned long base = mm->brk;
|
|
|
unsigned long ret;
|
|
|
|
|
|
-#ifdef CONFIG_PPC_STD_MMU_64
|
|
|
+#ifdef CONFIG_PPC_BOOK3S_64
|
|
|
/*
|
|
|
* If we are using 1TB segments and we are allowed to randomise
|
|
|
* the heap, we can put it above 1TB so it is backed by a 1TB
|