|
@@ -91,7 +91,6 @@ static inline void write_cr8(unsigned long x)
|
|
|
PVOP_VCALL1(cpu.write_cr8, x);
|
|
|
}
|
|
|
#endif
|
|
|
-#endif
|
|
|
|
|
|
static inline void arch_safe_halt(void)
|
|
|
{
|
|
@@ -103,7 +102,6 @@ static inline void halt(void)
|
|
|
PVOP_VCALL0(irq.halt);
|
|
|
}
|
|
|
|
|
|
-#ifdef CONFIG_PARAVIRT_XXL
|
|
|
static inline void wbinvd(void)
|
|
|
{
|
|
|
PVOP_VCALL0(cpu.wbinvd);
|
|
@@ -770,6 +768,7 @@ bool __raw_callee_save___native_vcpu_is_preempted(long cpu);
|
|
|
#define __PV_IS_CALLEE_SAVE(func) \
|
|
|
((struct paravirt_callee_save) { func })
|
|
|
|
|
|
+#ifdef CONFIG_PARAVIRT_XXL
|
|
|
static inline notrace unsigned long arch_local_save_flags(void)
|
|
|
{
|
|
|
return PVOP_CALLEE0(unsigned long, irq.save_fl);
|
|
@@ -798,6 +797,7 @@ static inline notrace unsigned long arch_local_irq_save(void)
|
|
|
arch_local_irq_disable();
|
|
|
return f;
|
|
|
}
|
|
|
+#endif
|
|
|
|
|
|
|
|
|
/* Make sure as little as possible of this mess escapes. */
|
|
@@ -884,7 +884,6 @@ extern void default_banner(void);
|
|
|
PARA_SITE(PARA_PATCH(PV_CPU_iret), \
|
|
|
ANNOTATE_RETPOLINE_SAFE; \
|
|
|
jmp PARA_INDIRECT(pv_ops+PV_CPU_iret);)
|
|
|
-#endif
|
|
|
|
|
|
#define DISABLE_INTERRUPTS(clobbers) \
|
|
|
PARA_SITE(PARA_PATCH(PV_IRQ_irq_disable), \
|
|
@@ -899,6 +898,7 @@ extern void default_banner(void);
|
|
|
ANNOTATE_RETPOLINE_SAFE; \
|
|
|
call PARA_INDIRECT(pv_ops+PV_IRQ_irq_enable); \
|
|
|
PV_RESTORE_REGS(clobbers | CLBR_CALLEE_SAVE);)
|
|
|
+#endif
|
|
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
#ifdef CONFIG_PARAVIRT_XXL
|