|
@@ -1,648 +0,0 @@
|
|
-#include <linux/irqchip/arm-gic.h>
|
|
|
|
-#include <asm/assembler.h>
|
|
|
|
-
|
|
|
|
-#define VCPU_USR_REG(_reg_nr) (VCPU_USR_REGS + (_reg_nr * 4))
|
|
|
|
-#define VCPU_USR_SP (VCPU_USR_REG(13))
|
|
|
|
-#define VCPU_USR_LR (VCPU_USR_REG(14))
|
|
|
|
-#define CP15_OFFSET(_cp15_reg_idx) (VCPU_CP15 + (_cp15_reg_idx * 4))
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Many of these macros need to access the VCPU structure, which is always
|
|
|
|
- * held in r0. These macros should never clobber r1, as it is used to hold the
|
|
|
|
- * exception code on the return path (except of course the macro that switches
|
|
|
|
- * all the registers before the final jump to the VM).
|
|
|
|
- */
|
|
|
|
-vcpu .req r0 @ vcpu pointer always in r0
|
|
|
|
-
|
|
|
|
-/* Clobbers {r2-r6} */
|
|
|
|
-.macro store_vfp_state vfp_base
|
|
|
|
- @ The VFPFMRX and VFPFMXR macros are the VMRS and VMSR instructions
|
|
|
|
- VFPFMRX r2, FPEXC
|
|
|
|
- @ Make sure VFP is enabled so we can touch the registers.
|
|
|
|
- orr r6, r2, #FPEXC_EN
|
|
|
|
- VFPFMXR FPEXC, r6
|
|
|
|
-
|
|
|
|
- VFPFMRX r3, FPSCR
|
|
|
|
- tst r2, #FPEXC_EX @ Check for VFP Subarchitecture
|
|
|
|
- beq 1f
|
|
|
|
- @ If FPEXC_EX is 0, then FPINST/FPINST2 reads are upredictable, so
|
|
|
|
- @ we only need to save them if FPEXC_EX is set.
|
|
|
|
- VFPFMRX r4, FPINST
|
|
|
|
- tst r2, #FPEXC_FP2V
|
|
|
|
- VFPFMRX r5, FPINST2, ne @ vmrsne
|
|
|
|
- bic r6, r2, #FPEXC_EX @ FPEXC_EX disable
|
|
|
|
- VFPFMXR FPEXC, r6
|
|
|
|
-1:
|
|
|
|
- VFPFSTMIA \vfp_base, r6 @ Save VFP registers
|
|
|
|
- stm \vfp_base, {r2-r5} @ Save FPEXC, FPSCR, FPINST, FPINST2
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/* Assume FPEXC_EN is on and FPEXC_EX is off, clobbers {r2-r6} */
|
|
|
|
-.macro restore_vfp_state vfp_base
|
|
|
|
- VFPFLDMIA \vfp_base, r6 @ Load VFP registers
|
|
|
|
- ldm \vfp_base, {r2-r5} @ Load FPEXC, FPSCR, FPINST, FPINST2
|
|
|
|
-
|
|
|
|
- VFPFMXR FPSCR, r3
|
|
|
|
- tst r2, #FPEXC_EX @ Check for VFP Subarchitecture
|
|
|
|
- beq 1f
|
|
|
|
- VFPFMXR FPINST, r4
|
|
|
|
- tst r2, #FPEXC_FP2V
|
|
|
|
- VFPFMXR FPINST2, r5, ne
|
|
|
|
-1:
|
|
|
|
- VFPFMXR FPEXC, r2 @ FPEXC (last, in case !EN)
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/* These are simply for the macros to work - value don't have meaning */
|
|
|
|
-.equ usr, 0
|
|
|
|
-.equ svc, 1
|
|
|
|
-.equ abt, 2
|
|
|
|
-.equ und, 3
|
|
|
|
-.equ irq, 4
|
|
|
|
-.equ fiq, 5
|
|
|
|
-
|
|
|
|
-.macro push_host_regs_mode mode
|
|
|
|
- mrs r2, SP_\mode
|
|
|
|
- mrs r3, LR_\mode
|
|
|
|
- mrs r4, SPSR_\mode
|
|
|
|
- push {r2, r3, r4}
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Store all host persistent registers on the stack.
|
|
|
|
- * Clobbers all registers, in all modes, except r0 and r1.
|
|
|
|
- */
|
|
|
|
-.macro save_host_regs
|
|
|
|
- /* Hyp regs. Only ELR_hyp (SPSR_hyp already saved) */
|
|
|
|
- mrs r2, ELR_hyp
|
|
|
|
- push {r2}
|
|
|
|
-
|
|
|
|
- /* usr regs */
|
|
|
|
- push {r4-r12} @ r0-r3 are always clobbered
|
|
|
|
- mrs r2, SP_usr
|
|
|
|
- mov r3, lr
|
|
|
|
- push {r2, r3}
|
|
|
|
-
|
|
|
|
- push_host_regs_mode svc
|
|
|
|
- push_host_regs_mode abt
|
|
|
|
- push_host_regs_mode und
|
|
|
|
- push_host_regs_mode irq
|
|
|
|
-
|
|
|
|
- /* fiq regs */
|
|
|
|
- mrs r2, r8_fiq
|
|
|
|
- mrs r3, r9_fiq
|
|
|
|
- mrs r4, r10_fiq
|
|
|
|
- mrs r5, r11_fiq
|
|
|
|
- mrs r6, r12_fiq
|
|
|
|
- mrs r7, SP_fiq
|
|
|
|
- mrs r8, LR_fiq
|
|
|
|
- mrs r9, SPSR_fiq
|
|
|
|
- push {r2-r9}
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-.macro pop_host_regs_mode mode
|
|
|
|
- pop {r2, r3, r4}
|
|
|
|
- msr SP_\mode, r2
|
|
|
|
- msr LR_\mode, r3
|
|
|
|
- msr SPSR_\mode, r4
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Restore all host registers from the stack.
|
|
|
|
- * Clobbers all registers, in all modes, except r0 and r1.
|
|
|
|
- */
|
|
|
|
-.macro restore_host_regs
|
|
|
|
- pop {r2-r9}
|
|
|
|
- msr r8_fiq, r2
|
|
|
|
- msr r9_fiq, r3
|
|
|
|
- msr r10_fiq, r4
|
|
|
|
- msr r11_fiq, r5
|
|
|
|
- msr r12_fiq, r6
|
|
|
|
- msr SP_fiq, r7
|
|
|
|
- msr LR_fiq, r8
|
|
|
|
- msr SPSR_fiq, r9
|
|
|
|
-
|
|
|
|
- pop_host_regs_mode irq
|
|
|
|
- pop_host_regs_mode und
|
|
|
|
- pop_host_regs_mode abt
|
|
|
|
- pop_host_regs_mode svc
|
|
|
|
-
|
|
|
|
- pop {r2, r3}
|
|
|
|
- msr SP_usr, r2
|
|
|
|
- mov lr, r3
|
|
|
|
- pop {r4-r12}
|
|
|
|
-
|
|
|
|
- pop {r2}
|
|
|
|
- msr ELR_hyp, r2
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Restore SP, LR and SPSR for a given mode. offset is the offset of
|
|
|
|
- * this mode's registers from the VCPU base.
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- *
|
|
|
|
- * Clobbers r1, r2, r3, r4.
|
|
|
|
- */
|
|
|
|
-.macro restore_guest_regs_mode mode, offset
|
|
|
|
- add r1, vcpu, \offset
|
|
|
|
- ldm r1, {r2, r3, r4}
|
|
|
|
- msr SP_\mode, r2
|
|
|
|
- msr LR_\mode, r3
|
|
|
|
- msr SPSR_\mode, r4
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Restore all guest registers from the vcpu struct.
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- *
|
|
|
|
- * Clobbers *all* registers.
|
|
|
|
- */
|
|
|
|
-.macro restore_guest_regs
|
|
|
|
- restore_guest_regs_mode svc, #VCPU_SVC_REGS
|
|
|
|
- restore_guest_regs_mode abt, #VCPU_ABT_REGS
|
|
|
|
- restore_guest_regs_mode und, #VCPU_UND_REGS
|
|
|
|
- restore_guest_regs_mode irq, #VCPU_IRQ_REGS
|
|
|
|
-
|
|
|
|
- add r1, vcpu, #VCPU_FIQ_REGS
|
|
|
|
- ldm r1, {r2-r9}
|
|
|
|
- msr r8_fiq, r2
|
|
|
|
- msr r9_fiq, r3
|
|
|
|
- msr r10_fiq, r4
|
|
|
|
- msr r11_fiq, r5
|
|
|
|
- msr r12_fiq, r6
|
|
|
|
- msr SP_fiq, r7
|
|
|
|
- msr LR_fiq, r8
|
|
|
|
- msr SPSR_fiq, r9
|
|
|
|
-
|
|
|
|
- @ Load return state
|
|
|
|
- ldr r2, [vcpu, #VCPU_PC]
|
|
|
|
- ldr r3, [vcpu, #VCPU_CPSR]
|
|
|
|
- msr ELR_hyp, r2
|
|
|
|
- msr SPSR_cxsf, r3
|
|
|
|
-
|
|
|
|
- @ Load user registers
|
|
|
|
- ldr r2, [vcpu, #VCPU_USR_SP]
|
|
|
|
- ldr r3, [vcpu, #VCPU_USR_LR]
|
|
|
|
- msr SP_usr, r2
|
|
|
|
- mov lr, r3
|
|
|
|
- add vcpu, vcpu, #(VCPU_USR_REGS)
|
|
|
|
- ldm vcpu, {r0-r12}
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Save SP, LR and SPSR for a given mode. offset is the offset of
|
|
|
|
- * this mode's registers from the VCPU base.
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- *
|
|
|
|
- * Clobbers r2, r3, r4, r5.
|
|
|
|
- */
|
|
|
|
-.macro save_guest_regs_mode mode, offset
|
|
|
|
- add r2, vcpu, \offset
|
|
|
|
- mrs r3, SP_\mode
|
|
|
|
- mrs r4, LR_\mode
|
|
|
|
- mrs r5, SPSR_\mode
|
|
|
|
- stm r2, {r3, r4, r5}
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Save all guest registers to the vcpu struct
|
|
|
|
- * Expects guest's r0, r1, r2 on the stack.
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- *
|
|
|
|
- * Clobbers r2, r3, r4, r5.
|
|
|
|
- */
|
|
|
|
-.macro save_guest_regs
|
|
|
|
- @ Store usr registers
|
|
|
|
- add r2, vcpu, #VCPU_USR_REG(3)
|
|
|
|
- stm r2, {r3-r12}
|
|
|
|
- add r2, vcpu, #VCPU_USR_REG(0)
|
|
|
|
- pop {r3, r4, r5} @ r0, r1, r2
|
|
|
|
- stm r2, {r3, r4, r5}
|
|
|
|
- mrs r2, SP_usr
|
|
|
|
- mov r3, lr
|
|
|
|
- str r2, [vcpu, #VCPU_USR_SP]
|
|
|
|
- str r3, [vcpu, #VCPU_USR_LR]
|
|
|
|
-
|
|
|
|
- @ Store return state
|
|
|
|
- mrs r2, ELR_hyp
|
|
|
|
- mrs r3, spsr
|
|
|
|
- str r2, [vcpu, #VCPU_PC]
|
|
|
|
- str r3, [vcpu, #VCPU_CPSR]
|
|
|
|
-
|
|
|
|
- @ Store other guest registers
|
|
|
|
- save_guest_regs_mode svc, #VCPU_SVC_REGS
|
|
|
|
- save_guest_regs_mode abt, #VCPU_ABT_REGS
|
|
|
|
- save_guest_regs_mode und, #VCPU_UND_REGS
|
|
|
|
- save_guest_regs_mode irq, #VCPU_IRQ_REGS
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/* Reads cp15 registers from hardware and stores them in memory
|
|
|
|
- * @store_to_vcpu: If 0, registers are written in-order to the stack,
|
|
|
|
- * otherwise to the VCPU struct pointed to by vcpup
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- *
|
|
|
|
- * Clobbers r2 - r12
|
|
|
|
- */
|
|
|
|
-.macro read_cp15_state store_to_vcpu
|
|
|
|
- mrc p15, 0, r2, c1, c0, 0 @ SCTLR
|
|
|
|
- mrc p15, 0, r3, c1, c0, 2 @ CPACR
|
|
|
|
- mrc p15, 0, r4, c2, c0, 2 @ TTBCR
|
|
|
|
- mrc p15, 0, r5, c3, c0, 0 @ DACR
|
|
|
|
- mrrc p15, 0, r6, r7, c2 @ TTBR 0
|
|
|
|
- mrrc p15, 1, r8, r9, c2 @ TTBR 1
|
|
|
|
- mrc p15, 0, r10, c10, c2, 0 @ PRRR
|
|
|
|
- mrc p15, 0, r11, c10, c2, 1 @ NMRR
|
|
|
|
- mrc p15, 2, r12, c0, c0, 0 @ CSSELR
|
|
|
|
-
|
|
|
|
- .if \store_to_vcpu == 0
|
|
|
|
- push {r2-r12} @ Push CP15 registers
|
|
|
|
- .else
|
|
|
|
- str r2, [vcpu, #CP15_OFFSET(c1_SCTLR)]
|
|
|
|
- str r3, [vcpu, #CP15_OFFSET(c1_CPACR)]
|
|
|
|
- str r4, [vcpu, #CP15_OFFSET(c2_TTBCR)]
|
|
|
|
- str r5, [vcpu, #CP15_OFFSET(c3_DACR)]
|
|
|
|
- add r2, vcpu, #CP15_OFFSET(c2_TTBR0)
|
|
|
|
- strd r6, r7, [r2]
|
|
|
|
- add r2, vcpu, #CP15_OFFSET(c2_TTBR1)
|
|
|
|
- strd r8, r9, [r2]
|
|
|
|
- str r10, [vcpu, #CP15_OFFSET(c10_PRRR)]
|
|
|
|
- str r11, [vcpu, #CP15_OFFSET(c10_NMRR)]
|
|
|
|
- str r12, [vcpu, #CP15_OFFSET(c0_CSSELR)]
|
|
|
|
- .endif
|
|
|
|
-
|
|
|
|
- mrc p15, 0, r2, c13, c0, 1 @ CID
|
|
|
|
- mrc p15, 0, r3, c13, c0, 2 @ TID_URW
|
|
|
|
- mrc p15, 0, r4, c13, c0, 3 @ TID_URO
|
|
|
|
- mrc p15, 0, r5, c13, c0, 4 @ TID_PRIV
|
|
|
|
- mrc p15, 0, r6, c5, c0, 0 @ DFSR
|
|
|
|
- mrc p15, 0, r7, c5, c0, 1 @ IFSR
|
|
|
|
- mrc p15, 0, r8, c5, c1, 0 @ ADFSR
|
|
|
|
- mrc p15, 0, r9, c5, c1, 1 @ AIFSR
|
|
|
|
- mrc p15, 0, r10, c6, c0, 0 @ DFAR
|
|
|
|
- mrc p15, 0, r11, c6, c0, 2 @ IFAR
|
|
|
|
- mrc p15, 0, r12, c12, c0, 0 @ VBAR
|
|
|
|
-
|
|
|
|
- .if \store_to_vcpu == 0
|
|
|
|
- push {r2-r12} @ Push CP15 registers
|
|
|
|
- .else
|
|
|
|
- str r2, [vcpu, #CP15_OFFSET(c13_CID)]
|
|
|
|
- str r3, [vcpu, #CP15_OFFSET(c13_TID_URW)]
|
|
|
|
- str r4, [vcpu, #CP15_OFFSET(c13_TID_URO)]
|
|
|
|
- str r5, [vcpu, #CP15_OFFSET(c13_TID_PRIV)]
|
|
|
|
- str r6, [vcpu, #CP15_OFFSET(c5_DFSR)]
|
|
|
|
- str r7, [vcpu, #CP15_OFFSET(c5_IFSR)]
|
|
|
|
- str r8, [vcpu, #CP15_OFFSET(c5_ADFSR)]
|
|
|
|
- str r9, [vcpu, #CP15_OFFSET(c5_AIFSR)]
|
|
|
|
- str r10, [vcpu, #CP15_OFFSET(c6_DFAR)]
|
|
|
|
- str r11, [vcpu, #CP15_OFFSET(c6_IFAR)]
|
|
|
|
- str r12, [vcpu, #CP15_OFFSET(c12_VBAR)]
|
|
|
|
- .endif
|
|
|
|
-
|
|
|
|
- mrc p15, 0, r2, c14, c1, 0 @ CNTKCTL
|
|
|
|
- mrrc p15, 0, r4, r5, c7 @ PAR
|
|
|
|
- mrc p15, 0, r6, c10, c3, 0 @ AMAIR0
|
|
|
|
- mrc p15, 0, r7, c10, c3, 1 @ AMAIR1
|
|
|
|
-
|
|
|
|
- .if \store_to_vcpu == 0
|
|
|
|
- push {r2,r4-r7}
|
|
|
|
- .else
|
|
|
|
- str r2, [vcpu, #CP15_OFFSET(c14_CNTKCTL)]
|
|
|
|
- add r12, vcpu, #CP15_OFFSET(c7_PAR)
|
|
|
|
- strd r4, r5, [r12]
|
|
|
|
- str r6, [vcpu, #CP15_OFFSET(c10_AMAIR0)]
|
|
|
|
- str r7, [vcpu, #CP15_OFFSET(c10_AMAIR1)]
|
|
|
|
- .endif
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Reads cp15 registers from memory and writes them to hardware
|
|
|
|
- * @read_from_vcpu: If 0, registers are read in-order from the stack,
|
|
|
|
- * otherwise from the VCPU struct pointed to by vcpup
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- */
|
|
|
|
-.macro write_cp15_state read_from_vcpu
|
|
|
|
- .if \read_from_vcpu == 0
|
|
|
|
- pop {r2,r4-r7}
|
|
|
|
- .else
|
|
|
|
- ldr r2, [vcpu, #CP15_OFFSET(c14_CNTKCTL)]
|
|
|
|
- add r12, vcpu, #CP15_OFFSET(c7_PAR)
|
|
|
|
- ldrd r4, r5, [r12]
|
|
|
|
- ldr r6, [vcpu, #CP15_OFFSET(c10_AMAIR0)]
|
|
|
|
- ldr r7, [vcpu, #CP15_OFFSET(c10_AMAIR1)]
|
|
|
|
- .endif
|
|
|
|
-
|
|
|
|
- mcr p15, 0, r2, c14, c1, 0 @ CNTKCTL
|
|
|
|
- mcrr p15, 0, r4, r5, c7 @ PAR
|
|
|
|
- mcr p15, 0, r6, c10, c3, 0 @ AMAIR0
|
|
|
|
- mcr p15, 0, r7, c10, c3, 1 @ AMAIR1
|
|
|
|
-
|
|
|
|
- .if \read_from_vcpu == 0
|
|
|
|
- pop {r2-r12}
|
|
|
|
- .else
|
|
|
|
- ldr r2, [vcpu, #CP15_OFFSET(c13_CID)]
|
|
|
|
- ldr r3, [vcpu, #CP15_OFFSET(c13_TID_URW)]
|
|
|
|
- ldr r4, [vcpu, #CP15_OFFSET(c13_TID_URO)]
|
|
|
|
- ldr r5, [vcpu, #CP15_OFFSET(c13_TID_PRIV)]
|
|
|
|
- ldr r6, [vcpu, #CP15_OFFSET(c5_DFSR)]
|
|
|
|
- ldr r7, [vcpu, #CP15_OFFSET(c5_IFSR)]
|
|
|
|
- ldr r8, [vcpu, #CP15_OFFSET(c5_ADFSR)]
|
|
|
|
- ldr r9, [vcpu, #CP15_OFFSET(c5_AIFSR)]
|
|
|
|
- ldr r10, [vcpu, #CP15_OFFSET(c6_DFAR)]
|
|
|
|
- ldr r11, [vcpu, #CP15_OFFSET(c6_IFAR)]
|
|
|
|
- ldr r12, [vcpu, #CP15_OFFSET(c12_VBAR)]
|
|
|
|
- .endif
|
|
|
|
-
|
|
|
|
- mcr p15, 0, r2, c13, c0, 1 @ CID
|
|
|
|
- mcr p15, 0, r3, c13, c0, 2 @ TID_URW
|
|
|
|
- mcr p15, 0, r4, c13, c0, 3 @ TID_URO
|
|
|
|
- mcr p15, 0, r5, c13, c0, 4 @ TID_PRIV
|
|
|
|
- mcr p15, 0, r6, c5, c0, 0 @ DFSR
|
|
|
|
- mcr p15, 0, r7, c5, c0, 1 @ IFSR
|
|
|
|
- mcr p15, 0, r8, c5, c1, 0 @ ADFSR
|
|
|
|
- mcr p15, 0, r9, c5, c1, 1 @ AIFSR
|
|
|
|
- mcr p15, 0, r10, c6, c0, 0 @ DFAR
|
|
|
|
- mcr p15, 0, r11, c6, c0, 2 @ IFAR
|
|
|
|
- mcr p15, 0, r12, c12, c0, 0 @ VBAR
|
|
|
|
-
|
|
|
|
- .if \read_from_vcpu == 0
|
|
|
|
- pop {r2-r12}
|
|
|
|
- .else
|
|
|
|
- ldr r2, [vcpu, #CP15_OFFSET(c1_SCTLR)]
|
|
|
|
- ldr r3, [vcpu, #CP15_OFFSET(c1_CPACR)]
|
|
|
|
- ldr r4, [vcpu, #CP15_OFFSET(c2_TTBCR)]
|
|
|
|
- ldr r5, [vcpu, #CP15_OFFSET(c3_DACR)]
|
|
|
|
- add r12, vcpu, #CP15_OFFSET(c2_TTBR0)
|
|
|
|
- ldrd r6, r7, [r12]
|
|
|
|
- add r12, vcpu, #CP15_OFFSET(c2_TTBR1)
|
|
|
|
- ldrd r8, r9, [r12]
|
|
|
|
- ldr r10, [vcpu, #CP15_OFFSET(c10_PRRR)]
|
|
|
|
- ldr r11, [vcpu, #CP15_OFFSET(c10_NMRR)]
|
|
|
|
- ldr r12, [vcpu, #CP15_OFFSET(c0_CSSELR)]
|
|
|
|
- .endif
|
|
|
|
-
|
|
|
|
- mcr p15, 0, r2, c1, c0, 0 @ SCTLR
|
|
|
|
- mcr p15, 0, r3, c1, c0, 2 @ CPACR
|
|
|
|
- mcr p15, 0, r4, c2, c0, 2 @ TTBCR
|
|
|
|
- mcr p15, 0, r5, c3, c0, 0 @ DACR
|
|
|
|
- mcrr p15, 0, r6, r7, c2 @ TTBR 0
|
|
|
|
- mcrr p15, 1, r8, r9, c2 @ TTBR 1
|
|
|
|
- mcr p15, 0, r10, c10, c2, 0 @ PRRR
|
|
|
|
- mcr p15, 0, r11, c10, c2, 1 @ NMRR
|
|
|
|
- mcr p15, 2, r12, c0, c0, 0 @ CSSELR
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Save the VGIC CPU state into memory
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- */
|
|
|
|
-.macro save_vgic_state
|
|
|
|
- /* Get VGIC VCTRL base into r2 */
|
|
|
|
- ldr r2, [vcpu, #VCPU_KVM]
|
|
|
|
- ldr r2, [r2, #KVM_VGIC_VCTRL]
|
|
|
|
- cmp r2, #0
|
|
|
|
- beq 2f
|
|
|
|
-
|
|
|
|
- /* Compute the address of struct vgic_cpu */
|
|
|
|
- add r11, vcpu, #VCPU_VGIC_CPU
|
|
|
|
-
|
|
|
|
- /* Save all interesting registers */
|
|
|
|
- ldr r4, [r2, #GICH_VMCR]
|
|
|
|
- ldr r5, [r2, #GICH_MISR]
|
|
|
|
- ldr r6, [r2, #GICH_EISR0]
|
|
|
|
- ldr r7, [r2, #GICH_EISR1]
|
|
|
|
- ldr r8, [r2, #GICH_ELRSR0]
|
|
|
|
- ldr r9, [r2, #GICH_ELRSR1]
|
|
|
|
- ldr r10, [r2, #GICH_APR]
|
|
|
|
-ARM_BE8(rev r4, r4 )
|
|
|
|
-ARM_BE8(rev r5, r5 )
|
|
|
|
-ARM_BE8(rev r6, r6 )
|
|
|
|
-ARM_BE8(rev r7, r7 )
|
|
|
|
-ARM_BE8(rev r8, r8 )
|
|
|
|
-ARM_BE8(rev r9, r9 )
|
|
|
|
-ARM_BE8(rev r10, r10 )
|
|
|
|
-
|
|
|
|
- str r4, [r11, #VGIC_V2_CPU_VMCR]
|
|
|
|
- str r5, [r11, #VGIC_V2_CPU_MISR]
|
|
|
|
-#ifdef CONFIG_CPU_ENDIAN_BE8
|
|
|
|
- str r6, [r11, #(VGIC_V2_CPU_EISR + 4)]
|
|
|
|
- str r7, [r11, #VGIC_V2_CPU_EISR]
|
|
|
|
- str r8, [r11, #(VGIC_V2_CPU_ELRSR + 4)]
|
|
|
|
- str r9, [r11, #VGIC_V2_CPU_ELRSR]
|
|
|
|
-#else
|
|
|
|
- str r6, [r11, #VGIC_V2_CPU_EISR]
|
|
|
|
- str r7, [r11, #(VGIC_V2_CPU_EISR + 4)]
|
|
|
|
- str r8, [r11, #VGIC_V2_CPU_ELRSR]
|
|
|
|
- str r9, [r11, #(VGIC_V2_CPU_ELRSR + 4)]
|
|
|
|
-#endif
|
|
|
|
- str r10, [r11, #VGIC_V2_CPU_APR]
|
|
|
|
-
|
|
|
|
- /* Clear GICH_HCR */
|
|
|
|
- mov r5, #0
|
|
|
|
- str r5, [r2, #GICH_HCR]
|
|
|
|
-
|
|
|
|
- /* Save list registers */
|
|
|
|
- add r2, r2, #GICH_LR0
|
|
|
|
- add r3, r11, #VGIC_V2_CPU_LR
|
|
|
|
- ldr r4, [r11, #VGIC_CPU_NR_LR]
|
|
|
|
-1: ldr r6, [r2], #4
|
|
|
|
-ARM_BE8(rev r6, r6 )
|
|
|
|
- str r6, [r3], #4
|
|
|
|
- subs r4, r4, #1
|
|
|
|
- bne 1b
|
|
|
|
-2:
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Restore the VGIC CPU state from memory
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- */
|
|
|
|
-.macro restore_vgic_state
|
|
|
|
- /* Get VGIC VCTRL base into r2 */
|
|
|
|
- ldr r2, [vcpu, #VCPU_KVM]
|
|
|
|
- ldr r2, [r2, #KVM_VGIC_VCTRL]
|
|
|
|
- cmp r2, #0
|
|
|
|
- beq 2f
|
|
|
|
-
|
|
|
|
- /* Compute the address of struct vgic_cpu */
|
|
|
|
- add r11, vcpu, #VCPU_VGIC_CPU
|
|
|
|
-
|
|
|
|
- /* We only restore a minimal set of registers */
|
|
|
|
- ldr r3, [r11, #VGIC_V2_CPU_HCR]
|
|
|
|
- ldr r4, [r11, #VGIC_V2_CPU_VMCR]
|
|
|
|
- ldr r8, [r11, #VGIC_V2_CPU_APR]
|
|
|
|
-ARM_BE8(rev r3, r3 )
|
|
|
|
-ARM_BE8(rev r4, r4 )
|
|
|
|
-ARM_BE8(rev r8, r8 )
|
|
|
|
-
|
|
|
|
- str r3, [r2, #GICH_HCR]
|
|
|
|
- str r4, [r2, #GICH_VMCR]
|
|
|
|
- str r8, [r2, #GICH_APR]
|
|
|
|
-
|
|
|
|
- /* Restore list registers */
|
|
|
|
- add r2, r2, #GICH_LR0
|
|
|
|
- add r3, r11, #VGIC_V2_CPU_LR
|
|
|
|
- ldr r4, [r11, #VGIC_CPU_NR_LR]
|
|
|
|
-1: ldr r6, [r3], #4
|
|
|
|
-ARM_BE8(rev r6, r6 )
|
|
|
|
- str r6, [r2], #4
|
|
|
|
- subs r4, r4, #1
|
|
|
|
- bne 1b
|
|
|
|
-2:
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-#define CNTHCTL_PL1PCTEN (1 << 0)
|
|
|
|
-#define CNTHCTL_PL1PCEN (1 << 1)
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Save the timer state onto the VCPU and allow physical timer/counter access
|
|
|
|
- * for the host.
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- * Clobbers r2-r5
|
|
|
|
- */
|
|
|
|
-.macro save_timer_state
|
|
|
|
- ldr r4, [vcpu, #VCPU_KVM]
|
|
|
|
- ldr r2, [r4, #KVM_TIMER_ENABLED]
|
|
|
|
- cmp r2, #0
|
|
|
|
- beq 1f
|
|
|
|
-
|
|
|
|
- mrc p15, 0, r2, c14, c3, 1 @ CNTV_CTL
|
|
|
|
- str r2, [vcpu, #VCPU_TIMER_CNTV_CTL]
|
|
|
|
-
|
|
|
|
- isb
|
|
|
|
-
|
|
|
|
- mrrc p15, 3, rr_lo_hi(r2, r3), c14 @ CNTV_CVAL
|
|
|
|
- ldr r4, =VCPU_TIMER_CNTV_CVAL
|
|
|
|
- add r5, vcpu, r4
|
|
|
|
- strd r2, r3, [r5]
|
|
|
|
-
|
|
|
|
- @ Ensure host CNTVCT == CNTPCT
|
|
|
|
- mov r2, #0
|
|
|
|
- mcrr p15, 4, r2, r2, c14 @ CNTVOFF
|
|
|
|
-
|
|
|
|
-1:
|
|
|
|
- mov r2, #0 @ Clear ENABLE
|
|
|
|
- mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL
|
|
|
|
-
|
|
|
|
- @ Allow physical timer/counter access for the host
|
|
|
|
- mrc p15, 4, r2, c14, c1, 0 @ CNTHCTL
|
|
|
|
- orr r2, r2, #(CNTHCTL_PL1PCEN | CNTHCTL_PL1PCTEN)
|
|
|
|
- mcr p15, 4, r2, c14, c1, 0 @ CNTHCTL
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/*
|
|
|
|
- * Load the timer state from the VCPU and deny physical timer/counter access
|
|
|
|
- * for the host.
|
|
|
|
- *
|
|
|
|
- * Assumes vcpu pointer in vcpu reg
|
|
|
|
- * Clobbers r2-r5
|
|
|
|
- */
|
|
|
|
-.macro restore_timer_state
|
|
|
|
- @ Disallow physical timer access for the guest
|
|
|
|
- @ Physical counter access is allowed
|
|
|
|
- mrc p15, 4, r2, c14, c1, 0 @ CNTHCTL
|
|
|
|
- orr r2, r2, #CNTHCTL_PL1PCTEN
|
|
|
|
- bic r2, r2, #CNTHCTL_PL1PCEN
|
|
|
|
- mcr p15, 4, r2, c14, c1, 0 @ CNTHCTL
|
|
|
|
-
|
|
|
|
- ldr r4, [vcpu, #VCPU_KVM]
|
|
|
|
- ldr r2, [r4, #KVM_TIMER_ENABLED]
|
|
|
|
- cmp r2, #0
|
|
|
|
- beq 1f
|
|
|
|
-
|
|
|
|
- ldr r2, [r4, #KVM_TIMER_CNTVOFF]
|
|
|
|
- ldr r3, [r4, #(KVM_TIMER_CNTVOFF + 4)]
|
|
|
|
- mcrr p15, 4, rr_lo_hi(r2, r3), c14 @ CNTVOFF
|
|
|
|
-
|
|
|
|
- ldr r4, =VCPU_TIMER_CNTV_CVAL
|
|
|
|
- add r5, vcpu, r4
|
|
|
|
- ldrd r2, r3, [r5]
|
|
|
|
- mcrr p15, 3, rr_lo_hi(r2, r3), c14 @ CNTV_CVAL
|
|
|
|
- isb
|
|
|
|
-
|
|
|
|
- ldr r2, [vcpu, #VCPU_TIMER_CNTV_CTL]
|
|
|
|
- and r2, r2, #3
|
|
|
|
- mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL
|
|
|
|
-1:
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-.equ vmentry, 0
|
|
|
|
-.equ vmexit, 1
|
|
|
|
-
|
|
|
|
-/* Configures the HSTR (Hyp System Trap Register) on entry/return
|
|
|
|
- * (hardware reset value is 0) */
|
|
|
|
-.macro set_hstr operation
|
|
|
|
- mrc p15, 4, r2, c1, c1, 3
|
|
|
|
- ldr r3, =HSTR_T(15)
|
|
|
|
- .if \operation == vmentry
|
|
|
|
- orr r2, r2, r3 @ Trap CR{15}
|
|
|
|
- .else
|
|
|
|
- bic r2, r2, r3 @ Don't trap any CRx accesses
|
|
|
|
- .endif
|
|
|
|
- mcr p15, 4, r2, c1, c1, 3
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/* Configures the HCPTR (Hyp Coprocessor Trap Register) on entry/return
|
|
|
|
- * (hardware reset value is 0). Keep previous value in r2.
|
|
|
|
- * An ISB is emited on vmexit/vmtrap, but executed on vmexit only if
|
|
|
|
- * VFP wasn't already enabled (always executed on vmtrap).
|
|
|
|
- * If a label is specified with vmexit, it is branched to if VFP wasn't
|
|
|
|
- * enabled.
|
|
|
|
- */
|
|
|
|
-.macro set_hcptr operation, mask, label = none
|
|
|
|
- mrc p15, 4, r2, c1, c1, 2
|
|
|
|
- ldr r3, =\mask
|
|
|
|
- .if \operation == vmentry
|
|
|
|
- orr r3, r2, r3 @ Trap coproc-accesses defined in mask
|
|
|
|
- .else
|
|
|
|
- bic r3, r2, r3 @ Don't trap defined coproc-accesses
|
|
|
|
- .endif
|
|
|
|
- mcr p15, 4, r3, c1, c1, 2
|
|
|
|
- .if \operation != vmentry
|
|
|
|
- .if \operation == vmexit
|
|
|
|
- tst r2, #(HCPTR_TCP(10) | HCPTR_TCP(11))
|
|
|
|
- beq 1f
|
|
|
|
- .endif
|
|
|
|
- isb
|
|
|
|
- .if \label != none
|
|
|
|
- b \label
|
|
|
|
- .endif
|
|
|
|
-1:
|
|
|
|
- .endif
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/* Configures the HDCR (Hyp Debug Configuration Register) on entry/return
|
|
|
|
- * (hardware reset value is 0) */
|
|
|
|
-.macro set_hdcr operation
|
|
|
|
- mrc p15, 4, r2, c1, c1, 1
|
|
|
|
- ldr r3, =(HDCR_TPM|HDCR_TPMCR)
|
|
|
|
- .if \operation == vmentry
|
|
|
|
- orr r2, r2, r3 @ Trap some perfmon accesses
|
|
|
|
- .else
|
|
|
|
- bic r2, r2, r3 @ Don't trap any perfmon accesses
|
|
|
|
- .endif
|
|
|
|
- mcr p15, 4, r2, c1, c1, 1
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-/* Enable/Disable: stage-2 trans., trap interrupts, trap wfi, trap smc */
|
|
|
|
-.macro configure_hyp_role operation
|
|
|
|
- .if \operation == vmentry
|
|
|
|
- ldr r2, [vcpu, #VCPU_HCR]
|
|
|
|
- ldr r3, [vcpu, #VCPU_IRQ_LINES]
|
|
|
|
- orr r2, r2, r3
|
|
|
|
- .else
|
|
|
|
- mov r2, #0
|
|
|
|
- .endif
|
|
|
|
- mcr p15, 4, r2, c1, c1, 0 @ HCR
|
|
|
|
-.endm
|
|
|
|
-
|
|
|
|
-.macro load_vcpu
|
|
|
|
- mrc p15, 4, vcpu, c13, c0, 2 @ HTPIDR
|
|
|
|
-.endm
|
|
|