|
|
@@ -812,8 +812,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_ARCH_206)
|
|
|
12: mtspr SPRN_SRR0, r10
|
|
|
mr r10,r0
|
|
|
mtspr SPRN_SRR1, r11
|
|
|
- li r11,(MSR_ME << 1) | 1 /* synthesize MSR_SF | MSR_ME */
|
|
|
- rotldi r11,r11,63
|
|
|
+ ld r11, VCPU_INTR_MSR(r4)
|
|
|
5:
|
|
|
|
|
|
/*
|
|
|
@@ -1551,8 +1550,7 @@ kvmppc_hdsi:
|
|
|
mtspr SPRN_SRR0, r10
|
|
|
mtspr SPRN_SRR1, r11
|
|
|
li r10, BOOK3S_INTERRUPT_DATA_STORAGE
|
|
|
- li r11, (MSR_ME << 1) | 1 /* synthesize MSR_SF | MSR_ME */
|
|
|
- rotldi r11, r11, 63
|
|
|
+ ld r11, VCPU_INTR_MSR(r9)
|
|
|
fast_interrupt_c_return:
|
|
|
6: ld r7, VCPU_CTR(r9)
|
|
|
lwz r8, VCPU_XER(r9)
|
|
|
@@ -1621,8 +1619,7 @@ kvmppc_hisi:
|
|
|
1: mtspr SPRN_SRR0, r10
|
|
|
mtspr SPRN_SRR1, r11
|
|
|
li r10, BOOK3S_INTERRUPT_INST_STORAGE
|
|
|
- li r11, (MSR_ME << 1) | 1 /* synthesize MSR_SF | MSR_ME */
|
|
|
- rotldi r11, r11, 63
|
|
|
+ ld r11, VCPU_INTR_MSR(r9)
|
|
|
b fast_interrupt_c_return
|
|
|
|
|
|
3: ld r6, VCPU_KVM(r9) /* not relocated, use VRMA */
|
|
|
@@ -1665,8 +1662,7 @@ sc_1_fast_return:
|
|
|
mtspr SPRN_SRR0,r10
|
|
|
mtspr SPRN_SRR1,r11
|
|
|
li r10, BOOK3S_INTERRUPT_SYSCALL
|
|
|
- li r11, (MSR_ME << 1) | 1 /* synthesize MSR_SF | MSR_ME */
|
|
|
- rotldi r11, r11, 63
|
|
|
+ ld r11, VCPU_INTR_MSR(r9)
|
|
|
mr r4,r9
|
|
|
b fast_guest_return
|
|
|
|
|
|
@@ -1994,8 +1990,7 @@ machine_check_realmode:
|
|
|
beq mc_cont
|
|
|
/* If not, deliver a machine check. SRR0/1 are already set */
|
|
|
li r10, BOOK3S_INTERRUPT_MACHINE_CHECK
|
|
|
- li r11, (MSR_ME << 1) | 1 /* synthesize MSR_SF | MSR_ME */
|
|
|
- rotldi r11, r11, 63
|
|
|
+ ld r11, VCPU_INTR_MSR(r9)
|
|
|
b fast_interrupt_c_return
|
|
|
|
|
|
/*
|