|
@@ -21,11 +21,13 @@
|
|
|
|
|
|
#include <asm/kvm_hyp.h>
|
|
|
|
|
|
+extern struct vgic_params vgic_v2_params;
|
|
|
+
|
|
|
static void __hyp_text save_maint_int_state(struct kvm_vcpu *vcpu,
|
|
|
void __iomem *base)
|
|
|
{
|
|
|
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
|
|
|
- int nr_lr = vcpu->arch.vgic_cpu.nr_lr;
|
|
|
+ int nr_lr = (kern_hyp_va(&vgic_v2_params))->nr_lr;
|
|
|
u32 eisr0, eisr1;
|
|
|
int i;
|
|
|
bool expect_mi;
|
|
@@ -67,7 +69,7 @@ static void __hyp_text save_maint_int_state(struct kvm_vcpu *vcpu,
|
|
|
static void __hyp_text save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base)
|
|
|
{
|
|
|
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
|
|
|
- int nr_lr = vcpu->arch.vgic_cpu.nr_lr;
|
|
|
+ int nr_lr = (kern_hyp_va(&vgic_v2_params))->nr_lr;
|
|
|
u32 elrsr0, elrsr1;
|
|
|
|
|
|
elrsr0 = readl_relaxed(base + GICH_ELRSR0);
|
|
@@ -86,7 +88,7 @@ static void __hyp_text save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base)
|
|
|
static void __hyp_text save_lrs(struct kvm_vcpu *vcpu, void __iomem *base)
|
|
|
{
|
|
|
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
|
|
|
- int nr_lr = vcpu->arch.vgic_cpu.nr_lr;
|
|
|
+ int nr_lr = (kern_hyp_va(&vgic_v2_params))->nr_lr;
|
|
|
int i;
|
|
|
|
|
|
for (i = 0; i < nr_lr; i++) {
|
|
@@ -141,13 +143,13 @@ void __hyp_text __vgic_v2_restore_state(struct kvm_vcpu *vcpu)
|
|
|
struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2;
|
|
|
struct vgic_dist *vgic = &kvm->arch.vgic;
|
|
|
void __iomem *base = kern_hyp_va(vgic->vctrl_base);
|
|
|
- int i, nr_lr;
|
|
|
+ int nr_lr = (kern_hyp_va(&vgic_v2_params))->nr_lr;
|
|
|
+ int i;
|
|
|
u64 live_lrs = 0;
|
|
|
|
|
|
if (!base)
|
|
|
return;
|
|
|
|
|
|
- nr_lr = vcpu->arch.vgic_cpu.nr_lr;
|
|
|
|
|
|
for (i = 0; i < nr_lr; i++)
|
|
|
if (cpu_if->vgic_lr[i] & GICH_LR_STATE)
|