|
@@ -9967,6 +9967,15 @@ static void prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12)
|
|
|
*/
|
|
|
vmx_set_constant_host_state(vmx);
|
|
|
|
|
|
+ /*
|
|
|
+ * Set the MSR load/store lists to match L0's settings.
|
|
|
+ */
|
|
|
+ vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0);
|
|
|
+ vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, vmx->msr_autoload.nr);
|
|
|
+ vmcs_write64(VM_EXIT_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.host));
|
|
|
+ vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, vmx->msr_autoload.nr);
|
|
|
+ vmcs_write64(VM_ENTRY_MSR_LOAD_ADDR, __pa(vmx->msr_autoload.guest));
|
|
|
+
|
|
|
/*
|
|
|
* HOST_RSP is normally set correctly in vmx_vcpu_run() just before
|
|
|
* entry, but only if the current (host) sp changed from the value
|
|
@@ -10799,6 +10808,8 @@ static void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 exit_reason,
|
|
|
load_vmcs12_host_state(vcpu, vmcs12);
|
|
|
|
|
|
/* Update any VMCS fields that might have changed while L2 ran */
|
|
|
+ vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, vmx->msr_autoload.nr);
|
|
|
+ vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, vmx->msr_autoload.nr);
|
|
|
vmcs_write64(TSC_OFFSET, vcpu->arch.tsc_offset);
|
|
|
if (vmx->hv_deadline_tsc == -1)
|
|
|
vmcs_clear_bits(PIN_BASED_VM_EXEC_CONTROL,
|