|
@@ -1424,6 +1424,16 @@ void s390_vcpu_unblock(struct kvm_vcpu *vcpu)
|
|
|
atomic_clear_mask(PROG_BLOCK_SIE, &vcpu->arch.sie_block->prog20);
|
|
|
}
|
|
|
|
|
|
+static void kvm_s390_vcpu_request(struct kvm_vcpu *vcpu)
|
|
|
+{
|
|
|
+ atomic_set_mask(PROG_REQUEST, &vcpu->arch.sie_block->prog20);
|
|
|
+}
|
|
|
+
|
|
|
+static void kvm_s390_vcpu_request_handled(struct kvm_vcpu *vcpu)
|
|
|
+{
|
|
|
+ atomic_clear_mask(PROG_REQUEST, &vcpu->arch.sie_block->prog20);
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* Kick a guest cpu out of SIE and wait until SIE is not running.
|
|
|
* If the CPU is not running (e.g. waiting as idle) the function will
|
|
@@ -1435,10 +1445,11 @@ void exit_sie(struct kvm_vcpu *vcpu)
|
|
|
cpu_relax();
|
|
|
}
|
|
|
|
|
|
-/* Kick a guest cpu out of SIE and prevent SIE-reentry */
|
|
|
-void exit_sie_sync(struct kvm_vcpu *vcpu)
|
|
|
+/* Kick a guest cpu out of SIE to process a request synchronously */
|
|
|
+void kvm_s390_sync_request(int req, struct kvm_vcpu *vcpu)
|
|
|
{
|
|
|
- s390_vcpu_block(vcpu);
|
|
|
+ kvm_make_request(req, vcpu);
|
|
|
+ kvm_s390_vcpu_request(vcpu);
|
|
|
exit_sie(vcpu);
|
|
|
}
|
|
|
|
|
@@ -1452,8 +1463,7 @@ static void kvm_gmap_notifier(struct gmap *gmap, unsigned long address)
|
|
|
/* match against both prefix pages */
|
|
|
if (kvm_s390_get_prefix(vcpu) == (address & ~0x1000UL)) {
|
|
|
VCPU_EVENT(vcpu, 2, "gmap notifier for %lx", address);
|
|
|
- kvm_make_request(KVM_REQ_MMU_RELOAD, vcpu);
|
|
|
- exit_sie_sync(vcpu);
|
|
|
+ kvm_s390_sync_request(KVM_REQ_MMU_RELOAD, vcpu);
|
|
|
}
|
|
|
}
|
|
|
}
|
|
@@ -1728,7 +1738,7 @@ static int kvm_s390_handle_requests(struct kvm_vcpu *vcpu)
|
|
|
if (!vcpu->requests)
|
|
|
return 0;
|
|
|
retry:
|
|
|
- s390_vcpu_unblock(vcpu);
|
|
|
+ kvm_s390_vcpu_request_handled(vcpu);
|
|
|
/*
|
|
|
* We use MMU_RELOAD just to re-arm the ipte notifier for the
|
|
|
* guest prefix page. gmap_ipte_notify will wait on the ptl lock.
|
|
@@ -2213,8 +2223,7 @@ int kvm_s390_vcpu_store_adtl_status(struct kvm_vcpu *vcpu, unsigned long addr)
|
|
|
static void __disable_ibs_on_vcpu(struct kvm_vcpu *vcpu)
|
|
|
{
|
|
|
kvm_check_request(KVM_REQ_ENABLE_IBS, vcpu);
|
|
|
- kvm_make_request(KVM_REQ_DISABLE_IBS, vcpu);
|
|
|
- exit_sie_sync(vcpu);
|
|
|
+ kvm_s390_sync_request(KVM_REQ_DISABLE_IBS, vcpu);
|
|
|
}
|
|
|
|
|
|
static void __disable_ibs_on_all_vcpus(struct kvm *kvm)
|
|
@@ -2230,8 +2239,7 @@ static void __disable_ibs_on_all_vcpus(struct kvm *kvm)
|
|
|
static void __enable_ibs_on_vcpu(struct kvm_vcpu *vcpu)
|
|
|
{
|
|
|
kvm_check_request(KVM_REQ_DISABLE_IBS, vcpu);
|
|
|
- kvm_make_request(KVM_REQ_ENABLE_IBS, vcpu);
|
|
|
- exit_sie_sync(vcpu);
|
|
|
+ kvm_s390_sync_request(KVM_REQ_ENABLE_IBS, vcpu);
|
|
|
}
|
|
|
|
|
|
void kvm_s390_vcpu_start(struct kvm_vcpu *vcpu)
|