|
@@ -61,6 +61,14 @@ static pfn_t kvm_pin_pages(struct kvm_memory_slot *slot, gfn_t gfn,
|
|
return pfn;
|
|
return pfn;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
+static void kvm_unpin_pages(struct kvm *kvm, pfn_t pfn, unsigned long npages)
|
|
|
|
+{
|
|
|
|
+ unsigned long i;
|
|
|
|
+
|
|
|
|
+ for (i = 0; i < npages; ++i)
|
|
|
|
+ kvm_release_pfn_clean(pfn + i);
|
|
|
|
+}
|
|
|
|
+
|
|
int kvm_iommu_map_pages(struct kvm *kvm, struct kvm_memory_slot *slot)
|
|
int kvm_iommu_map_pages(struct kvm *kvm, struct kvm_memory_slot *slot)
|
|
{
|
|
{
|
|
gfn_t gfn, end_gfn;
|
|
gfn_t gfn, end_gfn;
|
|
@@ -123,6 +131,7 @@ int kvm_iommu_map_pages(struct kvm *kvm, struct kvm_memory_slot *slot)
|
|
if (r) {
|
|
if (r) {
|
|
printk(KERN_ERR "kvm_iommu_map_address:"
|
|
printk(KERN_ERR "kvm_iommu_map_address:"
|
|
"iommu failed to map pfn=%llx\n", pfn);
|
|
"iommu failed to map pfn=%llx\n", pfn);
|
|
|
|
+ kvm_unpin_pages(kvm, pfn, page_size);
|
|
goto unmap_pages;
|
|
goto unmap_pages;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -134,7 +143,7 @@ int kvm_iommu_map_pages(struct kvm *kvm, struct kvm_memory_slot *slot)
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
unmap_pages:
|
|
unmap_pages:
|
|
- kvm_iommu_put_pages(kvm, slot->base_gfn, gfn);
|
|
|
|
|
|
+ kvm_iommu_put_pages(kvm, slot->base_gfn, gfn - slot->base_gfn);
|
|
return r;
|
|
return r;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -266,14 +275,6 @@ out_unlock:
|
|
return r;
|
|
return r;
|
|
}
|
|
}
|
|
|
|
|
|
-static void kvm_unpin_pages(struct kvm *kvm, pfn_t pfn, unsigned long npages)
|
|
|
|
-{
|
|
|
|
- unsigned long i;
|
|
|
|
-
|
|
|
|
- for (i = 0; i < npages; ++i)
|
|
|
|
- kvm_release_pfn_clean(pfn + i);
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
static void kvm_iommu_put_pages(struct kvm *kvm,
|
|
static void kvm_iommu_put_pages(struct kvm *kvm,
|
|
gfn_t base_gfn, unsigned long npages)
|
|
gfn_t base_gfn, unsigned long npages)
|
|
{
|
|
{
|