|
@@ -463,8 +463,6 @@ void kvm_exit(void);
|
|
|
|
|
|
|
|
void kvm_get_kvm(struct kvm *kvm);
|
|
void kvm_get_kvm(struct kvm *kvm);
|
|
|
void kvm_put_kvm(struct kvm *kvm);
|
|
void kvm_put_kvm(struct kvm *kvm);
|
|
|
-void update_memslots(struct kvm_memslots *slots, struct kvm_memory_slot *new,
|
|
|
|
|
- u64 last_generation);
|
|
|
|
|
|
|
|
|
|
static inline struct kvm_memslots *kvm_memslots(struct kvm *kvm)
|
|
static inline struct kvm_memslots *kvm_memslots(struct kvm *kvm)
|
|
|
{
|
|
{
|
|
@@ -537,7 +535,6 @@ unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
|
|
|
unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
|
|
unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
|
|
|
void kvm_release_page_clean(struct page *page);
|
|
void kvm_release_page_clean(struct page *page);
|
|
|
void kvm_release_page_dirty(struct page *page);
|
|
void kvm_release_page_dirty(struct page *page);
|
|
|
-void kvm_set_page_dirty(struct page *page);
|
|
|
|
|
void kvm_set_page_accessed(struct page *page);
|
|
void kvm_set_page_accessed(struct page *page);
|
|
|
|
|
|
|
|
pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn);
|
|
pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn);
|
|
@@ -549,7 +546,6 @@ pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
|
|
|
pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
|
|
pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
|
|
|
pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn);
|
|
pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn);
|
|
|
|
|
|
|
|
-void kvm_release_pfn_dirty(pfn_t pfn);
|
|
|
|
|
void kvm_release_pfn_clean(pfn_t pfn);
|
|
void kvm_release_pfn_clean(pfn_t pfn);
|
|
|
void kvm_set_pfn_dirty(pfn_t pfn);
|
|
void kvm_set_pfn_dirty(pfn_t pfn);
|
|
|
void kvm_set_pfn_accessed(pfn_t pfn);
|
|
void kvm_set_pfn_accessed(pfn_t pfn);
|
|
@@ -576,8 +572,6 @@ struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
|
|
|
int kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
|
|
int kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
|
|
|
unsigned long kvm_host_page_size(struct kvm *kvm, gfn_t gfn);
|
|
unsigned long kvm_host_page_size(struct kvm *kvm, gfn_t gfn);
|
|
|
void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
|
|
void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
|
|
|
-void mark_page_dirty_in_slot(struct kvm *kvm, struct kvm_memory_slot *memslot,
|
|
|
|
|
- gfn_t gfn);
|
|
|
|
|
|
|
|
|
|
void kvm_vcpu_block(struct kvm_vcpu *vcpu);
|
|
void kvm_vcpu_block(struct kvm_vcpu *vcpu);
|
|
|
void kvm_vcpu_kick(struct kvm_vcpu *vcpu);
|
|
void kvm_vcpu_kick(struct kvm_vcpu *vcpu);
|
|
@@ -604,8 +598,6 @@ int kvm_get_dirty_log(struct kvm *kvm,
|
|
|
int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm,
|
|
int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm,
|
|
|
struct kvm_dirty_log *log);
|
|
struct kvm_dirty_log *log);
|
|
|
|
|
|
|
|
-int kvm_vm_ioctl_set_memory_region(struct kvm *kvm,
|
|
|
|
|
- struct kvm_userspace_memory_region *mem);
|
|
|
|
|
int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_level,
|
|
int kvm_vm_ioctl_irq_line(struct kvm *kvm, struct kvm_irq_level *irq_level,
|
|
|
bool line_status);
|
|
bool line_status);
|
|
|
long kvm_arch_vm_ioctl(struct file *filp,
|
|
long kvm_arch_vm_ioctl(struct file *filp,
|
|
@@ -653,8 +645,6 @@ void kvm_arch_check_processor_compat(void *rtn);
|
|
|
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu);
|
|
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu);
|
|
|
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu);
|
|
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
-void kvm_free_physmem(struct kvm *kvm);
|
|
|
|
|
-
|
|
|
|
|
void *kvm_kvzalloc(unsigned long size);
|
|
void *kvm_kvzalloc(unsigned long size);
|
|
|
void kvm_kvfree(const void *addr);
|
|
void kvm_kvfree(const void *addr);
|
|
|
|
|
|
|
@@ -1097,12 +1087,6 @@ static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val)
|
|
|
static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val)
|
|
static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val)
|
|
|
{
|
|
{
|
|
|
}
|
|
}
|
|
|
-
|
|
|
|
|
-static inline bool kvm_vcpu_eligible_for_directed_yield(struct kvm_vcpu *vcpu)
|
|
|
|
|
-{
|
|
|
|
|
- return true;
|
|
|
|
|
-}
|
|
|
|
|
-
|
|
|
|
|
#endif /* CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT */
|
|
#endif /* CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT */
|
|
|
#endif
|
|
#endif
|
|
|
|
|
|