KVM: Introduce kvm_vcpu_destroy()
authorSean Christopherson <sean.j.christopherson@intel.com>
Wed, 18 Dec 2019 21:55:14 +0000 (13:55 -0800)
committerPaolo Bonzini <pbonzini@redhat.com>
Fri, 24 Jan 2020 08:19:11 +0000 (09:19 +0100)
Add kvm_vcpu_destroy() and wire up all architectures to call the common
function instead of their arch specific implementation.  The common
destruction function will be used by future patches to move allocation
and initialization of vCPUs to common KVM code, i.e. to free resources
that are allocated by arch agnostic code.

No functional change intended.

Acked-by: Christoffer Dall <christoffer.dall@arm.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Reviewed-by: Cornelia Huck <cohuck@redhat.com>
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
arch/mips/kvm/mips.c
arch/powerpc/kvm/powerpc.c
arch/s390/kvm/kvm-s390.c
arch/x86/kvm/x86.c
include/linux/kvm_host.h
virt/kvm/arm/arm.c
virt/kvm/kvm_main.c

index 73360e0..8546bc6 100644 (file)
@@ -156,7 +156,7 @@ void kvm_mips_free_vcpus(struct kvm *kvm)
        struct kvm_vcpu *vcpu;
 
        kvm_for_each_vcpu(i, vcpu, kvm) {
-               kvm_arch_vcpu_destroy(vcpu);
+               kvm_vcpu_destroy(vcpu);
        }
 
        mutex_lock(&kvm->lock);
index 998ef60..e3e2b88 100644 (file)
@@ -475,7 +475,7 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
 #endif
 
        kvm_for_each_vcpu(i, vcpu, kvm)
-               kvm_arch_vcpu_destroy(vcpu);
+               kvm_vcpu_destroy(vcpu);
 
        mutex_lock(&kvm->lock);
        for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
index 1f8ba07..8543d33 100644 (file)
@@ -2541,7 +2541,7 @@ static void kvm_free_vcpus(struct kvm *kvm)
        struct kvm_vcpu *vcpu;
 
        kvm_for_each_vcpu(i, vcpu, kvm)
-               kvm_arch_vcpu_destroy(vcpu);
+               kvm_vcpu_destroy(vcpu);
 
        mutex_lock(&kvm->lock);
        for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
index 335762a..42b9149 100644 (file)
@@ -9680,7 +9680,7 @@ static void kvm_free_vcpus(struct kvm *kvm)
                kvm_unload_vcpu_mmu(vcpu);
        }
        kvm_for_each_vcpu(i, vcpu, kvm)
-               kvm_arch_vcpu_destroy(vcpu);
+               kvm_vcpu_destroy(vcpu);
 
        mutex_lock(&kvm->lock);
        for (i = 0; i < atomic_read(&kvm->online_vcpus); i++)
index 59ac534..432827a 100644 (file)
@@ -581,6 +581,7 @@ static inline int kvm_vcpu_get_idx(struct kvm_vcpu *vcpu)
 
 int kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id);
 void kvm_vcpu_uninit(struct kvm_vcpu *vcpu);
+void kvm_vcpu_destroy(struct kvm_vcpu *vcpu);
 
 void vcpu_load(struct kvm_vcpu *vcpu);
 void vcpu_put(struct kvm_vcpu *vcpu);
index af3ce2b..0d8fb69 100644 (file)
@@ -194,7 +194,7 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
 
        for (i = 0; i < KVM_MAX_VCPUS; ++i) {
                if (kvm->vcpus[i]) {
-                       kvm_arch_vcpu_destroy(kvm->vcpus[i]);
+                       kvm_vcpu_destroy(kvm->vcpus[i]);
                        kvm->vcpus[i] = NULL;
                }
        }
index 7b52207..62ba25e 100644 (file)
@@ -375,6 +375,12 @@ void kvm_vcpu_uninit(struct kvm_vcpu *vcpu)
 }
 EXPORT_SYMBOL_GPL(kvm_vcpu_uninit);
 
+void kvm_vcpu_destroy(struct kvm_vcpu *vcpu)
+{
+       kvm_arch_vcpu_destroy(vcpu);
+}
+EXPORT_SYMBOL_GPL(kvm_vcpu_destroy);
+
 #if defined(CONFIG_MMU_NOTIFIER) && defined(KVM_ARCH_WANT_MMU_NOTIFIER)
 static inline struct kvm *mmu_notifier_to_kvm(struct mmu_notifier *mn)
 {