Commit a81d95ae authored by Paolo Bonzini's avatar Paolo Bonzini

Merge tag 'kvm-x86-asyncpf-6.9' of https://github.com/kvm-x86/linux into HEAD

KVM async page fault changes for 6.9:

 - Always flush the async page fault workqueue when a work item is being
   removed, especially during vCPU destruction, to ensure that there are no
   workers running in KVM code when all references to KVM-the-module are gone,
   i.e. to prevent a use-after-free if kvm.ko is unloaded.

 - Grab a reference to the VM's mm_struct in the async #PF worker itself instead
   of gifting the worker a reference, e.g. so that there's no need to remember
   to *conditionally* clean up after the worker.
parents 4d4c0285 c2744ed2
...@@ -238,7 +238,6 @@ struct kvm_async_pf { ...@@ -238,7 +238,6 @@ struct kvm_async_pf {
struct list_head link; struct list_head link;
struct list_head queue; struct list_head queue;
struct kvm_vcpu *vcpu; struct kvm_vcpu *vcpu;
struct mm_struct *mm;
gpa_t cr2_or_gpa; gpa_t cr2_or_gpa;
unsigned long addr; unsigned long addr;
struct kvm_arch_async_pf arch; struct kvm_arch_async_pf arch;
......
...@@ -46,8 +46,8 @@ static void async_pf_execute(struct work_struct *work) ...@@ -46,8 +46,8 @@ static void async_pf_execute(struct work_struct *work)
{ {
struct kvm_async_pf *apf = struct kvm_async_pf *apf =
container_of(work, struct kvm_async_pf, work); container_of(work, struct kvm_async_pf, work);
struct mm_struct *mm = apf->mm;
struct kvm_vcpu *vcpu = apf->vcpu; struct kvm_vcpu *vcpu = apf->vcpu;
struct mm_struct *mm = vcpu->kvm->mm;
unsigned long addr = apf->addr; unsigned long addr = apf->addr;
gpa_t cr2_or_gpa = apf->cr2_or_gpa; gpa_t cr2_or_gpa = apf->cr2_or_gpa;
int locked = 1; int locked = 1;
...@@ -56,15 +56,24 @@ static void async_pf_execute(struct work_struct *work) ...@@ -56,15 +56,24 @@ static void async_pf_execute(struct work_struct *work)
might_sleep(); might_sleep();
/* /*
* This work is run asynchronously to the task which owns * Attempt to pin the VM's host address space, and simply skip gup() if
* mm and might be done in another context, so we must * acquiring a pin fail, i.e. if the process is exiting. Note, KVM
* access remotely. * holds a reference to its associated mm_struct until the very end of
* kvm_destroy_vm(), i.e. the struct itself won't be freed before this
* work item is fully processed.
*/ */
mmap_read_lock(mm); if (mmget_not_zero(mm)) {
get_user_pages_remote(mm, addr, 1, FOLL_WRITE, NULL, &locked); mmap_read_lock(mm);
if (locked) get_user_pages_remote(mm, addr, 1, FOLL_WRITE, NULL, &locked);
mmap_read_unlock(mm); if (locked)
mmap_read_unlock(mm);
mmput(mm);
}
/*
* Notify and kick the vCPU even if faulting in the page failed, e.g.
* so that the vCPU can retry the fault synchronously.
*/
if (IS_ENABLED(CONFIG_KVM_ASYNC_PF_SYNC)) if (IS_ENABLED(CONFIG_KVM_ASYNC_PF_SYNC))
kvm_arch_async_page_present(vcpu, apf); kvm_arch_async_page_present(vcpu, apf);
...@@ -74,20 +83,39 @@ static void async_pf_execute(struct work_struct *work) ...@@ -74,20 +83,39 @@ static void async_pf_execute(struct work_struct *work)
apf->vcpu = NULL; apf->vcpu = NULL;
spin_unlock(&vcpu->async_pf.lock); spin_unlock(&vcpu->async_pf.lock);
if (!IS_ENABLED(CONFIG_KVM_ASYNC_PF_SYNC) && first)
kvm_arch_async_page_present_queued(vcpu);
/* /*
* apf may be freed by kvm_check_async_pf_completion() after * The apf struct may be freed by kvm_check_async_pf_completion() as
* this point * soon as the lock is dropped. Nullify it to prevent improper usage.
*/ */
apf = NULL;
if (!IS_ENABLED(CONFIG_KVM_ASYNC_PF_SYNC) && first)
kvm_arch_async_page_present_queued(vcpu);
trace_kvm_async_pf_completed(addr, cr2_or_gpa); trace_kvm_async_pf_completed(addr, cr2_or_gpa);
__kvm_vcpu_wake_up(vcpu); __kvm_vcpu_wake_up(vcpu);
}
mmput(mm); static void kvm_flush_and_free_async_pf_work(struct kvm_async_pf *work)
kvm_put_kvm(vcpu->kvm); {
/*
* The async #PF is "done", but KVM must wait for the work item itself,
* i.e. async_pf_execute(), to run to completion. If KVM is a module,
* KVM must ensure *no* code owned by the KVM (the module) can be run
* after the last call to module_put(). Note, flushing the work item
* is always required when the item is taken off the completion queue.
* E.g. even if the vCPU handles the item in the "normal" path, the VM
* could be terminated before async_pf_execute() completes.
*
* Wake all events skip the queue and go straight done, i.e. don't
* need to be flushed (but sanity check that the work wasn't queued).
*/
if (work->wakeup_all)
WARN_ON_ONCE(work->work.func);
else
flush_work(&work->work);
kmem_cache_free(async_pf_cache, work);
} }
void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu) void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu)
...@@ -112,11 +140,8 @@ void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu) ...@@ -112,11 +140,8 @@ void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu)
#ifdef CONFIG_KVM_ASYNC_PF_SYNC #ifdef CONFIG_KVM_ASYNC_PF_SYNC
flush_work(&work->work); flush_work(&work->work);
#else #else
if (cancel_work_sync(&work->work)) { if (cancel_work_sync(&work->work))
mmput(work->mm);
kvm_put_kvm(vcpu->kvm); /* == work->vcpu->kvm */
kmem_cache_free(async_pf_cache, work); kmem_cache_free(async_pf_cache, work);
}
#endif #endif
spin_lock(&vcpu->async_pf.lock); spin_lock(&vcpu->async_pf.lock);
} }
...@@ -126,7 +151,10 @@ void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu) ...@@ -126,7 +151,10 @@ void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu)
list_first_entry(&vcpu->async_pf.done, list_first_entry(&vcpu->async_pf.done,
typeof(*work), link); typeof(*work), link);
list_del(&work->link); list_del(&work->link);
kmem_cache_free(async_pf_cache, work);
spin_unlock(&vcpu->async_pf.lock);
kvm_flush_and_free_async_pf_work(work);
spin_lock(&vcpu->async_pf.lock);
} }
spin_unlock(&vcpu->async_pf.lock); spin_unlock(&vcpu->async_pf.lock);
...@@ -151,7 +179,7 @@ void kvm_check_async_pf_completion(struct kvm_vcpu *vcpu) ...@@ -151,7 +179,7 @@ void kvm_check_async_pf_completion(struct kvm_vcpu *vcpu)
list_del(&work->queue); list_del(&work->queue);
vcpu->async_pf.queued--; vcpu->async_pf.queued--;
kmem_cache_free(async_pf_cache, work); kvm_flush_and_free_async_pf_work(work);
} }
} }
...@@ -184,9 +212,6 @@ bool kvm_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, ...@@ -184,9 +212,6 @@ bool kvm_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
work->cr2_or_gpa = cr2_or_gpa; work->cr2_or_gpa = cr2_or_gpa;
work->addr = hva; work->addr = hva;
work->arch = *arch; work->arch = *arch;
work->mm = current->mm;
mmget(work->mm);
kvm_get_kvm(work->vcpu->kvm);
INIT_WORK(&work->work, async_pf_execute); INIT_WORK(&work->work, async_pf_execute);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment