Commit 0547758a authored by Sean Christopherson's avatar Sean Christopherson Committed by Paolo Bonzini

x86/kvm: Alloc dummy async #PF token outside of raw spinlock

Drop the raw spinlock in kvm_async_pf_task_wake() before allocating the
the dummy async #PF token, the allocator is preemptible on PREEMPT_RT
kernels and must not be called from truly atomic contexts.

Opportunistically document why it's ok to loop on allocation failure,
i.e. why the function won't get stuck in an infinite loop.
Reported-by: default avatarYajun Deng <yajun.deng@linux.dev>
Cc: stable@vger.kernel.org
Signed-off-by: default avatarSean Christopherson <seanjc@google.com>
Signed-off-by: default avatarPaolo Bonzini <pbonzini@redhat.com>
parent fee060cd
...@@ -191,7 +191,7 @@ void kvm_async_pf_task_wake(u32 token) ...@@ -191,7 +191,7 @@ void kvm_async_pf_task_wake(u32 token)
{ {
u32 key = hash_32(token, KVM_TASK_SLEEP_HASHBITS); u32 key = hash_32(token, KVM_TASK_SLEEP_HASHBITS);
struct kvm_task_sleep_head *b = &async_pf_sleepers[key]; struct kvm_task_sleep_head *b = &async_pf_sleepers[key];
struct kvm_task_sleep_node *n; struct kvm_task_sleep_node *n, *dummy = NULL;
if (token == ~0) { if (token == ~0) {
apf_task_wake_all(); apf_task_wake_all();
...@@ -203,28 +203,41 @@ void kvm_async_pf_task_wake(u32 token) ...@@ -203,28 +203,41 @@ void kvm_async_pf_task_wake(u32 token)
n = _find_apf_task(b, token); n = _find_apf_task(b, token);
if (!n) { if (!n) {
/* /*
* async PF was not yet handled. * Async #PF not yet handled, add a dummy entry for the token.
* Add dummy entry for the token. * Allocating the token must be down outside of the raw lock
* as the allocator is preemptible on PREEMPT_RT kernels.
*/ */
n = kzalloc(sizeof(*n), GFP_ATOMIC); if (!dummy) {
if (!n) { raw_spin_unlock(&b->lock);
dummy = kzalloc(sizeof(*dummy), GFP_KERNEL);
/* /*
* Allocation failed! Busy wait while other cpu * Continue looping on allocation failure, eventually
* handles async PF. * the async #PF will be handled and allocating a new
* node will be unnecessary.
*/
if (!dummy)
cpu_relax();
/*
* Recheck for async #PF completion before enqueueing
* the dummy token to avoid duplicate list entries.
*/ */
raw_spin_unlock(&b->lock);
cpu_relax();
goto again; goto again;
} }
n->token = token; dummy->token = token;
n->cpu = smp_processor_id(); dummy->cpu = smp_processor_id();
init_swait_queue_head(&n->wq); init_swait_queue_head(&dummy->wq);
hlist_add_head(&n->link, &b->list); hlist_add_head(&dummy->link, &b->list);
dummy = NULL;
} else { } else {
apf_task_wake_one(n); apf_task_wake_one(n);
} }
raw_spin_unlock(&b->lock); raw_spin_unlock(&b->lock);
return;
/* A dummy token might be allocated and ultimately not used. */
if (dummy)
kfree(dummy);
} }
EXPORT_SYMBOL_GPL(kvm_async_pf_task_wake); EXPORT_SYMBOL_GPL(kvm_async_pf_task_wake);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment