Re: [PATCH] KVM: x86/mmu: fix potential races when walking host page table
From: Paolo Bonzini
Date: Fri Apr 29 2022 - 10:45:17 EST
On 4/29/22 16:35, Sean Christopherson wrote:
On Fri, Apr 29, 2022, Paolo Bonzini wrote:
+out:
+ local_irq_restore(flags);
+ return level;
+}
+EXPORT_SYMBOL_GPL(kvm_lookup_address_level_in_mm);
Exporting is not needed.
Thanks for writing the walk code though. I'll adapt it and integrate the
patch.
But why are we fixing this only in KVM? I liked the idea of stealing perf's
implementation because it was a seemlingly perfect fit and wouldn't introduce
new code (ignoring wrappers, etc...).
We _know_ that at least one subsystem is misusing lookup_address_in_pgd() and
given that its wrappers are exported, I highly doubt KVM is the only offender.
It really feels like we're passing the buck here by burying the fix in KVM.
There are two ways to do it:
* having a generic function in mm/. The main issue there is the lack of
a PG_LEVEL_{P4D,PUD,PMD,PTE} enum at the mm/ level. We could use
(ctz(x) - 12) / 9 to go from size to level, but it's ugly and there
could be architectures with heterogeneous page table sizes.
* having a generic function in arch/x86/. In this case KVM seems to be
the odd one that doesn't need the PTE. For example vc_slow_virt_to_phys
needs the PTE, and needs the size rather than the "level" per se.
So for now I punted, while keeping open the door for moving code from
arch/x86/kvm/ to mm/ if anyone else (even other KVM ports) need the same
logic.
Paolo