Re: [PATCH 00/12] SEV Live Migration Patchset.
From: Andy Lutomirski
Date: Thu Feb 13 2020 - 00:43:57 EST
On Wed, Feb 12, 2020 at 5:14 PM Ashish Kalra <Ashish.Kalra@xxxxxxx> wrote:
>
> From: Ashish Kalra <ashish.kalra@xxxxxxx>
>
> This patchset adds support for SEV Live Migration on KVM/QEMU.
I skimmed this all and I don't see any description of how this all works.
Does any of this address the mess in svm_register_enc_region()? Right
now, when QEMU (or a QEMU alternative) wants to allocate some memory
to be used for guest encrypted pages, it mmap()s some memory and the
kernel does get_user_pages_fast() on it. The pages are kept pinned
for the lifetime of the mapping. This is not at all okay. Let's see:
- The memory is pinned and it doesn't play well with the Linux memory
management code. You just wrote a big patch set to migrate the pages
to a whole different machines, but we apparently can't even migrate
them to a different NUMA node or even just a different address. And
good luck swapping it out.
- The memory is still mapped in the QEMU process, and that mapping is
incoherent with actual guest access to the memory. It's nice that KVM
clflushes it so that, in principle, everything might actually work,
but this is gross. We should not be exposing incoherent mappings to
userspace.
Perhaps all this fancy infrastructure you're writing for migration and
all this new API surface could also teach the kernel how to migrate
pages from a guest *to the same guest* so we don't need to pin pages
forever. And perhaps you could put some thought into how to improve
the API so that it doesn't involve nonsensical incoherent mappings.
(To be blunt: if I had noticed how the SEV code worked before it was
merged, I would have NAKed it. It's too late now to retroactively
remove it from the kernel, but perhaps we could try not to pile more
complexity on top of the unfortunate foundation we have.)