Re: Re: [PATCH 40/70] x86/sev-es: Setup per-cpu GHCBs for the runtime handler
From: Bo Gan
Date: Wed Apr 22 2020 - 21:47:49 EST
On 4/15/20 8:53 AM, Joerg Roedel wrote:
Hi Mike,
On Tue, Apr 14, 2020 at 07:03:44PM +0000, Mike Stunes wrote:
set_memory_decrypted needs to check the return value. I see it
consistently return ENOMEM. I've traced that back to split_large_page
in arch/x86/mm/pat/set_memory.c.
I agree that the return code needs to be checked. But I wonder why this
happens. The split_large_page() function returns -ENOMEM when
alloc_pages() fails. Do you boot the guest with minal RAM assigned?
Regards,
Joerg
I just want to add some context around this. The call path that lead to
the failure is like the following:
__alloc_pages_slowpath
__alloc_pages_nodemask
alloc_pages_current
alloc_pages
split_large_page
__change_page_attr
__change_page_attr_set_clr
__set_memory_enc_dec
set_memory_decrypted
sev_es_init_ghcbs
trap_init -> before mm_init (in init/main.c)
start_kernel
x86_64_start_reservations
x86_64_start_kernel
secondary_startup_64
At this time, mem_init hasn't been called yet (which would be called by
mm_init). Thus, the free pages are still owned by memblock. It's in
mem_init (x86/mm/init_64.c) that memblock_free_all gets called and free
pages are released.
During testing, I've also noticed that debug_pagealloc=1 will make the
issue disappear. That's because with debug_pagealloc=1,
probe_page_size_mask in x86/mm/init.c will not allow large pages
(2M/1G). Therefore, no split_large_page would happen. Similarly, if CPU
doesn't have X86_FEATURE_PSE, there won't be large pages either.
Any thoughts? Maybe split_large_page should get pages from memblock at
early boot?
Bo