Re: [PATCHv5] x86/kdump: bugfix, make the behavior of crashkernel=X consistent with kaslr
From: Mike Rapoport
Date: Tue Jan 08 2019 - 10:49:10 EST
On Tue, Jan 08, 2019 at 05:01:38PM +0800, Baoquan He wrote:
> Hi Mike,
>
> On 01/08/19 at 10:05am, Mike Rapoport wrote:
> > I'm not thrilled by duplicating this code (yet again).
> > I liked the v3 of this patch [1] more, assuming we allow bottom-up mode to
> > allocate [0, kernel_start) unconditionally.
> > I'd just replace you first patch in v3 [2] with something like:
>
> In initmem_init(), we will restore the top-down allocation style anyway.
> While reserve_crashkernel() is called after initmem_init(), it's not
> appropriate to adjust memblock_find_in_range_node(), and we really want
> to find region bottom up for crashkernel reservation, no matter where
> kernel is loaded, better call __memblock_find_range_bottom_up().
>
> Create a wrapper to do the necessary handling, then call
> __memblock_find_range_bottom_up() directly, looks better.
What bothers me is 'the necessary handling' which is already done in
several places in memblock in a similar, but yet slightly different way.
memblock_find_in_range() and memblock_phys_alloc_nid() retry with different
MEMBLOCK_MIRROR, but memblock_phys_alloc_try_nid() does that only when
allocating from the specified node and does not retry when it falls back to
any node. And memblock_alloc_internal() has yet another set of fallbacks.
So what should be the necessary handling in the wrapper for
__memblock_find_range_bottom_up() ?
BTW, even without any memblock modifications, retrying allocation in
reserve_crashkerenel() for different ranges, like the proposal at [1] would
also work, wouldn't it?
[1] http://lists.infradead.org/pipermail/kexec/2017-October/019571.html
> Thanks
> Baoquan
>
> >
> > diff --git a/mm/memblock.c b/mm/memblock.c
> > index 7df468c..d1b30b9 100644
> > --- a/mm/memblock.c
> > +++ b/mm/memblock.c
> > @@ -274,24 +274,14 @@ phys_addr_t __init_memblock memblock_find_in_range_node(phys_addr_t size,
> > * try bottom-up allocation only when bottom-up mode
> > * is set and @end is above the kernel image.
> > */
> > - if (memblock_bottom_up() && end > kernel_end) {
> > - phys_addr_t bottom_up_start;
> > -
> > - /* make sure we will allocate above the kernel */
> > - bottom_up_start = max(start, kernel_end);
> > -
> > + if (memblock_bottom_up()) {
> > /* ok, try bottom-up allocation first */
> > - ret = __memblock_find_range_bottom_up(bottom_up_start, end,
> > + ret = __memblock_find_range_bottom_up(start, end,
> > size, align, nid, flags);
> > if (ret)
> > return ret;
> >
> > /*
> > - * we always limit bottom-up allocation above the kernel,
> > - * but top-down allocation doesn't have the limit, so
> > - * retrying top-down allocation may succeed when bottom-up
> > - * allocation failed.
> > - *
> > * bottom-up allocation is expected to be fail very rarely,
> > * so we use WARN_ONCE() here to see the stack trace if
> > * fail happens.
> >
> > [1] https://lore.kernel.org/lkml/1545966002-3075-3-git-send-email-kernelfans@xxxxxxxxx/
> > [2] https://lore.kernel.org/lkml/1545966002-3075-2-git-send-email-kernelfans@xxxxxxxxx/
> >
> > > +
> > > + return ret;
> > > +}
> > > +
> > > /**
> > > * __memblock_find_range_top_down - find free area utility, in top-down
> > > * @start: start of candidate range
> > > --
> > > 2.7.4
> > >
> >
> > --
> > Sincerely yours,
> > Mike.
> >
>
--
Sincerely yours,
Mike.