Re: [PATCH] mm/mmu_notifier.c: micro-optimization substitute kzalloc with kmalloc

From: Mike Rapoport
Date: Tue Sep 08 2020 - 02:42:57 EST


On Sun, Sep 06, 2020 at 06:06:39PM +0200, Mateusz Nosek wrote:
> Hi,
>
> I performed simple benchmarks using custom kernel module with the code
> fragment in question 'copy-pasted' in there in both versions. In case of 1k,
> 10k and 100k iterations the average time for kzalloc version was 5.1 and for
> kmalloc 3.9, for each iterations number.
> The time was measured using 'ktime_get(void)' function and the results given
> here are in ktime_t units.
> The machine I use has 4 core Intel(R) Core(TM) i5-3470 CPU @ 3.20GHz CPU.
>
> The performance increase happens, but as you wrote it is probably not really
> noticeable.

I don't think that saving a few cylces of memset() in a function that
called only on the initialization path in very particular cases is worth
risking uninitialized variables when somebody will add a new field to
the 'struct mmu_notifier_subscriptions' and will forget to explicitly
set it.

> I have found 3 other places in kernel code with similar kzalloc related
> issues, none of which seems to be 'hot' code.
> I leave the decision if this patch and potential others I would send
> regarding this issue, are worth applying to the community and maintainers.
>
> Best regards,
> Mateusz Nosek
>
> On 9/6/2020 4:26 PM, Mike Rapoport wrote:
> > Hi,
> >
> > On Sun, Sep 06, 2020 at 01:43:21PM +0200, mateusznosek0@xxxxxxxxx wrote:
> > > From: Mateusz Nosek <mateusznosek0@xxxxxxxxx>
> > >
> > > Most fields in struct pointed by 'subscriptions' are initialized explicitly
> > > after the allocation. By changing kzalloc to kmalloc the call to memset
> > > is avoided. As the only new code consists of 2 simple memory accesses,
> > > the performance is increased.
> >
> > Is there a measurable performance increase?
> >
> > The __mmu_notifier_register() is not used that frequently to trade off
> > robustness of kzalloc() for slight (if visible at all) performance gain.
> >
> > > Signed-off-by: Mateusz Nosek <mateusznosek0@xxxxxxxxx>
> > > ---
> > > mm/mmu_notifier.c | 4 +++-
> > > 1 file changed, 3 insertions(+), 1 deletion(-)
> > >
> > > diff --git a/mm/mmu_notifier.c b/mm/mmu_notifier.c
> > > index 4fc918163dd3..190e198dc5be 100644
> > > --- a/mm/mmu_notifier.c +++ b/mm/mmu_notifier.c @@ -625,7 +625,7
> > > @@ int __mmu_notifier_register(struct mmu_notifier *subscription,
> > > * know that mm->notifier_subscriptions can't change while we *
> > > hold the write side of the mmap_lock. */
> > > - subscriptions = kzalloc(
> > > + subscriptions = kmalloc(
> > > sizeof(struct mmu_notifier_subscriptions), GFP_KERNEL);
> > > if (!subscriptions)
> > > return -ENOMEM;
> > > @@ -636,6 +636,8 @@ int __mmu_notifier_register(struct mmu_notifier *subscription,
> > > subscriptions->itree = RB_ROOT_CACHED;
> > > init_waitqueue_head(&subscriptions->wq);
> > > INIT_HLIST_HEAD(&subscriptions->deferred_list);
> > > + subscriptions->active_invalidate_ranges = 0;
> > > + subscriptions->has_itree = false;
> > > }
> > > ret = mm_take_all_locks(mm);
> > > --
> > > 2.20.1
> > >
> > >
> >

--
Sincerely yours,
Mike.