Re: [PATCH 1/8] mm: cma: introduce cma_release_nowait()

From: David Hildenbrand
Date: Thu Mar 25 2021 - 05:58:10 EST


On 25.03.21 01:28, Mike Kravetz wrote:
From: Roman Gushchin <guro@xxxxxx>

cma_release() has to lock the cma_lock mutex to clear the cma bitmap.
It makes it a blocking function, which complicates its usage from
non-blocking contexts. For instance, hugetlbfs code is temporarily
dropping the hugetlb_lock spinlock to call cma_release().

This patch introduces a non-blocking cma_release_nowait(), which
postpones the cma bitmap clearance. It's done later from a work
context. The first page in the cma allocation is used to store
the work struct. Because CMA allocations and de-allocations are
usually not that frequent, a single global workqueue is used.

To make sure that subsequent cma_alloc() call will pass, cma_alloc()
flushes the cma_release_wq workqueue. To avoid a performance
regression in the case when only cma_release() is used, gate it
by a per-cma area flag, which is set by the first call
of cma_release_nowait().

Signed-off-by: Roman Gushchin <guro@xxxxxx>
[mike.kravetz@xxxxxxxxxx: rebased to v5.12-rc3-mmotm-2021-03-17-22-24]
Signed-off-by: Mike Kravetz <mike.kravetz@xxxxxxxxxx>
---


1. Is there a real reason this is a mutex and not a spin lock? It seems to only protect the bitmap. Are bitmaps that huge that we spend a significant amount of time in there?

Because I also read "Because CMA allocations and de-allocations are
usually not that frequent".

With a spinlock, you would no longer be sleeping, but obviously you might end up waiting for the lock ;) Not sure if that would help.

2. IIUC, if we would do the clearing completely lockless and use atomic bitmap ops instead, only cma_debug_show_areas() would see slight inconsistencies. As long as the setting code (-> allocation code) holds the lock, I think this should be fine (-> no double allocations).

(sorry if that has already been discussed)

--
Thanks,

David / dhildenb