On Thu, 3 Apr 2025 09:20:00 +0200
Thomas Zimmermann <tzimmermann@xxxxxxx> wrote:
HiOkay, so it looks more like a naming issue then. The intent here is to
Am 02.04.25 um 15:21 schrieb Boris Brezillon:
On Wed, 2 Apr 2025 15:58:55 +0300I once tried to add pin as part of vmap, so that pages stay in place.
Dmitry Osipenko <dmitry.osipenko@xxxxxxxxxxxxx> wrote:
On 4/2/25 15:47, Thomas Zimmermann wrote:Hm, is this problematic though? If you want to vmap() inside a section
Hi
Am 22.03.25 um 22:26 schrieb Dmitry Osipenko:
The vmapped pages shall be pinned in memory and previously get/I've meanwhile rediscovered this patch and I'm sure this is not correct.
put_pages()
were implicitly hard-pinning/unpinning the pages. This will no longer be
the case with addition of memory shrinker because pages_use_count > 0
won't
determine anymore whether pages are hard-pinned (they will be soft-
pinned),
while the new pages_pin_count will do the hard-pinning. Switch the
vmap/vunmap() to use pin/unpin() functions in a preparation of addition
of the memory shrinker support to drm-shmem.
Vmap should not pin AFAIK. It is possible to vmap if the buffer has been
pinned, but that's not automatic. For other vmaps it is necessary to
hold the reservation lock to prevent the buffer from moving.
that's protected by the resv lock, you can
- drm_gem_shmem_vmap_locked()
- do whatever you need to do with the vaddr,
- drm_gem_shmem_vunmap_locked()
and the {pin,page_use}_count will be back to their original values.
Those are just ref counters, and I doubt the overhead of
incrementing/decrementing them makes a difference compared to the heavy
page-allocation/vmap operations...
Christian was very clear about not doing this. I found this made a lot
of sense: vmap means "make the memory available to the CPU". The memory
location doesn't matter much here. Pin means something like "make the
memory available to the GPU". But which GPU depends on the caller: calls
via GEM refer to the local GPU, calls via dma-buf usually refer to the
importer's GPU. That GPU uncertainty makes pin problematic already.
make sure the page array doesn't disappear while we have a kernel
mapping active (address returned by vmap()). The reason we went from
pages_count to pages_use_count+pin_count is because we have two kind of
references in drm_gem_shmem:
- weak references (tracked with pages_use_count). Those are
usually held by GPU VMs, and they are weak in the sense they
shouldn't prevent the shrinker to reclaim them if the GPU VM is idle.
The other user of weak references is userspace mappings of GEM
objects (mmap()), because then we can repopulate those with our fault
handler.
- hard references (tracked with pin_count) which are used to prevent
the shrinker from even considering the GEM as reclaimable. And clearly
kernel mappings fall in that case, because otherwise we could reclaim
pages that might be dereferenced by the CPU later on. It's also used
to implement drm_gem_pin because it's the same mechanism really,
hence the name
In your case, vmap an pin both intent to hold the shmem pages in memory.But that's not what we do, is it? Sure, in drm_gem_shmem_vmap_locked(),
They might be build on top of the same implementation, but one should
not be implemented with the other because of their different meanings.
we call drm_gem_shmem_pin_locked(), but that's an internal function to
make sure the pages are allocated and stay around until
drm_gem_shmem_vunmap_locked() is called.
I guess we could rename pin_count into hard_refcount or
page_residency_count or xxx_count, and change the pin/unpin_locked()
function names accordingly, but that's just a naming detail, it doesn't
force you to call drm_gem_pin() to vmap() your GEM, it's something we
do internally.
More generally speaking, I've meanwhile come to the conclusion that pinWell, yes and no. There are situations where you simply can't move
should not even exist in the GEM interface. It's an internal operation
of TTM and reveals too much about what happens within the
implementation. Instead GEM should be free to move buffers around.
things around if there are active users, and vmap() is one of those
AFAICT.
Dma-buf importers should only tell exporters to make buffers availableAnd that's what they do, no? attach() tells the exporter to give the
to them, but not how to do this. AFAIK that's what dma-buf's
attach/detach is for.
importer a way to access those buffers, and given the exporter has no
clue about when/how the exporter will access those, there's no other way
but to pin the pages. Am I missing something here?