Re: [PATCH 14/26] drm/xe/eudebug: implement userptr_vma access
From: Simona Vetter
Date: Tue Dec 10 2024 - 06:19:40 EST
On Mon, Dec 09, 2024 at 04:42:32PM +0100, Christian König wrote:
> Am 09.12.24 um 16:31 schrieb Simona Vetter:
> > On Mon, Dec 09, 2024 at 03:03:04PM +0100, Christian König wrote:
> > > Am 09.12.24 um 14:33 schrieb Mika Kuoppala:
> > > > From: Andrzej Hajda <andrzej.hajda@xxxxxxxxx>
> > > >
> > > > Debugger needs to read/write program's vmas including userptr_vma.
> > > > Since hmm_range_fault is used to pin userptr vmas, it is possible
> > > > to map those vmas from debugger context.
> > > Oh, this implementation is extremely questionable as well. Adding the LKML
> > > and the MM list as well.
> > >
> > > First of all hmm_range_fault() does *not* pin anything!
> > >
> > > In other words you don't have a page reference when the function returns,
> > > but rather just a sequence number you can check for modifications.
> > I think it's all there, holds the invalidation lock during the critical
> > access/section, drops it when reacquiring pages, retries until it works.
> >
> > I think the issue is more that everyone hand-rolls userptr.
>
> Well that is part of the issue.
Yeah I ignored the other part, because that didn't seem super interesting
really.
Like for compute you can make the architectural assumption that gpu
addresses match cpu addresses, and this all becomes easy. Or at least
easier, there's still the issue of having to call into the driver for gpu
side flushes.
But for 3d userptr that's not the case, and you get two options:
- Have some tracking structure that umd and debugger agree on, so stable
abi fun and all that, so you can find the mapping. And I think in some
cases this would need to be added first.
- Just ask the kernel, which already knows.
Like for cpu mmaps we also don't inject tracking code into mmap/munmap, we
just ask the kernel through /proc/pid/maps. This sounds like the same
question, probably should have a similar answer.
I guess you can do a bit a bikeshed about whether the kernel should only
do the address translation and then you do a ptrace poke/peek for the
actual access. But again you need some flushes, so this might be a bit
silly.
But fundamentally this makes sense to me to ask the entity that actually
knows how userptr areas map to memory.
-Sima
>
> The general problem here is that the eudebug interface tries to simulate the
> memory accesses as they would have happened by the hardware.
>
> What the debugger should probably do is to cleanly attach to the
> application, get the information which CPU address is mapped to which GPU
> address and then use the standard ptrace interfaces.
>
> The whole interface re-invents a lot of functionality which is already there
> just because you don't like the idea to attach to the debugged application
> in userspace.
>
> As far as I can see this whole idea is extremely questionable. This looks
> like re-inventing the wheel in a different color.
>
> Regards,
> Christian.
>
> > Probably time
> > we standardize that and put it into gpuvm as an optional part, with
> > consistent locking, naming (like not calling it _pin_pages when it's
> > unpinnged userptr), kerneldoc and all the nice things so that we
> > stop consistently getting confused by other driver's userptr code.
> >
> > I think that was on the plan originally as an eventual step, I guess time
> > to pump that up. Matt/Thomas, thoughts?
> > -Sima
> >
> > > > v2: pin pages vs notifier, move to vm.c (Matthew)
> > > > v3: - iterate over system pages instead of DMA, fixes iommu enabled
> > > > - s/xe_uvma_access/xe_vm_uvma_access/ (Matt)
> > > >
> > > > Signed-off-by: Andrzej Hajda <andrzej.hajda@xxxxxxxxx>
> > > > Signed-off-by: Maciej Patelczyk <maciej.patelczyk@xxxxxxxxx>
> > > > Signed-off-by: Mika Kuoppala <mika.kuoppala@xxxxxxxxxxxxxxx>
> > > > Reviewed-by: Jonathan Cavitt <jonathan.cavitt@xxxxxxxxx> #v1
> > > > ---
> > > > drivers/gpu/drm/xe/xe_eudebug.c | 3 ++-
> > > > drivers/gpu/drm/xe/xe_vm.c | 47 +++++++++++++++++++++++++++++++++
> > > > drivers/gpu/drm/xe/xe_vm.h | 3 +++
> > > > 3 files changed, 52 insertions(+), 1 deletion(-)
> > > >
> > > > diff --git a/drivers/gpu/drm/xe/xe_eudebug.c b/drivers/gpu/drm/xe/xe_eudebug.c
> > > > index 9d87df75348b..e5949e4dcad8 100644
> > > > --- a/drivers/gpu/drm/xe/xe_eudebug.c
> > > > +++ b/drivers/gpu/drm/xe/xe_eudebug.c
> > > > @@ -3076,7 +3076,8 @@ static int xe_eudebug_vma_access(struct xe_vma *vma, u64 offset_in_vma,
> > > > return ret;
> > > > }
> > > > - return -EINVAL;
> > > > + return xe_vm_userptr_access(to_userptr_vma(vma), offset_in_vma,
> > > > + buf, bytes, write);
> > > > }
> > > > static int xe_eudebug_vm_access(struct xe_vm *vm, u64 offset,
> > > > diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> > > > index 0f17bc8b627b..224ff9e16941 100644
> > > > --- a/drivers/gpu/drm/xe/xe_vm.c
> > > > +++ b/drivers/gpu/drm/xe/xe_vm.c
> > > > @@ -3414,3 +3414,50 @@ void xe_vm_snapshot_free(struct xe_vm_snapshot *snap)
> > > > }
> > > > kvfree(snap);
> > > > }
> > > > +
> > > > +int xe_vm_userptr_access(struct xe_userptr_vma *uvma, u64 offset,
> > > > + void *buf, u64 len, bool write)
> > > > +{
> > > > + struct xe_vm *vm = xe_vma_vm(&uvma->vma);
> > > > + struct xe_userptr *up = &uvma->userptr;
> > > > + struct xe_res_cursor cur = {};
> > > > + int cur_len, ret = 0;
> > > > +
> > > > + while (true) {
> > > > + down_read(&vm->userptr.notifier_lock);
> > > > + if (!xe_vma_userptr_check_repin(uvma))
> > > > + break;
> > > > +
> > > > + spin_lock(&vm->userptr.invalidated_lock);
> > > > + list_del_init(&uvma->userptr.invalidate_link);
> > > > + spin_unlock(&vm->userptr.invalidated_lock);
> > > > +
> > > > + up_read(&vm->userptr.notifier_lock);
> > > > + ret = xe_vma_userptr_pin_pages(uvma);
> > > > + if (ret)
> > > > + return ret;
> > > > + }
> > > > +
> > > > + if (!up->sg) {
> > > > + ret = -EINVAL;
> > > > + goto out_unlock_notifier;
> > > > + }
> > > > +
> > > > + for (xe_res_first_sg_system(up->sg, offset, len, &cur); cur.remaining;
> > > > + xe_res_next(&cur, cur_len)) {
> > > > + void *ptr = kmap_local_page(sg_page(cur.sgl)) + cur.start;
> > > The interface basically creates a side channel to access userptrs in the way
> > > an userspace application would do without actually going through userspace.
> > >
> > > That is generally not something a device driver should ever do as far as I
> > > can see.
> > >
> > > > +
> > > > + cur_len = min(cur.size, cur.remaining);
> > > > + if (write)
> > > > + memcpy(ptr, buf, cur_len);
> > > > + else
> > > > + memcpy(buf, ptr, cur_len);
> > > > + kunmap_local(ptr);
> > > > + buf += cur_len;
> > > > + }
> > > > + ret = len;
> > > > +
> > > > +out_unlock_notifier:
> > > > + up_read(&vm->userptr.notifier_lock);
> > > I just strongly hope that this will prevent the mapping from changing.
> > >
> > > Regards,
> > > Christian.
> > >
> > > > + return ret;
> > > > +}
> > > > diff --git a/drivers/gpu/drm/xe/xe_vm.h b/drivers/gpu/drm/xe/xe_vm.h
> > > > index 23adb7442881..372ad40ad67f 100644
> > > > --- a/drivers/gpu/drm/xe/xe_vm.h
> > > > +++ b/drivers/gpu/drm/xe/xe_vm.h
> > > > @@ -280,3 +280,6 @@ struct xe_vm_snapshot *xe_vm_snapshot_capture(struct xe_vm *vm);
> > > > void xe_vm_snapshot_capture_delayed(struct xe_vm_snapshot *snap);
> > > > void xe_vm_snapshot_print(struct xe_vm_snapshot *snap, struct drm_printer *p);
> > > > void xe_vm_snapshot_free(struct xe_vm_snapshot *snap);
> > > > +
> > > > +int xe_vm_userptr_access(struct xe_userptr_vma *uvma, u64 offset,
> > > > + void *buf, u64 len, bool write);
>
--
Simona Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch