[PATCH v5 10/11] drm/panthor: Pass vm_bind_op to vm_prepare_map_op_ctx

From: Adrián Larumbe

Date: Fri Mar 13 2026 - 11:22:05 EST


Instead of passing its constituent elements, pass the whole struct to
simplify the function prototype.

Signed-off-by: Adrián Larumbe <adrian.larumbe@xxxxxxxxxxxxx>
---
drivers/gpu/drm/panthor/panthor_mmu.c | 57 ++++++++++++++-------------
1 file changed, 30 insertions(+), 27 deletions(-)

diff --git a/drivers/gpu/drm/panthor/panthor_mmu.c b/drivers/gpu/drm/panthor/panthor_mmu.c
index ba322e2029b9..a62ac715265b 100644
--- a/drivers/gpu/drm/panthor/panthor_mmu.c
+++ b/drivers/gpu/drm/panthor/panthor_mmu.c
@@ -1216,10 +1216,7 @@ panthor_vm_op_ctx_prealloc_vmas(struct panthor_vm_op_ctx *op_ctx)
static int panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx *op_ctx,
struct panthor_vm *vm,
struct panthor_gem_object *bo,
- u64 offset,
- u64 repeat_range,
- u64 size, u64 va,
- u32 flags)
+ const struct drm_panthor_vm_bind_op *op)
{
struct drm_gpuvm_bo *preallocated_vm_bo;
struct sg_table *sgt = NULL;
@@ -1229,30 +1226,32 @@ static int panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx *op_ctx,
if (!bo)
return -EINVAL;

- if ((flags & ~PANTHOR_VM_BIND_OP_MAP_FLAGS) ||
- (flags & DRM_PANTHOR_VM_BIND_OP_TYPE_MASK) != DRM_PANTHOR_VM_BIND_OP_TYPE_MAP)
+ if ((op->flags & ~PANTHOR_VM_BIND_OP_MAP_FLAGS) ||
+ (op->flags & DRM_PANTHOR_VM_BIND_OP_TYPE_MASK) != DRM_PANTHOR_VM_BIND_OP_TYPE_MAP)
return -EINVAL;

- if (!(flags & DRM_PANTHOR_VM_BIND_OP_MAP_REPEAT)) {
+ if (!(op->flags & DRM_PANTHOR_VM_BIND_OP_MAP_REPEAT)) {
/* Make sure the VA and size are in-bounds. */
- if (size > bo->base.base.size || offset > bo->base.base.size - size)
+ if (op->size > bo->base.base.size || op->bo_offset > bo->base.base.size - op->size)
return -EINVAL;
} else {
/* Current drm api uses 32-bit for repeat range, */
- if (repeat_range > U32_MAX)
+ if (op->bo_repeat_range > U32_MAX)
return -EINVAL;

/* Make sure the repeat_range is in-bounds. */
- if (repeat_range > bo->base.base.size || offset > bo->base.base.size - repeat_range)
+ if (op->bo_repeat_range > bo->base.base.size ||
+ op->bo_offset > bo->base.base.size - op->bo_repeat_range)
return -EINVAL;

/* Repeat range must a multiple of the minimum GPU page size */
- if (repeat_range & ((1u << (ffs(vm->ptdev->mmu_info.page_size_bitmap) - 1)) - 1))
+ if (op->bo_repeat_range &
+ ((1u << (ffs(vm->ptdev->mmu_info.page_size_bitmap) - 1)) - 1))
return -EINVAL;

- u64 repeat_count = size;
+ u64 repeat_count = op->size;

- if (do_div(repeat_count, repeat_range))
+ if (do_div(repeat_count, op->bo_repeat_range))
return -EINVAL;
}

@@ -1262,9 +1261,9 @@ static int panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx *op_ctx,
return -EINVAL;

memset(op_ctx, 0, sizeof(*op_ctx));
- op_ctx->flags = flags;
- op_ctx->va.range = size;
- op_ctx->va.addr = va;
+ op_ctx->flags = op->flags;
+ op_ctx->va.range = op->size;
+ op_ctx->va.addr = op->va;

ret = panthor_vm_op_ctx_prealloc_vmas(op_ctx);
if (ret)
@@ -1303,17 +1302,17 @@ static int panthor_vm_prepare_map_op_ctx(struct panthor_vm_op_ctx *op_ctx,

op_ctx->map.vm_bo = drm_gpuvm_bo_obtain_prealloc(preallocated_vm_bo);

- op_ctx->map.bo_offset = offset;
- op_ctx->map.bo_repeat_range = repeat_range;
+ op_ctx->map.bo_offset = op->bo_offset;
+ op_ctx->map.bo_repeat_range = op->bo_repeat_range;

/* L1, L2 and L3 page tables.
* We could optimize L3 allocation by iterating over the sgt and merging
* 2M contiguous blocks, but it's simpler to over-provision and return
* the pages if they're not used.
*/
- pt_count = ((ALIGN(va + size, 1ull << 39) - ALIGN_DOWN(va, 1ull << 39)) >> 39) +
- ((ALIGN(va + size, 1ull << 30) - ALIGN_DOWN(va, 1ull << 30)) >> 30) +
- ((ALIGN(va + size, 1ull << 21) - ALIGN_DOWN(va, 1ull << 21)) >> 21);
+ pt_count = ((ALIGN(op->va + op->size, 1ull << 39) - ALIGN_DOWN(op->va, 1ull << 39)) >> 39) +
+ ((ALIGN(op->va + op->size, 1ull << 30) - ALIGN_DOWN(op->va, 1ull << 30)) >> 30) +
+ ((ALIGN(op->va + op->size, 1ull << 21) - ALIGN_DOWN(op->va, 1ull << 21)) >> 21);

op_ctx->rsvd_page_tables.pages = kcalloc(pt_count,
sizeof(*op_ctx->rsvd_page_tables.pages),
@@ -2645,11 +2644,7 @@ panthor_vm_bind_prepare_op_ctx(struct drm_file *file,
gem = drm_gem_object_lookup(file, op->bo_handle);
ret = panthor_vm_prepare_map_op_ctx(op_ctx, vm,
gem ? to_panthor_bo(gem) : NULL,
- op->bo_offset,
- op->bo_repeat_range,
- op->size,
- op->va,
- op->flags);
+ op);
drm_gem_object_put(gem);
return ret;

@@ -2845,13 +2840,21 @@ int panthor_vm_bind_exec_sync_op(struct drm_file *file,
int panthor_vm_map_bo_range(struct panthor_vm *vm, struct panthor_gem_object *bo,
u64 offset, u64 size, u64 va, u32 flags)
{
+ struct drm_panthor_vm_bind_op op = {0};
struct panthor_vm_op_ctx op_ctx;
int ret;

if (drm_WARN_ON(&vm->ptdev->base, flags & DRM_PANTHOR_VM_BIND_OP_MAP_REPEAT))
return -EINVAL;

- ret = panthor_vm_prepare_map_op_ctx(&op_ctx, vm, bo, offset, 0, size, va, flags);
+ op = (struct drm_panthor_vm_bind_op){
+ .bo_offset = offset,
+ .size = size,
+ .va = va,
+ .flags = flags,
+ };
+
+ ret = panthor_vm_prepare_map_op_ctx(&op_ctx, vm, bo, &op);
if (ret)
return ret;

--
2.53.0