drm/xe: Add xe_vm_pgtable_update_op to xe_vma_ops
authorMatthew Brost <matthew.brost@intel.com>
Thu, 4 Jul 2024 04:16:47 +0000 (21:16 -0700)
committerMatthew Brost <matthew.brost@intel.com>
Thu, 4 Jul 2024 05:27:00 +0000 (22:27 -0700)
Each xe_vma_op resolves to 0-3 pt_ops. Add storage for the pt_ops to
xe_vma_ops which is dynamically allocated based the number and types of
xe_vma_op in the xe_vma_ops list. Allocation only implemented in this
patch.

This will help with converting xe_vma_ops (multiple xe_vma_op) in a
atomic update unit.

Signed-off-by: Matthew Brost <matthew.brost@intel.com>
Reviewed-by: Jonathan Cavitt <jonathan.cavitt@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20240704041652.272920-3-matthew.brost@intel.com
drivers/gpu/drm/xe/xe_pt_types.h
drivers/gpu/drm/xe/xe_vm.c
drivers/gpu/drm/xe/xe_vm_types.h

index cee70cb..2093150 100644 (file)
@@ -74,4 +74,16 @@ struct xe_vm_pgtable_update {
        u32 flags;
 };
 
+/** struct xe_vm_pgtable_update_op - Page table update operation */
+struct xe_vm_pgtable_update_op {
+       /** @entries: entries to update for this operation */
+       struct xe_vm_pgtable_update entries[XE_VM_MAX_LEVEL * 2 + 1];
+       /** @num_entries: number of entries for this update operation */
+       u32 num_entries;
+       /** @bind: is a bind */
+       bool bind;
+       /** @rebind: is a rebind */
+       bool rebind;
+};
+
 #endif
index 0c76464..6677874 100644 (file)
@@ -718,6 +718,42 @@ int xe_vm_userptr_check_repin(struct xe_vm *vm)
                list_empty_careful(&vm->userptr.invalidated)) ? 0 : -EAGAIN;
 }
 
+static int xe_vma_ops_alloc(struct xe_vma_ops *vops)
+{
+       int i;
+
+       for (i = 0; i < XE_MAX_TILES_PER_DEVICE; ++i) {
+               if (!vops->pt_update_ops[i].num_ops)
+                       continue;
+
+               vops->pt_update_ops[i].ops =
+                       kmalloc_array(vops->pt_update_ops[i].num_ops,
+                                     sizeof(*vops->pt_update_ops[i].ops),
+                                     GFP_KERNEL);
+               if (!vops->pt_update_ops[i].ops)
+                       return -ENOMEM;
+       }
+
+       return 0;
+}
+
+static void xe_vma_ops_fini(struct xe_vma_ops *vops)
+{
+       int i;
+
+       for (i = 0; i < XE_MAX_TILES_PER_DEVICE; ++i)
+               kfree(vops->pt_update_ops[i].ops);
+}
+
+static void xe_vma_ops_incr_pt_update_ops(struct xe_vma_ops *vops, u8 tile_mask)
+{
+       int i;
+
+       for (i = 0; i < XE_MAX_TILES_PER_DEVICE; ++i)
+               if (BIT(i) & tile_mask)
+                       ++vops->pt_update_ops[i].num_ops;
+}
+
 static void xe_vm_populate_rebind(struct xe_vma_op *op, struct xe_vma *vma,
                                  u8 tile_mask)
 {
@@ -745,6 +781,7 @@ static int xe_vm_ops_add_rebind(struct xe_vma_ops *vops, struct xe_vma *vma,
 
        xe_vm_populate_rebind(op, vma, tile_mask);
        list_add_tail(&op->link, &vops->list);
+       xe_vma_ops_incr_pt_update_ops(vops, tile_mask);
 
        return 0;
 }
@@ -785,6 +822,10 @@ int xe_vm_rebind(struct xe_vm *vm, bool rebind_worker)
                        goto free_ops;
        }
 
+       err = xe_vma_ops_alloc(&vops);
+       if (err)
+               goto free_ops;
+
        fence = ops_execute(vm, &vops);
        if (IS_ERR(fence)) {
                err = PTR_ERR(fence);
@@ -799,6 +840,7 @@ free_ops:
                list_del(&op->link);
                kfree(op);
        }
+       xe_vma_ops_fini(&vops);
 
        return err;
 }
@@ -820,12 +862,20 @@ struct dma_fence *xe_vma_rebind(struct xe_vm *vm, struct xe_vma *vma, u8 tile_ma
        if (err)
                return ERR_PTR(err);
 
+       err = xe_vma_ops_alloc(&vops);
+       if (err) {
+               fence = ERR_PTR(err);
+               goto free_ops;
+       }
+
        fence = ops_execute(vm, &vops);
 
+free_ops:
        list_for_each_entry_safe(op, next_op, &vops.list, link) {
                list_del(&op->link);
                kfree(op);
        }
+       xe_vma_ops_fini(&vops);
 
        return fence;
 }
@@ -2287,7 +2337,6 @@ static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op)
        return err;
 }
 
-
 static int vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct xe_exec_queue *q,
                                   struct drm_gpuva_ops *ops,
                                   struct xe_sync_entry *syncs, u32 num_syncs,
@@ -2339,6 +2388,9 @@ static int vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct xe_exec_queue *q,
                                return PTR_ERR(vma);
 
                        op->map.vma = vma;
+                       if (op->map.immediate || !xe_vm_in_fault_mode(vm))
+                               xe_vma_ops_incr_pt_update_ops(vops,
+                                                             op->tile_mask);
                        break;
                }
                case DRM_GPUVA_OP_REMAP:
@@ -2383,6 +2435,8 @@ static int vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct xe_exec_queue *q,
                                        vm_dbg(&xe->drm, "REMAP:SKIP_PREV: addr=0x%016llx, range=0x%016llx",
                                               (ULL)op->remap.start,
                                               (ULL)op->remap.range);
+                               } else {
+                                       xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask);
                                }
                        }
 
@@ -2419,13 +2473,16 @@ static int vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct xe_exec_queue *q,
                                        vm_dbg(&xe->drm, "REMAP:SKIP_NEXT: addr=0x%016llx, range=0x%016llx",
                                               (ULL)op->remap.start,
                                               (ULL)op->remap.range);
+                               } else {
+                                       xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask);
                                }
                        }
+                       xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask);
                        break;
                }
                case DRM_GPUVA_OP_UNMAP:
                case DRM_GPUVA_OP_PREFETCH:
-                       /* Nothing to do */
+                       xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask);
                        break;
                default:
                        drm_warn(&vm->xe->drm, "NOT POSSIBLE");
@@ -3272,11 +3329,16 @@ int xe_vm_bind_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
                goto unwind_ops;
        }
 
+       err = xe_vma_ops_alloc(&vops);
+       if (err)
+               goto unwind_ops;
+
        err = vm_bind_ioctl_ops_execute(vm, &vops);
 
 unwind_ops:
        if (err && err != -ENODATA)
                vm_bind_ioctl_ops_unwind(vm, ops, args->num_binds);
+       xe_vma_ops_fini(&vops);
        for (i = args->num_binds - 1; i >= 0; --i)
                if (ops[i])
                        drm_gpuva_ops_free(&vm->gpuvm, ops[i]);
index ce1a63a..211c888 100644 (file)
@@ -21,6 +21,7 @@ struct xe_bo;
 struct xe_sync_entry;
 struct xe_user_fence;
 struct xe_vm;
+struct xe_vm_pgtable_update_op;
 
 #define XE_VMA_READ_ONLY       DRM_GPUVA_USERBITS
 #define XE_VMA_DESTROYED       (DRM_GPUVA_USERBITS << 1)
@@ -368,6 +369,13 @@ struct xe_vma_ops {
        struct xe_sync_entry *syncs;
        /** @num_syncs: number of syncs */
        u32 num_syncs;
+       /** @pt_update_ops: page table update operations */
+       struct {
+               /** @ops: operations */
+               struct xe_vm_pgtable_update_op *ops;
+               /** @num_ops: number of operations */
+               u32 num_ops;
+       } pt_update_ops[XE_MAX_TILES_PER_DEVICE];
 };
 
 #endif