Commit 5ef091fc authored by Matthew Brost's avatar Matthew Brost Committed by Rodrigo Vivi

drm/xe: Fixup unwind on VM ops errors

Remap ops have 3 parts: unmap, prev, and next. The commit step can fail
on any of these. Add a flag for each to these so the unwind is only done
the steps that have been committed.

v2: (Rodrigo) Use bit macros
Reviewed-by: default avatarRodrigo Vivi <rodrigo.vivi@intel.com>
Signed-off-by: default avatarMatthew Brost <matthew.brost@intel.com>
Signed-off-by: default avatarRodrigo Vivi <rodrigo.vivi@intel.com>
parent 35dfb484
...@@ -2622,18 +2622,25 @@ static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op) ...@@ -2622,18 +2622,25 @@ static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op)
switch (op->base.op) { switch (op->base.op) {
case DRM_GPUVA_OP_MAP: case DRM_GPUVA_OP_MAP:
err |= xe_vm_insert_vma(vm, op->map.vma); err |= xe_vm_insert_vma(vm, op->map.vma);
if (!err)
op->flags |= XE_VMA_OP_COMMITTED;
break; break;
case DRM_GPUVA_OP_REMAP: case DRM_GPUVA_OP_REMAP:
prep_vma_destroy(vm, gpuva_to_vma(op->base.remap.unmap->va), prep_vma_destroy(vm, gpuva_to_vma(op->base.remap.unmap->va),
true); true);
op->flags |= XE_VMA_OP_COMMITTED;
if (op->remap.prev) { if (op->remap.prev) {
err |= xe_vm_insert_vma(vm, op->remap.prev); err |= xe_vm_insert_vma(vm, op->remap.prev);
if (!err)
op->flags |= XE_VMA_OP_PREV_COMMITTED;
if (!err && op->remap.skip_prev) if (!err && op->remap.skip_prev)
op->remap.prev = NULL; op->remap.prev = NULL;
} }
if (op->remap.next) { if (op->remap.next) {
err |= xe_vm_insert_vma(vm, op->remap.next); err |= xe_vm_insert_vma(vm, op->remap.next);
if (!err)
op->flags |= XE_VMA_OP_NEXT_COMMITTED;
if (!err && op->remap.skip_next) if (!err && op->remap.skip_next)
op->remap.next = NULL; op->remap.next = NULL;
} }
...@@ -2646,15 +2653,15 @@ static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op) ...@@ -2646,15 +2653,15 @@ static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op)
break; break;
case DRM_GPUVA_OP_UNMAP: case DRM_GPUVA_OP_UNMAP:
prep_vma_destroy(vm, gpuva_to_vma(op->base.unmap.va), true); prep_vma_destroy(vm, gpuva_to_vma(op->base.unmap.va), true);
op->flags |= XE_VMA_OP_COMMITTED;
break; break;
case DRM_GPUVA_OP_PREFETCH: case DRM_GPUVA_OP_PREFETCH:
/* Nothing to do */ op->flags |= XE_VMA_OP_COMMITTED;
break; break;
default: default:
XE_WARN_ON("NOT POSSIBLE"); XE_WARN_ON("NOT POSSIBLE");
} }
op->flags |= XE_VMA_OP_COMMITTED;
return err; return err;
} }
...@@ -2859,7 +2866,8 @@ static void xe_vma_op_cleanup(struct xe_vm *vm, struct xe_vma_op *op) ...@@ -2859,7 +2866,8 @@ static void xe_vma_op_cleanup(struct xe_vm *vm, struct xe_vma_op *op)
} }
static void xe_vma_op_unwind(struct xe_vm *vm, struct xe_vma_op *op, static void xe_vma_op_unwind(struct xe_vm *vm, struct xe_vma_op *op,
bool post_commit) bool post_commit, bool prev_post_commit,
bool next_post_commit)
{ {
lockdep_assert_held_write(&vm->lock); lockdep_assert_held_write(&vm->lock);
...@@ -2886,11 +2894,11 @@ static void xe_vma_op_unwind(struct xe_vm *vm, struct xe_vma_op *op, ...@@ -2886,11 +2894,11 @@ static void xe_vma_op_unwind(struct xe_vm *vm, struct xe_vma_op *op,
struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va); struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va);
if (op->remap.prev) { if (op->remap.prev) {
prep_vma_destroy(vm, op->remap.prev, post_commit); prep_vma_destroy(vm, op->remap.prev, prev_post_commit);
xe_vma_destroy_unlocked(op->remap.prev); xe_vma_destroy_unlocked(op->remap.prev);
} }
if (op->remap.next) { if (op->remap.next) {
prep_vma_destroy(vm, op->remap.next, post_commit); prep_vma_destroy(vm, op->remap.next, next_post_commit);
xe_vma_destroy_unlocked(op->remap.next); xe_vma_destroy_unlocked(op->remap.next);
} }
down_read(&vm->userptr.notifier_lock); down_read(&vm->userptr.notifier_lock);
...@@ -3029,7 +3037,9 @@ static int vm_bind_ioctl_ops_commit(struct xe_vm *vm, ...@@ -3029,7 +3037,9 @@ static int vm_bind_ioctl_ops_commit(struct xe_vm *vm,
unwind: unwind:
list_for_each_entry_reverse(op, ops_list, link) list_for_each_entry_reverse(op, ops_list, link)
xe_vma_op_unwind(vm, op, op->flags & XE_VMA_OP_COMMITTED); xe_vma_op_unwind(vm, op, op->flags & XE_VMA_OP_COMMITTED,
op->flags & XE_VMA_OP_PREV_COMMITTED,
op->flags & XE_VMA_OP_NEXT_COMMITTED);
list_for_each_entry_safe(op, next, ops_list, link) list_for_each_entry_safe(op, next, ops_list, link)
xe_vma_op_cleanup(vm, op); xe_vma_op_cleanup(vm, op);
...@@ -3056,7 +3066,7 @@ static void vm_bind_ioctl_ops_unwind(struct xe_vm *vm, ...@@ -3056,7 +3066,7 @@ static void vm_bind_ioctl_ops_unwind(struct xe_vm *vm,
drm_gpuva_for_each_op(__op, __ops) { drm_gpuva_for_each_op(__op, __ops) {
struct xe_vma_op *op = gpuva_op_to_vma_op(__op); struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
xe_vma_op_unwind(vm, op, false); xe_vma_op_unwind(vm, op, false, false, false);
} }
} }
} }
......
...@@ -370,11 +370,15 @@ struct xe_vma_op_prefetch { ...@@ -370,11 +370,15 @@ struct xe_vma_op_prefetch {
/** enum xe_vma_op_flags - flags for VMA operation */ /** enum xe_vma_op_flags - flags for VMA operation */
enum xe_vma_op_flags { enum xe_vma_op_flags {
/** @XE_VMA_OP_FIRST: first VMA operation for a set of syncs */ /** @XE_VMA_OP_FIRST: first VMA operation for a set of syncs */
XE_VMA_OP_FIRST = BIT(0), XE_VMA_OP_FIRST = BIT(0),
/** @XE_VMA_OP_LAST: last VMA operation for a set of syncs */ /** @XE_VMA_OP_LAST: last VMA operation for a set of syncs */
XE_VMA_OP_LAST = BIT(1), XE_VMA_OP_LAST = BIT(1),
/** @XE_VMA_OP_COMMITTED: VMA operation committed */ /** @XE_VMA_OP_COMMITTED: VMA operation committed */
XE_VMA_OP_COMMITTED = BIT(2), XE_VMA_OP_COMMITTED = BIT(2),
/** @XE_VMA_OP_PREV_COMMITTED: Previous VMA operation committed */
XE_VMA_OP_PREV_COMMITTED = BIT(3),
/** @XE_VMA_OP_NEXT_COMMITTED: Next VMA operation committed */
XE_VMA_OP_NEXT_COMMITTED = BIT(4),
}; };
/** struct xe_vma_op - VMA operation */ /** struct xe_vma_op - VMA operation */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment