Commit 08f72008 authored by Matthew Brost's avatar Matthew Brost Committed by Thomas Hellström

drm/xe: Decouple job seqno and lrc seqno

Tightly coupling these seqno presents problems if alternative fences for
jobs are used. Decouple these for correctness.

v2:
- Slightly reword commit message (Thomas)
- Make sure the lrc fence ops are used in comparison (Thomas)
- Assume seqno is unsigned rather than signed in format string (Thomas)

Cc: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Signed-off-by: default avatarMatthew Brost <matthew.brost@intel.com>
Signed-off-by: default avatarThomas Hellström <thomas.hellstrom@linux.intel.com>
Reviewed-by: default avatarRodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20240527135912.152156-2-thomas.hellstrom@linux.intel.com
parent d79e8cab
...@@ -98,7 +98,7 @@ static struct xe_exec_queue *__xe_exec_queue_alloc(struct xe_device *xe, ...@@ -98,7 +98,7 @@ static struct xe_exec_queue *__xe_exec_queue_alloc(struct xe_device *xe,
if (xe_exec_queue_is_parallel(q)) { if (xe_exec_queue_is_parallel(q)) {
q->parallel.composite_fence_ctx = dma_fence_context_alloc(1); q->parallel.composite_fence_ctx = dma_fence_context_alloc(1);
q->parallel.composite_fence_seqno = XE_FENCE_INITIAL_SEQNO; q->parallel.composite_fence_seqno = 0;
} }
return q; return q;
......
...@@ -940,8 +940,9 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job) ...@@ -940,8 +940,9 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
return DRM_GPU_SCHED_STAT_NOMINAL; return DRM_GPU_SCHED_STAT_NOMINAL;
} }
drm_notice(&xe->drm, "Timedout job: seqno=%u, guc_id=%d, flags=0x%lx", drm_notice(&xe->drm, "Timedout job: seqno=%u, lrc_seqno=%u, guc_id=%d, flags=0x%lx",
xe_sched_job_seqno(job), q->guc->id, q->flags); xe_sched_job_seqno(job), xe_sched_job_lrc_seqno(job),
q->guc->id, q->flags);
xe_gt_WARN(q->gt, q->flags & EXEC_QUEUE_FLAG_KERNEL, xe_gt_WARN(q->gt, q->flags & EXEC_QUEUE_FLAG_KERNEL,
"Kernel-submitted job timed out\n"); "Kernel-submitted job timed out\n");
xe_gt_WARN(q->gt, q->flags & EXEC_QUEUE_FLAG_VM && !exec_queue_killed(q), xe_gt_WARN(q->gt, q->flags & EXEC_QUEUE_FLAG_VM && !exec_queue_killed(q),
......
...@@ -398,7 +398,7 @@ static void emit_job_gen12_gsc(struct xe_sched_job *job) ...@@ -398,7 +398,7 @@ static void emit_job_gen12_gsc(struct xe_sched_job *job)
__emit_job_gen12_simple(job, job->q->lrc, __emit_job_gen12_simple(job, job->q->lrc,
job->batch_addr[0], job->batch_addr[0],
xe_sched_job_seqno(job)); xe_sched_job_lrc_seqno(job));
} }
static void emit_job_gen12_copy(struct xe_sched_job *job) static void emit_job_gen12_copy(struct xe_sched_job *job)
...@@ -407,14 +407,14 @@ static void emit_job_gen12_copy(struct xe_sched_job *job) ...@@ -407,14 +407,14 @@ static void emit_job_gen12_copy(struct xe_sched_job *job)
if (xe_sched_job_is_migration(job->q)) { if (xe_sched_job_is_migration(job->q)) {
emit_migration_job_gen12(job, job->q->lrc, emit_migration_job_gen12(job, job->q->lrc,
xe_sched_job_seqno(job)); xe_sched_job_lrc_seqno(job));
return; return;
} }
for (i = 0; i < job->q->width; ++i) for (i = 0; i < job->q->width; ++i)
__emit_job_gen12_simple(job, job->q->lrc + i, __emit_job_gen12_simple(job, job->q->lrc + i,
job->batch_addr[i], job->batch_addr[i],
xe_sched_job_seqno(job)); xe_sched_job_lrc_seqno(job));
} }
static void emit_job_gen12_video(struct xe_sched_job *job) static void emit_job_gen12_video(struct xe_sched_job *job)
...@@ -425,7 +425,7 @@ static void emit_job_gen12_video(struct xe_sched_job *job) ...@@ -425,7 +425,7 @@ static void emit_job_gen12_video(struct xe_sched_job *job)
for (i = 0; i < job->q->width; ++i) for (i = 0; i < job->q->width; ++i)
__emit_job_gen12_video(job, job->q->lrc + i, __emit_job_gen12_video(job, job->q->lrc + i,
job->batch_addr[i], job->batch_addr[i],
xe_sched_job_seqno(job)); xe_sched_job_lrc_seqno(job));
} }
static void emit_job_gen12_render_compute(struct xe_sched_job *job) static void emit_job_gen12_render_compute(struct xe_sched_job *job)
...@@ -435,7 +435,7 @@ static void emit_job_gen12_render_compute(struct xe_sched_job *job) ...@@ -435,7 +435,7 @@ static void emit_job_gen12_render_compute(struct xe_sched_job *job)
for (i = 0; i < job->q->width; ++i) for (i = 0; i < job->q->width; ++i)
__emit_job_gen12_render_compute(job, job->q->lrc + i, __emit_job_gen12_render_compute(job, job->q->lrc + i,
job->batch_addr[i], job->batch_addr[i],
xe_sched_job_seqno(job)); xe_sched_job_lrc_seqno(job));
} }
static const struct xe_ring_ops ring_ops_gen12_gsc = { static const struct xe_ring_ops ring_ops_gen12_gsc = {
......
...@@ -117,6 +117,7 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q, ...@@ -117,6 +117,7 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q,
err = PTR_ERR(job->fence); err = PTR_ERR(job->fence);
goto err_sched_job; goto err_sched_job;
} }
job->lrc_seqno = job->fence->seqno;
} else { } else {
struct dma_fence_array *cf; struct dma_fence_array *cf;
...@@ -132,6 +133,8 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q, ...@@ -132,6 +133,8 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q,
err = PTR_ERR(fences[j]); err = PTR_ERR(fences[j]);
goto err_fences; goto err_fences;
} }
if (!j)
job->lrc_seqno = fences[0]->seqno;
} }
cf = dma_fence_array_create(q->width, fences, cf = dma_fence_array_create(q->width, fences,
...@@ -144,10 +147,6 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q, ...@@ -144,10 +147,6 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q,
goto err_fences; goto err_fences;
} }
/* Sanity check */
for (j = 0; j < q->width; ++j)
xe_assert(job_to_xe(job), cf->base.seqno == fences[j]->seqno);
job->fence = &cf->base; job->fence = &cf->base;
} }
...@@ -229,9 +228,9 @@ bool xe_sched_job_started(struct xe_sched_job *job) ...@@ -229,9 +228,9 @@ bool xe_sched_job_started(struct xe_sched_job *job)
{ {
struct xe_lrc *lrc = job->q->lrc; struct xe_lrc *lrc = job->q->lrc;
return !__dma_fence_is_later(xe_sched_job_seqno(job), return !__dma_fence_is_later(xe_sched_job_lrc_seqno(job),
xe_lrc_start_seqno(lrc), xe_lrc_start_seqno(lrc),
job->fence->ops); dma_fence_array_first(job->fence)->ops);
} }
bool xe_sched_job_completed(struct xe_sched_job *job) bool xe_sched_job_completed(struct xe_sched_job *job)
...@@ -243,8 +242,9 @@ bool xe_sched_job_completed(struct xe_sched_job *job) ...@@ -243,8 +242,9 @@ bool xe_sched_job_completed(struct xe_sched_job *job)
* parallel handshake is done. * parallel handshake is done.
*/ */
return !__dma_fence_is_later(xe_sched_job_seqno(job), xe_lrc_seqno(lrc), return !__dma_fence_is_later(xe_sched_job_lrc_seqno(job),
job->fence->ops); xe_lrc_seqno(lrc),
dma_fence_array_first(job->fence)->ops);
} }
void xe_sched_job_arm(struct xe_sched_job *job) void xe_sched_job_arm(struct xe_sched_job *job)
......
...@@ -73,6 +73,11 @@ static inline u32 xe_sched_job_seqno(struct xe_sched_job *job) ...@@ -73,6 +73,11 @@ static inline u32 xe_sched_job_seqno(struct xe_sched_job *job)
return job->fence->seqno; return job->fence->seqno;
} }
static inline u32 xe_sched_job_lrc_seqno(struct xe_sched_job *job)
{
return job->lrc_seqno;
}
static inline void static inline void
xe_sched_job_add_migrate_flush(struct xe_sched_job *job, u32 flags) xe_sched_job_add_migrate_flush(struct xe_sched_job *job, u32 flags)
{ {
......
...@@ -37,6 +37,8 @@ struct xe_sched_job { ...@@ -37,6 +37,8 @@ struct xe_sched_job {
/** @user_fence.value: write back value */ /** @user_fence.value: write back value */
u64 value; u64 value;
} user_fence; } user_fence;
/** @lrc_seqno: LRC seqno */
u32 lrc_seqno;
/** @migrate_flush_flags: Additional flush flags for migration jobs */ /** @migrate_flush_flags: Additional flush flags for migration jobs */
u32 migrate_flush_flags; u32 migrate_flush_flags;
/** @ring_ops_flush_tlb: The ring ops need to flush TLB before payload. */ /** @ring_ops_flush_tlb: The ring ops need to flush TLB before payload. */
......
...@@ -254,6 +254,7 @@ DECLARE_EVENT_CLASS(xe_sched_job, ...@@ -254,6 +254,7 @@ DECLARE_EVENT_CLASS(xe_sched_job,
TP_STRUCT__entry( TP_STRUCT__entry(
__field(u32, seqno) __field(u32, seqno)
__field(u32, lrc_seqno)
__field(u16, guc_id) __field(u16, guc_id)
__field(u32, guc_state) __field(u32, guc_state)
__field(u32, flags) __field(u32, flags)
...@@ -264,6 +265,7 @@ DECLARE_EVENT_CLASS(xe_sched_job, ...@@ -264,6 +265,7 @@ DECLARE_EVENT_CLASS(xe_sched_job,
TP_fast_assign( TP_fast_assign(
__entry->seqno = xe_sched_job_seqno(job); __entry->seqno = xe_sched_job_seqno(job);
__entry->lrc_seqno = xe_sched_job_lrc_seqno(job);
__entry->guc_id = job->q->guc->id; __entry->guc_id = job->q->guc->id;
__entry->guc_state = __entry->guc_state =
atomic_read(&job->q->guc->state); atomic_read(&job->q->guc->state);
...@@ -273,8 +275,9 @@ DECLARE_EVENT_CLASS(xe_sched_job, ...@@ -273,8 +275,9 @@ DECLARE_EVENT_CLASS(xe_sched_job,
__entry->batch_addr = (u64)job->batch_addr[0]; __entry->batch_addr = (u64)job->batch_addr[0];
), ),
TP_printk("fence=%p, seqno=%u, guc_id=%d, batch_addr=0x%012llx, guc_state=0x%x, flags=0x%x, error=%d", TP_printk("fence=%p, seqno=%u, lrc_seqno=%u, guc_id=%d, batch_addr=0x%012llx, guc_state=0x%x, flags=0x%x, error=%d",
__entry->fence, __entry->seqno, __entry->guc_id, __entry->fence, __entry->seqno,
__entry->lrc_seqno, __entry->guc_id,
__entry->batch_addr, __entry->guc_state, __entry->batch_addr, __entry->guc_state,
__entry->flags, __entry->error) __entry->flags, __entry->error)
); );
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment