Commit ddf07be7 authored by Chris Wilson's avatar Chris Wilson

drm/i915: Simplify calling engine->sync_to

Since requests can no longer be generated as a side-effect of
intel_ring_begin(), we know that the seqno will be unchanged during
ring-emission. This predicatablity then means we do not have to check
for the seqno wrapping around whilst emitting the semaphore for
engine->sync_to().
Signed-off-by: default avatarChris Wilson <chris@chris-wilson.co.uk>
Reviewed-by: default avatarJoonas Lahtinen <joonas.lahtinen@linux.intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1469432687-22756-31-git-send-email-chris@chris-wilson.co.uk
Link: http://patchwork.freedesktop.org/patch/msgid/1470174640-18242-22-git-send-email-chris@chris-wilson.co.uk
parent 5b043f4e
...@@ -1757,7 +1757,7 @@ struct drm_i915_private { ...@@ -1757,7 +1757,7 @@ struct drm_i915_private {
struct i915_gem_context *kernel_context; struct i915_gem_context *kernel_context;
struct intel_engine_cs engine[I915_NUM_ENGINES]; struct intel_engine_cs engine[I915_NUM_ENGINES];
struct drm_i915_gem_object *semaphore_obj; struct drm_i915_gem_object *semaphore_obj;
uint32_t last_seqno, next_seqno; u32 next_seqno;
struct drm_dma_handle *status_page_dmah; struct drm_dma_handle *status_page_dmah;
struct resource mch_res; struct resource mch_res;
......
...@@ -2867,22 +2867,15 @@ __i915_gem_object_sync(struct drm_i915_gem_object *obj, ...@@ -2867,22 +2867,15 @@ __i915_gem_object_sync(struct drm_i915_gem_object *obj,
i915_gem_object_retire_request(obj, from); i915_gem_object_retire_request(obj, from);
} else { } else {
int idx = intel_engine_sync_index(from->engine, to->engine); int idx = intel_engine_sync_index(from->engine, to->engine);
u32 seqno = i915_gem_request_get_seqno(from); if (from->fence.seqno <= from->engine->semaphore.sync_seqno[idx])
if (seqno <= from->engine->semaphore.sync_seqno[idx])
return 0; return 0;
trace_i915_gem_ring_sync_to(to, from); trace_i915_gem_ring_sync_to(to, from);
ret = to->engine->semaphore.sync_to(to, from->engine, seqno); ret = to->engine->semaphore.sync_to(to, from);
if (ret) if (ret)
return ret; return ret;
/* We use last_read_req because sync_to() from->engine->semaphore.sync_seqno[idx] = from->fence.seqno;
* might have just caused seqno wrap under
* the radar.
*/
from->engine->semaphore.sync_seqno[idx] =
i915_gem_request_get_seqno(obj->last_read_req[from->engine->id]);
} }
return 0; return 0;
......
...@@ -264,14 +264,7 @@ int i915_gem_set_seqno(struct drm_device *dev, u32 seqno) ...@@ -264,14 +264,7 @@ int i915_gem_set_seqno(struct drm_device *dev, u32 seqno)
if (ret) if (ret)
return ret; return ret;
/* Carefully set the last_seqno value so that wrap
* detection still works
*/
dev_priv->next_seqno = seqno; dev_priv->next_seqno = seqno;
dev_priv->last_seqno = seqno - 1;
if (dev_priv->last_seqno == 0)
dev_priv->last_seqno--;
return 0; return 0;
} }
...@@ -288,7 +281,7 @@ static int i915_gem_get_seqno(struct drm_i915_private *dev_priv, u32 *seqno) ...@@ -288,7 +281,7 @@ static int i915_gem_get_seqno(struct drm_i915_private *dev_priv, u32 *seqno)
dev_priv->next_seqno = 1; dev_priv->next_seqno = 1;
} }
*seqno = dev_priv->last_seqno = dev_priv->next_seqno++; *seqno = dev_priv->next_seqno++;
return 0; return 0;
} }
......
...@@ -1496,12 +1496,6 @@ static int gen8_render_emit_request(struct drm_i915_gem_request *req) ...@@ -1496,12 +1496,6 @@ static int gen8_render_emit_request(struct drm_i915_gem_request *req)
return 0; return 0;
} }
static inline bool i915_gem_has_seqno_wrapped(struct drm_i915_private *dev_priv,
u32 seqno)
{
return dev_priv->last_seqno < seqno;
}
/** /**
* intel_ring_sync - sync the waiter to the signaller on seqno * intel_ring_sync - sync the waiter to the signaller on seqno
* *
...@@ -1511,24 +1505,23 @@ static inline bool i915_gem_has_seqno_wrapped(struct drm_i915_private *dev_priv, ...@@ -1511,24 +1505,23 @@ static inline bool i915_gem_has_seqno_wrapped(struct drm_i915_private *dev_priv,
*/ */
static int static int
gen8_ring_sync(struct drm_i915_gem_request *waiter_req, gen8_ring_sync(struct drm_i915_gem_request *wait,
struct intel_engine_cs *signaller, struct drm_i915_gem_request *signal)
u32 seqno)
{ {
struct intel_ring *waiter = waiter_req->ring; struct intel_ring *waiter = wait->ring;
struct drm_i915_private *dev_priv = waiter_req->i915; struct drm_i915_private *dev_priv = wait->i915;
u64 offset = GEN8_WAIT_OFFSET(waiter_req->engine, signaller->id); u64 offset = GEN8_WAIT_OFFSET(wait->engine, signal->engine->id);
struct i915_hw_ppgtt *ppgtt; struct i915_hw_ppgtt *ppgtt;
int ret; int ret;
ret = intel_ring_begin(waiter_req, 4); ret = intel_ring_begin(wait, 4);
if (ret) if (ret)
return ret; return ret;
intel_ring_emit(waiter, MI_SEMAPHORE_WAIT | intel_ring_emit(waiter, MI_SEMAPHORE_WAIT |
MI_SEMAPHORE_GLOBAL_GTT | MI_SEMAPHORE_GLOBAL_GTT |
MI_SEMAPHORE_SAD_GTE_SDD); MI_SEMAPHORE_SAD_GTE_SDD);
intel_ring_emit(waiter, seqno); intel_ring_emit(waiter, signal->fence.seqno);
intel_ring_emit(waiter, lower_32_bits(offset)); intel_ring_emit(waiter, lower_32_bits(offset));
intel_ring_emit(waiter, upper_32_bits(offset)); intel_ring_emit(waiter, upper_32_bits(offset));
intel_ring_advance(waiter); intel_ring_advance(waiter);
...@@ -1538,48 +1531,37 @@ gen8_ring_sync(struct drm_i915_gem_request *waiter_req, ...@@ -1538,48 +1531,37 @@ gen8_ring_sync(struct drm_i915_gem_request *waiter_req,
* We do this on the i915_switch_context() following the wait and * We do this on the i915_switch_context() following the wait and
* before the dispatch. * before the dispatch.
*/ */
ppgtt = waiter_req->ctx->ppgtt; ppgtt = wait->ctx->ppgtt;
if (ppgtt && waiter_req->engine->id != RCS) if (ppgtt && wait->engine->id != RCS)
ppgtt->pd_dirty_rings |= intel_engine_flag(waiter_req->engine); ppgtt->pd_dirty_rings |= intel_engine_flag(wait->engine);
return 0; return 0;
} }
static int static int
gen6_ring_sync(struct drm_i915_gem_request *waiter_req, gen6_ring_sync(struct drm_i915_gem_request *wait,
struct intel_engine_cs *signaller, struct drm_i915_gem_request *signal)
u32 seqno)
{ {
struct intel_ring *waiter = waiter_req->ring; struct intel_ring *waiter = wait->ring;
u32 dw1 = MI_SEMAPHORE_MBOX | u32 dw1 = MI_SEMAPHORE_MBOX |
MI_SEMAPHORE_COMPARE | MI_SEMAPHORE_COMPARE |
MI_SEMAPHORE_REGISTER; MI_SEMAPHORE_REGISTER;
u32 wait_mbox = signaller->semaphore.mbox.wait[waiter_req->engine->id]; u32 wait_mbox = signal->engine->semaphore.mbox.wait[wait->engine->id];
int ret; int ret;
/* Throughout all of the GEM code, seqno passed implies our current
* seqno is >= the last seqno executed. However for hardware the
* comparison is strictly greater than.
*/
seqno -= 1;
WARN_ON(wait_mbox == MI_SEMAPHORE_SYNC_INVALID); WARN_ON(wait_mbox == MI_SEMAPHORE_SYNC_INVALID);
ret = intel_ring_begin(waiter_req, 4); ret = intel_ring_begin(wait, 4);
if (ret) if (ret)
return ret; return ret;
/* If seqno wrap happened, omit the wait with no-ops */ intel_ring_emit(waiter, dw1 | wait_mbox);
if (likely(!i915_gem_has_seqno_wrapped(waiter_req->i915, seqno))) { /* Throughout all of the GEM code, seqno passed implies our current
intel_ring_emit(waiter, dw1 | wait_mbox); * seqno is >= the last seqno executed. However for hardware the
intel_ring_emit(waiter, seqno); * comparison is strictly greater than.
intel_ring_emit(waiter, 0); */
intel_ring_emit(waiter, MI_NOOP); intel_ring_emit(waiter, signal->fence.seqno - 1);
} else { intel_ring_emit(waiter, 0);
intel_ring_emit(waiter, MI_NOOP); intel_ring_emit(waiter, MI_NOOP);
intel_ring_emit(waiter, MI_NOOP);
intel_ring_emit(waiter, MI_NOOP);
intel_ring_emit(waiter, MI_NOOP);
}
intel_ring_advance(waiter); intel_ring_advance(waiter);
return 0; return 0;
......
...@@ -277,9 +277,8 @@ struct intel_engine_cs { ...@@ -277,9 +277,8 @@ struct intel_engine_cs {
}; };
/* AKA wait() */ /* AKA wait() */
int (*sync_to)(struct drm_i915_gem_request *to_req, int (*sync_to)(struct drm_i915_gem_request *to,
struct intel_engine_cs *from, struct drm_i915_gem_request *from);
u32 seqno);
int (*signal)(struct drm_i915_gem_request *signaller_req); int (*signal)(struct drm_i915_gem_request *signaller_req);
} semaphore; } semaphore;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment