Commit ce07fe9e authored by Dave Airlie's avatar Dave Airlie

Merge tag 'drm-intel-fixes-2018-12-12-1' of...

Merge tag 'drm-intel-fixes-2018-12-12-1' of git://anongit.freedesktop.org/drm/drm-intel into drm-fixes

- Two fixes to avoid GPU hangs (on Braswell and Gen3)
Signed-off-by: default avatarDave Airlie <airlied@redhat.com>
From: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212134010.GA18900@jlahtine-desk.ger.corp.intel.com
parents 71fb5539 5b2e3120
...@@ -2150,6 +2150,8 @@ struct drm_i915_private { ...@@ -2150,6 +2150,8 @@ struct drm_i915_private {
struct delayed_work idle_work; struct delayed_work idle_work;
ktime_t last_init_time; ktime_t last_init_time;
struct i915_vma *scratch;
} gt; } gt;
/* perform PHY state sanity checks? */ /* perform PHY state sanity checks? */
...@@ -3872,4 +3874,9 @@ static inline int intel_hws_csb_write_index(struct drm_i915_private *i915) ...@@ -3872,4 +3874,9 @@ static inline int intel_hws_csb_write_index(struct drm_i915_private *i915)
return I915_HWS_CSB_WRITE_INDEX; return I915_HWS_CSB_WRITE_INDEX;
} }
static inline u32 i915_scratch_offset(const struct drm_i915_private *i915)
{
return i915_ggtt_offset(i915->gt.scratch);
}
#endif #endif
...@@ -5500,6 +5500,44 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915) ...@@ -5500,6 +5500,44 @@ static int __intel_engines_record_defaults(struct drm_i915_private *i915)
goto out_ctx; goto out_ctx;
} }
static int
i915_gem_init_scratch(struct drm_i915_private *i915, unsigned int size)
{
struct drm_i915_gem_object *obj;
struct i915_vma *vma;
int ret;
obj = i915_gem_object_create_stolen(i915, size);
if (!obj)
obj = i915_gem_object_create_internal(i915, size);
if (IS_ERR(obj)) {
DRM_ERROR("Failed to allocate scratch page\n");
return PTR_ERR(obj);
}
vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL);
if (IS_ERR(vma)) {
ret = PTR_ERR(vma);
goto err_unref;
}
ret = i915_vma_pin(vma, 0, 0, PIN_GLOBAL | PIN_HIGH);
if (ret)
goto err_unref;
i915->gt.scratch = vma;
return 0;
err_unref:
i915_gem_object_put(obj);
return ret;
}
static void i915_gem_fini_scratch(struct drm_i915_private *i915)
{
i915_vma_unpin_and_release(&i915->gt.scratch, 0);
}
int i915_gem_init(struct drm_i915_private *dev_priv) int i915_gem_init(struct drm_i915_private *dev_priv)
{ {
int ret; int ret;
...@@ -5546,12 +5584,19 @@ int i915_gem_init(struct drm_i915_private *dev_priv) ...@@ -5546,12 +5584,19 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
goto err_unlock; goto err_unlock;
} }
ret = i915_gem_contexts_init(dev_priv); ret = i915_gem_init_scratch(dev_priv,
IS_GEN2(dev_priv) ? SZ_256K : PAGE_SIZE);
if (ret) { if (ret) {
GEM_BUG_ON(ret == -EIO); GEM_BUG_ON(ret == -EIO);
goto err_ggtt; goto err_ggtt;
} }
ret = i915_gem_contexts_init(dev_priv);
if (ret) {
GEM_BUG_ON(ret == -EIO);
goto err_scratch;
}
ret = intel_engines_init(dev_priv); ret = intel_engines_init(dev_priv);
if (ret) { if (ret) {
GEM_BUG_ON(ret == -EIO); GEM_BUG_ON(ret == -EIO);
...@@ -5624,6 +5669,8 @@ int i915_gem_init(struct drm_i915_private *dev_priv) ...@@ -5624,6 +5669,8 @@ int i915_gem_init(struct drm_i915_private *dev_priv)
err_context: err_context:
if (ret != -EIO) if (ret != -EIO)
i915_gem_contexts_fini(dev_priv); i915_gem_contexts_fini(dev_priv);
err_scratch:
i915_gem_fini_scratch(dev_priv);
err_ggtt: err_ggtt:
err_unlock: err_unlock:
intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL); intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL);
...@@ -5675,6 +5722,7 @@ void i915_gem_fini(struct drm_i915_private *dev_priv) ...@@ -5675,6 +5722,7 @@ void i915_gem_fini(struct drm_i915_private *dev_priv)
intel_uc_fini(dev_priv); intel_uc_fini(dev_priv);
i915_gem_cleanup_engines(dev_priv); i915_gem_cleanup_engines(dev_priv);
i915_gem_contexts_fini(dev_priv); i915_gem_contexts_fini(dev_priv);
i915_gem_fini_scratch(dev_priv);
mutex_unlock(&dev_priv->drm.struct_mutex); mutex_unlock(&dev_priv->drm.struct_mutex);
intel_wa_list_free(&dev_priv->gt_wa_list); intel_wa_list_free(&dev_priv->gt_wa_list);
......
...@@ -1268,7 +1268,7 @@ relocate_entry(struct i915_vma *vma, ...@@ -1268,7 +1268,7 @@ relocate_entry(struct i915_vma *vma,
else if (gen >= 4) else if (gen >= 4)
len = 4; len = 4;
else else
len = 6; len = 3;
batch = reloc_gpu(eb, vma, len); batch = reloc_gpu(eb, vma, len);
if (IS_ERR(batch)) if (IS_ERR(batch))
...@@ -1309,11 +1309,6 @@ relocate_entry(struct i915_vma *vma, ...@@ -1309,11 +1309,6 @@ relocate_entry(struct i915_vma *vma,
*batch++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL; *batch++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL;
*batch++ = addr; *batch++ = addr;
*batch++ = target_offset; *batch++ = target_offset;
/* And again for good measure (blb/pnv) */
*batch++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL;
*batch++ = addr;
*batch++ = target_offset;
} }
goto out; goto out;
......
...@@ -1495,7 +1495,7 @@ static void gem_record_rings(struct i915_gpu_state *error) ...@@ -1495,7 +1495,7 @@ static void gem_record_rings(struct i915_gpu_state *error)
if (HAS_BROKEN_CS_TLB(i915)) if (HAS_BROKEN_CS_TLB(i915))
ee->wa_batchbuffer = ee->wa_batchbuffer =
i915_error_object_create(i915, i915_error_object_create(i915,
engine->scratch); i915->gt.scratch);
request_record_user_bo(request, ee); request_record_user_bo(request, ee);
ee->ctx = ee->ctx =
......
...@@ -490,46 +490,6 @@ void intel_engine_setup_common(struct intel_engine_cs *engine) ...@@ -490,46 +490,6 @@ void intel_engine_setup_common(struct intel_engine_cs *engine)
intel_engine_init_cmd_parser(engine); intel_engine_init_cmd_parser(engine);
} }
int intel_engine_create_scratch(struct intel_engine_cs *engine,
unsigned int size)
{
struct drm_i915_gem_object *obj;
struct i915_vma *vma;
int ret;
WARN_ON(engine->scratch);
obj = i915_gem_object_create_stolen(engine->i915, size);
if (!obj)
obj = i915_gem_object_create_internal(engine->i915, size);
if (IS_ERR(obj)) {
DRM_ERROR("Failed to allocate scratch page\n");
return PTR_ERR(obj);
}
vma = i915_vma_instance(obj, &engine->i915->ggtt.vm, NULL);
if (IS_ERR(vma)) {
ret = PTR_ERR(vma);
goto err_unref;
}
ret = i915_vma_pin(vma, 0, 0, PIN_GLOBAL | PIN_HIGH);
if (ret)
goto err_unref;
engine->scratch = vma;
return 0;
err_unref:
i915_gem_object_put(obj);
return ret;
}
void intel_engine_cleanup_scratch(struct intel_engine_cs *engine)
{
i915_vma_unpin_and_release(&engine->scratch, 0);
}
static void cleanup_status_page(struct intel_engine_cs *engine) static void cleanup_status_page(struct intel_engine_cs *engine)
{ {
if (HWS_NEEDS_PHYSICAL(engine->i915)) { if (HWS_NEEDS_PHYSICAL(engine->i915)) {
...@@ -704,8 +664,6 @@ void intel_engine_cleanup_common(struct intel_engine_cs *engine) ...@@ -704,8 +664,6 @@ void intel_engine_cleanup_common(struct intel_engine_cs *engine)
{ {
struct drm_i915_private *i915 = engine->i915; struct drm_i915_private *i915 = engine->i915;
intel_engine_cleanup_scratch(engine);
cleanup_status_page(engine); cleanup_status_page(engine);
intel_engine_fini_breadcrumbs(engine); intel_engine_fini_breadcrumbs(engine);
......
...@@ -442,8 +442,13 @@ static u64 execlists_update_context(struct i915_request *rq) ...@@ -442,8 +442,13 @@ static u64 execlists_update_context(struct i915_request *rq)
* may not be visible to the HW prior to the completion of the UC * may not be visible to the HW prior to the completion of the UC
* register write and that we may begin execution from the context * register write and that we may begin execution from the context
* before its image is complete leading to invalid PD chasing. * before its image is complete leading to invalid PD chasing.
*
* Furthermore, Braswell, at least, wants a full mb to be sure that
* the writes are coherent in memory (visible to the GPU) prior to
* execution, and not just visible to other CPUs (as is the result of
* wmb).
*/ */
wmb(); mb();
return ce->lrc_desc; return ce->lrc_desc;
} }
...@@ -1443,9 +1448,10 @@ static int execlists_request_alloc(struct i915_request *request) ...@@ -1443,9 +1448,10 @@ static int execlists_request_alloc(struct i915_request *request)
static u32 * static u32 *
gen8_emit_flush_coherentl3_wa(struct intel_engine_cs *engine, u32 *batch) gen8_emit_flush_coherentl3_wa(struct intel_engine_cs *engine, u32 *batch)
{ {
/* NB no one else is allowed to scribble over scratch + 256! */
*batch++ = MI_STORE_REGISTER_MEM_GEN8 | MI_SRM_LRM_GLOBAL_GTT; *batch++ = MI_STORE_REGISTER_MEM_GEN8 | MI_SRM_LRM_GLOBAL_GTT;
*batch++ = i915_mmio_reg_offset(GEN8_L3SQCREG4); *batch++ = i915_mmio_reg_offset(GEN8_L3SQCREG4);
*batch++ = i915_ggtt_offset(engine->scratch) + 256; *batch++ = i915_scratch_offset(engine->i915) + 256;
*batch++ = 0; *batch++ = 0;
*batch++ = MI_LOAD_REGISTER_IMM(1); *batch++ = MI_LOAD_REGISTER_IMM(1);
...@@ -1459,7 +1465,7 @@ gen8_emit_flush_coherentl3_wa(struct intel_engine_cs *engine, u32 *batch) ...@@ -1459,7 +1465,7 @@ gen8_emit_flush_coherentl3_wa(struct intel_engine_cs *engine, u32 *batch)
*batch++ = MI_LOAD_REGISTER_MEM_GEN8 | MI_SRM_LRM_GLOBAL_GTT; *batch++ = MI_LOAD_REGISTER_MEM_GEN8 | MI_SRM_LRM_GLOBAL_GTT;
*batch++ = i915_mmio_reg_offset(GEN8_L3SQCREG4); *batch++ = i915_mmio_reg_offset(GEN8_L3SQCREG4);
*batch++ = i915_ggtt_offset(engine->scratch) + 256; *batch++ = i915_scratch_offset(engine->i915) + 256;
*batch++ = 0; *batch++ = 0;
return batch; return batch;
...@@ -1496,7 +1502,7 @@ static u32 *gen8_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch) ...@@ -1496,7 +1502,7 @@ static u32 *gen8_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch)
PIPE_CONTROL_GLOBAL_GTT_IVB | PIPE_CONTROL_GLOBAL_GTT_IVB |
PIPE_CONTROL_CS_STALL | PIPE_CONTROL_CS_STALL |
PIPE_CONTROL_QW_WRITE, PIPE_CONTROL_QW_WRITE,
i915_ggtt_offset(engine->scratch) + i915_scratch_offset(engine->i915) +
2 * CACHELINE_BYTES); 2 * CACHELINE_BYTES);
*batch++ = MI_ARB_ON_OFF | MI_ARB_ENABLE; *batch++ = MI_ARB_ON_OFF | MI_ARB_ENABLE;
...@@ -1573,7 +1579,7 @@ static u32 *gen9_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch) ...@@ -1573,7 +1579,7 @@ static u32 *gen9_init_indirectctx_bb(struct intel_engine_cs *engine, u32 *batch)
PIPE_CONTROL_GLOBAL_GTT_IVB | PIPE_CONTROL_GLOBAL_GTT_IVB |
PIPE_CONTROL_CS_STALL | PIPE_CONTROL_CS_STALL |
PIPE_CONTROL_QW_WRITE, PIPE_CONTROL_QW_WRITE,
i915_ggtt_offset(engine->scratch) i915_scratch_offset(engine->i915)
+ 2 * CACHELINE_BYTES); + 2 * CACHELINE_BYTES);
} }
...@@ -2141,7 +2147,7 @@ static int gen8_emit_flush_render(struct i915_request *request, ...@@ -2141,7 +2147,7 @@ static int gen8_emit_flush_render(struct i915_request *request,
{ {
struct intel_engine_cs *engine = request->engine; struct intel_engine_cs *engine = request->engine;
u32 scratch_addr = u32 scratch_addr =
i915_ggtt_offset(engine->scratch) + 2 * CACHELINE_BYTES; i915_scratch_offset(engine->i915) + 2 * CACHELINE_BYTES;
bool vf_flush_wa = false, dc_flush_wa = false; bool vf_flush_wa = false, dc_flush_wa = false;
u32 *cs, flags = 0; u32 *cs, flags = 0;
int len; int len;
...@@ -2478,10 +2484,6 @@ int logical_render_ring_init(struct intel_engine_cs *engine) ...@@ -2478,10 +2484,6 @@ int logical_render_ring_init(struct intel_engine_cs *engine)
if (ret) if (ret)
return ret; return ret;
ret = intel_engine_create_scratch(engine, PAGE_SIZE);
if (ret)
goto err_cleanup_common;
ret = intel_init_workaround_bb(engine); ret = intel_init_workaround_bb(engine);
if (ret) { if (ret) {
/* /*
...@@ -2496,10 +2498,6 @@ int logical_render_ring_init(struct intel_engine_cs *engine) ...@@ -2496,10 +2498,6 @@ int logical_render_ring_init(struct intel_engine_cs *engine)
intel_engine_init_workarounds(engine); intel_engine_init_workarounds(engine);
return 0; return 0;
err_cleanup_common:
intel_engine_cleanup_common(engine);
return ret;
} }
int logical_xcs_ring_init(struct intel_engine_cs *engine) int logical_xcs_ring_init(struct intel_engine_cs *engine)
......
...@@ -69,19 +69,28 @@ unsigned int intel_ring_update_space(struct intel_ring *ring) ...@@ -69,19 +69,28 @@ unsigned int intel_ring_update_space(struct intel_ring *ring)
static int static int
gen2_render_ring_flush(struct i915_request *rq, u32 mode) gen2_render_ring_flush(struct i915_request *rq, u32 mode)
{ {
unsigned int num_store_dw;
u32 cmd, *cs; u32 cmd, *cs;
cmd = MI_FLUSH; cmd = MI_FLUSH;
num_store_dw = 0;
if (mode & EMIT_INVALIDATE) if (mode & EMIT_INVALIDATE)
cmd |= MI_READ_FLUSH; cmd |= MI_READ_FLUSH;
if (mode & EMIT_FLUSH)
num_store_dw = 4;
cs = intel_ring_begin(rq, 2); cs = intel_ring_begin(rq, 2 + 3 * num_store_dw);
if (IS_ERR(cs)) if (IS_ERR(cs))
return PTR_ERR(cs); return PTR_ERR(cs);
*cs++ = cmd; *cs++ = cmd;
*cs++ = MI_NOOP; while (num_store_dw--) {
*cs++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL;
*cs++ = i915_scratch_offset(rq->i915);
*cs++ = 0;
}
*cs++ = MI_FLUSH | MI_NO_WRITE_FLUSH;
intel_ring_advance(rq, cs); intel_ring_advance(rq, cs);
return 0; return 0;
...@@ -150,8 +159,7 @@ gen4_render_ring_flush(struct i915_request *rq, u32 mode) ...@@ -150,8 +159,7 @@ gen4_render_ring_flush(struct i915_request *rq, u32 mode)
*/ */
if (mode & EMIT_INVALIDATE) { if (mode & EMIT_INVALIDATE) {
*cs++ = GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE; *cs++ = GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE;
*cs++ = i915_ggtt_offset(rq->engine->scratch) | *cs++ = i915_scratch_offset(rq->i915) | PIPE_CONTROL_GLOBAL_GTT;
PIPE_CONTROL_GLOBAL_GTT;
*cs++ = 0; *cs++ = 0;
*cs++ = 0; *cs++ = 0;
...@@ -159,8 +167,7 @@ gen4_render_ring_flush(struct i915_request *rq, u32 mode) ...@@ -159,8 +167,7 @@ gen4_render_ring_flush(struct i915_request *rq, u32 mode)
*cs++ = MI_FLUSH; *cs++ = MI_FLUSH;
*cs++ = GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE; *cs++ = GFX_OP_PIPE_CONTROL(4) | PIPE_CONTROL_QW_WRITE;
*cs++ = i915_ggtt_offset(rq->engine->scratch) | *cs++ = i915_scratch_offset(rq->i915) | PIPE_CONTROL_GLOBAL_GTT;
PIPE_CONTROL_GLOBAL_GTT;
*cs++ = 0; *cs++ = 0;
*cs++ = 0; *cs++ = 0;
} }
...@@ -212,8 +219,7 @@ gen4_render_ring_flush(struct i915_request *rq, u32 mode) ...@@ -212,8 +219,7 @@ gen4_render_ring_flush(struct i915_request *rq, u32 mode)
static int static int
intel_emit_post_sync_nonzero_flush(struct i915_request *rq) intel_emit_post_sync_nonzero_flush(struct i915_request *rq)
{ {
u32 scratch_addr = u32 scratch_addr = i915_scratch_offset(rq->i915) + 2 * CACHELINE_BYTES;
i915_ggtt_offset(rq->engine->scratch) + 2 * CACHELINE_BYTES;
u32 *cs; u32 *cs;
cs = intel_ring_begin(rq, 6); cs = intel_ring_begin(rq, 6);
...@@ -246,8 +252,7 @@ intel_emit_post_sync_nonzero_flush(struct i915_request *rq) ...@@ -246,8 +252,7 @@ intel_emit_post_sync_nonzero_flush(struct i915_request *rq)
static int static int
gen6_render_ring_flush(struct i915_request *rq, u32 mode) gen6_render_ring_flush(struct i915_request *rq, u32 mode)
{ {
u32 scratch_addr = u32 scratch_addr = i915_scratch_offset(rq->i915) + 2 * CACHELINE_BYTES;
i915_ggtt_offset(rq->engine->scratch) + 2 * CACHELINE_BYTES;
u32 *cs, flags = 0; u32 *cs, flags = 0;
int ret; int ret;
...@@ -316,8 +321,7 @@ gen7_render_ring_cs_stall_wa(struct i915_request *rq) ...@@ -316,8 +321,7 @@ gen7_render_ring_cs_stall_wa(struct i915_request *rq)
static int static int
gen7_render_ring_flush(struct i915_request *rq, u32 mode) gen7_render_ring_flush(struct i915_request *rq, u32 mode)
{ {
u32 scratch_addr = u32 scratch_addr = i915_scratch_offset(rq->i915) + 2 * CACHELINE_BYTES;
i915_ggtt_offset(rq->engine->scratch) + 2 * CACHELINE_BYTES;
u32 *cs, flags = 0; u32 *cs, flags = 0;
/* /*
...@@ -971,7 +975,7 @@ i965_emit_bb_start(struct i915_request *rq, ...@@ -971,7 +975,7 @@ i965_emit_bb_start(struct i915_request *rq,
} }
/* Just userspace ABI convention to limit the wa batch bo to a resonable size */ /* Just userspace ABI convention to limit the wa batch bo to a resonable size */
#define I830_BATCH_LIMIT (256*1024) #define I830_BATCH_LIMIT SZ_256K
#define I830_TLB_ENTRIES (2) #define I830_TLB_ENTRIES (2)
#define I830_WA_SIZE max(I830_TLB_ENTRIES*4096, I830_BATCH_LIMIT) #define I830_WA_SIZE max(I830_TLB_ENTRIES*4096, I830_BATCH_LIMIT)
static int static int
...@@ -979,7 +983,9 @@ i830_emit_bb_start(struct i915_request *rq, ...@@ -979,7 +983,9 @@ i830_emit_bb_start(struct i915_request *rq,
u64 offset, u32 len, u64 offset, u32 len,
unsigned int dispatch_flags) unsigned int dispatch_flags)
{ {
u32 *cs, cs_offset = i915_ggtt_offset(rq->engine->scratch); u32 *cs, cs_offset = i915_scratch_offset(rq->i915);
GEM_BUG_ON(rq->i915->gt.scratch->size < I830_WA_SIZE);
cs = intel_ring_begin(rq, 6); cs = intel_ring_begin(rq, 6);
if (IS_ERR(cs)) if (IS_ERR(cs))
...@@ -1437,7 +1443,6 @@ static int intel_init_ring_buffer(struct intel_engine_cs *engine) ...@@ -1437,7 +1443,6 @@ static int intel_init_ring_buffer(struct intel_engine_cs *engine)
{ {
struct i915_timeline *timeline; struct i915_timeline *timeline;
struct intel_ring *ring; struct intel_ring *ring;
unsigned int size;
int err; int err;
intel_engine_setup_common(engine); intel_engine_setup_common(engine);
...@@ -1462,21 +1467,12 @@ static int intel_init_ring_buffer(struct intel_engine_cs *engine) ...@@ -1462,21 +1467,12 @@ static int intel_init_ring_buffer(struct intel_engine_cs *engine)
GEM_BUG_ON(engine->buffer); GEM_BUG_ON(engine->buffer);
engine->buffer = ring; engine->buffer = ring;
size = PAGE_SIZE;
if (HAS_BROKEN_CS_TLB(engine->i915))
size = I830_WA_SIZE;
err = intel_engine_create_scratch(engine, size);
if (err)
goto err_unpin;
err = intel_engine_init_common(engine); err = intel_engine_init_common(engine);
if (err) if (err)
goto err_scratch; goto err_unpin;
return 0; return 0;
err_scratch:
intel_engine_cleanup_scratch(engine);
err_unpin: err_unpin:
intel_ring_unpin(ring); intel_ring_unpin(ring);
err_ring: err_ring:
...@@ -1550,7 +1546,7 @@ static int flush_pd_dir(struct i915_request *rq) ...@@ -1550,7 +1546,7 @@ static int flush_pd_dir(struct i915_request *rq)
/* Stall until the page table load is complete */ /* Stall until the page table load is complete */
*cs++ = MI_STORE_REGISTER_MEM | MI_SRM_LRM_GLOBAL_GTT; *cs++ = MI_STORE_REGISTER_MEM | MI_SRM_LRM_GLOBAL_GTT;
*cs++ = i915_mmio_reg_offset(RING_PP_DIR_BASE(engine)); *cs++ = i915_mmio_reg_offset(RING_PP_DIR_BASE(engine));
*cs++ = i915_ggtt_offset(engine->scratch); *cs++ = i915_scratch_offset(rq->i915);
*cs++ = MI_NOOP; *cs++ = MI_NOOP;
intel_ring_advance(rq, cs); intel_ring_advance(rq, cs);
...@@ -1659,7 +1655,7 @@ static inline int mi_set_context(struct i915_request *rq, u32 flags) ...@@ -1659,7 +1655,7 @@ static inline int mi_set_context(struct i915_request *rq, u32 flags)
/* Insert a delay before the next switch! */ /* Insert a delay before the next switch! */
*cs++ = MI_STORE_REGISTER_MEM | MI_SRM_LRM_GLOBAL_GTT; *cs++ = MI_STORE_REGISTER_MEM | MI_SRM_LRM_GLOBAL_GTT;
*cs++ = i915_mmio_reg_offset(last_reg); *cs++ = i915_mmio_reg_offset(last_reg);
*cs++ = i915_ggtt_offset(engine->scratch); *cs++ = i915_scratch_offset(rq->i915);
*cs++ = MI_NOOP; *cs++ = MI_NOOP;
} }
*cs++ = MI_ARB_ON_OFF | MI_ARB_ENABLE; *cs++ = MI_ARB_ON_OFF | MI_ARB_ENABLE;
......
...@@ -442,7 +442,6 @@ struct intel_engine_cs { ...@@ -442,7 +442,6 @@ struct intel_engine_cs {
struct intel_hw_status_page status_page; struct intel_hw_status_page status_page;
struct i915_ctx_workarounds wa_ctx; struct i915_ctx_workarounds wa_ctx;
struct i915_wa_list wa_list; struct i915_wa_list wa_list;
struct i915_vma *scratch;
u32 irq_keep_mask; /* always keep these interrupts */ u32 irq_keep_mask; /* always keep these interrupts */
u32 irq_enable_mask; /* bitmask to enable ring interrupt */ u32 irq_enable_mask; /* bitmask to enable ring interrupt */
...@@ -900,10 +899,6 @@ void intel_engine_setup_common(struct intel_engine_cs *engine); ...@@ -900,10 +899,6 @@ void intel_engine_setup_common(struct intel_engine_cs *engine);
int intel_engine_init_common(struct intel_engine_cs *engine); int intel_engine_init_common(struct intel_engine_cs *engine);
void intel_engine_cleanup_common(struct intel_engine_cs *engine); void intel_engine_cleanup_common(struct intel_engine_cs *engine);
int intel_engine_create_scratch(struct intel_engine_cs *engine,
unsigned int size);
void intel_engine_cleanup_scratch(struct intel_engine_cs *engine);
int intel_init_render_ring_buffer(struct intel_engine_cs *engine); int intel_init_render_ring_buffer(struct intel_engine_cs *engine);
int intel_init_bsd_ring_buffer(struct intel_engine_cs *engine); int intel_init_bsd_ring_buffer(struct intel_engine_cs *engine);
int intel_init_blt_ring_buffer(struct intel_engine_cs *engine); int intel_init_blt_ring_buffer(struct intel_engine_cs *engine);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment