提交 59b449d5 编写于 作者: O Oscar Mateo 提交者: Chris Wilson

drm/i915: Split out functions for different kinds of workarounds

There are different kind of workarounds (those that modify registers that
live in the context image, those that modify global registers, those that
whitelist registers, etc...) and they have different requirements in terms
of where they are applied and how. Also, by splitting them apart, it should
be easier to decide where a new workaround should go.

v2:
  - Add multiple MISSING_CASE
  - Rebased

v3:
  - Rename mmio_workarounds to gt_workarounds (Chris, Mika)
  - Create empty placeholders for BDW and CHV GT WAs
  - Rebased

v4: Rebased

v5:
 - Rebased
 - FORCE_TO_NONPRIV register exists since BDW, so make a path
   for it to achieve universality, even if empty (Chris)
Signed-off-by: NOscar Mateo <oscar.mateo@intel.com>
Cc: Mika Kuoppala <mika.kuoppala@linux.intel.com>
Cc: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: NChris Wilson <chris@chris-wilson.co.uk>
[ickle: appease checkpatch]
Signed-off-by: NChris Wilson <chris@chris-wilson.co.uk>
Link: https://patchwork.freedesktop.org/patch/msgid/1523376767-18480-2-git-send-email-oscar.mateo@intel.com
上级 7d3c425f
...@@ -35,6 +35,7 @@ ...@@ -35,6 +35,7 @@
#include "intel_drv.h" #include "intel_drv.h"
#include "intel_frontbuffer.h" #include "intel_frontbuffer.h"
#include "intel_mocs.h" #include "intel_mocs.h"
#include "intel_workarounds.h"
#include "i915_gemfs.h" #include "i915_gemfs.h"
#include <linux/dma-fence-array.h> #include <linux/dma-fence-array.h>
#include <linux/kthread.h> #include <linux/kthread.h>
...@@ -5191,6 +5192,8 @@ int i915_gem_init_hw(struct drm_i915_private *dev_priv) ...@@ -5191,6 +5192,8 @@ int i915_gem_init_hw(struct drm_i915_private *dev_priv)
} }
} }
intel_gt_workarounds_apply(dev_priv);
i915_gem_init_swizzling(dev_priv); i915_gem_init_swizzling(dev_priv);
/* /*
......
...@@ -90,6 +90,7 @@ ...@@ -90,6 +90,7 @@
#include <drm/i915_drm.h> #include <drm/i915_drm.h>
#include "i915_drv.h" #include "i915_drv.h"
#include "i915_trace.h" #include "i915_trace.h"
#include "intel_workarounds.h"
#define ALL_L3_SLICES(dev) (1 << NUM_L3_SLICES(dev)) - 1 #define ALL_L3_SLICES(dev) (1 << NUM_L3_SLICES(dev)) - 1
...@@ -459,11 +460,16 @@ static bool needs_preempt_context(struct drm_i915_private *i915) ...@@ -459,11 +460,16 @@ static bool needs_preempt_context(struct drm_i915_private *i915)
int i915_gem_contexts_init(struct drm_i915_private *dev_priv) int i915_gem_contexts_init(struct drm_i915_private *dev_priv)
{ {
struct i915_gem_context *ctx; struct i915_gem_context *ctx;
int ret;
/* Reassure ourselves we are only called once */ /* Reassure ourselves we are only called once */
GEM_BUG_ON(dev_priv->kernel_context); GEM_BUG_ON(dev_priv->kernel_context);
GEM_BUG_ON(dev_priv->preempt_context); GEM_BUG_ON(dev_priv->preempt_context);
ret = intel_ctx_workarounds_init(dev_priv);
if (ret)
return ret;
INIT_LIST_HEAD(&dev_priv->contexts.list); INIT_LIST_HEAD(&dev_priv->contexts.list);
INIT_WORK(&dev_priv->contexts.free_work, contexts_free_worker); INIT_WORK(&dev_priv->contexts.free_work, contexts_free_worker);
init_llist_head(&dev_priv->contexts.free_list); init_llist_head(&dev_priv->contexts.free_list);
......
...@@ -1744,6 +1744,10 @@ static int gen8_init_render_ring(struct intel_engine_cs *engine) ...@@ -1744,6 +1744,10 @@ static int gen8_init_render_ring(struct intel_engine_cs *engine)
if (ret) if (ret)
return ret; return ret;
ret = intel_whitelist_workarounds_apply(engine);
if (ret)
return ret;
/* We need to disable the AsyncFlip performance optimisations in order /* We need to disable the AsyncFlip performance optimisations in order
* to use MI_WAIT_FOR_EVENT within the CS. It should already be * to use MI_WAIT_FOR_EVENT within the CS. It should already be
* programmed to '1' on all products. * programmed to '1' on all products.
...@@ -1754,7 +1758,7 @@ static int gen8_init_render_ring(struct intel_engine_cs *engine) ...@@ -1754,7 +1758,7 @@ static int gen8_init_render_ring(struct intel_engine_cs *engine)
I915_WRITE(INSTPM, _MASKED_BIT_ENABLE(INSTPM_FORCE_ORDERING)); I915_WRITE(INSTPM, _MASKED_BIT_ENABLE(INSTPM_FORCE_ORDERING));
return init_workarounds_ring(engine); return 0;
} }
static int gen9_init_render_ring(struct intel_engine_cs *engine) static int gen9_init_render_ring(struct intel_engine_cs *engine)
...@@ -1765,7 +1769,11 @@ static int gen9_init_render_ring(struct intel_engine_cs *engine) ...@@ -1765,7 +1769,11 @@ static int gen9_init_render_ring(struct intel_engine_cs *engine)
if (ret) if (ret)
return ret; return ret;
return init_workarounds_ring(engine); ret = intel_whitelist_workarounds_apply(engine);
if (ret)
return ret;
return 0;
} }
static void reset_common_ring(struct intel_engine_cs *engine, static void reset_common_ring(struct intel_engine_cs *engine,
...@@ -2090,7 +2098,7 @@ static int gen8_init_rcs_context(struct i915_request *rq) ...@@ -2090,7 +2098,7 @@ static int gen8_init_rcs_context(struct i915_request *rq)
{ {
int ret; int ret;
ret = intel_ring_workarounds_emit(rq); ret = intel_ctx_workarounds_emit(rq);
if (ret) if (ret)
return ret; return ret;
......
...@@ -600,7 +600,7 @@ static int intel_rcs_ctx_init(struct i915_request *rq) ...@@ -600,7 +600,7 @@ static int intel_rcs_ctx_init(struct i915_request *rq)
{ {
int ret; int ret;
ret = intel_ring_workarounds_emit(rq); ret = intel_ctx_workarounds_emit(rq);
if (ret != 0) if (ret != 0)
return ret; return ret;
...@@ -618,6 +618,10 @@ static int init_render_ring(struct intel_engine_cs *engine) ...@@ -618,6 +618,10 @@ static int init_render_ring(struct intel_engine_cs *engine)
if (ret) if (ret)
return ret; return ret;
ret = intel_whitelist_workarounds_apply(engine);
if (ret)
return ret;
/* WaTimedSingleVertexDispatch:cl,bw,ctg,elk,ilk,snb */ /* WaTimedSingleVertexDispatch:cl,bw,ctg,elk,ilk,snb */
if (IS_GEN(dev_priv, 4, 6)) if (IS_GEN(dev_priv, 4, 6))
I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH)); I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH));
...@@ -659,7 +663,7 @@ static int init_render_ring(struct intel_engine_cs *engine) ...@@ -659,7 +663,7 @@ static int init_render_ring(struct intel_engine_cs *engine)
if (INTEL_GEN(dev_priv) >= 6) if (INTEL_GEN(dev_priv) >= 6)
I915_WRITE_IMR(engine, ~engine->irq_keep_mask); I915_WRITE_IMR(engine, ~engine->irq_keep_mask);
return init_workarounds_ring(engine); return 0;
} }
static u32 *gen6_signal(struct i915_request *rq, u32 *cs) static u32 *gen6_signal(struct i915_request *rq, u32 *cs)
......
...@@ -81,27 +81,8 @@ static int wa_add(struct drm_i915_private *dev_priv, ...@@ -81,27 +81,8 @@ static int wa_add(struct drm_i915_private *dev_priv,
#define WA_SET_FIELD_MASKED(addr, mask, value) \ #define WA_SET_FIELD_MASKED(addr, mask, value) \
WA_REG(addr, (mask), _MASKED_FIELD(mask, value)) WA_REG(addr, (mask), _MASKED_FIELD(mask, value))
static int wa_ring_whitelist_reg(struct intel_engine_cs *engine, static int gen8_ctx_workarounds_init(struct drm_i915_private *dev_priv)
i915_reg_t reg)
{ {
struct drm_i915_private *dev_priv = engine->i915;
struct i915_workarounds *wa = &dev_priv->workarounds;
const unsigned int index = wa->hw_whitelist_count[engine->id];
if (WARN_ON(index >= RING_MAX_NONPRIV_SLOTS))
return -EINVAL;
I915_WRITE(RING_FORCE_TO_NONPRIV(engine->mmio_base, index),
i915_mmio_reg_offset(reg));
wa->hw_whitelist_count[engine->id]++;
return 0;
}
static int gen8_init_workarounds(struct intel_engine_cs *engine)
{
struct drm_i915_private *dev_priv = engine->i915;
WA_SET_BIT_MASKED(INSTPM, INSTPM_FORCE_ORDERING); WA_SET_BIT_MASKED(INSTPM, INSTPM_FORCE_ORDERING);
/* WaDisableAsyncFlipPerfMode:bdw,chv */ /* WaDisableAsyncFlipPerfMode:bdw,chv */
...@@ -149,12 +130,11 @@ static int gen8_init_workarounds(struct intel_engine_cs *engine) ...@@ -149,12 +130,11 @@ static int gen8_init_workarounds(struct intel_engine_cs *engine)
return 0; return 0;
} }
static int bdw_init_workarounds(struct intel_engine_cs *engine) static int bdw_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
ret = gen8_init_workarounds(engine); ret = gen8_ctx_workarounds_init(dev_priv);
if (ret) if (ret)
return ret; return ret;
...@@ -181,12 +161,11 @@ static int bdw_init_workarounds(struct intel_engine_cs *engine) ...@@ -181,12 +161,11 @@ static int bdw_init_workarounds(struct intel_engine_cs *engine)
return 0; return 0;
} }
static int chv_init_workarounds(struct intel_engine_cs *engine) static int chv_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
ret = gen8_init_workarounds(engine); ret = gen8_ctx_workarounds_init(dev_priv);
if (ret) if (ret)
return ret; return ret;
...@@ -199,25 +178,8 @@ static int chv_init_workarounds(struct intel_engine_cs *engine) ...@@ -199,25 +178,8 @@ static int chv_init_workarounds(struct intel_engine_cs *engine)
return 0; return 0;
} }
static int gen9_init_workarounds(struct intel_engine_cs *engine) static int gen9_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret;
/* WaConextSwitchWithConcurrentTLBInvalidate:skl,bxt,kbl,glk,cfl */
I915_WRITE(GEN9_CSFE_CHICKEN1_RCS,
_MASKED_BIT_ENABLE(GEN9_PREEMPT_GPGPU_SYNC_SWITCH_DISABLE));
/* WaEnableLbsSlaRetryTimerDecrement:skl,bxt,kbl,glk,cfl */
I915_WRITE(BDW_SCRATCH1,
I915_READ(BDW_SCRATCH1) |
GEN9_LBS_SLA_RETRY_TIMER_DECREMENT_ENABLE);
/* WaDisableKillLogic:bxt,skl,kbl */
if (!IS_COFFEELAKE(dev_priv))
I915_WRITE(GAM_ECOCHK,
I915_READ(GAM_ECOCHK) | ECOCHK_DIS_TLB);
if (HAS_LLC(dev_priv)) { if (HAS_LLC(dev_priv)) {
/* WaCompressedResourceSamplerPbeMediaNewHashMode:skl,kbl /* WaCompressedResourceSamplerPbeMediaNewHashMode:skl,kbl
* *
...@@ -228,11 +190,6 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine) ...@@ -228,11 +190,6 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine)
GEN9_PBE_COMPRESSED_HASH_SELECTION); GEN9_PBE_COMPRESSED_HASH_SELECTION);
WA_SET_BIT_MASKED(GEN9_HALF_SLICE_CHICKEN7, WA_SET_BIT_MASKED(GEN9_HALF_SLICE_CHICKEN7,
GEN9_SAMPLER_HASH_COMPRESSED_READ_ADDR); GEN9_SAMPLER_HASH_COMPRESSED_READ_ADDR);
I915_WRITE(MMCD_MISC_CTRL,
I915_READ(MMCD_MISC_CTRL) |
MMCD_PCLA |
MMCD_HOTSPOT_EN);
} }
/* WaClearFlowControlGpgpuContextSave:skl,bxt,kbl,glk,cfl */ /* WaClearFlowControlGpgpuContextSave:skl,bxt,kbl,glk,cfl */
...@@ -284,10 +241,6 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine) ...@@ -284,10 +241,6 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine)
WA_SET_BIT_MASKED(HDC_CHICKEN0, WA_SET_BIT_MASKED(HDC_CHICKEN0,
HDC_FORCE_NON_COHERENT); HDC_FORCE_NON_COHERENT);
/* WaDisableHDCInvalidation:skl,bxt,kbl,cfl */
I915_WRITE(GAM_ECOCHK,
I915_READ(GAM_ECOCHK) | BDW_DISABLE_HDC_INVALIDATION);
/* WaDisableSamplerPowerBypassForSOPingPong:skl,bxt,kbl,cfl */ /* WaDisableSamplerPowerBypassForSOPingPong:skl,bxt,kbl,cfl */
if (IS_SKYLAKE(dev_priv) || if (IS_SKYLAKE(dev_priv) ||
IS_KABYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) ||
...@@ -298,19 +251,6 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine) ...@@ -298,19 +251,6 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine)
/* WaDisableSTUnitPowerOptimization:skl,bxt,kbl,glk,cfl */ /* WaDisableSTUnitPowerOptimization:skl,bxt,kbl,glk,cfl */
WA_SET_BIT_MASKED(HALF_SLICE_CHICKEN2, GEN8_ST_PO_DISABLE); WA_SET_BIT_MASKED(HALF_SLICE_CHICKEN2, GEN8_ST_PO_DISABLE);
/* WaProgramL3SqcReg1DefaultForPerf:bxt,glk */
if (IS_GEN9_LP(dev_priv)) {
u32 val = I915_READ(GEN8_L3SQCREG1);
val &= ~L3_PRIO_CREDITS_MASK;
val |= L3_GENERAL_PRIO_CREDITS(62) | L3_HIGH_PRIO_CREDITS(2);
I915_WRITE(GEN8_L3SQCREG1, val);
}
/* WaOCLCoherentLineFlush:skl,bxt,kbl,cfl */
I915_WRITE(GEN8_L3SQCREG4,
I915_READ(GEN8_L3SQCREG4) | GEN8_LQSC_FLUSH_COHERENT_LINES);
/* /*
* Supporting preemption with fine-granularity requires changes in the * Supporting preemption with fine-granularity requires changes in the
* batch buffer programming. Since we can't break old userspace, we * batch buffer programming. Since we can't break old userspace, we
...@@ -330,29 +270,11 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine) ...@@ -330,29 +270,11 @@ static int gen9_init_workarounds(struct intel_engine_cs *engine)
GEN9_PREEMPT_GPGPU_LEVEL_MASK, GEN9_PREEMPT_GPGPU_LEVEL_MASK,
GEN9_PREEMPT_GPGPU_COMMAND_LEVEL); GEN9_PREEMPT_GPGPU_COMMAND_LEVEL);
/* WaVFEStateAfterPipeControlwithMediaStateClear:skl,bxt,glk,cfl */
ret = wa_ring_whitelist_reg(engine, GEN9_CTX_PREEMPT_REG);
if (ret)
return ret;
/* WaEnablePreemptionGranularityControlByUMD:skl,bxt,kbl,cfl,[cnl] */
I915_WRITE(GEN7_FF_SLICE_CS_CHICKEN1,
_MASKED_BIT_ENABLE(GEN9_FFSC_PERCTX_PREEMPT_CTRL));
ret = wa_ring_whitelist_reg(engine, GEN8_CS_CHICKEN1);
if (ret)
return ret;
/* WaAllowUMDToModifyHDCChicken1:skl,bxt,kbl,glk,cfl */
ret = wa_ring_whitelist_reg(engine, GEN8_HDC_CHICKEN1);
if (ret)
return ret;
return 0; return 0;
} }
static int skl_tune_iz_hashing(struct intel_engine_cs *engine) static int skl_tune_iz_hashing(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
u8 vals[3] = { 0, 0, 0 }; u8 vals[3] = { 0, 0, 0 };
unsigned int i; unsigned int i;
...@@ -391,77 +313,97 @@ static int skl_tune_iz_hashing(struct intel_engine_cs *engine) ...@@ -391,77 +313,97 @@ static int skl_tune_iz_hashing(struct intel_engine_cs *engine)
return 0; return 0;
} }
static int skl_init_workarounds(struct intel_engine_cs *engine) static int skl_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
ret = gen9_init_workarounds(engine); ret = gen9_ctx_workarounds_init(dev_priv);
if (ret) if (ret)
return ret; return ret;
/* WaEnableGapsTsvCreditFix:skl */ return skl_tune_iz_hashing(dev_priv);
I915_WRITE(GEN8_GARBCNTL, }
I915_READ(GEN8_GARBCNTL) | GEN9_GAPS_TSV_CREDIT_DISABLE);
/* WaDisableGafsUnitClkGating:skl */
I915_WRITE(GEN7_UCGCTL4,
I915_READ(GEN7_UCGCTL4) | GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
/* WaInPlaceDecompressionHang:skl */ static int bxt_ctx_workarounds_init(struct drm_i915_private *dev_priv)
if (IS_SKL_REVID(dev_priv, SKL_REVID_H0, REVID_FOREVER)) {
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA, int ret;
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
/* WaDisableLSQCROPERFforOCL:skl */ ret = gen9_ctx_workarounds_init(dev_priv);
ret = wa_ring_whitelist_reg(engine, GEN8_L3SQCREG4);
if (ret) if (ret)
return ret; return ret;
return skl_tune_iz_hashing(engine); /* WaDisableThreadStallDopClockGating:bxt */
WA_SET_BIT_MASKED(GEN8_ROW_CHICKEN,
STALL_DOP_GATING_DISABLE);
/* WaToEnableHwFixForPushConstHWBug:bxt */
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2,
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
return 0;
} }
static int bxt_init_workarounds(struct intel_engine_cs *engine) static int kbl_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
ret = gen9_init_workarounds(engine); ret = gen9_ctx_workarounds_init(dev_priv);
if (ret) if (ret)
return ret; return ret;
/* WaDisableThreadStallDopClockGating:bxt */ /* WaDisableFenceDestinationToSLM:kbl (pre-prod) */
WA_SET_BIT_MASKED(GEN8_ROW_CHICKEN, if (IS_KBL_REVID(dev_priv, KBL_REVID_A0, KBL_REVID_A0))
STALL_DOP_GATING_DISABLE); WA_SET_BIT_MASKED(HDC_CHICKEN0,
HDC_FENCE_DEST_SLM_DISABLE);
/* WaDisablePooledEuLoadBalancingFix:bxt */
I915_WRITE(FF_SLICE_CS_CHICKEN2,
_MASKED_BIT_ENABLE(GEN9_POOLED_EU_LOAD_BALANCING_FIX_DISABLE));
/* WaToEnableHwFixForPushConstHWBug:bxt */ /* WaToEnableHwFixForPushConstHWBug:kbl */
if (IS_KBL_REVID(dev_priv, KBL_REVID_C0, REVID_FOREVER))
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2, WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2,
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION); GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
/* WaInPlaceDecompressionHang:bxt */ /* WaDisableSbeCacheDispatchPortSharing:kbl */
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA, WA_SET_BIT_MASKED(GEN7_HALF_SLICE_CHICKEN1,
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) | GEN7_SBE_SS_CACHE_DISPATCH_PORT_SHARING_DISABLE);
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
return 0; return 0;
} }
static int cnl_init_workarounds(struct intel_engine_cs *engine) static int glk_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
/* WaDisableI2mCycleOnWRPort:cnl (pre-prod) */ ret = gen9_ctx_workarounds_init(dev_priv);
if (IS_CNL_REVID(dev_priv, CNL_REVID_B0, CNL_REVID_B0)) if (ret)
I915_WRITE(GAMT_CHKN_BIT_REG, return ret;
I915_READ(GAMT_CHKN_BIT_REG) |
GAMT_CHKN_DISABLE_I2M_CYCLE_ON_WR_PORT); /* WaToEnableHwFixForPushConstHWBug:glk */
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2,
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
return 0;
}
static int cfl_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{
int ret;
ret = gen9_ctx_workarounds_init(dev_priv);
if (ret)
return ret;
/* WaToEnableHwFixForPushConstHWBug:cfl */
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2,
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
/* WaDisableSbeCacheDispatchPortSharing:cfl */
WA_SET_BIT_MASKED(GEN7_HALF_SLICE_CHICKEN1,
GEN7_SBE_SS_CACHE_DISPATCH_PORT_SHARING_DISABLE);
return 0;
}
static int cnl_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{
/* WaForceContextSaveRestoreNonCoherent:cnl */ /* WaForceContextSaveRestoreNonCoherent:cnl */
WA_SET_BIT_MASKED(CNL_HDC_CHICKEN0, WA_SET_BIT_MASKED(CNL_HDC_CHICKEN0,
HDC_FORCE_CONTEXT_SAVE_RESTORE_NON_COHERENT); HDC_FORCE_CONTEXT_SAVE_RESTORE_NON_COHERENT);
...@@ -479,15 +421,10 @@ static int cnl_init_workarounds(struct intel_engine_cs *engine) ...@@ -479,15 +421,10 @@ static int cnl_init_workarounds(struct intel_engine_cs *engine)
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2, WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2,
GEN8_CSC2_SBE_VUE_CACHE_CONSERVATIVE); GEN8_CSC2_SBE_VUE_CACHE_CONSERVATIVE);
/* WaInPlaceDecompressionHang:cnl */
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA,
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
/* WaPushConstantDereferenceHoldDisable:cnl */ /* WaPushConstantDereferenceHoldDisable:cnl */
WA_SET_BIT_MASKED(GEN7_ROW_CHICKEN2, PUSH_CONSTANT_DEREF_DISABLE); WA_SET_BIT_MASKED(GEN7_ROW_CHICKEN2, PUSH_CONSTANT_DEREF_DISABLE);
/* FtrEnableFastAnisoL1BankingFix: cnl */ /* FtrEnableFastAnisoL1BankingFix:cnl */
WA_SET_BIT_MASKED(HALF_SLICE_CHICKEN3, CNL_FAST_ANISO_L1_BANKING_FIX); WA_SET_BIT_MASKED(HALF_SLICE_CHICKEN3, CNL_FAST_ANISO_L1_BANKING_FIX);
/* WaDisable3DMidCmdPreemption:cnl */ /* WaDisable3DMidCmdPreemption:cnl */
...@@ -498,28 +435,173 @@ static int cnl_init_workarounds(struct intel_engine_cs *engine) ...@@ -498,28 +435,173 @@ static int cnl_init_workarounds(struct intel_engine_cs *engine)
GEN9_PREEMPT_GPGPU_LEVEL_MASK, GEN9_PREEMPT_GPGPU_LEVEL_MASK,
GEN9_PREEMPT_GPGPU_COMMAND_LEVEL); GEN9_PREEMPT_GPGPU_COMMAND_LEVEL);
/* WaEnablePreemptionGranularityControlByUMD:cnl */
I915_WRITE(GEN7_FF_SLICE_CS_CHICKEN1,
_MASKED_BIT_ENABLE(GEN9_FFSC_PERCTX_PREEMPT_CTRL));
ret = wa_ring_whitelist_reg(engine, GEN8_CS_CHICKEN1);
if (ret)
return ret;
/* WaDisableEarlyEOT:cnl */ /* WaDisableEarlyEOT:cnl */
WA_SET_BIT_MASKED(GEN8_ROW_CHICKEN, DISABLE_EARLY_EOT); WA_SET_BIT_MASKED(GEN8_ROW_CHICKEN, DISABLE_EARLY_EOT);
return 0; return 0;
} }
static int kbl_init_workarounds(struct intel_engine_cs *engine) int intel_ctx_workarounds_init(struct drm_i915_private *dev_priv)
{ {
struct drm_i915_private *dev_priv = engine->i915; int err = 0;
int ret;
ret = gen9_init_workarounds(engine); dev_priv->workarounds.count = 0;
if (INTEL_GEN(dev_priv) < 8)
err = 0;
else if (IS_BROADWELL(dev_priv))
err = bdw_ctx_workarounds_init(dev_priv);
else if (IS_CHERRYVIEW(dev_priv))
err = chv_ctx_workarounds_init(dev_priv);
else if (IS_SKYLAKE(dev_priv))
err = skl_ctx_workarounds_init(dev_priv);
else if (IS_BROXTON(dev_priv))
err = bxt_ctx_workarounds_init(dev_priv);
else if (IS_KABYLAKE(dev_priv))
err = kbl_ctx_workarounds_init(dev_priv);
else if (IS_GEMINILAKE(dev_priv))
err = glk_ctx_workarounds_init(dev_priv);
else if (IS_COFFEELAKE(dev_priv))
err = cfl_ctx_workarounds_init(dev_priv);
else if (IS_CANNONLAKE(dev_priv))
err = cnl_ctx_workarounds_init(dev_priv);
else
MISSING_CASE(INTEL_GEN(dev_priv));
if (err)
return err;
DRM_DEBUG_DRIVER("Number of context specific w/a: %d\n",
dev_priv->workarounds.count);
return 0;
}
int intel_ctx_workarounds_emit(struct i915_request *rq)
{
struct i915_workarounds *w = &rq->i915->workarounds;
u32 *cs;
int ret, i;
if (w->count == 0)
return 0;
ret = rq->engine->emit_flush(rq, EMIT_BARRIER);
if (ret) if (ret)
return ret; return ret;
cs = intel_ring_begin(rq, (w->count * 2 + 2));
if (IS_ERR(cs))
return PTR_ERR(cs);
*cs++ = MI_LOAD_REGISTER_IMM(w->count);
for (i = 0; i < w->count; i++) {
*cs++ = i915_mmio_reg_offset(w->reg[i].addr);
*cs++ = w->reg[i].value;
}
*cs++ = MI_NOOP;
intel_ring_advance(rq, cs);
ret = rq->engine->emit_flush(rq, EMIT_BARRIER);
if (ret)
return ret;
return 0;
}
static void bdw_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
}
static void chv_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
}
static void gen9_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
/* WaContextSwitchWithConcurrentTLBInvalidate:skl,bxt,kbl,glk,cfl */
I915_WRITE(GEN9_CSFE_CHICKEN1_RCS,
_MASKED_BIT_ENABLE(GEN9_PREEMPT_GPGPU_SYNC_SWITCH_DISABLE));
/* WaEnableLbsSlaRetryTimerDecrement:skl,bxt,kbl,glk,cfl */
I915_WRITE(BDW_SCRATCH1, I915_READ(BDW_SCRATCH1) |
GEN9_LBS_SLA_RETRY_TIMER_DECREMENT_ENABLE);
/* WaDisableKillLogic:bxt,skl,kbl */
if (!IS_COFFEELAKE(dev_priv))
I915_WRITE(GAM_ECOCHK, I915_READ(GAM_ECOCHK) |
ECOCHK_DIS_TLB);
if (HAS_LLC(dev_priv)) {
/* WaCompressedResourceSamplerPbeMediaNewHashMode:skl,kbl
*
* Must match Display Engine. See
* WaCompressedResourceDisplayNewHashMode.
*/
I915_WRITE(MMCD_MISC_CTRL,
I915_READ(MMCD_MISC_CTRL) |
MMCD_PCLA |
MMCD_HOTSPOT_EN);
}
/* WaDisableHDCInvalidation:skl,bxt,kbl,cfl */
I915_WRITE(GAM_ECOCHK, I915_READ(GAM_ECOCHK) |
BDW_DISABLE_HDC_INVALIDATION);
/* WaProgramL3SqcReg1DefaultForPerf:bxt,glk */
if (IS_GEN9_LP(dev_priv)) {
u32 val = I915_READ(GEN8_L3SQCREG1);
val &= ~L3_PRIO_CREDITS_MASK;
val |= L3_GENERAL_PRIO_CREDITS(62) | L3_HIGH_PRIO_CREDITS(2);
I915_WRITE(GEN8_L3SQCREG1, val);
}
/* WaOCLCoherentLineFlush:skl,bxt,kbl,cfl */
I915_WRITE(GEN8_L3SQCREG4,
I915_READ(GEN8_L3SQCREG4) | GEN8_LQSC_FLUSH_COHERENT_LINES);
/* WaEnablePreemptionGranularityControlByUMD:skl,bxt,kbl,cfl,[cnl] */
I915_WRITE(GEN7_FF_SLICE_CS_CHICKEN1,
_MASKED_BIT_ENABLE(GEN9_FFSC_PERCTX_PREEMPT_CTRL));
}
static void skl_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
gen9_gt_workarounds_apply(dev_priv);
/* WaEnableGapsTsvCreditFix:skl */
I915_WRITE(GEN8_GARBCNTL,
I915_READ(GEN8_GARBCNTL) | GEN9_GAPS_TSV_CREDIT_DISABLE);
/* WaDisableGafsUnitClkGating:skl */
I915_WRITE(GEN7_UCGCTL4,
I915_READ(GEN7_UCGCTL4) | GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
/* WaInPlaceDecompressionHang:skl */
if (IS_SKL_REVID(dev_priv, SKL_REVID_H0, REVID_FOREVER))
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA,
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
}
static void bxt_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
gen9_gt_workarounds_apply(dev_priv);
/* WaDisablePooledEuLoadBalancingFix:bxt */
I915_WRITE(FF_SLICE_CS_CHICKEN2,
_MASKED_BIT_ENABLE(GEN9_POOLED_EU_LOAD_BALANCING_FIX_DISABLE));
/* WaInPlaceDecompressionHang:bxt */
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA,
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
}
static void kbl_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
gen9_gt_workarounds_apply(dev_priv);
/* WaEnableGapsTsvCreditFix:kbl */ /* WaEnableGapsTsvCreditFix:kbl */
I915_WRITE(GEN8_GARBCNTL, I915_WRITE(GEN8_GARBCNTL,
I915_READ(GEN8_GARBCNTL) | GEN9_GAPS_TSV_CREDIT_DISABLE); I915_READ(GEN8_GARBCNTL) | GEN9_GAPS_TSV_CREDIT_DISABLE);
...@@ -530,30 +612,139 @@ static int kbl_init_workarounds(struct intel_engine_cs *engine) ...@@ -530,30 +612,139 @@ static int kbl_init_workarounds(struct intel_engine_cs *engine)
I915_READ(GAMT_CHKN_BIT_REG) | I915_READ(GAMT_CHKN_BIT_REG) |
GAMT_CHKN_DISABLE_DYNAMIC_CREDIT_SHARING); GAMT_CHKN_DISABLE_DYNAMIC_CREDIT_SHARING);
/* WaDisableFenceDestinationToSLM:kbl (pre-prod) */ /* WaDisableGafsUnitClkGating:kbl */
if (IS_KBL_REVID(dev_priv, KBL_REVID_A0, KBL_REVID_A0)) I915_WRITE(GEN7_UCGCTL4,
WA_SET_BIT_MASKED(HDC_CHICKEN0, I915_READ(GEN7_UCGCTL4) | GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
HDC_FENCE_DEST_SLM_DISABLE);
/* WaToEnableHwFixForPushConstHWBug:kbl */ /* WaInPlaceDecompressionHang:kbl */
if (IS_KBL_REVID(dev_priv, KBL_REVID_C0, REVID_FOREVER)) I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA,
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2, I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION); GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
}
/* WaDisableGafsUnitClkGating:kbl */ static void glk_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
gen9_gt_workarounds_apply(dev_priv);
}
static void cfl_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
gen9_gt_workarounds_apply(dev_priv);
/* WaEnableGapsTsvCreditFix:cfl */
I915_WRITE(GEN8_GARBCNTL,
I915_READ(GEN8_GARBCNTL) | GEN9_GAPS_TSV_CREDIT_DISABLE);
/* WaDisableGafsUnitClkGating:cfl */
I915_WRITE(GEN7_UCGCTL4, I915_WRITE(GEN7_UCGCTL4,
I915_READ(GEN7_UCGCTL4) | GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE); I915_READ(GEN7_UCGCTL4) | GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
/* WaDisableSbeCacheDispatchPortSharing:kbl */ /* WaInPlaceDecompressionHang:cfl */
WA_SET_BIT_MASKED(GEN7_HALF_SLICE_CHICKEN1, I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA,
GEN7_SBE_SS_CACHE_DISPATCH_PORT_SHARING_DISABLE); I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
}
/* WaInPlaceDecompressionHang:kbl */ static void cnl_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
/* WaDisableI2mCycleOnWRPort:cnl (pre-prod) */
if (IS_CNL_REVID(dev_priv, CNL_REVID_B0, CNL_REVID_B0))
I915_WRITE(GAMT_CHKN_BIT_REG,
I915_READ(GAMT_CHKN_BIT_REG) |
GAMT_CHKN_DISABLE_I2M_CYCLE_ON_WR_PORT);
/* WaInPlaceDecompressionHang:cnl */
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA, I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA,
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) | I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS); GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
/* WaDisableLSQCROPERFforOCL:kbl */ /* WaEnablePreemptionGranularityControlByUMD:cnl */
I915_WRITE(GEN7_FF_SLICE_CS_CHICKEN1,
_MASKED_BIT_ENABLE(GEN9_FFSC_PERCTX_PREEMPT_CTRL));
}
void intel_gt_workarounds_apply(struct drm_i915_private *dev_priv)
{
if (INTEL_GEN(dev_priv) < 8)
return;
else if (IS_BROADWELL(dev_priv))
bdw_gt_workarounds_apply(dev_priv);
else if (IS_CHERRYVIEW(dev_priv))
chv_gt_workarounds_apply(dev_priv);
else if (IS_SKYLAKE(dev_priv))
skl_gt_workarounds_apply(dev_priv);
else if (IS_BROXTON(dev_priv))
bxt_gt_workarounds_apply(dev_priv);
else if (IS_KABYLAKE(dev_priv))
kbl_gt_workarounds_apply(dev_priv);
else if (IS_GEMINILAKE(dev_priv))
glk_gt_workarounds_apply(dev_priv);
else if (IS_COFFEELAKE(dev_priv))
cfl_gt_workarounds_apply(dev_priv);
else if (IS_CANNONLAKE(dev_priv))
cnl_gt_workarounds_apply(dev_priv);
else
MISSING_CASE(INTEL_GEN(dev_priv));
}
static int wa_ring_whitelist_reg(struct intel_engine_cs *engine,
i915_reg_t reg)
{
struct drm_i915_private *dev_priv = engine->i915;
struct i915_workarounds *wa = &dev_priv->workarounds;
const unsigned int index = wa->hw_whitelist_count[engine->id];
if (WARN_ON(index >= RING_MAX_NONPRIV_SLOTS))
return -EINVAL;
I915_WRITE(RING_FORCE_TO_NONPRIV(engine->mmio_base, index),
i915_mmio_reg_offset(reg));
wa->hw_whitelist_count[engine->id]++;
return 0;
}
static int bdw_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{
return 0;
}
static int chv_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{
return 0;
}
static int gen9_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{
int ret;
/* WaVFEStateAfterPipeControlwithMediaStateClear:skl,bxt,glk,cfl */
ret = wa_ring_whitelist_reg(engine, GEN9_CTX_PREEMPT_REG);
if (ret)
return ret;
/* WaEnablePreemptionGranularityControlByUMD:skl,bxt,kbl,cfl,[cnl] */
ret = wa_ring_whitelist_reg(engine, GEN8_CS_CHICKEN1);
if (ret)
return ret;
/* WaAllowUMDToModifyHDCChicken1:skl,bxt,kbl,glk,cfl */
ret = wa_ring_whitelist_reg(engine, GEN8_HDC_CHICKEN1);
if (ret)
return ret;
return 0;
}
static int skl_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{
int ret;
ret = gen9_whitelist_workarounds_apply(engine);
if (ret)
return ret;
/* WaDisableLSQCROPERFforOCL:skl */
ret = wa_ring_whitelist_reg(engine, GEN8_L3SQCREG4); ret = wa_ring_whitelist_reg(engine, GEN8_L3SQCREG4);
if (ret) if (ret)
return ret; return ret;
...@@ -561,126 +752,105 @@ static int kbl_init_workarounds(struct intel_engine_cs *engine) ...@@ -561,126 +752,105 @@ static int kbl_init_workarounds(struct intel_engine_cs *engine)
return 0; return 0;
} }
static int glk_init_workarounds(struct intel_engine_cs *engine) static int bxt_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
ret = gen9_init_workarounds(engine); ret = gen9_whitelist_workarounds_apply(engine);
if (ret) if (ret)
return ret; return ret;
/* WA #0862: Userspace has to set "Barrier Mode" to avoid hangs. */ return 0;
ret = wa_ring_whitelist_reg(engine, GEN9_SLICE_COMMON_ECO_CHICKEN1); }
static int kbl_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{
int ret;
ret = gen9_whitelist_workarounds_apply(engine);
if (ret) if (ret)
return ret; return ret;
/* WaToEnableHwFixForPushConstHWBug:glk */ /* WaDisableLSQCROPERFforOCL:kbl */
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2, ret = wa_ring_whitelist_reg(engine, GEN8_L3SQCREG4);
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION); if (ret)
return ret;
return 0; return 0;
} }
static int cfl_init_workarounds(struct intel_engine_cs *engine) static int glk_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{ {
struct drm_i915_private *dev_priv = engine->i915;
int ret; int ret;
ret = gen9_init_workarounds(engine); ret = gen9_whitelist_workarounds_apply(engine);
if (ret) if (ret)
return ret; return ret;
/* WaEnableGapsTsvCreditFix:cfl */ /* WA #0862: Userspace has to set "Barrier Mode" to avoid hangs. */
I915_WRITE(GEN8_GARBCNTL, ret = wa_ring_whitelist_reg(engine, GEN9_SLICE_COMMON_ECO_CHICKEN1);
I915_READ(GEN8_GARBCNTL) | GEN9_GAPS_TSV_CREDIT_DISABLE); if (ret)
return ret;
/* WaToEnableHwFixForPushConstHWBug:cfl */ return 0;
WA_SET_BIT_MASKED(COMMON_SLICE_CHICKEN2, }
GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
/* WaDisableGafsUnitClkGating:cfl */ static int cfl_whitelist_workarounds_apply(struct intel_engine_cs *engine)
I915_WRITE(GEN7_UCGCTL4, {
I915_READ(GEN7_UCGCTL4) | GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE); int ret;
/* WaDisableSbeCacheDispatchPortSharing:cfl */ ret = gen9_whitelist_workarounds_apply(engine);
WA_SET_BIT_MASKED(GEN7_HALF_SLICE_CHICKEN1, if (ret)
GEN7_SBE_SS_CACHE_DISPATCH_PORT_SHARING_DISABLE); return ret;
/* WaInPlaceDecompressionHang:cfl */ return 0;
I915_WRITE(GEN9_GAMT_ECO_REG_RW_IA, }
I915_READ(GEN9_GAMT_ECO_REG_RW_IA) |
GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS); static int cnl_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{
int ret;
/* WaEnablePreemptionGranularityControlByUMD:cnl */
ret = wa_ring_whitelist_reg(engine, GEN8_CS_CHICKEN1);
if (ret)
return ret;
return 0; return 0;
} }
int init_workarounds_ring(struct intel_engine_cs *engine) int intel_whitelist_workarounds_apply(struct intel_engine_cs *engine)
{ {
struct drm_i915_private *dev_priv = engine->i915; struct drm_i915_private *dev_priv = engine->i915;
int err; int err = 0;
if (GEM_WARN_ON(engine->id != RCS)) WARN_ON(engine->id != RCS);
return -EINVAL;
dev_priv->workarounds.count = 0;
dev_priv->workarounds.hw_whitelist_count[engine->id] = 0; dev_priv->workarounds.hw_whitelist_count[engine->id] = 0;
if (IS_BROADWELL(dev_priv)) if (INTEL_GEN(dev_priv) < 8)
err = bdw_init_workarounds(engine); err = 0;
else if (IS_BROADWELL(dev_priv))
err = bdw_whitelist_workarounds_apply(engine);
else if (IS_CHERRYVIEW(dev_priv)) else if (IS_CHERRYVIEW(dev_priv))
err = chv_init_workarounds(engine); err = chv_whitelist_workarounds_apply(engine);
else if (IS_SKYLAKE(dev_priv)) else if (IS_SKYLAKE(dev_priv))
err = skl_init_workarounds(engine); err = skl_whitelist_workarounds_apply(engine);
else if (IS_BROXTON(dev_priv)) else if (IS_BROXTON(dev_priv))
err = bxt_init_workarounds(engine); err = bxt_whitelist_workarounds_apply(engine);
else if (IS_KABYLAKE(dev_priv)) else if (IS_KABYLAKE(dev_priv))
err = kbl_init_workarounds(engine); err = kbl_whitelist_workarounds_apply(engine);
else if (IS_GEMINILAKE(dev_priv)) else if (IS_GEMINILAKE(dev_priv))
err = glk_init_workarounds(engine); err = glk_whitelist_workarounds_apply(engine);
else if (IS_COFFEELAKE(dev_priv)) else if (IS_COFFEELAKE(dev_priv))
err = cfl_init_workarounds(engine); err = cfl_whitelist_workarounds_apply(engine);
else if (IS_CANNONLAKE(dev_priv)) else if (IS_CANNONLAKE(dev_priv))
err = cnl_init_workarounds(engine); err = cnl_whitelist_workarounds_apply(engine);
else else
err = 0; MISSING_CASE(INTEL_GEN(dev_priv));
if (err) if (err)
return err; return err;
DRM_DEBUG_DRIVER("%s: Number of context specific w/a: %d\n", DRM_DEBUG_DRIVER("%s: Number of whitelist w/a: %d\n", engine->name,
engine->name, dev_priv->workarounds.count); dev_priv->workarounds.hw_whitelist_count[engine->id]);
return 0;
}
int intel_ring_workarounds_emit(struct i915_request *rq)
{
struct i915_workarounds *w = &rq->i915->workarounds;
u32 *cs;
int ret, i;
if (w->count == 0)
return 0;
ret = rq->engine->emit_flush(rq, EMIT_BARRIER);
if (ret)
return ret;
cs = intel_ring_begin(rq, w->count * 2 + 2);
if (IS_ERR(cs))
return PTR_ERR(cs);
*cs++ = MI_LOAD_REGISTER_IMM(w->count);
for (i = 0; i < w->count; i++) {
*cs++ = i915_mmio_reg_offset(w->reg[i].addr);
*cs++ = w->reg[i].value;
}
*cs++ = MI_NOOP;
intel_ring_advance(rq, cs);
ret = rq->engine->emit_flush(rq, EMIT_BARRIER);
if (ret)
return ret;
return 0; return 0;
} }
...@@ -7,7 +7,11 @@ ...@@ -7,7 +7,11 @@
#ifndef _I915_WORKAROUNDS_H_ #ifndef _I915_WORKAROUNDS_H_
#define _I915_WORKAROUNDS_H_ #define _I915_WORKAROUNDS_H_
int init_workarounds_ring(struct intel_engine_cs *engine); int intel_ctx_workarounds_init(struct drm_i915_private *dev_priv);
int intel_ring_workarounds_emit(struct i915_request *rq); int intel_ctx_workarounds_emit(struct i915_request *rq);
void intel_gt_workarounds_apply(struct drm_i915_private *dev_priv);
int intel_whitelist_workarounds_apply(struct intel_engine_cs *engine);
#endif #endif
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册