intel_workarounds.c 58.3 KB
Newer Older
C
Chris Wilson 已提交
1
// SPDX-License-Identifier: MIT
2 3 4 5 6
/*
 * Copyright © 2014-2018 Intel Corporation
 */

#include "i915_drv.h"
7
#include "intel_context.h"
8
#include "intel_engine_pm.h"
9
#include "intel_gpu_commands.h"
10
#include "intel_gt.h"
11
#include "intel_ring.h"
12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43
#include "intel_workarounds.h"

/**
 * DOC: Hardware workarounds
 *
 * This file is intended as a central place to implement most [1]_ of the
 * required workarounds for hardware to work as originally intended. They fall
 * in five basic categories depending on how/when they are applied:
 *
 * - Workarounds that touch registers that are saved/restored to/from the HW
 *   context image. The list is emitted (via Load Register Immediate commands)
 *   everytime a new context is created.
 * - GT workarounds. The list of these WAs is applied whenever these registers
 *   revert to default values (on GPU reset, suspend/resume [2]_, etc..).
 * - Display workarounds. The list is applied during display clock-gating
 *   initialization.
 * - Workarounds that whitelist a privileged register, so that UMDs can manage
 *   them directly. This is just a special case of a MMMIO workaround (as we
 *   write the list of these to/be-whitelisted registers to some special HW
 *   registers).
 * - Workaround batchbuffers, that get executed automatically by the hardware
 *   on every HW context restore.
 *
 * .. [1] Please notice that there are other WAs that, due to their nature,
 *    cannot be applied from a central place. Those are peppered around the rest
 *    of the code, as needed.
 *
 * .. [2] Technically, some registers are powercontext saved & restored, so they
 *    survive a suspend/resume. In practice, writing them again is not too
 *    costly and simplifies things. We can revisit this in the future.
 *
 * Layout
44
 * ~~~~~~
45 46 47 48 49 50 51 52 53 54
 *
 * Keep things in this file ordered by WA type, as per the above (context, GT,
 * display, register whitelist, batchbuffer). Then, inside each type, keep the
 * following order:
 *
 * - Infrastructure functions and macros
 * - WAs per platform in standard gen/chrono order
 * - Public functions to init or apply the given workaround type.
 */

55
static void wa_init_start(struct i915_wa_list *wal, const char *name, const char *engine_name)
56 57
{
	wal->name = name;
58
	wal->engine_name = engine_name;
59 60
}

61 62
#define WA_LIST_CHUNK (1 << 4)

63 64
static void wa_init_finish(struct i915_wa_list *wal)
{
65 66 67 68 69 70 71 72 73 74 75 76
	/* Trim unused entries. */
	if (!IS_ALIGNED(wal->count, WA_LIST_CHUNK)) {
		struct i915_wa *list = kmemdup(wal->list,
					       wal->count * sizeof(*list),
					       GFP_KERNEL);

		if (list) {
			kfree(wal->list);
			wal->list = list;
		}
	}

77 78 79
	if (!wal->count)
		return;

80 81
	DRM_DEBUG_DRIVER("Initialized %u %s workarounds on %s\n",
			 wal->wa_count, wal->name, wal->engine_name);
82 83
}

84
static void _wa_add(struct i915_wa_list *wal, const struct i915_wa *wa)
85
{
86 87
	unsigned int addr = i915_mmio_reg_offset(wa->reg);
	unsigned int start = 0, end = wal->count;
88
	const unsigned int grow = WA_LIST_CHUNK;
89 90 91 92 93 94 95 96 97 98 99 100 101 102
	struct i915_wa *wa_;

	GEM_BUG_ON(!is_power_of_2(grow));

	if (IS_ALIGNED(wal->count, grow)) { /* Either uninitialized or full. */
		struct i915_wa *list;

		list = kmalloc_array(ALIGN(wal->count + 1, grow), sizeof(*wa),
				     GFP_KERNEL);
		if (!list) {
			DRM_ERROR("No space for workaround init!\n");
			return;
		}

103
		if (wal->list) {
104
			memcpy(list, wal->list, sizeof(*wa) * wal->count);
105 106
			kfree(wal->list);
		}
107 108 109

		wal->list = list;
	}
110 111 112 113

	while (start < end) {
		unsigned int mid = start + (end - start) / 2;

114
		if (i915_mmio_reg_offset(wal->list[mid].reg) < addr) {
115
			start = mid + 1;
116
		} else if (i915_mmio_reg_offset(wal->list[mid].reg) > addr) {
117 118
			end = mid;
		} else {
119
			wa_ = &wal->list[mid];
120

121 122
			if ((wa->clr | wa_->clr) && !(wa->clr & ~wa_->clr)) {
				DRM_ERROR("Discarding overwritten w/a for reg %04x (clear: %08x, set: %08x)\n",
123
					  i915_mmio_reg_offset(wa_->reg),
124
					  wa_->clr, wa_->set);
125

126
				wa_->set &= ~wa->clr;
127 128
			}

129
			wal->wa_count++;
130 131
			wa_->set |= wa->set;
			wa_->clr |= wa->clr;
132
			wa_->read |= wa->read;
133 134 135
			return;
		}
	}
136

137 138 139
	wal->wa_count++;
	wa_ = &wal->list[wal->count++];
	*wa_ = *wa;
140

141 142 143 144 145
	while (wa_-- > wal->list) {
		GEM_BUG_ON(i915_mmio_reg_offset(wa_[0].reg) ==
			   i915_mmio_reg_offset(wa_[1].reg));
		if (i915_mmio_reg_offset(wa_[1].reg) >
		    i915_mmio_reg_offset(wa_[0].reg))
146
			break;
147

148
		swap(wa_[1], wa_[0]);
149
	}
150 151
}

152
static void wa_add(struct i915_wa_list *wal, i915_reg_t reg,
153
		   u32 clear, u32 set, u32 read_mask, bool masked_reg)
154 155
{
	struct i915_wa wa = {
156
		.reg  = reg,
157 158
		.clr  = clear,
		.set  = set,
159
		.read = read_mask,
160
		.masked_reg = masked_reg,
161 162 163 164 165
	};

	_wa_add(wal, &wa);
}

166
static void
167
wa_write_clr_set(struct i915_wa_list *wal, i915_reg_t reg, u32 clear, u32 set)
168
{
169
	wa_add(wal, reg, clear, set, clear, false);
170 171
}

172
static void
173 174
wa_write(struct i915_wa_list *wal, i915_reg_t reg, u32 set)
{
175
	wa_write_clr_set(wal, reg, ~0, set);
176 177 178 179
}

static void
wa_write_or(struct i915_wa_list *wal, i915_reg_t reg, u32 set)
180
{
181
	wa_write_clr_set(wal, reg, set, set);
182 183
}

184 185 186
static void
wa_write_clr(struct i915_wa_list *wal, i915_reg_t reg, u32 clr)
{
187
	wa_write_clr_set(wal, reg, clr, 0);
188 189
}

190 191 192 193 194 195 196 197 198 199 200
/*
 * WA operations on "masked register". A masked register has the upper 16 bits
 * documented as "masked" in b-spec. Its purpose is to allow writing to just a
 * portion of the register without a rmw: you simply write in the upper 16 bits
 * the mask of bits you are going to modify.
 *
 * The wa_masked_* family of functions already does the necessary operations to
 * calculate the mask based on the parameters passed, so user only has to
 * provide the lower 16 bits of that register.
 */

201
static void
202
wa_masked_en(struct i915_wa_list *wal, i915_reg_t reg, u32 val)
203
{
204
	wa_add(wal, reg, 0, _MASKED_BIT_ENABLE(val), val, true);
205 206 207
}

static void
208
wa_masked_dis(struct i915_wa_list *wal, i915_reg_t reg, u32 val)
209
{
210
	wa_add(wal, reg, 0, _MASKED_BIT_DISABLE(val), val, true);
211 212
}

213 214 215 216
static void
wa_masked_field_set(struct i915_wa_list *wal, i915_reg_t reg,
		    u32 mask, u32 val)
{
217
	wa_add(wal, reg, 0, _MASKED_FIELD(mask, val), mask, true);
218
}
219

220 221 222
static void gen6_ctx_workarounds_init(struct intel_engine_cs *engine,
				      struct i915_wa_list *wal)
{
223
	wa_masked_en(wal, INSTPM, INSTPM_FORCE_ORDERING);
224 225 226 227 228
}

static void gen7_ctx_workarounds_init(struct intel_engine_cs *engine,
				      struct i915_wa_list *wal)
{
229
	wa_masked_en(wal, INSTPM, INSTPM_FORCE_ORDERING);
230 231
}

232 233
static void gen8_ctx_workarounds_init(struct intel_engine_cs *engine,
				      struct i915_wa_list *wal)
234
{
235
	wa_masked_en(wal, INSTPM, INSTPM_FORCE_ORDERING);
236 237

	/* WaDisableAsyncFlipPerfMode:bdw,chv */
238
	wa_masked_en(wal, MI_MODE, ASYNC_FLIP_PERF_DISABLE);
239 240

	/* WaDisablePartialInstShootdown:bdw,chv */
241 242
	wa_masked_en(wal, GEN8_ROW_CHICKEN,
		     PARTIAL_INSTRUCTION_SHOOTDOWN_DISABLE);
243 244

	/* Use Force Non-Coherent whenever executing a 3D context. This is a
245
	 * workaround for a possible hang in the unlikely event a TLB
246 247 248 249
	 * invalidation occurs during a PSD flush.
	 */
	/* WaForceEnableNonCoherent:bdw,chv */
	/* WaHdcDisableFetchWhenMasked:bdw,chv */
250 251 252
	wa_masked_en(wal, HDC_CHICKEN0,
		     HDC_DONOT_FETCH_MEM_WHEN_MASKED |
		     HDC_FORCE_NON_COHERENT);
253 254 255 256 257 258 259 260 261

	/* From the Haswell PRM, Command Reference: Registers, CACHE_MODE_0:
	 * "The Hierarchical Z RAW Stall Optimization allows non-overlapping
	 *  polygons in the same 8x4 pixel/sample area to be processed without
	 *  stalling waiting for the earlier ones to write to Hierarchical Z
	 *  buffer."
	 *
	 * This optimization is off by default for BDW and CHV; turn it on.
	 */
262
	wa_masked_dis(wal, CACHE_MODE_0_GEN7, HIZ_RAW_STALL_OPT_DISABLE);
263 264

	/* Wa4x4STCOptimizationDisable:bdw,chv */
265
	wa_masked_en(wal, CACHE_MODE_1, GEN8_4x4_STC_OPTIMIZATION_DISABLE);
266 267 268 269 270 271 272 273 274

	/*
	 * BSpec recommends 8x4 when MSAA is used,
	 * however in practice 16x4 seems fastest.
	 *
	 * Note that PS/WM thread counts depend on the WIZ hashing
	 * disable bit, which we don't touch here, but it's good
	 * to keep in mind (see 3DSTATE_PS and 3DSTATE_WM).
	 */
275
	wa_masked_field_set(wal, GEN7_GT_MODE,
276 277 278 279
			    GEN6_WIZ_HASHING_MASK,
			    GEN6_WIZ_HASHING_16x4);
}

280 281
static void bdw_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
282
{
283
	struct drm_i915_private *i915 = engine->i915;
284

285
	gen8_ctx_workarounds_init(engine, wal);
286 287

	/* WaDisableThreadStallDopClockGating:bdw (pre-production) */
288
	wa_masked_en(wal, GEN8_ROW_CHICKEN, STALL_DOP_GATING_DISABLE);
289 290 291

	/* WaDisableDopClockGating:bdw
	 *
292
	 * Also see the related UCGTCL1 write in bdw_init_clock_gating()
293 294
	 * to disable EUTC clock gating.
	 */
295 296
	wa_masked_en(wal, GEN7_ROW_CHICKEN2,
		     DOP_CLOCK_GATING_DISABLE);
297

298 299
	wa_masked_en(wal, HALF_SLICE_CHICKEN3,
		     GEN8_SAMPLER_POWER_BYPASS_DIS);
300

301 302 303 304 305
	wa_masked_en(wal, HDC_CHICKEN0,
		     /* WaForceContextSaveRestoreNonCoherent:bdw */
		     HDC_FORCE_CONTEXT_SAVE_RESTORE_NON_COHERENT |
		     /* WaDisableFenceDestinationToSLM:bdw (pre-prod) */
		     (IS_BDW_GT3(i915) ? HDC_FENCE_DEST_SLM_DISABLE : 0));
306 307
}

308 309
static void chv_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
310
{
311
	gen8_ctx_workarounds_init(engine, wal);
312 313

	/* WaDisableThreadStallDopClockGating:chv */
314
	wa_masked_en(wal, GEN8_ROW_CHICKEN, STALL_DOP_GATING_DISABLE);
315 316

	/* Improve HiZ throughput on CHV. */
317
	wa_masked_en(wal, HIZ_CHICKEN, CHV_HZ_8X8_MODE_IN_1X);
318 319
}

320 321
static void gen9_ctx_workarounds_init(struct intel_engine_cs *engine,
				      struct i915_wa_list *wal)
322
{
323 324 325
	struct drm_i915_private *i915 = engine->i915;

	if (HAS_LLC(i915)) {
326 327 328 329 330
		/* WaCompressedResourceSamplerPbeMediaNewHashMode:skl,kbl
		 *
		 * Must match Display Engine. See
		 * WaCompressedResourceDisplayNewHashMode.
		 */
331 332 333 334
		wa_masked_en(wal, COMMON_SLICE_CHICKEN2,
			     GEN9_PBE_COMPRESSED_HASH_SELECTION);
		wa_masked_en(wal, GEN9_HALF_SLICE_CHICKEN7,
			     GEN9_SAMPLER_HASH_COMPRESSED_READ_ADDR);
335 336 337 338
	}

	/* WaClearFlowControlGpgpuContextSave:skl,bxt,kbl,glk,cfl */
	/* WaDisablePartialInstShootdown:skl,bxt,kbl,glk,cfl */
339 340 341
	wa_masked_en(wal, GEN8_ROW_CHICKEN,
		     FLOW_CONTROL_ENABLE |
		     PARTIAL_INSTRUCTION_SHOOTDOWN_DISABLE);
342 343 344

	/* WaEnableYV12BugFixInHalfSliceChicken7:skl,bxt,kbl,glk,cfl */
	/* WaEnableSamplerGPGPUPreemptionSupport:skl,bxt,kbl,cfl */
345 346 347
	wa_masked_en(wal, GEN9_HALF_SLICE_CHICKEN7,
		     GEN9_ENABLE_YV12_BUGFIX |
		     GEN9_ENABLE_GPGPU_PREEMPTION);
348 349 350

	/* Wa4x4STCOptimizationDisable:skl,bxt,kbl,glk,cfl */
	/* WaDisablePartialResolveInVc:skl,bxt,kbl,cfl */
351 352 353
	wa_masked_en(wal, CACHE_MODE_1,
		     GEN8_4x4_STC_OPTIMIZATION_DISABLE |
		     GEN9_PARTIAL_RESOLVE_IN_VC_DISABLE);
354 355

	/* WaCcsTlbPrefetchDisable:skl,bxt,kbl,glk,cfl */
356 357
	wa_masked_dis(wal, GEN9_HALF_SLICE_CHICKEN5,
		      GEN9_CCS_TLB_PREFETCH_ENABLE);
358 359

	/* WaForceContextSaveRestoreNonCoherent:skl,bxt,kbl,cfl */
360 361 362
	wa_masked_en(wal, HDC_CHICKEN0,
		     HDC_FORCE_CONTEXT_SAVE_RESTORE_NON_COHERENT |
		     HDC_FORCE_CSR_NON_COHERENT_OVR_DISABLE);
363 364 365 366 367 368 369 370 371 372 373 374 375 376 377

	/* WaForceEnableNonCoherent and WaDisableHDCInvalidation are
	 * both tied to WaForceContextSaveRestoreNonCoherent
	 * in some hsds for skl. We keep the tie for all gen9. The
	 * documentation is a bit hazy and so we want to get common behaviour,
	 * even though there is no clear evidence we would need both on kbl/bxt.
	 * This area has been source of system hangs so we play it safe
	 * and mimic the skl regardless of what bspec says.
	 *
	 * Use Force Non-Coherent whenever executing a 3D context. This
	 * is a workaround for a possible hang in the unlikely event
	 * a TLB invalidation occurs during a PSD flush.
	 */

	/* WaForceEnableNonCoherent:skl,bxt,kbl,cfl */
378 379
	wa_masked_en(wal, HDC_CHICKEN0,
		     HDC_FORCE_NON_COHERENT);
380 381

	/* WaDisableSamplerPowerBypassForSOPingPong:skl,bxt,kbl,cfl */
382 383 384 385
	if (IS_SKYLAKE(i915) ||
	    IS_KABYLAKE(i915) ||
	    IS_COFFEELAKE(i915) ||
	    IS_COMETLAKE(i915))
386 387
		wa_masked_en(wal, HALF_SLICE_CHICKEN3,
			     GEN8_SAMPLER_POWER_BYPASS_DIS);
388 389

	/* WaDisableSTUnitPowerOptimization:skl,bxt,kbl,glk,cfl */
390
	wa_masked_en(wal, HALF_SLICE_CHICKEN2, GEN8_ST_PO_DISABLE);
391 392 393 394 395 396 397 398 399 400 401 402 403

	/*
	 * Supporting preemption with fine-granularity requires changes in the
	 * batch buffer programming. Since we can't break old userspace, we
	 * need to set our default preemption level to safe value. Userspace is
	 * still able to use more fine-grained preemption levels, since in
	 * WaEnablePreemptionGranularityControlByUMD we're whitelisting the
	 * per-ctx register. As such, WaDisable{3D,GPGPU}MidCmdPreemption are
	 * not real HW workarounds, but merely a way to start using preemption
	 * while maintaining old contract with userspace.
	 */

	/* WaDisable3DMidCmdPreemption:skl,bxt,glk,cfl,[cnl] */
404
	wa_masked_dis(wal, GEN8_CS_CHICKEN1, GEN9_PREEMPT_3D_OBJECT_LEVEL);
405 406

	/* WaDisableGPGPUMidCmdPreemption:skl,bxt,blk,cfl,[cnl] */
407
	wa_masked_field_set(wal, GEN8_CS_CHICKEN1,
408 409 410
			    GEN9_PREEMPT_GPGPU_LEVEL_MASK,
			    GEN9_PREEMPT_GPGPU_COMMAND_LEVEL);

411
	/* WaClearHIZ_WM_CHICKEN3:bxt,glk */
412
	if (IS_GEN9_LP(i915))
413
		wa_masked_en(wal, GEN9_WM_CHICKEN3, GEN9_FACTOR_IN_CLR_VAL_HIZ);
414 415
}

416 417
static void skl_tune_iz_hashing(struct intel_engine_cs *engine,
				struct i915_wa_list *wal)
418
{
419
	struct intel_gt *gt = engine->gt;
420 421 422 423 424 425 426 427 428 429
	u8 vals[3] = { 0, 0, 0 };
	unsigned int i;

	for (i = 0; i < 3; i++) {
		u8 ss;

		/*
		 * Only consider slices where one, and only one, subslice has 7
		 * EUs
		 */
430
		if (!is_power_of_2(gt->info.sseu.subslice_7eu[i]))
431 432 433 434 435 436 437 438
			continue;

		/*
		 * subslice_7eu[i] != 0 (because of the check above) and
		 * ss_max == 4 (maximum number of subslices possible per slice)
		 *
		 * ->    0 <= ss <= 3;
		 */
439
		ss = ffs(gt->info.sseu.subslice_7eu[i]) - 1;
440 441 442 443
		vals[i] = 3 - ss;
	}

	if (vals[0] == 0 && vals[1] == 0 && vals[2] == 0)
444
		return;
445 446

	/* Tune IZ hashing. See intel_device_info_runtime_init() */
447
	wa_masked_field_set(wal, GEN7_GT_MODE,
448 449 450 451 452 453 454 455
			    GEN9_IZ_HASHING_MASK(2) |
			    GEN9_IZ_HASHING_MASK(1) |
			    GEN9_IZ_HASHING_MASK(0),
			    GEN9_IZ_HASHING(2, vals[2]) |
			    GEN9_IZ_HASHING(1, vals[1]) |
			    GEN9_IZ_HASHING(0, vals[0]));
}

456 457
static void skl_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
458
{
459 460
	gen9_ctx_workarounds_init(engine, wal);
	skl_tune_iz_hashing(engine, wal);
461
}
462

463 464
static void bxt_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
465
{
466
	gen9_ctx_workarounds_init(engine, wal);
467

468
	/* WaDisableThreadStallDopClockGating:bxt */
469 470
	wa_masked_en(wal, GEN8_ROW_CHICKEN,
		     STALL_DOP_GATING_DISABLE);
471 472

	/* WaToEnableHwFixForPushConstHWBug:bxt */
473 474
	wa_masked_en(wal, COMMON_SLICE_CHICKEN2,
		     GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
475 476
}

477 478
static void kbl_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
479
{
480
	struct drm_i915_private *i915 = engine->i915;
481

482
	gen9_ctx_workarounds_init(engine, wal);
483

484
	/* WaToEnableHwFixForPushConstHWBug:kbl */
485
	if (IS_KBL_GT_STEP(i915, STEP_C0, STEP_FOREVER))
486 487
		wa_masked_en(wal, COMMON_SLICE_CHICKEN2,
			     GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
488

489
	/* WaDisableSbeCacheDispatchPortSharing:kbl */
490 491
	wa_masked_en(wal, GEN7_HALF_SLICE_CHICKEN1,
		     GEN7_SBE_SS_CACHE_DISPATCH_PORT_SHARING_DISABLE);
492 493
}

494 495
static void glk_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
496
{
497
	gen9_ctx_workarounds_init(engine, wal);
498 499

	/* WaToEnableHwFixForPushConstHWBug:glk */
500 501
	wa_masked_en(wal, COMMON_SLICE_CHICKEN2,
		     GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
502 503
}

504 505
static void cfl_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
506
{
507
	gen9_ctx_workarounds_init(engine, wal);
508 509

	/* WaToEnableHwFixForPushConstHWBug:cfl */
510 511
	wa_masked_en(wal, COMMON_SLICE_CHICKEN2,
		     GEN8_SBE_DISABLE_REPLAY_BUF_OPTIMIZATION);
512

513
	/* WaDisableSbeCacheDispatchPortSharing:cfl */
514 515
	wa_masked_en(wal, GEN7_HALF_SLICE_CHICKEN1,
		     GEN7_SBE_SS_CACHE_DISPATCH_PORT_SHARING_DISABLE);
516 517
}

518 519
static void icl_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
520
{
521
	/* Wa_1406697149 (WaDisableBankHangMode:icl) */
522 523 524 525 526
	wa_write(wal,
		 GEN8_L3CNTLREG,
		 intel_uncore_read(engine->uncore, GEN8_L3CNTLREG) |
		 GEN8_ERRDETBCTRL);

527 528 529 530 531 532 533
	/* WaForceEnableNonCoherent:icl
	 * This is not the same workaround as in early Gen9 platforms, where
	 * lacking this could cause system hangs, but coherency performance
	 * overhead is high and only a few compute workloads really need it
	 * (the register is whitelisted in hardware now, so UMDs can opt in
	 * for coherency if they have a good reason).
	 */
534
	wa_masked_en(wal, ICL_HDC_MODE, HDC_FORCE_NON_COHERENT);
535

536
	/* WaEnableFloatBlendOptimization:icl */
537 538 539 540
	wa_add(wal, GEN10_CACHE_MODE_SS, 0,
	       _MASKED_BIT_ENABLE(FLOAT_BLEND_OPTIMIZATION_ENABLE),
	       0 /* write-only, so skip validation */,
	       true);
541 542

	/* WaDisableGPGPUMidThreadPreemption:icl */
543
	wa_masked_field_set(wal, GEN8_CS_CHICKEN1,
544 545
			    GEN9_PREEMPT_GPGPU_LEVEL_MASK,
			    GEN9_PREEMPT_GPGPU_THREAD_GROUP_LEVEL);
546 547

	/* allow headerless messages for preemptible GPGPU context */
548 549
	wa_masked_en(wal, GEN10_SAMPLER_MODE,
		     GEN11_SAMPLER_ENABLE_HEADLESS_MSG);
M
Matt Roper 已提交
550 551 552

	/* Wa_1604278689:icl,ehl */
	wa_write(wal, IVB_FBC_RT_BASE, 0xFFFFFFFF & ~ILK_FBC_RT_VALID);
553 554 555
	wa_write_clr_set(wal, IVB_FBC_RT_BASE_UPPER,
			 0, /* write-only register; skip validation */
			 0xFFFFFFFF);
M
Matt Roper 已提交
556 557 558

	/* Wa_1406306137:icl,ehl */
	wa_masked_en(wal, GEN9_ROW_CHICKEN4, GEN11_DIS_PICK_2ND_EU);
559 560
}

561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584
/*
 * These settings aren't actually workarounds, but general tuning settings that
 * need to be programmed on several platforms.
 */
static void gen12_ctx_gt_tuning_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
{
	/*
	 * Although some platforms refer to it as Wa_1604555607, we need to
	 * program it even on those that don't explicitly list that
	 * workaround.
	 *
	 * Note that the programming of this register is further modified
	 * according to the FF_MODE2 guidance given by Wa_1608008084:gen12.
	 * Wa_1608008084 tells us the FF_MODE2 register will return the wrong
	 * value when read. The default value for this register is zero for all
	 * fields and there are no bit masks. So instead of doing a RMW we
	 * should just write TDS timer value. For the same reason read
	 * verification is ignored.
	 */
	wa_add(wal,
	       FF_MODE2,
	       FF_MODE2_TDS_TIMER_MASK,
	       FF_MODE2_TDS_TIMER_128,
585
	       0, false);
586 587
}

588 589
static void gen12_ctx_workarounds_init(struct intel_engine_cs *engine,
				       struct i915_wa_list *wal)
590
{
591 592
	gen12_ctx_gt_tuning_init(engine, wal);

593
	/*
594 595 596 597 598 599 600 601 602 603
	 * Wa_1409142259:tgl,dg1,adl-p
	 * Wa_1409347922:tgl,dg1,adl-p
	 * Wa_1409252684:tgl,dg1,adl-p
	 * Wa_1409217633:tgl,dg1,adl-p
	 * Wa_1409207793:tgl,dg1,adl-p
	 * Wa_1409178076:tgl,dg1,adl-p
	 * Wa_1408979724:tgl,dg1,adl-p
	 * Wa_14010443199:tgl,rkl,dg1,adl-p
	 * Wa_14010698770:tgl,rkl,dg1,adl-s,adl-p
	 * Wa_1409342910:tgl,rkl,dg1,adl-s,adl-p
604
	 */
605 606
	wa_masked_en(wal, GEN11_COMMON_SLICE_CHICKEN3,
		     GEN12_DISABLE_CPS_AWARE_COLOR_PIPE);
607

608
	/* WaDisableGPGPUMidThreadPreemption:gen12 */
609
	wa_masked_field_set(wal, GEN8_CS_CHICKEN1,
610 611 612
			    GEN9_PREEMPT_GPGPU_LEVEL_MASK,
			    GEN9_PREEMPT_GPGPU_THREAD_GROUP_LEVEL);

613
	/*
614
	 * Wa_16011163337
615
	 *
616 617
	 * Like in gen12_ctx_gt_tuning_init(), read verification is ignored due
	 * to Wa_1608008084.
618
	 */
619 620
	wa_add(wal,
	       FF_MODE2,
621 622
	       FF_MODE2_GS_TIMER_MASK,
	       FF_MODE2_GS_TIMER_224,
623
	       0, false);
624 625 626 627 628 629 630

	/*
	 * Wa_14012131227:dg1
	 * Wa_1508744258:tgl,rkl,dg1,adl-s,adl-p
	 */
	wa_masked_en(wal, GEN7_COMMON_SLICE_CHICKEN1,
		     GEN9_RHWO_OPTIMIZATION_DISABLE);
631 632
}

633 634 635 636 637 638
static void dg1_ctx_workarounds_init(struct intel_engine_cs *engine,
				     struct i915_wa_list *wal)
{
	gen12_ctx_workarounds_init(engine, wal);

	/* Wa_1409044764 */
639 640
	wa_masked_dis(wal, GEN11_COMMON_SLICE_CHICKEN3,
		      DG1_FLOAT_POINT_BLEND_OPT_STRICT_MODE_EN);
641 642

	/* Wa_22010493298 */
643 644
	wa_masked_en(wal, HIZ_CHICKEN,
		     DG1_HZ_READ_SUPPRESSION_OPTIMIZATION_DISABLE);
645 646
}

647 648 649 650
static void
__intel_engine_init_ctx_wa(struct intel_engine_cs *engine,
			   struct i915_wa_list *wal,
			   const char *name)
651
{
652 653
	struct drm_i915_private *i915 = engine->i915;

654 655 656
	if (engine->class != RENDER_CLASS)
		return;

657
	wa_init_start(wal, name, engine->name);
658

659 660
	if (IS_DG1(i915))
		dg1_ctx_workarounds_init(engine, wal);
661
	else if (GRAPHICS_VER(i915) == 12)
662
		gen12_ctx_workarounds_init(engine, wal);
663
	else if (GRAPHICS_VER(i915) == 11)
664
		icl_ctx_workarounds_init(engine, wal);
665
	else if (IS_COFFEELAKE(i915) || IS_COMETLAKE(i915))
666
		cfl_ctx_workarounds_init(engine, wal);
667
	else if (IS_GEMINILAKE(i915))
668
		glk_ctx_workarounds_init(engine, wal);
669
	else if (IS_KABYLAKE(i915))
670
		kbl_ctx_workarounds_init(engine, wal);
671
	else if (IS_BROXTON(i915))
672
		bxt_ctx_workarounds_init(engine, wal);
673
	else if (IS_SKYLAKE(i915))
674
		skl_ctx_workarounds_init(engine, wal);
675
	else if (IS_CHERRYVIEW(i915))
676
		chv_ctx_workarounds_init(engine, wal);
677
	else if (IS_BROADWELL(i915))
678
		bdw_ctx_workarounds_init(engine, wal);
679
	else if (GRAPHICS_VER(i915) == 7)
680
		gen7_ctx_workarounds_init(engine, wal);
681
	else if (GRAPHICS_VER(i915) == 6)
682
		gen6_ctx_workarounds_init(engine, wal);
683
	else if (GRAPHICS_VER(i915) < 8)
684
		;
685
	else
686
		MISSING_CASE(GRAPHICS_VER(i915));
687

688
	wa_init_finish(wal);
689 690
}

691 692 693 694 695
void intel_engine_init_ctx_wa(struct intel_engine_cs *engine)
{
	__intel_engine_init_ctx_wa(engine, &engine->ctx_wa_list, "context");
}

696
int intel_engine_emit_ctx_wa(struct i915_request *rq)
697
{
698 699 700
	struct i915_wa_list *wal = &rq->engine->ctx_wa_list;
	struct i915_wa *wa;
	unsigned int i;
701
	u32 *cs;
702
	int ret;
703

704
	if (wal->count == 0)
705 706 707
		return 0;

	ret = rq->engine->emit_flush(rq, EMIT_BARRIER);
708 709 710
	if (ret)
		return ret;

711
	cs = intel_ring_begin(rq, (wal->count * 2 + 2));
712 713 714
	if (IS_ERR(cs))
		return PTR_ERR(cs);

715 716 717
	*cs++ = MI_LOAD_REGISTER_IMM(wal->count);
	for (i = 0, wa = wal->list; i < wal->count; i++, wa++) {
		*cs++ = i915_mmio_reg_offset(wa->reg);
718
		*cs++ = wa->set;
719 720 721 722 723 724 725 726 727 728 729 730
	}
	*cs++ = MI_NOOP;

	intel_ring_advance(rq, cs);

	ret = rq->engine->emit_flush(rq, EMIT_BARRIER);
	if (ret)
		return ret;

	return 0;
}

731
static void
732 733
gen4_gt_workarounds_init(struct drm_i915_private *i915,
			 struct i915_wa_list *wal)
734
{
735 736 737 738 739 740 741 742
	/* WaDisable_RenderCache_OperationalFlush:gen4,ilk */
	wa_masked_dis(wal, CACHE_MODE_0, RC_OP_FLUSH_ENABLE);
}

static void
g4x_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	gen4_gt_workarounds_init(i915, wal);
743

744
	/* WaDisableRenderCachePipelinedFlush:g4x,ilk */
745
	wa_masked_en(wal, CACHE_MODE_0, CM0_PIPELINED_RENDER_FLUSH_DISABLE);
746
}
747

748 749 750 751 752 753
static void
ilk_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	g4x_gt_workarounds_init(i915, wal);

	wa_masked_en(wal, _3D_CHICKEN2, _3D_CHICKEN2_WM_READ_PIPELINED);
754 755
}

756 757 758 759 760
static void
snb_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
}

761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776
static void
ivb_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	/* Apply the WaDisableRHWOOptimizationForRenderHang:ivb workaround. */
	wa_masked_dis(wal,
		      GEN7_COMMON_SLICE_CHICKEN1,
		      GEN7_CSC1_RHWO_OPT_DISABLE_IN_RCC);

	/* WaApplyL3ControlAndL3ChickenMode:ivb */
	wa_write(wal, GEN7_L3CNTLREG1, GEN7_WA_FOR_GEN7_L3_CONTROL);
	wa_write(wal, GEN7_L3_CHICKEN_MODE_REGISTER, GEN7_WA_L3_CHICKEN_MODE);

	/* WaForceL3Serialization:ivb */
	wa_write_clr(wal, GEN7_L3SQCREG4, L3SQ_URB_READ_CAM_MATCH_DISABLE);
}

777 778 779 780 781 782 783 784 785 786 787 788 789
static void
vlv_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	/* WaForceL3Serialization:vlv */
	wa_write_clr(wal, GEN7_L3SQCREG4, L3SQ_URB_READ_CAM_MATCH_DISABLE);

	/*
	 * WaIncreaseL3CreditsForVLVB0:vlv
	 * This is the hardware default actually.
	 */
	wa_write(wal, GEN7_L3SQCREG1, VLV_B0_WA_L3SQCREG1_VALUE);
}

790 791 792 793 794 795 796 797 798
static void
hsw_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	/* L3 caching of data atomics doesn't work -- disable it. */
	wa_write(wal, HSW_SCRATCH1, HSW_SCRATCH1_L3_DATA_ATOMICS_DISABLE);

	wa_add(wal,
	       HSW_ROW_CHICKEN3, 0,
	       _MASKED_BIT_ENABLE(HSW_ROW_CHICKEN3_L3_GLOBAL_ATOMICS_DISABLE),
799
	       0 /* XXX does this reg exist? */, true);
800 801 802 803 804

	/* WaVSRefCountFullforceMissDisable:hsw */
	wa_write_clr(wal, GEN7_FF_THREAD_MODE, GEN7_FF_VS_REF_CNT_FFME);
}

805 806
static void
gen9_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
807
{
808
	/* WaDisableKillLogic:bxt,skl,kbl */
809
	if (!IS_COFFEELAKE(i915) && !IS_COMETLAKE(i915))
810 811 812
		wa_write_or(wal,
			    GAM_ECOCHK,
			    ECOCHK_DIS_TLB);
813

814
	if (HAS_LLC(i915)) {
815 816 817 818 819
		/* WaCompressedResourceSamplerPbeMediaNewHashMode:skl,kbl
		 *
		 * Must match Display Engine. See
		 * WaCompressedResourceDisplayNewHashMode.
		 */
820 821 822
		wa_write_or(wal,
			    MMCD_MISC_CTRL,
			    MMCD_PCLA | MMCD_HOTSPOT_EN);
823 824 825
	}

	/* WaDisableHDCInvalidation:skl,bxt,kbl,cfl */
826 827 828
	wa_write_or(wal,
		    GAM_ECOCHK,
		    BDW_DISABLE_HDC_INVALIDATION);
829 830
}

831 832
static void
skl_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
833
{
834
	gen9_gt_workarounds_init(i915, wal);
835 836

	/* WaDisableGafsUnitClkGating:skl */
837 838 839
	wa_write_or(wal,
		    GEN7_UCGCTL4,
		    GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
840 841

	/* WaInPlaceDecompressionHang:skl */
842
	if (IS_SKL_GT_STEP(i915, STEP_A0, STEP_H0))
843 844 845
		wa_write_or(wal,
			    GEN9_GAMT_ECO_REG_RW_IA,
			    GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
846 847
}

848 849
static void
kbl_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
850
{
851
	gen9_gt_workarounds_init(i915, wal);
852

853
	/* WaDisableDynamicCreditSharing:kbl */
854
	if (IS_KBL_GT_STEP(i915, 0, STEP_C0))
855 856 857
		wa_write_or(wal,
			    GAMT_CHKN_BIT_REG,
			    GAMT_CHKN_DISABLE_DYNAMIC_CREDIT_SHARING);
858

859
	/* WaDisableGafsUnitClkGating:kbl */
860 861 862
	wa_write_or(wal,
		    GEN7_UCGCTL4,
		    GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
863

864
	/* WaInPlaceDecompressionHang:kbl */
865 866 867
	wa_write_or(wal,
		    GEN9_GAMT_ECO_REG_RW_IA,
		    GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
868
}
869

870 871
static void
glk_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
872
{
873
	gen9_gt_workarounds_init(i915, wal);
874 875
}

876 877
static void
cfl_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
878
{
879
	gen9_gt_workarounds_init(i915, wal);
880 881

	/* WaDisableGafsUnitClkGating:cfl */
882 883 884
	wa_write_or(wal,
		    GEN7_UCGCTL4,
		    GEN8_EU_GAUNIT_CLOCK_GATE_DISABLE);
885

886
	/* WaInPlaceDecompressionHang:cfl */
887 888 889
	wa_write_or(wal,
		    GEN9_GAMT_ECO_REG_RW_IA,
		    GAMT_ECO_ENABLE_IN_PLACE_DECOMPRESS);
890
}
891

892
static void
893
icl_wa_init_mcr(struct drm_i915_private *i915, struct i915_wa_list *wal)
894
{
895
	const struct sseu_dev_info *sseu = &i915->gt.info.sseu;
896
	unsigned int slice, subslice;
897
	u32 mcr, mcr_mask;
898

899 900 901
	GEM_BUG_ON(GRAPHICS_VER(i915) < 11);
	GEM_BUG_ON(hweight8(sseu->slice_mask) > 1);
	slice = 0;
902

903
	/*
904 905 906 907 908 909 910
	 * Although a platform may have subslices, we need to always steer
	 * reads to the lowest instance that isn't fused off.  When Render
	 * Power Gating is enabled, grabbing forcewake will only power up a
	 * single subslice (the "minconfig") if there isn't a real workload
	 * that needs to be run; this means that if we steer register reads to
	 * one of the higher subslices, we run the risk of reading back 0's or
	 * random garbage.
911
	 */
912
	subslice = __ffs(intel_sseu_get_subslices(sseu, slice));
913

914 915 916 917 918 919 920
	/*
	 * If the subslice we picked above also steers us to a valid L3 bank,
	 * then we can just rely on the default steering and won't need to
	 * worry about explicitly re-steering L3BANK reads later.
	 */
	if (i915->gt.info.l3bank_mask & BIT(subslice))
		i915->gt.steering_table[L3BANK] = NULL;
921

922 923
	mcr = GEN11_MCR_SLICE(slice) | GEN11_MCR_SUBSLICE(subslice);
	mcr_mask = GEN11_MCR_SLICE_MASK | GEN11_MCR_SUBSLICE_MASK;
924

925
	drm_dbg(&i915->drm, "MCR slice/subslice = %x\n", mcr);
926

927
	wa_write_clr_set(wal, GEN8_MCR_SELECTOR, mcr_mask, mcr);
928 929
}

930 931
static void
icl_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
932
{
933
	icl_wa_init_mcr(i915, wal);
934

935
	/* WaModifyGamTlbPartitioning:icl */
936 937 938 939
	wa_write_clr_set(wal,
			 GEN11_GACB_PERF_CTRL,
			 GEN11_HASH_CTRL_MASK,
			 GEN11_HASH_CTRL_BIT0 | GEN11_HASH_CTRL_BIT4);
O
Oscar Mateo 已提交
940

O
Oscar Mateo 已提交
941 942 943
	/* Wa_1405766107:icl
	 * Formerly known as WaCL2SFHalfMaxAlloc
	 */
944 945 946 947
	wa_write_or(wal,
		    GEN11_LSN_UNSLCVC,
		    GEN11_LSN_UNSLCVC_GAFS_HALF_SF_MAXALLOC |
		    GEN11_LSN_UNSLCVC_GAFS_HALF_CL2_MAXALLOC);
O
Oscar Mateo 已提交
948 949 950 951

	/* Wa_220166154:icl
	 * Formerly known as WaDisCtxReload
	 */
952 953 954
	wa_write_or(wal,
		    GEN8_GAMW_ECO_DEV_RW_IA,
		    GAMW_ECO_DEV_CTX_RELOAD_DISABLE);
O
Oscar Mateo 已提交
955

O
Oscar Mateo 已提交
956 957 958
	/* Wa_1406463099:icl
	 * Formerly known as WaGamTlbPendError
	 */
959 960 961
	wa_write_or(wal,
		    GAMT_CHKN_BIT_REG,
		    GAMT_CHKN_DISABLE_L3_COH_PIPE);
M
Mika Kuoppala 已提交
962

963 964
	/* Wa_1607087056:icl,ehl,jsl */
	if (IS_ICELAKE(i915) ||
965
	    IS_JSL_EHL_GT_STEP(i915, STEP_A0, STEP_B0))
966 967 968
		wa_write_or(wal,
			    SLICE_UNIT_LEVEL_CLKGATE,
			    L3_CLKGATE_DIS | L3_CR2X_CLKGATE_DIS);
969 970 971 972 973 974

	/*
	 * This is not a documented workaround, but rather an optimization
	 * to reduce sampler power.
	 */
	wa_write_clr(wal, GEN10_DFR_RATIO_EN_AND_CHICKEN, DFR_DISABLE);
975 976
}

977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999
/*
 * Though there are per-engine instances of these registers,
 * they retain their value through engine resets and should
 * only be provided on the GT workaround list rather than
 * the engine-specific workaround list.
 */
static void
wa_14011060649(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	struct intel_engine_cs *engine;
	struct intel_gt *gt = &i915->gt;
	int id;

	for_each_engine(engine, gt, id) {
		if (engine->class != VIDEO_DECODE_CLASS ||
		    (engine->instance % 2))
			continue;

		wa_write_or(wal, VDBOX_CGCTL3F10(engine->mmio_base),
			    IECPUNIT_CLKGATE_DIS);
	}
}

1000
static void
1001 1002
gen12_gt_workarounds_init(struct drm_i915_private *i915,
			  struct i915_wa_list *wal)
1003
{
1004
	icl_wa_init_mcr(i915, wal);
1005

1006
	/* Wa_14011060649:tgl,rkl,dg1,adl-s,adl-p */
1007
	wa_14011060649(i915, wal);
1008 1009 1010

	/* Wa_14011059788:tgl,rkl,adl-s,dg1,adl-p */
	wa_write_or(wal, GEN10_DFR_RATIO_EN_AND_CHICKEN, DFR_DISABLE);
1011 1012 1013 1014 1015 1016
}

static void
tgl_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	gen12_gt_workarounds_init(i915, wal);
1017

M
Mika Kuoppala 已提交
1018
	/* Wa_1409420604:tgl */
1019
	if (IS_TGL_UY_GT_STEP(i915, STEP_A0, STEP_B0))
M
Mika Kuoppala 已提交
1020 1021 1022
		wa_write_or(wal,
			    SUBSLICE_UNIT_LEVEL_CLKGATE2,
			    CPSSUNIT_CLKGATE_DIS);
M
Mika Kuoppala 已提交
1023

1024
	/* Wa_1607087056:tgl also know as BUG:1409180338 */
1025
	if (IS_TGL_UY_GT_STEP(i915, STEP_A0, STEP_B0))
M
Mika Kuoppala 已提交
1026 1027 1028
		wa_write_or(wal,
			    SLICE_UNIT_LEVEL_CLKGATE,
			    L3_CLKGATE_DIS | L3_CR2X_CLKGATE_DIS);
1029 1030

	/* Wa_1408615072:tgl[a0] */
1031
	if (IS_TGL_UY_GT_STEP(i915, STEP_A0, STEP_B0))
1032 1033
		wa_write_or(wal, UNSLICE_UNIT_LEVEL_CLKGATE2,
			    VSUNIT_CLKGATE_DIS_TGL);
1034 1035
}

1036 1037 1038 1039 1040 1041
static void
dg1_gt_workarounds_init(struct drm_i915_private *i915, struct i915_wa_list *wal)
{
	gen12_gt_workarounds_init(i915, wal);

	/* Wa_1607087056:dg1 */
1042
	if (IS_DG1_GT_STEP(i915, STEP_A0, STEP_B0))
1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059
		wa_write_or(wal,
			    SLICE_UNIT_LEVEL_CLKGATE,
			    L3_CLKGATE_DIS | L3_CR2X_CLKGATE_DIS);

	/* Wa_1409420604:dg1 */
	if (IS_DG1(i915))
		wa_write_or(wal,
			    SUBSLICE_UNIT_LEVEL_CLKGATE2,
			    CPSSUNIT_CLKGATE_DIS);

	/* Wa_1408615072:dg1 */
	/* Empirical testing shows this register is unaffected by engine reset. */
	if (IS_DG1(i915))
		wa_write_or(wal, UNSLICE_UNIT_LEVEL_CLKGATE2,
			    VSUNIT_CLKGATE_DIS_TGL);
}

1060 1061
static void
gt_init_workarounds(struct drm_i915_private *i915, struct i915_wa_list *wal)
1062
{
1063 1064 1065
	if (IS_DG1(i915))
		dg1_gt_workarounds_init(i915, wal);
	else if (IS_TIGERLAKE(i915))
1066
		tgl_gt_workarounds_init(i915, wal);
1067
	else if (GRAPHICS_VER(i915) == 12)
1068
		gen12_gt_workarounds_init(i915, wal);
1069
	else if (GRAPHICS_VER(i915) == 11)
1070
		icl_gt_workarounds_init(i915, wal);
1071
	else if (IS_COFFEELAKE(i915) || IS_COMETLAKE(i915))
1072 1073 1074 1075 1076 1077
		cfl_gt_workarounds_init(i915, wal);
	else if (IS_GEMINILAKE(i915))
		glk_gt_workarounds_init(i915, wal);
	else if (IS_KABYLAKE(i915))
		kbl_gt_workarounds_init(i915, wal);
	else if (IS_BROXTON(i915))
1078
		gen9_gt_workarounds_init(i915, wal);
1079 1080
	else if (IS_SKYLAKE(i915))
		skl_gt_workarounds_init(i915, wal);
1081 1082
	else if (IS_HASWELL(i915))
		hsw_gt_workarounds_init(i915, wal);
1083 1084
	else if (IS_VALLEYVIEW(i915))
		vlv_gt_workarounds_init(i915, wal);
1085 1086
	else if (IS_IVYBRIDGE(i915))
		ivb_gt_workarounds_init(i915, wal);
1087
	else if (GRAPHICS_VER(i915) == 6)
1088
		snb_gt_workarounds_init(i915, wal);
1089
	else if (GRAPHICS_VER(i915) == 5)
1090
		ilk_gt_workarounds_init(i915, wal);
1091 1092
	else if (IS_G4X(i915))
		g4x_gt_workarounds_init(i915, wal);
1093
	else if (GRAPHICS_VER(i915) == 4)
1094
		gen4_gt_workarounds_init(i915, wal);
1095
	else if (GRAPHICS_VER(i915) <= 8)
1096
		;
1097
	else
1098
		MISSING_CASE(GRAPHICS_VER(i915));
1099 1100 1101 1102 1103
}

void intel_gt_init_workarounds(struct drm_i915_private *i915)
{
	struct i915_wa_list *wal = &i915->gt_wa_list;
1104

1105
	wa_init_start(wal, "GT", "global");
1106
	gt_init_workarounds(i915, wal);
1107 1108 1109 1110
	wa_init_finish(wal);
}

static enum forcewake_domains
1111
wal_get_fw_for_rmw(struct intel_uncore *uncore, const struct i915_wa_list *wal)
1112 1113 1114 1115 1116 1117
{
	enum forcewake_domains fw = 0;
	struct i915_wa *wa;
	unsigned int i;

	for (i = 0, wa = wal->list; i < wal->count; i++, wa++)
1118
		fw |= intel_uncore_forcewake_for_reg(uncore,
1119 1120 1121 1122 1123 1124 1125
						     wa->reg,
						     FW_REG_READ |
						     FW_REG_WRITE);

	return fw;
}

1126 1127 1128
static bool
wa_verify(const struct i915_wa *wa, u32 cur, const char *name, const char *from)
{
1129
	if ((cur ^ wa->set) & wa->read) {
1130
		DRM_ERROR("%s workaround lost on %s! (reg[%x]=0x%x, relevant bits were 0x%x vs expected 0x%x)\n",
1131
			  name, from, i915_mmio_reg_offset(wa->reg),
1132
			  cur, cur & wa->read, wa->set & wa->read);
1133 1134 1135 1136 1137 1138 1139

		return false;
	}

	return true;
}

1140
static void
1141
wa_list_apply(struct intel_gt *gt, const struct i915_wa_list *wal)
1142
{
1143
	struct intel_uncore *uncore = gt->uncore;
1144 1145 1146 1147 1148 1149 1150 1151
	enum forcewake_domains fw;
	unsigned long flags;
	struct i915_wa *wa;
	unsigned int i;

	if (!wal->count)
		return;

1152
	fw = wal_get_fw_for_rmw(uncore, wal);
1153

1154 1155
	spin_lock_irqsave(&uncore->lock, flags);
	intel_uncore_forcewake_get__locked(uncore, fw);
1156 1157

	for (i = 0, wa = wal->list; i < wal->count; i++, wa++) {
1158 1159 1160 1161 1162 1163 1164 1165
		u32 val, old = 0;

		/* open-coded rmw due to steering */
		old = wa->clr ? intel_gt_read_register_fw(gt, wa->reg) : 0;
		val = (old & ~wa->clr) | wa->set;
		if (val != old || !wa->clr)
			intel_uncore_write_fw(uncore, wa->reg, val);

1166
		if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM))
1167
			wa_verify(wa, intel_gt_read_register_fw(gt, wa->reg),
1168
				  wal->name, "application");
1169 1170
	}

1171 1172
	intel_uncore_forcewake_put__locked(uncore, fw);
	spin_unlock_irqrestore(&uncore->lock, flags);
1173 1174
}

1175
void intel_gt_apply_workarounds(struct intel_gt *gt)
1176
{
1177
	wa_list_apply(gt, &gt->i915->gt_wa_list);
1178 1179
}

1180
static bool wa_list_verify(struct intel_gt *gt,
1181 1182 1183
			   const struct i915_wa_list *wal,
			   const char *from)
{
1184
	struct intel_uncore *uncore = gt->uncore;
1185
	struct i915_wa *wa;
1186 1187
	enum forcewake_domains fw;
	unsigned long flags;
1188 1189 1190
	unsigned int i;
	bool ok = true;

1191 1192 1193 1194 1195
	fw = wal_get_fw_for_rmw(uncore, wal);

	spin_lock_irqsave(&uncore->lock, flags);
	intel_uncore_forcewake_get__locked(uncore, fw);

1196
	for (i = 0, wa = wal->list; i < wal->count; i++, wa++)
1197
		ok &= wa_verify(wa,
1198
				intel_gt_read_register_fw(gt, wa->reg),
1199
				wal->name, from);
1200

1201 1202 1203
	intel_uncore_forcewake_put__locked(uncore, fw);
	spin_unlock_irqrestore(&uncore->lock, flags);

1204 1205 1206
	return ok;
}

1207
bool intel_gt_verify_workarounds(struct intel_gt *gt, const char *from)
1208
{
1209
	return wa_list_verify(gt, &gt->i915->gt_wa_list, from);
1210 1211
}

1212
__maybe_unused
C
Chris Wilson 已提交
1213
static bool is_nonpriv_flags_valid(u32 flags)
1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226
{
	/* Check only valid flag bits are set */
	if (flags & ~RING_FORCE_TO_NONPRIV_MASK_VALID)
		return false;

	/* NB: Only 3 out of 4 enum values are valid for access field */
	if ((flags & RING_FORCE_TO_NONPRIV_ACCESS_MASK) ==
	    RING_FORCE_TO_NONPRIV_ACCESS_INVALID)
		return false;

	return true;
}

1227
static void
1228
whitelist_reg_ext(struct i915_wa_list *wal, i915_reg_t reg, u32 flags)
1229
{
1230 1231 1232
	struct i915_wa wa = {
		.reg = reg
	};
1233

1234 1235
	if (GEM_DEBUG_WARN_ON(wal->count >= RING_MAX_NONPRIV_SLOTS))
		return;
1236

1237 1238 1239
	if (GEM_DEBUG_WARN_ON(!is_nonpriv_flags_valid(flags)))
		return;

1240
	wa.reg.reg |= flags;
1241
	_wa_add(wal, &wa);
1242 1243
}

1244 1245 1246
static void
whitelist_reg(struct i915_wa_list *wal, i915_reg_t reg)
{
1247
	whitelist_reg_ext(wal, reg, RING_FORCE_TO_NONPRIV_ACCESS_RW);
1248 1249
}

1250
static void gen9_whitelist_build(struct i915_wa_list *w)
1251 1252
{
	/* WaVFEStateAfterPipeControlwithMediaStateClear:skl,bxt,glk,cfl */
1253
	whitelist_reg(w, GEN9_CTX_PREEMPT_REG);
1254 1255

	/* WaEnablePreemptionGranularityControlByUMD:skl,bxt,kbl,cfl,[cnl] */
1256
	whitelist_reg(w, GEN8_CS_CHICKEN1);
1257 1258

	/* WaAllowUMDToModifyHDCChicken1:skl,bxt,kbl,glk,cfl */
1259
	whitelist_reg(w, GEN8_HDC_CHICKEN1);
1260 1261 1262

	/* WaSendPushConstantsFromMMIO:skl,bxt */
	whitelist_reg(w, COMMON_SLICE_CHICKEN2);
1263 1264
}

1265
static void skl_whitelist_build(struct intel_engine_cs *engine)
1266
{
1267 1268 1269 1270 1271
	struct i915_wa_list *w = &engine->whitelist;

	if (engine->class != RENDER_CLASS)
		return;

1272
	gen9_whitelist_build(w);
1273 1274

	/* WaDisableLSQCROPERFforOCL:skl */
1275
	whitelist_reg(w, GEN8_L3SQCREG4);
1276 1277
}

1278
static void bxt_whitelist_build(struct intel_engine_cs *engine)
1279
{
1280 1281 1282 1283
	if (engine->class != RENDER_CLASS)
		return;

	gen9_whitelist_build(&engine->whitelist);
1284 1285
}

1286
static void kbl_whitelist_build(struct intel_engine_cs *engine)
1287
{
1288 1289 1290 1291 1292
	struct i915_wa_list *w = &engine->whitelist;

	if (engine->class != RENDER_CLASS)
		return;

1293
	gen9_whitelist_build(w);
1294

1295
	/* WaDisableLSQCROPERFforOCL:kbl */
1296
	whitelist_reg(w, GEN8_L3SQCREG4);
1297 1298
}

1299
static void glk_whitelist_build(struct intel_engine_cs *engine)
1300
{
1301 1302 1303 1304 1305
	struct i915_wa_list *w = &engine->whitelist;

	if (engine->class != RENDER_CLASS)
		return;

1306
	gen9_whitelist_build(w);
1307

1308
	/* WA #0862: Userspace has to set "Barrier Mode" to avoid hangs. */
1309
	whitelist_reg(w, GEN9_SLICE_COMMON_ECO_CHICKEN1);
1310
}
1311

1312
static void cfl_whitelist_build(struct intel_engine_cs *engine)
1313
{
1314 1315
	struct i915_wa_list *w = &engine->whitelist;

1316 1317 1318
	if (engine->class != RENDER_CLASS)
		return;

1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330
	gen9_whitelist_build(w);

	/*
	 * WaAllowPMDepthAndInvocationCountAccessFromUMD:cfl,whl,cml,aml
	 *
	 * This covers 4 register which are next to one another :
	 *   - PS_INVOCATION_COUNT
	 *   - PS_INVOCATION_COUNT_UDW
	 *   - PS_DEPTH_COUNT
	 *   - PS_DEPTH_COUNT_UDW
	 */
	whitelist_reg_ext(w, PS_INVOCATION_COUNT,
1331
			  RING_FORCE_TO_NONPRIV_ACCESS_RD |
1332
			  RING_FORCE_TO_NONPRIV_RANGE_4);
1333 1334
}

1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346
static void cml_whitelist_build(struct intel_engine_cs *engine)
{
	struct i915_wa_list *w = &engine->whitelist;

	if (engine->class != RENDER_CLASS)
		whitelist_reg_ext(w,
				  RING_CTX_TIMESTAMP(engine->mmio_base),
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);

	cfl_whitelist_build(engine);
}

1347
static void icl_whitelist_build(struct intel_engine_cs *engine)
1348
{
1349 1350
	struct i915_wa_list *w = &engine->whitelist;

1351 1352 1353 1354 1355 1356 1357 1358 1359 1360
	switch (engine->class) {
	case RENDER_CLASS:
		/* WaAllowUMDToModifyHalfSliceChicken7:icl */
		whitelist_reg(w, GEN9_HALF_SLICE_CHICKEN7);

		/* WaAllowUMDToModifySamplerMode:icl */
		whitelist_reg(w, GEN10_SAMPLER_MODE);

		/* WaEnableStateCacheRedirectToCS:icl */
		whitelist_reg(w, GEN9_SLICE_COMMON_ECO_CHICKEN1);
1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371

		/*
		 * WaAllowPMDepthAndInvocationCountAccessFromUMD:icl
		 *
		 * This covers 4 register which are next to one another :
		 *   - PS_INVOCATION_COUNT
		 *   - PS_INVOCATION_COUNT_UDW
		 *   - PS_DEPTH_COUNT
		 *   - PS_DEPTH_COUNT_UDW
		 */
		whitelist_reg_ext(w, PS_INVOCATION_COUNT,
1372
				  RING_FORCE_TO_NONPRIV_ACCESS_RD |
1373
				  RING_FORCE_TO_NONPRIV_RANGE_4);
1374 1375 1376 1377 1378
		break;

	case VIDEO_DECODE_CLASS:
		/* hucStatusRegOffset */
		whitelist_reg_ext(w, _MMIO(0x2000 + engine->mmio_base),
1379
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
1380 1381
		/* hucUKernelHdrInfoRegOffset */
		whitelist_reg_ext(w, _MMIO(0x2014 + engine->mmio_base),
1382
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
1383 1384
		/* hucStatus2RegOffset */
		whitelist_reg_ext(w, _MMIO(0x23B0 + engine->mmio_base),
1385
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
1386 1387 1388
		whitelist_reg_ext(w,
				  RING_CTX_TIMESTAMP(engine->mmio_base),
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
1389 1390 1391
		break;

	default:
1392 1393 1394
		whitelist_reg_ext(w,
				  RING_CTX_TIMESTAMP(engine->mmio_base),
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
1395 1396
		break;
	}
1397 1398
}

1399 1400
static void tgl_whitelist_build(struct intel_engine_cs *engine)
{
1401 1402 1403 1404 1405 1406
	struct i915_wa_list *w = &engine->whitelist;

	switch (engine->class) {
	case RENDER_CLASS:
		/*
		 * WaAllowPMDepthAndInvocationCountAccessFromUMD:tgl
1407
		 * Wa_1408556865:tgl
1408 1409 1410 1411 1412 1413 1414 1415 1416 1417
		 *
		 * This covers 4 registers which are next to one another :
		 *   - PS_INVOCATION_COUNT
		 *   - PS_INVOCATION_COUNT_UDW
		 *   - PS_DEPTH_COUNT
		 *   - PS_DEPTH_COUNT_UDW
		 */
		whitelist_reg_ext(w, PS_INVOCATION_COUNT,
				  RING_FORCE_TO_NONPRIV_ACCESS_RD |
				  RING_FORCE_TO_NONPRIV_RANGE_4);
1418 1419 1420

		/* Wa_1808121037:tgl */
		whitelist_reg(w, GEN7_COMMON_SLICE_CHICKEN1);
1421 1422 1423

		/* Wa_1806527549:tgl */
		whitelist_reg(w, HIZ_CHICKEN);
1424 1425
		break;
	default:
1426 1427 1428
		whitelist_reg_ext(w,
				  RING_CTX_TIMESTAMP(engine->mmio_base),
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
1429 1430
		break;
	}
1431 1432
}

1433 1434 1435 1436 1437 1438 1439
static void dg1_whitelist_build(struct intel_engine_cs *engine)
{
	struct i915_wa_list *w = &engine->whitelist;

	tgl_whitelist_build(engine);

	/* GEN:BUG:1409280441:dg1 */
1440
	if (IS_DG1_GT_STEP(engine->i915, STEP_A0, STEP_B0) &&
1441 1442 1443 1444 1445 1446
	    (engine->class == RENDER_CLASS ||
	     engine->class == COPY_ENGINE_CLASS))
		whitelist_reg_ext(w, RING_ID(engine->mmio_base),
				  RING_FORCE_TO_NONPRIV_ACCESS_RD);
}

1447
void intel_engine_init_whitelist(struct intel_engine_cs *engine)
1448 1449
{
	struct drm_i915_private *i915 = engine->i915;
1450
	struct i915_wa_list *w = &engine->whitelist;
1451

1452
	wa_init_start(w, "whitelist", engine->name);
1453

1454 1455
	if (IS_DG1(i915))
		dg1_whitelist_build(engine);
1456
	else if (GRAPHICS_VER(i915) == 12)
1457
		tgl_whitelist_build(engine);
1458
	else if (GRAPHICS_VER(i915) == 11)
1459
		icl_whitelist_build(engine);
1460 1461 1462
	else if (IS_COMETLAKE(i915))
		cml_whitelist_build(engine);
	else if (IS_COFFEELAKE(i915))
1463
		cfl_whitelist_build(engine);
1464
	else if (IS_GEMINILAKE(i915))
1465
		glk_whitelist_build(engine);
1466
	else if (IS_KABYLAKE(i915))
1467
		kbl_whitelist_build(engine);
1468
	else if (IS_BROXTON(i915))
1469
		bxt_whitelist_build(engine);
1470
	else if (IS_SKYLAKE(i915))
1471
		skl_whitelist_build(engine);
1472
	else if (GRAPHICS_VER(i915) <= 8)
1473
		;
1474
	else
1475
		MISSING_CASE(GRAPHICS_VER(i915));
1476

1477
	wa_init_finish(w);
1478 1479
}

1480
void intel_engine_apply_whitelist(struct intel_engine_cs *engine)
1481
{
1482
	const struct i915_wa_list *wal = &engine->whitelist;
1483
	struct intel_uncore *uncore = engine->uncore;
1484
	const u32 base = engine->mmio_base;
1485
	struct i915_wa *wa;
1486 1487
	unsigned int i;

1488
	if (!wal->count)
1489
		return;
1490

1491
	for (i = 0, wa = wal->list; i < wal->count; i++, wa++)
1492 1493 1494
		intel_uncore_write(uncore,
				   RING_FORCE_TO_NONPRIV(base, i),
				   i915_mmio_reg_offset(wa->reg));
1495

1496 1497
	/* And clear the rest just in case of garbage */
	for (; i < RING_MAX_NONPRIV_SLOTS; i++)
1498 1499 1500
		intel_uncore_write(uncore,
				   RING_FORCE_TO_NONPRIV(base, i),
				   i915_mmio_reg_offset(RING_NOPID(base)));
1501 1502
}

1503 1504
static void
rcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal)
1505 1506 1507
{
	struct drm_i915_private *i915 = engine->i915;

1508 1509
	if (IS_DG1_GT_STEP(i915, STEP_A0, STEP_B0) ||
	    IS_TGL_UY_GT_STEP(i915, STEP_A0, STEP_B0)) {
1510
		/*
1511 1512
		 * Wa_1607138336:tgl[a0],dg1[a0]
		 * Wa_1607063988:tgl[a0],dg1[a0]
1513
		 */
M
Mika Kuoppala 已提交
1514 1515 1516
		wa_write_or(wal,
			    GEN9_CTX_PREEMPT_REG,
			    GEN12_DISABLE_POSH_BUSY_FF_DOP_CG);
1517
	}
1518

1519
	if (IS_TGL_UY_GT_STEP(i915, STEP_A0, STEP_B0)) {
R
Radhakrishna Sripada 已提交
1520 1521 1522 1523 1524 1525 1526
		/*
		 * Wa_1606679103:tgl
		 * (see also Wa_1606682166:icl)
		 */
		wa_write_or(wal,
			    GEN7_SARCHKMD,
			    GEN7_DISABLE_SAMPLER_PREFETCH);
1527 1528
	}

1529
	if (IS_ALDERLAKE_P(i915) || IS_ALDERLAKE_S(i915) || IS_DG1(i915) ||
1530
	    IS_ROCKETLAKE(i915) || IS_TIGERLAKE(i915)) {
1531
		/* Wa_1606931601:tgl,rkl,dg1,adl-s,adl-p */
1532 1533
		wa_masked_en(wal, GEN7_ROW_CHICKEN2, GEN12_DISABLE_EARLY_READ);

1534 1535 1536 1537
		/*
		 * Wa_1407928979:tgl A*
		 * Wa_18011464164:tgl[B0+],dg1[B0+]
		 * Wa_22010931296:tgl[B0+],dg1[B0+]
1538
		 * Wa_14010919138:rkl,dg1,adl-s,adl-p
1539 1540 1541
		 */
		wa_write_or(wal, GEN7_FF_THREAD_MODE,
			    GEN12_FF_TESSELATION_DOP_GATE_DISABLE);
1542 1543

		/*
1544 1545 1546
		 * Wa_1606700617:tgl,dg1,adl-p
		 * Wa_22010271021:tgl,rkl,dg1,adl-s,adl-p
		 * Wa_14010826681:tgl,dg1,rkl,adl-p
1547 1548 1549 1550
		 */
		wa_masked_en(wal,
			     GEN9_CS_DEBUG_MODE1,
			     FF_DOP_CLOCK_GATE_DISABLE);
1551 1552
	}

1553
	if (IS_ALDERLAKE_P(i915) || IS_ALDERLAKE_S(i915) ||
1554
	    IS_DG1_GT_STEP(i915, STEP_A0, STEP_B0) ||
1555
	    IS_ROCKETLAKE(i915) || IS_TIGERLAKE(i915)) {
1556
		/* Wa_1409804808:tgl,rkl,dg1[a0],adl-s,adl-p */
1557 1558
		wa_masked_en(wal, GEN7_ROW_CHICKEN2,
			     GEN12_PUSH_CONST_DEREF_HOLD_DIS);
1559

1560 1561
		/*
		 * Wa_1409085225:tgl
1562
		 * Wa_14010229206:tgl,rkl,dg1[a0],adl-s,adl-p
1563 1564
		 */
		wa_masked_en(wal, GEN9_ROW_CHICKEN4, GEN12_DISABLE_TDL_PUSH);
1565 1566
	}

1567

1568
	if (IS_DG1_GT_STEP(i915, STEP_A0, STEP_B0) ||
1569
	    IS_ROCKETLAKE(i915) || IS_TIGERLAKE(i915)) {
1570 1571 1572
		/*
		 * Wa_1607030317:tgl
		 * Wa_1607186500:tgl
1573 1574 1575 1576 1577 1578
		 * Wa_1607297627:tgl,rkl,dg1[a0]
		 *
		 * On TGL and RKL there are multiple entries for this WA in the
		 * BSpec; some indicate this is an A0-only WA, others indicate
		 * it applies to all steppings so we trust the "all steppings."
		 * For DG1 this only applies to A0.
1579 1580 1581 1582 1583
		 */
		wa_masked_en(wal,
			     GEN6_RC_SLEEP_PSMI_CONTROL,
			     GEN12_WAIT_FOR_EVENT_POWER_DOWN_DISABLE |
			     GEN8_RC_SEMA_IDLE_MSG_DISABLE);
1584 1585
	}

1586
	if (IS_DG1(i915) || IS_ROCKETLAKE(i915) || IS_TIGERLAKE(i915) ||
1587 1588
	    IS_ALDERLAKE_S(i915) || IS_ALDERLAKE_P(i915)) {
		/* Wa_1406941453:tgl,rkl,dg1,adl-s,adl-p */
1589 1590 1591 1592 1593
		wa_masked_en(wal,
			     GEN10_SAMPLER_MODE,
			     ENABLE_SMALLPL);
	}

1594
	if (GRAPHICS_VER(i915) == 11) {
1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611
		/* This is not an Wa. Enable for better image quality */
		wa_masked_en(wal,
			     _3D_CHICKEN3,
			     _3D_CHICKEN3_AA_LINE_QUALITY_FIX_ENABLE);

		/*
		 * Wa_1405543622:icl
		 * Formerly known as WaGAPZPriorityScheme
		 */
		wa_write_or(wal,
			    GEN8_GARBCNTL,
			    GEN11_ARBITRATION_PRIO_ORDER_MASK);

		/*
		 * Wa_1604223664:icl
		 * Formerly known as WaL3BankAddressHashing
		 */
1612 1613 1614 1615 1616 1617 1618 1619
		wa_write_clr_set(wal,
				 GEN8_GARBCNTL,
				 GEN11_HASH_CTRL_EXCL_MASK,
				 GEN11_HASH_CTRL_EXCL_BIT0);
		wa_write_clr_set(wal,
				 GEN11_GLBLINVL,
				 GEN11_BANK_HASH_ADDR_EXCL_MASK,
				 GEN11_BANK_HASH_ADDR_EXCL_BIT0);
1620 1621 1622 1623 1624

		/*
		 * Wa_1405733216:icl
		 * Formerly known as WaDisableCleanEvicts
		 */
1625 1626 1627
		wa_write_or(wal,
			    GEN8_L3SQCREG4,
			    GEN11_LQSC_CLEAN_EVICT_DISABLE);
1628

1629 1630 1631 1632
		/* Wa_1606682166:icl */
		wa_write_or(wal,
			    GEN7_SARCHKMD,
			    GEN7_DISABLE_SAMPLER_PREFETCH);
T
Tvrtko Ursulin 已提交
1633 1634

		/* Wa_1409178092:icl */
1635 1636 1637 1638
		wa_write_clr_set(wal,
				 GEN11_SCRATCH2,
				 GEN11_COHERENT_PARTIAL_WRITE_MERGE_ENABLE,
				 0);
1639 1640 1641 1642 1643 1644 1645 1646 1647

		/* WaEnable32PlaneMode:icl */
		wa_masked_en(wal, GEN9_CSFE_CHICKEN1_RCS,
			     GEN11_ENABLE_32_PLANE_MODE);

		/*
		 * Wa_1408615072:icl,ehl  (vsunit)
		 * Wa_1407596294:icl,ehl  (hsunit)
		 */
1648 1649
		wa_write_or(wal, UNSLICE_UNIT_LEVEL_CLKGATE,
			    VSUNIT_CLKGATE_DIS | HSUNIT_CLKGATE_DIS);
1650 1651

		/* Wa_1407352427:icl,ehl */
1652 1653
		wa_write_or(wal, UNSLICE_UNIT_LEVEL_CLKGATE2,
			    PSDUNIT_CLKGATE_DIS);
1654 1655 1656 1657 1658

		/* Wa_1406680159:icl,ehl */
		wa_write_or(wal,
			    SUBSLICE_UNIT_LEVEL_CLKGATE,
			    GWUNIT_CLKGATE_DIS);
1659 1660 1661 1662 1663 1664 1665 1666

		/*
		 * Wa_1408767742:icl[a2..forever],ehl[all]
		 * Wa_1605460711:icl[a0..c0]
		 */
		wa_write_or(wal,
			    GEN7_FF_THREAD_MODE,
			    GEN12_FF_TESSELATION_DOP_GATE_DISABLE);
M
Matt Atwood 已提交
1667

1668 1669 1670 1671
		/* Wa_22010271021 */
		wa_masked_en(wal,
			     GEN9_CS_DEBUG_MODE1,
			     FF_DOP_CLOCK_GATE_DISABLE);
1672 1673
	}

1674
	if (IS_GRAPHICS_VER(i915, 9, 12)) {
1675
		/* FtrPerCtxtPreemptionGranularityControl:skl,bxt,kbl,cfl,cnl,icl,tgl */
1676 1677 1678 1679 1680
		wa_masked_en(wal,
			     GEN7_FF_SLICE_CS_CHICKEN1,
			     GEN9_FFSC_PERCTX_PREEMPT_CTRL);
	}

1681 1682 1683 1684
	if (IS_SKYLAKE(i915) ||
	    IS_KABYLAKE(i915) ||
	    IS_COFFEELAKE(i915) ||
	    IS_COMETLAKE(i915)) {
1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697
		/* WaEnableGapsTsvCreditFix:skl,kbl,cfl */
		wa_write_or(wal,
			    GEN8_GARBCNTL,
			    GEN9_GAPS_TSV_CREDIT_DISABLE);
	}

	if (IS_BROXTON(i915)) {
		/* WaDisablePooledEuLoadBalancingFix:bxt */
		wa_masked_en(wal,
			     FF_SLICE_CS_CHICKEN2,
			     GEN9_POOLED_EU_LOAD_BALANCING_FIX_DISABLE);
	}

1698
	if (GRAPHICS_VER(i915) == 9) {
1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710
		/* WaContextSwitchWithConcurrentTLBInvalidate:skl,bxt,kbl,glk,cfl */
		wa_masked_en(wal,
			     GEN9_CSFE_CHICKEN1_RCS,
			     GEN9_PREEMPT_GPGPU_SYNC_SWITCH_DISABLE);

		/* WaEnableLbsSlaRetryTimerDecrement:skl,bxt,kbl,glk,cfl */
		wa_write_or(wal,
			    BDW_SCRATCH1,
			    GEN9_LBS_SLA_RETRY_TIMER_DECREMENT_ENABLE);

		/* WaProgramL3SqcReg1DefaultForPerf:bxt,glk */
		if (IS_GEN9_LP(i915))
1711 1712 1713 1714 1715
			wa_write_clr_set(wal,
					 GEN8_L3SQCREG1,
					 L3_PRIO_CREDITS_MASK,
					 L3_GENERAL_PRIO_CREDITS(62) |
					 L3_HIGH_PRIO_CREDITS(2));
1716 1717 1718 1719 1720

		/* WaOCLCoherentLineFlush:skl,bxt,kbl,cfl */
		wa_write_or(wal,
			    GEN8_L3SQCREG4,
			    GEN8_LQSC_FLUSH_COHERENT_LINES);
1721 1722 1723 1724 1725 1726 1727 1728

		/* Disable atomics in L3 to prevent unrecoverable hangs */
		wa_write_clr_set(wal, GEN9_SCRATCH_LNCF1,
				 GEN9_LNCF_NONIA_COHERENT_ATOMICS_ENABLE, 0);
		wa_write_clr_set(wal, GEN8_L3SQCREG4,
				 GEN8_LQSQ_NONIA_COHERENT_ATOMICS_ENABLE, 0);
		wa_write_clr_set(wal, GEN9_SCRATCH1,
				 EVICTION_PERF_FIX_ENABLE, 0);
1729
	}
1730

1731 1732 1733 1734 1735 1736 1737 1738 1739
	if (IS_HASWELL(i915)) {
		/* WaSampleCChickenBitEnable:hsw */
		wa_masked_en(wal,
			     HALF_SLICE_CHICKEN3, HSW_SAMPLE_C_PERFORMANCE);

		wa_masked_dis(wal,
			      CACHE_MODE_0_GEN7,
			      /* enable HiZ Raw Stall Optimization */
			      HIZ_RAW_STALL_OPT_DISABLE);
1740 1741 1742 1743 1744 1745 1746
	}

	if (IS_VALLEYVIEW(i915)) {
		/* WaDisableEarlyCull:vlv */
		wa_masked_en(wal,
			     _3D_CHICKEN3,
			     _3D_CHICKEN_SF_DISABLE_OBJEND_CULL);
1747 1748

		/*
1749
		 * WaVSThreadDispatchOverride:ivb,vlv
1750
		 *
1751 1752
		 * This actually overrides the dispatch
		 * mode for all thread types.
1753
		 */
1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766
		wa_write_clr_set(wal,
				 GEN7_FF_THREAD_MODE,
				 GEN7_FF_SCHED_MASK,
				 GEN7_FF_TS_SCHED_HW |
				 GEN7_FF_VS_SCHED_HW |
				 GEN7_FF_DS_SCHED_HW);

		/* WaPsdDispatchEnable:vlv */
		/* WaDisablePSDDualDispatchEnable:vlv */
		wa_masked_en(wal,
			     GEN7_HALF_SLICE_CHICKEN1,
			     GEN7_MAX_PS_THREAD_DEP |
			     GEN7_PSD_SINGLE_PORT_DISPATCH_ENABLE);
1767 1768
	}

1769 1770
	if (IS_IVYBRIDGE(i915)) {
		/* WaDisableEarlyCull:ivb */
1771 1772 1773 1774
		wa_masked_en(wal,
			     _3D_CHICKEN3,
			     _3D_CHICKEN_SF_DISABLE_OBJEND_CULL);

1775 1776 1777 1778 1779 1780 1781
		if (0) { /* causes HiZ corruption on ivb:gt1 */
			/* enable HiZ Raw Stall Optimization */
			wa_masked_dis(wal,
				      CACHE_MODE_0_GEN7,
				      HIZ_RAW_STALL_OPT_DISABLE);
		}

1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794
		/*
		 * WaVSThreadDispatchOverride:ivb,vlv
		 *
		 * This actually overrides the dispatch
		 * mode for all thread types.
		 */
		wa_write_clr_set(wal,
				 GEN7_FF_THREAD_MODE,
				 GEN7_FF_SCHED_MASK,
				 GEN7_FF_TS_SCHED_HW |
				 GEN7_FF_VS_SCHED_HW |
				 GEN7_FF_DS_SCHED_HW);

1795 1796 1797 1798 1799 1800 1801
		/* WaDisablePSDDualDispatchEnable:ivb */
		if (IS_IVB_GT1(i915))
			wa_masked_en(wal,
				     GEN7_HALF_SLICE_CHICKEN1,
				     GEN7_PSD_SINGLE_PORT_DISPATCH_ENABLE);
	}

1802
	if (GRAPHICS_VER(i915) == 7) {
1803 1804 1805 1806 1807 1808
		/* WaBCSVCSTlbInvalidationMode:ivb,vlv,hsw */
		wa_masked_en(wal,
			     GFX_MODE_GEN7,
			     GFX_TLB_INVALIDATE_EXPLICIT | GFX_REPLAY_MODE);

		/* WaDisable_RenderCache_OperationalFlush:ivb,vlv,hsw */
1809 1810 1811 1812
		wa_masked_dis(wal, CACHE_MODE_0_GEN7, RC_OP_FLUSH_ENABLE);

		/*
		 * BSpec says this must be set, even though
1813
		 * WaDisable4x2SubspanOptimization:ivb,hsw
1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827
		 * WaDisable4x2SubspanOptimization isn't listed for VLV.
		 */
		wa_masked_en(wal,
			     CACHE_MODE_1,
			     PIXEL_SUBSPAN_COLLECT_OPT_DISABLE);

		/*
		 * BSpec recommends 8x4 when MSAA is used,
		 * however in practice 16x4 seems fastest.
		 *
		 * Note that PS/WM thread counts depend on the WIZ hashing
		 * disable bit, which we don't touch here, but it's good
		 * to keep in mind (see 3DSTATE_PS and 3DSTATE_WM).
		 */
1828 1829 1830 1831
		wa_masked_field_set(wal,
				    GEN7_GT_MODE,
				    GEN6_WIZ_HASHING_MASK,
				    GEN6_WIZ_HASHING_16x4);
1832 1833
	}

1834
	if (IS_GRAPHICS_VER(i915, 6, 7))
1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845
		/*
		 * We need to disable the AsyncFlip performance optimisations in
		 * order to use MI_WAIT_FOR_EVENT within the CS. It should
		 * already be programmed to '1' on all products.
		 *
		 * WaDisableAsyncFlipPerfMode:snb,ivb,hsw,vlv
		 */
		wa_masked_en(wal,
			     MI_MODE,
			     ASYNC_FLIP_PERF_DISABLE);

1846
	if (GRAPHICS_VER(i915) == 6) {
1847 1848 1849 1850 1851 1852 1853 1854 1855
		/*
		 * Required for the hardware to program scanline values for
		 * waiting
		 * WaEnableFlushTlbInvalidationMode:snb
		 */
		wa_masked_en(wal,
			     GFX_MODE,
			     GFX_TLB_INVALIDATE_EXPLICIT);

1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880
		/* WaDisableHiZPlanesWhenMSAAEnabled:snb */
		wa_masked_en(wal,
			     _3D_CHICKEN,
			     _3D_CHICKEN_HIZ_PLANE_DISABLE_MSAA_4X_SNB);

		wa_masked_en(wal,
			     _3D_CHICKEN3,
			     /* WaStripsFansDisableFastClipPerformanceFix:snb */
			     _3D_CHICKEN3_SF_DISABLE_FASTCLIP_CULL |
			     /*
			      * Bspec says:
			      * "This bit must be set if 3DSTATE_CLIP clip mode is set
			      * to normal and 3DSTATE_SF number of SF output attributes
			      * is more than 16."
			      */
			     _3D_CHICKEN3_SF_DISABLE_PIPELINED_ATTR_FETCH);

		/*
		 * BSpec recommends 8x4 when MSAA is used,
		 * however in practice 16x4 seems fastest.
		 *
		 * Note that PS/WM thread counts depend on the WIZ hashing
		 * disable bit, which we don't touch here, but it's good
		 * to keep in mind (see 3DSTATE_PS and 3DSTATE_WM).
		 */
1881 1882 1883 1884
		wa_masked_field_set(wal,
				    GEN6_GT_MODE,
				    GEN6_WIZ_HASHING_MASK,
				    GEN6_WIZ_HASHING_16x4);
1885 1886 1887 1888

		/* WaDisable_RenderCache_OperationalFlush:snb */
		wa_masked_dis(wal, CACHE_MODE_0, RC_OP_FLUSH_ENABLE);

1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899
		/*
		 * From the Sandybridge PRM, volume 1 part 3, page 24:
		 * "If this bit is set, STCunit will have LRA as replacement
		 *  policy. [...] This bit must be reset. LRA replacement
		 *  policy is not supported."
		 */
		wa_masked_dis(wal,
			      CACHE_MODE_0,
			      CM0_STC_EVICT_DISABLE_LRA_SNB);
	}

1900
	if (IS_GRAPHICS_VER(i915, 4, 6))
1901 1902 1903 1904
		/* WaTimedSingleVertexDispatch:cl,bw,ctg,elk,ilk,snb */
		wa_add(wal, MI_MODE,
		       0, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH),
		       /* XXX bit doesn't stick on Broadwater */
1905
		       IS_I965G(i915) ? 0 : VS_TIMER_DISPATCH, true);
1906

1907
	if (GRAPHICS_VER(i915) == 4)
1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919
		/*
		 * Disable CONSTANT_BUFFER before it is loaded from the context
		 * image. For as it is loaded, it is executed and the stored
		 * address may no longer be valid, leading to a GPU hang.
		 *
		 * This imposes the requirement that userspace reload their
		 * CONSTANT_BUFFER on every batch, fortunately a requirement
		 * they are already accustomed to from before contexts were
		 * enabled.
		 */
		wa_add(wal, ECOSKPD,
		       0, _MASKED_BIT_ENABLE(ECO_CONSTANT_BUFFER_SR_DISABLE),
1920 1921
		       0 /* XXX bit doesn't stick on Broadwater */,
		       true);
1922 1923
}

1924 1925
static void
xcs_engine_wa_init(struct intel_engine_cs *engine, struct i915_wa_list *wal)
1926 1927 1928 1929
{
	struct drm_i915_private *i915 = engine->i915;

	/* WaKBLVECSSemaphoreWaitPoll:kbl */
1930
	if (IS_KBL_GT_STEP(i915, STEP_A0, STEP_F0)) {
1931 1932 1933 1934 1935 1936
		wa_write(wal,
			 RING_SEMA_WAIT_POLL(engine->mmio_base),
			 1);
	}
}

1937 1938 1939
static void
engine_init_workarounds(struct intel_engine_cs *engine, struct i915_wa_list *wal)
{
1940
	if (I915_SELFTEST_ONLY(GRAPHICS_VER(engine->i915) < 4))
1941 1942
		return;

1943
	if (engine->class == RENDER_CLASS)
1944 1945 1946 1947 1948
		rcs_engine_wa_init(engine, wal);
	else
		xcs_engine_wa_init(engine, wal);
}

1949 1950 1951 1952
void intel_engine_init_workarounds(struct intel_engine_cs *engine)
{
	struct i915_wa_list *wal = &engine->wa_list;

1953
	if (GRAPHICS_VER(engine->i915) < 4)
1954 1955
		return;

1956
	wa_init_start(wal, "engine", engine->name);
1957
	engine_init_workarounds(engine, wal);
1958 1959 1960 1961 1962
	wa_init_finish(wal);
}

void intel_engine_apply_workarounds(struct intel_engine_cs *engine)
{
1963
	wa_list_apply(engine->gt, &engine->wa_list);
1964 1965
}

1966
struct mcr_range {
M
Matt Roper 已提交
1967 1968
	u32 start;
	u32 end;
1969 1970 1971
};

static const struct mcr_range mcr_ranges_gen8[] = {
M
Matt Roper 已提交
1972 1973 1974 1975 1976 1977 1978 1979
	{ .start = 0x5500, .end = 0x55ff },
	{ .start = 0x7000, .end = 0x7fff },
	{ .start = 0x9400, .end = 0x97ff },
	{ .start = 0xb000, .end = 0xb3ff },
	{ .start = 0xe000, .end = 0xe7ff },
	{},
};

1980 1981 1982 1983 1984 1985 1986 1987 1988
static const struct mcr_range mcr_ranges_gen12[] = {
	{ .start =  0x8150, .end =  0x815f },
	{ .start =  0x9520, .end =  0x955f },
	{ .start =  0xb100, .end =  0xb3ff },
	{ .start =  0xde80, .end =  0xe8ff },
	{ .start = 0x24a00, .end = 0x24a7f },
	{},
};

1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004
static const struct mcr_range mcr_ranges_xehp[] = {
	{ .start =  0x4000, .end =  0x4aff },
	{ .start =  0x5200, .end =  0x52ff },
	{ .start =  0x5400, .end =  0x7fff },
	{ .start =  0x8140, .end =  0x815f },
	{ .start =  0x8c80, .end =  0x8dff },
	{ .start =  0x94d0, .end =  0x955f },
	{ .start =  0x9680, .end =  0x96ff },
	{ .start =  0xb000, .end =  0xb3ff },
	{ .start =  0xc800, .end =  0xcfff },
	{ .start =  0xd800, .end =  0xd8ff },
	{ .start =  0xdc00, .end =  0xffff },
	{ .start = 0x17000, .end = 0x17fff },
	{ .start = 0x24a00, .end = 0x24a7f },
};

2005 2006
static bool mcr_range(struct drm_i915_private *i915, u32 offset)
{
2007
	const struct mcr_range *mcr_ranges;
M
Matt Roper 已提交
2008 2009
	int i;

2010 2011 2012
	if (GRAPHICS_VER_FULL(i915) >= IP_VER(12, 50))
		mcr_ranges = mcr_ranges_xehp;
	else if (GRAPHICS_VER(i915) >= 12)
2013
		mcr_ranges = mcr_ranges_gen12;
2014
	else if (GRAPHICS_VER(i915) >= 8)
2015 2016
		mcr_ranges = mcr_ranges_gen8;
	else
M
Matt Roper 已提交
2017 2018
		return false;

2019
	/*
M
Matt Roper 已提交
2020
	 * Registers in these ranges are affected by the MCR selector
2021 2022 2023
	 * which only controls CPU initiated MMIO. Routing does not
	 * work for CS access so we cannot verify them on this path.
	 */
2024 2025 2026
	for (i = 0; mcr_ranges[i].start; i++)
		if (offset >= mcr_ranges[i].start &&
		    offset <= mcr_ranges[i].end)
M
Matt Roper 已提交
2027
			return true;
2028 2029 2030 2031

	return false;
}

2032 2033 2034 2035 2036
static int
wa_list_srm(struct i915_request *rq,
	    const struct i915_wa_list *wal,
	    struct i915_vma *vma)
{
2037
	struct drm_i915_private *i915 = rq->engine->i915;
2038
	unsigned int i, count = 0;
2039 2040 2041 2042
	const struct i915_wa *wa;
	u32 srm, *cs;

	srm = MI_STORE_REGISTER_MEM | MI_SRM_LRM_GLOBAL_GTT;
2043
	if (GRAPHICS_VER(i915) >= 8)
2044 2045
		srm++;

2046 2047 2048 2049 2050 2051
	for (i = 0, wa = wal->list; i < wal->count; i++, wa++) {
		if (!mcr_range(i915, i915_mmio_reg_offset(wa->reg)))
			count++;
	}

	cs = intel_ring_begin(rq, 4 * count);
2052 2053 2054 2055
	if (IS_ERR(cs))
		return PTR_ERR(cs);

	for (i = 0, wa = wal->list; i < wal->count; i++, wa++) {
2056 2057 2058 2059 2060
		u32 offset = i915_mmio_reg_offset(wa->reg);

		if (mcr_range(i915, offset))
			continue;

2061
		*cs++ = srm;
2062
		*cs++ = offset;
2063 2064 2065 2066 2067 2068 2069 2070
		*cs++ = i915_ggtt_offset(vma) + sizeof(u32) * i;
		*cs++ = 0;
	}
	intel_ring_advance(rq, cs);

	return 0;
}

2071
static int engine_wa_list_verify(struct intel_context *ce,
2072 2073 2074 2075 2076 2077
				 const struct i915_wa_list * const wal,
				 const char *from)
{
	const struct i915_wa *wa;
	struct i915_request *rq;
	struct i915_vma *vma;
2078
	struct i915_gem_ww_ctx ww;
2079 2080 2081 2082 2083 2084 2085
	unsigned int i;
	u32 *results;
	int err;

	if (!wal->count)
		return 0;

2086 2087
	vma = __vm_create_scratch_for_read(&ce->engine->gt->ggtt->vm,
					   wal->count * sizeof(u32));
2088 2089 2090
	if (IS_ERR(vma))
		return PTR_ERR(vma);

2091
	intel_engine_pm_get(ce->engine);
2092 2093 2094 2095 2096 2097 2098 2099
	i915_gem_ww_ctx_init(&ww, false);
retry:
	err = i915_gem_object_lock(vma->obj, &ww);
	if (err == 0)
		err = intel_context_pin_ww(ce, &ww);
	if (err)
		goto err_pm;

2100 2101 2102 2103 2104
	err = i915_vma_pin_ww(vma, &ww, 0, 0,
			   i915_vma_is_ggtt(vma) ? PIN_GLOBAL : PIN_USER);
	if (err)
		goto err_unpin;

2105
	rq = i915_request_create(ce);
2106 2107
	if (IS_ERR(rq)) {
		err = PTR_ERR(rq);
2108
		goto err_vma;
2109 2110
	}

2111 2112 2113
	err = i915_request_await_object(rq, vma->obj, true);
	if (err == 0)
		err = i915_vma_move_to_active(vma, rq, EXEC_OBJECT_WRITE);
2114 2115
	if (err == 0)
		err = wa_list_srm(rq, wal, vma);
2116

2117
	i915_request_get(rq);
2118 2119
	if (err)
		i915_request_set_error_once(rq, err);
2120
	i915_request_add(rq);
2121 2122 2123 2124

	if (err)
		goto err_rq;

2125
	if (i915_request_wait(rq, 0, HZ / 5) < 0) {
2126
		err = -ETIME;
2127
		goto err_rq;
2128 2129 2130 2131 2132
	}

	results = i915_gem_object_pin_map(vma->obj, I915_MAP_WB);
	if (IS_ERR(results)) {
		err = PTR_ERR(results);
2133
		goto err_rq;
2134 2135 2136
	}

	err = 0;
2137
	for (i = 0, wa = wal->list; i < wal->count; i++, wa++) {
2138
		if (mcr_range(rq->engine->i915, i915_mmio_reg_offset(wa->reg)))
2139 2140
			continue;

2141 2142
		if (!wa_verify(wa, results[i], wal->name, from))
			err = -ENXIO;
2143
	}
2144 2145 2146

	i915_gem_object_unpin_map(vma->obj);

2147 2148
err_rq:
	i915_request_put(rq);
2149 2150
err_vma:
	i915_vma_unpin(vma);
2151 2152 2153 2154 2155 2156 2157 2158 2159 2160
err_unpin:
	intel_context_unpin(ce);
err_pm:
	if (err == -EDEADLK) {
		err = i915_gem_ww_ctx_backoff(&ww);
		if (!err)
			goto retry;
	}
	i915_gem_ww_ctx_fini(&ww);
	intel_engine_pm_put(ce->engine);
2161 2162 2163 2164 2165 2166 2167
	i915_vma_put(vma);
	return err;
}

int intel_engine_verify_workarounds(struct intel_engine_cs *engine,
				    const char *from)
{
2168 2169 2170
	return engine_wa_list_verify(engine->kernel_context,
				     &engine->wa_list,
				     from);
2171 2172
}

2173
#if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
2174
#include "selftest_workarounds.c"
2175
#endif