intel_guc_loader.c 23.8 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33
/*
 * Copyright © 2014 Intel Corporation
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice (including the next
 * paragraph) shall be included in all copies or substantial portions of the
 * Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
 * IN THE SOFTWARE.
 *
 * Authors:
 *    Vinit Azad <vinit.azad@intel.com>
 *    Ben Widawsky <ben@bwidawsk.net>
 *    Dave Gordon <david.s.gordon@intel.com>
 *    Alex Dai <yu.dai@intel.com>
 */
#include <linux/firmware.h>
#include "i915_drv.h"
#include "intel_guc.h"

/**
A
Alex Dai 已提交
34
 * DOC: GuC-specific firmware loader
35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61
 *
 * intel_guc:
 * Top level structure of guc. It handles firmware loading and manages client
 * pool and doorbells. intel_guc owns a i915_guc_client to replace the legacy
 * ExecList submission.
 *
 * Firmware versioning:
 * The firmware build process will generate a version header file with major and
 * minor version defined. The versions are built into CSS header of firmware.
 * i915 kernel driver set the minimal firmware version required per platform.
 * The firmware installation package will install (symbolic link) proper version
 * of firmware.
 *
 * GuC address space:
 * GuC does not allow any gfx GGTT address that falls into range [0, WOPCM_TOP),
 * which is reserved for Boot ROM, SRAM and WOPCM. Currently this top address is
 * 512K. In order to exclude 0-512K address space from GGTT, all gfx objects
 * used by GuC is pinned with PIN_OFFSET_BIAS along with size of WOPCM.
 *
 * Firmware log:
 * Firmware log is enabled by setting i915.guc_log_level to non-negative level.
 * Log data is printed out via reading debugfs i915_guc_log_dump. Reading from
 * i915_guc_load_status will print out firmware loading status and scratch
 * registers value.
 *
 */

62 63 64 65 66 67 68 69 70 71 72 73 74
#define SKL_FW_MAJOR 6
#define SKL_FW_MINOR 1

#define BXT_FW_MAJOR 8
#define BXT_FW_MINOR 7

#define KBL_FW_MAJOR 9
#define KBL_FW_MINOR 14

#define GUC_FW_PATH(platform, major, minor) \
       "i915/" __stringify(platform) "_guc_ver" __stringify(major) "_" __stringify(minor) ".bin"

#define I915_SKL_GUC_UCODE GUC_FW_PATH(skl, SKL_FW_MAJOR, SKL_FW_MINOR)
75 76
MODULE_FIRMWARE(I915_SKL_GUC_UCODE);

77
#define I915_BXT_GUC_UCODE GUC_FW_PATH(bxt, BXT_FW_MAJOR, BXT_FW_MINOR)
78 79
MODULE_FIRMWARE(I915_BXT_GUC_UCODE);

80
#define I915_KBL_GUC_UCODE GUC_FW_PATH(kbl, KBL_FW_MAJOR, KBL_FW_MINOR)
81 82
MODULE_FIRMWARE(I915_KBL_GUC_UCODE);

83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99
/* User-friendly representation of an enum */
const char *intel_guc_fw_status_repr(enum intel_guc_fw_status status)
{
	switch (status) {
	case GUC_FIRMWARE_FAIL:
		return "FAIL";
	case GUC_FIRMWARE_NONE:
		return "NONE";
	case GUC_FIRMWARE_PENDING:
		return "PENDING";
	case GUC_FIRMWARE_SUCCESS:
		return "SUCCESS";
	default:
		return "UNKNOWN!";
	}
};

100
static void guc_interrupts_release(struct drm_i915_private *dev_priv)
101
{
102
	struct intel_engine_cs *engine;
103
	enum intel_engine_id id;
104
	int irqs;
105

106
	/* tell all command streamers NOT to forward interrupts or vblank to GuC */
107 108
	irqs = _MASKED_FIELD(GFX_FORWARD_VBLANK_MASK, GFX_FORWARD_VBLANK_NEVER);
	irqs |= _MASKED_BIT_DISABLE(GFX_INTERRUPT_STEERING);
109
	for_each_engine(engine, dev_priv, id)
110
		I915_WRITE(RING_MODE_GEN7(engine), irqs);
111 112 113 114 115 116 117

	/* route all GT interrupts to the host */
	I915_WRITE(GUC_BCS_RCS_IER, 0);
	I915_WRITE(GUC_VCS2_VCS1_IER, 0);
	I915_WRITE(GUC_WD_VECS_IER, 0);
}

118
static void guc_interrupts_capture(struct drm_i915_private *dev_priv)
119
{
120
	struct intel_engine_cs *engine;
121
	enum intel_engine_id id;
122
	int irqs;
123
	u32 tmp;
124

125 126
	/* tell all command streamers to forward interrupts (but not vblank) to GuC */
	irqs = _MASKED_BIT_ENABLE(GFX_INTERRUPT_STEERING);
127
	for_each_engine(engine, dev_priv, id)
128
		I915_WRITE(RING_MODE_GEN7(engine), irqs);
129 130 131 132 133 134 135 136

	/* route USER_INTERRUPT to Host, all others are sent to GuC. */
	irqs = GT_RENDER_USER_INTERRUPT << GEN8_RCS_IRQ_SHIFT |
	       GT_RENDER_USER_INTERRUPT << GEN8_BCS_IRQ_SHIFT;
	/* These three registers have the same bit definitions */
	I915_WRITE(GUC_BCS_RCS_IER, ~irqs);
	I915_WRITE(GUC_VCS2_VCS1_IER, ~irqs);
	I915_WRITE(GUC_WD_VECS_IER, ~irqs);
137 138

	/*
139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156
	 * The REDIRECT_TO_GUC bit of the PMINTRMSK register directs all
	 * (unmasked) PM interrupts to the GuC. All other bits of this
	 * register *disable* generation of a specific interrupt.
	 *
	 * 'pm_intr_keep' indicates bits that are NOT to be set when
	 * writing to the PM interrupt mask register, i.e. interrupts
	 * that must not be disabled.
	 *
	 * If the GuC is handling these interrupts, then we must not let
	 * the PM code disable ANY interrupt that the GuC is expecting.
	 * So for each ENABLED (0) bit in this register, we must SET the
	 * bit in pm_intr_keep so that it's left enabled for the GuC.
	 *
	 * OTOH the REDIRECT_TO_GUC bit is initially SET in pm_intr_keep
	 * (so interrupts go to the DISPLAY unit at first); but here we
	 * need to CLEAR that bit, which will result in the register bit
	 * being left SET!
	 */
157
	tmp = I915_READ(GEN6_PMINTRMSK);
158 159 160
	if (tmp & GEN8_PMINTR_REDIRECT_TO_GUC) {
		dev_priv->rps.pm_intr_keep |= ~tmp;
		dev_priv->rps.pm_intr_keep &= ~GEN8_PMINTR_REDIRECT_TO_GUC;
161
	}
162 163
}

164 165 166 167 168 169 170 171
static u32 get_gttype(struct drm_i915_private *dev_priv)
{
	/* XXX: GT type based on PCI device ID? field seems unused by fw */
	return 0;
}

static u32 get_core_family(struct drm_i915_private *dev_priv)
{
172 173 174
	u32 gen = INTEL_GEN(dev_priv);

	switch (gen) {
175 176 177 178
	case 9:
		return GFXCORE_FAMILY_GEN9;

	default:
179
		WARN(1, "GEN%d does not support GuC operation!\n", gen);
180 181 182 183
		return GFXCORE_FAMILY_UNKNOWN;
	}
}

184 185 186 187 188 189
/*
 * Initialise the GuC parameter block before starting the firmware
 * transfer. These parameters are read by the firmware on startup
 * and cannot be changed thereafter.
 */
static void guc_params_init(struct drm_i915_private *dev_priv)
190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213
{
	struct intel_guc *guc = &dev_priv->guc;
	u32 params[GUC_CTL_MAX_DWORDS];
	int i;

	memset(&params, 0, sizeof(params));

	params[GUC_CTL_DEVICE_INFO] |=
		(get_gttype(dev_priv) << GUC_CTL_GTTYPE_SHIFT) |
		(get_core_family(dev_priv) << GUC_CTL_COREFAMILY_SHIFT);

	/*
	 * GuC ARAT increment is 10 ns. GuC default scheduler quantum is one
	 * second. This ARAR is calculated by:
	 * Scheduler-Quantum-in-ns / ARAT-increment-in-ns = 1000000000 / 10
	 */
	params[GUC_CTL_ARAT_HIGH] = 0;
	params[GUC_CTL_ARAT_LOW] = 100000000;

	params[GUC_CTL_WA] |= GUC_CTL_WA_UK_BY_DRIVER;

	params[GUC_CTL_FEATURE] |= GUC_CTL_DISABLE_SCHEDULER |
			GUC_CTL_VCS2_ENABLED;

214
	params[GUC_CTL_LOG_PARAMS] = guc->log.flags;
215

216 217 218
	if (i915.guc_log_level >= 0) {
		params[GUC_CTL_DEBUG] =
			i915.guc_log_level << GUC_LOG_VERBOSITY_SHIFT;
219 220
	} else
		params[GUC_CTL_DEBUG] = GUC_LOG_DISABLED;
221

222
	if (guc->ads_vma) {
223
		u32 ads = i915_ggtt_offset(guc->ads_vma) >> PAGE_SHIFT;
224 225 226 227
		params[GUC_CTL_DEBUG] |= ads << GUC_ADS_ADDR_SHIFT;
		params[GUC_CTL_DEBUG] |= GUC_ADS_ENABLED;
	}

228 229
	/* If GuC submission is enabled, set up additional parameters here */
	if (i915.enable_guc_submission) {
230
		u32 pgs = i915_ggtt_offset(dev_priv->guc.ctx_pool_vma);
231 232 233 234 235 236 237 238 239 240 241 242
		u32 ctx_in_16 = GUC_MAX_GPU_CONTEXTS / 16;

		pgs >>= PAGE_SHIFT;
		params[GUC_CTL_CTXINFO] = (pgs << GUC_CTL_BASE_ADDR_SHIFT) |
			(ctx_in_16 << GUC_CTL_CTXNUM_IN16_SHIFT);

		params[GUC_CTL_FEATURE] |= GUC_CTL_KERNEL_SUBMISSIONS;

		/* Unmask this bit to enable the GuC's internal scheduler */
		params[GUC_CTL_FEATURE] &= ~GUC_CTL_DISABLE_SCHEDULER;
	}

243 244 245 246 247 248 249 250 251 252 253 254
	I915_WRITE(SOFT_SCRATCH(0), 0);

	for (i = 0; i < GUC_CTL_MAX_DWORDS; i++)
		I915_WRITE(SOFT_SCRATCH(1 + i), params[i]);
}

/*
 * Read the GuC status register (GUC_STATUS) and store it in the
 * specified location; then return a boolean indicating whether
 * the value matches either of two values representing completion
 * of the GuC boot process.
 *
255
 * This is used for polling the GuC status in a wait_for()
256 257 258 259 260 261
 * loop below.
 */
static inline bool guc_ucode_response(struct drm_i915_private *dev_priv,
				      u32 *status)
{
	u32 val = I915_READ(GUC_STATUS);
262
	u32 uk_val = val & GS_UKERNEL_MASK;
263
	*status = val;
264 265
	return (uk_val == GS_UKERNEL_READY ||
		((val & GS_MIA_CORE_STATE) && uk_val == GS_UKERNEL_LAPIC_DONE));
266 267 268 269 270 271 272 273 274 275 276 277
}

/*
 * Transfer the firmware image to RAM for execution by the microcontroller.
 *
 * Architecturally, the DMA engine is bidirectional, and can potentially even
 * transfer between GTT locations. This functionality is left out of the API
 * for now as there is no need for it.
 *
 * Note that GuC needs the CSS header plus uKernel code to be copied by the
 * DMA engine in one operation, whereas the RSA signature is loaded via MMIO.
 */
C
Chris Wilson 已提交
278 279
static int guc_ucode_xfer_dma(struct drm_i915_private *dev_priv,
			      struct i915_vma *vma)
280 281 282
{
	struct intel_guc_fw *guc_fw = &dev_priv->guc.guc_fw;
	unsigned long offset;
C
Chris Wilson 已提交
283
	struct sg_table *sg = vma->pages;
A
Alex Dai 已提交
284
	u32 status, rsa[UOS_RSA_SCRATCH_MAX_COUNT];
285 286
	int i, ret = 0;

A
Alex Dai 已提交
287 288
	/* where RSA signature starts */
	offset = guc_fw->rsa_offset;
289 290

	/* Copy RSA signature from the fw image to HW for verification */
A
Alex Dai 已提交
291 292
	sg_pcopy_to_buffer(sg->sgl, sg->nents, rsa, sizeof(rsa), offset);
	for (i = 0; i < UOS_RSA_SCRATCH_MAX_COUNT; i++)
293
		I915_WRITE(UOS_RSA_SCRATCH(i), rsa[i]);
294

A
Alex Dai 已提交
295 296 297 298
	/* The header plus uCode will be copied to WOPCM via DMA, excluding any
	 * other components */
	I915_WRITE(DMA_COPY_SIZE, guc_fw->header_size + guc_fw->ucode_size);

299
	/* Set the source address for the new blob */
300
	offset = i915_ggtt_offset(vma) + guc_fw->header_offset;
301 302 303 304 305 306 307 308 309 310 311 312 313 314
	I915_WRITE(DMA_ADDR_0_LOW, lower_32_bits(offset));
	I915_WRITE(DMA_ADDR_0_HIGH, upper_32_bits(offset) & 0xFFFF);

	/*
	 * Set the DMA destination. Current uCode expects the code to be
	 * loaded at 8k; locations below this are used for the stack.
	 */
	I915_WRITE(DMA_ADDR_1_LOW, 0x2000);
	I915_WRITE(DMA_ADDR_1_HIGH, DMA_ADDRESS_SPACE_WOPCM);

	/* Finally start the DMA */
	I915_WRITE(DMA_CTRL, _MASKED_BIT_ENABLE(UOS_MOVE | START_DMA));

	/*
315
	 * Wait for the DMA to complete & the GuC to start up.
316 317 318 319 320 321
	 * NB: Docs recommend not using the interrupt for completion.
	 * Measurements indicate this should take no more than 20ms, so a
	 * timeout here indicates that the GuC has failed and is unusable.
	 * (Higher levels of the driver will attempt to fall back to
	 * execlist mode if this happens.)
	 */
322
	ret = wait_for(guc_ucode_response(dev_priv, &status), 100);
323 324 325 326 327 328 329 330 331 332 333 334 335 336

	DRM_DEBUG_DRIVER("DMA status 0x%x, GuC status 0x%x\n",
			I915_READ(DMA_CTRL), status);

	if ((status & GS_BOOTROM_MASK) == GS_BOOTROM_RSA_FAILED) {
		DRM_ERROR("GuC firmware signature verification failed\n");
		ret = -ENOEXEC;
	}

	DRM_DEBUG_DRIVER("returning %d\n", ret);

	return ret;
}

337 338 339 340 341 342 343 344 345 346 347
static u32 guc_wopcm_size(struct drm_i915_private *dev_priv)
{
	u32 wopcm_size = GUC_WOPCM_TOP;

	/* On BXT, the top of WOPCM is reserved for RC6 context */
	if (IS_BROXTON(dev_priv))
		wopcm_size -= BXT_GUC_WOPCM_RC6_RESERVED;

	return wopcm_size;
}

348 349 350 351 352 353
/*
 * Load the GuC firmware blob into the MinuteIA.
 */
static int guc_ucode_xfer(struct drm_i915_private *dev_priv)
{
	struct intel_guc_fw *guc_fw = &dev_priv->guc.guc_fw;
C
Chris Wilson 已提交
354
	struct i915_vma *vma;
355 356 357 358 359 360 361 362
	int ret;

	ret = i915_gem_object_set_to_gtt_domain(guc_fw->guc_fw_obj, false);
	if (ret) {
		DRM_DEBUG_DRIVER("set-domain failed %d\n", ret);
		return ret;
	}

C
Chris Wilson 已提交
363 364 365 366
	vma = i915_gem_object_ggtt_pin(guc_fw->guc_fw_obj, NULL, 0, 0, 0);
	if (IS_ERR(vma)) {
		DRM_DEBUG_DRIVER("pin failed %d\n", (int)PTR_ERR(vma));
		return PTR_ERR(vma);
367 368 369 370 371 372 373 374
	}

	/* Invalidate GuC TLB to let GuC take the latest updates to GTT. */
	I915_WRITE(GEN8_GTCR, GEN8_GTCR_INVALIDATE);

	intel_uncore_forcewake_get(dev_priv, FORCEWAKE_ALL);

	/* init WOPCM */
375
	I915_WRITE(GUC_WOPCM_SIZE, guc_wopcm_size(dev_priv));
376 377 378 379 380
	I915_WRITE(DMA_GUC_WOPCM_OFFSET, GUC_WOPCM_OFFSET_VALUE);

	/* Enable MIA caching. GuC clock gating is disabled. */
	I915_WRITE(GUC_SHIM_CONTROL, GUC_SHIM_CONTROL_VALUE);

381
	/* WaDisableMinuteIaClockGating:bxt */
382
	if (IS_BXT_REVID(dev_priv, 0, BXT_REVID_A1)) {
383 384 385 386
		I915_WRITE(GUC_SHIM_CONTROL, (I915_READ(GUC_SHIM_CONTROL) &
					      ~GUC_ENABLE_MIA_CLOCK_GATING));
	}

387
	/* WaC6DisallowByGfxPause:bxt */
388
	if (IS_BXT_REVID(dev_priv, 0, BXT_REVID_B0))
389
		I915_WRITE(GEN6_GFXPAUSE, 0x30FFF);
390

391
	if (IS_BROXTON(dev_priv))
392 393 394 395
		I915_WRITE(GEN9LP_GT_PM_CONFIG, GT_DOORBELL_ENABLE);
	else
		I915_WRITE(GEN9_GT_PM_CONFIG, GT_DOORBELL_ENABLE);

396
	if (IS_GEN9(dev_priv)) {
397 398 399 400
		/* DOP Clock Gating Enable for GuC clocks */
		I915_WRITE(GEN7_MISCCPCTL, (GEN8_DOP_CLOCK_GATE_GUC_ENABLE |
					    I915_READ(GEN7_MISCCPCTL)));

401
		/* allows for 5us (in 10ns units) before GT can go to RC6 */
402 403 404
		I915_WRITE(GUC_ARAT_C6DIS, 0x1FF);
	}

405
	guc_params_init(dev_priv);
406

C
Chris Wilson 已提交
407
	ret = guc_ucode_xfer_dma(dev_priv, vma);
408 409 410 411 412 413 414

	intel_uncore_forcewake_put(dev_priv, FORCEWAKE_ALL);

	/*
	 * We keep the object pages for reuse during resume. But we can unpin it
	 * now that DMA has completed, so it doesn't continue to take up space.
	 */
C
Chris Wilson 已提交
415
	i915_vma_unpin(vma);
416 417 418 419

	return ret;
}

420
static int guc_hw_reset(struct drm_i915_private *dev_priv)
421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437
{
	int ret;
	u32 guc_status;

	ret = intel_guc_reset(dev_priv);
	if (ret) {
		DRM_ERROR("GuC reset failed, ret = %d\n", ret);
		return ret;
	}

	guc_status = I915_READ(GUC_STATUS);
	WARN(!(guc_status & GS_MIA_IN_RESET),
	     "GuC status: 0x%x, MIA core expected to be in reset\n", guc_status);

	return ret;
}

438
/**
439
 * intel_guc_setup() - finish preparing the GuC for activity
440 441 442 443
 * @dev:	drm device
 *
 * Called from gem_init_hw() during driver loading and also after a GPU reset.
 *
444
 * The main action required here it to load the GuC uCode into the device.
445
 * The firmware image should have already been fetched into memory by the
446 447
 * earlier call to intel_guc_init(), so here we need only check that worked,
 * and then transfer the image to the h/w.
448 449 450
 *
 * Return:	non-zero code on error
 */
451
int intel_guc_setup(struct drm_device *dev)
452
{
453
	struct drm_i915_private *dev_priv = to_i915(dev);
454
	struct intel_guc_fw *guc_fw = &dev_priv->guc.guc_fw;
455 456
	const char *fw_path = guc_fw->guc_fw_path;
	int retries, ret, err;
457

458 459
	DRM_DEBUG_DRIVER("GuC fw status: path %s, fetch %s, load %s\n",
		fw_path,
460 461 462
		intel_guc_fw_status_repr(guc_fw->guc_fw_fetch_status),
		intel_guc_fw_status_repr(guc_fw->guc_fw_load_status));

463 464 465 466
	/* Loading forbidden, or no firmware to load? */
	if (!i915.enable_guc_loading) {
		err = 0;
		goto fail;
467 468 469 470 471 472
	} else if (fw_path == NULL) {
		/* Device is known to have no uCode (e.g. no GuC) */
		err = -ENXIO;
		goto fail;
	} else if (*fw_path == '\0') {
		/* Device has a GuC but we don't know what f/w to load? */
473
		WARN(1, "No GuC firmware known for this platform!\n");
474 475 476
		err = -ENODEV;
		goto fail;
	}
477

478 479
	/* Fetch failed, or already fetched but failed to load? */
	if (guc_fw->guc_fw_fetch_status != GUC_FIRMWARE_SUCCESS) {
480 481
		err = -EIO;
		goto fail;
482 483
	} else if (guc_fw->guc_fw_load_status == GUC_FIRMWARE_FAIL) {
		err = -ENOEXEC;
484 485 486
		goto fail;
	}

487
	guc_interrupts_release(dev_priv);
488
	gen9_reset_guc_interrupts(dev_priv);
489 490 491 492 493 494 495

	guc_fw->guc_fw_load_status = GUC_FIRMWARE_PENDING;

	DRM_DEBUG_DRIVER("GuC fw status: fetch %s, load %s\n",
		intel_guc_fw_status_repr(guc_fw->guc_fw_fetch_status),
		intel_guc_fw_status_repr(guc_fw->guc_fw_load_status));

496
	err = i915_guc_submission_init(dev_priv);
497 498 499
	if (err)
		goto fail;

500 501 502 503 504 505
	/*
	 * WaEnableuKernelHeaderValidFix:skl,bxt
	 * For BXT, this is only upto B0 but below WA is required for later
	 * steppings also so this is extended as well.
	 */
	/* WaEnableGuCBootHashCheckNotSet:skl,bxt */
506 507 508 509 510
	for (retries = 3; ; ) {
		/*
		 * Always reset the GuC just before (re)loading, so
		 * that the state and timing are fairly predictable
		 */
511
		err = guc_hw_reset(dev_priv);
512
		if (err)
513
			goto fail;
514 515 516 517 518 519 520 521

		err = guc_ucode_xfer(dev_priv);
		if (!err)
			break;

		if (--retries == 0)
			goto fail;

522 523
		DRM_INFO("GuC fw load failed: %d; will reset and "
			 "retry %d more time(s)\n", err, retries);
524
	}
525 526 527 528 529 530 531

	guc_fw->guc_fw_load_status = GUC_FIRMWARE_SUCCESS;

	DRM_DEBUG_DRIVER("GuC fw status: fetch %s, load %s\n",
		intel_guc_fw_status_repr(guc_fw->guc_fw_fetch_status),
		intel_guc_fw_status_repr(guc_fw->guc_fw_load_status));

532
	if (i915.enable_guc_submission) {
533 534 535
		if (i915.guc_log_level >= 0)
			gen9_enable_guc_interrupts(dev_priv);

536
		err = i915_guc_submission_enable(dev_priv);
537 538
		if (err)
			goto fail;
539
		guc_interrupts_capture(dev_priv);
540 541
	}

542 543 544 545 546 547
	return 0;

fail:
	if (guc_fw->guc_fw_load_status == GUC_FIRMWARE_PENDING)
		guc_fw->guc_fw_load_status = GUC_FIRMWARE_FAIL;

548
	guc_interrupts_release(dev_priv);
549 550
	i915_guc_submission_disable(dev_priv);
	i915_guc_submission_fini(dev_priv);
551

552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568
	/*
	 * We've failed to load the firmware :(
	 *
	 * Decide whether to disable GuC submission and fall back to
	 * execlist mode, and whether to hide the error by returning
	 * zero or to return -EIO, which the caller will treat as a
	 * nonfatal error (i.e. it doesn't prevent driver load, but
	 * marks the GPU as wedged until reset).
	 */
	if (i915.enable_guc_loading > 1) {
		ret = -EIO;
	} else if (i915.enable_guc_submission > 1) {
		ret = -EIO;
	} else {
		ret = 0;
	}

569
	if (err == 0 && !HAS_GUC_UCODE(dev_priv))
570 571
		;	/* Don't mention the GuC! */
	else if (err == 0)
572
		DRM_INFO("GuC firmware load skipped\n");
573
	else if (ret != -EIO)
574
		DRM_NOTE("GuC firmware load failed: %d\n", err);
575
	else
576
		DRM_WARN("GuC firmware load failed: %d\n", err);
577 578 579 580 581

	if (i915.enable_guc_submission) {
		if (fw_path == NULL)
			DRM_INFO("GuC submission without firmware not supported\n");
		if (ret == 0)
582
			DRM_NOTE("Falling back from GuC submission to execlist mode\n");
583 584 585 586 587 588
		else
			DRM_ERROR("GuC init failed: %d\n", ret);
	}
	i915.enable_guc_submission = 0;

	return ret;
589 590 591 592
}

static void guc_fw_fetch(struct drm_device *dev, struct intel_guc_fw *guc_fw)
{
D
David Weinehall 已提交
593
	struct pci_dev *pdev = dev->pdev;
594 595
	struct drm_i915_gem_object *obj;
	const struct firmware *fw;
A
Alex Dai 已提交
596 597
	struct guc_css_header *css;
	size_t size;
598 599 600 601 602
	int err;

	DRM_DEBUG_DRIVER("before requesting firmware: GuC fw fetch status %s\n",
		intel_guc_fw_status_repr(guc_fw->guc_fw_fetch_status));

D
David Weinehall 已提交
603
	err = request_firmware(&fw, guc_fw->guc_fw_path, &pdev->dev);
604 605 606 607 608 609 610 611
	if (err)
		goto fail;
	if (!fw)
		goto fail;

	DRM_DEBUG_DRIVER("fetch GuC fw from %s succeeded, fw %p\n",
		guc_fw->guc_fw_path, fw);

A
Alex Dai 已提交
612 613
	/* Check the size of the blob before examining buffer contents */
	if (fw->size < sizeof(struct guc_css_header)) {
614
		DRM_NOTE("Firmware header is missing\n");
615
		goto fail;
A
Alex Dai 已提交
616 617 618 619 620 621 622 623 624 625
	}

	css = (struct guc_css_header *)fw->data;

	/* Firmware bits always start from header */
	guc_fw->header_offset = 0;
	guc_fw->header_size = (css->header_size_dw - css->modulus_size_dw -
		css->key_size_dw - css->exponent_size_dw) * sizeof(u32);

	if (guc_fw->header_size != sizeof(struct guc_css_header)) {
626
		DRM_NOTE("CSS header definition mismatch\n");
A
Alex Dai 已提交
627 628 629 630 631 632 633 634 635
		goto fail;
	}

	/* then, uCode */
	guc_fw->ucode_offset = guc_fw->header_offset + guc_fw->header_size;
	guc_fw->ucode_size = (css->size_dw - css->header_size_dw) * sizeof(u32);

	/* now RSA */
	if (css->key_size_dw != UOS_RSA_SCRATCH_MAX_COUNT) {
636
		DRM_NOTE("RSA key size is bad\n");
A
Alex Dai 已提交
637 638 639 640 641 642 643 644
		goto fail;
	}
	guc_fw->rsa_offset = guc_fw->ucode_offset + guc_fw->ucode_size;
	guc_fw->rsa_size = css->key_size_dw * sizeof(u32);

	/* At least, it should have header, uCode and RSA. Size of all three. */
	size = guc_fw->header_size + guc_fw->ucode_size + guc_fw->rsa_size;
	if (fw->size < size) {
645
		DRM_NOTE("Missing firmware components\n");
A
Alex Dai 已提交
646 647 648 649 650
		goto fail;
	}

	/* Header and uCode will be loaded to WOPCM. Size of the two. */
	size = guc_fw->header_size + guc_fw->ucode_size;
651
	if (size > guc_wopcm_size(to_i915(dev))) {
652
		DRM_NOTE("Firmware is too large to fit in WOPCM\n");
A
Alex Dai 已提交
653 654
		goto fail;
	}
655 656 657 658 659 660 661

	/*
	 * The GuC firmware image has the version number embedded at a well-known
	 * offset within the firmware blob; note that major / minor version are
	 * TWO bytes each (i.e. u16), although all pointers and offsets are defined
	 * in terms of bytes (u8).
	 */
A
Alex Dai 已提交
662 663
	guc_fw->guc_fw_major_found = css->guc_sw_version >> 16;
	guc_fw->guc_fw_minor_found = css->guc_sw_version & 0xFFFF;
664 665 666

	if (guc_fw->guc_fw_major_found != guc_fw->guc_fw_major_wanted ||
	    guc_fw->guc_fw_minor_found < guc_fw->guc_fw_minor_wanted) {
667
		DRM_NOTE("GuC firmware version %d.%d, required %d.%d\n",
668 669 670 671 672 673 674 675 676 677
			guc_fw->guc_fw_major_found, guc_fw->guc_fw_minor_found,
			guc_fw->guc_fw_major_wanted, guc_fw->guc_fw_minor_wanted);
		err = -ENOEXEC;
		goto fail;
	}

	DRM_DEBUG_DRIVER("firmware version %d.%d OK (minimum %d.%d)\n",
			guc_fw->guc_fw_major_found, guc_fw->guc_fw_minor_found,
			guc_fw->guc_fw_major_wanted, guc_fw->guc_fw_minor_wanted);

678
	mutex_lock(&dev->struct_mutex);
679
	obj = i915_gem_object_create_from_data(dev, fw->data, fw->size);
680
	mutex_unlock(&dev->struct_mutex);
681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696
	if (IS_ERR_OR_NULL(obj)) {
		err = obj ? PTR_ERR(obj) : -ENOMEM;
		goto fail;
	}

	guc_fw->guc_fw_obj = obj;
	guc_fw->guc_fw_size = fw->size;

	DRM_DEBUG_DRIVER("GuC fw fetch status SUCCESS, obj %p\n",
			guc_fw->guc_fw_obj);

	release_firmware(fw);
	guc_fw->guc_fw_fetch_status = GUC_FIRMWARE_SUCCESS;
	return;

fail:
697 698
	DRM_WARN("Failed to fetch valid GuC firmware from %s (error %d)\n",
		 guc_fw->guc_fw_path, err);
699 700 701
	DRM_DEBUG_DRIVER("GuC fw fetch status FAIL; err %d, fw %p, obj %p\n",
		err, fw, guc_fw->guc_fw_obj);

702
	mutex_lock(&dev->struct_mutex);
703 704
	obj = guc_fw->guc_fw_obj;
	if (obj)
705
		i915_gem_object_put(obj);
706
	guc_fw->guc_fw_obj = NULL;
707
	mutex_unlock(&dev->struct_mutex);
708 709 710 711 712 713

	release_firmware(fw);		/* OK even if fw is NULL */
	guc_fw->guc_fw_fetch_status = GUC_FIRMWARE_FAIL;
}

/**
714
 * intel_guc_init() - define parameters and fetch firmware
715 716 717 718 719
 * @dev:	drm device
 *
 * Called early during driver load, but after GEM is initialised.
 *
 * The firmware will be transferred to the GuC's memory later,
720
 * when intel_guc_setup() is called.
721
 */
722
void intel_guc_init(struct drm_device *dev)
723
{
724
	struct drm_i915_private *dev_priv = to_i915(dev);
725 726 727
	struct intel_guc_fw *guc_fw = &dev_priv->guc.guc_fw;
	const char *fw_path;

728
	if (!HAS_GUC(dev_priv)) {
729 730 731 732 733
		i915.enable_guc_loading = 0;
		i915.enable_guc_submission = 0;
	} else {
		/* A negative value means "use platform default" */
		if (i915.enable_guc_loading < 0)
734
			i915.enable_guc_loading = HAS_GUC_UCODE(dev_priv);
735
		if (i915.enable_guc_submission < 0)
736
			i915.enable_guc_submission = HAS_GUC_SCHED(dev_priv);
737
	}
738

739
	if (!HAS_GUC_UCODE(dev_priv)) {
740
		fw_path = NULL;
741
	} else if (IS_SKYLAKE(dev_priv)) {
742
		fw_path = I915_SKL_GUC_UCODE;
743 744
		guc_fw->guc_fw_major_wanted = SKL_FW_MAJOR;
		guc_fw->guc_fw_minor_wanted = SKL_FW_MINOR;
745
	} else if (IS_BROXTON(dev_priv)) {
746
		fw_path = I915_BXT_GUC_UCODE;
747 748
		guc_fw->guc_fw_major_wanted = BXT_FW_MAJOR;
		guc_fw->guc_fw_minor_wanted = BXT_FW_MINOR;
749
	} else if (IS_KABYLAKE(dev_priv)) {
750
		fw_path = I915_KBL_GUC_UCODE;
751 752
		guc_fw->guc_fw_major_wanted = KBL_FW_MAJOR;
		guc_fw->guc_fw_minor_wanted = KBL_FW_MINOR;
753 754 755 756 757 758 759 760 761
	} else {
		fw_path = "";	/* unknown device */
	}

	guc_fw->guc_dev = dev;
	guc_fw->guc_fw_path = fw_path;
	guc_fw->guc_fw_fetch_status = GUC_FIRMWARE_NONE;
	guc_fw->guc_fw_load_status = GUC_FIRMWARE_NONE;

762 763 764
	/* Early (and silent) return if GuC loading is disabled */
	if (!i915.enable_guc_loading)
		return;
765 766
	if (fw_path == NULL)
		return;
767
	if (*fw_path == '\0')
768 769 770 771 772 773 774 775 776
		return;

	guc_fw->guc_fw_fetch_status = GUC_FIRMWARE_PENDING;
	DRM_DEBUG_DRIVER("GuC firmware pending, path %s\n", fw_path);
	guc_fw_fetch(dev, guc_fw);
	/* status must now be FAIL or SUCCESS */
}

/**
777
 * intel_guc_fini() - clean up all allocated resources
778 779
 * @dev:	drm device
 */
780
void intel_guc_fini(struct drm_device *dev)
781
{
782
	struct drm_i915_private *dev_priv = to_i915(dev);
783 784
	struct intel_guc_fw *guc_fw = &dev_priv->guc.guc_fw;

785
	mutex_lock(&dev->struct_mutex);
786
	guc_interrupts_release(dev_priv);
787 788
	i915_guc_submission_disable(dev_priv);
	i915_guc_submission_fini(dev_priv);
789

790
	if (guc_fw->guc_fw_obj)
791
		i915_gem_object_put(guc_fw->guc_fw_obj);
792
	guc_fw->guc_fw_obj = NULL;
793
	mutex_unlock(&dev->struct_mutex);
794 795 796

	guc_fw->guc_fw_fetch_status = GUC_FIRMWARE_NONE;
}