i915_irq.c 40.7 KB
Newer Older
D
Dave Airlie 已提交
1
/* i915_irq.c -- IRQ support for the I915 -*- linux-c -*-
L
Linus Torvalds 已提交
2
 */
D
Dave Airlie 已提交
3
/*
L
Linus Torvalds 已提交
4 5
 * Copyright 2003 Tungsten Graphics, Inc., Cedar Park, Texas.
 * All Rights Reserved.
6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the
 * "Software"), to deal in the Software without restriction, including
 * without limitation the rights to use, copy, modify, merge, publish,
 * distribute, sub license, and/or sell copies of the Software, and to
 * permit persons to whom the Software is furnished to do so, subject to
 * the following conditions:
 *
 * The above copyright notice and this permission notice (including the
 * next paragraph) shall be included in all copies or substantial portions
 * of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 *
D
Dave Airlie 已提交
27
 */
L
Linus Torvalds 已提交
28

29
#include <linux/sysrq.h>
L
Linus Torvalds 已提交
30 31 32 33
#include "drmP.h"
#include "drm.h"
#include "i915_drm.h"
#include "i915_drv.h"
C
Chris Wilson 已提交
34
#include "i915_trace.h"
J
Jesse Barnes 已提交
35
#include "intel_drv.h"
L
Linus Torvalds 已提交
36 37 38

#define MAX_NOPID ((u32)~0)

39 40 41 42 43 44 45
/**
 * Interrupts that are always left unmasked.
 *
 * Since pipe events are edge-triggered from the PIPESTAT register to IIR,
 * we leave them always unmasked in IMR and then control enabling them through
 * PIPESTAT alone.
 */
46 47 48 49 50 51 52
#define I915_INTERRUPT_ENABLE_FIX			\
	(I915_ASLE_INTERRUPT |				\
	 I915_DISPLAY_PIPE_A_EVENT_INTERRUPT |		\
	 I915_DISPLAY_PIPE_B_EVENT_INTERRUPT |		\
	 I915_DISPLAY_PLANE_A_FLIP_PENDING_INTERRUPT |	\
	 I915_DISPLAY_PLANE_B_FLIP_PENDING_INTERRUPT |	\
	 I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT)
53 54 55 56

/** Interrupts that we mask and unmask at runtime. */
#define I915_INTERRUPT_ENABLE_VAR (I915_USER_INTERRUPT)

J
Jesse Barnes 已提交
57 58 59 60 61 62 63 64 65
#define I915_PIPE_VBLANK_STATUS	(PIPE_START_VBLANK_INTERRUPT_STATUS |\
				 PIPE_VBLANK_INTERRUPT_STATUS)

#define I915_PIPE_VBLANK_ENABLE	(PIPE_START_VBLANK_INTERRUPT_ENABLE |\
				 PIPE_VBLANK_INTERRUPT_ENABLE)

#define DRM_I915_VBLANK_PIPE_ALL	(DRM_I915_VBLANK_PIPE_A | \
					 DRM_I915_VBLANK_PIPE_B)

66
void
67
ironlake_enable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask)
68 69 70 71 72 73 74 75 76
{
	if ((dev_priv->gt_irq_mask_reg & mask) != 0) {
		dev_priv->gt_irq_mask_reg &= ~mask;
		I915_WRITE(GTIMR, dev_priv->gt_irq_mask_reg);
		(void) I915_READ(GTIMR);
	}
}

static inline void
77
ironlake_disable_graphics_irq(drm_i915_private_t *dev_priv, u32 mask)
78 79 80 81 82 83 84 85 86 87
{
	if ((dev_priv->gt_irq_mask_reg & mask) != mask) {
		dev_priv->gt_irq_mask_reg |= mask;
		I915_WRITE(GTIMR, dev_priv->gt_irq_mask_reg);
		(void) I915_READ(GTIMR);
	}
}

/* For display hotplug interrupt */
void
88
ironlake_enable_display_irq(drm_i915_private_t *dev_priv, u32 mask)
89 90 91 92 93 94 95 96 97
{
	if ((dev_priv->irq_mask_reg & mask) != 0) {
		dev_priv->irq_mask_reg &= ~mask;
		I915_WRITE(DEIMR, dev_priv->irq_mask_reg);
		(void) I915_READ(DEIMR);
	}
}

static inline void
98
ironlake_disable_display_irq(drm_i915_private_t *dev_priv, u32 mask)
99 100 101 102 103 104 105 106
{
	if ((dev_priv->irq_mask_reg & mask) != mask) {
		dev_priv->irq_mask_reg |= mask;
		I915_WRITE(DEIMR, dev_priv->irq_mask_reg);
		(void) I915_READ(DEIMR);
	}
}

107
void
108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126
i915_enable_irq(drm_i915_private_t *dev_priv, u32 mask)
{
	if ((dev_priv->irq_mask_reg & mask) != 0) {
		dev_priv->irq_mask_reg &= ~mask;
		I915_WRITE(IMR, dev_priv->irq_mask_reg);
		(void) I915_READ(IMR);
	}
}

static inline void
i915_disable_irq(drm_i915_private_t *dev_priv, u32 mask)
{
	if ((dev_priv->irq_mask_reg & mask) != mask) {
		dev_priv->irq_mask_reg |= mask;
		I915_WRITE(IMR, dev_priv->irq_mask_reg);
		(void) I915_READ(IMR);
	}
}

127 128 129 130 131 132 133
static inline u32
i915_pipestat(int pipe)
{
	if (pipe == 0)
		return PIPEASTAT;
	if (pipe == 1)
		return PIPEBSTAT;
134
	BUG();
135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161
}

void
i915_enable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask)
{
	if ((dev_priv->pipestat[pipe] & mask) != mask) {
		u32 reg = i915_pipestat(pipe);

		dev_priv->pipestat[pipe] |= mask;
		/* Enable the interrupt, clear any pending status */
		I915_WRITE(reg, dev_priv->pipestat[pipe] | (mask >> 16));
		(void) I915_READ(reg);
	}
}

void
i915_disable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask)
{
	if ((dev_priv->pipestat[pipe] & mask) != 0) {
		u32 reg = i915_pipestat(pipe);

		dev_priv->pipestat[pipe] &= ~mask;
		I915_WRITE(reg, dev_priv->pipestat[pipe]);
		(void) I915_READ(reg);
	}
}

162 163 164 165 166 167 168
/**
 * intel_enable_asle - enable ASLE interrupt for OpRegion
 */
void intel_enable_asle (struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;

169 170
	if (IS_IRONLAKE(dev))
		ironlake_enable_display_irq(dev_priv, DE_GSE);
171 172 173 174 175
	else
		i915_enable_pipestat(dev_priv, 1,
				     I915_LEGACY_BLC_EVENT_ENABLE);
}

176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196
/**
 * i915_pipe_enabled - check if a pipe is enabled
 * @dev: DRM device
 * @pipe: pipe to check
 *
 * Reading certain registers when the pipe is disabled can hang the chip.
 * Use this routine to make sure the PLL is running and the pipe is active
 * before reading such registers if unsure.
 */
static int
i915_pipe_enabled(struct drm_device *dev, int pipe)
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	unsigned long pipeconf = pipe ? PIPEBCONF : PIPEACONF;

	if (I915_READ(pipeconf) & PIPEACONF_ENABLE)
		return 1;

	return 0;
}

197 198 199 200
/* Called from drm generic code, passed a 'crtc', which
 * we use as a pipe index
 */
u32 i915_get_vblank_counter(struct drm_device *dev, int pipe)
201 202 203 204 205 206 207 208 209 210
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	unsigned long high_frame;
	unsigned long low_frame;
	u32 high1, high2, low, count;

	high_frame = pipe ? PIPEBFRAMEHIGH : PIPEAFRAMEHIGH;
	low_frame = pipe ? PIPEBFRAMEPIXEL : PIPEAFRAMEPIXEL;

	if (!i915_pipe_enabled(dev, pipe)) {
211 212
		DRM_DEBUG_DRIVER("trying to get vblank count for disabled "
				"pipe %d\n", pipe);
213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234
		return 0;
	}

	/*
	 * High & low register fields aren't synchronized, so make sure
	 * we get a low value that's stable across two reads of the high
	 * register.
	 */
	do {
		high1 = ((I915_READ(high_frame) & PIPE_FRAME_HIGH_MASK) >>
			 PIPE_FRAME_HIGH_SHIFT);
		low =  ((I915_READ(low_frame) & PIPE_FRAME_LOW_MASK) >>
			PIPE_FRAME_LOW_SHIFT);
		high2 = ((I915_READ(high_frame) & PIPE_FRAME_HIGH_MASK) >>
			 PIPE_FRAME_HIGH_SHIFT);
	} while (high1 != high2);

	count = (high1 << 8) | low;

	return count;
}

235 236 237 238 239 240
u32 gm45_get_vblank_counter(struct drm_device *dev, int pipe)
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	int reg = pipe ? PIPEB_FRMCOUNT_GM45 : PIPEA_FRMCOUNT_GM45;

	if (!i915_pipe_enabled(dev, pipe)) {
241 242
		DRM_DEBUG_DRIVER("trying to get vblank count for disabled "
					"pipe %d\n", pipe);
243 244 245 246 247 248
		return 0;
	}

	return I915_READ(reg);
}

249 250 251 252 253 254 255 256
/*
 * Handle hotplug events outside the interrupt handler proper.
 */
static void i915_hotplug_work_func(struct work_struct *work)
{
	drm_i915_private_t *dev_priv = container_of(work, drm_i915_private_t,
						    hotplug_work);
	struct drm_device *dev = dev_priv->dev;
257 258 259 260 261 262 263 264 265 266 267
	struct drm_mode_config *mode_config = &dev->mode_config;
	struct drm_connector *connector;

	if (mode_config->num_connector) {
		list_for_each_entry(connector, &mode_config->connector_list, head) {
			struct intel_output *intel_output = to_intel_output(connector);
	
			if (intel_output->hot_plug)
				(*intel_output->hot_plug) (intel_output);
		}
	}
268 269 270 271
	/* Just fire off a uevent and let userspace tell us what to do */
	drm_sysfs_hotplug_event(dev);
}

272 273 274
static void i915_handle_rps_change(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
275
	u32 busy_up, busy_down, max_avg, min_avg;
276 277 278 279
	u16 rgvswctl;
	u8 new_delay = dev_priv->cur_delay;

	I915_WRITE(MEMINTRSTS, I915_READ(MEMINTRSTS) & ~MEMINT_EVAL_CHG);
280 281
	busy_up = I915_READ(RCPREVBSYTUPAVG);
	busy_down = I915_READ(RCPREVBSYTDNAVG);
282 283 284 285
	max_avg = I915_READ(RCBMAXAVG);
	min_avg = I915_READ(RCBMINAVG);

	/* Handle RCS change request from hw */
286
	if (busy_up > max_avg) {
287 288 289 290
		if (dev_priv->cur_delay != dev_priv->max_delay)
			new_delay = dev_priv->cur_delay - 1;
		if (new_delay < dev_priv->max_delay)
			new_delay = dev_priv->max_delay;
291
	} else if (busy_down < min_avg) {
292 293 294 295 296 297 298 299 300 301 302
		if (dev_priv->cur_delay != dev_priv->min_delay)
			new_delay = dev_priv->cur_delay + 1;
		if (new_delay > dev_priv->min_delay)
			new_delay = dev_priv->min_delay;
	}

	DRM_DEBUG("rps change requested: %d -> %d\n",
		  dev_priv->cur_delay, new_delay);

	rgvswctl = I915_READ(MEMSWCTL);
	if (rgvswctl & MEMCTL_CMD_STS) {
303 304
		DRM_ERROR("gpu busy, RCS change rejected\n");
		return; /* still busy with another command */
305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322
	}

	/* Program the new state */
	rgvswctl = (MEMCTL_CMD_CHFREQ << MEMCTL_CMD_SHIFT) |
		(new_delay << MEMCTL_FREQ_SHIFT) | MEMCTL_SFCAVM;
	I915_WRITE(MEMSWCTL, rgvswctl);
	POSTING_READ(MEMSWCTL);

	rgvswctl |= MEMCTL_CMD_STS;
	I915_WRITE(MEMSWCTL, rgvswctl);

	dev_priv->cur_delay = new_delay;

	DRM_DEBUG("rps changed\n");

	return;
}

323
irqreturn_t ironlake_irq_handler(struct drm_device *dev)
324 325 326
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	int ret = IRQ_NONE;
327
	u32 de_iir, gt_iir, de_ier, pch_iir;
328 329
	struct drm_i915_master_private *master_priv;

330 331 332 333 334
	/* disable master interrupt before clearing iir  */
	de_ier = I915_READ(DEIER);
	I915_WRITE(DEIER, de_ier & ~DE_MASTER_IRQ_CONTROL);
	(void)I915_READ(DEIER);

335 336
	de_iir = I915_READ(DEIIR);
	gt_iir = I915_READ(GTIIR);
337
	pch_iir = I915_READ(SDEIIR);
338

339 340
	if (de_iir == 0 && gt_iir == 0 && pch_iir == 0)
		goto done;
341

342
	ret = IRQ_HANDLED;
343

344 345 346 347 348 349
	if (dev->primary->master) {
		master_priv = dev->primary->master->driver_priv;
		if (master_priv->sarea_priv)
			master_priv->sarea_priv->last_dispatch =
				READ_BREADCRUMB(dev_priv);
	}
350

351 352 353 354 355 356 357 358
	if (gt_iir & GT_USER_INTERRUPT) {
		u32 seqno = i915_get_gem_seqno(dev);
		dev_priv->mm.irq_gem_seqno = seqno;
		trace_i915_gem_request_complete(dev, seqno);
		DRM_WAKEUP(&dev_priv->irq_queue);
		dev_priv->hangcheck_count = 0;
		mod_timer(&dev_priv->hangcheck_timer, jiffies + DRM_I915_HANGCHECK_PERIOD);
	}
359

360 361
	if (de_iir & DE_GSE)
		ironlake_opregion_gse_intr(dev);
362

363
	if (de_iir & DE_PLANEA_FLIP_DONE) {
364
		intel_prepare_page_flip(dev, 0);
365 366
		intel_finish_page_flip(dev, 0);
	}
367

368
	if (de_iir & DE_PLANEB_FLIP_DONE) {
369
		intel_prepare_page_flip(dev, 1);
370 371
		intel_finish_page_flip(dev, 1);
	}
372

373
	if (de_iir & DE_PIPEA_VBLANK)
374 375
		drm_handle_vblank(dev, 0);

376
	if (de_iir & DE_PIPEB_VBLANK)
377 378
		drm_handle_vblank(dev, 1);

379 380 381 382
	/* check event from PCH */
	if ((de_iir & DE_PCH_EVENT) &&
	    (pch_iir & SDE_HOTPLUG_MASK)) {
		queue_work(dev_priv->wq, &dev_priv->hotplug_work);
383 384
	}

385 386 387 388 389
	if (de_iir & DE_PCU_EVENT) {
		I915_WRITE(MEMINTRSTS, I915_READ(MEMINTRSTS));
		i915_handle_rps_change(dev);
	}

390 391 392 393 394 395
	/* should clear PCH hotplug event before clear CPU irq */
	I915_WRITE(SDEIIR, pch_iir);
	I915_WRITE(GTIIR, gt_iir);
	I915_WRITE(DEIIR, de_iir);

done:
396 397 398
	I915_WRITE(DEIER, de_ier);
	(void)I915_READ(DEIER);

399 400 401
	return ret;
}

402 403 404 405 406 407 408 409 410 411 412 413
/**
 * i915_error_work_func - do process context error handling work
 * @work: work struct
 *
 * Fire an error uevent so userspace can see that a hang or error
 * was detected.
 */
static void i915_error_work_func(struct work_struct *work)
{
	drm_i915_private_t *dev_priv = container_of(work, drm_i915_private_t,
						    error_work);
	struct drm_device *dev = dev_priv->dev;
414 415 416
	char *error_event[] = { "ERROR=1", NULL };
	char *reset_event[] = { "RESET=1", NULL };
	char *reset_done_event[] = { "ERROR=0", NULL };
417

418
	DRM_DEBUG_DRIVER("generating error event\n");
419 420
	kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, error_event);

421
	if (atomic_read(&dev_priv->mm.wedged)) {
422
		if (IS_I965G(dev)) {
423
			DRM_DEBUG_DRIVER("resetting chip\n");
424 425
			kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, reset_event);
			if (!i965_reset(dev, GDRST_RENDER)) {
426
				atomic_set(&dev_priv->mm.wedged, 0);
427 428 429
				kobject_uevent_env(&dev->primary->kdev.kobj, KOBJ_CHANGE, reset_done_event);
			}
		} else {
430
			DRM_DEBUG_DRIVER("reboot required\n");
431 432
		}
	}
433 434
}

435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549
static struct drm_i915_error_object *
i915_error_object_create(struct drm_device *dev,
			 struct drm_gem_object *src)
{
	struct drm_i915_error_object *dst;
	struct drm_i915_gem_object *src_priv;
	int page, page_count;

	if (src == NULL)
		return NULL;

	src_priv = src->driver_private;
	if (src_priv->pages == NULL)
		return NULL;

	page_count = src->size / PAGE_SIZE;

	dst = kmalloc(sizeof(*dst) + page_count * sizeof (u32 *), GFP_ATOMIC);
	if (dst == NULL)
		return NULL;

	for (page = 0; page < page_count; page++) {
		void *s, *d = kmalloc(PAGE_SIZE, GFP_ATOMIC);
		if (d == NULL)
			goto unwind;
		s = kmap_atomic(src_priv->pages[page], KM_USER0);
		memcpy(d, s, PAGE_SIZE);
		kunmap_atomic(s, KM_USER0);
		dst->pages[page] = d;
	}
	dst->page_count = page_count;
	dst->gtt_offset = src_priv->gtt_offset;

	return dst;

unwind:
	while (page--)
		kfree(dst->pages[page]);
	kfree(dst);
	return NULL;
}

static void
i915_error_object_free(struct drm_i915_error_object *obj)
{
	int page;

	if (obj == NULL)
		return;

	for (page = 0; page < obj->page_count; page++)
		kfree(obj->pages[page]);

	kfree(obj);
}

static void
i915_error_state_free(struct drm_device *dev,
		      struct drm_i915_error_state *error)
{
	i915_error_object_free(error->batchbuffer[0]);
	i915_error_object_free(error->batchbuffer[1]);
	i915_error_object_free(error->ringbuffer);
	kfree(error->active_bo);
	kfree(error);
}

static u32
i915_get_bbaddr(struct drm_device *dev, u32 *ring)
{
	u32 cmd;

	if (IS_I830(dev) || IS_845G(dev))
		cmd = MI_BATCH_BUFFER;
	else if (IS_I965G(dev))
		cmd = (MI_BATCH_BUFFER_START | (2 << 6) |
		       MI_BATCH_NON_SECURE_I965);
	else
		cmd = (MI_BATCH_BUFFER_START | (2 << 6));

	return ring[0] == cmd ? ring[1] : 0;
}

static u32
i915_ringbuffer_last_batch(struct drm_device *dev)
{
	struct drm_i915_private *dev_priv = dev->dev_private;
	u32 head, bbaddr;
	u32 *ring;

	/* Locate the current position in the ringbuffer and walk back
	 * to find the most recently dispatched batch buffer.
	 */
	bbaddr = 0;
	head = I915_READ(PRB0_HEAD) & HEAD_ADDR;
	ring = (u32 *)(dev_priv->ring.virtual_start + head);

	while (--ring >= (u32 *)dev_priv->ring.virtual_start) {
		bbaddr = i915_get_bbaddr(dev, ring);
		if (bbaddr)
			break;
	}

	if (bbaddr == 0) {
		ring = (u32 *)(dev_priv->ring.virtual_start + dev_priv->ring.Size);
		while (--ring >= (u32 *)dev_priv->ring.virtual_start) {
			bbaddr = i915_get_bbaddr(dev, ring);
			if (bbaddr)
				break;
		}
	}

	return bbaddr;
}

550 551 552 553 554 555 556 557 558
/**
 * i915_capture_error_state - capture an error record for later analysis
 * @dev: drm device
 *
 * Should be called when an error is detected (either a hang or an error
 * interrupt) to capture error state from the time of the error.  Fills
 * out a structure which becomes available in debugfs for user level tools
 * to pick up.
 */
559 560 561
static void i915_capture_error_state(struct drm_device *dev)
{
	struct drm_i915_private *dev_priv = dev->dev_private;
562
	struct drm_i915_gem_object *obj_priv;
563
	struct drm_i915_error_state *error;
564
	struct drm_gem_object *batchbuffer[2];
565
	unsigned long flags;
566 567
	u32 bbaddr;
	int count;
568 569

	spin_lock_irqsave(&dev_priv->error_lock, flags);
570 571 572 573
	error = dev_priv->first_error;
	spin_unlock_irqrestore(&dev_priv->error_lock, flags);
	if (error)
		return;
574 575 576

	error = kmalloc(sizeof(*error), GFP_ATOMIC);
	if (!error) {
577 578
		DRM_DEBUG_DRIVER("out of memory, not capturing error state\n");
		return;
579 580
	}

581
	error->seqno = i915_get_gem_seqno(dev);
582 583 584 585 586 587 588 589 590 591
	error->eir = I915_READ(EIR);
	error->pgtbl_er = I915_READ(PGTBL_ER);
	error->pipeastat = I915_READ(PIPEASTAT);
	error->pipebstat = I915_READ(PIPEBSTAT);
	error->instpm = I915_READ(INSTPM);
	if (!IS_I965G(dev)) {
		error->ipeir = I915_READ(IPEIR);
		error->ipehr = I915_READ(IPEHR);
		error->instdone = I915_READ(INSTDONE);
		error->acthd = I915_READ(ACTHD);
592
		error->bbaddr = 0;
593 594 595 596 597 598 599
	} else {
		error->ipeir = I915_READ(IPEIR_I965);
		error->ipehr = I915_READ(IPEHR_I965);
		error->instdone = I915_READ(INSTDONE_I965);
		error->instps = I915_READ(INSTPS);
		error->instdone1 = I915_READ(INSTDONE1);
		error->acthd = I915_READ(ACTHD_I965);
600
		error->bbaddr = I915_READ64(BB_ADDR);
601 602
	}

603
	bbaddr = i915_ringbuffer_last_batch(dev);
604

605 606 607 608 609 610
	/* Grab the current batchbuffer, most likely to have crashed. */
	batchbuffer[0] = NULL;
	batchbuffer[1] = NULL;
	count = 0;
	list_for_each_entry(obj_priv, &dev_priv->mm.active_list, list) {
		struct drm_gem_object *obj = obj_priv->obj;
611

612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676
		if (batchbuffer[0] == NULL &&
		    bbaddr >= obj_priv->gtt_offset &&
		    bbaddr < obj_priv->gtt_offset + obj->size)
			batchbuffer[0] = obj;

		if (batchbuffer[1] == NULL &&
		    error->acthd >= obj_priv->gtt_offset &&
		    error->acthd < obj_priv->gtt_offset + obj->size &&
		    batchbuffer[0] != obj)
			batchbuffer[1] = obj;

		count++;
	}

	/* We need to copy these to an anonymous buffer as the simplest
	 * method to avoid being overwritten by userpace.
	 */
	error->batchbuffer[0] = i915_error_object_create(dev, batchbuffer[0]);
	error->batchbuffer[1] = i915_error_object_create(dev, batchbuffer[1]);

	/* Record the ringbuffer */
	error->ringbuffer = i915_error_object_create(dev, dev_priv->ring.ring_obj);

	/* Record buffers on the active list. */
	error->active_bo = NULL;
	error->active_bo_count = 0;

	if (count)
		error->active_bo = kmalloc(sizeof(*error->active_bo)*count,
					   GFP_ATOMIC);

	if (error->active_bo) {
		int i = 0;
		list_for_each_entry(obj_priv, &dev_priv->mm.active_list, list) {
			struct drm_gem_object *obj = obj_priv->obj;

			error->active_bo[i].size = obj->size;
			error->active_bo[i].name = obj->name;
			error->active_bo[i].seqno = obj_priv->last_rendering_seqno;
			error->active_bo[i].gtt_offset = obj_priv->gtt_offset;
			error->active_bo[i].read_domains = obj->read_domains;
			error->active_bo[i].write_domain = obj->write_domain;
			error->active_bo[i].fence_reg = obj_priv->fence_reg;
			error->active_bo[i].pinned = 0;
			if (obj_priv->pin_count > 0)
				error->active_bo[i].pinned = 1;
			if (obj_priv->user_pin_count > 0)
				error->active_bo[i].pinned = -1;
			error->active_bo[i].tiling = obj_priv->tiling_mode;
			error->active_bo[i].dirty = obj_priv->dirty;
			error->active_bo[i].purgeable = obj_priv->madv != I915_MADV_WILLNEED;

			if (++i == count)
				break;
		}
		error->active_bo_count = i;
	}

	do_gettimeofday(&error->time);

	spin_lock_irqsave(&dev_priv->error_lock, flags);
	if (dev_priv->first_error == NULL) {
		dev_priv->first_error = error;
		error = NULL;
	}
677
	spin_unlock_irqrestore(&dev_priv->error_lock, flags);
678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694

	if (error)
		i915_error_state_free(dev, error);
}

void i915_destroy_error_state(struct drm_device *dev)
{
	struct drm_i915_private *dev_priv = dev->dev_private;
	struct drm_i915_error_state *error;

	spin_lock(&dev_priv->error_lock);
	error = dev_priv->first_error;
	dev_priv->first_error = NULL;
	spin_unlock(&dev_priv->error_lock);

	if (error)
		i915_error_state_free(dev, error);
695 696
}

697 698 699 700 701 702 703 704 705 706
/**
 * i915_handle_error - handle an error interrupt
 * @dev: drm device
 *
 * Do some basic checking of regsiter state at error interrupt time and
 * dump it to the syslog.  Also call i915_capture_error_state() to make
 * sure we get a record and make it available in debugfs.  Fire a uevent
 * so userspace knows something bad happened (should trigger collection
 * of a ring dump etc.).
 */
707
static void i915_handle_error(struct drm_device *dev, bool wedged)
708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816
{
	struct drm_i915_private *dev_priv = dev->dev_private;
	u32 eir = I915_READ(EIR);
	u32 pipea_stats = I915_READ(PIPEASTAT);
	u32 pipeb_stats = I915_READ(PIPEBSTAT);

	i915_capture_error_state(dev);

	printk(KERN_ERR "render error detected, EIR: 0x%08x\n",
	       eir);

	if (IS_G4X(dev)) {
		if (eir & (GM45_ERROR_MEM_PRIV | GM45_ERROR_CP_PRIV)) {
			u32 ipeir = I915_READ(IPEIR_I965);

			printk(KERN_ERR "  IPEIR: 0x%08x\n",
			       I915_READ(IPEIR_I965));
			printk(KERN_ERR "  IPEHR: 0x%08x\n",
			       I915_READ(IPEHR_I965));
			printk(KERN_ERR "  INSTDONE: 0x%08x\n",
			       I915_READ(INSTDONE_I965));
			printk(KERN_ERR "  INSTPS: 0x%08x\n",
			       I915_READ(INSTPS));
			printk(KERN_ERR "  INSTDONE1: 0x%08x\n",
			       I915_READ(INSTDONE1));
			printk(KERN_ERR "  ACTHD: 0x%08x\n",
			       I915_READ(ACTHD_I965));
			I915_WRITE(IPEIR_I965, ipeir);
			(void)I915_READ(IPEIR_I965);
		}
		if (eir & GM45_ERROR_PAGE_TABLE) {
			u32 pgtbl_err = I915_READ(PGTBL_ER);
			printk(KERN_ERR "page table error\n");
			printk(KERN_ERR "  PGTBL_ER: 0x%08x\n",
			       pgtbl_err);
			I915_WRITE(PGTBL_ER, pgtbl_err);
			(void)I915_READ(PGTBL_ER);
		}
	}

	if (IS_I9XX(dev)) {
		if (eir & I915_ERROR_PAGE_TABLE) {
			u32 pgtbl_err = I915_READ(PGTBL_ER);
			printk(KERN_ERR "page table error\n");
			printk(KERN_ERR "  PGTBL_ER: 0x%08x\n",
			       pgtbl_err);
			I915_WRITE(PGTBL_ER, pgtbl_err);
			(void)I915_READ(PGTBL_ER);
		}
	}

	if (eir & I915_ERROR_MEMORY_REFRESH) {
		printk(KERN_ERR "memory refresh error\n");
		printk(KERN_ERR "PIPEASTAT: 0x%08x\n",
		       pipea_stats);
		printk(KERN_ERR "PIPEBSTAT: 0x%08x\n",
		       pipeb_stats);
		/* pipestat has already been acked */
	}
	if (eir & I915_ERROR_INSTRUCTION) {
		printk(KERN_ERR "instruction error\n");
		printk(KERN_ERR "  INSTPM: 0x%08x\n",
		       I915_READ(INSTPM));
		if (!IS_I965G(dev)) {
			u32 ipeir = I915_READ(IPEIR);

			printk(KERN_ERR "  IPEIR: 0x%08x\n",
			       I915_READ(IPEIR));
			printk(KERN_ERR "  IPEHR: 0x%08x\n",
			       I915_READ(IPEHR));
			printk(KERN_ERR "  INSTDONE: 0x%08x\n",
			       I915_READ(INSTDONE));
			printk(KERN_ERR "  ACTHD: 0x%08x\n",
			       I915_READ(ACTHD));
			I915_WRITE(IPEIR, ipeir);
			(void)I915_READ(IPEIR);
		} else {
			u32 ipeir = I915_READ(IPEIR_I965);

			printk(KERN_ERR "  IPEIR: 0x%08x\n",
			       I915_READ(IPEIR_I965));
			printk(KERN_ERR "  IPEHR: 0x%08x\n",
			       I915_READ(IPEHR_I965));
			printk(KERN_ERR "  INSTDONE: 0x%08x\n",
			       I915_READ(INSTDONE_I965));
			printk(KERN_ERR "  INSTPS: 0x%08x\n",
			       I915_READ(INSTPS));
			printk(KERN_ERR "  INSTDONE1: 0x%08x\n",
			       I915_READ(INSTDONE1));
			printk(KERN_ERR "  ACTHD: 0x%08x\n",
			       I915_READ(ACTHD_I965));
			I915_WRITE(IPEIR_I965, ipeir);
			(void)I915_READ(IPEIR_I965);
		}
	}

	I915_WRITE(EIR, eir);
	(void)I915_READ(EIR);
	eir = I915_READ(EIR);
	if (eir) {
		/*
		 * some errors might have become stuck,
		 * mask them.
		 */
		DRM_ERROR("EIR stuck: 0x%08x, masking\n", eir);
		I915_WRITE(EMR, I915_READ(EMR) | eir);
		I915_WRITE(IIR, I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT);
	}

817 818 819
	if (wedged) {
		atomic_set(&dev_priv->mm.wedged, 1);

820 821 822 823 824 825
		/*
		 * Wakeup waiting processes so they don't hang
		 */
		DRM_WAKEUP(&dev_priv->irq_queue);
	}

826
	queue_work(dev_priv->wq, &dev_priv->error_work);
827 828
}

L
Linus Torvalds 已提交
829 830
irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS)
{
831
	struct drm_device *dev = (struct drm_device *) arg;
L
Linus Torvalds 已提交
832
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
833
	struct drm_i915_master_private *master_priv;
834 835
	u32 iir, new_iir;
	u32 pipea_stats, pipeb_stats;
836 837
	u32 vblank_status;
	u32 vblank_enable;
838
	int vblank = 0;
839
	unsigned long irqflags;
840 841
	int irq_received;
	int ret = IRQ_NONE;
842

843 844
	atomic_inc(&dev_priv->irq_received);

845 846
	if (IS_IRONLAKE(dev))
		return ironlake_irq_handler(dev);
847

848
	iir = I915_READ(IIR);
849

850 851 852 853 854 855 856
	if (IS_I965G(dev)) {
		vblank_status = I915_START_VBLANK_INTERRUPT_STATUS;
		vblank_enable = PIPE_START_VBLANK_INTERRUPT_ENABLE;
	} else {
		vblank_status = I915_VBLANK_INTERRUPT_STATUS;
		vblank_enable = I915_VBLANK_INTERRUPT_ENABLE;
	}
857

858 859 860 861 862 863 864 865 866 867 868
	for (;;) {
		irq_received = iir != 0;

		/* Can't rely on pipestat interrupt bit in iir as it might
		 * have been cleared after the pipestat interrupt was received.
		 * It doesn't set the bit in iir again, but it still produces
		 * interrupts (for non-MSI).
		 */
		spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
		pipea_stats = I915_READ(PIPEASTAT);
		pipeb_stats = I915_READ(PIPEBSTAT);
J
Jesse Barnes 已提交
869

870
		if (iir & I915_RENDER_COMMAND_PARSER_ERROR_INTERRUPT)
871
			i915_handle_error(dev, false);
872

873 874 875
		/*
		 * Clear the PIPE(A|B)STAT regs before the IIR
		 */
876
		if (pipea_stats & 0x8000ffff) {
877
			if (pipea_stats &  PIPE_FIFO_UNDERRUN_STATUS)
878
				DRM_DEBUG_DRIVER("pipe a underrun\n");
879
			I915_WRITE(PIPEASTAT, pipea_stats);
880
			irq_received = 1;
881
		}
L
Linus Torvalds 已提交
882

883
		if (pipeb_stats & 0x8000ffff) {
884
			if (pipeb_stats &  PIPE_FIFO_UNDERRUN_STATUS)
885
				DRM_DEBUG_DRIVER("pipe b underrun\n");
886
			I915_WRITE(PIPEBSTAT, pipeb_stats);
887
			irq_received = 1;
888
		}
889 890 891 892 893 894
		spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);

		if (!irq_received)
			break;

		ret = IRQ_HANDLED;
895

896 897 898 899 900
		/* Consume port.  Then clear IIR or we'll miss events */
		if ((I915_HAS_HOTPLUG(dev)) &&
		    (iir & I915_DISPLAY_PORT_INTERRUPT)) {
			u32 hotplug_status = I915_READ(PORT_HOTPLUG_STAT);

901
			DRM_DEBUG_DRIVER("hotplug event received, stat 0x%08x\n",
902 903
				  hotplug_status);
			if (hotplug_status & dev_priv->hotplug_supported_mask)
904 905
				queue_work(dev_priv->wq,
					   &dev_priv->hotplug_work);
906 907 908 909 910

			I915_WRITE(PORT_HOTPLUG_STAT, hotplug_status);
			I915_READ(PORT_HOTPLUG_STAT);
		}

911 912
		I915_WRITE(IIR, iir);
		new_iir = I915_READ(IIR); /* Flush posted writes */
913

914 915 916 917 918 919
		if (dev->primary->master) {
			master_priv = dev->primary->master->driver_priv;
			if (master_priv->sarea_priv)
				master_priv->sarea_priv->last_dispatch =
					READ_BREADCRUMB(dev_priv);
		}
920

921
		if (iir & I915_USER_INTERRUPT) {
C
Chris Wilson 已提交
922 923 924
			u32 seqno = i915_get_gem_seqno(dev);
			dev_priv->mm.irq_gem_seqno = seqno;
			trace_i915_gem_request_complete(dev, seqno);
925
			DRM_WAKEUP(&dev_priv->irq_queue);
B
Ben Gamari 已提交
926 927
			dev_priv->hangcheck_count = 0;
			mod_timer(&dev_priv->hangcheck_timer, jiffies + DRM_I915_HANGCHECK_PERIOD);
928
		}
929

930 931 932 933 934 935
		if (iir & I915_DISPLAY_PLANE_A_FLIP_PENDING_INTERRUPT)
			intel_prepare_page_flip(dev, 0);

		if (iir & I915_DISPLAY_PLANE_B_FLIP_PENDING_INTERRUPT)
			intel_prepare_page_flip(dev, 1);

936
		if (pipea_stats & vblank_status) {
937 938
			vblank++;
			drm_handle_vblank(dev, 0);
939
			intel_finish_page_flip(dev, 0);
940
		}
941

942
		if (pipeb_stats & vblank_status) {
943 944
			vblank++;
			drm_handle_vblank(dev, 1);
945
			intel_finish_page_flip(dev, 1);
946
		}
947

948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967
		if ((pipeb_stats & I915_LEGACY_BLC_EVENT_STATUS) ||
		    (iir & I915_ASLE_INTERRUPT))
			opregion_asle_intr(dev);

		/* With MSI, interrupts are only generated when iir
		 * transitions from zero to nonzero.  If another bit got
		 * set while we were handling the existing iir bits, then
		 * we would never get another interrupt.
		 *
		 * This is fine on non-MSI as well, as if we hit this path
		 * we avoid exiting the interrupt handler only to generate
		 * another one.
		 *
		 * Note that for MSI this could cause a stray interrupt report
		 * if an interrupt landed in the time between writing IIR and
		 * the posting read.  This should be rare enough to never
		 * trigger the 99% of 100,000 interrupts test for disabling
		 * stray interrupts.
		 */
		iir = new_iir;
968
	}
969

970
	return ret;
L
Linus Torvalds 已提交
971 972
}

973
static int i915_emit_irq(struct drm_device * dev)
L
Linus Torvalds 已提交
974 975
{
	drm_i915_private_t *dev_priv = dev->dev_private;
976
	struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
L
Linus Torvalds 已提交
977 978 979 980
	RING_LOCALS;

	i915_kernel_lost_context(dev);

981
	DRM_DEBUG_DRIVER("\n");
L
Linus Torvalds 已提交
982

983
	dev_priv->counter++;
984
	if (dev_priv->counter > 0x7FFFFFFFUL)
985
		dev_priv->counter = 1;
986 987
	if (master_priv->sarea_priv)
		master_priv->sarea_priv->last_enqueue = dev_priv->counter;
988

989
	BEGIN_LP_RING(4);
990
	OUT_RING(MI_STORE_DWORD_INDEX);
991
	OUT_RING(I915_BREADCRUMB_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
992
	OUT_RING(dev_priv->counter);
993
	OUT_RING(MI_USER_INTERRUPT);
L
Linus Torvalds 已提交
994
	ADVANCE_LP_RING();
D
Dave Airlie 已提交
995

996
	return dev_priv->counter;
L
Linus Torvalds 已提交
997 998
}

999
void i915_user_irq_get(struct drm_device *dev)
1000 1001
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1002
	unsigned long irqflags;
1003

1004
	spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
1005
	if (dev->irq_enabled && (++dev_priv->user_irq_refcount == 1)) {
1006 1007
		if (IS_IRONLAKE(dev))
			ironlake_enable_graphics_irq(dev_priv, GT_USER_INTERRUPT);
1008 1009 1010
		else
			i915_enable_irq(dev_priv, I915_USER_INTERRUPT);
	}
1011
	spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);
1012 1013
}

1014
void i915_user_irq_put(struct drm_device *dev)
1015 1016
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1017
	unsigned long irqflags;
1018

1019
	spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
1020
	BUG_ON(dev->irq_enabled && dev_priv->user_irq_refcount <= 0);
1021
	if (dev->irq_enabled && (--dev_priv->user_irq_refcount == 0)) {
1022 1023
		if (IS_IRONLAKE(dev))
			ironlake_disable_graphics_irq(dev_priv, GT_USER_INTERRUPT);
1024 1025 1026
		else
			i915_disable_irq(dev_priv, I915_USER_INTERRUPT);
	}
1027
	spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);
1028 1029
}

1030 1031 1032 1033 1034 1035 1036 1037 1038 1039
void i915_trace_irq_get(struct drm_device *dev, u32 seqno)
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;

	if (dev_priv->trace_irq_seqno == 0)
		i915_user_irq_get(dev);

	dev_priv->trace_irq_seqno = seqno;
}

1040
static int i915_wait_irq(struct drm_device * dev, int irq_nr)
L
Linus Torvalds 已提交
1041 1042
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1043
	struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
L
Linus Torvalds 已提交
1044 1045
	int ret = 0;

1046
	DRM_DEBUG_DRIVER("irq_nr=%d breadcrumb=%d\n", irq_nr,
L
Linus Torvalds 已提交
1047 1048
		  READ_BREADCRUMB(dev_priv));

1049
	if (READ_BREADCRUMB(dev_priv) >= irq_nr) {
1050 1051
		if (master_priv->sarea_priv)
			master_priv->sarea_priv->last_dispatch = READ_BREADCRUMB(dev_priv);
L
Linus Torvalds 已提交
1052
		return 0;
1053
	}
L
Linus Torvalds 已提交
1054

1055 1056
	if (master_priv->sarea_priv)
		master_priv->sarea_priv->perf_boxes |= I915_BOX_WAIT;
L
Linus Torvalds 已提交
1057

1058
	i915_user_irq_get(dev);
L
Linus Torvalds 已提交
1059 1060
	DRM_WAIT_ON(ret, dev_priv->irq_queue, 3 * DRM_HZ,
		    READ_BREADCRUMB(dev_priv) >= irq_nr);
1061
	i915_user_irq_put(dev);
L
Linus Torvalds 已提交
1062

E
Eric Anholt 已提交
1063
	if (ret == -EBUSY) {
1064
		DRM_ERROR("EBUSY -- rec: %d emitted: %d\n",
L
Linus Torvalds 已提交
1065 1066 1067
			  READ_BREADCRUMB(dev_priv), (int)dev_priv->counter);
	}

1068 1069 1070
	return ret;
}

L
Linus Torvalds 已提交
1071 1072
/* Needs the lock as it touches the ring.
 */
1073 1074
int i915_irq_emit(struct drm_device *dev, void *data,
			 struct drm_file *file_priv)
L
Linus Torvalds 已提交
1075 1076
{
	drm_i915_private_t *dev_priv = dev->dev_private;
1077
	drm_i915_irq_emit_t *emit = data;
L
Linus Torvalds 已提交
1078 1079
	int result;

1080
	if (!dev_priv || !dev_priv->ring.virtual_start) {
1081
		DRM_ERROR("called with no initialization\n");
E
Eric Anholt 已提交
1082
		return -EINVAL;
L
Linus Torvalds 已提交
1083
	}
1084 1085 1086

	RING_LOCK_TEST_WITH_RETURN(dev, file_priv);

1087
	mutex_lock(&dev->struct_mutex);
L
Linus Torvalds 已提交
1088
	result = i915_emit_irq(dev);
1089
	mutex_unlock(&dev->struct_mutex);
L
Linus Torvalds 已提交
1090

1091
	if (DRM_COPY_TO_USER(emit->irq_seq, &result, sizeof(int))) {
L
Linus Torvalds 已提交
1092
		DRM_ERROR("copy_to_user\n");
E
Eric Anholt 已提交
1093
		return -EFAULT;
L
Linus Torvalds 已提交
1094 1095 1096 1097 1098 1099 1100
	}

	return 0;
}

/* Doesn't need the hardware lock.
 */
1101 1102
int i915_irq_wait(struct drm_device *dev, void *data,
			 struct drm_file *file_priv)
L
Linus Torvalds 已提交
1103 1104
{
	drm_i915_private_t *dev_priv = dev->dev_private;
1105
	drm_i915_irq_wait_t *irqwait = data;
L
Linus Torvalds 已提交
1106 1107

	if (!dev_priv) {
1108
		DRM_ERROR("called with no initialization\n");
E
Eric Anholt 已提交
1109
		return -EINVAL;
L
Linus Torvalds 已提交
1110 1111
	}

1112
	return i915_wait_irq(dev, irqwait->irq_seq);
L
Linus Torvalds 已提交
1113 1114
}

1115 1116 1117 1118
/* Called from drm generic code, passed 'crtc' which
 * we use as a pipe index
 */
int i915_enable_vblank(struct drm_device *dev, int pipe)
1119 1120
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1121
	unsigned long irqflags;
1122 1123 1124 1125 1126 1127
	int pipeconf_reg = (pipe == 0) ? PIPEACONF : PIPEBCONF;
	u32 pipeconf;

	pipeconf = I915_READ(pipeconf_reg);
	if (!(pipeconf & PIPEACONF_ENABLE))
		return -EINVAL;
1128

1129
	spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
1130 1131 1132 1133
	if (IS_IRONLAKE(dev))
		ironlake_enable_display_irq(dev_priv, (pipe == 0) ? 
					    DE_PIPEA_VBLANK: DE_PIPEB_VBLANK);
	else if (IS_I965G(dev))
1134 1135
		i915_enable_pipestat(dev_priv, pipe,
				     PIPE_START_VBLANK_INTERRUPT_ENABLE);
1136
	else
1137 1138
		i915_enable_pipestat(dev_priv, pipe,
				     PIPE_VBLANK_INTERRUPT_ENABLE);
1139
	spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);
1140 1141 1142
	return 0;
}

1143 1144 1145 1146
/* Called from drm generic code, passed 'crtc' which
 * we use as a pipe index
 */
void i915_disable_vblank(struct drm_device *dev, int pipe)
1147 1148
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1149
	unsigned long irqflags;
1150

1151
	spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
1152 1153 1154 1155 1156 1157 1158
	if (IS_IRONLAKE(dev))
		ironlake_disable_display_irq(dev_priv, (pipe == 0) ? 
					     DE_PIPEA_VBLANK: DE_PIPEB_VBLANK);
	else
		i915_disable_pipestat(dev_priv, pipe,
				      PIPE_VBLANK_INTERRUPT_ENABLE |
				      PIPE_START_VBLANK_INTERRUPT_ENABLE);
1159
	spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);
1160 1161
}

J
Jesse Barnes 已提交
1162 1163 1164
void i915_enable_interrupt (struct drm_device *dev)
{
	struct drm_i915_private *dev_priv = dev->dev_private;
1165

1166
	if (!IS_IRONLAKE(dev))
1167
		opregion_enable_asle(dev);
J
Jesse Barnes 已提交
1168 1169 1170 1171
	dev_priv->irq_enabled = 1;
}


1172 1173
/* Set the vblank monitor pipe
 */
1174 1175
int i915_vblank_pipe_set(struct drm_device *dev, void *data,
			 struct drm_file *file_priv)
1176 1177 1178 1179
{
	drm_i915_private_t *dev_priv = dev->dev_private;

	if (!dev_priv) {
1180
		DRM_ERROR("called with no initialization\n");
E
Eric Anholt 已提交
1181
		return -EINVAL;
1182 1183
	}

1184
	return 0;
1185 1186
}

1187 1188
int i915_vblank_pipe_get(struct drm_device *dev, void *data,
			 struct drm_file *file_priv)
1189 1190
{
	drm_i915_private_t *dev_priv = dev->dev_private;
1191
	drm_i915_vblank_pipe_t *pipe = data;
1192 1193

	if (!dev_priv) {
1194
		DRM_ERROR("called with no initialization\n");
E
Eric Anholt 已提交
1195
		return -EINVAL;
1196 1197
	}

1198
	pipe->pipe = DRM_I915_VBLANK_PIPE_A | DRM_I915_VBLANK_PIPE_B;
1199

1200 1201 1202
	return 0;
}

1203 1204 1205
/**
 * Schedule buffer swap at given vertical blank.
 */
1206 1207
int i915_vblank_swap(struct drm_device *dev, void *data,
		     struct drm_file *file_priv)
1208
{
1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221
	/* The delayed swap mechanism was fundamentally racy, and has been
	 * removed.  The model was that the client requested a delayed flip/swap
	 * from the kernel, then waited for vblank before continuing to perform
	 * rendering.  The problem was that the kernel might wake the client
	 * up before it dispatched the vblank swap (since the lock has to be
	 * held while touching the ringbuffer), in which case the client would
	 * clear and start the next frame before the swap occurred, and
	 * flicker would occur in addition to likely missing the vblank.
	 *
	 * In the absence of this ioctl, userland falls back to a correct path
	 * of waiting for a vblank, then dispatching the swap on its own.
	 * Context switching to userland and back is plenty fast enough for
	 * meeting the requirements of vblank swapping.
1222
	 */
1223
	return -EINVAL;
1224 1225
}

B
Ben Gamari 已提交
1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256
struct drm_i915_gem_request *i915_get_tail_request(struct drm_device *dev) {
	drm_i915_private_t *dev_priv = dev->dev_private;
	return list_entry(dev_priv->mm.request_list.prev, struct drm_i915_gem_request, list);
}

/**
 * This is called when the chip hasn't reported back with completed
 * batchbuffers in a long time. The first time this is called we simply record
 * ACTHD. If ACTHD hasn't changed by the time the hangcheck timer elapses
 * again, we assume the chip is wedged and try to fix it.
 */
void i915_hangcheck_elapsed(unsigned long data)
{
	struct drm_device *dev = (struct drm_device *)data;
	drm_i915_private_t *dev_priv = dev->dev_private;
	uint32_t acthd;
       
	if (!IS_I965G(dev))
		acthd = I915_READ(ACTHD);
	else
		acthd = I915_READ(ACTHD_I965);

	/* If all work is done then ACTHD clearly hasn't advanced. */
	if (list_empty(&dev_priv->mm.request_list) ||
		       i915_seqno_passed(i915_get_gem_seqno(dev), i915_get_tail_request(dev)->seqno)) {
		dev_priv->hangcheck_count = 0;
		return;
	}

	if (dev_priv->last_acthd == acthd && dev_priv->hangcheck_count > 0) {
		DRM_ERROR("Hangcheck timer elapsed... GPU hung\n");
1257
		i915_handle_error(dev, true);
B
Ben Gamari 已提交
1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271
		return;
	} 

	/* Reset timer case chip hangs without another request being added */
	mod_timer(&dev_priv->hangcheck_timer, jiffies + DRM_I915_HANGCHECK_PERIOD);

	if (acthd != dev_priv->last_acthd)
		dev_priv->hangcheck_count = 0;
	else
		dev_priv->hangcheck_count++;

	dev_priv->last_acthd = acthd;
}

L
Linus Torvalds 已提交
1272 1273
/* drm_dma.h hooks
*/
1274
static void ironlake_irq_preinstall(struct drm_device *dev)
1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;

	I915_WRITE(HWSTAM, 0xeffe);

	/* XXX hotplug from PCH */

	I915_WRITE(DEIMR, 0xffffffff);
	I915_WRITE(DEIER, 0x0);
	(void) I915_READ(DEIER);

	/* and GT */
	I915_WRITE(GTIMR, 0xffffffff);
	I915_WRITE(GTIER, 0x0);
	(void) I915_READ(GTIER);
1290 1291 1292 1293 1294

	/* south display irq */
	I915_WRITE(SDEIMR, 0xffffffff);
	I915_WRITE(SDEIER, 0x0);
	(void) I915_READ(SDEIER);
1295 1296
}

1297
static int ironlake_irq_postinstall(struct drm_device *dev)
1298 1299 1300
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	/* enable kind of interrupts always enabled */
1301 1302
	u32 display_mask = DE_MASTER_IRQ_CONTROL | DE_GSE | DE_PCH_EVENT |
			   DE_PLANEA_FLIP_DONE | DE_PLANEB_FLIP_DONE;
1303
	u32 render_mask = GT_USER_INTERRUPT;
1304 1305
	u32 hotplug_mask = SDE_CRT_HOTPLUG | SDE_PORTB_HOTPLUG |
			   SDE_PORTC_HOTPLUG | SDE_PORTD_HOTPLUG;
1306 1307

	dev_priv->irq_mask_reg = ~display_mask;
1308
	dev_priv->de_irq_enable_reg = display_mask | DE_PIPEA_VBLANK | DE_PIPEB_VBLANK;
1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324

	/* should always can generate irq */
	I915_WRITE(DEIIR, I915_READ(DEIIR));
	I915_WRITE(DEIMR, dev_priv->irq_mask_reg);
	I915_WRITE(DEIER, dev_priv->de_irq_enable_reg);
	(void) I915_READ(DEIER);

	/* user interrupt should be enabled, but masked initial */
	dev_priv->gt_irq_mask_reg = 0xffffffff;
	dev_priv->gt_irq_enable_reg = render_mask;

	I915_WRITE(GTIIR, I915_READ(GTIIR));
	I915_WRITE(GTIMR, dev_priv->gt_irq_mask_reg);
	I915_WRITE(GTIER, dev_priv->gt_irq_enable_reg);
	(void) I915_READ(GTIER);

1325 1326 1327 1328 1329 1330 1331 1332
	dev_priv->pch_irq_mask_reg = ~hotplug_mask;
	dev_priv->pch_irq_enable_reg = hotplug_mask;

	I915_WRITE(SDEIIR, I915_READ(SDEIIR));
	I915_WRITE(SDEIMR, dev_priv->pch_irq_mask_reg);
	I915_WRITE(SDEIER, dev_priv->pch_irq_enable_reg);
	(void) I915_READ(SDEIER);

1333 1334 1335 1336 1337 1338 1339
	if (IS_IRONLAKE_M(dev)) {
		/* Clear & enable PCU event interrupts */
		I915_WRITE(DEIIR, DE_PCU_EVENT);
		I915_WRITE(DEIER, I915_READ(DEIER) | DE_PCU_EVENT);
		ironlake_enable_display_irq(dev_priv, DE_PCU_EVENT);
	}

1340 1341 1342
	return 0;
}

1343
void i915_driver_irq_preinstall(struct drm_device * dev)
L
Linus Torvalds 已提交
1344 1345 1346
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;

J
Jesse Barnes 已提交
1347 1348
	atomic_set(&dev_priv->irq_received, 0);

1349
	INIT_WORK(&dev_priv->hotplug_work, i915_hotplug_work_func);
1350
	INIT_WORK(&dev_priv->error_work, i915_error_work_func);
1351

1352 1353
	if (IS_IRONLAKE(dev)) {
		ironlake_irq_preinstall(dev);
1354 1355 1356
		return;
	}

1357 1358 1359 1360 1361
	if (I915_HAS_HOTPLUG(dev)) {
		I915_WRITE(PORT_HOTPLUG_EN, 0);
		I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
	}

1362
	I915_WRITE(HWSTAM, 0xeffe);
1363 1364
	I915_WRITE(PIPEASTAT, 0);
	I915_WRITE(PIPEBSTAT, 0);
1365
	I915_WRITE(IMR, 0xffffffff);
1366
	I915_WRITE(IER, 0x0);
1367
	(void) I915_READ(IER);
L
Linus Torvalds 已提交
1368 1369
}

1370 1371 1372 1373
/*
 * Must be called after intel_modeset_init or hotplug interrupts won't be
 * enabled correctly.
 */
1374
int i915_driver_irq_postinstall(struct drm_device *dev)
L
Linus Torvalds 已提交
1375 1376
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1377
	u32 enable_mask = I915_INTERRUPT_ENABLE_FIX | I915_INTERRUPT_ENABLE_VAR;
1378
	u32 error_mask;
1379

1380 1381
	DRM_INIT_WAITQUEUE(&dev_priv->irq_queue);

1382 1383
	dev_priv->vblank_pipe = DRM_I915_VBLANK_PIPE_A | DRM_I915_VBLANK_PIPE_B;

1384 1385
	if (IS_IRONLAKE(dev))
		return ironlake_irq_postinstall(dev);
1386

1387 1388 1389 1390 1391 1392
	/* Unmask the interrupts that we always want on. */
	dev_priv->irq_mask_reg = ~I915_INTERRUPT_ENABLE_FIX;

	dev_priv->pipestat[0] = 0;
	dev_priv->pipestat[1] = 0;

1393 1394 1395
	if (I915_HAS_HOTPLUG(dev)) {
		u32 hotplug_en = I915_READ(PORT_HOTPLUG_EN);

1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410
		/* Note HDMI and DP share bits */
		if (dev_priv->hotplug_supported_mask & HDMIB_HOTPLUG_INT_STATUS)
			hotplug_en |= HDMIB_HOTPLUG_INT_EN;
		if (dev_priv->hotplug_supported_mask & HDMIC_HOTPLUG_INT_STATUS)
			hotplug_en |= HDMIC_HOTPLUG_INT_EN;
		if (dev_priv->hotplug_supported_mask & HDMID_HOTPLUG_INT_STATUS)
			hotplug_en |= HDMID_HOTPLUG_INT_EN;
		if (dev_priv->hotplug_supported_mask & SDVOC_HOTPLUG_INT_STATUS)
			hotplug_en |= SDVOC_HOTPLUG_INT_EN;
		if (dev_priv->hotplug_supported_mask & SDVOB_HOTPLUG_INT_STATUS)
			hotplug_en |= SDVOB_HOTPLUG_INT_EN;
		if (dev_priv->hotplug_supported_mask & CRT_HOTPLUG_INT_STATUS)
			hotplug_en |= CRT_HOTPLUG_INT_EN;
		/* Ignore TV since it's buggy */

1411 1412 1413 1414 1415 1416 1417 1418
		I915_WRITE(PORT_HOTPLUG_EN, hotplug_en);

		/* Enable in IER... */
		enable_mask |= I915_DISPLAY_PORT_INTERRUPT;
		/* and unmask in IMR */
		i915_enable_irq(dev_priv, I915_DISPLAY_PORT_INTERRUPT);
	}

1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433
	/*
	 * Enable some error detection, note the instruction error mask
	 * bit is reserved, so we leave it masked.
	 */
	if (IS_G4X(dev)) {
		error_mask = ~(GM45_ERROR_PAGE_TABLE |
			       GM45_ERROR_MEM_PRIV |
			       GM45_ERROR_CP_PRIV |
			       I915_ERROR_MEMORY_REFRESH);
	} else {
		error_mask = ~(I915_ERROR_PAGE_TABLE |
			       I915_ERROR_MEMORY_REFRESH);
	}
	I915_WRITE(EMR, error_mask);

1434 1435 1436 1437 1438
	/* Disable pipe interrupt enables, clear pending pipe status */
	I915_WRITE(PIPEASTAT, I915_READ(PIPEASTAT) & 0x8000ffff);
	I915_WRITE(PIPEBSTAT, I915_READ(PIPEBSTAT) & 0x8000ffff);
	/* Clear pending interrupt status */
	I915_WRITE(IIR, I915_READ(IIR));
1439

1440
	I915_WRITE(IER, enable_mask);
1441
	I915_WRITE(IMR, dev_priv->irq_mask_reg);
1442 1443
	(void) I915_READ(IER);

1444
	opregion_enable_asle(dev);
1445 1446

	return 0;
L
Linus Torvalds 已提交
1447 1448
}

1449
static void ironlake_irq_uninstall(struct drm_device *dev)
1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	I915_WRITE(HWSTAM, 0xffffffff);

	I915_WRITE(DEIMR, 0xffffffff);
	I915_WRITE(DEIER, 0x0);
	I915_WRITE(DEIIR, I915_READ(DEIIR));

	I915_WRITE(GTIMR, 0xffffffff);
	I915_WRITE(GTIER, 0x0);
	I915_WRITE(GTIIR, I915_READ(GTIIR));
}

1463
void i915_driver_irq_uninstall(struct drm_device * dev)
L
Linus Torvalds 已提交
1464 1465
{
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
1466

L
Linus Torvalds 已提交
1467 1468 1469
	if (!dev_priv)
		return;

1470 1471
	dev_priv->vblank_pipe = 0;

1472 1473
	if (IS_IRONLAKE(dev)) {
		ironlake_irq_uninstall(dev);
1474 1475 1476
		return;
	}

1477 1478 1479 1480 1481
	if (I915_HAS_HOTPLUG(dev)) {
		I915_WRITE(PORT_HOTPLUG_EN, 0);
		I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT));
	}

1482
	I915_WRITE(HWSTAM, 0xffffffff);
1483 1484
	I915_WRITE(PIPEASTAT, 0);
	I915_WRITE(PIPEBSTAT, 0);
1485
	I915_WRITE(IMR, 0xffffffff);
1486
	I915_WRITE(IER, 0x0);
1487

1488 1489 1490
	I915_WRITE(PIPEASTAT, I915_READ(PIPEASTAT) & 0x8000ffff);
	I915_WRITE(PIPEBSTAT, I915_READ(PIPEBSTAT) & 0x8000ffff);
	I915_WRITE(IIR, I915_READ(IIR));
L
Linus Torvalds 已提交
1491
}