intel_ringbuffer.c 31.9 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32
/*
 * Copyright © 2008-2010 Intel Corporation
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice (including the next
 * paragraph) shall be included in all copies or substantial portions of the
 * Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
 * IN THE SOFTWARE.
 *
 * Authors:
 *    Eric Anholt <eric@anholt.net>
 *    Zou Nan hai <nanhai.zou@intel.com>
 *    Xiang Hai hao<haihao.xiang@intel.com>
 *
 */

#include "drmP.h"
#include "drm.h"
#include "i915_drv.h"
33
#include "i915_drm.h"
34
#include "i915_trace.h"
35
#include "intel_drv.h"
36

37 38 39 40 41 42 43 44 45 46 47 48 49 50
static u32 i915_gem_get_seqno(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
	u32 seqno;

	seqno = dev_priv->next_seqno;

	/* reserve 0 for non-seqno */
	if (++dev_priv->next_seqno == 0)
		dev_priv->next_seqno = 1;

	return seqno;
}

51
static int
52
render_ring_flush(struct intel_ring_buffer *ring,
53 54
		  u32	invalidate_domains,
		  u32	flush_domains)
55
{
56
	struct drm_device *dev = ring->dev;
57 58
	drm_i915_private_t *dev_priv = dev->dev_private;
	u32 cmd;
59
	int ret;
60

61 62 63 64
#if WATCH_EXEC
	DRM_INFO("%s: invalidate %08x flush %08x\n", __func__,
		  invalidate_domains, flush_domains);
#endif
65 66

	trace_i915_gem_request_flush(dev, dev_priv->next_seqno,
67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101
				     invalidate_domains, flush_domains);

	if ((invalidate_domains | flush_domains) & I915_GEM_GPU_DOMAINS) {
		/*
		 * read/write caches:
		 *
		 * I915_GEM_DOMAIN_RENDER is always invalidated, but is
		 * only flushed if MI_NO_WRITE_FLUSH is unset.  On 965, it is
		 * also flushed at 2d versus 3d pipeline switches.
		 *
		 * read-only caches:
		 *
		 * I915_GEM_DOMAIN_SAMPLER is flushed on pre-965 if
		 * MI_READ_FLUSH is set, and is always flushed on 965.
		 *
		 * I915_GEM_DOMAIN_COMMAND may not exist?
		 *
		 * I915_GEM_DOMAIN_INSTRUCTION, which exists on 965, is
		 * invalidated when MI_EXE_FLUSH is set.
		 *
		 * I915_GEM_DOMAIN_VERTEX, which exists on 965, is
		 * invalidated with every MI_FLUSH.
		 *
		 * TLBs:
		 *
		 * On 965, TLBs associated with I915_GEM_DOMAIN_COMMAND
		 * and I915_GEM_DOMAIN_CPU in are invalidated at PTE write and
		 * I915_GEM_DOMAIN_RENDER and I915_GEM_DOMAIN_SAMPLER
		 * are flushed at any MI_FLUSH.
		 */

		cmd = MI_FLUSH | MI_NO_WRITE_FLUSH;
		if ((invalidate_domains|flush_domains) &
		    I915_GEM_DOMAIN_RENDER)
			cmd &= ~MI_NO_WRITE_FLUSH;
102
		if (INTEL_INFO(dev)->gen < 4) {
103 104 105 106 107 108 109 110 111 112
			/*
			 * On the 965, the sampler cache always gets flushed
			 * and this bit is reserved.
			 */
			if (invalidate_domains & I915_GEM_DOMAIN_SAMPLER)
				cmd |= MI_READ_FLUSH;
		}
		if (invalidate_domains & I915_GEM_DOMAIN_INSTRUCTION)
			cmd |= MI_EXE_FLUSH;

113 114 115 116
		if (invalidate_domains & I915_GEM_DOMAIN_COMMAND &&
		    (IS_G4X(dev) || IS_GEN5(dev)))
			cmd |= MI_INVALIDATE_ISP;

117 118 119
#if WATCH_EXEC
		DRM_INFO("%s: queue flush %08x to ring\n", __func__, cmd);
#endif
120 121 122 123 124 125 126
		ret = intel_ring_begin(ring, 2);
		if (ret)
			return ret;

		intel_ring_emit(ring, cmd);
		intel_ring_emit(ring, MI_NOOP);
		intel_ring_advance(ring);
127
	}
128 129

	return 0;
130 131
}

132
static void ring_write_tail(struct intel_ring_buffer *ring,
133
			    u32 value)
134
{
135
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
136
	I915_WRITE_TAIL(ring, value);
137 138
}

139
u32 intel_ring_get_active_head(struct intel_ring_buffer *ring)
140
{
141 142
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
	u32 acthd_reg = INTEL_INFO(ring->dev)->gen >= 4 ?
D
Daniel Vetter 已提交
143
			RING_ACTHD(ring->mmio_base) : ACTHD;
144 145 146 147

	return I915_READ(acthd_reg);
}

148
static int init_ring_common(struct intel_ring_buffer *ring)
149
{
150
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
151
	struct drm_i915_gem_object *obj = ring->obj;
152 153 154
	u32 head;

	/* Stop the ring if it's running. */
155
	I915_WRITE_CTL(ring, 0);
156
	I915_WRITE_HEAD(ring, 0);
157
	ring->write_tail(ring, 0);
158 159

	/* Initialize the ring. */
160
	I915_WRITE_START(ring, obj->gtt_offset);
161
	head = I915_READ_HEAD(ring) & HEAD_ADDR;
162 163 164

	/* G45 ring initialization fails to reset head to zero */
	if (head != 0) {
165 166 167 168 169 170 171
		DRM_DEBUG_KMS("%s head not reset to zero "
			      "ctl %08x head %08x tail %08x start %08x\n",
			      ring->name,
			      I915_READ_CTL(ring),
			      I915_READ_HEAD(ring),
			      I915_READ_TAIL(ring),
			      I915_READ_START(ring));
172

173
		I915_WRITE_HEAD(ring, 0);
174

175 176 177 178 179 180 181 182 183
		if (I915_READ_HEAD(ring) & HEAD_ADDR) {
			DRM_ERROR("failed to set %s head to zero "
				  "ctl %08x head %08x tail %08x start %08x\n",
				  ring->name,
				  I915_READ_CTL(ring),
				  I915_READ_HEAD(ring),
				  I915_READ_TAIL(ring),
				  I915_READ_START(ring));
		}
184 185
	}

186
	I915_WRITE_CTL(ring,
187
			((ring->size - PAGE_SIZE) & RING_NR_PAGES)
188
			| RING_REPORT_64K | RING_VALID);
189 190

	/* If the head is still not zero, the ring is dead */
191
	if ((I915_READ_CTL(ring) & RING_VALID) == 0 ||
192
	    I915_READ_START(ring) != obj->gtt_offset ||
193
	    (I915_READ_HEAD(ring) & HEAD_ADDR) != 0) {
194 195 196 197 198 199 200 201
		DRM_ERROR("%s initialization failed "
				"ctl %08x head %08x tail %08x start %08x\n",
				ring->name,
				I915_READ_CTL(ring),
				I915_READ_HEAD(ring),
				I915_READ_TAIL(ring),
				I915_READ_START(ring));
		return -EIO;
202 203
	}

204 205
	if (!drm_core_check_feature(ring->dev, DRIVER_MODESET))
		i915_kernel_lost_context(ring->dev);
206
	else {
207
		ring->head = I915_READ_HEAD(ring) & HEAD_ADDR;
208
		ring->tail = I915_READ_TAIL(ring) & TAIL_ADDR;
209 210 211 212
		ring->space = ring->head - (ring->tail + 8);
		if (ring->space < 0)
			ring->space += ring->size;
	}
213

214 215 216
	return 0;
}

217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288
/*
 * 965+ support PIPE_CONTROL commands, which provide finer grained control
 * over cache flushing.
 */
struct pipe_control {
	struct drm_i915_gem_object *obj;
	volatile u32 *cpu_page;
	u32 gtt_offset;
};

static int
init_pipe_control(struct intel_ring_buffer *ring)
{
	struct pipe_control *pc;
	struct drm_i915_gem_object *obj;
	int ret;

	if (ring->private)
		return 0;

	pc = kmalloc(sizeof(*pc), GFP_KERNEL);
	if (!pc)
		return -ENOMEM;

	obj = i915_gem_alloc_object(ring->dev, 4096);
	if (obj == NULL) {
		DRM_ERROR("Failed to allocate seqno page\n");
		ret = -ENOMEM;
		goto err;
	}
	obj->agp_type = AGP_USER_CACHED_MEMORY;

	ret = i915_gem_object_pin(obj, 4096, true);
	if (ret)
		goto err_unref;

	pc->gtt_offset = obj->gtt_offset;
	pc->cpu_page =  kmap(obj->pages[0]);
	if (pc->cpu_page == NULL)
		goto err_unpin;

	pc->obj = obj;
	ring->private = pc;
	return 0;

err_unpin:
	i915_gem_object_unpin(obj);
err_unref:
	drm_gem_object_unreference(&obj->base);
err:
	kfree(pc);
	return ret;
}

static void
cleanup_pipe_control(struct intel_ring_buffer *ring)
{
	struct pipe_control *pc = ring->private;
	struct drm_i915_gem_object *obj;

	if (!ring->private)
		return;

	obj = pc->obj;
	kunmap(obj->pages[0]);
	i915_gem_object_unpin(obj);
	drm_gem_object_unreference(&obj->base);

	kfree(pc);
	ring->private = NULL;
}

289
static int init_render_ring(struct intel_ring_buffer *ring)
290
{
291
	struct drm_device *dev = ring->dev;
292
	struct drm_i915_private *dev_priv = dev->dev_private;
293
	int ret = init_ring_common(ring);
294

295
	if (INTEL_INFO(dev)->gen > 3) {
296
		int mode = VS_TIMER_DISPATCH << 16 | VS_TIMER_DISPATCH;
297 298 299
		if (IS_GEN6(dev))
			mode |= MI_FLUSH_ENABLE << 16 | MI_FLUSH_ENABLE;
		I915_WRITE(MI_MODE, mode);
300
	}
301

302 303 304 305 306 307 308
	if (INTEL_INFO(dev)->gen >= 6) {
	} else if (IS_GEN5(dev)) {
		ret = init_pipe_control(ring);
		if (ret)
			return ret;
	}

309 310 311
	return ret;
}

312 313 314 315 316 317 318 319
static void render_ring_cleanup(struct intel_ring_buffer *ring)
{
	if (!ring->private)
		return;

	cleanup_pipe_control(ring);
}

320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393
static void
update_semaphore(struct intel_ring_buffer *ring, int i, u32 seqno)
{
	struct drm_device *dev = ring->dev;
	struct drm_i915_private *dev_priv = dev->dev_private;
	int id;

	/*
	 * cs -> 1 = vcs, 0 = bcs
	 * vcs -> 1 = bcs, 0 = cs,
	 * bcs -> 1 = cs, 0 = vcs.
	 */
	id = ring - dev_priv->ring;
	id += 2 - i;
	id %= 3;

	intel_ring_emit(ring,
			MI_SEMAPHORE_MBOX |
			MI_SEMAPHORE_REGISTER |
			MI_SEMAPHORE_UPDATE);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring,
			RING_SYNC_0(dev_priv->ring[id].mmio_base) + 4*i);
}

static int
gen6_add_request(struct intel_ring_buffer *ring,
		 u32 *result)
{
	u32 seqno;
	int ret;

	ret = intel_ring_begin(ring, 10);
	if (ret)
		return ret;

	seqno = i915_gem_get_seqno(ring->dev);
	update_semaphore(ring, 0, seqno);
	update_semaphore(ring, 1, seqno);

	intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
	intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, MI_USER_INTERRUPT);
	intel_ring_advance(ring);

	*result = seqno;
	return 0;
}

int
intel_ring_sync(struct intel_ring_buffer *ring,
		struct intel_ring_buffer *to,
		u32 seqno)
{
	int ret;

	ret = intel_ring_begin(ring, 4);
	if (ret)
		return ret;

	intel_ring_emit(ring,
			MI_SEMAPHORE_MBOX |
			MI_SEMAPHORE_REGISTER |
			intel_ring_sync_index(ring, to) << 17 |
			MI_SEMAPHORE_COMPARE);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, 0);
	intel_ring_emit(ring, MI_NOOP);
	intel_ring_advance(ring);

	return 0;
}

394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452
#define PIPE_CONTROL_FLUSH(ring__, addr__)					\
do {									\
	intel_ring_emit(ring__, GFX_OP_PIPE_CONTROL | PIPE_CONTROL_QW_WRITE |		\
		 PIPE_CONTROL_DEPTH_STALL | 2);				\
	intel_ring_emit(ring__, (addr__) | PIPE_CONTROL_GLOBAL_GTT);			\
	intel_ring_emit(ring__, 0);							\
	intel_ring_emit(ring__, 0);							\
} while (0)

static int
pc_render_add_request(struct intel_ring_buffer *ring,
		      u32 *result)
{
	struct drm_device *dev = ring->dev;
	u32 seqno = i915_gem_get_seqno(dev);
	struct pipe_control *pc = ring->private;
	u32 scratch_addr = pc->gtt_offset + 128;
	int ret;

	/* For Ironlake, MI_USER_INTERRUPT was deprecated and apparently
	 * incoherent with writes to memory, i.e. completely fubar,
	 * so we need to use PIPE_NOTIFY instead.
	 *
	 * However, we also need to workaround the qword write
	 * incoherence by flushing the 6 PIPE_NOTIFY buffers out to
	 * memory before requesting an interrupt.
	 */
	ret = intel_ring_begin(ring, 32);
	if (ret)
		return ret;

	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL | PIPE_CONTROL_QW_WRITE |
			PIPE_CONTROL_WC_FLUSH | PIPE_CONTROL_TC_FLUSH);
	intel_ring_emit(ring, pc->gtt_offset | PIPE_CONTROL_GLOBAL_GTT);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, 0);
	PIPE_CONTROL_FLUSH(ring, scratch_addr);
	scratch_addr += 128; /* write to separate cachelines */
	PIPE_CONTROL_FLUSH(ring, scratch_addr);
	scratch_addr += 128;
	PIPE_CONTROL_FLUSH(ring, scratch_addr);
	scratch_addr += 128;
	PIPE_CONTROL_FLUSH(ring, scratch_addr);
	scratch_addr += 128;
	PIPE_CONTROL_FLUSH(ring, scratch_addr);
	scratch_addr += 128;
	PIPE_CONTROL_FLUSH(ring, scratch_addr);
	intel_ring_emit(ring, GFX_OP_PIPE_CONTROL | PIPE_CONTROL_QW_WRITE |
			PIPE_CONTROL_WC_FLUSH | PIPE_CONTROL_TC_FLUSH |
			PIPE_CONTROL_NOTIFY);
	intel_ring_emit(ring, pc->gtt_offset | PIPE_CONTROL_GLOBAL_GTT);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, 0);
	intel_ring_advance(ring);

	*result = seqno;
	return 0;
}

453 454 455 456 457 458 459
static int
render_ring_add_request(struct intel_ring_buffer *ring,
			u32 *result)
{
	struct drm_device *dev = ring->dev;
	u32 seqno = i915_gem_get_seqno(dev);
	int ret;
460

461 462 463
	ret = intel_ring_begin(ring, 4);
	if (ret)
		return ret;
464

465 466 467 468
	intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
	intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, MI_USER_INTERRUPT);
469
	intel_ring_advance(ring);
470

471 472
	*result = seqno;
	return 0;
473 474
}

475
static u32
476
ring_get_seqno(struct intel_ring_buffer *ring)
477
{
478 479 480
	return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
}

481 482 483 484 485 486 487
static u32
pc_render_get_seqno(struct intel_ring_buffer *ring)
{
	struct pipe_control *pc = ring->private;
	return pc->cpu_page[0];
}

488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519
static void
ironlake_enable_irq(drm_i915_private_t *dev_priv, u32 mask)
{
	dev_priv->gt_irq_mask &= ~mask;
	I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
	POSTING_READ(GTIMR);
}

static void
ironlake_disable_irq(drm_i915_private_t *dev_priv, u32 mask)
{
	dev_priv->gt_irq_mask |= mask;
	I915_WRITE(GTIMR, dev_priv->gt_irq_mask);
	POSTING_READ(GTIMR);
}

static void
i915_enable_irq(drm_i915_private_t *dev_priv, u32 mask)
{
	dev_priv->irq_mask &= ~mask;
	I915_WRITE(IMR, dev_priv->irq_mask);
	POSTING_READ(IMR);
}

static void
i915_disable_irq(drm_i915_private_t *dev_priv, u32 mask)
{
	dev_priv->irq_mask |= mask;
	I915_WRITE(IMR, dev_priv->irq_mask);
	POSTING_READ(IMR);
}

520
static bool
521
render_ring_get_irq(struct intel_ring_buffer *ring)
522
{
523
	struct drm_device *dev = ring->dev;
524
	drm_i915_private_t *dev_priv = dev->dev_private;
525

526 527 528
	if (!dev->irq_enabled)
		return false;

529
	spin_lock(&ring->irq_lock);
530
	if (ring->irq_refcount++ == 0) {
531
		if (HAS_PCH_SPLIT(dev))
532 533
			ironlake_enable_irq(dev_priv,
					    GT_PIPE_NOTIFY | GT_USER_INTERRUPT);
534 535 536
		else
			i915_enable_irq(dev_priv, I915_USER_INTERRUPT);
	}
537
	spin_unlock(&ring->irq_lock);
538 539

	return true;
540 541
}

542
static void
543
render_ring_put_irq(struct intel_ring_buffer *ring)
544
{
545
	struct drm_device *dev = ring->dev;
546
	drm_i915_private_t *dev_priv = dev->dev_private;
547

548
	spin_lock(&ring->irq_lock);
549
	if (--ring->irq_refcount == 0) {
550
		if (HAS_PCH_SPLIT(dev))
551 552 553
			ironlake_disable_irq(dev_priv,
					     GT_USER_INTERRUPT |
					     GT_PIPE_NOTIFY);
554 555 556
		else
			i915_disable_irq(dev_priv, I915_USER_INTERRUPT);
	}
557
	spin_unlock(&ring->irq_lock);
558 559
}

560
void intel_ring_setup_status_page(struct intel_ring_buffer *ring)
561
{
562 563 564 565 566 567
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
	u32 mmio = IS_GEN6(ring->dev) ?
		RING_HWS_PGA_GEN6(ring->mmio_base) :
		RING_HWS_PGA(ring->mmio_base);
	I915_WRITE(mmio, (u32)ring->status_page.gfx_addr);
	POSTING_READ(mmio);
568 569
}

570
static int
571 572 573
bsd_ring_flush(struct intel_ring_buffer *ring,
	       u32     invalidate_domains,
	       u32     flush_domains)
574
{
575 576
	int ret;

577
	if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0)
578
		return 0;
579

580 581 582 583 584 585 586 587
	ret = intel_ring_begin(ring, 2);
	if (ret)
		return ret;

	intel_ring_emit(ring, MI_FLUSH);
	intel_ring_emit(ring, MI_NOOP);
	intel_ring_advance(ring);
	return 0;
588 589
}

590
static int
591
ring_add_request(struct intel_ring_buffer *ring,
592
		 u32 *result)
593 594
{
	u32 seqno;
595 596 597 598 599
	int ret;

	ret = intel_ring_begin(ring, 4);
	if (ret)
		return ret;
600

601
	seqno = i915_gem_get_seqno(ring->dev);
602

603 604 605 606 607
	intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
	intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, MI_USER_INTERRUPT);
	intel_ring_advance(ring);
608 609

	DRM_DEBUG_DRIVER("%s %d\n", ring->name, seqno);
610 611
	*result = seqno;
	return 0;
612 613
}

614
static bool
615
ring_get_irq(struct intel_ring_buffer *ring, u32 flag)
616
{
617
	struct drm_device *dev = ring->dev;
618
	drm_i915_private_t *dev_priv = dev->dev_private;
619

620 621 622
	if (!dev->irq_enabled)
	       return false;

623
	spin_lock(&ring->irq_lock);
624
	if (ring->irq_refcount++ == 0)
625
		ironlake_enable_irq(dev_priv, flag);
626
	spin_unlock(&ring->irq_lock);
627 628

	return true;
629
}
630

631
static void
632
ring_put_irq(struct intel_ring_buffer *ring, u32 flag)
633
{
634
	struct drm_device *dev = ring->dev;
635
	drm_i915_private_t *dev_priv = dev->dev_private;
636

637
	spin_lock(&ring->irq_lock);
638
	if (--ring->irq_refcount == 0)
639
		ironlake_disable_irq(dev_priv, flag);
640
	spin_unlock(&ring->irq_lock);
641 642 643 644 645 646
}

static bool
gen6_ring_get_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag)
{
	struct drm_device *dev = ring->dev;
647
	drm_i915_private_t *dev_priv = dev->dev_private;
648 649 650 651

	if (!dev->irq_enabled)
	       return false;

652
	spin_lock(&ring->irq_lock);
653
	if (ring->irq_refcount++ == 0) {
654 655 656 657
		ring->irq_mask &= ~rflag;
		I915_WRITE_IMR(ring, ring->irq_mask);
		ironlake_enable_irq(dev_priv, gflag);
	}
658
	spin_unlock(&ring->irq_lock);
659 660 661 662 663 664 665 666

	return true;
}

static void
gen6_ring_put_irq(struct intel_ring_buffer *ring, u32 gflag, u32 rflag)
{
	struct drm_device *dev = ring->dev;
667
	drm_i915_private_t *dev_priv = dev->dev_private;
668

669
	spin_lock(&ring->irq_lock);
670
	if (--ring->irq_refcount == 0) {
671 672 673
		ring->irq_mask |= rflag;
		I915_WRITE_IMR(ring, ring->irq_mask);
		ironlake_disable_irq(dev_priv, gflag);
674
	}
675
	spin_unlock(&ring->irq_lock);
676 677
}

678
static bool
679
bsd_ring_get_irq(struct intel_ring_buffer *ring)
680
{
681
	return ring_get_irq(ring, GT_BSD_USER_INTERRUPT);
682 683 684 685
}
static void
bsd_ring_put_irq(struct intel_ring_buffer *ring)
{
686
	ring_put_irq(ring, GT_BSD_USER_INTERRUPT);
687 688 689
}

static int
690
ring_dispatch_execbuffer(struct intel_ring_buffer *ring, u32 offset, u32 length)
691
{
692
	int ret;
693

694 695 696 697
	ret = intel_ring_begin(ring, 2);
	if (ret)
		return ret;

698
	intel_ring_emit(ring,
699
			MI_BATCH_BUFFER_START | (2 << 6) |
700
			MI_BATCH_NON_SECURE_I965);
701
	intel_ring_emit(ring, offset);
702 703
	intel_ring_advance(ring);

704 705 706
	return 0;
}

707
static int
708
render_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
709
				u32 offset, u32 len)
710
{
711
	struct drm_device *dev = ring->dev;
712
	drm_i915_private_t *dev_priv = dev->dev_private;
713
	int ret;
714

715
	trace_i915_gem_request_submit(dev, dev_priv->next_seqno + 1);
716

717 718 719 720
	if (IS_I830(dev) || IS_845G(dev)) {
		ret = intel_ring_begin(ring, 4);
		if (ret)
			return ret;
721

722 723 724 725 726 727 728 729
		intel_ring_emit(ring, MI_BATCH_BUFFER);
		intel_ring_emit(ring, offset | MI_BATCH_NON_SECURE);
		intel_ring_emit(ring, offset + len - 8);
		intel_ring_emit(ring, 0);
	} else {
		ret = intel_ring_begin(ring, 2);
		if (ret)
			return ret;
730

731 732 733 734 735
		if (INTEL_INFO(dev)->gen >= 4) {
			intel_ring_emit(ring,
					MI_BATCH_BUFFER_START | (2 << 6) |
					MI_BATCH_NON_SECURE_I965);
			intel_ring_emit(ring, offset);
736
		} else {
737 738 739
			intel_ring_emit(ring,
					MI_BATCH_BUFFER_START | (2 << 6));
			intel_ring_emit(ring, offset | MI_BATCH_NON_SECURE);
740 741
		}
	}
742
	intel_ring_advance(ring);
743 744 745 746

	return 0;
}

747
static void cleanup_status_page(struct intel_ring_buffer *ring)
748
{
749
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
750
	struct drm_i915_gem_object *obj;
751

752 753
	obj = ring->status_page.obj;
	if (obj == NULL)
754 755
		return;

756
	kunmap(obj->pages[0]);
757
	i915_gem_object_unpin(obj);
758
	drm_gem_object_unreference(&obj->base);
759
	ring->status_page.obj = NULL;
760 761 762 763

	memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
}

764
static int init_status_page(struct intel_ring_buffer *ring)
765
{
766
	struct drm_device *dev = ring->dev;
767
	drm_i915_private_t *dev_priv = dev->dev_private;
768
	struct drm_i915_gem_object *obj;
769 770 771 772 773 774 775 776
	int ret;

	obj = i915_gem_alloc_object(dev, 4096);
	if (obj == NULL) {
		DRM_ERROR("Failed to allocate status page\n");
		ret = -ENOMEM;
		goto err;
	}
777
	obj->agp_type = AGP_USER_CACHED_MEMORY;
778

779
	ret = i915_gem_object_pin(obj, 4096, true);
780 781 782 783
	if (ret != 0) {
		goto err_unref;
	}

784 785
	ring->status_page.gfx_addr = obj->gtt_offset;
	ring->status_page.page_addr = kmap(obj->pages[0]);
786
	if (ring->status_page.page_addr == NULL) {
787 788 789
		memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
		goto err_unpin;
	}
790 791
	ring->status_page.obj = obj;
	memset(ring->status_page.page_addr, 0, PAGE_SIZE);
792

793
	intel_ring_setup_status_page(ring);
794 795
	DRM_DEBUG_DRIVER("%s hws offset: 0x%08x\n",
			ring->name, ring->status_page.gfx_addr);
796 797 798 799 800 801

	return 0;

err_unpin:
	i915_gem_object_unpin(obj);
err_unref:
802
	drm_gem_object_unreference(&obj->base);
803
err:
804
	return ret;
805 806
}

807
int intel_init_ring_buffer(struct drm_device *dev,
808
			   struct intel_ring_buffer *ring)
809
{
810
	struct drm_i915_gem_object *obj;
811 812
	int ret;

813
	ring->dev = dev;
814 815
	INIT_LIST_HEAD(&ring->active_list);
	INIT_LIST_HEAD(&ring->request_list);
816
	INIT_LIST_HEAD(&ring->gpu_write_list);
817 818

	spin_lock_init(&ring->irq_lock);
819
	ring->irq_mask = ~0;
820

821
	if (I915_NEED_GFX_HWS(dev)) {
822
		ret = init_status_page(ring);
823 824 825
		if (ret)
			return ret;
	}
826

827
	obj = i915_gem_alloc_object(dev, ring->size);
828 829
	if (obj == NULL) {
		DRM_ERROR("Failed to allocate ringbuffer\n");
830
		ret = -ENOMEM;
831
		goto err_hws;
832 833
	}

834
	ring->obj = obj;
835

836
	ret = i915_gem_object_pin(obj, PAGE_SIZE, true);
837 838
	if (ret)
		goto err_unref;
839

840
	ring->map.size = ring->size;
841
	ring->map.offset = dev->agp->base + obj->gtt_offset;
842 843 844 845 846 847 848
	ring->map.type = 0;
	ring->map.flags = 0;
	ring->map.mtrr = 0;

	drm_core_ioremap_wc(&ring->map, dev);
	if (ring->map.handle == NULL) {
		DRM_ERROR("Failed to map ringbuffer.\n");
849
		ret = -EINVAL;
850
		goto err_unpin;
851 852
	}

853
	ring->virtual_start = ring->map.handle;
854
	ret = ring->init(ring);
855 856
	if (ret)
		goto err_unmap;
857

858 859 860 861 862 863 864 865
	/* Workaround an erratum on the i830 which causes a hang if
	 * the TAIL pointer points to within the last 2 cachelines
	 * of the buffer.
	 */
	ring->effective_size = ring->size;
	if (IS_I830(ring->dev))
		ring->effective_size -= 128;

866
	return 0;
867 868 869 870 871 872

err_unmap:
	drm_core_ioremapfree(&ring->map, dev);
err_unpin:
	i915_gem_object_unpin(obj);
err_unref:
873 874
	drm_gem_object_unreference(&obj->base);
	ring->obj = NULL;
875
err_hws:
876
	cleanup_status_page(ring);
877
	return ret;
878 879
}

880
void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring)
881
{
882 883 884
	struct drm_i915_private *dev_priv;
	int ret;

885
	if (ring->obj == NULL)
886 887
		return;

888 889 890 891 892
	/* Disable the ring buffer. The ring must be idle at this point */
	dev_priv = ring->dev->dev_private;
	ret = intel_wait_ring_buffer(ring, ring->size - 8);
	I915_WRITE_CTL(ring, 0);

893
	drm_core_ioremapfree(&ring->map, ring->dev);
894

895 896 897
	i915_gem_object_unpin(ring->obj);
	drm_gem_object_unreference(&ring->obj->base);
	ring->obj = NULL;
898

Z
Zou Nan hai 已提交
899 900 901
	if (ring->cleanup)
		ring->cleanup(ring);

902
	cleanup_status_page(ring);
903 904
}

905
static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring)
906
{
907
	unsigned int *virt;
908
	int rem = ring->size - ring->tail;
909

910
	if (ring->space < rem) {
911
		int ret = intel_wait_ring_buffer(ring, rem);
912 913 914 915
		if (ret)
			return ret;
	}

916
	virt = (unsigned int *)(ring->virtual_start + ring->tail);
917 918
	rem /= 8;
	while (rem--) {
919
		*virt++ = MI_NOOP;
920 921
		*virt++ = MI_NOOP;
	}
922

923
	ring->tail = 0;
924
	ring->space = ring->head - 8;
925 926 927 928

	return 0;
}

929
int intel_wait_ring_buffer(struct intel_ring_buffer *ring, int n)
930
{
931
	struct drm_device *dev = ring->dev;
932
	struct drm_i915_private *dev_priv = dev->dev_private;
933
	unsigned long end;
934 935
	u32 head;

936
	trace_i915_ring_wait_begin (dev);
937 938
	end = jiffies + 3 * HZ;
	do {
939 940 941 942 943 944 945 946
		/* If the reported head position has wrapped or hasn't advanced,
		 * fallback to the slow and accurate path.
		 */
		head = intel_read_status_page(ring, 4);
		if (head < ring->actual_head)
			head = I915_READ_HEAD(ring);
		ring->actual_head = head;
		ring->head = head & HEAD_ADDR;
947 948
		ring->space = ring->head - (ring->tail + 8);
		if (ring->space < 0)
949
			ring->space += ring->size;
950
		if (ring->space >= n) {
951
			trace_i915_ring_wait_end(dev);
952 953 954 955 956 957 958 959
			return 0;
		}

		if (dev->primary->master) {
			struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
			if (master_priv->sarea_priv)
				master_priv->sarea_priv->perf_boxes |= I915_BOX_WAIT;
		}
960

961
		msleep(1);
962 963
		if (atomic_read(&dev_priv->mm.wedged))
			return -EAGAIN;
964 965 966 967
	} while (!time_after(jiffies, end));
	trace_i915_ring_wait_end (dev);
	return -EBUSY;
}
968

969 970
int intel_ring_begin(struct intel_ring_buffer *ring,
		     int num_dwords)
971
{
972
	int n = 4*num_dwords;
973
	int ret;
974

975
	if (unlikely(ring->tail + n > ring->effective_size)) {
976 977 978 979
		ret = intel_wrap_ring_buffer(ring);
		if (unlikely(ret))
			return ret;
	}
980

981 982 983 984 985
	if (unlikely(ring->space < n)) {
		ret = intel_wait_ring_buffer(ring, n);
		if (unlikely(ret))
			return ret;
	}
986 987

	ring->space -= n;
988
	return 0;
989
}
990

991
void intel_ring_advance(struct intel_ring_buffer *ring)
992
{
993
	ring->tail &= ring->size - 1;
994
	ring->write_tail(ring, ring->tail);
995
}
996

997
static const struct intel_ring_buffer render_ring = {
998
	.name			= "render ring",
999
	.id			= RING_RENDER,
1000
	.mmio_base		= RENDER_RING_BASE,
1001 1002
	.size			= 32 * PAGE_SIZE,
	.init			= init_render_ring,
1003
	.write_tail		= ring_write_tail,
1004 1005
	.flush			= render_ring_flush,
	.add_request		= render_ring_add_request,
1006 1007 1008
	.get_seqno		= ring_get_seqno,
	.irq_get		= render_ring_get_irq,
	.irq_put		= render_ring_put_irq,
1009
	.dispatch_execbuffer	= render_ring_dispatch_execbuffer,
1010
       .cleanup			= render_ring_cleanup,
1011
};
1012 1013 1014

/* ring buffer for bit-stream decoder */

1015
static const struct intel_ring_buffer bsd_ring = {
1016
	.name                   = "bsd ring",
1017
	.id			= RING_BSD,
1018
	.mmio_base		= BSD_RING_BASE,
1019
	.size			= 32 * PAGE_SIZE,
1020
	.init			= init_ring_common,
1021
	.write_tail		= ring_write_tail,
1022
	.flush			= bsd_ring_flush,
1023
	.add_request		= ring_add_request,
1024 1025 1026
	.get_seqno		= ring_get_seqno,
	.irq_get		= bsd_ring_get_irq,
	.irq_put		= bsd_ring_put_irq,
1027
	.dispatch_execbuffer	= ring_dispatch_execbuffer,
1028
};
1029

1030

1031
static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring,
1032
				     u32 value)
1033
{
1034
       drm_i915_private_t *dev_priv = ring->dev->dev_private;
1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046

       /* Every tail move must follow the sequence below */
       I915_WRITE(GEN6_BSD_SLEEP_PSMI_CONTROL,
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_MODIFY_MASK |
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_DISABLE);
       I915_WRITE(GEN6_BSD_RNCID, 0x0);

       if (wait_for((I915_READ(GEN6_BSD_SLEEP_PSMI_CONTROL) &
                               GEN6_BSD_SLEEP_PSMI_CONTROL_IDLE_INDICATOR) == 0,
                       50))
               DRM_ERROR("timed out waiting for IDLE Indicator\n");

1047
       I915_WRITE_TAIL(ring, value);
1048 1049 1050 1051 1052
       I915_WRITE(GEN6_BSD_SLEEP_PSMI_CONTROL,
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_MODIFY_MASK |
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_ENABLE);
}

1053 1054 1055
static int gen6_ring_flush(struct intel_ring_buffer *ring,
			   u32 invalidate_domains,
			   u32 flush_domains)
1056
{
1057 1058
	int ret;

1059
	if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0)
1060
		return 0;
1061

1062 1063 1064 1065 1066 1067 1068 1069 1070 1071
	ret = intel_ring_begin(ring, 4);
	if (ret)
		return ret;

	intel_ring_emit(ring, MI_FLUSH_DW);
	intel_ring_emit(ring, 0);
	intel_ring_emit(ring, 0);
	intel_ring_emit(ring, 0);
	intel_ring_advance(ring);
	return 0;
1072 1073 1074
}

static int
1075
gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
1076
			      u32 offset, u32 len)
1077
{
1078
       int ret;
1079

1080 1081 1082 1083
       ret = intel_ring_begin(ring, 2);
       if (ret)
	       return ret;

1084
       intel_ring_emit(ring, MI_BATCH_BUFFER_START | MI_BATCH_NON_SECURE_I965);
1085
       /* bit0-7 is the length on GEN6+ */
1086
       intel_ring_emit(ring, offset);
1087
       intel_ring_advance(ring);
1088

1089 1090 1091
       return 0;
}

1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107
static bool
gen6_render_ring_get_irq(struct intel_ring_buffer *ring)
{
	return gen6_ring_get_irq(ring,
				 GT_USER_INTERRUPT,
				 GEN6_RENDER_USER_INTERRUPT);
}

static void
gen6_render_ring_put_irq(struct intel_ring_buffer *ring)
{
	return gen6_ring_put_irq(ring,
				 GT_USER_INTERRUPT,
				 GEN6_RENDER_USER_INTERRUPT);
}

1108
static bool
1109 1110
gen6_bsd_ring_get_irq(struct intel_ring_buffer *ring)
{
1111 1112 1113
	return gen6_ring_get_irq(ring,
				 GT_GEN6_BSD_USER_INTERRUPT,
				 GEN6_BSD_USER_INTERRUPT);
1114 1115 1116 1117 1118
}

static void
gen6_bsd_ring_put_irq(struct intel_ring_buffer *ring)
{
1119 1120 1121
	return gen6_ring_put_irq(ring,
				 GT_GEN6_BSD_USER_INTERRUPT,
				 GEN6_BSD_USER_INTERRUPT);
1122 1123
}

1124
/* ring buffer for Video Codec for Gen6+ */
1125
static const struct intel_ring_buffer gen6_bsd_ring = {
1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137
	.name			= "gen6 bsd ring",
	.id			= RING_BSD,
	.mmio_base		= GEN6_BSD_RING_BASE,
	.size			= 32 * PAGE_SIZE,
	.init			= init_ring_common,
	.write_tail		= gen6_bsd_ring_write_tail,
	.flush			= gen6_ring_flush,
	.add_request		= gen6_add_request,
	.get_seqno		= ring_get_seqno,
	.irq_get		= gen6_bsd_ring_get_irq,
	.irq_put		= gen6_bsd_ring_put_irq,
	.dispatch_execbuffer	= gen6_ring_dispatch_execbuffer,
1138 1139 1140 1141
};

/* Blitter support (SandyBridge+) */

1142
static bool
1143
blt_ring_get_irq(struct intel_ring_buffer *ring)
1144
{
1145 1146 1147
	return gen6_ring_get_irq(ring,
				 GT_BLT_USER_INTERRUPT,
				 GEN6_BLITTER_USER_INTERRUPT);
1148
}
1149

1150
static void
1151
blt_ring_put_irq(struct intel_ring_buffer *ring)
1152
{
1153 1154 1155
	gen6_ring_put_irq(ring,
			  GT_BLT_USER_INTERRUPT,
			  GEN6_BLITTER_USER_INTERRUPT);
1156 1157
}

Z
Zou Nan hai 已提交
1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176

/* Workaround for some stepping of SNB,
 * each time when BLT engine ring tail moved,
 * the first command in the ring to be parsed
 * should be MI_BATCH_BUFFER_START
 */
#define NEED_BLT_WORKAROUND(dev) \
	(IS_GEN6(dev) && (dev->pdev->revision < 8))

static inline struct drm_i915_gem_object *
to_blt_workaround(struct intel_ring_buffer *ring)
{
	return ring->private;
}

static int blt_ring_init(struct intel_ring_buffer *ring)
{
	if (NEED_BLT_WORKAROUND(ring->dev)) {
		struct drm_i915_gem_object *obj;
1177
		u32 *ptr;
Z
Zou Nan hai 已提交
1178 1179
		int ret;

1180
		obj = i915_gem_alloc_object(ring->dev, 4096);
Z
Zou Nan hai 已提交
1181 1182 1183
		if (obj == NULL)
			return -ENOMEM;

1184
		ret = i915_gem_object_pin(obj, 4096, true);
Z
Zou Nan hai 已提交
1185 1186 1187 1188 1189 1190
		if (ret) {
			drm_gem_object_unreference(&obj->base);
			return ret;
		}

		ptr = kmap(obj->pages[0]);
1191 1192
		*ptr++ = MI_BATCH_BUFFER_END;
		*ptr++ = MI_NOOP;
Z
Zou Nan hai 已提交
1193 1194
		kunmap(obj->pages[0]);

1195
		ret = i915_gem_object_set_to_gtt_domain(obj, false);
Z
Zou Nan hai 已提交
1196
		if (ret) {
1197
			i915_gem_object_unpin(obj);
Z
Zou Nan hai 已提交
1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223
			drm_gem_object_unreference(&obj->base);
			return ret;
		}

		ring->private = obj;
	}

	return init_ring_common(ring);
}

static int blt_ring_begin(struct intel_ring_buffer *ring,
			  int num_dwords)
{
	if (ring->private) {
		int ret = intel_ring_begin(ring, num_dwords+2);
		if (ret)
			return ret;

		intel_ring_emit(ring, MI_BATCH_BUFFER_START);
		intel_ring_emit(ring, to_blt_workaround(ring)->gtt_offset);

		return 0;
	} else
		return intel_ring_begin(ring, 4);
}

1224
static int blt_ring_flush(struct intel_ring_buffer *ring,
Z
Zou Nan hai 已提交
1225 1226 1227
			   u32 invalidate_domains,
			   u32 flush_domains)
{
1228 1229
	int ret;

1230
	if ((flush_domains & I915_GEM_DOMAIN_RENDER) == 0)
1231
		return 0;
1232

1233 1234 1235 1236 1237 1238 1239 1240 1241 1242
	ret = blt_ring_begin(ring, 4);
	if (ret)
		return ret;

	intel_ring_emit(ring, MI_FLUSH_DW);
	intel_ring_emit(ring, 0);
	intel_ring_emit(ring, 0);
	intel_ring_emit(ring, 0);
	intel_ring_advance(ring);
	return 0;
Z
Zou Nan hai 已提交
1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254
}

static void blt_ring_cleanup(struct intel_ring_buffer *ring)
{
	if (!ring->private)
		return;

	i915_gem_object_unpin(ring->private);
	drm_gem_object_unreference(ring->private);
	ring->private = NULL;
}

1255 1256 1257 1258 1259
static const struct intel_ring_buffer gen6_blt_ring = {
       .name			= "blt ring",
       .id			= RING_BLT,
       .mmio_base		= BLT_RING_BASE,
       .size			= 32 * PAGE_SIZE,
Z
Zou Nan hai 已提交
1260
       .init			= blt_ring_init,
1261
       .write_tail		= ring_write_tail,
Z
Zou Nan hai 已提交
1262
       .flush			= blt_ring_flush,
1263 1264 1265 1266
       .add_request		= gen6_add_request,
       .get_seqno		= ring_get_seqno,
       .irq_get			= blt_ring_get_irq,
       .irq_put			= blt_ring_put_irq,
1267
       .dispatch_execbuffer	= gen6_ring_dispatch_execbuffer,
Z
Zou Nan hai 已提交
1268
       .cleanup			= blt_ring_cleanup,
1269 1270
};

1271 1272 1273
int intel_init_render_ring_buffer(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
1274
	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];
1275

1276 1277 1278
	*ring = render_ring;
	if (INTEL_INFO(dev)->gen >= 6) {
		ring->add_request = gen6_add_request;
1279 1280
		ring->irq_get = gen6_render_ring_get_irq;
		ring->irq_put = gen6_render_ring_put_irq;
1281 1282 1283
	} else if (IS_GEN5(dev)) {
		ring->add_request = pc_render_add_request;
		ring->get_seqno = pc_render_get_seqno;
1284
	}
1285 1286

	if (!I915_NEED_GFX_HWS(dev)) {
1287 1288
		ring->status_page.page_addr = dev_priv->status_page_dmah->vaddr;
		memset(ring->status_page.page_addr, 0, PAGE_SIZE);
1289 1290
	}

1291
	return intel_init_ring_buffer(dev, ring);
1292 1293
}

1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335
int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
	struct intel_ring_buffer *ring = &dev_priv->ring[RCS];

	*ring = render_ring;
	if (INTEL_INFO(dev)->gen >= 6) {
		ring->add_request = gen6_add_request;
		ring->irq_get = gen6_render_ring_get_irq;
		ring->irq_put = gen6_render_ring_put_irq;
	} else if (IS_GEN5(dev)) {
		ring->add_request = pc_render_add_request;
		ring->get_seqno = pc_render_get_seqno;
	}

	ring->dev = dev;
	INIT_LIST_HEAD(&ring->active_list);
	INIT_LIST_HEAD(&ring->request_list);
	INIT_LIST_HEAD(&ring->gpu_write_list);

	ring->size = size;
	ring->effective_size = ring->size;
	if (IS_I830(ring->dev))
		ring->effective_size -= 128;

	ring->map.offset = start;
	ring->map.size = size;
	ring->map.type = 0;
	ring->map.flags = 0;
	ring->map.mtrr = 0;

	drm_core_ioremap_wc(&ring->map, dev);
	if (ring->map.handle == NULL) {
		DRM_ERROR("can not ioremap virtual address for"
			  " ring buffer\n");
		return -ENOMEM;
	}

	ring->virtual_start = (void __force __iomem *)ring->map.handle;
	return 0;
}

1336 1337 1338
int intel_init_bsd_ring_buffer(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
1339
	struct intel_ring_buffer *ring = &dev_priv->ring[VCS];
1340

1341
	if (IS_GEN6(dev))
1342
		*ring = gen6_bsd_ring;
1343
	else
1344
		*ring = bsd_ring;
1345

1346
	return intel_init_ring_buffer(dev, ring);
1347
}
1348 1349 1350 1351

int intel_init_blt_ring_buffer(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
1352
	struct intel_ring_buffer *ring = &dev_priv->ring[BCS];
1353

1354
	*ring = gen6_blt_ring;
1355

1356
	return intel_init_ring_buffer(dev, ring);
1357
}