intel_ringbuffer.c 23.9 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32
/*
 * Copyright © 2008-2010 Intel Corporation
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice (including the next
 * paragraph) shall be included in all copies or substantial portions of the
 * Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
 * IN THE SOFTWARE.
 *
 * Authors:
 *    Eric Anholt <eric@anholt.net>
 *    Zou Nan hai <nanhai.zou@intel.com>
 *    Xiang Hai hao<haihao.xiang@intel.com>
 *
 */

#include "drmP.h"
#include "drm.h"
#include "i915_drv.h"
33
#include "i915_drm.h"
34
#include "i915_trace.h"
35
#include "intel_drv.h"
36

37 38 39 40 41 42 43 44 45 46 47 48 49 50
static u32 i915_gem_get_seqno(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;
	u32 seqno;

	seqno = dev_priv->next_seqno;

	/* reserve 0 for non-seqno */
	if (++dev_priv->next_seqno == 0)
		dev_priv->next_seqno = 1;

	return seqno;
}

51
static void
52
render_ring_flush(struct intel_ring_buffer *ring,
53 54
		  u32	invalidate_domains,
		  u32	flush_domains)
55
{
56
	struct drm_device *dev = ring->dev;
57 58 59
	drm_i915_private_t *dev_priv = dev->dev_private;
	u32 cmd;

60 61 62 63
#if WATCH_EXEC
	DRM_INFO("%s: invalidate %08x flush %08x\n", __func__,
		  invalidate_domains, flush_domains);
#endif
64 65

	trace_i915_gem_request_flush(dev, dev_priv->next_seqno,
66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100
				     invalidate_domains, flush_domains);

	if ((invalidate_domains | flush_domains) & I915_GEM_GPU_DOMAINS) {
		/*
		 * read/write caches:
		 *
		 * I915_GEM_DOMAIN_RENDER is always invalidated, but is
		 * only flushed if MI_NO_WRITE_FLUSH is unset.  On 965, it is
		 * also flushed at 2d versus 3d pipeline switches.
		 *
		 * read-only caches:
		 *
		 * I915_GEM_DOMAIN_SAMPLER is flushed on pre-965 if
		 * MI_READ_FLUSH is set, and is always flushed on 965.
		 *
		 * I915_GEM_DOMAIN_COMMAND may not exist?
		 *
		 * I915_GEM_DOMAIN_INSTRUCTION, which exists on 965, is
		 * invalidated when MI_EXE_FLUSH is set.
		 *
		 * I915_GEM_DOMAIN_VERTEX, which exists on 965, is
		 * invalidated with every MI_FLUSH.
		 *
		 * TLBs:
		 *
		 * On 965, TLBs associated with I915_GEM_DOMAIN_COMMAND
		 * and I915_GEM_DOMAIN_CPU in are invalidated at PTE write and
		 * I915_GEM_DOMAIN_RENDER and I915_GEM_DOMAIN_SAMPLER
		 * are flushed at any MI_FLUSH.
		 */

		cmd = MI_FLUSH | MI_NO_WRITE_FLUSH;
		if ((invalidate_domains|flush_domains) &
		    I915_GEM_DOMAIN_RENDER)
			cmd &= ~MI_NO_WRITE_FLUSH;
101
		if (INTEL_INFO(dev)->gen < 4) {
102 103 104 105 106 107 108 109 110 111 112 113 114
			/*
			 * On the 965, the sampler cache always gets flushed
			 * and this bit is reserved.
			 */
			if (invalidate_domains & I915_GEM_DOMAIN_SAMPLER)
				cmd |= MI_READ_FLUSH;
		}
		if (invalidate_domains & I915_GEM_DOMAIN_INSTRUCTION)
			cmd |= MI_EXE_FLUSH;

#if WATCH_EXEC
		DRM_INFO("%s: queue flush %08x to ring\n", __func__, cmd);
#endif
115 116 117 118 119
		if (intel_ring_begin(ring, 2) == 0) {
			intel_ring_emit(ring, cmd);
			intel_ring_emit(ring, MI_NOOP);
			intel_ring_advance(ring);
		}
120
	}
121 122
}

123
static void ring_write_tail(struct intel_ring_buffer *ring,
124
			    u32 value)
125
{
126
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
127
	I915_WRITE_TAIL(ring, value);
128 129
}

130
u32 intel_ring_get_active_head(struct intel_ring_buffer *ring)
131
{
132 133
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
	u32 acthd_reg = INTEL_INFO(ring->dev)->gen >= 4 ?
D
Daniel Vetter 已提交
134
			RING_ACTHD(ring->mmio_base) : ACTHD;
135 136 137 138

	return I915_READ(acthd_reg);
}

139
static int init_ring_common(struct intel_ring_buffer *ring)
140
{
141 142
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
	struct drm_i915_gem_object *obj_priv = to_intel_bo(ring->gem_object);
143 144 145
	u32 head;

	/* Stop the ring if it's running. */
146
	I915_WRITE_CTL(ring, 0);
147
	I915_WRITE_HEAD(ring, 0);
148
	ring->write_tail(ring, 0);
149 150

	/* Initialize the ring. */
151
	I915_WRITE_START(ring, obj_priv->gtt_offset);
152
	head = I915_READ_HEAD(ring) & HEAD_ADDR;
153 154 155 156 157 158

	/* G45 ring initialization fails to reset head to zero */
	if (head != 0) {
		DRM_ERROR("%s head not reset to zero "
				"ctl %08x head %08x tail %08x start %08x\n",
				ring->name,
159
				I915_READ_CTL(ring),
160
				I915_READ_HEAD(ring),
161
				I915_READ_TAIL(ring),
162
				I915_READ_START(ring));
163

164
		I915_WRITE_HEAD(ring, 0);
165 166 167 168

		DRM_ERROR("%s head forced to zero "
				"ctl %08x head %08x tail %08x start %08x\n",
				ring->name,
169
				I915_READ_CTL(ring),
170
				I915_READ_HEAD(ring),
171
				I915_READ_TAIL(ring),
172
				I915_READ_START(ring));
173 174
	}

175
	I915_WRITE_CTL(ring,
176 177 178 179
			((ring->gem_object->size - PAGE_SIZE) & RING_NR_PAGES)
			| RING_NO_REPORT | RING_VALID);

	/* If the head is still not zero, the ring is dead */
180 181 182
	if ((I915_READ_CTL(ring) & RING_VALID) == 0 ||
	    I915_READ_START(ring) != obj_priv->gtt_offset ||
	    (I915_READ_HEAD(ring) & HEAD_ADDR) != 0) {
183 184 185
		DRM_ERROR("%s initialization failed "
				"ctl %08x head %08x tail %08x start %08x\n",
				ring->name,
186
				I915_READ_CTL(ring),
187
				I915_READ_HEAD(ring),
188
				I915_READ_TAIL(ring),
189
				I915_READ_START(ring));
190 191 192
		return -EIO;
	}

193 194
	if (!drm_core_check_feature(ring->dev, DRIVER_MODESET))
		i915_kernel_lost_context(ring->dev);
195
	else {
196
		ring->head = I915_READ_HEAD(ring) & HEAD_ADDR;
197
		ring->tail = I915_READ_TAIL(ring) & TAIL_ADDR;
198 199 200 201 202 203 204
		ring->space = ring->head - (ring->tail + 8);
		if (ring->space < 0)
			ring->space += ring->size;
	}
	return 0;
}

205
static int init_render_ring(struct intel_ring_buffer *ring)
206
{
207 208
	struct drm_device *dev = ring->dev;
	int ret = init_ring_common(ring);
209

210
	if (INTEL_INFO(dev)->gen > 3) {
211 212
		drm_i915_private_t *dev_priv = dev->dev_private;
		int mode = VS_TIMER_DISPATCH << 16 | VS_TIMER_DISPATCH;
213 214 215
		if (IS_GEN6(dev))
			mode |= MI_FLUSH_ENABLE << 16 | MI_FLUSH_ENABLE;
		I915_WRITE(MI_MODE, mode);
216
	}
217

218 219 220
	return ret;
}

221
#define PIPE_CONTROL_FLUSH(ring__, addr__)					\
222
do {									\
223
	intel_ring_emit(ring__, GFX_OP_PIPE_CONTROL | PIPE_CONTROL_QW_WRITE |		\
224
		 PIPE_CONTROL_DEPTH_STALL | 2);				\
225 226 227
	intel_ring_emit(ring__, (addr__) | PIPE_CONTROL_GLOBAL_GTT);			\
	intel_ring_emit(ring__, 0);							\
	intel_ring_emit(ring__, 0);							\
228
} while (0)
229 230 231 232 233 234 235 236 237

/**
 * Creates a new sequence number, emitting a write of it to the status page
 * plus an interrupt, which will trigger i915_user_interrupt_handler.
 *
 * Must be called with struct_lock held.
 *
 * Returned sequence numbers are nonzero on success.
 */
238
static int
239
render_ring_add_request(struct intel_ring_buffer *ring,
240
			u32 *result)
241
{
242
	struct drm_device *dev = ring->dev;
243
	drm_i915_private_t *dev_priv = dev->dev_private;
244 245
	u32 seqno = i915_gem_get_seqno(dev);
	int ret;
246 247

	if (IS_GEN6(dev)) {
248 249 250 251 252 253 254 255 256 257 258 259
		ret = intel_ring_begin(ring, 6);
		if (ret)
		    return ret;

		intel_ring_emit(ring, GFX_OP_PIPE_CONTROL | 3);
		intel_ring_emit(ring, PIPE_CONTROL_QW_WRITE |
				PIPE_CONTROL_WC_FLUSH | PIPE_CONTROL_IS_FLUSH |
				PIPE_CONTROL_NOTIFY);
		intel_ring_emit(ring, dev_priv->seqno_gfx_addr | PIPE_CONTROL_GLOBAL_GTT);
		intel_ring_emit(ring, seqno);
		intel_ring_emit(ring, 0);
		intel_ring_emit(ring, 0);
260
	} else if (HAS_PIPE_CONTROL(dev)) {
261 262 263 264 265 266 267
		u32 scratch_addr = dev_priv->seqno_gfx_addr + 128;

		/*
		 * Workaround qword write incoherence by flushing the
		 * PIPE_NOTIFY buffers out to memory before requesting
		 * an interrupt.
		 */
268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293
		ret = intel_ring_begin(ring, 32);
		if (ret)
			return ret;

		intel_ring_emit(ring, GFX_OP_PIPE_CONTROL | PIPE_CONTROL_QW_WRITE |
				PIPE_CONTROL_WC_FLUSH | PIPE_CONTROL_TC_FLUSH);
		intel_ring_emit(ring, dev_priv->seqno_gfx_addr | PIPE_CONTROL_GLOBAL_GTT);
		intel_ring_emit(ring, seqno);
		intel_ring_emit(ring, 0);
		PIPE_CONTROL_FLUSH(ring, scratch_addr);
		scratch_addr += 128; /* write to separate cachelines */
		PIPE_CONTROL_FLUSH(ring, scratch_addr);
		scratch_addr += 128;
		PIPE_CONTROL_FLUSH(ring, scratch_addr);
		scratch_addr += 128;
		PIPE_CONTROL_FLUSH(ring, scratch_addr);
		scratch_addr += 128;
		PIPE_CONTROL_FLUSH(ring, scratch_addr);
		scratch_addr += 128;
		PIPE_CONTROL_FLUSH(ring, scratch_addr);
		intel_ring_emit(ring, GFX_OP_PIPE_CONTROL | PIPE_CONTROL_QW_WRITE |
				PIPE_CONTROL_WC_FLUSH | PIPE_CONTROL_TC_FLUSH |
				PIPE_CONTROL_NOTIFY);
		intel_ring_emit(ring, dev_priv->seqno_gfx_addr | PIPE_CONTROL_GLOBAL_GTT);
		intel_ring_emit(ring, seqno);
		intel_ring_emit(ring, 0);
294
	} else {
295 296 297
		ret = intel_ring_begin(ring, 4);
		if (ret)
		    return ret;
298

299 300 301 302 303
		intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
		intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
		intel_ring_emit(ring, seqno);

		intel_ring_emit(ring, MI_USER_INTERRUPT);
304
	}
305 306 307 308

	intel_ring_advance(ring);
	*result = seqno;
	return 0;
309 310
}

311
static u32
312
render_ring_get_seqno(struct intel_ring_buffer *ring)
313
{
314
	struct drm_device *dev = ring->dev;
315 316 317 318 319 320 321 322
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	if (HAS_PIPE_CONTROL(dev))
		return ((volatile u32 *)(dev_priv->seqno_page))[0];
	else
		return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
}

static void
323
render_ring_get_user_irq(struct intel_ring_buffer *ring)
324
{
325
	struct drm_device *dev = ring->dev;
326 327 328 329
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	unsigned long irqflags;

	spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
330
	if (dev->irq_enabled && (++ring->user_irq_refcount == 1)) {
331 332 333 334 335 336 337 338
		if (HAS_PCH_SPLIT(dev))
			ironlake_enable_graphics_irq(dev_priv, GT_PIPE_NOTIFY);
		else
			i915_enable_irq(dev_priv, I915_USER_INTERRUPT);
	}
	spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);
}

339
static void
340
render_ring_put_user_irq(struct intel_ring_buffer *ring)
341
{
342
	struct drm_device *dev = ring->dev;
343 344 345 346
	drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
	unsigned long irqflags;

	spin_lock_irqsave(&dev_priv->user_irq_lock, irqflags);
347 348
	BUG_ON(dev->irq_enabled && ring->user_irq_refcount <= 0);
	if (dev->irq_enabled && (--ring->user_irq_refcount == 0)) {
349 350 351 352 353 354 355 356
		if (HAS_PCH_SPLIT(dev))
			ironlake_disable_graphics_irq(dev_priv, GT_PIPE_NOTIFY);
		else
			i915_disable_irq(dev_priv, I915_USER_INTERRUPT);
	}
	spin_unlock_irqrestore(&dev_priv->user_irq_lock, irqflags);
}

357
void intel_ring_setup_status_page(struct intel_ring_buffer *ring)
358
{
359 360 361 362 363 364
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
	u32 mmio = IS_GEN6(ring->dev) ?
		RING_HWS_PGA_GEN6(ring->mmio_base) :
		RING_HWS_PGA(ring->mmio_base);
	I915_WRITE(mmio, (u32)ring->status_page.gfx_addr);
	POSTING_READ(mmio);
365 366
}

367
static void
368 369 370
bsd_ring_flush(struct intel_ring_buffer *ring,
	       u32     invalidate_domains,
	       u32     flush_domains)
371
{
372 373 374 375 376
	if (intel_ring_begin(ring, 2) == 0) {
		intel_ring_emit(ring, MI_FLUSH);
		intel_ring_emit(ring, MI_NOOP);
		intel_ring_advance(ring);
	}
377 378
}

379
static int
380
ring_add_request(struct intel_ring_buffer *ring,
381
		 u32 *result)
382 383
{
	u32 seqno;
384 385 386 387 388
	int ret;

	ret = intel_ring_begin(ring, 4);
	if (ret)
		return ret;
389

390
	seqno = i915_gem_get_seqno(ring->dev);
391

392 393 394 395 396
	intel_ring_emit(ring, MI_STORE_DWORD_INDEX);
	intel_ring_emit(ring, I915_GEM_HWS_INDEX << MI_STORE_DWORD_INDEX_SHIFT);
	intel_ring_emit(ring, seqno);
	intel_ring_emit(ring, MI_USER_INTERRUPT);
	intel_ring_advance(ring);
397 398

	DRM_DEBUG_DRIVER("%s %d\n", ring->name, seqno);
399 400
	*result = seqno;
	return 0;
401 402 403
}

static void
404
bsd_ring_get_user_irq(struct intel_ring_buffer *ring)
405 406 407 408
{
	/* do nothing */
}
static void
409
bsd_ring_put_user_irq(struct intel_ring_buffer *ring)
410 411 412 413 414
{
	/* do nothing */
}

static u32
415
ring_status_page_get_seqno(struct intel_ring_buffer *ring)
416 417 418 419 420
{
	return intel_read_status_page(ring, I915_GEM_HWS_INDEX);
}

static int
421 422 423 424
ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
			 struct drm_i915_gem_execbuffer2 *exec,
			 struct drm_clip_rect *cliprects,
			 uint64_t exec_offset)
425 426
{
	uint32_t exec_start;
427
	int ret;
428

429
	exec_start = (uint32_t) exec_offset + exec->batch_start_offset;
430

431 432 433 434
	ret = intel_ring_begin(ring, 2);
	if (ret)
		return ret;

435 436 437 438 439 440 441
	intel_ring_emit(ring,
			MI_BATCH_BUFFER_START |
			(2 << 6) |
			MI_BATCH_NON_SECURE_I965);
	intel_ring_emit(ring, exec_start);
	intel_ring_advance(ring);

442 443 444
	return 0;
}

445
static int
446 447 448 449
render_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
				struct drm_i915_gem_execbuffer2 *exec,
				struct drm_clip_rect *cliprects,
				uint64_t exec_offset)
450
{
451
	struct drm_device *dev = ring->dev;
452 453 454
	drm_i915_private_t *dev_priv = dev->dev_private;
	int nbox = exec->num_cliprects;
	uint32_t exec_start, exec_len;
455
	int i, count, ret;
456

457 458 459
	exec_start = (uint32_t) exec_offset + exec->batch_start_offset;
	exec_len = (uint32_t) exec->batch_len;

460
	trace_i915_gem_request_submit(dev, dev_priv->next_seqno + 1);
461 462 463 464

	count = nbox ? nbox : 1;
	for (i = 0; i < count; i++) {
		if (i < nbox) {
465 466
			ret = i915_emit_box(dev, cliprects, i,
					    exec->DR1, exec->DR4);
467 468 469 470 471
			if (ret)
				return ret;
		}

		if (IS_I830(dev) || IS_845G(dev)) {
472 473 474 475
			ret = intel_ring_begin(ring, 4);
			if (ret)
				return ret;

476 477 478 479
			intel_ring_emit(ring, MI_BATCH_BUFFER);
			intel_ring_emit(ring, exec_start | MI_BATCH_NON_SECURE);
			intel_ring_emit(ring, exec_start + exec_len - 4);
			intel_ring_emit(ring, 0);
480
		} else {
481 482 483 484
			ret = intel_ring_begin(ring, 2);
			if (ret)
				return ret;

485
			if (INTEL_INFO(dev)->gen >= 4) {
486
				intel_ring_emit(ring,
487 488
						MI_BATCH_BUFFER_START | (2 << 6)
						| MI_BATCH_NON_SECURE_I965);
489
				intel_ring_emit(ring, exec_start);
490
			} else {
491
				intel_ring_emit(ring, MI_BATCH_BUFFER_START
492
						| (2 << 6));
493
				intel_ring_emit(ring, exec_start |
494
						MI_BATCH_NON_SECURE);
495 496
			}
		}
497
		intel_ring_advance(ring);
498 499
	}

500
	if (IS_G4X(dev) || IS_GEN5(dev)) {
501 502 503 504 505 506 507
		if (intel_ring_begin(ring, 2) == 0) {
			intel_ring_emit(ring, MI_FLUSH |
					MI_NO_WRITE_FLUSH |
					MI_INVALIDATE_ISP );
			intel_ring_emit(ring, MI_NOOP);
			intel_ring_advance(ring);
		}
508
	}
509
	/* XXX breadcrumb */
510

511 512 513
	return 0;
}

514
static void cleanup_status_page(struct intel_ring_buffer *ring)
515
{
516
	drm_i915_private_t *dev_priv = ring->dev->dev_private;
517 518 519
	struct drm_gem_object *obj;
	struct drm_i915_gem_object *obj_priv;

520 521
	obj = ring->status_page.obj;
	if (obj == NULL)
522 523 524 525 526 527
		return;
	obj_priv = to_intel_bo(obj);

	kunmap(obj_priv->pages[0]);
	i915_gem_object_unpin(obj);
	drm_gem_object_unreference(obj);
528
	ring->status_page.obj = NULL;
529 530 531 532

	memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
}

533
static int init_status_page(struct intel_ring_buffer *ring)
534
{
535
	struct drm_device *dev = ring->dev;
536 537 538 539 540 541 542 543 544 545 546 547 548 549
	drm_i915_private_t *dev_priv = dev->dev_private;
	struct drm_gem_object *obj;
	struct drm_i915_gem_object *obj_priv;
	int ret;

	obj = i915_gem_alloc_object(dev, 4096);
	if (obj == NULL) {
		DRM_ERROR("Failed to allocate status page\n");
		ret = -ENOMEM;
		goto err;
	}
	obj_priv = to_intel_bo(obj);
	obj_priv->agp_type = AGP_USER_CACHED_MEMORY;

550
	ret = i915_gem_object_pin(obj, 4096, true, false);
551 552 553 554
	if (ret != 0) {
		goto err_unref;
	}

555 556 557
	ring->status_page.gfx_addr = obj_priv->gtt_offset;
	ring->status_page.page_addr = kmap(obj_priv->pages[0]);
	if (ring->status_page.page_addr == NULL) {
558 559 560
		memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
		goto err_unpin;
	}
561 562
	ring->status_page.obj = obj;
	memset(ring->status_page.page_addr, 0, PAGE_SIZE);
563

564
	intel_ring_setup_status_page(ring);
565 566
	DRM_DEBUG_DRIVER("%s hws offset: 0x%08x\n",
			ring->name, ring->status_page.gfx_addr);
567 568 569 570 571 572 573 574

	return 0;

err_unpin:
	i915_gem_object_unpin(obj);
err_unref:
	drm_gem_object_unreference(obj);
err:
575
	return ret;
576 577
}

578
int intel_init_ring_buffer(struct drm_device *dev,
579
			   struct intel_ring_buffer *ring)
580
{
581 582
	struct drm_i915_gem_object *obj_priv;
	struct drm_gem_object *obj;
583 584
	int ret;

585
	ring->dev = dev;
586 587
	INIT_LIST_HEAD(&ring->active_list);
	INIT_LIST_HEAD(&ring->request_list);
588
	INIT_LIST_HEAD(&ring->gpu_write_list);
589

590
	if (I915_NEED_GFX_HWS(dev)) {
591
		ret = init_status_page(ring);
592 593 594
		if (ret)
			return ret;
	}
595

596
	obj = i915_gem_alloc_object(dev, ring->size);
597 598
	if (obj == NULL) {
		DRM_ERROR("Failed to allocate ringbuffer\n");
599
		ret = -ENOMEM;
600
		goto err_hws;
601 602
	}

603 604
	ring->gem_object = obj;

605
	ret = i915_gem_object_pin(obj, PAGE_SIZE, true, false);
606 607
	if (ret)
		goto err_unref;
608

609 610
	obj_priv = to_intel_bo(obj);
	ring->map.size = ring->size;
611 612 613 614 615 616 617 618
	ring->map.offset = dev->agp->base + obj_priv->gtt_offset;
	ring->map.type = 0;
	ring->map.flags = 0;
	ring->map.mtrr = 0;

	drm_core_ioremap_wc(&ring->map, dev);
	if (ring->map.handle == NULL) {
		DRM_ERROR("Failed to map ringbuffer.\n");
619
		ret = -EINVAL;
620
		goto err_unpin;
621 622
	}

623
	ring->virtual_start = ring->map.handle;
624
	ret = ring->init(ring);
625 626
	if (ret)
		goto err_unmap;
627

628
	return 0;
629 630 631 632 633 634 635 636 637

err_unmap:
	drm_core_ioremapfree(&ring->map, dev);
err_unpin:
	i915_gem_object_unpin(obj);
err_unref:
	drm_gem_object_unreference(obj);
	ring->gem_object = NULL;
err_hws:
638
	cleanup_status_page(ring);
639
	return ret;
640 641
}

642
void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring)
643
{
644 645 646
	struct drm_i915_private *dev_priv;
	int ret;

647
	if (ring->gem_object == NULL)
648 649
		return;

650 651 652 653 654
	/* Disable the ring buffer. The ring must be idle at this point */
	dev_priv = ring->dev->dev_private;
	ret = intel_wait_ring_buffer(ring, ring->size - 8);
	I915_WRITE_CTL(ring, 0);

655
	drm_core_ioremapfree(&ring->map, ring->dev);
656

657 658 659
	i915_gem_object_unpin(ring->gem_object);
	drm_gem_object_unreference(ring->gem_object);
	ring->gem_object = NULL;
660 661

	cleanup_status_page(ring);
662 663
}

664
static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring)
665
{
666
	unsigned int *virt;
667
	int rem;
668
	rem = ring->size - ring->tail;
669

670
	if (ring->space < rem) {
671
		int ret = intel_wait_ring_buffer(ring, rem);
672 673 674 675
		if (ret)
			return ret;
	}

676
	virt = (unsigned int *)(ring->virtual_start + ring->tail);
677 678
	rem /= 8;
	while (rem--) {
679
		*virt++ = MI_NOOP;
680 681
		*virt++ = MI_NOOP;
	}
682

683
	ring->tail = 0;
684
	ring->space = ring->head - 8;
685 686 687 688

	return 0;
}

689
int intel_wait_ring_buffer(struct intel_ring_buffer *ring, int n)
690
{
691
	struct drm_device *dev = ring->dev;
692
	drm_i915_private_t *dev_priv = dev->dev_private;
693
	unsigned long end;
694 695

	trace_i915_ring_wait_begin (dev);
696 697
	end = jiffies + 3 * HZ;
	do {
698
		ring->head = I915_READ_HEAD(ring) & HEAD_ADDR;
699 700
		ring->space = ring->head - (ring->tail + 8);
		if (ring->space < 0)
701
			ring->space += ring->size;
702
		if (ring->space >= n) {
703
			trace_i915_ring_wait_end(dev);
704 705 706 707 708 709 710 711
			return 0;
		}

		if (dev->primary->master) {
			struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv;
			if (master_priv->sarea_priv)
				master_priv->sarea_priv->perf_boxes |= I915_BOX_WAIT;
		}
712

713
		msleep(1);
714 715
		if (atomic_read(&dev_priv->mm.wedged))
			return -EAGAIN;
716 717 718 719
	} while (!time_after(jiffies, end));
	trace_i915_ring_wait_end (dev);
	return -EBUSY;
}
720

721 722
int intel_ring_begin(struct intel_ring_buffer *ring,
		     int num_dwords)
723
{
724
	int n = 4*num_dwords;
725
	int ret;
726

727 728 729 730 731
	if (unlikely(ring->tail + n > ring->size)) {
		ret = intel_wrap_ring_buffer(ring);
		if (unlikely(ret))
			return ret;
	}
732

733 734 735 736 737
	if (unlikely(ring->space < n)) {
		ret = intel_wait_ring_buffer(ring, n);
		if (unlikely(ret))
			return ret;
	}
738 739

	ring->space -= n;
740
	return 0;
741
}
742

743
void intel_ring_advance(struct intel_ring_buffer *ring)
744
{
745
	ring->tail &= ring->size - 1;
746
	ring->write_tail(ring, ring->tail);
747
}
748

749
static const struct intel_ring_buffer render_ring = {
750
	.name			= "render ring",
751
	.id			= RING_RENDER,
752
	.mmio_base		= RENDER_RING_BASE,
753 754
	.size			= 32 * PAGE_SIZE,
	.init			= init_render_ring,
755
	.write_tail		= ring_write_tail,
756 757
	.flush			= render_ring_flush,
	.add_request		= render_ring_add_request,
758
	.get_seqno		= render_ring_get_seqno,
759 760
	.user_irq_get		= render_ring_get_user_irq,
	.user_irq_put		= render_ring_put_user_irq,
761
	.dispatch_execbuffer	= render_ring_dispatch_execbuffer,
762
};
763 764 765

/* ring buffer for bit-stream decoder */

766
static const struct intel_ring_buffer bsd_ring = {
767
	.name                   = "bsd ring",
768
	.id			= RING_BSD,
769
	.mmio_base		= BSD_RING_BASE,
770
	.size			= 32 * PAGE_SIZE,
771
	.init			= init_ring_common,
772
	.write_tail		= ring_write_tail,
773
	.flush			= bsd_ring_flush,
774 775
	.add_request		= ring_add_request,
	.get_seqno		= ring_status_page_get_seqno,
776 777
	.user_irq_get		= bsd_ring_get_user_irq,
	.user_irq_put		= bsd_ring_put_user_irq,
778
	.dispatch_execbuffer	= ring_dispatch_execbuffer,
779
};
780

781

782
static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring,
783
				     u32 value)
784
{
785
       drm_i915_private_t *dev_priv = ring->dev->dev_private;
786 787 788 789 790 791 792 793 794 795 796 797

       /* Every tail move must follow the sequence below */
       I915_WRITE(GEN6_BSD_SLEEP_PSMI_CONTROL,
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_MODIFY_MASK |
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_DISABLE);
       I915_WRITE(GEN6_BSD_RNCID, 0x0);

       if (wait_for((I915_READ(GEN6_BSD_SLEEP_PSMI_CONTROL) &
                               GEN6_BSD_SLEEP_PSMI_CONTROL_IDLE_INDICATOR) == 0,
                       50))
               DRM_ERROR("timed out waiting for IDLE Indicator\n");

798
       I915_WRITE_TAIL(ring, value);
799 800 801 802 803
       I915_WRITE(GEN6_BSD_SLEEP_PSMI_CONTROL,
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_MODIFY_MASK |
	       GEN6_BSD_SLEEP_PSMI_CONTROL_RC_ILDL_MESSAGE_ENABLE);
}

804
static void gen6_ring_flush(struct intel_ring_buffer *ring,
805 806
			    u32 invalidate_domains,
			    u32 flush_domains)
807
{
808 809 810 811 812 813 814
	if (intel_ring_begin(ring, 4) == 0) {
		intel_ring_emit(ring, MI_FLUSH_DW);
		intel_ring_emit(ring, 0);
		intel_ring_emit(ring, 0);
		intel_ring_emit(ring, 0);
		intel_ring_advance(ring);
	}
815 816 817
}

static int
818 819 820 821
gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring,
			      struct drm_i915_gem_execbuffer2 *exec,
			      struct drm_clip_rect *cliprects,
			      uint64_t exec_offset)
822 823
{
       uint32_t exec_start;
824
       int ret;
825

826
       exec_start = (uint32_t) exec_offset + exec->batch_start_offset;
827

828 829 830 831
       ret = intel_ring_begin(ring, 2);
       if (ret)
	       return ret;

832
       intel_ring_emit(ring, MI_BATCH_BUFFER_START | MI_BATCH_NON_SECURE_I965);
833
       /* bit0-7 is the length on GEN6+ */
834 835
       intel_ring_emit(ring, exec_start);
       intel_ring_advance(ring);
836

837 838 839 840
       return 0;
}

/* ring buffer for Video Codec for Gen6+ */
841
static const struct intel_ring_buffer gen6_bsd_ring = {
842 843
       .name			= "gen6 bsd ring",
       .id			= RING_BSD,
844
       .mmio_base		= GEN6_BSD_RING_BASE,
845
       .size			= 32 * PAGE_SIZE,
846
       .init			= init_ring_common,
847
       .write_tail		= gen6_bsd_ring_write_tail,
848 849 850
       .flush			= gen6_ring_flush,
       .add_request		= ring_add_request,
       .get_seqno		= ring_status_page_get_seqno,
851 852
       .user_irq_get		= bsd_ring_get_user_irq,
       .user_irq_put		= bsd_ring_put_user_irq,
853
       .dispatch_execbuffer	= gen6_ring_dispatch_execbuffer,
854 855 856 857 858
};

/* Blitter support (SandyBridge+) */

static void
859
blt_ring_get_user_irq(struct intel_ring_buffer *ring)
860 861 862 863
{
	/* do nothing */
}
static void
864
blt_ring_put_user_irq(struct intel_ring_buffer *ring)
865 866 867 868 869 870 871 872 873 874
{
	/* do nothing */
}

static const struct intel_ring_buffer gen6_blt_ring = {
       .name			= "blt ring",
       .id			= RING_BLT,
       .mmio_base		= BLT_RING_BASE,
       .size			= 32 * PAGE_SIZE,
       .init			= init_ring_common,
875
       .write_tail		= ring_write_tail,
876 877 878 879 880
       .flush			= gen6_ring_flush,
       .add_request		= ring_add_request,
       .get_seqno		= ring_status_page_get_seqno,
       .user_irq_get		= blt_ring_get_user_irq,
       .user_irq_put		= blt_ring_put_user_irq,
881
       .dispatch_execbuffer	= gen6_ring_dispatch_execbuffer,
882 883
};

884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903
int intel_init_render_ring_buffer(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;

	dev_priv->render_ring = render_ring;

	if (!I915_NEED_GFX_HWS(dev)) {
		dev_priv->render_ring.status_page.page_addr
			= dev_priv->status_page_dmah->vaddr;
		memset(dev_priv->render_ring.status_page.page_addr,
				0, PAGE_SIZE);
	}

	return intel_init_ring_buffer(dev, &dev_priv->render_ring);
}

int intel_init_bsd_ring_buffer(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;

904 905 906 907
	if (IS_GEN6(dev))
		dev_priv->bsd_ring = gen6_bsd_ring;
	else
		dev_priv->bsd_ring = bsd_ring;
908 909 910

	return intel_init_ring_buffer(dev, &dev_priv->bsd_ring);
}
911 912 913 914 915 916 917 918 919

int intel_init_blt_ring_buffer(struct drm_device *dev)
{
	drm_i915_private_t *dev_priv = dev->dev_private;

	dev_priv->blt_ring = gen6_blt_ring;

	return intel_init_ring_buffer(dev, &dev_priv->blt_ring);
}