nouveau_fence.c 14.3 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29
/*
 * Copyright (C) 2007 Ben Skeggs.
 * All Rights Reserved.
 *
 * Permission is hereby granted, free of charge, to any person obtaining
 * a copy of this software and associated documentation files (the
 * "Software"), to deal in the Software without restriction, including
 * without limitation the rights to use, copy, modify, merge, publish,
 * distribute, sublicense, and/or sell copies of the Software, and to
 * permit persons to whom the Software is furnished to do so, subject to
 * the following conditions:
 *
 * The above copyright notice and this permission notice (including the
 * next paragraph) shall be included in all copies or substantial
 * portions of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 *
 */

#include "drmP.h"
#include "drm.h"

30 31 32
#include <linux/ktime.h>
#include <linux/hrtimer.h>

33
#include "nouveau_drv.h"
34
#include "nouveau_ramht.h"
35 36
#include "nouveau_dma.h"

37
#define USE_REFCNT(dev) (nouveau_private(dev)->chipset >= 0x10)
38
#define USE_SEMA(dev) (nouveau_private(dev)->chipset >= 0x17)
39 40 41 42 43 44 45 46

struct nouveau_fence {
	struct nouveau_channel *channel;
	struct kref refcount;
	struct list_head entry;

	uint32_t sequence;
	bool signalled;
47 48 49

	void (*work)(void *priv, bool signalled);
	void *priv;
50 51
};

52 53 54 55 56 57
struct nouveau_semaphore {
	struct kref ref;
	struct drm_device *dev;
	struct drm_mm_node *mem;
};

58 59 60 61 62 63 64 65 66 67 68 69
static inline struct nouveau_fence *
nouveau_fence(void *sync_obj)
{
	return (struct nouveau_fence *)sync_obj;
}

static void
nouveau_fence_del(struct kref *ref)
{
	struct nouveau_fence *fence =
		container_of(ref, struct nouveau_fence, refcount);

70
	nouveau_channel_ref(NULL, &fence->channel);
71 72 73 74 75 76
	kfree(fence);
}

void
nouveau_fence_update(struct nouveau_channel *chan)
{
77 78
	struct drm_device *dev = chan->dev;
	struct nouveau_fence *tmp, *fence;
79 80
	uint32_t sequence;

81 82
	spin_lock(&chan->fence.lock);

83 84 85 86 87 88 89 90 91 92 93
	/* Fetch the last sequence if the channel is still up and running */
	if (likely(!list_empty(&chan->fence.pending))) {
		if (USE_REFCNT(dev))
			sequence = nvchan_rd32(chan, 0x48);
		else
			sequence = atomic_read(&chan->fence.last_sequence_irq);

		if (chan->fence.sequence_ack == sequence)
			goto out;
		chan->fence.sequence_ack = sequence;
	}
94

95
	list_for_each_entry_safe(fence, tmp, &chan->fence.pending, entry) {
96 97 98
		if (fence->sequence > chan->fence.sequence_ack)
			break;

99 100
		fence->signalled = true;
		list_del(&fence->entry);
101
		if (fence->work)
102 103
			fence->work(fence->priv, true);

104 105
		kref_put(&fence->refcount, nouveau_fence_del);
	}
106

107
out:
108
	spin_unlock(&chan->fence.lock);
109 110 111 112 113 114 115 116 117 118 119 120 121
}

int
nouveau_fence_new(struct nouveau_channel *chan, struct nouveau_fence **pfence,
		  bool emit)
{
	struct nouveau_fence *fence;
	int ret = 0;

	fence = kzalloc(sizeof(*fence), GFP_KERNEL);
	if (!fence)
		return -ENOMEM;
	kref_init(&fence->refcount);
122
	nouveau_channel_ref(chan, &fence->channel);
123 124 125 126 127

	if (emit)
		ret = nouveau_fence_emit(fence);

	if (ret)
128
		nouveau_fence_unref(&fence);
129 130 131 132 133 134 135
	*pfence = fence;
	return ret;
}

struct nouveau_channel *
nouveau_fence_channel(struct nouveau_fence *fence)
{
136
	return fence ? nouveau_channel_get_unlocked(fence->channel) : NULL;
137 138 139 140 141 142
}

int
nouveau_fence_emit(struct nouveau_fence *fence)
{
	struct nouveau_channel *chan = fence->channel;
143
	struct drm_device *dev = chan->dev;
B
Ben Skeggs 已提交
144
	struct drm_nouveau_private *dev_priv = dev->dev_private;
145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160
	int ret;

	ret = RING_SPACE(chan, 2);
	if (ret)
		return ret;

	if (unlikely(chan->fence.sequence == chan->fence.sequence_ack - 1)) {
		nouveau_fence_update(chan);

		BUG_ON(chan->fence.sequence ==
		       chan->fence.sequence_ack - 1);
	}

	fence->sequence = ++chan->fence.sequence;

	kref_get(&fence->refcount);
161
	spin_lock(&chan->fence.lock);
162
	list_add_tail(&fence->entry, &chan->fence.pending);
163
	spin_unlock(&chan->fence.lock);
164

B
Ben Skeggs 已提交
165 166
	if (USE_REFCNT(dev)) {
		if (dev_priv->card_type < NV_C0)
167
			BEGIN_RING(chan, 0, NV10_SUBCHAN_REF_CNT, 1);
B
Ben Skeggs 已提交
168
		else
169
			BEGIN_NVC0(chan, 2, 0, NV10_SUBCHAN_REF_CNT, 1);
B
Ben Skeggs 已提交
170 171 172 173
	} else {
		BEGIN_RING(chan, NvSubSw, 0x0150, 1);
	}
	OUT_RING (chan, fence->sequence);
174 175 176 177 178
	FIRE_RING(chan);

	return 0;
}

179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197
void
nouveau_fence_work(struct nouveau_fence *fence,
		   void (*work)(void *priv, bool signalled),
		   void *priv)
{
	BUG_ON(fence->work);

	spin_lock(&fence->channel->fence.lock);

	if (fence->signalled) {
		work(priv, true);
	} else {
		fence->work = work;
		fence->priv = priv;
	}

	spin_unlock(&fence->channel->fence.lock);
}

198
void
199
__nouveau_fence_unref(void **sync_obj)
200 201 202 203 204 205 206 207 208
{
	struct nouveau_fence *fence = nouveau_fence(*sync_obj);

	if (fence)
		kref_put(&fence->refcount, nouveau_fence_del);
	*sync_obj = NULL;
}

void *
209
__nouveau_fence_ref(void *sync_obj)
210 211 212 213 214 215 216 217
{
	struct nouveau_fence *fence = nouveau_fence(sync_obj);

	kref_get(&fence->refcount);
	return sync_obj;
}

bool
218
__nouveau_fence_signalled(void *sync_obj, void *sync_arg)
219 220 221 222 223 224 225 226 227 228 229 230
{
	struct nouveau_fence *fence = nouveau_fence(sync_obj);
	struct nouveau_channel *chan = fence->channel;

	if (fence->signalled)
		return true;

	nouveau_fence_update(chan);
	return fence->signalled;
}

int
231
__nouveau_fence_wait(void *sync_obj, void *sync_arg, bool lazy, bool intr)
232 233
{
	unsigned long timeout = jiffies + (3 * DRM_HZ);
234 235
	unsigned long sleep_time = NSEC_PER_MSEC / 1000;
	ktime_t t;
236 237 238
	int ret = 0;

	while (1) {
239
		if (__nouveau_fence_signalled(sync_obj, sync_arg))
240 241 242 243 244 245 246
			break;

		if (time_after_eq(jiffies, timeout)) {
			ret = -EBUSY;
			break;
		}

247 248
		__set_current_state(intr ? TASK_INTERRUPTIBLE
			: TASK_UNINTERRUPTIBLE);
249 250 251 252 253 254 255
		if (lazy) {
			t = ktime_set(0, sleep_time);
			schedule_hrtimeout(&t, HRTIMER_MODE_REL);
			sleep_time *= 2;
			if (sleep_time > NSEC_PER_MSEC)
				sleep_time = NSEC_PER_MSEC;
		}
256 257

		if (intr && signal_pending(current)) {
B
Ben Skeggs 已提交
258
			ret = -ERESTARTSYS;
259 260 261 262 263 264 265 266 267
			break;
		}
	}

	__set_current_state(TASK_RUNNING);

	return ret;
}

268
static struct nouveau_semaphore *
269
semaphore_alloc(struct drm_device *dev)
270 271 272
{
	struct drm_nouveau_private *dev_priv = dev->dev_private;
	struct nouveau_semaphore *sema;
273 274
	int size = (dev_priv->chipset < 0x84) ? 4 : 16;
	int ret, i;
275 276 277 278 279 280 281 282

	if (!USE_SEMA(dev))
		return NULL;

	sema = kmalloc(sizeof(*sema), GFP_KERNEL);
	if (!sema)
		goto fail;

283 284 285 286
	ret = drm_mm_pre_get(&dev_priv->fence.heap);
	if (ret)
		goto fail;

287
	spin_lock(&dev_priv->fence.lock);
288
	sema->mem = drm_mm_search_free(&dev_priv->fence.heap, size, 0, 0);
289
	if (sema->mem)
290
		sema->mem = drm_mm_get_block_atomic(sema->mem, size, 0);
291 292 293 294 295 296 297
	spin_unlock(&dev_priv->fence.lock);

	if (!sema->mem)
		goto fail;

	kref_init(&sema->ref);
	sema->dev = dev;
298 299
	for (i = sema->mem->start; i < sema->mem->start + size; i += 4)
		nouveau_bo_wr32(dev_priv->fence.bo, i / 4, 0);
300 301 302 303 304 305 306 307

	return sema;
fail:
	kfree(sema);
	return NULL;
}

static void
308
semaphore_free(struct kref *ref)
309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329
{
	struct nouveau_semaphore *sema =
		container_of(ref, struct nouveau_semaphore, ref);
	struct drm_nouveau_private *dev_priv = sema->dev->dev_private;

	spin_lock(&dev_priv->fence.lock);
	drm_mm_put_block(sema->mem);
	spin_unlock(&dev_priv->fence.lock);

	kfree(sema);
}

static void
semaphore_work(void *priv, bool signalled)
{
	struct nouveau_semaphore *sema = priv;
	struct drm_nouveau_private *dev_priv = sema->dev->dev_private;

	if (unlikely(!signalled))
		nouveau_bo_wr32(dev_priv->fence.bo, sema->mem->start / 4, 1);

330
	kref_put(&sema->ref, semaphore_free);
331 332 333
}

static int
334
semaphore_acquire(struct nouveau_channel *chan, struct nouveau_semaphore *sema)
335
{
336 337
	struct drm_nouveau_private *dev_priv = chan->dev->dev_private;
	struct nouveau_fence *fence = NULL;
338
	u64 offset = chan->fence.vma.offset + sema->mem->start;
339 340
	int ret;

341
	if (dev_priv->chipset < 0x84) {
342
		ret = RING_SPACE(chan, 4);
343 344
		if (ret)
			return ret;
345

346
		BEGIN_RING(chan, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 3);
347
		OUT_RING  (chan, NvSema);
348
		OUT_RING  (chan, offset);
349
		OUT_RING  (chan, 1);
350 351
	} else
	if (dev_priv->chipset < 0xc0) {
352
		ret = RING_SPACE(chan, 7);
353 354 355
		if (ret)
			return ret;

356
		BEGIN_RING(chan, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 1);
357
		OUT_RING  (chan, chan->vram_handle);
358
		BEGIN_RING(chan, 0, NV84_SUBCHAN_SEMAPHORE_ADDRESS_HIGH, 4);
359 360
		OUT_RING  (chan, upper_32_bits(offset));
		OUT_RING  (chan, lower_32_bits(offset));
361 362
		OUT_RING  (chan, 1);
		OUT_RING  (chan, 1); /* ACQUIRE_EQ */
363 364 365 366 367
	} else {
		ret = RING_SPACE(chan, 5);
		if (ret)
			return ret;

368
		BEGIN_NVC0(chan, 2, 0, NV84_SUBCHAN_SEMAPHORE_ADDRESS_HIGH, 4);
369 370 371 372
		OUT_RING  (chan, upper_32_bits(offset));
		OUT_RING  (chan, lower_32_bits(offset));
		OUT_RING  (chan, 1);
		OUT_RING  (chan, 0x1001); /* ACQUIRE_EQ */
373 374
	}

375 376 377 378
	/* Delay semaphore destruction until its work is done */
	ret = nouveau_fence_new(chan, &fence, true);
	if (ret)
		return ret;
379

380 381 382 383 384 385 386 387 388 389 390
	kref_get(&sema->ref);
	nouveau_fence_work(fence, semaphore_work, sema);
	nouveau_fence_unref(&fence);
	return 0;
}

static int
semaphore_release(struct nouveau_channel *chan, struct nouveau_semaphore *sema)
{
	struct drm_nouveau_private *dev_priv = chan->dev->dev_private;
	struct nouveau_fence *fence = NULL;
391
	u64 offset = chan->fence.vma.offset + sema->mem->start;
392 393 394
	int ret;

	if (dev_priv->chipset < 0x84) {
395
		ret = RING_SPACE(chan, 5);
396 397 398
		if (ret)
			return ret;

399
		BEGIN_RING(chan, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 2);
400
		OUT_RING  (chan, NvSema);
401
		OUT_RING  (chan, offset);
402
		BEGIN_RING(chan, 0, NV11_SUBCHAN_SEMAPHORE_RELEASE, 1);
403
		OUT_RING  (chan, 1);
404 405
	} else
	if (dev_priv->chipset < 0xc0) {
406
		ret = RING_SPACE(chan, 7);
407 408 409
		if (ret)
			return ret;

410
		BEGIN_RING(chan, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 1);
411
		OUT_RING  (chan, chan->vram_handle);
412
		BEGIN_RING(chan, 0, NV84_SUBCHAN_SEMAPHORE_ADDRESS_HIGH, 4);
413 414
		OUT_RING  (chan, upper_32_bits(offset));
		OUT_RING  (chan, lower_32_bits(offset));
415 416
		OUT_RING  (chan, 1);
		OUT_RING  (chan, 2); /* RELEASE */
417 418 419 420 421
	} else {
		ret = RING_SPACE(chan, 5);
		if (ret)
			return ret;

422
		BEGIN_NVC0(chan, 2, 0, NV84_SUBCHAN_SEMAPHORE_ADDRESS_HIGH, 4);
423 424 425 426
		OUT_RING  (chan, upper_32_bits(offset));
		OUT_RING  (chan, lower_32_bits(offset));
		OUT_RING  (chan, 1);
		OUT_RING  (chan, 0x1002); /* RELEASE */
427 428
	}

429 430 431 432 433 434 435
	/* Delay semaphore destruction until its work is done */
	ret = nouveau_fence_new(chan, &fence, true);
	if (ret)
		return ret;

	kref_get(&sema->ref);
	nouveau_fence_work(fence, semaphore_work, sema);
436
	nouveau_fence_unref(&fence);
437 438 439
	return 0;
}

440 441 442 443 444
int
nouveau_fence_sync(struct nouveau_fence *fence,
		   struct nouveau_channel *wchan)
{
	struct nouveau_channel *chan = nouveau_fence_channel(fence);
445 446
	struct drm_device *dev = wchan->dev;
	struct nouveau_semaphore *sema;
447
	int ret = 0;
448

449
	if (likely(!chan || chan == wchan ||
450
		   nouveau_fence_signalled(fence)))
451
		goto out;
452

453
	sema = semaphore_alloc(dev);
454 455 456
	if (!sema) {
		/* Early card or broken userspace, fall back to
		 * software sync. */
457
		ret = nouveau_fence_wait(fence, true, false);
458
		goto out;
459 460
	}

461
	/* try to take chan's mutex, if we can't take it right away
462 463 464
	 * we have to fallback to software sync to prevent locking
	 * order issues
	 */
465
	if (!mutex_trylock(&chan->mutex)) {
466
		ret = nouveau_fence_wait(fence, true, false);
467
		goto out_unref;
468 469
	}

470
	/* Make wchan wait until it gets signalled */
471
	ret = semaphore_acquire(wchan, sema);
472
	if (ret)
473
		goto out_unlock;
474

475
	/* Signal the semaphore from chan */
476
	ret = semaphore_release(chan, sema);
477 478

out_unlock:
479
	mutex_unlock(&chan->mutex);
480
out_unref:
481
	kref_put(&sema->ref, semaphore_free);
482 483 484
out:
	if (chan)
		nouveau_channel_put_unlocked(&chan);
485
	return ret;
486 487
}

488
int
489
__nouveau_fence_flush(void *sync_obj, void *sync_arg)
490 491 492 493 494
{
	return 0;
}

int
495
nouveau_fence_channel_init(struct nouveau_channel *chan)
496
{
497 498
	struct drm_device *dev = chan->dev;
	struct drm_nouveau_private *dev_priv = dev->dev_private;
499 500 501
	struct nouveau_gpuobj *obj = NULL;
	int ret;

502 503 504 505 506
	if (dev_priv->card_type < NV_C0) {
		/* Create an NV_SW object for various sync purposes */
		ret = nouveau_gpuobj_gr_new(chan, NvSw, NV_SW);
		if (ret)
			return ret;
507

508 509 510
		ret = RING_SPACE(chan, 2);
		if (ret)
			return ret;
511

512
		BEGIN_RING(chan, NvSubSw, NV01_SUBCHAN_OBJECT, 1);
513 514 515
		OUT_RING  (chan, NvSw);
		FIRE_RING (chan);
	}
516

517
	/* Setup area of memory shared between all channels for x-chan sync */
518
	if (USE_SEMA(dev) && dev_priv->chipset < 0x84) {
519
		struct ttm_mem_reg *mem = &dev_priv->fence.bo->bo.mem;
520

521
		ret = nouveau_gpuobj_dma_new(chan, NV_CLASS_DMA_FROM_MEMORY,
522
					     mem->start << PAGE_SHIFT,
523
					     mem->size, NV_MEM_ACCESS_RW,
524
					     NV_MEM_TARGET_VRAM, &obj);
525 526 527 528 529 530 531
		if (ret)
			return ret;

		ret = nouveau_ramht_insert(chan, NvSema, obj);
		nouveau_gpuobj_ref(NULL, &obj);
		if (ret)
			return ret;
532 533
	} else
	if (USE_SEMA(dev)) {
534 535 536 537 538
		/* map fence bo into channel's vm */
		ret = nouveau_bo_vma_add(dev_priv->fence.bo, chan->vm,
					 &chan->fence.vma);
		if (ret)
			return ret;
539 540
	}

541
	atomic_set(&chan->fence.last_sequence_irq, 0);
542 543 544 545
	return 0;
}

void
546
nouveau_fence_channel_fini(struct nouveau_channel *chan)
547
{
548
	struct drm_nouveau_private *dev_priv = chan->dev->dev_private;
549
	struct nouveau_fence *tmp, *fence;
550

551
	spin_lock(&chan->fence.lock);
552
	list_for_each_entry_safe(fence, tmp, &chan->fence.pending, entry) {
553 554
		fence->signalled = true;
		list_del(&fence->entry);
555 556 557 558

		if (unlikely(fence->work))
			fence->work(fence->priv, false);

559 560
		kref_put(&fence->refcount, nouveau_fence_del);
	}
561
	spin_unlock(&chan->fence.lock);
562 563

	nouveau_bo_vma_del(dev_priv->fence.bo, &chan->fence.vma);
564 565
}

566 567 568 569
int
nouveau_fence_init(struct drm_device *dev)
{
	struct drm_nouveau_private *dev_priv = dev->dev_private;
570
	int size = (dev_priv->chipset < 0x84) ? 4096 : 16384;
571 572 573 574
	int ret;

	/* Create a shared VRAM heap for cross-channel sync. */
	if (USE_SEMA(dev)) {
575
		ret = nouveau_bo_new(dev, size, 0, TTM_PL_FLAG_VRAM,
576
				     0, 0, &dev_priv->fence.bo);
577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614
		if (ret)
			return ret;

		ret = nouveau_bo_pin(dev_priv->fence.bo, TTM_PL_FLAG_VRAM);
		if (ret)
			goto fail;

		ret = nouveau_bo_map(dev_priv->fence.bo);
		if (ret)
			goto fail;

		ret = drm_mm_init(&dev_priv->fence.heap, 0,
				  dev_priv->fence.bo->bo.mem.size);
		if (ret)
			goto fail;

		spin_lock_init(&dev_priv->fence.lock);
	}

	return 0;
fail:
	nouveau_bo_unmap(dev_priv->fence.bo);
	nouveau_bo_ref(NULL, &dev_priv->fence.bo);
	return ret;
}

void
nouveau_fence_fini(struct drm_device *dev)
{
	struct drm_nouveau_private *dev_priv = dev->dev_private;

	if (USE_SEMA(dev)) {
		drm_mm_takedown(&dev_priv->fence.heap);
		nouveau_bo_unmap(dev_priv->fence.bo);
		nouveau_bo_unpin(dev_priv->fence.bo);
		nouveau_bo_ref(NULL, &dev_priv->fence.bo);
	}
}