nouveau_display.c 22.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
/*
 * Copyright (C) 2008 Maarten Maathuis.
 * All Rights Reserved.
 *
 * Permission is hereby granted, free of charge, to any person obtaining
 * a copy of this software and associated documentation files (the
 * "Software"), to deal in the Software without restriction, including
 * without limitation the rights to use, copy, modify, merge, publish,
 * distribute, sublicense, and/or sell copies of the Software, and to
 * permit persons to whom the Software is furnished to do so, subject to
 * the following conditions:
 *
 * The above copyright notice and this permission notice (including the
 * next paragraph) shall be included in all copies or substantial
 * portions of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 *
 */

27 28
#include <drm/drmP.h>
#include <drm/drm_crtc_helper.h>
29

30
#include "nouveau_fbcon.h"
31
#include "dispnv04/hw.h"
32 33
#include "nouveau_crtc.h"
#include "nouveau_dma.h"
34
#include "nouveau_gem.h"
35
#include "nouveau_connector.h"
36
#include "nv50_display.h"
37

38 39
#include "nouveau_fence.h"

40
#include <engine/disp.h>
41

42 43
#include <core/class.h>

44
static int
45
nouveau_display_vblank_handler(void *data, u32 type, int head)
46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70
{
	struct nouveau_drm *drm = data;
	drm_handle_vblank(drm->dev, head);
	return NVKM_EVENT_KEEP;
}

int
nouveau_display_vblank_enable(struct drm_device *dev, int head)
{
	struct nouveau_display *disp = nouveau_display(dev);
	if (disp) {
		nouveau_event_get(disp->vblank[head]);
		return 0;
	}
	return -EIO;
}

void
nouveau_display_vblank_disable(struct drm_device *dev, int head)
{
	struct nouveau_display *disp = nouveau_display(dev);
	if (disp)
		nouveau_event_put(disp->vblank[head]);
}

71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107
static inline int
calc(int blanks, int blanke, int total, int line)
{
	if (blanke >= blanks) {
		if (line >= blanks)
			line -= total;
	} else {
		if (line >= blanks)
			line -= total;
		line -= blanke + 1;
	}
	return line;
}

int
nouveau_display_scanoutpos_head(struct drm_crtc *crtc, int *vpos, int *hpos,
				ktime_t *stime, ktime_t *etime)
{
	const u32 mthd = NV04_DISP_SCANOUTPOS + nouveau_crtc(crtc)->index;
	struct nouveau_display *disp = nouveau_display(crtc->dev);
	struct nv04_display_scanoutpos args;
	int ret, retry = 1;

	do {
		ret = nv_exec(disp->core, mthd, &args, sizeof(args));
		if (ret != 0)
			return 0;

		if (args.vline) {
			ret |= DRM_SCANOUTPOS_ACCURATE;
			ret |= DRM_SCANOUTPOS_VALID;
			break;
		}

		if (retry) ndelay(crtc->linedur_ns);
	} while (retry--);

108
	*hpos = args.hline;
109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150
	*vpos = calc(args.vblanks, args.vblanke, args.vtotal, args.vline);
	if (stime) *stime = ns_to_ktime(args.time[0]);
	if (etime) *etime = ns_to_ktime(args.time[1]);

	if (*vpos < 0)
		ret |= DRM_SCANOUTPOS_INVBL;
	return ret;
}

int
nouveau_display_scanoutpos(struct drm_device *dev, int head, unsigned int flags,
			   int *vpos, int *hpos, ktime_t *stime, ktime_t *etime)
{
	struct drm_crtc *crtc;

	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		if (nouveau_crtc(crtc)->index == head) {
			return nouveau_display_scanoutpos_head(crtc, vpos, hpos,
							       stime, etime);
		}
	}

	return 0;
}

int
nouveau_display_vblstamp(struct drm_device *dev, int head, int *max_error,
			 struct timeval *time, unsigned flags)
{
	struct drm_crtc *crtc;

	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		if (nouveau_crtc(crtc)->index == head) {
			return drm_calc_vbltimestamp_from_scanoutpos(dev,
					head, max_error, time, flags, crtc,
					&crtc->hwmode);
		}
	}

	return -EINVAL;
}

151 152 153 154 155 156
static void
nouveau_display_vblank_fini(struct drm_device *dev)
{
	struct nouveau_display *disp = nouveau_display(dev);
	int i;

157 158
	drm_vblank_cleanup(dev);

159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180
	if (disp->vblank) {
		for (i = 0; i < dev->mode_config.num_crtc; i++)
			nouveau_event_ref(NULL, &disp->vblank[i]);
		kfree(disp->vblank);
		disp->vblank = NULL;
	}
}

static int
nouveau_display_vblank_init(struct drm_device *dev)
{
	struct nouveau_display *disp = nouveau_display(dev);
	struct nouveau_drm *drm = nouveau_drm(dev);
	struct nouveau_disp *pdisp = nouveau_disp(drm->device);
	int ret, i;

	disp->vblank = kzalloc(dev->mode_config.num_crtc *
			       sizeof(*disp->vblank), GFP_KERNEL);
	if (!disp->vblank)
		return -ENOMEM;

	for (i = 0; i < dev->mode_config.num_crtc; i++) {
181
		ret = nouveau_event_new(pdisp->vblank, 1, i,
182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198
					nouveau_display_vblank_handler,
					drm, &disp->vblank[i]);
		if (ret) {
			nouveau_display_vblank_fini(dev);
			return ret;
		}
	}

	ret = drm_vblank_init(dev, dev->mode_config.num_crtc);
	if (ret) {
		nouveau_display_vblank_fini(dev);
		return ret;
	}

	return 0;
}

199 200 201 202 203
static void
nouveau_user_framebuffer_destroy(struct drm_framebuffer *drm_fb)
{
	struct nouveau_framebuffer *fb = nouveau_framebuffer(drm_fb);

204
	if (fb->nvbo)
205
		drm_gem_object_unreference_unlocked(&fb->nvbo->gem);
206 207 208 209 210 211 212 213 214 215 216 217

	drm_framebuffer_cleanup(drm_fb);
	kfree(fb);
}

static int
nouveau_user_framebuffer_create_handle(struct drm_framebuffer *drm_fb,
				       struct drm_file *file_priv,
				       unsigned int *handle)
{
	struct nouveau_framebuffer *fb = nouveau_framebuffer(drm_fb);

218
	return drm_gem_handle_create(file_priv, &fb->nvbo->gem, handle);
219 220 221 222 223 224 225
}

static const struct drm_framebuffer_funcs nouveau_framebuffer_funcs = {
	.destroy = nouveau_user_framebuffer_destroy,
	.create_handle = nouveau_user_framebuffer_create_handle,
};

226
int
227 228
nouveau_framebuffer_init(struct drm_device *dev,
			 struct nouveau_framebuffer *nv_fb,
229
			 struct drm_mode_fb_cmd2 *mode_cmd,
230
			 struct nouveau_bo *nvbo)
231
{
232
	struct nouveau_drm *drm = nouveau_drm(dev);
233
	struct drm_framebuffer *fb = &nv_fb->base;
234 235
	int ret;

236 237 238
	drm_helper_mode_fill_fb_struct(fb, mode_cmd);
	nv_fb->nvbo = nvbo;

239
	if (nv_device(drm->device)->card_type >= NV_50) {
240 241 242 243 244 245 246 247 248 249 250
		u32 tile_flags = nouveau_bo_tile_layout(nvbo);
		if (tile_flags == 0x7a00 ||
		    tile_flags == 0xfe00)
			nv_fb->r_dma = NvEvoFB32;
		else
		if (tile_flags == 0x7000)
			nv_fb->r_dma = NvEvoFB16;
		else
			nv_fb->r_dma = NvEvoVRAM_LP;

		switch (fb->depth) {
251 252 253
		case  8: nv_fb->r_format = 0x1e00; break;
		case 15: nv_fb->r_format = 0xe900; break;
		case 16: nv_fb->r_format = 0xe800; break;
254
		case 24:
255 256
		case 32: nv_fb->r_format = 0xcf00; break;
		case 30: nv_fb->r_format = 0xd100; break;
257
		default:
258
			 NV_ERROR(drm, "unknown depth %d\n", fb->depth);
259 260 261
			 return -EINVAL;
		}

262 263 264 265 266
		if (nvbo->tile_flags & NOUVEAU_GEM_TILE_NONCONTIG) {
			NV_ERROR(drm, "framebuffer requires contiguous bo\n");
			return -EINVAL;
		}

267
		if (nv_device(drm->device)->chipset == 0x50)
268 269
			nv_fb->r_format |= (tile_flags << 8);

270
		if (!tile_flags) {
271
			if (nv_device(drm->device)->card_type < NV_D0)
272
				nv_fb->r_pitch = 0x00100000 | fb->pitches[0];
273
			else
274
				nv_fb->r_pitch = 0x01000000 | fb->pitches[0];
275
		} else {
276
			u32 mode = nvbo->tile_mode;
277
			if (nv_device(drm->device)->card_type >= NV_C0)
278
				mode >>= 4;
279
			nv_fb->r_pitch = ((fb->pitches[0] / 4) << 4) | mode;
280 281 282
		}
	}

283 284 285 286 287
	ret = drm_framebuffer_init(dev, fb, &nouveau_framebuffer_funcs);
	if (ret) {
		return ret;
	}

288
	return 0;
289 290 291 292 293
}

static struct drm_framebuffer *
nouveau_user_framebuffer_create(struct drm_device *dev,
				struct drm_file *file_priv,
294
				struct drm_mode_fb_cmd2 *mode_cmd)
295
{
296
	struct nouveau_framebuffer *nouveau_fb;
297
	struct drm_gem_object *gem;
298
	int ret = -ENOMEM;
299

300
	gem = drm_gem_object_lookup(dev, file_priv, mode_cmd->handles[0]);
301
	if (!gem)
302
		return ERR_PTR(-ENOENT);
303

304 305
	nouveau_fb = kzalloc(sizeof(struct nouveau_framebuffer), GFP_KERNEL);
	if (!nouveau_fb)
306
		goto err_unref;
307 308

	ret = nouveau_framebuffer_init(dev, nouveau_fb, mode_cmd, nouveau_gem_object(gem));
309 310
	if (ret)
		goto err;
311

312
	return &nouveau_fb->base;
313 314 315 316 317 318

err:
	kfree(nouveau_fb);
err_unref:
	drm_gem_object_unreference(gem);
	return ERR_PTR(ret);
319 320
}

321
static const struct drm_mode_config_funcs nouveau_mode_config_funcs = {
322
	.fb_create = nouveau_user_framebuffer_create,
323
	.output_poll_changed = nouveau_fbcon_output_poll_changed,
324 325
};

326

327
struct nouveau_drm_prop_enum_list {
328
	u8 gen_mask;
329 330 331 332
	int type;
	char *name;
};

333
static struct nouveau_drm_prop_enum_list underscan[] = {
334 335 336
	{ 6, UNDERSCAN_AUTO, "auto" },
	{ 6, UNDERSCAN_OFF, "off" },
	{ 6, UNDERSCAN_ON, "on" },
337
	{}
338 339
};

340
static struct nouveau_drm_prop_enum_list dither_mode[] = {
341 342 343 344 345 346 347 348 349
	{ 7, DITHERING_MODE_AUTO, "auto" },
	{ 7, DITHERING_MODE_OFF, "off" },
	{ 1, DITHERING_MODE_ON, "on" },
	{ 6, DITHERING_MODE_STATIC2X2, "static 2x2" },
	{ 6, DITHERING_MODE_DYNAMIC2X2, "dynamic 2x2" },
	{ 4, DITHERING_MODE_TEMPORAL, "temporal" },
	{}
};

350
static struct nouveau_drm_prop_enum_list dither_depth[] = {
351 352 353 354 355 356 357
	{ 6, DITHERING_DEPTH_AUTO, "auto" },
	{ 6, DITHERING_DEPTH_6BPC, "6 bpc" },
	{ 6, DITHERING_DEPTH_8BPC, "8 bpc" },
	{}
};

#define PROP_ENUM(p,gen,n,list) do {                                           \
358
	struct nouveau_drm_prop_enum_list *l = (list);                         \
359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378
	int c = 0;                                                             \
	while (l->gen_mask) {                                                  \
		if (l->gen_mask & (1 << (gen)))                                \
			c++;                                                   \
		l++;                                                           \
	}                                                                      \
	if (c) {                                                               \
		p = drm_property_create(dev, DRM_MODE_PROP_ENUM, n, c);        \
		l = (list);                                                    \
		c = 0;                                                         \
		while (p && l->gen_mask) {                                     \
			if (l->gen_mask & (1 << (gen))) {                      \
				drm_property_add_enum(p, c, l->type, l->name); \
				c++;                                           \
			}                                                      \
			l++;                                                   \
		}                                                              \
	}                                                                      \
} while(0)

379 380 381
int
nouveau_display_init(struct drm_device *dev)
{
382
	struct nouveau_display *disp = nouveau_display(dev);
383
	struct drm_connector *connector;
384 385 386
	int ret;

	ret = disp->init(dev);
387 388 389
	if (ret)
		return ret;

390
	/* enable polling for external displays */
391 392 393 394 395
	drm_kms_helper_poll_enable(dev);

	/* enable hotplug interrupts */
	list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
		struct nouveau_connector *conn = nouveau_connector(connector);
396
		if (conn->hpd) nouveau_event_get(conn->hpd);
397 398 399 400 401 402 403 404
	}

	return ret;
}

void
nouveau_display_fini(struct drm_device *dev)
{
405
	struct nouveau_display *disp = nouveau_display(dev);
406 407 408 409 410
	struct drm_connector *connector;

	/* disable hotplug interrupts */
	list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
		struct nouveau_connector *conn = nouveau_connector(connector);
411
		if (conn->hpd) nouveau_event_put(conn->hpd);
412
	}
413 414 415 416 417

	drm_kms_helper_poll_disable(dev);
	disp->fini(dev);
}

418 419 420
int
nouveau_display_create(struct drm_device *dev)
{
421
	struct nouveau_drm *drm = nouveau_drm(dev);
A
Alexandre Courbot 已提交
422
	struct nouveau_device *device = nouveau_dev(dev);
423
	struct nouveau_display *disp;
424
	int ret, gen;
425

426 427 428 429
	disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL);
	if (!disp)
		return -ENOMEM;

430 431
	drm_mode_config_init(dev);
	drm_mode_create_scaling_mode_property(dev);
432
	drm_mode_create_dvi_i_properties(dev);
433

434
	if (nv_device(drm->device)->card_type < NV_50)
435 436
		gen = 0;
	else
437
	if (nv_device(drm->device)->card_type < NV_D0)
438 439 440 441 442 443 444
		gen = 1;
	else
		gen = 2;

	PROP_ENUM(disp->dithering_mode, gen, "dithering mode", dither_mode);
	PROP_ENUM(disp->dithering_depth, gen, "dithering depth", dither_depth);
	PROP_ENUM(disp->underscan_property, gen, "underscan", underscan);
445 446

	disp->underscan_hborder_property =
447
		drm_property_create_range(dev, 0, "underscan hborder", 0, 128);
448 449

	disp->underscan_vborder_property =
450
		drm_property_create_range(dev, 0, "underscan vborder", 0, 128);
451

452
	if (gen >= 1) {
453
		/* -90..+90 */
454
		disp->vibrant_hue_property =
455
			drm_property_create_range(dev, 0, "vibrant hue", 0, 180);
456

457
		/* -100..+100 */
458
		disp->color_vibrance_property =
459
			drm_property_create_range(dev, 0, "color vibrance", 0, 200);
460 461
	}

462
	dev->mode_config.funcs = &nouveau_mode_config_funcs;
A
Alexandre Courbot 已提交
463
	dev->mode_config.fb_base = nv_device_resource_start(device, 1);
464 465 466

	dev->mode_config.min_width = 0;
	dev->mode_config.min_height = 0;
467
	if (nv_device(drm->device)->card_type < NV_10) {
468 469 470
		dev->mode_config.max_width = 2048;
		dev->mode_config.max_height = 2048;
	} else
471
	if (nv_device(drm->device)->card_type < NV_50) {
472 473 474 475 476 477 478
		dev->mode_config.max_width = 4096;
		dev->mode_config.max_height = 4096;
	} else {
		dev->mode_config.max_width = 8192;
		dev->mode_config.max_height = 8192;
	}

479 480 481
	dev->mode_config.preferred_depth = 24;
	dev->mode_config.prefer_shadow = 1;

482 483 484 485 486
	if (nv_device(drm->device)->chipset < 0x11)
		dev->mode_config.async_page_flip = false;
	else
		dev->mode_config.async_page_flip = true;

487 488 489
	drm_kms_helper_poll_init(dev);
	drm_kms_helper_poll_disable(dev);

490
	if (drm->vbios.dcb.entries) {
491
		static const u16 oclass[] = {
492
			GM107_DISP_CLASS,
493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516
			NVF0_DISP_CLASS,
			NVE0_DISP_CLASS,
			NVD0_DISP_CLASS,
			NVA3_DISP_CLASS,
			NV94_DISP_CLASS,
			NVA0_DISP_CLASS,
			NV84_DISP_CLASS,
			NV50_DISP_CLASS,
			NV04_DISP_CLASS,
		};
		int i;

		for (i = 0, ret = -ENODEV; ret && i < ARRAY_SIZE(oclass); i++) {
			ret = nouveau_object_new(nv_object(drm), NVDRM_DEVICE,
						 NVDRM_DISPLAY, oclass[i],
						 NULL, 0, &disp->core);
		}

		if (ret == 0) {
			if (nv_mclass(disp->core) < NV50_DISP_CLASS)
				ret = nv04_display_create(dev);
			else
				ret = nv50_display_create(dev);
		}
517 518 519
	} else {
		ret = 0;
	}
520

521 522
	if (ret)
		goto disp_create_err;
523

524
	if (dev->mode_config.num_crtc) {
525
		ret = nouveau_display_vblank_init(dev);
526 527
		if (ret)
			goto vblank_err;
528 529
	}

530
	nouveau_backlight_init(dev);
531 532 533
	return 0;

vblank_err:
534
	disp->dtor(dev);
535 536 537
disp_create_err:
	drm_kms_helper_poll_fini(dev);
	drm_mode_config_cleanup(dev);
538
	return ret;
539 540 541 542 543
}

void
nouveau_display_destroy(struct drm_device *dev)
{
544
	struct nouveau_display *disp = nouveau_display(dev);
545
	struct nouveau_drm *drm = nouveau_drm(dev);
546

547
	nouveau_backlight_exit(dev);
548
	nouveau_display_vblank_fini(dev);
549

550 551 552
	drm_kms_helper_poll_fini(dev);
	drm_mode_config_cleanup(dev);

553 554
	if (disp->dtor)
		disp->dtor(dev);
555

556 557
	nouveau_object_del(nv_object(drm), NVDRM_DEVICE, NVDRM_DISPLAY);

558 559 560 561 562 563 564 565 566 567 568 569
	nouveau_drm(dev)->display = NULL;
	kfree(disp);
}

int
nouveau_display_suspend(struct drm_device *dev)
{
	struct nouveau_drm *drm = nouveau_drm(dev);
	struct drm_crtc *crtc;

	nouveau_display_fini(dev);

570
	NV_INFO(drm, "unpinning framebuffer(s)...\n");
571 572 573
	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		struct nouveau_framebuffer *nouveau_fb;

574
		nouveau_fb = nouveau_framebuffer(crtc->primary->fb);
575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591
		if (!nouveau_fb || !nouveau_fb->nvbo)
			continue;

		nouveau_bo_unpin(nouveau_fb->nvbo);
	}

	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);

		nouveau_bo_unmap(nv_crtc->cursor.nvbo);
		nouveau_bo_unpin(nv_crtc->cursor.nvbo);
	}

	return 0;
}

void
592
nouveau_display_repin(struct drm_device *dev)
593 594 595 596 597 598 599 600
{
	struct nouveau_drm *drm = nouveau_drm(dev);
	struct drm_crtc *crtc;
	int ret;

	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		struct nouveau_framebuffer *nouveau_fb;

601
		nouveau_fb = nouveau_framebuffer(crtc->primary->fb);
602 603 604 605 606 607 608 609 610 611 612 613 614 615 616
		if (!nouveau_fb || !nouveau_fb->nvbo)
			continue;

		nouveau_bo_pin(nouveau_fb->nvbo, TTM_PL_FLAG_VRAM);
	}

	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);

		ret = nouveau_bo_pin(nv_crtc->cursor.nvbo, TTM_PL_FLAG_VRAM);
		if (!ret)
			ret = nouveau_bo_map(nv_crtc->cursor.nvbo);
		if (ret)
			NV_ERROR(drm, "Could not pin/map cursor.\n");
	}
617
}
618

619 620 621 622
void
nouveau_display_resume(struct drm_device *dev)
{
	struct drm_crtc *crtc;
623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641
	nouveau_display_init(dev);

	/* Force CLUT to get re-loaded during modeset */
	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);

		nv_crtc->lut.depth = 0;
	}

	drm_helper_resume_force_mode(dev);

	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) {
		struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc);
		u32 offset = nv_crtc->cursor.nvbo->bo.offset;

		nv_crtc->cursor.set_offset(nv_crtc, offset);
		nv_crtc->cursor.set_pos(nv_crtc, nv_crtc->cursor_saved_x,
						 nv_crtc->cursor_saved_y);
	}
642 643
}

644 645 646 647 648 649 650
static int
nouveau_page_flip_emit(struct nouveau_channel *chan,
		       struct nouveau_bo *old_bo,
		       struct nouveau_bo *new_bo,
		       struct nouveau_page_flip_state *s,
		       struct nouveau_fence **pfence)
{
651
	struct nouveau_fence_chan *fctx = chan->fence;
652 653
	struct nouveau_drm *drm = chan->drm;
	struct drm_device *dev = drm->dev;
654 655 656 657 658
	unsigned long flags;
	int ret;

	/* Queue it to the pending list */
	spin_lock_irqsave(&dev->event_lock, flags);
659
	list_add_tail(&s->head, &fctx->flip);
660 661 662 663 664 665 666 667
	spin_unlock_irqrestore(&dev->event_lock, flags);

	/* Synchronize with the old framebuffer */
	ret = nouveau_fence_sync(old_bo->bo.sync_obj, chan);
	if (ret)
		goto fail;

	/* Emit the pageflip */
668
	ret = RING_SPACE(chan, 2);
669 670 671
	if (ret)
		goto fail;

672
	if (nv_device(drm->device)->card_type < NV_C0)
673
		BEGIN_NV04(chan, NvSubSw, NV_SW_PAGE_FLIP, 1);
674 675 676
	else
		BEGIN_NVC0(chan, FermiSw, NV_SW_PAGE_FLIP, 1);
	OUT_RING  (chan, 0x00000000);
677
	FIRE_RING (chan);
678

679
	ret = nouveau_fence_new(chan, false, pfence);
680 681 682 683 684 685 686 687 688 689 690 691 692
	if (ret)
		goto fail;

	return 0;
fail:
	spin_lock_irqsave(&dev->event_lock, flags);
	list_del(&s->head);
	spin_unlock_irqrestore(&dev->event_lock, flags);
	return ret;
}

int
nouveau_crtc_page_flip(struct drm_crtc *crtc, struct drm_framebuffer *fb,
693
		       struct drm_pending_vblank_event *event, u32 flags)
694
{
695
	const int swap_interval = (flags & DRM_MODE_PAGE_FLIP_ASYNC) ? 0 : 1;
696
	struct drm_device *dev = crtc->dev;
697
	struct nouveau_drm *drm = nouveau_drm(dev);
698
	struct nouveau_bo *old_bo = nouveau_framebuffer(crtc->primary->fb)->nvbo;
699 700
	struct nouveau_bo *new_bo = nouveau_framebuffer(fb)->nvbo;
	struct nouveau_page_flip_state *s;
701
	struct nouveau_channel *chan = drm->channel;
702 703 704
	struct nouveau_fence *fence;
	int ret;

705
	if (!drm->channel)
706 707 708 709 710 711
		return -ENODEV;

	s = kzalloc(sizeof(*s), GFP_KERNEL);
	if (!s)
		return -ENOMEM;

712 713 714 715 716 717 718 719
	if (new_bo != old_bo) {
		ret = nouveau_bo_pin(new_bo, TTM_PL_FLAG_VRAM);
		if (ret)
			goto fail_free;
	}

	mutex_lock(&chan->cli->mutex);

720 721 722 723 724
	/* synchronise rendering channel with the kernel's channel */
	spin_lock(&new_bo->bo.bdev->fence_lock);
	fence = nouveau_fence_ref(new_bo->bo.sync_obj);
	spin_unlock(&new_bo->bo.bdev->fence_lock);
	ret = nouveau_fence_sync(fence, chan);
725
	nouveau_fence_unref(&fence);
726
	if (ret)
727
		goto fail_unpin;
728

729
	ret = ttm_bo_reserve(&old_bo->bo, true, false, false, NULL);
730 731
	if (ret)
		goto fail_unpin;
732 733 734 735 736 737 738

	/* Initialize a page flip struct */
	*s = (struct nouveau_page_flip_state)
		{ { }, event, nouveau_crtc(crtc)->index,
		  fb->bits_per_pixel, fb->pitches[0], crtc->x, crtc->y,
		  new_bo->bo.offset };

739 740 741
	/* Keep vblanks on during flip, for the target crtc of this flip */
	drm_vblank_get(dev, nouveau_crtc(crtc)->index);

742
	/* Emit a page flip */
743
	if (nv_device(drm->device)->card_type >= NV_50) {
744
		ret = nv50_display_flip_next(crtc, fb, chan, swap_interval);
745
		if (ret)
B
Ben Skeggs 已提交
746
			goto fail_unreserve;
747 748
	} else {
		struct nv04_display *dispnv04 = nv04_display(dev);
749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766
		int head = nouveau_crtc(crtc)->index;

		if (swap_interval) {
			ret = RING_SPACE(chan, 8);
			if (ret)
				goto fail_unreserve;

			BEGIN_NV04(chan, NvSubImageBlit, 0x012c, 1);
			OUT_RING  (chan, 0);
			BEGIN_NV04(chan, NvSubImageBlit, 0x0134, 1);
			OUT_RING  (chan, head);
			BEGIN_NV04(chan, NvSubImageBlit, 0x0100, 1);
			OUT_RING  (chan, 0);
			BEGIN_NV04(chan, NvSubImageBlit, 0x0130, 1);
			OUT_RING  (chan, 0);
		}

		nouveau_bo_ref(new_bo, &dispnv04->image[head]);
B
Ben Skeggs 已提交
767 768
	}

769 770 771
	ret = nouveau_page_flip_emit(chan, old_bo, new_bo, s, &fence);
	if (ret)
		goto fail_unreserve;
772
	mutex_unlock(&chan->cli->mutex);
773 774

	/* Update the crtc struct and cleanup */
775
	crtc->primary->fb = fb;
776

777 778
	nouveau_bo_fence(old_bo, fence);
	ttm_bo_unreserve(&old_bo->bo);
779
	if (old_bo != new_bo)
780
		nouveau_bo_unpin(old_bo);
781 782 783 784
	nouveau_fence_unref(&fence);
	return 0;

fail_unreserve:
785
	drm_vblank_put(dev, nouveau_crtc(crtc)->index);
786
	ttm_bo_unreserve(&old_bo->bo);
787 788 789
fail_unpin:
	mutex_unlock(&chan->cli->mutex);
	if (old_bo != new_bo)
790
		nouveau_bo_unpin(new_bo);
791 792 793 794 795 796 797 798 799
fail_free:
	kfree(s);
	return ret;
}

int
nouveau_finish_page_flip(struct nouveau_channel *chan,
			 struct nouveau_page_flip_state *ps)
{
800
	struct nouveau_fence_chan *fctx = chan->fence;
801 802
	struct nouveau_drm *drm = chan->drm;
	struct drm_device *dev = drm->dev;
803 804
	struct nouveau_page_flip_state *s;
	unsigned long flags;
805
	int crtcid = -1;
806 807 808

	spin_lock_irqsave(&dev->event_lock, flags);

809
	if (list_empty(&fctx->flip)) {
810
		NV_ERROR(drm, "unexpected pageflip\n");
811 812 813 814
		spin_unlock_irqrestore(&dev->event_lock, flags);
		return -EINVAL;
	}

815
	s = list_first_entry(&fctx->flip, struct nouveau_page_flip_state, head);
816 817 818 819 820 821 822
	if (s->event) {
		/* Vblank timestamps/counts are only correct on >= NV-50 */
		if (nv_device(drm->device)->card_type >= NV_50)
			crtcid = s->crtc;

		drm_send_vblank_event(dev, crtcid, s->event);
	}
823

824 825 826
	/* Give up ownership of vblank for page-flipped crtc */
	drm_vblank_put(dev, s->crtc);

827
	list_del(&s->head);
B
Ben Skeggs 已提交
828 829
	if (ps)
		*ps = *s;
830 831 832 833 834
	kfree(s);

	spin_unlock_irqrestore(&dev->event_lock, flags);
	return 0;
}
B
Ben Skeggs 已提交
835

836 837 838 839
int
nouveau_flip_complete(void *data)
{
	struct nouveau_channel *chan = data;
840
	struct nouveau_drm *drm = chan->drm;
841 842 843
	struct nouveau_page_flip_state state;

	if (!nouveau_finish_page_flip(chan, &state)) {
844 845
		if (nv_device(drm->device)->card_type < NV_50) {
			nv_set_crtc_base(drm->dev, state.crtc, state.offset +
846 847 848 849 850 851 852 853
					 state.y * state.pitch +
					 state.x * state.bpp / 8);
		}
	}

	return 0;
}

B
Ben Skeggs 已提交
854 855 856 857 858 859 860 861 862 863 864
int
nouveau_display_dumb_create(struct drm_file *file_priv, struct drm_device *dev,
			    struct drm_mode_create_dumb *args)
{
	struct nouveau_bo *bo;
	int ret;

	args->pitch = roundup(args->width * (args->bpp / 8), 256);
	args->size = args->pitch * args->height;
	args->size = roundup(args->size, PAGE_SIZE);

865
	ret = nouveau_gem_new(dev, args->size, 0, NOUVEAU_GEM_DOMAIN_VRAM, 0, 0, &bo);
B
Ben Skeggs 已提交
866 867 868
	if (ret)
		return ret;

869 870
	ret = drm_gem_handle_create(file_priv, &bo->gem, &args->handle);
	drm_gem_object_unreference_unlocked(&bo->gem);
B
Ben Skeggs 已提交
871 872 873 874 875 876 877 878 879 880 881 882
	return ret;
}

int
nouveau_display_dumb_map_offset(struct drm_file *file_priv,
				struct drm_device *dev,
				uint32_t handle, uint64_t *poffset)
{
	struct drm_gem_object *gem;

	gem = drm_gem_object_lookup(dev, file_priv, handle);
	if (gem) {
883
		struct nouveau_bo *bo = nouveau_gem_object(gem);
884
		*poffset = drm_vma_node_offset_addr(&bo->bo.vma_node);
B
Ben Skeggs 已提交
885 886 887 888 889 890
		drm_gem_object_unreference_unlocked(gem);
		return 0;
	}

	return -ENOENT;
}