kfd_device_queue_manager.c 32.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
/*
 * Copyright 2014 Advanced Micro Devices, Inc.
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice shall be included in
 * all copies or substantial portions of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
 * OTHER DEALINGS IN THE SOFTWARE.
 *
 */

#include <linux/slab.h>
#include <linux/list.h>
#include <linux/types.h>
#include <linux/printk.h>
#include <linux/bitops.h>
29
#include <linux/sched.h>
30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45
#include "kfd_priv.h"
#include "kfd_device_queue_manager.h"
#include "kfd_mqd_manager.h"
#include "cik_regs.h"
#include "kfd_kernel_queue.h"

/* Size of the per-pipe EOP queue */
#define CIK_HPD_EOP_BYTES_LOG2 11
#define CIK_HPD_EOP_BYTES (1U << CIK_HPD_EOP_BYTES_LOG2)

static int set_pasid_vmid_mapping(struct device_queue_manager *dqm,
					unsigned int pasid, unsigned int vmid);

static int create_compute_queue_nocpsch(struct device_queue_manager *dqm,
					struct queue *q,
					struct qcm_process_device *qpd);
46

47 48 49
static int execute_queues_cpsch(struct device_queue_manager *dqm,
				enum kfd_unmap_queues_filter filter,
				uint32_t filter_param);
50
static int unmap_queues_cpsch(struct device_queue_manager *dqm,
51 52
				enum kfd_unmap_queues_filter filter,
				uint32_t filter_param);
53

F
Felix Kuehling 已提交
54 55
static int map_queues_cpsch(struct device_queue_manager *dqm);

56 57 58 59 60 61
static int create_sdma_queue_nocpsch(struct device_queue_manager *dqm,
					struct queue *q,
					struct qcm_process_device *qpd);

static void deallocate_sdma_queue(struct device_queue_manager *dqm,
				unsigned int sdma_queue_id);
62

63 64
static inline
enum KFD_MQD_TYPE get_mqd_type_from_queue_type(enum kfd_queue_type type)
65
{
66
	if (type == KFD_QUEUE_TYPE_SDMA)
67 68
		return KFD_MQD_TYPE_SDMA;
	return KFD_MQD_TYPE_CP;
69 70
}

71 72 73 74 75 76 77 78 79 80 81 82 83 84 85
static bool is_pipe_enabled(struct device_queue_manager *dqm, int mec, int pipe)
{
	int i;
	int pipe_offset = mec * dqm->dev->shared_resources.num_pipe_per_mec
		+ pipe * dqm->dev->shared_resources.num_queue_per_pipe;

	/* queue is available for KFD usage if bit is 1 */
	for (i = 0; i <  dqm->dev->shared_resources.num_queue_per_pipe; ++i)
		if (test_bit(pipe_offset + i,
			      dqm->dev->shared_resources.queue_bitmap))
			return true;
	return false;
}

unsigned int get_queues_num(struct device_queue_manager *dqm)
86
{
87 88
	return bitmap_weight(dqm->dev->shared_resources.queue_bitmap,
				KGD_MAX_QUEUES);
89 90
}

91
unsigned int get_queues_per_pipe(struct device_queue_manager *dqm)
92
{
93 94 95 96 97 98
	return dqm->dev->shared_resources.num_queue_per_pipe;
}

unsigned int get_pipes_per_mec(struct device_queue_manager *dqm)
{
	return dqm->dev->shared_resources.num_pipe_per_mec;
99 100
}

101
void program_sh_mem_settings(struct device_queue_manager *dqm,
102 103
					struct qcm_process_device *qpd)
{
104 105
	return dqm->dev->kfd2kgd->program_sh_mem_settings(
						dqm->dev->kgd, qpd->vmid,
106 107 108 109 110 111 112 113 114 115 116 117 118 119 120
						qpd->sh_mem_config,
						qpd->sh_mem_ape1_base,
						qpd->sh_mem_ape1_limit,
						qpd->sh_mem_bases);
}

static int allocate_vmid(struct device_queue_manager *dqm,
			struct qcm_process_device *qpd,
			struct queue *q)
{
	int bit, allocated_vmid;

	if (dqm->vmid_bitmap == 0)
		return -ENOMEM;

121 122
	bit = find_first_bit((unsigned long *)&dqm->vmid_bitmap,
				dqm->dev->vm_info.vmid_num_kfd);
123 124
	clear_bit(bit, (unsigned long *)&dqm->vmid_bitmap);

125
	allocated_vmid = bit + dqm->dev->vm_info.first_vmid_kfd;
126
	pr_debug("vmid allocation %d\n", allocated_vmid);
127 128 129 130 131 132 133 134 135 136 137 138 139
	qpd->vmid = allocated_vmid;
	q->properties.vmid = allocated_vmid;

	set_pasid_vmid_mapping(dqm, q->process->pasid, q->properties.vmid);
	program_sh_mem_settings(dqm, qpd);

	return 0;
}

static void deallocate_vmid(struct device_queue_manager *dqm,
				struct qcm_process_device *qpd,
				struct queue *q)
{
140
	int bit = qpd->vmid - dqm->dev->vm_info.first_vmid_kfd;
141

142 143 144
	/* Release the vmid mapping */
	set_pasid_vmid_mapping(dqm, 0, qpd->vmid);

145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160
	set_bit(bit, (unsigned long *)&dqm->vmid_bitmap);
	qpd->vmid = 0;
	q->properties.vmid = 0;
}

static int create_queue_nocpsch(struct device_queue_manager *dqm,
				struct queue *q,
				struct qcm_process_device *qpd,
				int *allocated_vmid)
{
	int retval;

	print_queue(q);

	mutex_lock(&dqm->lock);

161
	if (dqm->total_queue_count >= max_num_of_queues_per_device) {
162
		pr_warn("Can't create new usermode queue because %d queues were already created\n",
163
				dqm->total_queue_count);
K
Kent Russell 已提交
164 165
		retval = -EPERM;
		goto out_unlock;
166 167
	}

168 169
	if (list_empty(&qpd->queues_list)) {
		retval = allocate_vmid(dqm, qpd, q);
K
Kent Russell 已提交
170 171
		if (retval)
			goto out_unlock;
172 173 174 175
	}
	*allocated_vmid = qpd->vmid;
	q->properties.vmid = qpd->vmid;

F
Felix Kuehling 已提交
176 177 178
	q->properties.tba_addr = qpd->tba_addr;
	q->properties.tma_addr = qpd->tma_addr;

179 180
	if (q->properties.type == KFD_QUEUE_TYPE_COMPUTE)
		retval = create_compute_queue_nocpsch(dqm, q, qpd);
K
Kent Russell 已提交
181
	else if (q->properties.type == KFD_QUEUE_TYPE_SDMA)
182
		retval = create_sdma_queue_nocpsch(dqm, q, qpd);
K
Kent Russell 已提交
183 184
	else
		retval = -EINVAL;
185

186
	if (retval) {
187 188 189 190
		if (list_empty(&qpd->queues_list)) {
			deallocate_vmid(dqm, qpd, q);
			*allocated_vmid = 0;
		}
K
Kent Russell 已提交
191
		goto out_unlock;
192 193 194
	}

	list_add(&q->list, &qpd->queues_list);
195
	qpd->queue_count++;
196 197
	if (q->properties.is_active)
		dqm->queue_count++;
198

199 200
	if (q->properties.type == KFD_QUEUE_TYPE_SDMA)
		dqm->sdma_queue_count++;
201

202 203 204 205 206 207 208 209
	/*
	 * Unconditionally increment this counter, regardless of the queue's
	 * type or whether the queue is active.
	 */
	dqm->total_queue_count++;
	pr_debug("Total of %d queues are accountable so far\n",
			dqm->total_queue_count);

K
Kent Russell 已提交
210
out_unlock:
211
	mutex_unlock(&dqm->lock);
K
Kent Russell 已提交
212
	return retval;
213 214 215 216 217
}

static int allocate_hqd(struct device_queue_manager *dqm, struct queue *q)
{
	bool set;
218
	int pipe, bit, i;
219 220 221

	set = false;

222 223
	for (pipe = dqm->next_pipe_to_allocate, i = 0;
			i < get_pipes_per_mec(dqm);
224 225 226 227 228
			pipe = ((pipe + 1) % get_pipes_per_mec(dqm)), ++i) {

		if (!is_pipe_enabled(dqm, 0, pipe))
			continue;

229 230 231
		if (dqm->allocated_queues[pipe] != 0) {
			bit = find_first_bit(
				(unsigned long *)&dqm->allocated_queues[pipe],
232
				get_queues_per_pipe(dqm));
233 234 235 236 237 238 239 240 241 242

			clear_bit(bit,
				(unsigned long *)&dqm->allocated_queues[pipe]);
			q->pipe = pipe;
			q->queue = bit;
			set = true;
			break;
		}
	}

243
	if (!set)
244 245
		return -EBUSY;

246
	pr_debug("hqd slot - pipe %d, queue %d\n", q->pipe, q->queue);
247
	/* horizontal hqd allocation */
248
	dqm->next_pipe_to_allocate = (pipe + 1) % get_pipes_per_mec(dqm);
249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265

	return 0;
}

static inline void deallocate_hqd(struct device_queue_manager *dqm,
				struct queue *q)
{
	set_bit(q->queue, (unsigned long *)&dqm->allocated_queues[q->pipe]);
}

static int create_compute_queue_nocpsch(struct device_queue_manager *dqm,
					struct queue *q,
					struct qcm_process_device *qpd)
{
	int retval;
	struct mqd_manager *mqd;

266
	mqd = dqm->ops.get_mqd_manager(dqm, KFD_MQD_TYPE_COMPUTE);
267
	if (!mqd)
268 269 270
		return -ENOMEM;

	retval = allocate_hqd(dqm, q);
271
	if (retval)
272 273 274 275
		return retval;

	retval = mqd->init_mqd(mqd, &q->mqd, &q->mqd_mem_obj,
				&q->gart_mqd_addr, &q->properties);
K
Kent Russell 已提交
276 277
	if (retval)
		goto out_deallocate_hqd;
278

279 280
	pr_debug("Loading mqd to hqd on pipe %d, queue %d\n",
			q->pipe, q->queue);
281

282 283 284
	dqm->dev->kfd2kgd->set_scratch_backing_va(
			dqm->dev->kgd, qpd->sh_hidden_private_base, qpd->vmid);

F
Felix Kuehling 已提交
285 286 287
	if (!q->properties.is_active)
		return 0;

288 289
	retval = mqd->load_mqd(mqd, q->mqd, q->pipe, q->queue, &q->properties,
			       q->process->mm);
K
Kent Russell 已提交
290 291
	if (retval)
		goto out_uninit_mqd;
292

293
	return 0;
K
Kent Russell 已提交
294 295 296 297 298 299 300

out_uninit_mqd:
	mqd->uninit_mqd(mqd, q->mqd, q->mqd_mem_obj);
out_deallocate_hqd:
	deallocate_hqd(dqm, q);

	return retval;
301 302
}

303 304 305 306
/* Access to DQM has to be locked before calling destroy_queue_nocpsch_locked
 * to avoid asynchronized access
 */
static int destroy_queue_nocpsch_locked(struct device_queue_manager *dqm,
307 308 309 310 311 312
				struct qcm_process_device *qpd,
				struct queue *q)
{
	int retval;
	struct mqd_manager *mqd;

313 314 315 316
	mqd = dqm->ops.get_mqd_manager(dqm,
		get_mqd_type_from_queue_type(q->properties.type));
	if (!mqd)
		return -ENOMEM;
317

318 319 320 321 322
	if (q->properties.type == KFD_QUEUE_TYPE_COMPUTE) {
		deallocate_hqd(dqm, q);
	} else if (q->properties.type == KFD_QUEUE_TYPE_SDMA) {
		dqm->sdma_queue_count--;
		deallocate_sdma_queue(dqm, q->sdma_id);
323
	} else {
324
		pr_debug("q->properties.type %d is invalid\n",
325
				q->properties.type);
326
		return -EINVAL;
327
	}
328
	dqm->total_queue_count--;
329 330

	retval = mqd->destroy_mqd(mqd, q->mqd,
331
				KFD_PREEMPT_TYPE_WAVEFRONT_RESET,
332
				KFD_UNMAP_LATENCY_MS,
333
				q->pipe, q->queue);
334 335
	if (retval == -ETIME)
		qpd->reset_wavefronts = true;
336 337 338 339

	mqd->uninit_mqd(mqd, q->mqd, q->mqd_mem_obj);

	list_del(&q->list);
340 341 342 343 344 345 346 347 348 349 350 351
	if (list_empty(&qpd->queues_list)) {
		if (qpd->reset_wavefronts) {
			pr_warn("Resetting wave fronts (nocpsch) on dev %p\n",
					dqm->dev);
			/* dbgdev_wave_reset_wavefronts has to be called before
			 * deallocate_vmid(), i.e. when vmid is still in use.
			 */
			dbgdev_wave_reset_wavefronts(dqm->dev,
					qpd->pqm->process);
			qpd->reset_wavefronts = false;
		}

352
		deallocate_vmid(dqm, qpd, q);
353
	}
354
	qpd->queue_count--;
355 356
	if (q->properties.is_active)
		dqm->queue_count--;
357

358 359
	return retval;
}
360

361 362 363 364 365 366 367 368
static int destroy_queue_nocpsch(struct device_queue_manager *dqm,
				struct qcm_process_device *qpd,
				struct queue *q)
{
	int retval;

	mutex_lock(&dqm->lock);
	retval = destroy_queue_nocpsch_locked(dqm, qpd, q);
369
	mutex_unlock(&dqm->lock);
370

371 372 373 374 375 376 377
	return retval;
}

static int update_queue(struct device_queue_manager *dqm, struct queue *q)
{
	int retval;
	struct mqd_manager *mqd;
378
	bool prev_active = false;
379 380

	mutex_lock(&dqm->lock);
O
Oded Gabbay 已提交
381 382
	mqd = dqm->ops.get_mqd_manager(dqm,
			get_mqd_type_from_queue_type(q->properties.type));
383
	if (!mqd) {
K
Kent Russell 已提交
384 385
		retval = -ENOMEM;
		goto out_unlock;
386 387
	}

F
Felix Kuehling 已提交
388 389 390 391 392 393 394
	/* Save previous activity state for counters */
	prev_active = q->properties.is_active;

	/* Make sure the queue is unmapped before updating the MQD */
	if (sched_policy != KFD_SCHED_POLICY_NO_HWS) {
		retval = unmap_queues_cpsch(dqm,
				KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0);
F
Felix Kuehling 已提交
395
		if (retval) {
F
Felix Kuehling 已提交
396 397 398
			pr_err("unmap queue failed\n");
			goto out_unlock;
		}
F
Felix Kuehling 已提交
399
	} else if (prev_active &&
F
Felix Kuehling 已提交
400 401 402 403 404 405 406 407 408 409 410 411 412
		   (q->properties.type == KFD_QUEUE_TYPE_COMPUTE ||
		    q->properties.type == KFD_QUEUE_TYPE_SDMA)) {
		retval = mqd->destroy_mqd(mqd, q->mqd,
				KFD_PREEMPT_TYPE_WAVEFRONT_DRAIN,
				KFD_UNMAP_LATENCY_MS, q->pipe, q->queue);
		if (retval) {
			pr_err("destroy mqd failed\n");
			goto out_unlock;
		}
	}

	retval = mqd->update_mqd(mqd, q->mqd, &q->properties);

413 414 415 416 417 418 419 420 421 422 423
	/*
	 * check active state vs. the previous state and modify
	 * counter accordingly. map_queues_cpsch uses the
	 * dqm->queue_count to determine whether a new runlist must be
	 * uploaded.
	 */
	if (q->properties.is_active && !prev_active)
		dqm->queue_count++;
	else if (!q->properties.is_active && prev_active)
		dqm->queue_count--;

F
Felix Kuehling 已提交
424 425
	if (sched_policy != KFD_SCHED_POLICY_NO_HWS)
		retval = map_queues_cpsch(dqm);
F
Felix Kuehling 已提交
426
	else if (q->properties.is_active &&
F
Felix Kuehling 已提交
427 428 429 430
		 (q->properties.type == KFD_QUEUE_TYPE_COMPUTE ||
		  q->properties.type == KFD_QUEUE_TYPE_SDMA))
		retval = mqd->load_mqd(mqd, q->mqd, q->pipe, q->queue,
				       &q->properties, q->process->mm);
431

K
Kent Russell 已提交
432
out_unlock:
433 434 435 436
	mutex_unlock(&dqm->lock);
	return retval;
}

437
static struct mqd_manager *get_mqd_manager(
438 439 440 441
		struct device_queue_manager *dqm, enum KFD_MQD_TYPE type)
{
	struct mqd_manager *mqd;

442 443
	if (WARN_ON(type >= KFD_MQD_TYPE_MAX))
		return NULL;
444

445
	pr_debug("mqd type %d\n", type);
446 447 448 449

	mqd = dqm->mqds[type];
	if (!mqd) {
		mqd = mqd_manager_init(type, dqm->dev);
450
		if (!mqd)
451
			pr_err("mqd manager is NULL");
452 453 454 455 456 457
		dqm->mqds[type] = mqd;
	}

	return mqd;
}

458
static int register_process(struct device_queue_manager *dqm,
459 460 461
					struct qcm_process_device *qpd)
{
	struct device_process_node *n;
462
	int retval;
463

464
	n = kzalloc(sizeof(*n), GFP_KERNEL);
465 466 467 468 469 470 471 472
	if (!n)
		return -ENOMEM;

	n->qpd = qpd;

	mutex_lock(&dqm->lock);
	list_add(&n->list, &dqm->queues);

473
	retval = dqm->asic_ops.update_qpd(dqm, qpd);
474

475 476 477 478
	dqm->processes_count++;

	mutex_unlock(&dqm->lock);

479
	return retval;
480 481
}

482
static int unregister_process(struct device_queue_manager *dqm,
483 484 485 486 487
					struct qcm_process_device *qpd)
{
	int retval;
	struct device_process_node *cur, *next;

488 489
	pr_debug("qpd->queues_list is %s\n",
			list_empty(&qpd->queues_list) ? "empty" : "not empty");
490 491 492 493 494 495 496

	retval = 0;
	mutex_lock(&dqm->lock);

	list_for_each_entry_safe(cur, next, &dqm->queues, list) {
		if (qpd == cur->qpd) {
			list_del(&cur->list);
497
			kfree(cur);
498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514
			dqm->processes_count--;
			goto out;
		}
	}
	/* qpd not found in dqm list */
	retval = 1;
out:
	mutex_unlock(&dqm->lock);
	return retval;
}

static int
set_pasid_vmid_mapping(struct device_queue_manager *dqm, unsigned int pasid,
			unsigned int vmid)
{
	uint32_t pasid_mapping;

515 516 517 518 519 520
	pasid_mapping = (pasid == 0) ? 0 :
		(uint32_t)pasid |
		ATC_VMID_PASID_MAPPING_VALID;

	return dqm->dev->kfd2kgd->set_pasid_vmid_mapping(
						dqm->dev->kgd, pasid_mapping,
521 522 523
						vmid);
}

524 525 526 527
static void init_interrupts(struct device_queue_manager *dqm)
{
	unsigned int i;

528 529 530
	for (i = 0 ; i < get_pipes_per_mec(dqm) ; i++)
		if (is_pipe_enabled(dqm, 0, i))
			dqm->dev->kfd2kgd->init_interrupts(dqm->dev->kgd, i);
531 532
}

533 534
static int initialize_nocpsch(struct device_queue_manager *dqm)
{
535
	int pipe, queue;
536

537
	pr_debug("num of pipes: %d\n", get_pipes_per_mec(dqm));
538

K
Kent Russell 已提交
539 540 541 542 543
	dqm->allocated_queues = kcalloc(get_pipes_per_mec(dqm),
					sizeof(unsigned int), GFP_KERNEL);
	if (!dqm->allocated_queues)
		return -ENOMEM;

544 545 546
	mutex_init(&dqm->lock);
	INIT_LIST_HEAD(&dqm->queues);
	dqm->queue_count = dqm->next_pipe_to_allocate = 0;
547
	dqm->sdma_queue_count = 0;
548

549 550 551 552 553 554 555 556
	for (pipe = 0; pipe < get_pipes_per_mec(dqm); pipe++) {
		int pipe_offset = pipe * get_queues_per_pipe(dqm);

		for (queue = 0; queue < get_queues_per_pipe(dqm); queue++)
			if (test_bit(pipe_offset + queue,
				     dqm->dev->shared_resources.queue_bitmap))
				dqm->allocated_queues[pipe] |= 1 << queue;
	}
557

558
	dqm->vmid_bitmap = (1 << dqm->dev->vm_info.vmid_num_kfd) - 1;
559
	dqm->sdma_bitmap = (1 << CIK_SDMA_QUEUES) - 1;
560 561 562 563

	return 0;
}

564
static void uninitialize(struct device_queue_manager *dqm)
565
{
566 567
	int i;

568
	WARN_ON(dqm->queue_count > 0 || dqm->processes_count > 0);
569 570

	kfree(dqm->allocated_queues);
571 572
	for (i = 0 ; i < KFD_MQD_TYPE_MAX ; i++)
		kfree(dqm->mqds[i]);
573
	mutex_destroy(&dqm->lock);
574
	kfd_gtt_sa_free(dqm->dev, dqm->pipeline_mem);
575 576 577 578
}

static int start_nocpsch(struct device_queue_manager *dqm)
{
579
	init_interrupts(dqm);
580 581 582 583 584 585 586 587
	return 0;
}

static int stop_nocpsch(struct device_queue_manager *dqm)
{
	return 0;
}

588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607
static int allocate_sdma_queue(struct device_queue_manager *dqm,
				unsigned int *sdma_queue_id)
{
	int bit;

	if (dqm->sdma_bitmap == 0)
		return -ENOMEM;

	bit = find_first_bit((unsigned long *)&dqm->sdma_bitmap,
				CIK_SDMA_QUEUES);

	clear_bit(bit, (unsigned long *)&dqm->sdma_bitmap);
	*sdma_queue_id = bit;

	return 0;
}

static void deallocate_sdma_queue(struct device_queue_manager *dqm,
				unsigned int sdma_queue_id)
{
608
	if (sdma_queue_id >= CIK_SDMA_QUEUES)
609 610 611 612 613 614 615 616 617 618 619
		return;
	set_bit(sdma_queue_id, (unsigned long *)&dqm->sdma_bitmap);
}

static int create_sdma_queue_nocpsch(struct device_queue_manager *dqm,
					struct queue *q,
					struct qcm_process_device *qpd)
{
	struct mqd_manager *mqd;
	int retval;

620
	mqd = dqm->ops.get_mqd_manager(dqm, KFD_MQD_TYPE_SDMA);
621 622 623 624
	if (!mqd)
		return -ENOMEM;

	retval = allocate_sdma_queue(dqm, &q->sdma_id);
625
	if (retval)
626 627
		return retval;

628 629
	q->properties.sdma_queue_id = q->sdma_id / CIK_SDMA_QUEUES_PER_ENGINE;
	q->properties.sdma_engine_id = q->sdma_id % CIK_SDMA_QUEUES_PER_ENGINE;
630

631 632 633
	pr_debug("SDMA id is:    %d\n", q->sdma_id);
	pr_debug("SDMA queue id: %d\n", q->properties.sdma_queue_id);
	pr_debug("SDMA engine id: %d\n", q->properties.sdma_engine_id);
634

635
	dqm->asic_ops.init_sdma_vm(dqm, q, qpd);
636 637
	retval = mqd->init_mqd(mqd, &q->mqd, &q->mqd_mem_obj,
				&q->gart_mqd_addr, &q->properties);
K
Kent Russell 已提交
638 639
	if (retval)
		goto out_deallocate_sdma_queue;
640

641
	retval = mqd->load_mqd(mqd, q->mqd, 0, 0, &q->properties, NULL);
K
Kent Russell 已提交
642 643
	if (retval)
		goto out_uninit_mqd;
644

645
	return 0;
K
Kent Russell 已提交
646 647 648 649 650 651 652

out_uninit_mqd:
	mqd->uninit_mqd(mqd, q->mqd, q->mqd_mem_obj);
out_deallocate_sdma_queue:
	deallocate_sdma_queue(dqm, q->sdma_id);

	return retval;
653 654
}

655 656 657 658 659 660
/*
 * Device Queue Manager implementation for cp scheduler
 */

static int set_sched_resources(struct device_queue_manager *dqm)
{
661
	int i, mec;
662 663
	struct scheduling_resources res;

664
	res.vmid_mask = dqm->dev->shared_resources.compute_vmid_bitmap;
665 666 667 668 669 670 671 672 673 674 675 676 677 678 679

	res.queue_mask = 0;
	for (i = 0; i < KGD_MAX_QUEUES; ++i) {
		mec = (i / dqm->dev->shared_resources.num_queue_per_pipe)
			/ dqm->dev->shared_resources.num_pipe_per_mec;

		if (!test_bit(i, dqm->dev->shared_resources.queue_bitmap))
			continue;

		/* only acquire queues from the first MEC */
		if (mec > 0)
			continue;

		/* This situation may be hit in the future if a new HW
		 * generation exposes more than 64 queues. If so, the
680 681
		 * definition of res.queue_mask needs updating
		 */
682
		if (WARN_ON(i >= (sizeof(res.queue_mask)*8))) {
683 684 685 686 687 688
			pr_err("Invalid queue enabled by amdgpu: %d\n", i);
			break;
		}

		res.queue_mask |= (1ull << i);
	}
689 690 691
	res.gws_mask = res.oac_mask = res.gds_heap_base =
						res.gds_heap_size = 0;

692 693 694
	pr_debug("Scheduling resources:\n"
			"vmid mask: 0x%8X\n"
			"queue mask: 0x%8llX\n",
695 696 697 698 699 700 701
			res.vmid_mask, res.queue_mask);

	return pm_send_set_resources(&dqm->packets, &res);
}

static int initialize_cpsch(struct device_queue_manager *dqm)
{
702
	pr_debug("num of pipes: %d\n", get_pipes_per_mec(dqm));
703 704 705 706

	mutex_init(&dqm->lock);
	INIT_LIST_HEAD(&dqm->queues);
	dqm->queue_count = dqm->processes_count = 0;
707
	dqm->sdma_queue_count = 0;
708
	dqm->active_runlist = false;
709
	dqm->sdma_bitmap = (1 << CIK_SDMA_QUEUES) - 1;
710

711
	return 0;
712 713 714 715 716 717 718 719 720
}

static int start_cpsch(struct device_queue_manager *dqm)
{
	int retval;

	retval = 0;

	retval = pm_init(&dqm->packets, dqm);
721
	if (retval)
722 723 724
		goto fail_packet_manager_init;

	retval = set_sched_resources(dqm);
725
	if (retval)
726 727
		goto fail_set_sched_resources;

728
	pr_debug("Allocating fence memory\n");
729 730

	/* allocate fence memory on the gart */
731 732
	retval = kfd_gtt_sa_allocate(dqm->dev, sizeof(*dqm->fence_addr),
					&dqm->fence_mem);
733

734
	if (retval)
735 736 737 738
		goto fail_allocate_vidmem;

	dqm->fence_addr = dqm->fence_mem->cpu_ptr;
	dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr;
739 740 741

	init_interrupts(dqm);

742
	mutex_lock(&dqm->lock);
743
	execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0);
744
	mutex_unlock(&dqm->lock);
745 746 747 748 749 750 751 752 753 754 755

	return 0;
fail_allocate_vidmem:
fail_set_sched_resources:
	pm_uninit(&dqm->packets);
fail_packet_manager_init:
	return retval;
}

static int stop_cpsch(struct device_queue_manager *dqm)
{
756
	mutex_lock(&dqm->lock);
757
	unmap_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES, 0);
758
	mutex_unlock(&dqm->lock);
759

760
	kfd_gtt_sa_free(dqm->dev, dqm->fence_mem);
761 762 763 764 765 766 767 768 769 770
	pm_uninit(&dqm->packets);

	return 0;
}

static int create_kernel_queue_cpsch(struct device_queue_manager *dqm,
					struct kernel_queue *kq,
					struct qcm_process_device *qpd)
{
	mutex_lock(&dqm->lock);
771
	if (dqm->total_queue_count >= max_num_of_queues_per_device) {
772
		pr_warn("Can't create new kernel queue because %d queues were already created\n",
773 774 775 776 777 778 779 780 781 782 783 784 785
				dqm->total_queue_count);
		mutex_unlock(&dqm->lock);
		return -EPERM;
	}

	/*
	 * Unconditionally increment this counter, regardless of the queue's
	 * type or whether the queue is active.
	 */
	dqm->total_queue_count++;
	pr_debug("Total of %d queues are accountable so far\n",
			dqm->total_queue_count);

786 787 788
	list_add(&kq->list, &qpd->priv_queue_list);
	dqm->queue_count++;
	qpd->is_debug = true;
789
	execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0);
790 791 792 793 794 795 796 797 798 799 800 801 802
	mutex_unlock(&dqm->lock);

	return 0;
}

static void destroy_kernel_queue_cpsch(struct device_queue_manager *dqm,
					struct kernel_queue *kq,
					struct qcm_process_device *qpd)
{
	mutex_lock(&dqm->lock);
	list_del(&kq->list);
	dqm->queue_count--;
	qpd->is_debug = false;
803
	execute_queues_cpsch(dqm, KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES, 0);
804 805 806 807
	/*
	 * Unconditionally decrement this counter, regardless of the queue's
	 * type.
	 */
808
	dqm->total_queue_count--;
809 810
	pr_debug("Total of %d queues are accountable so far\n",
			dqm->total_queue_count);
811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826
	mutex_unlock(&dqm->lock);
}

static int create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q,
			struct qcm_process_device *qpd, int *allocate_vmid)
{
	int retval;
	struct mqd_manager *mqd;

	retval = 0;

	if (allocate_vmid)
		*allocate_vmid = 0;

	mutex_lock(&dqm->lock);

827
	if (dqm->total_queue_count >= max_num_of_queues_per_device) {
828
		pr_warn("Can't create new usermode queue because %d queues were already created\n",
829 830 831 832 833
				dqm->total_queue_count);
		retval = -EPERM;
		goto out;
	}

834 835
	if (q->properties.type == KFD_QUEUE_TYPE_SDMA) {
		retval = allocate_sdma_queue(dqm, &q->sdma_id);
F
Felix Kuehling 已提交
836
		if (retval)
837 838 839 840 841 842
			goto out;
		q->properties.sdma_queue_id =
			q->sdma_id / CIK_SDMA_QUEUES_PER_ENGINE;
		q->properties.sdma_engine_id =
			q->sdma_id % CIK_SDMA_QUEUES_PER_ENGINE;
	}
843
	mqd = dqm->ops.get_mqd_manager(dqm,
844 845
			get_mqd_type_from_queue_type(q->properties.type));

846
	if (!mqd) {
K
Kent Russell 已提交
847 848
		retval = -ENOMEM;
		goto out;
849 850
	}

851
	dqm->asic_ops.init_sdma_vm(dqm, q, qpd);
F
Felix Kuehling 已提交
852 853 854

	q->properties.tba_addr = qpd->tba_addr;
	q->properties.tma_addr = qpd->tma_addr;
855 856
	retval = mqd->init_mqd(mqd, &q->mqd, &q->mqd_mem_obj,
				&q->gart_mqd_addr, &q->properties);
857
	if (retval)
858 859 860
		goto out;

	list_add(&q->list, &qpd->queues_list);
861
	qpd->queue_count++;
862 863
	if (q->properties.is_active) {
		dqm->queue_count++;
864 865
		retval = execute_queues_cpsch(dqm,
				KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0);
866 867
	}

868
	if (q->properties.type == KFD_QUEUE_TYPE_SDMA)
869
		dqm->sdma_queue_count++;
870 871 872 873 874 875 876 877 878
	/*
	 * Unconditionally increment this counter, regardless of the queue's
	 * type or whether the queue is active.
	 */
	dqm->total_queue_count++;

	pr_debug("Total of %d queues are accountable so far\n",
			dqm->total_queue_count);

879 880 881 882 883
out:
	mutex_unlock(&dqm->lock);
	return retval;
}

884
int amdkfd_fence_wait_timeout(unsigned int *fence_addr,
885
				unsigned int fence_value,
886
				unsigned int timeout_ms)
887
{
888
	unsigned long end_jiffies = msecs_to_jiffies(timeout_ms) + jiffies;
889 890

	while (*fence_addr != fence_value) {
891
		if (time_after(jiffies, end_jiffies)) {
892
			pr_err("qcm fence wait loop timeout expired\n");
893 894
			return -ETIME;
		}
895
		schedule();
896 897 898 899 900
	}

	return 0;
}

901
static int unmap_sdma_queues(struct device_queue_manager *dqm,
902 903 904
				unsigned int sdma_engine)
{
	return pm_send_unmap_queue(&dqm->packets, KFD_QUEUE_TYPE_SDMA,
905
			KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0, false,
906 907 908
			sdma_engine);
}

F
Felix Kuehling 已提交
909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929
/* dqm->lock mutex has to be locked before calling this function */
static int map_queues_cpsch(struct device_queue_manager *dqm)
{
	int retval;

	if (dqm->queue_count <= 0 || dqm->processes_count <= 0)
		return 0;

	if (dqm->active_runlist)
		return 0;

	retval = pm_send_runlist(&dqm->packets, &dqm->queues);
	if (retval) {
		pr_err("failed to execute runlist\n");
		return retval;
	}
	dqm->active_runlist = true;

	return retval;
}

930
/* dqm->lock mutex has to be locked before calling this function */
931
static int unmap_queues_cpsch(struct device_queue_manager *dqm,
932 933
				enum kfd_unmap_queues_filter filter,
				uint32_t filter_param)
934
{
935
	int retval = 0;
936

937
	if (!dqm->active_runlist)
938
		return retval;
939

940
	pr_debug("Before destroying queues, sdma queue count is : %u\n",
941 942 943
		dqm->sdma_queue_count);

	if (dqm->sdma_queue_count > 0) {
944 945
		unmap_sdma_queues(dqm, 0);
		unmap_sdma_queues(dqm, 1);
946 947
	}

948
	retval = pm_send_unmap_queue(&dqm->packets, KFD_QUEUE_TYPE_COMPUTE,
949
			filter, filter_param, false, 0);
950
	if (retval)
951
		return retval;
952 953 954 955 956

	*dqm->fence_addr = KFD_FENCE_INIT;
	pm_send_query_status(&dqm->packets, dqm->fence_gpu_addr,
				KFD_FENCE_COMPLETED);
	/* should be timed out */
957
	retval = amdkfd_fence_wait_timeout(dqm->fence_addr, KFD_FENCE_COMPLETED,
958
				QUEUE_PREEMPT_DEFAULT_TIMEOUT_MS);
959
	if (retval)
960
		return retval;
961

962 963 964 965 966 967
	pm_release_ib(&dqm->packets);
	dqm->active_runlist = false;

	return retval;
}

968
/* dqm->lock mutex has to be locked before calling this function */
969 970 971
static int execute_queues_cpsch(struct device_queue_manager *dqm,
				enum kfd_unmap_queues_filter filter,
				uint32_t filter_param)
972 973 974
{
	int retval;

975
	retval = unmap_queues_cpsch(dqm, filter, filter_param);
976
	if (retval) {
977
		pr_err("The cp might be in an unrecoverable state due to an unsuccessful queues preemption\n");
978
		return retval;
979 980
	}

F
Felix Kuehling 已提交
981
	return map_queues_cpsch(dqm);
982 983 984 985 986 987 988 989
}

static int destroy_queue_cpsch(struct device_queue_manager *dqm,
				struct qcm_process_device *qpd,
				struct queue *q)
{
	int retval;
	struct mqd_manager *mqd;
990
	bool preempt_all_queues;
991

992 993
	preempt_all_queues = false;

994 995 996 997
	retval = 0;

	/* remove queue from list to prevent rescheduling after preemption */
	mutex_lock(&dqm->lock);
998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008

	if (qpd->is_debug) {
		/*
		 * error, currently we do not allow to destroy a queue
		 * of a currently debugged process
		 */
		retval = -EBUSY;
		goto failed_try_destroy_debugged_queue;

	}

1009
	mqd = dqm->ops.get_mqd_manager(dqm,
1010
			get_mqd_type_from_queue_type(q->properties.type));
1011 1012 1013 1014 1015
	if (!mqd) {
		retval = -ENOMEM;
		goto failed;
	}

1016
	if (q->properties.type == KFD_QUEUE_TYPE_SDMA) {
1017
		dqm->sdma_queue_count--;
1018 1019
		deallocate_sdma_queue(dqm, q->sdma_id);
	}
1020

1021
	list_del(&q->list);
1022
	qpd->queue_count--;
1023 1024
	if (q->properties.is_active)
		dqm->queue_count--;
1025

1026 1027 1028 1029
	retval = execute_queues_cpsch(dqm,
				KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES, 0);
	if (retval == -ETIME)
		qpd->reset_wavefronts = true;
1030 1031

	mqd->uninit_mqd(mqd, q->mqd, q->mqd_mem_obj);
1032 1033 1034 1035 1036 1037 1038 1039

	/*
	 * Unconditionally decrement this counter, regardless of the queue's
	 * type
	 */
	dqm->total_queue_count--;
	pr_debug("Total of %d queues are accountable so far\n",
			dqm->total_queue_count);
1040 1041 1042 1043 1044 1045

	mutex_unlock(&dqm->lock);

	return 0;

failed:
1046 1047
failed_try_destroy_debugged_queue:

1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066
	mutex_unlock(&dqm->lock);
	return retval;
}

/*
 * Low bits must be 0000/FFFF as required by HW, high bits must be 0 to
 * stay in user mode.
 */
#define APE1_FIXED_BITS_MASK 0xFFFF80000000FFFFULL
/* APE1 limit is inclusive and 64K aligned. */
#define APE1_LIMIT_ALIGNMENT 0xFFFF

static bool set_cache_memory_policy(struct device_queue_manager *dqm,
				   struct qcm_process_device *qpd,
				   enum cache_policy default_policy,
				   enum cache_policy alternate_policy,
				   void __user *alternate_aperture_base,
				   uint64_t alternate_aperture_size)
{
1067
	bool retval;
1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088

	mutex_lock(&dqm->lock);

	if (alternate_aperture_size == 0) {
		/* base > limit disables APE1 */
		qpd->sh_mem_ape1_base = 1;
		qpd->sh_mem_ape1_limit = 0;
	} else {
		/*
		 * In FSA64, APE1_Base[63:0] = { 16{SH_MEM_APE1_BASE[31]},
		 *			SH_MEM_APE1_BASE[31:0], 0x0000 }
		 * APE1_Limit[63:0] = { 16{SH_MEM_APE1_LIMIT[31]},
		 *			SH_MEM_APE1_LIMIT[31:0], 0xFFFF }
		 * Verify that the base and size parameters can be
		 * represented in this format and convert them.
		 * Additionally restrict APE1 to user-mode addresses.
		 */

		uint64_t base = (uintptr_t)alternate_aperture_base;
		uint64_t limit = base + alternate_aperture_size - 1;

K
Kent Russell 已提交
1089 1090 1091
		if (limit <= base || (base & APE1_FIXED_BITS_MASK) != 0 ||
		   (limit & APE1_FIXED_BITS_MASK) != APE1_LIMIT_ALIGNMENT) {
			retval = false;
1092
			goto out;
K
Kent Russell 已提交
1093
		}
1094 1095 1096 1097 1098

		qpd->sh_mem_ape1_base = base >> 16;
		qpd->sh_mem_ape1_limit = limit >> 16;
	}

1099
	retval = dqm->asic_ops.set_cache_memory_policy(
1100 1101 1102 1103 1104 1105
			dqm,
			qpd,
			default_policy,
			alternate_policy,
			alternate_aperture_base,
			alternate_aperture_size);
1106 1107 1108 1109

	if ((sched_policy == KFD_SCHED_POLICY_NO_HWS) && (qpd->vmid != 0))
		program_sh_mem_settings(dqm, qpd);

1110
	pr_debug("sh_mem_config: 0x%x, ape1_base: 0x%x, ape1_limit: 0x%x\n",
1111 1112 1113 1114 1115
		qpd->sh_mem_config, qpd->sh_mem_ape1_base,
		qpd->sh_mem_ape1_limit);

out:
	mutex_unlock(&dqm->lock);
K
Kent Russell 已提交
1116
	return retval;
1117 1118
}

1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138
static int set_trap_handler(struct device_queue_manager *dqm,
				struct qcm_process_device *qpd,
				uint64_t tba_addr,
				uint64_t tma_addr)
{
	uint64_t *tma;

	if (dqm->dev->cwsr_enabled) {
		/* Jump from CWSR trap handler to user trap */
		tma = (uint64_t *)(qpd->cwsr_kaddr + KFD_CWSR_TMA_OFFSET);
		tma[0] = tba_addr;
		tma[1] = tma_addr;
	} else {
		qpd->tba_addr = tba_addr;
		qpd->tma_addr = tma_addr;
	}

	return 0;
}

1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232
static int process_termination_nocpsch(struct device_queue_manager *dqm,
		struct qcm_process_device *qpd)
{
	struct queue *q, *next;
	struct device_process_node *cur, *next_dpn;
	int retval = 0;

	mutex_lock(&dqm->lock);

	/* Clear all user mode queues */
	list_for_each_entry_safe(q, next, &qpd->queues_list, list) {
		int ret;

		ret = destroy_queue_nocpsch_locked(dqm, qpd, q);
		if (ret)
			retval = ret;
	}

	/* Unregister process */
	list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) {
		if (qpd == cur->qpd) {
			list_del(&cur->list);
			kfree(cur);
			dqm->processes_count--;
			break;
		}
	}

	mutex_unlock(&dqm->lock);
	return retval;
}


static int process_termination_cpsch(struct device_queue_manager *dqm,
		struct qcm_process_device *qpd)
{
	int retval;
	struct queue *q, *next;
	struct kernel_queue *kq, *kq_next;
	struct mqd_manager *mqd;
	struct device_process_node *cur, *next_dpn;
	enum kfd_unmap_queues_filter filter =
		KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES;

	retval = 0;

	mutex_lock(&dqm->lock);

	/* Clean all kernel queues */
	list_for_each_entry_safe(kq, kq_next, &qpd->priv_queue_list, list) {
		list_del(&kq->list);
		dqm->queue_count--;
		qpd->is_debug = false;
		dqm->total_queue_count--;
		filter = KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES;
	}

	/* Clear all user mode queues */
	list_for_each_entry(q, &qpd->queues_list, list) {
		if (q->properties.type == KFD_QUEUE_TYPE_SDMA)
			dqm->sdma_queue_count--;

		if (q->properties.is_active)
			dqm->queue_count--;

		dqm->total_queue_count--;
	}

	/* Unregister process */
	list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) {
		if (qpd == cur->qpd) {
			list_del(&cur->list);
			kfree(cur);
			dqm->processes_count--;
			break;
		}
	}

	retval = execute_queues_cpsch(dqm, filter, 0);
	if (retval || qpd->reset_wavefronts) {
		pr_warn("Resetting wave fronts (cpsch) on dev %p\n", dqm->dev);
		dbgdev_wave_reset_wavefronts(dqm->dev, qpd->pqm->process);
		qpd->reset_wavefronts = false;
	}

	/* lastly, free mqd resources */
	list_for_each_entry_safe(q, next, &qpd->queues_list, list) {
		mqd = dqm->ops.get_mqd_manager(dqm,
			get_mqd_type_from_queue_type(q->properties.type));
		if (!mqd) {
			retval = -ENOMEM;
			goto out;
		}
		list_del(&q->list);
1233
		qpd->queue_count--;
1234 1235 1236 1237 1238 1239 1240 1241
		mqd->uninit_mqd(mqd, q->mqd, q->mqd_mem_obj);
	}

out:
	mutex_unlock(&dqm->lock);
	return retval;
}

1242 1243 1244 1245
struct device_queue_manager *device_queue_manager_init(struct kfd_dev *dev)
{
	struct device_queue_manager *dqm;

1246
	pr_debug("Loading device queue manager\n");
1247

1248
	dqm = kzalloc(sizeof(*dqm), GFP_KERNEL);
1249 1250 1251 1252 1253 1254 1255 1256
	if (!dqm)
		return NULL;

	dqm->dev = dev;
	switch (sched_policy) {
	case KFD_SCHED_POLICY_HWS:
	case KFD_SCHED_POLICY_HWS_NO_OVERSUBSCRIPTION:
		/* initialize dqm for cp scheduling */
1257 1258 1259 1260 1261 1262
		dqm->ops.create_queue = create_queue_cpsch;
		dqm->ops.initialize = initialize_cpsch;
		dqm->ops.start = start_cpsch;
		dqm->ops.stop = stop_cpsch;
		dqm->ops.destroy_queue = destroy_queue_cpsch;
		dqm->ops.update_queue = update_queue;
1263 1264 1265 1266
		dqm->ops.get_mqd_manager = get_mqd_manager;
		dqm->ops.register_process = register_process;
		dqm->ops.unregister_process = unregister_process;
		dqm->ops.uninitialize = uninitialize;
1267 1268 1269
		dqm->ops.create_kernel_queue = create_kernel_queue_cpsch;
		dqm->ops.destroy_kernel_queue = destroy_kernel_queue_cpsch;
		dqm->ops.set_cache_memory_policy = set_cache_memory_policy;
1270
		dqm->ops.set_trap_handler = set_trap_handler;
1271
		dqm->ops.process_termination = process_termination_cpsch;
1272 1273 1274
		break;
	case KFD_SCHED_POLICY_NO_HWS:
		/* initialize dqm for no cp scheduling */
1275 1276 1277 1278 1279
		dqm->ops.start = start_nocpsch;
		dqm->ops.stop = stop_nocpsch;
		dqm->ops.create_queue = create_queue_nocpsch;
		dqm->ops.destroy_queue = destroy_queue_nocpsch;
		dqm->ops.update_queue = update_queue;
1280 1281 1282
		dqm->ops.get_mqd_manager = get_mqd_manager;
		dqm->ops.register_process = register_process;
		dqm->ops.unregister_process = unregister_process;
1283
		dqm->ops.initialize = initialize_nocpsch;
1284
		dqm->ops.uninitialize = uninitialize;
1285
		dqm->ops.set_cache_memory_policy = set_cache_memory_policy;
1286
		dqm->ops.set_trap_handler = set_trap_handler;
1287
		dqm->ops.process_termination = process_termination_nocpsch;
1288 1289
		break;
	default:
1290 1291
		pr_err("Invalid scheduling policy %d\n", sched_policy);
		goto out_free;
1292 1293
	}

1294 1295
	switch (dev->device_info->asic_family) {
	case CHIP_CARRIZO:
1296
		device_queue_manager_init_vi(&dqm->asic_ops);
1297 1298
		break;

1299
	case CHIP_KAVERI:
1300
		device_queue_manager_init_cik(&dqm->asic_ops);
1301
		break;
1302 1303 1304 1305
	default:
		WARN(1, "Unexpected ASIC family %u",
		     dev->device_info->asic_family);
		goto out_free;
1306 1307
	}

1308 1309
	if (!dqm->ops.initialize(dqm))
		return dqm;
1310

1311 1312 1313
out_free:
	kfree(dqm);
	return NULL;
1314 1315 1316 1317
}

void device_queue_manager_uninit(struct device_queue_manager *dqm)
{
1318
	dqm->ops.uninitialize(dqm);
1319 1320
	kfree(dqm);
}