booke.c 18.3 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
/*
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License, version 2, as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
 *
 * Copyright IBM Corp. 2007
 *
 * Authors: Hollis Blanchard <hollisb@us.ibm.com>
 *          Christian Ehrhardt <ehrhardt@linux.vnet.ibm.com>
 */

#include <linux/errno.h>
#include <linux/err.h>
#include <linux/kvm_host.h>
24
#include <linux/gfp.h>
25 26 27
#include <linux/module.h>
#include <linux/vmalloc.h>
#include <linux/fs.h>
28

29 30 31
#include <asm/cputable.h>
#include <asm/uaccess.h>
#include <asm/kvm_ppc.h>
32
#include "timing.h"
33
#include <asm/cacheflush.h>
34

35
#include "booke.h"
36

37 38
unsigned long kvmppc_booke_handlers;

39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55
#define VM_STAT(x) offsetof(struct kvm, stat.x), KVM_STAT_VM
#define VCPU_STAT(x) offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU

struct kvm_stats_debugfs_item debugfs_entries[] = {
	{ "mmio",       VCPU_STAT(mmio_exits) },
	{ "dcr",        VCPU_STAT(dcr_exits) },
	{ "sig",        VCPU_STAT(signal_exits) },
	{ "itlb_r",     VCPU_STAT(itlb_real_miss_exits) },
	{ "itlb_v",     VCPU_STAT(itlb_virt_miss_exits) },
	{ "dtlb_r",     VCPU_STAT(dtlb_real_miss_exits) },
	{ "dtlb_v",     VCPU_STAT(dtlb_virt_miss_exits) },
	{ "sysc",       VCPU_STAT(syscall_exits) },
	{ "isi",        VCPU_STAT(isi_exits) },
	{ "dsi",        VCPU_STAT(dsi_exits) },
	{ "inst_emu",   VCPU_STAT(emulated_inst_exits) },
	{ "dec",        VCPU_STAT(dec_exits) },
	{ "ext_intr",   VCPU_STAT(ext_intr_exits) },
56
	{ "halt_wakeup", VCPU_STAT(halt_wakeup) },
57 58 59 60 61 62 63 64
	{ NULL }
};

/* TODO: use vcpu_printf() */
void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu)
{
	int i;

65
	printk("pc:   %08lx msr:  %08llx\n", vcpu->arch.pc, vcpu->arch.shared->msr);
66
	printk("lr:   %08lx ctr:  %08lx\n", vcpu->arch.lr, vcpu->arch.ctr);
67 68
	printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0,
					    vcpu->arch.shared->srr1);
69 70 71 72

	printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions);

	for (i = 0; i < 32; i += 4) {
73
		printk("gpr%02d: %08lx %08lx %08lx %08lx\n", i,
74 75 76 77
		       kvmppc_get_gpr(vcpu, i),
		       kvmppc_get_gpr(vcpu, i+1),
		       kvmppc_get_gpr(vcpu, i+2),
		       kvmppc_get_gpr(vcpu, i+3));
78 79 80
	}
}

81 82
static void kvmppc_booke_queue_irqprio(struct kvm_vcpu *vcpu,
                                       unsigned int priority)
83 84 85 86
{
	set_bit(priority, &vcpu->arch.pending_exceptions);
}

87 88
static void kvmppc_core_queue_dtlb_miss(struct kvm_vcpu *vcpu,
                                        ulong dear_flags, ulong esr_flags)
89
{
90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112
	vcpu->arch.queued_dear = dear_flags;
	vcpu->arch.queued_esr = esr_flags;
	kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_DTLB_MISS);
}

static void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu,
                                           ulong dear_flags, ulong esr_flags)
{
	vcpu->arch.queued_dear = dear_flags;
	vcpu->arch.queued_esr = esr_flags;
	kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_DATA_STORAGE);
}

static void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu,
                                           ulong esr_flags)
{
	vcpu->arch.queued_esr = esr_flags;
	kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_INST_STORAGE);
}

void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong esr_flags)
{
	vcpu->arch.queued_esr = esr_flags;
113
	kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_PROGRAM);
114 115 116 117
}

void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu)
{
118
	kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_DECREMENTER);
119 120 121 122
}

int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu)
{
123
	return test_bit(BOOKE_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions);
124 125
}

126 127 128 129 130
void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu)
{
	clear_bit(BOOKE_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions);
}

131 132 133
void kvmppc_core_queue_external(struct kvm_vcpu *vcpu,
                                struct kvm_interrupt *irq)
{
134
	kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_EXTERNAL);
135 136
}

137 138 139 140 141 142
void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu,
                                  struct kvm_interrupt *irq)
{
	clear_bit(BOOKE_IRQPRIO_EXTERNAL, &vcpu->arch.pending_exceptions);
}

143 144 145
/* Deliver the interrupt of the corresponding priority, if possible. */
static int kvmppc_booke_irqprio_deliver(struct kvm_vcpu *vcpu,
                                        unsigned int priority)
146
{
147
	int allowed = 0;
148
	ulong uninitialized_var(msr_mask);
149
	bool update_esr = false, update_dear = false;
150 151 152 153 154 155 156 157 158 159 160 161 162 163
	ulong crit_raw = vcpu->arch.shared->critical;
	ulong crit_r1 = kvmppc_get_gpr(vcpu, 1);
	bool crit;

	/* Truncate crit indicators in 32 bit mode */
	if (!(vcpu->arch.shared->msr & MSR_SF)) {
		crit_raw &= 0xffffffff;
		crit_r1 &= 0xffffffff;
	}

	/* Critical section when crit == r1 */
	crit = (crit_raw == crit_r1);
	/* ... and we're in supervisor mode */
	crit = crit && !(vcpu->arch.shared->msr & MSR_PR);
164 165 166 167

	switch (priority) {
	case BOOKE_IRQPRIO_DTLB_MISS:
	case BOOKE_IRQPRIO_DATA_STORAGE:
168 169
		update_dear = true;
		/* fall through */
170
	case BOOKE_IRQPRIO_INST_STORAGE:
171 172 173 174 175
	case BOOKE_IRQPRIO_PROGRAM:
		update_esr = true;
		/* fall through */
	case BOOKE_IRQPRIO_ITLB_MISS:
	case BOOKE_IRQPRIO_SYSCALL:
176
	case BOOKE_IRQPRIO_FP_UNAVAIL:
177 178 179
	case BOOKE_IRQPRIO_SPE_UNAVAIL:
	case BOOKE_IRQPRIO_SPE_FP_DATA:
	case BOOKE_IRQPRIO_SPE_FP_ROUND:
180 181 182 183
	case BOOKE_IRQPRIO_AP_UNAVAIL:
	case BOOKE_IRQPRIO_ALIGNMENT:
		allowed = 1;
		msr_mask = MSR_CE|MSR_ME|MSR_DE;
184
		break;
185 186
	case BOOKE_IRQPRIO_CRITICAL:
	case BOOKE_IRQPRIO_WATCHDOG:
187
		allowed = vcpu->arch.shared->msr & MSR_CE;
188
		msr_mask = MSR_ME;
189
		break;
190
	case BOOKE_IRQPRIO_MACHINE_CHECK:
191
		allowed = vcpu->arch.shared->msr & MSR_ME;
192
		msr_mask = 0;
193
		break;
194 195 196
	case BOOKE_IRQPRIO_EXTERNAL:
	case BOOKE_IRQPRIO_DECREMENTER:
	case BOOKE_IRQPRIO_FIT:
197
		allowed = vcpu->arch.shared->msr & MSR_EE;
198
		allowed = allowed && !crit;
199
		msr_mask = MSR_CE|MSR_ME|MSR_DE;
200
		break;
201
	case BOOKE_IRQPRIO_DEBUG:
202
		allowed = vcpu->arch.shared->msr & MSR_DE;
203
		msr_mask = MSR_ME;
204 205 206
		break;
	}

207
	if (allowed) {
208 209
		vcpu->arch.shared->srr0 = vcpu->arch.pc;
		vcpu->arch.shared->srr1 = vcpu->arch.shared->msr;
210
		vcpu->arch.pc = vcpu->arch.ivpr | vcpu->arch.ivor[priority];
211 212 213
		if (update_esr == true)
			vcpu->arch.esr = vcpu->arch.queued_esr;
		if (update_dear == true)
214
			vcpu->arch.shared->dar = vcpu->arch.queued_dear;
215
		kvmppc_set_msr(vcpu, vcpu->arch.shared->msr & msr_mask);
216

217
		clear_bit(priority, &vcpu->arch.pending_exceptions);
218 219
	}

220
	return allowed;
221 222 223
}

/* Check pending exceptions and deliver one, if possible. */
224
void kvmppc_core_deliver_interrupts(struct kvm_vcpu *vcpu)
225 226
{
	unsigned long *pending = &vcpu->arch.pending_exceptions;
227
	unsigned long old_pending = vcpu->arch.pending_exceptions;
228 229
	unsigned int priority;

230
	priority = __ffs(*pending);
231
	while (priority <= BOOKE_IRQPRIO_MAX) {
232
		if (kvmppc_booke_irqprio_deliver(vcpu, priority))
233 234 235 236 237 238
			break;

		priority = find_next_bit(pending,
		                         BITS_PER_BYTE * sizeof(*pending),
		                         priority + 1);
	}
239 240 241 242 243 244

	/* Tell the guest about our interrupt status */
	if (*pending)
		vcpu->arch.shared->int_pending = 1;
	else if (old_pending)
		vcpu->arch.shared->int_pending = 0;
245 246 247 248 249 250 251 252 253 254 255 256 257
}

/**
 * kvmppc_handle_exit
 *
 * Return value is in the form (errcode<<2 | RESUME_FLAG_HOST | RESUME_FLAG_NV)
 */
int kvmppc_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu,
                       unsigned int exit_nr)
{
	enum emulation_result er;
	int r = RESUME_HOST;

258 259 260
	/* update before a new last_exit_type is rewritten */
	kvmppc_update_timing_stats(vcpu);

261 262 263 264 265 266 267 268 269 270 271 272 273
	local_irq_enable();

	run->exit_reason = KVM_EXIT_UNKNOWN;
	run->ready_for_interrupt_injection = 1;

	switch (exit_nr) {
	case BOOKE_INTERRUPT_MACHINE_CHECK:
		printk("MACHINE CHECK: %lx\n", mfspr(SPRN_MCSR));
		kvmppc_dump_vcpu(vcpu);
		r = RESUME_HOST;
		break;

	case BOOKE_INTERRUPT_EXTERNAL:
274
		kvmppc_account_exit(vcpu, EXT_INTR_EXITS);
275 276 277 278 279
		if (need_resched())
			cond_resched();
		r = RESUME_GUEST;
		break;

280 281 282 283
	case BOOKE_INTERRUPT_DECREMENTER:
		/* Since we switched IVPR back to the host's value, the host
		 * handled this interrupt the moment we enabled interrupts.
		 * Now we just offer it a chance to reschedule the guest. */
284
		kvmppc_account_exit(vcpu, DEC_EXITS);
285 286 287 288 289 290
		if (need_resched())
			cond_resched();
		r = RESUME_GUEST;
		break;

	case BOOKE_INTERRUPT_PROGRAM:
291
		if (vcpu->arch.shared->msr & MSR_PR) {
292 293
			/* Program traps generated by user-level software must be handled
			 * by the guest kernel. */
294
			kvmppc_core_queue_program(vcpu, vcpu->arch.fault_esr);
295
			r = RESUME_GUEST;
296
			kvmppc_account_exit(vcpu, USR_PR_INST);
297 298 299 300 301 302
			break;
		}

		er = kvmppc_emulate_instruction(run, vcpu);
		switch (er) {
		case EMULATE_DONE:
303
			/* don't overwrite subtypes, just account kvm_stats */
304
			kvmppc_account_exit_stat(vcpu, EMULATED_INST_EXITS);
305 306 307 308 309 310 311 312 313 314
			/* Future optimization: only reload non-volatiles if
			 * they were actually modified by emulation. */
			r = RESUME_GUEST_NV;
			break;
		case EMULATE_DO_DCR:
			run->exit_reason = KVM_EXIT_DCR;
			r = RESUME_HOST;
			break;
		case EMULATE_FAIL:
			/* XXX Deliver Program interrupt to guest. */
315
			printk(KERN_CRIT "%s: emulation at %lx failed (%08x)\n",
316 317 318 319 320 321 322 323 324 325 326 327
			       __func__, vcpu->arch.pc, vcpu->arch.last_inst);
			/* For debugging, encode the failing instruction and
			 * report it to userspace. */
			run->hw.hardware_exit_reason = ~0ULL << 32;
			run->hw.hardware_exit_reason |= vcpu->arch.last_inst;
			r = RESUME_HOST;
			break;
		default:
			BUG();
		}
		break;

328
	case BOOKE_INTERRUPT_FP_UNAVAIL:
329
		kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_FP_UNAVAIL);
330
		kvmppc_account_exit(vcpu, FP_UNAVAIL);
331 332 333
		r = RESUME_GUEST;
		break;

334 335 336 337 338 339 340 341 342 343 344 345 346 347 348
	case BOOKE_INTERRUPT_SPE_UNAVAIL:
		kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_SPE_UNAVAIL);
		r = RESUME_GUEST;
		break;

	case BOOKE_INTERRUPT_SPE_FP_DATA:
		kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_SPE_FP_DATA);
		r = RESUME_GUEST;
		break;

	case BOOKE_INTERRUPT_SPE_FP_ROUND:
		kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_SPE_FP_ROUND);
		r = RESUME_GUEST;
		break;

349
	case BOOKE_INTERRUPT_DATA_STORAGE:
350 351
		kvmppc_core_queue_data_storage(vcpu, vcpu->arch.fault_dear,
		                               vcpu->arch.fault_esr);
352
		kvmppc_account_exit(vcpu, DSI_EXITS);
353 354 355 356
		r = RESUME_GUEST;
		break;

	case BOOKE_INTERRUPT_INST_STORAGE:
357
		kvmppc_core_queue_inst_storage(vcpu, vcpu->arch.fault_esr);
358
		kvmppc_account_exit(vcpu, ISI_EXITS);
359 360 361 362
		r = RESUME_GUEST;
		break;

	case BOOKE_INTERRUPT_SYSCALL:
363 364 365 366 367 368 369 370 371
		if (!(vcpu->arch.shared->msr & MSR_PR) &&
		    (((u32)kvmppc_get_gpr(vcpu, 0)) == KVM_SC_MAGIC_R0)) {
			/* KVM PV hypercalls */
			kvmppc_set_gpr(vcpu, 3, kvmppc_kvm_pv(vcpu));
			r = RESUME_GUEST;
		} else {
			/* Guest syscalls */
			kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_SYSCALL);
		}
372
		kvmppc_account_exit(vcpu, SYSCALL_EXITS);
373 374 375 376 377
		r = RESUME_GUEST;
		break;

	case BOOKE_INTERRUPT_DTLB_MISS: {
		unsigned long eaddr = vcpu->arch.fault_dear;
378
		int gtlb_index;
379
		gpa_t gpaddr;
380 381 382
		gfn_t gfn;

		/* Check the guest TLB. */
383
		gtlb_index = kvmppc_mmu_dtlb_index(vcpu, eaddr);
384
		if (gtlb_index < 0) {
385
			/* The guest didn't have a mapping for it. */
386 387 388
			kvmppc_core_queue_dtlb_miss(vcpu,
			                            vcpu->arch.fault_dear,
			                            vcpu->arch.fault_esr);
389
			kvmppc_mmu_dtlb_miss(vcpu);
390
			kvmppc_account_exit(vcpu, DTLB_REAL_MISS_EXITS);
391 392 393 394
			r = RESUME_GUEST;
			break;
		}

395
		gpaddr = kvmppc_mmu_xlate(vcpu, gtlb_index, eaddr);
396
		gfn = gpaddr >> PAGE_SHIFT;
397 398 399 400 401 402 403 404

		if (kvm_is_visible_gfn(vcpu->kvm, gfn)) {
			/* The guest TLB had a mapping, but the shadow TLB
			 * didn't, and it is RAM. This could be because:
			 * a) the entry is mapping the host kernel, or
			 * b) the guest used a large mapping which we're faking
			 * Either way, we need to satisfy the fault without
			 * invoking the guest. */
405
			kvmppc_mmu_map(vcpu, eaddr, gpaddr, gtlb_index);
406
			kvmppc_account_exit(vcpu, DTLB_VIRT_MISS_EXITS);
407 408 409 410
			r = RESUME_GUEST;
		} else {
			/* Guest has mapped and accessed a page which is not
			 * actually RAM. */
411
			vcpu->arch.paddr_accessed = gpaddr;
412
			r = kvmppc_emulate_mmio(run, vcpu);
413
			kvmppc_account_exit(vcpu, MMIO_EXITS);
414 415 416 417 418 419 420
		}

		break;
	}

	case BOOKE_INTERRUPT_ITLB_MISS: {
		unsigned long eaddr = vcpu->arch.pc;
421
		gpa_t gpaddr;
422
		gfn_t gfn;
423
		int gtlb_index;
424 425 426 427

		r = RESUME_GUEST;

		/* Check the guest TLB. */
428
		gtlb_index = kvmppc_mmu_itlb_index(vcpu, eaddr);
429
		if (gtlb_index < 0) {
430
			/* The guest didn't have a mapping for it. */
431
			kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_ITLB_MISS);
432
			kvmppc_mmu_itlb_miss(vcpu);
433
			kvmppc_account_exit(vcpu, ITLB_REAL_MISS_EXITS);
434 435 436
			break;
		}

437
		kvmppc_account_exit(vcpu, ITLB_VIRT_MISS_EXITS);
438

439
		gpaddr = kvmppc_mmu_xlate(vcpu, gtlb_index, eaddr);
440
		gfn = gpaddr >> PAGE_SHIFT;
441 442 443 444 445 446 447 448

		if (kvm_is_visible_gfn(vcpu->kvm, gfn)) {
			/* The guest TLB had a mapping, but the shadow TLB
			 * didn't. This could be because:
			 * a) the entry is mapping the host kernel, or
			 * b) the guest used a large mapping which we're faking
			 * Either way, we need to satisfy the fault without
			 * invoking the guest. */
449
			kvmppc_mmu_map(vcpu, eaddr, gpaddr, gtlb_index);
450 451
		} else {
			/* Guest mapped and leaped at non-RAM! */
452
			kvmppc_booke_queue_irqprio(vcpu, BOOKE_IRQPRIO_MACHINE_CHECK);
453 454 455 456 457
		}

		break;
	}

458 459 460 461 462 463 464 465 466 467 468
	case BOOKE_INTERRUPT_DEBUG: {
		u32 dbsr;

		vcpu->arch.pc = mfspr(SPRN_CSRR0);

		/* clear IAC events in DBSR register */
		dbsr = mfspr(SPRN_DBSR);
		dbsr &= DBSR_IAC1 | DBSR_IAC2 | DBSR_IAC3 | DBSR_IAC4;
		mtspr(SPRN_DBSR, dbsr);

		run->exit_reason = KVM_EXIT_DEBUG;
469
		kvmppc_account_exit(vcpu, DEBUG_EXITS);
470 471 472 473
		r = RESUME_HOST;
		break;
	}

474 475 476 477 478 479 480
	default:
		printk(KERN_EMERG "exit_nr %d\n", exit_nr);
		BUG();
	}

	local_irq_disable();

481
	kvmppc_core_deliver_interrupts(vcpu);
482 483 484 485 486 487 488 489

	if (!(r & RESUME_HOST)) {
		/* To avoid clobbering exit_reason, only check for signals if
		 * we aren't already exiting to userspace for some other
		 * reason. */
		if (signal_pending(current)) {
			run->exit_reason = KVM_EXIT_INTR;
			r = (-EINTR << 2) | RESUME_HOST | (r & RESUME_FLAG_NV);
490
			kvmppc_account_exit(vcpu, SIGNAL_EXITS);
491 492 493 494 495 496 497 498 499 500
		}
	}

	return r;
}

/* Initial guest state: 16MB mapping 0 -> 0, PC = 0, MSR = 0, R1 = 16MB */
int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu)
{
	vcpu->arch.pc = 0;
501
	vcpu->arch.shared->msr = 0;
502
	kvmppc_set_gpr(vcpu, 1, (16<<20) - 8); /* -8 for the callee-save LR slot */
503

504 505
	vcpu->arch.shadow_pid = 1;

506 507 508 509
	/* Eye-catching number so we know if the guest takes an interrupt
	 * before it's programmed its own IVPR. */
	vcpu->arch.ivpr = 0x55550000;

510 511
	kvmppc_init_timing_stats(vcpu);

512
	return kvmppc_core_vcpu_setup(vcpu);
513 514 515 516 517 518 519
}

int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
{
	int i;

	regs->pc = vcpu->arch.pc;
520
	regs->cr = kvmppc_get_cr(vcpu);
521 522
	regs->ctr = vcpu->arch.ctr;
	regs->lr = vcpu->arch.lr;
523
	regs->xer = kvmppc_get_xer(vcpu);
524
	regs->msr = vcpu->arch.shared->msr;
525 526
	regs->srr0 = vcpu->arch.shared->srr0;
	regs->srr1 = vcpu->arch.shared->srr1;
527
	regs->pid = vcpu->arch.pid;
528 529 530 531
	regs->sprg0 = vcpu->arch.shared->sprg0;
	regs->sprg1 = vcpu->arch.shared->sprg1;
	regs->sprg2 = vcpu->arch.shared->sprg2;
	regs->sprg3 = vcpu->arch.shared->sprg3;
532 533 534 535 536
	regs->sprg5 = vcpu->arch.sprg4;
	regs->sprg6 = vcpu->arch.sprg5;
	regs->sprg7 = vcpu->arch.sprg6;

	for (i = 0; i < ARRAY_SIZE(regs->gpr); i++)
537
		regs->gpr[i] = kvmppc_get_gpr(vcpu, i);
538 539 540 541 542 543 544 545 546

	return 0;
}

int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
{
	int i;

	vcpu->arch.pc = regs->pc;
547
	kvmppc_set_cr(vcpu, regs->cr);
548 549
	vcpu->arch.ctr = regs->ctr;
	vcpu->arch.lr = regs->lr;
550
	kvmppc_set_xer(vcpu, regs->xer);
551
	kvmppc_set_msr(vcpu, regs->msr);
552 553
	vcpu->arch.shared->srr0 = regs->srr0;
	vcpu->arch.shared->srr1 = regs->srr1;
554 555 556 557
	vcpu->arch.shared->sprg0 = regs->sprg0;
	vcpu->arch.shared->sprg1 = regs->sprg1;
	vcpu->arch.shared->sprg2 = regs->sprg2;
	vcpu->arch.shared->sprg3 = regs->sprg3;
558 559 560 561
	vcpu->arch.sprg5 = regs->sprg4;
	vcpu->arch.sprg6 = regs->sprg5;
	vcpu->arch.sprg7 = regs->sprg6;

562 563
	for (i = 0; i < ARRAY_SIZE(regs->gpr); i++)
		kvmppc_set_gpr(vcpu, i, regs->gpr[i]);
564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592

	return 0;
}

int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
                                  struct kvm_sregs *sregs)
{
	return -ENOTSUPP;
}

int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
                                  struct kvm_sregs *sregs)
{
	return -ENOTSUPP;
}

int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
{
	return -ENOTSUPP;
}

int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
{
	return -ENOTSUPP;
}

int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
                                  struct kvm_translation *tr)
{
593 594 595 596
	int r;

	r = kvmppc_core_vcpu_translate(vcpu, tr);
	return r;
597
}
598

599 600 601 602 603
int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
{
	return -ENOTSUPP;
}

604
int __init kvmppc_booke_init(void)
605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648
{
	unsigned long ivor[16];
	unsigned long max_ivor = 0;
	int i;

	/* We install our own exception handlers by hijacking IVPR. IVPR must
	 * be 16-bit aligned, so we need a 64KB allocation. */
	kvmppc_booke_handlers = __get_free_pages(GFP_KERNEL | __GFP_ZERO,
	                                         VCPU_SIZE_ORDER);
	if (!kvmppc_booke_handlers)
		return -ENOMEM;

	/* XXX make sure our handlers are smaller than Linux's */

	/* Copy our interrupt handlers to match host IVORs. That way we don't
	 * have to swap the IVORs on every guest/host transition. */
	ivor[0] = mfspr(SPRN_IVOR0);
	ivor[1] = mfspr(SPRN_IVOR1);
	ivor[2] = mfspr(SPRN_IVOR2);
	ivor[3] = mfspr(SPRN_IVOR3);
	ivor[4] = mfspr(SPRN_IVOR4);
	ivor[5] = mfspr(SPRN_IVOR5);
	ivor[6] = mfspr(SPRN_IVOR6);
	ivor[7] = mfspr(SPRN_IVOR7);
	ivor[8] = mfspr(SPRN_IVOR8);
	ivor[9] = mfspr(SPRN_IVOR9);
	ivor[10] = mfspr(SPRN_IVOR10);
	ivor[11] = mfspr(SPRN_IVOR11);
	ivor[12] = mfspr(SPRN_IVOR12);
	ivor[13] = mfspr(SPRN_IVOR13);
	ivor[14] = mfspr(SPRN_IVOR14);
	ivor[15] = mfspr(SPRN_IVOR15);

	for (i = 0; i < 16; i++) {
		if (ivor[i] > max_ivor)
			max_ivor = ivor[i];

		memcpy((void *)kvmppc_booke_handlers + ivor[i],
		       kvmppc_handlers_start + i * kvmppc_handler_len,
		       kvmppc_handler_len);
	}
	flush_icache_range(kvmppc_booke_handlers,
	                   kvmppc_booke_handlers + max_ivor + kvmppc_handler_len);

649
	return 0;
650 651
}

652
void __exit kvmppc_booke_exit(void)
653 654 655 656
{
	free_pages(kvmppc_booke_handlers, VCPU_SIZE_ORDER);
	kvm_exit();
}