book3s.c 27.3 KB
Newer Older
A
Alexander Graf 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
/*
 * Copyright (C) 2009. SUSE Linux Products GmbH. All rights reserved.
 *
 * Authors:
 *    Alexander Graf <agraf@suse.de>
 *    Kevin Wolf <mail@kevin-wolf.de>
 *
 * Description:
 * This file is derived from arch/powerpc/kvm/44x.c,
 * by Hollis Blanchard <hollisb@us.ibm.com>.
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License, version 2, as
 * published by the Free Software Foundation.
 */

#include <linux/kvm_host.h>
#include <linux/err.h>
19
#include <linux/export.h>
20
#include <linux/slab.h>
21 22
#include <linux/module.h>
#include <linux/miscdevice.h>
23 24 25 26
#include <linux/gfp.h>
#include <linux/sched.h>
#include <linux/vmalloc.h>
#include <linux/highmem.h>
A
Alexander Graf 已提交
27 28 29 30

#include <asm/reg.h>
#include <asm/cputable.h>
#include <asm/cacheflush.h>
31
#include <linux/uaccess.h>
A
Alexander Graf 已提交
32 33 34 35
#include <asm/io.h>
#include <asm/kvm_ppc.h>
#include <asm/kvm_book3s.h>
#include <asm/mmu_context.h>
36
#include <asm/page.h>
37
#include <asm/xive.h>
A
Alexander Graf 已提交
38

39
#include "book3s.h"
40 41
#include "trace.h"

42
#define VM_STAT(x) offsetof(struct kvm, stat.x), KVM_STAT_VM
A
Alexander Graf 已提交
43 44 45
#define VCPU_STAT(x) offsetof(struct kvm_vcpu, stat.x), KVM_STAT_VCPU

/* #define EXIT_DEBUG */
46

A
Alexander Graf 已提交
47 48 49 50 51 52 53 54 55
struct kvm_stats_debugfs_item debugfs_entries[] = {
	{ "exits",       VCPU_STAT(sum_exits) },
	{ "mmio",        VCPU_STAT(mmio_exits) },
	{ "sig",         VCPU_STAT(signal_exits) },
	{ "sysc",        VCPU_STAT(syscall_exits) },
	{ "inst_emu",    VCPU_STAT(emulated_inst_exits) },
	{ "dec",         VCPU_STAT(dec_exits) },
	{ "ext_intr",    VCPU_STAT(ext_intr_exits) },
	{ "queue_intr",  VCPU_STAT(queue_intr) },
56 57 58
	{ "halt_poll_success_ns",	VCPU_STAT(halt_poll_success_ns) },
	{ "halt_poll_fail_ns",		VCPU_STAT(halt_poll_fail_ns) },
	{ "halt_wait_ns",		VCPU_STAT(halt_wait_ns) },
59
	{ "halt_successful_poll", VCPU_STAT(halt_successful_poll), },
60
	{ "halt_attempted_poll", VCPU_STAT(halt_attempted_poll), },
61
	{ "halt_successful_wait",	VCPU_STAT(halt_successful_wait) },
62
	{ "halt_poll_invalid", VCPU_STAT(halt_poll_invalid) },
A
Alexander Graf 已提交
63 64 65 66 67 68 69 70 71
	{ "halt_wakeup", VCPU_STAT(halt_wakeup) },
	{ "pf_storage",  VCPU_STAT(pf_storage) },
	{ "sp_storage",  VCPU_STAT(sp_storage) },
	{ "pf_instruc",  VCPU_STAT(pf_instruc) },
	{ "sp_instruc",  VCPU_STAT(sp_instruc) },
	{ "ld",          VCPU_STAT(ld) },
	{ "ld_slow",     VCPU_STAT(ld_slow) },
	{ "st",          VCPU_STAT(st) },
	{ "st_slow",     VCPU_STAT(st_slow) },
72 73 74
	{ "pthru_all",       VCPU_STAT(pthru_all) },
	{ "pthru_host",      VCPU_STAT(pthru_host) },
	{ "pthru_bad_aff",   VCPU_STAT(pthru_bad_aff) },
75 76
	{ "largepages_2M",    VM_STAT(num_2M_pages) },
	{ "largepages_1G",    VM_STAT(num_1G_pages) },
A
Alexander Graf 已提交
77 78 79
	{ NULL }
};

80 81 82 83
void kvmppc_unfixup_split_real(struct kvm_vcpu *vcpu)
{
	if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) {
		ulong pc = kvmppc_get_pc(vcpu);
84
		ulong lr = kvmppc_get_lr(vcpu);
85 86
		if ((pc & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS)
			kvmppc_set_pc(vcpu, pc & ~SPLIT_HACK_MASK);
87 88
		if ((lr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS)
			kvmppc_set_lr(vcpu, lr & ~SPLIT_HACK_MASK);
89 90 91 92 93
		vcpu->arch.hflags &= ~BOOK3S_HFLAG_SPLIT_HACK;
	}
}
EXPORT_SYMBOL_GPL(kvmppc_unfixup_split_real);

94 95
static inline unsigned long kvmppc_interrupt_offset(struct kvm_vcpu *vcpu)
{
96
	if (!is_kvmppc_hv_enabled(vcpu->kvm))
97 98 99 100 101 102 103
		return to_book3s(vcpu)->hior;
	return 0;
}

static inline void kvmppc_update_int_pending(struct kvm_vcpu *vcpu,
			unsigned long pending_now, unsigned long old_pending)
{
104
	if (is_kvmppc_hv_enabled(vcpu->kvm))
105 106
		return;
	if (pending_now)
107
		kvmppc_set_int_pending(vcpu, 1);
108
	else if (old_pending)
109
		kvmppc_set_int_pending(vcpu, 0);
110 111 112 113 114 115 116 117
}

static inline bool kvmppc_critical_section(struct kvm_vcpu *vcpu)
{
	ulong crit_raw;
	ulong crit_r1;
	bool crit;

118
	if (is_kvmppc_hv_enabled(vcpu->kvm))
119 120
		return false;

121
	crit_raw = kvmppc_get_critical(vcpu);
122 123 124
	crit_r1 = kvmppc_get_gpr(vcpu, 1);

	/* Truncate crit indicators in 32 bit mode */
125
	if (!(kvmppc_get_msr(vcpu) & MSR_SF)) {
126 127 128 129 130 131 132
		crit_raw &= 0xffffffff;
		crit_r1 &= 0xffffffff;
	}

	/* Critical section when crit == r1 */
	crit = (crit_raw == crit_r1);
	/* ... and we're in supervisor mode */
133
	crit = crit && !(kvmppc_get_msr(vcpu) & MSR_PR);
134 135 136 137

	return crit;
}

A
Alexander Graf 已提交
138 139
void kvmppc_inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 flags)
{
140
	kvmppc_unfixup_split_real(vcpu);
141
	kvmppc_set_srr0(vcpu, kvmppc_get_pc(vcpu));
142
	kvmppc_set_srr1(vcpu, (kvmppc_get_msr(vcpu) & ~0x783f0000ul) | flags);
143
	kvmppc_set_pc(vcpu, kvmppc_interrupt_offset(vcpu) + vec);
A
Alexander Graf 已提交
144 145 146
	vcpu->arch.mmu.reset_msr(vcpu);
}

147
static int kvmppc_book3s_vec2irqprio(unsigned int vec)
A
Alexander Graf 已提交
148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166
{
	unsigned int prio;

	switch (vec) {
	case 0x100: prio = BOOK3S_IRQPRIO_SYSTEM_RESET;		break;
	case 0x200: prio = BOOK3S_IRQPRIO_MACHINE_CHECK;	break;
	case 0x300: prio = BOOK3S_IRQPRIO_DATA_STORAGE;		break;
	case 0x380: prio = BOOK3S_IRQPRIO_DATA_SEGMENT;		break;
	case 0x400: prio = BOOK3S_IRQPRIO_INST_STORAGE;		break;
	case 0x480: prio = BOOK3S_IRQPRIO_INST_SEGMENT;		break;
	case 0x500: prio = BOOK3S_IRQPRIO_EXTERNAL;		break;
	case 0x600: prio = BOOK3S_IRQPRIO_ALIGNMENT;		break;
	case 0x700: prio = BOOK3S_IRQPRIO_PROGRAM;		break;
	case 0x800: prio = BOOK3S_IRQPRIO_FP_UNAVAIL;		break;
	case 0x900: prio = BOOK3S_IRQPRIO_DECREMENTER;		break;
	case 0xc00: prio = BOOK3S_IRQPRIO_SYSCALL;		break;
	case 0xd00: prio = BOOK3S_IRQPRIO_DEBUG;		break;
	case 0xf20: prio = BOOK3S_IRQPRIO_ALTIVEC;		break;
	case 0xf40: prio = BOOK3S_IRQPRIO_VSX;			break;
167
	case 0xf60: prio = BOOK3S_IRQPRIO_FAC_UNAVAIL;		break;
A
Alexander Graf 已提交
168 169 170
	default:    prio = BOOK3S_IRQPRIO_MAX;			break;
	}

171 172 173
	return prio;
}

174
void kvmppc_book3s_dequeue_irqprio(struct kvm_vcpu *vcpu,
175 176
					  unsigned int vec)
{
177 178
	unsigned long old_pending = vcpu->arch.pending_exceptions;

179 180
	clear_bit(kvmppc_book3s_vec2irqprio(vec),
		  &vcpu->arch.pending_exceptions);
181

182 183
	kvmppc_update_int_pending(vcpu, vcpu->arch.pending_exceptions,
				  old_pending);
184 185
}

186 187 188 189 190 191
void kvmppc_book3s_queue_irqprio(struct kvm_vcpu *vcpu, unsigned int vec)
{
	vcpu->stat.queue_intr++;

	set_bit(kvmppc_book3s_vec2irqprio(vec),
		&vcpu->arch.pending_exceptions);
A
Alexander Graf 已提交
192 193 194 195
#ifdef EXIT_DEBUG
	printk(KERN_INFO "Queueing interrupt %x\n", vec);
#endif
}
196
EXPORT_SYMBOL_GPL(kvmppc_book3s_queue_irqprio);
A
Alexander Graf 已提交
197

198 199 200 201 202 203 204
void kvmppc_core_queue_machine_check(struct kvm_vcpu *vcpu, ulong flags)
{
	/* might as well deliver this straight away */
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_MACHINE_CHECK, flags);
}
EXPORT_SYMBOL_GPL(kvmppc_core_queue_machine_check);

205
void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong flags)
A
Alexander Graf 已提交
206
{
207 208
	/* might as well deliver this straight away */
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_PROGRAM, flags);
A
Alexander Graf 已提交
209
}
210
EXPORT_SYMBOL_GPL(kvmppc_core_queue_program);
A
Alexander Graf 已提交
211

212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229
void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu)
{
	/* might as well deliver this straight away */
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_FP_UNAVAIL, 0);
}

void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu)
{
	/* might as well deliver this straight away */
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_ALTIVEC, 0);
}

void kvmppc_core_queue_vsx_unavail(struct kvm_vcpu *vcpu)
{
	/* might as well deliver this straight away */
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_VSX, 0);
}

A
Alexander Graf 已提交
230 231 232 233
void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu)
{
	kvmppc_book3s_queue_irqprio(vcpu, BOOK3S_INTERRUPT_DECREMENTER);
}
234
EXPORT_SYMBOL_GPL(kvmppc_core_queue_dec);
A
Alexander Graf 已提交
235 236 237

int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu)
{
238
	return test_bit(BOOK3S_IRQPRIO_DECREMENTER, &vcpu->arch.pending_exceptions);
A
Alexander Graf 已提交
239
}
240
EXPORT_SYMBOL_GPL(kvmppc_core_pending_dec);
A
Alexander Graf 已提交
241

242 243 244 245
void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu)
{
	kvmppc_book3s_dequeue_irqprio(vcpu, BOOK3S_INTERRUPT_DECREMENTER);
}
246
EXPORT_SYMBOL_GPL(kvmppc_core_dequeue_dec);
247

A
Alexander Graf 已提交
248 249 250
void kvmppc_core_queue_external(struct kvm_vcpu *vcpu,
                                struct kvm_interrupt *irq)
{
251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274
	/*
	 * This case (KVM_INTERRUPT_SET) should never actually arise for
	 * a pseries guest (because pseries guests expect their interrupt
	 * controllers to continue asserting an external interrupt request
	 * until it is acknowledged at the interrupt controller), but is
	 * included to avoid ABI breakage and potentially for other
	 * sorts of guest.
	 *
	 * There is a subtlety here: HV KVM does not test the
	 * external_oneshot flag in the code that synthesizes
	 * external interrupts for the guest just before entering
	 * the guest.  That is OK even if userspace did do a
	 * KVM_INTERRUPT_SET on a pseries guest vcpu, because the
	 * caller (kvm_vcpu_ioctl_interrupt) does a kvm_vcpu_kick()
	 * which ends up doing a smp_send_reschedule(), which will
	 * pull the guest all the way out to the host, meaning that
	 * we will call kvmppc_core_prepare_to_enter() before entering
	 * the guest again, and that will handle the external_oneshot
	 * flag correctly.
	 */
	if (irq->irq == KVM_INTERRUPT_SET)
		vcpu->arch.external_oneshot = 1;

	kvmppc_book3s_queue_irqprio(vcpu, BOOK3S_INTERRUPT_EXTERNAL);
A
Alexander Graf 已提交
275 276
}

277
void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu)
278 279 280 281
{
	kvmppc_book3s_dequeue_irqprio(vcpu, BOOK3S_INTERRUPT_EXTERNAL);
}

282 283 284 285 286
void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu, ulong dar,
				    ulong flags)
{
	kvmppc_set_dar(vcpu, dar);
	kvmppc_set_dsisr(vcpu, flags);
287
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_DATA_STORAGE, 0);
288
}
289
EXPORT_SYMBOL_GPL(kvmppc_core_queue_data_storage);
290 291 292

void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu, ulong flags)
{
293
	kvmppc_inject_interrupt(vcpu, BOOK3S_INTERRUPT_INST_STORAGE, flags);
294
}
295
EXPORT_SYMBOL_GPL(kvmppc_core_queue_inst_storage);
296

T
Thomas Huth 已提交
297 298
static int kvmppc_book3s_irqprio_deliver(struct kvm_vcpu *vcpu,
					 unsigned int priority)
A
Alexander Graf 已提交
299 300 301
{
	int deliver = 1;
	int vec = 0;
302
	bool crit = kvmppc_critical_section(vcpu);
A
Alexander Graf 已提交
303 304 305

	switch (priority) {
	case BOOK3S_IRQPRIO_DECREMENTER:
306
		deliver = (kvmppc_get_msr(vcpu) & MSR_EE) && !crit;
A
Alexander Graf 已提交
307 308 309
		vec = BOOK3S_INTERRUPT_DECREMENTER;
		break;
	case BOOK3S_IRQPRIO_EXTERNAL:
310
		deliver = (kvmppc_get_msr(vcpu) & MSR_EE) && !crit;
A
Alexander Graf 已提交
311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354
		vec = BOOK3S_INTERRUPT_EXTERNAL;
		break;
	case BOOK3S_IRQPRIO_SYSTEM_RESET:
		vec = BOOK3S_INTERRUPT_SYSTEM_RESET;
		break;
	case BOOK3S_IRQPRIO_MACHINE_CHECK:
		vec = BOOK3S_INTERRUPT_MACHINE_CHECK;
		break;
	case BOOK3S_IRQPRIO_DATA_STORAGE:
		vec = BOOK3S_INTERRUPT_DATA_STORAGE;
		break;
	case BOOK3S_IRQPRIO_INST_STORAGE:
		vec = BOOK3S_INTERRUPT_INST_STORAGE;
		break;
	case BOOK3S_IRQPRIO_DATA_SEGMENT:
		vec = BOOK3S_INTERRUPT_DATA_SEGMENT;
		break;
	case BOOK3S_IRQPRIO_INST_SEGMENT:
		vec = BOOK3S_INTERRUPT_INST_SEGMENT;
		break;
	case BOOK3S_IRQPRIO_ALIGNMENT:
		vec = BOOK3S_INTERRUPT_ALIGNMENT;
		break;
	case BOOK3S_IRQPRIO_PROGRAM:
		vec = BOOK3S_INTERRUPT_PROGRAM;
		break;
	case BOOK3S_IRQPRIO_VSX:
		vec = BOOK3S_INTERRUPT_VSX;
		break;
	case BOOK3S_IRQPRIO_ALTIVEC:
		vec = BOOK3S_INTERRUPT_ALTIVEC;
		break;
	case BOOK3S_IRQPRIO_FP_UNAVAIL:
		vec = BOOK3S_INTERRUPT_FP_UNAVAIL;
		break;
	case BOOK3S_IRQPRIO_SYSCALL:
		vec = BOOK3S_INTERRUPT_SYSCALL;
		break;
	case BOOK3S_IRQPRIO_DEBUG:
		vec = BOOK3S_INTERRUPT_TRACE;
		break;
	case BOOK3S_IRQPRIO_PERFORMANCE_MONITOR:
		vec = BOOK3S_INTERRUPT_PERFMON;
		break;
355 356 357
	case BOOK3S_IRQPRIO_FAC_UNAVAIL:
		vec = BOOK3S_INTERRUPT_FAC_UNAVAIL;
		break;
A
Alexander Graf 已提交
358 359 360 361 362 363 364 365 366 367 368
	default:
		deliver = 0;
		printk(KERN_ERR "KVM: Unknown interrupt: 0x%x\n", priority);
		break;
	}

#if 0
	printk(KERN_INFO "Deliver interrupt 0x%x? %x\n", vec, deliver);
#endif

	if (deliver)
369
		kvmppc_inject_interrupt(vcpu, vec, 0);
A
Alexander Graf 已提交
370 371 372 373

	return deliver;
}

374 375 376 377 378 379 380 381 382
/*
 * This function determines if an irqprio should be cleared once issued.
 */
static bool clear_irqprio(struct kvm_vcpu *vcpu, unsigned int priority)
{
	switch (priority) {
		case BOOK3S_IRQPRIO_DECREMENTER:
			/* DEC interrupts get cleared by mtdec */
			return false;
383 384 385 386 387 388 389 390 391 392
		case BOOK3S_IRQPRIO_EXTERNAL:
			/*
			 * External interrupts get cleared by userspace
			 * except when set by the KVM_INTERRUPT ioctl with
			 * KVM_INTERRUPT_SET (not KVM_INTERRUPT_SET_LEVEL).
			 */
			if (vcpu->arch.external_oneshot) {
				vcpu->arch.external_oneshot = 0;
				return true;
			}
393 394 395 396 397 398
			return false;
	}

	return true;
}

399
int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu)
A
Alexander Graf 已提交
400 401
{
	unsigned long *pending = &vcpu->arch.pending_exceptions;
402
	unsigned long old_pending = vcpu->arch.pending_exceptions;
A
Alexander Graf 已提交
403 404 405 406 407 408 409
	unsigned int priority;

#ifdef EXIT_DEBUG
	if (vcpu->arch.pending_exceptions)
		printk(KERN_EMERG "KVM: Check pending: %lx\n", vcpu->arch.pending_exceptions);
#endif
	priority = __ffs(*pending);
A
Alexander Graf 已提交
410
	while (priority < BOOK3S_IRQPRIO_MAX) {
411
		if (kvmppc_book3s_irqprio_deliver(vcpu, priority) &&
412
		    clear_irqprio(vcpu, priority)) {
A
Alexander Graf 已提交
413 414 415 416 417 418 419 420
			clear_bit(priority, &vcpu->arch.pending_exceptions);
			break;
		}

		priority = find_next_bit(pending,
					 BITS_PER_BYTE * sizeof(*pending),
					 priority + 1);
	}
421 422

	/* Tell the guest about our interrupt status */
423
	kvmppc_update_int_pending(vcpu, *pending, old_pending);
424 425

	return 0;
A
Alexander Graf 已提交
426
}
427
EXPORT_SYMBOL_GPL(kvmppc_core_prepare_to_enter);
A
Alexander Graf 已提交
428

D
Dan Williams 已提交
429
kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing,
430
			bool *writable)
431
{
432 433
	ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM;
	gfn_t gfn = gpa >> PAGE_SHIFT;
434

435
	if (!(kvmppc_get_msr(vcpu) & MSR_SF))
436 437
		mp_pa = (uint32_t)mp_pa;

438
	/* Magic page override */
439 440
	gpa &= ~0xFFFULL;
	if (unlikely(mp_pa) && unlikely((gpa & KVM_PAM) == mp_pa)) {
441
		ulong shared_page = ((ulong)vcpu->arch.shared) & PAGE_MASK;
D
Dan Williams 已提交
442
		kvm_pfn_t pfn;
443

D
Dan Williams 已提交
444
		pfn = (kvm_pfn_t)virt_to_phys((void*)shared_page) >> PAGE_SHIFT;
445
		get_page(pfn_to_page(pfn));
446 447
		if (writable)
			*writable = true;
448 449 450
		return pfn;
	}

451
	return gfn_to_pfn_prot(vcpu->kvm, gfn, writing, writable);
452
}
453
EXPORT_SYMBOL_GPL(kvmppc_gpa_to_pfn);
454

455 456
int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, enum xlate_instdata xlid,
		 enum xlate_readwrite xlrw, struct kvmppc_pte *pte)
A
Alexander Graf 已提交
457
{
458 459
	bool data = (xlid == XLATE_DATA);
	bool iswrite = (xlrw == XLATE_WRITE);
460
	int relocated = (kvmppc_get_msr(vcpu) & (data ? MSR_DR : MSR_IR));
A
Alexander Graf 已提交
461 462 463
	int r;

	if (relocated) {
464
		r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite);
A
Alexander Graf 已提交
465 466
	} else {
		pte->eaddr = eaddr;
A
Alexander Graf 已提交
467
		pte->raddr = eaddr & KVM_PAM;
468
		pte->vpage = VSID_REAL | eaddr >> 12;
A
Alexander Graf 已提交
469 470 471 472
		pte->may_read = true;
		pte->may_write = true;
		pte->may_execute = true;
		r = 0;
473 474 475 476 477 478 479

		if ((kvmppc_get_msr(vcpu) & (MSR_IR | MSR_DR)) == MSR_DR &&
		    !data) {
			if ((vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) &&
			    ((eaddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS))
			pte->raddr &= ~SPLIT_HACK_MASK;
		}
A
Alexander Graf 已提交
480 481 482 483 484
	}

	return r;
}

485 486
int kvmppc_load_last_inst(struct kvm_vcpu *vcpu,
		enum instruction_fetch_type type, u32 *inst)
487 488 489 490 491 492 493 494 495 496 497 498 499 500 501
{
	ulong pc = kvmppc_get_pc(vcpu);
	int r;

	if (type == INST_SC)
		pc -= 4;

	r = kvmppc_ld(vcpu, &pc, sizeof(u32), inst, false);
	if (r == EMULATE_DONE)
		return r;
	else
		return EMULATE_AGAIN;
}
EXPORT_SYMBOL_GPL(kvmppc_load_last_inst);

A
Alexander Graf 已提交
502 503 504 505 506
int kvm_arch_vcpu_setup(struct kvm_vcpu *vcpu)
{
	return 0;
}

507 508 509 510 511 512 513 514 515
int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu)
{
	return 0;
}

void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu)
{
}

516 517 518
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
				  struct kvm_sregs *sregs)
{
519 520 521 522 523 524 525
	int ret;

	vcpu_load(vcpu);
	ret = vcpu->kvm->arch.kvm_ops->get_sregs(vcpu, sregs);
	vcpu_put(vcpu);

	return ret;
526 527 528 529 530
}

int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
				  struct kvm_sregs *sregs)
{
531 532 533 534 535 536 537
	int ret;

	vcpu_load(vcpu);
	ret = vcpu->kvm->arch.kvm_ops->set_sregs(vcpu, sregs);
	vcpu_put(vcpu);

	return ret;
538 539
}

A
Alexander Graf 已提交
540 541 542 543
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
{
	int i;

544
	regs->pc = kvmppc_get_pc(vcpu);
545
	regs->cr = kvmppc_get_cr(vcpu);
546 547
	regs->ctr = kvmppc_get_ctr(vcpu);
	regs->lr = kvmppc_get_lr(vcpu);
548
	regs->xer = kvmppc_get_xer(vcpu);
549 550 551
	regs->msr = kvmppc_get_msr(vcpu);
	regs->srr0 = kvmppc_get_srr0(vcpu);
	regs->srr1 = kvmppc_get_srr1(vcpu);
A
Alexander Graf 已提交
552
	regs->pid = vcpu->arch.pid;
553 554 555 556 557 558 559 560
	regs->sprg0 = kvmppc_get_sprg0(vcpu);
	regs->sprg1 = kvmppc_get_sprg1(vcpu);
	regs->sprg2 = kvmppc_get_sprg2(vcpu);
	regs->sprg3 = kvmppc_get_sprg3(vcpu);
	regs->sprg4 = kvmppc_get_sprg4(vcpu);
	regs->sprg5 = kvmppc_get_sprg5(vcpu);
	regs->sprg6 = kvmppc_get_sprg6(vcpu);
	regs->sprg7 = kvmppc_get_sprg7(vcpu);
A
Alexander Graf 已提交
561 562

	for (i = 0; i < ARRAY_SIZE(regs->gpr); i++)
563
		regs->gpr[i] = kvmppc_get_gpr(vcpu, i);
A
Alexander Graf 已提交
564 565 566 567 568 569 570 571

	return 0;
}

int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
{
	int i;

572
	kvmppc_set_pc(vcpu, regs->pc);
573
	kvmppc_set_cr(vcpu, regs->cr);
574 575
	kvmppc_set_ctr(vcpu, regs->ctr);
	kvmppc_set_lr(vcpu, regs->lr);
576
	kvmppc_set_xer(vcpu, regs->xer);
A
Alexander Graf 已提交
577
	kvmppc_set_msr(vcpu, regs->msr);
578 579 580 581 582 583 584 585 586 587
	kvmppc_set_srr0(vcpu, regs->srr0);
	kvmppc_set_srr1(vcpu, regs->srr1);
	kvmppc_set_sprg0(vcpu, regs->sprg0);
	kvmppc_set_sprg1(vcpu, regs->sprg1);
	kvmppc_set_sprg2(vcpu, regs->sprg2);
	kvmppc_set_sprg3(vcpu, regs->sprg3);
	kvmppc_set_sprg4(vcpu, regs->sprg4);
	kvmppc_set_sprg5(vcpu, regs->sprg5);
	kvmppc_set_sprg6(vcpu, regs->sprg6);
	kvmppc_set_sprg7(vcpu, regs->sprg7);
A
Alexander Graf 已提交
588

589 590
	for (i = 0; i < ARRAY_SIZE(regs->gpr); i++)
		kvmppc_set_gpr(vcpu, i, regs->gpr[i]);
A
Alexander Graf 已提交
591 592 593 594 595 596 597 598 599 600 601 602 603 604

	return 0;
}

int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
{
	return -ENOTSUPP;
}

int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
{
	return -ENOTSUPP;
}

605 606
int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id,
			union kvmppc_one_reg *val)
607
{
608
	int r = 0;
609
	long int i;
610

611
	r = vcpu->kvm->arch.kvm_ops->get_one_reg(vcpu, id, val);
612 613
	if (r == -EINVAL) {
		r = 0;
614
		switch (id) {
615
		case KVM_REG_PPC_DAR:
616
			*val = get_reg_val(id, kvmppc_get_dar(vcpu));
617 618
			break;
		case KVM_REG_PPC_DSISR:
619
			*val = get_reg_val(id, kvmppc_get_dsisr(vcpu));
620
			break;
621
		case KVM_REG_PPC_FPR0 ... KVM_REG_PPC_FPR31:
622 623
			i = id - KVM_REG_PPC_FPR0;
			*val = get_reg_val(id, VCPU_FPR(vcpu, i));
624 625
			break;
		case KVM_REG_PPC_FPSCR:
626
			*val = get_reg_val(id, vcpu->arch.fp.fpscr);
627
			break;
628 629 630
#ifdef CONFIG_VSX
		case KVM_REG_PPC_VSR0 ... KVM_REG_PPC_VSR31:
			if (cpu_has_feature(CPU_FTR_VSX)) {
631 632 633
				i = id - KVM_REG_PPC_VSR0;
				val->vsxval[0] = vcpu->arch.fp.fpr[i][0];
				val->vsxval[1] = vcpu->arch.fp.fpr[i][1];
634 635 636 637 638
			} else {
				r = -ENXIO;
			}
			break;
#endif /* CONFIG_VSX */
639 640
		case KVM_REG_PPC_DEBUG_INST:
			*val = get_reg_val(id, INS_TW);
641
			break;
642 643
#ifdef CONFIG_KVM_XICS
		case KVM_REG_PPC_ICP_STATE:
644
			if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) {
645 646 647
				r = -ENXIO;
				break;
			}
648
			if (xics_on_xive())
649 650 651
				*val = get_reg_val(id, kvmppc_xive_get_icp(vcpu));
			else
				*val = get_reg_val(id, kvmppc_xics_get_icp(vcpu));
652 653
			break;
#endif /* CONFIG_KVM_XICS */
654
		case KVM_REG_PPC_FSCR:
655
			*val = get_reg_val(id, vcpu->arch.fscr);
656
			break;
657
		case KVM_REG_PPC_TAR:
658
			*val = get_reg_val(id, vcpu->arch.tar);
659
			break;
660
		case KVM_REG_PPC_EBBHR:
661
			*val = get_reg_val(id, vcpu->arch.ebbhr);
662 663
			break;
		case KVM_REG_PPC_EBBRR:
664
			*val = get_reg_val(id, vcpu->arch.ebbrr);
665 666
			break;
		case KVM_REG_PPC_BESCR:
667
			*val = get_reg_val(id, vcpu->arch.bescr);
668
			break;
669
		case KVM_REG_PPC_IC:
670
			*val = get_reg_val(id, vcpu->arch.ic);
671
			break;
672 673 674 675 676 677 678 679 680
		default:
			r = -EINVAL;
			break;
		}
	}

	return r;
}

681 682
int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id,
			union kvmppc_one_reg *val)
683
{
684
	int r = 0;
685
	long int i;
686

687
	r = vcpu->kvm->arch.kvm_ops->set_one_reg(vcpu, id, val);
688 689
	if (r == -EINVAL) {
		r = 0;
690
		switch (id) {
691
		case KVM_REG_PPC_DAR:
692
			kvmppc_set_dar(vcpu, set_reg_val(id, *val));
693 694
			break;
		case KVM_REG_PPC_DSISR:
695
			kvmppc_set_dsisr(vcpu, set_reg_val(id, *val));
696
			break;
697
		case KVM_REG_PPC_FPR0 ... KVM_REG_PPC_FPR31:
698 699
			i = id - KVM_REG_PPC_FPR0;
			VCPU_FPR(vcpu, i) = set_reg_val(id, *val);
700 701
			break;
		case KVM_REG_PPC_FPSCR:
702
			vcpu->arch.fp.fpscr = set_reg_val(id, *val);
703
			break;
704 705 706
#ifdef CONFIG_VSX
		case KVM_REG_PPC_VSR0 ... KVM_REG_PPC_VSR31:
			if (cpu_has_feature(CPU_FTR_VSX)) {
707 708 709
				i = id - KVM_REG_PPC_VSR0;
				vcpu->arch.fp.fpr[i][0] = val->vsxval[0];
				vcpu->arch.fp.fpr[i][1] = val->vsxval[1];
710 711 712 713 714
			} else {
				r = -ENXIO;
			}
			break;
#endif /* CONFIG_VSX */
715 716
#ifdef CONFIG_KVM_XICS
		case KVM_REG_PPC_ICP_STATE:
717
			if (!vcpu->arch.icp && !vcpu->arch.xive_vcpu) {
718 719 720
				r = -ENXIO;
				break;
			}
721
			if (xics_on_xive())
722 723 724
				r = kvmppc_xive_set_icp(vcpu, set_reg_val(id, *val));
			else
				r = kvmppc_xics_set_icp(vcpu, set_reg_val(id, *val));
725 726
			break;
#endif /* CONFIG_KVM_XICS */
727
		case KVM_REG_PPC_FSCR:
728
			vcpu->arch.fscr = set_reg_val(id, *val);
729
			break;
730
		case KVM_REG_PPC_TAR:
731
			vcpu->arch.tar = set_reg_val(id, *val);
732
			break;
733
		case KVM_REG_PPC_EBBHR:
734
			vcpu->arch.ebbhr = set_reg_val(id, *val);
735 736
			break;
		case KVM_REG_PPC_EBBRR:
737
			vcpu->arch.ebbrr = set_reg_val(id, *val);
738 739
			break;
		case KVM_REG_PPC_BESCR:
740
			vcpu->arch.bescr = set_reg_val(id, *val);
741
			break;
742
		case KVM_REG_PPC_IC:
743
			vcpu->arch.ic = set_reg_val(id, *val);
744
			break;
745 746 747 748 749 750 751 752 753
		default:
			r = -EINVAL;
			break;
		}
	}

	return r;
}

754 755
void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
{
756
	vcpu->kvm->arch.kvm_ops->vcpu_load(vcpu, cpu);
757 758 759 760
}

void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu)
{
761
	vcpu->kvm->arch.kvm_ops->vcpu_put(vcpu);
762 763 764 765
}

void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 msr)
{
766
	vcpu->kvm->arch.kvm_ops->set_msr(vcpu, msr);
767
}
768
EXPORT_SYMBOL_GPL(kvmppc_set_msr);
769 770 771

int kvmppc_vcpu_run(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
{
772
	return vcpu->kvm->arch.kvm_ops->vcpu_run(kvm_run, vcpu);
773 774
}

A
Alexander Graf 已提交
775 776 777 778 779 780
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
                                  struct kvm_translation *tr)
{
	return 0;
}

781 782 783
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
					struct kvm_guest_debug *dbg)
{
784
	vcpu_load(vcpu);
785
	vcpu->guest_debug = dbg->control;
786
	vcpu_put(vcpu);
787
	return 0;
788 789
}

790
void kvmppc_decrementer_func(struct kvm_vcpu *vcpu)
791 792 793 794
{
	kvmppc_core_queue_dec(vcpu);
	kvm_vcpu_kick(vcpu);
}
795 796 797

struct kvm_vcpu *kvmppc_core_vcpu_create(struct kvm *kvm, unsigned int id)
{
798
	return kvm->arch.kvm_ops->vcpu_create(kvm, id);
799 800 801 802
}

void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu)
{
803
	vcpu->kvm->arch.kvm_ops->vcpu_free(vcpu);
804 805 806 807
}

int kvmppc_core_check_requests(struct kvm_vcpu *vcpu)
{
808
	return vcpu->kvm->arch.kvm_ops->check_requests(vcpu);
809 810 811 812
}

int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log)
{
813
	return kvm->arch.kvm_ops->get_dirty_log(kvm, log);
814 815
}

816
void kvmppc_core_free_memslot(struct kvm *kvm, struct kvm_memory_slot *free,
817 818
			      struct kvm_memory_slot *dont)
{
819
	kvm->arch.kvm_ops->free_memslot(free, dont);
820 821
}

822
int kvmppc_core_create_memslot(struct kvm *kvm, struct kvm_memory_slot *slot,
823 824
			       unsigned long npages)
{
825
	return kvm->arch.kvm_ops->create_memslot(slot, npages);
826 827 828 829
}

void kvmppc_core_flush_memslot(struct kvm *kvm, struct kvm_memory_slot *memslot)
{
830
	kvm->arch.kvm_ops->flush_memslot(kvm, memslot);
831 832 833 834
}

int kvmppc_core_prepare_memory_region(struct kvm *kvm,
				struct kvm_memory_slot *memslot,
835
				const struct kvm_userspace_memory_region *mem)
836
{
837
	return kvm->arch.kvm_ops->prepare_memory_region(kvm, memslot, mem);
838 839 840
}

void kvmppc_core_commit_memory_region(struct kvm *kvm,
841
				const struct kvm_userspace_memory_region *mem,
842
				const struct kvm_memory_slot *old,
843 844
				const struct kvm_memory_slot *new,
				enum kvm_mr_change change)
845
{
846
	kvm->arch.kvm_ops->commit_memory_region(kvm, mem, old, new, change);
847 848 849 850
}

int kvm_unmap_hva_range(struct kvm *kvm, unsigned long start, unsigned long end)
{
851
	return kvm->arch.kvm_ops->unmap_hva_range(kvm, start, end);
852 853
}

A
Andres Lagar-Cavilla 已提交
854
int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end)
855
{
A
Andres Lagar-Cavilla 已提交
856
	return kvm->arch.kvm_ops->age_hva(kvm, start, end);
857 858 859 860
}

int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
{
861
	return kvm->arch.kvm_ops->test_age_hva(kvm, hva);
862 863
}

864
int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
865
{
866
	kvm->arch.kvm_ops->set_spte_hva(kvm, hva, pte);
867
	return 0;
868 869 870 871
}

void kvmppc_mmu_destroy(struct kvm_vcpu *vcpu)
{
872
	vcpu->kvm->arch.kvm_ops->mmu_destroy(vcpu);
873 874 875 876 877 878
}

int kvmppc_core_init_vm(struct kvm *kvm)
{

#ifdef CONFIG_PPC64
879
	INIT_LIST_HEAD_RCU(&kvm->arch.spapr_tce_tables);
880 881 882
	INIT_LIST_HEAD(&kvm->arch.rtas_tokens);
#endif

883
	return kvm->arch.kvm_ops->init_vm(kvm);
884 885 886 887
}

void kvmppc_core_destroy_vm(struct kvm *kvm)
{
888
	kvm->arch.kvm_ops->destroy_vm(kvm);
889 890 891 892 893 894 895

#ifdef CONFIG_PPC64
	kvmppc_rtas_tokens_free(kvm);
	WARN_ON(!list_empty(&kvm->arch.spapr_tce_tables));
#endif
}

896 897 898 899 900
int kvmppc_h_logical_ci_load(struct kvm_vcpu *vcpu)
{
	unsigned long size = kvmppc_get_gpr(vcpu, 4);
	unsigned long addr = kvmppc_get_gpr(vcpu, 5);
	u64 buf;
901
	int srcu_idx;
902 903 904 905 906
	int ret;

	if (!is_power_of_2(size) || (size > sizeof(buf)))
		return H_TOO_HARD;

907
	srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
908
	ret = kvm_io_bus_read(vcpu, KVM_MMIO_BUS, addr, size, &buf);
909
	srcu_read_unlock(&vcpu->kvm->srcu, srcu_idx);
910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943
	if (ret != 0)
		return H_TOO_HARD;

	switch (size) {
	case 1:
		kvmppc_set_gpr(vcpu, 4, *(u8 *)&buf);
		break;

	case 2:
		kvmppc_set_gpr(vcpu, 4, be16_to_cpu(*(__be16 *)&buf));
		break;

	case 4:
		kvmppc_set_gpr(vcpu, 4, be32_to_cpu(*(__be32 *)&buf));
		break;

	case 8:
		kvmppc_set_gpr(vcpu, 4, be64_to_cpu(*(__be64 *)&buf));
		break;

	default:
		BUG();
	}

	return H_SUCCESS;
}
EXPORT_SYMBOL_GPL(kvmppc_h_logical_ci_load);

int kvmppc_h_logical_ci_store(struct kvm_vcpu *vcpu)
{
	unsigned long size = kvmppc_get_gpr(vcpu, 4);
	unsigned long addr = kvmppc_get_gpr(vcpu, 5);
	unsigned long val = kvmppc_get_gpr(vcpu, 6);
	u64 buf;
944
	int srcu_idx;
945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967
	int ret;

	switch (size) {
	case 1:
		*(u8 *)&buf = val;
		break;

	case 2:
		*(__be16 *)&buf = cpu_to_be16(val);
		break;

	case 4:
		*(__be32 *)&buf = cpu_to_be32(val);
		break;

	case 8:
		*(__be64 *)&buf = cpu_to_be64(val);
		break;

	default:
		return H_TOO_HARD;
	}

968
	srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
969
	ret = kvm_io_bus_write(vcpu, KVM_MMIO_BUS, addr, size, &buf);
970
	srcu_read_unlock(&vcpu->kvm->srcu, srcu_idx);
971 972 973 974 975 976 977
	if (ret != 0)
		return H_TOO_HARD;

	return H_SUCCESS;
}
EXPORT_SYMBOL_GPL(kvmppc_h_logical_ci_store);

978 979
int kvmppc_core_check_processor_compat(void)
{
980 981
	/*
	 * We always return 0 for book3s. We check
982
	 * for compatibility while loading the HV
983 984 985 986 987
	 * or PR module
	 */
	return 0;
}

988 989 990 991 992
int kvmppc_book3s_hcall_implemented(struct kvm *kvm, unsigned long hcall)
{
	return kvm->arch.kvm_ops->hcall_implemented(hcall);
}

993 994 995 996
#ifdef CONFIG_KVM_XICS
int kvm_set_irq(struct kvm *kvm, int irq_source_id, u32 irq, int level,
		bool line_status)
{
997
	if (xics_on_xive())
998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036
		return kvmppc_xive_set_irq(kvm, irq_source_id, irq, level,
					   line_status);
	else
		return kvmppc_xics_set_irq(kvm, irq_source_id, irq, level,
					   line_status);
}

int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *irq_entry,
			      struct kvm *kvm, int irq_source_id,
			      int level, bool line_status)
{
	return kvm_set_irq(kvm, irq_source_id, irq_entry->gsi,
			   level, line_status);
}
static int kvmppc_book3s_set_irq(struct kvm_kernel_irq_routing_entry *e,
				 struct kvm *kvm, int irq_source_id, int level,
				 bool line_status)
{
	return kvm_set_irq(kvm, irq_source_id, e->gsi, level, line_status);
}

int kvm_irq_map_gsi(struct kvm *kvm,
		    struct kvm_kernel_irq_routing_entry *entries, int gsi)
{
	entries->gsi = gsi;
	entries->type = KVM_IRQ_ROUTING_IRQCHIP;
	entries->set = kvmppc_book3s_set_irq;
	entries->irqchip.irqchip = 0;
	entries->irqchip.pin = gsi;
	return 1;
}

int kvm_irq_map_chip_pin(struct kvm *kvm, unsigned irqchip, unsigned pin)
{
	return pin;
}

#endif /* CONFIG_KVM_XICS */

1037 1038 1039 1040 1041 1042 1043
static int kvmppc_book3s_init(void)
{
	int r;

	r = kvm_init(NULL, sizeof(struct kvm_vcpu), 0, THIS_MODULE);
	if (r)
		return r;
1044
#ifdef CONFIG_KVM_BOOK3S_32_HANDLER
1045 1046 1047
	r = kvmppc_book3s_init_pr();
#endif

1048 1049
#ifdef CONFIG_KVM_XICS
#ifdef CONFIG_KVM_XIVE
1050
	if (xics_on_xive()) {
1051 1052 1053 1054 1055 1056 1057
		kvmppc_xive_init_module();
		kvm_register_device_ops(&kvm_xive_ops, KVM_DEV_TYPE_XICS);
	} else
#endif
		kvm_register_device_ops(&kvm_xics_ops, KVM_DEV_TYPE_XICS);
#endif
	return r;
1058 1059 1060 1061
}

static void kvmppc_book3s_exit(void)
{
1062
#ifdef CONFIG_KVM_XICS
1063
	if (xics_on_xive())
1064 1065
		kvmppc_xive_exit_module();
#endif
1066
#ifdef CONFIG_KVM_BOOK3S_32_HANDLER
1067 1068 1069
	kvmppc_book3s_exit_pr();
#endif
	kvm_exit();
1070
}
1071 1072 1073

module_init(kvmppc_book3s_init);
module_exit(kvmppc_book3s_exit);
1074 1075

/* On 32bit this is our one and only kernel module */
1076
#ifdef CONFIG_KVM_BOOK3S_32_HANDLER
1077 1078 1079
MODULE_ALIAS_MISCDEV(KVM_MINOR);
MODULE_ALIAS("devname:kvm");
#endif