svm.c 52.3 KB
Newer Older
A
Avi Kivity 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
/*
 * Kernel-based Virtual Machine driver for Linux
 *
 * AMD SVM support
 *
 * Copyright (C) 2006 Qumranet, Inc.
 *
 * Authors:
 *   Yaniv Kamay  <yaniv@qumranet.com>
 *   Avi Kivity   <avi@qumranet.com>
 *
 * This work is licensed under the terms of the GNU GPL, version 2.  See
 * the COPYING file in the top-level directory.
 *
 */
16 17
#include <linux/kvm_host.h>

A
Avi Kivity 已提交
18
#include "kvm_svm.h"
19
#include "irq.h"
20
#include "mmu.h"
21
#include "kvm_cache_regs.h"
A
Avi Kivity 已提交
22

A
Avi Kivity 已提交
23
#include <linux/module.h>
24
#include <linux/kernel.h>
A
Avi Kivity 已提交
25 26
#include <linux/vmalloc.h>
#include <linux/highmem.h>
A
Alexey Dobriyan 已提交
27
#include <linux/sched.h>
A
Avi Kivity 已提交
28

A
Avi Kivity 已提交
29
#include <asm/desc.h>
A
Avi Kivity 已提交
30

31 32
#include <asm/virtext.h>

33 34
#define __ex(x) __kvm_handle_fault_on_reboot(x)

A
Avi Kivity 已提交
35 36 37 38 39 40 41 42 43 44 45 46
MODULE_AUTHOR("Qumranet");
MODULE_LICENSE("GPL");

#define IOPM_ALLOC_ORDER 2
#define MSRPM_ALLOC_ORDER 1

#define DR7_GD_MASK (1 << 13)
#define DR6_BD_MASK (1 << 13)

#define SEG_TYPE_LDT 2
#define SEG_TYPE_BUSY_TSS16 3

47 48
#define SVM_FEATURE_NPT  (1 << 0)
#define SVM_FEATURE_LBRV (1 << 1)
A
Amit Shah 已提交
49
#define SVM_FEATURE_SVML (1 << 2)
50

51 52
#define DEBUGCTL_RESERVED_BITS (~(0x3fULL))

53 54 55 56 57 58 59 60 61
/* Turn on to get debugging output*/
/* #define NESTED_DEBUG */

#ifdef NESTED_DEBUG
#define nsvm_printk(fmt, args...) printk(KERN_INFO fmt, ## args)
#else
#define nsvm_printk(fmt, args...) do {} while(0)
#endif

62 63 64 65
/* enable NPT for AMD64 and X86 with PAE */
#if defined(CONFIG_X86_64) || defined(CONFIG_X86_PAE)
static bool npt_enabled = true;
#else
66
static bool npt_enabled = false;
67
#endif
68 69 70
static int npt = 1;

module_param(npt, int, S_IRUGO);
71

72
static void kvm_reput_irq(struct vcpu_svm *svm);
73
static void svm_flush_tlb(struct kvm_vcpu *vcpu);
74

75 76
static inline struct vcpu_svm *to_svm(struct kvm_vcpu *vcpu)
{
R
Rusty Russell 已提交
77
	return container_of(vcpu, struct vcpu_svm, vcpu);
78 79
}

80
static unsigned long iopm_base;
A
Avi Kivity 已提交
81 82 83 84 85 86 87 88 89 90 91 92 93

struct kvm_ldttss_desc {
	u16 limit0;
	u16 base0;
	unsigned base1 : 8, type : 5, dpl : 2, p : 1;
	unsigned limit1 : 4, zero0 : 3, g : 1, base2 : 8;
	u32 base3;
	u32 zero1;
} __attribute__((packed));

struct svm_cpu_data {
	int cpu;

A
Avi Kivity 已提交
94 95 96
	u64 asid_generation;
	u32 max_asid;
	u32 next_asid;
A
Avi Kivity 已提交
97 98 99 100 101 102
	struct kvm_ldttss_desc *tss_desc;

	struct page *save_area;
};

static DEFINE_PER_CPU(struct svm_cpu_data *, svm_data);
103
static uint32_t svm_features;
A
Avi Kivity 已提交
104 105 106 107 108 109 110 111

struct svm_init_data {
	int cpu;
	int r;
};

static u32 msrpm_ranges[] = {0, 0xc0000000, 0xc0010000};

112
#define NUM_MSR_MAPS ARRAY_SIZE(msrpm_ranges)
A
Avi Kivity 已提交
113 114 115 116 117
#define MSRS_RANGE_SIZE 2048
#define MSRS_IN_RANGE (MSRS_RANGE_SIZE * 8 / 2)

#define MAX_INST_SIZE 15

118 119 120 121 122
static inline u32 svm_has(u32 feat)
{
	return svm_features & feat;
}

A
Avi Kivity 已提交
123 124
static inline u8 pop_irq(struct kvm_vcpu *vcpu)
{
125 126
	int word_index = __ffs(vcpu->arch.irq_summary);
	int bit_index = __ffs(vcpu->arch.irq_pending[word_index]);
A
Avi Kivity 已提交
127 128
	int irq = word_index * BITS_PER_LONG + bit_index;

129 130 131
	clear_bit(bit_index, &vcpu->arch.irq_pending[word_index]);
	if (!vcpu->arch.irq_pending[word_index])
		clear_bit(word_index, &vcpu->arch.irq_summary);
A
Avi Kivity 已提交
132 133 134 135 136
	return irq;
}

static inline void push_irq(struct kvm_vcpu *vcpu, u8 irq)
{
137 138
	set_bit(irq, vcpu->arch.irq_pending);
	set_bit(irq / BITS_PER_LONG, &vcpu->arch.irq_summary);
A
Avi Kivity 已提交
139 140 141 142
}

static inline void clgi(void)
{
143
	asm volatile (__ex(SVM_CLGI));
A
Avi Kivity 已提交
144 145 146 147
}

static inline void stgi(void)
{
148
	asm volatile (__ex(SVM_STGI));
A
Avi Kivity 已提交
149 150 151 152
}

static inline void invlpga(unsigned long addr, u32 asid)
{
153
	asm volatile (__ex(SVM_INVLPGA) :: "a"(addr), "c"(asid));
A
Avi Kivity 已提交
154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196
}

static inline unsigned long kvm_read_cr2(void)
{
	unsigned long cr2;

	asm volatile ("mov %%cr2, %0" : "=r" (cr2));
	return cr2;
}

static inline void kvm_write_cr2(unsigned long val)
{
	asm volatile ("mov %0, %%cr2" :: "r" (val));
}

static inline unsigned long read_dr6(void)
{
	unsigned long dr6;

	asm volatile ("mov %%dr6, %0" : "=r" (dr6));
	return dr6;
}

static inline void write_dr6(unsigned long val)
{
	asm volatile ("mov %0, %%dr6" :: "r" (val));
}

static inline unsigned long read_dr7(void)
{
	unsigned long dr7;

	asm volatile ("mov %%dr7, %0" : "=r" (dr7));
	return dr7;
}

static inline void write_dr7(unsigned long val)
{
	asm volatile ("mov %0, %%dr7" :: "r" (val));
}

static inline void force_new_asid(struct kvm_vcpu *vcpu)
{
197
	to_svm(vcpu)->asid_generation--;
A
Avi Kivity 已提交
198 199 200 201 202 203 204 205 206
}

static inline void flush_guest_tlb(struct kvm_vcpu *vcpu)
{
	force_new_asid(vcpu);
}

static void svm_set_efer(struct kvm_vcpu *vcpu, u64 efer)
{
207
	if (!npt_enabled && !(efer & EFER_LMA))
208
		efer &= ~EFER_LME;
A
Avi Kivity 已提交
209

210
	to_svm(vcpu)->vmcb->save.efer = efer | EFER_SVME;
211
	vcpu->arch.shadow_efer = efer;
A
Avi Kivity 已提交
212 213
}

214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232
static void svm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr,
				bool has_error_code, u32 error_code)
{
	struct vcpu_svm *svm = to_svm(vcpu);

	svm->vmcb->control.event_inj = nr
		| SVM_EVTINJ_VALID
		| (has_error_code ? SVM_EVTINJ_VALID_ERR : 0)
		| SVM_EVTINJ_TYPE_EXEPT;
	svm->vmcb->control.event_inj_err = error_code;
}

static bool svm_exception_injected(struct kvm_vcpu *vcpu)
{
	struct vcpu_svm *svm = to_svm(vcpu);

	return !(svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_VALID);
}

A
Avi Kivity 已提交
233 234 235 236 237 238 239 240
static int is_external_interrupt(u32 info)
{
	info &= SVM_EVTINJ_TYPE_MASK | SVM_EVTINJ_VALID;
	return info == (SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_INTR);
}

static void skip_emulated_instruction(struct kvm_vcpu *vcpu)
{
241 242 243
	struct vcpu_svm *svm = to_svm(vcpu);

	if (!svm->next_rip) {
244
		printk(KERN_DEBUG "%s: NOP\n", __func__);
A
Avi Kivity 已提交
245 246
		return;
	}
247 248 249
	if (svm->next_rip - kvm_rip_read(vcpu) > MAX_INST_SIZE)
		printk(KERN_ERR "%s: ip 0x%lx next 0x%llx\n",
		       __func__, kvm_rip_read(vcpu), svm->next_rip);
A
Avi Kivity 已提交
250

251
	kvm_rip_write(vcpu, svm->next_rip);
252
	svm->vmcb->control.int_state &= ~SVM_INTERRUPT_SHADOW_MASK;
253

254
	vcpu->arch.interrupt_window_open = (svm->vcpu.arch.hflags & HF_GIF_MASK);
A
Avi Kivity 已提交
255 256 257 258
}

static int has_svm(void)
{
259
	const char *msg;
A
Avi Kivity 已提交
260

261 262
	if (!cpu_has_svm(&msg)) {
		printk(KERN_INFO "has_svn: %s\n", msg);
A
Avi Kivity 已提交
263 264 265 266 267 268 269 270
		return 0;
	}

	return 1;
}

static void svm_hardware_disable(void *garbage)
{
271
	cpu_svm_disable();
A
Avi Kivity 已提交
272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298
}

static void svm_hardware_enable(void *garbage)
{

	struct svm_cpu_data *svm_data;
	uint64_t efer;
	struct desc_ptr gdt_descr;
	struct desc_struct *gdt;
	int me = raw_smp_processor_id();

	if (!has_svm()) {
		printk(KERN_ERR "svm_cpu_init: err EOPNOTSUPP on %d\n", me);
		return;
	}
	svm_data = per_cpu(svm_data, me);

	if (!svm_data) {
		printk(KERN_ERR "svm_cpu_init: svm_data is NULL on %d\n",
		       me);
		return;
	}

	svm_data->asid_generation = 1;
	svm_data->max_asid = cpuid_ebx(SVM_CPUID_FUNC) - 1;
	svm_data->next_asid = svm_data->max_asid + 1;

M
Mike Day 已提交
299
	asm volatile ("sgdt %0" : "=m"(gdt_descr));
A
Avi Kivity 已提交
300 301 302 303
	gdt = (struct desc_struct *)gdt_descr.address;
	svm_data->tss_desc = (struct kvm_ldttss_desc *)(gdt + GDT_ENTRY_TSS);

	rdmsrl(MSR_EFER, efer);
304
	wrmsrl(MSR_EFER, efer | EFER_SVME);
A
Avi Kivity 已提交
305 306 307 308 309

	wrmsrl(MSR_VM_HSAVE_PA,
	       page_to_pfn(svm_data->save_area) << PAGE_SHIFT);
}

310 311 312 313 314 315 316 317 318 319 320 321 322
static void svm_cpu_uninit(int cpu)
{
	struct svm_cpu_data *svm_data
		= per_cpu(svm_data, raw_smp_processor_id());

	if (!svm_data)
		return;

	per_cpu(svm_data, raw_smp_processor_id()) = NULL;
	__free_page(svm_data->save_area);
	kfree(svm_data);
}

A
Avi Kivity 已提交
323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346
static int svm_cpu_init(int cpu)
{
	struct svm_cpu_data *svm_data;
	int r;

	svm_data = kzalloc(sizeof(struct svm_cpu_data), GFP_KERNEL);
	if (!svm_data)
		return -ENOMEM;
	svm_data->cpu = cpu;
	svm_data->save_area = alloc_page(GFP_KERNEL);
	r = -ENOMEM;
	if (!svm_data->save_area)
		goto err_1;

	per_cpu(svm_data, cpu) = svm_data;

	return 0;

err_1:
	kfree(svm_data);
	return r;

}

347 348
static void set_msr_interception(u32 *msrpm, unsigned msr,
				 int read, int write)
A
Avi Kivity 已提交
349 350 351 352 353 354 355 356 357 358 359 360 361 362
{
	int i;

	for (i = 0; i < NUM_MSR_MAPS; i++) {
		if (msr >= msrpm_ranges[i] &&
		    msr < msrpm_ranges[i] + MSRS_IN_RANGE) {
			u32 msr_offset = (i * MSRS_IN_RANGE + msr -
					  msrpm_ranges[i]) * 2;

			u32 *base = msrpm + (msr_offset / 32);
			u32 msr_shift = msr_offset % 32;
			u32 mask = ((write) ? 0 : 2) | ((read) ? 0 : 1);
			*base = (*base & ~(0x3 << msr_shift)) |
				(mask << msr_shift);
363
			return;
A
Avi Kivity 已提交
364 365
		}
	}
366
	BUG();
A
Avi Kivity 已提交
367 368
}

369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386
static void svm_vcpu_init_msrpm(u32 *msrpm)
{
	memset(msrpm, 0xff, PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER));

#ifdef CONFIG_X86_64
	set_msr_interception(msrpm, MSR_GS_BASE, 1, 1);
	set_msr_interception(msrpm, MSR_FS_BASE, 1, 1);
	set_msr_interception(msrpm, MSR_KERNEL_GS_BASE, 1, 1);
	set_msr_interception(msrpm, MSR_LSTAR, 1, 1);
	set_msr_interception(msrpm, MSR_CSTAR, 1, 1);
	set_msr_interception(msrpm, MSR_SYSCALL_MASK, 1, 1);
#endif
	set_msr_interception(msrpm, MSR_K6_STAR, 1, 1);
	set_msr_interception(msrpm, MSR_IA32_SYSENTER_CS, 1, 1);
	set_msr_interception(msrpm, MSR_IA32_SYSENTER_ESP, 1, 1);
	set_msr_interception(msrpm, MSR_IA32_SYSENTER_EIP, 1, 1);
}

387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408
static void svm_enable_lbrv(struct vcpu_svm *svm)
{
	u32 *msrpm = svm->msrpm;

	svm->vmcb->control.lbr_ctl = 1;
	set_msr_interception(msrpm, MSR_IA32_LASTBRANCHFROMIP, 1, 1);
	set_msr_interception(msrpm, MSR_IA32_LASTBRANCHTOIP, 1, 1);
	set_msr_interception(msrpm, MSR_IA32_LASTINTFROMIP, 1, 1);
	set_msr_interception(msrpm, MSR_IA32_LASTINTTOIP, 1, 1);
}

static void svm_disable_lbrv(struct vcpu_svm *svm)
{
	u32 *msrpm = svm->msrpm;

	svm->vmcb->control.lbr_ctl = 0;
	set_msr_interception(msrpm, MSR_IA32_LASTBRANCHFROMIP, 0, 0);
	set_msr_interception(msrpm, MSR_IA32_LASTBRANCHTOIP, 0, 0);
	set_msr_interception(msrpm, MSR_IA32_LASTINTFROMIP, 0, 0);
	set_msr_interception(msrpm, MSR_IA32_LASTINTTOIP, 0, 0);
}

A
Avi Kivity 已提交
409 410 411 412
static __init int svm_hardware_setup(void)
{
	int cpu;
	struct page *iopm_pages;
413
	void *iopm_va;
A
Avi Kivity 已提交
414 415 416 417 418 419
	int r;

	iopm_pages = alloc_pages(GFP_KERNEL, IOPM_ALLOC_ORDER);

	if (!iopm_pages)
		return -ENOMEM;
420 421 422 423

	iopm_va = page_address(iopm_pages);
	memset(iopm_va, 0xff, PAGE_SIZE * (1 << IOPM_ALLOC_ORDER));
	clear_bit(0x80, iopm_va); /* allow direct access to PC debug port */
A
Avi Kivity 已提交
424 425
	iopm_base = page_to_pfn(iopm_pages) << PAGE_SHIFT;

426 427 428
	if (boot_cpu_has(X86_FEATURE_NX))
		kvm_enable_efer_bits(EFER_NX);

A
Avi Kivity 已提交
429 430 431
	for_each_online_cpu(cpu) {
		r = svm_cpu_init(cpu);
		if (r)
432
			goto err;
A
Avi Kivity 已提交
433
	}
434 435 436

	svm_features = cpuid_edx(SVM_CPUID_FUNC);

437 438 439
	if (!svm_has(SVM_FEATURE_NPT))
		npt_enabled = false;

440 441 442 443 444
	if (npt_enabled && !npt) {
		printk(KERN_INFO "kvm: Nested Paging disabled\n");
		npt_enabled = false;
	}

445
	if (npt_enabled) {
446
		printk(KERN_INFO "kvm: Nested Paging enabled\n");
447
		kvm_enable_tdp();
448 449
	} else
		kvm_disable_tdp();
450

A
Avi Kivity 已提交
451 452
	return 0;

453
err:
A
Avi Kivity 已提交
454 455 456 457 458 459 460
	__free_pages(iopm_pages, IOPM_ALLOC_ORDER);
	iopm_base = 0;
	return r;
}

static __exit void svm_hardware_unsetup(void)
{
461 462 463 464 465
	int cpu;

	for_each_online_cpu(cpu)
		svm_cpu_uninit(cpu);

A
Avi Kivity 已提交
466
	__free_pages(pfn_to_page(iopm_base >> PAGE_SHIFT), IOPM_ALLOC_ORDER);
467
	iopm_base = 0;
A
Avi Kivity 已提交
468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486
}

static void init_seg(struct vmcb_seg *seg)
{
	seg->selector = 0;
	seg->attrib = SVM_SELECTOR_P_MASK | SVM_SELECTOR_S_MASK |
		SVM_SELECTOR_WRITE_MASK; /* Read/Write Data Segment */
	seg->limit = 0xffff;
	seg->base = 0;
}

static void init_sys_seg(struct vmcb_seg *seg, uint32_t type)
{
	seg->selector = 0;
	seg->attrib = SVM_SELECTOR_P_MASK | type;
	seg->limit = 0xffff;
	seg->base = 0;
}

487
static void init_vmcb(struct vcpu_svm *svm)
A
Avi Kivity 已提交
488
{
489 490
	struct vmcb_control_area *control = &svm->vmcb->control;
	struct vmcb_save_area *save = &svm->vmcb->save;
A
Avi Kivity 已提交
491 492 493

	control->intercept_cr_read = 	INTERCEPT_CR0_MASK |
					INTERCEPT_CR3_MASK |
494
					INTERCEPT_CR4_MASK;
A
Avi Kivity 已提交
495 496 497

	control->intercept_cr_write = 	INTERCEPT_CR0_MASK |
					INTERCEPT_CR3_MASK |
498 499
					INTERCEPT_CR4_MASK |
					INTERCEPT_CR8_MASK;
A
Avi Kivity 已提交
500 501 502 503 504 505 506 507 508 509 510 511 512

	control->intercept_dr_read = 	INTERCEPT_DR0_MASK |
					INTERCEPT_DR1_MASK |
					INTERCEPT_DR2_MASK |
					INTERCEPT_DR3_MASK;

	control->intercept_dr_write = 	INTERCEPT_DR0_MASK |
					INTERCEPT_DR1_MASK |
					INTERCEPT_DR2_MASK |
					INTERCEPT_DR3_MASK |
					INTERCEPT_DR5_MASK |
					INTERCEPT_DR7_MASK;

513
	control->intercept_exceptions = (1 << PF_VECTOR) |
514 515
					(1 << UD_VECTOR) |
					(1 << MC_VECTOR);
A
Avi Kivity 已提交
516 517 518 519


	control->intercept = 	(1ULL << INTERCEPT_INTR) |
				(1ULL << INTERCEPT_NMI) |
520
				(1ULL << INTERCEPT_SMI) |
A
Avi Kivity 已提交
521
				(1ULL << INTERCEPT_CPUID) |
522
				(1ULL << INTERCEPT_INVD) |
A
Avi Kivity 已提交
523
				(1ULL << INTERCEPT_HLT) |
M
Marcelo Tosatti 已提交
524
				(1ULL << INTERCEPT_INVLPG) |
A
Avi Kivity 已提交
525 526 527 528
				(1ULL << INTERCEPT_INVLPGA) |
				(1ULL << INTERCEPT_IOIO_PROT) |
				(1ULL << INTERCEPT_MSR_PROT) |
				(1ULL << INTERCEPT_TASK_SWITCH) |
529
				(1ULL << INTERCEPT_SHUTDOWN) |
A
Avi Kivity 已提交
530 531 532 533 534 535
				(1ULL << INTERCEPT_VMRUN) |
				(1ULL << INTERCEPT_VMMCALL) |
				(1ULL << INTERCEPT_VMLOAD) |
				(1ULL << INTERCEPT_VMSAVE) |
				(1ULL << INTERCEPT_STGI) |
				(1ULL << INTERCEPT_CLGI) |
536
				(1ULL << INTERCEPT_SKINIT) |
537
				(1ULL << INTERCEPT_WBINVD) |
538 539
				(1ULL << INTERCEPT_MONITOR) |
				(1ULL << INTERCEPT_MWAIT);
A
Avi Kivity 已提交
540 541

	control->iopm_base_pa = iopm_base;
542
	control->msrpm_base_pa = __pa(svm->msrpm);
543
	control->tsc_offset = 0;
A
Avi Kivity 已提交
544 545 546 547 548 549 550 551 552 553 554 555 556
	control->int_ctl = V_INTR_MASKING_MASK;

	init_seg(&save->es);
	init_seg(&save->ss);
	init_seg(&save->ds);
	init_seg(&save->fs);
	init_seg(&save->gs);

	save->cs.selector = 0xf000;
	/* Executable/Readable Code Segment */
	save->cs.attrib = SVM_SELECTOR_READ_MASK | SVM_SELECTOR_P_MASK |
		SVM_SELECTOR_S_MASK | SVM_SELECTOR_CODE_MASK;
	save->cs.limit = 0xffff;
557 558 559 560 561 562 563
	/*
	 * cs.base should really be 0xffff0000, but vmx can't handle that, so
	 * be consistent with it.
	 *
	 * Replace when we have real mode working for vmx.
	 */
	save->cs.base = 0xf0000;
A
Avi Kivity 已提交
564 565 566 567 568 569 570

	save->gdtr.limit = 0xffff;
	save->idtr.limit = 0xffff;

	init_sys_seg(&save->ldtr, SEG_TYPE_LDT);
	init_sys_seg(&save->tr, SEG_TYPE_BUSY_TSS16);

571
	save->efer = EFER_SVME;
M
Mike Day 已提交
572
	save->dr6 = 0xffff0ff0;
A
Avi Kivity 已提交
573 574 575
	save->dr7 = 0x400;
	save->rflags = 2;
	save->rip = 0x0000fff0;
576
	svm->vcpu.arch.regs[VCPU_REGS_RIP] = save->rip;
A
Avi Kivity 已提交
577 578 579 580 581

	/*
	 * cr0 val on cpu init should be 0x60000010, we enable cpu
	 * cache by default. the orderly way is to enable cache in bios.
	 */
582
	save->cr0 = 0x00000010 | X86_CR0_PG | X86_CR0_WP;
583
	save->cr4 = X86_CR4_PAE;
A
Avi Kivity 已提交
584
	/* rdx = ?? */
585 586 587 588

	if (npt_enabled) {
		/* Setup VMCB for Nested Paging */
		control->nested_ctl = 1;
M
Marcelo Tosatti 已提交
589 590
		control->intercept &= ~((1ULL << INTERCEPT_TASK_SWITCH) |
					(1ULL << INTERCEPT_INVLPG));
591 592 593 594 595 596 597 598 599 600 601
		control->intercept_exceptions &= ~(1 << PF_VECTOR);
		control->intercept_cr_read &= ~(INTERCEPT_CR0_MASK|
						INTERCEPT_CR3_MASK);
		control->intercept_cr_write &= ~(INTERCEPT_CR0_MASK|
						 INTERCEPT_CR3_MASK);
		save->g_pat = 0x0007040600070406ULL;
		/* enable caching because the QEMU Bios doesn't enable it */
		save->cr0 = X86_CR0_ET;
		save->cr3 = 0;
		save->cr4 = 0;
	}
602
	force_new_asid(&svm->vcpu);
603 604

	svm->vcpu.arch.hflags = HF_GIF_MASK;
A
Avi Kivity 已提交
605 606
}

607
static int svm_vcpu_reset(struct kvm_vcpu *vcpu)
608 609 610
{
	struct vcpu_svm *svm = to_svm(vcpu);

611
	init_vmcb(svm);
A
Avi Kivity 已提交
612 613

	if (vcpu->vcpu_id != 0) {
614
		kvm_rip_write(vcpu, 0);
615 616
		svm->vmcb->save.cs.base = svm->vcpu.arch.sipi_vector << 12;
		svm->vmcb->save.cs.selector = svm->vcpu.arch.sipi_vector << 8;
A
Avi Kivity 已提交
617
	}
618 619
	vcpu->arch.regs_avail = ~0;
	vcpu->arch.regs_dirty = ~0;
620 621

	return 0;
622 623
}

R
Rusty Russell 已提交
624
static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id)
A
Avi Kivity 已提交
625
{
626
	struct vcpu_svm *svm;
A
Avi Kivity 已提交
627
	struct page *page;
628
	struct page *msrpm_pages;
A
Alexander Graf 已提交
629
	struct page *hsave_page;
R
Rusty Russell 已提交
630
	int err;
A
Avi Kivity 已提交
631

632
	svm = kmem_cache_zalloc(kvm_vcpu_cache, GFP_KERNEL);
R
Rusty Russell 已提交
633 634 635 636 637 638 639 640 641
	if (!svm) {
		err = -ENOMEM;
		goto out;
	}

	err = kvm_vcpu_init(&svm->vcpu, kvm, id);
	if (err)
		goto free_svm;

A
Avi Kivity 已提交
642
	page = alloc_page(GFP_KERNEL);
R
Rusty Russell 已提交
643 644 645 646
	if (!page) {
		err = -ENOMEM;
		goto uninit;
	}
A
Avi Kivity 已提交
647

648 649 650 651 652 653 654
	err = -ENOMEM;
	msrpm_pages = alloc_pages(GFP_KERNEL, MSRPM_ALLOC_ORDER);
	if (!msrpm_pages)
		goto uninit;
	svm->msrpm = page_address(msrpm_pages);
	svm_vcpu_init_msrpm(svm->msrpm);

A
Alexander Graf 已提交
655 656 657 658 659
	hsave_page = alloc_page(GFP_KERNEL);
	if (!hsave_page)
		goto uninit;
	svm->hsave = page_address(hsave_page);

660 661 662 663 664
	svm->vmcb = page_address(page);
	clear_page(svm->vmcb);
	svm->vmcb_pa = page_to_pfn(page) << PAGE_SHIFT;
	svm->asid_generation = 0;
	memset(svm->db_regs, 0, sizeof(svm->db_regs));
665
	init_vmcb(svm);
666

R
Rusty Russell 已提交
667 668
	fx_init(&svm->vcpu);
	svm->vcpu.fpu_active = 1;
669
	svm->vcpu.arch.apic_base = 0xfee00000 | MSR_IA32_APICBASE_ENABLE;
R
Rusty Russell 已提交
670
	if (svm->vcpu.vcpu_id == 0)
671
		svm->vcpu.arch.apic_base |= MSR_IA32_APICBASE_BSP;
A
Avi Kivity 已提交
672

R
Rusty Russell 已提交
673
	return &svm->vcpu;
674

R
Rusty Russell 已提交
675 676 677
uninit:
	kvm_vcpu_uninit(&svm->vcpu);
free_svm:
678
	kmem_cache_free(kvm_vcpu_cache, svm);
R
Rusty Russell 已提交
679 680
out:
	return ERR_PTR(err);
A
Avi Kivity 已提交
681 682 683 684
}

static void svm_free_vcpu(struct kvm_vcpu *vcpu)
{
685 686
	struct vcpu_svm *svm = to_svm(vcpu);

R
Rusty Russell 已提交
687
	__free_page(pfn_to_page(svm->vmcb_pa >> PAGE_SHIFT));
688
	__free_pages(virt_to_page(svm->msrpm), MSRPM_ALLOC_ORDER);
A
Alexander Graf 已提交
689
	__free_page(virt_to_page(svm->hsave));
R
Rusty Russell 已提交
690
	kvm_vcpu_uninit(vcpu);
691
	kmem_cache_free(kvm_vcpu_cache, svm);
A
Avi Kivity 已提交
692 693
}

694
static void svm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
A
Avi Kivity 已提交
695
{
696
	struct vcpu_svm *svm = to_svm(vcpu);
697
	int i;
698 699 700 701 702 703 704 705 706

	if (unlikely(cpu != vcpu->cpu)) {
		u64 tsc_this, delta;

		/*
		 * Make sure that the guest sees a monotonically
		 * increasing TSC.
		 */
		rdtscll(tsc_this);
707
		delta = vcpu->arch.host_tsc - tsc_this;
708
		svm->vmcb->control.tsc_offset += delta;
709
		vcpu->cpu = cpu;
M
Marcelo Tosatti 已提交
710
		kvm_migrate_timers(vcpu);
711
	}
712 713

	for (i = 0; i < NR_HOST_SAVE_USER_MSRS; i++)
714
		rdmsrl(host_save_user_msrs[i], svm->host_user_msrs[i]);
A
Avi Kivity 已提交
715 716 717 718
}

static void svm_vcpu_put(struct kvm_vcpu *vcpu)
{
719
	struct vcpu_svm *svm = to_svm(vcpu);
720 721
	int i;

722
	++vcpu->stat.host_state_reload;
723
	for (i = 0; i < NR_HOST_SAVE_USER_MSRS; i++)
724
		wrmsrl(host_save_user_msrs[i], svm->host_user_msrs[i]);
725

726
	rdtscll(vcpu->arch.host_tsc);
A
Avi Kivity 已提交
727 728 729 730
}

static unsigned long svm_get_rflags(struct kvm_vcpu *vcpu)
{
731
	return to_svm(vcpu)->vmcb->save.rflags;
A
Avi Kivity 已提交
732 733 734 735
}

static void svm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
{
736
	to_svm(vcpu)->vmcb->save.rflags = rflags;
A
Avi Kivity 已提交
737 738
}

739 740 741 742 743 744 745 746 747 748
static void svm_set_vintr(struct vcpu_svm *svm)
{
	svm->vmcb->control.intercept |= 1ULL << INTERCEPT_VINTR;
}

static void svm_clear_vintr(struct vcpu_svm *svm)
{
	svm->vmcb->control.intercept &= ~(1ULL << INTERCEPT_VINTR);
}

A
Avi Kivity 已提交
749 750
static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg)
{
751
	struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save;
A
Avi Kivity 已提交
752 753 754 755 756 757 758 759 760 761 762 763

	switch (seg) {
	case VCPU_SREG_CS: return &save->cs;
	case VCPU_SREG_DS: return &save->ds;
	case VCPU_SREG_ES: return &save->es;
	case VCPU_SREG_FS: return &save->fs;
	case VCPU_SREG_GS: return &save->gs;
	case VCPU_SREG_SS: return &save->ss;
	case VCPU_SREG_TR: return &save->tr;
	case VCPU_SREG_LDTR: return &save->ldtr;
	}
	BUG();
A
Al Viro 已提交
764
	return NULL;
A
Avi Kivity 已提交
765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789
}

static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg)
{
	struct vmcb_seg *s = svm_seg(vcpu, seg);

	return s->base;
}

static void svm_get_segment(struct kvm_vcpu *vcpu,
			    struct kvm_segment *var, int seg)
{
	struct vmcb_seg *s = svm_seg(vcpu, seg);

	var->base = s->base;
	var->limit = s->limit;
	var->selector = s->selector;
	var->type = s->attrib & SVM_SELECTOR_TYPE_MASK;
	var->s = (s->attrib >> SVM_SELECTOR_S_SHIFT) & 1;
	var->dpl = (s->attrib >> SVM_SELECTOR_DPL_SHIFT) & 3;
	var->present = (s->attrib >> SVM_SELECTOR_P_SHIFT) & 1;
	var->avl = (s->attrib >> SVM_SELECTOR_AVL_SHIFT) & 1;
	var->l = (s->attrib >> SVM_SELECTOR_L_SHIFT) & 1;
	var->db = (s->attrib >> SVM_SELECTOR_DB_SHIFT) & 1;
	var->g = (s->attrib >> SVM_SELECTOR_G_SHIFT) & 1;
790 791 792 793 794 795 796 797 798

	/*
	 * SVM always stores 0 for the 'G' bit in the CS selector in
	 * the VMCB on a VMEXIT. This hurts cross-vendor migration:
	 * Intel's VMENTRY has a check on the 'G' bit.
	 */
	if (seg == VCPU_SREG_CS)
		var->g = s->limit > 0xfffff;

799 800 801 802 803 804 805
	/*
	 * Work around a bug where the busy flag in the tr selector
	 * isn't exposed
	 */
	if (seg == VCPU_SREG_TR)
		var->type |= 0x2;

A
Avi Kivity 已提交
806 807 808
	var->unusable = !var->present;
}

809 810 811 812 813 814 815
static int svm_get_cpl(struct kvm_vcpu *vcpu)
{
	struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save;

	return save->cpl;
}

A
Avi Kivity 已提交
816 817
static void svm_get_idt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
{
818 819 820 821
	struct vcpu_svm *svm = to_svm(vcpu);

	dt->limit = svm->vmcb->save.idtr.limit;
	dt->base = svm->vmcb->save.idtr.base;
A
Avi Kivity 已提交
822 823 824 825
}

static void svm_set_idt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
{
826 827 828 829
	struct vcpu_svm *svm = to_svm(vcpu);

	svm->vmcb->save.idtr.limit = dt->limit;
	svm->vmcb->save.idtr.base = dt->base ;
A
Avi Kivity 已提交
830 831 832 833
}

static void svm_get_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
{
834 835 836 837
	struct vcpu_svm *svm = to_svm(vcpu);

	dt->limit = svm->vmcb->save.gdtr.limit;
	dt->base = svm->vmcb->save.gdtr.base;
A
Avi Kivity 已提交
838 839 840 841
}

static void svm_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
{
842 843 844 845
	struct vcpu_svm *svm = to_svm(vcpu);

	svm->vmcb->save.gdtr.limit = dt->limit;
	svm->vmcb->save.gdtr.base = dt->base ;
A
Avi Kivity 已提交
846 847
}

848
static void svm_decache_cr4_guest_bits(struct kvm_vcpu *vcpu)
849 850 851
{
}

A
Avi Kivity 已提交
852 853
static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
{
854 855
	struct vcpu_svm *svm = to_svm(vcpu);

856
#ifdef CONFIG_X86_64
857
	if (vcpu->arch.shadow_efer & EFER_LME) {
858
		if (!is_paging(vcpu) && (cr0 & X86_CR0_PG)) {
859
			vcpu->arch.shadow_efer |= EFER_LMA;
860
			svm->vmcb->save.efer |= EFER_LMA | EFER_LME;
A
Avi Kivity 已提交
861 862
		}

M
Mike Day 已提交
863
		if (is_paging(vcpu) && !(cr0 & X86_CR0_PG)) {
864
			vcpu->arch.shadow_efer &= ~EFER_LMA;
865
			svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME);
A
Avi Kivity 已提交
866 867 868
		}
	}
#endif
869 870 871
	if (npt_enabled)
		goto set;

872
	if ((vcpu->arch.cr0 & X86_CR0_TS) && !(cr0 & X86_CR0_TS)) {
873
		svm->vmcb->control.intercept_exceptions &= ~(1 << NM_VECTOR);
A
Anthony Liguori 已提交
874 875 876
		vcpu->fpu_active = 1;
	}

877
	vcpu->arch.cr0 = cr0;
878
	cr0 |= X86_CR0_PG | X86_CR0_WP;
879 880
	if (!vcpu->fpu_active) {
		svm->vmcb->control.intercept_exceptions |= (1 << NM_VECTOR);
J
Joerg Roedel 已提交
881
		cr0 |= X86_CR0_TS;
882
	}
883 884 885 886 887 888 889
set:
	/*
	 * re-enable caching here because the QEMU bios
	 * does not do it - this results in some delay at
	 * reboot
	 */
	cr0 &= ~(X86_CR0_CD | X86_CR0_NW);
890
	svm->vmcb->save.cr0 = cr0;
A
Avi Kivity 已提交
891 892 893 894
}

static void svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
{
895
	unsigned long host_cr4_mce = read_cr4() & X86_CR4_MCE;
896 897 898 899
	unsigned long old_cr4 = to_svm(vcpu)->vmcb->save.cr4;

	if (npt_enabled && ((old_cr4 ^ cr4) & X86_CR4_PGE))
		force_new_asid(vcpu);
900

901 902 903
	vcpu->arch.cr4 = cr4;
	if (!npt_enabled)
		cr4 |= X86_CR4_PAE;
904
	cr4 |= host_cr4_mce;
905
	to_svm(vcpu)->vmcb->save.cr4 = cr4;
A
Avi Kivity 已提交
906 907 908 909 910
}

static void svm_set_segment(struct kvm_vcpu *vcpu,
			    struct kvm_segment *var, int seg)
{
911
	struct vcpu_svm *svm = to_svm(vcpu);
A
Avi Kivity 已提交
912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929
	struct vmcb_seg *s = svm_seg(vcpu, seg);

	s->base = var->base;
	s->limit = var->limit;
	s->selector = var->selector;
	if (var->unusable)
		s->attrib = 0;
	else {
		s->attrib = (var->type & SVM_SELECTOR_TYPE_MASK);
		s->attrib |= (var->s & 1) << SVM_SELECTOR_S_SHIFT;
		s->attrib |= (var->dpl & 3) << SVM_SELECTOR_DPL_SHIFT;
		s->attrib |= (var->present & 1) << SVM_SELECTOR_P_SHIFT;
		s->attrib |= (var->avl & 1) << SVM_SELECTOR_AVL_SHIFT;
		s->attrib |= (var->l & 1) << SVM_SELECTOR_L_SHIFT;
		s->attrib |= (var->db & 1) << SVM_SELECTOR_DB_SHIFT;
		s->attrib |= (var->g & 1) << SVM_SELECTOR_G_SHIFT;
	}
	if (seg == VCPU_SREG_CS)
930 931
		svm->vmcb->save.cpl
			= (svm->vmcb->save.cs.attrib
A
Avi Kivity 已提交
932 933 934 935 936 937 938 939 940
			   >> SVM_SELECTOR_DPL_SHIFT) & 3;

}

static int svm_guest_debug(struct kvm_vcpu *vcpu, struct kvm_debug_guest *dbg)
{
	return -EOPNOTSUPP;
}

E
Eddie Dong 已提交
941 942 943 944 945 946 947 948 949 950
static int svm_get_irq(struct kvm_vcpu *vcpu)
{
	struct vcpu_svm *svm = to_svm(vcpu);
	u32 exit_int_info = svm->vmcb->control.exit_int_info;

	if (is_external_interrupt(exit_int_info))
		return exit_int_info & SVM_EVTINJ_VEC_MASK;
	return -1;
}

A
Avi Kivity 已提交
951 952
static void load_host_msrs(struct kvm_vcpu *vcpu)
{
953
#ifdef CONFIG_X86_64
954
	wrmsrl(MSR_GS_BASE, to_svm(vcpu)->host_gs_base);
955
#endif
A
Avi Kivity 已提交
956 957 958 959
}

static void save_host_msrs(struct kvm_vcpu *vcpu)
{
960
#ifdef CONFIG_X86_64
961
	rdmsrl(MSR_GS_BASE, to_svm(vcpu)->host_gs_base);
962
#endif
A
Avi Kivity 已提交
963 964
}

R
Rusty Russell 已提交
965
static void new_asid(struct vcpu_svm *svm, struct svm_cpu_data *svm_data)
A
Avi Kivity 已提交
966 967 968 969
{
	if (svm_data->next_asid > svm_data->max_asid) {
		++svm_data->asid_generation;
		svm_data->next_asid = 1;
970
		svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ALL_ASID;
A
Avi Kivity 已提交
971 972
	}

R
Rusty Russell 已提交
973
	svm->vcpu.cpu = svm_data->cpu;
974 975
	svm->asid_generation = svm_data->asid_generation;
	svm->vmcb->control.asid = svm_data->next_asid++;
A
Avi Kivity 已提交
976 977 978 979
}

static unsigned long svm_get_dr(struct kvm_vcpu *vcpu, int dr)
{
980 981 982
	unsigned long val = to_svm(vcpu)->db_regs[dr];
	KVMTRACE_2D(DR_READ, vcpu, (u32)dr, (u32)val, handler);
	return val;
A
Avi Kivity 已提交
983 984 985 986 987
}

static void svm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long value,
		       int *exception)
{
988 989
	struct vcpu_svm *svm = to_svm(vcpu);

A
Avi Kivity 已提交
990 991
	*exception = 0;

992 993 994
	if (svm->vmcb->save.dr7 & DR7_GD_MASK) {
		svm->vmcb->save.dr7 &= ~DR7_GD_MASK;
		svm->vmcb->save.dr6 |= DR6_BD_MASK;
A
Avi Kivity 已提交
995 996 997 998 999 1000
		*exception = DB_VECTOR;
		return;
	}

	switch (dr) {
	case 0 ... 3:
1001
		svm->db_regs[dr] = value;
A
Avi Kivity 已提交
1002 1003
		return;
	case 4 ... 5:
1004
		if (vcpu->arch.cr4 & X86_CR4_DE) {
A
Avi Kivity 已提交
1005 1006 1007 1008 1009 1010 1011 1012
			*exception = UD_VECTOR;
			return;
		}
	case 7: {
		if (value & ~((1ULL << 32) - 1)) {
			*exception = GP_VECTOR;
			return;
		}
1013
		svm->vmcb->save.dr7 = value;
A
Avi Kivity 已提交
1014 1015 1016 1017
		return;
	}
	default:
		printk(KERN_DEBUG "%s: unexpected dr %u\n",
1018
		       __func__, dr);
A
Avi Kivity 已提交
1019 1020 1021 1022 1023
		*exception = UD_VECTOR;
		return;
	}
}

R
Rusty Russell 已提交
1024
static int pf_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1025
{
1026
	u32 exit_int_info = svm->vmcb->control.exit_int_info;
R
Rusty Russell 已提交
1027
	struct kvm *kvm = svm->vcpu.kvm;
A
Avi Kivity 已提交
1028 1029
	u64 fault_address;
	u32 error_code;
1030
	bool event_injection = false;
A
Avi Kivity 已提交
1031

1032
	if (!irqchip_in_kernel(kvm) &&
1033 1034
	    is_external_interrupt(exit_int_info)) {
		event_injection = true;
R
Rusty Russell 已提交
1035
		push_irq(&svm->vcpu, exit_int_info & SVM_EVTINJ_VEC_MASK);
1036
	}
A
Avi Kivity 已提交
1037

1038 1039
	fault_address  = svm->vmcb->control.exit_info_2;
	error_code = svm->vmcb->control.exit_info_1;
1040 1041 1042 1043 1044

	if (!npt_enabled)
		KVMTRACE_3D(PAGE_FAULT, &svm->vcpu, error_code,
			    (u32)fault_address, (u32)(fault_address >> 32),
			    handler);
1045 1046 1047 1048
	else
		KVMTRACE_3D(TDP_FAULT, &svm->vcpu, error_code,
			    (u32)fault_address, (u32)(fault_address >> 32),
			    handler);
1049 1050 1051 1052 1053 1054 1055
	/*
	 * FIXME: Tis shouldn't be necessary here, but there is a flush
	 * missing in the MMU code. Until we find this bug, flush the
	 * complete TLB here on an NPF
	 */
	if (npt_enabled)
		svm_flush_tlb(&svm->vcpu);
1056

1057
	if (!npt_enabled && event_injection)
1058
		kvm_mmu_unprotect_page_virt(&svm->vcpu, fault_address);
1059
	return kvm_mmu_page_fault(&svm->vcpu, fault_address, error_code);
A
Avi Kivity 已提交
1060 1061
}

1062 1063 1064 1065
static int ud_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	int er;

1066
	er = emulate_instruction(&svm->vcpu, kvm_run, 0, 0, EMULTYPE_TRAP_UD);
1067
	if (er != EMULATE_DONE)
1068
		kvm_queue_exception(&svm->vcpu, UD_VECTOR);
1069 1070 1071
	return 1;
}

R
Rusty Russell 已提交
1072
static int nm_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Anthony Liguori 已提交
1073
{
1074
	svm->vmcb->control.intercept_exceptions &= ~(1 << NM_VECTOR);
1075
	if (!(svm->vcpu.arch.cr0 & X86_CR0_TS))
1076
		svm->vmcb->save.cr0 &= ~X86_CR0_TS;
R
Rusty Russell 已提交
1077
	svm->vcpu.fpu_active = 1;
1078 1079

	return 1;
A
Anthony Liguori 已提交
1080 1081
}

1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094
static int mc_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	/*
	 * On an #MC intercept the MCE handler is not called automatically in
	 * the host. So do it by hand here.
	 */
	asm volatile (
		"int $0x12\n");
	/* not sure if we ever come back to this point */

	return 1;
}

R
Rusty Russell 已提交
1095
static int shutdown_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
1096 1097 1098 1099 1100
{
	/*
	 * VMCB is undefined after a SHUTDOWN intercept
	 * so reinitialize it.
	 */
1101
	clear_page(svm->vmcb);
1102
	init_vmcb(svm);
1103 1104 1105 1106 1107

	kvm_run->exit_reason = KVM_EXIT_SHUTDOWN;
	return 0;
}

R
Rusty Russell 已提交
1108
static int io_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1109
{
M
Mike Day 已提交
1110
	u32 io_info = svm->vmcb->control.exit_info_1; /* address size bug? */
1111 1112
	int size, down, in, string, rep;
	unsigned port;
A
Avi Kivity 已提交
1113

R
Rusty Russell 已提交
1114
	++svm->vcpu.stat.io_exits;
A
Avi Kivity 已提交
1115

1116
	svm->next_rip = svm->vmcb->control.exit_info_2;
A
Avi Kivity 已提交
1117

1118 1119 1120
	string = (io_info & SVM_IOIO_STR_MASK) != 0;

	if (string) {
1121 1122
		if (emulate_instruction(&svm->vcpu,
					kvm_run, 0, 0, 0) == EMULATE_DO_MMIO)
1123 1124 1125 1126
			return 0;
		return 1;
	}

1127 1128 1129 1130
	in = (io_info & SVM_IOIO_TYPE_MASK) != 0;
	port = io_info >> 16;
	size = (io_info & SVM_IOIO_SIZE_MASK) >> SVM_IOIO_SIZE_SHIFT;
	rep = (io_info & SVM_IOIO_REP_MASK) != 0;
1131
	down = (svm->vmcb->save.rflags & X86_EFLAGS_DF) != 0;
A
Avi Kivity 已提交
1132

1133
	skip_emulated_instruction(&svm->vcpu);
L
Laurent Vivier 已提交
1134
	return kvm_emulate_pio(&svm->vcpu, kvm_run, in, size, port);
A
Avi Kivity 已提交
1135 1136
}

1137 1138
static int nmi_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
1139
	KVMTRACE_0D(NMI, &svm->vcpu, handler);
1140 1141 1142
	return 1;
}

1143 1144 1145
static int intr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	++svm->vcpu.stat.irq_exits;
1146
	KVMTRACE_0D(INTR, &svm->vcpu, handler);
1147 1148 1149
	return 1;
}

R
Rusty Russell 已提交
1150
static int nop_on_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1151 1152 1153 1154
{
	return 1;
}

R
Rusty Russell 已提交
1155
static int halt_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1156
{
1157
	svm->next_rip = kvm_rip_read(&svm->vcpu) + 1;
R
Rusty Russell 已提交
1158 1159
	skip_emulated_instruction(&svm->vcpu);
	return kvm_emulate_halt(&svm->vcpu);
A
Avi Kivity 已提交
1160 1161
}

R
Rusty Russell 已提交
1162
static int vmmcall_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
1163
{
1164
	svm->next_rip = kvm_rip_read(&svm->vcpu) + 3;
R
Rusty Russell 已提交
1165
	skip_emulated_instruction(&svm->vcpu);
1166 1167
	kvm_emulate_hypercall(&svm->vcpu);
	return 1;
1168 1169
}

1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245
static int nested_svm_check_permissions(struct vcpu_svm *svm)
{
	if (!(svm->vcpu.arch.shadow_efer & EFER_SVME)
	    || !is_paging(&svm->vcpu)) {
		kvm_queue_exception(&svm->vcpu, UD_VECTOR);
		return 1;
	}

	if (svm->vmcb->save.cpl) {
		kvm_inject_gp(&svm->vcpu, 0);
		return 1;
	}

       return 0;
}

static struct page *nested_svm_get_page(struct vcpu_svm *svm, u64 gpa)
{
	struct page *page;

	down_read(&current->mm->mmap_sem);
	page = gfn_to_page(svm->vcpu.kvm, gpa >> PAGE_SHIFT);
	up_read(&current->mm->mmap_sem);

	if (is_error_page(page)) {
		printk(KERN_INFO "%s: could not find page at 0x%llx\n",
		       __func__, gpa);
		kvm_release_page_clean(page);
		kvm_inject_gp(&svm->vcpu, 0);
		return NULL;
	}
	return page;
}

static int nested_svm_do(struct vcpu_svm *svm,
			 u64 arg1_gpa, u64 arg2_gpa, void *opaque,
			 int (*handler)(struct vcpu_svm *svm,
					void *arg1,
					void *arg2,
					void *opaque))
{
	struct page *arg1_page;
	struct page *arg2_page = NULL;
	void *arg1;
	void *arg2 = NULL;
	int retval;

	arg1_page = nested_svm_get_page(svm, arg1_gpa);
	if(arg1_page == NULL)
		return 1;

	if (arg2_gpa) {
		arg2_page = nested_svm_get_page(svm, arg2_gpa);
		if(arg2_page == NULL) {
			kvm_release_page_clean(arg1_page);
			return 1;
		}
	}

	arg1 = kmap_atomic(arg1_page, KM_USER0);
	if (arg2_gpa)
		arg2 = kmap_atomic(arg2_page, KM_USER1);

	retval = handler(svm, arg1, arg2, opaque);

	kunmap_atomic(arg1, KM_USER0);
	if (arg2_gpa)
		kunmap_atomic(arg2, KM_USER1);

	kvm_release_page_dirty(arg1_page);
	if (arg2_gpa)
		kvm_release_page_dirty(arg2_page);

	return retval;
}

1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275
static int stgi_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	if (nested_svm_check_permissions(svm))
		return 1;

	svm->next_rip = kvm_rip_read(&svm->vcpu) + 3;
	skip_emulated_instruction(&svm->vcpu);

	svm->vcpu.arch.hflags |= HF_GIF_MASK;

	return 1;
}

static int clgi_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	if (nested_svm_check_permissions(svm))
		return 1;

	svm->next_rip = kvm_rip_read(&svm->vcpu) + 3;
	skip_emulated_instruction(&svm->vcpu);

	svm->vcpu.arch.hflags &= ~HF_GIF_MASK;

	/* After a CLGI no interrupts should come */
	svm_clear_vintr(svm);
	svm->vmcb->control.int_ctl &= ~V_IRQ_MASK;

	return 1;
}

R
Rusty Russell 已提交
1276 1277
static int invalid_op_interception(struct vcpu_svm *svm,
				   struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1278
{
1279
	kvm_queue_exception(&svm->vcpu, UD_VECTOR);
A
Avi Kivity 已提交
1280 1281 1282
	return 1;
}

R
Rusty Russell 已提交
1283 1284
static int task_switch_interception(struct vcpu_svm *svm,
				    struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1285
{
1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297
	u16 tss_selector;

	tss_selector = (u16)svm->vmcb->control.exit_info_1;
	if (svm->vmcb->control.exit_info_2 &
	    (1ULL << SVM_EXITINFOSHIFT_TS_REASON_IRET))
		return kvm_task_switch(&svm->vcpu, tss_selector,
				       TASK_SWITCH_IRET);
	if (svm->vmcb->control.exit_info_2 &
	    (1ULL << SVM_EXITINFOSHIFT_TS_REASON_JMP))
		return kvm_task_switch(&svm->vcpu, tss_selector,
				       TASK_SWITCH_JMP);
	return kvm_task_switch(&svm->vcpu, tss_selector, TASK_SWITCH_CALL);
A
Avi Kivity 已提交
1298 1299
}

R
Rusty Russell 已提交
1300
static int cpuid_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1301
{
1302
	svm->next_rip = kvm_rip_read(&svm->vcpu) + 2;
R
Rusty Russell 已提交
1303
	kvm_emulate_cpuid(&svm->vcpu);
1304
	return 1;
A
Avi Kivity 已提交
1305 1306
}

M
Marcelo Tosatti 已提交
1307 1308 1309 1310 1311 1312 1313
static int invlpg_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	if (emulate_instruction(&svm->vcpu, kvm_run, 0, 0, 0) != EMULATE_DONE)
		pr_unimpl(&svm->vcpu, "%s: failed\n", __func__);
	return 1;
}

R
Rusty Russell 已提交
1314 1315
static int emulate_on_interception(struct vcpu_svm *svm,
				   struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1316
{
1317
	if (emulate_instruction(&svm->vcpu, NULL, 0, 0, 0) != EMULATE_DONE)
1318
		pr_unimpl(&svm->vcpu, "%s: failed\n", __func__);
A
Avi Kivity 已提交
1319 1320 1321
	return 1;
}

1322 1323 1324 1325 1326 1327 1328 1329 1330
static int cr8_write_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
{
	emulate_instruction(&svm->vcpu, NULL, 0, 0, 0);
	if (irqchip_in_kernel(svm->vcpu.kvm))
		return 1;
	kvm_run->exit_reason = KVM_EXIT_SET_TPR;
	return 0;
}

A
Avi Kivity 已提交
1331 1332
static int svm_get_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 *data)
{
1333 1334
	struct vcpu_svm *svm = to_svm(vcpu);

A
Avi Kivity 已提交
1335 1336 1337 1338 1339
	switch (ecx) {
	case MSR_IA32_TIME_STAMP_COUNTER: {
		u64 tsc;

		rdtscll(tsc);
1340
		*data = svm->vmcb->control.tsc_offset + tsc;
A
Avi Kivity 已提交
1341 1342
		break;
	}
1343
	case MSR_K6_STAR:
1344
		*data = svm->vmcb->save.star;
A
Avi Kivity 已提交
1345
		break;
1346
#ifdef CONFIG_X86_64
A
Avi Kivity 已提交
1347
	case MSR_LSTAR:
1348
		*data = svm->vmcb->save.lstar;
A
Avi Kivity 已提交
1349 1350
		break;
	case MSR_CSTAR:
1351
		*data = svm->vmcb->save.cstar;
A
Avi Kivity 已提交
1352 1353
		break;
	case MSR_KERNEL_GS_BASE:
1354
		*data = svm->vmcb->save.kernel_gs_base;
A
Avi Kivity 已提交
1355 1356
		break;
	case MSR_SYSCALL_MASK:
1357
		*data = svm->vmcb->save.sfmask;
A
Avi Kivity 已提交
1358 1359 1360
		break;
#endif
	case MSR_IA32_SYSENTER_CS:
1361
		*data = svm->vmcb->save.sysenter_cs;
A
Avi Kivity 已提交
1362 1363
		break;
	case MSR_IA32_SYSENTER_EIP:
1364
		*data = svm->vmcb->save.sysenter_eip;
A
Avi Kivity 已提交
1365 1366
		break;
	case MSR_IA32_SYSENTER_ESP:
1367
		*data = svm->vmcb->save.sysenter_esp;
A
Avi Kivity 已提交
1368
		break;
1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386
	/* Nobody will change the following 5 values in the VMCB so
	   we can safely return them on rdmsr. They will always be 0
	   until LBRV is implemented. */
	case MSR_IA32_DEBUGCTLMSR:
		*data = svm->vmcb->save.dbgctl;
		break;
	case MSR_IA32_LASTBRANCHFROMIP:
		*data = svm->vmcb->save.br_from;
		break;
	case MSR_IA32_LASTBRANCHTOIP:
		*data = svm->vmcb->save.br_to;
		break;
	case MSR_IA32_LASTINTFROMIP:
		*data = svm->vmcb->save.last_excp_from;
		break;
	case MSR_IA32_LASTINTTOIP:
		*data = svm->vmcb->save.last_excp_to;
		break;
A
Alexander Graf 已提交
1387 1388 1389
	case MSR_VM_HSAVE_PA:
		*data = svm->hsave_msr;
		break;
A
Avi Kivity 已提交
1390
	default:
1391
		return kvm_get_msr_common(vcpu, ecx, data);
A
Avi Kivity 已提交
1392 1393 1394 1395
	}
	return 0;
}

R
Rusty Russell 已提交
1396
static int rdmsr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1397
{
1398
	u32 ecx = svm->vcpu.arch.regs[VCPU_REGS_RCX];
A
Avi Kivity 已提交
1399 1400
	u64 data;

R
Rusty Russell 已提交
1401
	if (svm_get_msr(&svm->vcpu, ecx, &data))
1402
		kvm_inject_gp(&svm->vcpu, 0);
A
Avi Kivity 已提交
1403
	else {
1404 1405 1406
		KVMTRACE_3D(MSR_READ, &svm->vcpu, ecx, (u32)data,
			    (u32)(data >> 32), handler);

1407
		svm->vcpu.arch.regs[VCPU_REGS_RAX] = data & 0xffffffff;
1408
		svm->vcpu.arch.regs[VCPU_REGS_RDX] = data >> 32;
1409
		svm->next_rip = kvm_rip_read(&svm->vcpu) + 2;
R
Rusty Russell 已提交
1410
		skip_emulated_instruction(&svm->vcpu);
A
Avi Kivity 已提交
1411 1412 1413 1414 1415 1416
	}
	return 1;
}

static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data)
{
1417 1418
	struct vcpu_svm *svm = to_svm(vcpu);

A
Avi Kivity 已提交
1419 1420 1421 1422 1423
	switch (ecx) {
	case MSR_IA32_TIME_STAMP_COUNTER: {
		u64 tsc;

		rdtscll(tsc);
1424
		svm->vmcb->control.tsc_offset = data - tsc;
A
Avi Kivity 已提交
1425 1426
		break;
	}
1427
	case MSR_K6_STAR:
1428
		svm->vmcb->save.star = data;
A
Avi Kivity 已提交
1429
		break;
1430
#ifdef CONFIG_X86_64
A
Avi Kivity 已提交
1431
	case MSR_LSTAR:
1432
		svm->vmcb->save.lstar = data;
A
Avi Kivity 已提交
1433 1434
		break;
	case MSR_CSTAR:
1435
		svm->vmcb->save.cstar = data;
A
Avi Kivity 已提交
1436 1437
		break;
	case MSR_KERNEL_GS_BASE:
1438
		svm->vmcb->save.kernel_gs_base = data;
A
Avi Kivity 已提交
1439 1440
		break;
	case MSR_SYSCALL_MASK:
1441
		svm->vmcb->save.sfmask = data;
A
Avi Kivity 已提交
1442 1443 1444
		break;
#endif
	case MSR_IA32_SYSENTER_CS:
1445
		svm->vmcb->save.sysenter_cs = data;
A
Avi Kivity 已提交
1446 1447
		break;
	case MSR_IA32_SYSENTER_EIP:
1448
		svm->vmcb->save.sysenter_eip = data;
A
Avi Kivity 已提交
1449 1450
		break;
	case MSR_IA32_SYSENTER_ESP:
1451
		svm->vmcb->save.sysenter_esp = data;
A
Avi Kivity 已提交
1452
		break;
1453
	case MSR_IA32_DEBUGCTLMSR:
1454 1455
		if (!svm_has(SVM_FEATURE_LBRV)) {
			pr_unimpl(vcpu, "%s: MSR_IA32_DEBUGCTL 0x%llx, nop\n",
1456
					__func__, data);
1457 1458 1459 1460 1461 1462 1463 1464 1465 1466
			break;
		}
		if (data & DEBUGCTL_RESERVED_BITS)
			return 1;

		svm->vmcb->save.dbgctl = data;
		if (data & (1ULL<<0))
			svm_enable_lbrv(svm);
		else
			svm_disable_lbrv(svm);
1467
		break;
1468 1469 1470 1471
	case MSR_K7_EVNTSEL0:
	case MSR_K7_EVNTSEL1:
	case MSR_K7_EVNTSEL2:
	case MSR_K7_EVNTSEL3:
1472 1473 1474 1475
	case MSR_K7_PERFCTR0:
	case MSR_K7_PERFCTR1:
	case MSR_K7_PERFCTR2:
	case MSR_K7_PERFCTR3:
1476
		/*
1477 1478 1479
		 * Just discard all writes to the performance counters; this
		 * should keep both older linux and windows 64-bit guests
		 * happy
1480
		 */
1481 1482
		pr_unimpl(vcpu, "unimplemented perfctr wrmsr: 0x%x data 0x%llx\n", ecx, data);

A
Alexander Graf 已提交
1483 1484 1485
		break;
	case MSR_VM_HSAVE_PA:
		svm->hsave_msr = data;
1486
		break;
A
Avi Kivity 已提交
1487
	default:
1488
		return kvm_set_msr_common(vcpu, ecx, data);
A
Avi Kivity 已提交
1489 1490 1491 1492
	}
	return 0;
}

R
Rusty Russell 已提交
1493
static int wrmsr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1494
{
1495
	u32 ecx = svm->vcpu.arch.regs[VCPU_REGS_RCX];
1496
	u64 data = (svm->vcpu.arch.regs[VCPU_REGS_RAX] & -1u)
1497
		| ((u64)(svm->vcpu.arch.regs[VCPU_REGS_RDX] & -1u) << 32);
1498 1499 1500 1501

	KVMTRACE_3D(MSR_WRITE, &svm->vcpu, ecx, (u32)data, (u32)(data >> 32),
		    handler);

1502
	svm->next_rip = kvm_rip_read(&svm->vcpu) + 2;
R
Rusty Russell 已提交
1503
	if (svm_set_msr(&svm->vcpu, ecx, data))
1504
		kvm_inject_gp(&svm->vcpu, 0);
A
Avi Kivity 已提交
1505
	else
R
Rusty Russell 已提交
1506
		skip_emulated_instruction(&svm->vcpu);
A
Avi Kivity 已提交
1507 1508 1509
	return 1;
}

R
Rusty Russell 已提交
1510
static int msr_interception(struct vcpu_svm *svm, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1511
{
R
Rusty Russell 已提交
1512 1513
	if (svm->vmcb->control.exit_info_1)
		return wrmsr_interception(svm, kvm_run);
A
Avi Kivity 已提交
1514
	else
R
Rusty Russell 已提交
1515
		return rdmsr_interception(svm, kvm_run);
A
Avi Kivity 已提交
1516 1517
}

R
Rusty Russell 已提交
1518
static int interrupt_window_interception(struct vcpu_svm *svm,
1519 1520
				   struct kvm_run *kvm_run)
{
1521 1522
	KVMTRACE_0D(PEND_INTR, &svm->vcpu, handler);

1523
	svm_clear_vintr(svm);
1524
	svm->vmcb->control.int_ctl &= ~V_IRQ_MASK;
1525 1526 1527 1528 1529
	/*
	 * If the user space waits to inject interrupts, exit as soon as
	 * possible
	 */
	if (kvm_run->request_interrupt_window &&
1530
	    !svm->vcpu.arch.irq_summary) {
R
Rusty Russell 已提交
1531
		++svm->vcpu.stat.irq_window_exits;
1532 1533 1534 1535 1536 1537 1538
		kvm_run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
		return 0;
	}

	return 1;
}

R
Rusty Russell 已提交
1539
static int (*svm_exit_handlers[])(struct vcpu_svm *svm,
A
Avi Kivity 已提交
1540 1541 1542 1543
				      struct kvm_run *kvm_run) = {
	[SVM_EXIT_READ_CR0]           		= emulate_on_interception,
	[SVM_EXIT_READ_CR3]           		= emulate_on_interception,
	[SVM_EXIT_READ_CR4]           		= emulate_on_interception,
1544
	[SVM_EXIT_READ_CR8]           		= emulate_on_interception,
A
Avi Kivity 已提交
1545 1546 1547 1548
	/* for now: */
	[SVM_EXIT_WRITE_CR0]          		= emulate_on_interception,
	[SVM_EXIT_WRITE_CR3]          		= emulate_on_interception,
	[SVM_EXIT_WRITE_CR4]          		= emulate_on_interception,
1549
	[SVM_EXIT_WRITE_CR8]          		= cr8_write_interception,
A
Avi Kivity 已提交
1550 1551 1552 1553 1554 1555 1556 1557 1558 1559
	[SVM_EXIT_READ_DR0] 			= emulate_on_interception,
	[SVM_EXIT_READ_DR1]			= emulate_on_interception,
	[SVM_EXIT_READ_DR2]			= emulate_on_interception,
	[SVM_EXIT_READ_DR3]			= emulate_on_interception,
	[SVM_EXIT_WRITE_DR0]			= emulate_on_interception,
	[SVM_EXIT_WRITE_DR1]			= emulate_on_interception,
	[SVM_EXIT_WRITE_DR2]			= emulate_on_interception,
	[SVM_EXIT_WRITE_DR3]			= emulate_on_interception,
	[SVM_EXIT_WRITE_DR5]			= emulate_on_interception,
	[SVM_EXIT_WRITE_DR7]			= emulate_on_interception,
1560
	[SVM_EXIT_EXCP_BASE + UD_VECTOR]	= ud_interception,
A
Avi Kivity 已提交
1561
	[SVM_EXIT_EXCP_BASE + PF_VECTOR] 	= pf_interception,
A
Anthony Liguori 已提交
1562
	[SVM_EXIT_EXCP_BASE + NM_VECTOR] 	= nm_interception,
1563
	[SVM_EXIT_EXCP_BASE + MC_VECTOR] 	= mc_interception,
1564
	[SVM_EXIT_INTR] 			= intr_interception,
1565
	[SVM_EXIT_NMI]				= nmi_interception,
A
Avi Kivity 已提交
1566 1567
	[SVM_EXIT_SMI]				= nop_on_interception,
	[SVM_EXIT_INIT]				= nop_on_interception,
1568
	[SVM_EXIT_VINTR]			= interrupt_window_interception,
A
Avi Kivity 已提交
1569 1570
	/* [SVM_EXIT_CR0_SEL_WRITE]		= emulate_on_interception, */
	[SVM_EXIT_CPUID]			= cpuid_interception,
1571
	[SVM_EXIT_INVD]                         = emulate_on_interception,
A
Avi Kivity 已提交
1572
	[SVM_EXIT_HLT]				= halt_interception,
M
Marcelo Tosatti 已提交
1573
	[SVM_EXIT_INVLPG]			= invlpg_interception,
A
Avi Kivity 已提交
1574 1575 1576 1577
	[SVM_EXIT_INVLPGA]			= invalid_op_interception,
	[SVM_EXIT_IOIO] 		  	= io_interception,
	[SVM_EXIT_MSR]				= msr_interception,
	[SVM_EXIT_TASK_SWITCH]			= task_switch_interception,
1578
	[SVM_EXIT_SHUTDOWN]			= shutdown_interception,
A
Avi Kivity 已提交
1579
	[SVM_EXIT_VMRUN]			= invalid_op_interception,
1580
	[SVM_EXIT_VMMCALL]			= vmmcall_interception,
A
Avi Kivity 已提交
1581 1582
	[SVM_EXIT_VMLOAD]			= invalid_op_interception,
	[SVM_EXIT_VMSAVE]			= invalid_op_interception,
1583 1584
	[SVM_EXIT_STGI]				= stgi_interception,
	[SVM_EXIT_CLGI]				= clgi_interception,
A
Avi Kivity 已提交
1585
	[SVM_EXIT_SKINIT]			= invalid_op_interception,
1586
	[SVM_EXIT_WBINVD]                       = emulate_on_interception,
1587 1588
	[SVM_EXIT_MONITOR]			= invalid_op_interception,
	[SVM_EXIT_MWAIT]			= invalid_op_interception,
1589
	[SVM_EXIT_NPF]				= pf_interception,
A
Avi Kivity 已提交
1590 1591
};

1592
static int handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
A
Avi Kivity 已提交
1593
{
1594
	struct vcpu_svm *svm = to_svm(vcpu);
1595
	u32 exit_code = svm->vmcb->control.exit_code;
A
Avi Kivity 已提交
1596

1597 1598 1599
	KVMTRACE_3D(VMEXIT, vcpu, exit_code, (u32)svm->vmcb->save.rip,
		    (u32)((u64)svm->vmcb->save.rip >> 32), entryexit);

1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619
	if (npt_enabled) {
		int mmu_reload = 0;
		if ((vcpu->arch.cr0 ^ svm->vmcb->save.cr0) & X86_CR0_PG) {
			svm_set_cr0(vcpu, svm->vmcb->save.cr0);
			mmu_reload = 1;
		}
		vcpu->arch.cr0 = svm->vmcb->save.cr0;
		vcpu->arch.cr3 = svm->vmcb->save.cr3;
		if (is_paging(vcpu) && is_pae(vcpu) && !is_long_mode(vcpu)) {
			if (!load_pdptrs(vcpu, vcpu->arch.cr3)) {
				kvm_inject_gp(vcpu, 0);
				return 1;
			}
		}
		if (mmu_reload) {
			kvm_mmu_reset_context(vcpu);
			kvm_mmu_load(vcpu);
		}
	}

1620 1621 1622 1623 1624 1625 1626 1627 1628
	kvm_reput_irq(svm);

	if (svm->vmcb->control.exit_code == SVM_EXIT_ERR) {
		kvm_run->exit_reason = KVM_EXIT_FAIL_ENTRY;
		kvm_run->fail_entry.hardware_entry_failure_reason
			= svm->vmcb->control.exit_code;
		return 0;
	}

1629
	if (is_external_interrupt(svm->vmcb->control.exit_int_info) &&
1630 1631
	    exit_code != SVM_EXIT_EXCP_BASE + PF_VECTOR &&
	    exit_code != SVM_EXIT_NPF)
A
Avi Kivity 已提交
1632 1633
		printk(KERN_ERR "%s: unexpected exit_ini_info 0x%x "
		       "exit_code 0x%x\n",
1634
		       __func__, svm->vmcb->control.exit_int_info,
A
Avi Kivity 已提交
1635 1636
		       exit_code);

1637
	if (exit_code >= ARRAY_SIZE(svm_exit_handlers)
J
Joe Perches 已提交
1638
	    || !svm_exit_handlers[exit_code]) {
A
Avi Kivity 已提交
1639
		kvm_run->exit_reason = KVM_EXIT_UNKNOWN;
1640
		kvm_run->hw.hardware_exit_reason = exit_code;
A
Avi Kivity 已提交
1641 1642 1643
		return 0;
	}

R
Rusty Russell 已提交
1644
	return svm_exit_handlers[exit_code](svm, kvm_run);
A
Avi Kivity 已提交
1645 1646 1647 1648 1649 1650 1651
}

static void reload_tss(struct kvm_vcpu *vcpu)
{
	int cpu = raw_smp_processor_id();

	struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu);
M
Mike Day 已提交
1652
	svm_data->tss_desc->type = 9; /* available 32/64-bit TSS */
A
Avi Kivity 已提交
1653 1654 1655
	load_TR_desc();
}

R
Rusty Russell 已提交
1656
static void pre_svm_run(struct vcpu_svm *svm)
A
Avi Kivity 已提交
1657 1658 1659 1660 1661
{
	int cpu = raw_smp_processor_id();

	struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu);

1662
	svm->vmcb->control.tlb_ctl = TLB_CONTROL_DO_NOTHING;
R
Rusty Russell 已提交
1663
	if (svm->vcpu.cpu != cpu ||
1664
	    svm->asid_generation != svm_data->asid_generation)
R
Rusty Russell 已提交
1665
		new_asid(svm, svm_data);
A
Avi Kivity 已提交
1666 1667 1668
}


1669
static inline void svm_inject_irq(struct vcpu_svm *svm, int irq)
A
Avi Kivity 已提交
1670 1671 1672
{
	struct vmcb_control_area *control;

1673 1674
	KVMTRACE_1D(INJ_VIRQ, &svm->vcpu, (u32)irq, handler);

1675
	++svm->vcpu.stat.irq_injections;
R
Rusty Russell 已提交
1676
	control = &svm->vmcb->control;
1677
	control->int_vector = irq;
A
Avi Kivity 已提交
1678 1679 1680 1681 1682
	control->int_ctl &= ~V_INTR_PRIO_MASK;
	control->int_ctl |= V_IRQ_MASK |
		((/*control->int_vector >> 4*/ 0xf) << V_INTR_PRIO_SHIFT);
}

E
Eddie Dong 已提交
1683 1684 1685 1686 1687 1688 1689
static void svm_set_irq(struct kvm_vcpu *vcpu, int irq)
{
	struct vcpu_svm *svm = to_svm(vcpu);

	svm_inject_irq(svm, irq);
}

1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710
static void update_cr8_intercept(struct kvm_vcpu *vcpu)
{
	struct vcpu_svm *svm = to_svm(vcpu);
	struct vmcb *vmcb = svm->vmcb;
	int max_irr, tpr;

	if (!irqchip_in_kernel(vcpu->kvm) || vcpu->arch.apic->vapic_addr)
		return;

	vmcb->control.intercept_cr_write &= ~INTERCEPT_CR8_MASK;

	max_irr = kvm_lapic_find_highest_irr(vcpu);
	if (max_irr == -1)
		return;

	tpr = kvm_lapic_get_cr8(vcpu) << 4;

	if (tpr >= (max_irr & 0xf0))
		vmcb->control.intercept_cr_write |= INTERCEPT_CR8_MASK;
}

1711
static void svm_intr_assist(struct kvm_vcpu *vcpu)
A
Avi Kivity 已提交
1712
{
1713
	struct vcpu_svm *svm = to_svm(vcpu);
1714 1715 1716 1717 1718 1719 1720 1721 1722
	struct vmcb *vmcb = svm->vmcb;
	int intr_vector = -1;

	if ((vmcb->control.exit_int_info & SVM_EVTINJ_VALID) &&
	    ((vmcb->control.exit_int_info & SVM_EVTINJ_TYPE_MASK) == 0)) {
		intr_vector = vmcb->control.exit_int_info &
			      SVM_EVTINJ_VEC_MASK;
		vmcb->control.exit_int_info = 0;
		svm_inject_irq(svm, intr_vector);
1723
		goto out;
1724 1725 1726
	}

	if (vmcb->control.int_ctl & V_IRQ_MASK)
1727
		goto out;
1728

1729
	if (!kvm_cpu_has_interrupt(vcpu))
1730
		goto out;
1731

1732 1733 1734
	if (!(svm->vcpu.arch.hflags & HF_GIF_MASK))
		goto out;

1735 1736 1737 1738
	if (!(vmcb->save.rflags & X86_EFLAGS_IF) ||
	    (vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK) ||
	    (vmcb->control.event_inj & SVM_EVTINJ_VALID)) {
		/* unable to deliver irq, set pending irq */
1739
		svm_set_vintr(svm);
1740
		svm_inject_irq(svm, 0x0);
1741
		goto out;
1742 1743
	}
	/* Okay, we can deliver the interrupt: grab it and update PIC state. */
1744
	intr_vector = kvm_cpu_get_interrupt(vcpu);
1745
	svm_inject_irq(svm, intr_vector);
1746 1747
out:
	update_cr8_intercept(vcpu);
1748 1749 1750 1751
}

static void kvm_reput_irq(struct vcpu_svm *svm)
{
R
Rusty Russell 已提交
1752
	struct vmcb_control_area *control = &svm->vmcb->control;
A
Avi Kivity 已提交
1753

1754 1755
	if ((control->int_ctl & V_IRQ_MASK)
	    && !irqchip_in_kernel(svm->vcpu.kvm)) {
A
Avi Kivity 已提交
1756
		control->int_ctl &= ~V_IRQ_MASK;
R
Rusty Russell 已提交
1757
		push_irq(&svm->vcpu, control->int_vector);
A
Avi Kivity 已提交
1758
	}
1759

1760
	svm->vcpu.arch.interrupt_window_open =
1761 1762
		!(control->int_state & SVM_INTERRUPT_SHADOW_MASK) &&
		 (svm->vcpu.arch.hflags & HF_GIF_MASK);
1763 1764
}

1765 1766 1767
static void svm_do_inject_vector(struct vcpu_svm *svm)
{
	struct kvm_vcpu *vcpu = &svm->vcpu;
1768 1769
	int word_index = __ffs(vcpu->arch.irq_summary);
	int bit_index = __ffs(vcpu->arch.irq_pending[word_index]);
1770 1771
	int irq = word_index * BITS_PER_LONG + bit_index;

1772 1773 1774
	clear_bit(bit_index, &vcpu->arch.irq_pending[word_index]);
	if (!vcpu->arch.irq_pending[word_index])
		clear_bit(word_index, &vcpu->arch.irq_summary);
1775 1776 1777
	svm_inject_irq(svm, irq);
}

1778
static void do_interrupt_requests(struct kvm_vcpu *vcpu,
1779 1780
				       struct kvm_run *kvm_run)
{
1781
	struct vcpu_svm *svm = to_svm(vcpu);
1782
	struct vmcb_control_area *control = &svm->vmcb->control;
1783

1784
	svm->vcpu.arch.interrupt_window_open =
1785
		(!(control->int_state & SVM_INTERRUPT_SHADOW_MASK) &&
1786 1787
		 (svm->vmcb->save.rflags & X86_EFLAGS_IF) &&
		 (svm->vcpu.arch.hflags & HF_GIF_MASK));
1788

1789
	if (svm->vcpu.arch.interrupt_window_open && svm->vcpu.arch.irq_summary)
1790 1791 1792
		/*
		 * If interrupts enabled, and not blocked by sti or mov ss. Good.
		 */
1793
		svm_do_inject_vector(svm);
1794 1795 1796 1797

	/*
	 * Interrupts blocked.  Wait for unblock.
	 */
1798 1799
	if (!svm->vcpu.arch.interrupt_window_open &&
	    (svm->vcpu.arch.irq_summary || kvm_run->request_interrupt_window))
1800 1801 1802
		svm_set_vintr(svm);
	else
		svm_clear_vintr(svm);
1803 1804
}

1805 1806 1807 1808 1809
static int svm_set_tss_addr(struct kvm *kvm, unsigned int addr)
{
	return 0;
}

A
Avi Kivity 已提交
1810 1811
static void save_db_regs(unsigned long *db_regs)
{
1812 1813 1814 1815
	asm volatile ("mov %%dr0, %0" : "=r"(db_regs[0]));
	asm volatile ("mov %%dr1, %0" : "=r"(db_regs[1]));
	asm volatile ("mov %%dr2, %0" : "=r"(db_regs[2]));
	asm volatile ("mov %%dr3, %0" : "=r"(db_regs[3]));
A
Avi Kivity 已提交
1816 1817 1818 1819
}

static void load_db_regs(unsigned long *db_regs)
{
1820 1821 1822 1823
	asm volatile ("mov %0, %%dr0" : : "r"(db_regs[0]));
	asm volatile ("mov %0, %%dr1" : : "r"(db_regs[1]));
	asm volatile ("mov %0, %%dr2" : : "r"(db_regs[2]));
	asm volatile ("mov %0, %%dr3" : : "r"(db_regs[3]));
A
Avi Kivity 已提交
1824 1825
}

1826 1827 1828 1829 1830
static void svm_flush_tlb(struct kvm_vcpu *vcpu)
{
	force_new_asid(vcpu);
}

1831 1832 1833 1834
static void svm_prepare_guest_switch(struct kvm_vcpu *vcpu)
{
}

1835 1836 1837 1838 1839 1840 1841 1842 1843 1844
static inline void sync_cr8_to_lapic(struct kvm_vcpu *vcpu)
{
	struct vcpu_svm *svm = to_svm(vcpu);

	if (!(svm->vmcb->control.intercept_cr_write & INTERCEPT_CR8_MASK)) {
		int cr8 = svm->vmcb->control.int_ctl & V_TPR_MASK;
		kvm_lapic_set_tpr(vcpu, cr8);
	}
}

1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857
static inline void sync_lapic_to_cr8(struct kvm_vcpu *vcpu)
{
	struct vcpu_svm *svm = to_svm(vcpu);
	u64 cr8;

	if (!irqchip_in_kernel(vcpu->kvm))
		return;

	cr8 = kvm_get_cr8(vcpu);
	svm->vmcb->control.int_ctl &= ~V_TPR_MASK;
	svm->vmcb->control.int_ctl |= cr8 & V_TPR_MASK;
}

1858 1859 1860 1861 1862 1863
#ifdef CONFIG_X86_64
#define R "r"
#else
#define R "e"
#endif

1864
static void svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
A
Avi Kivity 已提交
1865
{
1866
	struct vcpu_svm *svm = to_svm(vcpu);
A
Avi Kivity 已提交
1867 1868 1869
	u16 fs_selector;
	u16 gs_selector;
	u16 ldt_selector;
1870

1871 1872 1873 1874
	svm->vmcb->save.rax = vcpu->arch.regs[VCPU_REGS_RAX];
	svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
	svm->vmcb->save.rip = vcpu->arch.regs[VCPU_REGS_RIP];

R
Rusty Russell 已提交
1875
	pre_svm_run(svm);
A
Avi Kivity 已提交
1876

1877 1878
	sync_lapic_to_cr8(vcpu);

A
Avi Kivity 已提交
1879
	save_host_msrs(vcpu);
1880 1881 1882
	fs_selector = kvm_read_fs();
	gs_selector = kvm_read_gs();
	ldt_selector = kvm_read_ldt();
1883 1884 1885
	svm->host_cr2 = kvm_read_cr2();
	svm->host_dr6 = read_dr6();
	svm->host_dr7 = read_dr7();
1886
	svm->vmcb->save.cr2 = vcpu->arch.cr2;
1887 1888 1889
	/* required for live migration with NPT */
	if (npt_enabled)
		svm->vmcb->save.cr3 = vcpu->arch.cr3;
A
Avi Kivity 已提交
1890

1891
	if (svm->vmcb->save.dr7 & 0xff) {
A
Avi Kivity 已提交
1892
		write_dr7(0);
1893 1894
		save_db_regs(svm->host_db_regs);
		load_db_regs(svm->db_regs);
A
Avi Kivity 已提交
1895
	}
1896

1897 1898 1899
	clgi();

	local_irq_enable();
1900

A
Avi Kivity 已提交
1901
	asm volatile (
1902 1903 1904 1905 1906 1907 1908
		"push %%"R"bp; \n\t"
		"mov %c[rbx](%[svm]), %%"R"bx \n\t"
		"mov %c[rcx](%[svm]), %%"R"cx \n\t"
		"mov %c[rdx](%[svm]), %%"R"dx \n\t"
		"mov %c[rsi](%[svm]), %%"R"si \n\t"
		"mov %c[rdi](%[svm]), %%"R"di \n\t"
		"mov %c[rbp](%[svm]), %%"R"bp \n\t"
1909
#ifdef CONFIG_X86_64
R
Rusty Russell 已提交
1910 1911 1912 1913 1914 1915 1916 1917
		"mov %c[r8](%[svm]),  %%r8  \n\t"
		"mov %c[r9](%[svm]),  %%r9  \n\t"
		"mov %c[r10](%[svm]), %%r10 \n\t"
		"mov %c[r11](%[svm]), %%r11 \n\t"
		"mov %c[r12](%[svm]), %%r12 \n\t"
		"mov %c[r13](%[svm]), %%r13 \n\t"
		"mov %c[r14](%[svm]), %%r14 \n\t"
		"mov %c[r15](%[svm]), %%r15 \n\t"
A
Avi Kivity 已提交
1918 1919 1920
#endif

		/* Enter guest mode */
1921 1922
		"push %%"R"ax \n\t"
		"mov %c[vmcb](%[svm]), %%"R"ax \n\t"
1923 1924 1925
		__ex(SVM_VMLOAD) "\n\t"
		__ex(SVM_VMRUN) "\n\t"
		__ex(SVM_VMSAVE) "\n\t"
1926
		"pop %%"R"ax \n\t"
A
Avi Kivity 已提交
1927 1928

		/* Save guest registers, load host registers */
1929 1930 1931 1932 1933 1934
		"mov %%"R"bx, %c[rbx](%[svm]) \n\t"
		"mov %%"R"cx, %c[rcx](%[svm]) \n\t"
		"mov %%"R"dx, %c[rdx](%[svm]) \n\t"
		"mov %%"R"si, %c[rsi](%[svm]) \n\t"
		"mov %%"R"di, %c[rdi](%[svm]) \n\t"
		"mov %%"R"bp, %c[rbp](%[svm]) \n\t"
1935
#ifdef CONFIG_X86_64
R
Rusty Russell 已提交
1936 1937 1938 1939 1940 1941 1942 1943
		"mov %%r8,  %c[r8](%[svm]) \n\t"
		"mov %%r9,  %c[r9](%[svm]) \n\t"
		"mov %%r10, %c[r10](%[svm]) \n\t"
		"mov %%r11, %c[r11](%[svm]) \n\t"
		"mov %%r12, %c[r12](%[svm]) \n\t"
		"mov %%r13, %c[r13](%[svm]) \n\t"
		"mov %%r14, %c[r14](%[svm]) \n\t"
		"mov %%r15, %c[r15](%[svm]) \n\t"
A
Avi Kivity 已提交
1944
#endif
1945
		"pop %%"R"bp"
A
Avi Kivity 已提交
1946
		:
R
Rusty Russell 已提交
1947
		: [svm]"a"(svm),
A
Avi Kivity 已提交
1948
		  [vmcb]"i"(offsetof(struct vcpu_svm, vmcb_pa)),
1949 1950 1951 1952 1953 1954
		  [rbx]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RBX])),
		  [rcx]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RCX])),
		  [rdx]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RDX])),
		  [rsi]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RSI])),
		  [rdi]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RDI])),
		  [rbp]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RBP]))
1955
#ifdef CONFIG_X86_64
1956 1957 1958 1959 1960 1961 1962 1963
		  , [r8]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R8])),
		  [r9]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R9])),
		  [r10]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R10])),
		  [r11]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R11])),
		  [r12]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R12])),
		  [r13]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R13])),
		  [r14]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R14])),
		  [r15]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_R15]))
A
Avi Kivity 已提交
1964
#endif
1965
		: "cc", "memory"
1966
		, R"bx", R"cx", R"dx", R"si", R"di"
1967 1968 1969 1970
#ifdef CONFIG_X86_64
		, "r8", "r9", "r10", "r11" , "r12", "r13", "r14", "r15"
#endif
		);
A
Avi Kivity 已提交
1971

1972 1973
	if ((svm->vmcb->save.dr7 & 0xff))
		load_db_regs(svm->host_db_regs);
A
Avi Kivity 已提交
1974

1975
	vcpu->arch.cr2 = svm->vmcb->save.cr2;
1976 1977 1978
	vcpu->arch.regs[VCPU_REGS_RAX] = svm->vmcb->save.rax;
	vcpu->arch.regs[VCPU_REGS_RSP] = svm->vmcb->save.rsp;
	vcpu->arch.regs[VCPU_REGS_RIP] = svm->vmcb->save.rip;
A
Avi Kivity 已提交
1979

1980 1981 1982
	write_dr6(svm->host_dr6);
	write_dr7(svm->host_dr7);
	kvm_write_cr2(svm->host_cr2);
A
Avi Kivity 已提交
1983

1984 1985 1986
	kvm_load_fs(fs_selector);
	kvm_load_gs(gs_selector);
	kvm_load_ldt(ldt_selector);
A
Avi Kivity 已提交
1987 1988 1989 1990
	load_host_msrs(vcpu);

	reload_tss(vcpu);

1991 1992 1993 1994
	local_irq_disable();

	stgi();

1995 1996
	sync_cr8_to_lapic(vcpu);

1997
	svm->next_rip = 0;
A
Avi Kivity 已提交
1998 1999
}

2000 2001
#undef R

A
Avi Kivity 已提交
2002 2003
static void svm_set_cr3(struct kvm_vcpu *vcpu, unsigned long root)
{
2004 2005
	struct vcpu_svm *svm = to_svm(vcpu);

2006 2007 2008 2009 2010 2011
	if (npt_enabled) {
		svm->vmcb->control.nested_cr3 = root;
		force_new_asid(vcpu);
		return;
	}

2012
	svm->vmcb->save.cr3 = root;
A
Avi Kivity 已提交
2013
	force_new_asid(vcpu);
A
Anthony Liguori 已提交
2014 2015

	if (vcpu->fpu_active) {
2016 2017
		svm->vmcb->control.intercept_exceptions |= (1 << NM_VECTOR);
		svm->vmcb->save.cr0 |= X86_CR0_TS;
A
Anthony Liguori 已提交
2018 2019
		vcpu->fpu_active = 0;
	}
A
Avi Kivity 已提交
2020 2021 2022 2023
}

static int is_disabled(void)
{
2024 2025 2026 2027 2028 2029
	u64 vm_cr;

	rdmsrl(MSR_VM_CR, vm_cr);
	if (vm_cr & (1 << SVM_VM_CR_SVM_DISABLE))
		return 1;

A
Avi Kivity 已提交
2030 2031 2032
	return 0;
}

I
Ingo Molnar 已提交
2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043
static void
svm_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall)
{
	/*
	 * Patch in the VMMCALL instruction:
	 */
	hypercall[0] = 0x0f;
	hypercall[1] = 0x01;
	hypercall[2] = 0xd9;
}

Y
Yang, Sheng 已提交
2044 2045 2046 2047 2048
static void svm_check_processor_compat(void *rtn)
{
	*(int *)rtn = 0;
}

2049 2050 2051 2052 2053
static bool svm_cpu_has_accelerated_tpr(void)
{
	return false;
}

2054 2055 2056 2057 2058 2059 2060 2061 2062
static int get_npt_level(void)
{
#ifdef CONFIG_X86_64
	return PT64_ROOT_LEVEL;
#else
	return PT32E_ROOT_LEVEL;
#endif
}

S
Sheng Yang 已提交
2063 2064 2065 2066 2067
static int svm_get_mt_mask_shift(void)
{
	return 0;
}

2068
static struct kvm_x86_ops svm_x86_ops = {
A
Avi Kivity 已提交
2069 2070 2071 2072
	.cpu_has_kvm_support = has_svm,
	.disabled_by_bios = is_disabled,
	.hardware_setup = svm_hardware_setup,
	.hardware_unsetup = svm_hardware_unsetup,
Y
Yang, Sheng 已提交
2073
	.check_processor_compatibility = svm_check_processor_compat,
A
Avi Kivity 已提交
2074 2075
	.hardware_enable = svm_hardware_enable,
	.hardware_disable = svm_hardware_disable,
2076
	.cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr,
A
Avi Kivity 已提交
2077 2078 2079

	.vcpu_create = svm_create_vcpu,
	.vcpu_free = svm_free_vcpu,
2080
	.vcpu_reset = svm_vcpu_reset,
A
Avi Kivity 已提交
2081

2082
	.prepare_guest_switch = svm_prepare_guest_switch,
A
Avi Kivity 已提交
2083 2084 2085 2086 2087 2088 2089 2090 2091
	.vcpu_load = svm_vcpu_load,
	.vcpu_put = svm_vcpu_put,

	.set_guest_debug = svm_guest_debug,
	.get_msr = svm_get_msr,
	.set_msr = svm_set_msr,
	.get_segment_base = svm_get_segment_base,
	.get_segment = svm_get_segment,
	.set_segment = svm_set_segment,
2092
	.get_cpl = svm_get_cpl,
2093
	.get_cs_db_l_bits = kvm_get_cs_db_l_bits,
2094
	.decache_cr4_guest_bits = svm_decache_cr4_guest_bits,
A
Avi Kivity 已提交
2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110
	.set_cr0 = svm_set_cr0,
	.set_cr3 = svm_set_cr3,
	.set_cr4 = svm_set_cr4,
	.set_efer = svm_set_efer,
	.get_idt = svm_get_idt,
	.set_idt = svm_set_idt,
	.get_gdt = svm_get_gdt,
	.set_gdt = svm_set_gdt,
	.get_dr = svm_get_dr,
	.set_dr = svm_set_dr,
	.get_rflags = svm_get_rflags,
	.set_rflags = svm_set_rflags,

	.tlb_flush = svm_flush_tlb,

	.run = svm_vcpu_run,
2111
	.handle_exit = handle_exit,
A
Avi Kivity 已提交
2112
	.skip_emulated_instruction = skip_emulated_instruction,
I
Ingo Molnar 已提交
2113
	.patch_hypercall = svm_patch_hypercall,
E
Eddie Dong 已提交
2114 2115
	.get_irq = svm_get_irq,
	.set_irq = svm_set_irq,
2116 2117
	.queue_exception = svm_queue_exception,
	.exception_injected = svm_exception_injected,
2118 2119
	.inject_pending_irq = svm_intr_assist,
	.inject_pending_vectors = do_interrupt_requests,
2120 2121

	.set_tss_addr = svm_set_tss_addr,
2122
	.get_tdp_level = get_npt_level,
S
Sheng Yang 已提交
2123
	.get_mt_mask_shift = svm_get_mt_mask_shift,
A
Avi Kivity 已提交
2124 2125 2126 2127
};

static int __init svm_init(void)
{
2128
	return kvm_init(&svm_x86_ops, sizeof(struct vcpu_svm),
2129
			      THIS_MODULE);
A
Avi Kivity 已提交
2130 2131 2132 2133
}

static void __exit svm_exit(void)
{
2134
	kvm_exit();
A
Avi Kivity 已提交
2135 2136 2137 2138
}

module_init(svm_init)
module_exit(svm_exit)