kvm.c 20.2 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
/*
 * KVM paravirt_ops implementation
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation; either version 2 of the License, or
 * (at your option) any later version.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
 *
 * Copyright (C) 2007, Red Hat, Inc., Ingo Molnar <mingo@redhat.com>
 * Copyright IBM Corporation, 2007
 *   Authors: Anthony Liguori <aliguori@us.ibm.com>
 */

23
#include <linux/context_tracking.h>
24
#include <linux/init.h>
25 26 27 28
#include <linux/kernel.h>
#include <linux/kvm_para.h>
#include <linux/cpu.h>
#include <linux/mm.h>
29
#include <linux/highmem.h>
30
#include <linux/hardirq.h>
31 32
#include <linux/notifier.h>
#include <linux/reboot.h>
G
Gleb Natapov 已提交
33 34 35 36
#include <linux/hash.h>
#include <linux/sched.h>
#include <linux/slab.h>
#include <linux/kprobes.h>
37
#include <linux/debugfs.h>
38
#include <linux/nmi.h>
39
#include <linux/swait.h>
40
#include <asm/timer.h>
41
#include <asm/cpu.h>
G
Gleb Natapov 已提交
42 43
#include <asm/traps.h>
#include <asm/desc.h>
44
#include <asm/tlbflush.h>
45 46
#include <asm/apic.h>
#include <asm/apicdef.h>
47
#include <asm/hypervisor.h>
48
#include <asm/tlb.h>
49

50 51
static int kvmapf = 1;

52
static int __init parse_no_kvmapf(char *arg)
53 54 55 56 57 58 59
{
        kvmapf = 0;
        return 0;
}

early_param("no-kvmapf", parse_no_kvmapf);

60
static int steal_acc = 1;
61
static int __init parse_no_stealacc(char *arg)
62 63 64 65 66 67 68
{
        steal_acc = 0;
        return 0;
}

early_param("no-steal-acc", parse_no_stealacc);

69 70
static DEFINE_PER_CPU_DECRYPTED(struct kvm_vcpu_pv_apf_data, apf_reason) __aligned(64);
static DEFINE_PER_CPU_DECRYPTED(struct kvm_steal_time, steal_time) __aligned(64);
71
static int has_steal_clock = 0;
72

73 74 75 76 77 78 79
/*
 * No need for any "IO delay" on KVM
 */
static void kvm_io_delay(void)
{
}

G
Gleb Natapov 已提交
80 81 82 83 84
#define KVM_TASK_SLEEP_HASHBITS 8
#define KVM_TASK_SLEEP_HASHSIZE (1<<KVM_TASK_SLEEP_HASHBITS)

struct kvm_task_sleep_node {
	struct hlist_node link;
85
	struct swait_queue_head wq;
G
Gleb Natapov 已提交
86 87
	u32 token;
	int cpu;
88
	bool halted;
G
Gleb Natapov 已提交
89 90 91
};

static struct kvm_task_sleep_head {
92
	raw_spinlock_t lock;
G
Gleb Natapov 已提交
93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110
	struct hlist_head list;
} async_pf_sleepers[KVM_TASK_SLEEP_HASHSIZE];

static struct kvm_task_sleep_node *_find_apf_task(struct kvm_task_sleep_head *b,
						  u32 token)
{
	struct hlist_node *p;

	hlist_for_each(p, &b->list) {
		struct kvm_task_sleep_node *n =
			hlist_entry(p, typeof(*n), link);
		if (n->token == token)
			return n;
	}

	return NULL;
}

111 112 113 114 115
/*
 * @interrupt_kernel: Is this called from a routine which interrupts the kernel
 * 		      (other than user space)?
 */
void kvm_async_pf_task_wait(u32 token, int interrupt_kernel)
G
Gleb Natapov 已提交
116 117 118 119
{
	u32 key = hash_32(token, KVM_TASK_SLEEP_HASHBITS);
	struct kvm_task_sleep_head *b = &async_pf_sleepers[key];
	struct kvm_task_sleep_node n, *e;
120
	DECLARE_SWAITQUEUE(wait);
G
Gleb Natapov 已提交
121

122 123
	rcu_irq_enter();

124
	raw_spin_lock(&b->lock);
G
Gleb Natapov 已提交
125 126 127 128 129
	e = _find_apf_task(b, token);
	if (e) {
		/* dummy entry exist -> wake up was delivered ahead of PF */
		hlist_del(&e->link);
		kfree(e);
130
		raw_spin_unlock(&b->lock);
131 132

		rcu_irq_exit();
G
Gleb Natapov 已提交
133 134 135 136 137
		return;
	}

	n.token = token;
	n.cpu = smp_processor_id();
138 139 140 141
	n.halted = is_idle_task(current) ||
		   (IS_ENABLED(CONFIG_PREEMPT_COUNT)
		    ? preempt_count() > 1 || rcu_preempt_depth()
		    : interrupt_kernel);
142
	init_swait_queue_head(&n.wq);
G
Gleb Natapov 已提交
143
	hlist_add_head(&n.link, &b->list);
144
	raw_spin_unlock(&b->lock);
G
Gleb Natapov 已提交
145 146

	for (;;) {
147
		if (!n.halted)
148
			prepare_to_swait_exclusive(&n.wq, &wait, TASK_UNINTERRUPTIBLE);
G
Gleb Natapov 已提交
149 150
		if (hlist_unhashed(&n.link))
			break;
151

152 153
		rcu_irq_exit();

154 155 156 157 158 159 160 161 162 163 164
		if (!n.halted) {
			local_irq_enable();
			schedule();
			local_irq_disable();
		} else {
			/*
			 * We cannot reschedule. So halt.
			 */
			native_safe_halt();
			local_irq_disable();
		}
165 166

		rcu_irq_enter();
G
Gleb Natapov 已提交
167
	}
168
	if (!n.halted)
169
		finish_swait(&n.wq, &wait);
G
Gleb Natapov 已提交
170

171
	rcu_irq_exit();
G
Gleb Natapov 已提交
172 173 174 175 176 177 178
	return;
}
EXPORT_SYMBOL_GPL(kvm_async_pf_task_wait);

static void apf_task_wake_one(struct kvm_task_sleep_node *n)
{
	hlist_del_init(&n->link);
179 180
	if (n->halted)
		smp_send_reschedule(n->cpu);
181
	else if (swq_has_sleeper(&n->wq))
182
		swake_up_one(&n->wq);
G
Gleb Natapov 已提交
183 184 185 186 187 188 189 190 191
}

static void apf_task_wake_all(void)
{
	int i;

	for (i = 0; i < KVM_TASK_SLEEP_HASHSIZE; i++) {
		struct hlist_node *p, *next;
		struct kvm_task_sleep_head *b = &async_pf_sleepers[i];
192
		raw_spin_lock(&b->lock);
G
Gleb Natapov 已提交
193 194 195 196 197 198
		hlist_for_each_safe(p, next, &b->list) {
			struct kvm_task_sleep_node *n =
				hlist_entry(p, typeof(*n), link);
			if (n->cpu == smp_processor_id())
				apf_task_wake_one(n);
		}
199
		raw_spin_unlock(&b->lock);
G
Gleb Natapov 已提交
200 201 202 203 204 205 206 207 208 209 210 211 212 213 214
	}
}

void kvm_async_pf_task_wake(u32 token)
{
	u32 key = hash_32(token, KVM_TASK_SLEEP_HASHBITS);
	struct kvm_task_sleep_head *b = &async_pf_sleepers[key];
	struct kvm_task_sleep_node *n;

	if (token == ~0) {
		apf_task_wake_all();
		return;
	}

again:
215
	raw_spin_lock(&b->lock);
G
Gleb Natapov 已提交
216 217 218 219 220 221
	n = _find_apf_task(b, token);
	if (!n) {
		/*
		 * async PF was not yet handled.
		 * Add dummy entry for the token.
		 */
222
		n = kzalloc(sizeof(*n), GFP_ATOMIC);
G
Gleb Natapov 已提交
223 224 225 226 227
		if (!n) {
			/*
			 * Allocation failed! Busy wait while other cpu
			 * handles async PF.
			 */
228
			raw_spin_unlock(&b->lock);
G
Gleb Natapov 已提交
229 230 231 232 233
			cpu_relax();
			goto again;
		}
		n->token = token;
		n->cpu = smp_processor_id();
234
		init_swait_queue_head(&n->wq);
G
Gleb Natapov 已提交
235 236 237
		hlist_add_head(&n->link, &b->list);
	} else
		apf_task_wake_one(n);
238
	raw_spin_unlock(&b->lock);
G
Gleb Natapov 已提交
239 240 241 242 243 244 245 246
	return;
}
EXPORT_SYMBOL_GPL(kvm_async_pf_task_wake);

u32 kvm_read_and_reset_pf_reason(void)
{
	u32 reason = 0;

247 248 249
	if (__this_cpu_read(apf_reason.enabled)) {
		reason = __this_cpu_read(apf_reason.reason);
		__this_cpu_write(apf_reason.reason, 0);
G
Gleb Natapov 已提交
250 251 252 253 254
	}

	return reason;
}
EXPORT_SYMBOL_GPL(kvm_read_and_reset_pf_reason);
255
NOKPROBE_SYMBOL(kvm_read_and_reset_pf_reason);
G
Gleb Natapov 已提交
256

257
dotraplinkage void
G
Gleb Natapov 已提交
258 259
do_async_page_fault(struct pt_regs *regs, unsigned long error_code)
{
260 261
	enum ctx_state prev_state;

G
Gleb Natapov 已提交
262 263
	switch (kvm_read_and_reset_pf_reason()) {
	default:
264
		do_page_fault(regs, error_code);
G
Gleb Natapov 已提交
265 266 267
		break;
	case KVM_PV_REASON_PAGE_NOT_PRESENT:
		/* page is swapped out by the host. */
268
		prev_state = exception_enter();
269
		kvm_async_pf_task_wait((u32)read_cr2(), !user_mode(regs));
270
		exception_exit(prev_state);
G
Gleb Natapov 已提交
271 272
		break;
	case KVM_PV_REASON_PAGE_READY:
273
		rcu_irq_enter();
G
Gleb Natapov 已提交
274
		kvm_async_pf_task_wake((u32)read_cr2());
275
		rcu_irq_exit();
G
Gleb Natapov 已提交
276 277 278
		break;
	}
}
279
NOKPROBE_SYMBOL(do_async_page_fault);
G
Gleb Natapov 已提交
280

281
static void __init paravirt_ops_setup(void)
282 283
{
	pv_info.name = "KVM";
284

285
	if (kvm_para_has_feature(KVM_FEATURE_NOP_IO_DELAY))
286
		pv_ops.cpu.io_delay = kvm_io_delay;
287

288 289 290
#ifdef CONFIG_X86_IO_APIC
	no_timer_check = 1;
#endif
291 292
}

293 294 295 296 297 298 299 300
static void kvm_register_steal_time(void)
{
	int cpu = smp_processor_id();
	struct kvm_steal_time *st = &per_cpu(steal_time, cpu);

	if (!has_steal_clock)
		return;

301
	wrmsrl(MSR_KVM_STEAL_TIME, (slow_virt_to_phys(st) | KVM_MSR_ENABLED));
302 303
	pr_info("kvm-stealtime: cpu %d, msr %llx\n",
		cpu, (unsigned long long) slow_virt_to_phys(st));
304 305
}

306
static DEFINE_PER_CPU_DECRYPTED(unsigned long, kvm_apic_eoi) = KVM_PV_EOI_DISABLED;
307

308
static notrace void kvm_guest_apic_eoi_write(u32 reg, u32 val)
309 310 311 312 313 314 315 316
{
	/**
	 * This relies on __test_and_clear_bit to modify the memory
	 * in a way that is atomic with respect to the local CPU.
	 * The hypervisor only accesses this memory from the local CPU so
	 * there's no need for lock or memory barriers.
	 * An optimization barrier is implied in apic write.
	 */
317
	if (__test_and_clear_bit(KVM_PV_EOI_BIT, this_cpu_ptr(&kvm_apic_eoi)))
318
		return;
319
	apic->native_eoi_write(APIC_EOI, APIC_EOI_ACK);
320 321
}

322
static void kvm_guest_cpu_init(void)
323 324 325 326 327
{
	if (!kvm_para_available())
		return;

	if (kvm_para_has_feature(KVM_FEATURE_ASYNC_PF) && kvmapf) {
328
		u64 pa = slow_virt_to_phys(this_cpu_ptr(&apf_reason));
329

330 331 332
#ifdef CONFIG_PREEMPT
		pa |= KVM_ASYNC_PF_SEND_ALWAYS;
#endif
333 334
		pa |= KVM_ASYNC_PF_ENABLED;

335 336 337 338
		if (kvm_para_has_feature(KVM_FEATURE_ASYNC_PF_VMEXIT))
			pa |= KVM_ASYNC_PF_DELIVERY_AS_PF_VMEXIT;

		wrmsrl(MSR_KVM_ASYNC_PF_EN, pa);
339
		__this_cpu_write(apf_reason.enabled, 1);
340 341 342
		printk(KERN_INFO"KVM setup async PF for cpu %d\n",
		       smp_processor_id());
	}
343

344 345 346 347
	if (kvm_para_has_feature(KVM_FEATURE_PV_EOI)) {
		unsigned long pa;
		/* Size alignment is implied but just to make it explicit. */
		BUILD_BUG_ON(__alignof__(kvm_apic_eoi) < 4);
348 349
		__this_cpu_write(kvm_apic_eoi, 0);
		pa = slow_virt_to_phys(this_cpu_ptr(&kvm_apic_eoi))
350
			| KVM_MSR_ENABLED;
351 352 353
		wrmsrl(MSR_KVM_PV_EOI_EN, pa);
	}

354 355
	if (has_steal_clock)
		kvm_register_steal_time();
356 357
}

358
static void kvm_pv_disable_apf(void)
359
{
360
	if (!__this_cpu_read(apf_reason.enabled))
361 362 363
		return;

	wrmsrl(MSR_KVM_ASYNC_PF_EN, 0);
364
	__this_cpu_write(apf_reason.enabled, 0);
365 366 367 368 369

	printk(KERN_INFO"Unregister pv shared memory for cpu %d\n",
	       smp_processor_id());
}

370 371 372 373 374 375 376 377 378 379
static void kvm_pv_guest_cpu_reboot(void *unused)
{
	/*
	 * We disable PV EOI before we load a new kernel by kexec,
	 * since MSR_KVM_PV_EOI_EN stores a pointer into old kernel's memory.
	 * New kernel can re-enable when it boots.
	 */
	if (kvm_para_has_feature(KVM_FEATURE_PV_EOI))
		wrmsrl(MSR_KVM_PV_EOI_EN, 0);
	kvm_pv_disable_apf();
380
	kvm_disable_steal_time();
381 382
}

383 384 385 386
static int kvm_pv_reboot_notify(struct notifier_block *nb,
				unsigned long code, void *unused)
{
	if (code == SYS_RESTART)
387
		on_each_cpu(kvm_pv_guest_cpu_reboot, NULL, 1);
388 389 390 391 392 393 394
	return NOTIFY_DONE;
}

static struct notifier_block kvm_pv_reboot_nb = {
	.notifier_call = kvm_pv_reboot_notify,
};

395 396 397 398 399 400 401 402 403
static u64 kvm_steal_clock(int cpu)
{
	u64 steal;
	struct kvm_steal_time *src;
	int version;

	src = &per_cpu(steal_time, cpu);
	do {
		version = src->version;
404
		virt_rmb();
405
		steal = src->steal;
406
		virt_rmb();
407 408 409 410 411 412 413 414 415 416 417 418 419
	} while ((version & 1) || (version != src->version));

	return steal;
}

void kvm_disable_steal_time(void)
{
	if (!has_steal_clock)
		return;

	wrmsr(MSR_KVM_STEAL_TIME, 0, 0);
}

420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446
static inline void __set_percpu_decrypted(void *ptr, unsigned long size)
{
	early_set_memory_decrypted((unsigned long) ptr, size);
}

/*
 * Iterate through all possible CPUs and map the memory region pointed
 * by apf_reason, steal_time and kvm_apic_eoi as decrypted at once.
 *
 * Note: we iterate through all possible CPUs to ensure that CPUs
 * hotplugged will have their per-cpu variable already mapped as
 * decrypted.
 */
static void __init sev_map_percpu_data(void)
{
	int cpu;

	if (!sev_active())
		return;

	for_each_possible_cpu(cpu) {
		__set_percpu_decrypted(&per_cpu(apf_reason, cpu), sizeof(apf_reason));
		__set_percpu_decrypted(&per_cpu(steal_time, cpu), sizeof(steal_time));
		__set_percpu_decrypted(&per_cpu(kvm_apic_eoi, cpu), sizeof(kvm_apic_eoi));
	}
}

447
#ifdef CONFIG_SMP
448 449 450 451 452 453 454 455 456 457 458 459
#define KVM_IPI_CLUSTER_SIZE	(2 * BITS_PER_LONG)

static void __send_ipi_mask(const struct cpumask *mask, int vector)
{
	unsigned long flags;
	int cpu, apic_id, icr;
	int min = 0, max = 0;
#ifdef CONFIG_X86_64
	__uint128_t ipi_bitmap = 0;
#else
	u64 ipi_bitmap = 0;
#endif
460
	long ret;
461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485

	if (cpumask_empty(mask))
		return;

	local_irq_save(flags);

	switch (vector) {
	default:
		icr = APIC_DM_FIXED | vector;
		break;
	case NMI_VECTOR:
		icr = APIC_DM_NMI;
		break;
	}

	for_each_cpu(cpu, mask) {
		apic_id = per_cpu(x86_cpu_to_apicid, cpu);
		if (!ipi_bitmap) {
			min = max = apic_id;
		} else if (apic_id < min && max - apic_id < KVM_IPI_CLUSTER_SIZE) {
			ipi_bitmap <<= min - apic_id;
			min = apic_id;
		} else if (apic_id < min + KVM_IPI_CLUSTER_SIZE) {
			max = apic_id < max ? max : apic_id;
		} else {
486
			ret = kvm_hypercall4(KVM_HC_SEND_IPI, (unsigned long)ipi_bitmap,
487
				(unsigned long)(ipi_bitmap >> BITS_PER_LONG), min, icr);
488
			WARN_ONCE(ret < 0, "KVM: failed to send PV IPI: %ld", ret);
489 490 491 492 493 494 495
			min = max = apic_id;
			ipi_bitmap = 0;
		}
		__set_bit(apic_id - min, (unsigned long *)&ipi_bitmap);
	}

	if (ipi_bitmap) {
496
		ret = kvm_hypercall4(KVM_HC_SEND_IPI, (unsigned long)ipi_bitmap,
497
			(unsigned long)(ipi_bitmap >> BITS_PER_LONG), min, icr);
498
		WARN_ONCE(ret < 0, "KVM: failed to send PV IPI: %ld", ret);
499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542
	}

	local_irq_restore(flags);
}

static void kvm_send_ipi_mask(const struct cpumask *mask, int vector)
{
	__send_ipi_mask(mask, vector);
}

static void kvm_send_ipi_mask_allbutself(const struct cpumask *mask, int vector)
{
	unsigned int this_cpu = smp_processor_id();
	struct cpumask new_mask;
	const struct cpumask *local_mask;

	cpumask_copy(&new_mask, mask);
	cpumask_clear_cpu(this_cpu, &new_mask);
	local_mask = &new_mask;
	__send_ipi_mask(local_mask, vector);
}

static void kvm_send_ipi_allbutself(int vector)
{
	kvm_send_ipi_mask_allbutself(cpu_online_mask, vector);
}

static void kvm_send_ipi_all(int vector)
{
	__send_ipi_mask(cpu_online_mask, vector);
}

/*
 * Set the IPI entry points
 */
static void kvm_setup_pv_ipi(void)
{
	apic->send_IPI_mask = kvm_send_ipi_mask;
	apic->send_IPI_mask_allbutself = kvm_send_ipi_mask_allbutself;
	apic->send_IPI_allbutself = kvm_send_ipi_allbutself;
	apic->send_IPI_all = kvm_send_ipi_all;
	pr_info("KVM setup pv IPIs\n");
}

543 544 545
static void __init kvm_smp_prepare_cpus(unsigned int max_cpus)
{
	native_smp_prepare_cpus(max_cpus);
546
	if (kvm_para_has_hint(KVM_HINTS_REALTIME))
547 548 549
		static_branch_disable(&virt_spin_lock_key);
}

550 551
static void __init kvm_smp_prepare_boot_cpu(void)
{
552 553 554 555 556 557
	/*
	 * Map the per-cpu variables as decrypted before kvm_guest_cpu_init()
	 * shares the guest physical address with the hypervisor.
	 */
	sev_map_percpu_data();

558
	kvm_guest_cpu_init();
559
	native_smp_prepare_boot_cpu();
560
	kvm_spinlock_init();
561
}
562

563
static void kvm_guest_cpu_offline(void)
564
{
565
	kvm_disable_steal_time();
566 567 568
	if (kvm_para_has_feature(KVM_FEATURE_PV_EOI))
		wrmsrl(MSR_KVM_PV_EOI_EN, 0);
	kvm_pv_disable_apf();
G
Gleb Natapov 已提交
569
	apf_task_wake_all();
570 571
}

572
static int kvm_cpu_online(unsigned int cpu)
573
{
574 575 576 577
	local_irq_disable();
	kvm_guest_cpu_init();
	local_irq_enable();
	return 0;
578 579
}

580 581 582 583 584 585 586
static int kvm_cpu_down_prepare(unsigned int cpu)
{
	local_irq_disable();
	kvm_guest_cpu_offline();
	local_irq_enable();
	return 0;
}
587 588
#endif

G
Gleb Natapov 已提交
589 590
static void __init kvm_apf_trap_init(void)
{
T
Thomas Gleixner 已提交
591
	update_intr_gate(X86_TRAP_PF, async_page_fault);
G
Gleb Natapov 已提交
592 593
}

594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621
static DEFINE_PER_CPU(cpumask_var_t, __pv_tlb_mask);

static void kvm_flush_tlb_others(const struct cpumask *cpumask,
			const struct flush_tlb_info *info)
{
	u8 state;
	int cpu;
	struct kvm_steal_time *src;
	struct cpumask *flushmask = this_cpu_cpumask_var_ptr(__pv_tlb_mask);

	cpumask_copy(flushmask, cpumask);
	/*
	 * We have to call flush only on online vCPUs. And
	 * queue flush_on_enter for pre-empted vCPUs
	 */
	for_each_cpu(cpu, flushmask) {
		src = &per_cpu(steal_time, cpu);
		state = READ_ONCE(src->preempted);
		if ((state & KVM_VCPU_PREEMPTED)) {
			if (try_cmpxchg(&src->preempted, &state,
					state | KVM_VCPU_FLUSH_TLB))
				__cpumask_clear_cpu(cpu, flushmask);
		}
	}

	native_flush_tlb_others(flushmask, info);
}

622
static void __init kvm_guest_init(void)
623
{
G
Gleb Natapov 已提交
624 625
	int i;

626 627 628 629
	if (!kvm_para_available())
		return;

	paravirt_ops_setup();
630
	register_reboot_notifier(&kvm_pv_reboot_nb);
G
Gleb Natapov 已提交
631
	for (i = 0; i < KVM_TASK_SLEEP_HASHSIZE; i++)
632
		raw_spin_lock_init(&async_pf_sleepers[i].lock);
G
Gleb Natapov 已提交
633 634 635
	if (kvm_para_has_feature(KVM_FEATURE_ASYNC_PF))
		x86_init.irqs.trap_init = kvm_apf_trap_init;

636 637
	if (kvm_para_has_feature(KVM_FEATURE_STEAL_TIME)) {
		has_steal_clock = 1;
638
		pv_ops.time.steal_clock = kvm_steal_clock;
639 640
	}

641
	if (kvm_para_has_feature(KVM_FEATURE_PV_TLB_FLUSH) &&
642
	    !kvm_para_has_hint(KVM_HINTS_REALTIME) &&
643
	    kvm_para_has_feature(KVM_FEATURE_STEAL_TIME)) {
644 645
		pv_ops.mmu.flush_tlb_others = kvm_flush_tlb_others;
		pv_ops.mmu.tlb_remove_table = tlb_remove_table;
646
	}
647

648 649
	if (kvm_para_has_feature(KVM_FEATURE_PV_EOI))
		apic_set_eoi_write(kvm_guest_apic_eoi_write);
650

651
#ifdef CONFIG_SMP
652
	smp_ops.smp_prepare_cpus = kvm_smp_prepare_cpus;
653
	smp_ops.smp_prepare_boot_cpu = kvm_smp_prepare_boot_cpu;
654 655 656
	if (cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN, "x86/kvm:online",
				      kvm_cpu_online, kvm_cpu_down_prepare) < 0)
		pr_err("kvm_guest: Failed to install cpu hotplug callbacks\n");
657
#else
658
	sev_map_percpu_data();
659
	kvm_guest_cpu_init();
660
#endif
661 662 663 664 665 666

	/*
	 * Hard lockup detection is enabled by default. Disable it, as guests
	 * can get false positives too easily, for example if the host is
	 * overcommitted.
	 */
667
	hardlockup_detector_disable();
668
}
669

670 671 672 673 674
static noinline uint32_t __kvm_cpuid_base(void)
{
	if (boot_cpu_data.cpuid_level < 0)
		return 0;	/* So we don't blow up on old processors */

675
	if (boot_cpu_has(X86_FEATURE_HYPERVISOR))
676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696
		return hypervisor_cpuid_base("KVMKVMKVM\0\0\0", 0);

	return 0;
}

static inline uint32_t kvm_cpuid_base(void)
{
	static int kvm_cpuid_base = -1;

	if (kvm_cpuid_base == -1)
		kvm_cpuid_base = __kvm_cpuid_base();

	return kvm_cpuid_base;
}

bool kvm_para_available(void)
{
	return kvm_cpuid_base() != 0;
}
EXPORT_SYMBOL_GPL(kvm_para_available);

697 698 699 700 701
unsigned int kvm_arch_para_features(void)
{
	return cpuid_eax(kvm_cpuid_base() | KVM_CPUID_FEATURES);
}

702 703 704 705 706
unsigned int kvm_arch_para_hints(void)
{
	return cpuid_edx(kvm_cpuid_base() | KVM_CPUID_FEATURES);
}

J
Jason Wang 已提交
707
static uint32_t __init kvm_detect(void)
708
{
J
Jason Wang 已提交
709
	return kvm_cpuid_base();
710 711
}

712 713
static void __init kvm_apic_init(void)
{
714 715 716 717
#if defined(CONFIG_SMP)
	if (kvm_para_has_feature(KVM_FEATURE_PV_SEND_IPI))
		kvm_setup_pv_ipi();
#endif
718 719 720 721
}

static void __init kvm_init_platform(void)
{
722
	kvmclock_init();
723 724 725
	x86_platform.apic_post_init = kvm_apic_init;
}

726
const __initconst struct hypervisor_x86 x86_hyper_kvm = {
727 728
	.name			= "KVM",
	.detect			= kvm_detect,
729
	.type			= X86_HYPER_KVM,
730
	.init.guest_late_init	= kvm_guest_init,
731
	.init.x2apic_available	= kvm_para_available,
732
	.init.init_platform	= kvm_init_platform,
733 734
};

735 736 737
static __init int activate_jump_labels(void)
{
	if (has_steal_clock) {
738
		static_key_slow_inc(&paravirt_steal_enabled);
739
		if (steal_acc)
740
			static_key_slow_inc(&paravirt_steal_rq_enabled);
741 742 743 744 745
	}

	return 0;
}
arch_initcall(activate_jump_labels);
746

747 748 749 750
static __init int kvm_setup_pv_tlb_flush(void)
{
	int cpu;

751
	if (kvm_para_has_feature(KVM_FEATURE_PV_TLB_FLUSH) &&
752
	    !kvm_para_has_hint(KVM_HINTS_REALTIME) &&
753
	    kvm_para_has_feature(KVM_FEATURE_STEAL_TIME)) {
754 755 756 757 758 759 760 761 762 763 764
		for_each_possible_cpu(cpu) {
			zalloc_cpumask_var_node(per_cpu_ptr(&__pv_tlb_mask, cpu),
				GFP_KERNEL, cpu_to_node(cpu));
		}
		pr_info("KVM setup pv remote TLB flush\n");
	}

	return 0;
}
arch_initcall(kvm_setup_pv_tlb_flush);

765 766 767
#ifdef CONFIG_PARAVIRT_SPINLOCKS

/* Kick a cpu by its apicid. Used to wake up a halted vcpu */
768
static void kvm_kick_cpu(int cpu)
769 770 771 772 773 774 775 776
{
	int apicid;
	unsigned long flags = 0;

	apicid = per_cpu(x86_cpu_to_apicid, cpu);
	kvm_hypercall2(KVM_HC_KICK_CPU, flags, apicid);
}

777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804
#include <asm/qspinlock.h>

static void kvm_wait(u8 *ptr, u8 val)
{
	unsigned long flags;

	if (in_nmi())
		return;

	local_irq_save(flags);

	if (READ_ONCE(*ptr) != val)
		goto out;

	/*
	 * halt until it's our turn and kicked. Note that we do safe halt
	 * for irq enabled case to avoid hang when lock info is overwritten
	 * in irq spinlock slowpath and no spurious interrupt occur to save us.
	 */
	if (arch_irqs_disabled_flags(flags))
		halt();
	else
		safe_halt();

out:
	local_irq_restore(flags);
}

805
#ifdef CONFIG_X86_32
806
__visible bool __kvm_vcpu_is_preempted(long cpu)
807 808 809
{
	struct kvm_steal_time *src = &per_cpu(steal_time, cpu);

W
Wanpeng Li 已提交
810
	return !!(src->preempted & KVM_VCPU_PREEMPTED);
811 812 813
}
PV_CALLEE_SAVE_REGS_THUNK(__kvm_vcpu_is_preempted);

814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836
#else

#include <asm/asm-offsets.h>

extern bool __raw_callee_save___kvm_vcpu_is_preempted(long);

/*
 * Hand-optimize version for x86-64 to avoid 8 64-bit register saving and
 * restoring to/from the stack.
 */
asm(
".pushsection .text;"
".global __raw_callee_save___kvm_vcpu_is_preempted;"
".type __raw_callee_save___kvm_vcpu_is_preempted, @function;"
"__raw_callee_save___kvm_vcpu_is_preempted:"
"movq	__per_cpu_offset(,%rdi,8), %rax;"
"cmpb	$0, " __stringify(KVM_STEAL_TIME_preempted) "+steal_time(%rax);"
"setne	%al;"
"ret;"
".popsection");

#endif

837 838 839 840 841 842 843 844 845 846 847
/*
 * Setup pv_lock_ops to exploit KVM_FEATURE_PV_UNHALT if present.
 */
void __init kvm_spinlock_init(void)
{
	if (!kvm_para_available())
		return;
	/* Does host kernel support KVM_FEATURE_PV_UNHALT? */
	if (!kvm_para_has_feature(KVM_FEATURE_PV_UNHALT))
		return;

848
	if (kvm_para_has_hint(KVM_HINTS_REALTIME))
849 850
		return;

851 852 853 854
	/* Don't use the pvqspinlock code if there is only 1 vCPU. */
	if (num_possible_cpus() == 1)
		return;

855
	__pv_init_lock_hash();
856 857 858 859 860
	pv_ops.lock.queued_spin_lock_slowpath = __pv_queued_spin_lock_slowpath;
	pv_ops.lock.queued_spin_unlock =
		PV_CALLEE_SAVE(__pv_queued_spin_unlock);
	pv_ops.lock.wait = kvm_wait;
	pv_ops.lock.kick = kvm_kick_cpu;
861 862

	if (kvm_para_has_feature(KVM_FEATURE_STEAL_TIME)) {
863
		pv_ops.lock.vcpu_is_preempted =
864 865
			PV_CALLEE_SAVE(__kvm_vcpu_is_preempted);
	}
866 867
}

868
#endif	/* CONFIG_PARAVIRT_SPINLOCKS */