提交 1218cfc2 编写于 作者: V Vitaly Kuznetsov 提交者: Zheng Zengkai

x86/kvm: Disable all PV features on crash

stable inclusion
from stable-5.10.43
commit 334c59d58de5faf449d9c9feaa8c50dd8b4046a7
bugzilla: 109284
CVE: NA

--------------------------------

commit 3d6b8413 upstream.

Crash shutdown handler only disables kvmclock and steal time, other PV
features remain active so we risk corrupting memory or getting some
side-effects in kdump kernel. Move crash handler to kvm.c and unify
with CPU offline.
Signed-off-by: NVitaly Kuznetsov <vkuznets@redhat.com>
Message-Id: <20210414123544.1060604-5-vkuznets@redhat.com>
Signed-off-by: NPaolo Bonzini <pbonzini@redhat.com>
Signed-off-by: NKrzysztof Kozlowski <krzysztof.kozlowski@canonical.com>
Signed-off-by: NGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Signed-off-by: NChen Jun <chenjun102@huawei.com>
Signed-off-by: NZheng Zengkai <zhengzengkai@huawei.com>
上级 44b95584
...@@ -92,7 +92,6 @@ unsigned int kvm_arch_para_hints(void); ...@@ -92,7 +92,6 @@ unsigned int kvm_arch_para_hints(void);
void kvm_async_pf_task_wait_schedule(u32 token); void kvm_async_pf_task_wait_schedule(u32 token);
void kvm_async_pf_task_wake(u32 token); void kvm_async_pf_task_wake(u32 token);
u32 kvm_read_and_reset_apf_flags(void); u32 kvm_read_and_reset_apf_flags(void);
void kvm_disable_steal_time(void);
bool __kvm_handle_async_pf(struct pt_regs *regs, u32 token); bool __kvm_handle_async_pf(struct pt_regs *regs, u32 token);
DECLARE_STATIC_KEY_FALSE(kvm_async_pf_enabled); DECLARE_STATIC_KEY_FALSE(kvm_async_pf_enabled);
...@@ -137,11 +136,6 @@ static inline u32 kvm_read_and_reset_apf_flags(void) ...@@ -137,11 +136,6 @@ static inline u32 kvm_read_and_reset_apf_flags(void)
return 0; return 0;
} }
static inline void kvm_disable_steal_time(void)
{
return;
}
static __always_inline bool kvm_handle_async_pf(struct pt_regs *regs, u32 token) static __always_inline bool kvm_handle_async_pf(struct pt_regs *regs, u32 token)
{ {
return false; return false;
......
...@@ -38,6 +38,7 @@ ...@@ -38,6 +38,7 @@
#include <asm/tlb.h> #include <asm/tlb.h>
#include <asm/cpuidle_haltpoll.h> #include <asm/cpuidle_haltpoll.h>
#include <asm/ptrace.h> #include <asm/ptrace.h>
#include <asm/reboot.h>
#include <asm/svm.h> #include <asm/svm.h>
DEFINE_STATIC_KEY_FALSE(kvm_async_pf_enabled); DEFINE_STATIC_KEY_FALSE(kvm_async_pf_enabled);
...@@ -375,6 +376,14 @@ static void kvm_pv_disable_apf(void) ...@@ -375,6 +376,14 @@ static void kvm_pv_disable_apf(void)
pr_info("Unregister pv shared memory for cpu %d\n", smp_processor_id()); pr_info("Unregister pv shared memory for cpu %d\n", smp_processor_id());
} }
static void kvm_disable_steal_time(void)
{
if (!has_steal_clock)
return;
wrmsr(MSR_KVM_STEAL_TIME, 0, 0);
}
static void kvm_pv_guest_cpu_reboot(void *unused) static void kvm_pv_guest_cpu_reboot(void *unused)
{ {
/* /*
...@@ -417,14 +426,6 @@ static u64 kvm_steal_clock(int cpu) ...@@ -417,14 +426,6 @@ static u64 kvm_steal_clock(int cpu)
return steal; return steal;
} }
void kvm_disable_steal_time(void)
{
if (!has_steal_clock)
return;
wrmsr(MSR_KVM_STEAL_TIME, 0, 0);
}
static inline void __set_percpu_decrypted(void *ptr, unsigned long size) static inline void __set_percpu_decrypted(void *ptr, unsigned long size)
{ {
early_set_memory_decrypted((unsigned long) ptr, size); early_set_memory_decrypted((unsigned long) ptr, size);
...@@ -461,12 +462,13 @@ static bool pv_tlb_flush_supported(void) ...@@ -461,12 +462,13 @@ static bool pv_tlb_flush_supported(void)
static DEFINE_PER_CPU(cpumask_var_t, __pv_cpu_mask); static DEFINE_PER_CPU(cpumask_var_t, __pv_cpu_mask);
static void kvm_guest_cpu_offline(void) static void kvm_guest_cpu_offline(bool shutdown)
{ {
kvm_disable_steal_time(); kvm_disable_steal_time();
if (kvm_para_has_feature(KVM_FEATURE_PV_EOI)) if (kvm_para_has_feature(KVM_FEATURE_PV_EOI))
wrmsrl(MSR_KVM_PV_EOI_EN, 0); wrmsrl(MSR_KVM_PV_EOI_EN, 0);
kvm_pv_disable_apf(); kvm_pv_disable_apf();
if (!shutdown)
apf_task_wake_all(); apf_task_wake_all();
kvmclock_disable(); kvmclock_disable();
} }
...@@ -613,7 +615,7 @@ static int kvm_cpu_down_prepare(unsigned int cpu) ...@@ -613,7 +615,7 @@ static int kvm_cpu_down_prepare(unsigned int cpu)
unsigned long flags; unsigned long flags;
local_irq_save(flags); local_irq_save(flags);
kvm_guest_cpu_offline(); kvm_guest_cpu_offline(false);
local_irq_restore(flags); local_irq_restore(flags);
return 0; return 0;
} }
...@@ -622,7 +624,7 @@ static int kvm_cpu_down_prepare(unsigned int cpu) ...@@ -622,7 +624,7 @@ static int kvm_cpu_down_prepare(unsigned int cpu)
static int kvm_suspend(void) static int kvm_suspend(void)
{ {
kvm_guest_cpu_offline(); kvm_guest_cpu_offline(false);
return 0; return 0;
} }
...@@ -637,6 +639,20 @@ static struct syscore_ops kvm_syscore_ops = { ...@@ -637,6 +639,20 @@ static struct syscore_ops kvm_syscore_ops = {
.resume = kvm_resume, .resume = kvm_resume,
}; };
/*
* After a PV feature is registered, the host will keep writing to the
* registered memory location. If the guest happens to shutdown, this memory
* won't be valid. In cases like kexec, in which you install a new kernel, this
* means a random memory location will be kept being written.
*/
#ifdef CONFIG_KEXEC_CORE
static void kvm_crash_shutdown(struct pt_regs *regs)
{
kvm_guest_cpu_offline(true);
native_machine_crash_shutdown(regs);
}
#endif
static void kvm_flush_tlb_others(const struct cpumask *cpumask, static void kvm_flush_tlb_others(const struct cpumask *cpumask,
const struct flush_tlb_info *info) const struct flush_tlb_info *info)
{ {
...@@ -705,6 +721,10 @@ static void __init kvm_guest_init(void) ...@@ -705,6 +721,10 @@ static void __init kvm_guest_init(void)
kvm_guest_cpu_init(); kvm_guest_cpu_init();
#endif #endif
#ifdef CONFIG_KEXEC_CORE
machine_ops.crash_shutdown = kvm_crash_shutdown;
#endif
register_syscore_ops(&kvm_syscore_ops); register_syscore_ops(&kvm_syscore_ops);
/* /*
......
...@@ -20,7 +20,6 @@ ...@@ -20,7 +20,6 @@
#include <asm/hypervisor.h> #include <asm/hypervisor.h>
#include <asm/mem_encrypt.h> #include <asm/mem_encrypt.h>
#include <asm/x86_init.h> #include <asm/x86_init.h>
#include <asm/reboot.h>
#include <asm/kvmclock.h> #include <asm/kvmclock.h>
static int kvmclock __initdata = 1; static int kvmclock __initdata = 1;
...@@ -204,23 +203,6 @@ static void kvm_setup_secondary_clock(void) ...@@ -204,23 +203,6 @@ static void kvm_setup_secondary_clock(void)
} }
#endif #endif
/*
* After the clock is registered, the host will keep writing to the
* registered memory location. If the guest happens to shutdown, this memory
* won't be valid. In cases like kexec, in which you install a new kernel, this
* means a random memory location will be kept being written. So before any
* kind of shutdown from our side, we unregister the clock by writing anything
* that does not have the 'enable' bit set in the msr
*/
#ifdef CONFIG_KEXEC_CORE
static void kvm_crash_shutdown(struct pt_regs *regs)
{
native_write_msr(msr_kvm_system_time, 0, 0);
kvm_disable_steal_time();
native_machine_crash_shutdown(regs);
}
#endif
void kvmclock_disable(void) void kvmclock_disable(void)
{ {
native_write_msr(msr_kvm_system_time, 0, 0); native_write_msr(msr_kvm_system_time, 0, 0);
...@@ -350,9 +332,6 @@ void __init kvmclock_init(void) ...@@ -350,9 +332,6 @@ void __init kvmclock_init(void)
#endif #endif
x86_platform.save_sched_clock_state = kvm_save_sched_clock_state; x86_platform.save_sched_clock_state = kvm_save_sched_clock_state;
x86_platform.restore_sched_clock_state = kvm_restore_sched_clock_state; x86_platform.restore_sched_clock_state = kvm_restore_sched_clock_state;
#ifdef CONFIG_KEXEC_CORE
machine_ops.crash_shutdown = kvm_crash_shutdown;
#endif
kvm_get_preset_lpj(); kvm_get_preset_lpj();
/* /*
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册