提交 573014c3 编写于 作者: V Vitaly Kuznetsov 提交者: Zheng Zengkai

x86/kvm: Teardown PV features on boot CPU as well

stable inclusion
from stable-5.10.43
commit 38b858da1c58ad46519a257764e059e663b59ff2
bugzilla: 109284
CVE: NA

--------------------------------

commit 8b79feff upstream.

Various PV features (Async PF, PV EOI, steal time) work through memory
shared with hypervisor and when we restore from hibernation we must
properly teardown all these features to make sure hypervisor doesn't
write to stale locations after we jump to the previously hibernated kernel
(which can try to place anything there). For secondary CPUs the job is
already done by kvm_cpu_down_prepare(), register syscore ops to do
the same for boot CPU.

Krzysztof:
This fixes memory corruption visible after second resume from
hibernation:

  BUG: Bad page state in process dbus-daemon  pfn:18b01
  page:ffffea000062c040 refcount:0 mapcount:0 mapping:0000000000000000 index:0x1 compound_mapcount: -30591
  flags: 0xfffffc0078141(locked|error|workingset|writeback|head|mappedtodisk|reclaim)
  raw: 000fffffc0078141 dead0000000002d0 dead000000000100 0000000000000000
  raw: 0000000000000001 0000000000000000 00000000ffffffff 0000000000000000
  page dumped because: PAGE_FLAGS_CHECK_AT_PREP flag set
  bad because of flags: 0x78141(locked|error|workingset|writeback|head|mappedtodisk|reclaim)
Signed-off-by: NVitaly Kuznetsov <vkuznets@redhat.com>
Message-Id: <20210414123544.1060604-3-vkuznets@redhat.com>
Signed-off-by: NPaolo Bonzini <pbonzini@redhat.com>
Signed-off-by: NAndrea Righi <andrea.righi@canonical.com>
[krzysztof: Extend the commit message, adjust for v5.10 context]
Signed-off-by: NKrzysztof Kozlowski <krzysztof.kozlowski@canonical.com>
Signed-off-by: NGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Signed-off-by: NChen Jun <chenjun102@huawei.com>
Signed-off-by: NZheng Zengkai <zhengzengkai@huawei.com>
上级 165437dc
...@@ -26,6 +26,7 @@ ...@@ -26,6 +26,7 @@
#include <linux/kprobes.h> #include <linux/kprobes.h>
#include <linux/nmi.h> #include <linux/nmi.h>
#include <linux/swait.h> #include <linux/swait.h>
#include <linux/syscore_ops.h>
#include <asm/timer.h> #include <asm/timer.h>
#include <asm/cpu.h> #include <asm/cpu.h>
#include <asm/traps.h> #include <asm/traps.h>
...@@ -460,6 +461,25 @@ static bool pv_tlb_flush_supported(void) ...@@ -460,6 +461,25 @@ static bool pv_tlb_flush_supported(void)
static DEFINE_PER_CPU(cpumask_var_t, __pv_cpu_mask); static DEFINE_PER_CPU(cpumask_var_t, __pv_cpu_mask);
static void kvm_guest_cpu_offline(void)
{
kvm_disable_steal_time();
if (kvm_para_has_feature(KVM_FEATURE_PV_EOI))
wrmsrl(MSR_KVM_PV_EOI_EN, 0);
kvm_pv_disable_apf();
apf_task_wake_all();
}
static int kvm_cpu_online(unsigned int cpu)
{
unsigned long flags;
local_irq_save(flags);
kvm_guest_cpu_init();
local_irq_restore(flags);
return 0;
}
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
static bool pv_ipi_supported(void) static bool pv_ipi_supported(void)
...@@ -587,31 +607,34 @@ static void __init kvm_smp_prepare_boot_cpu(void) ...@@ -587,31 +607,34 @@ static void __init kvm_smp_prepare_boot_cpu(void)
kvm_spinlock_init(); kvm_spinlock_init();
} }
static void kvm_guest_cpu_offline(void) static int kvm_cpu_down_prepare(unsigned int cpu)
{ {
kvm_disable_steal_time(); unsigned long flags;
if (kvm_para_has_feature(KVM_FEATURE_PV_EOI))
wrmsrl(MSR_KVM_PV_EOI_EN, 0);
kvm_pv_disable_apf();
apf_task_wake_all();
}
static int kvm_cpu_online(unsigned int cpu) local_irq_save(flags);
{ kvm_guest_cpu_offline();
local_irq_disable(); local_irq_restore(flags);
kvm_guest_cpu_init();
local_irq_enable();
return 0; return 0;
} }
static int kvm_cpu_down_prepare(unsigned int cpu) #endif
static int kvm_suspend(void)
{ {
local_irq_disable();
kvm_guest_cpu_offline(); kvm_guest_cpu_offline();
local_irq_enable();
return 0; return 0;
} }
#endif
static void kvm_resume(void)
{
kvm_cpu_online(raw_smp_processor_id());
}
static struct syscore_ops kvm_syscore_ops = {
.suspend = kvm_suspend,
.resume = kvm_resume,
};
static void kvm_flush_tlb_others(const struct cpumask *cpumask, static void kvm_flush_tlb_others(const struct cpumask *cpumask,
const struct flush_tlb_info *info) const struct flush_tlb_info *info)
...@@ -681,6 +704,8 @@ static void __init kvm_guest_init(void) ...@@ -681,6 +704,8 @@ static void __init kvm_guest_init(void)
kvm_guest_cpu_init(); kvm_guest_cpu_init();
#endif #endif
register_syscore_ops(&kvm_syscore_ops);
/* /*
* Hard lockup detection is enabled by default. Disable it, as guests * Hard lockup detection is enabled by default. Disable it, as guests
* can get false positives too easily, for example if the host is * can get false positives too easily, for example if the host is
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册