提交 6129ed87 编写于 作者: S Sean Christopherson 提交者: Paolo Bonzini

KVM: x86/mmu: Set mmio_value to '0' if reserved #PF can't be generated

Set the mmio_value to '0' instead of simply clearing the present bit to
squash a benign warning in kvm_mmu_set_mmio_spte_mask() that complains
about the mmio_value overlapping the lower GFN mask on systems with 52
bits of PA space.

Opportunistically clean up the code and comments.

Cc: stable@vger.kernel.org
Fixes: d43e2675 ("KVM: x86: only do L1TF workaround on affected processors")
Signed-off-by: NSean Christopherson <sean.j.christopherson@intel.com>
Message-Id: <20200527084909.23492-1-sean.j.christopherson@intel.com>
Signed-off-by: NPaolo Bonzini <pbonzini@redhat.com>
上级 d43e2675
...@@ -6143,25 +6143,16 @@ static void kvm_set_mmio_spte_mask(void) ...@@ -6143,25 +6143,16 @@ static void kvm_set_mmio_spte_mask(void)
u64 mask; u64 mask;
/* /*
* Set the reserved bits and the present bit of an paging-structure * Set a reserved PA bit in MMIO SPTEs to generate page faults with
* entry to generate page fault with PFER.RSV = 1. * PFEC.RSVD=1 on MMIO accesses. 64-bit PTEs (PAE, x86-64, and EPT
* paging) support a maximum of 52 bits of PA, i.e. if the CPU supports
* 52-bit physical addresses then there are no reserved PA bits in the
* PTEs and so the reserved PA approach must be disabled.
*/ */
if (shadow_phys_bits < 52)
/* mask = BIT_ULL(51) | PT_PRESENT_MASK;
* Mask the uppermost physical address bit, which would be reserved as else
* long as the supported physical address width is less than 52. mask = 0;
*/
mask = 1ull << 51;
/* Set the present bit. */
mask |= 1ull;
/*
* If reserved bit is not supported, clear the present bit to disable
* mmio page fault.
*/
if (shadow_phys_bits == 52)
mask &= ~1ull;
kvm_mmu_set_mmio_spte_mask(mask, mask, ACC_WRITE_MASK | ACC_USER_MASK); kvm_mmu_set_mmio_spte_mask(mask, mask, ACC_WRITE_MASK | ACC_USER_MASK);
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册