提交 05b3e0c2 编写于 作者: A Avi Kivity 提交者: Linus Torvalds

[PATCH] KVM: Replace __x86_64__ with CONFIG_X86_64

As per akpm's request.
Signed-off-by: NAvi Kivity <avi@qumranet.com>
Signed-off-by: NAndrew Morton <akpm@osdl.org>
Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
上级 5aff458e
...@@ -140,7 +140,7 @@ enum { ...@@ -140,7 +140,7 @@ enum {
VCPU_REGS_RBP = 5, VCPU_REGS_RBP = 5,
VCPU_REGS_RSI = 6, VCPU_REGS_RSI = 6,
VCPU_REGS_RDI = 7, VCPU_REGS_RDI = 7,
#ifdef __x86_64__ #ifdef CONFIG_X86_64
VCPU_REGS_R8 = 8, VCPU_REGS_R8 = 8,
VCPU_REGS_R9 = 9, VCPU_REGS_R9 = 9,
VCPU_REGS_R10 = 10, VCPU_REGS_R10 = 10,
...@@ -375,7 +375,7 @@ void set_cr4(struct kvm_vcpu *vcpu, unsigned long cr0); ...@@ -375,7 +375,7 @@ void set_cr4(struct kvm_vcpu *vcpu, unsigned long cr0);
void set_cr8(struct kvm_vcpu *vcpu, unsigned long cr0); void set_cr8(struct kvm_vcpu *vcpu, unsigned long cr0);
void lmsw(struct kvm_vcpu *vcpu, unsigned long msw); void lmsw(struct kvm_vcpu *vcpu, unsigned long msw);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
void set_efer(struct kvm_vcpu *vcpu, u64 efer); void set_efer(struct kvm_vcpu *vcpu, u64 efer);
#endif #endif
...@@ -485,7 +485,7 @@ static inline unsigned long read_tr_base(void) ...@@ -485,7 +485,7 @@ static inline unsigned long read_tr_base(void)
return segment_base(tr); return segment_base(tr);
} }
#ifdef __x86_64__ #ifdef CONFIG_X86_64
static inline unsigned long read_msr(unsigned long msr) static inline unsigned long read_msr(unsigned long msr)
{ {
u64 value; u64 value;
...@@ -533,7 +533,7 @@ static inline u32 get_rdx_init_val(void) ...@@ -533,7 +533,7 @@ static inline u32 get_rdx_init_val(void)
#define TSS_REDIRECTION_SIZE (256 / 8) #define TSS_REDIRECTION_SIZE (256 / 8)
#define RMODE_TSS_SIZE (TSS_BASE_SIZE + TSS_REDIRECTION_SIZE + TSS_IOPB_SIZE + 1) #define RMODE_TSS_SIZE (TSS_BASE_SIZE + TSS_REDIRECTION_SIZE + TSS_IOPB_SIZE + 1)
#ifdef __x86_64__ #ifdef CONFIG_X86_64
/* /*
* When emulating 32-bit mode, cr3 is only 32 bits even on x86_64. Therefore * When emulating 32-bit mode, cr3 is only 32 bits even on x86_64. Therefore
......
...@@ -83,7 +83,7 @@ struct vmx_msr_entry *find_msr_entry(struct kvm_vcpu *vcpu, u32 msr) ...@@ -83,7 +83,7 @@ struct vmx_msr_entry *find_msr_entry(struct kvm_vcpu *vcpu, u32 msr)
} }
EXPORT_SYMBOL_GPL(find_msr_entry); EXPORT_SYMBOL_GPL(find_msr_entry);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
// LDT or TSS descriptor in the GDT. 16 bytes. // LDT or TSS descriptor in the GDT. 16 bytes.
struct segment_descriptor_64 { struct segment_descriptor_64 {
struct segment_descriptor s; struct segment_descriptor s;
...@@ -115,7 +115,7 @@ unsigned long segment_base(u16 selector) ...@@ -115,7 +115,7 @@ unsigned long segment_base(u16 selector)
} }
d = (struct segment_descriptor *)(table_base + (selector & ~7)); d = (struct segment_descriptor *)(table_base + (selector & ~7));
v = d->base_low | ((ul)d->base_mid << 16) | ((ul)d->base_high << 24); v = d->base_low | ((ul)d->base_mid << 16) | ((ul)d->base_high << 24);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
if (d->system == 0 if (d->system == 0
&& (d->type == 2 || d->type == 9 || d->type == 11)) && (d->type == 2 || d->type == 9 || d->type == 11))
v |= ((ul)((struct segment_descriptor_64 *)d)->base_higher) << 32; v |= ((ul)((struct segment_descriptor_64 *)d)->base_higher) << 32;
...@@ -351,7 +351,7 @@ void set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) ...@@ -351,7 +351,7 @@ void set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
} }
if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK)) { if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK)) {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
if ((vcpu->shadow_efer & EFER_LME)) { if ((vcpu->shadow_efer & EFER_LME)) {
int cs_db, cs_l; int cs_db, cs_l;
...@@ -1120,7 +1120,7 @@ static int get_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata) ...@@ -1120,7 +1120,7 @@ static int get_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata)
return kvm_arch_ops->get_msr(vcpu, msr_index, pdata); return kvm_arch_ops->get_msr(vcpu, msr_index, pdata);
} }
#ifdef __x86_64__ #ifdef CONFIG_X86_64
void set_efer(struct kvm_vcpu *vcpu, u64 efer) void set_efer(struct kvm_vcpu *vcpu, u64 efer)
{ {
...@@ -1243,7 +1243,7 @@ static int kvm_dev_ioctl_get_regs(struct kvm *kvm, struct kvm_regs *regs) ...@@ -1243,7 +1243,7 @@ static int kvm_dev_ioctl_get_regs(struct kvm *kvm, struct kvm_regs *regs)
regs->rdi = vcpu->regs[VCPU_REGS_RDI]; regs->rdi = vcpu->regs[VCPU_REGS_RDI];
regs->rsp = vcpu->regs[VCPU_REGS_RSP]; regs->rsp = vcpu->regs[VCPU_REGS_RSP];
regs->rbp = vcpu->regs[VCPU_REGS_RBP]; regs->rbp = vcpu->regs[VCPU_REGS_RBP];
#ifdef __x86_64__ #ifdef CONFIG_X86_64
regs->r8 = vcpu->regs[VCPU_REGS_R8]; regs->r8 = vcpu->regs[VCPU_REGS_R8];
regs->r9 = vcpu->regs[VCPU_REGS_R9]; regs->r9 = vcpu->regs[VCPU_REGS_R9];
regs->r10 = vcpu->regs[VCPU_REGS_R10]; regs->r10 = vcpu->regs[VCPU_REGS_R10];
...@@ -1287,7 +1287,7 @@ static int kvm_dev_ioctl_set_regs(struct kvm *kvm, struct kvm_regs *regs) ...@@ -1287,7 +1287,7 @@ static int kvm_dev_ioctl_set_regs(struct kvm *kvm, struct kvm_regs *regs)
vcpu->regs[VCPU_REGS_RDI] = regs->rdi; vcpu->regs[VCPU_REGS_RDI] = regs->rdi;
vcpu->regs[VCPU_REGS_RSP] = regs->rsp; vcpu->regs[VCPU_REGS_RSP] = regs->rsp;
vcpu->regs[VCPU_REGS_RBP] = regs->rbp; vcpu->regs[VCPU_REGS_RBP] = regs->rbp;
#ifdef __x86_64__ #ifdef CONFIG_X86_64
vcpu->regs[VCPU_REGS_R8] = regs->r8; vcpu->regs[VCPU_REGS_R8] = regs->r8;
vcpu->regs[VCPU_REGS_R9] = regs->r9; vcpu->regs[VCPU_REGS_R9] = regs->r9;
vcpu->regs[VCPU_REGS_R10] = regs->r10; vcpu->regs[VCPU_REGS_R10] = regs->r10;
...@@ -1401,7 +1401,7 @@ static int kvm_dev_ioctl_set_sregs(struct kvm *kvm, struct kvm_sregs *sregs) ...@@ -1401,7 +1401,7 @@ static int kvm_dev_ioctl_set_sregs(struct kvm *kvm, struct kvm_sregs *sregs)
vcpu->cr8 = sregs->cr8; vcpu->cr8 = sregs->cr8;
mmu_reset_needed |= vcpu->shadow_efer != sregs->efer; mmu_reset_needed |= vcpu->shadow_efer != sregs->efer;
#ifdef __x86_64__ #ifdef CONFIG_X86_64
kvm_arch_ops->set_efer(vcpu, sregs->efer); kvm_arch_ops->set_efer(vcpu, sregs->efer);
#endif #endif
vcpu->apic_base = sregs->apic_base; vcpu->apic_base = sregs->apic_base;
...@@ -1434,7 +1434,7 @@ static int kvm_dev_ioctl_set_sregs(struct kvm *kvm, struct kvm_sregs *sregs) ...@@ -1434,7 +1434,7 @@ static int kvm_dev_ioctl_set_sregs(struct kvm *kvm, struct kvm_sregs *sregs)
static u32 msrs_to_save[] = { static u32 msrs_to_save[] = {
MSR_IA32_SYSENTER_CS, MSR_IA32_SYSENTER_ESP, MSR_IA32_SYSENTER_EIP, MSR_IA32_SYSENTER_CS, MSR_IA32_SYSENTER_ESP, MSR_IA32_SYSENTER_EIP,
MSR_K6_STAR, MSR_K6_STAR,
#ifdef __x86_64__ #ifdef CONFIG_X86_64
MSR_CSTAR, MSR_KERNEL_GS_BASE, MSR_SYSCALL_MASK, MSR_LSTAR, MSR_CSTAR, MSR_KERNEL_GS_BASE, MSR_SYSCALL_MASK, MSR_LSTAR,
#endif #endif
MSR_IA32_TIME_STAMP_COUNTER, MSR_IA32_TIME_STAMP_COUNTER,
......
...@@ -9,7 +9,7 @@ ...@@ -9,7 +9,7 @@
#include "kvm.h" #include "kvm.h"
static const u32 host_save_msrs[] = { static const u32 host_save_msrs[] = {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
MSR_STAR, MSR_LSTAR, MSR_CSTAR, MSR_SYSCALL_MASK, MSR_KERNEL_GS_BASE, MSR_STAR, MSR_LSTAR, MSR_CSTAR, MSR_SYSCALL_MASK, MSR_KERNEL_GS_BASE,
MSR_FS_BASE, MSR_GS_BASE, MSR_FS_BASE, MSR_GS_BASE,
#endif #endif
......
#ifndef __KVM_VMX_H #ifndef __KVM_VMX_H
#define __KVM_VMX_H #define __KVM_VMX_H
#ifdef __x86_64__ #ifdef CONFIG_X86_64
/* /*
* avoid save/load MSR_SYSCALL_MASK and MSR_LSTAR by std vt * avoid save/load MSR_SYSCALL_MASK and MSR_LSTAR by std vt
* mechanism (cpu bug AA24) * mechanism (cpu bug AA24)
......
...@@ -287,7 +287,7 @@ static void svm_hardware_enable(void *garbage) ...@@ -287,7 +287,7 @@ static void svm_hardware_enable(void *garbage)
struct svm_cpu_data *svm_data; struct svm_cpu_data *svm_data;
uint64_t efer; uint64_t efer;
#ifdef __x86_64__ #ifdef CONFIG_X86_64
struct desc_ptr gdt_descr; struct desc_ptr gdt_descr;
#else #else
struct Xgt_desc_struct gdt_descr; struct Xgt_desc_struct gdt_descr;
...@@ -397,7 +397,7 @@ static __init int svm_hardware_setup(void) ...@@ -397,7 +397,7 @@ static __init int svm_hardware_setup(void)
memset(msrpm_va, 0xff, PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER)); memset(msrpm_va, 0xff, PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER));
msrpm_base = page_to_pfn(msrpm_pages) << PAGE_SHIFT; msrpm_base = page_to_pfn(msrpm_pages) << PAGE_SHIFT;
#ifdef __x86_64__ #ifdef CONFIG_X86_64
set_msr_interception(msrpm_va, MSR_GS_BASE, 1, 1); set_msr_interception(msrpm_va, MSR_GS_BASE, 1, 1);
set_msr_interception(msrpm_va, MSR_FS_BASE, 1, 1); set_msr_interception(msrpm_va, MSR_FS_BASE, 1, 1);
set_msr_interception(msrpm_va, MSR_KERNEL_GS_BASE, 1, 1); set_msr_interception(msrpm_va, MSR_KERNEL_GS_BASE, 1, 1);
...@@ -704,7 +704,7 @@ static void svm_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt) ...@@ -704,7 +704,7 @@ static void svm_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) static void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
{ {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
if (vcpu->shadow_efer & KVM_EFER_LME) { if (vcpu->shadow_efer & KVM_EFER_LME) {
if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK)) { if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK)) {
vcpu->shadow_efer |= KVM_EFER_LMA; vcpu->shadow_efer |= KVM_EFER_LMA;
...@@ -1097,7 +1097,7 @@ static int svm_get_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 *data) ...@@ -1097,7 +1097,7 @@ static int svm_get_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 *data)
case MSR_IA32_APICBASE: case MSR_IA32_APICBASE:
*data = vcpu->apic_base; *data = vcpu->apic_base;
break; break;
#ifdef __x86_64__ #ifdef CONFIG_X86_64
case MSR_STAR: case MSR_STAR:
*data = vcpu->svm->vmcb->save.star; *data = vcpu->svm->vmcb->save.star;
break; break;
...@@ -1149,7 +1149,7 @@ static int rdmsr_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1149,7 +1149,7 @@ static int rdmsr_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data) static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data)
{ {
switch (ecx) { switch (ecx) {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
case MSR_EFER: case MSR_EFER:
set_efer(vcpu, data); set_efer(vcpu, data);
break; break;
...@@ -1172,7 +1172,7 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data) ...@@ -1172,7 +1172,7 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data)
case MSR_IA32_APICBASE: case MSR_IA32_APICBASE:
vcpu->apic_base = data; vcpu->apic_base = data;
break; break;
#ifdef __x86_64___ #ifdef CONFIG_X86_64_
case MSR_STAR: case MSR_STAR:
vcpu->svm->vmcb->save.star = data; vcpu->svm->vmcb->save.star = data;
break; break;
...@@ -1387,7 +1387,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1387,7 +1387,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
load_db_regs(vcpu->svm->db_regs); load_db_regs(vcpu->svm->db_regs);
} }
asm volatile ( asm volatile (
#ifdef __x86_64__ #ifdef CONFIG_X86_64
"push %%rbx; push %%rcx; push %%rdx;" "push %%rbx; push %%rcx; push %%rdx;"
"push %%rsi; push %%rdi; push %%rbp;" "push %%rsi; push %%rdi; push %%rbp;"
"push %%r8; push %%r9; push %%r10; push %%r11;" "push %%r8; push %%r9; push %%r10; push %%r11;"
...@@ -1397,7 +1397,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1397,7 +1397,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
"push %%esi; push %%edi; push %%ebp;" "push %%esi; push %%edi; push %%ebp;"
#endif #endif
#ifdef __x86_64__ #ifdef CONFIG_X86_64
"mov %c[rbx](%[vcpu]), %%rbx \n\t" "mov %c[rbx](%[vcpu]), %%rbx \n\t"
"mov %c[rcx](%[vcpu]), %%rcx \n\t" "mov %c[rcx](%[vcpu]), %%rcx \n\t"
"mov %c[rdx](%[vcpu]), %%rdx \n\t" "mov %c[rdx](%[vcpu]), %%rdx \n\t"
...@@ -1421,7 +1421,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1421,7 +1421,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
"mov %c[rbp](%[vcpu]), %%ebp \n\t" "mov %c[rbp](%[vcpu]), %%ebp \n\t"
#endif #endif
#ifdef __x86_64__ #ifdef CONFIG_X86_64
/* Enter guest mode */ /* Enter guest mode */
"push %%rax \n\t" "push %%rax \n\t"
"mov %c[svm](%[vcpu]), %%rax \n\t" "mov %c[svm](%[vcpu]), %%rax \n\t"
...@@ -1442,7 +1442,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1442,7 +1442,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
#endif #endif
/* Save guest registers, load host registers */ /* Save guest registers, load host registers */
#ifdef __x86_64__ #ifdef CONFIG_X86_64
"mov %%rbx, %c[rbx](%[vcpu]) \n\t" "mov %%rbx, %c[rbx](%[vcpu]) \n\t"
"mov %%rcx, %c[rcx](%[vcpu]) \n\t" "mov %%rcx, %c[rcx](%[vcpu]) \n\t"
"mov %%rdx, %c[rdx](%[vcpu]) \n\t" "mov %%rdx, %c[rdx](%[vcpu]) \n\t"
...@@ -1483,7 +1483,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1483,7 +1483,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
[rsi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RSI])), [rsi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RSI])),
[rdi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RDI])), [rdi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RDI])),
[rbp]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RBP])) [rbp]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RBP]))
#ifdef __x86_64__ #ifdef CONFIG_X86_64
,[r8 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R8 ])), ,[r8 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R8 ])),
[r9 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R9 ])), [r9 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R9 ])),
[r10]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R10])), [r10]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R10])),
......
...@@ -34,7 +34,7 @@ MODULE_LICENSE("GPL"); ...@@ -34,7 +34,7 @@ MODULE_LICENSE("GPL");
static DEFINE_PER_CPU(struct vmcs *, vmxarea); static DEFINE_PER_CPU(struct vmcs *, vmxarea);
static DEFINE_PER_CPU(struct vmcs *, current_vmcs); static DEFINE_PER_CPU(struct vmcs *, current_vmcs);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
#define HOST_IS_64 1 #define HOST_IS_64 1
#else #else
#define HOST_IS_64 0 #define HOST_IS_64 0
...@@ -71,7 +71,7 @@ static struct kvm_vmx_segment_field { ...@@ -71,7 +71,7 @@ static struct kvm_vmx_segment_field {
}; };
static const u32 vmx_msr_index[] = { static const u32 vmx_msr_index[] = {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
MSR_SYSCALL_MASK, MSR_LSTAR, MSR_CSTAR, MSR_KERNEL_GS_BASE, MSR_SYSCALL_MASK, MSR_LSTAR, MSR_CSTAR, MSR_KERNEL_GS_BASE,
#endif #endif
MSR_EFER, MSR_K6_STAR, MSR_EFER, MSR_K6_STAR,
...@@ -138,7 +138,7 @@ static u32 vmcs_read32(unsigned long field) ...@@ -138,7 +138,7 @@ static u32 vmcs_read32(unsigned long field)
static u64 vmcs_read64(unsigned long field) static u64 vmcs_read64(unsigned long field)
{ {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
return vmcs_readl(field); return vmcs_readl(field);
#else #else
return vmcs_readl(field) | ((u64)vmcs_readl(field+1) << 32); return vmcs_readl(field) | ((u64)vmcs_readl(field+1) << 32);
...@@ -168,7 +168,7 @@ static void vmcs_write32(unsigned long field, u32 value) ...@@ -168,7 +168,7 @@ static void vmcs_write32(unsigned long field, u32 value)
static void vmcs_write64(unsigned long field, u64 value) static void vmcs_write64(unsigned long field, u64 value)
{ {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
vmcs_writel(field, value); vmcs_writel(field, value);
#else #else
vmcs_writel(field, value); vmcs_writel(field, value);
...@@ -297,7 +297,7 @@ static void guest_write_tsc(u64 guest_tsc) ...@@ -297,7 +297,7 @@ static void guest_write_tsc(u64 guest_tsc)
static void reload_tss(void) static void reload_tss(void)
{ {
#ifndef __x86_64__ #ifndef CONFIG_X86_64
/* /*
* VT restores TR but not its size. Useless. * VT restores TR but not its size. Useless.
...@@ -328,7 +328,7 @@ static int vmx_get_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata) ...@@ -328,7 +328,7 @@ static int vmx_get_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata)
} }
switch (msr_index) { switch (msr_index) {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
case MSR_FS_BASE: case MSR_FS_BASE:
data = vmcs_readl(GUEST_FS_BASE); data = vmcs_readl(GUEST_FS_BASE);
break; break;
...@@ -391,7 +391,7 @@ static int vmx_set_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data) ...@@ -391,7 +391,7 @@ static int vmx_set_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data)
{ {
struct vmx_msr_entry *msr; struct vmx_msr_entry *msr;
switch (msr_index) { switch (msr_index) {
#ifdef __x86_64__ #ifdef CONFIG_X86_64
case MSR_FS_BASE: case MSR_FS_BASE:
vmcs_writel(GUEST_FS_BASE, data); vmcs_writel(GUEST_FS_BASE, data);
break; break;
...@@ -726,7 +726,7 @@ static void enter_rmode(struct kvm_vcpu *vcpu) ...@@ -726,7 +726,7 @@ static void enter_rmode(struct kvm_vcpu *vcpu)
fix_rmode_seg(VCPU_SREG_FS, &vcpu->rmode.fs); fix_rmode_seg(VCPU_SREG_FS, &vcpu->rmode.fs);
} }
#ifdef __x86_64__ #ifdef CONFIG_X86_64
static void enter_lmode(struct kvm_vcpu *vcpu) static void enter_lmode(struct kvm_vcpu *vcpu)
{ {
...@@ -768,7 +768,7 @@ static void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) ...@@ -768,7 +768,7 @@ static void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
if (!vcpu->rmode.active && !(cr0 & CR0_PE_MASK)) if (!vcpu->rmode.active && !(cr0 & CR0_PE_MASK))
enter_rmode(vcpu); enter_rmode(vcpu);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
if (vcpu->shadow_efer & EFER_LME) { if (vcpu->shadow_efer & EFER_LME) {
if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK)) if (!is_paging(vcpu) && (cr0 & CR0_PG_MASK))
enter_lmode(vcpu); enter_lmode(vcpu);
...@@ -809,7 +809,7 @@ static void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) ...@@ -809,7 +809,7 @@ static void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
vcpu->cr4 = cr4; vcpu->cr4 = cr4;
} }
#ifdef __x86_64__ #ifdef CONFIG_X86_64
static void vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) static void vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer)
{ {
...@@ -1096,7 +1096,7 @@ static int vmx_vcpu_setup(struct kvm_vcpu *vcpu) ...@@ -1096,7 +1096,7 @@ static int vmx_vcpu_setup(struct kvm_vcpu *vcpu)
vmcs_write16(HOST_FS_SELECTOR, read_fs()); /* 22.2.4 */ vmcs_write16(HOST_FS_SELECTOR, read_fs()); /* 22.2.4 */
vmcs_write16(HOST_GS_SELECTOR, read_gs()); /* 22.2.4 */ vmcs_write16(HOST_GS_SELECTOR, read_gs()); /* 22.2.4 */
vmcs_write16(HOST_SS_SELECTOR, __KERNEL_DS); /* 22.2.4 */ vmcs_write16(HOST_SS_SELECTOR, __KERNEL_DS); /* 22.2.4 */
#ifdef __x86_64__ #ifdef CONFIG_X86_64
rdmsrl(MSR_FS_BASE, a); rdmsrl(MSR_FS_BASE, a);
vmcs_writel(HOST_FS_BASE, a); /* 22.2.4 */ vmcs_writel(HOST_FS_BASE, a); /* 22.2.4 */
rdmsrl(MSR_GS_BASE, a); rdmsrl(MSR_GS_BASE, a);
...@@ -1174,7 +1174,7 @@ static int vmx_vcpu_setup(struct kvm_vcpu *vcpu) ...@@ -1174,7 +1174,7 @@ static int vmx_vcpu_setup(struct kvm_vcpu *vcpu)
vcpu->cr0 = 0x60000010; vcpu->cr0 = 0x60000010;
vmx_set_cr0(vcpu, vcpu->cr0); // enter rmode vmx_set_cr0(vcpu, vcpu->cr0); // enter rmode
vmx_set_cr4(vcpu, 0); vmx_set_cr4(vcpu, 0);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
vmx_set_efer(vcpu, 0); vmx_set_efer(vcpu, 0);
#endif #endif
...@@ -1690,7 +1690,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1690,7 +1690,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
vmcs_write16(HOST_GS_SELECTOR, 0); vmcs_write16(HOST_GS_SELECTOR, 0);
} }
#ifdef __x86_64__ #ifdef CONFIG_X86_64
vmcs_writel(HOST_FS_BASE, read_msr(MSR_FS_BASE)); vmcs_writel(HOST_FS_BASE, read_msr(MSR_FS_BASE));
vmcs_writel(HOST_GS_BASE, read_msr(MSR_GS_BASE)); vmcs_writel(HOST_GS_BASE, read_msr(MSR_GS_BASE));
#else #else
...@@ -1714,7 +1714,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1714,7 +1714,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
asm ( asm (
/* Store host registers */ /* Store host registers */
"pushf \n\t" "pushf \n\t"
#ifdef __x86_64__ #ifdef CONFIG_X86_64
"push %%rax; push %%rbx; push %%rdx;" "push %%rax; push %%rbx; push %%rdx;"
"push %%rsi; push %%rdi; push %%rbp;" "push %%rsi; push %%rdi; push %%rbp;"
"push %%r8; push %%r9; push %%r10; push %%r11;" "push %%r8; push %%r9; push %%r10; push %%r11;"
...@@ -1728,7 +1728,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1728,7 +1728,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
/* Check if vmlaunch of vmresume is needed */ /* Check if vmlaunch of vmresume is needed */
"cmp $0, %1 \n\t" "cmp $0, %1 \n\t"
/* Load guest registers. Don't clobber flags. */ /* Load guest registers. Don't clobber flags. */
#ifdef __x86_64__ #ifdef CONFIG_X86_64
"mov %c[cr2](%3), %%rax \n\t" "mov %c[cr2](%3), %%rax \n\t"
"mov %%rax, %%cr2 \n\t" "mov %%rax, %%cr2 \n\t"
"mov %c[rax](%3), %%rax \n\t" "mov %c[rax](%3), %%rax \n\t"
...@@ -1765,7 +1765,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1765,7 +1765,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
".globl kvm_vmx_return \n\t" ".globl kvm_vmx_return \n\t"
"kvm_vmx_return: " "kvm_vmx_return: "
/* Save guest registers, load host registers, keep flags */ /* Save guest registers, load host registers, keep flags */
#ifdef __x86_64__ #ifdef CONFIG_X86_64
"xchg %3, 0(%%rsp) \n\t" "xchg %3, 0(%%rsp) \n\t"
"mov %%rax, %c[rax](%3) \n\t" "mov %%rax, %c[rax](%3) \n\t"
"mov %%rbx, %c[rbx](%3) \n\t" "mov %%rbx, %c[rbx](%3) \n\t"
...@@ -1817,7 +1817,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1817,7 +1817,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
[rsi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RSI])), [rsi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RSI])),
[rdi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RDI])), [rdi]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RDI])),
[rbp]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RBP])), [rbp]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_RBP])),
#ifdef __x86_64__ #ifdef CONFIG_X86_64
[r8 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R8 ])), [r8 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R8 ])),
[r9 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R9 ])), [r9 ]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R9 ])),
[r10]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R10])), [r10]"i"(offsetof(struct kvm_vcpu, regs[VCPU_REGS_R10])),
...@@ -1838,7 +1838,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1838,7 +1838,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
fx_save(vcpu->guest_fx_image); fx_save(vcpu->guest_fx_image);
fx_restore(vcpu->host_fx_image); fx_restore(vcpu->host_fx_image);
#ifndef __x86_64__ #ifndef CONFIG_X86_64
asm ("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS)); asm ("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS));
#endif #endif
...@@ -1856,7 +1856,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) ...@@ -1856,7 +1856,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
*/ */
local_irq_disable(); local_irq_disable();
load_gs(gs_sel); load_gs(gs_sel);
#ifdef __x86_64__ #ifdef CONFIG_X86_64
wrmsrl(MSR_GS_BASE, vmcs_readl(HOST_GS_BASE)); wrmsrl(MSR_GS_BASE, vmcs_readl(HOST_GS_BASE));
#endif #endif
local_irq_enable(); local_irq_enable();
...@@ -1966,7 +1966,7 @@ static struct kvm_arch_ops vmx_arch_ops = { ...@@ -1966,7 +1966,7 @@ static struct kvm_arch_ops vmx_arch_ops = {
.set_cr0_no_modeswitch = vmx_set_cr0_no_modeswitch, .set_cr0_no_modeswitch = vmx_set_cr0_no_modeswitch,
.set_cr3 = vmx_set_cr3, .set_cr3 = vmx_set_cr3,
.set_cr4 = vmx_set_cr4, .set_cr4 = vmx_set_cr4,
#ifdef __x86_64__ #ifdef CONFIG_X86_64
.set_efer = vmx_set_efer, .set_efer = vmx_set_efer,
#endif #endif
.get_idt = vmx_get_idt, .get_idt = vmx_get_idt,
......
...@@ -238,7 +238,7 @@ struct operand { ...@@ -238,7 +238,7 @@ struct operand {
* any modified flags. * any modified flags.
*/ */
#if defined(__x86_64__) #if defined(CONFIG_X86_64)
#define _LO32 "k" /* force 32-bit operand */ #define _LO32 "k" /* force 32-bit operand */
#define _STK "%%rsp" /* stack pointer */ #define _STK "%%rsp" /* stack pointer */
#elif defined(__i386__) #elif defined(__i386__)
...@@ -385,7 +385,7 @@ struct operand { ...@@ -385,7 +385,7 @@ struct operand {
} while (0) } while (0)
/* Emulate an instruction with quadword operands (x86/64 only). */ /* Emulate an instruction with quadword operands (x86/64 only). */
#if defined(__x86_64__) #if defined(CONFIG_X86_64)
#define __emulate_2op_8byte(_op, _src, _dst, _eflags, _qx, _qy) \ #define __emulate_2op_8byte(_op, _src, _dst, _eflags, _qx, _qy) \
do { \ do { \
__asm__ __volatile__ ( \ __asm__ __volatile__ ( \
...@@ -495,7 +495,7 @@ x86_emulate_memop(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops) ...@@ -495,7 +495,7 @@ x86_emulate_memop(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
case X86EMUL_MODE_PROT32: case X86EMUL_MODE_PROT32:
op_bytes = ad_bytes = 4; op_bytes = ad_bytes = 4;
break; break;
#ifdef __x86_64__ #ifdef CONFIG_X86_64
case X86EMUL_MODE_PROT64: case X86EMUL_MODE_PROT64:
op_bytes = 4; op_bytes = 4;
ad_bytes = 8; ad_bytes = 8;
...@@ -1341,7 +1341,7 @@ x86_emulate_memop(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops) ...@@ -1341,7 +1341,7 @@ x86_emulate_memop(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
} }
break; break;
} }
#elif defined(__x86_64__) #elif defined(CONFIG_X86_64)
{ {
unsigned long old, new; unsigned long old, new;
if ((rc = ops->read_emulated(cr2, &old, 8, ctxt)) != 0) if ((rc = ops->read_emulated(cr2, &old, 8, ctxt)) != 0)
......
...@@ -162,7 +162,7 @@ struct x86_emulate_ctxt { ...@@ -162,7 +162,7 @@ struct x86_emulate_ctxt {
/* Host execution mode. */ /* Host execution mode. */
#if defined(__i386__) #if defined(__i386__)
#define X86EMUL_MODE_HOST X86EMUL_MODE_PROT32 #define X86EMUL_MODE_HOST X86EMUL_MODE_PROT32
#elif defined(__x86_64__) #elif defined(CONFIG_X86_64)
#define X86EMUL_MODE_HOST X86EMUL_MODE_PROT64 #define X86EMUL_MODE_HOST X86EMUL_MODE_PROT64
#endif #endif
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册