kvm_emulate.h 8.6 KB
Newer Older
A
Avi Kivity 已提交
1 2 3 4 5 6 7 8 9 10
/******************************************************************************
 * x86_emulate.h
 *
 * Generic x86 (32-bit and 64-bit) instruction decoder and emulator.
 *
 * Copyright (c) 2005 Keir Fraser
 *
 * From: xen-unstable 10676:af9809f51f81a3c43f276f00c81a52ef558afda4
 */

H
H. Peter Anvin 已提交
11 12
#ifndef _ASM_X86_KVM_X86_EMULATE_H
#define _ASM_X86_KVM_X86_EMULATE_H
A
Avi Kivity 已提交
13

14 15
#include <asm/desc_defs.h>

A
Avi Kivity 已提交
16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53
struct x86_emulate_ctxt;

/*
 * x86_emulate_ops:
 *
 * These operations represent the instruction emulator's interface to memory.
 * There are two categories of operation: those that act on ordinary memory
 * regions (*_std), and those that act on memory regions known to require
 * special treatment or emulation (*_emulated).
 *
 * The emulator assumes that an instruction accesses only one 'emulated memory'
 * location, that this location is the given linear faulting address (cr2), and
 * that this is one of the instruction's data operands. Instruction fetches and
 * stack operations are assumed never to access emulated memory. The emulator
 * automatically deduces which operand of a string-move operation is accessing
 * emulated memory, and assumes that the other operand accesses normal memory.
 *
 * NOTES:
 *  1. The emulator isn't very smart about emulated vs. standard memory.
 *     'Emulated memory' access addresses should be checked for sanity.
 *     'Normal memory' accesses may fault, and the caller must arrange to
 *     detect and handle reentrancy into the emulator via recursive faults.
 *     Accesses may be unaligned and may cross page boundaries.
 *  2. If the access fails (cannot emulate, or a standard access faults) then
 *     it is up to the memop to propagate the fault to the guest VM via
 *     some out-of-band mechanism, unknown to the emulator. The memop signals
 *     failure by returning X86EMUL_PROPAGATE_FAULT to the emulator, which will
 *     then immediately bail.
 *  3. Valid access sizes are 1, 2, 4 and 8 bytes. On x86/32 systems only
 *     cmpxchg8b_emulated need support 8-byte accesses.
 *  4. The emulator cannot handle 64-bit mode emulation on an x86/32 system.
 */
/* Access completed successfully: continue emulation as normal. */
#define X86EMUL_CONTINUE        0
/* Access is unhandleable: bail from emulation and return error to caller. */
#define X86EMUL_UNHANDLEABLE    1
/* Terminate emulation but return success to the caller. */
#define X86EMUL_PROPAGATE_FAULT 2 /* propagate a generated fault to guest */
54 55 56
#define X86EMUL_RETRY_INSTR     3 /* retry the instruction for some reason */
#define X86EMUL_CMPXCHG_FAILED  4 /* cmpxchg did not see expected value */

A
Avi Kivity 已提交
57 58 59
struct x86_emulate_ops {
	/*
	 * read_std: Read bytes of standard (non-emulated/special) memory.
60
	 *           Used for descriptor reading.
A
Avi Kivity 已提交
61 62 63 64
	 *  @addr:  [IN ] Linear address from which to read.
	 *  @val:   [OUT] Value read from memory, zero-extended to 'u_long'.
	 *  @bytes: [IN ] Number of bytes to read from memory.
	 */
65
	int (*read_std)(unsigned long addr, void *val,
66 67
			unsigned int bytes, struct kvm_vcpu *vcpu, u32 *error);

68 69 70 71 72 73 74 75 76
	/*
	 * write_std: Write bytes of standard (non-emulated/special) memory.
	 *            Used for descriptor writing.
	 *  @addr:  [IN ] Linear address to which to write.
	 *  @val:   [OUT] Value write to memory, zero-extended to 'u_long'.
	 *  @bytes: [IN ] Number of bytes to write to memory.
	 */
	int (*write_std)(unsigned long addr, void *val,
			 unsigned int bytes, struct kvm_vcpu *vcpu, u32 *error);
77 78 79 80 81 82 83 84 85
	/*
	 * fetch: Read bytes of standard (non-emulated/special) memory.
	 *        Used for instruction fetch.
	 *  @addr:  [IN ] Linear address from which to read.
	 *  @val:   [OUT] Value read from memory, zero-extended to 'u_long'.
	 *  @bytes: [IN ] Number of bytes to read from memory.
	 */
	int (*fetch)(unsigned long addr, void *val,
			unsigned int bytes, struct kvm_vcpu *vcpu, u32 *error);
A
Avi Kivity 已提交
86 87 88 89 90 91 92

	/*
	 * read_emulated: Read bytes from emulated/special memory area.
	 *  @addr:  [IN ] Linear address from which to read.
	 *  @val:   [OUT] Value read from memory, zero-extended to 'u_long'.
	 *  @bytes: [IN ] Number of bytes to read from memory.
	 */
93 94 95 96
	int (*read_emulated)(unsigned long addr,
			     void *val,
			     unsigned int bytes,
			     struct kvm_vcpu *vcpu);
A
Avi Kivity 已提交
97 98

	/*
99
	 * write_emulated: Write bytes to emulated/special memory area.
A
Avi Kivity 已提交
100 101 102 103 104
	 *  @addr:  [IN ] Linear address to which to write.
	 *  @val:   [IN ] Value to write to memory (low-order bytes used as
	 *                required).
	 *  @bytes: [IN ] Number of bytes to write to memory.
	 */
105 106 107 108
	int (*write_emulated)(unsigned long addr,
			      const void *val,
			      unsigned int bytes,
			      struct kvm_vcpu *vcpu);
A
Avi Kivity 已提交
109 110 111 112 113 114 115 116 117

	/*
	 * cmpxchg_emulated: Emulate an atomic (LOCKed) CMPXCHG operation on an
	 *                   emulated/special memory area.
	 *  @addr:  [IN ] Linear address to access.
	 *  @old:   [IN ] Value expected to be current at @addr.
	 *  @new:   [IN ] Value to write to @addr.
	 *  @bytes: [IN ] Number of bytes to access using CMPXCHG.
	 */
118 119 120 121 122
	int (*cmpxchg_emulated)(unsigned long addr,
				const void *old,
				const void *new,
				unsigned int bytes,
				struct kvm_vcpu *vcpu);
123 124 125 126 127 128 129

	int (*pio_in_emulated)(int size, unsigned short port, void *val,
			       unsigned int count, struct kvm_vcpu *vcpu);

	int (*pio_out_emulated)(int size, unsigned short port, const void *val,
				unsigned int count, struct kvm_vcpu *vcpu);

130 131 132 133 134 135
	bool (*get_cached_descriptor)(struct desc_struct *desc,
				      int seg, struct kvm_vcpu *vcpu);
	void (*set_cached_descriptor)(struct desc_struct *desc,
				      int seg, struct kvm_vcpu *vcpu);
	u16 (*get_segment_selector)(int seg, struct kvm_vcpu *vcpu);
	void (*set_segment_selector)(u16 sel, int seg, struct kvm_vcpu *vcpu);
136
	unsigned long (*get_cached_segment_base)(int seg, struct kvm_vcpu *vcpu);
137
	void (*get_gdt)(struct desc_ptr *dt, struct kvm_vcpu *vcpu);
138
	ulong (*get_cr)(int cr, struct kvm_vcpu *vcpu);
139
	int (*set_cr)(int cr, ulong val, struct kvm_vcpu *vcpu);
140
	int (*cpl)(struct kvm_vcpu *vcpu);
141
	void (*set_rflags)(struct kvm_vcpu *vcpu, unsigned long rflags);
142 143
	int (*get_dr)(int dr, unsigned long *dest, struct kvm_vcpu *vcpu);
	int (*set_dr)(int dr, unsigned long value, struct kvm_vcpu *vcpu);
144 145
	int (*set_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 data);
	int (*get_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata);
A
Avi Kivity 已提交
146 147
};

148 149
/* Type, address-of, and value of an instruction's operand. */
struct operand {
150
	enum { OP_REG, OP_MEM, OP_IMM, OP_NONE } type;
151
	unsigned int bytes;
152 153 154 155 156
	unsigned long orig_val, *ptr;
	union {
		unsigned long val;
		char valptr[sizeof(unsigned long) + 2];
	};
157 158
};

159 160 161 162 163 164
struct fetch_cache {
	u8 data[15];
	unsigned long start;
	unsigned long end;
};

165 166 167 168 169 170
struct read_cache {
	u8 data[1024];
	unsigned long pos;
	unsigned long end;
};

171 172 173 174 175 176 177
struct decode_cache {
	u8 twobyte;
	u8 b;
	u8 lock_prefix;
	u8 rep_prefix;
	u8 op_bytes;
	u8 ad_bytes;
178
	u8 rex_prefix;
179
	struct operand src;
180
	struct operand src2;
181
	struct operand dst;
182 183
	bool has_seg_override;
	u8 seg_override;
184 185
	unsigned int d;
	unsigned long regs[NR_VCPU_REGS];
186
	unsigned long eip;
187 188 189 190 191 192
	/* modrm */
	u8 modrm;
	u8 modrm_mod;
	u8 modrm_reg;
	u8 modrm_rm;
	u8 use_modrm_ea;
193
	bool rip_relative;
194
	unsigned long modrm_ea;
195
	void *modrm_ptr;
196
	unsigned long modrm_val;
197
	struct fetch_cache fetch;
198
	struct read_cache io_read;
199
	struct read_cache mem_read;
200 201
};

A
Avi Kivity 已提交
202 203 204 205 206
struct x86_emulate_ctxt {
	/* Register state before/after emulation. */
	struct kvm_vcpu *vcpu;

	unsigned long eflags;
207
	unsigned long eip; /* eip before instruction emulation */
A
Avi Kivity 已提交
208 209
	/* Emulated execution mode, represented by an X86EMUL_MODE value. */
	int mode;
210
	u32 cs_base;
211

212 213 214
	/* interruptibility state, as a result of execution of STI or MOV SS */
	int interruptibility;

215
	bool restart; /* restart string instruction after writeback */
216 217
	/* decode cache */
	struct decode_cache decode;
A
Avi Kivity 已提交
218 219
};

220
/* Repeat String Operation Prefix */
S
Sheng Yang 已提交
221 222
#define REPE_PREFIX	1
#define REPNE_PREFIX	2
223

A
Avi Kivity 已提交
224 225
/* Execution mode, passed to the emulator. */
#define X86EMUL_MODE_REAL     0	/* Real mode.             */
226
#define X86EMUL_MODE_VM86     1	/* Virtual 8086 mode.     */
A
Avi Kivity 已提交
227 228 229 230 231
#define X86EMUL_MODE_PROT16   2	/* 16-bit protected mode. */
#define X86EMUL_MODE_PROT32   4	/* 32-bit protected mode. */
#define X86EMUL_MODE_PROT64   8	/* 64-bit (long) mode.    */

/* Host execution mode. */
S
Sheng Yang 已提交
232
#if defined(CONFIG_X86_32)
A
Avi Kivity 已提交
233
#define X86EMUL_MODE_HOST X86EMUL_MODE_PROT32
234
#elif defined(CONFIG_X86_64)
A
Avi Kivity 已提交
235 236 237
#define X86EMUL_MODE_HOST X86EMUL_MODE_PROT64
#endif

238 239 240 241
int x86_decode_insn(struct x86_emulate_ctxt *ctxt,
		    struct x86_emulate_ops *ops);
int x86_emulate_insn(struct x86_emulate_ctxt *ctxt,
		     struct x86_emulate_ops *ops);
242 243
int emulator_task_switch(struct x86_emulate_ctxt *ctxt,
			 struct x86_emulate_ops *ops,
244 245
			 u16 tss_selector, int reason,
			 bool has_error_code, u32 error_code);
A
Avi Kivity 已提交
246

H
H. Peter Anvin 已提交
247
#endif /* _ASM_X86_KVM_X86_EMULATE_H */