paravirt.h 35.0 KB
Newer Older
1 2 3 4
#ifndef __ASM_PARAVIRT_H
#define __ASM_PARAVIRT_H
/* Various instructions on x86 need to be replaced for
 * para-virtualization: those hooks are defined here. */
5 6

#ifdef CONFIG_PARAVIRT
7
#include <asm/page.h>
8
#include <asm/asm.h>
9

10
/* Bitmask of what can be clobbered: usually at least eax. */
11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28
#define CLBR_NONE 0
#define CLBR_EAX  (1 << 0)
#define CLBR_ECX  (1 << 1)
#define CLBR_EDX  (1 << 2)

#ifdef CONFIG_X86_64
#define CLBR_RSI  (1 << 3)
#define CLBR_RDI  (1 << 4)
#define CLBR_R8   (1 << 5)
#define CLBR_R9   (1 << 6)
#define CLBR_R10  (1 << 7)
#define CLBR_R11  (1 << 8)
#define CLBR_ANY  ((1 << 9) - 1)
#include <asm/desc_defs.h>
#else
/* CLBR_ANY should match all regs platform has. For i386, that's just it */
#define CLBR_ANY  ((1 << 3) - 1)
#endif /* X86_64 */
29

30
#ifndef __ASSEMBLY__
31
#include <linux/types.h>
32
#include <linux/cpumask.h>
33
#include <asm/kmap_types.h>
34
#include <asm/desc_defs.h>
35

36
struct page;
37
struct thread_struct;
38
struct desc_ptr;
39
struct tss_struct;
40
struct mm_struct;
41
struct desc_struct;
42

43 44
/* general info */
struct pv_info {
45
	unsigned int kernel_rpl;
46
	int shared_kernel_pmd;
47
	int paravirt_enabled;
48
	const char *name;
49
};
50

51
struct pv_init_ops {
52
	/*
53 54 55 56 57 58
	 * Patch may replace one of the defined code sequences with
	 * arbitrary code, subject to the same register constraints.
	 * This generally means the code is not free to clobber any
	 * registers other than EAX.  The patch function should return
	 * the number of bytes of code generated, as we nop pad the
	 * rest in generic code.
59
	 */
60 61
	unsigned (*patch)(u8 type, u16 clobber, void *insnbuf,
			  unsigned long addr, unsigned len);
62

63
	/* Basic arch-specific setup */
64 65
	void (*arch_setup)(void);
	char *(*memory_setup)(void);
66 67
	void (*post_allocator_init)(void);

68
	/* Print a banner to identify the environment */
69
	void (*banner)(void);
70 71 72
};


73
struct pv_lazy_ops {
74
	/* Set deferred update mode, used for batching operations. */
75 76
	void (*enter)(void);
	void (*leave)(void);
77 78 79 80
};

struct pv_time_ops {
	void (*time_init)(void);
81

82
	/* Set and set time of day */
83 84 85
	unsigned long (*get_wallclock)(void);
	int (*set_wallclock)(unsigned long);

86 87 88
	unsigned long long (*sched_clock)(void);
	unsigned long (*get_cpu_khz)(void);
};
89

90
struct pv_cpu_ops {
91
	/* hooks for various privileged instructions */
92 93
	unsigned long (*get_debugreg)(int regno);
	void (*set_debugreg)(int regno, unsigned long value);
94

95
	void (*clts)(void);
96

97 98
	unsigned long (*read_cr0)(void);
	void (*write_cr0)(unsigned long);
99

100 101 102
	unsigned long (*read_cr4_safe)(void);
	unsigned long (*read_cr4)(void);
	void (*write_cr4)(unsigned long);
103

104
	/* Segment descriptor handling */
105
	void (*load_tr_desc)(void);
106 107 108 109
	void (*load_gdt)(const struct desc_ptr *);
	void (*load_idt)(const struct desc_ptr *);
	void (*store_gdt)(struct desc_ptr *);
	void (*store_idt)(struct desc_ptr *);
110 111 112
	void (*set_ldt)(const void *desc, unsigned entries);
	unsigned long (*store_tr)(void);
	void (*load_tls)(struct thread_struct *t, unsigned int cpu);
113 114
	void (*write_ldt_entry)(struct desc_struct *ldt, int entrynum,
				const void *desc);
115
	void (*write_gdt_entry)(struct desc_struct *,
116
				int entrynum, const void *desc, int size);
117 118
	void (*write_idt_entry)(gate_desc *,
				int entrynum, const gate_desc *gate);
119
	void (*load_sp0)(struct tss_struct *tss, struct thread_struct *t);
120

121
	void (*set_iopl_mask)(unsigned mask);
122 123

	void (*wbinvd)(void);
124
	void (*io_delay)(void);
125

126 127 128 129 130 131 132
	/* cpuid emulation, mostly so that caps bits can be disabled */
	void (*cpuid)(unsigned int *eax, unsigned int *ebx,
		      unsigned int *ecx, unsigned int *edx);

	/* MSR, PMC and TSR operations.
	   err = 0/-EFAULT.  wrmsr returns 0/-EFAULT. */
	u64 (*read_msr)(unsigned int msr, int *err);
133
	int (*write_msr)(unsigned int msr, unsigned low, unsigned high);
134 135

	u64 (*read_tsc)(void);
136
	u64 (*read_pmc)(int counter);
137
	unsigned long long (*read_tscp)(unsigned int *aux);
138 139

	/* These two are jmp to, not actually called. */
140
	void (*irq_enable_syscall_ret)(void);
141
	void (*iret)(void);
142

143 144
	void (*swapgs)(void);

145
	struct pv_lazy_ops lazy_mode;
146 147 148 149 150
};

struct pv_irq_ops {
	void (*init_IRQ)(void);

151
	/*
152 153 154 155
	 * Get/set interrupt state.  save_fl and restore_fl are only
	 * expected to use X86_EFLAGS_IF; all other bits
	 * returned from save_fl are undefined, and may be ignored by
	 * restore_fl.
156
	 */
157 158 159 160 161 162 163
	unsigned long (*save_fl)(void);
	void (*restore_fl)(unsigned long);
	void (*irq_disable)(void);
	void (*irq_enable)(void);
	void (*safe_halt)(void);
	void (*halt)(void);
};
164

165
struct pv_apic_ops {
166
#ifdef CONFIG_X86_LOCAL_APIC
167 168 169 170
	/*
	 * Direct APIC operations, principally for VMI.  Ideally
	 * these shouldn't be in this interface.
	 */
171 172 173
	void (*apic_write)(unsigned long reg, u32 v);
	void (*apic_write_atomic)(unsigned long reg, u32 v);
	u32 (*apic_read)(unsigned long reg);
Z
Zachary Amsden 已提交
174 175
	void (*setup_boot_clock)(void);
	void (*setup_secondary_clock)(void);
176 177 178 179

	void (*startup_ipi_hook)(int phys_apicid,
				 unsigned long start_eip,
				 unsigned long start_esp);
180
#endif
181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208
};

struct pv_mmu_ops {
	/*
	 * Called before/after init_mm pagetable setup. setup_start
	 * may reset %cr3, and may pre-install parts of the pagetable;
	 * pagetable setup is expected to preserve any existing
	 * mapping.
	 */
	void (*pagetable_setup_start)(pgd_t *pgd_base);
	void (*pagetable_setup_done)(pgd_t *pgd_base);

	unsigned long (*read_cr2)(void);
	void (*write_cr2)(unsigned long);

	unsigned long (*read_cr3)(void);
	void (*write_cr3)(unsigned long);

	/*
	 * Hooks for intercepting the creation/use/destruction of an
	 * mm_struct.
	 */
	void (*activate_mm)(struct mm_struct *prev,
			    struct mm_struct *next);
	void (*dup_mmap)(struct mm_struct *oldmm,
			 struct mm_struct *mm);
	void (*exit_mmap)(struct mm_struct *mm);

209

210
	/* TLB operations */
211 212
	void (*flush_tlb_user)(void);
	void (*flush_tlb_kernel)(void);
213
	void (*flush_tlb_single)(unsigned long addr);
214 215
	void (*flush_tlb_others)(const cpumask_t *cpus, struct mm_struct *mm,
				 unsigned long va);
216

217
	/* Hooks for allocating/releasing pagetable pages */
218
	void (*alloc_pt)(struct mm_struct *mm, u32 pfn);
219 220 221 222 223
	void (*alloc_pd)(u32 pfn);
	void (*alloc_pd_clone)(u32 pfn, u32 clonepfn, u32 start, u32 count);
	void (*release_pt)(u32 pfn);
	void (*release_pd)(u32 pfn);

224
	/* Pagetable manipulation functions */
225
	void (*set_pte)(pte_t *ptep, pte_t pteval);
226 227
	void (*set_pte_at)(struct mm_struct *mm, unsigned long addr,
			   pte_t *ptep, pte_t pteval);
228
	void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval);
229
	void (*pte_update)(struct mm_struct *mm, unsigned long addr, pte_t *ptep);
230 231
	void (*pte_update_defer)(struct mm_struct *mm,
				 unsigned long addr, pte_t *ptep);
232

233
#ifdef CONFIG_X86_PAE
234
	void (*set_pte_atomic)(pte_t *ptep, pte_t pteval);
235 236
	void (*set_pte_present)(struct mm_struct *mm, unsigned long addr,
				pte_t *ptep, pte_t pte);
237
	void (*set_pud)(pud_t *pudp, pud_t pudval);
238
	void (*pte_clear)(struct mm_struct *mm, unsigned long addr, pte_t *ptep);
239
	void (*pmd_clear)(pmd_t *pmdp);
240 241 242 243 244 245 246 247 248 249 250 251 252 253

	unsigned long long (*pte_val)(pte_t);
	unsigned long long (*pmd_val)(pmd_t);
	unsigned long long (*pgd_val)(pgd_t);

	pte_t (*make_pte)(unsigned long long pte);
	pmd_t (*make_pmd)(unsigned long long pmd);
	pgd_t (*make_pgd)(unsigned long long pgd);
#else
	unsigned long (*pte_val)(pte_t);
	unsigned long (*pgd_val)(pgd_t);

	pte_t (*make_pte)(unsigned long pte);
	pgd_t (*make_pgd)(unsigned long pgd);
254 255
#endif

256 257 258
#ifdef CONFIG_HIGHPTE
	void *(*kmap_atomic_pte)(struct page *page, enum km_type type);
#endif
259 260

	struct pv_lazy_ops lazy_mode;
261
};
262

263 264 265 266 267 268 269 270 271 272 273
/* This contains all the paravirt structures: we get a convenient
 * number for each function using the offset which we use to indicate
 * what to patch. */
struct paravirt_patch_template
{
	struct pv_init_ops pv_init_ops;
	struct pv_time_ops pv_time_ops;
	struct pv_cpu_ops pv_cpu_ops;
	struct pv_irq_ops pv_irq_ops;
	struct pv_apic_ops pv_apic_ops;
	struct pv_mmu_ops pv_mmu_ops;
274 275
};

276 277 278 279 280 281 282
extern struct pv_info pv_info;
extern struct pv_init_ops pv_init_ops;
extern struct pv_time_ops pv_time_ops;
extern struct pv_cpu_ops pv_cpu_ops;
extern struct pv_irq_ops pv_irq_ops;
extern struct pv_apic_ops pv_apic_ops;
extern struct pv_mmu_ops pv_mmu_ops;
283

284
#define PARAVIRT_PATCH(x)					\
285
	(offsetof(struct paravirt_patch_template, x) / sizeof(void *))
286

287 288 289
#define paravirt_type(op)				\
	[paravirt_typenum] "i" (PARAVIRT_PATCH(op)),	\
	[paravirt_opptr] "m" (op)
290 291 292
#define paravirt_clobber(clobber)		\
	[paravirt_clobber] "i" (clobber)

293 294 295 296
/*
 * Generate some code, and mark it as patchable by the
 * apply_paravirt() alternate instruction patcher.
 */
297 298 299
#define _paravirt_alt(insn_string, type, clobber)	\
	"771:\n\t" insn_string "\n" "772:\n"		\
	".pushsection .parainstructions,\"a\"\n"	\
300 301
	_ASM_ALIGN "\n"					\
	_ASM_PTR " 771b\n"				\
302 303 304 305 306
	"  .byte " type "\n"				\
	"  .byte 772b-771b\n"				\
	"  .short " clobber "\n"			\
	".popsection\n"

307
/* Generate patchable code, with the default asm parameters. */
308
#define paravirt_alt(insn_string)					\
309 310
	_paravirt_alt(insn_string, "%c[paravirt_typenum]", "%c[paravirt_clobber]")

311 312 313 314 315
/* Simple instruction patching code. */
#define DEF_NATIVE(ops, name, code) 					\
	extern const char start_##ops##_##name[], end_##ops##_##name[];	\
	asm("start_" #ops "_" #name ": " code "; end_" #ops "_" #name ":")

316 317
unsigned paravirt_patch_nop(void);
unsigned paravirt_patch_ignore(unsigned len);
318 319 320
unsigned paravirt_patch_call(void *insnbuf,
			     const void *target, u16 tgt_clobbers,
			     unsigned long addr, u16 site_clobbers,
321
			     unsigned len);
322
unsigned paravirt_patch_jmp(void *insnbuf, const void *target,
323 324 325
			    unsigned long addr, unsigned len);
unsigned paravirt_patch_default(u8 type, u16 clobbers, void *insnbuf,
				unsigned long addr, unsigned len);
326

327
unsigned paravirt_patch_insns(void *insnbuf, unsigned len,
328 329
			      const char *start, const char *end);

330 331 332
unsigned native_patch(u8 type, u16 clobbers, void *ibuf,
		      unsigned long addr, unsigned len);

333
int paravirt_disable_iospace(void);
334

335 336 337
/*
 * This generates an indirect call based on the operation type number.
 * The type number, computed in PARAVIRT_PATCH, is derived from the
338 339
 * offset into the paravirt_patch_template structure, and can therefore be
 * freely converted back into a structure offset.
340
 */
341
#define PARAVIRT_CALL	"call *%[paravirt_opptr];"
342 343

/*
344 345
 * These macros are intended to wrap calls through one of the paravirt
 * ops structs, so that they can be later identified and patched at
346 347 348
 * runtime.
 *
 * Normally, a call to a pv_op function is a simple indirect call:
349
 * (pv_op_struct.operations)(args...).
350 351 352 353 354 355 356 357 358
 *
 * Unfortunately, this is a relatively slow operation for modern CPUs,
 * because it cannot necessarily determine what the destination
 * address is.  In this case, the address is a runtime constant, so at
 * the very least we can patch the call to e a simple direct call, or
 * ideally, patch an inline implementation into the callsite.  (Direct
 * calls are essentially free, because the call and return addresses
 * are completely predictable.)
 *
359
 * For i386, these macros rely on the standard gcc "regparm(3)" calling
360 361 362 363
 * convention, in which the first three arguments are placed in %eax,
 * %edx, %ecx (in that order), and the remaining arguments are placed
 * on the stack.  All caller-save registers (eax,edx,ecx) are expected
 * to be modified (either clobbered or used for return values).
364 365 366 367 368 369
 * X86_64, on the other hand, already specifies a register-based calling
 * conventions, returning at %rax, with parameteres going on %rdi, %rsi,
 * %rdx, and %rcx. Note that for this reason, x86_64 does not need any
 * special handling for dealing with 4 arguments, unlike i386.
 * However, x86_64 also have to clobber all caller saved registers, which
 * unfortunately, are quite a bit (r8 - r11)
370 371 372 373
 *
 * The call instruction itself is marked by placing its start address
 * and size into the .parainstructions section, so that
 * apply_paravirt() in arch/i386/kernel/alternative.c can do the
374
 * appropriate patching under the control of the backend pv_init_ops
375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391
 * implementation.
 *
 * Unfortunately there's no way to get gcc to generate the args setup
 * for the call, and then allow the call itself to be generated by an
 * inline asm.  Because of this, we must do the complete arg setup and
 * return value handling from within these macros.  This is fairly
 * cumbersome.
 *
 * There are 5 sets of PVOP_* macros for dealing with 0-4 arguments.
 * It could be extended to more arguments, but there would be little
 * to be gained from that.  For each number of arguments, there are
 * the two VCALL and CALL variants for void and non-void functions.
 *
 * When there is a return value, the invoker of the macro must specify
 * the return type.  The macro then uses sizeof() on that type to
 * determine whether its a 32 or 64 bit value, and places the return
 * in the right register(s) (just %eax for 32-bit, and %edx:%eax for
392 393
 * 64-bit). For x86_64 machines, it just returns at %rax regardless of
 * the return value size.
394 395
 *
 * 64-bit arguments are passed as a pair of adjacent 32-bit arguments
396 397
 * i386 also passes 64-bit arguments as a pair of adjacent 32-bit arguments
 * in low,high order
398 399 400 401 402 403 404 405 406
 *
 * Small structures are passed and returned in registers.  The macro
 * calling convention can't directly deal with this, so the wrapper
 * functions must do this.
 *
 * These PVOP_* macros are only defined within this header.  This
 * means that all uses must be wrapped in inline functions.  This also
 * makes sure the incoming and outgoing types are always correct.
 */
407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427
#ifdef CONFIG_X86_32
#define PVOP_VCALL_ARGS			unsigned long __eax, __edx, __ecx
#define PVOP_CALL_ARGS			PVOP_VCALL_ARGS
#define PVOP_VCALL_CLOBBERS		"=a" (__eax), "=d" (__edx),	\
					"=c" (__ecx)
#define PVOP_CALL_CLOBBERS		PVOP_VCALL_CLOBBERS
#define EXTRA_CLOBBERS
#define VEXTRA_CLOBBERS
#else
#define PVOP_VCALL_ARGS		unsigned long __edi, __esi, __edx, __ecx
#define PVOP_CALL_ARGS		PVOP_VCALL_ARGS, __eax
#define PVOP_VCALL_CLOBBERS	"=D" (__edi),				\
				"=S" (__esi), "=d" (__edx),		\
				"=c" (__ecx)

#define PVOP_CALL_CLOBBERS	PVOP_VCALL_CLOBBERS, "=a" (__eax)

#define EXTRA_CLOBBERS	 , "r8", "r9", "r10", "r11"
#define VEXTRA_CLOBBERS	 , "rax", "r8", "r9", "r10", "r11"
#endif

428
#define __PVOP_CALL(rettype, op, pre, post, ...)			\
429
	({								\
430
		rettype __ret;						\
431 432 433
		PVOP_CALL_ARGS;					\
		/* This is 32-bit specific, but is okay in 64-bit */	\
		/* since this condition will never hold */		\
434 435 436 437
		if (sizeof(rettype) > sizeof(unsigned long)) {		\
			asm volatile(pre				\
				     paravirt_alt(PARAVIRT_CALL)	\
				     post				\
438
				     : PVOP_CALL_CLOBBERS		\
439 440 441
				     : paravirt_type(op),		\
				       paravirt_clobber(CLBR_ANY),	\
				       ##__VA_ARGS__			\
442
				     : "memory", "cc" EXTRA_CLOBBERS);	\
443
			__ret = (rettype)((((u64)__edx) << 32) | __eax); \
444
		} else {						\
445
			asm volatile(pre				\
446
				     paravirt_alt(PARAVIRT_CALL)	\
447
				     post				\
448
				     : PVOP_CALL_CLOBBERS		\
449 450 451
				     : paravirt_type(op),		\
				       paravirt_clobber(CLBR_ANY),	\
				       ##__VA_ARGS__			\
452
				     : "memory", "cc" EXTRA_CLOBBERS);	\
453
			__ret = (rettype)__eax;				\
454 455 456
		}							\
		__ret;							\
	})
457
#define __PVOP_VCALL(op, pre, post, ...)				\
458
	({								\
459
		PVOP_VCALL_ARGS;					\
460
		asm volatile(pre					\
461
			     paravirt_alt(PARAVIRT_CALL)		\
462
			     post					\
463
			     : PVOP_VCALL_CLOBBERS			\
464 465 466
			     : paravirt_type(op),			\
			       paravirt_clobber(CLBR_ANY),		\
			       ##__VA_ARGS__				\
467
			     : "memory", "cc" VEXTRA_CLOBBERS);		\
468 469
	})

470 471 472 473 474 475
#define PVOP_CALL0(rettype, op)						\
	__PVOP_CALL(rettype, op, "", "")
#define PVOP_VCALL0(op)							\
	__PVOP_VCALL(op, "", "")

#define PVOP_CALL1(rettype, op, arg1)					\
476
	__PVOP_CALL(rettype, op, "", "", "0" ((unsigned long)(arg1)))
477
#define PVOP_VCALL1(op, arg1)						\
478
	__PVOP_VCALL(op, "", "", "0" ((unsigned long)(arg1)))
479 480

#define PVOP_CALL2(rettype, op, arg1, arg2)				\
481 482
	__PVOP_CALL(rettype, op, "", "", "0" ((unsigned long)(arg1)), 	\
	"1" ((unsigned long)(arg2)))
483
#define PVOP_VCALL2(op, arg1, arg2)					\
484 485
	__PVOP_VCALL(op, "", "", "0" ((unsigned long)(arg1)), 		\
	"1" ((unsigned long)(arg2)))
486 487

#define PVOP_CALL3(rettype, op, arg1, arg2, arg3)			\
488 489
	__PVOP_CALL(rettype, op, "", "", "0" ((unsigned long)(arg1)),	\
	"1"((unsigned long)(arg2)), "2"((unsigned long)(arg3)))
490
#define PVOP_VCALL3(op, arg1, arg2, arg3)				\
491 492
	__PVOP_VCALL(op, "", "", "0" ((unsigned long)(arg1)),		\
	"1"((unsigned long)(arg2)), "2"((unsigned long)(arg3)))
493

494 495
/* This is the only difference in x86_64. We can make it much simpler */
#ifdef CONFIG_X86_32
496 497 498 499 500 501 502 503 504 505
#define PVOP_CALL4(rettype, op, arg1, arg2, arg3, arg4)			\
	__PVOP_CALL(rettype, op,					\
		    "push %[_arg4];", "lea 4(%%esp),%%esp;",		\
		    "0" ((u32)(arg1)), "1" ((u32)(arg2)),		\
		    "2" ((u32)(arg3)), [_arg4] "mr" ((u32)(arg4)))
#define PVOP_VCALL4(op, arg1, arg2, arg3, arg4)				\
	__PVOP_VCALL(op,						\
		    "push %[_arg4];", "lea 4(%%esp),%%esp;",		\
		    "0" ((u32)(arg1)), "1" ((u32)(arg2)),		\
		    "2" ((u32)(arg3)), [_arg4] "mr" ((u32)(arg4)))
506 507 508 509 510 511 512 513 514 515
#else
#define PVOP_CALL4(rettype, op, arg1, arg2, arg3, arg4)			\
	__PVOP_CALL(rettype, op, "", "", "0" ((unsigned long)(arg1)),	\
	"1"((unsigned long)(arg2)), "2"((unsigned long)(arg3)),		\
	"3"((unsigned long)(arg4)))
#define PVOP_VCALL4(op, arg1, arg2, arg3, arg4)				\
	__PVOP_VCALL(op, "", "", "0" ((unsigned long)(arg1)),		\
	"1"((unsigned long)(arg2)), "2"((unsigned long)(arg3)),		\
	"3"((unsigned long)(arg4)))
#endif
516

517 518
static inline int paravirt_enabled(void)
{
519
	return pv_info.paravirt_enabled;
520
}
521

522
static inline void load_sp0(struct tss_struct *tss,
523 524
			     struct thread_struct *thread)
{
525
	PVOP_VCALL2(pv_cpu_ops.load_sp0, tss, thread);
526 527
}

528
#define ARCH_SETUP			pv_init_ops.arch_setup();
529 530
static inline unsigned long get_wallclock(void)
{
531
	return PVOP_CALL0(unsigned long, pv_time_ops.get_wallclock);
532 533 534 535
}

static inline int set_wallclock(unsigned long nowtime)
{
536
	return PVOP_CALL1(int, pv_time_ops.set_wallclock, nowtime);
537 538
}

Z
Zachary Amsden 已提交
539
static inline void (*choose_time_init(void))(void)
540
{
541
	return pv_time_ops.time_init;
542 543 544 545 546 547
}

/* The paravirtualized CPUID instruction. */
static inline void __cpuid(unsigned int *eax, unsigned int *ebx,
			   unsigned int *ecx, unsigned int *edx)
{
548
	PVOP_VCALL4(pv_cpu_ops.cpuid, eax, ebx, ecx, edx);
549 550 551 552 553
}

/*
 * These special macros can be used to get or set a debugging register
 */
554 555
static inline unsigned long paravirt_get_debugreg(int reg)
{
556
	return PVOP_CALL1(unsigned long, pv_cpu_ops.get_debugreg, reg);
557 558 559 560
}
#define get_debugreg(var, reg) var = paravirt_get_debugreg(reg)
static inline void set_debugreg(unsigned long val, int reg)
{
561
	PVOP_VCALL2(pv_cpu_ops.set_debugreg, reg, val);
562
}
563

564 565
static inline void clts(void)
{
566
	PVOP_VCALL0(pv_cpu_ops.clts);
567
}
568

569 570
static inline unsigned long read_cr0(void)
{
571
	return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr0);
572
}
573

574 575
static inline void write_cr0(unsigned long x)
{
576
	PVOP_VCALL1(pv_cpu_ops.write_cr0, x);
577 578 579 580
}

static inline unsigned long read_cr2(void)
{
581
	return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr2);
582 583 584 585
}

static inline void write_cr2(unsigned long x)
{
586
	PVOP_VCALL1(pv_mmu_ops.write_cr2, x);
587 588 589 590
}

static inline unsigned long read_cr3(void)
{
591
	return PVOP_CALL0(unsigned long, pv_mmu_ops.read_cr3);
592
}
593

594 595
static inline void write_cr3(unsigned long x)
{
596
	PVOP_VCALL1(pv_mmu_ops.write_cr3, x);
597
}
598

599 600
static inline unsigned long read_cr4(void)
{
601
	return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr4);
602 603 604
}
static inline unsigned long read_cr4_safe(void)
{
605
	return PVOP_CALL0(unsigned long, pv_cpu_ops.read_cr4_safe);
606
}
607

608 609
static inline void write_cr4(unsigned long x)
{
610
	PVOP_VCALL1(pv_cpu_ops.write_cr4, x);
611
}
612

613 614
static inline void raw_safe_halt(void)
{
615
	PVOP_VCALL0(pv_irq_ops.safe_halt);
616 617 618 619
}

static inline void halt(void)
{
620
	PVOP_VCALL0(pv_irq_ops.safe_halt);
621 622 623 624
}

static inline void wbinvd(void)
{
625
	PVOP_VCALL0(pv_cpu_ops.wbinvd);
626 627
}

628
#define get_kernel_rpl()  (pv_info.kernel_rpl)
629

630 631
static inline u64 paravirt_read_msr(unsigned msr, int *err)
{
632
	return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err);
633 634 635
}
static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high)
{
636
	return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high);
637 638
}

639
/* These should all do BUG_ON(_err), but our headers are too tangled. */
640 641 642 643 644
#define rdmsr(msr,val1,val2) do {		\
	int _err;				\
	u64 _l = paravirt_read_msr(msr, &_err);	\
	val1 = (u32)_l;				\
	val2 = _l >> 32;			\
645 646
} while(0)

647 648
#define wrmsr(msr,val1,val2) do {		\
	paravirt_write_msr(msr, val1, val2);	\
649 650
} while(0)

651 652 653
#define rdmsrl(msr,val) do {			\
	int _err;				\
	val = paravirt_read_msr(msr, &_err);	\
654 655
} while(0)

656
#define wrmsrl(msr,val)		wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32)
657
#define wrmsr_safe(msr,a,b)	paravirt_write_msr(msr, a, b)
658 659

/* rdmsr with exception handling */
660 661 662 663 664
#define rdmsr_safe(msr,a,b) ({			\
	int _err;				\
	u64 _l = paravirt_read_msr(msr, &_err);	\
	(*a) = (u32)_l;				\
	(*b) = _l >> 32;			\
665 666
	_err; })

667 668 669

static inline u64 paravirt_read_tsc(void)
{
670
	return PVOP_CALL0(u64, pv_cpu_ops.read_tsc);
671
}
672

673 674 675
#define rdtscl(low) do {			\
	u64 _l = paravirt_read_tsc();		\
	low = (int)_l;				\
676 677
} while(0)

678
#define rdtscll(val) (val = paravirt_read_tsc())
679

680 681
static inline unsigned long long paravirt_sched_clock(void)
{
682
	return PVOP_CALL0(unsigned long long, pv_time_ops.sched_clock);
683
}
684
#define calculate_cpu_khz() (pv_time_ops.get_cpu_khz())
685

686 687
static inline unsigned long long paravirt_read_pmc(int counter)
{
688
	return PVOP_CALL1(u64, pv_cpu_ops.read_pmc, counter);
689
}
690

691 692 693 694 695
#define rdpmc(counter,low,high) do {		\
	u64 _l = paravirt_read_pmc(counter);	\
	low = (u32)_l;				\
	high = _l >> 32;			\
} while(0)
696

697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717
static inline unsigned long long paravirt_rdtscp(unsigned int *aux)
{
	return PVOP_CALL1(u64, pv_cpu_ops.read_tscp, aux);
}

#define rdtscp(low, high, aux)				\
do {							\
	int __aux;					\
	unsigned long __val = paravirt_rdtscp(&__aux);	\
	(low) = (u32)__val;				\
	(high) = (u32)(__val >> 32);			\
	(aux) = __aux;					\
} while (0)

#define rdtscpll(val, aux)				\
do {							\
	unsigned long __aux; 				\
	val = paravirt_rdtscp(&__aux);			\
	(aux) = __aux;					\
} while (0)

718 719
static inline void load_TR_desc(void)
{
720
	PVOP_VCALL0(pv_cpu_ops.load_tr_desc);
721
}
722
static inline void load_gdt(const struct desc_ptr *dtr)
723
{
724
	PVOP_VCALL1(pv_cpu_ops.load_gdt, dtr);
725
}
726
static inline void load_idt(const struct desc_ptr *dtr)
727
{
728
	PVOP_VCALL1(pv_cpu_ops.load_idt, dtr);
729 730 731
}
static inline void set_ldt(const void *addr, unsigned entries)
{
732
	PVOP_VCALL2(pv_cpu_ops.set_ldt, addr, entries);
733
}
734
static inline void store_gdt(struct desc_ptr *dtr)
735
{
736
	PVOP_VCALL1(pv_cpu_ops.store_gdt, dtr);
737
}
738
static inline void store_idt(struct desc_ptr *dtr)
739
{
740
	PVOP_VCALL1(pv_cpu_ops.store_idt, dtr);
741 742 743
}
static inline unsigned long paravirt_store_tr(void)
{
744
	return PVOP_CALL0(unsigned long, pv_cpu_ops.store_tr);
745 746 747 748
}
#define store_tr(tr)	((tr) = paravirt_store_tr())
static inline void load_TLS(struct thread_struct *t, unsigned cpu)
{
749
	PVOP_VCALL2(pv_cpu_ops.load_tls, t, cpu);
750
}
751 752 753

static inline void write_ldt_entry(struct desc_struct *dt, int entry,
				   const void *desc)
754
{
755
	PVOP_VCALL3(pv_cpu_ops.write_ldt_entry, dt, entry, desc);
756
}
757 758 759

static inline void write_gdt_entry(struct desc_struct *dt, int entry,
				   void *desc, int type)
760
{
761
	PVOP_VCALL4(pv_cpu_ops.write_gdt_entry, dt, entry, desc, type);
762
}
763

764
static inline void write_idt_entry(gate_desc *dt, int entry, const gate_desc *g)
765
{
766
	PVOP_VCALL3(pv_cpu_ops.write_idt_entry, dt, entry, g);
767 768 769
}
static inline void set_iopl_mask(unsigned mask)
{
770
	PVOP_VCALL1(pv_cpu_ops.set_iopl_mask, mask);
771
}
772

773 774
/* The paravirtualized I/O functions */
static inline void slow_down_io(void) {
775
	pv_cpu_ops.io_delay();
776
#ifdef REALLY_SLOW_IO
777 778 779
	pv_cpu_ops.io_delay();
	pv_cpu_ops.io_delay();
	pv_cpu_ops.io_delay();
780 781 782
#endif
}

783 784 785 786
#ifdef CONFIG_X86_LOCAL_APIC
/*
 * Basic functions accessing APICs.
 */
787
static inline void apic_write(unsigned long reg, u32 v)
788
{
789
	PVOP_VCALL2(pv_apic_ops.apic_write, reg, v);
790 791
}

792
static inline void apic_write_atomic(unsigned long reg, u32 v)
793
{
794
	PVOP_VCALL2(pv_apic_ops.apic_write_atomic, reg, v);
795 796
}

797
static inline u32 apic_read(unsigned long reg)
798
{
799
	return PVOP_CALL1(unsigned long, pv_apic_ops.apic_read, reg);
800
}
Z
Zachary Amsden 已提交
801 802 803

static inline void setup_boot_clock(void)
{
804
	PVOP_VCALL0(pv_apic_ops.setup_boot_clock);
Z
Zachary Amsden 已提交
805 806 807 808
}

static inline void setup_secondary_clock(void)
{
809
	PVOP_VCALL0(pv_apic_ops.setup_secondary_clock);
Z
Zachary Amsden 已提交
810
}
811 812
#endif

813 814
static inline void paravirt_post_allocator_init(void)
{
815 816
	if (pv_init_ops.post_allocator_init)
		(*pv_init_ops.post_allocator_init)();
817 818
}

819 820
static inline void paravirt_pagetable_setup_start(pgd_t *base)
{
821
	(*pv_mmu_ops.pagetable_setup_start)(base);
822 823 824 825
}

static inline void paravirt_pagetable_setup_done(pgd_t *base)
{
826
	(*pv_mmu_ops.pagetable_setup_done)(base);
827
}
828

829 830 831 832
#ifdef CONFIG_SMP
static inline void startup_ipi_hook(int phys_apicid, unsigned long start_eip,
				    unsigned long start_esp)
{
833 834
	PVOP_VCALL3(pv_apic_ops.startup_ipi_hook,
		    phys_apicid, start_eip, start_esp);
835 836
}
#endif
837

838 839 840
static inline void paravirt_activate_mm(struct mm_struct *prev,
					struct mm_struct *next)
{
841
	PVOP_VCALL2(pv_mmu_ops.activate_mm, prev, next);
842 843 844 845 846
}

static inline void arch_dup_mmap(struct mm_struct *oldmm,
				 struct mm_struct *mm)
{
847
	PVOP_VCALL2(pv_mmu_ops.dup_mmap, oldmm, mm);
848 849 850 851
}

static inline void arch_exit_mmap(struct mm_struct *mm)
{
852
	PVOP_VCALL1(pv_mmu_ops.exit_mmap, mm);
853 854
}

855 856
static inline void __flush_tlb(void)
{
857
	PVOP_VCALL0(pv_mmu_ops.flush_tlb_user);
858 859 860
}
static inline void __flush_tlb_global(void)
{
861
	PVOP_VCALL0(pv_mmu_ops.flush_tlb_kernel);
862 863 864
}
static inline void __flush_tlb_single(unsigned long addr)
{
865
	PVOP_VCALL1(pv_mmu_ops.flush_tlb_single, addr);
866
}
867

868 869 870
static inline void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm,
				    unsigned long va)
{
871
	PVOP_VCALL3(pv_mmu_ops.flush_tlb_others, &cpumask, mm, va);
872 873
}

874
static inline void paravirt_alloc_pt(struct mm_struct *mm, unsigned pfn)
875
{
876
	PVOP_VCALL2(pv_mmu_ops.alloc_pt, mm, pfn);
877 878 879
}
static inline void paravirt_release_pt(unsigned pfn)
{
880
	PVOP_VCALL1(pv_mmu_ops.release_pt, pfn);
881
}
882

883 884
static inline void paravirt_alloc_pd(unsigned pfn)
{
885
	PVOP_VCALL1(pv_mmu_ops.alloc_pd, pfn);
886
}
887

888 889 890
static inline void paravirt_alloc_pd_clone(unsigned pfn, unsigned clonepfn,
					   unsigned start, unsigned count)
{
891
	PVOP_VCALL4(pv_mmu_ops.alloc_pd_clone, pfn, clonepfn, start, count);
892 893
}
static inline void paravirt_release_pd(unsigned pfn)
894
{
895
	PVOP_VCALL1(pv_mmu_ops.release_pd, pfn);
896 897
}

898 899 900 901
#ifdef CONFIG_HIGHPTE
static inline void *kmap_atomic_pte(struct page *page, enum km_type type)
{
	unsigned long ret;
902
	ret = PVOP_CALL2(unsigned long, pv_mmu_ops.kmap_atomic_pte, page, type);
903 904 905 906
	return (void *)ret;
}
#endif

907 908
static inline void pte_update(struct mm_struct *mm, unsigned long addr,
			      pte_t *ptep)
909
{
910
	PVOP_VCALL3(pv_mmu_ops.pte_update, mm, addr, ptep);
911 912
}

913 914
static inline void pte_update_defer(struct mm_struct *mm, unsigned long addr,
				    pte_t *ptep)
915
{
916
	PVOP_VCALL3(pv_mmu_ops.pte_update_defer, mm, addr, ptep);
917 918
}

919 920
#ifdef CONFIG_X86_PAE
static inline pte_t __pte(unsigned long long val)
921
{
922 923
	unsigned long long ret = PVOP_CALL2(unsigned long long,
					    pv_mmu_ops.make_pte,
924
					    val, val >> 32);
925
	return (pte_t) { .pte = ret };
926 927
}

928
static inline pmd_t __pmd(unsigned long long val)
929
{
930 931
	return (pmd_t) { PVOP_CALL2(unsigned long long, pv_mmu_ops.make_pmd,
				    val, val >> 32) };
932 933 934 935
}

static inline pgd_t __pgd(unsigned long long val)
{
936 937
	return (pgd_t) { PVOP_CALL2(unsigned long long, pv_mmu_ops.make_pgd,
				    val, val >> 32) };
938 939 940 941
}

static inline unsigned long long pte_val(pte_t x)
{
942 943
	return PVOP_CALL2(unsigned long long, pv_mmu_ops.pte_val,
			  x.pte_low, x.pte_high);
944 945 946 947
}

static inline unsigned long long pmd_val(pmd_t x)
{
948 949
	return PVOP_CALL2(unsigned long long, pv_mmu_ops.pmd_val,
			  x.pmd, x.pmd >> 32);
950 951 952 953
}

static inline unsigned long long pgd_val(pgd_t x)
{
954 955
	return PVOP_CALL2(unsigned long long, pv_mmu_ops.pgd_val,
			  x.pgd, x.pgd >> 32);
956 957 958 959
}

static inline void set_pte(pte_t *ptep, pte_t pteval)
{
960
	PVOP_VCALL3(pv_mmu_ops.set_pte, ptep, pteval.pte_low, pteval.pte_high);
961 962 963 964 965 966
}

static inline void set_pte_at(struct mm_struct *mm, unsigned long addr,
			      pte_t *ptep, pte_t pteval)
{
	/* 5 arg words */
967
	pv_mmu_ops.set_pte_at(mm, addr, ptep, pteval);
968 969 970 971
}

static inline void set_pte_atomic(pte_t *ptep, pte_t pteval)
{
972 973
	PVOP_VCALL3(pv_mmu_ops.set_pte_atomic, ptep,
		    pteval.pte_low, pteval.pte_high);
974 975
}

976 977
static inline void set_pte_present(struct mm_struct *mm, unsigned long addr,
				   pte_t *ptep, pte_t pte)
978
{
979
	/* 5 arg words */
980
	pv_mmu_ops.set_pte_present(mm, addr, ptep, pte);
981 982
}

983 984
static inline void set_pmd(pmd_t *pmdp, pmd_t pmdval)
{
985 986
	PVOP_VCALL3(pv_mmu_ops.set_pmd, pmdp,
		    pmdval.pmd, pmdval.pmd >> 32);
987 988
}

989 990
static inline void set_pud(pud_t *pudp, pud_t pudval)
{
991 992
	PVOP_VCALL3(pv_mmu_ops.set_pud, pudp,
		    pudval.pgd.pgd, pudval.pgd.pgd >> 32);
993 994 995 996
}

static inline void pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep)
{
997
	PVOP_VCALL3(pv_mmu_ops.pte_clear, mm, addr, ptep);
998 999 1000 1001
}

static inline void pmd_clear(pmd_t *pmdp)
{
1002
	PVOP_VCALL1(pv_mmu_ops.pmd_clear, pmdp);
1003 1004 1005
}

#else  /* !CONFIG_X86_PAE */
1006

1007 1008
static inline pte_t __pte(unsigned long val)
{
1009
	return (pte_t) { PVOP_CALL1(unsigned long, pv_mmu_ops.make_pte, val) };
1010
}
1011 1012 1013

static inline pgd_t __pgd(unsigned long val)
{
1014
	return (pgd_t) { PVOP_CALL1(unsigned long, pv_mmu_ops.make_pgd, val) };
1015 1016 1017 1018
}

static inline unsigned long pte_val(pte_t x)
{
1019
	return PVOP_CALL1(unsigned long, pv_mmu_ops.pte_val, x.pte_low);
1020 1021 1022 1023
}

static inline unsigned long pgd_val(pgd_t x)
{
1024
	return PVOP_CALL1(unsigned long, pv_mmu_ops.pgd_val, x.pgd);
1025 1026 1027 1028
}

static inline void set_pte(pte_t *ptep, pte_t pteval)
{
1029
	PVOP_VCALL2(pv_mmu_ops.set_pte, ptep, pteval.pte_low);
1030 1031 1032 1033 1034
}

static inline void set_pte_at(struct mm_struct *mm, unsigned long addr,
			      pte_t *ptep, pte_t pteval)
{
1035
	PVOP_VCALL4(pv_mmu_ops.set_pte_at, mm, addr, ptep, pteval.pte_low);
1036 1037 1038 1039
}

static inline void set_pmd(pmd_t *pmdp, pmd_t pmdval)
{
1040
	PVOP_VCALL2(pv_mmu_ops.set_pmd, pmdp, pmdval.pud.pgd.pgd);
1041 1042
}
#endif	/* CONFIG_X86_PAE */
1043

1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057
/* Lazy mode for batching updates / context switch */
enum paravirt_lazy_mode {
	PARAVIRT_LAZY_NONE,
	PARAVIRT_LAZY_MMU,
	PARAVIRT_LAZY_CPU,
};

enum paravirt_lazy_mode paravirt_get_lazy_mode(void);
void paravirt_enter_lazy_cpu(void);
void paravirt_leave_lazy_cpu(void);
void paravirt_enter_lazy_mmu(void);
void paravirt_leave_lazy_mmu(void);
void paravirt_leave_lazy(enum paravirt_lazy_mode mode);

1058
#define  __HAVE_ARCH_ENTER_LAZY_CPU_MODE
1059 1060
static inline void arch_enter_lazy_cpu_mode(void)
{
1061
	PVOP_VCALL0(pv_cpu_ops.lazy_mode.enter);
1062 1063 1064 1065
}

static inline void arch_leave_lazy_cpu_mode(void)
{
1066
	PVOP_VCALL0(pv_cpu_ops.lazy_mode.leave);
1067 1068 1069 1070
}

static inline void arch_flush_lazy_cpu_mode(void)
{
1071 1072 1073 1074
	if (unlikely(paravirt_get_lazy_mode() == PARAVIRT_LAZY_CPU)) {
		arch_leave_lazy_cpu_mode();
		arch_enter_lazy_cpu_mode();
	}
1075 1076
}

1077 1078

#define  __HAVE_ARCH_ENTER_LAZY_MMU_MODE
1079 1080
static inline void arch_enter_lazy_mmu_mode(void)
{
1081
	PVOP_VCALL0(pv_mmu_ops.lazy_mode.enter);
1082 1083 1084 1085
}

static inline void arch_leave_lazy_mmu_mode(void)
{
1086
	PVOP_VCALL0(pv_mmu_ops.lazy_mode.leave);
1087 1088 1089 1090
}

static inline void arch_flush_lazy_mmu_mode(void)
{
1091 1092 1093 1094
	if (unlikely(paravirt_get_lazy_mode() == PARAVIRT_LAZY_MMU)) {
		arch_leave_lazy_mmu_mode();
		arch_enter_lazy_mmu_mode();
	}
1095
}
1096

1097 1098 1099
void _paravirt_nop(void);
#define paravirt_nop	((void *)_paravirt_nop)

1100
/* These all sit in the .parainstructions section to tell us what to patch. */
1101
struct paravirt_patch_site {
1102 1103 1104 1105 1106 1107
	u8 *instr; 		/* original instructions */
	u8 instrtype;		/* type of this instruction */
	u8 len;			/* length of original instruction */
	u16 clobbers;		/* what registers you may clobber */
};

1108 1109 1110
extern struct paravirt_patch_site __parainstructions[],
	__parainstructions_end[];

1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126
#ifdef CONFIG_X86_32
#define PV_SAVE_REGS "pushl %%ecx; pushl %%edx;"
#define PV_RESTORE_REGS "popl %%edx; popl %%ecx"
#define PV_FLAGS_ARG "0"
#define PV_EXTRA_CLOBBERS
#define PV_VEXTRA_CLOBBERS
#else
/* We save some registers, but all of them, that's too much. We clobber all
 * caller saved registers but the argument parameter */
#define PV_SAVE_REGS "pushq %%rdi;"
#define PV_RESTORE_REGS "popq %%rdi;"
#define PV_EXTRA_CLOBBERS EXTRA_CLOBBERS, "rcx" , "rdx"
#define PV_VEXTRA_CLOBBERS EXTRA_CLOBBERS, "rdi", "rcx" , "rdx"
#define PV_FLAGS_ARG "D"
#endif

1127 1128 1129 1130
static inline unsigned long __raw_local_save_flags(void)
{
	unsigned long f;

1131
	asm volatile(paravirt_alt(PV_SAVE_REGS
1132
				  PARAVIRT_CALL
1133
				  PV_RESTORE_REGS)
1134
		     : "=a"(f)
1135
		     : paravirt_type(pv_irq_ops.save_fl),
1136
		       paravirt_clobber(CLBR_EAX)
1137
		     : "memory", "cc" PV_VEXTRA_CLOBBERS);
1138 1139 1140 1141 1142
	return f;
}

static inline void raw_local_irq_restore(unsigned long f)
{
1143
	asm volatile(paravirt_alt(PV_SAVE_REGS
1144
				  PARAVIRT_CALL
1145
				  PV_RESTORE_REGS)
1146
		     : "=a"(f)
1147
		     : PV_FLAGS_ARG(f),
1148
		       paravirt_type(pv_irq_ops.restore_fl),
1149
		       paravirt_clobber(CLBR_EAX)
1150
		     : "memory", "cc" PV_EXTRA_CLOBBERS);
1151 1152 1153 1154
}

static inline void raw_local_irq_disable(void)
{
1155
	asm volatile(paravirt_alt(PV_SAVE_REGS
1156
				  PARAVIRT_CALL
1157
				  PV_RESTORE_REGS)
1158
		     :
1159
		     : paravirt_type(pv_irq_ops.irq_disable),
1160
		       paravirt_clobber(CLBR_EAX)
1161
		     : "memory", "eax", "cc" PV_EXTRA_CLOBBERS);
1162 1163 1164 1165
}

static inline void raw_local_irq_enable(void)
{
1166
	asm volatile(paravirt_alt(PV_SAVE_REGS
1167
				  PARAVIRT_CALL
1168
				  PV_RESTORE_REGS)
1169
		     :
1170
		     : paravirt_type(pv_irq_ops.irq_enable),
1171
		       paravirt_clobber(CLBR_EAX)
1172
		     : "memory", "eax", "cc" PV_EXTRA_CLOBBERS);
1173 1174 1175 1176 1177 1178
}

static inline unsigned long __raw_local_irq_save(void)
{
	unsigned long f;

1179 1180
	f = __raw_local_save_flags();
	raw_local_irq_disable();
1181 1182 1183
	return f;
}

1184
/* Make sure as little as possible of this mess escapes. */
1185
#undef PARAVIRT_CALL
1186 1187
#undef __PVOP_CALL
#undef __PVOP_VCALL
1188 1189 1190 1191 1192 1193 1194 1195 1196 1197
#undef PVOP_VCALL0
#undef PVOP_CALL0
#undef PVOP_VCALL1
#undef PVOP_CALL1
#undef PVOP_VCALL2
#undef PVOP_CALL2
#undef PVOP_VCALL3
#undef PVOP_CALL3
#undef PVOP_VCALL4
#undef PVOP_CALL4
1198

1199 1200
#else  /* __ASSEMBLY__ */

1201
#define _PVSITE(ptype, clobbers, ops, word, algn)	\
1202 1203 1204 1205
771:;						\
	ops;					\
772:;						\
	.pushsection .parainstructions,"a";	\
1206 1207
	 .align	algn;				\
	 word 771b;				\
1208 1209 1210 1211 1212
	 .byte ptype;				\
	 .byte 772b-771b;			\
	 .short clobbers;			\
	.popsection

1213 1214

#ifdef CONFIG_X86_64
1215 1216 1217
#define PV_SAVE_REGS   pushq %rax; pushq %rdi; pushq %rcx; pushq %rdx
#define PV_RESTORE_REGS popq %rdx; popq %rcx; popq %rdi; popq %rax
#define PARA_PATCH(struct, off)        ((PARAVIRT_PATCH_##struct + (off)) / 8)
1218 1219
#define PARA_SITE(ptype, clobbers, ops) _PVSITE(ptype, clobbers, ops, .quad, 8)
#else
1220 1221 1222
#define PV_SAVE_REGS   pushl %eax; pushl %edi; pushl %ecx; pushl %edx
#define PV_RESTORE_REGS popl %edx; popl %ecx; popl %edi; popl %eax
#define PARA_PATCH(struct, off)        ((PARAVIRT_PATCH_##struct + (off)) / 4)
1223 1224 1225
#define PARA_SITE(ptype, clobbers, ops) _PVSITE(ptype, clobbers, ops, .long, 4)
#endif

1226 1227 1228
#define INTERRUPT_RETURN						\
	PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_iret), CLBR_NONE,	\
		  jmp *%cs:pv_cpu_ops+PV_CPU_iret)
1229 1230

#define DISABLE_INTERRUPTS(clobbers)					\
1231
	PARA_SITE(PARA_PATCH(pv_irq_ops, PV_IRQ_irq_disable), clobbers, \
1232
		  PV_SAVE_REGS;			\
1233
		  call *%cs:pv_irq_ops+PV_IRQ_irq_disable;		\
1234
		  PV_RESTORE_REGS;)			\
1235 1236

#define ENABLE_INTERRUPTS(clobbers)					\
1237
	PARA_SITE(PARA_PATCH(pv_irq_ops, PV_IRQ_irq_enable), clobbers,	\
1238
		  PV_SAVE_REGS;			\
1239
		  call *%cs:pv_irq_ops+PV_IRQ_irq_enable;		\
1240
		  PV_RESTORE_REGS;)
1241

1242 1243 1244 1245
#define ENABLE_INTERRUPTS_SYSCALL_RET					\
	PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_irq_enable_syscall_ret),\
		  CLBR_NONE,						\
		  jmp *%cs:pv_cpu_ops+PV_CPU_irq_enable_syscall_ret)
1246

1247

1248
#ifdef CONFIG_X86_32
1249
#define GET_CR0_INTO_EAX			\
1250
	push %ecx; push %edx;			\
1251
	call *pv_cpu_ops+PV_CPU_read_cr0;	\
1252
	pop %edx; pop %ecx
1253
#else
1254 1255 1256 1257 1258 1259 1260
#define SWAPGS								\
	PARA_SITE(PARA_PATCH(pv_cpu_ops, PV_CPU_swapgs), CLBR_NONE,	\
		  PV_SAVE_REGS;						\
		  call *pv_cpu_ops+PV_CPU_swapgs;			\
		  PV_RESTORE_REGS					\
		 )

1261 1262 1263 1264 1265
#define GET_CR2_INTO_RCX			\
	call *pv_mmu_ops+PV_MMU_read_cr2;	\
	movq %rax, %rcx;			\
	xorq %rax, %rax;

1266
#endif
1267

1268 1269 1270
#endif /* __ASSEMBLY__ */
#endif /* CONFIG_PARAVIRT */
#endif	/* __ASM_PARAVIRT_H */