xen-asm_64.S 2.7 KB
Newer Older
1
/*
2 3 4 5 6 7 8 9 10 11
 * Asm versions of Xen pv-ops, suitable for either direct use or
 * inlining.  The inline versions are the same as the direct-use
 * versions, with the pre- and post-amble chopped off.
 *
 * This code is encoded for size rather than absolute efficiency, with
 * a view to being able to inline as much as possible.
 *
 * We only bother with direct forms (ie, vcpu in pda) of the
 * operations here; the indirect forms are better handled in C, since
 * they're generally too large to inline anyway.
12 13
 */

14
#include <asm/errno.h>
15
#include <asm/percpu.h>
16 17
#include <asm/processor-flags.h>
#include <asm/segment.h>
18
#include <asm/asm-offsets.h>
19
#include <asm/thread_info.h>
20 21 22

#include <xen/interface/xen.h>

23
#include "xen-asm.h"
24

25
ENTRY(xen_adjust_exception_frame)
26 27
	mov 8+0(%rsp), %rcx
	mov 8+8(%rsp), %r11
28
	ret $16
29
ENDPROC(xen_adjust_exception_frame)
30

31 32
hypercall_iret = hypercall_page + __HYPERVISOR_iret * 32
/*
33 34 35 36 37 38 39 40 41 42 43 44 45
 * Xen64 iret frame:
 *
 *	ss
 *	rsp
 *	rflags
 *	cs
 *	rip		<-- standard iret frame
 *
 *	flags
 *
 *	rcx		}
 *	r11		}<-- pushed by hypercall page
 * rsp->rax		}
46
 */
47 48
ENTRY(xen_iret)
	pushq $0
49 50 51
1:	jmp hypercall_iret
ENDPATCH(xen_iret)
RELOC(xen_iret, 1b+1)
52

53
ENTRY(xen_sysret64)
54 55 56 57
	/*
	 * We're already on the usermode stack at this point, but
	 * still with the kernel gs, so we can easily switch back
	 */
58
	movq %rsp, PER_CPU_VAR(rsp_scratch)
59
	movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
60 61

	pushq $__USER_DS
62
	pushq PER_CPU_VAR(rsp_scratch)
63 64 65 66 67 68 69 70 71 72
	pushq %r11
	pushq $__USER_CS
	pushq %rcx

	pushq $VGCF_in_syscall
1:	jmp hypercall_iret
ENDPATCH(xen_sysret64)
RELOC(xen_sysret64, 1b+1)

/*
73 74 75 76 77 78 79 80 81 82 83 84 85 86 87
 * Xen handles syscall callbacks much like ordinary exceptions, which
 * means we have:
 * - kernel gs
 * - kernel rsp
 * - an iret-like stack frame on the stack (including rcx and r11):
 *	ss
 *	rsp
 *	rflags
 *	cs
 *	rip
 *	r11
 * rsp->rcx
 *
 * In all the entrypoints, we undo all that to make it look like a
 * CPU-generated syscall/sysenter and jump to the normal entrypoint.
88 89 90
 */

.macro undo_xen_syscall
91 92 93
	mov 0*8(%rsp), %rcx
	mov 1*8(%rsp), %r11
	mov 5*8(%rsp), %rsp
94 95 96 97 98
.endm

/* Normal 64-bit system call target */
ENTRY(xen_syscall_target)
	undo_xen_syscall
99
	jmp entry_SYSCALL_64_after_swapgs
100 101 102 103 104 105 106
ENDPROC(xen_syscall_target)

#ifdef CONFIG_IA32_EMULATION

/* 32-bit compat syscall target */
ENTRY(xen_syscall32_target)
	undo_xen_syscall
107
	jmp entry_SYSCALL_compat
108 109 110 111 112
ENDPROC(xen_syscall32_target)

/* 32-bit compat sysenter target */
ENTRY(xen_sysenter_target)
	undo_xen_syscall
113
	jmp entry_SYSENTER_compat
114 115 116 117 118 119
ENDPROC(xen_sysenter_target)

#else /* !CONFIG_IA32_EMULATION */

ENTRY(xen_syscall32_target)
ENTRY(xen_sysenter_target)
120
	lea 16(%rsp), %rsp	/* strip %rcx, %r11 */
121
	mov $-ENOSYS, %rax
122
	pushq $0
123 124 125 126 127
	jmp hypercall_iret
ENDPROC(xen_syscall32_target)
ENDPROC(xen_sysenter_target)

#endif	/* CONFIG_IA32_EMULATION */