calling.h 7.5 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
2
#include <linux/jump_label.h>
3
#include <asm/unwind_hints.h>
4 5
#include <asm/cpufeatures.h>
#include <asm/page_types.h>
6

7
/*
8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54

 x86 function call convention, 64-bit:
 -------------------------------------
  arguments           |  callee-saved      | extra caller-saved | return
 [callee-clobbered]   |                    | [callee-clobbered] |
 ---------------------------------------------------------------------------
 rdi rsi rdx rcx r8-9 | rbx rbp [*] r12-15 | r10-11             | rax, rdx [**]

 ( rsp is obviously invariant across normal function calls. (gcc can 'merge'
   functions when it sees tail-call optimization possibilities) rflags is
   clobbered. Leftover arguments are passed over the stack frame.)

 [*]  In the frame-pointers case rbp is fixed to the stack frame.

 [**] for struct return values wider than 64 bits the return convention is a
      bit more complex: up to 128 bits width we return small structures
      straight in rax, rdx. For structures larger than that (3 words or
      larger) the caller puts a pointer to an on-stack return struct
      [allocated in the caller's stack frame] into the first argument - i.e.
      into rdi. All other arguments shift up by one in this case.
      Fortunately this case is rare in the kernel.

For 32-bit we have the following conventions - kernel is built with
-mregparm=3 and -freg-struct-return:

 x86 function calling convention, 32-bit:
 ----------------------------------------
  arguments         | callee-saved        | extra caller-saved | return
 [callee-clobbered] |                     | [callee-clobbered] |
 -------------------------------------------------------------------------
 eax edx ecx        | ebx edi esi ebp [*] | <none>             | eax, edx [**]

 ( here too esp is obviously invariant across normal function calls. eflags
   is clobbered. Leftover arguments are passed over the stack frame. )

 [*]  In the frame-pointers case ebp is fixed to the stack frame.

 [**] We build with -freg-struct-return, which on 32-bit means similar
      semantics as on 64-bit: edx can be used for a second return value
      (i.e. covering integer and structure sizes up to 64 bits) - after that
      it gets more complex and more expensive: 3-word or larger struct returns
      get done in the caller's frame and the pointer to the return struct goes
      into regparm0, i.e. eax - the other arguments shift up and the
      function's register parameters degenerate to regparm=2 in essence.

*/

55 56
#ifdef CONFIG_X86_64

57
/*
T
Tao Guo 已提交
58 59
 * 64-bit system call stack frame layout defines and helpers,
 * for assembly code:
60
 */
L
Linus Torvalds 已提交
61

62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94
/* The layout forms the "struct pt_regs" on the stack: */
/*
 * C ABI says these regs are callee-preserved. They aren't saved on kernel entry
 * unless syscall needs a complete, fully filled "struct pt_regs".
 */
#define R15		0*8
#define R14		1*8
#define R13		2*8
#define R12		3*8
#define RBP		4*8
#define RBX		5*8
/* These regs are callee-clobbered. Always saved on kernel entry. */
#define R11		6*8
#define R10		7*8
#define R9		8*8
#define R8		9*8
#define RAX		10*8
#define RCX		11*8
#define RDX		12*8
#define RSI		13*8
#define RDI		14*8
/*
 * On syscall entry, this is syscall#. On CPU exception, this is error code.
 * On hw interrupt, it's IRQ number:
 */
#define ORIG_RAX	15*8
/* Return frame for iretq */
#define RIP		16*8
#define CS		17*8
#define EFLAGS		18*8
#define RSP		19*8
#define SS		20*8

95 96
#define SIZEOF_PTREGS	21*8

97 98
	.macro ALLOC_PT_GPREGS_ON_STACK
	addq	$-(15*8), %rsp
99
	.endm
100

101 102
	.macro SAVE_C_REGS_HELPER offset=0 rax=1 rcx=1 r8910=1 r11=1
	.if \r11
103
	movq %r11, 6*8+\offset(%rsp)
104 105
	.endif
	.if \r8910
106 107 108
	movq %r10, 7*8+\offset(%rsp)
	movq %r9,  8*8+\offset(%rsp)
	movq %r8,  9*8+\offset(%rsp)
109
	.endif
110
	.if \rax
111
	movq %rax, 10*8+\offset(%rsp)
112 113
	.endif
	.if \rcx
114
	movq %rcx, 11*8+\offset(%rsp)
L
Linus Torvalds 已提交
115
	.endif
116 117 118
	movq %rdx, 12*8+\offset(%rsp)
	movq %rsi, 13*8+\offset(%rsp)
	movq %rdi, 14*8+\offset(%rsp)
119
	UNWIND_HINT_REGS offset=\offset extra=0
120 121
	.endm
	.macro SAVE_C_REGS offset=0
122
	SAVE_C_REGS_HELPER \offset, 1, 1, 1, 1
123 124
	.endm
	.macro SAVE_C_REGS_EXCEPT_RAX_RCX offset=0
125
	SAVE_C_REGS_HELPER \offset, 0, 0, 1, 1
126 127
	.endm
	.macro SAVE_C_REGS_EXCEPT_R891011
128
	SAVE_C_REGS_HELPER 0, 1, 1, 0, 0
129 130
	.endm
	.macro SAVE_C_REGS_EXCEPT_RCX_R891011
131 132 133 134
	SAVE_C_REGS_HELPER 0, 1, 0, 0, 0
	.endm
	.macro SAVE_C_REGS_EXCEPT_RAX_RCX_R11
	SAVE_C_REGS_HELPER 0, 0, 0, 1, 0
135
	.endm
136

137
	.macro SAVE_EXTRA_REGS offset=0
138 139 140 141 142 143
	movq %r15, 0*8+\offset(%rsp)
	movq %r14, 1*8+\offset(%rsp)
	movq %r13, 2*8+\offset(%rsp)
	movq %r12, 3*8+\offset(%rsp)
	movq %rbp, 4*8+\offset(%rsp)
	movq %rbx, 5*8+\offset(%rsp)
144
	UNWIND_HINT_REGS offset=\offset
145
	.endm
L
Linus Torvalds 已提交
146

147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165
	.macro POP_EXTRA_REGS
	popq %r15
	popq %r14
	popq %r13
	popq %r12
	popq %rbp
	popq %rbx
	.endm

	.macro POP_C_REGS
	popq %r11
	popq %r10
	popq %r9
	popq %r8
	popq %rax
	popq %rcx
	popq %rdx
	popq %rsi
	popq %rdi
L
Linus Torvalds 已提交
166 167 168 169 170
	.endm

	.macro icebp
	.byte 0xf1
	.endm
171

172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191
/*
 * This is a sneaky trick to help the unwinder find pt_regs on the stack.  The
 * frame pointer is replaced with an encoded pointer to pt_regs.  The encoding
 * is just setting the LSB, which makes it an invalid stack address and is also
 * a signal to the unwinder that it's a pt_regs pointer in disguise.
 *
 * NOTE: This macro must be used *after* SAVE_EXTRA_REGS because it corrupts
 * the original rbp.
 */
.macro ENCODE_FRAME_POINTER ptregs_offset=0
#ifdef CONFIG_FRAME_POINTER
	.if \ptregs_offset
		leaq \ptregs_offset(%rsp), %rbp
	.else
		mov %rsp, %rbp
	.endif
	orq	$0x1, %rbp
#endif
.endm

192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207
#ifdef CONFIG_PAGE_TABLE_ISOLATION

/* PAGE_TABLE_ISOLATION PGDs are 8k.  Flip bit 12 to switch between the two halves: */
#define PTI_SWITCH_MASK (1<<PAGE_SHIFT)

.macro ADJUST_KERNEL_CR3 reg:req
	/* Clear "PAGE_TABLE_ISOLATION bit", point CR3 at kernel pagetables: */
	andq	$(~PTI_SWITCH_MASK), \reg
.endm

.macro ADJUST_USER_CR3 reg:req
	/* Move CR3 up a page to the user page tables: */
	orq	$(PTI_SWITCH_MASK), \reg
.endm

.macro SWITCH_TO_KERNEL_CR3 scratch_reg:req
208
	ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI
209 210 211
	mov	%cr3, \scratch_reg
	ADJUST_KERNEL_CR3 \scratch_reg
	mov	\scratch_reg, %cr3
212
.Lend_\@:
213 214 215
.endm

.macro SWITCH_TO_USER_CR3 scratch_reg:req
216
	ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI
217 218 219
	mov	%cr3, \scratch_reg
	ADJUST_USER_CR3 \scratch_reg
	mov	\scratch_reg, %cr3
220
.Lend_\@:
221 222 223
.endm

.macro SAVE_AND_SWITCH_TO_KERNEL_CR3 scratch_reg:req save_reg:req
224
	ALTERNATIVE "jmp .Ldone_\@", "", X86_FEATURE_PTI
225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240
	movq	%cr3, \scratch_reg
	movq	\scratch_reg, \save_reg
	/*
	 * Is the switch bit zero?  This means the address is
	 * up in real PAGE_TABLE_ISOLATION patches in a moment.
	 */
	testq	$(PTI_SWITCH_MASK), \scratch_reg
	jz	.Ldone_\@

	ADJUST_KERNEL_CR3 \scratch_reg
	movq	\scratch_reg, %cr3

.Ldone_\@:
.endm

.macro RESTORE_CR3 save_reg:req
241
	ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI
242 243 244 245 246
	/*
	 * The CR3 write could be avoided when not changing its value,
	 * but would require a CR3 read *and* a scratch register.
	 */
	movq	\save_reg, %cr3
247
.Lend_\@:
248 249 250 251 252 253 254 255 256 257 258 259 260 261 262
.endm

#else /* CONFIG_PAGE_TABLE_ISOLATION=n: */

.macro SWITCH_TO_KERNEL_CR3 scratch_reg:req
.endm
.macro SWITCH_TO_USER_CR3 scratch_reg:req
.endm
.macro SAVE_AND_SWITCH_TO_KERNEL_CR3 scratch_reg:req save_reg:req
.endm
.macro RESTORE_CR3 save_reg:req
.endm

#endif

263 264
#endif /* CONFIG_X86_64 */

265 266 267 268 269 270 271 272 273 274 275 276 277
/*
 * This does 'call enter_from_user_mode' unless we can avoid it based on
 * kernel config or using the static jump infrastructure.
 */
.macro CALL_enter_from_user_mode
#ifdef CONFIG_CONTEXT_TRACKING
#ifdef HAVE_JUMP_LABEL
	STATIC_JUMP_IF_FALSE .Lafter_call_\@, context_tracking_enabled, def=0
#endif
	call enter_from_user_mode
.Lafter_call_\@:
#endif
.endm