hyp-entry.S 5.6 KB
Newer Older
M
Marc Zyngier 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
/*
 * Copyright (C) 2015 - ARM Ltd
 * Author: Marc Zyngier <marc.zyngier@arm.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program.  If not, see <http://www.gnu.org/licenses/>.
 */

18
#include <linux/arm-smccc.h>
M
Marc Zyngier 已提交
19 20 21 22 23 24 25 26 27 28 29 30
#include <linux/linkage.h>

#include <asm/alternative.h>
#include <asm/assembler.h>
#include <asm/cpufeature.h>
#include <asm/kvm_arm.h>
#include <asm/kvm_asm.h>
#include <asm/kvm_mmu.h>

	.text
	.pushsection	.hyp.text, "ax"

31 32 33 34 35
.macro do_el2_call
	/*
	 * Shuffle the parameters before calling the function
	 * pointed to in x0. Assumes parameters in x[1,2,3].
	 */
36
	str	lr, [sp, #-16]!
37 38 39 40 41
	mov	lr, x0
	mov	x0, x1
	mov	x1, x2
	mov	x2, x3
	blr	lr
42
	ldr	lr, [sp], #16
43 44 45 46 47 48 49 50 51 52 53 54 55
.endm

ENTRY(__vhe_hyp_call)
	do_el2_call
	/*
	 * We used to rely on having an exception return to get
	 * an implicit isb. In the E2H case, we don't have it anymore.
	 * rather than changing all the leaf functions, just do it here
	 * before returning to the rest of the kernel.
	 */
	isb
	ret
ENDPROC(__vhe_hyp_call)
56

M
Marc Zyngier 已提交
57 58
el1_sync:				// Guest trapped into EL2

59 60
	mrs	x0, esr_el2
	lsr	x0, x0, #ESR_ELx_EC_SHIFT
61
	cmp	x0, #ESR_ELx_EC_HVC64
62
	ccmp	x0, #ESR_ELx_EC_HVC32, #4, ne
M
Marc Zyngier 已提交
63 64
	b.ne	el1_trap

65 66
	mrs	x1, vttbr_el2		// If vttbr is valid, the guest
	cbnz	x1, el1_hvc_guest	// called HVC
M
Marc Zyngier 已提交
67 68

	/* Here, we're pretty sure the host called HVC. */
69
	ldp	x0, x1, [sp], #16
M
Marc Zyngier 已提交
70

71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88
	/* Check for a stub HVC call */
	cmp	x0, #HVC_STUB_HCALL_NR
	b.hs	1f

	/*
	 * Compute the idmap address of __kvm_handle_stub_hvc and
	 * jump there. Since we use kimage_voffset, do not use the
	 * HYP VA for __kvm_handle_stub_hvc, but the kernel VA instead
	 * (by loading it from the constant pool).
	 *
	 * Preserve x0-x4, which may contain stub parameters.
	 */
	ldr	x5, =__kvm_handle_stub_hvc
	ldr_l	x6, kimage_voffset

	/* x5 = __pa(x5) */
	sub	x5, x5, x6
	br	x5
M
Marc Zyngier 已提交
89

90
1:
M
Marc Zyngier 已提交
91
	/*
92
	 * Perform the EL2 call
M
Marc Zyngier 已提交
93 94
	 */
	kern_hyp_va	x0
95
	do_el2_call
M
Marc Zyngier 已提交
96

97
	eret
M
Marc Zyngier 已提交
98

99 100 101 102 103 104 105 106 107 108 109 110 111 112
el1_hvc_guest:
	/*
	 * Fastest possible path for ARM_SMCCC_ARCH_WORKAROUND_1.
	 * The workaround has already been applied on the host,
	 * so let's quickly get back to the guest. We don't bother
	 * restoring x1, as it can be clobbered anyway.
	 */
	ldr	x1, [sp]				// Guest's x0
	eor	w1, w1, #ARM_SMCCC_ARCH_WORKAROUND_1
	cbnz	w1, el1_trap
	mov	x0, x1
	add	sp, sp, #16
	eret

M
Marc Zyngier 已提交
113
el1_trap:
114 115 116 117
	get_vcpu_ptr	x1, x0

	mrs		x0, esr_el2
	lsr		x0, x0, #ESR_ELx_EC_SHIFT
M
Marc Zyngier 已提交
118
	/*
119
	 * x0: ESR_EC
120
	 * x1: vcpu pointer
M
Marc Zyngier 已提交
121 122
	 */

123 124 125 126 127 128 129
	/*
	 * We trap the first access to the FP/SIMD to save the host context
	 * and restore the guest context lazily.
	 * If FP/SIMD is not implemented, handle the trap and inject an
	 * undefined instruction exception to the guest.
	 */
alternative_if_not ARM64_HAS_NO_FPSIMD
130
	cmp	x0, #ESR_ELx_EC_FP_ASIMD
M
Marc Zyngier 已提交
131
	b.eq	__fpsimd_guest_restore
132
alternative_else_nop_endif
M
Marc Zyngier 已提交
133

134
	mov	x0, #ARM_EXCEPTION_TRAP
M
Marc Zyngier 已提交
135 136 137
	b	__guest_exit

el1_irq:
138
	get_vcpu_ptr	x1, x0
139
	mov	x0, #ARM_EXCEPTION_IRQ
M
Marc Zyngier 已提交
140 141
	b	__guest_exit

142
el1_error:
143
	get_vcpu_ptr	x1, x0
144 145 146
	mov	x0, #ARM_EXCEPTION_EL1_SERROR
	b	__guest_exit

147
el2_error:
148 149
	ldp	x0, x1, [sp], #16

150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170
	/*
	 * Only two possibilities:
	 * 1) Either we come from the exit path, having just unmasked
	 *    PSTATE.A: change the return code to an EL2 fault, and
	 *    carry on, as we're already in a sane state to handle it.
	 * 2) Or we come from anywhere else, and that's a bug: we panic.
	 *
	 * For (1), x0 contains the original return code and x1 doesn't
	 * contain anything meaningful at that stage. We can reuse them
	 * as temp registers.
	 * For (2), who cares?
	 */
	mrs	x0, elr_el2
	adr	x1, abort_guest_exit_start
	cmp	x0, x1
	adr	x1, abort_guest_exit_end
	ccmp	x0, x1, #4, ne
	b.ne	__hyp_panic
	mov	x0, #(1 << ARM_EXIT_WITH_SERROR_BIT)
	eret

M
Marc Zyngier 已提交
171 172 173 174 175 176 177 178 179
ENTRY(__hyp_do_panic)
	mov	lr, #(PSR_F_BIT | PSR_I_BIT | PSR_A_BIT | PSR_D_BIT |\
		      PSR_MODE_EL1h)
	msr	spsr_el2, lr
	ldr	lr, =panic
	msr	elr_el2, lr
	eret
ENDPROC(__hyp_do_panic)

180
ENTRY(__hyp_panic)
181
	get_host_ctxt x0, x1
182 183 184
	b	hyp_panic
ENDPROC(__hyp_panic)

M
Marc Zyngier 已提交
185
.macro invalid_vector	label, target = __hyp_panic
M
Marc Zyngier 已提交
186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205
	.align	2
\label:
	b \target
ENDPROC(\label)
.endm

	/* None of these should ever happen */
	invalid_vector	el2t_sync_invalid
	invalid_vector	el2t_irq_invalid
	invalid_vector	el2t_fiq_invalid
	invalid_vector	el2t_error_invalid
	invalid_vector	el2h_sync_invalid
	invalid_vector	el2h_irq_invalid
	invalid_vector	el2h_fiq_invalid
	invalid_vector	el1_fiq_invalid

	.ltorg

	.align 11

206 207 208 209 210 211 212 213 214 215 216
.macro valid_vect target
	.align 7
	stp	x0, x1, [sp, #-16]!
	b	\target
.endm

.macro invalid_vect target
	.align 7
	b	\target
.endm

217
ENTRY(__kvm_hyp_vector)
218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236
	invalid_vect	el2t_sync_invalid	// Synchronous EL2t
	invalid_vect	el2t_irq_invalid	// IRQ EL2t
	invalid_vect	el2t_fiq_invalid	// FIQ EL2t
	invalid_vect	el2t_error_invalid	// Error EL2t

	invalid_vect	el2h_sync_invalid	// Synchronous EL2h
	invalid_vect	el2h_irq_invalid	// IRQ EL2h
	invalid_vect	el2h_fiq_invalid	// FIQ EL2h
	valid_vect	el2_error		// Error EL2h

	valid_vect	el1_sync		// Synchronous 64-bit EL1
	valid_vect	el1_irq			// IRQ 64-bit EL1
	invalid_vect	el1_fiq_invalid		// FIQ 64-bit EL1
	valid_vect	el1_error		// Error 64-bit EL1

	valid_vect	el1_sync		// Synchronous 32-bit EL1
	valid_vect	el1_irq			// IRQ 32-bit EL1
	invalid_vect	el1_fiq_invalid		// FIQ 32-bit EL1
	valid_vect	el1_error		// Error 32-bit EL1
237
ENDPROC(__kvm_hyp_vector)