提交 57f4959b 编写于 作者: J James Morse 提交者: Catalin Marinas

arm64: kernel: Add support for User Access Override

'User Access Override' is a new ARMv8.2 feature which allows the
unprivileged load and store instructions to be overridden to behave in
the normal way.

This patch converts {get,put}_user() and friends to use ldtr*/sttr*
instructions - so that they can only access EL0 memory, then enables
UAO when fs==KERNEL_DS so that these functions can access kernel memory.

This allows user space's read/write permissions to be checked against the
page tables, instead of testing addr<USER_DS, then using the kernel's
read/write permissions.
Signed-off-by: NJames Morse <james.morse@arm.com>
[catalin.marinas@arm.com: move uao_thread_switch() above dsb()]
Signed-off-by: NCatalin Marinas <catalin.marinas@arm.com>
上级 406e3087
...@@ -756,6 +756,27 @@ config ARM64_LSE_ATOMICS ...@@ -756,6 +756,27 @@ config ARM64_LSE_ATOMICS
endmenu endmenu
config ARM64_UAO
bool "Enable support for User Access Override (UAO)"
default y
help
User Access Override (UAO; part of the ARMv8.2 Extensions)
causes the 'unprivileged' variant of the load/store instructions to
be overriden to be privileged.
This option changes get_user() and friends to use the 'unprivileged'
variant of the load/store instructions. This ensures that user-space
really did have access to the supplied memory. When addr_limit is
set to kernel memory the UAO bit will be set, allowing privileged
access to kernel memory.
Choosing this option will cause copy_to_user() et al to use user-space
memory permissions.
The feature is detected at runtime, the kernel will use the
regular load/store instructions if the cpu does not implement the
feature.
endmenu endmenu
menu "Boot options" menu "Boot options"
......
#ifndef __ASM_ALTERNATIVE_H #ifndef __ASM_ALTERNATIVE_H
#define __ASM_ALTERNATIVE_H #define __ASM_ALTERNATIVE_H
#include <asm/cpufeature.h>
#ifndef __ASSEMBLY__ #ifndef __ASSEMBLY__
#include <linux/init.h> #include <linux/init.h>
...@@ -63,6 +65,8 @@ void apply_alternatives(void *start, size_t length); ...@@ -63,6 +65,8 @@ void apply_alternatives(void *start, size_t length);
#else #else
#include <asm/assembler.h>
.macro altinstruction_entry orig_offset alt_offset feature orig_len alt_len .macro altinstruction_entry orig_offset alt_offset feature orig_len alt_len
.word \orig_offset - . .word \orig_offset - .
.word \alt_offset - . .word \alt_offset - .
...@@ -136,6 +140,74 @@ void apply_alternatives(void *start, size_t length); ...@@ -136,6 +140,74 @@ void apply_alternatives(void *start, size_t length);
alternative_insn insn1, insn2, cap, IS_ENABLED(cfg) alternative_insn insn1, insn2, cap, IS_ENABLED(cfg)
/*
* Generate the assembly for UAO alternatives with exception table entries.
* This is complicated as there is no post-increment or pair versions of the
* unprivileged instructions, and USER() only works for single instructions.
*/
#ifdef CONFIG_ARM64_UAO
.macro uao_ldp l, reg1, reg2, addr, post_inc
alternative_if_not ARM64_HAS_UAO
8888: ldp \reg1, \reg2, [\addr], \post_inc;
8889: nop;
nop;
alternative_else
ldtr \reg1, [\addr];
ldtr \reg2, [\addr, #8];
add \addr, \addr, \post_inc;
alternative_endif
.section __ex_table,"a";
.align 3;
.quad 8888b,\l;
.quad 8889b,\l;
.previous;
.endm
.macro uao_stp l, reg1, reg2, addr, post_inc
alternative_if_not ARM64_HAS_UAO
8888: stp \reg1, \reg2, [\addr], \post_inc;
8889: nop;
nop;
alternative_else
sttr \reg1, [\addr];
sttr \reg2, [\addr, #8];
add \addr, \addr, \post_inc;
alternative_endif
.section __ex_table,"a";
.align 3;
.quad 8888b,\l;
.quad 8889b,\l;
.previous
.endm
.macro uao_user_alternative l, inst, alt_inst, reg, addr, post_inc
alternative_if_not ARM64_HAS_UAO
8888: \inst \reg, [\addr], \post_inc;
nop;
alternative_else
\alt_inst \reg, [\addr];
add \addr, \addr, \post_inc;
alternative_endif
.section __ex_table,"a";
.align 3;
.quad 8888b,\l;
.previous
.endm
#else
.macro uao_ldp l, reg1, reg2, addr, post_inc
USER(\l, ldp \reg1, \reg2, [\addr], \post_inc)
.endm
.macro uao_stp l, reg1, reg2, addr, post_inc
USER(\l, stp \reg1, \reg2, [\addr], \post_inc)
.endm
.macro uao_user_alternative l, inst, alt_inst, reg, addr, post_inc
USER(\l, \inst \reg, [\addr], \post_inc)
.endm
#endif
#endif /* __ASSEMBLY__ */ #endif /* __ASSEMBLY__ */
/* /*
......
...@@ -31,8 +31,9 @@ ...@@ -31,8 +31,9 @@
#define ARM64_WORKAROUND_CAVIUM_23154 6 #define ARM64_WORKAROUND_CAVIUM_23154 6
#define ARM64_WORKAROUND_834220 7 #define ARM64_WORKAROUND_834220 7
#define ARM64_HAS_NO_HW_PREFETCH 8 #define ARM64_HAS_NO_HW_PREFETCH 8
#define ARM64_HAS_UAO 9
#define ARM64_NCAPS 9 #define ARM64_NCAPS 10
#ifndef __ASSEMBLY__ #ifndef __ASSEMBLY__
......
...@@ -191,5 +191,6 @@ static inline void spin_lock_prefetch(const void *ptr) ...@@ -191,5 +191,6 @@ static inline void spin_lock_prefetch(const void *ptr)
#endif #endif
void cpu_enable_pan(void *__unused); void cpu_enable_pan(void *__unused);
void cpu_enable_uao(void *__unused);
#endif /* __ASM_PROCESSOR_H */ #endif /* __ASM_PROCESSOR_H */
...@@ -79,9 +79,12 @@ ...@@ -79,9 +79,12 @@
#define SYS_DCZID_EL0 sys_reg(3, 3, 0, 0, 7) #define SYS_DCZID_EL0 sys_reg(3, 3, 0, 0, 7)
#define REG_PSTATE_PAN_IMM sys_reg(0, 0, 4, 0, 4) #define REG_PSTATE_PAN_IMM sys_reg(0, 0, 4, 0, 4)
#define REG_PSTATE_UAO_IMM sys_reg(0, 0, 4, 0, 3)
#define SET_PSTATE_PAN(x) __inst_arm(0xd5000000 | REG_PSTATE_PAN_IMM |\ #define SET_PSTATE_PAN(x) __inst_arm(0xd5000000 | REG_PSTATE_PAN_IMM |\
(!!x)<<8 | 0x1f) (!!x)<<8 | 0x1f)
#define SET_PSTATE_UAO(x) __inst_arm(0xd5000000 | REG_PSTATE_UAO_IMM |\
(!!x)<<8 | 0x1f)
/* SCTLR_EL1 */ /* SCTLR_EL1 */
#define SCTLR_EL1_CP15BEN (0x1 << 5) #define SCTLR_EL1_CP15BEN (0x1 << 5)
......
...@@ -85,6 +85,12 @@ static inline struct thread_info *current_thread_info(void) ...@@ -85,6 +85,12 @@ static inline struct thread_info *current_thread_info(void)
return (struct thread_info *)sp_el0; return (struct thread_info *)sp_el0;
} }
/* Access struct thread_info of another thread */
static inline struct thread_info *get_thread_info(unsigned long thread_stack)
{
return (struct thread_info *)(thread_stack & ~(THREAD_SIZE - 1));
}
#define thread_saved_pc(tsk) \ #define thread_saved_pc(tsk) \
((unsigned long)(tsk->thread.cpu_context.pc)) ((unsigned long)(tsk->thread.cpu_context.pc))
#define thread_saved_sp(tsk) \ #define thread_saved_sp(tsk) \
......
...@@ -64,6 +64,16 @@ extern int fixup_exception(struct pt_regs *regs); ...@@ -64,6 +64,16 @@ extern int fixup_exception(struct pt_regs *regs);
static inline void set_fs(mm_segment_t fs) static inline void set_fs(mm_segment_t fs)
{ {
current_thread_info()->addr_limit = fs; current_thread_info()->addr_limit = fs;
/*
* Enable/disable UAO so that copy_to_user() etc can access
* kernel memory with the unprivileged instructions.
*/
if (IS_ENABLED(CONFIG_ARM64_UAO) && fs == KERNEL_DS)
asm(ALTERNATIVE("nop", SET_PSTATE_UAO(1), ARM64_HAS_UAO));
else
asm(ALTERNATIVE("nop", SET_PSTATE_UAO(0), ARM64_HAS_UAO,
CONFIG_ARM64_UAO));
} }
#define segment_eq(a, b) ((a) == (b)) #define segment_eq(a, b) ((a) == (b))
...@@ -113,9 +123,10 @@ static inline void set_fs(mm_segment_t fs) ...@@ -113,9 +123,10 @@ static inline void set_fs(mm_segment_t fs)
* The "__xxx_error" versions set the third argument to -EFAULT if an error * The "__xxx_error" versions set the third argument to -EFAULT if an error
* occurs, and leave it unchanged on success. * occurs, and leave it unchanged on success.
*/ */
#define __get_user_asm(instr, reg, x, addr, err) \ #define __get_user_asm(instr, alt_instr, reg, x, addr, err, feature) \
asm volatile( \ asm volatile( \
"1: " instr " " reg "1, [%2]\n" \ "1:"ALTERNATIVE(instr " " reg "1, [%2]\n", \
alt_instr " " reg "1, [%2]\n", feature) \
"2:\n" \ "2:\n" \
" .section .fixup, \"ax\"\n" \ " .section .fixup, \"ax\"\n" \
" .align 2\n" \ " .align 2\n" \
...@@ -138,16 +149,20 @@ do { \ ...@@ -138,16 +149,20 @@ do { \
CONFIG_ARM64_PAN)); \ CONFIG_ARM64_PAN)); \
switch (sizeof(*(ptr))) { \ switch (sizeof(*(ptr))) { \
case 1: \ case 1: \
__get_user_asm("ldrb", "%w", __gu_val, (ptr), (err)); \ __get_user_asm("ldrb", "ldtrb", "%w", __gu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
case 2: \ case 2: \
__get_user_asm("ldrh", "%w", __gu_val, (ptr), (err)); \ __get_user_asm("ldrh", "ldtrh", "%w", __gu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
case 4: \ case 4: \
__get_user_asm("ldr", "%w", __gu_val, (ptr), (err)); \ __get_user_asm("ldr", "ldtr", "%w", __gu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
case 8: \ case 8: \
__get_user_asm("ldr", "%", __gu_val, (ptr), (err)); \ __get_user_asm("ldr", "ldtr", "%", __gu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
default: \ default: \
BUILD_BUG(); \ BUILD_BUG(); \
...@@ -181,9 +196,10 @@ do { \ ...@@ -181,9 +196,10 @@ do { \
((x) = 0, -EFAULT); \ ((x) = 0, -EFAULT); \
}) })
#define __put_user_asm(instr, reg, x, addr, err) \ #define __put_user_asm(instr, alt_instr, reg, x, addr, err, feature) \
asm volatile( \ asm volatile( \
"1: " instr " " reg "1, [%2]\n" \ "1:"ALTERNATIVE(instr " " reg "1, [%2]\n", \
alt_instr " " reg "1, [%2]\n", feature) \
"2:\n" \ "2:\n" \
" .section .fixup,\"ax\"\n" \ " .section .fixup,\"ax\"\n" \
" .align 2\n" \ " .align 2\n" \
...@@ -205,16 +221,20 @@ do { \ ...@@ -205,16 +221,20 @@ do { \
CONFIG_ARM64_PAN)); \ CONFIG_ARM64_PAN)); \
switch (sizeof(*(ptr))) { \ switch (sizeof(*(ptr))) { \
case 1: \ case 1: \
__put_user_asm("strb", "%w", __pu_val, (ptr), (err)); \ __put_user_asm("strb", "sttrb", "%w", __pu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
case 2: \ case 2: \
__put_user_asm("strh", "%w", __pu_val, (ptr), (err)); \ __put_user_asm("strh", "sttrh", "%w", __pu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
case 4: \ case 4: \
__put_user_asm("str", "%w", __pu_val, (ptr), (err)); \ __put_user_asm("str", "sttr", "%w", __pu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
case 8: \ case 8: \
__put_user_asm("str", "%", __pu_val, (ptr), (err)); \ __put_user_asm("str", "sttr", "%", __pu_val, (ptr), \
(err), ARM64_HAS_UAO); \
break; \ break; \
default: \ default: \
BUILD_BUG(); \ BUILD_BUG(); \
......
...@@ -45,6 +45,7 @@ ...@@ -45,6 +45,7 @@
#define PSR_A_BIT 0x00000100 #define PSR_A_BIT 0x00000100
#define PSR_D_BIT 0x00000200 #define PSR_D_BIT 0x00000200
#define PSR_PAN_BIT 0x00400000 #define PSR_PAN_BIT 0x00400000
#define PSR_UAO_BIT 0x00800000
#define PSR_Q_BIT 0x08000000 #define PSR_Q_BIT 0x08000000
#define PSR_V_BIT 0x10000000 #define PSR_V_BIT 0x10000000
#define PSR_C_BIT 0x20000000 #define PSR_C_BIT 0x20000000
......
...@@ -677,6 +677,17 @@ static const struct arm64_cpu_capabilities arm64_features[] = { ...@@ -677,6 +677,17 @@ static const struct arm64_cpu_capabilities arm64_features[] = {
.capability = ARM64_HAS_NO_HW_PREFETCH, .capability = ARM64_HAS_NO_HW_PREFETCH,
.matches = has_no_hw_prefetch, .matches = has_no_hw_prefetch,
}, },
#ifdef CONFIG_ARM64_UAO
{
.desc = "User Access Override",
.capability = ARM64_HAS_UAO,
.matches = has_cpuid_feature,
.sys_reg = SYS_ID_AA64MMFR2_EL1,
.field_pos = ID_AA64MMFR2_UAO_SHIFT,
.min_field_value = 1,
.enable = cpu_enable_uao,
},
#endif /* CONFIG_ARM64_UAO */
{}, {},
}; };
......
...@@ -46,6 +46,7 @@ ...@@ -46,6 +46,7 @@
#include <linux/notifier.h> #include <linux/notifier.h>
#include <trace/events/power.h> #include <trace/events/power.h>
#include <asm/alternative.h>
#include <asm/compat.h> #include <asm/compat.h>
#include <asm/cacheflush.h> #include <asm/cacheflush.h>
#include <asm/fpsimd.h> #include <asm/fpsimd.h>
...@@ -280,6 +281,9 @@ int copy_thread(unsigned long clone_flags, unsigned long stack_start, ...@@ -280,6 +281,9 @@ int copy_thread(unsigned long clone_flags, unsigned long stack_start,
} else { } else {
memset(childregs, 0, sizeof(struct pt_regs)); memset(childregs, 0, sizeof(struct pt_regs));
childregs->pstate = PSR_MODE_EL1h; childregs->pstate = PSR_MODE_EL1h;
if (IS_ENABLED(CONFIG_ARM64_UAO) &&
cpus_have_cap(ARM64_HAS_UAO))
childregs->pstate |= PSR_UAO_BIT;
p->thread.cpu_context.x19 = stack_start; p->thread.cpu_context.x19 = stack_start;
p->thread.cpu_context.x20 = stk_sz; p->thread.cpu_context.x20 = stk_sz;
} }
...@@ -308,6 +312,20 @@ static void tls_thread_switch(struct task_struct *next) ...@@ -308,6 +312,20 @@ static void tls_thread_switch(struct task_struct *next)
: : "r" (tpidr), "r" (tpidrro)); : : "r" (tpidr), "r" (tpidrro));
} }
/* Restore the UAO state depending on next's addr_limit */
static void uao_thread_switch(struct task_struct *next)
{
unsigned long next_sp = next->thread.cpu_context.sp;
if (IS_ENABLED(CONFIG_ARM64_UAO) &&
get_thread_info(next_sp)->addr_limit == KERNEL_DS)
asm(ALTERNATIVE("nop", SET_PSTATE_UAO(1), ARM64_HAS_UAO,
CONFIG_ARM64_UAO));
else
asm(ALTERNATIVE("nop", SET_PSTATE_UAO(0), ARM64_HAS_UAO,
CONFIG_ARM64_UAO));
}
/* /*
* Thread switching. * Thread switching.
*/ */
...@@ -320,6 +338,7 @@ struct task_struct *__switch_to(struct task_struct *prev, ...@@ -320,6 +338,7 @@ struct task_struct *__switch_to(struct task_struct *prev,
tls_thread_switch(next); tls_thread_switch(next);
hw_breakpoint_thread_switch(next); hw_breakpoint_thread_switch(next);
contextidr_thread_switch(next); contextidr_thread_switch(next);
uao_thread_switch(next);
/* /*
* Complete any pending TLB or cache maintenance on this CPU in case * Complete any pending TLB or cache maintenance on this CPU in case
......
...@@ -39,20 +39,20 @@ ALTERNATIVE("nop", __stringify(SET_PSTATE_PAN(0)), ARM64_HAS_PAN, \ ...@@ -39,20 +39,20 @@ ALTERNATIVE("nop", __stringify(SET_PSTATE_PAN(0)), ARM64_HAS_PAN, \
subs x1, x1, #8 subs x1, x1, #8
b.mi 2f b.mi 2f
1: 1:
USER(9f, str xzr, [x0], #8 ) uao_user_alternative 9f, str, sttr, xzr, x0, 8
subs x1, x1, #8 subs x1, x1, #8
b.pl 1b b.pl 1b
2: adds x1, x1, #4 2: adds x1, x1, #4
b.mi 3f b.mi 3f
USER(9f, str wzr, [x0], #4 ) uao_user_alternative 9f, str, sttr, wzr, x0, 4
sub x1, x1, #4 sub x1, x1, #4
3: adds x1, x1, #2 3: adds x1, x1, #2
b.mi 4f b.mi 4f
USER(9f, strh wzr, [x0], #2 ) uao_user_alternative 9f, strh, sttrh, wzr, x0, 2
sub x1, x1, #2 sub x1, x1, #2
4: adds x1, x1, #1 4: adds x1, x1, #1
b.mi 5f b.mi 5f
USER(9f, strb wzr, [x0] ) uao_user_alternative 9f, strb, sttrb, wzr, x0, 0
5: mov x0, #0 5: mov x0, #0
ALTERNATIVE("nop", __stringify(SET_PSTATE_PAN(1)), ARM64_HAS_PAN, \ ALTERNATIVE("nop", __stringify(SET_PSTATE_PAN(1)), ARM64_HAS_PAN, \
CONFIG_ARM64_PAN) CONFIG_ARM64_PAN)
......
...@@ -34,7 +34,7 @@ ...@@ -34,7 +34,7 @@
*/ */
.macro ldrb1 ptr, regB, val .macro ldrb1 ptr, regB, val
USER(9998f, ldrb \ptr, [\regB], \val) uao_user_alternative 9998f, ldrb, ldtrb, \ptr, \regB, \val
.endm .endm
.macro strb1 ptr, regB, val .macro strb1 ptr, regB, val
...@@ -42,7 +42,7 @@ ...@@ -42,7 +42,7 @@
.endm .endm
.macro ldrh1 ptr, regB, val .macro ldrh1 ptr, regB, val
USER(9998f, ldrh \ptr, [\regB], \val) uao_user_alternative 9998f, ldrh, ldtrh, \ptr, \regB, \val
.endm .endm
.macro strh1 ptr, regB, val .macro strh1 ptr, regB, val
...@@ -50,7 +50,7 @@ ...@@ -50,7 +50,7 @@
.endm .endm
.macro ldr1 ptr, regB, val .macro ldr1 ptr, regB, val
USER(9998f, ldr \ptr, [\regB], \val) uao_user_alternative 9998f, ldr, ldtr, \ptr, \regB, \val
.endm .endm
.macro str1 ptr, regB, val .macro str1 ptr, regB, val
...@@ -58,7 +58,7 @@ ...@@ -58,7 +58,7 @@
.endm .endm
.macro ldp1 ptr, regB, regC, val .macro ldp1 ptr, regB, regC, val
USER(9998f, ldp \ptr, \regB, [\regC], \val) uao_ldp 9998f, \ptr, \regB, \regC, \val
.endm .endm
.macro stp1 ptr, regB, regC, val .macro stp1 ptr, regB, regC, val
......
...@@ -35,35 +35,35 @@ ...@@ -35,35 +35,35 @@
* x0 - bytes not copied * x0 - bytes not copied
*/ */
.macro ldrb1 ptr, regB, val .macro ldrb1 ptr, regB, val
USER(9998f, ldrb \ptr, [\regB], \val) uao_user_alternative 9998f, ldrb, ldtrb, \ptr, \regB, \val
.endm .endm
.macro strb1 ptr, regB, val .macro strb1 ptr, regB, val
USER(9998f, strb \ptr, [\regB], \val) uao_user_alternative 9998f, strb, sttrb, \ptr, \regB, \val
.endm .endm
.macro ldrh1 ptr, regB, val .macro ldrh1 ptr, regB, val
USER(9998f, ldrh \ptr, [\regB], \val) uao_user_alternative 9998f, ldrh, ldtrh, \ptr, \regB, \val
.endm .endm
.macro strh1 ptr, regB, val .macro strh1 ptr, regB, val
USER(9998f, strh \ptr, [\regB], \val) uao_user_alternative 9998f, strh, sttrh, \ptr, \regB, \val
.endm .endm
.macro ldr1 ptr, regB, val .macro ldr1 ptr, regB, val
USER(9998f, ldr \ptr, [\regB], \val) uao_user_alternative 9998f, ldr, ldtr, \ptr, \regB, \val
.endm .endm
.macro str1 ptr, regB, val .macro str1 ptr, regB, val
USER(9998f, str \ptr, [\regB], \val) uao_user_alternative 9998f, str, sttr, \ptr, \regB, \val
.endm .endm
.macro ldp1 ptr, regB, regC, val .macro ldp1 ptr, regB, regC, val
USER(9998f, ldp \ptr, \regB, [\regC], \val) uao_ldp 9998f, \ptr, \regB, \regC, \val
.endm .endm
.macro stp1 ptr, regB, regC, val .macro stp1 ptr, regB, regC, val
USER(9998f, stp \ptr, \regB, [\regC], \val) uao_stp 9998f, \ptr, \regB, \regC, \val
.endm .endm
end .req x5 end .req x5
......
...@@ -37,7 +37,7 @@ ...@@ -37,7 +37,7 @@
.endm .endm
.macro strb1 ptr, regB, val .macro strb1 ptr, regB, val
USER(9998f, strb \ptr, [\regB], \val) uao_user_alternative 9998f, strb, sttrb, \ptr, \regB, \val
.endm .endm
.macro ldrh1 ptr, regB, val .macro ldrh1 ptr, regB, val
...@@ -45,7 +45,7 @@ ...@@ -45,7 +45,7 @@
.endm .endm
.macro strh1 ptr, regB, val .macro strh1 ptr, regB, val
USER(9998f, strh \ptr, [\regB], \val) uao_user_alternative 9998f, strh, sttrh, \ptr, \regB, \val
.endm .endm
.macro ldr1 ptr, regB, val .macro ldr1 ptr, regB, val
...@@ -53,7 +53,7 @@ ...@@ -53,7 +53,7 @@
.endm .endm
.macro str1 ptr, regB, val .macro str1 ptr, regB, val
USER(9998f, str \ptr, [\regB], \val) uao_user_alternative 9998f, str, sttr, \ptr, \regB, \val
.endm .endm
.macro ldp1 ptr, regB, regC, val .macro ldp1 ptr, regB, regC, val
...@@ -61,7 +61,7 @@ ...@@ -61,7 +61,7 @@
.endm .endm
.macro stp1 ptr, regB, regC, val .macro stp1 ptr, regB, regC, val
USER(9998f, stp \ptr, \regB, [\regC], \val) uao_stp 9998f, \ptr, \regB, \regC, \val
.endm .endm
end .req x5 end .req x5
......
...@@ -192,6 +192,14 @@ static int __do_page_fault(struct mm_struct *mm, unsigned long addr, ...@@ -192,6 +192,14 @@ static int __do_page_fault(struct mm_struct *mm, unsigned long addr,
return fault; return fault;
} }
static inline int permission_fault(unsigned int esr)
{
unsigned int ec = (esr & ESR_ELx_EC_MASK) >> ESR_ELx_EC_SHIFT;
unsigned int fsc_type = esr & ESR_ELx_FSC_TYPE;
return (ec == ESR_ELx_EC_DABT_CUR && fsc_type == ESR_ELx_FSC_PERM);
}
static int __kprobes do_page_fault(unsigned long addr, unsigned int esr, static int __kprobes do_page_fault(unsigned long addr, unsigned int esr,
struct pt_regs *regs) struct pt_regs *regs)
{ {
...@@ -225,12 +233,10 @@ static int __kprobes do_page_fault(unsigned long addr, unsigned int esr, ...@@ -225,12 +233,10 @@ static int __kprobes do_page_fault(unsigned long addr, unsigned int esr,
mm_flags |= FAULT_FLAG_WRITE; mm_flags |= FAULT_FLAG_WRITE;
} }
/* if (permission_fault(esr) && (addr < USER_DS)) {
* PAN bit set implies the fault happened in kernel space, but not if (!search_exception_tables(regs->pc))
* in the arch's user access functions. panic("Accessing user space memory outside uaccess.h routines");
*/ }
if (IS_ENABLED(CONFIG_ARM64_PAN) && (regs->pstate & PSR_PAN_BIT))
goto no_context;
/* /*
* As per x86, we may deadlock here. However, since the kernel only * As per x86, we may deadlock here. However, since the kernel only
...@@ -561,3 +567,16 @@ void cpu_enable_pan(void *__unused) ...@@ -561,3 +567,16 @@ void cpu_enable_pan(void *__unused)
config_sctlr_el1(SCTLR_EL1_SPAN, 0); config_sctlr_el1(SCTLR_EL1_SPAN, 0);
} }
#endif /* CONFIG_ARM64_PAN */ #endif /* CONFIG_ARM64_PAN */
#ifdef CONFIG_ARM64_UAO
/*
* Kernel threads have fs=KERNEL_DS by default, and don't need to call
* set_fs(), devtmpfs in particular relies on this behaviour.
* We need to enable the feature at runtime (instead of adding it to
* PSR_MODE_EL1h) as the feature may not be implemented by the cpu.
*/
void cpu_enable_uao(void *__unused)
{
asm(SET_PSTATE_UAO(1));
}
#endif /* CONFIG_ARM64_UAO */
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册