提交 44a65ef0 编写于 作者: A Ard Biesheuvel 提交者: Zheng Zengkai

ARM: 9058/1: cache-v7: refactor v7_invalidate_l1 to avoid clobbering r5/r6

mainline inclusion
from mainline-v5.13-rc1
commit f9e7a99f
category: bugfix
bugzilla: https://gitee.com/openeuler/kernel/issues/I634EK
CVE: NA

Reference: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=f9e7a99fb6b86aa6a00e53b34ee6973840e005aa

--------------------------------

The cache invalidation code in v7_invalidate_l1 can be tweaked to
re-read the associativity from CCSIDR, and keep the way identifier
component in a single register that is assigned in the outer loop. This
way, we need 2 registers less.

Given that the number of sets is typically much larger than the
associativity, rearrange the code so that the outer loop has the fewer
number of iterations, ensuring that the re-read of CCSIDR only occurs a
handful of times in practice.

Fix the whitespace while at it, and update the comment to indicate that
this code is no longer a clone of anything else.
Acked-by: NNicolas Pitre <nico@fluxnic.net>
Signed-off-by: NArd Biesheuvel <ardb@kernel.org>
Signed-off-by: NRussell King <rmk+kernel@armlinux.org.uk>
Signed-off-by: NZhang Jianhua <chris.zjh@huawei.com>
Reviewed-by: NLiao Chang <liaochang1@huawei.com>
Signed-off-by: NZheng Zengkai <zhengzengkai@huawei.com>
上级 1489b74d
...@@ -33,9 +33,8 @@ icache_size: ...@@ -33,9 +33,8 @@ icache_size:
* processor. We fix this by performing an invalidate, rather than a * processor. We fix this by performing an invalidate, rather than a
* clean + invalidate, before jumping into the kernel. * clean + invalidate, before jumping into the kernel.
* *
* This function is cloned from arch/arm/mach-tegra/headsmp.S, and needs * This function needs to be called for both secondary cores startup and
* to be called for both secondary cores startup and primary core resume * primary core resume procedures.
* procedures.
*/ */
ENTRY(v7_invalidate_l1) ENTRY(v7_invalidate_l1)
mov r0, #0 mov r0, #0
...@@ -43,32 +42,32 @@ ENTRY(v7_invalidate_l1) ...@@ -43,32 +42,32 @@ ENTRY(v7_invalidate_l1)
isb isb
mrc p15, 1, r0, c0, c0, 0 @ read cache geometry from CCSIDR mrc p15, 1, r0, c0, c0, 0 @ read cache geometry from CCSIDR
movw r1, #0x7fff movw r3, #0x3ff
and r2, r1, r0, lsr #13 and r3, r3, r0, lsr #3 @ 'Associativity' in CCSIDR[12:3]
clz r1, r3 @ WayShift
mov r2, #1
mov r3, r3, lsl r1 @ NumWays-1 shifted into bits [31:...]
movs r1, r2, lsl r1 @ #1 shifted left by same amount
moveq r1, #1 @ r1 needs value > 0 even if only 1 way
movw r1, #0x3ff and r2, r0, #0x7
add r2, r2, #4 @ SetShift
and r3, r1, r0, lsr #3 @ NumWays - 1 1: movw r4, #0x7fff
add r2, r2, #1 @ NumSets and r0, r4, r0, lsr #13 @ 'NumSets' in CCSIDR[27:13]
and r0, r0, #0x7 2: mov r4, r0, lsl r2 @ NumSet << SetShift
add r0, r0, #4 @ SetShift orr r4, r4, r3 @ Reg = (Temp<<WayShift)|(NumSets<<SetShift)
mcr p15, 0, r4, c7, c6, 2
clz r1, r3 @ WayShift subs r0, r0, #1 @ Set--
add r4, r3, #1 @ NumWays bpl 2b
1: sub r2, r2, #1 @ NumSets-- subs r3, r3, r1 @ Way--
mov r3, r4 @ Temp = NumWays bcc 3f
2: subs r3, r3, #1 @ Temp-- mrc p15, 1, r0, c0, c0, 0 @ re-read cache geometry from CCSIDR
mov r5, r3, lsl r1 b 1b
mov r6, r2, lsl r0 3: dsb st
orr r5, r5, r6 @ Reg = (Temp<<WayShift)|(NumSets<<SetShift) isb
mcr p15, 0, r5, c7, c6, 2 ret lr
bgt 2b
cmp r2, #0
bgt 1b
dsb st
isb
ret lr
ENDPROC(v7_invalidate_l1) ENDPROC(v7_invalidate_l1)
/* /*
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册