提交 4a8d67c4 编写于 作者: A Ard Biesheuvel 提交者: Zheng Zengkai

crypto: arm64/sha512-ce - simplify NEON yield

mainline inclusion
from mainline-v5.12-rc1
commit 5f6cb2e6
category: bugfix
bugzilla: 172149 https://gitee.com/openeuler/kernel/issues/I4CZ7H
CVE: NA

Reference: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=5f6cb2e61768

-----------------------------------------------

Instead of calling into kernel_neon_end() and kernel_neon_begin() (and
potentially into schedule()) from the assembler code when running in
task mode and a reschedule is pending, perform only the preempt count
check in assembler, but simply return early in this case, and let the C
code deal with the consequences.

This reverts commit 6caf7adc.
Signed-off-by: NArd Biesheuvel <ardb@kernel.org>
Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
Signed-off-by: NWei Li <liwei391@huawei.com>
Reviewed-by: NJason Yan <yanaijie@huawei.com>
Reviewed-by: NHanjun Guo <guohanjun@huawei.com>
Signed-off-by: NChen Jun <chenjun102@huawei.com>
Signed-off-by: NZheng Zengkai <zhengzengkai@huawei.com>
上级 180fa19b
...@@ -107,23 +107,17 @@ ...@@ -107,23 +107,17 @@
*/ */
.text .text
SYM_FUNC_START(sha512_ce_transform) SYM_FUNC_START(sha512_ce_transform)
frame_push 3
mov x19, x0
mov x20, x1
mov x21, x2
/* load state */ /* load state */
0: ld1 {v8.2d-v11.2d}, [x19] ld1 {v8.2d-v11.2d}, [x0]
/* load first 4 round constants */ /* load first 4 round constants */
adr_l x3, .Lsha512_rcon adr_l x3, .Lsha512_rcon
ld1 {v20.2d-v23.2d}, [x3], #64 ld1 {v20.2d-v23.2d}, [x3], #64
/* load input */ /* load input */
1: ld1 {v12.2d-v15.2d}, [x20], #64 0: ld1 {v12.2d-v15.2d}, [x1], #64
ld1 {v16.2d-v19.2d}, [x20], #64 ld1 {v16.2d-v19.2d}, [x1], #64
sub w21, w21, #1 sub w2, w2, #1
CPU_LE( rev64 v12.16b, v12.16b ) CPU_LE( rev64 v12.16b, v12.16b )
CPU_LE( rev64 v13.16b, v13.16b ) CPU_LE( rev64 v13.16b, v13.16b )
...@@ -201,19 +195,12 @@ CPU_LE( rev64 v19.16b, v19.16b ) ...@@ -201,19 +195,12 @@ CPU_LE( rev64 v19.16b, v19.16b )
add v10.2d, v10.2d, v2.2d add v10.2d, v10.2d, v2.2d
add v11.2d, v11.2d, v3.2d add v11.2d, v11.2d, v3.2d
cond_yield 3f, x4
/* handled all input blocks? */ /* handled all input blocks? */
cbz w21, 3f cbnz w2, 0b
if_will_cond_yield_neon
st1 {v8.2d-v11.2d}, [x19]
do_cond_yield_neon
b 0b
endif_yield_neon
b 1b
/* store new state */ /* store new state */
3: st1 {v8.2d-v11.2d}, [x19] 3: st1 {v8.2d-v11.2d}, [x0]
frame_pop mov w0, w2
ret ret
SYM_FUNC_END(sha512_ce_transform) SYM_FUNC_END(sha512_ce_transform)
...@@ -26,11 +26,25 @@ MODULE_LICENSE("GPL v2"); ...@@ -26,11 +26,25 @@ MODULE_LICENSE("GPL v2");
MODULE_ALIAS_CRYPTO("sha384"); MODULE_ALIAS_CRYPTO("sha384");
MODULE_ALIAS_CRYPTO("sha512"); MODULE_ALIAS_CRYPTO("sha512");
asmlinkage void sha512_ce_transform(struct sha512_state *sst, u8 const *src, asmlinkage int sha512_ce_transform(struct sha512_state *sst, u8 const *src,
int blocks); int blocks);
asmlinkage void sha512_block_data_order(u64 *digest, u8 const *src, int blocks); asmlinkage void sha512_block_data_order(u64 *digest, u8 const *src, int blocks);
static void __sha512_ce_transform(struct sha512_state *sst, u8 const *src,
int blocks)
{
while (blocks) {
int rem;
kernel_neon_begin();
rem = sha512_ce_transform(sst, src, blocks);
kernel_neon_end();
src += (blocks - rem) * SHA512_BLOCK_SIZE;
blocks = rem;
}
}
static void __sha512_block_data_order(struct sha512_state *sst, u8 const *src, static void __sha512_block_data_order(struct sha512_state *sst, u8 const *src,
int blocks) int blocks)
{ {
...@@ -40,45 +54,30 @@ static void __sha512_block_data_order(struct sha512_state *sst, u8 const *src, ...@@ -40,45 +54,30 @@ static void __sha512_block_data_order(struct sha512_state *sst, u8 const *src,
static int sha512_ce_update(struct shash_desc *desc, const u8 *data, static int sha512_ce_update(struct shash_desc *desc, const u8 *data,
unsigned int len) unsigned int len)
{ {
if (!crypto_simd_usable()) sha512_block_fn *fn = crypto_simd_usable() ? __sha512_ce_transform
return sha512_base_do_update(desc, data, len, : __sha512_block_data_order;
__sha512_block_data_order);
kernel_neon_begin();
sha512_base_do_update(desc, data, len, sha512_ce_transform);
kernel_neon_end();
sha512_base_do_update(desc, data, len, fn);
return 0; return 0;
} }
static int sha512_ce_finup(struct shash_desc *desc, const u8 *data, static int sha512_ce_finup(struct shash_desc *desc, const u8 *data,
unsigned int len, u8 *out) unsigned int len, u8 *out)
{ {
if (!crypto_simd_usable()) { sha512_block_fn *fn = crypto_simd_usable() ? __sha512_ce_transform
if (len) : __sha512_block_data_order;
sha512_base_do_update(desc, data, len,
__sha512_block_data_order);
sha512_base_do_finalize(desc, __sha512_block_data_order);
return sha512_base_finish(desc, out);
}
kernel_neon_begin(); sha512_base_do_update(desc, data, len, fn);
sha512_base_do_update(desc, data, len, sha512_ce_transform); sha512_base_do_finalize(desc, fn);
sha512_base_do_finalize(desc, sha512_ce_transform);
kernel_neon_end();
return sha512_base_finish(desc, out); return sha512_base_finish(desc, out);
} }
static int sha512_ce_final(struct shash_desc *desc, u8 *out) static int sha512_ce_final(struct shash_desc *desc, u8 *out)
{ {
if (!crypto_simd_usable()) { sha512_block_fn *fn = crypto_simd_usable() ? __sha512_ce_transform
sha512_base_do_finalize(desc, __sha512_block_data_order); : __sha512_block_data_order;
return sha512_base_finish(desc, out);
}
kernel_neon_begin(); sha512_base_do_finalize(desc, fn);
sha512_base_do_finalize(desc, sha512_ce_transform);
kernel_neon_end();
return sha512_base_finish(desc, out); return sha512_base_finish(desc, out);
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册