Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenHarmony
Third Party Openssl
提交
e822c756
T
Third Party Openssl
项目概览
OpenHarmony
/
Third Party Openssl
1 年多 前同步成功
通知
10
Star
18
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
T
Third Party Openssl
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
e822c756
编写于
11月 29, 2010
作者:
A
Andy Polyakov
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
s390x assembler pack: adapt for -m31 build, see commentary in Configure
for more details.
上级
300b1d76
变更
9
隐藏空白更改
内联
并排
Showing
9 changed file
with
329 addition
and
140 deletion
+329
-140
Configure
Configure
+16
-2
config
config
+12
-1
crypto/aes/asm/aes-s390x.pl
crypto/aes/asm/aes-s390x.pl
+95
-72
crypto/bn/asm/s390x-mont.pl
crypto/bn/asm/s390x-mont.pl
+74
-25
crypto/modes/asm/ghash-s390x.pl
crypto/modes/asm/ghash-s390x.pl
+26
-3
crypto/rc4/asm/rc4-s390x.pl
crypto/rc4/asm/rc4-s390x.pl
+30
-5
crypto/s390xcpuid.S
crypto/s390xcpuid.S
+3
-0
crypto/sha/asm/sha1-s390x.pl
crypto/sha/asm/sha1-s390x.pl
+33
-13
crypto/sha/asm/sha512-s390x.pl
crypto/sha/asm/sha512-s390x.pl
+40
-19
未找到文件。
Configure
浏览文件 @
e822c756
...
@@ -134,7 +134,7 @@ my $sparcv8_asm=":sparcv8.o:des_enc-sparc.o fcrypt_b.o::::::::::::void";
...
@@ -134,7 +134,7 @@ my $sparcv8_asm=":sparcv8.o:des_enc-sparc.o fcrypt_b.o::::::::::::void";
my $alpha_asm="alphacpuid.o:bn_asm.o alpha-mont.o:::::sha1-alpha.o:::::::ghash-alpha.o:void";
my $alpha_asm="alphacpuid.o:bn_asm.o alpha-mont.o:::::sha1-alpha.o:::::::ghash-alpha.o:void";
my $mips32_asm=":bn-mips.o::aes_cbc.o aes-mips.o:::sha1-mips.o sha256-mips.o:::::::";
my $mips32_asm=":bn-mips.o::aes_cbc.o aes-mips.o:::sha1-mips.o sha256-mips.o:::::::";
my $mips64_asm=":bn-mips.o mips-mont.o::aes_cbc.o aes-mips.o:::sha1-mips.o sha256-mips.o sha512-mips.o:::::::";
my $mips64_asm=":bn-mips.o mips-mont.o::aes_cbc.o aes-mips.o:::sha1-mips.o sha256-mips.o sha512-mips.o:::::::";
my $s390x_asm="s390xcap.o s390xcpuid.o:bn-s390x.o s390x-mont.o::aes_ctr.o aes-s390x.o:::sha1-s390x.o sha256-s390x.o sha512-s390x.o::rc4-s390x.o:::::ghash-s390x.o
:void
";
my $s390x_asm="s390xcap.o s390xcpuid.o:bn-s390x.o s390x-mont.o::aes_ctr.o aes-s390x.o:::sha1-s390x.o sha256-s390x.o sha512-s390x.o::rc4-s390x.o:::::ghash-s390x.o";
my $armv4_asm=":bn_asm.o armv4-mont.o::aes_cbc.o aes-armv4.o:::sha1-armv4-large.o sha256-armv4.o sha512-armv4.o:::::::ghash-armv4.o:void";
my $armv4_asm=":bn_asm.o armv4-mont.o::aes_cbc.o aes-armv4.o:::sha1-armv4-large.o sha256-armv4.o sha512-armv4.o:::::::ghash-armv4.o:void";
my $parisc11_asm="pariscid.o:bn_asm.o parisc-mont.o::aes_core.o aes_cbc.o aes-parisc.o:::sha1-parisc.o sha256-parisc.o sha512-parisc.o::rc4-parisc.o:::::ghash-parisc.o:32";
my $parisc11_asm="pariscid.o:bn_asm.o parisc-mont.o::aes_core.o aes_cbc.o aes-parisc.o:::sha1-parisc.o sha256-parisc.o sha512-parisc.o::rc4-parisc.o:::::ghash-parisc.o:32";
my $parisc20_asm="pariscid.o:pa-risc2W.o parisc-mont.o::aes_core.o aes_cbc.o aes-parisc.o:::sha1-parisc.o sha256-parisc.o sha512-parisc.o::rc4-parisc.o:::::ghash-parisc.o:64";
my $parisc20_asm="pariscid.o:pa-risc2W.o parisc-mont.o::aes_core.o aes_cbc.o aes-parisc.o:::sha1-parisc.o sha256-parisc.o sha512-parisc.o::rc4-parisc.o:::::ghash-parisc.o:64";
...
@@ -356,7 +356,21 @@ my %table=(
...
@@ -356,7 +356,21 @@ my %table=(
"linux-ia64-ecc","ecc:-DL_ENDIAN -DTERMIO -O2 -Wall -no_cpprt::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHUNK DES_INT:${ia64_asm}:dlfcn:linux-shared:-fPIC::.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR)",
"linux-ia64-ecc","ecc:-DL_ENDIAN -DTERMIO -O2 -Wall -no_cpprt::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHUNK DES_INT:${ia64_asm}:dlfcn:linux-shared:-fPIC::.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR)",
"linux-ia64-icc","icc:-DL_ENDIAN -DTERMIO -O2 -Wall -no_cpprt::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHUNK DES_RISC1 DES_INT:${ia64_asm}:dlfcn:linux-shared:-fPIC::.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR)",
"linux-ia64-icc","icc:-DL_ENDIAN -DTERMIO -O2 -Wall -no_cpprt::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHUNK DES_RISC1 DES_INT:${ia64_asm}:dlfcn:linux-shared:-fPIC::.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR)",
"linux-x86_64", "gcc:-m64 -DL_ENDIAN -DTERMIO -O3 -Wall::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHUNK DES_INT DES_UNROLL:${x86_64_asm}:elf:dlfcn:linux-shared:-fPIC:-m64:.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR):::64",
"linux-x86_64", "gcc:-m64 -DL_ENDIAN -DTERMIO -O3 -Wall::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHUNK DES_INT DES_UNROLL:${x86_64_asm}:elf:dlfcn:linux-shared:-fPIC:-m64:.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR):::64",
"linux-s390x", "gcc:-m64 -DB_ENDIAN -DTERMIO -O3 -Wall::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHAR RC4_CHUNK DES_INT DES_UNROLL:${s390x_asm}:dlfcn:linux-shared:-fPIC:-m64:.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR):::64",
"linux64-s390x", "gcc:-m64 -DB_ENDIAN -DTERMIO -O3 -Wall::-D_REENTRANT::-ldl:SIXTY_FOUR_BIT_LONG RC4_CHAR RC4_CHUNK DES_INT DES_UNROLL:${s390x_asm}:64:dlfcn:linux-shared:-fPIC:-m64:.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR):::64",
#### So called "highgprs" target for z/Architecture CPUs
# "Highgprs" is kernel feature first implemented in Linux 2.6.32, see
# /proc/cpuinfo. The idea is to preserve most significant bits of
# general purpose registers not only upon 32-bit process context
# switch, but even on asynchronous signal delivery to such process.
# This makes it possible to deploy 64-bit instructions even in legacy
# application context and achieve better [or should we say adequate]
# performance. The build is binary compatible with linux-generic32,
# and the idea is to be able to install the resulting libcrypto.so
# alongside generic one, e.g. as /lib/highgprs/libcrypto.so.x.y, for
# ldconfig and run-time linker to autodiscover. Unfortunately it
# doesn't work just yet, because of couple of bugs in glibc
# sysdep/s390/dl-procinfo.c affecting ldconfig and ld.so.1...
"linux32-s390x", "gcc:-m31 -Wa,-mzarch -DB_ENDIAN -DTERMIO -O3 -Wall::-D_REENTRANT::-ldl:BN_LLONG RC4_CHAR RC4_CHUNK DES_INT DES_UNROLL:s390xcap.o s390xcpuid.o:bn_asm.o s390x-mont.o::aes_ctr.o aes-s390x.o:::sha1-s390x.o sha256-s390x.o sha512-s390x.o::rc4-s390x.o:::::ghash-s390x.o:31:dlfcn:linux-shared:-fPIC:-m31:.so.\$(SHLIB_MAJOR).\$(SHLIB_MINOR):::/highgprs",
#### SPARC Linux setups
#### SPARC Linux setups
# Ray Miller <ray.miller@computing-services.oxford.ac.uk> has patiently
# Ray Miller <ray.miller@computing-services.oxford.ac.uk> has patiently
# assisted with debugging of following two configs.
# assisted with debugging of following two configs.
...
...
config
浏览文件 @
e822c756
...
@@ -629,7 +629,18 @@ case "$GUESSOS" in
...
@@ -629,7 +629,18 @@ case "$GUESSOS" in
sh
*
-
*
-linux2
)
OUT
=
"linux-generic32"
;
options
=
"
$options
-DL_ENDIAN"
;;
sh
*
-
*
-linux2
)
OUT
=
"linux-generic32"
;
options
=
"
$options
-DL_ENDIAN"
;;
m68k
*
-
*
-linux2
)
OUT
=
"linux-generic32"
;
options
=
"
$options
-DB_ENDIAN"
;;
m68k
*
-
*
-linux2
)
OUT
=
"linux-generic32"
;
options
=
"
$options
-DB_ENDIAN"
;;
s390-
*
-linux2
)
OUT
=
"linux-generic32"
;
options
=
"
$options
-DB_ENDIAN"
;;
s390-
*
-linux2
)
OUT
=
"linux-generic32"
;
options
=
"
$options
-DB_ENDIAN"
;;
s390x-
*
-linux2
)
OUT
=
"linux-s390x"
;;
s390x-
*
-linux2
)
# To be uncommented when glibc bug is fixed, see Configure...
#if egrep -e '^features.* highgprs' /proc/cpuinfo >/dev/null ; then
# echo "WARNING! If you wish to build \"highgprs\" 32-bit library, then you"
# echo " have to invoke './Configure linux32-s390x' *manually*."
# if [ "$TEST" = "false" -a -t -1 ]; then
# echo " You have about 5 seconds to press Ctrl-C to abort."
# (trap "stty `stty -g`" 2 0; stty -icanon min 0 time 50; read waste) <&1
# fi
#fi
OUT
=
"linux64-s390x"
;;
x86_64-
*
-linux
?
)
OUT
=
"linux-x86_64"
;;
x86_64-
*
-linux
?
)
OUT
=
"linux-x86_64"
;;
*
86-
*
-linux2
)
OUT
=
"linux-elf"
*
86-
*
-linux2
)
OUT
=
"linux-elf"
if
[
"
$GCCVER
"
-gt
28
]
;
then
if
[
"
$GCCVER
"
-gt
28
]
;
then
...
...
crypto/aes/asm/aes-s390x.pl
浏览文件 @
e822c756
...
@@ -60,6 +60,26 @@
...
@@ -60,6 +60,26 @@
# maximum, but *on average* it would be as much as ~98%. Meaning that
# maximum, but *on average* it would be as much as ~98%. Meaning that
# worst case is unlike, it's like hitting ravine on plateau.
# worst case is unlike, it's like hitting ravine on plateau.
# November 2010.
#
# Adapt for -m31 build. If kernel supports what's called "highgprs"
# feature on Linux [see /proc/cpuinfo], it's possible to use 64-bit
# instructions and achieve "64-bit" performance even in 31-bit legacy
# application context. The feature is not specific to any particular
# processor, as long as it's "z-CPU". Latter implies that the code
# remains z/Architecture specific. On z990 it was measured to perform
# 2x better than code generated by gcc 4.3.
$flavour
=
shift
;
if
(
$flavour
=~
/3[12]/
)
{
$SIZE_T
=
4
;
$g
=
"";
}
else
{
$SIZE_T
=
8
;
$g
=
"
g
";
}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
open
STDOUT
,"
>
$output
";
open
STDOUT
,"
>
$output
";
...
@@ -82,6 +102,8 @@ $rounds="%r13";
...
@@ -82,6 +102,8 @@ $rounds="%r13";
$ra
=
"
%r14
";
$ra
=
"
%r14
";
$sp
=
"
%r15
";
$sp
=
"
%r15
";
$stdframe
=
16
*$SIZE_T
+
4
*
8
;
sub
_data_word
()
sub
_data_word
()
{
my
$i
;
{
my
$i
;
while
(
defined
(
$i
=
shift
))
{
$code
.=
sprintf
"
.long
\t
0x%08x,0x%08x
\n
",
$i
,
$i
;
}
while
(
defined
(
$i
=
shift
))
{
$code
.=
sprintf
"
.long
\t
0x%08x,0x%08x
\n
",
$i
,
$i
;
}
...
@@ -223,7 +245,7 @@ $code.=<<___ if (!$softonly);
...
@@ -223,7 +245,7 @@ $code.=<<___ if (!$softonly);
.
Lesoft:
.
Lesoft:
___
___
$code
.=
<<___;
$code
.=
<<___;
stm
g %r3,$ra,24
($sp)
stm
${g} %r3,$ra,3*$SIZE_T
($sp)
llgf $s0,0($inp)
llgf $s0,0($inp)
llgf $s1,4($inp)
llgf $s1,4($inp)
...
@@ -233,20 +255,20 @@ $code.=<<___;
...
@@ -233,20 +255,20 @@ $code.=<<___;
larl $tbl,AES_Te
larl $tbl,AES_Te
bras $ra,_s390x_AES_encrypt
bras $ra,_s390x_AES_encrypt
l
g $out,24
($sp)
l
${g} $out,3*$SIZE_T
($sp)
st $s0,0($out)
st $s0,0($out)
st $s1,4($out)
st $s1,4($out)
st $s2,8($out)
st $s2,8($out)
st $s3,12($out)
st $s3,12($out)
lm
g %r6,$ra,48
($sp)
lm
${g} %r6,$ra,6*$SIZE_T
($sp)
br $ra
br $ra
.size AES_encrypt,.-AES_encrypt
.size AES_encrypt,.-AES_encrypt
.type _s390x_AES_encrypt,\@function
.type _s390x_AES_encrypt,\@function
.align 16
.align 16
_s390x_AES_encrypt:
_s390x_AES_encrypt:
st
g $ra,152
($sp)
st
${g} $ra,`$stdframe-$SIZE_T`
($sp)
x $s0,0($key)
x $s0,0($key)
x $s1,4($key)
x $s1,4($key)
x $s2,8($key)
x $s2,8($key)
...
@@ -410,7 +432,7 @@ _s390x_AES_encrypt:
...
@@ -410,7 +432,7 @@ _s390x_AES_encrypt:
or $s2,$i3
or $s2,$i3
or $s3,$t3
or $s3,$t3
l
g $ra,152
($sp)
l
${g} $ra,`$stdframe-$SIZE_T`
($sp)
xr $s0,$t0
xr $s0,$t0
xr $s1,$t2
xr $s1,$t2
x $s2,24($key)
x $s2,24($key)
...
@@ -549,7 +571,7 @@ $code.=<<___ if (!$softonly);
...
@@ -549,7 +571,7 @@ $code.=<<___ if (!$softonly);
.
Ldsoft:
.
Ldsoft:
___
___
$code
.=
<<___;
$code
.=
<<___;
stm
g %r3,$ra,24
($sp)
stm
${g} %r3,$ra,3*$SIZE_T
($sp)
llgf $s0,0($inp)
llgf $s0,0($inp)
llgf $s1,4($inp)
llgf $s1,4($inp)
...
@@ -559,20 +581,20 @@ $code.=<<___;
...
@@ -559,20 +581,20 @@ $code.=<<___;
larl $tbl,AES_Td
larl $tbl,AES_Td
bras $ra,_s390x_AES_decrypt
bras $ra,_s390x_AES_decrypt
l
g $out,24
($sp)
l
${g} $out,3*$SIZE_T
($sp)
st $s0,0($out)
st $s0,0($out)
st $s1,4($out)
st $s1,4($out)
st $s2,8($out)
st $s2,8($out)
st $s3,12($out)
st $s3,12($out)
lm
g %r6,$ra,48
($sp)
lm
${g} %r6,$ra,6*$SIZE_T
($sp)
br $ra
br $ra
.size AES_decrypt,.-AES_decrypt
.size AES_decrypt,.-AES_decrypt
.type _s390x_AES_decrypt,\@function
.type _s390x_AES_decrypt,\@function
.align 16
.align 16
_s390x_AES_decrypt:
_s390x_AES_decrypt:
st
g $ra,152
($sp)
st
${g} $ra,`$stdframe-$SIZE_T`
($sp)
x $s0,0($key)
x $s0,0($key)
x $s1,4($key)
x $s1,4($key)
x $s2,8($key)
x $s2,8($key)
...
@@ -716,7 +738,7 @@ _s390x_AES_decrypt:
...
@@ -716,7 +738,7 @@ _s390x_AES_decrypt:
nr $i1,$mask
nr $i1,$mask
nr $i2,$mask
nr $i2,$mask
l
g $ra,152
($sp)
l
${g} $ra,`$stdframe-$SIZE_T`
($sp)
or $s1,$t1
or $s1,$t1
l $t0,16($key)
l $t0,16($key)
l $t1,20($key)
l $t1,20($key)
...
@@ -750,9 +772,9 @@ $code.=<<___;
...
@@ -750,9 +772,9 @@ $code.=<<___;
.align 16
.align 16
AES_set_encrypt_key:
AES_set_encrypt_key:
lghi $t0,0
lghi $t0,0
cl
g
r $inp,$t0
cl
${g}
r $inp,$t0
je .Lminus1
je .Lminus1
cl
g
r $key,$t0
cl
${g}
r $key,$t0
je .Lminus1
je .Lminus1
lghi $t0,128
lghi $t0,128
...
@@ -810,7 +832,7 @@ ___
...
@@ -810,7 +832,7 @@ ___
$code
.=
<<___;
$code
.=
<<___;
.align 16
.align 16
.Lekey_internal:
.Lekey_internal:
stm
g %r6,%r13,48
($sp) # all non-volatile regs
stm
${g} %r6,%r13,6*$SIZE_T
($sp) # all non-volatile regs
larl $tbl,AES_Te+2048
larl $tbl,AES_Te+2048
...
@@ -871,7 +893,7 @@ $code.=<<___;
...
@@ -871,7 +893,7 @@ $code.=<<___;
la $t3,4($t3) # i++
la $t3,4($t3) # i++
brct $rounds,.L128_loop
brct $rounds,.L128_loop
lghi %r2,0
lghi %r2,0
lm
g %r6,%r13,48
($sp)
lm
${g} %r6,%r13,6*$SIZE_T
($sp)
br $ra
br $ra
.align 16
.align 16
...
@@ -919,7 +941,7 @@ $code.=<<___;
...
@@ -919,7 +941,7 @@ $code.=<<___;
st $s3,36($key)
st $s3,36($key)
brct $rounds,.L192_continue
brct $rounds,.L192_continue
lghi %r2,0
lghi %r2,0
lm
g %r6,%r13,48
($sp)
lm
${g} %r6,%r13,6*$SIZE_T
($sp)
br $ra
br $ra
.align 16
.align 16
...
@@ -981,7 +1003,7 @@ $code.=<<___;
...
@@ -981,7 +1003,7 @@ $code.=<<___;
st $s3,44($key)
st $s3,44($key)
brct $rounds,.L256_continue
brct $rounds,.L256_continue
lghi %r2,0
lghi %r2,0
lm
g %r6,%r13,48
($sp)
lm
${g} %r6,%r13,6*$SIZE_T
($sp)
br $ra
br $ra
.align 16
.align 16
...
@@ -1032,11 +1054,11 @@ $code.=<<___;
...
@@ -1032,11 +1054,11 @@ $code.=<<___;
.type AES_set_decrypt_key,\@function
.type AES_set_decrypt_key,\@function
.align 16
.align 16
AES_set_decrypt_key:
AES_set_decrypt_key:
st
g $key,32($sp)
# I rely on AES_set_encrypt_key to
st
${g} $key,4*$SIZE_T($sp)
# I rely on AES_set_encrypt_key to
st
g $ra,112($sp)
# save non-volatile registers!
st
${g} $ra,14*$SIZE_T($sp)
# save non-volatile registers!
bras $ra,AES_set_encrypt_key
bras $ra,AES_set_encrypt_key
l
g $key,32
($sp)
l
${g} $key,4*$SIZE_T
($sp)
l
g $ra,112
($sp)
l
${g} $ra,14*$SIZE_T
($sp)
ltgr %r2,%r2
ltgr %r2,%r2
bnzr $ra
bnzr $ra
___
___
...
@@ -1051,11 +1073,11 @@ $code.=<<___ if (!$softonly);
...
@@ -1051,11 +1073,11 @@ $code.=<<___ if (!$softonly);
.
align
16
.
align
16
.
Ldkey_internal:
.
Ldkey_internal:
st
g
$key
,
32
(
$sp
)
st
$
{
g
}
$key
,
4
*$SIZE_T
(
$sp
)
st
g
$ra
,
40
(
$sp
)
st
$
{
g
}
$ra
,
14
*$SIZE_T
(
$sp
)
bras
$ra
,
.
Lekey_internal
bras
$ra
,
.
Lekey_internal
l
g
$key
,
32
(
$sp
)
l
$
{
g
}
$key
,
4
*$SIZE_T
(
$sp
)
l
g
$ra
,
40
(
$sp
)
l
$
{
g
}
$ra
,
14
*$SIZE_T
(
$sp
)
___
___
$code
.=
<<___;
$code
.=
<<___;
...
@@ -1136,7 +1158,7 @@ $code.=<<___;
...
@@ -1136,7 +1158,7 @@ $code.=<<___;
la $key,4($key)
la $key,4($key)
brct $rounds,.Lmix
brct $rounds,.Lmix
lm
g %r6,%r13,48
($sp)# as was saved by AES_set_encrypt_key!
lm
${g} %r6,%r13,6*$SIZE_T
($sp)# as was saved by AES_set_encrypt_key!
lghi %r2,0
lghi %r2,0
br $ra
br $ra
.size AES_set_decrypt_key,.-AES_set_decrypt_key
.size AES_set_decrypt_key,.-AES_set_decrypt_key
...
@@ -1176,7 +1198,7 @@ $code.=<<___ if (!$softonly);
...
@@ -1176,7 +1198,7 @@ $code.=<<___ if (!$softonly);
l
%r0
,
240
(
$key
)
# load kmc code
l
%r0
,
240
(
$key
)
# load kmc code
lghi
$key
,
15
# res=len%16, len-=res;
lghi
$key
,
15
# res=len%16, len-=res;
ngr
$key
,
$len
ngr
$key
,
$len
sl
g
r
$len
,
$key
sl
$
{
g
}
r
$len
,
$key
la
%r1
,
16
(
$sp
)
# parameter block - ivec || key
la
%r1
,
16
(
$sp
)
# parameter block - ivec || key
jz
.
Lkmc_truncated
jz
.
Lkmc_truncated
.
long
0xb92f0042
# kmc %r4,%r2
.
long
0xb92f0042
# kmc %r4,%r2
...
@@ -1194,34 +1216,34 @@ $code.=<<___ if (!$softonly);
...
@@ -1194,34 +1216,34 @@ $code.=<<___ if (!$softonly);
tmll
%r0
,
0x80
tmll
%r0
,
0x80
jnz
.
Lkmc_truncated_dec
jnz
.
Lkmc_truncated_dec
lghi
%r1
,
0
lghi
%r1
,
0
stg
%r1
,
1
28
(
$sp
)
stg
%r1
,
1
6
*$SIZE_T
(
$sp
)
stg
%r1
,
1
36
(
$sp
)
stg
%r1
,
1
6
*$SIZE_T
+
8
(
$sp
)
bras
%r1
,
1
f
bras
%r1
,
1
f
mvc
1
28
(
1
,
$sp
),
0
(
$inp
)
mvc
1
6
*$SIZE_T
(
1
,
$sp
),
0
(
$inp
)
1
:
ex
$key
,
0
(
%r1
)
1
:
ex
$key
,
0
(
%r1
)
la
%r1
,
16
(
$sp
)
# restore parameter block
la
%r1
,
16
(
$sp
)
# restore parameter block
la
$inp
,
1
28
(
$sp
)
la
$inp
,
1
6
*$SIZE_T
(
$sp
)
lghi
$len
,
16
lghi
$len
,
16
.
long
0xb92f0042
# kmc %r4,%r2
.
long
0xb92f0042
# kmc %r4,%r2
j
.
Lkmc_done
j
.
Lkmc_done
.
align
16
.
align
16
.
Lkmc_truncated_dec:
.
Lkmc_truncated_dec:
st
g
$out
,
64
(
$sp
)
st
$
{
g
}
$out
,
4
*$SIZE_T
(
$sp
)
la
$out
,
1
28
(
$sp
)
la
$out
,
1
6
*$SIZE_T
(
$sp
)
lghi
$len
,
16
lghi
$len
,
16
.
long
0xb92f0042
# kmc %r4,%r2
.
long
0xb92f0042
# kmc %r4,%r2
l
g
$out
,
64
(
$sp
)
l
$
{
g
}
$out
,
4
*$SIZE_T
(
$sp
)
bras
%r1
,
2
f
bras
%r1
,
2
f
mvc
0
(
1
,
$out
),
1
28
(
$sp
)
mvc
0
(
1
,
$out
),
1
6
*$SIZE_T
(
$sp
)
2
:
ex
$key
,
0
(
%r1
)
2
:
ex
$key
,
0
(
%r1
)
j
.
Lkmc_done
j
.
Lkmc_done
.
align
16
.
align
16
.
Lcbc_software:
.
Lcbc_software:
___
___
$code
.=
<<___;
$code
.=
<<___;
stm
g $key,$ra,40
($sp)
stm
${g} $key,$ra,5*$SIZE_T
($sp)
lhi %r0,0
lhi %r0,0
cl %r0,
164
($sp)
cl %r0,
`$stdframe+$SIZE_T-4`
($sp)
je .Lcbc_decrypt
je .Lcbc_decrypt
larl $tbl,AES_Te
larl $tbl,AES_Te
...
@@ -1232,10 +1254,10 @@ $code.=<<___;
...
@@ -1232,10 +1254,10 @@ $code.=<<___;
llgf $s3,12($ivp)
llgf $s3,12($ivp)
lghi $t0,16
lghi $t0,16
sl
g
r $len,$t0
sl
${g}
r $len,$t0
brc 4,.Lcbc_enc_tail # if borrow
brc 4,.Lcbc_enc_tail # if borrow
.Lcbc_enc_loop:
.Lcbc_enc_loop:
stm
g $inp,$out,16
($sp)
stm
${g} $inp,$out,2*$SIZE_T
($sp)
x $s0,0($inp)
x $s0,0($inp)
x $s1,4($inp)
x $s1,4($inp)
x $s2,8($inp)
x $s2,8($inp)
...
@@ -1244,7 +1266,7 @@ $code.=<<___;
...
@@ -1244,7 +1266,7 @@ $code.=<<___;
bras $ra,_s390x_AES_encrypt
bras $ra,_s390x_AES_encrypt
lm
g $inp,$key,16
($sp)
lm
${g} $inp,$key,2*$SIZE_T
($sp)
st $s0,0($out)
st $s0,0($out)
st $s1,4($out)
st $s1,4($out)
st $s2,8($out)
st $s2,8($out)
...
@@ -1253,33 +1275,33 @@ $code.=<<___;
...
@@ -1253,33 +1275,33 @@ $code.=<<___;
la $inp,16($inp)
la $inp,16($inp)
la $out,16($out)
la $out,16($out)
lghi $t0,16
lghi $t0,16
lt
g
r $len,$len
lt
${g}
r $len,$len
jz .Lcbc_enc_done
jz .Lcbc_enc_done
sl
g
r $len,$t0
sl
${g}
r $len,$t0
brc 4,.Lcbc_enc_tail # if borrow
brc 4,.Lcbc_enc_tail # if borrow
j .Lcbc_enc_loop
j .Lcbc_enc_loop
.align 16
.align 16
.Lcbc_enc_done:
.Lcbc_enc_done:
l
g $ivp,48
($sp)
l
${g} $ivp,6*$SIZE_T
($sp)
st $s0,0($ivp)
st $s0,0($ivp)
st $s1,4($ivp)
st $s1,4($ivp)
st $s2,8($ivp)
st $s2,8($ivp)
st $s3,12($ivp)
st $s3,12($ivp)
lm
g %r7,$ra,56
($sp)
lm
${g} %r7,$ra,7*$SIZE_T
($sp)
br $ra
br $ra
.align 16
.align 16
.Lcbc_enc_tail:
.Lcbc_enc_tail:
aghi $len,15
aghi $len,15
lghi $t0,0
lghi $t0,0
stg $t0,1
28
($sp)
stg $t0,1
6*$SIZE_T
($sp)
stg $t0,1
36
($sp)
stg $t0,1
6*$SIZE_T+8
($sp)
bras $t1,3f
bras $t1,3f
mvc 1
28
(1,$sp),0($inp)
mvc 1
6*$SIZE_T
(1,$sp),0($inp)
3: ex $len,0($t1)
3: ex $len,0($t1)
lghi $len,0
lghi $len,0
la $inp,1
28
($sp)
la $inp,1
6*$SIZE_T
($sp)
j .Lcbc_enc_loop
j .Lcbc_enc_loop
.align 16
.align 16
...
@@ -1288,10 +1310,10 @@ $code.=<<___;
...
@@ -1288,10 +1310,10 @@ $code.=<<___;
lg $t0,0($ivp)
lg $t0,0($ivp)
lg $t1,8($ivp)
lg $t1,8($ivp)
stmg $t0,$t1,1
28
($sp)
stmg $t0,$t1,1
6*$SIZE_T
($sp)
.Lcbc_dec_loop:
.Lcbc_dec_loop:
stm
g $inp,$out,16
($sp)
stm
${g} $inp,$out,2*$SIZE_T
($sp)
llgf $s0,0($inp)
llgf $s0,0($inp)
llgf $s1,4($inp)
llgf $s1,4($inp)
llgf $s2,8($inp)
llgf $s2,8($inp)
...
@@ -1300,7 +1322,7 @@ $code.=<<___;
...
@@ -1300,7 +1322,7 @@ $code.=<<___;
bras $ra,_s390x_AES_decrypt
bras $ra,_s390x_AES_decrypt
lm
g $inp,$key,16
($sp)
lm
${g} $inp,$key,2*$SIZE_T
($sp)
sllg $s0,$s0,32
sllg $s0,$s0,32
sllg $s2,$s2,32
sllg $s2,$s2,32
lr $s0,$s1
lr $s0,$s1
...
@@ -1308,15 +1330,15 @@ $code.=<<___;
...
@@ -1308,15 +1330,15 @@ $code.=<<___;
lg $t0,0($inp)
lg $t0,0($inp)
lg $t1,8($inp)
lg $t1,8($inp)
xg $s0,1
28
($sp)
xg $s0,1
6*$SIZE_T
($sp)
xg $s2,1
36
($sp)
xg $s2,1
6*$SIZE_T+8
($sp)
lghi $s1,16
lghi $s1,16
sl
g
r $len,$s1
sl
${g}
r $len,$s1
brc 4,.Lcbc_dec_tail # if borrow
brc 4,.Lcbc_dec_tail # if borrow
brc 2,.Lcbc_dec_done # if zero
brc 2,.Lcbc_dec_done # if zero
stg $s0,0($out)
stg $s0,0($out)
stg $s2,8($out)
stg $s2,8($out)
stmg $t0,$t1,1
28
($sp)
stmg $t0,$t1,1
6*$SIZE_T
($sp)
la $inp,16($inp)
la $inp,16($inp)
la $out,16($out)
la $out,16($out)
...
@@ -1326,7 +1348,7 @@ $code.=<<___;
...
@@ -1326,7 +1348,7 @@ $code.=<<___;
stg $s0,0($out)
stg $s0,0($out)
stg $s2,8($out)
stg $s2,8($out)
.Lcbc_dec_exit:
.Lcbc_dec_exit:
lm
g $ivp,$ra,48
($sp)
lm
${g} %r6,$ra,6*$SIZE_T
($sp)
stmg $t0,$t1,0($ivp)
stmg $t0,$t1,0($ivp)
br $ra
br $ra
...
@@ -1334,10 +1356,10 @@ $code.=<<___;
...
@@ -1334,10 +1356,10 @@ $code.=<<___;
.align 16
.align 16
.Lcbc_dec_tail:
.Lcbc_dec_tail:
aghi $len,15
aghi $len,15
stg $s0,1
28
($sp)
stg $s0,1
6*$SIZE_T
($sp)
stg $s2,1
36
($sp)
stg $s2,1
6*$SIZE_T+8
($sp)
bras $s1,4f
bras $s1,4f
mvc 0(1,$out),1
28
($sp)
mvc 0(1,$out),1
6*$SIZE_T
($sp)
4: ex $len,0($s1)
4: ex $len,0($s1)
j .Lcbc_dec_exit
j .Lcbc_dec_exit
.size AES_cbc_encrypt,.-AES_cbc_encrypt
.size AES_cbc_encrypt,.-AES_cbc_encrypt
...
@@ -1359,6 +1381,7 @@ $code.=<<___;
...
@@ -1359,6 +1381,7 @@ $code.=<<___;
.type AES_ctr32_encrypt,\@function
.type AES_ctr32_encrypt,\@function
.align 16
.align 16
AES_ctr32_encrypt:
AES_ctr32_encrypt:
llgfr $len,$len # safe in ctr32 subroutine even in 64-bit case
___
___
$code
.=<<
___
if
(
!
$softonly
);
$code
.=<<
___
if
(
!
$softonly
);
l
%r0
,
240
(
$key
)
l
%r0
,
240
(
$key
)
...
@@ -1366,7 +1389,7 @@ $code.=<<___ if (!$softonly);
...
@@ -1366,7 +1389,7 @@ $code.=<<___ if (!$softonly);
clr
%r0
,
%r1
clr
%r0
,
%r1
jl
.
Lctr32_software
jl
.
Lctr32_software
stm
g
%r6
,
$s3
,
48
(
$sp
)
stm
$
{
g
}
%r6
,
$s3
,
6
*$SIZE_T
(
$sp
)
slgr
$out
,
$inp
slgr
$out
,
$inp
la
%r1
,
0
(
$key
)
# %r1 is permanent copy of $key
la
%r1
,
0
(
$key
)
# %r1 is permanent copy of $key
...
@@ -1388,14 +1411,14 @@ $code.=<<___ if (!$softonly);
...
@@ -1388,14 +1411,14 @@ $code.=<<___ if (!$softonly);
la
$sp
,
1024
(
$s0
)
# alloca
la
$sp
,
1024
(
$s0
)
# alloca
srlg
$fp
,
$fp
,
4
# convert bytes to blocks, minimum 16
srlg
$fp
,
$fp
,
4
# convert bytes to blocks, minimum 16
st
g
$s2
,
0
(
$sp
)
# back-chain
st
$
{
g
}
$s2
,
0
(
$sp
)
# back-chain
st
g
$fp
,
8
(
$sp
)
st
$
{
g
}
$fp
,
$SIZE_T
(
$sp
)
slgr
$len
,
$fp
slgr
$len
,
$fp
brc
1
,
.
Lctr32_hw_loop
# not zero, no borrow
brc
1
,
.
Lctr32_hw_loop
# not zero, no borrow
algr
$fp
,
$len
# input is shorter than allocated buffer
algr
$fp
,
$len
# input is shorter than allocated buffer
lghi
$len
,
0
lghi
$len
,
0
st
g
$fp
,
8
(
$sp
)
st
$
{
g
}
$fp
,
$SIZE_T
(
$sp
)
.
Lctr32_hw_loop:
.
Lctr32_hw_loop:
la
$s2
,
16
(
$sp
)
la
$s2
,
16
(
$sp
)
...
@@ -1432,8 +1455,8 @@ $code.=<<___ if (!$softonly);
...
@@ -1432,8 +1455,8 @@ $code.=<<___ if (!$softonly);
lghi
$len
,
0
lghi
$len
,
0
brc
4
+
1
,
.
Lctr32_hw_loop
# not zero
brc
4
+
1
,
.
Lctr32_hw_loop
# not zero
l
g
$s0
,
0
(
$sp
)
l
$
{
g
}
$s0
,
0
(
$sp
)
l
g
$s1
,
8
(
$sp
)
l
$
{
g
}
$s1
,
$SIZE_T
(
$sp
)
la
$s2
,
16
(
$sp
)
la
$s2
,
16
(
$sp
)
.
Lctr32_hw_zap:
.
Lctr32_hw_zap:
stg
$s0
,
0
(
$s2
)
stg
$s0
,
0
(
$s2
)
...
@@ -1442,30 +1465,30 @@ $code.=<<___ if (!$softonly);
...
@@ -1442,30 +1465,30 @@ $code.=<<___ if (!$softonly);
brct
$s1
,
.
Lctr32_hw_zap
brct
$s1
,
.
Lctr32_hw_zap
la
$sp
,
0
(
$s0
)
la
$sp
,
0
(
$s0
)
lm
g
%r6
,
$s3
,
48
(
$sp
)
lm
$
{
g
}
%r6
,
$s3
,
6
*$SIZE_T
(
$sp
)
br
$ra
br
$ra
.
align
16
.
align
16
.
Lctr32_software:
.
Lctr32_software:
___
___
$code
.=
<<___;
$code
.=
<<___;
stm
g $key,$ra,40
($sp)
stm
${g} $key,$ra,5*$SIZE_T
($sp)
sl
g
r $out,$inp
sl
${g}
r $out,$inp
larl $tbl,AES_Te
larl $tbl,AES_Te
llgf $t1,12($ivp)
llgf $t1,12($ivp)
.Lctr32_loop:
.Lctr32_loop:
stm
g $inp,$len,16
($sp)
stm
${g} $inp,$len,2*$SIZE_T
($sp)
llgf $s0,0($ivp)
llgf $s0,0($ivp)
llgf $s1,4($ivp)
llgf $s1,4($ivp)
llgf $s2,8($ivp)
llgf $s2,8($ivp)
lgr $s3,$t1
lgr $s3,$t1
st $t1,1
28
($sp)
st $t1,1
6*$SIZE_T
($sp)
lgr %r4,$key
lgr %r4,$key
bras $ra,_s390x_AES_encrypt
bras $ra,_s390x_AES_encrypt
lm
g $inp,$ivp,16
($sp)
lm
${g} $inp,$ivp,2*$SIZE_T
($sp)
llgf $t1,1
28
($sp)
llgf $t1,1
6*$SIZE_T
($sp)
x $s0,0($inp)
x $s0,0($inp)
x $s1,4($inp)
x $s1,4($inp)
x $s2,8($inp)
x $s2,8($inp)
...
@@ -1479,7 +1502,7 @@ $code.=<<___;
...
@@ -1479,7 +1502,7 @@ $code.=<<___;
ahi $t1,1 # 32-bit increment
ahi $t1,1 # 32-bit increment
brct $len,.Lctr32_loop
brct $len,.Lctr32_loop
lm
g %r6,$ra,48
($sp)
lm
${g} %r6,$ra,6*$SIZE_T
($sp)
br $ra
br $ra
.size AES_ctr32_encrypt,.-AES_ctr32_encrypt
.size AES_ctr32_encrypt,.-AES_ctr32_encrypt
___
___
...
...
crypto/bn/asm/s390x-mont.pl
浏览文件 @
e822c756
...
@@ -32,9 +32,33 @@
...
@@ -32,9 +32,33 @@
# Reschedule to minimize/avoid Address Generation Interlock hazard,
# Reschedule to minimize/avoid Address Generation Interlock hazard,
# make inner loops counter-based.
# make inner loops counter-based.
# November 2010.
#
# Adapt for -m31 build. If kernel supports what's called "highgprs"
# feature on Linux [see /proc/cpuinfo], it's possible to use 64-bit
# instructions and achieve "64-bit" performance even in 31-bit legacy
# application context. The feature is not specific to any particular
# processor, as long as it's "z-CPU". Latter implies that the code
# remains z/Architecture specific. Compatibility with 32-bit BN_ULONG
# is achieved by swapping words after 64-bit loads, follow _dswap-s.
# On z990 it was measured to perform 2.6-2.2 times better, less for
# longer keys...
$flavour
=
shift
;
if
(
$flavour
=~
/3[12]/
)
{
$SIZE_T
=
4
;
$g
=
"";
}
else
{
$SIZE_T
=
8
;
$g
=
"
g
";
}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
open
STDOUT
,"
>
$output
";
open
STDOUT
,"
>
$output
";
$stdframe
=
16
*$SIZE_T
+
4
*
8
;
$mn0
=
"
%r0
";
$mn0
=
"
%r0
";
$num
=
"
%r1
";
$num
=
"
%r1
";
...
@@ -63,34 +87,44 @@ $code.=<<___;
...
@@ -63,34 +87,44 @@ $code.=<<___;
.globl bn_mul_mont
.globl bn_mul_mont
.type bn_mul_mont,\@function
.type bn_mul_mont,\@function
bn_mul_mont:
bn_mul_mont:
lgf $num,
164
($sp) # pull $num
lgf $num,
`$stdframe+$SIZE_T-4`
($sp) # pull $num
sla $num,
3
# $num to enumerate bytes
sla $num,
`log($SIZE_T)/log(2)`
# $num to enumerate bytes
la $bp,0($num,$bp)
la $bp,0($num,$bp)
st
g %r2,16
($sp)
st
${g} %r2,2*$SIZE_T
($sp)
cghi $num,16 #
cghi $num,16 #
lghi %r2,0 #
lghi %r2,0 #
blr %r14 # if($num<16) return 0;
blr %r14 # if($num<16) return 0;
___
$code
.=<<
___
if
(
$flavour
=~
/3[12]/
);
tmll
$num
,
4
bnzr
%r14
# if ($num&1) return 0;
___
$code
.=<<
___
if
(
$flavour
!~
/3[12]/
);
cghi
$num
,
128
#
cghi
$num
,
128
#
bhr
%r14
# if($num>128) return 0;
bhr
%r14
# if($num>128) return 0;
___
$code
.=
<<___;
stm${g} %r3,%r15,3*$SIZE_T($sp)
stmg %r3,%r15,24($sp)
lghi $rp,-$stdframe-8 # leave room for carry bit
lghi $rp,-160-8 # leave room for carry bit
lcgr $j,$num # -$num
lcgr $j,$num # -$num
lgr %r0,$sp
lgr %r0,$sp
la $rp,0($rp,$sp)
la $rp,0($rp,$sp)
la $sp,0($j,$rp) # alloca
la $sp,0($j,$rp) # alloca
st
g
%r0,0($sp) # back chain
st
${g}
%r0,0($sp) # back chain
sra $num,3 # restore $num
sra $num,3 # restore $num
la $bp,0($j,$bp) # restore $bp
la $bp,0($j,$bp) # restore $bp
ahi $num,-1 # adjust $num for inner loop
ahi $num,-1 # adjust $num for inner loop
lg $n0,0($n0) # pull n0
lg $n0,0($n0) # pull n0
_dswap $n0
lg $bi,0($bp)
lg $bi,0($bp)
_dswap $bi
lg $alo,0($ap)
lg $alo,0($ap)
_dswap $alo
mlgr $ahi,$bi # ap[0]*bp[0]
mlgr $ahi,$bi # ap[0]*bp[0]
lgr $AHI,$ahi
lgr $AHI,$ahi
...
@@ -98,6 +132,7 @@ bn_mul_mont:
...
@@ -98,6 +132,7 @@ bn_mul_mont:
msgr $mn0,$n0
msgr $mn0,$n0
lg $nlo,0($np) #
lg $nlo,0($np) #
_dswap $nlo
mlgr $nhi,$mn0 # np[0]*m1
mlgr $nhi,$mn0 # np[0]*m1
algr $nlo,$alo # +="tp[0]"
algr $nlo,$alo # +="tp[0]"
lghi $NHI,0
lghi $NHI,0
...
@@ -109,12 +144,14 @@ bn_mul_mont:
...
@@ -109,12 +144,14 @@ bn_mul_mont:
.align 16
.align 16
.L1st:
.L1st:
lg $alo,0($j,$ap)
lg $alo,0($j,$ap)
_dswap $alo
mlgr $ahi,$bi # ap[j]*bp[0]
mlgr $ahi,$bi # ap[j]*bp[0]
algr $alo,$AHI
algr $alo,$AHI
lghi $AHI,0
lghi $AHI,0
alcgr $AHI,$ahi
alcgr $AHI,$ahi
lg $nlo,0($j,$np)
lg $nlo,0($j,$np)
_dswap $nlo
mlgr $nhi,$mn0 # np[j]*m1
mlgr $nhi,$mn0 # np[j]*m1
algr $nlo,$NHI
algr $nlo,$NHI
lghi $NHI,0
lghi $NHI,0
...
@@ -122,22 +159,24 @@ bn_mul_mont:
...
@@ -122,22 +159,24 @@ bn_mul_mont:
algr $nlo,$alo
algr $nlo,$alo
alcgr $NHI,$nhi
alcgr $NHI,$nhi
stg $nlo,
160
-8($j,$sp) # tp[j-1]=
stg $nlo,
$stdframe
-8($j,$sp) # tp[j-1]=
la $j,8($j) # j++
la $j,8($j) # j++
brct $count,.L1st
brct $count,.L1st
algr $NHI,$AHI
algr $NHI,$AHI
lghi $AHI,0
lghi $AHI,0
alcgr $AHI,$AHI # upmost overflow bit
alcgr $AHI,$AHI # upmost overflow bit
stg $NHI,
160
-8($j,$sp)
stg $NHI,
$stdframe
-8($j,$sp)
stg $AHI,
160
($j,$sp)
stg $AHI,
$stdframe
($j,$sp)
la $bp,8($bp) # bp++
la $bp,8($bp) # bp++
.Louter:
.Louter:
lg $bi,0($bp) # bp[i]
lg $bi,0($bp) # bp[i]
_dswap $bi
lg $alo,0($ap)
lg $alo,0($ap)
_dswap $alo
mlgr $ahi,$bi # ap[0]*bp[i]
mlgr $ahi,$bi # ap[0]*bp[i]
alg $alo,
160
($sp) # +=tp[0]
alg $alo,
$stdframe
($sp) # +=tp[0]
lghi $AHI,0
lghi $AHI,0
alcgr $AHI,$ahi
alcgr $AHI,$ahi
...
@@ -145,6 +184,7 @@ bn_mul_mont:
...
@@ -145,6 +184,7 @@ bn_mul_mont:
msgr $mn0,$n0 # tp[0]*n0
msgr $mn0,$n0 # tp[0]*n0
lg $nlo,0($np) # np[0]
lg $nlo,0($np) # np[0]
_dswap $nlo
mlgr $nhi,$mn0 # np[0]*m1
mlgr $nhi,$mn0 # np[0]*m1
algr $nlo,$alo # +="tp[0]"
algr $nlo,$alo # +="tp[0]"
lghi $NHI,0
lghi $NHI,0
...
@@ -156,14 +196,16 @@ bn_mul_mont:
...
@@ -156,14 +196,16 @@ bn_mul_mont:
.align 16
.align 16
.Linner:
.Linner:
lg $alo,0($j,$ap)
lg $alo,0($j,$ap)
_dswap $alo
mlgr $ahi,$bi # ap[j]*bp[i]
mlgr $ahi,$bi # ap[j]*bp[i]
algr $alo,$AHI
algr $alo,$AHI
lghi $AHI,0
lghi $AHI,0
alcgr $ahi,$AHI
alcgr $ahi,$AHI
alg $alo,
160
($j,$sp)# +=tp[j]
alg $alo,
$stdframe
($j,$sp)# +=tp[j]
alcgr $AHI,$ahi
alcgr $AHI,$ahi
lg $nlo,0($j,$np)
lg $nlo,0($j,$np)
_dswap $nlo
mlgr $nhi,$mn0 # np[j]*m1
mlgr $nhi,$mn0 # np[j]*m1
algr $nlo,$NHI
algr $nlo,$NHI
lghi $NHI,0
lghi $NHI,0
...
@@ -171,31 +213,33 @@ bn_mul_mont:
...
@@ -171,31 +213,33 @@ bn_mul_mont:
algr $nlo,$alo # +="tp[j]"
algr $nlo,$alo # +="tp[j]"
alcgr $NHI,$nhi
alcgr $NHI,$nhi
stg $nlo,
160
-8($j,$sp) # tp[j-1]=
stg $nlo,
$stdframe
-8($j,$sp) # tp[j-1]=
la $j,8($j) # j++
la $j,8($j) # j++
brct $count,.Linner
brct $count,.Linner
algr $NHI,$AHI
algr $NHI,$AHI
lghi $AHI,0
lghi $AHI,0
alcgr $AHI,$AHI
alcgr $AHI,$AHI
alg $NHI,
160
($j,$sp)# accumulate previous upmost overflow bit
alg $NHI,
$stdframe
($j,$sp)# accumulate previous upmost overflow bit
lghi $ahi,0
lghi $ahi,0
alcgr $AHI,$ahi # new upmost overflow bit
alcgr $AHI,$ahi # new upmost overflow bit
stg $NHI,
160
-8($j,$sp)
stg $NHI,
$stdframe
-8($j,$sp)
stg $AHI,
160
($j,$sp)
stg $AHI,
$stdframe
($j,$sp)
la $bp,8($bp) # bp++
la $bp,8($bp) # bp++
cl
g $bp,160+8+32
($j,$sp) # compare to &bp[num]
cl
${g} $bp,`$stdframe+8+4*$SIZE_T`
($j,$sp) # compare to &bp[num]
jne .Louter
jne .Louter
l
g $rp,160+8+16
($j,$sp) # reincarnate rp
l
${g} $rp,`$stdframe+8+2*$SIZE_T`
($j,$sp) # reincarnate rp
la $ap,
160
($sp)
la $ap,
$stdframe
($sp)
ahi $num,1 # restore $num, incidentally clears "borrow"
ahi $num,1 # restore $num, incidentally clears "borrow"
la $j,0(%r0)
la $j,0(%r0)
lr $count,$num
lr $count,$num
.Lsub: lg $alo,0($j,$ap)
.Lsub: lg $alo,0($j,$ap)
slbg $alo,0($j,$np)
lg $nlo,0($j,$np)
_dswap $nlo
slbgr $alo,$nlo
stg $alo,0($j,$rp)
stg $alo,0($j,$rp)
la $j,8($j)
la $j,8($j)
brct $count,.Lsub
brct $count,.Lsub
...
@@ -210,19 +254,24 @@ bn_mul_mont:
...
@@ -210,19 +254,24 @@ bn_mul_mont:
la $j,0(%r0)
la $j,0(%r0)
lgr $count,$num
lgr $count,$num
.Lcopy: lg $alo,0($j,$ap) # copy or in-place refresh
.Lcopy: lg $alo,0($j,$ap) # copy or in-place refresh
stg $j,160($j,$sp) # zap tp
_dswap $alo
stg $j,$stdframe($j,$sp) # zap tp
stg $alo,0($j,$rp)
stg $alo,0($j,$rp)
la $j,8($j)
la $j,8($j)
brct $count,.Lcopy
brct $count,.Lcopy
la %r1,
160+8+48
($j,$sp)
la %r1,
`$stdframe+8+6*$SIZE_T`
($j,$sp)
lm
g
%r6,%r15,0(%r1)
lm
${g}
%r6,%r15,0(%r1)
lghi %r2,1 # signal "processed"
lghi %r2,1 # signal "processed"
br %r14
br %r14
.size bn_mul_mont,.-bn_mul_mont
.size bn_mul_mont,.-bn_mul_mont
.string "Montgomery Multiplication for s390x, CRYPTOGAMS by <appro\@openssl.org>"
.string "Montgomery Multiplication for s390x, CRYPTOGAMS by <appro\@openssl.org>"
___
___
print
$code
;
foreach
(
split
("
\n
",
$code
))
{
s/\`([^\`]*)\`/eval $1/g
e
;
s/_dswap\s+(%r[0-9]+)/sprintf("rllg\t%s,%s,32",$1,$1) if($SIZE_T==4)/
e
;
print
$_
,"
\n
";
}
close
STDOUT
;
close
STDOUT
;
crypto/modes/asm/ghash-s390x.pl
浏览文件 @
e822c756
...
@@ -18,6 +18,26 @@
...
@@ -18,6 +18,26 @@
# and the result should be close to 12. In the lack of instruction-
# and the result should be close to 12. In the lack of instruction-
# level profiling data it's impossible to tell why...
# level profiling data it's impossible to tell why...
# November 2010.
#
# Adapt for -m31 build. If kernel supports what's called "highgprs"
# feature on Linux [see /proc/cpuinfo], it's possible to use 64-bit
# instructions and achieve "64-bit" performance even in 31-bit legacy
# application context. The feature is not specific to any particular
# processor, as long as it's "z-CPU". Latter implies that the code
# remains z/Architecture specific. On z990 it was measured to perform
# 2.8x better than 32-bit code generated by gcc 4.3.
$flavour
=
shift
;
if
(
$flavour
=~
/3[12]/
)
{
$SIZE_T
=
4
;
$g
=
"";
}
else
{
$SIZE_T
=
8
;
$g
=
"
g
";
}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
open
STDOUT
,"
>
$output
";
open
STDOUT
,"
>
$output
";
...
@@ -74,7 +94,7 @@ $code.=<<___ if(!$softonly);
...
@@ -74,7 +94,7 @@ $code.=<<___ if(!$softonly);
.
Lsoft_gmult:
.
Lsoft_gmult:
___
___
$code
.=
<<___;
$code
.=
<<___;
stm
g %r6,%r14,48
($sp)
stm
${g} %r6,%r14,6*$SIZE_T
($sp)
aghi $Xi,-1
aghi $Xi,-1
lghi $len,1
lghi $len,1
...
@@ -109,8 +129,11 @@ $code.=<<___ if(!$softonly);
...
@@ -109,8 +129,11 @@ $code.=<<___ if(!$softonly);
.
align
32
.
align
32
.
Lsoft_ghash:
.
Lsoft_ghash:
___
___
$cdoe
.=<<
___
if
(
$flavour
=~
/3[12]/
);
llgfr
$len
,
$len
___
$code
.=
<<___;
$code
.=
<<___;
stm
g %r6,%r14,48
($sp)
stm
${g} %r6,%r14,6*$SIZE_T
($sp)
aghi $Xi,-1
aghi $Xi,-1
srlg $len,$len,4
srlg $len,$len,4
...
@@ -209,7 +232,7 @@ $code.=<<___;
...
@@ -209,7 +232,7 @@ $code.=<<___;
xgr $Zhi,$tmp
xgr $Zhi,$tmp
stg $Zlo,8+1($Xi)
stg $Zlo,8+1($Xi)
stg $Zhi,0+1($Xi)
stg $Zhi,0+1($Xi)
lm
g %r6,%r14,48
($sp)
lm
${g} %r6,%r14,6*$SIZE_T
($sp)
br %r14
br %r14
.type gcm_ghash_4bit,\@function
.type gcm_ghash_4bit,\@function
.size gcm_ghash_4bit,(.-gcm_ghash_4bit)
.size gcm_ghash_4bit,(.-gcm_ghash_4bit)
...
...
crypto/rc4/asm/rc4-s390x.pl
浏览文件 @
e822c756
...
@@ -13,6 +13,26 @@
...
@@ -13,6 +13,26 @@
# "cluster" Address Generation Interlocks, so that one pipeline stall
# "cluster" Address Generation Interlocks, so that one pipeline stall
# resolves several dependencies.
# resolves several dependencies.
# November 2010.
#
# Adapt for -m31 build. If kernel supports what's called "highgprs"
# feature on Linux [see /proc/cpuinfo], it's possible to use 64-bit
# instructions and achieve "64-bit" performance even in 31-bit legacy
# application context. The feature is not specific to any particular
# processor, as long as it's "z-CPU". Latter implies that the code
# remains z/Architecture specific. On z990 it was measured to perform
# 50% better than code generated by gcc 4.3.
$flavour
=
shift
;
if
(
$flavour
=~
/3[12]/
)
{
$SIZE_T
=
4
;
$g
=
"";
}
else
{
$SIZE_T
=
8
;
$g
=
"
g
";
}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
open
STDOUT
,"
>
$output
";
open
STDOUT
,"
>
$output
";
...
@@ -42,7 +62,12 @@ $code.=<<___;
...
@@ -42,7 +62,12 @@ $code.=<<___;
.type RC4,\@function
.type RC4,\@function
.align 64
.align 64
RC4:
RC4:
stmg %r6,%r11,48($sp)
stm${g} %r6,%r11,6*$SIZE_T($sp)
___
$code
.=<<
___
if
(
$flavour
=~
/3[12]/
);
llgfr
$len
,
$len
___
$code
.=
<<___;
llgc $XX[0],0($key)
llgc $XX[0],0($key)
llgc $YY,1($key)
llgc $YY,1($key)
la $XX[0],1($XX[0])
la $XX[0],1($XX[0])
...
@@ -93,7 +118,7 @@ $code.=<<___;
...
@@ -93,7 +118,7 @@ $code.=<<___;
xgr $acc,$TX[1]
xgr $acc,$TX[1]
stg $acc,0($out)
stg $acc,0($out)
la $out,8($out)
la $out,8($out)
brct $cnt,.Loop8
brct
g
$cnt,.Loop8
.Lshort:
.Lshort:
lghi $acc,7
lghi $acc,7
...
@@ -125,7 +150,7 @@ $code.=<<___;
...
@@ -125,7 +150,7 @@ $code.=<<___;
ahi $XX[0],-1
ahi $XX[0],-1
stc $XX[0],0($key)
stc $XX[0],0($key)
stc $YY,1($key)
stc $YY,1($key)
lm
g %r6,%r11,48
($sp)
lm
${g} %r6,%r11,6*$SIZE_T
($sp)
br $rp
br $rp
.size RC4,.-RC4
.size RC4,.-RC4
.string "RC4 for s390x, CRYPTOGAMS by <appro\@openssl.org>"
.string "RC4 for s390x, CRYPTOGAMS by <appro\@openssl.org>"
...
@@ -150,7 +175,7 @@ $code.=<<___;
...
@@ -150,7 +175,7 @@ $code.=<<___;
.type RC4_set_key,\@function
.type RC4_set_key,\@function
.align 64
.align 64
RC4_set_key:
RC4_set_key:
stm
g %r6,%r8,48
($sp)
stm
${g} %r6,%r8,6*$SIZE_T
($sp)
lhi $cnt,256
lhi $cnt,256
la $idx,0(%r0)
la $idx,0(%r0)
sth $idx,0($key)
sth $idx,0($key)
...
@@ -183,7 +208,7 @@ RC4_set_key:
...
@@ -183,7 +208,7 @@ RC4_set_key:
la $iinp,0(%r0)
la $iinp,0(%r0)
j .L2ndloop
j .L2ndloop
.Ldone:
.Ldone:
lm
g %r6,%r8,48
($sp)
lm
${g} %r6,%r8,6*$SIZE_T
($sp)
br $rp
br $rp
.size RC4_set_key,.-RC4_set_key
.size RC4_set_key,.-RC4_set_key
...
...
crypto/s390xcpuid.S
浏览文件 @
e822c756
...
@@ -62,6 +62,9 @@ OPENSSL_wipe_cpu:
...
@@ -62,6 +62,9 @@ OPENSSL_wipe_cpu:
.
type
OPENSSL_cleanse
,@
function
.
type
OPENSSL_cleanse
,@
function
.
align
16
.
align
16
OPENSSL_cleanse
:
OPENSSL_cleanse
:
#if !defined(__s390x__) && !defined(__s390x)
llgfr
%
r3
,%
r3
#endif
lghi
%
r4
,
15
lghi
%
r4
,
15
lghi
%
r0
,
0
lghi
%
r0
,
0
clgr
%
r3
,%
r4
clgr
%
r3
,%
r4
...
...
crypto/sha/asm/sha1-s390x.pl
浏览文件 @
e822c756
...
@@ -21,8 +21,27 @@
...
@@ -21,8 +21,27 @@
# instructions to favour dual-issue z10 pipeline. On z10 hardware is
# instructions to favour dual-issue z10 pipeline. On z10 hardware is
# "only" ~2.3x faster than software.
# "only" ~2.3x faster than software.
# November 2010.
#
# Adapt for -m31 build. If kernel supports what's called "highgprs"
# feature on Linux [see /proc/cpuinfo], it's possible to use 64-bit
# instructions and achieve "64-bit" performance even in 31-bit legacy
# application context. The feature is not specific to any particular
# processor, as long as it's "z-CPU". Latter implies that the code
# remains z/Architecture specific.
$kimdfunc
=
1
;
# magic function code for kimd instruction
$kimdfunc
=
1
;
# magic function code for kimd instruction
$flavour
=
shift
;
if
(
$flavour
=~
/3[12]/
)
{
$SIZE_T
=
4
;
$g
=
"";
}
else
{
$SIZE_T
=
8
;
$g
=
"
g
";
}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
while
((
$output
=
shift
)
&&
(
$output
!~
/^\w[\w\-]*\.\w+$/
))
{}
open
STDOUT
,"
>
$output
";
open
STDOUT
,"
>
$output
";
...
@@ -42,13 +61,14 @@ $t1="%r11";
...
@@ -42,13 +61,14 @@ $t1="%r11";
@X
=
("
%r12
","
%r13
","
%r14
");
@X
=
("
%r12
","
%r13
","
%r14
");
$sp
=
"
%r15
";
$sp
=
"
%r15
";
$frame
=
160
+
16
*
4
;
$stdframe
=
16
*$SIZE_T
+
4
*
8
;
$frame
=
$stdframe
+
16
*
4
;
sub
Xupdate
{
sub
Xupdate
{
my
$i
=
shift
;
my
$i
=
shift
;
$code
.=<<
___
if
(
$i
==
15
);
$code
.=<<
___
if
(
$i
==
15
);
lg
$prefetch
,
160
(
$sp
)
### Xupdate(16) warm-up
lg
$prefetch
,
$stdframe
(
$sp
)
### Xupdate(16) warm-up
lr
$X
[
0
],
$X
[
2
]
lr
$X
[
0
],
$X
[
2
]
___
___
return
if
(
$i
&
1
);
# Xupdate is vectorized and executed every 2nd cycle
return
if
(
$i
&
1
);
# Xupdate is vectorized and executed every 2nd cycle
...
@@ -58,8 +78,8 @@ $code.=<<___ if ($i<16);
...
@@ -58,8 +78,8 @@ $code.=<<___ if ($i<16);
___
___
$code
.=<<
___
if
(
$i
>=
16
);
$code
.=<<
___
if
(
$i
>=
16
);
xgr
$X
[
0
],
$prefetch
### Xupdate($i)
xgr
$X
[
0
],
$prefetch
### Xupdate($i)
lg
$prefetch
,`
160
+4*((
$i
+2)%16)
`(
$sp
)
lg
$prefetch
,`
$stdframe
+4*((
$i
+2)%16)
`(
$sp
)
xg
$X
[
0
],`
160
+4*((
$i
+8)%16)
`(
$sp
)
xg
$X
[
0
],`
$stdframe
+4*((
$i
+8)%16)
`(
$sp
)
xgr
$X
[
0
],
$prefetch
xgr
$X
[
0
],
$prefetch
rll
$X
[
0
],
$X
[
0
],
1
rll
$X
[
0
],
$X
[
0
],
1
rllg
$X
[
1
],
$X
[
0
],
32
rllg
$X
[
1
],
$X
[
0
],
32
...
@@ -68,7 +88,7 @@ $code.=<<___ if ($i>=16);
...
@@ -68,7 +88,7 @@ $code.=<<___ if ($i>=16);
lr
$X
[
2
],
$X
[
1
]
# feedback
lr
$X
[
2
],
$X
[
1
]
# feedback
___
___
$code
.=<<
___
if
(
$i
<=
70
);
$code
.=<<
___
if
(
$i
<=
70
);
stg
$X
[
0
],`
160
+4*(
$i
%16)
`(
$sp
)
stg
$X
[
0
],`
$stdframe
+4*(
$i
%16)
`(
$sp
)
___
___
unshift
(
@X
,
pop
(
@X
));
unshift
(
@X
,
pop
(
@X
));
}
}
...
@@ -148,9 +168,9 @@ $code.=<<___ if ($kimdfunc);
...
@@ -148,9 +168,9 @@ $code.=<<___ if ($kimdfunc);
tmhl
%r0
,
0x4000
# check for message-security assist
tmhl
%r0
,
0x4000
# check for message-security assist
jz
.
Lsoftware
jz
.
Lsoftware
lghi
%r0
,
0
lghi
%r0
,
0
la
%r1
,
16
(
$sp
)
la
%r1
,
`
2*
$SIZE_T
`
(
$sp
)
.
long
0xb93e0002
# kimd %r0,%r2
.
long
0xb93e0002
# kimd %r0,%r2
lg
%r0
,
16
(
$sp
)
lg
%r0
,
`
2*
$SIZE_T
`
(
$sp
)
tmhh
%r0
,`
0x8000>>
$kimdfunc
`
tmhh
%r0
,`
0x8000>>
$kimdfunc
`
jz
.
Lsoftware
jz
.
Lsoftware
lghi
%r0
,
$kimdfunc
lghi
%r0
,
$kimdfunc
...
@@ -165,11 +185,11 @@ $code.=<<___ if ($kimdfunc);
...
@@ -165,11 +185,11 @@ $code.=<<___ if ($kimdfunc);
___
___
$code
.=
<<___;
$code
.=
<<___;
lghi %r1,-$frame
lghi %r1,-$frame
st
g $ctx,16
($sp)
st
${g} $ctx,`2*$SIZE_T`
($sp)
stm
g %r6,%r15,48
($sp)
stm
${g} %r6,%r15,`6*$SIZE_T`
($sp)
lgr %r0,$sp
lgr %r0,$sp
la $sp,0(%r1,$sp)
la $sp,0(%r1,$sp)
st
g
%r0,0($sp)
st
${g}
%r0,0($sp)
larl $t0,Ktable
larl $t0,Ktable
llgf $A,0($ctx)
llgf $A,0($ctx)
...
@@ -199,7 +219,7 @@ ___
...
@@ -199,7 +219,7 @@ ___
for
(;
$i
<
80
;
$i
++
)
{
&BODY_20_39
(
$i
,
@V
);
unshift
(
@V
,
pop
(
@V
));
}
for
(;
$i
<
80
;
$i
++
)
{
&BODY_20_39
(
$i
,
@V
);
unshift
(
@V
,
pop
(
@V
));
}
$code
.=
<<___;
$code
.=
<<___;
l
g $ctx,`$frame+16
`($sp)
l
${g} $ctx,`$frame+2*$SIZE_T
`($sp)
la $inp,64($inp)
la $inp,64($inp)
al $A,0($ctx)
al $A,0($ctx)
al $B,4($ctx)
al $B,4($ctx)
...
@@ -211,9 +231,9 @@ $code.=<<___;
...
@@ -211,9 +231,9 @@ $code.=<<___;
st $C,8($ctx)
st $C,8($ctx)
st $D,12($ctx)
st $D,12($ctx)
st $E,16($ctx)
st $E,16($ctx)
brct
$len,.Lloop
brct
${g}
$len,.Lloop
lm
g %r6,%r15,`$frame+48
`($sp)
lm
${g} %r6,%r15,`$frame+6*$SIZE_T
`($sp)
br %r14
br %r14
.size sha1_block_data_order,.-sha1_block_data_order
.size sha1_block_data_order,.-sha1_block_data_order
.string "SHA1 block transform for s390x, CRYPTOGAMS by <appro\@openssl.org>"
.string "SHA1 block transform for s390x, CRYPTOGAMS by <appro\@openssl.org>"
...
...
crypto/sha/asm/sha512-s390x.pl
浏览文件 @
e822c756
...
@@ -26,6 +26,26 @@
...
@@ -26,6 +26,26 @@
# favour dual-issue z10 pipeline. Hardware SHA256/512 is ~4.7x faster
# favour dual-issue z10 pipeline. Hardware SHA256/512 is ~4.7x faster
# than software.
# than software.
# November 2010.
#
# Adapt for -m31 build. If kernel supports what's called "highgprs"
# feature on Linux [see /proc/cpuinfo], it's possible to use 64-bit
# instructions and achieve "64-bit" performance even in 31-bit legacy
# application context. The feature is not specific to any particular
# processor, as long as it's "z-CPU". Latter implies that the code
# remains z/Architecture specific. On z900 SHA256 was measured to
# perform 2.4x and SHA512 - 13x better than code generated by gcc 4.3.
$flavour
=
shift
;
if
(
$flavour
=~
/3[12]/
)
{
$SIZE_T
=
4
;
$g
=
"";
}
else
{
$SIZE_T
=
8
;
$g
=
"
g
";
}
$t0
=
"
%r0
";
$t0
=
"
%r0
";
$t1
=
"
%r1
";
$t1
=
"
%r1
";
$ctx
=
"
%r2
";
$t2
=
"
%r2
";
$ctx
=
"
%r2
";
$t2
=
"
%r2
";
...
@@ -78,7 +98,8 @@ if ($output =~ /512/) {
...
@@ -78,7 +98,8 @@ if ($output =~ /512/) {
}
}
$Func
=
"
sha
${label}
_block_data_order
";
$Func
=
"
sha
${label}
_block_data_order
";
$Table
=
"
K
${label}
";
$Table
=
"
K
${label}
";
$frame
=
160
+
16
*$SZ
;
$stdframe
=
16
*$SIZE_T
+
4
*
8
;
$frame
=
$stdframe
+
16
*$SZ
;
sub
BODY_00_15
{
sub
BODY_00_15
{
my
(
$i
,
$a
,
$b
,
$c
,
$d
,
$e
,
$f
,
$g
,
$h
)
=
@_
;
my
(
$i
,
$a
,
$b
,
$c
,
$d
,
$e
,
$f
,
$g
,
$h
)
=
@_
;
...
@@ -93,9 +114,9 @@ $code.=<<___;
...
@@ -93,9 +114,9 @@ $code.=<<___;
xgr $t0,$t1
xgr $t0,$t1
$ROT $t1,$t1,`$Sigma1[2]-$Sigma1[1]`
$ROT $t1,$t1,`$Sigma1[2]-$Sigma1[1]`
xgr $t2,$g
xgr $t2,$g
$ST $T1,`
160
+$SZ*($i%16)`($sp)
$ST $T1,`
$stdframe
+$SZ*($i%16)`($sp)
xgr $t0,$t1 # Sigma1(e)
xgr $t0,$t1 # Sigma1(e)
la $T1,0($T1,$h)
# T1+=h
algr $T1,$h
# T1+=h
ngr $t2,$e
ngr $t2,$e
lgr $t1,$a
lgr $t1,$a
algr $T1,$t0 # T1+=Sigma1(e)
algr $T1,$t0 # T1+=Sigma1(e)
...
@@ -113,7 +134,7 @@ $code.=<<___;
...
@@ -113,7 +134,7 @@ $code.=<<___;
ngr $t2,$b
ngr $t2,$b
algr $h,$T1 # h+=T1
algr $h,$T1 # h+=T1
ogr $t2,$t1 # Maj(a,b,c)
ogr $t2,$t1 # Maj(a,b,c)
la $d,0($d,$T1)
# d+=T1
algr $d,$T1
# d+=T1
algr $h,$t2 # h+=Maj(a,b,c)
algr $h,$t2 # h+=Maj(a,b,c)
___
___
}
}
...
@@ -122,19 +143,19 @@ sub BODY_16_XX {
...
@@ -122,19 +143,19 @@ sub BODY_16_XX {
my
(
$i
,
$a
,
$b
,
$c
,
$d
,
$e
,
$f
,
$g
,
$h
)
=
@_
;
my
(
$i
,
$a
,
$b
,
$c
,
$d
,
$e
,
$f
,
$g
,
$h
)
=
@_
;
$code
.=
<<___;
$code
.=
<<___;
$LD $T1,`
160
+$SZ*(($i+1)%16)`($sp) ### $i
$LD $T1,`
$stdframe
+$SZ*(($i+1)%16)`($sp) ### $i
$LD $t1,`
160
+$SZ*(($i+14)%16)`($sp)
$LD $t1,`
$stdframe
+$SZ*(($i+14)%16)`($sp)
$ROT $t0,$T1,$sigma0[0]
$ROT $t0,$T1,$sigma0[0]
$SHR $T1,$sigma0[2]
$SHR $T1,$sigma0[2]
$ROT $t2,$t0,`$sigma0[1]-$sigma0[0]`
$ROT $t2,$t0,`$sigma0[1]-$sigma0[0]`
xgr $T1,$t0
xgr $T1,$t0
$ROT $t0,$t1,$sigma1[0]
$ROT $t0,$t1,$sigma1[0]
xgr $T1,$t2 # sigma0(X[i+1])
xgr $T1,$t2
# sigma0(X[i+1])
$SHR $t1,$sigma1[2]
$SHR $t1,$sigma1[2]
$ADD $T1,`
160
+$SZ*($i%16)`($sp) # +=X[i]
$ADD $T1,`
$stdframe
+$SZ*($i%16)`($sp) # +=X[i]
xgr $t1,$t0
xgr $t1,$t0
$ROT $t0,$t0,`$sigma1[1]-$sigma1[0]`
$ROT $t0,$t0,`$sigma1[1]-$sigma1[0]`
$ADD $T1,`
160
+$SZ*(($i+9)%16)`($sp) # +=X[i+9]
$ADD $T1,`
$stdframe
+$SZ*(($i+9)%16)`($sp) # +=X[i+9]
xgr $t1,$t0 # sigma1(X[i+14])
xgr $t1,$t0 # sigma1(X[i+14])
algr $T1,$t1 # +=sigma1(X[i+14])
algr $T1,$t1 # +=sigma1(X[i+14])
___
___
...
@@ -212,6 +233,7 @@ $code.=<<___;
...
@@ -212,6 +233,7 @@ $code.=<<___;
.globl $Func
.globl $Func
.type $Func,\@function
.type $Func,\@function
$Func:
$Func:
sllg $len,$len,`log(16*$SZ)/log(2)`
___
___
$code
.=<<
___
if
(
$kimdfunc
);
$code
.=<<
___
if
(
$kimdfunc
);
larl
%r1
,
OPENSSL_s390xcap_P
larl
%r1
,
OPENSSL_s390xcap_P
...
@@ -219,15 +241,15 @@ $code.=<<___ if ($kimdfunc);
...
@@ -219,15 +241,15 @@ $code.=<<___ if ($kimdfunc);
tmhl
%r0
,
0x4000
# check for message-security assist
tmhl
%r0
,
0x4000
# check for message-security assist
jz
.
Lsoftware
jz
.
Lsoftware
lghi
%r0
,
0
lghi
%r0
,
0
la
%r1
,
16
(
$sp
)
la
%r1
,
`
2*
$SIZE_T
`
(
$sp
)
.
long
0xb93e0002
# kimd %r0,%r2
.
long
0xb93e0002
# kimd %r0,%r2
lg
%r0
,
16
(
$sp
)
lg
%r0
,
`
2*
$SIZE_T
`
(
$sp
)
tmhh
%r0
,`
0x8000>>
$kimdfunc
`
tmhh
%r0
,`
0x8000>>
$kimdfunc
`
jz
.
Lsoftware
jz
.
Lsoftware
lghi
%r0
,
$kimdfunc
lghi
%r0
,
$kimdfunc
lgr
%r1
,
$ctx
lgr
%r1
,
$ctx
lgr
%r2
,
$inp
lgr
%r2
,
$inp
sllg
%r3
,
$len
,`
log(16*
$SZ
)/log(2)
`
lgr
%r3
,
$len
.
long
0xb93e0002
# kimd %r0,%r2
.
long
0xb93e0002
# kimd %r0,%r2
brc
1
,
.-
4
# pay attention to "partial completion"
brc
1
,
.-
4
# pay attention to "partial completion"
br
%r14
br
%r14
...
@@ -235,13 +257,12 @@ $code.=<<___ if ($kimdfunc);
...
@@ -235,13 +257,12 @@ $code.=<<___ if ($kimdfunc);
.
Lsoftware:
.
Lsoftware:
___
___
$code
.=
<<___;
$code
.=
<<___;
sllg $len,$len,`log(16*$SZ)/log(2)`
lghi %r1,-$frame
lghi %r1,-$frame
agr $len,$inp
la $len,0($len,$inp)
stm
g $ctx,%r15,16
($sp)
stm
${g} $ctx,%r15,`2*$SIZE_T`
($sp)
lgr %r0,$sp
lgr %r0,$sp
la $sp,0(%r1,$sp)
la $sp,0(%r1,$sp)
st
g
%r0,0($sp)
st
${g}
%r0,0($sp)
larl $tbl,$Table
larl $tbl,$Table
$LD $A,`0*$SZ`($ctx)
$LD $A,`0*$SZ`($ctx)
...
@@ -265,7 +286,7 @@ $code.=<<___;
...
@@ -265,7 +286,7 @@ $code.=<<___;
clgr $len,$t0
clgr $len,$t0
jne .Lrounds_16_xx
jne .Lrounds_16_xx
l
g $ctx,`$frame+16
`($sp)
l
${g} $ctx,`$frame+2*$SIZE_T
`($sp)
la $inp,`16*$SZ`($inp)
la $inp,`16*$SZ`($inp)
$ADD $A,`0*$SZ`($ctx)
$ADD $A,`0*$SZ`($ctx)
$ADD $B,`1*$SZ`($ctx)
$ADD $B,`1*$SZ`($ctx)
...
@@ -283,10 +304,10 @@ $code.=<<___;
...
@@ -283,10 +304,10 @@ $code.=<<___;
$ST $F,`5*$SZ`($ctx)
$ST $F,`5*$SZ`($ctx)
$ST $G,`6*$SZ`($ctx)
$ST $G,`6*$SZ`($ctx)
$ST $H,`7*$SZ`($ctx)
$ST $H,`7*$SZ`($ctx)
cl
g $inp,`$frame+32
`($sp)
cl
${g} $inp,`$frame+4*$SIZE_T
`($sp)
jne .Lloop
jne .Lloop
lm
g %r6,%r15,`$frame+48
`($sp)
lm
${g} %r6,%r15,`$frame+6*$SIZE_T
`($sp)
br %r14
br %r14
.size $Func,.-$Func
.size $Func,.-$Func
.string "SHA${label} block transform for s390x, CRYPTOGAMS by <appro\@openssl.org>"
.string "SHA${label} block transform for s390x, CRYPTOGAMS by <appro\@openssl.org>"
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录