Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenHarmony
Third Party Openssl
提交
24798c5e
T
Third Party Openssl
项目概览
OpenHarmony
/
Third Party Openssl
11 个月 前同步成功
通知
8
Star
18
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
T
Third Party Openssl
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
24798c5e
编写于
11月 05, 2012
作者:
A
Andy Polyakov
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
ghash-sparcv9.pl: 22% improvement on T4.
上级
5b0e3daf
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
154 addition
and
118 deletion
+154
-118
crypto/modes/asm/ghash-sparcv9.pl
crypto/modes/asm/ghash-sparcv9.pl
+152
-118
crypto/modes/gcm128.c
crypto/modes/gcm128.c
+2
-0
未找到文件。
crypto/modes/asm/ghash-sparcv9.pl
浏览文件 @
24798c5e
...
...
@@ -41,8 +41,10 @@
#
# Add VIS3 lookup-table-free implementation using polynomial
# multiplication xmulx[hi] and extended addition addxc[cc]
# instructions. 3.96/6.26x improvement on T3/T4 or in absolute
# terms 9.02/2.61 cycles per byte.
# instructions. 4.22/7.63x improvement on T3/T4 or in absolute
# terms 8.45/2.14 cycles per byte. On T4 multi-process benchmark
# saturates at ~15x single-process result on 8-core processor, or
# ~19.7GBps per 2.85GHz socket.
$bits
=
32
;
for
(
@ARGV
)
{
$bits
=
64
if
(
/\-m64/
||
/\-xarch\=v9/
);
}
...
...
@@ -335,74 +337,103 @@ gcm_gmult_4bit:
___
{{{
# Straightforward
64-bits-at-a-time approach with pair of 128x64-bit
#
multiplications followed by 64-bit reductions. While it might be
#
suboptimal with regard to sheer amount of multiplications, other
# m
ethods would require larger amount of 64-bit registers, which we
#
don't have in 32-bit application. Also, they [alternative methods
#
such as aggregated reduction] kind of thrive on fast 128-bit SIMD
#
instructions and these are not option on SPARC..
.
# Straightforward
128x128-bit multiplication using Karatsuba algorithm
#
followed by pair of 64-bit reductions [with a shortcut in first one,
#
which allowed to break dependency between reductions and remove one
# m
ulitplication from critical path]. While it might be suboptimal
#
with regard to sheer number of multiplications, other methods [such
#
as aggregate reduction] would require more 64-bit registers, which
#
we don't have in 32-bit application context
.
(
$Xip
,
$Htable
,
$inp
,
$len
)
=
map
("
%i
$_
",(
0
..
3
));
(
$xE1
,
$Hhi
,
$Hlo
,
$Rhi
,
$Rlo
,
$M0hi
,
$M0lo
,
$M1hi
,
$M1lo
,
$Zhi
,
$Zlo
,
$X
)
=
(
map
("
%g
$_
",(
1
..
5
)),
map
("
%o
$_
",(
0
..
5
,
7
)));
(
$shl
,
$shr
)
=
map
("
%l
$_
",(
0
..
7
));
(
$Hhl
,
$Hlo
,
$Hhi
,
$Xlo
,
$Xhi
,
$xE1
,
$x384
,
$C0
,
$C1
,
$C2
,
$C3
,
$V
)
=
(
map
("
%o
$_
",(
0
..
5
,
7
)),
map
("
%g
$_
",(
1
..
5
)));
(
$shl
,
$shr
,
$sqr
)
=
map
("
%l
$_
",(
0
..
7
));
# For details regarding "twisted H" see ghash-x86.pl.
$code
.=
<<___;
.globl gcm_
gmul
t_vis3
.globl gcm_
ini
t_vis3
.align 32
gcm_
gmul
t_vis3:
gcm_
ini
t_vis3:
save %sp,-$frame,%sp
ldx [$Xip+8],$X ! load X.lo
ldx [$Htable-8], $Hlo ! load H
ldx [$Htable-16],$Hhi
mov 0xE1,$xE1
sllx $xE1,57,$xE1
xmulx $X,$Hlo,$M0lo ! HX.lo
xmulxhi $X,$Hlo,$M0hi
xmulx $X,$Hhi,$M1lo
xmulxhi $X,$Hhi,$M1hi
ldx [$Xip+0],$X ! load X.hi
addcc $M0lo,$M0lo,$M0lo ! (HX.lo)<<1
xor $M0hi,$M1lo,$M1lo
xmulx $xE1,$M0lo,$Rlo ! res=Z.lo(0xE1<<57)
xmulxhi $xE1,$M0lo,$Rhi
addxccc $M1lo,$M1lo,$Zlo ! Z=((HX.lo)<<1)>>64
addxc $M1hi,$M1hi,$Zhi
xor $M0lo,$Zhi,$Zhi ! overflow bit from 0xE1<<57
xmulx $X,$Hlo,$M0lo ! HX.hi
xmulxhi $X,$Hlo,$M0hi
xmulx $X,$Hhi,$M1lo
xmulxhi $X,$Hhi,$M1hi
ldx [%i1+0],$Hhi
ldx [%i1+8],$Hlo
mov 0xE1,$Xhi
mov 1,$Xlo
sllx $Xhi,57,$Xhi
srax $Hhi,63,$C0 ! carry
addcc $Hlo,$Hlo,$Hlo ! H<<=1
addxc $Hhi,$Hhi,$Hhi
and $Xlo,$C0,$Xlo
and $Xhi,$C0,$Xhi
xor $Xlo,$Hlo,$Hlo
xor $Xhi,$Hhi,$Hhi
stx $Hlo,[%i0+8] ! save twisted H
stx $Hhi,[%i0+0]
xor $Rlo,$Zlo,$Zlo ! Z^=res
xor $Rhi,$Zhi,$Zhi
addcc $M0lo,$M0lo,$M0lo ! (HX.lo)<<1
xor $Zlo, $M0lo,$M0lo
xor $M0hi,$M1lo,$M1lo
xmulx $xE1,$M0lo,$Rlo ! res=Z.lo(0xE1<<57)
xmulxhi $xE1,$M0lo,$Rhi
addxccc $M1lo,$M1lo,$M1lo
addxc $M1hi,$M1hi,$M1hi
xor $M1lo,$Zhi,$Zlo ! Z=(Z^(HX.hi)<<1)>>64
xor $M0lo,$M1hi,$Zhi ! overflow bit from 0xE1<<57
ret
restore
.type gcm_init_vis3,#function
.size gcm_init_vis3,.-gcm_init_vis3
xor $Rlo,$Zlo,$Zlo ! Z^=res
xor $Rhi,$Zhi,$Zhi
.globl gcm_gmult_vis3
.align 32
gcm_gmult_vis3:
save %sp,-$frame,%sp
stx $Zlo,[$Xip+8] ! save Xi
stx $Zhi,[$Xip+0]
ldx [$Xip+8],$Xlo ! load Xi
ldx [$Xip+0],$Xhi
ldx [$Htable+8],$Hlo ! load twisted H
ldx [$Htable+0],$Hhi
sethi %hi(0xA0406080),$V
sethi %hi(0x20C0E000),%l0
or $V,%lo(0xA0406080),$V
or %l0,%lo(0x20C0E000),%l0
sllx $V,32,$V
mov 0xE1,%l1
or %l0,$V,$V ! (0xE0i)&0xff=0xA040608020C0E000
sllx %l1,57,$xE1 ! 57 is not a typo
sllx %l1,50,$x384
xor $Hhi,$Hlo,$Hhl ! Karatsuba pre-processing
xmulx $Xlo,$Hlo,$C0
xor $Xlo,$Xhi,$C2 ! Karatsuba pre-processing
xmulx $C2,$Hhl,$C1
xmulxhi $Xlo,$Hlo,$Xlo
xmulxhi $C2,$Hhl,$C2
xmulxhi $Xhi,$Hhi,$C3
xmulx $Xhi,$Hhi,$Xhi
sll $C0,3,$sqr
srlx $V,$sqr,$sqr ! 0xE0 [implicit &(7<<3)]
xor $C0,$sqr,$sqr
and $sqr,0x7f,$sqr
xor $C0,$C1,$C1 ! Karatsuba post-processing
xor $Xlo,$C2,$C2
xor $Xhi,$C1,$C1
xor $C3,$C2,$C2
xor $Xlo,$C1,$C1
xmulxhi $C0,$xE1,$Xlo ! 0xE1<<1<<56
xor $Xhi,$C2,$C2
xmulx $sqr,$x384,$Xhi ! 0xE1<<2<<48
xor $C0,$C2,$C2
xmulx $C1,$xE1,$C0
xor $C1,$C3,$C3
xmulxhi $C1,$xE1,$C1
xor $Xlo,$C2,$C2
xor $Xhi,$C3,$C3
xor $C0,$C2,$C2
xor $C1,$C3,$C3
stx $C2,[$Xip+8] ! save Xi
stx $C3,[$Xip+0]
ret
restore
...
...
@@ -414,80 +445,83 @@ gcm_gmult_vis3:
gcm_ghash_vis3:
save %sp,-$frame,%sp
ldx [$Xip+0],$Zhi ! load X.hi
ldx [$Xip+8],$Zlo ! load X.lo
ldx [$Xip+8],$C2 ! load Xi
ldx [$Xip+0],$C3
ldx [$Htable+8],$Hlo ! load twisted H
ldx [$Htable+0],$Hhi
sethi %hi(0xA0406080),$V
sethi %hi(0x20C0E000),%l6
or $V,%lo(0xA0406080),$V
or %l6,%lo(0x20C0E000),%l6
sllx $V,32,$V
mov 0xE1,%l7
or %l6,$V,$V ! (0xE0i)&0xff=0xA040608020C0E000
sllx %l7,57,$xE1 ! 57 is not a typo
sllx %l7,50,$x384
and $inp,7,$shl
andn $inp,7,$inp
ldx [$Htable-8], $Hlo ! load H
ldx [$Htable-16],$Hhi
sll $shl,3,$shl
prefetch [$inp+63], 20
mov 0xE1,$xE1
sub %g0,$shl,$shr
sllx $xE1,57,$xE1
xor $Hhi,$Hlo,$Hhl ! Karatsuba pre-processing
.Loop:
ldx [$inp+8],$
Rlo ! load *inp
ldx [$inp+8],$
Xlo
brz,pt $shl,1f
ldx [$inp+0],$Rhi
ldx [$inp+16],$X ! align data
srlx $Rlo,$shr,$M0lo
sllx $Rlo,$shl,$Rlo
sllx $Rhi,$shl,$Rhi
srlx $X,$shr,$X
or $M0lo,$Rhi,$Rhi
or $X,$Rlo,$Rlo
ldx [$inp+0],$Xhi
ldx [$inp+16],$C1 ! align data
srlx $Xlo,$shr,$C0
sllx $Xlo,$shl,$Xlo
sllx $Xhi,$shl,$Xhi
srlx $C1,$shr,$C1
or $C0,$Xhi,$Xhi
or $C1,$Xlo,$Xlo
1:
add $inp,16,$inp
sub $len,16,$len
xor $Rlo,$Zlo,$X
xor $C2,$Xlo,$Xlo
xor $C3,$Xhi,$Xhi
prefetch [$inp+63], 20
xmulx $X,$Hlo,$M0lo ! HX.lo
xmulxhi $X,$Hlo,$M0hi
xmulx $X,$Hhi,$M1lo
xmulxhi $X,$Hhi,$M1hi
xor $Rhi,$Zhi,$X
addcc $M0lo,$M0lo,$M0lo ! (HX.lo)<<1
xor $M0hi,$M1lo,$M1lo
xmulx $xE1,$M0lo,$Rlo ! res=Z.lo(0xE1<<57)
xmulxhi $xE1,$M0lo,$Rhi
addxccc $M1lo,$M1lo,$Zlo ! Z=((HX.lo)<<1)>>64
addxc $M1hi,$M1hi,$Zhi
xor $M0lo,$Zhi,$Zhi ! overflow bit from 0xE1<<57
xmulx $X,$Hlo,$M0lo ! HX.hi
xmulxhi $X,$Hlo,$M0hi
xmulx $X,$Hhi,$M1lo
xmulxhi $X,$Hhi,$M1hi
xor $Rlo,$Zlo,$Zlo ! Z^=res
xor $Rhi,$Zhi,$Zhi
addcc $M0lo,$M0lo,$M0lo ! (HX.lo)<<1
xor $Zlo, $M0lo,$M0lo
xor $M0hi,$M1lo,$M1lo
xmulx $xE1,$M0lo,$Rlo ! res=Z.lo(0xE1<<57)
xmulxhi $xE1,$M0lo,$Rhi
addxccc $M1lo,$M1lo,$M1lo
addxc $M1hi,$M1hi,$M1hi
xor $M1lo,$Zhi,$Zlo ! Z=(Z^(HX.hi)<<1)>>64
xor $M0lo,$M1hi,$Zhi ! overflow bit from 0xE1<<57
xor $Rlo,$Zlo,$Zlo ! Z^=res
xmulx $Xlo,$Hlo,$C0
xor $Xlo,$Xhi,$C2 ! Karatsuba pre-processing
xmulx $C2,$Hhl,$C1
xmulxhi $Xlo,$Hlo,$Xlo
xmulxhi $C2,$Hhl,$C2
xmulxhi $Xhi,$Hhi,$C3
xmulx $Xhi,$Hhi,$Xhi
sll $C0,3,$sqr
srlx $V,$sqr,$sqr ! 0xE0 [implicit &(7<<3)]
xor $C0,$sqr,$sqr
and $sqr,0x7f,$sqr
xor $C0,$C1,$C1 ! Karatsuba post-processing
xor $Xlo,$C2,$C2
xor $Xhi,$C1,$C1
xor $C3,$C2,$C2
xor $Xlo,$C1,$C1
xmulxhi $C0,$xE1,$Xlo ! 0xE1<<1<<56
xor $Xhi,$C2,$C2
xmulx $sqr,$x384,$Xhi ! 0xE1<<2<<48
xor $C0,$C2,$C2
xmulx $C1,$xE1,$C0
xor $C1,$C3,$C3
xmulxhi $C1,$xE1,$C1
xor $Xlo,$C2,$C2
xor $Xhi,$C3,$C3
xor $C0,$C2,$C2
brnz,pt $len,.Loop
xor $
Rhi,$Zhi,$Zhi
xor $
C1,$C3,$C3
stx $
Zlo
,[$Xip+8] ! save Xi
stx $
Zhi
,[$Xip+0]
stx $
C2
,[$Xip+8] ! save Xi
stx $
C3
,[$Xip+0]
ret
restore
...
...
crypto/modes/gcm128.c
浏览文件 @
24798c5e
...
...
@@ -679,6 +679,7 @@ void gcm_ghash_neon(u64 Xi[2],const u128 Htable[16],const u8 *inp,size_t len);
# define GHASH_ASM_SPARC
# define GCM_FUNCREF_4BIT
extern
unsigned
int
OPENSSL_sparcv9cap_P
[];
void
gcm_init_vis3
(
u128
Htable
[
16
],
const
u64
Xi
[
2
]);
void
gcm_gmult_vis3
(
u64
Xi
[
2
],
const
u128
Htable
[
16
]);
void
gcm_ghash_vis3
(
u64
Xi
[
2
],
const
u128
Htable
[
16
],
const
u8
*
inp
,
size_t
len
);
# endif
...
...
@@ -759,6 +760,7 @@ void CRYPTO_gcm128_init(GCM128_CONTEXT *ctx,void *key,block128_f block)
}
# elif defined(GHASH_ASM_SPARC)
if
(
OPENSSL_sparcv9cap_P
[
0
]
&
SPARCV9_VIS3
)
{
gcm_init_vis3
(
ctx
->
Htable
,
ctx
->
H
.
u
);
ctx
->
gmult
=
gcm_gmult_vis3
;
ctx
->
ghash
=
gcm_ghash_vis3
;
}
else
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录