Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
OpenHarmony
Third Party Openssl
提交
b83d09f5
T
Third Party Openssl
项目概览
OpenHarmony
/
Third Party Openssl
大约 1 年 前同步成功
通知
9
Star
18
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
T
Third Party Openssl
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
b83d09f5
编写于
5月 23, 2014
作者:
A
Andy Polyakov
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
vpaes-ppc.pl: comply with ABI.
上级
27da9394
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
114 addition
and
109 deletion
+114
-109
crypto/aes/asm/vpaes-ppc.pl
crypto/aes/asm/vpaes-ppc.pl
+114
-109
未找到文件。
crypto/aes/asm/vpaes-ppc.pl
浏览文件 @
b83d09f5
...
@@ -35,12 +35,14 @@ if ($flavour =~ /64/) {
...
@@ -35,12 +35,14 @@ if ($flavour =~ /64/) {
$STU
=
"
stdu
";
$STU
=
"
stdu
";
$POP
=
"
ld
";
$POP
=
"
ld
";
$PUSH
=
"
std
";
$PUSH
=
"
std
";
$UCMP
=
"
cmpld
";
}
elsif
(
$flavour
=~
/32/
)
{
}
elsif
(
$flavour
=~
/32/
)
{
$SIZE_T
=
4
;
$SIZE_T
=
4
;
$LRSAVE
=
$SIZE_T
;
$LRSAVE
=
$SIZE_T
;
$STU
=
"
stwu
";
$STU
=
"
stwu
";
$POP
=
"
lwz
";
$POP
=
"
lwz
";
$PUSH
=
"
stw
";
$PUSH
=
"
stw
";
$UCMP
=
"
cmplw
";
}
else
{
die
"
nonsense
$flavour
";
}
}
else
{
die
"
nonsense
$flavour
";
}
$sp
=
"
r1
";
$sp
=
"
r1
";
...
@@ -302,28 +304,28 @@ Lenc_entry:
...
@@ -302,28 +304,28 @@ Lenc_entry:
mflr r6
mflr r6
mfspr r7, 256 # save vrsave
mfspr r7, 256 # save vrsave
stvx v20,r10,$sp
stvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v21,r11,$sp
stvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v22,r10,$sp
stvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v23,r11,$sp
stvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v24,r10,$sp
stvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v25,r11,$sp
stvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v26,r10,$sp
stvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v27,r11,$sp
stvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v28,r10,$sp
stvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v29,r11,$sp
stvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v30,r10,$sp
stvx v30,r10,$sp
stvx v31,r11,$sp
stvx v31,r11,$sp
lwz
r7,`$FRAME-4`($sp) # save vrsave
stw
r7,`$FRAME-4`($sp) # save vrsave
li r0, -1
li r0, -1
$PUSH r6,`$FRAME+$LRSAVE`($sp)
$PUSH r6,`$FRAME+$LRSAVE`($sp)
mtspr 256, r0 # preserve all AltiVec registers
mtspr 256, r0 # preserve all AltiVec registers
...
@@ -359,25 +361,25 @@ Lenc_entry:
...
@@ -359,25 +361,25 @@ Lenc_entry:
mtlr r6
mtlr r6
mtspr 256, r7 # restore vrsave
mtspr 256, r7 # restore vrsave
lvx v20,r10,$sp
lvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v21,r11,$sp
lvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v22,r10,$sp
lvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v23,r11,$sp
lvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v24,r10,$sp
lvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v25,r11,$sp
lvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v26,r10,$sp
lvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v27,r11,$sp
lvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v28,r10,$sp
lvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v29,r11,$sp
lvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v30,r10,$sp
lvx v30,r10,$sp
lvx v31,r11,$sp
lvx v31,r11,$sp
addi $sp,$sp,$FRAME
addi $sp,$sp,$FRAME
...
@@ -531,28 +533,28 @@ Ldec_entry:
...
@@ -531,28 +533,28 @@ Ldec_entry:
mflr r6
mflr r6
mfspr r7, 256 # save vrsave
mfspr r7, 256 # save vrsave
stvx v20,r10,$sp
stvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v21,r11,$sp
stvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v22,r10,$sp
stvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v23,r11,$sp
stvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v24,r10,$sp
stvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v25,r11,$sp
stvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v26,r10,$sp
stvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v27,r11,$sp
stvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v28,r10,$sp
stvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v29,r11,$sp
stvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v30,r10,$sp
stvx v30,r10,$sp
stvx v31,r11,$sp
stvx v31,r11,$sp
lwz
r7,`$FRAME-4`($sp) # save vrsave
stw
r7,`$FRAME-4`($sp) # save vrsave
li r0, -1
li r0, -1
$PUSH r6,`$FRAME+$LRSAVE`($sp)
$PUSH r6,`$FRAME+$LRSAVE`($sp)
mtspr 256, r0 # preserve all AltiVec registers
mtspr 256, r0 # preserve all AltiVec registers
...
@@ -588,25 +590,25 @@ Ldec_entry:
...
@@ -588,25 +590,25 @@ Ldec_entry:
mtlr r6
mtlr r6
mtspr 256, r7 # restore vrsave
mtspr 256, r7 # restore vrsave
lvx v20,r10,$sp
lvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v21,r11,$sp
lvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v22,r10,$sp
lvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v23,r11,$sp
lvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v24,r10,$sp
lvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v25,r11,$sp
lvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v26,r10,$sp
lvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v27,r11,$sp
lvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v28,r10,$sp
lvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v29,r11,$sp
lvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v30,r10,$sp
lvx v30,r10,$sp
lvx v31,r11,$sp
lvx v31,r11,$sp
addi $sp,$sp,$FRAME
addi $sp,$sp,$FRAME
...
@@ -619,40 +621,43 @@ Ldec_entry:
...
@@ -619,40 +621,43 @@ Ldec_entry:
.globl .vpaes_cbc_encrypt
.globl .vpaes_cbc_encrypt
.align 5
.align 5
.vpaes_cbc_encrypt:
.vpaes_cbc_encrypt:
${UCMP}i r5,16
bltlr-
$STU $sp,-`($FRAME+2*$SIZE_T)`($sp)
$STU $sp,-`($FRAME+2*$SIZE_T)`($sp)
mflr r0
mflr r0
li r10,`15+6*$SIZE_T`
li r10,`15+6*$SIZE_T`
li r11,`31+6*$SIZE_T`
li r11,`31+6*$SIZE_T`
mfspr r12, 256
mfspr r12, 256
stvx v20,r10,$sp
stvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v21,r11,$sp
stvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v22,r10,$sp
stvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v23,r11,$sp
stvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v24,r10,$sp
stvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v25,r11,$sp
stvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v26,r10,$sp
stvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v27,r11,$sp
stvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v28,r10,$sp
stvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v29,r11,$sp
stvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v30,r10,$sp
stvx v30,r10,$sp
stvx v31,r11,$sp
stvx v31,r11,$sp
lwz
r12,`$FRAME-4`($sp) # save vrsave
stw
r12,`$FRAME-4`($sp) # save vrsave
$PUSH r30,`$FRAME+$SIZE_T*0`($sp)
$PUSH r30,`$FRAME+$SIZE_T*0`($sp)
$PUSH r31,`$FRAME+$SIZE_T*1`($sp)
$PUSH r31,`$FRAME+$SIZE_T*1`($sp)
li r9, 16
li r9,
-
16
$PUSH r0, `$FRAME+$SIZE_T*2+$LRSAVE`($sp)
$PUSH r0, `$FRAME+$SIZE_T*2+$LRSAVE`($sp)
sub. r30, r5, r9 # copy length
-16
and r30, r5, r9 # copy length&
-16
mr r5, r6 # copy pointer to key
mr r5, r6 # copy pointer to key
mr r31, r7 # copy pointer to iv
mr r31, r7 # copy pointer to iv
blt Lcbc_abort
blt Lcbc_abort
...
@@ -699,7 +704,7 @@ Lcbc_enc_loop:
...
@@ -699,7 +704,7 @@ Lcbc_enc_loop:
vmr $outhead, v0
vmr $outhead, v0
stvx v1, 0, $out
stvx v1, 0, $out
addi $out, $out, 16
addi $out, $out, 16
b
g
e Lcbc_enc_loop
b
n
e Lcbc_enc_loop
b Lcbc_done
b Lcbc_done
...
@@ -725,7 +730,7 @@ Lcbc_dec_loop:
...
@@ -725,7 +730,7 @@ Lcbc_dec_loop:
vmr $outhead, v0
vmr $outhead, v0
stvx v1, 0, $out
stvx v1, 0, $out
addi $out, $out, 16
addi $out, $out, 16
b
g
e Lcbc_dec_loop
b
n
e Lcbc_dec_loop
Lcbc_done:
Lcbc_done:
addi $out, $out, -1
addi $out, $out, -1
...
@@ -750,25 +755,25 @@ Lcbc_done:
...
@@ -750,25 +755,25 @@ Lcbc_done:
li r10,`15+6*$SIZE_T`
li r10,`15+6*$SIZE_T`
li r11,`31+6*$SIZE_T`
li r11,`31+6*$SIZE_T`
lvx v20,r10,$sp
lvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v21,r11,$sp
lvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v22,r10,$sp
lvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v23,r11,$sp
lvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v24,r10,$sp
lvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v25,r11,$sp
lvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v26,r10,$sp
lvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v27,r11,$sp
lvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v28,r10,$sp
lvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v29,r11,$sp
lvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v30,r10,$sp
lvx v30,r10,$sp
lvx v31,r11,$sp
lvx v31,r11,$sp
Lcbc_abort:
Lcbc_abort:
...
@@ -1306,28 +1311,28 @@ Lschedule_mangle_dec:
...
@@ -1306,28 +1311,28 @@ Lschedule_mangle_dec:
mflr r0
mflr r0
mfspr r6, 256 # save vrsave
mfspr r6, 256 # save vrsave
stvx v20,r10,$sp
stvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v21,r11,$sp
stvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v22,r10,$sp
stvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v23,r11,$sp
stvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v24,r10,$sp
stvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v25,r11,$sp
stvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v26,r10,$sp
stvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v27,r11,$sp
stvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v28,r10,$sp
stvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v29,r11,$sp
stvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v30,r10,$sp
stvx v30,r10,$sp
stvx v31,r11,$sp
stvx v31,r11,$sp
lwz
r6,`$FRAME-4`($sp) # save vrsave
stw
r6,`$FRAME-4`($sp) # save vrsave
li r7, -1
li r7, -1
$PUSH r0, `$FRAME+$LRSAVE`($sp)
$PUSH r0, `$FRAME+$LRSAVE`($sp)
mtspr 256, r7 # preserve all AltiVec registers
mtspr 256, r7 # preserve all AltiVec registers
...
@@ -1347,25 +1352,25 @@ Lschedule_mangle_dec:
...
@@ -1347,25 +1352,25 @@ Lschedule_mangle_dec:
mtlr r0
mtlr r0
xor r3, r3, r3
xor r3, r3, r3
lvx v20,r10,$sp
lvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v21,r11,$sp
lvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v22,r10,$sp
lvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v23,r11,$sp
lvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v24,r10,$sp
lvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v25,r11,$sp
lvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v26,r10,$sp
lvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v27,r11,$sp
lvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v28,r10,$sp
lvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v29,r11,$sp
lvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v30,r10,$sp
lvx v30,r10,$sp
lvx v31,r11,$sp
lvx v31,r11,$sp
addi $sp,$sp,$FRAME
addi $sp,$sp,$FRAME
...
@@ -1384,28 +1389,28 @@ Lschedule_mangle_dec:
...
@@ -1384,28 +1389,28 @@ Lschedule_mangle_dec:
mflr r0
mflr r0
mfspr r6, 256 # save vrsave
mfspr r6, 256 # save vrsave
stvx v20,r10,$sp
stvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v21,r11,$sp
stvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v22,r10,$sp
stvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v23,r11,$sp
stvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v24,r10,$sp
stvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v25,r11,$sp
stvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v26,r10,$sp
stvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v27,r11,$sp
stvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v28,r10,$sp
stvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
stvx v29,r11,$sp
stvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
stvx v30,r10,$sp
stvx v30,r10,$sp
stvx v31,r11,$sp
stvx v31,r11,$sp
lwz
r6,`$FRAME-4`($sp) # save vrsave
stw
r6,`$FRAME-4`($sp) # save vrsave
li r7, -1
li r7, -1
$PUSH r0, `$FRAME+$LRSAVE`($sp)
$PUSH r0, `$FRAME+$LRSAVE`($sp)
mtspr 256, r7 # preserve all AltiVec registers
mtspr 256, r7 # preserve all AltiVec registers
...
@@ -1430,25 +1435,25 @@ Lschedule_mangle_dec:
...
@@ -1430,25 +1435,25 @@ Lschedule_mangle_dec:
mtlr r0
mtlr r0
xor r3, r3, r3
xor r3, r3, r3
lvx v20,r10,$sp
lvx v20,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v21,r11,$sp
lvx v21,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v22,r10,$sp
lvx v22,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v23,r11,$sp
lvx v23,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v24,r10,$sp
lvx v24,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v25,r11,$sp
lvx v25,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v26,r10,$sp
lvx v26,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v27,r11,$sp
lvx v27,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v28,r10,$sp
lvx v28,r10,$sp
addi r10,r10,
16
addi r10,r10,
32
lvx v29,r11,$sp
lvx v29,r11,$sp
addi r11,r11,
16
addi r11,r11,
32
lvx v30,r10,$sp
lvx v30,r10,$sp
lvx v31,r11,$sp
lvx v31,r11,$sp
addi $sp,$sp,$FRAME
addi $sp,$sp,$FRAME
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录