Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
2e101df7
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
2e101df7
编写于
12月 19, 2017
作者:
T
tensor-tang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
enable gtest for MKLPackedRecurrentLayer
上级
624e3e52
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
159 addition
and
6 deletion
+159
-6
paddle/gserver/tests/test_RecurrentLayer.cpp
paddle/gserver/tests/test_RecurrentLayer.cpp
+159
-6
未找到文件。
paddle/gserver/tests/test_RecurrentLayer.cpp
浏览文件 @
2e101df7
...
@@ -420,12 +420,165 @@ TEST(Layer, LstmLayer) {
...
@@ -420,12 +420,165 @@ TEST(Layer, LstmLayer) {
}
}
}
}
#ifdef PADDLE_WITH_MKLML
LayerPtr
initMKLPackedLayer
(
LayerConfig
layerConfig
,
bool
reversed
,
int
layerSize
,
LayerPtr
dataLayer
,
ParameterPtr
para
,
ParameterPtr
bias
=
nullptr
)
{
LayerMap
layerMap
;
ParameterMap
parameterMap
;
layerMap
[
dataLayer
->
getName
()]
=
dataLayer
;
parameterMap
[
para
->
getName
()]
=
para
;
if
(
bias
)
{
parameterMap
[
bias
->
getName
()]
=
bias
;
layerConfig
.
set_bias_parameter_name
(
"bias_0"
);
}
layerConfig
.
set_size
(
layerSize
);
layerConfig
.
set_reversed
(
reversed
);
layerConfig
.
add_inputs
();
LayerInputConfig
&
input
=
*
(
layerConfig
.
mutable_inputs
(
0
));
input
.
set_input_layer_name
(
"layer_0"
);
input
.
set_input_parameter_name
(
"para_0"
);
LayerPtr
testLayer
=
Layer
::
create
(
layerConfig
);
layerMap
[
testLayer
->
getName
()]
=
testLayer
;
testLayer
->
init
(
layerMap
,
parameterMap
);
testLayer
->
setNeedGradient
(
true
);
return
testLayer
;
}
void
checkMKLPackedLayer
(
LayerPtr
testLayer1
,
LayerPtr
testLayer2
)
{
const
VectorPtr
&
weightGrad
=
(
testLayer1
->
getParameters
()[
0
])
->
getBuf
(
PARAMETER_GRADIENT
);
const
MatrixPtr
&
inputGrad
=
testLayer1
->
getPrev
(
0
)
->
getOutputGrad
();
CpuVector
wgt_grad1
(
weightGrad
->
getSize
());
CpuVector
wgt_grad2
(
weightGrad
->
getSize
());
CpuMatrix
input_grad1
(
inputGrad
->
getHeight
(),
inputGrad
->
getWidth
());
CpuMatrix
input_grad2
(
inputGrad
->
getHeight
(),
inputGrad
->
getWidth
());
CpuMatrix
outputGrad
(
inputGrad
->
getHeight
(),
inputGrad
->
getWidth
());
outputGrad
.
randomizeUniform
();
for
(
int
i
=
0
;
i
<
2
;
i
++
)
{
FLAGS_rnn_use_batch
=
true
;
testLayer1
->
forward
(
PASS_GC
);
testLayer1
->
getOutputGrad
()
->
copyFrom
(
outputGrad
);
weightGrad
->
zero
();
inputGrad
->
zero
();
testLayer1
->
backward
(
nullptr
);
wgt_grad1
.
copyFrom
(
*
weightGrad
);
input_grad1
.
copyFrom
(
*
inputGrad
);
FLAGS_rnn_use_batch
=
true
;
testLayer2
->
forward
(
PASS_GC
);
testLayer2
->
getOutputGrad
()
->
copyFrom
(
outputGrad
);
weightGrad
->
zero
();
inputGrad
->
zero
();
testLayer2
->
backward
(
nullptr
);
wgt_grad2
.
copyFrom
(
*
weightGrad
);
input_grad2
.
copyFrom
(
*
inputGrad
);
checkError
(
*
testLayer1
->
getOutputValue
(),
*
testLayer2
->
getOutputValue
());
checkError
(
wgt_grad1
,
wgt_grad2
);
checkError
(
input_grad1
,
input_grad2
);
}
for
(
int
i
=
0
;
i
<
2
;
i
++
)
{
CpuMatrix
outputValue
(
testLayer2
->
getOutputValue
()
->
getHeight
(),
testLayer2
->
getOutputValue
()
->
getWidth
());
FLAGS_rnn_use_batch
=
true
;
testLayer2
->
forward
(
PASS_GC
);
outputValue
.
copyFrom
(
*
testLayer2
->
getOutputValue
());
testLayer2
->
getOutputGrad
()
->
copyFrom
(
outputGrad
);
weightGrad
->
zero
();
inputGrad
->
zero
();
testLayer2
->
backward
(
nullptr
);
wgt_grad1
.
copyFrom
(
*
weightGrad
);
input_grad1
.
copyFrom
(
*
inputGrad
);
FLAGS_rnn_use_batch
=
false
;
testLayer2
->
getOutputValue
()
->
zero
();
testLayer2
->
forward
(
PASS_GC
);
testLayer2
->
getOutputGrad
()
->
copyFrom
(
outputGrad
);
weightGrad
->
zero
();
inputGrad
->
zero
();
testLayer2
->
backward
(
nullptr
);
wgt_grad2
.
copyFrom
(
*
weightGrad
);
input_grad2
.
copyFrom
(
*
inputGrad
);
checkError
(
outputValue
,
*
testLayer2
->
getOutputValue
());
checkError
(
wgt_grad1
,
wgt_grad2
);
checkError
(
input_grad1
,
input_grad2
);
}
}
TEST
(
MKLPackedLayer
,
RecurrentLayer
)
{
LayerConfig
layerConfig1
;
LayerConfig
layerConfig2
;
layerConfig1
.
set_name
(
"paddle-rnn"
);
layerConfig1
.
set_type
(
"recurrent"
);
layerConfig1
.
set_active_type
(
"relu"
);
layerConfig2
.
set_name
(
"mkl-packed-rnn"
);
layerConfig2
.
set_type
(
"mkl_packed_recurrent"
);
layerConfig2
.
set_active_type
(
"relu"
);
for
(
auto
layerSize
:
{
32
,
64
,
128
,
256
,
512
})
{
for
(
auto
batchSize
:
{
1
,
5
,
100
,
500
})
{
for
(
auto
reversed
:
{
true
,
false
})
{
LOG
(
INFO
)
<<
" layerSize="
<<
layerSize
<<
" batchSize="
<<
batchSize
<<
" reversed="
<<
reversed
;
LayerPtr
dataLayer
=
creatDataLayer
(
"layer_0"
,
batchSize
,
layerSize
,
false
);
ParameterPtr
para
=
creatParameter
(
"para_0"
,
0
,
layerSize
*
layerSize
,
false
);
LayerPtr
testLayer1
=
initMKLPackedLayer
(
layerConfig1
,
reversed
,
layerSize
,
dataLayer
,
para
);
LayerPtr
testLayer2
=
initMKLPackedLayer
(
layerConfig2
,
reversed
,
layerSize
,
dataLayer
,
para
);
checkMKLPackedLayer
(
testLayer1
,
testLayer2
);
}
}
}
}
#endif
int
main
(
int
argc
,
char
**
argv
)
{
int
main
(
int
argc
,
char
**
argv
)
{
if
(
version
::
isWithGpu
())
{
testing
::
InitGoogleTest
(
&
argc
,
argv
);
testing
::
InitGoogleTest
(
&
argc
,
argv
);
initMain
(
argc
,
argv
);
initMain
(
argc
,
argv
);
if
(
!
version
::
isWithGpu
())
{
return
RUN_ALL_TESTS
();
testing
::
GTEST_FLAG
(
filter
)
=
"-Layer.*"
;
}
else
{
return
0
;
}
}
return
RUN_ALL_TESTS
();
}
}
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录