Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
0c951176
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 2 年 前同步成功
通知
2325
Star
20933
Fork
5424
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
0c951176
编写于
8月 07, 2017
作者:
T
tensor-tang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
pass mkldnn gtest
上级
ec9009f3
变更
6
隐藏空白更改
内联
并排
Showing
6 changed file
with
112 addition
and
54 deletion
+112
-54
paddle/gserver/layers/MkldnnFcLayer.cpp
paddle/gserver/layers/MkldnnFcLayer.cpp
+20
-4
paddle/gserver/layers/MkldnnFcLayer.h
paddle/gserver/layers/MkldnnFcLayer.h
+8
-3
paddle/gserver/layers/MkldnnLayer.cpp
paddle/gserver/layers/MkldnnLayer.cpp
+40
-22
paddle/gserver/layers/MkldnnLayer.h
paddle/gserver/layers/MkldnnLayer.h
+26
-1
paddle/gserver/tests/MkldnnTester.cpp
paddle/gserver/tests/MkldnnTester.cpp
+12
-18
paddle/gserver/tests/test_Mkldnn.cpp
paddle/gserver/tests/test_Mkldnn.cpp
+6
-6
未找到文件。
paddle/gserver/layers/MkldnnFcLayer.cpp
浏览文件 @
0c951176
...
@@ -42,7 +42,6 @@ bool MkldnnFcLayer::init(const LayerMap& layerMap,
...
@@ -42,7 +42,6 @@ bool MkldnnFcLayer::init(const LayerMap& layerMap,
// create weight
// create weight
weight_
=
weight_
=
std
::
unique_ptr
<
Weight
>
(
new
Weight
(
oc_
,
iLayerSize_
,
parameters_
[
0
],
0
));
std
::
unique_ptr
<
Weight
>
(
new
Weight
(
oc_
,
iLayerSize_
,
parameters_
[
0
],
0
));
initWgt
();
// create biases
// create biases
if
(
biasParameter_
.
get
()
!=
NULL
)
{
if
(
biasParameter_
.
get
()
!=
NULL
)
{
...
@@ -51,20 +50,36 @@ bool MkldnnFcLayer::init(const LayerMap& layerMap,
...
@@ -51,20 +50,36 @@ bool MkldnnFcLayer::init(const LayerMap& layerMap,
return
true
;
return
true
;
}
}
void
MkldnnFcLayer
::
initWgt
()
{
void
MkldnnFcLayer
::
cvtWgtFromPaddle
()
{
if
(
hasInitedWgt_
)
{
return
;
}
// The weight_ is transposed from initial paddle weight
// The weight_ is transposed from initial paddle weight
MatrixPtr
paddleWgt
=
Matrix
::
create
(
MatrixPtr
paddleWgt
=
Matrix
::
create
(
weight_
->
getW
()
->
getData
(),
iLayerSize_
,
oc_
,
false
,
false
);
weight_
->
getW
()
->
getData
(),
iLayerSize_
,
oc_
,
false
,
false
);
std
::
ostringstream
ostr
;
std
::
ostringstream
ostr
;
paddleWgt
->
print
(
ostr
);
paddleWgt
->
print
(
ostr
);
VLOG
(
DNN_
BASE
)
<<
ostr
.
str
();
VLOG
(
DNN_
ALL
)
<<
"Initial Weight from paddle: "
<<
std
::
endl
<<
ostr
.
str
();
//
Firstly in mkldnn, the matrix is transposed from initial paddle weight
//
The mkldnn weight is transposed from initial paddle matrix
MatrixPtr
paddleWgtT
;
MatrixPtr
paddleWgtT
;
paddleWgt
->
transpose
(
paddleWgtT
,
true
);
paddleWgt
->
transpose
(
paddleWgtT
,
true
);
weight_
->
getW
()
->
copyFrom
(
*
paddleWgtT
);
weight_
->
getW
()
->
copyFrom
(
*
paddleWgtT
);
hasInitedWgt_
=
true
;
}
void
MkldnnFcLayer
::
cvtWgtToPaddle
()
{
MatrixPtr
dnnWgt
=
weight_
->
getW
();
MatrixPtr
paddleWgt
;
dnnWgt
->
transpose
(
paddleWgt
,
true
);
// copy paddle weight and override on weight_
MatrixPtr
dnnWgtT
=
Matrix
::
create
(
dnnWgt
->
getData
(),
dnnWgt
->
getWidth
(),
dnnWgt
->
getHeight
(),
false
,
false
);
dnnWgtT
->
copyFrom
(
*
paddleWgt
);
}
}
void
MkldnnFcLayer
::
reshape
()
{
void
MkldnnFcLayer
::
reshape
()
{
...
@@ -86,6 +101,7 @@ void MkldnnFcLayer::reshape() {
...
@@ -86,6 +101,7 @@ void MkldnnFcLayer::reshape() {
ic_
=
iLayerSize_
/
(
ih_
*
iw_
);
ic_
=
iLayerSize_
/
(
ih_
*
iw_
);
CHECK_EQ
(
size_t
(
ic_
*
ih_
*
iw_
),
iLayerSize_
)
<<
"not divisible"
;
CHECK_EQ
(
size_t
(
ic_
*
ih_
*
iw_
),
iLayerSize_
)
<<
"not divisible"
;
CHECK_EQ
(
size_t
(
oc_
),
getSize
());
CHECK_EQ
(
size_t
(
oc_
),
getSize
());
printSizeInfo
();
// reset output
// reset output
output_
.
setFrameHeight
(
oh_
);
output_
.
setFrameHeight
(
oh_
);
...
...
paddle/gserver/layers/MkldnnFcLayer.h
浏览文件 @
0c951176
...
@@ -29,25 +29,30 @@ protected:
...
@@ -29,25 +29,30 @@ protected:
// input layer size, can not be change after init
// input layer size, can not be change after init
size_t
iLayerSize_
;
// == ic * ih * iw
size_t
iLayerSize_
;
// == ic * ih * iw
bool
hasInitedWgt_
;
// fc weight and bias
// fc weight and bias
std
::
unique_ptr
<
Weight
>
weight_
;
std
::
unique_ptr
<
Weight
>
weight_
;
std
::
unique_ptr
<
Weight
>
biases_
;
std
::
unique_ptr
<
Weight
>
biases_
;
public:
public:
explicit
MkldnnFcLayer
(
const
LayerConfig
&
config
)
:
MkldnnLayer
(
config
)
{}
explicit
MkldnnFcLayer
(
const
LayerConfig
&
config
)
:
MkldnnLayer
(
config
),
hasInitedWgt_
(
false
)
{}
~
MkldnnFcLayer
()
{}
~
MkldnnFcLayer
()
{}
bool
init
(
const
LayerMap
&
layerMap
,
bool
init
(
const
LayerMap
&
layerMap
,
const
ParameterMap
&
parameterMap
)
override
;
const
ParameterMap
&
parameterMap
)
override
;
void
initWgt
()
;
void
cvtWgtFromPaddle
()
override
;
void
reshape
()
;
void
cvtWgtToPaddle
()
override
;
void
forward
(
PassType
passType
)
override
;
void
forward
(
PassType
passType
)
override
;
void
backward
(
const
UpdateCallback
&
callback
)
override
;
void
backward
(
const
UpdateCallback
&
callback
)
override
;
void
reshape
();
};
};
}
// namespace paddle
}
// namespace paddle
paddle/gserver/layers/MkldnnLayer.cpp
浏览文件 @
0c951176
...
@@ -25,11 +25,18 @@ namespace paddle {
...
@@ -25,11 +25,18 @@ namespace paddle {
bool
MkldnnLayer
::
init
(
const
LayerMap
&
layerMap
,
bool
MkldnnLayer
::
init
(
const
LayerMap
&
layerMap
,
const
ParameterMap
&
parameterMap
)
{
const
ParameterMap
&
parameterMap
)
{
if
(
!
Layer
::
init
(
layerMap
,
parameterMap
))
{
return
false
;
}
CHECK
(
FLAGS_use_mkldnn
)
<<
"MkldnnLayers only support use_mkldnn."
CHECK
(
FLAGS_use_mkldnn
)
<<
"MkldnnLayers only support use_mkldnn."
<<
"Please set WITH_MKLDNN=ON "
<<
"Please set WITH_MKLDNN=ON "
<<
"and set use_mkldnn=True"
;
<<
"and set use_mkldnn=True"
;
stream_
.
reset
(
new
MkldnnStream
());
engine_
=
CpuEngine
::
Instance
().
getEngine
();
// TODO(TJ): deivecId
// TODO(TJ): deivecId
return
Layer
::
init
(
layerMap
,
parameterMap
)
;
return
true
;
}
}
void
MkldnnLayer
::
resetForwardFC
(
int
bs
,
void
MkldnnLayer
::
resetForwardFC
(
int
bs
,
...
@@ -42,7 +49,6 @@ void MkldnnLayer::resetForwardFC(int bs,
...
@@ -42,7 +49,6 @@ void MkldnnLayer::resetForwardFC(int bs,
real
*
wgtData
,
real
*
wgtData
,
real
*
biasData
)
{
real
*
biasData
)
{
bool
hasSpatial
=
ih
==
1
&&
iw
==
1
?
false
:
true
;
bool
hasSpatial
=
ih
==
1
&&
iw
==
1
?
false
:
true
;
engine_
=
CpuEngine
::
Instance
().
getEngine
();
mem
::
desc
botMD
=
hasSpatial
?
createMD
({
bs
,
ic
,
ih
,
iw
},
format
::
nchw
)
mem
::
desc
botMD
=
hasSpatial
?
createMD
({
bs
,
ic
,
ih
,
iw
},
format
::
nchw
)
:
createMD
({
bs
,
ic
},
format
::
nc
);
:
createMD
({
bs
,
ic
},
format
::
nc
);
...
@@ -52,21 +58,21 @@ void MkldnnLayer::resetForwardFC(int bs,
...
@@ -52,21 +58,21 @@ void MkldnnLayer::resetForwardFC(int bs,
:
createMD
({},
format
::
format_undef
);
:
createMD
({},
format
::
format_undef
);
mem
::
desc
topMD
=
createMD
({
bs
,
oc
},
format
::
nc
);
mem
::
desc
topMD
=
createMD
({
bs
,
oc
},
format
::
nc
);
inVal_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
botMD
,
engine_
),
botData
));
wgtVal_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
wgtMD
,
engine_
),
wgtData
));
outVal_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
topMD
,
engine_
),
topData
));
mkldnn
::
prop_kind
pk
=
mkldnn
::
prop_kind
::
forward
;
mkldnn
::
prop_kind
pk
=
mkldnn
::
prop_kind
::
forward
;
fc_fwd
::
desc
fwdDesc
=
biasData
!=
NULL
fc_fwd
::
desc
fwdDesc
=
biasData
!=
NULL
?
fc_fwd
::
desc
(
pk
,
botMD
,
wgtMD
,
biasMD
,
topMD
)
?
fc_fwd
::
desc
(
pk
,
botMD
,
wgtMD
,
biasMD
,
topMD
)
:
fc_fwd
::
desc
(
pk
,
botMD
,
wgtMD
,
topMD
);
:
fc_fwd
::
desc
(
pk
,
botMD
,
wgtMD
,
topMD
);
fc_fwd
::
primitive_desc
fwdPD
=
fc_fwd
::
primitive_desc
(
fwdDesc
,
engine_
);
fc_fwd
::
primitive_desc
fwdPD
=
fc_fwd
::
primitive_desc
(
fwdDesc
,
engine_
);
mem
bot
=
mem
(
mem
::
primitive_desc
(
botMD
,
engine_
),
botData
);
mem
wgt
=
mem
(
mem
::
primitive_desc
(
wgtMD
,
engine_
),
wgtData
);
mem
top
=
mem
(
mem
::
primitive_desc
(
topMD
,
engine_
),
topData
);
if
(
biasData
!=
NULL
)
{
if
(
biasData
!=
NULL
)
{
mem
bias
=
mem
(
mem
::
primitive_desc
(
biasMD
,
engine_
),
biasData
);
biasVal_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
biasMD
,
engine_
),
biasData
)
);
fwd_
.
reset
(
new
fc_fwd
(
fwdPD
,
bot
,
wgt
,
bias
,
top
));
fwd_
.
reset
(
new
fc_fwd
(
fwdPD
,
*
inVal_
,
*
wgtVal_
,
*
biasVal_
,
*
outVal_
));
}
else
{
}
else
{
fwd_
.
reset
(
new
fc_fwd
(
fwdPD
,
bot
,
wgt
,
top
));
fwd_
.
reset
(
new
fc_fwd
(
fwdPD
,
*
inVal_
,
*
wgtVal_
,
*
outVal_
));
}
}
pipelineFwd_
.
clear
();
pipelineFwd_
.
clear
();
pipelineFwd_
.
push_back
(
*
fwd_
);
pipelineFwd_
.
push_back
(
*
fwd_
);
...
@@ -84,8 +90,12 @@ void MkldnnLayer::mkldnnForwardFC(int bs,
...
@@ -84,8 +90,12 @@ void MkldnnLayer::mkldnnForwardFC(int bs,
// if input size changed, reset it
// if input size changed, reset it
resetForwardFC
(
bs
,
ic
,
ih
,
iw
,
botData
,
oc
,
topData
,
wgtData
,
biasData
);
resetForwardFC
(
bs
,
ic
,
ih
,
iw
,
botData
,
oc
,
topData
,
wgtData
,
biasData
);
this
->
cvtWgtFromPaddle
();
// update input, since the data might be changed if this is after data layer
inVal_
->
set_data_handle
(
botData
);
// just forward
// just forward
// update botdata
stream_
->
submit
(
pipelineFwd_
);
stream_
->
submit
(
pipelineFwd_
);
}
}
...
@@ -112,6 +122,10 @@ void MkldnnLayer::resetBackwardFC(int bs,
...
@@ -112,6 +122,10 @@ void MkldnnLayer::resetBackwardFC(int bs,
mem
::
desc
biasMD
=
biasDiff
!=
NULL
?
createMD
({
oc
},
format
::
x
)
mem
::
desc
biasMD
=
biasDiff
!=
NULL
?
createMD
({
oc
},
format
::
x
)
:
createMD
({},
format
::
format_undef
);
:
createMD
({},
format
::
format_undef
);
inVal_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
botMD
,
engine_
),
botData
));
wgtGrad_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
wgtMD
,
engine_
),
wgtDiff
));
outGrad_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
topMD
,
engine_
),
topDiff
));
fc_fwd
::
desc
fwdDesc
=
fc_fwd
::
desc
fwdDesc
=
fc_fwd
::
desc
(
mkldnn
::
prop_kind
::
forward
,
botMD
,
wgtMD
,
topMD
);
fc_fwd
::
desc
(
mkldnn
::
prop_kind
::
forward
,
botMD
,
wgtMD
,
topMD
);
fc_fwd
::
primitive_desc
fwdPD
=
fc_fwd
::
primitive_desc
(
fwdDesc
,
engine_
);
fc_fwd
::
primitive_desc
fwdPD
=
fc_fwd
::
primitive_desc
(
fwdDesc
,
engine_
);
...
@@ -121,15 +135,12 @@ void MkldnnLayer::resetBackwardFC(int bs,
...
@@ -121,15 +135,12 @@ void MkldnnLayer::resetBackwardFC(int bs,
fc_bwdWgt
::
primitive_desc
bwdWgtPD
=
fc_bwdWgt
::
primitive_desc
bwdWgtPD
=
fc_bwdWgt
::
primitive_desc
(
bwdWgtDesc
,
engine_
,
fwdPD
);
fc_bwdWgt
::
primitive_desc
(
bwdWgtDesc
,
engine_
,
fwdPD
);
mem
botVal
=
mem
(
mem
::
primitive_desc
(
botMD
,
engine_
),
botData
);
mem
wgtGrad
=
mem
(
mem
::
primitive_desc
(
wgtMD
,
engine_
),
wgtDiff
);
mem
topGrad
=
mem
(
mem
::
primitive_desc
(
topMD
,
engine_
),
topDiff
);
if
(
biasDiff
!=
NULL
)
{
if
(
biasDiff
!=
NULL
)
{
mem
biasGrad
=
mem
(
mem
::
primitive_desc
(
biasMD
,
engine_
),
biasDiff
);
biasGrad_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
biasMD
,
engine_
),
biasDiff
));
bwdWgt_
.
reset
(
new
fc_bwdWgt
(
bwdWgtPD
,
botVal
,
topGrad
,
wgtGrad
,
biasGrad
));
bwdWgt_
.
reset
(
new
fc_bwdWgt
(
bwdWgtPD
,
*
inVal_
,
*
outGrad_
,
*
wgtGrad_
,
*
biasGrad_
));
}
else
{
}
else
{
bwdWgt_
.
reset
(
new
fc_bwdWgt
(
bwdWgtPD
,
botVal
,
topGrad
,
wgtGrad
));
bwdWgt_
.
reset
(
new
fc_bwdWgt
(
bwdWgtPD
,
*
inVal_
,
*
outGrad_
,
*
wgtGrad_
));
}
}
pipelineBwd_
.
clear
();
pipelineBwd_
.
clear
();
pipelineBwd_
.
push_back
(
*
bwdWgt_
);
pipelineBwd_
.
push_back
(
*
bwdWgt_
);
...
@@ -142,9 +153,9 @@ void MkldnnLayer::resetBackwardFC(int bs,
...
@@ -142,9 +153,9 @@ void MkldnnLayer::resetBackwardFC(int bs,
fc_bwdData
::
desc
bwdDataDesc
=
fc_bwdData
::
desc
(
botMD
,
wgtMD
,
topMD
);
fc_bwdData
::
desc
bwdDataDesc
=
fc_bwdData
::
desc
(
botMD
,
wgtMD
,
topMD
);
fc_bwdData
::
primitive_desc
bwdDataPD
=
fc_bwdData
::
primitive_desc
bwdDataPD
=
fc_bwdData
::
primitive_desc
(
bwdDataDesc
,
engine_
,
fwdPD
);
fc_bwdData
::
primitive_desc
(
bwdDataDesc
,
engine_
,
fwdPD
);
mem
botGrad
=
mem
(
mem
::
primitive_desc
(
botMD
,
engine_
),
botDiff
);
inGrad_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
botMD
,
engine_
),
botDiff
)
);
mem
wgtVal
=
mem
(
mem
::
primitive_desc
(
wgtMD
,
engine_
),
wgtData
);
wgtVal_
.
reset
(
new
mem
(
mem
::
primitive_desc
(
wgtMD
,
engine_
),
wgtData
)
);
bwdData_
.
reset
(
new
fc_bwdData
(
bwdDataPD
,
topGrad
,
wgtVal
,
botGrad
));
bwdData_
.
reset
(
new
fc_bwdData
(
bwdDataPD
,
*
outGrad_
,
*
wgtVal_
,
*
inGrad_
));
pipelineBwd_
.
push_back
(
*
bwdData_
);
pipelineBwd_
.
push_back
(
*
bwdData_
);
}
}
...
@@ -172,11 +183,18 @@ void MkldnnLayer::mkldnnBackwardFC(int bs,
...
@@ -172,11 +183,18 @@ void MkldnnLayer::mkldnnBackwardFC(int bs,
wgtData
,
wgtData
,
biasDiff
);
biasDiff
);
// just forward
// update data
// update botdata
outGrad_
->
set_data_handle
(
topDiff
);
stream_
->
submit
(
pipelineBwd_
);
stream_
->
submit
(
pipelineBwd_
);
}
}
void
MkldnnLayer
::
printSizeInfo
()
{
VLOG
(
DNN_SIZES
)
<<
"bs: "
<<
bs_
<<
", ic: "
<<
ic_
<<
", ih: "
<<
ih_
<<
", iw: "
<<
iw_
<<
", oc: "
<<
oc_
<<
", oh: "
<<
oh_
<<
", ow: "
<<
ow_
;
}
mem
::
desc
MkldnnLayer
::
createMD
(
mem
::
dims
dims
,
mem
::
desc
MkldnnLayer
::
createMD
(
mem
::
dims
dims
,
mem
::
format
fmt
,
mem
::
format
fmt
,
mem
::
data_type
type
)
{
mem
::
data_type
type
)
{
...
...
paddle/gserver/layers/MkldnnLayer.h
浏览文件 @
0c951176
...
@@ -40,13 +40,24 @@ protected:
...
@@ -40,13 +40,24 @@ protected:
// mkldnn engine, stream and primivtives
// mkldnn engine, stream and primivtives
mkldnn
::
engine
engine_
;
mkldnn
::
engine
engine_
;
std
::
shared_ptr
<
MkldnnStream
>
stream_
;
std
::
shared_ptr
<
MkldnnStream
>
stream_
;
std
::
shared_ptr
<
mkldnn
::
primitive
>
fwd_
;
std
::
shared_ptr
<
mkldnn
::
primitive
>
fwd_
;
std
::
shared_ptr
<
mkldnn
::
primitive
>
bwdWgt_
;
std
::
shared_ptr
<
mkldnn
::
primitive
>
bwdWgt_
;
std
::
shared_ptr
<
mkldnn
::
primitive
>
bwdData_
;
std
::
shared_ptr
<
mkldnn
::
primitive
>
bwdData_
;
std
::
vector
<
mkldnn
::
primitive
>
pipelineFwd_
;
std
::
vector
<
mkldnn
::
primitive
>
pipelineFwd_
;
std
::
vector
<
mkldnn
::
primitive
>
pipelineBwd_
;
std
::
vector
<
mkldnn
::
primitive
>
pipelineBwd_
;
// TODO(TJ): change below memory as MkldnnMatrixPtr type
// input == bottom, output == top
// value == data, grad == diff
std
::
shared_ptr
<
mkldnn
::
memory
>
inVal_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
inGrad_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
outVal_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
outGrad_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
wgtVal_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
wgtGrad_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
biasVal_
;
std
::
shared_ptr
<
mkldnn
::
memory
>
biasGrad_
;
public:
public:
explicit
MkldnnLayer
(
const
LayerConfig
&
config
)
explicit
MkldnnLayer
(
const
LayerConfig
&
config
)
:
Layer
(
config
),
:
Layer
(
config
),
...
@@ -67,6 +78,20 @@ public:
...
@@ -67,6 +78,20 @@ public:
virtual
bool
init
(
const
LayerMap
&
layerMap
,
const
ParameterMap
&
parameterMap
);
virtual
bool
init
(
const
LayerMap
&
layerMap
,
const
ParameterMap
&
parameterMap
);
virtual
void
printSizeInfo
();
/**
* convert weight from paddle format to mkldnn format
* weight_ will be override
*/
virtual
void
cvtWgtFromPaddle
()
{
;
}
/**
* convert mkldnn weight to paddle format
* weight_ will be override
*/
virtual
void
cvtWgtToPaddle
()
{
;
}
void
resetForwardFC
(
int
bs
,
void
resetForwardFC
(
int
bs
,
int
ic
,
int
ic
,
int
ih
,
int
ih
,
...
...
paddle/gserver/tests/MkldnnTester.cpp
浏览文件 @
0c951176
...
@@ -14,6 +14,7 @@ limitations under the License. */
...
@@ -14,6 +14,7 @@ limitations under the License. */
#include "MkldnnTester.h"
#include "MkldnnTester.h"
#include "paddle/gserver/layers/MkldnnBase.h"
#include "paddle/gserver/layers/MkldnnBase.h"
#include "paddle/gserver/layers/MkldnnLayer.h"
namespace
paddle
{
namespace
paddle
{
...
@@ -145,7 +146,10 @@ void MkldnnTester::checkBackwardWgts() {
...
@@ -145,7 +146,10 @@ void MkldnnTester::checkBackwardWgts() {
vector
<
VectorPtr
>
dnnWgts
;
// used to temply save mkldnn weights
vector
<
VectorPtr
>
dnnWgts
;
// used to temply save mkldnn weights
saveWgt
(
parameters_
[
DNN
],
dnnWgts
);
saveWgt
(
parameters_
[
DNN
],
dnnWgts
);
// TODO(TJ): cvtWgtToPaddle
const
MkldnnLayerPtr
dnnlayer
=
std
::
dynamic_pointer_cast
<
MkldnnLayer
>
(
dnnLayer_
);
CHECK
(
dnnlayer
);
dnnlayer
->
cvtWgtToPaddle
();
for
(
size_t
i
=
0
;
i
<
parameters_
[
DNN
].
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
parameters_
[
DNN
].
size
();
++
i
)
{
const
VectorPtr
&
dnn
=
parameters_
[
DNN
][
i
]
->
getBuf
(
PARAMETER_VALUE
);
const
VectorPtr
&
dnn
=
parameters_
[
DNN
][
i
]
->
getBuf
(
PARAMETER_VALUE
);
const
VectorPtr
&
ref
=
parameters_
[
REF
][
i
]
->
getBuf
(
PARAMETER_VALUE
);
const
VectorPtr
&
ref
=
parameters_
[
REF
][
i
]
->
getBuf
(
PARAMETER_VALUE
);
...
@@ -233,11 +237,10 @@ void MkldnnTester::printMatrix(const MatrixPtr& m) {
...
@@ -233,11 +237,10 @@ void MkldnnTester::printMatrix(const MatrixPtr& m) {
if
(
!
log_
)
{
if
(
!
log_
)
{
return
;
return
;
}
}
#ifdef _DEBUG
std
::
ostream
str
;
std
::
ostringstream
ostr
;
m
->
print
(
str
);
m
->
print
(
ostr
);
VLOG
(
lvl_
)
<<
str
;
VLOG
(
lvl_
)
<<
std
::
endl
<<
ostr
.
str
();
#endif
}
}
void
MkldnnTester
::
printVector
(
const
VectorPtr
&
v
)
{
void
MkldnnTester
::
printVector
(
const
VectorPtr
&
v
)
{
...
@@ -245,15 +248,9 @@ void MkldnnTester::printVector(const VectorPtr& v) {
...
@@ -245,15 +248,9 @@ void MkldnnTester::printVector(const VectorPtr& v) {
return
;
return
;
}
}
CHECK
(
v
);
std
::
ostringstream
ostr
;
CHECK
(
v
->
getData
());
v
->
print
(
ostr
,
v
->
getSize
());
const
real
*
pd
=
v
->
getData
();
VLOG
(
lvl_
)
<<
std
::
endl
<<
ostr
.
str
();
const
size_t
sz
=
v
->
getSize
();
std
::
stringstream
row
;
for
(
size_t
i
=
0
;
i
<
sz
;
++
i
)
{
row
<<
pd
[
i
]
<<
", "
;
}
VLOG
(
lvl_
)
<<
row
.
str
();
}
}
double
MkldnnTester
::
getDelta
(
const
real
*
d1
,
double
MkldnnTester
::
getDelta
(
const
real
*
d1
,
...
@@ -335,7 +332,6 @@ void MkldnnTester::run(const TestConfig& dnn,
...
@@ -335,7 +332,6 @@ void MkldnnTester::run(const TestConfig& dnn,
// Firstly always set flag false to initial from paddle weight
// Firstly always set flag false to initial from paddle weight
TestConfig
first
=
dnn
;
TestConfig
first
=
dnn
;
// first.layerConfig.set_init_wgt_from_mkldnn(false);
// reset and run once
// reset and run once
reset
(
first
,
ref
,
batchSize
);
reset
(
first
,
ref
,
batchSize
);
...
@@ -348,8 +344,6 @@ void MkldnnTester::run(const TestConfig& dnn,
...
@@ -348,8 +344,6 @@ void MkldnnTester::run(const TestConfig& dnn,
// firstly get the flag
// firstly get the flag
bool
initWgtFromMkldnn
=
false
;
bool
initWgtFromMkldnn
=
false
;
// dnn.layerConfig.has_init_wgt_from_mkldnn() &&
// dnn.layerConfig.init_wgt_from_mkldnn();
if
(
initWgtFromMkldnn
)
{
if
(
initWgtFromMkldnn
)
{
// after run once the mkldnn weight has been stored in dnnlayer
// after run once the mkldnn weight has been stored in dnnlayer
...
...
paddle/gserver/tests/test_Mkldnn.cpp
浏览文件 @
0c951176
...
@@ -55,12 +55,12 @@ void testFcLayer(const testFCDesc& pm) {
...
@@ -55,12 +55,12 @@ void testFcLayer(const testFCDesc& pm) {
}
}
TEST
(
MkldnnLayer
,
fcLayer
)
{
TEST
(
MkldnnLayer
,
fcLayer
)
{
testFcLayer
({
2
,
2
,
3
,
1
,
1
});
/*
testFcLayer
({
2
,
2
,
3
,
1
,
1
});
testFcLayer({16, 32, 64
, 1, 1});
testFcLayer
({
3
,
7
,
19
,
1
,
1
});
testFcLayer({8, 16, 32, 13, 13});
testFcLayer
({
8
,
16
,
32
,
13
,
13
});
testFcLayer({4, 12, 18, 13, 11});
testFcLayer
({
4
,
12
,
18
,
13
,
11
});
testFcLayer({2, 64, 32, 16, 16});
testFcLayer
({
2
,
64
,
32
,
16
,
16
});
testFcLayer({15, 3, 6, 16, 16});*/
testFcLayer
({
15
,
3
,
6
,
16
,
16
});
}
}
// TODO(TJ): add branch test
// TODO(TJ): add branch test
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录