Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
s920243400
PaddleDetection
提交
74cd9a75
P
PaddleDetection
项目概览
s920243400
/
PaddleDetection
与 Fork 源项目一致
Fork自
PaddlePaddle / PaddleDetection
通知
2
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
74cd9a75
编写于
7月 28, 2017
作者:
D
dongzhihong
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
"fix unittest"
上级
72839a76
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
6 addition
and
6 deletion
+6
-6
paddle/framework/backward.cc
paddle/framework/backward.cc
+1
-1
paddle/framework/backward_test.cc
paddle/framework/backward_test.cc
+5
-5
未找到文件。
paddle/framework/backward.cc
浏览文件 @
74cd9a75
...
@@ -79,11 +79,11 @@ static std::shared_ptr<OperatorBase> BackwardImpl(
...
@@ -79,11 +79,11 @@ static std::shared_ptr<OperatorBase> BackwardImpl(
std
::
unordered_map
<
std
::
string
/*var name*/
,
std
::
unordered_map
<
std
::
string
/*var name*/
,
std
::
vector
<
size_t
>
/*op offset*/
>
std
::
vector
<
size_t
>
/*op offset*/
>
dup_output_ops
;
dup_output_ops
;
size_t
local_op_id
=
0
;
// Because it is a net op, it can static_cast.
// Because it is a net op, it can static_cast.
auto
&
forwardNet
=
static_cast
<
const
NetOp
&>
(
forwardOp
);
auto
&
forwardNet
=
static_cast
<
const
NetOp
&>
(
forwardOp
);
// travesal subnet/op
// travesal subnet/op
size_t
local_op_id
=
0
;
for
(
auto
it
=
forwardNet
.
ops_
.
rbegin
();
it
!=
forwardNet
.
ops_
.
rend
();
for
(
auto
it
=
forwardNet
.
ops_
.
rbegin
();
it
!=
forwardNet
.
ops_
.
rend
();
++
it
)
{
++
it
)
{
auto
fwd
=
*
it
;
auto
fwd
=
*
it
;
...
...
paddle/framework/backward_test.cc
浏览文件 @
74cd9a75
...
@@ -149,7 +149,6 @@ TEST(Backward, simple_op_grad) {
...
@@ -149,7 +149,6 @@ TEST(Backward, simple_op_grad) {
auto
fwd
=
f
::
OpRegistry
::
CreateOp
(
"rowwise_add"
,
{
"X"
,
"b"
},
{
"Out"
},
{});
auto
fwd
=
f
::
OpRegistry
::
CreateOp
(
"rowwise_add"
,
{
"X"
,
"b"
},
{
"Out"
},
{});
ASSERT_NE
(
fwd
,
nullptr
);
ASSERT_NE
(
fwd
,
nullptr
);
auto
gop
=
f
::
OpRegistry
::
CreateGradOp
(
*
fwd
);
auto
gop
=
f
::
OpRegistry
::
CreateGradOp
(
*
fwd
);
LOG
(
INFO
)
<<
gop
->
DebugString
();
ASSERT_EQ
(
1UL
,
gop
->
inputs_
.
size
());
ASSERT_EQ
(
1UL
,
gop
->
inputs_
.
size
());
ASSERT_EQ
(
"Out"
+
f
::
OperatorBase
::
GRAD_VAR_SUFFIX
(),
gop
->
inputs_
[
0
]);
ASSERT_EQ
(
"Out"
+
f
::
OperatorBase
::
GRAD_VAR_SUFFIX
(),
gop
->
inputs_
[
0
]);
ASSERT_EQ
(
"rowwise_add_grad"
,
gop
->
type_
);
ASSERT_EQ
(
"rowwise_add_grad"
,
gop
->
type_
);
...
@@ -161,18 +160,19 @@ TEST(Backward, simple_op_grad) {
...
@@ -161,18 +160,19 @@ TEST(Backward, simple_op_grad) {
// LOG(INFO) << gop->Output("X" + "@GRAD");
// LOG(INFO) << gop->Output("X" + "@GRAD");
}
}
TEST
(
Backward
,
simple_
net
_grad
)
{
TEST
(
Backward
,
simple_
op_not_need
_grad
)
{
auto
fwd
=
f
::
OpRegistry
::
CreateOp
(
"rowwise_add"
,
{
"
X"
,
"b"
},
{
"O
ut"
},
{});
auto
fwd
=
f
::
OpRegistry
::
CreateOp
(
"rowwise_add"
,
{
"
x"
,
"b"
},
{
"o
ut"
},
{});
ASSERT_NE
(
fwd
,
nullptr
);
ASSERT_NE
(
fwd
,
nullptr
);
auto
gop
=
f
::
Backward
(
*
fwd
,
{});
auto
gop
=
f
::
Backward
(
*
fwd
,
{
"x"
});
LOG
(
INFO
)
<<
gop
->
DebugString
();
LOG
(
INFO
)
<<
gop
->
DebugString
();
ASSERT_NE
(
gop
->
outputs_
.
find
(
"x"
+
f
::
OperatorBase
::
GRAD_VAR_SUFFIX
()),
gop
->
outputs_
.
end
());
}
}
TEST
(
Backward
,
net_fc_backward_normal
)
{
TEST
(
Backward
,
net_fc_backward_normal
)
{
std
::
shared_ptr
<
f
::
OperatorBase
>
fwd
=
f
::
OpRegistry
::
CreateOp
(
std
::
shared_ptr
<
f
::
OperatorBase
>
fwd
=
f
::
OpRegistry
::
CreateOp
(
"fc"
,
{
"X"
,
"w"
,
"b"
},
{
"mul_result"
,
"add_result"
,
"out"
},
{});
"fc"
,
{
"X"
,
"w"
,
"b"
},
{
"mul_result"
,
"add_result"
,
"out"
},
{});
ASSERT_NE
(
fwd
,
nullptr
);
ASSERT_NE
(
fwd
,
nullptr
);
LOG
(
INFO
)
<<
fwd
->
DebugString
();
std
::
shared_ptr
<
f
::
OperatorBase
>
gop
=
f
::
Backward
(
*
fwd
,
{});
std
::
shared_ptr
<
f
::
OperatorBase
>
gop
=
f
::
Backward
(
*
fwd
,
{});
ASSERT_TRUE
(
gop
->
IsNetOp
());
ASSERT_TRUE
(
gop
->
IsNetOp
());
auto
net
=
static_cast
<
f
::
NetOp
*>
(
gop
.
get
());
auto
net
=
static_cast
<
f
::
NetOp
*>
(
gop
.
get
());
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录