Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
a4b17225
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
a4b17225
编写于
10月 15, 2017
作者:
Q
Qiao Longfei
提交者:
GitHub
10月 15, 2017
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #4809 from jacquesqiao/backward-return-map
Backward return map
上级
17b4cea4
2befb9f9
变更
4
显示空白变更内容
内联
并排
Showing
4 changed file
with
38 addition
and
17 deletion
+38
-17
paddle/framework/backward.cc
paddle/framework/backward.cc
+10
-9
paddle/framework/backward.h
paddle/framework/backward.h
+6
-5
paddle/pybind/protobuf.cc
paddle/pybind/protobuf.cc
+13
-1
python/paddle/v2/framework/tests/test_program.py
python/paddle/v2/framework/tests/test_program.py
+9
-2
未找到文件。
paddle/framework/backward.cc
浏览文件 @
a4b17225
...
@@ -274,9 +274,10 @@ static bool AllGradInSet(const std::vector<std::string>& names,
...
@@ -274,9 +274,10 @@ static bool AllGradInSet(const std::vector<std::string>& names,
}
}
static
void
CreateGradVarInBlock
(
static
void
CreateGradVarInBlock
(
std
::
unordered_map
<
std
::
string
,
GradVarInfo
>*
grad_var_record
,
size_t
grad_op_start_index
,
BlockDescBind
*
block_desc
,
size_t
grad_op_start_index
,
const
std
::
unordered_map
<
std
::
string
,
std
::
string
>&
param_name_map
,
const
std
::
unordered_map
<
std
::
string
,
std
::
string
>&
param_name_map
)
{
BlockDescBind
*
block_desc
,
std
::
unordered_map
<
std
::
string
,
GradVarInfo
>*
grad_var_record
)
{
auto
ops
=
block_desc
->
AllOps
();
auto
ops
=
block_desc
->
AllOps
();
for
(
size_t
op_index
=
grad_op_start_index
;
op_index
<
ops
.
size
();
for
(
size_t
op_index
=
grad_op_start_index
;
op_index
<
ops
.
size
();
++
op_index
)
{
++
op_index
)
{
...
@@ -422,8 +423,8 @@ std::vector<std::unique_ptr<OpDescBind>> MakeBlockBackward(
...
@@ -422,8 +423,8 @@ std::vector<std::unique_ptr<OpDescBind>> MakeBlockBackward(
return
backward_descs
;
return
backward_descs
;
}
}
std
::
unordered_map
<
std
::
string
/*fwd_var_name*/
,
GradVarInfo
/*grad_var_info*/
>
ParamGradInfoMap
AppendBackward
(
AppendBackward
(
ProgramDescBind
&
program_desc
,
const
VarDescBind
&
target
,
ProgramDescBind
&
program_desc
,
const
VarDescBind
&
target
,
const
std
::
unordered_set
<
std
::
string
>&
no_grad_vars
)
{
const
std
::
unordered_set
<
std
::
string
>&
no_grad_vars
)
{
std
::
unordered_set
<
std
::
string
>
no_grad_var_names
;
std
::
unordered_set
<
std
::
string
>
no_grad_var_names
;
no_grad_var_names
.
reserve
(
no_grad_vars
.
size
()
+
1
);
no_grad_var_names
.
reserve
(
no_grad_vars
.
size
()
+
1
);
...
@@ -461,11 +462,11 @@ AppendBackward(ProgramDescBind& program_desc, const VarDescBind& target,
...
@@ -461,11 +462,11 @@ AppendBackward(ProgramDescBind& program_desc, const VarDescBind& target,
root_block
->
Var
(
fill_one_op_out
);
root_block
->
Var
(
fill_one_op_out
);
// create grad_var for all blocks in this program
// create grad_var for all blocks in this program
CreateGradVarInBlock
(
&
retv
,
root_block
,
forward_op_num
,
grad_to_var
);
CreateGradVarInBlock
(
forward_op_num
,
grad_to_var
,
root_block
,
&
retv
);
for
(
size_t
block_index
=
forward_block_num
;
for
(
size_t
block_index
=
forward_block_num
;
block_index
<
program_desc
.
Size
();
++
block_index
)
{
block_index
<
program_desc
.
Size
();
++
block_index
)
{
CreateGradVarInBlock
(
&
retv
,
program_desc
.
Block
(
block_index
),
0
,
CreateGradVarInBlock
(
0
,
grad_to_var
,
program_desc
.
Block
(
block_index
)
,
grad_to_var
);
&
retv
);
}
}
return
retv
;
return
retv
;
}
}
...
...
paddle/framework/backward.h
浏览文件 @
a4b17225
...
@@ -36,10 +36,11 @@ struct GradVarInfo {
...
@@ -36,10 +36,11 @@ struct GradVarInfo {
int
op_idx_
;
int
op_idx_
;
};
};
// TODO(jiayi): Add target as parameter and generate backward op
using
ParamGradInfoMap
=
std
::
unordered_map
<
std
::
string
/*fwd_var_name*/
,
// according to target.
GradVarInfo
/*grad_var_info*/
>
;
std
::
unordered_map
<
std
::
string
/*fwd_var_name*/
,
GradVarInfo
/*grad_var_info*/
>
AppendBackward
(
ProgramDescBind
&
program_desc
,
const
VarDescBind
&
target
,
ParamGradInfoMap
AppendBackward
(
ProgramDescBind
&
program_desc
,
const
VarDescBind
&
target
,
const
std
::
unordered_set
<
std
::
string
>&
no_grad_vars
);
const
std
::
unordered_set
<
std
::
string
>&
no_grad_vars
);
}
// namespace framework
}
// namespace framework
...
...
paddle/pybind/protobuf.cc
浏览文件 @
a4b17225
...
@@ -120,7 +120,19 @@ void BindProgramDesc(py::module &m) {
...
@@ -120,7 +120,19 @@ void BindProgramDesc(py::module &m) {
.
def
(
"append_backward"
,
.
def
(
"append_backward"
,
[](
ProgramDescBind
&
program_desc
,
const
VarDescBind
&
target
,
[](
ProgramDescBind
&
program_desc
,
const
VarDescBind
&
target
,
const
std
::
unordered_set
<
std
::
string
>
&
no_grad_vars
)
{
const
std
::
unordered_set
<
std
::
string
>
&
no_grad_vars
)
{
ParamGradInfoMap
param_grad_map
=
AppendBackward
(
program_desc
,
target
,
no_grad_vars
);
AppendBackward
(
program_desc
,
target
,
no_grad_vars
);
std
::
unordered_map
<
std
::
string
,
std
::
tuple
<
std
::
string
/* grad_var_name */
,
int
/* block_idx */
,
int
/* op_idx */
>>
retv
;
for
(
auto
it
=
param_grad_map
.
begin
();
it
!=
param_grad_map
.
end
();
++
it
)
{
const
auto
&
grad_info
=
it
->
second
;
retv
[
it
->
first
]
=
std
::
make_tuple
(
grad_info
.
name_
,
grad_info
.
block_idx_
,
grad_info
.
op_idx_
);
}
return
retv
;
})
})
.
def
(
"block"
,
&
ProgramDescBind
::
Block
,
py
::
return_value_policy
::
reference
)
.
def
(
"block"
,
&
ProgramDescBind
::
Block
,
py
::
return_value_policy
::
reference
)
.
def
(
"num_blocks"
,
&
ProgramDescBind
::
Size
)
.
def
(
"num_blocks"
,
&
ProgramDescBind
::
Size
)
...
...
python/paddle/v2/framework/tests/test_program.py
浏览文件 @
a4b17225
...
@@ -57,11 +57,18 @@ class TestProgram(unittest.TestCase):
...
@@ -57,11 +57,18 @@ class TestProgram(unittest.TestCase):
"mul"
,
"elementwise_add"
,
"fill_constant"
,
"elementwise_add_grad"
,
"mul"
,
"elementwise_add"
,
"fill_constant"
,
"elementwise_add_grad"
,
"mul_grad"
"mul_grad"
]
]
def
grad_name
(
name
):
return
name
+
"@GRAD"
actual_ops
=
[]
actual_ops
=
[]
prog
.
append_backward
(
target
,
set
())
param_to_grad
=
prog
.
append_backward
(
target
,
set
())
for
var_name
in
(
"x1"
,
"y1"
,
"out1"
,
"b1"
):
self
.
assertEqual
(
param_to_grad
[
var_name
][
0
],
grad_name
(
var_name
))
self
.
assertEqual
(
param_to_grad
[
var_name
][
1
],
0
)
for
op
in
block
.
all_ops
():
for
op
in
block
.
all_ops
():
actual_ops
.
append
(
op
.
type
())
actual_ops
.
append
(
op
.
type
())
print
(
actual_ops
)
self
.
assertEqual
(
actual_ops
,
expect_ops
)
self
.
assertEqual
(
actual_ops
,
expect_ops
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录