Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
2e1bca99
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
2e1bca99
编写于
10月 12, 2020
作者:
G
guofei
提交者:
GitHub
10月 12, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Refine the gradient calculation errors caused by renaming in while_grad (#27814)
test=develop
上级
8fa4c098
变更
5
隐藏空白更改
内联
并排
Showing
5 changed file
with
93 addition
and
23 deletion
+93
-23
paddle/fluid/operators/controlflow/while_op.cc
paddle/fluid/operators/controlflow/while_op.cc
+61
-9
paddle/fluid/operators/controlflow/while_op_helper.cc
paddle/fluid/operators/controlflow/while_op_helper.cc
+11
-0
paddle/fluid/operators/controlflow/while_op_helper.h
paddle/fluid/operators/controlflow/while_op_helper.h
+4
-0
python/paddle/fluid/tests/unittests/test_while_loop_op.py
python/paddle/fluid/tests/unittests/test_while_loop_op.py
+15
-14
python/paddle/fluid/tests/unittests/test_while_op.py
python/paddle/fluid/tests/unittests/test_while_op.py
+2
-0
未找到文件。
paddle/fluid/operators/controlflow/while_op.cc
浏览文件 @
2e1bca99
...
@@ -12,6 +12,8 @@
...
@@ -12,6 +12,8 @@
// See the License for the specific language governing permissions and
// See the License for the specific language governing permissions and
// limitations under the License.
// limitations under the License.
#include <set>
#include "paddle/fluid/framework/executor.h"
#include "paddle/fluid/framework/executor.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/operator.h"
#include "paddle/fluid/framework/operator.h"
...
@@ -70,6 +72,23 @@ class WhileOp : public framework::OperatorBase {
...
@@ -70,6 +72,23 @@ class WhileOp : public framework::OperatorBase {
auto
*
block
=
Attr
<
framework
::
BlockDesc
*>
(
kStepBlock
);
auto
*
block
=
Attr
<
framework
::
BlockDesc
*>
(
kStepBlock
);
auto
*
program
=
block
->
Program
();
auto
*
program
=
block
->
Program
();
bool
is_test
=
Attr
<
bool
>
(
"is_test"
);
std
::
set
<
std
::
string
>
no_copy_var_names
;
if
(
!
is_test
)
{
const
std
::
vector
<
framework
::
OpDesc
*>
&
all_ops
=
block
->
AllOps
();
for
(
const
framework
::
OpDesc
*
op
:
all_ops
)
{
const
framework
::
VariableNameMap
&
input_var_names
=
op
->
Inputs
();
const
framework
::
VariableNameMap
&
output_var_names
=
op
->
Outputs
();
for
(
auto
&
ipt
:
input_var_names
)
{
for
(
const
std
::
string
&
var_name
:
ipt
.
second
)
{
if
(
StrInVaraiableNameMap
(
var_name
,
output_var_names
))
{
no_copy_var_names
.
insert
(
var_name
);
}
}
}
}
}
auto
step_scopes
=
auto
step_scopes
=
scope
.
FindVar
(
Output
(
kStepScopes
))
->
GetMutable
<
StepScopeVar
>
();
scope
.
FindVar
(
Output
(
kStepScopes
))
->
GetMutable
<
StepScopeVar
>
();
...
@@ -89,7 +108,6 @@ class WhileOp : public framework::OperatorBase {
...
@@ -89,7 +108,6 @@ class WhileOp : public framework::OperatorBase {
"The Output(StepScope) of WhileOp should be empty."
));
"The Output(StepScope) of WhileOp should be empty."
));
bool
cond_data
=
GetCondData
(
cond
);
bool
cond_data
=
GetCondData
(
cond
);
bool
is_test
=
Attr
<
bool
>
(
"is_test"
);
auto
&
skip_vars
=
Attr
<
std
::
vector
<
std
::
string
>>
(
kSkipEagerDeletionVars
);
auto
&
skip_vars
=
Attr
<
std
::
vector
<
std
::
string
>>
(
kSkipEagerDeletionVars
);
VLOG
(
2
)
<<
GetSkipEagerDeletionVarsDebugString
(
skip_vars
);
VLOG
(
2
)
<<
GetSkipEagerDeletionVarsDebugString
(
skip_vars
);
...
@@ -98,8 +116,32 @@ class WhileOp : public framework::OperatorBase {
...
@@ -98,8 +116,32 @@ class WhileOp : public framework::OperatorBase {
while
(
cond_data
)
{
while
(
cond_data
)
{
auto
&
current_scope
=
scope
.
NewScope
();
auto
&
current_scope
=
scope
.
NewScope
();
step_scopes
->
push_back
(
&
current_scope
);
step_scopes
->
push_back
(
&
current_scope
);
std
::
vector
<
std
::
string
>
rename_vars
;
for
(
const
std
::
string
&
input_var_name
:
Inputs
(
kX
))
{
if
(
no_copy_var_names
.
find
(
input_var_name
)
==
no_copy_var_names
.
end
())
{
std
::
string
input_var_rename
=
input_var_name
+
kSuffix
;
framework
::
Variable
*
input_var
=
scope
.
FindVar
(
input_var_name
);
if
(
input_var
->
IsType
<
framework
::
LoDTensor
>
())
{
rename_vars
.
push_back
(
input_var_rename
);
auto
input_var_tensor
=
input_var
->
Get
<
LoDTensor
>
();
auto
*
rename_input_var_tensor
=
current_scope
.
Var
(
input_var_rename
)
->
GetMutable
<
LoDTensor
>
();
framework
::
TensorCopy
(
input_var_tensor
,
dev_place
,
rename_input_var_tensor
);
rename_input_var_tensor
->
set_lod
(
input_var_tensor
.
lod
());
}
}
}
executor
.
RunPreparedContext
(
ctx
.
get
(),
&
current_scope
,
false
,
true
,
executor
.
RunPreparedContext
(
ctx
.
get
(),
&
current_scope
,
false
,
true
,
true
);
true
);
for
(
auto
&
var_rename
:
rename_vars
)
{
std
::
string
input_var_name
=
var_rename
.
substr
(
0
,
var_rename
.
size
()
-
strlen
(
kSuffix
));
current_scope
.
Rename
(
var_rename
,
input_var_name
);
}
cond_data
=
cond_data
=
GetCondData
(
scope
.
FindVar
(
Input
(
kCondition
))
->
Get
<
LoDTensor
>
());
GetCondData
(
scope
.
FindVar
(
Input
(
kCondition
))
->
Get
<
LoDTensor
>
());
}
}
...
@@ -312,6 +354,10 @@ class WhileGradOp : public framework::OperatorBase {
...
@@ -312,6 +354,10 @@ class WhileGradOp : public framework::OperatorBase {
// continue;
// continue;
// }
// }
auto
var_iter
=
std
::
find
(
outside_og_names
.
begin
(),
outside_og_names
.
end
(),
pg_ig_names
[
param_id
]);
// zero gradient variable in step 0
// zero gradient variable in step 0
if
(
cur_scope_iter
==
step_scopes
->
rbegin
())
{
if
(
cur_scope_iter
==
step_scopes
->
rbegin
())
{
auto
*
var
=
(
*
cur_scope_iter
)
->
FindVar
(
inside_grad_name
);
auto
*
var
=
(
*
cur_scope_iter
)
->
FindVar
(
inside_grad_name
);
...
@@ -326,7 +372,8 @@ class WhileGradOp : public framework::OperatorBase {
...
@@ -326,7 +372,8 @@ class WhileGradOp : public framework::OperatorBase {
"or LoDTensor, but the received var[%s] is %s."
,
"or LoDTensor, but the received var[%s] is %s."
,
inside_grad_name
,
framework
::
ToTypeName
(
var
->
Type
())));
inside_grad_name
,
framework
::
ToTypeName
(
var
->
Type
())));
if
(
var
->
IsType
<
LoDTensor
>
())
{
if
((
var_iter
==
outside_og_names
.
end
())
&&
var
->
IsType
<
LoDTensor
>
())
{
auto
&
inside_tensor
=
var
->
Get
<
framework
::
LoDTensor
>
();
auto
&
inside_tensor
=
var
->
Get
<
framework
::
LoDTensor
>
();
framework
::
AttributeMap
attrs
;
framework
::
AttributeMap
attrs
;
attrs
[
"dtype"
]
=
inside_tensor
.
type
();
attrs
[
"dtype"
]
=
inside_tensor
.
type
();
...
@@ -343,13 +390,18 @@ class WhileGradOp : public framework::OperatorBase {
...
@@ -343,13 +390,18 @@ class WhileGradOp : public framework::OperatorBase {
->
set_lod
(
inside_tensor
.
lod
());
->
set_lod
(
inside_tensor
.
lod
());
}
}
}
}
auto
new_inside_name
=
cur_scope
.
Rename
(
inside_grad_name
);
auto
var_outside
=
scope
.
FindVar
(
pg_ig_names
[
param_id
]);
auto
sum_op
=
framework
::
OpRegistry
::
CreateOp
(
if
((
var_iter
==
outside_og_names
.
end
())
||
"sum"
,
{{
"X"
,
{
pg_ig_names
[
param_id
],
new_inside_name
}}},
((
var_iter
!=
outside_og_names
.
end
())
&&
{{
"Out"
,
{
pg_ig_names
[
param_id
]}}},
var_outside
->
IsType
<
framework
::
LoDTensorArray
>
()))
{
framework
::
AttributeMap
{{
"use_mkldnn"
,
{
false
}}});
auto
new_inside_name
=
cur_scope
.
Rename
(
inside_grad_name
);
sum_op
->
Run
(
cur_scope
,
dev_place
);
auto
sum_op
=
framework
::
OpRegistry
::
CreateOp
(
cur_scope
.
Rename
(
new_inside_name
,
inside_grad_name
);
"sum"
,
{{
"X"
,
{
pg_ig_names
[
param_id
],
new_inside_name
}}},
{{
"Out"
,
{
pg_ig_names
[
param_id
]}}},
framework
::
AttributeMap
{{
"use_mkldnn"
,
{
false
}}});
sum_op
->
Run
(
cur_scope
,
dev_place
);
cur_scope
.
Rename
(
new_inside_name
,
inside_grad_name
);
}
}
}
dev_ctx
.
Wait
();
dev_ctx
.
Wait
();
const_cast
<
framework
::
Scope
&>
(
scope
).
DeleteScope
(
&
cur_scope
);
const_cast
<
framework
::
Scope
&>
(
scope
).
DeleteScope
(
&
cur_scope
);
...
...
paddle/fluid/operators/controlflow/while_op_helper.cc
浏览文件 @
2e1bca99
...
@@ -232,5 +232,16 @@ bool GetCondData(const framework::LoDTensor &cond) {
...
@@ -232,5 +232,16 @@ bool GetCondData(const framework::LoDTensor &cond) {
return
cpu_cond
->
data
<
bool
>
()[
0
];
return
cpu_cond
->
data
<
bool
>
()[
0
];
}
}
bool
StrInVaraiableNameMap
(
const
std
::
string
&
name
,
const
framework
::
VariableNameMap
&
var_names
)
{
for
(
auto
&
ipt
:
var_names
)
{
if
(
std
::
find
(
ipt
.
second
.
begin
(),
ipt
.
second
.
end
(),
name
)
!=
ipt
.
second
.
end
())
{
return
true
;
}
}
return
false
;
}
}
// namespace operators
}
// namespace operators
}
// namespace paddle
}
// namespace paddle
paddle/fluid/operators/controlflow/while_op_helper.h
浏览文件 @
2e1bca99
...
@@ -38,6 +38,7 @@ static constexpr char kX[] = "X";
...
@@ -38,6 +38,7 @@ static constexpr char kX[] = "X";
static
constexpr
char
kXGRAD
[]
=
"X@GRAD"
;
static
constexpr
char
kXGRAD
[]
=
"X@GRAD"
;
static
constexpr
char
kOutputs
[]
=
"Out"
;
static
constexpr
char
kOutputs
[]
=
"Out"
;
static
constexpr
char
kSkipEagerDeletionVars
[]
=
"skip_eager_deletion_vars"
;
static
constexpr
char
kSkipEagerDeletionVars
[]
=
"skip_eager_deletion_vars"
;
static
constexpr
char
kSuffix
[]
=
"@TMP_COPY"
;
void
PrepareSafeEagerDeletionOnWhileOpAndWhileGradOp
(
void
PrepareSafeEagerDeletionOnWhileOpAndWhileGradOp
(
const
framework
::
ProgramDesc
&
program
,
int
block_id
,
const
framework
::
ProgramDesc
&
program
,
int
block_id
,
...
@@ -50,5 +51,8 @@ void PrepareSafeEagerDeletionOnWhileOpAndWhileGradOp(
...
@@ -50,5 +51,8 @@ void PrepareSafeEagerDeletionOnWhileOpAndWhileGradOp(
bool
GetCondData
(
const
framework
::
LoDTensor
&
cond
);
bool
GetCondData
(
const
framework
::
LoDTensor
&
cond
);
bool
StrInVaraiableNameMap
(
const
std
::
string
&
,
const
framework
::
VariableNameMap
&
);
}
// namespace operators
}
// namespace operators
}
// namespace paddle
}
// namespace paddle
python/paddle/fluid/tests/unittests/test_while_loop_op.py
浏览文件 @
2e1bca99
...
@@ -16,6 +16,7 @@ from __future__ import print_function
...
@@ -16,6 +16,7 @@ from __future__ import print_function
import
numpy
as
np
import
numpy
as
np
import
unittest
import
unittest
import
paddle
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
import
paddle.fluid.core
as
core
import
paddle.fluid.core
as
core
import
paddle.fluid.layers
as
layers
import
paddle.fluid.layers
as
layers
...
@@ -24,6 +25,8 @@ from paddle.fluid.executor import Executor
...
@@ -24,6 +25,8 @@ from paddle.fluid.executor import Executor
from
paddle.fluid.framework
import
Program
,
program_guard
from
paddle.fluid.framework
import
Program
,
program_guard
from
paddle.fluid.backward
import
append_backward
from
paddle.fluid.backward
import
append_backward
paddle
.
enable_static
()
class
TestApiWhileLoop
(
unittest
.
TestCase
):
class
TestApiWhileLoop
(
unittest
.
TestCase
):
def
test_var_tuple
(
self
):
def
test_var_tuple
(
self
):
...
@@ -199,16 +202,10 @@ class TestApiWhileLoop_Backward(unittest.TestCase):
...
@@ -199,16 +202,10 @@ class TestApiWhileLoop_Backward(unittest.TestCase):
def
cond
(
i
,
x
):
def
cond
(
i
,
x
):
return
layers
.
less_than
(
i
,
eleven
)
return
layers
.
less_than
(
i
,
eleven
)
def
body
(
j
,
x
):
def
body
(
i
,
x
):
# TODO: In while block, if the var created in parent block
# participates in the calculation of gradient, the result of gradient
# is incorrect because each step scope always returns the same value
# generated by last step.
# Here we call `assign` op in while block to avoid this bug, and working on fixing it in next PR.
i
=
layers
.
assign
(
j
)
x
=
layers
.
elementwise_mul
(
x
=
i
,
y
=
i
)
x
=
layers
.
elementwise_mul
(
x
=
i
,
y
=
i
)
j
=
layers
.
increment
(
j
)
i
=
layers
.
increment
(
i
)
return
[
j
,
x
]
return
[
i
,
x
]
main_program
=
Program
()
main_program
=
Program
()
startup_program
=
Program
()
startup_program
=
Program
()
...
@@ -244,10 +241,10 @@ class TestApiWhileLoop_Backward(unittest.TestCase):
...
@@ -244,10 +241,10 @@ class TestApiWhileLoop_Backward(unittest.TestCase):
def
test_while_loop_backward2
(
self
):
def
test_while_loop_backward2
(
self
):
def
cond
(
i
,
x
):
def
cond
(
i
,
x
):
return
i
<
5
return
i
<
3
def
body
(
i
,
x
):
def
body
(
i
,
x
):
x
=
x
+
i
x
=
x
*
i
i
=
i
+
1
i
=
i
+
1
return
[
i
,
x
]
return
[
i
,
x
]
...
@@ -269,17 +266,21 @@ class TestApiWhileLoop_Backward(unittest.TestCase):
...
@@ -269,17 +266,21 @@ class TestApiWhileLoop_Backward(unittest.TestCase):
feed_i
=
np
.
ones
(
1
).
astype
(
'float32'
)
feed_i
=
np
.
ones
(
1
).
astype
(
'float32'
)
feed_x
=
np
.
ones
(
1
).
astype
(
'float32'
)
feed_x
=
np
.
ones
(
1
).
astype
(
'float32'
)
data
=
np
.
asarray
([
11
]).
astype
(
'float32'
)
data
=
np
.
asarray
([
2
]).
astype
(
'float32'
)
i_grad
=
np
.
asarray
([
1
]).
astype
(
'float32'
)
i_grad
=
np
.
asarray
([
3
]).
astype
(
'float32'
)
x_grad
=
np
.
asarray
([
2
]).
astype
(
'float32'
)
res
=
exe
.
run
(
main_program
,
res
=
exe
.
run
(
main_program
,
feed
=
{
'i'
:
feed_i
,
feed
=
{
'i'
:
feed_i
,
'x'
:
feed_x
},
'x'
:
feed_x
},
fetch_list
=
[
mean
.
name
,
i
.
grad_name
])
fetch_list
=
[
mean
.
name
,
i
.
grad_name
,
x
.
grad_name
])
self
.
assertTrue
(
np
.
allclose
(
np
.
asarray
(
res
[
0
]),
data
))
self
.
assertTrue
(
np
.
allclose
(
np
.
asarray
(
res
[
0
]),
data
))
self
.
assertTrue
(
self
.
assertTrue
(
np
.
allclose
(
np
.
asarray
(
res
[
1
]),
i_grad
),
np
.
allclose
(
np
.
asarray
(
res
[
1
]),
i_grad
),
msg
=
"
\n
res =
\n
{}
\n\n
ans =
\n
{}"
.
format
(
res
[
1
],
i_grad
))
msg
=
"
\n
res =
\n
{}
\n\n
ans =
\n
{}"
.
format
(
res
[
1
],
i_grad
))
self
.
assertTrue
(
np
.
allclose
(
np
.
asarray
(
res
[
2
]),
x_grad
),
msg
=
"
\n
res =
\n
{}
\n\n
ans =
\n
{}"
.
format
(
res
[
2
],
x_grad
))
class
TestApiWhileLoop_NestedWithBackwardAndLoDTensorArray
(
unittest
.
TestCase
):
class
TestApiWhileLoop_NestedWithBackwardAndLoDTensorArray
(
unittest
.
TestCase
):
...
...
python/paddle/fluid/tests/unittests/test_while_op.py
浏览文件 @
2e1bca99
...
@@ -24,6 +24,8 @@ from paddle.fluid.backward import append_backward
...
@@ -24,6 +24,8 @@ from paddle.fluid.backward import append_backward
import
numpy
import
numpy
from
paddle.fluid
import
compiler
,
Program
,
program_guard
from
paddle.fluid
import
compiler
,
Program
,
program_guard
paddle
.
enable_static
()
class
TestWhileOp
(
unittest
.
TestCase
):
class
TestWhileOp
(
unittest
.
TestCase
):
def
simple_net
(
self
):
def
simple_net
(
self
):
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录