Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
8c43c0fe
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
8c43c0fe
编写于
7月 29, 2022
作者:
J
Jiabin Yang
提交者:
GitHub
7月 29, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Support backward final hook (#44686)
上级
b7496bcb
变更
17
隐藏空白更改
内联
并排
Showing
17 changed file
with
259 addition
and
130 deletion
+259
-130
paddle/fluid/distributed/collective/reducer.cc
paddle/fluid/distributed/collective/reducer.cc
+1
-1
paddle/fluid/eager/accumulation/accumulation_node.cc
paddle/fluid/eager/accumulation/accumulation_node.cc
+1
-1
paddle/fluid/eager/accumulation/accumulation_node.h
paddle/fluid/eager/accumulation/accumulation_node.h
+2
-2
paddle/fluid/eager/api/utils/global_utils.h
paddle/fluid/eager/api/utils/global_utils.h
+18
-1
paddle/fluid/eager/api/utils/hook_utils.cc
paddle/fluid/eager/api/utils/hook_utils.cc
+12
-6
paddle/fluid/eager/api/utils/hook_utils.h
paddle/fluid/eager/api/utils/hook_utils.h
+5
-2
paddle/fluid/eager/backward.cc
paddle/fluid/eager/backward.cc
+6
-0
paddle/fluid/eager/hooks.h
paddle/fluid/eager/hooks.h
+7
-6
paddle/fluid/eager/tests/data_structure_tests/accumulation_node_test.cc
...ager/tests/data_structure_tests/accumulation_node_test.cc
+2
-4
paddle/fluid/eager/tests/task_tests/fwd_bwd_joint_test.cc
paddle/fluid/eager/tests/task_tests/fwd_bwd_joint_test.cc
+6
-6
paddle/fluid/eager/tests/task_tests/hook_test.cc
paddle/fluid/eager/tests/task_tests/hook_test.cc
+4
-8
paddle/fluid/eager/tests/task_tests/hook_test_intermidiate.cc
...le/fluid/eager/tests/task_tests/hook_test_intermidiate.cc
+64
-12
paddle/fluid/pybind/eager_functions.cc
paddle/fluid/pybind/eager_functions.cc
+15
-0
paddle/fluid/pybind/eager_method.cc
paddle/fluid/pybind/eager_method.cc
+1
-81
paddle/fluid/pybind/eager_utils.cc
paddle/fluid/pybind/eager_utils.cc
+50
-0
paddle/fluid/pybind/eager_utils.h
paddle/fluid/pybind/eager_utils.h
+35
-0
python/paddle/fluid/tests/unittests/test_tensor_register_hook.py
...paddle/fluid/tests/unittests/test_tensor_register_hook.py
+30
-0
未找到文件。
paddle/fluid/distributed/collective/reducer.cc
浏览文件 @
8c43c0fe
...
@@ -321,7 +321,7 @@ EagerReducer::EagerReducer(
...
@@ -321,7 +321,7 @@ EagerReducer::EagerReducer(
const
auto
&
accumulation_grad_node
=
const
auto
&
accumulation_grad_node
=
std
::
dynamic_pointer_cast
<
egr
::
GradNodeAccumulation
>
(
grad_node
);
std
::
dynamic_pointer_cast
<
egr
::
GradNodeAccumulation
>
(
grad_node
);
accumulation_grad_node
->
RegisterReduceHook
(
accumulation_grad_node
->
RegisterReduceHook
(
std
::
make_shared
<
egr
::
Cpp
Tensor
VoidHook
>
(
reduce_hook
));
std
::
make_shared
<
egr
::
CppVoidHook
>
(
reduce_hook
));
gradnode_index_map_
[
grad_node
.
get
()]
=
global_var_index
;
gradnode_index_map_
[
grad_node
.
get
()]
=
global_var_index
;
}
}
...
...
paddle/fluid/eager/accumulation/accumulation_node.cc
浏览文件 @
8c43c0fe
...
@@ -127,7 +127,7 @@ GradNodeAccumulation::operator()(
...
@@ -127,7 +127,7 @@ GradNodeAccumulation::operator()(
}
}
void
GradNodeAccumulation
::
RegisterReduceHook
(
void
GradNodeAccumulation
::
RegisterReduceHook
(
std
::
shared_ptr
<
Tensor
VoidHook
>&&
hook
)
{
std
::
shared_ptr
<
VoidHook
>&&
hook
)
{
reduce_hooks_
.
emplace_back
(
std
::
move
(
hook
));
reduce_hooks_
.
emplace_back
(
std
::
move
(
hook
));
}
}
...
...
paddle/fluid/eager/accumulation/accumulation_node.h
浏览文件 @
8c43c0fe
...
@@ -51,7 +51,7 @@ class GradNodeAccumulation : public GradNodeBase {
...
@@ -51,7 +51,7 @@ class GradNodeAccumulation : public GradNodeBase {
/**
/**
* Register ReduceHook
* Register ReduceHook
* **/
* **/
void
RegisterReduceHook
(
std
::
shared_ptr
<
Tensor
VoidHook
>&&
hook
);
void
RegisterReduceHook
(
std
::
shared_ptr
<
VoidHook
>&&
hook
);
/**
/**
* Apply ReduceHook here
* Apply ReduceHook here
...
@@ -70,7 +70,7 @@ class GradNodeAccumulation : public GradNodeBase {
...
@@ -70,7 +70,7 @@ class GradNodeAccumulation : public GradNodeBase {
// TODO(Jiabin): remove this when we make our clear gradient really cleared;
// TODO(Jiabin): remove this when we make our clear gradient really cleared;
bool
is_fake_empty_
=
{
false
};
bool
is_fake_empty_
=
{
false
};
std
::
weak_ptr
<
paddle
::
experimental
::
Tensor
>
weak_grad_
;
std
::
weak_ptr
<
paddle
::
experimental
::
Tensor
>
weak_grad_
;
std
::
vector
<
std
::
shared_ptr
<
Tensor
VoidHook
>>
reduce_hooks_
;
std
::
vector
<
std
::
shared_ptr
<
VoidHook
>>
reduce_hooks_
;
std
::
function
<
paddle
::
experimental
::
Tensor
(
std
::
function
<
paddle
::
experimental
::
Tensor
(
const
paddle
::
experimental
::
Tensor
&
)
>
const
paddle
::
experimental
::
Tensor
&
)
>
retain_grad_hook_
;
retain_grad_hook_
;
...
...
paddle/fluid/eager/api/utils/global_utils.h
浏览文件 @
8c43c0fe
...
@@ -18,11 +18,11 @@
...
@@ -18,11 +18,11 @@
#include <atomic>
#include <atomic>
#include <memory>
#include <memory>
#include "paddle/fluid/eager/hooks.h"
#include "paddle/fluid/eager/type_defs.h"
#include "paddle/fluid/eager/type_defs.h"
#include "paddle/fluid/imperative/tracer.h"
#include "paddle/fluid/imperative/tracer.h"
#include "paddle/phi/api/ext/op_meta_info.h"
#include "paddle/phi/api/ext/op_meta_info.h"
#include "paddle/utils/small_vector.h"
#include "paddle/utils/small_vector.h"
namespace
egr
{
namespace
egr
{
class
UniqueNameGenerator
{
class
UniqueNameGenerator
{
public:
public:
...
@@ -85,6 +85,22 @@ class Controller {
...
@@ -85,6 +85,22 @@ class Controller {
GetCustomEdgesSlotMap
()
{
GetCustomEdgesSlotMap
()
{
return
custom_edges_slot_map_
;
return
custom_edges_slot_map_
;
}
}
// For Cpp Hook
void
RegisterBackwardFinalHook
(
const
std
::
function
<
void
()
>&
call_back
)
{
VLOG
(
6
)
<<
"RegisterBackwardFinalHook"
;
final_backward_hooks_
.
emplace_back
(
std
::
make_shared
<
CppVoidHook
>
(
std
::
move
(
call_back
)));
VLOG
(
6
)
<<
"Size: "
<<
final_backward_hooks_
.
size
();
}
// For Python hook
void
RegisterBackwardFinalHook
(
const
std
::
shared_ptr
<
VoidHook
>&
call_back
)
{
final_backward_hooks_
.
emplace_back
(
call_back
);
}
const
std
::
vector
<
std
::
shared_ptr
<
VoidHook
>>&
FinalBackwardHooks
()
const
{
return
final_backward_hooks_
;
}
void
ClearFinalBackwardHooks
()
{
final_backward_hooks_
.
clear
();
}
private:
private:
Controller
()
=
default
;
Controller
()
=
default
;
...
@@ -98,6 +114,7 @@ class Controller {
...
@@ -98,6 +114,7 @@ class Controller {
std
::
unordered_map
<
std
::
string
,
std
::
unordered_map
<
std
::
string
,
std
::
vector
<
std
::
vector
<
std
::
unordered_map
<
int
,
int
>>>>
std
::
vector
<
std
::
vector
<
std
::
unordered_map
<
int
,
int
>>>>
custom_edges_slot_map_
;
custom_edges_slot_map_
;
std
::
vector
<
std
::
shared_ptr
<
VoidHook
>>
final_backward_hooks_
;
DISABLE_COPY_AND_ASSIGN
(
Controller
);
DISABLE_COPY_AND_ASSIGN
(
Controller
);
};
};
...
...
paddle/fluid/eager/api/utils/hook_utils.cc
浏览文件 @
8c43c0fe
...
@@ -25,17 +25,20 @@ namespace egr_utils_api {
...
@@ -25,17 +25,20 @@ namespace egr_utils_api {
int64_t
RegisterGradientHookForTensor
(
int64_t
RegisterGradientHookForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
,
const
paddle
::
experimental
::
Tensor
&
tensor
,
std
::
shared_ptr
<
egr
::
TensorHook
>&&
hook
)
{
const
std
::
function
<
paddle
::
experimental
::
Tensor
(
const
paddle
::
experimental
::
Tensor
&
)
>&
hook
)
{
// Find grad_node and out_rank from AutogradMeta
// Find grad_node and out_rank from AutogradMeta
std
::
shared_ptr
<
GradNodeBase
>
grad_node
=
EagerUtils
::
grad_node
(
tensor
);
std
::
shared_ptr
<
GradNodeBase
>
grad_node
=
EagerUtils
::
grad_node
(
tensor
);
auto
rank_info
=
EagerUtils
::
unsafe_autograd_meta
(
tensor
)
->
OutRankInfo
();
auto
rank_info
=
EagerUtils
::
unsafe_autograd_meta
(
tensor
)
->
OutRankInfo
();
return
grad_node
->
RegisterGradientHook
(
return
grad_node
->
RegisterGradientHook
(
rank_info
.
first
,
rank_info
.
second
,
std
::
move
(
hook
));
rank_info
.
first
,
rank_info
.
second
,
std
::
move
(
std
::
make_shared
<
CppTensorHook
>
(
hook
)));
}
}
void
RegisterReduceHookForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
,
void
RegisterReduceHookForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
,
std
::
shared_ptr
<
egr
::
TensorVoidHook
>&
&
hook
)
{
const
std
::
function
<
void
()
>
&
hook
)
{
if
(
IsLeafTensor
(
tensor
))
{
if
(
IsLeafTensor
(
tensor
))
{
VLOG
(
6
)
<<
"Register ReduceHook for leaf tensor"
;
VLOG
(
6
)
<<
"Register ReduceHook for leaf tensor"
;
std
::
shared_ptr
<
GradNodeBase
>
grad_node
=
EagerUtils
::
grad_node
(
tensor
);
std
::
shared_ptr
<
GradNodeBase
>
grad_node
=
EagerUtils
::
grad_node
(
tensor
);
...
@@ -46,7 +49,8 @@ void RegisterReduceHookForTensor(const paddle::experimental::Tensor& tensor,
...
@@ -46,7 +49,8 @@ void RegisterReduceHookForTensor(const paddle::experimental::Tensor& tensor,
"with type: GradNodeAccumulation"
));
"with type: GradNodeAccumulation"
));
auto
accumulation_grad_node
=
auto
accumulation_grad_node
=
std
::
dynamic_pointer_cast
<
GradNodeAccumulation
>
(
grad_node
);
std
::
dynamic_pointer_cast
<
GradNodeAccumulation
>
(
grad_node
);
accumulation_grad_node
->
RegisterReduceHook
(
std
::
move
(
hook
));
accumulation_grad_node
->
RegisterReduceHook
(
std
::
move
(
std
::
make_shared
<
CppVoidHook
>
(
hook
)));
}
else
{
}
else
{
PADDLE_THROW
(
paddle
::
platform
::
errors
::
Fatal
(
PADDLE_THROW
(
paddle
::
platform
::
errors
::
Fatal
(
"Only can register reduce hook for leaf Tensor."
));
"Only can register reduce hook for leaf Tensor."
));
...
@@ -90,10 +94,12 @@ void RetainGradForTensor(const paddle::experimental::Tensor& tensor) {
...
@@ -90,10 +94,12 @@ void RetainGradForTensor(const paddle::experimental::Tensor& tensor) {
};
};
// Append to GradientHooks
// Append to GradientHooks
RegisterGradientHookForTensor
(
tensor
,
RegisterGradientHookForTensor
(
tensor
,
hook
);
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook
));
}
}
}
}
void
RegisterBackwardFinalHook
(
const
std
::
function
<
void
()
>&
hook
)
{
Controller
::
Instance
().
RegisterBackwardFinalHook
(
hook
);
}
}
// namespace egr_utils_api
}
// namespace egr_utils_api
}
// namespace egr
}
// namespace egr
paddle/fluid/eager/api/utils/hook_utils.h
浏览文件 @
8c43c0fe
...
@@ -23,11 +23,14 @@ namespace egr_utils_api {
...
@@ -23,11 +23,14 @@ namespace egr_utils_api {
int64_t
RegisterGradientHookForTensor
(
int64_t
RegisterGradientHookForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
,
const
paddle
::
experimental
::
Tensor
&
tensor
,
std
::
shared_ptr
<
egr
::
TensorHook
>&&
hook
);
const
std
::
function
<
paddle
::
experimental
::
Tensor
(
const
paddle
::
experimental
::
Tensor
&
)
>&
hook
);
void
RegisterReduceHookForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
,
void
RegisterReduceHookForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
,
std
::
shared_ptr
<
egr
::
TensorVoidHook
>&
&
hook
);
const
std
::
function
<
void
()
>
&
hook
);
void
RetainGradForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
);
void
RetainGradForTensor
(
const
paddle
::
experimental
::
Tensor
&
tensor
);
void
RegisterBackwardFinalHook
(
const
std
::
function
<
void
()
>&
hook
);
}
// namespace egr_utils_api
}
// namespace egr_utils_api
}
// namespace egr
}
// namespace egr
paddle/fluid/eager/backward.cc
浏览文件 @
8c43c0fe
...
@@ -371,6 +371,12 @@ std::vector<paddle::experimental::Tensor> RunBackward(
...
@@ -371,6 +371,12 @@ std::vector<paddle::experimental::Tensor> RunBackward(
}
}
}
}
VLOG
(
6
)
<<
"Run Backward Final hook size: "
<<
egr
::
Controller
::
Instance
().
FinalBackwardHooks
().
size
();
for
(
auto
&
hook
:
egr
::
Controller
::
Instance
().
FinalBackwardHooks
())
{
(
*
hook
)();
}
egr
::
Controller
::
Instance
().
ClearFinalBackwardHooks
();
if
(
!
is_general_grad
)
return
{};
if
(
!
is_general_grad
)
return
{};
return
GeneralGrad
::
Instance
().
GetResults
(
inputs
,
allow_unused
,
create_graph
);
return
GeneralGrad
::
Instance
().
GetResults
(
inputs
,
allow_unused
,
create_graph
);
}
}
...
...
paddle/fluid/eager/hooks.h
浏览文件 @
8c43c0fe
...
@@ -29,16 +29,16 @@ class TensorHook {
...
@@ -29,16 +29,16 @@ class TensorHook {
const
paddle
::
experimental
::
Tensor
&
var
)
=
0
;
const
paddle
::
experimental
::
Tensor
&
var
)
=
0
;
};
};
class
Tensor
VoidHook
{
class
VoidHook
{
public:
public:
virtual
~
Tensor
VoidHook
()
=
default
;
virtual
~
VoidHook
()
=
default
;
virtual
void
operator
()()
=
0
;
virtual
void
operator
()()
=
0
;
};
};
class
CppTensorHook
:
public
TensorHook
{
class
CppTensorHook
:
public
TensorHook
{
public:
public:
explicit
CppTensorHook
(
std
::
function
<
paddle
::
experimental
::
Tensor
(
explicit
CppTensorHook
(
const
std
::
function
<
paddle
::
experimental
::
Tensor
(
const
paddle
::
experimental
::
Tensor
&
)
>&
&
fn
)
const
paddle
::
experimental
::
Tensor
&
)
>&
fn
)
:
fn_
(
std
::
move
(
fn
))
{}
:
fn_
(
std
::
move
(
fn
))
{}
paddle
::
experimental
::
Tensor
operator
()(
paddle
::
experimental
::
Tensor
operator
()(
...
@@ -52,13 +52,14 @@ class CppTensorHook : public TensorHook {
...
@@ -52,13 +52,14 @@ class CppTensorHook : public TensorHook {
fn_
;
fn_
;
};
};
class
Cpp
TensorVoidHook
:
public
Tensor
VoidHook
{
class
Cpp
VoidHook
:
public
VoidHook
{
public:
public:
explicit
Cpp
TensorVoidHook
(
std
::
function
<
void
()
>&
&
fn
)
:
fn_
(
std
::
move
(
fn
))
{}
explicit
Cpp
VoidHook
(
const
std
::
function
<
void
()
>
&
fn
)
:
fn_
(
std
::
move
(
fn
))
{}
void
operator
()()
override
{
return
fn_
();
}
void
operator
()()
override
{
return
fn_
();
}
private:
private:
std
::
function
<
void
()
>
fn_
;
std
::
function
<
void
()
>
fn_
;
};
};
}
// namespace egr
}
// namespace egr
paddle/fluid/eager/tests/data_structure_tests/accumulation_node_test.cc
浏览文件 @
8c43c0fe
...
@@ -328,8 +328,7 @@ TEST(AccumulationNode, Tensor) {
...
@@ -328,8 +328,7 @@ TEST(AccumulationNode, Tensor) {
VLOG
(
6
)
<<
"Running Reduce Hook"
;
VLOG
(
6
)
<<
"Running Reduce Hook"
;
};
};
node
->
RegisterReduceHook
(
node
->
RegisterReduceHook
(
std
::
make_shared
<
egr
::
CppVoidHook
>
(
reduce_hook_1
));
std
::
make_shared
<
egr
::
CppTensorVoidHook
>
(
reduce_hook_1
));
// operator()
// operator()
paddle
::
experimental
::
Tensor
_ret
=
node
->
operator
()(
et0_vec
)[
0
][
0
];
paddle
::
experimental
::
Tensor
_ret
=
node
->
operator
()(
et0_vec
)[
0
][
0
];
...
@@ -354,8 +353,7 @@ TEST(AccumulationNode, Tensor) {
...
@@ -354,8 +353,7 @@ TEST(AccumulationNode, Tensor) {
ret_et0_ptr
[
0
]
=
100.0
;
// set to 100.0
ret_et0_ptr
[
0
]
=
100.0
;
// set to 100.0
VLOG
(
6
)
<<
"Running Reduce Hook"
;
VLOG
(
6
)
<<
"Running Reduce Hook"
;
};
};
node
->
RegisterReduceHook
(
node
->
RegisterReduceHook
(
std
::
make_shared
<
egr
::
CppVoidHook
>
(
reduce_hook_2
));
std
::
make_shared
<
egr
::
CppTensorVoidHook
>
(
reduce_hook_2
));
node
->
ApplyReduceHooks
();
node
->
ApplyReduceHooks
();
// Check ApplyReduceHooks result
// Check ApplyReduceHooks result
...
...
paddle/fluid/eager/tests/task_tests/fwd_bwd_joint_test.cc
浏览文件 @
8c43c0fe
...
@@ -256,8 +256,8 @@ TEST(FwdBwdJoint, GradientHook) {
...
@@ -256,8 +256,8 @@ TEST(FwdBwdJoint, GradientHook) {
true
/*bias_after_scale*/
,
true
/*bias_after_scale*/
,
true
/*trace_backward*/
);
true
/*trace_backward*/
);
egr_utils_api
::
RetainGradForTensor
(
out0
);
// hook: +5
egr_utils_api
::
RetainGradForTensor
(
out0
);
// hook: +5
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
out0
,
out0
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
)
);
// hook: +5
hook_function
);
// hook: +5
// Run Forward Node 1
// Run Forward Node 1
float
scale1
=
5.0
;
float
scale1
=
5.0
;
...
@@ -265,8 +265,8 @@ TEST(FwdBwdJoint, GradientHook) {
...
@@ -265,8 +265,8 @@ TEST(FwdBwdJoint, GradientHook) {
paddle
::
experimental
::
Tensor
out1
=
egr
::
scale
(
paddle
::
experimental
::
Tensor
out1
=
egr
::
scale
(
out0
,
scale1
,
bias1
,
true
/*bias_after_scale*/
,
true
/*trace_backward*/
);
out0
,
scale1
,
bias1
,
true
/*bias_after_scale*/
,
true
/*trace_backward*/
);
egr_utils_api
::
RetainGradForTensor
(
out1
);
// hook: +5
egr_utils_api
::
RetainGradForTensor
(
out1
);
// hook: +5
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
out1
,
out1
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
)
);
// hook: +5
hook_function
);
// hook: +5
// Run Forward Node 2
// Run Forward Node 2
float
scale2
=
10.0
;
float
scale2
=
10.0
;
...
@@ -274,8 +274,8 @@ TEST(FwdBwdJoint, GradientHook) {
...
@@ -274,8 +274,8 @@ TEST(FwdBwdJoint, GradientHook) {
paddle
::
experimental
::
Tensor
out2
=
egr
::
scale
(
paddle
::
experimental
::
Tensor
out2
=
egr
::
scale
(
out0
,
scale2
,
bias2
,
true
/*bias_after_scale*/
,
true
/*trace_backward*/
);
out0
,
scale2
,
bias2
,
true
/*bias_after_scale*/
,
true
/*trace_backward*/
);
egr_utils_api
::
RetainGradForTensor
(
out2
);
// hook: +5
egr_utils_api
::
RetainGradForTensor
(
out2
);
// hook: +5
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
out2
,
out2
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
)
);
// hook: +5
hook_function
);
// hook: +5
// 4. Run Backward
// 4. Run Backward
std
::
vector
<
paddle
::
experimental
::
Tensor
>
outs
=
{
out1
,
out2
};
std
::
vector
<
paddle
::
experimental
::
Tensor
>
outs
=
{
out1
,
out2
};
...
...
paddle/fluid/eager/tests/task_tests/hook_test.cc
浏览文件 @
8c43c0fe
...
@@ -95,8 +95,7 @@ TEST(RetainGrad, HookBeforeRetainGrad) {
...
@@ -95,8 +95,7 @@ TEST(RetainGrad, HookBeforeRetainGrad) {
std
::
dynamic_pointer_cast
<
paddle
::
experimental
::
AbstractAutogradMeta
>
(
std
::
dynamic_pointer_cast
<
paddle
::
experimental
::
AbstractAutogradMeta
>
(
auto_grad_meta
));
auto_grad_meta
));
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
target_tensor
,
hook_function
);
target_tensor
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
));
egr_utils_api
::
RetainGradForTensor
(
egr_utils_api
::
RetainGradForTensor
(
target_tensor
);
// result: 1.0 + 3.0 = 4.0
target_tensor
);
// result: 1.0 + 3.0 = 4.0
egr_utils_api
::
RetainGradForTensor
(
egr_utils_api
::
RetainGradForTensor
(
...
@@ -122,8 +121,7 @@ TEST(RetainGrad, HookBeforeRetainGrad) {
...
@@ -122,8 +121,7 @@ TEST(RetainGrad, HookBeforeRetainGrad) {
std
::
dynamic_pointer_cast
<
paddle
::
experimental
::
AbstractAutogradMeta
>
(
std
::
dynamic_pointer_cast
<
paddle
::
experimental
::
AbstractAutogradMeta
>
(
tmp_tensor0
.
mutable_autograd_meta
()));
tmp_tensor0
.
mutable_autograd_meta
()));
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
leaf_tensor
,
hook_function
);
leaf_tensor
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
));
egr_utils_api
::
RetainGradForTensor
(
egr_utils_api
::
RetainGradForTensor
(
leaf_tensor
);
// result: 4.0*5.0 + 3.0 = 23.0
leaf_tensor
);
// result: 4.0*5.0 + 3.0 = 23.0
}
}
...
@@ -173,8 +171,7 @@ TEST(RetainGrad, HookAfterRetainGrad) {
...
@@ -173,8 +171,7 @@ TEST(RetainGrad, HookAfterRetainGrad) {
auto_grad_meta
));
auto_grad_meta
));
egr_utils_api
::
RetainGradForTensor
(
target_tensor
);
// result: 1.0
egr_utils_api
::
RetainGradForTensor
(
target_tensor
);
// result: 1.0
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
target_tensor
,
hook_function
);
target_tensor
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
));
}
}
// Retain Grad for leaf tensor1
// Retain Grad for leaf tensor1
...
@@ -193,8 +190,7 @@ TEST(RetainGrad, HookAfterRetainGrad) {
...
@@ -193,8 +190,7 @@ TEST(RetainGrad, HookAfterRetainGrad) {
std
::
dynamic_pointer_cast
<
paddle
::
experimental
::
AbstractAutogradMeta
>
(
std
::
dynamic_pointer_cast
<
paddle
::
experimental
::
AbstractAutogradMeta
>
(
tmp_tensor0
.
mutable_autograd_meta
()));
tmp_tensor0
.
mutable_autograd_meta
()));
egr_utils_api
::
RegisterGradientHookForTensor
(
egr_utils_api
::
RegisterGradientHookForTensor
(
leaf_tensor
,
hook_function
);
leaf_tensor
,
std
::
make_shared
<
egr
::
CppTensorHook
>
(
hook_function
));
}
}
Backward
(
target_tensors
,
{});
Backward
(
target_tensors
,
{});
...
...
paddle/fluid/eager/tests/task_tests/hook_test_intermidiate.cc
浏览文件 @
8c43c0fe
...
@@ -89,12 +89,11 @@ void test_sigmoid(bool is_remove_gradient_hook) {
...
@@ -89,12 +89,11 @@ void test_sigmoid(bool is_remove_gradient_hook) {
egr_utils_api
::
RetainGradForTensor
(
tensor
);
egr_utils_api
::
RetainGradForTensor
(
tensor
);
VLOG
(
6
)
<<
"Register GradientHook for Tensor"
;
VLOG
(
6
)
<<
"Register GradientHook for Tensor"
;
int64_t
hook_id
=
egr_utils_api
::
RegisterGradientHookForTensor
(
int64_t
hook_id
=
tensor
,
std
::
make_shared
<
CppTensorHook
>
(
hook_function
)
);
egr_utils_api
::
RegisterGradientHookForTensor
(
tensor
,
hook_function
);
VLOG
(
6
)
<<
"Register ReduceHook for Tensor"
;
VLOG
(
6
)
<<
"Register ReduceHook for Tensor"
;
egr_utils_api
::
RegisterReduceHookForTensor
(
egr_utils_api
::
RegisterReduceHookForTensor
(
tensor
,
reduce_hook
);
tensor
,
std
::
make_shared
<
CppTensorVoidHook
>
(
reduce_hook
));
VLOG
(
6
)
<<
"Runing Forward"
;
VLOG
(
6
)
<<
"Runing Forward"
;
auto
output_tensor
=
sigmoid_dygraph_function
(
tensor
,
{});
auto
output_tensor
=
sigmoid_dygraph_function
(
tensor
,
{});
...
@@ -161,10 +160,9 @@ void test_elementwiseAdd(bool is_remove_gradient_hook) {
...
@@ -161,10 +160,9 @@ void test_elementwiseAdd(bool is_remove_gradient_hook) {
};
};
egr_utils_api
::
RetainGradForTensor
(
Y
);
egr_utils_api
::
RetainGradForTensor
(
Y
);
int64_t
hook_id
=
egr_utils_api
::
RegisterGradientHookForTensor
(
int64_t
hook_id
=
Y
,
std
::
make_shared
<
CppTensorHook
>
(
hook_function
));
egr_utils_api
::
RegisterGradientHookForTensor
(
Y
,
hook_function
);
egr_utils_api
::
RegisterReduceHookForTensor
(
egr_utils_api
::
RegisterReduceHookForTensor
(
Y
,
reduce_hook
);
Y
,
std
::
make_shared
<
CppTensorVoidHook
>
(
reduce_hook
));
auto
output_tensor
=
elementwise_add_dygraph_function
(
X
,
Y
,
{});
auto
output_tensor
=
elementwise_add_dygraph_function
(
X
,
Y
,
{});
...
@@ -226,10 +224,9 @@ void test_matmul(bool is_remove_gradient_hook) {
...
@@ -226,10 +224,9 @@ void test_matmul(bool is_remove_gradient_hook) {
};
};
egr_utils_api
::
RetainGradForTensor
(
Y
);
egr_utils_api
::
RetainGradForTensor
(
Y
);
int64_t
hook_id
=
egr_utils_api
::
RegisterGradientHookForTensor
(
int64_t
hook_id
=
Y
,
std
::
make_shared
<
CppTensorHook
>
(
hook_function
));
egr_utils_api
::
RegisterGradientHookForTensor
(
Y
,
hook_function
);
egr_utils_api
::
RegisterReduceHookForTensor
(
egr_utils_api
::
RegisterReduceHookForTensor
(
Y
,
reduce_hook
);
Y
,
std
::
make_shared
<
CppTensorVoidHook
>
(
reduce_hook
));
auto
output_tensor
=
matmul_v2_dygraph_function
(
auto
output_tensor
=
matmul_v2_dygraph_function
(
X
,
Y
,
{{
"trans_x"
,
false
},
{
"trans_y"
,
false
}});
X
,
Y
,
{{
"trans_x"
,
false
},
{
"trans_y"
,
false
}});
...
@@ -256,6 +253,59 @@ void test_matmul(bool is_remove_gradient_hook) {
...
@@ -256,6 +253,59 @@ void test_matmul(bool is_remove_gradient_hook) {
}
}
}
}
void
test_backward_final_hooks
()
{
// Prepare Device Contexts
VLOG
(
6
)
<<
"Init Env"
;
eager_test
::
InitEnv
(
paddle
::
platform
::
CPUPlace
());
VLOG
(
6
)
<<
"Make paddle::experimental::Tensor"
;
paddle
::
framework
::
DDim
ddimX
=
phi
::
make_ddim
({
4
,
16
});
paddle
::
experimental
::
Tensor
X
=
egr_utils_api
::
CreateTensorWithValue
(
ddimX
,
paddle
::
platform
::
CPUPlace
(),
phi
::
DataType
::
FLOAT32
,
phi
::
DataLayout
::
NCHW
,
3.0
,
true
);
paddle
::
framework
::
DDim
ddimY
=
phi
::
make_ddim
({
16
,
20
});
egr_utils_api
::
RetainGradForTensor
(
X
);
paddle
::
experimental
::
Tensor
Y
=
egr_utils_api
::
CreateTensorWithValue
(
ddimY
,
paddle
::
platform
::
CPUPlace
(),
phi
::
DataType
::
FLOAT32
,
phi
::
DataLayout
::
NCHW
,
2.0
,
true
);
VLOG
(
6
)
<<
"Make ReduceHook function"
;
auto
backward_final_hook
=
[
&
](
void
)
->
void
{
auto
*
t_ptr
=
std
::
dynamic_pointer_cast
<
phi
::
DenseTensor
>
(
X
.
impl
())
->
data
<
float
>
();
VLOG
(
6
)
<<
"Run Target Backward Hook"
;
for
(
int
i
=
0
;
i
<
X
.
numel
();
i
++
)
{
t_ptr
[
i
]
=
100.0
;
// set to 100.0
}
};
VLOG
(
6
)
<<
"Register Backward Final Hook"
;
egr_utils_api
::
RegisterBackwardFinalHook
(
backward_final_hook
);
VLOG
(
6
)
<<
"Runing Forward"
;
auto
output_tensor
=
matmul_v2_dygraph_function
(
X
,
Y
,
{{
"trans_x"
,
false
},
{
"trans_y"
,
false
}});
auto
res
=
sigmoid_dygraph_function
(
output_tensor
,
{});
VLOG
(
6
)
<<
"Finish Forward"
;
eager_test
::
CompareTensorWithValue
<
float
>
(
X
,
3.0
);
std
::
vector
<
paddle
::
experimental
::
Tensor
>
target_tensors
=
{
output_tensor
};
VLOG
(
6
)
<<
"Runing Backward"
;
Backward
(
target_tensors
,
{});
VLOG
(
6
)
<<
"Finish Backward"
;
eager_test
::
CompareTensorWithValue
<
float
>
(
X
,
100.0
);
}
TEST
(
Hook_intermidiate
,
Sigmoid
)
{
TEST
(
Hook_intermidiate
,
Sigmoid
)
{
// True or false represents whether to call RemoveGradientHook
// True or false represents whether to call RemoveGradientHook
test_sigmoid
(
true
);
test_sigmoid
(
true
);
...
@@ -271,6 +321,8 @@ TEST(Hook_intermidiate, Matmul_v2) {
...
@@ -271,6 +321,8 @@ TEST(Hook_intermidiate, Matmul_v2) {
test_matmul
(
true
);
test_matmul
(
true
);
test_matmul
(
false
);
test_matmul
(
false
);
}
}
TEST
(
Hook_intermidiate
,
BackwardFinal
)
{
test_backward_final_hooks
();
}
}
// namespace egr
}
// namespace egr
USE_OP_ITSELF
(
sigmoid
);
USE_OP_ITSELF
(
sigmoid
);
...
...
paddle/fluid/pybind/eager_functions.cc
浏览文件 @
8c43c0fe
...
@@ -907,12 +907,27 @@ static PyObject* eager_api_to_uva_tensor(PyObject* self,
...
@@ -907,12 +907,27 @@ static PyObject* eager_api_to_uva_tensor(PyObject* self,
}
}
#endif
#endif
static
PyObject
*
eager_api__add_backward_final_hook
(
PyObject
*
self
,
PyObject
*
args
,
PyObject
*
kwargs
)
{
EAGER_TRY
PyObject
*
hook_func
=
PyTuple_GET_ITEM
(
args
,
0
);
egr
::
Controller
::
Instance
().
RegisterBackwardFinalHook
(
std
::
make_shared
<
PyVoidHook
>
(
hook_func
));
RETURN_PY_NONE
EAGER_CATCH_AND_THROW_RETURN_NULL
}
PyMethodDef
variable_functions
[]
=
{
PyMethodDef
variable_functions
[]
=
{
// TODO(jiabin): Remove scale when we have final state tests
// TODO(jiabin): Remove scale when we have final state tests
{
"scale"
,
{
"scale"
,
(
PyCFunction
)(
void
(
*
)(
void
))
eager_api_scale
,
(
PyCFunction
)(
void
(
*
)(
void
))
eager_api_scale
,
METH_VARARGS
|
METH_KEYWORDS
,
METH_VARARGS
|
METH_KEYWORDS
,
NULL
},
NULL
},
{
"_add_backward_final_hook"
,
(
PyCFunction
)(
void
(
*
)(
void
))
eager_api__add_backward_final_hook
,
METH_VARARGS
|
METH_KEYWORDS
,
NULL
},
{
"run_backward"
,
{
"run_backward"
,
(
PyCFunction
)(
void
(
*
)(
void
))
eager_api_run_backward
,
(
PyCFunction
)(
void
(
*
)(
void
))
eager_api_run_backward
,
METH_VARARGS
|
METH_KEYWORDS
,
METH_VARARGS
|
METH_KEYWORDS
,
...
...
paddle/fluid/pybind/eager_method.cc
浏览文件 @
8c43c0fe
...
@@ -57,86 +57,6 @@ typedef SSIZE_T ssize_t;
...
@@ -57,86 +57,6 @@ typedef SSIZE_T ssize_t;
namespace
paddle
{
namespace
paddle
{
namespace
pybind
{
namespace
pybind
{
namespace
py
=
::
pybind11
;
class
PyTensorHook
:
public
egr
::
TensorHook
{
public:
explicit
PyTensorHook
(
PyObject
*
func
)
:
py_func_
(
func
)
{
Py_INCREF
(
py_func_
);
}
~
PyTensorHook
()
{
py
::
gil_scoped_acquire
gil
;
Py_DECREF
(
py_func_
);
}
paddle
::
experimental
::
Tensor
operator
()(
const
paddle
::
experimental
::
Tensor
&
var
)
override
{
py
::
gil_scoped_acquire
gil
;
VLOG
(
3
)
<<
"Call PyTensorHook for var "
<<
var
.
name
();
PyObject
*
res
=
nullptr
;
try
{
PyObject
*
p_tmp_var
=
ToPyObject
(
var
);
res
=
PyObject_CallFunctionObjArgs
(
py_func_
,
p_tmp_var
,
nullptr
);
Py_DECREF
(
p_tmp_var
);
}
catch
(
platform
::
EnforceNotMet
&
e
)
{
throw
std
::
move
(
e
);
}
catch
(
std
::
exception
&
e
)
{
PADDLE_THROW
(
platform
::
errors
::
Unavailable
(
"Hook function of Tensor raises an exception: %s."
,
e
.
what
()));
}
catch
(...)
{
PADDLE_THROW
(
platform
::
errors
::
Fatal
(
"Hook function of Tensor raises an unknown exception."
));
}
PADDLE_ENFORCE_NOT_NULL
(
res
,
platform
::
errors
::
Unavailable
(
"Hook function of Tensor return a nullptr."
));
if
(
res
==
Py_None
)
{
return
var
;
}
auto
res_tensor
=
reinterpret_cast
<
TensorObject
*>
(
res
)
->
tensor
;
Py_DECREF
(
res
);
return
res_tensor
;
}
private:
PyObject
*
py_func_
;
};
class
PyTensorVoidHook
:
public
egr
::
TensorVoidHook
{
public:
explicit
PyTensorVoidHook
(
PyObject
*
func
)
:
py_func_
(
func
)
{
Py_INCREF
(
py_func_
);
}
~
PyTensorVoidHook
()
{
py
::
gil_scoped_acquire
gil
;
Py_DECREF
(
py_func_
);
}
void
operator
()()
override
{
py
::
gil_scoped_acquire
gil
;
VLOG
(
3
)
<<
"Call PyTensorVoidHook"
;
try
{
PyObject_CallFunctionObjArgs
(
py_func_
,
nullptr
);
}
catch
(
platform
::
EnforceNotMet
&
e
)
{
throw
std
::
move
(
e
);
}
catch
(
std
::
exception
&
e
)
{
PADDLE_THROW
(
platform
::
errors
::
Unavailable
(
"Hook function of Tensor raises an exception: %s."
,
e
.
what
()));
}
catch
(...)
{
PADDLE_THROW
(
platform
::
errors
::
Fatal
(
"Hook function of Tensor raises an unknown exception."
));
}
}
private:
PyObject
*
py_func_
;
};
extern
void
InitTensorWithNumpyValue
(
TensorObject
*
self
,
extern
void
InitTensorWithNumpyValue
(
TensorObject
*
self
,
const
pybind11
::
object
&
array
,
const
pybind11
::
object
&
array
,
const
paddle
::
platform
::
Place
&
place
,
const
paddle
::
platform
::
Place
&
place
,
...
@@ -1363,7 +1283,7 @@ static PyObject* tensor_register_reduce_hook(TensorObject* self,
...
@@ -1363,7 +1283,7 @@ static PyObject* tensor_register_reduce_hook(TensorObject* self,
auto
accumulation_grad_node
=
auto
accumulation_grad_node
=
std
::
dynamic_pointer_cast
<
egr
::
GradNodeAccumulation
>
(
grad_node
);
std
::
dynamic_pointer_cast
<
egr
::
GradNodeAccumulation
>
(
grad_node
);
accumulation_grad_node
->
RegisterReduceHook
(
accumulation_grad_node
->
RegisterReduceHook
(
std
::
make_shared
<
Py
Tensor
VoidHook
>
(
hook_func
));
std
::
make_shared
<
PyVoidHook
>
(
hook_func
));
RETURN_PY_NONE
RETURN_PY_NONE
...
...
paddle/fluid/pybind/eager_utils.cc
浏览文件 @
8c43c0fe
...
@@ -18,6 +18,7 @@ limitations under the License. */
...
@@ -18,6 +18,7 @@ limitations under the License. */
#include "paddle/fluid/eager/api/all.h"
#include "paddle/fluid/eager/api/all.h"
#include "paddle/fluid/eager/autograd_meta.h"
#include "paddle/fluid/eager/autograd_meta.h"
#include "paddle/fluid/eager/hooks.h"
#include "paddle/fluid/framework/convert_utils.h"
#include "paddle/fluid/framework/convert_utils.h"
#include "paddle/fluid/framework/scope.h"
#include "paddle/fluid/framework/scope.h"
#include "paddle/fluid/framework/scope_guard.h"
#include "paddle/fluid/framework/scope_guard.h"
...
@@ -1427,5 +1428,54 @@ paddle::DataType CastPyArg2DataType(PyObject* obj,
...
@@ -1427,5 +1428,54 @@ paddle::DataType CastPyArg2DataType(PyObject* obj,
framework
::
proto
::
VarType
::
Type
type
=
CastPyArg2ProtoType
(
obj
,
arg_pos
);
framework
::
proto
::
VarType
::
Type
type
=
CastPyArg2ProtoType
(
obj
,
arg_pos
);
return
framework
::
TransToPhiDataType
(
type
);
return
framework
::
TransToPhiDataType
(
type
);
}
}
paddle
::
experimental
::
Tensor
PyTensorHook
::
operator
()(
const
paddle
::
experimental
::
Tensor
&
var
)
{
py
::
gil_scoped_acquire
gil
;
VLOG
(
3
)
<<
"Call PyTensorHook for var "
<<
var
.
name
();
PyObject
*
res
=
nullptr
;
try
{
PyObject
*
p_tmp_var
=
ToPyObject
(
var
);
res
=
PyObject_CallFunctionObjArgs
(
py_func_
,
p_tmp_var
,
nullptr
);
Py_DECREF
(
p_tmp_var
);
}
catch
(
platform
::
EnforceNotMet
&
e
)
{
throw
std
::
move
(
e
);
}
catch
(
std
::
exception
&
e
)
{
PADDLE_THROW
(
platform
::
errors
::
Unavailable
(
"Hook function of Tensor raises an exception: %s."
,
e
.
what
()));
}
catch
(...)
{
PADDLE_THROW
(
platform
::
errors
::
Fatal
(
"Hook function of Tensor raises an unknown exception."
));
}
PADDLE_ENFORCE_NOT_NULL
(
res
,
platform
::
errors
::
Unavailable
(
"Hook function of Tensor return a nullptr."
));
if
(
res
==
Py_None
)
{
return
var
;
}
auto
res_tensor
=
reinterpret_cast
<
TensorObject
*>
(
res
)
->
tensor
;
Py_DECREF
(
res
);
return
res_tensor
;
}
void
PyVoidHook
::
operator
()()
{
py
::
gil_scoped_acquire
gil
;
VLOG
(
3
)
<<
"Call PyVoidHook"
;
try
{
PyObject_CallFunctionObjArgs
(
py_func_
,
nullptr
);
}
catch
(
platform
::
EnforceNotMet
&
e
)
{
throw
std
::
move
(
e
);
}
catch
(
std
::
exception
&
e
)
{
PADDLE_THROW
(
platform
::
errors
::
Unavailable
(
"Hook function of Tensor raises an exception: %s."
,
e
.
what
()));
}
catch
(...)
{
PADDLE_THROW
(
platform
::
errors
::
Fatal
(
"Hook function of Tensor raises an unknown exception."
));
}
}
}
// namespace pybind
}
// namespace pybind
}
// namespace paddle
}
// namespace paddle
paddle/fluid/pybind/eager_utils.h
浏览文件 @
8c43c0fe
...
@@ -17,6 +17,7 @@ typedef SSIZE_T ssize_t;
...
@@ -17,6 +17,7 @@ typedef SSIZE_T ssize_t;
#include <Python.h>
#include <Python.h>
#include "paddle/fluid/eager/hooks.h"
#include "paddle/fluid/framework/lod_tensor.h"
#include "paddle/fluid/framework/lod_tensor.h"
#include "paddle/fluid/framework/tensor.h"
#include "paddle/fluid/framework/tensor.h"
#include "paddle/fluid/jit/base_function.h"
#include "paddle/fluid/jit/base_function.h"
...
@@ -36,6 +37,7 @@ class Scope;
...
@@ -36,6 +37,7 @@ class Scope;
}
}
namespace
pybind
{
namespace
pybind
{
namespace
py
=
::
pybind11
;
#define RETURN_PY_NONE \
#define RETURN_PY_NONE \
Py_INCREF(Py_None); \
Py_INCREF(Py_None); \
return Py_None;
return Py_None;
...
@@ -110,6 +112,39 @@ PyObject* ToPyObject(
...
@@ -110,6 +112,39 @@ PyObject* ToPyObject(
const
std
::
unordered_map
<
std
::
string
,
std
::
vector
<
std
::
string
>>&
value
);
const
std
::
unordered_map
<
std
::
string
,
std
::
vector
<
std
::
string
>>&
value
);
PyObject
*
ToPyObject
(
const
std
::
unordered_map
<
std
::
wstring
,
int
>&
value
);
PyObject
*
ToPyObject
(
const
std
::
unordered_map
<
std
::
wstring
,
int
>&
value
);
class
PyTensorHook
:
public
egr
::
TensorHook
{
public:
explicit
PyTensorHook
(
PyObject
*
func
)
:
py_func_
(
func
)
{
Py_INCREF
(
py_func_
);
}
~
PyTensorHook
()
{
py
::
gil_scoped_acquire
gil
;
Py_DECREF
(
py_func_
);
}
paddle
::
experimental
::
Tensor
operator
()(
const
paddle
::
experimental
::
Tensor
&
var
)
override
;
private:
PyObject
*
py_func_
;
};
class
PyVoidHook
:
public
egr
::
VoidHook
{
public:
explicit
PyVoidHook
(
PyObject
*
func
)
:
py_func_
(
func
)
{
Py_INCREF
(
py_func_
);
}
~
PyVoidHook
()
{
py
::
gil_scoped_acquire
gil
;
Py_DECREF
(
py_func_
);
}
void
operator
()()
override
;
private:
PyObject
*
py_func_
;
};
template
<
typename
Tuple
,
size_t
N
>
template
<
typename
Tuple
,
size_t
N
>
struct
TupleTensorResult
{
struct
TupleTensorResult
{
static
void
Run
(
const
Tuple
&
out
,
PyObject
*
result
)
{
static
void
Run
(
const
Tuple
&
out
,
PyObject
*
result
)
{
...
...
python/paddle/fluid/tests/unittests/test_tensor_register_hook.py
浏览文件 @
8c43c0fe
...
@@ -639,5 +639,35 @@ class TestTensorRegisterBackwardHook(unittest.TestCase):
...
@@ -639,5 +639,35 @@ class TestTensorRegisterBackwardHook(unittest.TestCase):
self
.
func_register_backward_hook_for_var_without_gradient
()
self
.
func_register_backward_hook_for_var_without_gradient
()
class
TestRegsiterBackwardFinalHook
(
unittest
.
TestCase
):
def
setUp
(
self
):
self
.
devices
=
[
"cpu"
]
if
paddle
.
is_compiled_with_cuda
():
self
.
devices
.
append
(
"gpu"
)
def
test_register_backward_hook
(
self
):
global
HOOK_INIT_VALUE
global
HOOK_IS_CALLED
for
device
in
self
.
devices
:
np_x
=
np
.
random
.
rand
(
4
,
16
).
astype
(
"float32"
)
np_y
=
np
.
random
.
rand
(
16
,
20
).
astype
(
"float32"
)
x
=
paddle
.
to_tensor
(
np_x
,
stop_gradient
=
False
)
y
=
paddle
.
to_tensor
(
np_y
,
stop_gradient
=
False
)
core
.
eager
.
_add_backward_final_hook
(
global_void_hook
)
out
=
paddle
.
matmul
(
x
,
y
)
out
=
paddle
.
sum
(
out
)
out
.
backward
()
self
.
assertEqual
(
HOOK_INIT_VALUE
,
20
)
self
.
assertTrue
(
HOOK_IS_CALLED
)
# reset initial value
HOOK_INIT_VALUE
=
10
HOOK_IS_CALLED
=
False
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
unittest
.
main
()
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录