Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
7c577754
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
7c577754
编写于
12月 04, 2022
作者:
W
Weilong Wu
提交者:
GitHub
12月 04, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
clear test_dropout_op (#48657)
上级
46371c53
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
7 addition
and
70 deletion
+7
-70
python/paddle/fluid/tests/unittests/test_dropout_op.py
python/paddle/fluid/tests/unittests/test_dropout_op.py
+7
-70
未找到文件。
python/paddle/fluid/tests/unittests/test_dropout_op.py
浏览文件 @
7c577754
...
...
@@ -23,7 +23,6 @@ import paddle.fluid.core as core
import
paddle.static
as
static
from
paddle
import
_C_ops
from
paddle.fluid
import
Program
,
program_guard
from
paddle.fluid.framework
import
_enable_legacy_dygraph
,
_test_eager_guard
class
TestDropoutOp
(
OpTest
):
...
...
@@ -1046,13 +1045,14 @@ class TestDropoutBackward(unittest.TestCase):
return
mask
.
astype
(
"float32"
)
def
test_backward_downscale_in_infer
(
self
):
_enable_legacy_dygraph
()
for
place
in
self
.
places
:
with
fluid
.
dygraph
.
guard
(
place
):
input
=
paddle
.
uniform
([
40
,
40
],
dtype
=
"float32"
)
input
.
stop_gradient
=
False
out
,
mask
=
core
.
ops
.
dropout
(
input
,
'dropout_prob'
,
0.5
)
out
,
mask
=
_C_ops
.
dropout
(
input
,
None
,
0.5
,
False
,
"downgrade_in_infer"
,
0
,
False
)
out
.
backward
()
np
.
testing
.
assert_array_equal
(
...
...
@@ -1060,35 +1060,15 @@ class TestDropoutBackward(unittest.TestCase):
self
.
cal_grad_downscale_in_infer
(
mask
.
numpy
()),
)
def
test_backward_downscale_in_infer_eager
(
self
):
for
place
in
self
.
places
:
with
fluid
.
dygraph
.
guard
(
place
):
with
_test_eager_guard
():
input
=
paddle
.
uniform
([
40
,
40
],
dtype
=
"float32"
)
input
.
stop_gradient
=
False
out
,
mask
=
_C_ops
.
dropout
(
input
,
None
,
0.5
,
False
,
"downgrade_in_infer"
,
0
,
False
)
out
.
backward
()
np
.
testing
.
assert_array_equal
(
input
.
gradient
(),
self
.
cal_grad_downscale_in_infer
(
mask
.
numpy
()),
)
def
test_backward_upscale_train
(
self
):
_enable_legacy_dygraph
()
for
place
in
self
.
places
:
with
fluid
.
dygraph
.
guard
(
place
):
prob
=
0.5
input
=
paddle
.
uniform
([
40
,
40
],
dtype
=
"float32"
)
input
.
stop_gradient
=
False
out
,
mask
=
core
.
ops
.
dropout
(
input
,
'dropout_prob'
,
prob
,
"dropout_implementation"
,
"upscale_in_train"
,
out
,
mask
=
_C_ops
.
dropout
(
input
,
None
,
0.5
,
False
,
"upscale_in_train"
,
0
,
False
)
out
.
backward
()
...
...
@@ -1098,38 +1078,15 @@ class TestDropoutBackward(unittest.TestCase):
rtol
=
1e-05
,
)
def
test_backward_upscale_train_eager
(
self
):
for
place
in
self
.
places
:
with
fluid
.
dygraph
.
guard
(
place
):
with
_test_eager_guard
():
prob
=
0.5
input
=
paddle
.
uniform
([
40
,
40
],
dtype
=
"float32"
)
input
.
stop_gradient
=
False
out
,
mask
=
_C_ops
.
dropout
(
input
,
None
,
0.5
,
False
,
"upscale_in_train"
,
0
,
False
)
out
.
backward
()
np
.
testing
.
assert_allclose
(
input
.
gradient
(),
self
.
cal_grad_upscale_train
(
mask
.
numpy
(),
prob
),
rtol
=
1e-05
,
)
def
test_backward_upscale_train_2
(
self
):
_enable_legacy_dygraph
()
for
place
in
self
.
places
:
with
fluid
.
dygraph
.
guard
(
place
):
prob
=
0.3
input
=
paddle
.
uniform
([
40
,
40
],
dtype
=
"float32"
)
input
.
stop_gradient
=
False
out
,
mask
=
core
.
ops
.
dropout
(
input
,
'dropout_prob'
,
prob
,
"dropout_implementation"
,
"upscale_in_train"
,
out
,
mask
=
_C_ops
.
dropout
(
input
,
None
,
0.3
,
False
,
"upscale_in_train"
,
0
,
False
)
out
.
backward
()
...
...
@@ -1139,26 +1096,6 @@ class TestDropoutBackward(unittest.TestCase):
rtol
=
1e-05
,
)
def
test_backward_upscale_train_2_eager
(
self
):
for
place
in
self
.
places
:
with
fluid
.
dygraph
.
guard
(
place
):
with
_test_eager_guard
():
prob
=
0.3
input
=
paddle
.
uniform
([
40
,
40
],
dtype
=
"float32"
)
input
.
stop_gradient
=
False
out
,
mask
=
_C_ops
.
dropout
(
input
,
None
,
0.3
,
False
,
"upscale_in_train"
,
0
,
False
)
out
.
backward
()
np
.
testing
.
assert_allclose
(
input
.
gradient
(),
self
.
cal_grad_upscale_train
(
mask
.
numpy
(),
prob
),
rtol
=
1e-05
,
)
class
TestDropOutWithProbTensor
(
unittest
.
TestCase
):
def
setUp
(
self
):
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录