Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
f3296eae
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
f3296eae
编写于
4月 11, 2022
作者:
X
xiongkun
提交者:
GitHub
4月 11, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Yaml] add unittest for prelu, gelu. (#41444) (#41620)
* add gelu pythonapi and unittest * fix prelu
上级
664199aa
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
27 addition
and
6 deletion
+27
-6
python/paddle/fluid/tests/unittests/test_gelu_op.py
python/paddle/fluid/tests/unittests/test_gelu_op.py
+5
-0
python/paddle/fluid/tests/unittests/test_prelu_op.py
python/paddle/fluid/tests/unittests/test_prelu_op.py
+18
-5
python/paddle/nn/functional/activation.py
python/paddle/nn/functional/activation.py
+4
-1
未找到文件。
python/paddle/fluid/tests/unittests/test_gelu_op.py
浏览文件 @
f3296eae
...
...
@@ -21,6 +21,7 @@ import paddle.fluid as fluid
import
paddle.fluid.dygraph
as
dg
import
paddle
import
paddle.nn.functional
as
F
from
paddle.fluid.framework
import
_test_eager_guard
def
gelu
(
x
,
approximate
):
...
...
@@ -91,6 +92,10 @@ class TestGeluOp(unittest.TestCase):
np
.
allclose
(
x_g_ref
,
x_g_fast_math
,
rtol
=
1e-5
,
atol
=
5e-4
))
def
test_fast_math_eager
(
self
):
with
_test_eager_guard
():
self
.
test_fast_math
()
if
__name__
==
'__main__'
:
unittest
.
main
()
python/paddle/fluid/tests/unittests/test_prelu_op.py
浏览文件 @
f3296eae
...
...
@@ -23,6 +23,7 @@ from paddle.fluid import Program, program_guard
from
op_test
import
OpTest
,
skip_check_grad_ci
import
paddle
import
paddle.nn.functional
as
F
from
paddle.fluid.framework
import
_test_eager_guard
def
ref_prelu
(
x
,
weight
):
...
...
@@ -76,6 +77,10 @@ class TestFunctionalPReluAPI(unittest.TestCase):
self
.
dygraph_check
(
self
.
weight_np_0
)
self
.
dygraph_check
(
self
.
weight_np_1
)
def
test_dygraph_api_eager
(
self
):
with
_test_eager_guard
():
self
.
test_dygraph_api
()
def
test_error
(
self
):
with
paddle
.
static
.
program_guard
(
paddle
.
static
.
Program
()):
weight_fp32
=
paddle
.
fluid
.
data
(
...
...
@@ -151,13 +156,19 @@ class TestNNPReluAPI(unittest.TestCase):
paddle
.
enable_static
()
def
prelu_api_wrapper
(
x
,
weight
,
data_format
=
"NCHW"
):
weight
=
weight
.
reshape
([
-
1
])
return
paddle
.
nn
.
functional
.
prelu
(
x
,
weight
,
data_format
,
name
=
None
)
class
PReluTest
(
OpTest
):
def
setUp
(
self
):
self
.
init_dtype
()
self
.
init_input_shape
()
self
.
eager_mode
=
True
self
.
init_attr
()
self
.
op_type
=
"prelu"
self
.
python_api
=
p
addle
.
nn
.
functional
.
prelu
self
.
python_api
=
p
relu_api_wrapper
x_np
=
np
.
random
.
uniform
(
-
1
,
1
,
self
.
x_shape
).
astype
(
self
.
dtype
)
# Since zero point in prelu is not differentiable, avoid randomize
...
...
@@ -178,6 +189,8 @@ class PReluTest(OpTest):
alpha_np
=
np
.
random
.
uniform
(
-
1
,
-
0.5
,
[
1
,
1
,
1
,
self
.
x_shape
[
-
1
]])
else
:
alpha_np
=
np
.
random
.
uniform
(
-
1
,
-
0.5
,
[
1
]
+
self
.
x_shape
[
1
:])
# eager check don't support mode = 'all'
self
.
eager_mode
=
False
alpha_np
=
alpha_np
.
astype
(
self
.
dtype
)
self
.
inputs
=
{
'X'
:
x_np
,
'Alpha'
:
alpha_np
}
...
...
@@ -208,10 +221,10 @@ class PReluTest(OpTest):
self
.
attrs
=
{
'mode'
:
"channel"
,
"data_format"
:
"NCHW"
}
def
test_check_output
(
self
):
self
.
check_output
(
check_eager
=
Fals
e
)
self
.
check_output
(
check_eager
=
self
.
eager_mod
e
)
def
test_check_grad
(
self
):
self
.
check_grad
([
'X'
,
'Alpha'
],
'Out'
,
check_eager
=
Fals
e
)
self
.
check_grad
([
'X'
,
'Alpha'
],
'Out'
,
check_eager
=
self
.
eager_mod
e
)
@
skip_check_grad_ci
(
...
...
@@ -375,7 +388,7 @@ def create_test_fp16_class(parent,
place
=
core
.
CUDAPlace
(
0
)
if
core
.
is_float16_supported
(
place
):
self
.
check_output_with_place
(
place
,
atol
=
atol
,
check_eager
=
Fals
e
)
place
,
atol
=
atol
,
check_eager
=
self
.
eager_mod
e
)
def
test_check_grad
(
self
):
place
=
core
.
CUDAPlace
(
0
)
...
...
@@ -384,7 +397,7 @@ def create_test_fp16_class(parent,
place
,
[
'X'
,
'Alpha'
],
'Out'
,
max_relative_error
=
max_relative_error
,
check_eager
=
Fals
e
)
check_eager
=
self
.
eager_mod
e
)
cls_name
=
"{0}_{1}"
.
format
(
parent
.
__name__
,
"Fp16Op"
)
TestPReluFp16Case
.
__name__
=
cls_name
...
...
python/paddle/nn/functional/activation.py
浏览文件 @
f3296eae
...
...
@@ -175,7 +175,10 @@ def gelu(x, approximate=False, name=None):
# [ 0.84119201, 1.39957154]]
"""
if
in_dynamic_mode
():
if
in_dygraph_mode
():
return
_C_ops
.
final_state_gelu
(
x
,
approximate
)
if
_in_legacy_dygraph
():
return
_C_ops
.
gelu
(
x
,
'approximate'
,
approximate
)
check_variable_and_dtype
(
x
,
'x'
,
[
'float16'
,
'float32'
,
'float64'
],
'gelu'
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录