Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
067107ad
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
067107ad
编写于
7月 28, 2022
作者:
z8hanghuan
提交者:
GitHub
7月 28, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
support log_grad op, *test=kunlun (#44662)
上级
acde295c
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
76 addition
and
6 deletion
+76
-6
paddle/fluid/operators/activation_op_xpu.cc
paddle/fluid/operators/activation_op_xpu.cc
+49
-2
paddle/fluid/platform/device/xpu/xpu2_op_list.h
paddle/fluid/platform/device/xpu/xpu2_op_list.h
+1
-0
python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py
...addle/fluid/tests/unittests/xpu/test_activation_op_xpu.py
+26
-4
未找到文件。
paddle/fluid/operators/activation_op_xpu.cc
浏览文件 @
067107ad
/* Copyright (c) 202
0
PaddlePaddle Authors. All Rights Reserved.
/* Copyright (c) 202
2
PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
you may not use this file except in compliance with the License.
...
@@ -358,6 +358,52 @@ struct XPULeakyReluGradFunctor : public BaseActivationFunctor<T> {
...
@@ -358,6 +358,52 @@ struct XPULeakyReluGradFunctor : public BaseActivationFunctor<T> {
}
}
};
};
template
<
typename
T
>
struct
XPULogGradFunctor
:
public
BaseActivationFunctor
<
T
>
{
using
XPUType
=
typename
XPUTypeTrait
<
T
>::
Type
;
void
operator
()(
const
framework
::
ExecutionContext
&
ctx
)
const
{
const
auto
*
x
=
ctx
.
Input
<
Tensor
>
(
"X"
);
auto
*
dOut
=
ctx
.
Input
<
framework
::
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
auto
*
dX
=
ctx
.
Output
<
framework
::
Tensor
>
(
framework
::
GradVarName
(
"X"
));
const
T
*
x_data
=
nullptr
;
const
T
*
y_grad
=
nullptr
;
if
(
x
!=
nullptr
)
x_data
=
x
->
data
<
T
>
();
if
(
dOut
!=
nullptr
)
y_grad
=
dOut
->
data
<
T
>
();
T
*
x_grad
=
dX
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
auto
dev_ctx
=
ctx
.
device_context
<
paddle
::
platform
::
XPUDeviceContext
>
().
x_context
();
const
auto
x_dims
=
x
->
dims
();
auto
xshape
=
phi
::
vectorize
<
int
>
(
x_dims
);
int
len
=
x
->
dims
()[
x_dims
.
size
()
-
1
];
std
::
vector
<
int
>
yshape
(
1
,
len
);
xpu
::
ctx_guard
RAII_GUARD
(
dev_ctx
);
T
*
y_data
=
RAII_GUARD
.
alloc_l3_or_gm
<
T
>
(
len
);
PADDLE_ENFORCE_XDNN_NOT_NULL
(
y_data
);
T
*
tmp_grad
=
RAII_GUARD
.
alloc_l3_or_gm
<
T
>
(
x
->
numel
());
PADDLE_ENFORCE_XDNN_NOT_NULL
(
tmp_grad
);
int
r
=
xpu
::
constant
<
T
>
(
dev_ctx
,
y_data
,
len
,
static_cast
<
T
>
(
1.0
));
PADDLE_ENFORCE_XDNN_SUCCESS
(
r
,
"constant"
);
// dx.device(d) = dout * (static_cast<T>(1) / x);
r
=
xpu
::
broadcast_div
(
dev_ctx
,
reinterpret_cast
<
const
float
*>
(
y_data
),
reinterpret_cast
<
const
float
*>
(
x_data
),
reinterpret_cast
<
float
*>
(
tmp_grad
),
yshape
,
xshape
);
PADDLE_ENFORCE_XDNN_SUCCESS
(
r
,
"broadcast_div"
);
r
=
xpu
::
broadcast_mul
(
dev_ctx
,
reinterpret_cast
<
const
float
*>
(
y_grad
),
reinterpret_cast
<
const
float
*>
(
tmp_grad
),
reinterpret_cast
<
float
*>
(
x_grad
),
xshape
,
xshape
);
PADDLE_ENFORCE_XDNN_SUCCESS
(
r
,
"broadcast_mul"
);
}
};
template
<
typename
T
>
template
<
typename
T
>
struct
XPUPowFunctor
:
public
BaseActivationFunctor
<
T
>
{
struct
XPUPowFunctor
:
public
BaseActivationFunctor
<
T
>
{
void
operator
()(
const
framework
::
ExecutionContext
&
ctx
)
const
{
void
operator
()(
const
framework
::
ExecutionContext
&
ctx
)
const
{
...
@@ -584,5 +630,6 @@ REGISTER_OP_XPU_KERNEL(exp,
...
@@ -584,5 +630,6 @@ REGISTER_OP_XPU_KERNEL(exp,
ops
::
XPUActivationKernel
<
ops
::
XPUExpFunctor
<
float
>>
);
ops
::
XPUActivationKernel
<
ops
::
XPUExpFunctor
<
float
>>
);
REGISTER_OP_XPU_KERNEL
(
log
,
REGISTER_OP_XPU_KERNEL
(
log
,
ops
::
XPUActivationKernel
<
ops
::
XPULogFunctor
<
float
>>
);
ops
::
XPUActivationKernel
<
ops
::
XPULogFunctor
<
float
>>
);
REGISTER_OP_XPU_KERNEL
(
log_grad
,
ops
::
XPUActivationGradKernel
<
ops
::
XPULogGradFunctor
<
float
>>
);
#endif // PADDLE_WITH_XPU
#endif // PADDLE_WITH_XPU
paddle/fluid/platform/device/xpu/xpu2_op_list.h
浏览文件 @
067107ad
...
@@ -276,6 +276,7 @@ XPUOpMap& get_kl2_ops() {
...
@@ -276,6 +276,7 @@ XPUOpMap& get_kl2_ops() {
pOpKernelType
(
vartype
::
INT32
,
XPUPlace
()),
pOpKernelType
(
vartype
::
INT32
,
XPUPlace
()),
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
{
"log"
,
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
{
"log"
,
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
{
"log_grad"
,
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
{
"log_softmax"
,
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
{
"log_softmax"
,
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
{
"log_softmax_grad"
,
{
"log_softmax_grad"
,
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
XPUKernelSet
({
pOpKernelType
(
vartype
::
FP32
,
XPUPlace
())})},
...
...
python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py
浏览文件 @
067107ad
...
@@ -34,12 +34,15 @@ class TestActivationOPBase(XPUOpTest):
...
@@ -34,12 +34,15 @@ class TestActivationOPBase(XPUOpTest):
def
setUp
(
self
):
def
setUp
(
self
):
self
.
place
=
paddle
.
XPUPlace
(
0
)
self
.
place
=
paddle
.
XPUPlace
(
0
)
self
.
init_dtype
()
self
.
init_dtype
()
self
.
set_shape
()
self
.
set_case
()
self
.
set_case
()
def
set_shape
(
self
):
self
.
shape
=
[
11
,
17
]
def
set_case
(
self
):
def
set_case
(
self
):
self
.
op_type
=
'exp'
self
.
op_type
=
'exp'
x
=
np
.
random
.
uniform
(
-
1
,
1
,
self
.
shape
).
astype
(
self
.
dtype
)
x
=
np
.
random
.
uniform
(
-
1
,
1
,
[
11
,
17
]).
astype
(
self
.
dtype
)
out
=
np
.
exp
(
x
)
out
=
np
.
exp
(
x
)
self
.
attrs
=
{
'use_xpu'
:
True
}
self
.
attrs
=
{
'use_xpu'
:
True
}
self
.
inputs
=
{
'X'
:
OpTest
.
np_dtype_to_fluid_dtype
(
x
)}
self
.
inputs
=
{
'X'
:
OpTest
.
np_dtype_to_fluid_dtype
(
x
)}
...
@@ -313,14 +316,33 @@ class XPUTestLogOP(XPUOpTestWrapper):
...
@@ -313,14 +316,33 @@ class XPUTestLogOP(XPUOpTestWrapper):
def
set_case
(
self
):
def
set_case
(
self
):
self
.
op_type
=
"log"
self
.
op_type
=
"log"
self
.
dtype
=
self
.
in_type
self
.
dtype
=
self
.
in_type
x
=
np
.
random
.
uniform
(
0.1
,
1
,
self
.
shape
).
astype
(
self
.
dtype
)
x
=
np
.
random
.
uniform
(
0.1
,
1
,
[
11
,
17
]).
astype
(
self
.
dtype
)
out
=
np
.
log
(
x
)
out
=
np
.
log
(
x
)
self
.
attrs
=
{
'use_xpu'
:
True
}
self
.
attrs
=
{
'use_xpu'
:
True
}
self
.
inputs
=
{
'X'
:
OpTest
.
np_dtype_to_fluid_dtype
(
x
)}
self
.
inputs
=
{
'X'
:
OpTest
.
np_dtype_to_fluid_dtype
(
x
)}
self
.
outputs
=
{
'Out'
:
out
}
self
.
outputs
=
{
'Out'
:
out
}
class
TestLogCase1
(
XPUTestLog
):
def
set_shape
(
self
):
self
.
shape
=
[
1
,
11
,
17
]
class
TestLogCase2
(
XPUTestLog
):
def
set_shape
(
self
):
self
.
shape
=
[
2
,
2
,
2
]
class
TestLogCase3
(
XPUTestLog
):
def
set_shape
(
self
):
self
.
shape
=
[
2
]
class
TestLogCase4
(
XPUTestLog
):
def
set_shape
(
self
):
self
.
shape
=
[
1
,
2
,
3
,
4
]
support_types
=
get_xpu_op_support_types
(
'log'
)
support_types
=
get_xpu_op_support_types
(
'log'
)
for
stype
in
support_types
:
for
stype
in
support_types
:
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录