Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
cbd15f7d
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
cbd15f7d
编写于
10月 18, 2021
作者:
Q
Qi Li
提交者:
GitHub
10月 18, 2021
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[NPU] add kernels for elementwise_add gather_nd tile, test=develop (#36464)
上级
8757fc5b
变更
6
隐藏空白更改
内联
并排
Showing
6 changed file
with
80 addition
and
48 deletion
+80
-48
paddle/fluid/operators/elementwise/elementwise_add_op_npu.cc
paddle/fluid/operators/elementwise/elementwise_add_op_npu.cc
+3
-0
paddle/fluid/operators/gather_nd_op_npu.cc
paddle/fluid/operators/gather_nd_op_npu.cc
+17
-19
paddle/fluid/operators/tile_op_npu.cc
paddle/fluid/operators/tile_op_npu.cc
+23
-15
python/paddle/fluid/tests/unittests/npu/test_elementwise_add_op_npu.py
.../fluid/tests/unittests/npu/test_elementwise_add_op_npu.py
+10
-5
python/paddle/fluid/tests/unittests/npu/test_gather_nd_op_npu.py
...paddle/fluid/tests/unittests/npu/test_gather_nd_op_npu.py
+8
-8
python/paddle/fluid/tests/unittests/npu/test_tile_op_npu.py
python/paddle/fluid/tests/unittests/npu/test_tile_op_npu.py
+19
-1
未找到文件。
paddle/fluid/operators/elementwise/elementwise_add_op_npu.cc
浏览文件 @
cbd15f7d
...
...
@@ -146,6 +146,9 @@ namespace ops = paddle::operators;
namespace
plat
=
paddle
::
platform
;
REGISTER_OP_NPU_KERNEL
(
elementwise_add
,
ops
::
ElementwiseAddNPUKernel
<
float
>
,
#ifdef PADDLE_WITH_ASCEND_INT64
ops
::
ElementwiseAddNPUKernel
<
int64_t
>
,
#endif
ops
::
ElementwiseAddNPUKernel
<
plat
::
float16
>
);
REGISTER_OP_NPU_KERNEL
(
elementwise_add_grad
,
...
...
paddle/fluid/operators/gather_nd_op_npu.cc
浏览文件 @
cbd15f7d
...
...
@@ -18,7 +18,10 @@ limitations under the License. */
namespace
paddle
{
namespace
operators
{
template
<
typename
DeviceContext
,
typename
T
>
using
Tensor
=
framework
::
Tensor
;
using
NPUDeviceContext
=
platform
::
NPUDeviceContext
;
template
<
typename
T
>
class
GatherNdNPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
...
...
@@ -49,14 +52,12 @@ class GatherNdNPUKernel : public framework::OpKernel<T> {
framework
::
proto
::
VarType
::
INT64
)));
const
auto
&
runner
=
NpuOpRunner
(
"GatherNd"
,
{
*
x
,
*
index
},
{
*
out
},
{});
auto
stream
=
ctx
.
template
device_context
<
paddle
::
platform
::
NPUDeviceContext
>()
.
stream
();
auto
stream
=
ctx
.
template
device_context
<
NPUDeviceContext
>().
stream
();
runner
.
Run
(
stream
);
}
};
template
<
typename
DeviceContext
,
typename
T
>
template
<
typename
T
>
class
GatherNdGradNPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
...
...
@@ -91,10 +92,7 @@ class GatherNdGradNPUKernel : public framework::OpKernel<T> {
dout
=
&
tmp_tensor2
;
}
auto
stream
=
ctx
.
template
device_context
<
paddle
::
platform
::
NPUDeviceContext
>()
.
stream
();
auto
stream
=
ctx
.
template
device_context
<
NPUDeviceContext
>().
stream
();
platform
::
NPUMemsetAsync
(
static_cast
<
void
*>
(
p
),
0
,
dx
->
numel
()
*
sizeof
(
T
),
stream
);
...
...
@@ -108,13 +106,13 @@ class GatherNdGradNPUKernel : public framework::OpKernel<T> {
}
// namespace paddle
namespace
ops
=
paddle
::
operators
;
REGISTER_OP_NPU_KERNEL
(
gather_nd
,
ops
::
GatherNdNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
paddle
::
platform
::
float16
>
,
ops
::
GatherNdNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
float
>
);
REGISTER_OP_NPU_KERNEL
(
gather_nd_grad
,
ops
::
GatherNdGradNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
paddle
::
platform
::
float16
>
,
ops
::
GatherNdGradNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
float
>
);
REGISTER_OP_NPU_KERNEL
(
gather_nd
,
ops
::
GatherNdNPUKernel
<
paddle
::
platform
::
float16
>
,
#ifdef PADDLE_WITH_ASCEND_INT64
ops
::
GatherNdNPUKernel
<
int64_t
>
,
#endif
ops
::
GatherNdNPUKernel
<
float
>
);
REGISTER_OP_NPU_KERNEL
(
gather_nd_grad
,
ops
::
GatherNdGradNPUKernel
<
paddle
::
platform
::
float16
>
,
ops
::
GatherNdGradNPUKernel
<
float
>
);
paddle/fluid/operators/tile_op_npu.cc
浏览文件 @
cbd15f7d
...
...
@@ -16,7 +16,11 @@ limitations under the License. */
namespace
paddle
{
namespace
operators
{
template
<
typename
DeviceContext
,
typename
T
>
using
Tensor
=
framework
::
Tensor
;
using
NPUDeviceContext
=
platform
::
NPUDeviceContext
;
template
<
typename
T
>
class
TileNPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
...
...
@@ -92,18 +96,21 @@ class TileNPUKernel : public framework::OpKernel<T> {
std
::
vector
<
int
>
temp
(
repeat_times
.
size
(),
1
);
if
(
repeat_times
==
temp
)
{
framework
::
TensorCopy
(
*
in0
,
context
.
GetPlace
(),
context
.
template
device_context
<
platform
::
DeviceContext
>(),
out0
);
framework
::
TensorCopy
(
*
in0
,
context
.
GetPlace
(),
context
.
template
device_context
<
NPUDeviceContext
>
(),
out0
);
return
;
}
const
auto
&
runner
=
NpuOpRunner
(
"TileD"
,
{
*
in0
},
{
*
out0
},
{{
"multiples"
,
repeat_times
}});
auto
stream
=
context
.
template
device_context
<
paddle
::
platform
::
NPUDeviceContext
>()
.
stream
();
runner
.
Run
(
stream
);
// const auto& runner =
// NpuOpRunner("TileD", {*in0}, {*out0}, {{"multiples", repeat_times}});
auto
stream
=
context
.
template
device_context
<
NPUDeviceContext
>().
stream
();
NpuOpRunner
runner
;
runner
.
SetType
(
"Tile"
)
.
AddInput
(
*
in0
)
.
AddInput
(
std
::
move
(
repeat_times
))
.
AddOutput
(
*
out0
)
.
Run
(
stream
);
}
};
...
...
@@ -111,8 +118,9 @@ class TileNPUKernel : public framework::OpKernel<T> {
}
// namespace paddle
namespace
ops
=
paddle
::
operators
;
REGISTER_OP_NPU_KERNEL
(
tile
,
ops
::
TileNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
float
>
,
ops
::
TileNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
int
>
,
ops
::
TileNPUKernel
<
paddle
::
platform
::
NPUDeviceContext
,
paddle
::
platform
::
float16
>
);
REGISTER_OP_NPU_KERNEL
(
tile
,
ops
::
TileNPUKernel
<
float
>
,
ops
::
TileNPUKernel
<
int
>
,
#ifdef PADDLE_WITH_ASCEND_INT64
ops
::
TileNPUKernel
<
int64_t
>
,
#endif
ops
::
TileNPUKernel
<
bool
>
,
ops
::
TileNPUKernel
<
paddle
::
platform
::
float16
>
);
python/paddle/fluid/tests/unittests/npu/test_elementwise_add_op_npu.py
浏览文件 @
cbd15f7d
...
...
@@ -65,7 +65,7 @@ class TestElementwiseAddOp(OpTest):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad_normal
(
self
):
if
self
.
dtype
==
np
.
float16
:
if
self
.
dtype
==
np
.
float16
or
self
.
dtype
==
np
.
int64
:
return
self
.
check_grad_with_place
(
...
...
@@ -75,7 +75,7 @@ class TestElementwiseAddOp(OpTest):
max_relative_error
=
0.006
,
)
def
test_check_grad_ingore_x
(
self
):
if
self
.
dtype
==
np
.
float16
:
if
self
.
dtype
==
np
.
float16
or
self
.
dtype
==
np
.
int64
:
return
self
.
check_grad_with_place
(
...
...
@@ -86,7 +86,7 @@ class TestElementwiseAddOp(OpTest):
max_relative_error
=
0.006
,
)
def
test_check_grad_ingore_y
(
self
):
if
self
.
dtype
==
np
.
float16
:
if
self
.
dtype
==
np
.
float16
or
self
.
dtype
==
np
.
int64
:
return
self
.
check_grad_with_place
(
...
...
@@ -102,6 +102,11 @@ class TestFP16ElementwiseAddOp(TestElementwiseAddOp):
self
.
dtype
=
np
.
float16
class
TestINT64ElementwiseAddOp
(
TestElementwiseAddOp
):
def
init_dtype
(
self
):
self
.
dtype
=
np
.
int64
@
skip_check_grad_ci
(
reason
=
"[skip shape check] Use y_shape(1) to test broadcast."
)
class
TestElementwiseAddOp_scalar
(
TestElementwiseAddOp
):
...
...
@@ -507,8 +512,8 @@ class TestAddApi(unittest.TestCase):
def
test_dygraph
(
self
):
with
fluid
.
dygraph
.
guard
(
paddle
.
NPUPlace
(
0
)):
np_x
=
np
.
array
([
2
,
3
,
4
]).
astype
(
'float
64
'
)
np_y
=
np
.
array
([
1
,
5
,
2
]).
astype
(
'float
64
'
)
np_x
=
np
.
array
([
2
,
3
,
4
]).
astype
(
'float
32
'
)
np_y
=
np
.
array
([
1
,
5
,
2
]).
astype
(
'float
32
'
)
x
=
fluid
.
dygraph
.
to_variable
(
np_x
)
y
=
fluid
.
dygraph
.
to_variable
(
np_y
)
z
=
self
.
_executed_api
(
x
,
y
)
...
...
python/paddle/fluid/tests/unittests/npu/test_gather_nd_op_npu.py
浏览文件 @
cbd15f7d
...
...
@@ -61,7 +61,7 @@ def test_class1(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
self
.
place
,
[
'X'
],
'Out'
)
...
...
@@ -88,7 +88,7 @@ def test_class2(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
self
.
place
,
[
'X'
],
'Out'
)
...
...
@@ -120,7 +120,7 @@ def test_class3(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
...
...
@@ -153,7 +153,7 @@ def test_class4(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
self
.
place
,
[
'X'
],
'Out'
)
...
...
@@ -184,7 +184,7 @@ def test_class5(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
self
.
place
,
[
'X'
],
'Out'
)
...
...
@@ -217,7 +217,7 @@ def test_class6(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
self
.
place
,
[
'X'
],
'Out'
)
...
...
@@ -252,7 +252,7 @@ def test_class7(op_type, typename):
self
.
check_output_with_place
(
self
.
place
)
def
test_check_grad
(
self
):
if
typename
==
"float16"
:
if
typename
==
"float16"
or
typename
==
"int64"
:
self
.
__class__
.
no_need_check_grad
=
True
else
:
self
.
check_grad_with_place
(
self
.
place
,
[
'X'
],
'Out'
)
...
...
@@ -276,7 +276,7 @@ class TestGatherNdAPI(unittest.TestCase):
paddle
.
enable_static
()
for
_typename
in
{
'float16'
,
'float32'
}:
for
_typename
in
{
'float16'
,
'float32'
,
'int64'
}:
test_class1
(
'gather_nd'
,
_typename
)
test_class2
(
'gather_nd'
,
_typename
)
test_class3
(
'gather_nd'
,
_typename
)
...
...
python/paddle/fluid/tests/unittests/npu/test_tile_op_npu.py
浏览文件 @
cbd15f7d
...
...
@@ -206,7 +206,7 @@ class TestTileOpInt64_t(OpTest):
self
.
op_type
=
"tile"
self
.
inputs
=
{
'X'
:
np
.
random
.
randint
(
10
,
size
=
(
2
,
4
,
5
)).
astype
(
"int
32
"
)
10
,
size
=
(
2
,
4
,
5
)).
astype
(
"int
64
"
)
}
self
.
attrs
=
{
'repeat_times'
:
[
2
,
1
,
4
]}
output
=
np
.
tile
(
self
.
inputs
[
'X'
],
(
2
,
1
,
4
))
...
...
@@ -219,6 +219,24 @@ class TestTileOpInt64_t(OpTest):
self
.
check_output_with_place
(
self
.
place
)
# Situation 6: input x is Bool
class
TestTileOpBool
(
OpTest
):
def
setUp
(
self
):
self
.
set_npu
()
self
.
place
=
paddle
.
NPUPlace
(
0
)
self
.
op_type
=
"tile"
self
.
inputs
=
{
'X'
:
np
.
random
.
randint
(
1
,
size
=
(
2
,
4
,
5
)).
astype
(
"bool"
)}
self
.
attrs
=
{
'repeat_times'
:
[
2
,
1
,
4
]}
output
=
np
.
tile
(
self
.
inputs
[
'X'
],
(
2
,
1
,
4
))
self
.
outputs
=
{
'Out'
:
output
}
def
set_npu
(
self
):
self
.
__class__
.
use_npu
=
True
def
test_check_output
(
self
):
self
.
check_output_with_place
(
self
.
place
)
# Test python API
class
TestTileAPI
(
unittest
.
TestCase
):
def
test_api
(
self
):
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录