Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
4d0ca02b
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
4d0ca02b
编写于
6月 15, 2022
作者:
F
fwenguang
提交者:
GitHub
6月 15, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[MLU] add size kernel for mlu (#43450)
上级
79dc32b4
变更
2
显示空白变更内容
内联
并排
Showing
2 changed file
with
164 addition
and
0 deletion
+164
-0
paddle/fluid/operators/size_op_mlu.cc
paddle/fluid/operators/size_op_mlu.cc
+43
-0
python/paddle/fluid/tests/unittests/mlu/test_size_op_mlu.py
python/paddle/fluid/tests/unittests/mlu/test_size_op_mlu.py
+121
-0
未找到文件。
paddle/fluid/operators/size_op_mlu.cc
0 → 100644
浏览文件 @
4d0ca02b
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/operators/mlu/mlu_baseop.h"
namespace
paddle
{
namespace
operators
{
template
<
typename
T
>
class
SizeMLUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
x
=
ctx
.
Input
<
framework
::
Tensor
>
(
"Input"
);
auto
*
out
=
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
);
out
->
mutable_data
<
int64_t
>
(
ctx
.
GetPlace
());
int64_t
size
=
x
->
numel
();
FillMLUTensorWithHostValue
<
int64_t
>
(
ctx
,
size
,
out
);
}
};
}
// namespace operators
}
// namespace paddle
namespace
ops
=
paddle
::
operators
;
REGISTER_OP_MLU_KERNEL
(
size
,
ops
::
SizeMLUKernel
<
int
>
,
ops
::
SizeMLUKernel
<
int64_t
>
,
ops
::
SizeMLUKernel
<
paddle
::
platform
::
float16
>
,
ops
::
SizeMLUKernel
<
float
>
,
ops
::
SizeMLUKernel
<
double
>
,
ops
::
SizeMLUKernel
<
bool
>
);
python/paddle/fluid/tests/unittests/mlu/test_size_op_mlu.py
0 → 100644
浏览文件 @
4d0ca02b
# Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import
unittest
import
numpy
as
np
import
paddle
import
paddle.fluid
as
fluid
import
sys
sys
.
path
.
append
(
'..'
)
from
op_test
import
OpTest
paddle
.
enable_static
()
class
TestSizeOp
(
OpTest
):
def
setUp
(
self
):
self
.
op_type
=
"size"
self
.
shape
=
[]
self
.
config
()
input
=
np
.
zeros
(
self
.
shape
,
dtype
=
'bool'
)
self
.
inputs
=
{
'Input'
:
input
}
self
.
outputs
=
{
'Out'
:
np
.
array
([
np
.
size
(
input
)],
dtype
=
'int64'
)}
def
config
(
self
):
pass
def
test_check_output
(
self
):
self
.
check_output_with_place
(
paddle
.
device
.
MLUPlace
(
0
))
class
TestRank1Tensor
(
TestSizeOp
):
def
config
(
self
):
self
.
shape
=
[
2
]
class
TestRank2Tensor
(
TestSizeOp
):
def
config
(
self
):
self
.
shape
=
[
2
,
3
]
class
TestRank3Tensor
(
TestSizeOp
):
def
config
(
self
):
self
.
shape
=
[
2
,
3
,
100
]
class
TestLargeTensor
(
TestSizeOp
):
def
config
(
self
):
self
.
shape
=
[
2
**
10
]
class
TestSizeAPI
(
unittest
.
TestCase
):
def
test_size_static
(
self
):
main_program
=
fluid
.
Program
()
startup_program
=
fluid
.
Program
()
with
fluid
.
program_guard
(
main_program
,
startup_program
):
shape1
=
[
2
,
1
,
4
,
5
]
shape2
=
[
1
,
4
,
5
]
x_1
=
paddle
.
fluid
.
data
(
shape
=
shape1
,
dtype
=
'int32'
,
name
=
'x_1'
)
x_2
=
paddle
.
fluid
.
data
(
shape
=
shape2
,
dtype
=
'int32'
,
name
=
'x_2'
)
input_1
=
np
.
random
.
random
(
shape1
).
astype
(
"int32"
)
input_2
=
np
.
random
.
random
(
shape2
).
astype
(
"int32"
)
out_1
=
paddle
.
fluid
.
layers
.
size
(
x_1
)
out_2
=
paddle
.
fluid
.
layers
.
size
(
x_2
)
exe
=
paddle
.
static
.
Executor
(
place
=
paddle
.
MLUPlace
(
0
))
res_1
,
res_2
=
exe
.
run
(
feed
=
{
"x_1"
:
input_1
,
"x_2"
:
input_2
,
},
fetch_list
=
[
out_1
,
out_2
])
assert
(
np
.
array_equal
(
res_1
,
np
.
array
([
np
.
size
(
input_1
)
]).
astype
(
"int64"
)))
assert
(
np
.
array_equal
(
res_2
,
np
.
array
([
np
.
size
(
input_2
)
]).
astype
(
"int64"
)))
def
test_size_imperative
(
self
):
paddle
.
disable_static
(
paddle
.
MLUPlace
(
0
))
input_1
=
np
.
random
.
random
([
2
,
1
,
4
,
5
]).
astype
(
"int32"
)
input_2
=
np
.
random
.
random
([
1
,
4
,
5
]).
astype
(
"int32"
)
x_1
=
paddle
.
to_tensor
(
input_1
)
x_2
=
paddle
.
to_tensor
(
input_2
)
out_1
=
paddle
.
fluid
.
layers
.
size
(
x_1
)
out_2
=
paddle
.
fluid
.
layers
.
size
(
x_2
)
assert
(
np
.
array_equal
(
out_1
.
numpy
().
item
(
0
),
np
.
size
(
input_1
)))
assert
(
np
.
array_equal
(
out_2
.
numpy
().
item
(
0
),
np
.
size
(
input_2
)))
paddle
.
enable_static
()
def
test_error
(
self
):
main_program
=
fluid
.
Program
()
startup_program
=
fluid
.
Program
()
with
fluid
.
program_guard
(
main_program
,
startup_program
):
def
test_x_type
():
shape
=
[
1
,
4
,
5
]
input_1
=
np
.
random
.
random
(
shape
).
astype
(
"int32"
)
out_1
=
paddle
.
fluid
.
layers
.
size
(
input_1
)
self
.
assertRaises
(
TypeError
,
test_x_type
)
if
__name__
==
'__main__'
:
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录