Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
69394d1a
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
69394d1a
编写于
2月 09, 2022
作者:
H
hong
提交者:
GitHub
2月 09, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Revert "Move trace op to pten (#39227)"
This reverts commit
d7dddf94
.
上级
1bd7a143
变更
11
显示空白变更内容
内联
并排
Showing
11 changed file
with
200 addition
and
360 deletion
+200
-360
paddle/fluid/operators/trace_op.cc
paddle/fluid/operators/trace_op.cc
+19
-1
paddle/fluid/operators/trace_op.cu
paddle/fluid/operators/trace_op.cu
+77
-0
paddle/fluid/operators/trace_op.h
paddle/fluid/operators/trace_op.h
+104
-87
paddle/pten/kernels/cpu/trace_grad_kernel.cc
paddle/pten/kernels/cpu/trace_grad_kernel.cc
+0
-31
paddle/pten/kernels/cpu/trace_kernel.cc
paddle/pten/kernels/cpu/trace_kernel.cc
+0
-58
paddle/pten/kernels/gpu/trace_grad_kernel.cu
paddle/pten/kernels/gpu/trace_grad_kernel.cu
+0
-31
paddle/pten/kernels/gpu/trace_kernel.cu
paddle/pten/kernels/gpu/trace_kernel.cu
+0
-57
paddle/pten/kernels/trace_grad_kernel.h
paddle/pten/kernels/trace_grad_kernel.h
+0
-30
paddle/pten/kernels/trace_kernel.h
paddle/pten/kernels/trace_kernel.h
+0
-29
paddle/pten/ops/compat/trace_sig.cc
paddle/pten/ops/compat/trace_sig.cc
+0
-34
python/paddle/fluid/tests/unittests/test_trace_op.py
python/paddle/fluid/tests/unittests/test_trace_op.py
+0
-2
未找到文件。
paddle/fluid/operators/trace_op.cc
浏览文件 @
69394d1a
...
...
@@ -12,7 +12,7 @@
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/
framework/op_registry
.h"
#include "paddle/fluid/
operators/trace_op
.h"
#include "paddle/fluid/framework/op_version_registry.h"
namespace
paddle
{
...
...
@@ -161,6 +161,24 @@ REGISTER_OPERATOR(trace, ops::TraceOp, ops::TraceOpMaker,
REGISTER_OPERATOR
(
trace_grad
,
ops
::
TraceOpGrad
,
ops
::
TraceGradNoNeedBufferVarsInferer
);
REGISTER_OP_CPU_KERNEL
(
trace
,
ops
::
TraceKernel
<
paddle
::
platform
::
CPUDeviceContext
,
int
>
,
ops
::
TraceKernel
<
paddle
::
platform
::
CPUDeviceContext
,
float
>
,
ops
::
TraceKernel
<
paddle
::
platform
::
CPUDeviceContext
,
double
>
,
ops
::
TraceKernel
<
paddle
::
platform
::
CPUDeviceContext
,
int64_t
>
,
ops
::
TraceKernel
<
paddle
::
platform
::
CPUDeviceContext
,
paddle
::
platform
::
complex
<
float
>>
,
ops
::
TraceKernel
<
paddle
::
platform
::
CPUDeviceContext
,
paddle
::
platform
::
complex
<
double
>>
);
REGISTER_OP_CPU_KERNEL
(
trace_grad
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CPUDeviceContext
,
int
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CPUDeviceContext
,
float
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CPUDeviceContext
,
double
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CPUDeviceContext
,
int64_t
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CPUDeviceContext
,
paddle
::
platform
::
complex
<
float
>>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CPUDeviceContext
,
paddle
::
platform
::
complex
<
double
>>
);
/* ========================== register checkpoint ===========================*/
REGISTER_OP_VERSION
(
trace
)
...
...
paddle/fluid/operators/trace_op.cu
0 → 100644
浏览文件 @
69394d1a
// Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include <thrust/device_vector.h>
#include <thrust/host_vector.h>
#include "paddle/fluid/operators/math/math_function.h"
#include "paddle/fluid/operators/reduce_ops/reduce_op.cu.h"
#include "paddle/fluid/operators/trace_op.h"
namespace
paddle
{
namespace
operators
{
template
<
typename
DeviceContext
,
typename
T
>
class
TraceCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
input
=
context
.
Input
<
framework
::
Tensor
>
(
"Input"
);
auto
*
out
=
context
.
Output
<
framework
::
Tensor
>
(
"Out"
);
const
int64_t
offset
=
context
.
Attr
<
int
>
(
"offset"
);
const
int64_t
dim1
=
context
.
Attr
<
int
>
(
"axis1"
);
const
int64_t
dim2
=
context
.
Attr
<
int
>
(
"axis2"
);
T
*
out_data
=
out
->
mutable_data
<
T
>
(
context
.
GetPlace
());
const
framework
::
Tensor
diag
=
Diagonal
<
DeviceContext
,
T
>
(
context
,
input
,
offset
,
dim1
,
dim2
);
if
(
diag
.
numel
()
>
0
)
{
auto
stream
=
context
.
cuda_device_context
().
stream
();
std
::
vector
<
int
>
reduce_dims
;
reduce_dims
.
push_back
(
out
->
dims
().
size
());
TensorReduceImpl
<
T
,
T
,
kps
::
AddFunctor
,
kps
::
IdentityFunctor
<
T
>>
(
context
.
cuda_device_context
(),
diag
,
out
,
kps
::
IdentityFunctor
<
T
>
(),
reduce_dims
,
stream
);
}
else
{
math
::
SetConstant
<
DeviceContext
,
T
>
functor
;
functor
(
context
.
device_context
<
DeviceContext
>
(),
out
,
static_cast
<
T
>
(
0
));
}
}
};
}
// namespace operators
}
// namespace paddle
namespace
ops
=
paddle
::
operators
;
namespace
platform
=
paddle
::
platform
;
REGISTER_OP_CUDA_KERNEL
(
trace
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
int
>
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
int64_t
>
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
platform
::
float16
>
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
float
>
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
double
>
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
paddle
::
platform
::
complex
<
float
>>
,
ops
::
TraceCUDAKernel
<
paddle
::
platform
::
CUDADeviceContext
,
paddle
::
platform
::
complex
<
double
>>
);
REGISTER_OP_CUDA_KERNEL
(
trace_grad
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
int
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
int64_t
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
platform
::
float16
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
float
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
double
>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
paddle
::
platform
::
complex
<
float
>>
,
ops
::
TraceGradKernel
<
paddle
::
platform
::
CUDADeviceContext
,
paddle
::
platform
::
complex
<
double
>>
);
paddle/
pten/kernels/impl/trace_kernel_impl
.h
→
paddle/
fluid/operators/trace_op
.h
浏览文件 @
69394d1a
// Copyright (c) 202
2
PaddlePaddle Authors. All Rights Reserved.
// Copyright (c) 202
0
PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
...
...
@@ -13,26 +13,20 @@
// limitations under the License.
#pragma once
#if defined(__NVCC__) || defined(__HIPCC__)
#include <thrust/device_vector.h>
#include <thrust/host_vector.h>
#endif
#include <algorithm>
#include <vector>
#include "paddle/fluid/framework/eigen.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/operators/math/math_function.h"
#include "paddle/fluid/platform/for_range.h"
namespace
pten
{
namespace
paddle
{
namespace
operators
{
template
<
typename
T
>
struct
DiagonalFunctor
{
DiagonalFunctor
(
const
T
*
input
,
const
int64_t
*
diag_stride
,
const
int64_t
*
ret_strides
,
int64_t
pos
,
int64_t
dim_size
,
DiagonalFunctor
(
const
T
*
input
,
const
int64_t
*
diag_stride
,
const
int64_t
*
ret_strides
,
int64_t
pos
,
int64_t
dim_size
,
T
*
diag
)
:
input_
(
input
),
diag_stride_
(
diag_stride
),
...
...
@@ -61,15 +55,9 @@ struct DiagonalFunctor {
template
<
typename
T
>
struct
TraceGradFunctor
{
TraceGradFunctor
(
const
T
*
d_out
,
const
int64_t
*
out_stride
,
const
int64_t
*
x_strides
,
int64_t
pos
,
int64_t
dim_size
,
int64_t
dim1
,
int64_t
dim2
,
int64_t
diag_size
,
T
*
d_x
)
TraceGradFunctor
(
const
T
*
d_out
,
const
int64_t
*
out_stride
,
const
int64_t
*
x_strides
,
int64_t
pos
,
int64_t
dim_size
,
int64_t
dim1
,
int64_t
dim2
,
int64_t
diag_size
,
T
*
d_x
)
:
d_out_
(
d_out
),
out_stride_
(
out_stride
),
x_strides_
(
x_strides
),
...
...
@@ -113,12 +101,10 @@ struct TraceGradFunctor {
T
*
d_x_
;
};
template
<
typename
T
,
typename
DeviceContext
>
DenseTensor
Diagonal
(
const
DeviceContext
&
context
,
const
DenseTensor
*
input
,
int64_t
offset
,
int64_t
dim1
,
int64_t
dim2
)
{
template
<
typename
DeviceContext
,
typename
T
>
framework
::
Tensor
Diagonal
(
const
framework
::
ExecutionContext
&
context
,
const
framework
::
Tensor
*
input
,
int64_t
offset
,
int64_t
dim1
,
int64_t
dim2
)
{
auto
*
input_data
=
input
->
data
<
T
>
();
auto
input_dims
=
input
->
dims
();
auto
input_stride
=
framework
::
stride
(
input_dims
);
...
...
@@ -152,7 +138,7 @@ DenseTensor Diagonal(const DeviceContext& context,
}
ret_strides
.
push_back
(
stride1
+
stride2
);
ret_dims
.
push_back
(
diag_size
);
Dense
Tensor
diag
;
framework
::
Tensor
diag
;
framework
::
DDim
diag_dims
=
framework
::
make_ddim
(
ret_dims
);
auto
dig_stride
=
framework
::
stride
(
diag_dims
);
auto
diag_data
=
diag
.
mutable_data
<
T
>
(
diag_dims
,
context
.
GetPlace
());
...
...
@@ -169,10 +155,10 @@ DenseTensor Diagonal(const DeviceContext& context,
const
auto
*
ret_arr
=
ret_strides
.
data
();
#endif
//
auto& dev_ctx = context.template device_context<DeviceContext>();
p
addle
::
platform
::
ForRange
<
DeviceContext
>
for_range
(
context
,
diag
.
numel
());
DiagonalFunctor
<
T
>
functor
(
input_data
,
diag_arr
,
ret_arr
,
pos
,
dim_size
,
diag_data
);
auto
&
dev_ctx
=
context
.
template
device_context
<
DeviceContext
>();
p
latform
::
ForRange
<
DeviceContext
>
for_range
(
dev_ctx
,
diag
.
numel
());
DiagonalFunctor
<
T
>
functor
(
input_data
,
diag_arr
,
ret_arr
,
pos
,
dim_size
,
diag_data
);
for_range
(
functor
);
return
diag
;
}
else
{
...
...
@@ -180,27 +166,62 @@ DenseTensor Diagonal(const DeviceContext& context,
}
}
template
<
typename
T
,
typename
Context
>
void
TraceGradKernel
(
const
Context
&
ctx
,
const
DenseTensor
&
out_grad
,
const
DenseTensor
&
x
,
int
offset
,
int
axis1
,
int
axis2
,
DenseTensor
*
in_grad
)
{
auto
input_dims
=
in_grad
->
dims
();
template
<
typename
DeviceContext
,
typename
T
>
class
TraceKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
input
=
context
.
Input
<
framework
::
Tensor
>
(
"Input"
);
auto
*
out
=
context
.
Output
<
framework
::
Tensor
>
(
"Out"
);
const
int64_t
offset
=
context
.
Attr
<
int
>
(
"offset"
);
const
int64_t
dim1
=
context
.
Attr
<
int
>
(
"axis1"
);
const
int64_t
dim2
=
context
.
Attr
<
int
>
(
"axis2"
);
auto
output_dims
=
out
->
dims
();
T
*
out_data
=
out
->
mutable_data
<
T
>
(
context
.
GetPlace
());
const
framework
::
Tensor
diag
=
Diagonal
<
DeviceContext
,
T
>
(
context
,
input
,
offset
,
dim1
,
dim2
);
if
(
diag
.
numel
()
>
0
)
{
auto
x
=
framework
::
EigenMatrix
<
T
>::
Reshape
(
diag
,
diag
.
dims
().
size
()
-
1
);
auto
output
=
framework
::
EigenVector
<
T
>::
Flatten
(
*
out
);
auto
&
place
=
*
context
.
template
device_context
<
DeviceContext
>().
eigen_device
();
auto
reduce_dim
=
Eigen
::
array
<
int
,
1
>
({
1
});
output
.
device
(
place
)
=
x
.
sum
(
reduce_dim
);
out
->
Resize
(
output_dims
);
}
else
{
std
::
fill
(
out_data
,
out_data
+
out
->
numel
(),
static_cast
<
T
>
(
0
));
}
}
};
template
<
typename
DeviceContext
,
typename
T
>
class
TraceGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
auto
*
d_out
=
context
.
Input
<
framework
::
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
auto
*
d_x
=
context
.
Output
<
framework
::
Tensor
>
(
framework
::
GradVarName
(
"Input"
));
int64_t
offset
=
context
.
Attr
<
int
>
(
"offset"
);
int64_t
dim1
=
context
.
Attr
<
int
>
(
"axis1"
);
int64_t
dim2
=
context
.
Attr
<
int
>
(
"axis2"
);
auto
input_dims
=
d_x
->
dims
();
auto
input_stride
=
framework
::
stride
(
input_dims
);
auto
output_dims
=
out_grad
.
dims
();
auto
output_dims
=
d_out
->
dims
();
auto
output_stride
=
framework
::
stride
(
output_dims
);
auto
*
out_data
=
out_grad
.
data
<
T
>
();
T
*
x_data
=
in_grad
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
auto
*
out_data
=
d_out
->
data
<
T
>
();
T
*
x_data
=
d_x
->
mutable_data
<
T
>
(
context
.
GetPlace
());
paddle
::
operators
::
math
::
SetConstant
<
Context
,
T
>
set_zero
;
math
::
SetConstant
<
DeviceContext
,
T
>
set_zero
;
auto
&
dev_ctx
=
context
.
template
device_context
<
DeviceContext
>();
set_zero
(
dev_ctx
,
d_x
,
static_cast
<
T
>
(
0.0
));
set_zero
(
ctx
,
in_grad
,
static_cast
<
T
>
(
0.0
));
auto
dim1
=
axis1
;
auto
dim2
=
axis2
;
auto
dim1_
=
dim1
<
0
?
input_dims
.
size
()
+
dim1
:
dim1
;
auto
dim2_
=
dim2
<
0
?
input_dims
.
size
()
+
dim2
:
dim2
;
auto
len1
=
input_dims
[
std
::
min
(
dim1_
,
dim2_
)];
...
...
@@ -230,18 +251,14 @@ void TraceGradKernel(const Context& ctx,
const
auto
*
input_arr
=
input_stride
.
Get
();
#endif
paddle
::
platform
::
ForRange
<
Context
>
for_range
(
ctx
,
in_grad
->
numel
());
TraceGradFunctor
<
T
>
functor
(
out_data
,
output_arr
,
input_arr
,
pos
,
input_dims
.
size
(),
dim1_
,
dim2_
,
diag_size
,
platform
::
ForRange
<
DeviceContext
>
for_range
(
dev_ctx
,
d_x
->
numel
());
TraceGradFunctor
<
T
>
functor
(
out_data
,
output_arr
,
input_arr
,
pos
,
input_dims
.
size
(),
dim1_
,
dim2_
,
diag_size
,
x_data
);
for_range
(
functor
);
}
}
}
};
}
// namespace pten
}
// namespace operators
}
// namespace paddle
paddle/pten/kernels/cpu/trace_grad_kernel.cc
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/pten/kernels/trace_grad_kernel.h"
#include "paddle/pten/kernels/impl/trace_kernel_impl.h"
#include "paddle/pten/backends/cpu/cpu_context.h"
#include "paddle/pten/core/kernel_registry.h"
PT_REGISTER_KERNEL
(
trace_grad
,
CPU
,
ALL_LAYOUT
,
pten
::
TraceGradKernel
,
float
,
double
,
int
,
int64_t
,
paddle
::
platform
::
float16
,
paddle
::
platform
::
complex
<
float
>
,
paddle
::
platform
::
complex
<
double
>
)
{}
paddle/pten/kernels/cpu/trace_kernel.cc
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/pten/kernels/trace_kernel.h"
#include "paddle/pten/backends/cpu/cpu_context.h"
#include "paddle/pten/core/kernel_registry.h"
#include "paddle/pten/kernels/impl/trace_kernel_impl.h"
namespace
pten
{
template
<
typename
T
,
typename
Context
>
void
TraceKernel
(
const
Context
&
ctx
,
const
DenseTensor
&
x
,
int
offset
,
int
axis1
,
int
axis2
,
DenseTensor
*
out
)
{
auto
output_dims
=
out
->
dims
();
T
*
out_data
=
out
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
const
DenseTensor
diag
=
Diagonal
<
T
,
Context
>
(
ctx
,
&
x
,
offset
,
axis1
,
axis2
);
if
(
diag
.
numel
()
>
0
)
{
auto
x
=
paddle
::
framework
::
EigenMatrix
<
T
>::
Reshape
(
diag
,
diag
.
dims
().
size
()
-
1
);
auto
output
=
paddle
::
framework
::
EigenVector
<
T
>::
Flatten
(
*
out
);
auto
reduce_dim
=
Eigen
::
array
<
int
,
1
>
({
1
});
output
.
device
(
*
ctx
.
eigen_device
())
=
x
.
sum
(
reduce_dim
);
out
->
Resize
(
output_dims
);
}
else
{
std
::
fill
(
out_data
,
out_data
+
out
->
numel
(),
static_cast
<
T
>
(
0
));
}
}
}
// namespace pten
PT_REGISTER_KERNEL
(
trace
,
CPU
,
ALL_LAYOUT
,
pten
::
TraceKernel
,
float
,
double
,
int
,
int64_t
,
paddle
::
platform
::
float16
,
paddle
::
platform
::
complex
<
float
>
,
paddle
::
platform
::
complex
<
double
>
)
{}
paddle/pten/kernels/gpu/trace_grad_kernel.cu
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/pten/kernels/impl/trace_kernel_impl.h"
#include "paddle/pten/kernels/trace_grad_kernel.h"
#include "paddle/pten/backends/cpu/cpu_context.h"
#include "paddle/pten/core/kernel_registry.h"
PT_REGISTER_KERNEL
(
trace_grad
,
GPU
,
ALL_LAYOUT
,
pten
::
TraceGradKernel
,
float
,
double
,
int
,
int64_t
,
paddle
::
platform
::
float16
,
paddle
::
platform
::
complex
<
float
>
,
paddle
::
platform
::
complex
<
double
>
)
{}
paddle/pten/kernels/gpu/trace_kernel.cu
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/pten/backends/gpu/gpu_context.h"
#include "paddle/pten/core/kernel_registry.h"
#include "paddle/pten/kernels/gpu/reduce.h"
#include "paddle/pten/kernels/impl/trace_kernel_impl.h"
#include "paddle/pten/kernels/trace_kernel.h"
namespace
pten
{
template
<
typename
T
,
typename
Context
>
void
TraceKernel
(
const
Context
&
ctx
,
const
DenseTensor
&
x
,
int
offset
,
int
axis1
,
int
axis2
,
DenseTensor
*
out
)
{
T
*
out_data
=
out
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
auto
diag
=
Diagonal
<
T
,
Context
>
(
ctx
,
&
x
,
offset
,
axis1
,
axis2
);
if
(
diag
.
numel
()
>
0
)
{
auto
stream
=
ctx
.
stream
();
std
::
vector
<
int
>
reduce_dims
;
reduce_dims
.
push_back
(
out
->
dims
().
size
());
kernels
::
TensorReduceFunctorImpl
<
T
,
T
,
kps
::
AddFunctor
,
kps
::
IdentityFunctor
<
T
>>
(
ctx
,
diag
,
out
,
kps
::
IdentityFunctor
<
T
>
(),
reduce_dims
,
stream
);
}
else
{
paddle
::
operators
::
math
::
SetConstant
<
Context
,
T
>
functor
;
functor
(
ctx
,
out
,
static_cast
<
T
>
(
0
));
}
}
}
// namespace pten
PT_REGISTER_KERNEL
(
trace
,
GPU
,
ALL_LAYOUT
,
pten
::
TraceKernel
,
float
,
double
,
int
,
int64_t
,
paddle
::
platform
::
float16
,
paddle
::
platform
::
complex
<
float
>
,
paddle
::
platform
::
complex
<
double
>
)
{}
paddle/pten/kernels/trace_grad_kernel.h
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#pragma once
#include "paddle/pten/core/dense_tensor.h"
namespace
pten
{
template
<
typename
T
,
typename
Context
>
void
TraceGradKernel
(
const
Context
&
ctx
,
const
DenseTensor
&
out_grad
,
const
DenseTensor
&
x
,
int
offset
,
int
axis1
,
int
axis2
,
DenseTensor
*
in_grad
);
}
// namespace pten
paddle/pten/kernels/trace_kernel.h
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#pragma once
#include "paddle/pten/core/dense_tensor.h"
namespace
pten
{
template
<
typename
T
,
typename
Context
>
void
TraceKernel
(
const
Context
&
ctx
,
const
DenseTensor
&
x
,
int
offset
,
int
axis1
,
int
axis2
,
DenseTensor
*
out
);
}
// namespace pten
paddle/pten/ops/compat/trace_sig.cc
已删除
100644 → 0
浏览文件 @
1bd7a143
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/pten/core/compat/op_utils.h"
namespace
pten
{
KernelSignature
TraceOpArgumentMapping
(
const
ArgumentMappingContext
&
ctx
)
{
return
KernelSignature
(
"trace"
,
{
"Input"
},
{
"offset"
,
"axis1"
,
"axis2"
},
{
"Out"
});
}
KernelSignature
TraceGradOpArgumentMapping
(
const
ArgumentMappingContext
&
ctx
)
{
return
KernelSignature
(
"trace_grad"
,
{
GradVarName
(
"Out"
),
"Input"
},
{
"offset"
,
"axis1"
,
"axis2"
},
{
GradVarName
(
"Input"
)});
}
}
// namespace pten
PT_REGISTER_ARG_MAPPING_FN
(
trace
,
pten
::
TraceOpArgumentMapping
);
PT_REGISTER_ARG_MAPPING_FN
(
trace_grad
,
pten
::
TraceGradOpArgumentMapping
);
python/paddle/fluid/tests/unittests/test_trace_op.py
浏览文件 @
69394d1a
...
...
@@ -21,7 +21,6 @@ import paddle.nn.functional as F
import
paddle.fluid
as
fluid
import
paddle.fluid.core
as
core
import
paddle.tensor
as
tensor
import
paddle
class
TestTraceOp
(
OpTest
):
...
...
@@ -87,5 +86,4 @@ class TestTraceAPICase(unittest.TestCase):
if
__name__
==
"__main__"
:
paddle
.
enable_static
()
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录