Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
3a5693e0
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 1 年 前同步成功
通知
2299
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
3a5693e0
编写于
9月 27, 2017
作者:
Y
Yu Yang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Add Skeleton of Double support
上级
d2edbe57
变更
60
显示空白变更内容
内联
并排
Showing
60 changed file
with
217 addition
and
129 deletion
+217
-129
paddle/framework/data_type.h
paddle/framework/data_type.h
+36
-0
paddle/framework/op_registry.h
paddle/framework/op_registry.h
+3
-2
paddle/framework/operator.h
paddle/framework/operator.h
+62
-15
paddle/framework/tensor.h
paddle/framework/tensor.h
+2
-10
paddle/operators/accuracy_op.cu
paddle/operators/accuracy_op.cu
+1
-1
paddle/operators/accuracy_op.h
paddle/operators/accuracy_op.h
+1
-1
paddle/operators/activation_op.h
paddle/operators/activation_op.h
+10
-10
paddle/operators/add_op.h
paddle/operators/add_op.h
+1
-1
paddle/operators/clip_op.h
paddle/operators/clip_op.h
+2
-2
paddle/operators/concat_op.h
paddle/operators/concat_op.h
+1
-1
paddle/operators/cos_sim_op.h
paddle/operators/cos_sim_op.h
+2
-2
paddle/operators/crop_op.h
paddle/operators/crop_op.h
+2
-2
paddle/operators/cross_entropy_op.cc
paddle/operators/cross_entropy_op.cc
+12
-0
paddle/operators/cross_entropy_op.cu
paddle/operators/cross_entropy_op.cu
+2
-2
paddle/operators/cross_entropy_op.h
paddle/operators/cross_entropy_op.h
+2
-2
paddle/operators/dropout_op.cu
paddle/operators/dropout_op.cu
+1
-1
paddle/operators/dropout_op.h
paddle/operators/dropout_op.h
+2
-2
paddle/operators/elementwise_add_op.h
paddle/operators/elementwise_add_op.h
+2
-2
paddle/operators/elementwise_div_op.h
paddle/operators/elementwise_div_op.h
+2
-2
paddle/operators/elementwise_mul_op.h
paddle/operators/elementwise_mul_op.h
+2
-2
paddle/operators/elementwise_sub_op.h
paddle/operators/elementwise_sub_op.h
+2
-2
paddle/operators/fill_zeros_like_op.h
paddle/operators/fill_zeros_like_op.h
+1
-1
paddle/operators/gather_op.h
paddle/operators/gather_op.h
+2
-2
paddle/operators/gaussian_random_op.cc
paddle/operators/gaussian_random_op.cc
+1
-1
paddle/operators/gaussian_random_op.cu
paddle/operators/gaussian_random_op.cu
+1
-1
paddle/operators/gemm_conv2d_op.h
paddle/operators/gemm_conv2d_op.h
+2
-2
paddle/operators/lookup_table_op.cu
paddle/operators/lookup_table_op.cu
+2
-2
paddle/operators/lookup_table_op.h
paddle/operators/lookup_table_op.h
+2
-2
paddle/operators/lstm_unit_op.cu
paddle/operators/lstm_unit_op.cu
+2
-2
paddle/operators/lstm_unit_op.h
paddle/operators/lstm_unit_op.h
+2
-2
paddle/operators/mean_op.h
paddle/operators/mean_op.h
+2
-2
paddle/operators/minus_op.h
paddle/operators/minus_op.h
+1
-1
paddle/operators/modified_huber_loss_op.cu
paddle/operators/modified_huber_loss_op.cu
+1
-1
paddle/operators/modified_huber_loss_op.h
paddle/operators/modified_huber_loss_op.h
+2
-2
paddle/operators/mul_op.h
paddle/operators/mul_op.h
+2
-2
paddle/operators/multiplex_op.cu
paddle/operators/multiplex_op.cu
+2
-2
paddle/operators/multiplex_op.h
paddle/operators/multiplex_op.h
+2
-2
paddle/operators/pad_op.h
paddle/operators/pad_op.h
+2
-2
paddle/operators/prelu_op.h
paddle/operators/prelu_op.h
+2
-2
paddle/operators/rank_loss_op.h
paddle/operators/rank_loss_op.h
+2
-2
paddle/operators/reshape_op.h
paddle/operators/reshape_op.h
+2
-2
paddle/operators/rowwise_add_op.h
paddle/operators/rowwise_add_op.h
+2
-2
paddle/operators/scale_op.h
paddle/operators/scale_op.h
+1
-1
paddle/operators/scatter_op.h
paddle/operators/scatter_op.h
+2
-2
paddle/operators/sequence_pool_op.h
paddle/operators/sequence_pool_op.h
+2
-2
paddle/operators/sgd_op.h
paddle/operators/sgd_op.h
+1
-1
paddle/operators/smooth_l1_loss_op.h
paddle/operators/smooth_l1_loss_op.h
+2
-2
paddle/operators/softmax_op.h
paddle/operators/softmax_op.h
+2
-2
paddle/operators/softmax_with_cross_entropy_op.cu
paddle/operators/softmax_with_cross_entropy_op.cu
+2
-2
paddle/operators/softmax_with_cross_entropy_op.h
paddle/operators/softmax_with_cross_entropy_op.h
+2
-2
paddle/operators/split_op.h
paddle/operators/split_op.h
+1
-1
paddle/operators/squared_l2_distance_op.h
paddle/operators/squared_l2_distance_op.h
+2
-2
paddle/operators/sum_op.h
paddle/operators/sum_op.h
+2
-2
paddle/operators/top_k_op.cu
paddle/operators/top_k_op.cu
+1
-1
paddle/operators/top_k_op.h
paddle/operators/top_k_op.h
+1
-1
paddle/operators/transpose_op.h
paddle/operators/transpose_op.h
+2
-2
paddle/operators/uniform_random_op.cc
paddle/operators/uniform_random_op.cc
+1
-1
paddle/operators/uniform_random_op.cu
paddle/operators/uniform_random_op.cu
+1
-1
paddle/platform/place.cc
paddle/platform/place.cc
+1
-1
paddle/pybind/tensor_py.h
paddle/pybind/tensor_py.h
+4
-4
未找到文件。
paddle/framework/data_type.h
0 → 100644
浏览文件 @
3a5693e0
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#include <typeindex>
#include "paddle/framework/framework.pb.h"
namespace
paddle
{
namespace
framework
{
inline
DataType
ToDataType
(
std
::
type_index
type
)
{
if
(
typeid
(
float
).
hash_code
()
==
type
.
hash_code
())
{
return
DataType
::
FP32
;
}
else
if
(
typeid
(
double
).
hash_code
()
==
type
.
hash_code
())
{
return
DataType
::
FP64
;
}
else
if
(
typeid
(
int
).
hash_code
()
==
type
.
hash_code
())
{
return
DataType
::
INT32
;
}
else
{
PADDLE_THROW
(
"Not supported"
);
return
static_cast
<
DataType
>
(
-
1
);
}
}
}
// namespace framework
}
// namespace paddle
paddle/framework/op_registry.h
浏览文件 @
3a5693e0
...
...
@@ -104,8 +104,9 @@ template <typename PlaceType, typename KernelType>
class
OpKernelRegistrar
:
public
Registrar
{
public:
explicit
OpKernelRegistrar
(
const
char
*
op_type
)
{
OperatorWithKernel
::
OpKernelKey
key
;
key
.
place_
=
PlaceType
();
using
T
=
typename
KernelType
::
ELEMENT_TYPE
;
OperatorWithKernel
::
OpKernelKey
key
(
ToDataType
(
std
::
type_index
(
typeid
(
T
))),
PlaceType
());
OperatorWithKernel
::
AllOpKernels
()[
op_type
][
key
].
reset
(
new
KernelType
);
}
};
...
...
paddle/framework/operator.h
浏览文件 @
3a5693e0
...
...
@@ -21,6 +21,7 @@ limitations under the License. */
#include "op_info.h"
#include "paddle/framework/attribute.h"
#include "paddle/framework/data_type.h"
#include "paddle/framework/framework.pb.h"
#include "paddle/framework/lod_tensor.h"
#include "paddle/framework/scope.h"
...
...
@@ -407,7 +408,7 @@ class RuntimeInferShapeContext : public InferShapeContextBase {
const
Scope
&
scope_
;
};
class
OpKernel
{
class
OpKernel
Base
{
public:
/**
* ExecutionContext is the only parameter of Kernel Run function.
...
...
@@ -418,33 +419,47 @@ class OpKernel {
virtual
void
Compute
(
const
ExecutionContext
&
context
)
const
=
0
;
virtual
~
OpKernel
()
{}
virtual
~
OpKernelBase
()
=
default
;
};
template
<
typename
T
>
class
OpKernel
:
public
OpKernelBase
{
public:
using
ELEMENT_TYPE
=
T
;
};
class
OperatorWithKernel
:
public
OperatorBase
{
public:
struct
OpKernelKey
{
platform
::
Place
place_
;
DataType
data_type_
;
OpKernelKey
()
=
default
;
explicit
OpKernelKey
(
const
platform
::
DeviceContext
&
dev_ctx
)
{
place_
=
dev_ctx
.
GetPlace
();
}
OpKernelKey
(
DataType
data_type
,
platform
::
Place
place
)
:
place_
(
place
),
data_type_
(
data_type
)
{}
OpKernelKey
(
DataType
data_type
,
const
platform
::
DeviceContext
&
dev_ctx
)
:
place_
(
dev_ctx
.
GetPlace
()),
data_type_
(
data_type
)
{}
bool
operator
==
(
const
OpKernelKey
&
o
)
const
{
return
platform
::
places_are_same_class
(
place_
,
o
.
place_
);
return
platform
::
places_are_same_class
(
place_
,
o
.
place_
)
&&
data_type_
==
o
.
data_type_
;
}
};
struct
OpKernelHash
{
std
::
hash
<
bool
>
hash_
;
std
::
hash
<
int
>
hash_
;
size_t
operator
()(
const
OpKernelKey
&
key
)
const
{
return
hash_
(
platform
::
is_gpu_place
(
key
.
place_
));
int
place
=
key
.
place_
.
which
();
int
data_type
=
static_cast
<
int
>
(
key
.
data_type_
);
// NOTE: Number of places limit to 16.
int
pre_hash
=
data_type
<<
4
|
(
place
&
0x0F
);
return
hash_
(
pre_hash
);
}
};
using
OpKernelMap
=
std
::
unordered_map
<
OpKernelKey
,
std
::
unique_ptr
<
OpKernel
>
,
OpKernelHash
>
;
std
::
unordered_map
<
OpKernelKey
,
std
::
unique_ptr
<
OpKernelBase
>
,
OpKernelHash
>
;
OperatorWithKernel
(
const
std
::
string
&
type
,
const
VariableNameMap
&
inputs
,
const
VariableNameMap
&
outputs
,
const
AttributeMap
&
attrs
)
...
...
@@ -458,8 +473,10 @@ class OperatorWithKernel : public OperatorBase {
void
Run
(
const
Scope
&
scope
,
const
platform
::
DeviceContext
&
dev_ctx
)
const
final
{
auto
&
opKernel
=
AllOpKernels
().
at
(
type_
).
at
(
OpKernelKey
(
dev_ctx
));
opKernel
->
Compute
(
ExecutionContext
(
*
this
,
scope
,
dev_ctx
));
ExecutionContext
ctx
(
*
this
,
scope
,
dev_ctx
);
auto
&
opKernel
=
AllOpKernels
().
at
(
type_
).
at
(
OpKernelKey
(
IndicateDataType
(
ctx
),
dev_ctx
));
opKernel
->
Compute
(
ctx
);
}
static
std
::
unordered_map
<
std
::
string
/* op_type */
,
OpKernelMap
>&
...
...
@@ -469,13 +486,43 @@ class OperatorWithKernel : public OperatorBase {
}
bool
SupportGPU
()
const
override
{
OperatorWithKernel
::
OpKernelKey
key
;
key
.
place_
=
platform
::
GPUPlace
();
return
OperatorWithKernel
::
AllOpKernels
().
at
(
type_
).
count
(
key
)
!=
0
;
auto
&
op_kernels
=
OperatorWithKernel
::
AllOpKernels
().
at
(
type_
);
return
std
::
any_of
(
op_kernels
.
begin
(),
op_kernels
.
end
(),
[](
OpKernelMap
::
const_reference
kern_pair
)
{
return
platform
::
is_gpu_place
(
kern_pair
.
first
.
place_
);
});
}
protected:
virtual
void
InferShape
(
InferShapeContextBase
*
ctx
)
const
=
0
;
// indicate kernel DataType by input data. Defaultly all input data must be
// same.
virtual
DataType
IndicateDataType
(
const
ExecutionContext
&
ctx
)
const
{
auto
&
scope
=
ctx
.
scope
();
int
data_type
=
-
1
;
for
(
auto
&
input
:
this
->
inputs_
)
{
for
(
auto
&
ipt_name
:
input
.
second
)
{
auto
*
var
=
scope
.
FindVar
(
ipt_name
);
if
(
var
!=
nullptr
)
{
const
Tensor
*
t
=
nullptr
;
if
(
var
->
IsType
<
Tensor
>
())
{
t
=
&
var
->
Get
<
Tensor
>
();
}
else
if
(
var
->
IsType
<
LoDTensor
>
())
{
t
=
&
var
->
Get
<
LoDTensor
>
();
}
if
(
t
!=
nullptr
)
{
int
tmp
=
static_cast
<
int
>
(
ToDataType
(
t
->
type
()));
PADDLE_ENFORCE
(
tmp
==
data_type
||
data_type
==
-
1
,
"DataType of Paddle Op must be same."
);
data_type
=
tmp
;
}
}
}
}
PADDLE_ENFORCE
(
data_type
!=
-
1
,
"DataType should be indicated by input"
);
return
static_cast
<
DataType
>
(
data_type
);
}
};
}
// namespace framework
...
...
paddle/framework/tensor.h
浏览文件 @
3a5693e0
...
...
@@ -29,20 +29,10 @@ limitations under the License. */
namespace
paddle
{
namespace
pybind
{
namespace
details
{
template
<
bool
less
,
size_t
i
,
typename
...
args
>
struct
CastToPyBufferImpl
;
}
}
// namespace pybind
namespace
framework
{
class
Tensor
{
public:
template
<
bool
less
,
size_t
i
,
typename
...
args
>
friend
struct
pybind
::
details
::
CastToPyBufferImpl
;
template
<
typename
T
,
size_t
D
,
int
MajorType
,
typename
IndexType
>
friend
struct
EigenTensor
;
...
...
@@ -119,6 +109,8 @@ class Tensor {
return
holder_
->
place
();
}
std
::
type_index
type
()
const
{
return
holder_
->
type
();
}
private:
template
<
typename
T
>
inline
void
check_memory_size
()
const
;
...
...
paddle/operators/accuracy_op.cu
浏览文件 @
3a5693e0
...
...
@@ -47,7 +47,7 @@ __global__ void AccuracyCudaKernel(const int N, const int D, const int* Xdata,
}
template
<
typename
T
>
class
AccuracyOpCUDAKernel
:
public
framework
::
OpKernel
{
class
AccuracyOpCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
...
...
paddle/operators/accuracy_op.h
浏览文件 @
3a5693e0
...
...
@@ -35,7 +35,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenScalar
=
framework
::
EigenScalar
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
AccuracyKernel
:
public
framework
::
OpKernel
{
class
AccuracyKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
inference
=
ctx
.
Input
<
Tensor
>
(
"Inference"
);
...
...
paddle/operators/activation_op.h
浏览文件 @
3a5693e0
...
...
@@ -20,7 +20,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
,
typename
Functor
>
class
ActivationKernel
:
public
framework
::
OpKernel
{
class
ActivationKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -36,7 +36,7 @@ class ActivationKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
Functor
>
class
ActivationGradKernel
:
public
framework
::
OpKernel
{
class
ActivationGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -202,7 +202,7 @@ struct SquareGradFunctor {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
BReluKernel
:
public
framework
::
OpKernel
{
class
BReluKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -219,7 +219,7 @@ class BReluKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
BReluGradKernel
:
public
framework
::
OpKernel
{
class
BReluGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -239,7 +239,7 @@ class BReluGradKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
SoftReluKernel
:
public
framework
::
OpKernel
{
class
SoftReluKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -256,7 +256,7 @@ class SoftReluKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
SoftReluGradKernel
:
public
framework
::
OpKernel
{
class
SoftReluGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -277,7 +277,7 @@ class SoftReluGradKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
PowKernel
:
public
framework
::
OpKernel
{
class
PowKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -293,7 +293,7 @@ class PowKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
PowGradKernel
:
public
framework
::
OpKernel
{
class
PowGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -312,7 +312,7 @@ class PowGradKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
STanhKernel
:
public
framework
::
OpKernel
{
class
STanhKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -329,7 +329,7 @@ class STanhKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
STanhGradKernel
:
public
framework
::
OpKernel
{
class
STanhGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
X
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
paddle/operators/add_op.h
浏览文件 @
3a5693e0
...
...
@@ -25,7 +25,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenVector
=
framework
::
EigenVector
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
AddKernel
:
public
framework
::
OpKernel
{
class
AddKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
input0
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
paddle/operators/clip_op.h
浏览文件 @
3a5693e0
...
...
@@ -56,7 +56,7 @@ class ClipGradFunctor {
};
template
<
typename
Place
,
typename
T
>
class
ClipKernel
:
public
framework
::
OpKernel
{
class
ClipKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
max
=
context
.
Attr
<
T
>
(
"max"
);
...
...
@@ -73,7 +73,7 @@ class ClipKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
ClipGradKernel
:
public
framework
::
OpKernel
{
class
ClipGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
max
=
context
.
Attr
<
T
>
(
"max"
);
...
...
paddle/operators/concat_op.h
浏览文件 @
3a5693e0
...
...
@@ -21,7 +21,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
ConcatKernel
:
public
framework
::
OpKernel
{
class
ConcatKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
ins
=
ctx
.
MultiInput
<
framework
::
Tensor
>
(
"X"
);
...
...
paddle/operators/cos_sim_op.h
浏览文件 @
3a5693e0
...
...
@@ -28,7 +28,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenVector
=
framework
::
EigenVector
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
CosSimKernel
:
public
framework
::
OpKernel
{
class
CosSimKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
// get Tensor
...
...
@@ -67,7 +67,7 @@ class CosSimKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
CosSimGradKernel
:
public
framework
::
OpKernel
{
class
CosSimGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
// get Tensor
...
...
paddle/operators/crop_op.h
浏览文件 @
3a5693e0
...
...
@@ -27,7 +27,7 @@ using EigenTensor = framework::EigenTensor<T, D, MajorType, IndexType>;
using
framework
::
Tensor
;
template
<
typename
T
>
class
CropKernel
:
public
framework
::
OpKernel
{
class
CropKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
x
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -69,7 +69,7 @@ void CropGradFunction(const framework::ExecutionContext& context) {
}
template
<
typename
Place
,
typename
T
>
class
CropGradKernel
:
public
framework
::
OpKernel
{
class
CropGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
size_t
rank
=
...
...
paddle/operators/cross_entropy_op.cc
浏览文件 @
3a5693e0
...
...
@@ -47,6 +47,12 @@ class CrossEntropyOp : public framework::OperatorWithKernel {
ctx
->
SetOutputDim
(
"Y"
,
{
x_dims
[
0
],
1
});
ctx
->
ShareLoD
(
"X"
,
/*->*/
"Y"
);
}
// CrossEntropy's data type just determined by "X"
framework
::
DataType
IndicateDataType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
return
framework
::
ToDataType
(
ctx
.
Input
<
Tensor
>
(
"X"
)
->
type
());
}
};
class
CrossEntropyGradientOp
:
public
framework
::
OperatorWithKernel
{
...
...
@@ -87,6 +93,12 @@ class CrossEntropyGradientOp : public framework::OperatorWithKernel {
}
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
x_dims
);
}
// CrossEntropy's data type just determined by "X"
framework
::
DataType
IndicateDataType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
return
framework
::
ToDataType
(
ctx
.
Input
<
Tensor
>
(
"X"
)
->
type
());
}
};
class
CrossEntropyOpMaker
:
public
framework
::
OpProtoAndCheckerMaker
{
...
...
paddle/operators/cross_entropy_op.cu
浏览文件 @
3a5693e0
...
...
@@ -53,7 +53,7 @@ __global__ void SoftCrossEntropyGradientKernel(T* dX, const T* dY, const T* X,
}
// namespace
template
<
typename
T
>
class
CrossEntropyOpCUDAKernel
:
public
framework
::
OpKernel
{
class
CrossEntropyOpCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
...
...
@@ -69,7 +69,7 @@ class CrossEntropyOpCUDAKernel : public framework::OpKernel {
};
template
<
typename
T
>
class
CrossEntropyGradientOpCUDAKernel
:
public
framework
::
OpKernel
{
class
CrossEntropyGradientOpCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
...
...
paddle/operators/cross_entropy_op.h
浏览文件 @
3a5693e0
...
...
@@ -26,7 +26,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
T
>
class
CrossEntropyOpKernel
:
public
framework
::
OpKernel
{
class
CrossEntropyOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_cpu_place
(
ctx
.
GetPlace
()),
...
...
@@ -42,7 +42,7 @@ class CrossEntropyOpKernel : public framework::OpKernel {
};
template
<
typename
T
>
class
CrossEntropyGradientOpKernel
:
public
framework
::
OpKernel
{
class
CrossEntropyGradientOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_cpu_place
(
ctx
.
GetPlace
()),
...
...
paddle/operators/dropout_op.cu
浏览文件 @
3a5693e0
...
...
@@ -47,7 +47,7 @@ struct MaskGenerator {
// Use std::random and thrust::random(thrust is a std library in CUDA) to
// implement uniform random.
template
<
typename
Place
,
typename
T
,
typename
AttrType
>
class
GPUDropoutKernel
:
public
framework
::
OpKernel
{
class
GPUDropoutKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
x
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
paddle/operators/dropout_op.h
浏览文件 @
3a5693e0
...
...
@@ -26,7 +26,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
,
typename
AttrType
>
class
CPUDropoutKernel
:
public
framework
::
OpKernel
{
class
CPUDropoutKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
x
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -62,7 +62,7 @@ class CPUDropoutKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
DropoutGradKernel
:
public
framework
::
OpKernel
{
class
DropoutGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
PADDLE_ENFORCE
(
context
.
Attr
<
bool
>
(
"is_training"
),
...
...
paddle/operators/elementwise_add_op.h
浏览文件 @
3a5693e0
...
...
@@ -20,7 +20,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
ElementwiseAddKernel
:
public
framework
::
OpKernel
{
class
ElementwiseAddKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseCompute
<
EigenAddFunctor
,
Place
,
T
>
(
ctx
);
...
...
@@ -101,7 +101,7 @@ struct ElementwiseAddBroadCast2GradFunctor {
};
template
<
typename
Place
,
typename
T
>
class
ElementwiseAddGradKernel
:
public
framework
::
OpKernel
{
class
ElementwiseAddGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseGradCompute
<
Place
,
T
,
ElementwiseAddGradFunctor
<
T
>
,
...
...
paddle/operators/elementwise_div_op.h
浏览文件 @
3a5693e0
...
...
@@ -20,7 +20,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
ElementwiseDivKernel
:
public
framework
::
OpKernel
{
class
ElementwiseDivKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseCompute
<
EigenDivFunctor
,
Place
,
T
>
(
ctx
);
...
...
@@ -103,7 +103,7 @@ struct ElementwiseDivBroadCast2GradFunctor {
};
template
<
typename
Place
,
typename
T
>
class
ElementwiseDivGradKernel
:
public
framework
::
OpKernel
{
class
ElementwiseDivGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseGradCompute
<
Place
,
T
,
ElementwiseDivGradFunctor
<
T
>
,
...
...
paddle/operators/elementwise_mul_op.h
浏览文件 @
3a5693e0
...
...
@@ -19,7 +19,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
ElementwiseMulKernel
:
public
framework
::
OpKernel
{
class
ElementwiseMulKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseCompute
<
EigenMulFunctor
,
Place
,
T
>
(
ctx
);
...
...
@@ -102,7 +102,7 @@ struct ElementwiseMulBroadCast2GradFunctor {
};
template
<
typename
Place
,
typename
T
>
class
ElementwiseMulGradKernel
:
public
framework
::
OpKernel
{
class
ElementwiseMulGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseGradCompute
<
Place
,
T
,
ElementwiseMulGradFunctor
<
T
>
,
...
...
paddle/operators/elementwise_sub_op.h
浏览文件 @
3a5693e0
...
...
@@ -19,7 +19,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
ElementwiseSubKernel
:
public
framework
::
OpKernel
{
class
ElementwiseSubKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseCompute
<
EigenSubFunctor
,
Place
,
T
>
(
ctx
);
...
...
@@ -102,7 +102,7 @@ struct ElementwiseSubBroadCast2GradFunctor {
};
template
<
typename
Place
,
typename
T
>
class
ElementwiseSubGradKernel
:
public
framework
::
OpKernel
{
class
ElementwiseSubGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
ElementwiseGradCompute
<
Place
,
T
,
ElementwiseSubGradFunctor
<
T
>
,
...
...
paddle/operators/fill_zeros_like_op.h
浏览文件 @
3a5693e0
...
...
@@ -20,7 +20,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
FillZerosLikeKernel
:
public
framework
::
OpKernel
{
class
FillZerosLikeKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
output
=
context
.
Output
<
framework
::
Tensor
>
(
"Y"
);
...
...
paddle/operators/gather_op.h
浏览文件 @
3a5693e0
...
...
@@ -24,7 +24,7 @@ namespace operators {
using
Tensor
=
framework
::
Tensor
;
template
<
typename
Place
,
typename
T
>
class
GatherOpKernel
:
public
framework
::
OpKernel
{
class
GatherOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
X
=
ctx
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -37,7 +37,7 @@ class GatherOpKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
GatherGradientOpKernel
:
public
framework
::
OpKernel
{
class
GatherGradientOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
Index
=
ctx
.
Input
<
Tensor
>
(
"Index"
);
...
...
paddle/operators/gaussian_random_op.cc
浏览文件 @
3a5693e0
...
...
@@ -16,7 +16,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
T
>
class
CPUGaussianRandomKernel
:
public
framework
::
OpKernel
{
class
CPUGaussianRandomKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
float
mean
=
context
.
Attr
<
float
>
(
"mean"
);
...
...
paddle/operators/gaussian_random_op.cu
浏览文件 @
3a5693e0
...
...
@@ -37,7 +37,7 @@ struct GaussianGenerator {
};
template
<
typename
T
>
class
GPUGaussianRandomKernel
:
public
framework
::
OpKernel
{
class
GPUGaussianRandomKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
tensor
=
context
.
Output
<
framework
::
Tensor
>
(
"Out"
);
...
...
paddle/operators/gemm_conv2d_op.h
浏览文件 @
3a5693e0
...
...
@@ -25,7 +25,7 @@ namespace operators {
using
Tensor
=
framework
::
Tensor
;
template
<
typename
Place
,
typename
T
>
class
GemmConv2DKernel
:
public
framework
::
OpKernel
{
class
GemmConv2DKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
Tensor
*
input
=
context
.
Input
<
Tensor
>
(
"Input"
);
...
...
@@ -98,7 +98,7 @@ class GemmConv2DKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
GemmConvGrad2DKernel
:
public
framework
::
OpKernel
{
class
GemmConvGrad2DKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
Tensor
*
input
=
context
.
Input
<
Tensor
>
(
"Input"
);
...
...
paddle/operators/lookup_table_op.cu
浏览文件 @
3a5693e0
...
...
@@ -61,7 +61,7 @@ __global__ void LookupTableGrad(T* table, const T* output, const int32_t* ids,
}
template
<
typename
T
>
class
LookupTableCUDAKernel
:
public
framework
::
OpKernel
{
class
LookupTableCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
table_t
=
context
.
Input
<
Tensor
>
(
"W"
);
...
...
@@ -85,7 +85,7 @@ class LookupTableCUDAKernel : public framework::OpKernel {
};
template
<
typename
T
>
class
LookupTableGradCUDAKernel
:
public
framework
::
OpKernel
{
class
LookupTableGradCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
ids_t
=
context
.
Input
<
Tensor
>
(
"Ids"
);
...
...
paddle/operators/lookup_table_op.h
浏览文件 @
3a5693e0
...
...
@@ -23,7 +23,7 @@ namespace operators {
using
Tensor
=
framework
::
Tensor
;
template
<
typename
T
>
class
LookupTableKernel
:
public
framework
::
OpKernel
{
class
LookupTableKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
table_t
=
context
.
Input
<
Tensor
>
(
"W"
);
// float tensor
...
...
@@ -44,7 +44,7 @@ class LookupTableKernel : public framework::OpKernel {
};
template
<
typename
T
>
class
LookupTableGradKernel
:
public
framework
::
OpKernel
{
class
LookupTableGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
ids_t
=
context
.
Input
<
Tensor
>
(
"Ids"
);
...
...
paddle/operators/lstm_unit_op.cu
浏览文件 @
3a5693e0
...
...
@@ -90,7 +90,7 @@ __global__ void LSTMUnitGradientKernel(const int nthreads, const int dim,
}
template
<
typename
T
,
typename
AttrType
=
T
>
class
LstmUnitOpCUDAKernel
:
public
framework
::
OpKernel
{
class
LstmUnitOpCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
...
...
@@ -121,7 +121,7 @@ class LstmUnitOpCUDAKernel : public framework::OpKernel {
};
template
<
typename
T
,
typename
AttrType
=
T
>
class
LstmUnitGradOpCUDAKernel
:
public
framework
::
OpKernel
{
class
LstmUnitGradOpCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
...
...
paddle/operators/lstm_unit_op.h
浏览文件 @
3a5693e0
...
...
@@ -33,7 +33,7 @@ inline T tanh(T x) {
}
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
LstmUnitKernel
:
public
framework
::
OpKernel
{
class
LstmUnitKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_cpu_place
(
ctx
.
GetPlace
()),
...
...
@@ -76,7 +76,7 @@ class LstmUnitKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
LstmUnitGradKernel
:
public
framework
::
OpKernel
{
class
LstmUnitGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_cpu_place
(
ctx
.
GetPlace
()),
...
...
paddle/operators/mean_op.h
浏览文件 @
3a5693e0
...
...
@@ -28,7 +28,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenVector
=
framework
::
EigenVector
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
MeanKernel
:
public
framework
::
OpKernel
{
class
MeanKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
input
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -45,7 +45,7 @@ class MeanKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
MeanGradKernel
:
public
framework
::
OpKernel
{
class
MeanGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
OG
=
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
...
...
paddle/operators/minus_op.h
浏览文件 @
3a5693e0
...
...
@@ -20,7 +20,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
MinusKernel
:
public
framework
::
OpKernel
{
class
MinusKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
left_tensor
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
paddle/operators/modified_huber_loss_op.cu
浏览文件 @
3a5693e0
...
...
@@ -39,7 +39,7 @@ struct ModifiedHuberLossBackward {
};
template
<
typename
T
>
class
ModifiedHuberLossGradGPUKernel
:
public
framework
::
OpKernel
{
class
ModifiedHuberLossGradGPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"Y"
);
...
...
paddle/operators/modified_huber_loss_op.h
浏览文件 @
3a5693e0
...
...
@@ -47,7 +47,7 @@ struct ModifiedHuberLossForward {
};
template
<
typename
Place
,
typename
T
>
class
ModifiedHuberLossKernel
:
public
framework
::
OpKernel
{
class
ModifiedHuberLossKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -73,7 +73,7 @@ class ModifiedHuberLossKernel : public framework::OpKernel {
// CPU backward kernel
template
<
typename
T
>
class
ModifiedHuberLossGradCPUKernel
:
public
framework
::
OpKernel
{
class
ModifiedHuberLossGradCPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"Y"
);
...
...
paddle/operators/mul_op.h
浏览文件 @
3a5693e0
...
...
@@ -28,7 +28,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
MulKernel
:
public
framework
::
OpKernel
{
class
MulKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
Tensor
*
x
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -52,7 +52,7 @@ class MulKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
MulGradKernel
:
public
framework
::
OpKernel
{
class
MulGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
int
x_num_col_dims
=
ctx
.
template
Attr
<
int
>(
"x_num_col_dims"
);
...
...
paddle/operators/multiplex_op.cu
浏览文件 @
3a5693e0
...
...
@@ -21,7 +21,7 @@ namespace operators {
using
Tensor
=
framework
::
Tensor
;
template
<
typename
Place
,
typename
T
>
class
MultiplexGPUKernel
:
public
framework
::
OpKernel
{
class
MultiplexGPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
ins
=
ctx
.
MultiInput
<
Tensor
>
(
"X"
);
...
...
@@ -51,7 +51,7 @@ class MultiplexGPUKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
MultiplexGradGPUKernel
:
public
framework
::
OpKernel
{
class
MultiplexGradGPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
*
d_out
=
ctx
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
...
...
paddle/operators/multiplex_op.h
浏览文件 @
3a5693e0
...
...
@@ -23,7 +23,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
MultiplexCPUKernel
:
public
framework
::
OpKernel
{
class
MultiplexCPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
ins
=
ctx
.
MultiInput
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -48,7 +48,7 @@ class MultiplexCPUKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
MultiplexGradCPUKernel
:
public
framework
::
OpKernel
{
class
MultiplexGradCPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
*
d_out
=
ctx
.
Input
<
framework
::
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
...
...
paddle/operators/pad_op.h
浏览文件 @
3a5693e0
...
...
@@ -47,7 +47,7 @@ void PadFunction(const framework::ExecutionContext& context) {
}
template
<
typename
Place
,
typename
T
>
class
PadKernel
:
public
framework
::
OpKernel
{
class
PadKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
int
rank
=
context
.
Input
<
Tensor
>
(
"X"
)
->
dims
().
size
();
...
...
@@ -97,7 +97,7 @@ void PadGradFunction(const framework::ExecutionContext& context) {
}
template
<
typename
Place
,
typename
T
>
class
PadGradKernel
:
public
framework
::
OpKernel
{
class
PadGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
size_t
rank
=
...
...
paddle/operators/prelu_op.h
浏览文件 @
3a5693e0
...
...
@@ -40,7 +40,7 @@ class PReluFunctor {
};
template
<
typename
Place
,
typename
T
>
class
PReluKernel
:
public
framework
::
OpKernel
{
class
PReluKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
x
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -77,7 +77,7 @@ class PReluGradFunctor {
};
template
<
typename
Place
,
typename
T
>
class
PReluGradKernel
:
public
framework
::
OpKernel
{
class
PReluGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
dx
=
context
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
...
...
paddle/operators/rank_loss_op.h
浏览文件 @
3a5693e0
...
...
@@ -21,7 +21,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
RankLossKernel
:
public
framework
::
OpKernel
{
class
RankLossKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
*
out_t
=
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
);
...
...
@@ -42,7 +42,7 @@ class RankLossKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
RankLossGradKernel
:
public
framework
::
OpKernel
{
class
RankLossGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
*
d_left_t
=
...
...
paddle/operators/reshape_op.h
浏览文件 @
3a5693e0
...
...
@@ -21,7 +21,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
ReshapeKernel
:
public
framework
::
OpKernel
{
class
ReshapeKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
*
out
=
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
);
...
...
@@ -39,7 +39,7 @@ class ReshapeKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
ReshapeGradKernel
:
public
framework
::
OpKernel
{
class
ReshapeGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
*
d_out
=
ctx
.
Input
<
framework
::
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
...
...
paddle/operators/rowwise_add_op.h
浏览文件 @
3a5693e0
...
...
@@ -28,7 +28,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
RowwiseAddKernel
:
public
framework
::
OpKernel
{
class
RowwiseAddKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
out
=
context
.
Output
<
Tensor
>
(
"Out"
);
...
...
@@ -50,7 +50,7 @@ class RowwiseAddKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
RowwiseAddGradKernel
:
public
framework
::
OpKernel
{
class
RowwiseAddGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
dout
=
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
...
...
paddle/operators/scale_op.h
浏览文件 @
3a5693e0
...
...
@@ -20,7 +20,7 @@
namespace
paddle
{
namespace
operators
{
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
ScaleKernel
:
public
framework
::
OpKernel
{
class
ScaleKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
virtual
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
{
auto
*
tensor
=
context
.
Output
<
framework
::
Tensor
>
(
"Out"
);
...
...
paddle/operators/scatter_op.h
浏览文件 @
3a5693e0
...
...
@@ -24,7 +24,7 @@ namespace operators {
using
Tensor
=
framework
::
Tensor
;
template
<
typename
Place
,
typename
T
>
class
ScatterOpKernel
:
public
framework
::
OpKernel
{
class
ScatterOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
Ref
=
ctx
.
Input
<
Tensor
>
(
"Ref"
);
...
...
@@ -40,7 +40,7 @@ class ScatterOpKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
ScatterGradientOpKernel
:
public
framework
::
OpKernel
{
class
ScatterGradientOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
dRef
=
ctx
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"Ref"
));
...
...
paddle/operators/sequence_pool_op.h
浏览文件 @
3a5693e0
...
...
@@ -38,7 +38,7 @@ enum SeqPoolType {
};
template
<
typename
Place
,
typename
T
>
class
SequencePoolKernel
:
public
framework
::
OpKernel
{
class
SequencePoolKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in
=
context
.
Input
<
LoDTensor
>
(
"X"
);
...
...
@@ -85,7 +85,7 @@ class SequencePoolKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
SequencePoolGradKernel
:
public
framework
::
OpKernel
{
class
SequencePoolGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in
=
context
.
Input
<
LoDTensor
>
(
"X"
);
...
...
paddle/operators/sgd_op.h
浏览文件 @
3a5693e0
...
...
@@ -25,7 +25,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenVector
=
framework
::
EigenVector
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
SGDOpKernel
:
public
framework
::
OpKernel
{
class
SGDOpKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
param
=
ctx
.
Input
<
Tensor
>
(
"param"
);
...
...
paddle/operators/smooth_l1_loss_op.h
浏览文件 @
3a5693e0
...
...
@@ -45,7 +45,7 @@ struct SmoothL1LossForward {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
SmoothL1LossKernel
:
public
framework
::
OpKernel
{
class
SmoothL1LossKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -115,7 +115,7 @@ struct SmoothL1LossBackward {
};
template
<
typename
Place
,
typename
T
,
typename
AttrType
=
T
>
class
SmoothL1LossGradKernel
:
public
framework
::
OpKernel
{
class
SmoothL1LossGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"InsideWeight"
);
...
...
paddle/operators/softmax_op.h
浏览文件 @
3a5693e0
...
...
@@ -26,7 +26,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
SoftmaxKernel
:
public
framework
::
OpKernel
{
class
SoftmaxKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
X
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -40,7 +40,7 @@ class SoftmaxKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
SoftmaxGradKernel
:
public
framework
::
OpKernel
{
class
SoftmaxGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
Y
=
context
.
Input
<
Tensor
>
(
"Y"
);
...
...
paddle/operators/softmax_with_cross_entropy_op.cu
浏览文件 @
3a5693e0
...
...
@@ -53,7 +53,7 @@ __global__ void SoftCrossEntropyGradientKernel(T* logit_grad,
}
// namespace
template
<
typename
T
>
class
SoftmaxWithCrossEntropyCUDAKernel
:
public
framework
::
OpKernel
{
class
SoftmaxWithCrossEntropyCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
context
.
GetPlace
()),
...
...
@@ -73,7 +73,7 @@ class SoftmaxWithCrossEntropyCUDAKernel : public framework::OpKernel {
};
template
<
typename
T
>
class
SoftmaxWithCrossEntropyGradCUDAKernel
:
public
framework
::
OpKernel
{
class
SoftmaxWithCrossEntropyGradCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
context
.
GetPlace
()),
...
...
paddle/operators/softmax_with_cross_entropy_op.h
浏览文件 @
3a5693e0
...
...
@@ -27,7 +27,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
T
>
class
SoftmaxWithCrossEntropyKernel
:
public
framework
::
OpKernel
{
class
SoftmaxWithCrossEntropyKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_cpu_place
(
context
.
GetPlace
()),
...
...
@@ -47,7 +47,7 @@ class SoftmaxWithCrossEntropyKernel : public framework::OpKernel {
};
template
<
typename
T
>
class
SoftmaxWithCrossEntropyGradKernel
:
public
framework
::
OpKernel
{
class
SoftmaxWithCrossEntropyGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
const
Tensor
*
out_grad
=
...
...
paddle/operators/split_op.h
浏览文件 @
3a5693e0
...
...
@@ -21,7 +21,7 @@ namespace paddle {
namespace
operators
{
template
<
typename
Place
,
typename
T
>
class
SplitKernel
:
public
framework
::
OpKernel
{
class
SplitKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
in
=
ctx
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
paddle/operators/squared_l2_distance_op.h
浏览文件 @
3a5693e0
...
...
@@ -28,7 +28,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
SquaredL2DistanceKernel
:
public
framework
::
OpKernel
{
class
SquaredL2DistanceKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"X"
);
...
...
@@ -68,7 +68,7 @@ class SquaredL2DistanceKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
SquaredL2DistanceGradKernel
:
public
framework
::
OpKernel
{
class
SquaredL2DistanceGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
in0
=
context
.
Input
<
Tensor
>
(
"sub_result"
);
...
...
paddle/operators/sum_op.h
浏览文件 @
3a5693e0
...
...
@@ -22,7 +22,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenVector
=
framework
::
EigenVector
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
SumKernel
:
public
framework
::
OpKernel
{
class
SumKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
ins
=
context
.
MultiInput
<
Tensor
>
(
"X"
);
...
...
@@ -43,7 +43,7 @@ class SumKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
SumGradKernel
:
public
framework
::
OpKernel
{
class
SumGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
input
=
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
...
...
paddle/operators/top_k_op.cu
浏览文件 @
3a5693e0
...
...
@@ -279,7 +279,7 @@ __global__ void KeMatrixTopK(T* output, int output_stride, int* indices,
}
template
<
typename
T
>
class
TopkOpCUDAKernel
:
public
framework
::
OpKernel
{
class
TopkOpCUDAKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
...
...
paddle/operators/top_k_op.h
浏览文件 @
3a5693e0
...
...
@@ -28,7 +28,7 @@ template <typename T, int MajorType = Eigen::RowMajor,
using
EigenMatrix
=
framework
::
EigenMatrix
<
T
,
MajorType
,
IndexType
>
;
template
<
typename
Place
,
typename
T
>
class
TopkKernel
:
public
framework
::
OpKernel
{
class
TopkKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
// Get the top k elements of each row of input tensor
...
...
paddle/operators/transpose_op.h
浏览文件 @
3a5693e0
...
...
@@ -38,7 +38,7 @@ void EigenTranspose(const framework::ExecutionContext& context,
}
template
<
typename
Place
,
typename
T
>
class
TransposeKernel
:
public
framework
::
OpKernel
{
class
TransposeKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
x
=
context
.
Input
<
framework
::
Tensor
>
(
"X"
);
...
...
@@ -73,7 +73,7 @@ class TransposeKernel : public framework::OpKernel {
};
template
<
typename
Place
,
typename
T
>
class
TransposeGradKernel
:
public
framework
::
OpKernel
{
class
TransposeGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
out_grad
=
...
...
paddle/operators/uniform_random_op.cc
浏览文件 @
3a5693e0
...
...
@@ -21,7 +21,7 @@ namespace operators {
// Use std::random and thrust::random(thrust is a std library in CUDA) to
// implement uniform random.
template
<
typename
T
>
class
CPUUniformRandomKernel
:
public
framework
::
OpKernel
{
class
CPUUniformRandomKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
*
tensor
=
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
);
...
...
paddle/operators/uniform_random_op.cu
浏览文件 @
3a5693e0
...
...
@@ -40,7 +40,7 @@ struct UniformGenerator {
// Use std::random and thrust::random(thrust is a std library in CUDA) to
// implement uniform random.
template
<
typename
T
>
class
GPUUniformRandomKernel
:
public
framework
::
OpKernel
{
class
GPUUniformRandomKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
auto
*
tensor
=
context
.
Output
<
framework
::
Tensor
>
(
"Out"
);
...
...
paddle/platform/place.cc
浏览文件 @
3a5693e0
...
...
@@ -47,7 +47,7 @@ bool is_cpu_place(const Place &p) {
}
bool
places_are_same_class
(
const
Place
&
p1
,
const
Place
&
p2
)
{
return
is_gpu_place
(
p1
)
==
is_gpu_place
(
p2
);
return
p1
.
which
()
==
p2
.
which
(
);
}
std
::
ostream
&
operator
<<
(
std
::
ostream
&
os
,
const
Place
&
p
)
{
...
...
paddle/pybind/tensor_py.h
浏览文件 @
3a5693e0
...
...
@@ -42,7 +42,7 @@ template <size_t I, typename... ARGS>
struct
CastToPyBufferImpl
<
true
,
I
,
ARGS
...
>
{
using
CUR_TYPE
=
typename
std
::
tuple_element
<
I
,
std
::
tuple
<
ARGS
...
>>::
type
;
py
::
buffer_info
operator
()(
framework
::
Tensor
&
tensor
)
{
if
(
std
::
type_index
(
typeid
(
CUR_TYPE
))
==
tensor
.
holder_
->
type
())
{
if
(
std
::
type_index
(
typeid
(
CUR_TYPE
))
==
tensor
.
type
())
{
auto
dim_vec
=
framework
::
vectorize
(
tensor
.
dims
());
std
::
vector
<
size_t
>
dims_outside
;
std
::
vector
<
size_t
>
strides
;
...
...
@@ -56,13 +56,13 @@ struct CastToPyBufferImpl<true, I, ARGS...> {
prod
*=
dims_outside
[
i
-
1
];
}
framework
::
Tensor
dst_tensor
;
if
(
paddle
::
platform
::
is_gpu_place
(
tensor
.
holder_
->
place
()))
{
if
(
paddle
::
platform
::
is_gpu_place
(
tensor
.
place
()))
{
dst_tensor
.
CopyFrom
<
CUR_TYPE
>
(
tensor
,
platform
::
CPUPlace
());
}
else
if
(
paddle
::
platform
::
is_cpu_place
(
tensor
.
holder_
->
place
()))
{
}
else
if
(
paddle
::
platform
::
is_cpu_place
(
tensor
.
place
()))
{
dst_tensor
=
tensor
;
}
return
py
::
buffer_info
(
dst_tensor
.
mutable_data
<
CUR_TYPE
>
(
dst_tensor
.
holder_
->
place
()),
dst_tensor
.
mutable_data
<
CUR_TYPE
>
(
dst_tensor
.
place
()),
sizeof
(
CUR_TYPE
),
py
::
format_descriptor
<
CUR_TYPE
>::
format
(),
(
size_t
)
framework
::
arity
(
dst_tensor
.
dims
()),
dims_outside
,
strides
);
}
else
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录