Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
a57a19ea
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
a57a19ea
编写于
11月 02, 2022
作者:
H
HongyuJia
提交者:
GitHub
11月 02, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Revert "[Kernel Selection] Remove hard code of PADDLE_WITH_CUDA (#47325)" (#47582)
This reverts commit
f9134045
.
上级
c061c082
变更
14
隐藏空白更改
内联
并排
Showing
14 changed file
with
124 addition
and
75 deletion
+124
-75
paddle/fluid/framework/new_executor/interpreter/data_transfer.cc
...fluid/framework/new_executor/interpreter/data_transfer.cc
+0
-9
paddle/fluid/framework/new_executor/interpreter/interpreter_util.cc
...id/framework/new_executor/interpreter/interpreter_util.cc
+0
-9
paddle/fluid/framework/operator.cc
paddle/fluid/framework/operator.cc
+0
-24
paddle/fluid/imperative/execution_context.h
paddle/fluid/imperative/execution_context.h
+1
-2
paddle/fluid/imperative/prepared_operator.cc
paddle/fluid/imperative/prepared_operator.cc
+0
-9
paddle/fluid/operators/activation_op.cc
paddle/fluid/operators/activation_op.cc
+0
-8
paddle/fluid/operators/affine_grid_op.cc
paddle/fluid/operators/affine_grid_op.cc
+19
-4
paddle/fluid/operators/conv_transpose_op.cc
paddle/fluid/operators/conv_transpose_op.cc
+27
-0
paddle/fluid/operators/grid_sampler_op.cc
paddle/fluid/operators/grid_sampler_op.cc
+22
-4
paddle/fluid/operators/pool_op.cc
paddle/fluid/operators/pool_op.cc
+19
-2
paddle/fluid/operators/sequence_ops/sequence_softmax_op.cc
paddle/fluid/operators/sequence_ops/sequence_softmax_op.cc
+16
-0
paddle/fluid/operators/softmax_op.cc
paddle/fluid/operators/softmax_op.cc
+16
-0
paddle/fluid/platform/device/gpu/cuda/cudnn_helper.h
paddle/fluid/platform/device/gpu/cuda/cudnn_helper.h
+2
-2
paddle/fluid/platform/device/gpu/rocm/miopen_helper.h
paddle/fluid/platform/device/gpu/rocm/miopen_helper.h
+2
-2
未找到文件。
paddle/fluid/framework/new_executor/interpreter/data_transfer.cc
浏览文件 @
a57a19ea
...
...
@@ -22,9 +22,6 @@
#ifdef PADDLE_WITH_MKLDNN
#include "paddle/phi/backends/onednn/onednn_context.h"
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_dnn.h"
#endif
namespace
paddle
{
namespace
framework
{
...
...
@@ -136,12 +133,6 @@ void DataTranferHelper::RunAndConstructOpFuncNode(
auto
*
dev_ctx
=
pool
.
Get
(
place_
);
auto
exec_ctx
=
ExecutionContext
(
*
op
,
Scope
(),
*
dev_ctx
,
runtime_context
);
auto
expected_kernel_key
=
op_with_kernel
->
GetExpectedKernelType
(
exec_ctx
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
!
op_with_kernel
->
DnnFallback
()
&&
paddle
::
platform
::
CanCUDNNBeUsed
(
exec_ctx
))
{
expected_kernel_key
.
library_type_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
VLOG
(
6
)
<<
"expected_kernel_key "
<<
expected_kernel_key
<<
"
\n
"
;
VLOG
(
6
)
<<
"op_with_kernel Type() "
<<
op_with_kernel
->
Type
()
<<
"
\n
"
;
...
...
paddle/fluid/framework/new_executor/interpreter/interpreter_util.cc
浏览文件 @
a57a19ea
...
...
@@ -32,9 +32,6 @@
#ifdef PADDLE_WITH_MKLDNN
#include "paddle/fluid/platform/mkldnn_helper.h"
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_dnn.h"
#endif
PADDLE_DEFINE_EXPORTED_bool
(
new_executor_serial_run
,
...
...
@@ -621,12 +618,6 @@ void BuildOpFuncList(const platform::Place& place,
*
op_with_kernel
,
*
runtime_scope
,
*
dev_ctx
,
runtime_context
);
auto
expected_kernel_key
=
op_with_kernel
->
GetExpectedKernelType
(
exec_ctx
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
!
op_with_kernel
->
DnnFallback
()
&&
paddle
::
platform
::
CanCUDNNBeUsed
(
exec_ctx
))
{
expected_kernel_key
.
library_type_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
VLOG
(
4
)
<<
"expected_kernel_key : "
<<
expected_kernel_key
;
// change device by the device_guard()
ApplyDeviceGuard
(
op
,
place
,
&
expected_kernel_key
);
...
...
paddle/fluid/framework/operator.cc
浏览文件 @
a57a19ea
...
...
@@ -58,10 +58,6 @@ class DenseTensor;
#include "paddle/fluid/platform/device/mlu/mlu_info.h"
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_dnn.h"
#endif
DECLARE_bool
(
benchmark
);
DECLARE_bool
(
check_nan_inf
);
DECLARE_bool
(
enable_unused_var_check
);
...
...
@@ -1413,14 +1409,6 @@ bool OperatorWithKernel::SupportsKernelType(
}
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
!
this
->
DnnFallback
()
&&
paddle
::
platform
::
CanCUDNNBeUsed
(
exe_ctx
))
{
auto
tmp_kernel_type
=
kernel_type
;
tmp_kernel_type
.
library_type_
=
framework
::
LibraryType
::
kCUDNN
;
return
kernels
.
find
(
tmp_kernel_type
)
!=
kernels
.
end
();
}
#endif
return
kernel_iter
!=
kernels
.
end
();
}
...
...
@@ -1601,12 +1589,6 @@ void OperatorWithKernel::RunImpl(const Scope& scope,
}
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
!
this
->
DnnFallback
()
&&
paddle
::
platform
::
CanCUDNNBeUsed
(
exe_ctx
))
{
kernel_type_
->
library_type_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
// NOTE(Liu-xiandong):In my ctest, this branch do not be executed,
// I can't understand it, it's really confusing.
// But we still need to keep this to avoid errors.
...
...
@@ -1850,12 +1832,6 @@ OpKernelType OperatorWithKernel::InnerGetExpectedKernelType(
}
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
!
this
->
DnnFallback
()
&&
paddle
::
platform
::
CanCUDNNBeUsed
(
ctx
))
{
expected_kernel_key
.
library_type_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
if
(
HasAttr
(
"op_device"
))
{
if
(
Attr
<
std
::
string
>
(
"op_device"
)
==
"cpu"
)
{
expected_kernel_key
.
place_
=
platform
::
CPUPlace
();
...
...
paddle/fluid/imperative/execution_context.h
浏览文件 @
a57a19ea
...
...
@@ -103,8 +103,7 @@ class DygraphExecutionContext : public framework::ExecutionContext {
bool
HasAttr
(
const
std
::
string
&
name
)
const
override
{
if
(
attrs_
.
find
(
name
)
==
attrs_
.
end
())
{
return
&
default_attrs_
!=
nullptr
&&
default_attrs_
.
find
(
name
)
!=
default_attrs_
.
end
();
return
default_attrs_
.
find
(
name
)
!=
default_attrs_
.
end
();
}
return
true
;
}
...
...
paddle/fluid/imperative/prepared_operator.cc
浏览文件 @
a57a19ea
...
...
@@ -28,9 +28,6 @@
#ifdef PADDLE_WITH_MKLDNN
#include "paddle/fluid/platform/mkldnn_op_list.h"
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_dnn.h"
#endif
#include "paddle/fluid/framework/library_type.h"
#include "paddle/fluid/platform/device/gpu/gpu_info.h"
#include "paddle/fluid/platform/profiler/event_tracing.h"
...
...
@@ -249,12 +246,6 @@ PreparedOp PrepareImpl(
}
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
!
op
.
DnnFallback
()
&&
paddle
::
platform
::
CanCUDNNBeUsed
(
dygraph_exe_ctx
))
{
expected_kernel_key
.
library_type_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
#if defined(PADDLE_WITH_XPU)
bool
is_xpu_unsupport
=
paddle
::
platform
::
is_xpu_place
(
expected_kernel_key
.
place_
)
&&
...
...
paddle/fluid/operators/activation_op.cc
浏览文件 @
a57a19ea
...
...
@@ -93,14 +93,6 @@ framework::OpKernelType GetKernelType(const framework::ExecutionContext& ctx,
// library = framework::LibraryType::kCUDNN;
// }
// #endif
// NOTE(jiahongyu): Activation ops have attribute use_cudnn, but cudnn kernels
// are temporarily disabled. Therefore, cudnn kernel also needs to fallback to
// plain GPU kernel temporarily. When above codes are uncommented, below
// fallback codes can be deleted safely.
if
(
paddle
::
platform
::
is_gpu_place
(
ctx
.
GetPlace
()))
{
oper
.
SetDnnFallback
(
true
);
}
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
}
...
...
paddle/fluid/operators/affine_grid_op.cc
浏览文件 @
a57a19ea
...
...
@@ -134,8 +134,15 @@ class AffineGridOp : public framework::OperatorWithKernel {
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
framework
::
LibraryType
library
{
framework
::
LibraryType
::
kPlain
};
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
library
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"Theta"
);
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
library
);
}
};
...
...
@@ -245,9 +252,17 @@ class AffineGridOpGrad : public framework::OperatorWithKernel {
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
framework
::
GradVarName
(
"Output"
));
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
framework
::
LibraryType
library_
{
framework
::
LibraryType
::
kPlain
};
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
library_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
return
framework
::
OpKernelType
(
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
framework
::
GradVarName
(
"Output"
)),
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
library_
);
}
};
...
...
paddle/fluid/operators/conv_transpose_op.cc
浏览文件 @
a57a19ea
...
...
@@ -28,6 +28,9 @@ limitations under the License. */
#ifdef PADDLE_WITH_MKLDNN
#include "paddle/fluid/platform/mkldnn_helper.h"
#endif
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
#include "paddle/fluid/platform/device/gpu/gpu_dnn.h"
#endif
namespace
paddle
{
namespace
operators
{
...
...
@@ -37,6 +40,14 @@ using DataLayout = phi::DataLayout;
framework
::
OpKernelType
ConvTransposeOp
::
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"Input"
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
}
...
...
@@ -257,6 +268,14 @@ Example:
framework
::
OpKernelType
ConvTransposeOpGrad
::
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"Input"
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
}
...
...
@@ -324,6 +343,14 @@ class ConvTransposeDoubleGradMaker : public framework::SingleGradOpMaker<T> {
framework
::
OpKernelType
ConvTransposeOpDoubleGrad
::
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"Input"
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
}
...
...
paddle/fluid/operators/grid_sampler_op.cc
浏览文件 @
a57a19ea
...
...
@@ -35,8 +35,17 @@ class GridSampleOp : public framework::OperatorWithKernel {
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
);
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
framework
::
LibraryType
library_
{
framework
::
LibraryType
::
kPlain
};
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
library_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
return
framework
::
OpKernelType
(
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
),
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
library_
);
}
};
...
...
@@ -137,8 +146,17 @@ class GridSampleOpGrad : public framework::OperatorWithKernel {
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
);
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
framework
::
LibraryType
library_
{
framework
::
LibraryType
::
kPlain
};
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
library_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
return
framework
::
OpKernelType
(
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
),
ctx
.
GetPlace
(),
phi
::
DataLayout
::
kAnyLayout
,
library_
);
}
};
...
...
paddle/fluid/operators/pool_op.cc
浏览文件 @
a57a19ea
...
...
@@ -44,13 +44,21 @@ bool CanMKLDNNSupportPool(const framework::ExecutionContext& ctx) {
framework
::
OpKernelType
PoolOp
::
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
framework
::
LibraryType
library_
{
framework
::
LibraryType
::
kPlain
};
phi
::
DataLayout
layout_
=
phi
::
DataLayout
::
kAnyLayout
;
auto
data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
library_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
// NOTE(jiahongyu): Below codes originally enclosed by PADDLE_WITH_MKLDNN
this
->
SetDnnFallback
(
!
CanMKLDNNSupportPool
(
ctx
));
// NOTE(jiahongyu) END: Above codes originally enclosed by PADDLE_WITH_MKLDNN
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
());
return
framework
::
OpKernelType
(
data_type
,
ctx
.
GetPlace
()
,
layout_
,
library_
);
}
framework
::
OpKernelType
PoolOp
::
GetKernelTypeForVar
(
...
...
@@ -78,13 +86,22 @@ framework::OpKernelType PoolOp::GetKernelTypeForVar(
framework
::
OpKernelType
PoolOpGrad
::
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
{
framework
::
LibraryType
library_
{
framework
::
LibraryType
::
kPlain
};
phi
::
DataLayout
layout_
=
phi
::
DataLayout
::
kAnyLayout
;
auto
input_data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
);
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
library_
=
framework
::
LibraryType
::
kCUDNN
;
}
#endif
// NOTE(jiahongyu): Below codes originally enclosed by PADDLE_WITH_MKLDNN
this
->
SetDnnFallback
(
!
CanMKLDNNSupportPool
(
ctx
));
// NOTE(jiahongyu): Above codes originally enclosed by PADDLE_WITH_MKLDNN
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
());
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
library_
);
}
framework
::
OpKernelType
PoolOpGrad
::
GetKernelTypeForVar
(
...
...
paddle/fluid/operators/sequence_ops/sequence_softmax_op.cc
浏览文件 @
a57a19ea
...
...
@@ -43,6 +43,14 @@ class SequenceSoftmaxOp : public framework::OperatorWithKernel {
if
(
ctx
.
HasAttr
(
"data_format"
))
{
layout_
=
phi
::
StringToDataLayout
(
ctx
.
Attr
<
std
::
string
>
(
"data_format"
));
}
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
);
}
};
...
...
@@ -127,6 +135,14 @@ class SequenceSoftmaxGradOp : public framework::OperatorWithKernel {
if
(
ctx
.
HasAttr
(
"data_format"
))
{
layout_
=
phi
::
StringToDataLayout
(
ctx
.
Attr
<
std
::
string
>
(
"data_format"
));
}
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
);
}
};
...
...
paddle/fluid/operators/softmax_op.cc
浏览文件 @
a57a19ea
...
...
@@ -48,6 +48,14 @@ class SoftmaxOp : public framework::OperatorWithKernel {
platform
::
errors
::
InvalidArgument
(
"float16 can only be used on GPU/NPU/XPU/MLU and custom place"
));
}
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
);
}
};
...
...
@@ -132,6 +140,14 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
PADDLE_THROW
(
platform
::
errors
::
InvalidArgument
(
"float16 can only be used on GPU/NPU/XPU/MLU and custom place"
));
}
#if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)
if
(
platform
::
CanCUDNNBeUsed
(
ctx
))
{
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
framework
::
LibraryType
::
kCUDNN
);
}
#endif
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
);
}
};
...
...
paddle/fluid/platform/device/gpu/cuda/cudnn_helper.h
浏览文件 @
a57a19ea
...
...
@@ -617,8 +617,8 @@ class ScopedActivationDescriptor {
};
inline
bool
CanCUDNNBeUsed
(
const
framework
::
ExecutionContext
&
ctx
)
{
bool
use_cudnn
=
paddle
::
platform
::
is_gpu_place
(
ctx
.
GetPlace
())
&&
ctx
.
HasAttr
(
"use_cudnn"
)
&&
ctx
.
Attr
<
bool
>
(
"use_cudnn"
);
bool
use_cudnn
=
ctx
.
HasAttr
(
"use_cudnn"
)
&&
ctx
.
Attr
<
bool
>
(
"use_cudnn"
);
use_cudnn
&=
paddle
::
platform
::
is_gpu_place
(
ctx
.
GetPlace
()
);
#ifdef PADDLE_WITH_CUDA
if
(
use_cudnn
)
{
auto
&
dev_ctx
=
ctx
.
device_context
<
phi
::
GPUContext
>
();
...
...
paddle/fluid/platform/device/gpu/rocm/miopen_helper.h
浏览文件 @
a57a19ea
...
...
@@ -554,8 +554,8 @@ class ScopedActivationDescriptor {
};
inline
bool
CanCUDNNBeUsed
(
const
framework
::
ExecutionContext
&
ctx
)
{
bool
use_cudnn
=
paddle
::
platform
::
is_gpu_place
(
ctx
.
GetPlace
())
&&
ctx
.
HasAttr
(
"use_cudnn"
)
&&
ctx
.
Attr
<
bool
>
(
"use_cudnn"
);
bool
use_cudnn
=
ctx
.
HasAttr
(
"use_cudnn"
)
&&
ctx
.
Attr
<
bool
>
(
"use_cudnn"
);
use_cudnn
&=
paddle
::
platform
::
is_gpu_place
(
ctx
.
GetPlace
()
);
#ifdef PADDLE_WITH_HIP
if
(
use_cudnn
)
{
auto
&
dev_ctx
=
ctx
.
device_context
<
phi
::
GPUContext
>
();
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录