Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
cad139a7
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
cad139a7
编写于
6月 08, 2022
作者:
X
xiaoxiaohehe001
提交者:
GitHub
6月 08, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
call_once (#43206)
上级
8dab2690
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
82 addition
and
40 deletion
+82
-40
paddle/phi/backends/gpu/gpu_context.cc
paddle/phi/backends/gpu/gpu_context.cc
+82
-40
未找到文件。
paddle/phi/backends/gpu/gpu_context.cc
浏览文件 @
cad139a7
...
...
@@ -214,23 +214,6 @@ struct GPUContext::Impl {
&
max_grid_dim_size_
);
phi
::
InitStream
(
&
stream_
);
InitEigenDevice
();
phi
::
InitBlasHandle
(
&
blas_handle_
,
stream_
);
#ifdef PADDLE_WITH_CUDA
#if CUDA_VERSION >= 9000
phi
::
InitBlasHandle
(
&
blas_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tensor_core_handle_
,
CUBLAS_TENSOR_OP_MATH
));
#endif
#if CUDA_VERSION >= 11000
phi
::
InitBlasHandle
(
&
blas_tf32_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tf32_tensor_core_handle_
,
CUBLAS_TF32_TENSOR_OP_MATH
));
#endif
#endif
phi
::
InitBlasLtHandle
(
&
blaslt_handle_
);
phi
::
InitDnnHandle
(
&
dnn_handle_
,
stream_
,
place_
);
phi
::
InitSolverHandle
(
&
solver_handle_
,
stream_
);
phi
::
InitSparseHandle
(
&
sparse_handle_
,
stream_
);
InitDnnWorkspace
();
}
...
...
@@ -246,23 +229,6 @@ struct GPUContext::Impl {
&
max_threads_per_block_
,
&
max_grid_dim_size_
);
phi
::
InitStream
(
&
stream_
);
phi
::
InitBlasHandle
(
&
blas_handle_
,
stream_
);
#ifdef PADDLE_WITH_CUDA
#if CUDA_VERSION >= 9000
phi
::
InitBlasHandle
(
&
blas_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tensor_core_handle_
,
CUBLAS_TENSOR_OP_MATH
));
#endif
#if CUDA_VERSION >= 11000
phi
::
InitBlasHandle
(
&
blas_tf32_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tf32_tensor_core_handle_
,
CUBLAS_TF32_TENSOR_OP_MATH
));
#endif
#endif
phi
::
InitBlasLtHandle
(
&
blaslt_handle_
);
phi
::
InitDnnHandle
(
&
dnn_handle_
,
stream_
,
place_
);
phi
::
InitSolverHandle
(
&
solver_handle_
,
stream_
);
phi
::
InitSparseHandle
(
&
sparse_handle_
,
stream_
);
}
void
PartialInitWithAllocator
()
{
...
...
@@ -356,7 +322,28 @@ struct GPUContext::Impl {
return
eigen_device_
;
}
blasHandle_t
GetBlasHandle
()
const
{
blasHandle_t
GetBlasHandle
()
{
std
::
call_once
(
flag_blas_
,
[
=
]()
{
if
(
!
blas_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_handle_
,
stream_
);
}
#ifdef PADDLE_WITH_CUDA
#if CUDA_VERSION >= 9000
if
(
!
blas_tensor_core_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tensor_core_handle_
,
CUBLAS_TENSOR_OP_MATH
));
}
#endif
#if CUDA_VERSION >= 11000
if
(
!
blas_tf32_tensor_core_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_tf32_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tf32_tensor_core_handle_
,
CUBLAS_TF32_TENSOR_OP_MATH
));
}
#endif
#endif
});
PD_CHECK
(
blas_handle_
!=
nullptr
,
"the gpu blas handle is nullptr."
);
return
blas_handle_
;
}
...
...
@@ -373,12 +360,18 @@ struct GPUContext::Impl {
void
SetBlasLtHandle
(
blasLtHandle_t
blaslt
)
{
blaslt_handle_
=
blaslt
;
}
blasLtHandle_t
GetBlasLtHandle
()
const
{
blasLtHandle_t
GetBlasLtHandle
()
{
std
::
call_once
(
flag_blaslt_
,
[
=
]()
{
if
(
!
blaslt_handle_
)
phi
::
InitBlasLtHandle
(
&
blaslt_handle_
);
});
PD_CHECK
(
blaslt_handle_
!=
nullptr
,
"the gpu blasLt handle is nullptr."
);
return
blaslt_handle_
;
}
dnnHandle_t
GetDnnHandle
()
{
std
::
call_once
(
flag_dnn_
,
[
=
]()
{
if
(
!
dnn_handle_
)
phi
::
InitDnnHandle
(
&
dnn_handle_
,
stream_
,
place_
);
});
PD_CHECK
(
dnn_handle_
!=
nullptr
,
"the gpu dnn handle is nullptr."
);
return
dnn_handle_
;
}
...
...
@@ -399,7 +392,10 @@ struct GPUContext::Impl {
void
SetDnnHandle
(
dnnHandle_t
handle
)
{
dnn_handle_
=
handle
;
}
solverHandle_t
GetSolverHandle
()
const
{
solverHandle_t
GetSolverHandle
()
{
std
::
call_once
(
flag_slover_
,
[
=
]()
{
if
(
!
solver_handle_
)
phi
::
InitSolverHandle
(
&
solver_handle_
,
stream_
);
});
PD_CHECK
(
solver_handle_
!=
nullptr
,
"the gpu solver handle is nullptr."
);
return
solver_handle_
;
}
...
...
@@ -461,8 +457,28 @@ struct GPUContext::Impl {
#endif
}
inline
void
CublasCall
(
const
std
::
function
<
void
(
blasHandle_t
)
>&
callback
)
const
{
inline
void
CublasCall
(
const
std
::
function
<
void
(
blasHandle_t
)
>&
callback
)
{
std
::
call_once
(
flag_cublas_
,
[
=
]()
{
if
(
!
blas_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_handle_
,
stream_
);
}
#ifdef PADDLE_WITH_CUDA
#if CUDA_VERSION >= 9000
if
(
!
blas_tensor_core_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tensor_core_handle_
,
CUBLAS_TENSOR_OP_MATH
));
}
#endif
#if CUDA_VERSION >= 11000
if
(
!
blas_tf32_tensor_core_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_tf32_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tf32_tensor_core_handle_
,
CUBLAS_TF32_TENSOR_OP_MATH
));
}
#endif
#endif
});
if
(
blas_tf32_tensor_core_handle_
!=
nullptr
)
{
std
::
lock_guard
<
std
::
mutex
>
guard
(
blas_tf32_mtx_
);
callback
(
blas_tf32_tensor_core_handle_
);
...
...
@@ -473,7 +489,26 @@ struct GPUContext::Impl {
}
inline
void
TensorCoreCublasCallIfAvailable
(
const
std
::
function
<
void
(
blasHandle_t
)
>&
callback
)
const
{
const
std
::
function
<
void
(
blasHandle_t
)
>&
callback
)
{
std
::
call_once
(
flag_tensorcore_cublas_
,
[
=
]()
{
if
(
!
blas_handle_
)
phi
::
InitBlasHandle
(
&
blas_handle_
,
stream_
);
#ifdef PADDLE_WITH_CUDA
#if CUDA_VERSION >= 9000
if
(
!
blas_tensor_core_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tensor_core_handle_
,
CUBLAS_TENSOR_OP_MATH
));
}
#endif
#if CUDA_VERSION >= 11000
if
(
!
blas_tf32_tensor_core_handle_
)
{
phi
::
InitBlasHandle
(
&
blas_tf32_tensor_core_handle_
,
stream_
);
PADDLE_RETRY_CUDA_SUCCESS
(
phi
::
dynload
::
cublasSetMathMode
(
blas_tf32_tensor_core_handle_
,
CUBLAS_TF32_TENSOR_OP_MATH
));
}
#endif
#endif
});
if
(
blas_tensor_core_handle_
!=
nullptr
)
{
std
::
lock_guard
<
std
::
mutex
>
guard
(
blas_tensor_core_mtx_
);
callback
(
blas_tensor_core_handle_
);
...
...
@@ -563,6 +598,13 @@ struct GPUContext::Impl {
sparseHandle_t
sparse_handle_
{
nullptr
};
DnnWorkspaceHandle
*
workspace_
{
nullptr
};
std
::
once_flag
flag_blas_
;
std
::
once_flag
flag_blaslt_
;
std
::
once_flag
flag_dnn_
;
std
::
once_flag
flag_slover_
;
std
::
once_flag
flag_cublas_
;
std
::
once_flag
flag_tensorcore_cublas_
;
#if defined(PADDLE_WITH_NCCL) || defined(PADDLE_WITH_RCCL)
// NCCL communicator (single process version) for NCCL collective operations.
// NCCL collective operations provides fast collectives over multiple GPUs
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录