Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
MegEngine 天元
MegEngine
提交
3a35827d
MegEngine
项目概览
MegEngine 天元
/
MegEngine
10 个月 前同步成功
通知
392
Star
4702
Fork
582
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
MegEngine
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
3a35827d
编写于
2月 15, 2023
作者:
M
Megvii Engine Team
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
feat(cuda): upgrade x86_64-cuda to 11.8
GitOrigin-RevId: 21f3ee70312eaa39454e30e7e698fce53ae7ac53
上级
0612b573
变更
6
展开全部
隐藏空白更改
内联
并排
Showing
6 changed file
with
8287 addition
and
5 deletion
+8287
-5
CMakeLists.txt
CMakeLists.txt
+10
-0
dnn/cuda-stub/src/libcuda-wrap_11.8.h
dnn/cuda-stub/src/libcuda-wrap_11.8.h
+8230
-0
dnn/cuda-stub/src/libcuda.cpp
dnn/cuda-stub/src/libcuda.cpp
+2
-0
dnn/src/cuda/kernel_common/diagnostic_prologue.cuh
dnn/src/cuda/kernel_common/diagnostic_prologue.cuh
+1
-0
scripts/whl/manylinux2014/build_wheel_common.sh
scripts/whl/manylinux2014/build_wheel_common.sh
+41
-3
src/tensorrt/include/megbrain/tensorrt/tensorrt_opr.h
src/tensorrt/include/megbrain/tensorrt/tensorrt_opr.h
+3
-2
未找到文件。
CMakeLists.txt
浏览文件 @
3a35827d
...
...
@@ -767,6 +767,16 @@ if(MGE_WITH_CUDA)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_61,code=sm_61"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_70,code=sm_70"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_75,code=sm_75"
)
elseif
(
${
CMAKE_CUDA_COMPILER_VERSION
}
VERSION_GREATER
"11.8.0"
OR
${
CMAKE_CUDA_COMPILER_VERSION
}
VERSION_EQUAL
"11.8.0"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_61,code=sm_61"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_70,code=sm_70"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_75,code=sm_75"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_80,code=sm_80"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_86,code=sm_86"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_89,code=sm_89"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_89,code=compute_89"
)
elseif
(
${
CMAKE_CUDA_COMPILER_VERSION
}
VERSION_GREATER
"11.1.0"
OR
${
CMAKE_CUDA_COMPILER_VERSION
}
VERSION_EQUAL
"11.1.0"
)
set
(
MGE_CUDA_GENCODE
"
${
MGE_CUDA_GENCODE
}
-gencode arch=compute_61,code=sm_61"
)
...
...
dnn/cuda-stub/src/libcuda-wrap_11.8.h
0 → 100644
浏览文件 @
3a35827d
此差异已折叠。
点击以展开。
dnn/cuda-stub/src/libcuda.cpp
浏览文件 @
3a35827d
...
...
@@ -59,6 +59,8 @@ CUresult on_init_failed(int func_idx) {
#include "./libcuda-wrap_11.2.h"
#elif CUDA_VERSION == 11040
#include "./libcuda-wrap_11.4.h"
#elif CUDA_VERSION == 11080
#include "./libcuda-wrap_11.8.h"
#else
#error "cuda stub not support this cuda version, you can close cuda stub to passby"
#endif
...
...
dnn/src/cuda/kernel_common/diagnostic_prologue.cuh
浏览文件 @
3a35827d
...
...
@@ -3,6 +3,7 @@
#else
#define MEGDNN_DIAGNOSTIC_PROLOGUE_INCLUDED
#endif
#include <cuda.h>
//! see
//! https://stackoverflow.com/questions/49836419/how-to-hide-nvccs-function-was-declared-but-never-referenced-warnings
...
...
scripts/whl/manylinux2014/build_wheel_common.sh
浏览文件 @
3a35827d
#!/bin/bash -e
set
-x
set
-
e
x
CWD
=
$(
dirname
$0
)
BASEDIR
=
$(
readlink
-f
${
CWD
}
/../../..
)
...
...
@@ -12,7 +12,7 @@ CUDA_LIB_DIR="/usr/local/cuda/lib64/"
TensorRT_LIB_DIR
=
"/opt/tensorrt/lib/"
SDK_NAME
=
"unknown"
x86_64_support_version
=
"cu101 cu111 cu112 cpu cu111_cudnn821_tensorRT825 cu114"
x86_64_support_version
=
"cu101 cu111 cu112 cpu cu111_cudnn821_tensorRT825 cu114
cu118
"
aarch64_support_version
=
"cu102_JetsonNano cu111 cpu"
if
[[
-z
${
IN_CI
}
]]
then
...
...
@@ -225,6 +225,44 @@ elif [ $SDK_NAME == "cu114" ];then
-gencode arch=compute_86,code=sm_86
\
-gencode arch=compute_86,code=compute_86
\"
"
elif
[
$SDK_NAME
==
"cu118"
]
;
then
BUILD_GCC8
=
"ON"
REQUIR_CUDA_VERSION
=
"11080"
REQUIR_CUDNN_VERSION
=
"8.6.0"
REQUIR_TENSORRT_VERSION
=
"8.5.3.1"
REQUIR_CUBLAS_VERSION
=
"11.11.3.6"
# override the default cuda/cudnn/trt lib dir
CUDNN_LIB_DIR
=
"/opt/cudnn/lib/"
CUDA_LIB_DIR
=
"/usr/local/cuda/targets/x86_64-linux/lib/"
TensorRT_LIB_DIR
=
"/opt/tensorrt/lib/"
CUDA_COPY_LIB_LIST
=
"
\
${
CUDA_LIB_DIR
}
/libnvrtc.so.11.2:
\
${
CUDA_LIB_DIR
}
/libcublasLt.so.11:
\
${
CUDA_LIB_DIR
}
/libcublas.so.11:
\
${
CUDNN_LIB_DIR
}
/libcudnn_adv_infer.so.8:
\
${
CUDNN_LIB_DIR
}
/libcudnn_adv_train.so.8:
\
${
CUDNN_LIB_DIR
}
/libcudnn_cnn_infer.so.8:
\
${
CUDNN_LIB_DIR
}
/libcudnn_cnn_train.so.8:
\
${
CUDNN_LIB_DIR
}
/libcudnn_ops_infer.so.8:
\
${
CUDNN_LIB_DIR
}
/libcudnn_ops_train.so.8:
\
${
CUDNN_LIB_DIR
}
/libcudnn.so.8:
\
${
TensorRT_LIB_DIR
}
/libnvinfer_plugin.so.8:
\
${
TensorRT_LIB_DIR
}
/libnvonnxparser.so.8
\
${
TensorRT_LIB_DIR
}
/libnvinfer_builder_resource.so.8.5.3:
\
${
TensorRT_LIB_DIR
}
/libnvparsers.so.8:
\
${
TensorRT_LIB_DIR
}
/libnvinfer.so.8"
EXTRA_CMAKE_FLAG
=
" -DMGE_WITH_CUDNN_SHARED=ON -DMGE_WITH_CUBLAS_SHARED=ON
\
-DMGE_CUDA_GENCODE=
\"
-gencode arch=compute_61,code=sm_61
\
-gencode arch=compute_70,code=sm_70
\
-gencode arch=compute_75,code=sm_75
\
-gencode arch=compute_80,code=sm_80
\
-gencode arch=compute_86,code=sm_86
\
-gencode arch=compute_89,code=sm_89
\
-gencode arch=compute_89,code=compute_89
\"
"
elif
[
$SDK_NAME
==
"cpu"
]
;
then
echo
"use
$SDK_NAME
without cuda support"
...
...
@@ -308,7 +346,7 @@ if [ ${BUILD_WHL_CPU_ONLY} = "OFF" ]; then
CUBLAS_VERSION_CONTEXT
=
$(
head
-150
${
CUBLAS_VERSION_PATH
}
)
CUDA_VERSION_CONTEXT
=
$(
head
-300
${
CUDA_VERSION_PATH
}
)
CUDNN_VERSION_CONTEXT
=
$(
head
-62
${
CUDNN_VERSION_PATH
}
)
TENSORRT_VERSION_CONTEXT
=
$(
tail
-
12
${
TENSORRT_VERSION_PATH
}
)
TENSORRT_VERSION_CONTEXT
=
$(
tail
-
20
${
TENSORRT_VERSION_PATH
}
)
if
[
"
$REQUIR_CUDA_VERSION
"
-ge
"11000"
]
;
then
CUDA_API_VERSION
=
$(
echo
$CUDA_VERSION_CONTEXT
|
grep
-Eo
"define CUDA_VERSION * +([0-9]+)"
)
...
...
src/tensorrt/include/megbrain/tensorrt/tensorrt_opr.h
浏览文件 @
3a35827d
...
...
@@ -36,8 +36,9 @@ enum class Empty : int32_t {};
#define TENSOR_RT_MANAGE_ALL_WORKSPACE 0
#endif
#if NV_TENSOR_RT_VERSION >= 8000
#error "if trt8 fix https://github.com/NVIDIA/TensorRT/issues/2290, try TENSOR_RT_MANAGE_ALL_WORKSPACE=0"
//! issue detail: https://github.com/NVIDIA/TensorRT/issues/2290
#if NV_TENSOR_RT_VERSION >= 8400
#define TENSOR_RT_MANAGE_ALL_WORKSPACE 0
#endif
namespace
mgb
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录