Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
6dd52c5b
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
6dd52c5b
编写于
1月 15, 2021
作者:
H
hutuxian
提交者:
GitHub
1月 15, 2021
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Ascend rc (#30483)
上级
715d8628
变更
30
展开全部
显示空白变更内容
内联
并排
Showing
30 changed file
with
1904 addition
and
24 deletion
+1904
-24
CMakeLists.txt
CMakeLists.txt
+5
-0
cmake/configure.cmake
cmake/configure.cmake
+4
-0
cmake/external/ascend.cmake
cmake/external/ascend.cmake
+61
-0
cmake/external/cryptopp.cmake
cmake/external/cryptopp.cmake
+2
-2
cmake/external/dlpack.cmake
cmake/external/dlpack.cmake
+1
-1
cmake/external/gflags.cmake
cmake/external/gflags.cmake
+3
-3
cmake/external/glog.cmake
cmake/external/glog.cmake
+3
-3
cmake/external/grpc.cmake
cmake/external/grpc.cmake
+1
-1
cmake/external/openblas.cmake
cmake/external/openblas.cmake
+1
-1
cmake/external/protobuf.cmake
cmake/external/protobuf.cmake
+4
-4
cmake/external/pybind11.cmake
cmake/external/pybind11.cmake
+2
-2
cmake/external/threadpool.cmake
cmake/external/threadpool.cmake
+1
-1
cmake/external/warpctc.cmake
cmake/external/warpctc.cmake
+3
-2
cmake/external/xbyak.cmake
cmake/external/xbyak.cmake
+1
-1
cmake/external/xxhash.cmake
cmake/external/xxhash.cmake
+1
-1
cmake/external/zlib.cmake
cmake/external/zlib.cmake
+2
-2
cmake/third_party.cmake
cmake/third_party.cmake
+5
-0
paddle/fluid/framework/fleet/CMakeLists.txt
paddle/fluid/framework/fleet/CMakeLists.txt
+4
-0
paddle/fluid/framework/fleet/ascend_wrapper.cc
paddle/fluid/framework/fleet/ascend_wrapper.cc
+22
-0
paddle/fluid/framework/fleet/ascend_wrapper.h
paddle/fluid/framework/fleet/ascend_wrapper.h
+183
-0
paddle/fluid/operators/CMakeLists.txt
paddle/fluid/operators/CMakeLists.txt
+3
-0
paddle/fluid/operators/ascend_trigger_op.cc
paddle/fluid/operators/ascend_trigger_op.cc
+52
-0
paddle/fluid/operators/ascend_trigger_op.h
paddle/fluid/operators/ascend_trigger_op.h
+46
-0
paddle/fluid/pybind/CMakeLists.txt
paddle/fluid/pybind/CMakeLists.txt
+5
-0
paddle/fluid/pybind/ascend_wrapper_py.cc
paddle/fluid/pybind/ascend_wrapper_py.cc
+694
-0
paddle/fluid/pybind/ascend_wrapper_py.h
paddle/fluid/pybind/ascend_wrapper_py.h
+31
-0
paddle/fluid/pybind/pybind.cc
paddle/fluid/pybind/pybind.cc
+7
-0
python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_optimizer.py
...tributed/fleet/meta_optimizers/ascend/ascend_optimizer.py
+179
-0
python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py
...distributed/fleet/meta_optimizers/ascend/ascend_parser.py
+529
-0
python/paddle/fluid/tests/unittests/test_ascend_trigger.py
python/paddle/fluid/tests/unittests/test_ascend_trigger.py
+49
-0
未找到文件。
CMakeLists.txt
浏览文件 @
6dd52c5b
...
...
@@ -31,9 +31,13 @@ option(WITH_GPU "Compile PaddlePaddle with NVIDIA GPU" ${CUDA_F
option
(
WITH_TENSORRT
"Compile PaddlePaddle with NVIDIA TensorRT"
OFF
)
option
(
WITH_XPU
"Compile PaddlePaddle with BAIDU KUNLUN XPU"
OFF
)
option
(
WITH_WIN_DUMP_DBG
"Compile with windows core dump debug mode"
OFF
)
option
(
WITH_ASCEND
"Compile PaddlePaddle with ASCEND"
OFF
)
if
(
WITH_GPU AND WITH_XPU
)
message
(
FATAL_ERROR
"Error when compile GPU and XPU at the same time"
)
endif
()
if
(
WITH_GPU AND WITH_ASCEND
)
message
(
FATAL_ERROR
"Error when compile GPU and ASCEND at the same time"
)
endif
()
# cmake 3.12, 3.13, 3.14 will append gcc link options to nvcc, and nvcc doesn't recognize them.
if
(
WITH_GPU
AND
(
${
CMAKE_VERSION
}
VERSION_GREATER_EQUAL 3.12
)
AND
(
${
CMAKE_VERSION
}
VERSION_LESS 3.15
))
message
(
FATAL_ERROR
"cmake
${
CMAKE_VERSION
}
is not supported when WITH_GPU=ON because of bug https://cmake.org/pipermail/cmake/2018-September/068195.html. "
...
...
@@ -322,6 +326,7 @@ set(PADDLE_PYTHON_BUILD_DIR "${CMAKE_CURRENT_BINARY_DIR}/python/build")
set
(
CMAKE_CXX_FLAGS_RELWITHDEBINFO
"-O3 -g -DNDEBUG"
)
set
(
CMAKE_C_FLAGS_RELWITHDEBINFO
"-O3 -g -DNDEBUG"
)
set
(
CMAKE_CXX_FLAGS
"
${
CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0"
)
if
(
ON_INFER
)
# you can trun off the paddle fluid and inference lib by set ON_INFER=OFF
...
...
cmake/configure.cmake
浏览文件 @
6dd52c5b
...
...
@@ -78,6 +78,10 @@ if(WITH_BOX_PS)
add_definitions
(
-DPADDLE_WITH_BOX_PS
)
endif
()
if
(
WITH_ASCEND
)
add_definitions
(
-DPADDLE_WITH_ASCEND
)
endif
()
if
(
WITH_XPU
)
message
(
STATUS
"Compile with XPU!"
)
add_definitions
(
-DPADDLE_WITH_XPU
)
...
...
cmake/external/ascend.cmake
0 → 100644
浏览文件 @
6dd52c5b
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
INCLUDE
(
ExternalProject
)
SET
(
ASCEND_PROJECT
"extern_ascend"
)
IF
((
NOT DEFINED ASCEND_VER
)
OR
(
NOT DEFINED ASCEND_URL
))
MESSAGE
(
STATUS
"use pre defined download url"
)
SET
(
ASCEND_VER
"0.1.1"
CACHE STRING
""
FORCE
)
SET
(
ASCEND_NAME
"ascend"
CACHE STRING
""
FORCE
)
SET
(
ASCEND_URL
"http://paddle-ascend.bj.bcebos.com/ascend.tar.gz"
CACHE STRING
""
FORCE
)
ENDIF
()
MESSAGE
(
STATUS
"ASCEND_NAME:
${
ASCEND_NAME
}
, ASCEND_URL:
${
ASCEND_URL
}
"
)
SET
(
ASCEND_SOURCE_DIR
"
${
THIRD_PARTY_PATH
}
/ascend"
)
SET
(
ASCEND_DOWNLOAD_DIR
"
${
ASCEND_SOURCE_DIR
}
/src/
${
ASCEND_PROJECT
}
"
)
SET
(
ASCEND_DST_DIR
"ascend"
)
SET
(
ASCEND_INSTALL_ROOT
"
${
THIRD_PARTY_PATH
}
/install"
)
SET
(
ASCEND_INSTALL_DIR
${
ASCEND_INSTALL_ROOT
}
/
${
ASCEND_DST_DIR
}
)
SET
(
ASCEND_ROOT
${
ASCEND_INSTALL_DIR
}
)
SET
(
ASCEND_INC_DIR
${
ASCEND_ROOT
}
/include
)
SET
(
ASCEND_LIB_DIR
${
ASCEND_ROOT
}
/lib
)
SET
(
ASCEND_LIB
${
ASCEND_LIB_DIR
}
/libge_runner.so
)
SET
(
ASCEND_GRAPH_LIB
${
ASCEND_LIB_DIR
}
/libgraph.so
)
SET
(
CMAKE_INSTALL_RPATH
"
${
CMAKE_INSTALL_RPATH
}
"
"
${
ASCEND_ROOT
}
/lib"
)
INCLUDE_DIRECTORIES
(
${
ASCEND_INC_DIR
}
)
FILE
(
WRITE
${
ASCEND_DOWNLOAD_DIR
}
/CMakeLists.txt
"PROJECT(ASCEND)
\n
"
"cmake_minimum_required(VERSION 3.0)
\n
"
"install(DIRECTORY
${
ASCEND_NAME
}
/include
${
ASCEND_NAME
}
/lib
\n
"
" DESTINATION
${
ASCEND_DST_DIR
}
)
\n
"
)
ExternalProject_Add
(
${
ASCEND_PROJECT
}
${
EXTERNAL_PROJECT_LOG_ARGS
}
PREFIX
${
ASCEND_SOURCE_DIR
}
DOWNLOAD_DIR
${
ASCEND_DOWNLOAD_DIR
}
DOWNLOAD_COMMAND wget --no-check-certificate
${
ASCEND_URL
}
-c -q -O
${
ASCEND_NAME
}
.tar.gz
&& tar zxvf
${
ASCEND_NAME
}
.tar.gz
DOWNLOAD_NO_PROGRESS 1
UPDATE_COMMAND
""
CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=
${
ASCEND_INSTALL_ROOT
}
CMAKE_CACHE_ARGS -DCMAKE_INSTALL_PREFIX:PATH=
${
ASCEND_INSTALL_ROOT
}
)
ADD_LIBRARY
(
ascend SHARED IMPORTED GLOBAL
)
SET_PROPERTY
(
TARGET ascend PROPERTY IMPORTED_LOCATION
${
ASCEND_LIB
}
)
ADD_LIBRARY
(
ascend_graph SHARED IMPORTED GLOBAL
)
SET_PROPERTY
(
TARGET ascend_graph PROPERTY IMPORTED_LOCATION
${
ASCEND_GRAPH_LIB
}
)
ADD_DEPENDENCIES
(
ascend ascend_graph
${
ASCEND_PROJECT
}
)
cmake/external/cryptopp.cmake
浏览文件 @
6dd52c5b
...
...
@@ -17,7 +17,7 @@ INCLUDE(ExternalProject)
SET
(
CRYPTOPP_PREFIX_DIR
${
THIRD_PARTY_PATH
}
/cryptopp
)
SET
(
CRYPTOPP_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/cryptopp
)
SET
(
CRYPTOPP_INCLUDE_DIR
"
${
CRYPTOPP_INSTALL_DIR
}
/include"
CACHE PATH
"cryptopp include directory."
FORCE
)
SET
(
CRYPTOPP_REPOSITORY
${
GIT_URL
}
/weidai11
/cryptopp.git
)
SET
(
CRYPTOPP_REPOSITORY
https://gitee.com/tianjianhe
/cryptopp.git
)
SET
(
CRYPTOPP_TAG CRYPTOPP_8_2_0
)
IF
(
WIN32
)
...
...
@@ -33,7 +33,7 @@ set(CRYPTOPP_CMAKE_ARGS ${COMMON_CMAKE_ARGS}
-DCMAKE_INSTALL_LIBDIR=
${
CRYPTOPP_INSTALL_DIR
}
/lib
-DCMAKE_INSTALL_PREFIX=
${
CRYPTOPP_INSTALL_DIR
}
-DCMAKE_BUILD_TYPE=
${
THIRD_PARTY_BUILD_TYPE
}
-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
"-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0"
-DCMAKE_CXX_FLAGS_RELEASE=
${
CMAKE_CXX_FLAGS_RELEASE
}
-DCMAKE_C_COMPILER=
${
CMAKE_C_COMPILER
}
-DCMAKE_CXX_COMPILER=
${
CMAKE_CXX_COMPILER
}
...
...
cmake/external/dlpack.cmake
浏览文件 @
6dd52c5b
...
...
@@ -17,7 +17,7 @@ include(ExternalProject)
set
(
DLPACK_PREFIX_DIR
${
THIRD_PARTY_PATH
}
/dlpack
)
set
(
DLPACK_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/dlpack/src/extern_dlpack
)
set
(
DLPACK_REPOSITORY
${
GIT_URL
}
/dmlc
/dlpack.git
)
set
(
DLPACK_REPOSITORY
https://gitee.com/tianjianhe
/dlpack.git
)
set
(
DLPACK_TAG v0.2
)
cache_third_party
(
extern_dlpack
...
...
cmake/external/gflags.cmake
浏览文件 @
6dd52c5b
...
...
@@ -18,8 +18,8 @@ SET(GFLAGS_PREFIX_DIR ${THIRD_PARTY_PATH}/gflags)
SET
(
GFLAGS_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/gflags/src/extern_gflags
)
SET
(
GFLAGS_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/gflags
)
SET
(
GFLAGS_INCLUDE_DIR
"
${
GFLAGS_INSTALL_DIR
}
/include"
CACHE PATH
"gflags include directory."
FORCE
)
set
(
GFLAGS_REPOSITORY
${
GIT_URL
}
/gflags
/gflags.git
)
set
(
GFLAGS_TAG
"v2.2.2"
)
set
(
GFLAGS_REPOSITORY
https://gitee.com/tianjianhe
/gflags.git
)
set
(
GFLAGS_TAG
77592648e3f3be87d6c7123eb81cbad75f9aef5a
)
IF
(
WIN32
)
set
(
GFLAGS_LIBRARIES
"
${
GFLAGS_INSTALL_DIR
}
/lib/gflags_static.lib"
CACHE FILEPATH
"GFLAGS_LIBRARIES"
FORCE
)
ELSE
(
WIN32
)
...
...
@@ -48,7 +48,7 @@ ExternalProject_Add(
INSTALL_COMMAND
${
INSTALL_COMMAND
}
CMAKE_ARGS -DCMAKE_CXX_COMPILER=
${
CMAKE_CXX_COMPILER
}
-DCMAKE_C_COMPILER=
${
CMAKE_C_COMPILER
}
-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
"-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0"
-DCMAKE_CXX_FLAGS_RELEASE=
${
CMAKE_CXX_FLAGS_RELEASE
}
-DCMAKE_CXX_FLAGS_DEBUG=
${
CMAKE_CXX_FLAGS_DEBUG
}
-DCMAKE_C_FLAGS=
${
CMAKE_C_FLAGS
}
...
...
cmake/external/glog.cmake
浏览文件 @
6dd52c5b
...
...
@@ -18,8 +18,8 @@ SET(GLOG_PREFIX_DIR ${THIRD_PARTY_PATH}/glog)
SET
(
GLOG_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/glog/src/extern_glog
)
SET
(
GLOG_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/glog
)
SET
(
GLOG_INCLUDE_DIR
"
${
GLOG_INSTALL_DIR
}
/include"
CACHE PATH
"glog include directory."
FORCE
)
SET
(
GLOG_REPOSITORY
${
GIT_URL
}
/googl
e/glog.git
)
SET
(
GLOG_TAG v0.
4.0
)
SET
(
GLOG_REPOSITORY
https://gitee.com/tianjianh
e/glog.git
)
SET
(
GLOG_TAG v0.
3.5
)
IF
(
WIN32
)
SET
(
GLOG_LIBRARIES
"
${
GLOG_INSTALL_DIR
}
/lib/glog.lib"
CACHE FILEPATH
"glog library."
FORCE
)
...
...
@@ -47,7 +47,7 @@ ExternalProject_Add(
SOURCE_DIR
${
GLOG_SOURCE_DIR
}
CMAKE_ARGS -DCMAKE_CXX_COMPILER=
${
CMAKE_CXX_COMPILER
}
-DCMAKE_C_COMPILER=
${
CMAKE_C_COMPILER
}
-DCMAKE_CXX_FLAGS=
${
GLOG_CMAKE_CXX_FLAGS
}
"-DCMAKE_CXX_FLAGS=
${
GLOG_CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0"
-DCMAKE_CXX_FLAGS_RELEASE=
${
CMAKE_CXX_FLAGS_RELEASE
}
-DCMAKE_CXX_FLAGS_DEBUG=
${
CMAKE_CXX_FLAGS_DEBUG
}
-DCMAKE_C_FLAGS=
${
CMAKE_C_FLAGS
}
...
...
cmake/external/grpc.cmake
浏览文件 @
6dd52c5b
...
...
@@ -28,7 +28,7 @@ IF(APPLE)
SET
(
GRPC_INSTALL_CMD make prefix=
${
GRPC_INSTALL_DIR
}
install
)
ELSE
()
SET
(
GRPC_CFLAGS
"-Wno-error -std=c11
${
CLFAGS
}
"
)
SET
(
GRPC_CXXFLAGS
"-Wno-error -std=c++11
${
CXXFLAGS
}
"
)
SET
(
GRPC_CXXFLAGS
"-Wno-error -std=c++11
${
CXXFLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0
"
)
SET
(
BUILD_CMD make CFLAGS=
${
GRPC_CFLAGS
}
CXXFLAGS=
${
GRPC_CXXFLAGS
}
HAS_SYSTEM_PROTOBUF=false -s -j
${
NUM_OF_PROCESSOR
}
static grpc_cpp_plugin
)
SET
(
GRPC_INSTALL_CMD make prefix=
${
GRPC_INSTALL_DIR
}
install CFLAGS=
${
GRPC_CFLAGS
}
CXXFLAGS=
${
GRPC_CXXFLAGS
}
)
ENDIF
()
...
...
cmake/external/openblas.cmake
浏览文件 @
6dd52c5b
...
...
@@ -17,7 +17,7 @@ INCLUDE(ExternalProject)
SET
(
CBLAS_PREFIX_DIR
${
THIRD_PARTY_PATH
}
/openblas
)
SET
(
CBLAS_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/openblas/src/extern_openblas
)
SET
(
CBLAS_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/openblas
)
SET
(
CBLAS_REPOSITORY
${
GIT_URL
}
/xianyi
/OpenBLAS.git
)
SET
(
CBLAS_REPOSITORY
https://gitee.com/tianjianhe
/OpenBLAS.git
)
SET
(
CBLAS_TAG v0.3.7
)
if
(
WITH_MIPS
)
SET
(
CBLAS_TAG v0.3.13
)
...
...
cmake/external/protobuf.cmake
浏览文件 @
6dd52c5b
...
...
@@ -183,7 +183,7 @@ FUNCTION(build_protobuf TARGET_NAME BUILD_FOR_HOST)
"-DCMAKE_C_FLAGS=
${
CMAKE_C_FLAGS
}
"
"-DCMAKE_C_FLAGS_DEBUG=
${
CMAKE_C_FLAGS_DEBUG
}
"
"-DCMAKE_C_FLAGS_RELEASE=
${
CMAKE_C_FLAGS_RELEASE
}
"
"-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
"
"-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0
"
"-DCMAKE_CXX_FLAGS_RELEASE=
${
CMAKE_CXX_FLAGS_RELEASE
}
"
"-DCMAKE_CXX_FLAGS_DEBUG=
${
CMAKE_CXX_FLAGS_DEBUG
}
"
"-Dprotobuf_WITH_ZLIB=ON"
...
...
@@ -198,8 +198,8 @@ FUNCTION(build_protobuf TARGET_NAME BUILD_FOR_HOST)
"-Dprotobuf_MSVC_STATIC_RUNTIME=
${
MSVC_STATIC_CRT
}
"
)
ENDIF
()
SET
(
PROTOBUF_REPOSITORY
${
GIT_URL
}
/protocolbuffers
/protobuf.git
)
SET
(
PROTOBUF_TAG
9f75c5aa851cd877fb0d93ccc31b8567a6706546
)
SET
(
PROTOBUF_REPOSITORY
https://gitee.com/tianjianhe
/protobuf.git
)
SET
(
PROTOBUF_TAG
v3.8.0
)
cache_third_party
(
${
TARGET_NAME
}
REPOSITORY
${
PROTOBUF_REPOSITORY
}
...
...
@@ -234,7 +234,7 @@ FUNCTION(build_protobuf TARGET_NAME BUILD_FOR_HOST)
)
ENDFUNCTION
()
SET
(
PROTOBUF_VERSION 3.1.0
)
#
SET(PROTOBUF_VERSION 3.1.0)
IF
(
NOT PROTOBUF_FOUND
)
build_protobuf
(
extern_protobuf FALSE
)
...
...
cmake/external/pybind11.cmake
浏览文件 @
6dd52c5b
...
...
@@ -16,8 +16,8 @@ include(ExternalProject)
set
(
PYBIND_PREFIX_DIR
${
THIRD_PARTY_PATH
}
/pybind
)
set
(
PYBIND_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/pybind/src/extern_pybind
)
SET
(
PYBIND_REPOSITORY
${
GIT_URL
}
/pybind
/pybind11.git
)
SET
(
PYBIND_TAG v2.
4.3
)
SET
(
PYBIND_REPOSITORY
https://gitee.com/tianjianhe
/pybind11.git
)
SET
(
PYBIND_TAG v2.
6.0
)
cache_third_party
(
extern_pybind
REPOSITORY
${
PYBIND_REPOSITORY
}
...
...
cmake/external/threadpool.cmake
浏览文件 @
6dd52c5b
...
...
@@ -16,7 +16,7 @@ INCLUDE(ExternalProject)
SET
(
THREADPOOL_PREFIX_DIR
${
THIRD_PARTY_PATH
}
/threadpool
)
SET
(
THREADPOOL_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/threadpool/src/extern_threadpool
)
SET
(
THREADPOOL_REPOSITORY
${
GIT_URL
}
/progschj
/ThreadPool.git
)
SET
(
THREADPOOL_REPOSITORY
https://gitee.com/tianjianhe
/ThreadPool.git
)
SET
(
THREADPOOL_TAG 9a42ec1329f259a5f4881a291db1dcb8f2ad9040
)
cache_third_party
(
extern_threadpool
...
...
cmake/external/warpctc.cmake
浏览文件 @
6dd52c5b
...
...
@@ -17,8 +17,9 @@ INCLUDE(ExternalProject)
SET
(
WARPCTC_PREFIX_DIR
${
THIRD_PARTY_PATH
}
/warpctc
)
SET
(
WARPCTC_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/warpctc/src/extern_warpctc
)
SET
(
WARPCTC_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/warpctc
)
set
(
WARPCTC_REPOSITORY
${
GIT_URL
}
/baidu-research
/warp-ctc.git
)
set
(
WARPCTC_REPOSITORY
https://gitee.com/tianjianhe
/warp-ctc.git
)
set
(
WARPCTC_TAG 95a461eddeabd51099ef059dcfada1117eb1bfb8
)
# set(WARPCTC_TAG bc29dcfff07ced1c7a19a4ecee48e5ad583cef8e)
SET
(
WARPCTC_INCLUDE_DIR
"
${
WARPCTC_INSTALL_DIR
}
/include"
CACHE PATH
"Warp-ctc Directory"
FORCE
)
...
...
@@ -52,7 +53,7 @@ ExternalProject_Add(
-DCMAKE_C_FLAGS=
${
CMAKE_C_FLAGS
}
-DCMAKE_C_FLAGS_DEBUG=
${
CMAKE_C_FLAGS_DEBUG
}
-DCMAKE_C_FLAGS_RELEASE=
${
CMAKE_C_FLAGS_RELEASE
}
-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
"-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0"
-DCMAKE_CXX_FLAGS_RELEASE=
${
CMAKE_CXX_FLAGS_RELEASE
}
-DCMAKE_CXX_FLAGS_DEBUG=
${
CMAKE_CXX_FLAGS_DEBUG
}
-DCMAKE_INSTALL_PREFIX=
${
WARPCTC_INSTALL_DIR
}
...
...
cmake/external/xbyak.cmake
浏览文件 @
6dd52c5b
...
...
@@ -19,7 +19,7 @@ set(XBYAK_PREFIX_DIR ${THIRD_PARTY_PATH}/xbyak)
SET
(
XBYAK_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/xbyak/src/extern_xbyak
)
set
(
XBYAK_INSTALL_ROOT
${
THIRD_PARTY_PATH
}
/install/xbyak
)
set
(
XBYAK_INC_DIR
${
XBYAK_INSTALL_ROOT
}
/include
)
set
(
XBYAK_REPOSITORY
${
GIT_URL
}
/herumi
/xbyak.git
)
set
(
XBYAK_REPOSITORY
https://gitee.com/tianjianhe
/xbyak.git
)
set
(
XBYAK_TAG v5.661
)
# Jul 26th
include_directories
(
${
XBYAK_INC_DIR
}
)
...
...
cmake/external/xxhash.cmake
浏览文件 @
6dd52c5b
...
...
@@ -18,7 +18,7 @@ set(XXHASH_PREFIX_DIR ${THIRD_PARTY_PATH}/xxhash)
set
(
XXHASH_SOURCE_DIR
${
THIRD_PARTY_PATH
}
/xxhash/src/extern_xxhash
)
set
(
XXHASH_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/xxhash
)
set
(
XXHASH_INCLUDE_DIR
"
${
XXHASH_INSTALL_DIR
}
/include"
)
set
(
XXHASH_REPOSITORY
${
GIT_URL
}
/Cyan4973
/xxHash.git
)
set
(
XXHASH_REPOSITORY
https://gitee.com/tianjianhe
/xxHash.git
)
set
(
XXHASH_TAG v0.6.5
)
cache_third_party
(
extern_xxhash
...
...
cmake/external/zlib.cmake
浏览文件 @
6dd52c5b
...
...
@@ -19,7 +19,7 @@ SET(ZLIB_SOURCE_DIR ${THIRD_PARTY_PATH}/zlib/src/extern_zlib)
SET
(
ZLIB_INSTALL_DIR
${
THIRD_PARTY_PATH
}
/install/zlib
)
SET
(
ZLIB_ROOT
${
ZLIB_INSTALL_DIR
}
CACHE FILEPATH
"zlib root directory."
FORCE
)
SET
(
ZLIB_INCLUDE_DIR
"
${
ZLIB_INSTALL_DIR
}
/include"
CACHE PATH
"zlib include directory."
FORCE
)
set
(
ZLIB_REPOSITORY
${
GIT_URL
}
/madler
/zlib.git
)
set
(
ZLIB_REPOSITORY
https://gitee.com/tianjianhe
/zlib.git
)
set
(
ZLIB_TAG v1.2.8
)
INCLUDE_DIRECTORIES
(
${
ZLIB_INCLUDE_DIR
}
)
# For zlib code to include its own headers.
...
...
@@ -41,7 +41,7 @@ ExternalProject_Add(
CMAKE_ARGS -DCMAKE_C_COMPILER=
${
CMAKE_C_COMPILER
}
-DCMAKE_CXX_COMPILER=
${
CMAKE_CXX_COMPILER
}
-DCMAKE_C_FLAGS=
${
CMAKE_C_FLAGS
}
-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
"-DCMAKE_CXX_FLAGS=
${
CMAKE_CXX_FLAGS
}
-D_GLIBCXX_USE_CXX11_ABI=0"
-DCMAKE_INSTALL_PREFIX=
${
ZLIB_INSTALL_DIR
}
-DBUILD_SHARED_LIBS=OFF
-DCMAKE_POSITION_INDEPENDENT_CODE=ON
...
...
cmake/third_party.cmake
浏览文件 @
6dd52c5b
...
...
@@ -274,6 +274,11 @@ if(WITH_BOX_PS)
list
(
APPEND third_party_deps extern_box_ps
)
endif
(
WITH_BOX_PS
)
if
(
WITH_ASCEND
)
include
(
external/ascend
)
list
(
APPEND third_party_deps extern_ascend
)
endif
(
WITH_ASCEND
)
if
(
WITH_PSCORE
)
include
(
external/snappy
)
list
(
APPEND third_party_deps extern_snappy
)
...
...
paddle/fluid/framework/fleet/CMakeLists.txt
浏览文件 @
6dd52c5b
...
...
@@ -31,3 +31,7 @@ endif(WITH_GLOO)
cc_library
(
heter_wrapper SRCS heter_wrapper.cc DEPS framework_proto device_context heter_service_proto
)
cc_test
(
test_fleet_cc SRCS test_fleet.cc DEPS fleet_wrapper gloo_wrapper fs shell
)
if
(
WITH_ASCEND
)
cc_library
(
ascend_wrapper SRCS ascend_wrapper.cc DEPS framework_proto lod_tensor ascend ascend_graph
)
endif
(
WITH_ASCEND
)
paddle/fluid/framework/fleet/ascend_wrapper.cc
0 → 100644
浏览文件 @
6dd52c5b
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#ifdef PADDLE_WITH_ASCEND
#include "paddle/fluid/framework/fleet/ascend_wrapper.h"
namespace
paddle
{
namespace
framework
{
std
::
shared_ptr
<
AscendInstance
>
AscendInstance
::
ascend_instance_
=
nullptr
;
}
// end namespace framework
}
// end namespace paddle
#endif
paddle/fluid/framework/fleet/ascend_wrapper.h
0 → 100644
浏览文件 @
6dd52c5b
/* Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#ifdef PADDLE_WITH_ASCEND
#include <glog/logging.h>
#include <map>
#include <memory>
#include <string>
#include <vector>
#include "paddle/fluid/framework/data_type.h"
#include "paddle/fluid/framework/lod_tensor.h"
#include "paddle/fluid/platform/gpu_info.h"
#include "paddle/fluid/platform/place.h"
#include "paddle/fluid/platform/timer.h"
#include "ge/ge_api.h"
#include "ge/ge_api_types.h"
#include "graph/attr_value.h"
#include "graph/tensor.h"
#include "graph/types.h"
namespace
paddle
{
namespace
framework
{
// typedef std::vector<std::string> AscendGraphDesc;
typedef
ge
::
Graph
AscendGraphDesc
;
class
AscendInstance
{
public:
virtual
~
AscendInstance
()
{}
AscendInstance
()
{}
std
::
map
<
std
::
string
,
std
::
string
>
GetDefaultInitSessionOptions
()
{
std
::
map
<
std
::
string
,
std
::
string
>
init_options
;
init_options
[
"a"
]
=
"b"
;
init_options
[
"ge.trainFlag"
]
=
"1"
;
return
init_options
;
}
// add other parameters here to init
void
InitGlobalResouces
()
{
session_
.
reset
(
new
ge
::
Session
(
GetDefaultInitSessionOptions
()));
VLOG
(
1
)
<<
"InitGlobalResouces Done"
;
}
static
std
::
shared_ptr
<
AscendInstance
>
GetInstance
()
{
if
(
nullptr
==
ascend_instance_
)
{
ascend_instance_
.
reset
(
new
paddle
::
framework
::
AscendInstance
());
VLOG
(
1
)
<<
"Initialize AscendInstance Done"
;
}
return
ascend_instance_
;
}
void
AddAscendSubgraph
(
int
graph_idx
,
const
AscendGraphDesc
&
graph
)
{
ge
::
Status
status
=
session_
->
AddGraph
(
graph_idx
,
graph
);
PADDLE_ENFORCE_EQ
(
status
,
ge
::
SUCCESS
,
paddle
::
platform
::
errors
::
PreconditionNotMet
(
"Calling addGraph of graph engine failed, please "
"check Ascend Log."
));
VLOG
(
1
)
<<
"AddAscendSubgraph "
<<
graph_idx
<<
" Done"
;
}
ge
::
DataType
VarTypeToGeType
(
proto
::
VarType
::
Type
type
)
{
if
(
type
==
proto
::
VarType
::
FP16
)
{
return
ge
::
DataType
::
DT_FLOAT16
;
}
else
if
(
type
==
proto
::
VarType
::
FP32
)
{
return
ge
::
DataType
::
DT_FLOAT
;
}
else
if
(
type
==
proto
::
VarType
::
FP64
)
{
return
ge
::
DataType
::
DT_DOUBLE
;
}
else
if
(
type
==
proto
::
VarType
::
INT32
)
{
return
ge
::
DataType
::
DT_INT32
;
}
else
if
(
type
==
proto
::
VarType
::
INT64
)
{
return
ge
::
DataType
::
DT_INT64
;
}
else
{
PADDLE_THROW
(
platform
::
errors
::
Unimplemented
(
"Not support %s as tensor type."
,
DataTypeToString
(
type
)));
}
}
int
GeTypeSize
(
proto
::
VarType
::
Type
type
)
{
if
(
type
==
proto
::
VarType
::
FP16
)
{
return
2
;
}
else
if
(
type
==
proto
::
VarType
::
FP32
)
{
return
4
;
}
else
if
(
type
==
proto
::
VarType
::
FP64
)
{
return
8
;
}
else
if
(
type
==
proto
::
VarType
::
INT32
)
{
return
4
;
}
else
if
(
type
==
proto
::
VarType
::
INT64
)
{
return
8
;
}
else
{
PADDLE_THROW
(
platform
::
errors
::
Unimplemented
(
"Not support %s as tensor type."
,
DataTypeToString
(
type
)));
}
}
ge
::
Tensor
ConvertToGeTensor
(
const
Tensor
*
tensor
)
{
auto
numel
=
tensor
->
numel
();
std
::
vector
<
int64_t
>
vec_dim
;
auto
dimen
=
arity
(
tensor
->
dims
());
for
(
auto
i
=
0
;
i
<
dimen
;
++
i
)
{
vec_dim
.
push_back
(
tensor
->
dims
()[
i
]);
}
// For Debug
// VLOG(1) << "input numel: " << numel << ", dimen is " << vec_dim.size() <<
// ", and shape is";
// for (const auto e : vec_dim) {
// VLOG(0) << e;
// }
ge
::
Shape
shape
(
vec_dim
);
ge
::
TensorDesc
tensor_desc
(
shape
,
ge
::
Format
::
FORMAT_ND
,
VarTypeToGeType
(
tensor
->
type
()));
tensor_desc
.
SetRealDimCnt
(
vec_dim
.
size
());
const
uint8_t
*
data
=
reinterpret_cast
<
const
uint8_t
*>
(
tensor
->
data
<
void
>
());
std
::
vector
<
uint8_t
>
dst
(
numel
*
GeTypeSize
(
tensor
->
type
()));
memcpy
(
dst
.
data
(),
data
,
GeTypeSize
(
tensor
->
type
())
*
numel
);
ge
::
Tensor
ge_tensor
(
tensor_desc
,
dst
);
return
ge_tensor
;
}
void
RunAscendSubgraph
(
int
graph_idx
,
const
std
::
vector
<
const
Tensor
*>
&
inputs
,
std
::
vector
<
Tensor
*>
*
outputs
)
{
VLOG
(
1
)
<<
"Ascend Graph["
<<
graph_idx
<<
"] is about to run."
;
// Convert paddle Tensor to GE Tensor
std
::
vector
<
ge
::
Tensor
>
ge_inputs
;
for
(
const
auto
&
e
:
inputs
)
{
ge_inputs
.
push_back
(
ConvertToGeTensor
(
e
));
}
// Run Graph
std
::
vector
<
ge
::
Tensor
>
ge_outputs
;
ge
::
Status
status
=
session_
->
RunGraph
(
graph_idx
,
ge_inputs
,
ge_outputs
);
PADDLE_ENFORCE_EQ
(
status
,
ge
::
SUCCESS
,
paddle
::
platform
::
errors
::
PreconditionNotMet
(
"Calling RunGraph of graph engine failed, please "
"check Ascend Log."
));
VLOG
(
1
)
<<
"Run Ascend Graph["
<<
graph_idx
<<
"] Done"
;
// change tensor back, note all tensor's type computed in GE is uint8
for
(
size_t
i
=
0
;
i
<
ge_outputs
.
size
();
++
i
)
{
const
uint8_t
*
ret_data
=
ge_outputs
[
i
].
GetData
();
size_t
size
=
ge_outputs
[
i
].
GetSize
();
VLOG
(
1
)
<<
"GE Tensor size of the "
<<
i
<<
"th output var is "
<<
size
;
auto
*
dst
=
(
*
outputs
)[
i
]
->
mutable_data
<
uint8_t
>
({(
int64_t
)
size
},
platform
::
CPUPlace
());
memcpy
(
dst
,
ret_data
,
size
);
// Following for debug:
// VLOG(0) << "output for " << i << " var: ";
// float *tmp = reinterpret_cast<float*>(dst);
// for (size_t j = 0; j < size / 4; ++j) {
// printf("%f ", tmp[j]);
// }
// printf("\n");
}
}
protected:
std
::
shared_ptr
<
ge
::
Session
>
session_
;
private:
static
std
::
shared_ptr
<
AscendInstance
>
ascend_instance_
;
};
}
// end namespace framework
}
// end namespace paddle
#endif
paddle/fluid/operators/CMakeLists.txt
浏览文件 @
6dd52c5b
...
...
@@ -115,6 +115,9 @@ set(COMMON_OP_DEPS ${COMMON_OP_DEPS} device_memory_aligment)
set
(
COMMON_OP_DEPS
${
COMMON_OP_DEPS
}
layer
)
set
(
COMMON_OP_DEPS
${
COMMON_OP_DEPS
}
tensor_formatter
)
set
(
COMMON_OP_DEPS
${
COMMON_OP_DEPS
}
op_version_registry
)
if
(
WITH_ASCEND
)
set
(
COMMON_OP_DEPS
${
COMMON_OP_DEPS
}
ascend_wrapper
)
endif
()
# FIXME(typhoonzero): operator deps may not needed.
# op_library(lod_tensor_to_array_op DEPS lod_rank_table_op)
...
...
paddle/fluid/operators/ascend_trigger_op.cc
0 → 100644
浏览文件 @
6dd52c5b
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "paddle/fluid/operators/ascend_trigger_op.h"
namespace
paddle
{
namespace
operators
{
class
AscendTriggerOp
:
public
framework
::
OperatorWithKernel
{
public:
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{}
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
return
framework
::
OpKernelType
(
framework
::
proto
::
VarType
::
FP32
,
ctx
.
device_context
());
}
};
class
AscendTriggerOpMaker
:
public
framework
::
OpProtoAndCheckerMaker
{
public:
void
Make
()
override
{
AddInput
(
"FeedList"
,
"FeedList of Ascend SubGraph"
).
AsDuplicable
();
AddOutput
(
"FetchList"
,
"FetchList of Ascend SubGraph"
).
AsDuplicable
();
AddAttr
<
int
>
(
"graph_idx"
,
"(int, the graph index"
).
SetDefault
(
-
1
);
AddComment
(
R"DOC(
Trigger Ascend SubGraph
)DOC"
);
}
};
}
// namespace operators
}
// namespace paddle
namespace
ops
=
paddle
::
operators
;
REGISTER_OPERATOR
(
ascend_trigger
,
ops
::
AscendTriggerOp
,
ops
::
AscendTriggerOpMaker
);
REGISTER_OP_CPU_KERNEL
(
ascend_trigger
,
ops
::
AscendTriggerCPUKernel
<
float
>
)
paddle/fluid/operators/ascend_trigger_op.h
0 → 100644
浏览文件 @
6dd52c5b
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#pragma once
#include <memory>
#include <vector>
#include "paddle/fluid/framework/op_registry.h"
#ifdef PADDLE_WITH_ASCEND
#include "paddle/fluid/framework/fleet/ascend_wrapper.h"
#include "paddle/fluid/framework/tensor.h"
#endif
namespace
paddle
{
namespace
operators
{
template
<
typename
T
>
class
AscendTriggerCPUKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
#ifdef PADDLE_WITH_ASCEND
auto
ascend_ptr
=
paddle
::
framework
::
AscendInstance
::
GetInstance
();
auto
graph_idx
=
ctx
.
Attr
<
int
>
(
"graph_idx"
);
VLOG
(
4
)
<<
"AscendTrigger Kernel, begin to run graph: "
<<
graph_idx
;
auto
inputs
=
ctx
.
MultiInput
<
framework
::
Tensor
>
(
"FeedList"
);
auto
outputs
=
ctx
.
MultiOutput
<
framework
::
Tensor
>
(
"FetchList"
);
ascend_ptr
->
RunAscendSubgraph
(
graph_idx
,
inputs
,
&
outputs
);
#else
PADDLE_THROW
(
platform
::
errors
::
PreconditionNotMet
(
"Please compile WITH_ASCEND option to enable ascend_trigger op"
));
#endif
}
};
}
// namespace operators
}
// namespace paddle
paddle/fluid/pybind/CMakeLists.txt
浏览文件 @
6dd52c5b
...
...
@@ -39,6 +39,11 @@ set(PYBIND_SRCS
compatible.cc
generator_py.cc
)
if
(
WITH_ASCEND
)
set
(
PYBIND_DEPS
${
PYBIND_DEPS
}
ascend_wrapper
)
set
(
PYBIND_SRCS
${
PYBIND_SRCS
}
ascend_wrapper_py.cc
)
endif
(
WITH_ASCEND
)
if
(
WITH_GLOO
)
set
(
PYBIND_DEPS
${
PYBIND_DEPS
}
gloo_context
)
set
(
PYBIND_SRCS
${
PYBIND_SRCS
}
gloo_context_py.cc
)
...
...
paddle/fluid/pybind/ascend_wrapper_py.cc
0 → 100644
浏览文件 @
6dd52c5b
此差异已折叠。
点击以展开。
paddle/fluid/pybind/ascend_wrapper_py.h
0 → 100644
浏览文件 @
6dd52c5b
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#pragma once
#ifdef PADDLE_WITH_ASCEND
#include "pybind11/pybind11.h"
#include "pybind11/stl.h"
namespace
py
=
pybind11
;
namespace
paddle
{
namespace
pybind
{
void
BindAscendGraph
(
py
::
module
*
m
);
void
BindAscendWrapper
(
py
::
module
*
m
);
}
// namespace pybind
}
// namespace paddle
#endif
paddle/fluid/pybind/pybind.cc
浏览文件 @
6dd52c5b
...
...
@@ -65,6 +65,9 @@ limitations under the License. */
#include "paddle/fluid/platform/monitor.h"
#include "paddle/fluid/platform/place.h"
#include "paddle/fluid/platform/profiler.h"
#ifdef PADDLE_WITH_ASCEND
#include "paddle/fluid/pybind/ascend_wrapper_py.h"
#endif
#include "paddle/fluid/pybind/box_helper_py.h"
#include "paddle/fluid/pybind/compatible.h"
#include "paddle/fluid/pybind/const_value.h"
...
...
@@ -2837,6 +2840,10 @@ All parameter, weight, gradient are variables in Paddle.
BindCompatible
(
&
m
);
BindDataset
(
&
m
);
BindGenerator
(
&
m
);
#ifdef PADDLE_WITH_ASCEND
BindAscendWrapper
(
&
m
);
BindAscendGraph
(
&
m
);
#endif
#ifdef PADDLE_WITH_CRYPTO
BindCrypto
(
&
m
);
#endif
...
...
python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_optimizer.py
0 → 100644
浏览文件 @
6dd52c5b
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import
paddle.fluid.framework
as
framework
from
paddle.fluid.optimizer
import
Optimizer
import
paddle.fluid.core
as
core
import
numpy
as
np
import
ascend_parser
class
AscendIRParser
(
object
):
def
__init__
(
self
):
self
.
graph_idx
=
0
def
_construct_input_map
(
self
,
input_varlist
):
ret_map
=
{}
ge_in_operator
=
[]
for
id
,
var
in
enumerate
(
input_varlist
):
if
var
.
is_data
:
# input data
ge_input
=
core
.
GEOperatorFactory
.
create_operator
(
var
.
name
,
"Data"
).
set_attr_int32
(
"index"
,
id
)
ret_map
[
var
.
name
]
=
ge_input
ge_in_operator
.
append
(
ge_input
)
else
:
# param, learning ...
ge_input
=
core
.
GEOperatorFactory
.
create_operator
(
var
.
name
,
"Variable"
)
ge_input
.
update_output_desc
(
"y"
,
core
.
GETensorDesc
(
core
.
GEShape
(
var
.
shape
),
core
.
GEFormat
.
FORMAT_ND
,
core
.
GEDataType
.
DT_FLOAT
))
ret_map
[
var
.
name
]
=
ge_input
return
ge_in_operator
,
ret_map
def
parse_op
(
self
,
op
):
if
op
.
type
in
ascend_parser
.
registerd_op
:
print
(
"Op[%s] has been registered, begin to parse it"
%
(
op
.
type
))
op_parser
=
self
.
parser_factory
.
create_parse
(
ascend_parser
.
registerd_op
[
op
.
type
])
op_parser
.
apply
(
op
)
else
:
print
(
"Op[%s] has not been registered, so we have to skip it"
%
(
op
.
type
))
def
_parse_program
(
self
,
graph_name
,
program
,
input_varlist
=
[],
fetch_list
=
[]):
begin_graph_idx
=
self
.
graph_idx
ge_in_operator
=
[]
ge_out_operator
=
[]
self
.
var2geop
=
{}
block
=
program
.
global_block
()
if
len
(
block
.
ops
)
==
0
:
print
(
"There is no ops in program %s"
%
(
graph_name
))
return
[]
graph
=
core
.
GEGraph
(
graph_name
)
ge_in_operator
,
self
.
var2geop
=
self
.
_construct_input_map
(
input_varlist
)
self
.
parser_factory
=
ascend_parser
.
AscendParserFactory
(
graph
,
self
.
var2geop
)
for
i
,
curop
in
list
(
enumerate
(
block
.
ops
)):
self
.
parse_op
(
curop
)
# Set fetch_var for GE
for
e
in
fetch_list
:
name
=
e
if
not
isinstance
(
e
,
str
):
name
=
e
.
name
ge_out_operator
.
append
(
self
.
var2geop
[
name
])
# (Debug) If you want to print back prop vars, append/assign the varname in ge_out_operator here, such as:
# if graph_name == "main":
# ge_out_operator.append(self.var2geop["reduce_sum_0.tmp_0@GRAD"])
# Add ops that may be input of a graph, such as const.
for
varname
,
geop
in
self
.
var2geop
.
items
():
if
varname
.
startswith
(
"geinput"
):
ge_in_operator
.
append
(
geop
)
graph
.
set_inputs
(
ge_in_operator
).
set_outputs
(
ge_out_operator
)
# Remove ops of origin program
op_num
=
len
(
block
.
ops
)
for
i
in
range
(
op_num
-
1
,
-
1
,
-
1
):
block
.
_remove_op
(
i
)
input_varlist
=
[
var
for
var
in
input_varlist
if
var
.
is_data
]
block
.
append_op
(
type
=
"ascend_trigger"
,
inputs
=
{
"FeedList"
:
input_varlist
},
outputs
=
{
"FetchList"
:
fetch_list
},
attrs
=
{
'graph_idx'
:
self
.
graph_idx
})
self
.
graph_idx
+=
1
return
graph
def
parse_program
(
self
,
startup_program
,
main_program
,
input_varlist
,
fetch_list
):
startup_graph
=
self
.
_parse_program
(
"startup"
,
startup_program
)
main_graph
=
self
.
_parse_program
(
"main"
,
main_program
,
input_varlist
,
fetch_list
)
return
startup_graph
,
main_graph
# AscendOptimizer is a wrapper for basic optimizer now
# We will make it part of fleet meta_optimizer in the future
class
AscendOptimizer
(
Optimizer
):
def
__init__
(
self
,
optimizer
,
fetch_list
=
[]):
self
.
inner_opt
=
optimizer
self
.
fetch_list
=
fetch_list
def
__del__
(
self
):
core
.
ge_finalize
()
def
_can_apply
(
self
):
if
not
self
.
user_defined_strategy
.
ascend
:
return
False
# TODO(hutuxian): other check here
return
True
def
_disable_strategy
(
self
,
dist_strategy
):
dist_strategy
.
ascend
=
False
dist_strategy
.
ascend_configs
=
{}
def
_get_input_varlist
(
program
):
ret_list
=
[]
for
var
in
program
.
list_vars
():
if
var
.
is_data
or
var
.
persistable
:
ret_list
.
append
(
var
)
return
ret_list
def
minimize
(
self
,
loss
,
startup_program
=
None
,
parameter_list
=
None
,
no_grad_set
=
None
):
minimized
=
self
.
inner_opt
.
minimize
(
loss
,
startup_program
=
startup_program
)
self
.
ascend_instance
=
core
.
AscendInstance
()
# Config about Graph Engine can be found in https://support.huaweicloud.com/
config
=
{
"ge.exec.deviceId"
:
"0"
,
"ge.graphRunMode"
:
"1"
,
"ge.exec.precision_mode"
:
"must_keep_origin_dtype"
}
core
.
ge_initialize
(
config
)
# Init Session
self
.
ascend_instance
.
init_global_resources
()
main_block
=
loss
.
block
self
.
parser
=
AscendIRParser
()
input_varlist
=
_get_input_varlist
(
main_block
.
program
)
startup_graph
,
main_graph
=
self
.
parser
.
parse_program
(
startup_program
,
main_block
.
program
,
input_varlist
,
self
.
fetch_list
)
self
.
ascend_instance
.
add_ascend_subgraph
(
0
,
startup_graph
)
self
.
ascend_instance
.
add_ascend_subgraph
(
1
,
main_graph
)
return
minimized
python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py
0 → 100644
浏览文件 @
6dd52c5b
此差异已折叠。
点击以展开。
python/paddle/fluid/tests/unittests/test_ascend_trigger.py
0 → 100644
浏览文件 @
6dd52c5b
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import
paddle
import
paddle.fluid
as
fluid
import
paddle.fluid.layers
as
layers
import
unittest
class
TestAscendTriggerOP
(
unittest
.
TestCase
):
""" TestCases for ascend_trigger op"""
def
test_ascend_trigger_op
(
self
):
paddle
.
enable_static
()
program
=
fluid
.
Program
()
block
=
program
.
global_block
()
with
fluid
.
program_guard
(
program
):
x
=
fluid
.
data
(
name
=
'x'
,
shape
=
[
1
],
dtype
=
'int64'
,
lod_level
=
0
)
y
=
fluid
.
data
(
name
=
'y'
,
shape
=
[
1
],
dtype
=
'int64'
,
lod_level
=
0
)
block
.
append_op
(
type
=
"ascend_trigger"
,
inputs
=
{
"FeedList"
:
[
x
]},
outputs
=
{
"FetchList"
:
[
y
]},
attrs
=
{
'graph_idx'
:
0
})
exe
=
paddle
.
static
.
Executor
(
paddle
.
CPUPlace
())
try
:
exe
.
run
(
program
)
except
RuntimeError
as
e
:
pass
except
:
self
.
assertTrue
(
False
)
paddle
.
disable_static
()
if
__name__
==
'__main__'
:
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录