未验证 提交 2b40434e 编写于 作者: 张春乔 提交者: GitHub

昇腾和寒武纪相关代码退场 WITH_ASCEND_CL (#52612)

* mv WITH_ASCEND_CL

* mv WITH_ASCEND

* rollback
上级 ed9bac2f
...@@ -53,7 +53,7 @@ cc_test_old( ...@@ -53,7 +53,7 @@ cc_test_old(
if(WITH_DISTRIBUTE if(WITH_DISTRIBUTE
AND NOT WITH_PSLIB AND NOT WITH_PSLIB
AND NOT (WITH_ASCEND OR WITH_ASCEND_CL)) AND NOT (WITH_ASCEND))
set_source_files_properties( set_source_files_properties(
interceptor_ping_pong_with_brpc_test.cc interceptor_ping_pong_with_brpc_test.cc
PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS})
......
...@@ -123,16 +123,6 @@ if(NOT WIN32) ...@@ -123,16 +123,6 @@ if(NOT WIN32)
SRCS reducer.cc SRCS reducer.cc
DEPS layer) DEPS layer)
endif() endif()
if(WITH_ASCEND_CL)
cc_library(
hccl_context
SRCS hccl_context.cc
DEPS collective_helper device_context tensor var_type_traits)
cc_library(
reducer
SRCS reducer.cc
DEPS layer)
endif()
if(WITH_CNCL) if(WITH_CNCL)
cc_library( cc_library(
cncl_context cncl_context
...@@ -145,8 +135,7 @@ if(NOT WIN32) ...@@ -145,8 +135,7 @@ if(NOT WIN32)
endif() endif()
if(WITH_NCCL if(WITH_NCCL
OR WITH_RCCL OR WITH_RCCL
OR WITH_XPU_BKCL OR WITH_XPU_BKCL)
OR WITH_ASCEND_CL)
cc_library( cc_library(
heter_ccl_context heter_ccl_context
SRCS heter_ccl_context.cc SRCS heter_ccl_context.cc
...@@ -167,7 +156,6 @@ if(WITH_GLOO) ...@@ -167,7 +156,6 @@ if(WITH_GLOO)
(WITH_NCCL (WITH_NCCL
OR WITH_RCCL OR WITH_RCCL
OR WITH_XPU_BKCL OR WITH_XPU_BKCL
OR WITH_ASCEND_CL
OR WITH_CNCL) OR WITH_CNCL)
)) ))
cc_library( cc_library(
...@@ -177,32 +165,16 @@ if(WITH_GLOO) ...@@ -177,32 +165,16 @@ if(WITH_GLOO)
endif() endif()
endif() endif()
if(NOT WITH_ASCEND_CL) cc_library(
cc_library( gradient_accumulator
gradient_accumulator SRCS gradient_accumulator.cc
SRCS gradient_accumulator.cc DEPS blas
DEPS blas operator
operator lod_tensor
lod_tensor selected_rows_utils
selected_rows_utils selected_rows_functor
selected_rows_functor var_type_traits
var_type_traits layer
layer math_function
math_function phi_tensor
phi_tensor ${MLU_DEPS})
${MLU_DEPS})
else()
cc_library(
gradient_accumulator
SRCS gradient_accumulator.cc
DEPS blas
operator
lod_tensor
selected_rows_utils
selected_rows_functor
var_type_traits
layer
math_function
npu_op_runner
phi_tensor)
endif()
...@@ -1212,7 +1212,7 @@ inference_analysis_test( ...@@ -1212,7 +1212,7 @@ inference_analysis_test(
if(WITH_DISTRIBUTE if(WITH_DISTRIBUTE
AND WITH_PSCORE AND WITH_PSCORE
AND NOT (WITH_ASCEND OR WITH_ASCEND_CL)) AND NOT (WITH_ASCEND))
inference_analysis_test( inference_analysis_test(
test_analyzer_dist_model test_analyzer_dist_model
SRCS SRCS
......
...@@ -117,10 +117,6 @@ if (WITH_GPU OR WITH_ROCM) ...@@ -117,10 +117,6 @@ if (WITH_GPU OR WITH_ROCM)
endif() endif()
endif() endif()
if (WITH_ASCEND_CL)
op_library(sync_batch_norm_op)
endif()
op_library(lstm_op DEPS ${OP_HEADER_DEPS} lstm_compute) op_library(lstm_op DEPS ${OP_HEADER_DEPS} lstm_compute)
op_library(recurrent_op DEPS ${OP_HEADER_DEPS}) op_library(recurrent_op DEPS ${OP_HEADER_DEPS})
...@@ -161,11 +157,6 @@ if (WITH_ASCEND) ...@@ -161,11 +157,6 @@ if (WITH_ASCEND)
set(COMMON_OP_DEPS ${COMMON_OP_DEPS} ascend_wrapper) set(COMMON_OP_DEPS ${COMMON_OP_DEPS} ascend_wrapper)
endif() endif()
if (WITH_ASCEND_CL)
cc_test(assign_op_npu_test SRCS assign_op_npu_test.cc DEPS generated_static_op)
set(COMMON_OP_DEPS ${COMMON_OP_DEPS} npu_op_runner)
endif()
# FIXME(typhoonzero): operator deps may not needed. # FIXME(typhoonzero): operator deps may not needed.
# op_library(unsqueeze_op DEPS reshape_op) # op_library(unsqueeze_op DEPS reshape_op)
# op_library(squeeze_op DEPS reshape_op) # op_library(squeeze_op DEPS reshape_op)
...@@ -200,18 +191,10 @@ if (WITH_PYTHON) ...@@ -200,18 +191,10 @@ if (WITH_PYTHON)
cc_library(py_func_op SRCS py_func_op.cc DEPS op_registry python pybind) cc_library(py_func_op SRCS py_func_op.cc DEPS op_registry python pybind)
endif() endif()
if (WITH_ASCEND_CL)
cc_test(range_op_npu_test SRCS range_op_npu_test.cc DEPS op_registry range_op scope device_context enforce executor)
cc_test(expand_op_npu_test SRCS expand_op_npu_test.cc DEPS op_registry expand_op eigen_function scope device_context enforce executor compare_op)
endif()
set(GLOB_OP_LIB ${OP_LIBRARY} CACHE INTERNAL "Global OP library") set(GLOB_OP_LIB ${OP_LIBRARY} CACHE INTERNAL "Global OP library")
add_subdirectory(benchmark) add_subdirectory(benchmark)
cc_test_old(op_debug_string_test SRCS op_debug_string_test.cc DEPS elementwise_add_op ${COMMON_OP_DEPS}) cc_test_old(op_debug_string_test SRCS op_debug_string_test.cc DEPS elementwise_add_op ${COMMON_OP_DEPS})
if (WITH_ASCEND_CL)
cc_test(transpose_op_npu_test SRCS transpose_op_npu_test.cc DEPS op_registry transpose_op scope device_context enforce executor)
endif()
if(WITH_MKLDNN) if(WITH_MKLDNN)
...@@ -227,7 +210,7 @@ if(WITH_UNITY_BUILD) ...@@ -227,7 +210,7 @@ if(WITH_UNITY_BUILD)
target_link_libraries(paddle_operators_unity ${OP_HEADER_DEPS} ${COMMON_OP_DEPS}) target_link_libraries(paddle_operators_unity ${OP_HEADER_DEPS} ${COMMON_OP_DEPS})
endif() endif()
if (WITH_GPU OR WITH_ASCEND_CL) if (WITH_GPU)
cc_test(copy_cross_scope_test SRCS copy_cross_scope_test.cc DEPS op_registry copy_cross_scope_op scope device_context enforce executor) cc_test(copy_cross_scope_test SRCS copy_cross_scope_test.cc DEPS op_registry copy_cross_scope_op scope device_context enforce executor)
endif() endif()
......
...@@ -4,11 +4,3 @@ if(WITH_UNITY_BUILD) ...@@ -4,11 +4,3 @@ if(WITH_UNITY_BUILD)
include(unity_build_rule.cmake) include(unity_build_rule.cmake)
endif() endif()
register_operators() register_operators()
if(WITH_ASCEND_CL)
cc_test(
check_finite_and_unscale_op_npu_test
SRCS check_finite_and_unscale_op_npu_test.cc
DEPS op_registry check_finite_and_unscale_op scope device_context enforce
executor)
endif()
...@@ -51,106 +51,9 @@ if(WITH_CNCL) ...@@ -51,106 +51,9 @@ if(WITH_CNCL)
op_library(c_gen_cncl_id_op DEPS ${COLLECTIVE_DEPS}) op_library(c_gen_cncl_id_op DEPS ${COLLECTIVE_DEPS})
endif() endif()
if(WITH_ASCEND_CL)
cc_library(
gen_hccl_id_op_helper
SRCS gen_hccl_id_op_helper.cc
DEPS dynload_warpctc dynamic_loader scope)
set(COLLECTIVE_DEPS ${COLLECTIVE_DEPS} collective_helper
gen_hccl_id_op_helper)
op_library(c_gen_hccl_id_op DEPS ${COLLECTIVE_DEPS})
op_library(gen_hccl_id_op DEPS ${COLLECTIVE_DEPS})
endif()
set(OPERATOR_DEPS set(OPERATOR_DEPS
${OPERATOR_DEPS} ${COLLECTIVE_DEPS} ${OPERATOR_DEPS} ${COLLECTIVE_DEPS}
PARENT_SCOPE) PARENT_SCOPE)
set(GLOB_COLLECTIVE_DEPS set(GLOB_COLLECTIVE_DEPS
${COLLECTIVE_DEPS} ${COLLECTIVE_DEPS}
CACHE INTERNAL "collective dependency") CACHE INTERNAL "collective dependency")
if(WITH_ASCEND_CL)
set(COMMON_TEST_DEPS_FOR_HCOM
c_comm_init_hccl_op
c_gen_hccl_id_op
gen_hccl_id_op_helper
gen_hccl_id_op
op_registry
ascend_hccl
flags
dynamic_loader
dynload_warpctc
scope
device_context
enforce
executor)
cc_test(
c_broadcast_op_npu_test
SRCS c_broadcast_op_npu_test.cc
DEPS c_broadcast_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
c_allreduce_sum_op_npu_test
SRCS c_allreduce_sum_op_npu_test.cc
DEPS c_allreduce_sum_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
c_reducescatter_op_npu_test
SRCS c_reducescatter_op_npu_test.cc
DEPS c_reducescatter_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
c_allgather_op_npu_test
SRCS c_allgather_op_npu_test.cc
DEPS c_allgather_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
c_reduce_sum_op_npu_test
SRCS c_reduce_sum_op_npu_test.cc
DEPS c_reduce_sum_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
c_allreduce_max_op_npu_test
SRCS c_allreduce_max_op_npu_test.cc
DEPS c_allreduce_max_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
send_v2_op_npu_test
SRCS send_v2_op_npu_test.cc
DEPS send_v2_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
recv_v2_op_npu_test
SRCS recv_v2_op_npu_test.cc
DEPS recv_v2_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
checknumeric
SRCS checknumeric_npu_test.cc
DEPS c_allreduce_sum_op ${COLLECTIVE_DEPS} ${COMMON_TEST_DEPS_FOR_HCOM})
cc_test(
c_sync_comm_stream_op_npu_test
SRCS c_sync_comm_stream_op_npu_test.cc
DEPS op_registry
c_broadcast_op
c_comm_init_hccl_op
c_sync_comm_stream_op
c_gen_hccl_id_op
gen_hccl_id_op_helper
${COLLECTIVE_DEPS}
ascend_hccl
dynamic_loader
dynload_warpctc
scope
device_context
enforce
executor)
cc_test(
c_sync_calc_stream_op_npu_test
SRCS c_sync_calc_stream_op_npu_test.cc
DEPS op_registry
elementwise_add_op
c_sync_calc_stream_op
c_gen_hccl_id_op
gen_hccl_id_op_helper
${COLLECTIVE_DEPS}
ascend_hccl
dynamic_loader
dynload_warpctc
scope
device_context
enforce
executor)
endif()
...@@ -17,16 +17,3 @@ cc_test( ...@@ -17,16 +17,3 @@ cc_test(
test_elementwise_add_grad_grad test_elementwise_add_grad_grad
SRCS test_elementwise_add_grad_grad.cc SRCS test_elementwise_add_grad_grad.cc
DEPS op_registry elementwise_add_op scope device_context enforce executor) DEPS op_registry elementwise_add_op scope device_context enforce executor)
if(WITH_ASCEND_CL)
cc_test(
elementwise_op_npu_test
SRCS elementwise_op_npu_test.cc
DEPS op_registry
elementwise_add_op
elementwise_sub_op
scope
device_context
enforce
executor)
endif()
...@@ -16,9 +16,7 @@ math_library(sampler DEPS generator) ...@@ -16,9 +16,7 @@ math_library(sampler DEPS generator)
# math_library(math_function DEPS blas dense_tensor tensor) # math_library(math_function DEPS blas dense_tensor tensor)
math_library(sequence_pooling DEPS math_function jit_kernel_helper) math_library(sequence_pooling DEPS math_function jit_kernel_helper)
if(WITH_ASCEND_CL) if(WITH_XPU)
math_library(beam_search DEPS math_function beam_search_npu)
elseif(WITH_XPU)
math_library(beam_search DEPS math_function beam_search_xpu) math_library(beam_search DEPS math_function beam_search_xpu)
else() else()
math_library(beam_search DEPS math_function) math_library(beam_search DEPS math_function)
......
...@@ -33,10 +33,3 @@ if(WITH_ROCM) ...@@ -33,10 +33,3 @@ if(WITH_ROCM)
SRCS check_reduce_rank_test.cu SRCS check_reduce_rank_test.cu
DEPS tensor) DEPS tensor)
endif() endif()
if(WITH_ASCEND_CL)
cc_test(
reduce_any_op_npu_test
SRCS reduce_any_op_npu_test.cc
DEPS op_registry reduce_any_op scope device_context enforce executor)
endif()
...@@ -74,17 +74,6 @@ else() ...@@ -74,17 +74,6 @@ else()
set(IPU_CTX_DEPS) set(IPU_CTX_DEPS)
endif() endif()
if(WITH_ASCEND_CL)
set(NPU_CTX_DEPS npu_stream npu_info)
endif()
if(WITH_ASCEND_CL)
cc_library(
stream_callback_manager
SRCS stream_callback_manager.cc
DEPS simple_threadpool enforce)
endif()
if(WITH_GPU) if(WITH_GPU)
nv_library( nv_library(
stream_callback_manager stream_callback_manager
...@@ -100,8 +89,6 @@ endif() ...@@ -100,8 +89,6 @@ endif()
if(WITH_GPU OR WITH_ROCM) if(WITH_GPU OR WITH_ROCM)
set(STREAM_CALLBACK_DEPS stream_callback_manager) set(STREAM_CALLBACK_DEPS stream_callback_manager)
elseif(WITH_ASCEND_CL)
set(STREAM_CALLBACK_DEPS stream_callback_manager)
else() else()
set(STREAM_CALLBACK_DEPS) set(STREAM_CALLBACK_DEPS)
endif() endif()
...@@ -151,9 +138,6 @@ cc_library( ...@@ -151,9 +138,6 @@ cc_library(
collective_helper collective_helper
SRCS collective_helper.cc gen_comm_id_helper.cc SRCS collective_helper.cc gen_comm_id_helper.cc
DEPS framework_proto device_context enforce) DEPS framework_proto device_context enforce)
if(WITH_ASCEND_CL)
target_link_libraries(collective_helper npu_collective_helper)
endif()
if(WITH_CNCL) if(WITH_CNCL)
target_link_libraries(collective_helper mlu_collective_helper) target_link_libraries(collective_helper mlu_collective_helper)
...@@ -163,10 +147,6 @@ if(WITH_GPU OR WITH_ROCM) ...@@ -163,10 +147,6 @@ if(WITH_GPU OR WITH_ROCM)
target_link_libraries(device_context gpu_resource_pool) target_link_libraries(device_context gpu_resource_pool)
endif() endif()
if(WITH_ASCEND_CL)
target_link_libraries(device_context npu_resource_pool)
endif()
if(WITH_XPU) if(WITH_XPU)
target_link_libraries(device_context xpu_resource_pool) target_link_libraries(device_context xpu_resource_pool)
endif() endif()
...@@ -190,16 +170,6 @@ set(DEVICE_EVENT_LIBS ...@@ -190,16 +170,6 @@ set(DEVICE_EVENT_LIBS
device_event_base device_event_base
CACHE INTERNAL "device event libs") CACHE INTERNAL "device event libs")
if(WITH_ASCEND_CL)
cc_library(
device_event_npu
SRCS device_event_npu.cc
DEPS device_event_base npu_resource_pool)
set(DEVICE_EVENT_LIBS
device_event_npu
CACHE INTERNAL "device event libs")
endif()
if(WITH_GPU) if(WITH_GPU)
nv_library( nv_library(
device_event_gpu device_event_gpu
......
...@@ -62,11 +62,6 @@ if(WITH_ROCM) ...@@ -62,11 +62,6 @@ if(WITH_ROCM)
dynload_warpctc dynload_warpctc
SRCS warpctc.cc SRCS warpctc.cc
DEPS dynamic_loader warpctc phi_dynload_warpctc) DEPS dynamic_loader warpctc phi_dynload_warpctc)
elseif(WITH_ASCEND_CL)
cc_library(
dynload_warpctc
SRCS warpctc.cc
DEPS dynamic_loader warpctc npu_hccl phi_dynload_warpctc)
else() else()
nv_library( nv_library(
dynload_cuda dynload_cuda
......
...@@ -91,12 +91,6 @@ if(WITH_XPU_BKCL) ...@@ -91,12 +91,6 @@ if(WITH_XPU_BKCL)
set(PYBIND_DEPS ${PYBIND_DEPS} heter_ccl_context) set(PYBIND_DEPS ${PYBIND_DEPS} heter_ccl_context)
endif() endif()
if(WITH_ASCEND_CL)
set(PYBIND_DEPS ${PYBIND_DEPS} reducer)
set(PYBIND_DEPS ${PYBIND_DEPS} hccl_context)
set(PYBIND_DEPS ${PYBIND_DEPS} heter_ccl_context)
endif()
if(WITH_CNCL) if(WITH_CNCL)
set(PYBIND_DEPS ${PYBIND_DEPS} reducer) set(PYBIND_DEPS ${PYBIND_DEPS} reducer)
set(PYBIND_DEPS ${PYBIND_DEPS} cncl_context) set(PYBIND_DEPS ${PYBIND_DEPS} cncl_context)
...@@ -249,28 +243,16 @@ endif() ...@@ -249,28 +243,16 @@ endif()
if(WITH_PYTHON) if(WITH_PYTHON)
# generate op pybind functions automatically for dygraph. # generate op pybind functions automatically for dygraph.
if(WITH_ASCEND_CL)
set(OP_FUNCTION_GENERETOR_DEPS set(OP_FUNCTION_GENERETOR_DEPS
pybind pybind
proto_desc proto_desc
executor executor
layer layer
tracer tracer
engine engine
imperative_profiler imperative_profiler
imperative_flag imperative_flag)
ascend_wrapper)
else()
set(OP_FUNCTION_GENERETOR_DEPS
pybind
proto_desc
executor
layer
tracer
engine
imperative_profiler
imperative_flag)
endif()
list(APPEND OP_FUNCTION_GENERETOR_DEPS ${GLOB_OP_LIB}) list(APPEND OP_FUNCTION_GENERETOR_DEPS ${GLOB_OP_LIB})
list(APPEND OP_FUNCTION_GENERETOR_DEPS ${GLOB_OPERATOR_DEPS}) list(APPEND OP_FUNCTION_GENERETOR_DEPS ${GLOB_OPERATOR_DEPS})
...@@ -282,10 +264,6 @@ if(WITH_PYTHON) ...@@ -282,10 +264,6 @@ if(WITH_PYTHON)
list(APPEND OP_FUNCTION_GENERETOR_DEPS bkcl_context) list(APPEND OP_FUNCTION_GENERETOR_DEPS bkcl_context)
endif() endif()
if(WITH_ASCEND_CL)
list(APPEND OP_FUNCTION_GENERETOR_DEPS hccl_context)
endif()
if(WITH_CNCL) if(WITH_CNCL)
list(APPEND OP_FUNCTION_GENERETOR_DEPS cncl_context) list(APPEND OP_FUNCTION_GENERETOR_DEPS cncl_context)
endif() endif()
......
...@@ -6,8 +6,6 @@ if(WITH_GPU) ...@@ -6,8 +6,6 @@ if(WITH_GPU)
set(PACKAGE_NAME "paddlepaddle-gpu") set(PACKAGE_NAME "paddlepaddle-gpu")
elseif(WITH_ROCM) elseif(WITH_ROCM)
set(PACKAGE_NAME "paddlepaddle-rocm") set(PACKAGE_NAME "paddlepaddle-rocm")
elseif(WITH_ASCEND_CL)
set(PACKAGE_NAME "paddlepaddle-npu")
elseif(WITH_XPU) elseif(WITH_XPU)
set(PACKAGE_NAME "paddlepaddle-xpu") set(PACKAGE_NAME "paddlepaddle-xpu")
elseif(WITH_IPU) elseif(WITH_IPU)
......
...@@ -17,7 +17,7 @@ string(REPLACE ".py" "" DIST_TEST_OPS "${DIST_TEST_OPS}") ...@@ -17,7 +17,7 @@ string(REPLACE ".py" "" DIST_TEST_OPS "${DIST_TEST_OPS}")
if((NOT WITH_GPU) if((NOT WITH_GPU)
AND (NOT WITH_XPU) AND (NOT WITH_XPU)
AND NOT (WITH_ASCEND OR WITH_ASCEND_CL)) AND NOT (WITH_ASCEND))
list(REMOVE_ITEM DIST_TEST_OPS "test_dist_mnist_batch_merge") list(REMOVE_ITEM DIST_TEST_OPS "test_dist_mnist_batch_merge")
endif() endif()
list(APPEND DIST_TEST_OPS test_parallel_dygraph_dataparallel) list(APPEND DIST_TEST_OPS test_parallel_dygraph_dataparallel)
...@@ -272,47 +272,25 @@ function(py_test_modules TARGET_NAME) ...@@ -272,47 +272,25 @@ function(py_test_modules TARGET_NAME)
if(WITH_COVERAGE AND NOT (WITH_INCREMENTAL_COVERAGE if(WITH_COVERAGE AND NOT (WITH_INCREMENTAL_COVERAGE
AND "$ENV{PADDLE_GIT_DIFF_PY_FILE}" STREQUAL "")) AND "$ENV{PADDLE_GIT_DIFF_PY_FILE}" STREQUAL ""))
if(WITH_ASCEND_CL)
add_test( add_test(
NAME ${TARGET_NAME} NAME ${TARGET_NAME}
COMMAND COMMAND
${CMAKE_COMMAND} -E env ${CMAKE_COMMAND} -E env PYTHONPATH=${PADDLE_BINARY_DIR}/python
PYTHONPATH=${PADDLE_BINARY_DIR}/python:$ENV{PYTHONPATH} ${py_test_modules_ENVS}
${py_test_modules_ENVS} COVERAGE_FILE=${PADDLE_BINARY_DIR}/python-coverage.data
COVERAGE_FILE=${PADDLE_BINARY_DIR}/python-coverage.data ${PYTHON_EXECUTABLE} -m coverage run --branch -p
${PYTHON_EXECUTABLE} -m coverage run --branch -p ${PADDLE_SOURCE_DIR}/tools/test_runner.py ${py_test_modules_MODULES}
${PADDLE_SOURCE_DIR}/tools/test_runner.py ${py_test_modules_MODULES} WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR})
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR})
else()
add_test(
NAME ${TARGET_NAME}
COMMAND
${CMAKE_COMMAND} -E env PYTHONPATH=${PADDLE_BINARY_DIR}/python
${py_test_modules_ENVS}
COVERAGE_FILE=${PADDLE_BINARY_DIR}/python-coverage.data
${PYTHON_EXECUTABLE} -m coverage run --branch -p
${PADDLE_SOURCE_DIR}/tools/test_runner.py ${py_test_modules_MODULES}
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR})
endif()
else() else()
if(WITH_ASCEND_CL) add_test(
add_test( NAME ${TARGET_NAME}
NAME ${TARGET_NAME} COMMAND
COMMAND ${CMAKE_COMMAND} -E env PYTHONPATH=${PADDLE_BINARY_DIR}/python
${CMAKE_COMMAND} -E env ${py_test_modules_ENVS} ${PYTHON_EXECUTABLE}
PYTHONPATH=${PADDLE_BINARY_DIR}/python:$ENV{PYTHONPATH} ${PADDLE_SOURCE_DIR}/tools/test_runner.py ${py_test_modules_MODULES}
${py_test_modules_ENVS} ${PYTHON_EXECUTABLE} WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR})
${PADDLE_SOURCE_DIR}/tools/test_runner.py ${py_test_modules_MODULES}
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR})
else()
add_test(
NAME ${TARGET_NAME}
COMMAND
${CMAKE_COMMAND} -E env PYTHONPATH=${PADDLE_BINARY_DIR}/python
${py_test_modules_ENVS} ${PYTHON_EXECUTABLE}
${PADDLE_SOURCE_DIR}/tools/test_runner.py ${py_test_modules_MODULES}
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR})
endif()
endif() endif()
if(py_test_modules_SERIAL) if(py_test_modules_SERIAL)
...@@ -517,7 +495,7 @@ endforeach() ...@@ -517,7 +495,7 @@ endforeach()
if((NOT WITH_GPU) if((NOT WITH_GPU)
AND (NOT WITH_XPU) AND (NOT WITH_XPU)
AND NOT (WITH_ASCEND OR WITH_ASCEND_CL)) AND NOT (WITH_ASCEND))
list(REMOVE_ITEM TEST_OPS "test_dist_mnist_batch_merge") list(REMOVE_ITEM TEST_OPS "test_dist_mnist_batch_merge")
endif() endif()
...@@ -530,7 +508,6 @@ py_test_modules(test_adam_op_multi_thread MODULES test_adam_op ENVS ...@@ -530,7 +508,6 @@ py_test_modules(test_adam_op_multi_thread MODULES test_adam_op ENVS
if(WITH_GPU if(WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND
OR WITH_ASCEND_CL
OR APPLE) OR APPLE)
py_test_modules(test_warpctc_op MODULES test_warpctc_op) py_test_modules(test_warpctc_op MODULES test_warpctc_op)
set_tests_properties(test_warpctc_op PROPERTIES TIMEOUT 120) set_tests_properties(test_warpctc_op PROPERTIES TIMEOUT 120)
...@@ -668,8 +645,7 @@ if(WITH_DISTRIBUTE) ...@@ -668,8 +645,7 @@ if(WITH_DISTRIBUTE)
PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR}) PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR})
if(WITH_GPU if(WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL)
bash_test_modules( bash_test_modules(
test_fleet_launch_nproc START_BASH test_fleet_launch_nproc.sh ENVS test_fleet_launch_nproc START_BASH test_fleet_launch_nproc.sh ENVS
PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR}) PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR})
...@@ -683,7 +659,7 @@ if(WITH_DISTRIBUTE) ...@@ -683,7 +659,7 @@ if(WITH_DISTRIBUTE)
test_fleet_launch_cloud START_BASH test_fleet_launch_cloud.sh ENVS test_fleet_launch_cloud START_BASH test_fleet_launch_cloud.sh ENVS
PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR}) PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR})
endif() endif()
if(WITH_ASCEND OR WITH_ASCEND_CL) if(WITH_ASCEND)
bash_test_modules( bash_test_modules(
test_fleet_launch_ascend START_BASH test_fleet_launch_ascend.sh ENVS test_fleet_launch_ascend START_BASH test_fleet_launch_ascend.sh ENVS
PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR}) PADDLE_BINARY_DIR=${PADDLE_BINARY_DIR})
......
...@@ -370,9 +370,7 @@ if((WITH_GPU OR WITH_ROCM) AND (LINUX)) ...@@ -370,9 +370,7 @@ if((WITH_GPU OR WITH_ROCM) AND (LINUX))
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_ROCM OR WITH_ROCM
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_gen_nccl_id_op test_gen_nccl_id_op
......
...@@ -6,9 +6,7 @@ set(LOCAL_ALL_ARCH ON) ...@@ -6,9 +6,7 @@ set(LOCAL_ALL_ARCH ON)
set(LOCAL_ALL_PLAT ON) set(LOCAL_ALL_PLAT ON)
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
py_test_modules( py_test_modules(
test_fleet_sharding_meta_optimizer MODULES test_fleet_sharding_meta_optimizer MODULES
...@@ -85,9 +83,7 @@ if(LOCAL_ALL_ARCH AND (LINUX OR WIN32)) ...@@ -85,9 +83,7 @@ if(LOCAL_ALL_ARCH AND (LINUX OR WIN32))
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_rnn_dp test_rnn_dp
...@@ -180,9 +176,7 @@ if((WITH_GPU) AND LOCAL_ALL_PLAT) ...@@ -180,9 +176,7 @@ if((WITH_GPU) AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND (LINUX)) AND (LINUX))
py_test_modules( py_test_modules(
test_fleet_localsgd_meta_optimizer MODULES test_fleet_localsgd_meta_optimizer MODULES
...@@ -276,9 +270,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT) ...@@ -276,9 +270,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_fleet_lars_meta_optimizer test_fleet_lars_meta_optimizer
...@@ -331,9 +323,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT) ...@@ -331,9 +323,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_fleet_pipeline_meta_optimizer test_fleet_pipeline_meta_optimizer
...@@ -347,9 +337,7 @@ if((WITH_GPU ...@@ -347,9 +337,7 @@ if((WITH_GPU
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
py_test_modules( py_test_modules(
test_fleet_gradient_merge_meta_optimizer MODULES test_fleet_gradient_merge_meta_optimizer MODULES
...@@ -368,9 +356,7 @@ if(LOCAL_ALL_ARCH AND (LINUX OR WIN32)) ...@@ -368,9 +356,7 @@ if(LOCAL_ALL_ARCH AND (LINUX OR WIN32))
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_fleet_raw_program_meta_optimizer test_fleet_raw_program_meta_optimizer
...@@ -449,9 +435,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT) ...@@ -449,9 +435,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND (LINUX)) AND (LINUX))
py_test_modules( py_test_modules(
test_fleet_lamb_meta_optimizer MODULES test_fleet_lamb_meta_optimizer ENVS test_fleet_lamb_meta_optimizer MODULES test_fleet_lamb_meta_optimizer ENVS
...@@ -479,9 +463,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT) ...@@ -479,9 +463,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_fleet_pipeline_meta_optimizer_with_recompute test_fleet_pipeline_meta_optimizer_with_recompute
...@@ -495,9 +477,7 @@ if((WITH_GPU ...@@ -495,9 +477,7 @@ if((WITH_GPU
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND (LINUX OR WIN32)) AND (LINUX OR WIN32))
py_test_modules( py_test_modules(
test_fleet_hybrid_meta_optimizer MODULES test_fleet_hybrid_meta_optimizer test_fleet_hybrid_meta_optimizer MODULES test_fleet_hybrid_meta_optimizer
...@@ -548,9 +528,7 @@ if((WITH_ROCM) AND LOCAL_ALL_PLAT) ...@@ -548,9 +528,7 @@ if((WITH_ROCM) AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
py_test_modules( py_test_modules(
test_fleet_amp_meta_optimizer MODULES test_fleet_amp_meta_optimizer ENVS test_fleet_amp_meta_optimizer MODULES test_fleet_amp_meta_optimizer ENVS
...@@ -605,9 +583,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT) ...@@ -605,9 +583,7 @@ if(LOCAL_ALL_ARCH AND LOCAL_ALL_PLAT)
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND (LINUX OR WIN32)) AND (LINUX OR WIN32))
py_test_modules( py_test_modules(
test_fleet_recompute_meta_optimizer MODULES test_fleet_recompute_meta_optimizer MODULES
...@@ -621,9 +597,7 @@ if(LOCAL_ALL_ARCH AND (LINUX OR WIN32)) ...@@ -621,9 +597,7 @@ if(LOCAL_ALL_ARCH AND (LINUX OR WIN32))
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND LOCAL_ALL_PLAT) AND LOCAL_ALL_PLAT)
bash_test_modules( bash_test_modules(
test_new_group test_new_group
...@@ -636,9 +610,7 @@ if((WITH_GPU ...@@ -636,9 +610,7 @@ if((WITH_GPU
endif() endif()
if((WITH_GPU if((WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL
)
AND (LINUX)) AND (LINUX))
bash_test_modules( bash_test_modules(
test_c_comm_init_op test_c_comm_init_op
......
...@@ -6,7 +6,7 @@ string(REPLACE ".py" "" TEST_OPS "${TEST_OPS}") ...@@ -6,7 +6,7 @@ string(REPLACE ".py" "" TEST_OPS "${TEST_OPS}")
if((NOT WITH_GPU) if((NOT WITH_GPU)
AND (NOT WITH_XPU) AND (NOT WITH_XPU)
AND NOT (WITH_ASCEND OR WITH_ASCEND_CL)) AND NOT (WITH_ASCEND))
list(REMOVE_ITEM TEST_OPS "test_dist_fuse_adam_pass") list(REMOVE_ITEM TEST_OPS "test_dist_fuse_adam_pass")
list(REMOVE_ITEM TEST_OPS "test_dist_fuse_all_reduce_pass") list(REMOVE_ITEM TEST_OPS "test_dist_fuse_all_reduce_pass")
list(REMOVE_ITEM TEST_OPS "test_dist_fuse_bn_act_pass") list(REMOVE_ITEM TEST_OPS "test_dist_fuse_bn_act_pass")
......
...@@ -26,8 +26,7 @@ if((WITH_DISTRIBUTE) ...@@ -26,8 +26,7 @@ if((WITH_DISTRIBUTE)
AND (NOT APPLE)) AND (NOT APPLE))
if(WITH_GPU if(WITH_GPU
OR WITH_XPU OR WITH_XPU
OR WITH_ASCEND OR WITH_ASCEND)
OR WITH_ASCEND_CL)
py_test_modules(test_fleet_with_asp_sharding MODULES py_test_modules(test_fleet_with_asp_sharding MODULES
test_fleet_with_asp_sharding ENVS ${dist_ENVS}) test_fleet_with_asp_sharding ENVS ${dist_ENVS})
endif() endif()
......
#!/usr/bin/env bash #!/usr/bin/env bash
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. # Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
# #
# Licensed under the Apache License, Version 2.0 (the "License"); # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License. # you may not use this file except in compliance with the License.
# You may obtain a copy of the License at # You may obtain a copy of the License at
# #
# http://www.apache.org/licenses/LICENSE-2.0 # http://www.apache.org/licenses/LICENSE-2.0
# #
# Unless required by applicable law or agreed to in writing, software # Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, # distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
...@@ -20,7 +20,7 @@ PADDLE_ROOT="$( cd "$( dirname "${BASH_SOURCE[0]}")/../../" && pwd )" ...@@ -20,7 +20,7 @@ PADDLE_ROOT="$( cd "$( dirname "${BASH_SOURCE[0]}")/../../" && pwd )"
# install lcov # install lcov
if [ ! -f "/root/.cache/lcov-1.14.tar.gz" ];then if [ ! -f "/root/.cache/lcov-1.14.tar.gz" ];then
wget -P /home https://paddle-ci.gz.bcebos.com/coverage/lcov-1.14.tar.gz --no-proxy --no-check-certificate || exit 101 wget -P /home https://paddle-ci.gz.bcebos.com/coverage/lcov-1.14.tar.gz --no-proxy --no-check-certificate || exit 101
cp /home/lcov-1.14.tar.gz /root/.cache/lcov-1.14.tar.gz cp /home/lcov-1.14.tar.gz /root/.cache/lcov-1.14.tar.gz
else else
cp /root/.cache/lcov-1.14.tar.gz /home/lcov-1.14.tar.gz cp /root/.cache/lcov-1.14.tar.gz /home/lcov-1.14.tar.gz
...@@ -113,8 +113,6 @@ function gen_full_html_report_npu() { ...@@ -113,8 +113,6 @@ function gen_full_html_report_npu() {
if [ ${WITH_XPU:-OFF} == "ON" ]; then if [ ${WITH_XPU:-OFF} == "ON" ]; then
gen_full_html_report_xpu || true gen_full_html_report_xpu || true
elif [ ${WITH_ASCEND_CL:-OFF} == "ON" ]; then
gen_full_html_report_npu || true
else else
gen_full_html_report || true gen_full_html_report || true
fi fi
...@@ -213,8 +211,6 @@ echo "Assert Python Diff Coverage" ...@@ -213,8 +211,6 @@ echo "Assert Python Diff Coverage"
if [ ${WITH_XPU:-OFF} == "ON" ]; then if [ ${WITH_XPU:-OFF} == "ON" ]; then
echo "XPU has no python coverage!" echo "XPU has no python coverage!"
elif [ ${WITH_ASCEND_CL:-OFF} == "ON" ]; then
echo "NPU has no python coverage!"
else else
if [[ "${NO_PYTHON_COVERAGE_DATA}" != "1" ]];then if [[ "${NO_PYTHON_COVERAGE_DATA}" != "1" ]];then
python3.7 ${PADDLE_ROOT}/tools/coverage/coverage_lines.py python-coverage-diff.info 0.9 || PYTHON_COVERAGE_LINES_ASSERT=1 python3.7 ${PADDLE_ROOT}/tools/coverage/coverage_lines.py python-coverage-diff.info 0.9 || PYTHON_COVERAGE_LINES_ASSERT=1
......
#!/usr/bin/env bash #!/usr/bin/env bash
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. # Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
# #
# Licensed under the Apache License, Version 2.0 (the "License"); # Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License. # you may not use this file except in compliance with the License.
# You may obtain a copy of the License at # You may obtain a copy of the License at
# #
# http://www.apache.org/licenses/LICENSE-2.0 # http://www.apache.org/licenses/LICENSE-2.0
# #
# Unless required by applicable law or agreed to in writing, software # Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, # distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
...@@ -21,7 +21,7 @@ PADDLE_ROOT="$( cd "$( dirname "${BASH_SOURCE[0]}")/../../" && pwd )" ...@@ -21,7 +21,7 @@ PADDLE_ROOT="$( cd "$( dirname "${BASH_SOURCE[0]}")/../../" && pwd )"
function lcov_init(){ function lcov_init(){
# install lcov # install lcov
if [ ! -f "/root/.cache/lcov-1.14.tar.gz" ];then if [ ! -f "/root/.cache/lcov-1.14.tar.gz" ];then
wget -P /home https://paddle-ci.gz.bcebos.com/coverage/lcov-1.14.tar.gz --no-proxy --no-check-certificate || exit 101 wget -P /home https://paddle-ci.gz.bcebos.com/coverage/lcov-1.14.tar.gz --no-proxy --no-check-certificate || exit 101
cp /home/lcov-1.14.tar.gz /root/.cache/lcov-1.14.tar.gz cp /home/lcov-1.14.tar.gz /root/.cache/lcov-1.14.tar.gz
else else
cp /root/.cache/lcov-1.14.tar.gz /home/lcov-1.14.tar.gz cp /root/.cache/lcov-1.14.tar.gz /home/lcov-1.14.tar.gz
...@@ -230,7 +230,7 @@ function covinfo_combine_full(){ ...@@ -230,7 +230,7 @@ function covinfo_combine_full(){
mv infer-python-coverage.info python-coverage.info mv infer-python-coverage.info python-coverage.info
else else
echo "Cannot found python coverage.info" echo "Cannot found python coverage.info"
fi fi
gen_python_full_html_report || true gen_python_full_html_report || true
gen_full_html_report || true gen_full_html_report || true
} }
...@@ -243,8 +243,6 @@ function cov_rate_judge(){ ...@@ -243,8 +243,6 @@ function cov_rate_judge(){
if [ ${WITH_XPU:-OFF} == "ON" ]; then if [ ${WITH_XPU:-OFF} == "ON" ]; then
echo "XPU has no python coverage!" echo "XPU has no python coverage!"
elif [ ${WITH_ASCEND_CL:-OFF} == "ON" ]; then
echo "NPU has no python coverage!"
else else
if [[ python-coverage-diff.info ]];then if [[ python-coverage-diff.info ]];then
python ${PADDLE_ROOT}/tools/coverage/coverage_lines.py python-coverage-diff.info 0.9 || PYTHON_COVERAGE_LINES_ASSERT=1 python ${PADDLE_ROOT}/tools/coverage/coverage_lines.py python-coverage-diff.info 0.9 || PYTHON_COVERAGE_LINES_ASSERT=1
...@@ -269,7 +267,7 @@ function print_usage() { ...@@ -269,7 +267,7 @@ function print_usage() {
function main () { function main () {
local CMD=$1 local CMD=$1
lcov_init lcov_init
case $CMD in case $CMD in
gen_cov_info) gen_cov_info)
gen_cpp_covinfo gen_cpp_covinfo
gen_py_covinfo gen_py_covinfo
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册