#windows treat symbolic file as a real file, which is different with unix #We create a hidden file and compile it instead of origin source file. function(windows_symbolic TARGET) set(oneValueArgs "") set(multiValueArgs SRCS PATH) cmake_parse_arguments(windows_symbolic "${options}" "${oneValueArgs}" "${multiValueArgs}" ${ARGN}) set(final_path ${CMAKE_CURRENT_SOURCE_DIR}/${windows_symbolic_PATH}) foreach(src ${windows_symbolic_SRCS}) get_filename_component(src ${src} NAME_WE) if(NOT EXISTS ${final_path}/${src}.cc OR NOT EXISTS ${final_path}/${src}.cu) message( FATAL " ${src}.cc and ${src}.cu must exsits, and ${src}.cu must be symbolic file." ) endif() file( GENERATE OUTPUT ${final_path}/.${src}.cu INPUT ${final_path}/${src}.cc) add_custom_command( OUTPUT ${final_path}/.${src}.cu COMMAND ${CMAKE_COMMAND} -E copy_if_different "${final_path}/${src}.cc" "${final_path}/.${src}.cu" COMMENT "create hidden file of ${src}.cu") add_custom_target(${TARGET} ALL DEPENDS ${final_path}/.${src}.cu) endforeach() endfunction() # Usage: pass_library(target inference) will append to paddle_inference_pass.h set(pass_file ${PADDLE_BINARY_DIR}/paddle/fluid/inference/api/paddle_inference_pass.h.tmp) set(pass_file_final ${PADDLE_BINARY_DIR}/paddle/fluid/inference/api/paddle_inference_pass.h) file( WRITE ${pass_file} "// Generated by the paddle/fluid/framework/ir/CMakeLists.txt. DO NOT EDIT!\n\n" ) file(APPEND ${pass_file} "\#pragma once\n") file(APPEND ${pass_file} "\#include \"paddle/fluid/framework/ir/pass.h\"\n") copy_if_different(${pass_file} ${pass_file_final}) function(pass_library TARGET DEST) set(options "") set(oneValueArgs "") set(multiValueArgs SRCS DEPS DIR) set(targetPrefix "") cmake_parse_arguments(pass_library "${options}" "${oneValueArgs}" "${multiValueArgs}" ${ARGN}) if(pass_library_DIR) cc_library( ${TARGET} SRCS ${pass_library_DIR}/${TARGET}.cc DEPS graph_pattern_detector pass fuse_pass_base op_version_registry ${pass_library_DEPS}) else() cc_library( ${TARGET} SRCS ${TARGET}.cc DEPS graph_pattern_detector pass fuse_pass_base op_version_registry ${pass_library_DEPS}) endif() # add more DEST here, such as train, dist and collect USE_PASS into a file automatically. if(${DEST} STREQUAL "base" OR ${DEST} STREQUAL "inference") if(NOT CMAKE_BUILD_TYPE STREQUAL "Release") message(STATUS "add pass ${TARGET} ${DEST}") endif() file(APPEND ${pass_file} "USE_PASS(${TARGET});\n") set(INFER_IR_PASSES ${INFER_IR_PASSES} ${TARGET} CACHE INTERNAL "") endif() endfunction() add_subdirectory(ir) add_subdirectory(details) add_subdirectory(fleet) add_subdirectory(io) add_subdirectory(new_executor) if(WITH_CINN) add_subdirectory(paddle2cinn) endif() #ddim lib proto_library(framework_proto SRCS framework.proto) proto_library(pass_desc_proto SRCS pass_desc.proto DEPS framework_proto) proto_library(op_def_proto SRCS op_def.proto DEPS framework_proto) cc_library( op_def_api SRCS op_def_api.cc DEPS op_def_proto) file(GLOB OP_DEF_FILES ${PADDLE_SOURCE_DIR}/paddle/fluid/operators/compat/*.pbtxt) file(WRITE ${CMAKE_CURRENT_BINARY_DIR}/op_def.pbtxt "namespace { \n" "const std::unordered_map op_def_map = { \n") foreach(OP_DEF_FILE ${OP_DEF_FILES}) file(READ ${OP_DEF_FILE} OP_DEF_CONTENT) get_filename_component(OP_NAME ${OP_DEF_FILE} NAME_WE) file(APPEND ${CMAKE_CURRENT_BINARY_DIR}/op_def.pbtxt "{\"${OP_NAME}\",R\"(${OP_DEF_CONTENT})\"},\n") endforeach() file(APPEND ${CMAKE_CURRENT_BINARY_DIR}/op_def.pbtxt "{\"\",\"\"}};\n}") proto_library(heter_service_proto SRCS heter_service.proto) proto_library(data_feed_proto SRCS data_feed.proto) proto_library(trainer_desc_proto SRCS trainer_desc.proto DEPS framework_proto data_feed_proto) cc_library( string_array SRCS string_array.cc DEPS utf8proc phi_enforce) cc_library( data_type SRCS data_type.cc DEPS framework_proto) cc_test( data_type_test SRCS data_type_test.cc DEPS data_type place tensor) cc_library( tensor SRCS tensor_util.cc DEPS place memory data_type device_context dense_tensor) cc_test( tensor_test SRCS tensor_test.cc DEPS tensor isfinite_op) if(WITH_GPU) nv_test( tensor_util_test SRCS tensor_util_test.cc tensor_util_test.cu DEPS tensor dlpack_tensor isfinite_op) elseif(WITH_ROCM) hip_test( tensor_util_test SRCS tensor_util_test.cc tensor_util_test.cu DEPS tensor dlpack_tensor isfinite_op) else() cc_test( tensor_util_test SRCS tensor_util_test.cc DEPS tensor dlpack_tensor isfinite_op) endif() cc_test( copy_same_tensor_test SRCS copy_same_tensor_test.cc DEPS tensor) cc_test( eigen_test SRCS eigen_test.cc DEPS tensor) cc_library( lod_tensor SRCS lod_tensor.cc DEPS ddim mixed_vector place tensor framework_proto version) cc_test( lod_tensor_test SRCS lod_tensor_test.cc DEPS lod_utils lod_tensor memory) if(WITH_GPU) nv_test( lod_tensor_gpu_test SRCS lod_tensor_test.cu DEPS lod_tensor) elseif(WITH_ROCM) hip_test( lod_tensor_gpu_test SRCS lod_tensor_test.cu DEPS lod_tensor) endif() cc_library( garbage_collector SRCS garbage_collector.cc DEPS device_context memory gflags glog) cc_library( reader SRCS reader.cc DEPS lod_tensor ddim) cc_test( reader_test SRCS reader_test.cc DEPS reader) cc_test( threadpool_test SRCS threadpool_test.cc DEPS threadpool) cc_library( var_type_traits SRCS var_type_traits.cc DEPS framework_proto scope tensor_array sparse_coo_tensor sparse_csr_tensor extended_tensor) if(WITH_GPU) target_link_libraries(var_type_traits dynload_cuda) endif() # every source file that includes "dnnl.h" must depends on mkldnn # or, the first one should depends on mkldnn if(WITH_MKLDNN) add_dependencies(var_type_traits mkldnn) endif() cc_test( var_type_traits_test SRCS var_type_traits_test.cc DEPS var_type_traits) set(BRPC_DEPS "") if(WITH_PSCORE) set(BRPC_DEPS brpc ssl crypto) endif() if(WITH_PSLIB) if(WITH_PSLIB_BRPC) set(BRPC_DEPS pslib_brpc) elseif(NOT WITH_HETERPS) set(BRPC_DEPS brpc ssl crypto) endif() if(WITH_ARM_BRPC) set(BRPC_DEPS arm_brpc) endif() endif() cc_library( scope SRCS scope.cc DEPS glog threadpool xxhash var_type_traits) cc_library( device_worker SRCS device_worker.cc DEPS trainer_desc_proto lod_tensor scope ${BRPC_DEPS}) cc_test( device_worker_test SRCS device_worker_test.cc DEPS device_worker) cc_library( scope_pool SRCS scope_pool.cc DEPS scope) cc_test( scope_test SRCS scope_test.cc DEPS scope) cc_test( variable_test SRCS variable_test.cc DEPS tensor var_type_traits) cc_library( data_device_transform SRCS data_device_transform.cc DEPS tensor) if(WITH_GPU) nv_test( data_device_transform_test SRCS data_device_transform_test.cu DEPS operator op_registry device_context math_function scope) elseif(WITH_ROCM) hip_test( data_device_transform_test SRCS data_device_transform_test.cu DEPS operator op_registry device_context math_function scope) endif() if(WITH_GPU) if(WIN32) #windows treat symbolic file as a real file, which is different with unix #We create a hidden file and compile it instead of origin source file. windows_symbolic(hidden_file SRCS data_type_transform.cu) nv_library( data_type_transform SRCS .data_type_transform.cu DEPS tensor) add_dependencies(data_type_transform hidden_file) else() nv_library( data_type_transform SRCS data_type_transform.cu DEPS tensor) endif() nv_test( data_type_transform_test SRCS data_type_transform_test.cc data_type_transform_test.cu DEPS data_type_transform) elseif(WITH_ROCM) hip_library( data_type_transform SRCS data_type_transform.cu DEPS tensor) hip_test( data_type_transform_test SRCS data_type_transform_test.cc data_type_transform_test.cu DEPS data_type_transform) elseif(WITH_XPU) cc_library( data_type_transform SRCS data_type_transform.cc DEPS tensor xpulib) cc_test( data_type_transform_test SRCS data_type_transform_test.cc DEPS data_type_transform) else() cc_library( data_type_transform SRCS data_type_transform.cc DEPS tensor) cc_test( data_type_transform_test SRCS data_type_transform_test.cc DEPS data_type_transform) endif() cc_library( data_layout_transform SRCS data_layout_transform.cc DEPS tensor math_function) cc_test( data_layout_transform_test SRCS data_layout_transform_test.cc DEPS data_layout_transform) cc_library( data_transform SRCS data_transform.cc DEPS math_function tensor framework_proto selected_rows_utils data_device_transform data_type_transform data_layout_transform) cc_library( attribute SRCS attribute.cc DEPS framework_proto enforce) cc_test( attribute_test SRCS attribute_test.cc DEPS attribute framework_proto proto_desc) cc_test( program_desc_test SRCS program_desc_test.cc DEPS proto_desc device_context) cc_library( op_version_proto SRCS op_version_proto.cc DEPS framework_proto) cc_library( op_version_registry SRCS op_version_registry.cc DEPS op_version_proto framework_proto) cc_test( op_version_registry_test SRCS op_version_registry_test.cc DEPS op_version_registry) cc_library( op_proto_maker SRCS op_proto_maker.cc DEPS framework_proto attribute ops_extra_info glog) cc_test( op_proto_maker_test SRCS op_proto_maker_test.cc DEPS op_proto_maker) cc_library( no_need_buffer_vars_inference SRCS no_need_buffer_vars_inference.cc DEPS attribute device_context) cc_library( op_info SRCS op_info.cc DEPS attribute framework_proto no_need_buffer_vars_inference) cc_library( shape_inference SRCS shape_inference.cc DEPS ddim attribute selected_rows_utils) # every source file that includes "dnnl.h" must depends on mkldnn # or, the first one should depends on mkldnn if(WITH_MKLDNN) add_dependencies(shape_inference mkldnn) endif() cc_test( no_need_buffer_vars_inference_test SRCS no_need_buffer_vars_inference_test.cc DEPS no_need_buffer_vars_inference layer) cc_library( transfer_scope_cache SRCS transfer_scope_cache.cc DEPS scope framework_proto device_context) cc_library( unused_var_check SRCS unused_var_check.cc DEPS glog no_need_buffer_vars_inference) cc_library( op_kernel_type SRCS op_kernel_type.cc DEPS device_context place) if(WITH_XPU) cc_library( phi_utils SRCS phi_utils.cc DEPS lod_tensor dense_tensor selected_rows_utils place phi var_type_traits op_info xpu_op_list convert_utils phi_api_utils) else() cc_library( phi_utils SRCS phi_utils.cc DEPS lod_tensor dense_tensor selected_rows_utils place phi var_type_traits op_info convert_utils phi_api_utils) endif() if(WITH_XPU) cc_library( operator SRCS operator.cc DEPS xpu_op_list op_info proto_desc device_context tensor scope glog trainer_desc_proto data_feed_proto shape_inference data_transform lod_tensor profiler transfer_scope_cache op_kernel_type op_call_stack unused_var_check nan_inf_utils phi_utils kernel_factory infershape_utils op_utils op_compat_infos) else() cc_library( operator SRCS operator.cc DEPS op_info proto_desc device_context tensor scope glog trainer_desc_proto data_feed_proto shape_inference data_transform lod_tensor profiler transfer_scope_cache op_kernel_type op_call_stack unused_var_check nan_inf_utils phi_utils kernel_factory infershape_utils op_utils op_compat_infos) endif() cc_test( operator_test SRCS operator_test.cc DEPS operator op_registry device_context) cc_test( operator_exception_test SRCS operator_exception_test.cc DEPS operator op_registry device_context) cc_library(version SRCS version.cc) cc_test( version_test SRCS version_test.cc DEPS version) cc_library( proto_desc SRCS var_desc.cc op_desc.cc block_desc.cc program_desc.cc DEPS attribute ops_extra_info shape_inference op_info operator glog version xxhash dist_attr) cc_library( op_registry SRCS op_registry.cc DEPS op_proto_maker op_info operator ops_extra_info glog proto_desc) cc_library( op_call_stack SRCS op_call_stack.cc DEPS op_proto_maker enforce) cc_test( op_call_stack_test SRCS op_call_stack_test.cc DEPS op_call_stack) cc_library( program_utils SRCS program_utils.cc DEPS proto_desc) cc_test( program_utils_test SRCS program_utils_test.cc DEPS proto_desc program_utils) if(WITH_GPU) nv_test( op_registry_test SRCS op_registry_test.cc DEPS op_registry) elseif(WITH_ROCM) hip_test( op_registry_test SRCS op_registry_test.cc DEPS op_registry) endif() if(WITH_PYTHON) py_proto_compile(framework_py_proto SRCS framework.proto data_feed.proto) py_proto_compile(trainer_py_proto SRCS trainer_desc.proto data_feed.proto) py_proto_compile(distributed_strategy_py_proto SRCS distributed_strategy.proto) py_proto_compile(pass_desc_py_proto SRCS pass_desc.proto) #Generate an empty \ #__init__.py to make framework_py_proto as a valid python module. file(MAKE_DIRECTORY ${PADDLE_BINARY_DIR}/python/paddle/distributed/fleet/proto) file(TOUCH ${PADDLE_BINARY_DIR}/python/paddle/distributed/fleet/proto/__init__.py) file(TOUCH ${CMAKE_CURRENT_BINARY_DIR}/__init__.py) add_dependencies( framework_py_proto trainer_py_proto distributed_strategy_py_proto pass_desc_py_proto ps_py_proto pslib_py_proto) if(NOT WIN32) add_custom_command( TARGET framework_py_proto POST_BUILD COMMAND ${CMAKE_COMMAND} -E make_directory ${PADDLE_BINARY_DIR}/python/paddle/fluid/proto COMMAND cp *.py ${PADDLE_BINARY_DIR}/python/paddle/fluid/proto/ COMMAND cp distributed_strategy_*.py ${PADDLE_BINARY_DIR}/python/paddle/distributed/fleet/proto COMMENT "Copy generated python proto into directory paddle/fluid/proto." WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}) if(NOT WITH_ROCM) add_custom_target( fleet_executor_proto_init ALL DEPENDS fleet_executor_desc_py_proto COMMAND cp ${PADDLE_BINARY_DIR}/paddle/fluid/distributed/fleet_executor/fleet_executor_*.py ${PADDLE_BINARY_DIR}/python/paddle/distributed/fleet/proto COMMENT "Copy generated python proto into directory paddle/distributed/fleet/proto." ) endif() else() string(REPLACE "/" "\\" proto_dstpath "${PADDLE_BINARY_DIR}/python/paddle/fluid/proto/") string( REPLACE "/" "\\" fleet_proto_dstpath "${PADDLE_BINARY_DIR}/python/paddle/distributed/fleet/proto/") add_custom_command( TARGET framework_py_proto POST_BUILD COMMAND ${CMAKE_COMMAND} -E make_directory ${PADDLE_BINARY_DIR}/python/paddle/fluid/proto COMMAND copy /Y *.py ${proto_dstpath} COMMAND copy /Y distributed_strategy_*.py ${fleet_proto_dstpath} COMMENT "Copy generated python proto into directory paddle/fluid/proto." COMMENT "Copy generated python proto into directory paddle/distributed/fleet/proto." WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}) endif() endif() if(WITH_PSCORE) add_custom_target( index_dataset_proto_init ALL DEPENDS index_dataset_py_proto COMMAND cp ${PADDLE_BINARY_DIR}/paddle/fluid/distributed/index_dataset/index_dataset_*.py ${PADDLE_BINARY_DIR}/python/paddle/distributed/fleet/proto COMMENT "Copy generated python proto into directory paddle/distributed/fleet/proto." ) endif() cc_library( lod_rank_table SRCS lod_rank_table.cc DEPS lod_tensor) cc_library( feed_fetch_method SRCS feed_fetch_method.cc DEPS lod_tensor scope glog) cc_library( variable_helper SRCS variable_helper.cc DEPS lod_tensor) if(TENSORRT_FOUND) cc_library( naive_executor SRCS naive_executor.cc DEPS op_registry denormal device_context scope framework_proto glog lod_rank_table feed_fetch_method graph_to_program_pass variable_helper tensorrt_engine_op) else() cc_library( naive_executor SRCS naive_executor.cc DEPS op_registry denormal device_context scope framework_proto glog lod_rank_table feed_fetch_method graph_to_program_pass variable_helper) endif() cc_library( executor_gc_helper SRCS executor_gc_helper.cc DEPS scope proto_desc operator garbage_collector op_registry while_op_helper recurrent_op_helper conditional_block_op_helper) if(WITH_DISTRIBUTE) if(WITH_PSLIB) cc_library( executor SRCS executor.cc multi_trainer.cc pipeline_trainer.cc dataset_factory.cc dist_multi_trainer.cc trainer_factory.cc trainer.cc data_feed_factory.cc heterxpu_trainer.cc data_feed.cc device_worker.cc hogwild_worker.cc hetercpu_worker.cc ps_gpu_worker.cc ps_gpu_trainer.cc downpour_worker.cc downpour_worker_opt.cc data_feed.cu pull_dense_worker.cc section_worker.cc device_worker_factory.cc data_set.cc DEPS op_registry device_context scope framework_proto trainer_desc_proto glog fs shell fleet_wrapper heter_wrapper ps_gpu_wrapper box_wrapper metrics lodtensor_printer lod_rank_table feed_fetch_method collective_helper ${GLOB_DISTRIBUTE_DEPS} graph_to_program_pass variable_helper data_feed_proto timer monitor heter_service_proto fleet_executor ${BRPC_DEP}) set(DISTRIBUTE_COMPILE_FLAGS "-Wno-non-virtual-dtor -Wno-error=non-virtual-dtor -Wno-error=delete-non-virtual-dtor -Wno-error=parentheses" ) if(CMAKE_CXX_COMPILER_VERSION VERSION_GREATER 7.0) set(DISTRIBUTE_COMPILE_FLAGS "${DISTRIBUTE_COMPILE_FLAGS} -faligned-new") endif() set_source_files_properties( executor.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( device_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( hetercpu_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( heterxpu_trainer.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) elseif(WITH_PSCORE) # cc_library(executor SRCS executor.cc multi_trainer.cc pipeline_trainer.cc dataset_factory.cc # dist_multi_trainer.cc trainer_factory.cc trainer.cc data_feed_factory.cc # heterxpu_trainer.cc heter_pipeline_trainer.cc # data_feed.cc device_worker.cc hogwild_worker.cc hetercpu_worker.cc # downpour_worker.cc downpour_lite_worker.cc downpour_worker_opt.cc data_feed.cu # pull_dense_worker.cc section_worker.cc heter_section_worker.cc device_worker_factory.cc data_set.cc DEPS op_registry # device_context scope framework_proto data_feed_proto heter_service_proto trainer_desc_proto glog # index_sampler index_wrapper sampler index_dataset_proto # lod_rank_table fs shell fleet_wrapper heter_wrapper box_wrapper metrics lodtensor_printer feed_fetch_method # graph_to_program_pass variable_helper timer monitor # heter_service_proto fleet heter_server brpc fleet_executor # graph_gpu_wrapper) cc_library( executor SRCS executor.cc multi_trainer.cc pipeline_trainer.cc dataset_factory.cc dist_multi_trainer.cc trainer_factory.cc trainer.cc data_feed_factory.cc heterxpu_trainer.cc heter_pipeline_trainer.cc data_feed.cc device_worker.cc hogwild_worker.cc hetercpu_worker.cc downpour_worker.cc downpour_lite_worker.cc downpour_worker_opt.cc data_feed.cu pull_dense_worker.cc section_worker.cc heter_section_worker.cc device_worker_factory.cc data_set.cc DEPS op_registry device_context scope framework_proto data_feed_proto heter_service_proto trainer_desc_proto glog index_sampler index_wrapper sampler index_dataset_proto lod_rank_table fs shell fleet_wrapper heter_wrapper box_wrapper metrics lodtensor_printer feed_fetch_method graph_to_program_pass variable_helper timer monitor heter_service_proto fleet heter_server brpc fleet_executor flags) set(DISTRIBUTE_COMPILE_FLAGS "-Wno-non-virtual-dtor -Wno-error=non-virtual-dtor -Wno-error=delete-non-virtual-dtor -Wno-error=parentheses" ) if(CMAKE_CXX_COMPILER_VERSION VERSION_GREATER 7.0) set(DISTRIBUTE_COMPILE_FLAGS "${DISTRIBUTE_COMPILE_FLAGS} -faligned-new") endif() set_source_files_properties( executor.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( device_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( multi_trainer.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( hogwild_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( downpour_lite_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( heter_section_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( heter_pipeline_trainer.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) else() cc_library( executor SRCS executor.cc multi_trainer.cc pipeline_trainer.cc dataset_factory.cc dist_multi_trainer.cc trainer_factory.cc trainer.cc data_feed_factory.cc heterxpu_trainer.cc data_feed.cc device_worker.cc hogwild_worker.cc hetercpu_worker.cc ps_gpu_worker.cc ps_gpu_trainer.cc downpour_worker.cc downpour_worker_opt.cc data_feed.cu pull_dense_worker.cc section_worker.cc device_worker_factory.cc data_set.cc DEPS op_registry device_context scope framework_proto data_feed_proto heter_service_proto trainer_desc_proto glog lod_rank_table fs shell fleet_wrapper heter_wrapper ps_gpu_wrapper box_wrapper metrics lodtensor_printer feed_fetch_method graph_to_program_pass variable_helper timer monitor fleet_executor) endif() elseif(WITH_PSLIB) set(DISTRIBUTE_COMPILE_FLAGS "-Wno-non-virtual-dtor -Wno-error=non-virtual-dtor -Wno-error=delete-non-virtual-dtor" ) if(CMAKE_CXX_COMPILER_VERSION VERSION_GREATER 7.0) set(DISTRIBUTE_COMPILE_FLAGS "${DISTRIBUTE_COMPILE_FLAGS} -faligned-new") endif() set_source_files_properties( executor.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( device_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( hetercpu_worker.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) set_source_files_properties( heterxpu_trainer.cc PROPERTIES COMPILE_FLAGS ${DISTRIBUTE_COMPILE_FLAGS}) cc_library( executor SRCS executor.cc multi_trainer.cc pipeline_trainer.cc dataset_factory.cc dist_multi_trainer.cc trainer_factory.cc trainer.cc data_feed_factory.cc heterxpu_trainer.cc data_feed.cc device_worker.cc hogwild_worker.cc hetercpu_worker.cc ps_gpu_worker.cc ps_gpu_trainer.cc downpour_worker.cc downpour_worker_opt.cc data_feed.cu pull_dense_worker.cc section_worker.cc device_worker_factory.cc data_set.cc DEPS op_registry device_context scope framework_proto data_feed_proto heter_service_proto trainer_desc_proto glog lod_rank_table fs shell fleet_wrapper heter_wrapper ps_gpu_wrapper box_wrapper lodtensor_printer feed_fetch_method graph_to_program_pass variable_helper timer monitor fleet_executor ${BRPC_DEP}) else() cc_library( executor SRCS executor.cc multi_trainer.cc pipeline_trainer.cc dataset_factory.cc dist_multi_trainer.cc trainer_factory.cc trainer.cc data_feed_factory.cc heterxpu_trainer.cc data_feed.cc device_worker.cc hogwild_worker.cc hetercpu_worker.cc ps_gpu_worker.cc ps_gpu_trainer.cc downpour_worker.cc downpour_worker_opt.cc data_feed.cu pull_dense_worker.cc section_worker.cc device_worker_factory.cc data_set.cc DEPS op_registry device_context scope framework_proto data_feed_proto heter_service_proto trainer_desc_proto glog lod_rank_table fs shell fleet_wrapper heter_wrapper ps_gpu_wrapper box_wrapper lodtensor_printer feed_fetch_method graph_to_program_pass variable_helper timer monitor fleet_executor) endif() target_link_libraries(executor while_op_helper executor_gc_helper recurrent_op_helper conditional_block_op_helper) cc_library( parallel_executor SRCS parallel_executor.cc DEPS threaded_ssa_graph_executor scope_buffered_ssa_graph_executor parallel_ssa_graph_executor async_ssa_graph_executor graph build_strategy bind_threaded_ssa_graph_executor collective_helper fast_threaded_ssa_graph_executor variable_helper) cc_library( executor_cache SRCS executor_cache.cc DEPS parallel_executor standalone_executor) if(WITH_PSCORE) get_property(RPC_DEPS GLOBAL PROPERTY RPC_DEPS) if(WITH_HETERPS) cc_test( dist_multi_trainer_test SRCS dist_multi_trainer_test.cc DEPS conditional_block_op executor gloo_wrapper ${RPC_DEPS} graph_gpu_wrapper) cc_test( heter_pipeline_trainer_test SRCS heter_pipeline_trainer_test.cc DEPS conditional_block_op generated_op heter_listen_and_serv_op executor heter_server gloo_wrapper eigen_function ${RPC_DEPS} graph_gpu_wrapper) else() cc_test( dist_multi_trainer_test SRCS dist_multi_trainer_test.cc DEPS conditional_block_op executor gloo_wrapper ${RPC_DEPS}) cc_test( heter_pipeline_trainer_test SRCS heter_pipeline_trainer_test.cc DEPS conditional_block_op generated_op heter_listen_and_serv_op executor heter_server gloo_wrapper eigen_function ${RPC_DEPS}) endif() else() cc_test( dist_multi_trainer_test SRCS dist_multi_trainer_test.cc DEPS conditional_block_op executor gloo_wrapper) endif() cc_library( prune SRCS prune.cc DEPS framework_proto) cc_test( prune_test SRCS prune_test.cc DEPS op_info prune recurrent_op device_context) cc_test( var_type_inference_test SRCS var_type_inference_test.cc DEPS op_registry proto_desc) cc_library( selected_rows_utils SRCS selected_rows_utils.cc DEPS selected_rows device_context) cc_test( selected_rows_utils_test SRCS selected_rows_utils_test.cc DEPS selected_rows_utils) cc_test( op_kernel_type_test SRCS op_kernel_type_test.cc DEPS place device_context framework_proto op_kernel_type) cc_test(cow_ptr_tests SRCS details/cow_ptr_test.cc) cc_test(tuple_test SRCS tuple_test.cc) cc_test(inlined_vector_test SRCS inlined_vector_test.cc) cc_library( dlpack_tensor SRCS dlpack_tensor.cc DEPS tensor dlpack) cc_test( dlpack_tensor_test SRCS dlpack_tensor_test.cc DEPS dlpack_tensor glog) cc_library( op_compatible_info SRCS op_compatible_info.cc DEPS string_helper proto_desc) cc_test_old( op_compatible_info_test SRCS op_compatible_info_test.cc DEPS op_compatible_info proto_desc string_helper glog) cc_library( infershape_utils SRCS infershape_utils.cc DEPS lod_tensor selected_rows_utils attribute place var_type_traits phi phi_utils op_info shape_inference sparse_coo_tensor) cc_test( infershape_utils_test SRCS infershape_utils_test.cc DEPS infershape_utils infermeta_utils meta_tensor) # Get the current working branch execute_process( COMMAND git rev-parse --abbrev-ref HEAD WORKING_DIRECTORY ${CMAKE_SOURCE_DIR} OUTPUT_VARIABLE PADDLE_BRANCH OUTPUT_STRIP_TRAILING_WHITESPACE) # Get the latest abbreviated commit hash of the working branch execute_process( COMMAND git log -1 --format=%h WORKING_DIRECTORY ${CMAKE_SOURCE_DIR} OUTPUT_VARIABLE PADDLE_COMMIT OUTPUT_STRIP_TRAILING_WHITESPACE) message(STATUS "commit: ${PADDLE_COMMIT}") message(STATUS "branch: ${PADDLE_BRANCH}") configure_file(commit.h.in commit.h) cc_library( custom_operator SRCS custom_operator.cc DEPS tensor attribute framework_proto op_registry operator dynamic_loader string_helper phi_tensor op_meta_info phi_api tensor_api phi_tensor_operants operants_manager) set(FLUID_FRAMEWORK_MODULES proto_desc memory lod_tensor executor data_feed_proto layer dynamic_loader custom_operator) cc_library(paddle_framework DEPS ${FLUID_FRAMEWORK_MODULES}) if(WITH_TESTING AND TEST selected_rows_utils_test) set_tests_properties(selected_rows_utils_test PROPERTIES TIMEOUT 120) endif() cc_test(scope_guard_test SRCS scope_guard_test.cc) cc_test( phi_utils_test SRCS phi_utils_test.cc DEPS phi_utils) cc_test(convert_utils_test SRCS convert_utils_test.cc)