1. 09 2月, 2023 2 次提交
    • Z
      remove paddle.fluid.dygraph.parallel.ParallelEnv (#50157) · 9dd1f4bf
      zqw_1997 提交于
      * remove dygraph.parallel.ParallelEnv
      
      * logger.py error: AttributeError: module 'paddle' has no attribute 'distributed'
      
      * move the implenmentation to the root folder
      
      * logger.py import ParallelEnv from paddle.parallel to avoid circular import
      
      * add the comment of why import ParallelEnv from paddle.parallel in logger.py and remove the api interface in the paddle/parallel.py
      
      * outdated Env and note removed
      
      * decouple the logger.py and ParallelEnv
      
      * remove another ref of parallel in init.py
      9dd1f4bf
    • Y
      Fix bugs in pass_base.py (#50136) · 5cae5fdd
      yuehuayingxueluo 提交于
      * fix the processing order of passes in pass_base.py
      
      * fix processing order
      
      * add _PASS_PROCESS_ORDER_LIST
      
      * delete some pass in _PASS_PROCESS_ORDER_LIST
      
      * add assert in pass_base.py
      
      * remove fuse_optimizer
      
      * add _fusion_opt_list_rule
      
      * add test_pass_base_list.py
      
      * fix some bug
      
      * add fused_attention
      
      * add some passes to list
      
      * fix ci bug
      
      * fix ci bug
      5cae5fdd
  2. 01 2月, 2023 1 次提交
  3. 30 1月, 2023 1 次提交
    • Z
      [Pglbox2.0] merge gpugraph to develop (#49946) · cb525d4e
      zmxdream 提交于
      * add set slot_num for psgpuwraper (#177)
      
      * add set slot_num_for_pull_feature for psgpuwarper
      
      * Add get_epoch_finish python interface (#182)
      
      * add get_epoch_finish interface
      
      * add return
      
      * delete return
      
      * add unzip op (#183)
      
      * fix miss key for error dataset (#186)
      
      * fix miss key for error dataset
      
      * fix miss key for error dataset
      Co-authored-by: Nyangjunchao <yangjunchao@baidu.com>
      
      * add excluded_train_pair and infer_node_type (#187)
      
      * support return of degree (#188)
      
      * fix task stuck in barrier (#189)
      Co-authored-by: Nyangjunchao <yangjunchao@baidu.com>
      
      * check node/feature format when loading (#190)
      
      * check node&feature format when loading
      
      * check node&feature format when loading (2£ (2)
      
      * degrade log (#191)
      
      * [PGLBOX]fix conflict
      
      * [PGLBOX]fix conflict
      
      * [PGLBOX]replace LodTensor with phi::DenseTensor
      
      * [PGLBOX]fix gpu_primitives.h include path
      
      * [PGLBOX]from platform::PADDLE_CUDA_NUM_THREADS to phi::PADDLE_CUDA_NUM_THREADS
      
      * [PGLBOX]fix unzip example code
      
      * [PGLBOX]fix unzip example code
      
      * [PGLBOX]fix unzip example code
      
      * [PGLBOX]fix unzip example code
      
      * [PGLBOX]fix unzip ut
      
      * [PGLBOX]fix unzip ut
      
      * [PGLBOX]fix code style
      
      * [PGLBOX]fix code style
      
      * [PGLBOX]fix code style
      
      * fix code style
      
      * fix code style
      
      * fix unzip ut
      
      * fix unzip ut
      
      * fix unzip ut
      
      * fix unzip
      
      * fix code stype
      
      * add ut
      
      * add c++ ut & fix train_mode_ set
      
      * fix load into memory
      
      * fix c++ ut
      
      * fix c++ ut
      
      * fix c++ ut
      
      * fix c++ ut
      
      * fix code style
      
      * fix collective
      
      * fix unzip_op.cc
      
      * fix barrier
      
      * fix code style
      
      * fix barrier
      
      * fix barrier
      
      * fix code styple
      
      * fix unzip
      
      * add unzip.py
      
      * add unzip.py
      
      * fix unzip.py
      
      ---------
      Co-authored-by: Nchao9527 <33347532+chao9527@users.noreply.github.com>
      Co-authored-by: NSiming Dai <908660116@qq.com>
      Co-authored-by: Nhuwei02 <53012141+huwei02@users.noreply.github.com>
      Co-authored-by: Nyangjunchao <yangjunchao@baidu.com>
      cb525d4e
  4. 16 1月, 2023 2 次提交
    • W
      [Fluid clean]clean distributed fluid API (#49795) · 7de9420a
      wangxiaoning 提交于
      7de9420a
    • Y
      [Auto Parallel] Clear some fluid APIs (#49793) · e70af91d
      Yulong Ao 提交于
      * [Auto Parallel] Rename methods of ProcessMesh
      
      * [Auto Parallel] Impl the python process_mesh by the c++ one
      
      * [Auto Parallel] Add some minor modifications
      
      * [Auto Parallel] Rename some methods
      
      * [Auto Parallel] Remove unnecessary codes
      
      * [Auto Parallel] Add back some removed files
      
      * [Auto Parallel] Fix bugs
      
      * [Auto Parallel] Fix a bug
      
      * Update process_mesh.cc
      
      * [Auto Parallel] Merge dist attrs of Python into C++
      
      * [Auto Parallel] Add back deleted importing
      
      * [Auto Parallel] Add back removed unittest
      
      * [Auto Parallel] Remove type qualifiers of return types
      
      * [Auto Parallel] Fix some bugs
      
      * [Auto Parallel] Fix a bug of the quant pass
      
      * [Auto Parallel] Fix the code style
      
      * [Auto Parallel] Clear some fluid APIs
      e70af91d
  5. 12 1月, 2023 1 次提交
  6. 10 1月, 2023 2 次提交
  7. 09 1月, 2023 1 次提交
  8. 07 1月, 2023 1 次提交
    • R
      Enable standalone executor for fleet training (#49293) · 67fc8e93
      Ruibiao Chen 提交于
      * Enable standalone executor for fleet training
      
      * Update code
      
      * Replace use_standalone_executor utils in auto parallel
      
      * Update code
      
      * Diable standalone executor for test_pass_sharding
      
      * Update code
      
      * Set sequential run for auto parallel
      
      * Fix dist_attr bug
      
      * Set sequential run for auto parallel
      67fc8e93
  9. 06 1月, 2023 2 次提交
    • G
    • Y
      [Auto Parallel] Merge dist attrs from python into c++ (#49214) · c7899074
      Yulong Ao 提交于
      * [Auto Parallel] Rename methods of ProcessMesh
      
      * [Auto Parallel] Impl the python process_mesh by the c++ one
      
      * [Auto Parallel] Add some minor modifications
      
      * [Auto Parallel] Rename some methods
      
      * [Auto Parallel] Remove unnecessary codes
      
      * [Auto Parallel] Add back some removed files
      
      * [Auto Parallel] Fix bugs
      
      * [Auto Parallel] Fix a bug
      
      * Update process_mesh.cc
      
      * [Auto Parallel] Merge dist attrs of Python into C++
      
      * [Auto Parallel] Add back deleted importing
      
      * [Auto Parallel] Add back removed unittest
      
      * [Auto Parallel] Remove type qualifiers of return types
      
      * [Auto Parallel] Fix some bugs
      
      * [Auto Parallel] Fix a bug of the quant pass
      
      * [Auto Parallel] Fix the code style
      c7899074
  10. 04 1月, 2023 1 次提交
    • J
      [Auto Parallel-Performance] Sharding Comm Optimization (#48604) · 5592f8ad
      JZ-LIANG 提交于
      * remove deps and prior comm
      
      * grad comm fuse
      
      * add deps for amp&global norm
      
      * stage2 broadcast prior deps
      
      * stage2 grad overlap
      
      * stream_analyzer bugfix
      
      * overlap enable
      
      * dep op namescope
      
      * depend support multiple inputs
      
      * check finite deps
      
      * stage2 param comm overlap
      
      * Set kD2HStream
      
      * grad comm hierarchical
      
      * grad comm hierarchical
      
      * new unitest
      Co-authored-by: Nchenruibiao <chenruibiao@baidu.com>
      5592f8ad
  11. 30 12月, 2022 2 次提交
  12. 29 12月, 2022 1 次提交
  13. 28 12月, 2022 1 次提交
  14. 27 12月, 2022 1 次提交
  15. 26 12月, 2022 1 次提交
    • Y
      [Auto Parallel] Merge the python and c++ impls of ProcessMesh (#47503) · 1c0afa79
      Yulong Ao 提交于
      * [Auto Parallel] Rename methods of ProcessMesh
      
      * [Auto Parallel] Impl the python process_mesh by the c++ one
      
      * [Auto Parallel] Add some minor modifications
      
      * [Auto Parallel] Rename some methods
      
      * [Auto Parallel] Remove unnecessary codes
      
      * [Auto Parallel] Add back some removed files
      
      * [Auto Parallel] Fix bugs
      
      * [Auto Parallel] Fix a bug
      
      * Update process_mesh.cc
      
      * [Auto Parallel] Fix a bug
      1c0afa79
  16. 23 12月, 2022 1 次提交
  17. 21 12月, 2022 1 次提交
  18. 14 12月, 2022 1 次提交
    • Z
      [AutoParallel] recompute tuning (#48608) · 170a31f9
      zhaoyingli 提交于
      * [AutoParallel] recompute tuning
      
      * fix conflict
      
      * update comment
      
      * bug fix
      
      * update rc algo
      
      * tiny fix
      
      * fix clear process_group
      
      * remove comment
      
      * update segment print
      
      * fix import OpRole
      
      * adapt amp pass and grad_clip pass for opt_tuner
      
      * update tuning config
      
      * fix import
      
      * annotate recompute info on ops and upgrade recompute pass
      
      * add op_namescope for seed op
      
      * record reserved vars
      
      * fix recompute var's dist_attr
      
      * fix strategy unittest
      
      * adapt for fp16
      
      * update unittest
      
      * revert copy opt
      
      * update unittest
      
      * rename set_recompute_segments
      
      * fix unittest
      170a31f9
  19. 12 12月, 2022 1 次提交
  20. 09 12月, 2022 1 次提交
  21. 02 12月, 2022 1 次提交
    • S
      Fix fuse_gemm_epilogue (#47805) · 6efc2888
      Shijie 提交于
      * Fix fuse_gemm_epilogue
      
      * update tests
      
      * Update CMakeLists.txt
      
      * Update CMakeLists.txt
      
      * Update CMakeLists.txt
      
      * fix random seed
      
      * use assert_allclose
      
      * Update test_dist_fuse_gemm_epilogue_pass.py
      
      * Update cpp_pass.py
      
      * Update test_dist_fuse_gemm_epilogue_pass.py
      
      * fix codestyle
      
      * update seed and atol
      6efc2888
  22. 01 12月, 2022 1 次提交
  23. 29 11月, 2022 2 次提交
  24. 28 11月, 2022 1 次提交
  25. 25 11月, 2022 1 次提交
  26. 22 11月, 2022 1 次提交
  27. 18 11月, 2022 2 次提交
  28. 14 11月, 2022 2 次提交
  29. 10 11月, 2022 1 次提交
  30. 09 11月, 2022 1 次提交
  31. 08 11月, 2022 2 次提交