1. 27 12月, 2022 1 次提交
  2. 26 12月, 2022 1 次提交
    • Y
      [Auto Parallel] Merge the python and c++ impls of ProcessMesh (#47503) · 1c0afa79
      Yulong Ao 提交于
      * [Auto Parallel] Rename methods of ProcessMesh
      
      * [Auto Parallel] Impl the python process_mesh by the c++ one
      
      * [Auto Parallel] Add some minor modifications
      
      * [Auto Parallel] Rename some methods
      
      * [Auto Parallel] Remove unnecessary codes
      
      * [Auto Parallel] Add back some removed files
      
      * [Auto Parallel] Fix bugs
      
      * [Auto Parallel] Fix a bug
      
      * Update process_mesh.cc
      
      * [Auto Parallel] Fix a bug
      1c0afa79
  3. 23 12月, 2022 1 次提交
  4. 21 12月, 2022 1 次提交
  5. 14 12月, 2022 1 次提交
    • Z
      [AutoParallel] recompute tuning (#48608) · 170a31f9
      zhaoyingli 提交于
      * [AutoParallel] recompute tuning
      
      * fix conflict
      
      * update comment
      
      * bug fix
      
      * update rc algo
      
      * tiny fix
      
      * fix clear process_group
      
      * remove comment
      
      * update segment print
      
      * fix import OpRole
      
      * adapt amp pass and grad_clip pass for opt_tuner
      
      * update tuning config
      
      * fix import
      
      * annotate recompute info on ops and upgrade recompute pass
      
      * add op_namescope for seed op
      
      * record reserved vars
      
      * fix recompute var's dist_attr
      
      * fix strategy unittest
      
      * adapt for fp16
      
      * update unittest
      
      * revert copy opt
      
      * update unittest
      
      * rename set_recompute_segments
      
      * fix unittest
      170a31f9
  6. 12 12月, 2022 1 次提交
  7. 09 12月, 2022 1 次提交
  8. 02 12月, 2022 1 次提交
    • S
      Fix fuse_gemm_epilogue (#47805) · 6efc2888
      Shijie 提交于
      * Fix fuse_gemm_epilogue
      
      * update tests
      
      * Update CMakeLists.txt
      
      * Update CMakeLists.txt
      
      * Update CMakeLists.txt
      
      * fix random seed
      
      * use assert_allclose
      
      * Update test_dist_fuse_gemm_epilogue_pass.py
      
      * Update cpp_pass.py
      
      * Update test_dist_fuse_gemm_epilogue_pass.py
      
      * fix codestyle
      
      * update seed and atol
      6efc2888
  9. 01 12月, 2022 1 次提交
  10. 29 11月, 2022 2 次提交
  11. 28 11月, 2022 1 次提交
  12. 25 11月, 2022 1 次提交
  13. 22 11月, 2022 1 次提交
  14. 18 11月, 2022 2 次提交
  15. 14 11月, 2022 2 次提交
  16. 10 11月, 2022 1 次提交
  17. 09 11月, 2022 1 次提交
  18. 08 11月, 2022 2 次提交
  19. 07 11月, 2022 2 次提交
  20. 03 11月, 2022 1 次提交
  21. 01 11月, 2022 1 次提交
  22. 23 10月, 2022 1 次提交
  23. 20 10月, 2022 1 次提交
  24. 19 10月, 2022 1 次提交
  25. 18 10月, 2022 2 次提交
    • N
      [CodeStyle][py2] remove `compat` module (to_text) (#47036) · ad4c773b
      Nyakku Shigure 提交于
      * [CodeStyle][py2] remove `compat` module (to_text)
      
      * remove some unnecessary decode
      
      * remove to_text definition and unittest
      
      * Revert "remove to_text definition and unittest"
      
      This reverts commit a6b69cb8dca8b9b031ce10ea32d1040e7e0dd267.
      
      * remove an assertion
      
      * empty commit
      ad4c773b
    • Z
      [AutoParallel] add callbacks (#47014) · 7c92177c
      zhaoyingli 提交于
      * [AutoParallel] add callbacks
      
      * fix unittest
      
      * fix dist_context
      
      * fix engine
      
      * fix cmakelist
      
      * fix unittest's returns
      
      * fix cmakelist
      7c92177c
  26. 12 10月, 2022 2 次提交
    • J
      bugfix (#46921) · acdaa4fb
      JZ-LIANG 提交于
      acdaa4fb
    • N
      [CodeStyle][F401] remove unused imports in python/paddle/distributed (#46758) · fe716a0b
      Nyakku Shigure 提交于
      * [CodeStyle][F401] remove unused import in python/paddle/distributed
      
      * remove pass
      
      * empty commit
      
      * Fix ValueError: list.remove(x): x not in list for meta_optimizer_names.
      
      Fix ValueError: list.remove(x): x not in list for meta_optimizer_names.
      
      * Fix split import.
      
      Fix split import.
      
      * add noqa after meta_optimizers in factory
      
      * restort collective ops
      
      * expand `import *`
      
      * add noqa after required imports
      
      * try to fix APIs without core.ops
      
      * Revert "try to fix APIs without core.ops"
      
      This reverts commit 6172beaf601e84bf61f2490c12c4739f0edaa5eb.
      
      * fix an increment
      
      * empty commit
      
      * add noqa after required imports
      
      * expand `import *`, fix ci error
      Co-authored-by: NShuangchi He <34329208+Yulv-git@users.noreply.github.com>
      fe716a0b
  27. 10 10月, 2022 1 次提交
  28. 29 9月, 2022 1 次提交
  29. 28 9月, 2022 2 次提交
  30. 26 9月, 2022 1 次提交
  31. 20 9月, 2022 2 次提交
    • Z
      Fl ps (#46258) · 5bbfca15
      ziyoujiyi 提交于
      * back fl
      
      * delete ssl cert
      
      * .
      
      * make warning
      
      * .
      
      * unittest paral degree
      
      * solve unittest
      
      * heter & multi cloud commm ready
      
      * .
      
      * .
      
      * fl-ps v1.0
      
      * .
      
      * support N + N mode
      
      * .
      
      * .
      
      * .
      
      * .
      
      * delete print
      
      * .
      
      * .
      
      * .
      
      * .
      
      * fix bug
      
      * .
      
      * .
      
      * fl-ps with coordinator ready
      
      * merge dev
      
      * update message parse only
      
      * update fl client scheduler
      
      * fix bug
      
      * update multithreads sync
      
      * fix ci errors
      
      * update role_maker.py
      
      * update role_maker.py
      
      * fix ci error: windows py import error
      
      * fix ci error: windows py import error
      
      * fix windows ci pylib import error
      
      * add dump fields & params
      
      * try to fix windows import fleet error
      
      * fix ps FLAGS error
      
      * fix logging risk
      
      * fix logging possible risk
      
      * write trainer_desc file
      
      * support split sparse params in local & remote
      
      * fix import paddle.fluid.core.PSGPU
      
      * fix import paddle.fluid.core.PSGPU
      
      * add remote_sparse & local_sparse config
      
      * fix unittest
      
      * fix test_dist_fleet_geo table error
      
      * fix PADDLE_ENFORCE error
      
      * fix other's pr conflict
      
      * forbidden ssd table
      
      * .
      
      * recover ssd table code
      
      * recover file mode
      
      * debug auc 0.5
      
      * adapt for nn fl-ps
      
      * adapt for nn fl-ps
      
      * add learning_rate_0 intializer op
      
      * recover ssd table
      
      * modify file mode
      
      * flps del fake-init op
      5bbfca15
    • J
      [Auto Parallel] performance improvement for Sharding-DP hybrid parallelism (#46180) · f769f850
      JZ-LIANG 提交于
      * remove no need grad allreduce communication when sharding-dp
      
      * remove no need grad allreduce communication when sharding-dp
      
      * bugfix
      
      * bugfix
      
      * bugfix
      f769f850