1. 18 10月, 2022 1 次提交
  2. 12 10月, 2022 1 次提交
    • N
      [CodeStyle][F401] remove unused imports in python/paddle/distributed (#46758) · fe716a0b
      Nyakku Shigure 提交于
      * [CodeStyle][F401] remove unused import in python/paddle/distributed
      
      * remove pass
      
      * empty commit
      
      * Fix ValueError: list.remove(x): x not in list for meta_optimizer_names.
      
      Fix ValueError: list.remove(x): x not in list for meta_optimizer_names.
      
      * Fix split import.
      
      Fix split import.
      
      * add noqa after meta_optimizers in factory
      
      * restort collective ops
      
      * expand `import *`
      
      * add noqa after required imports
      
      * try to fix APIs without core.ops
      
      * Revert "try to fix APIs without core.ops"
      
      This reverts commit 6172beaf601e84bf61f2490c12c4739f0edaa5eb.
      
      * fix an increment
      
      * empty commit
      
      * add noqa after required imports
      
      * expand `import *`, fix ci error
      Co-authored-by: NShuangchi He <34329208+Yulv-git@users.noreply.github.com>
      fe716a0b
  3. 10 10月, 2022 1 次提交
  4. 15 9月, 2022 1 次提交
    • Y
      [Auto Parallel] Improve the APIs (#45776) · b042a3b1
      Yulong Ao 提交于
      * [Auto Parallel] Use c++ dist attr in the completion process
      
      * [Auto Parallel] Add minor changes
      
      * [Auto Parallel] Use c++ dist attr in the completion process
      
      * [Auto Parallel] Add minor changes
      
      * [Auto Parallel] Add the serialization process for dist attrs
      
      * [Auto Parallel] Remove unnecessary comments
      
      * [Auto Parallel] Fix some bugs
      
      * [Auto Parallel] Fix the code style
      
      * [Auto Parallel] Remove unnecessary impls
      
      * [Auto Parallel] Fix the importing error
      
      * [Auto Parallel] Fix the copy from bugs of op dist attr
      
      * [Auto Parallel] Replace the use of constexpr if
      
      * [Auto Parallel] Redesign the shard_tensor, shard_op and ProcessMesh
      
      * [Auto Parallel] Change API of the completion unittest
      
      * [Auto Parallel] Fix the bug when set_attr an int
      
      * [Auto Parallel] Add the unittest for the serialization
      
      * [Auto Parallel] Add some unit tests
      
      * [Auto Paralle] Unify the strategy
      
      * [Auto Parallel] Improve the engine api
      
      * [Auto Parallel] Reset the changes made to the framework
      
      * [Auto Parallel] Change the engine unittest
      
      * [Auto Parallel] Update API of the completion and partitioner
      
      * [Auto Parallel] Update unit tests using engine api
      
      * update shard annotation
      
      * [Auto Parallel] Remove the modifications of other modules
      
      * [Auto Parallel] Add docs for APIs
      
      * add new strategy
      
      * [Auto Parallel] Replace the logger
      
      * [Auto Parallel] Restore the test_program.py
      
      * [Auto Parallel] Change the import rules
      
      * [Auto Parallel] Add the examples for Engine
      
      * [Auto Parallel] Do some minor changes
      
      * [Auto Parallel] Remove yaml dependency
      
      * [Auto Parallel] Fix the unittests
      
      * add valid after train
      
      * bug fix
      Co-authored-by: Nzhaoyingli <zhaoyingli@baidu.com>
      Co-authored-by: Ncaozhou <caozhou@radi.ac.cn>
      Co-authored-by: Ncaozhou <48191911+Caozhou1995@users.noreply.github.com>
      b042a3b1
  5. 25 8月, 2022 1 次提交
  6. 18 8月, 2022 1 次提交
  7. 29 7月, 2022 1 次提交
  8. 11 7月, 2022 1 次提交
  9. 24 6月, 2022 1 次提交
    • Y
      [Auto Parallel] Use a fast completion for data parallelism (#43585) · e64823c1
      Yulong Ao 提交于
      * [Auto Parallel] Use a fast completion for data parallelism
      
      * remove unuse cuSparse function
      
      * [Auto Parallel] Fix some bugs of the fast dp completion
      
      * [Auto Parallel] Add the cmake statements
      
      * [Auto Parallel] Make the unittest adapt to the new interface
      
      * [Auto Parallel] Modify the timeout of the unittest
      
      * [Auto Parallel] Remove unnecessary comments
      Co-authored-by: Nzhouwei25 <zhouwei25@baidu.com>
      e64823c1
  10. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  11. 02 6月, 2022 1 次提交
  12. 01 6月, 2022 1 次提交
    • Y
      [Auto Parallel] Add miscellaneous improvements (#43108) · 010aba33
      Yulong Ao 提交于
      * [Auto Parallel] Add the parallel tuner
      
      * [Auto Parallel] Improve the parallel tuner and fix some bugs
      
      * upodate cost model
      
      * update import Resharder by dist op
      
      * update cost model
      
      * fix comp cost bug
      
      * update cost model
      
      * [Auto Parallel] Amend the dist attr for #processses=1
      
      * update cost model and tuner
      
      * update cost model and tuner
      
      * update cost model and tuner
      
      * update cluster
      
      * update reshard
      
      * [Auto Parallel] Add the estimation from the cost model
      
      * [Auto Parallel] Reimplement the backup and restore functions
      
      * [Auto Parallel] Fix the bugs of the parallel tuner
      
      * [Auto Parallel] Update the engine api and dist context
      
      * [Auto Parallel] Work around the high order grad problem
      
      * [Auto Parallel] Add some miscellaneous improvements
      
      * [Auto Parallel] Add a unittest for DistributedContext
      Co-authored-by: Ncaozhou <caozhou@radi.ac.cn>
      010aba33
  13. 30 5月, 2022 1 次提交
  14. 19 5月, 2022 1 次提交
  15. 10 5月, 2022 1 次提交
  16. 07 5月, 2022 1 次提交
  17. 06 5月, 2022 1 次提交
    • Z
      [AutoParallel] adapt for 2d laplace (#41601) · c043a21b
      zhaoyingli 提交于
      * add default_ctx in backward.py
      
      * record grad_var_to_var with grad_times
      
      * fix backward
      
      * update annotation
      
      * add complete_high_order_grad in complete_forward
      
      * add dist slice op
      
      * update grad_var_to_var type
      
      * update partition_block init mapping before loss op
      
      * update compatible for 'XShape' & update 'allreduce_vars'
      
      * add dist reshape op when input dim equal to output dim
      
      * update 'set_grad_var_shape' with grad_var_to_var
      
      * fix dist slice
      
      * fix set_grad_var_shape
      
      * add dist pnorm op
      
      * fix dist pnorm dist_attr
      
      * fix engine startprogram & adapt highorder grad
      
      * fix set_grad_var_shape when mp
      
      * update unittest
      
      * update cmakelist
      
      * default strategy in engine: dp
      
      * bug fix
      
      * tiny fix
      
      * flatten outputs
      
      * fix default strategy
      
      * init default ctx
      
      * tiny fix
      
      * test=allcase
      c043a21b
  18. 16 3月, 2022 1 次提交
    • Y
      [Auto Parallel] Add the support for the auto completion of while_op (#39939) · ec6b8fbd
      Yulong Ao 提交于
      * [Auto Parallel] Support the auto completion of while_op
      
      * [Auto Parallel] Improve the completion algorithms
      
      * [Auto Parallel] Fix bugs for ernie inference
      
      * [Auto Parallel] Remove attrs which cannot be pickled
      
      * [Auto Parallel] make the dims_mappings of LodTensorArray vars empty
      
      * [Auto Parallel] Fix bugs for the ernie inference in the pipeline parallel
      
      * [Auto Parallel] Remove unncessary comments
      
      * [Auto Parallel] Fix a bug of the CMakeLists
      
      * [Auto Parallel] Use the newest APIs to write the unit test
      
      * [Auto Parallel] Remove unnecessary statements
      ec6b8fbd
  19. 22 2月, 2022 2 次提交
  20. 21 1月, 2022 1 次提交
    • Y
      [Auto Parallel] Use the new completion algorithm (#39086) · e5cda6fa
      Yulong Ao 提交于
      * Add the backward support for QR
      
      * Remove unnecessary comments
      
      * [Auto Parallel] Improve the dist op interface and compatible computation
      
      * Remove unnecessary modification
      
      * Recover some modifications
      
      * Add lost files
      
      * Fix a minor bug
      
      * Fix the bug of the planner
      
      * Fix the format problem
      
      * [Auto Parallel] Update the completion algorithm
      
      * Fix the bug of auto_searcher unittest
      e5cda6fa
  21. 20 1月, 2022 1 次提交
  22. 18 1月, 2022 1 次提交
  23. 11 1月, 2022 1 次提交
  24. 31 12月, 2021 1 次提交
  25. 30 12月, 2021 1 次提交
  26. 24 12月, 2021 1 次提交
  27. 24 11月, 2021 1 次提交
  28. 29 10月, 2021 1 次提交
    • Y
      [Auto Parallel] Improve the interface and the underlying mechanisms (#36617) · a02532b5
      Yulong Ao 提交于
      * default dist op
      
      * add dist_attr for dist op
      
      * add unitest
      
      * update inputname
      
      * update function name
      
      * add unitest
      
      * update CMakeLists.txt for CI
      
      * fix dis_matmul
      
      * fix compile error
      
      * update matmul to matmul_v2
      
      * unify api
      
      * unify api
      
      * todo
      
      * update distop forward func
      
      * update distop forward func
      
      * auto parallel backward
      
      * update dist op
      
      * autoparallel backward
      
      * add backward for embedding
      
      * temp1
      
      * temp2
      
      * temp3
      
      * temp4
      
      * backward done1
      
      * backward done2
      
      * backward done3
      
      * dist embedding remove mp mode
      
      * dist matmul remove mp mode
      
      * update dist embedding
      『
      
      * dist op init1
      
      * dist op init 2
      
      * update unitest
      
      * context remove parallel mode
      
      * partitioner remove parallel mode
      
      * update unitest
      
      * a more general method to support varying mesh in pipeline parallel
      
      * support varying mesh in pipeline parallel
      
      * embedding support varying mesh in pipeline parallel
      
      * matmul support varying mesh in pipeline parallel
      
      * default dist op support varying mesh in pipeline parallel
      
      * dist attribute for startup program
      
      * default dist op support varying mesh in pipeline parallel 2
      
      * partitoner support varying mesh in pipeline parallel
      
      * revise logic for auto compeletion
      
      * revise framework.py
      
      * revise reshard unitest
      
      * revise unitest for parallelize
      
      * chmod
      
      * fixed bug for dist embedding name mapping
      
      * Improve the interface and the underlying mechanisms of auto parallel
      
      * revise completion for backward
      
      * revise completion for update
      
      * revise completion for update
      
      * update unitest
      
      * chmod
      
      * bugfix for grad_op output var's mesh
      
      * Modify codes for pr 36744
      
      * Remove unnecessary comments in framework.py
      
      * Remove unnecessary comments in completion.py
      Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
      Co-authored-by: Nzhaoyingli <zhaoyingli@baidu.com>
      Co-authored-by: NJZ-LIANG <38102074+JZ-LIANG@users.noreply.github.com>
      a02532b5