1. 24 11月, 2022 1 次提交
    • W
      [Fluid clean] (#48105) · 43b92b63
      wangxiaoning 提交于
      * add index sample fp16 support
      
      * remove fluid APIs in distributed_strategy.py and role_maker.py
      
      * Revert "remove fluid APIs in distributed_strategy.py and role_maker.py"
      
      This reverts commit 223bbee990d3bf69e252fc3c0f19e3873550a264.
      
      * remove fluid APIs in distributed_strategy.py and role_maker.py
      
      * remove index sample op changes
      
      * remove fluid APIs under fleet.base
      
      * remove fluid APIs under fleet.layers.mpu
      
      * remove fluid APIs under fleet.meta_optimizers
      
      * fix fluid error
      
      * fix util_factory.py
      
      * reset fluid.io.load_inference_model API
      43b92b63
  2. 22 11月, 2022 1 次提交
    • U
      [Docs]fix math api en docs issue (#47448) · 94c6ec86
      ustiniankw 提交于
      * fix_docx_stanh
      
      * fix einsum api en docs issue
      
      * fix model api en docs issue
      
      * for codestyle
      
      * fix_einsum.py_einsum, test=document_fix
      
      * fix_model.py_Model, test=ducument_fix
      
      * fix_creation.py_meshgrid, test=document_fix
      
      * fix_linalg.py_slogdet, test=document_fix
      
      * fix_loss.py_SoftMarginLoss_CrossEntropyLoss_NLLLoss_BCELoss, test=document_fix
      
      * norm.py_SyncBatchNorm, test=document-fix
      
      * norm.py_SyncBatchNorm, test=document_fix
      
      * norm.py_SyncBatchNorm, test=document_fix
      
      * list18-30, test=document_fix
      
      * refix_list1-15, test=document_fix
      
      * deletefiles, test=document_fix
      
      * fixedapi_pre-commit, test=document_fix
      
      * fix_list31-45, test=document_fix
      
      * list111, test=document_fix
      
      * some_fix, test=document_fix
      
      * some_fix, test=document_fix
      
      * somefix, test=document_fix
      
      * somefix, test=document_fix
      
      * refix, test=document_fix
      
      * refix, test=document_fix
      
      * refix, test=document_fix
      
      * refix, test=document_fix
      
      * rerfix, test=document_fix
      Co-authored-by: 梦柳's avatarLigoml <limengliu@tiaozhan.com>
      94c6ec86
  3. 09 11月, 2022 1 次提交
  4. 08 11月, 2022 1 次提交
  5. 23 10月, 2022 1 次提交
  6. 12 10月, 2022 1 次提交
    • N
      [CodeStyle][F401] remove unused imports in python/paddle/distributed (#46758) · fe716a0b
      Nyakku Shigure 提交于
      * [CodeStyle][F401] remove unused import in python/paddle/distributed
      
      * remove pass
      
      * empty commit
      
      * Fix ValueError: list.remove(x): x not in list for meta_optimizer_names.
      
      Fix ValueError: list.remove(x): x not in list for meta_optimizer_names.
      
      * Fix split import.
      
      Fix split import.
      
      * add noqa after meta_optimizers in factory
      
      * restort collective ops
      
      * expand `import *`
      
      * add noqa after required imports
      
      * try to fix APIs without core.ops
      
      * Revert "try to fix APIs without core.ops"
      
      This reverts commit 6172beaf601e84bf61f2490c12c4739f0edaa5eb.
      
      * fix an increment
      
      * empty commit
      
      * add noqa after required imports
      
      * expand `import *`, fix ci error
      Co-authored-by: NShuangchi He <34329208+Yulv-git@users.noreply.github.com>
      fe716a0b
  7. 14 9月, 2022 1 次提交
  8. 07 9月, 2022 1 次提交
  9. 01 9月, 2022 1 次提交
  10. 23 8月, 2022 1 次提交
  11. 01 8月, 2022 1 次提交
  12. 26 7月, 2022 1 次提交
    • Z
      add horizontal federation learning ps feature (#44327) · 4bc22b69
      ziyoujiyi 提交于
      * back fl
      
      * delete ssl cert
      
      * .
      
      * make warning
      
      * .
      
      * unittest paral degree
      
      * solve unittest
      
      * heter & multi cloud commm ready
      
      * .
      
      * .
      
      * fl-ps v1.0
      
      * .
      
      * support N + N mode
      
      * .
      
      * .
      
      * .
      
      * .
      
      * delete print
      
      * .
      
      * .
      
      * .
      
      * .
      
      * fix bug
      
      * .
      
      * .
      
      * fl-ps with coordinator ready
      
      * merge dev
      
      * update message parse only
      
      * update fl client scheduler
      
      * fix bug
      
      * update multithreads sync
      
      * fix ci errors
      
      * update role_maker.py
      
      * update role_maker.py
      
      * fix ci error: windows py import error
      
      * fix ci error: windows py import error
      
      * fix windows ci pylib import error
      
      * add dump fields & params
      
      * try to fix windows import fleet error
      
      * fix ps FLAGS error
      4bc22b69
  13. 20 7月, 2022 1 次提交
  14. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  15. 02 6月, 2022 1 次提交
  16. 19 5月, 2022 1 次提交
  17. 12 5月, 2022 1 次提交
  18. 19 4月, 2022 1 次提交
  19. 13 4月, 2022 1 次提交
    • W
      the one ps proto (#41659) · b12af9e1
      wangguanqun 提交于
      * the one ps proto
      
      * the one ps proto
      
      * fix
      
      * fix
      
      * fix
      
      * fix windows ci
      
      * fix windows ci
      
      * add dependency
      
      * add dependency
      b12af9e1
  20. 31 3月, 2022 1 次提交
  21. 17 1月, 2022 1 次提交
  22. 09 12月, 2021 1 次提交
  23. 06 12月, 2021 1 次提交
  24. 30 11月, 2021 1 次提交
  25. 26 11月, 2021 1 次提交
  26. 24 11月, 2021 1 次提交
  27. 08 9月, 2021 1 次提交
    • Y
      [Auto Parallel] Integrate all modules (#35483) · 12155358
      Yulong Ao 提交于
      * add auto_parallel dir
      
      * mv to paddle.distributed
      
      * add shard_xx api
      
      * add distributed attrs for var
      
      * add ut, test=develop
      
      * add dist
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update, test=develop
      
      * update, test=develop
      
      * update
      
      * update
      
      * delete unused proto
      
      * resotre op_desc
      
      * restore type_defs
      
      * update var_desc
      
      * remove dimss_mapping for proto_pybind
      
      * update interface.py
      
      * update framework.py
      
      * update
      
      * update
      
      * add auto_parallel dir
      
      * mv to paddle.distributed
      
      * add shard_xx api
      
      * add distributed attrs for var
      
      * add ut, test=develop
      
      * [WIP] Add the auto completion feature and related codes
      
      * [WIP] Improve the auto completion and related codes
      
      * [WIP] Make the auto completion to support data-parallel
      
      * [WIP] Make the completion support mp and dp+mp
      
      * [WIP] Refactor auto completion unit test for MLP
      
      * [WIP] Refactor the implementation of DistributedOperatorImpl
      
      * [WIP] Improve dims_mapping update rule and fix a bug
      
      * [WIP] Support auto completion for one transformer decoder layer
      
      * [WIP] Add a minor change
      
      * [WIP] Fix a bug within the uint test
      
      * Shard XShape tensor, add embedding completion and refactor code
      
      * Add the distributed_operators dir to setup.py.in
      
      * Improve the completion process and add the unittest for gpt
      
      * fix process_mesh ut
      
      * fix process_mesh ut
      
      * update
      
      * update, test=develop
      
      * Add support for automatically completing distributed attrs of special ops
      
      * update
      
      * update
      
      * update
      
      * fix doc sample codes, test=develop
      
      * improve coverage, test=develop
      
      * add static_mode check, test=develop
      
      * Model the cluster for cost model and physical mapping
      
      * update, test=develop
      
      * add set_placement, test=develop
      
      * Add the check to make sure the candidate tensors' size is great than zero
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update, test=develop
      
      * Auto mark dist attrs annotated by user
      
      * update ndarray to nested list, test=develop
      
      * update, test=develop
      
      * Add auto-completion module for auto-parallel (based on PR#33804)
      
      * Remove unnecessary files
      
      * Remove unrelated files for the auto completion pr
      
      * Update the unit test to improve the coverage
      
      * Modify codes based on reviews
      
      * Minor changes for CI
      
      * Improve some codes based on new comments
      
      * Fix bugs caused by shallow copy in attributes.py
      * Imporve amend_distributed_attr_for_program in context.py
      * Other changes for weihang's comments
      
      * support shard reader
      
      * support shard reader
      
      * add parallel mode
      
      * update process mesh
      
      * add method to compute comm_group
      
      * implement dist_embedding forward func
      
      * implement dist matmul forward func
      
      * implement dist reshape forward func
      
      * add transpiler framework
      
      * add transpiler forward
      
      * implement transpiler forward
      
      * implement transpiler backward & update
      
      * add process
      
      * add unitest
      
      * chmod
      
      * chmod
      
      * chmod
      
      * update unitest
      
      * add unitest for gpt
      
      * remove unused print
      
      * rename transpiler --> partitioner
      
      * rename transpiler --> partitioner
      
      * chmod
      
      * chmod
      
      * bug fixed
      
      * remove amp function
      
      * update case for dp mode
      
      * update case for dp mode
      
      * [Auto Parallel] Integrate all parts with the newest code
      
      * Integrate all parts of auto parallel and improve codes
      
      * Integrate all parts by AutoParallelizer
      * Add unit test for AutoParallelizer
      * Improve auto completion module for pipeline parallel
      * Add support for matmul_v2 in dist_matmul
      * Correct the typo "stratergy" to "strategy"
      
      * Modify distributed_strategy.proto to conform the main stream
      
      * Restore parts of distributed_strategy to conform the develop branch
      Co-authored-by: Nsandyhouse <lilong12@baidu.com>
      Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
      12155358
  28. 20 8月, 2021 1 次提交
  29. 18 8月, 2021 1 次提交
  30. 30 7月, 2021 1 次提交
  31. 08 7月, 2021 1 次提交
  32. 01 7月, 2021 1 次提交
  33. 21 6月, 2021 1 次提交
  34. 10 6月, 2021 1 次提交
  35. 09 6月, 2021 1 次提交
  36. 07 6月, 2021 1 次提交
  37. 26 5月, 2021 1 次提交
  38. 17 5月, 2021 1 次提交
  39. 11 5月, 2021 1 次提交
  40. 08 5月, 2021 1 次提交