- 13 10月, 2021 1 次提交
-
-
由 Guoxia Wang 提交于
-
- 13 9月, 2021 1 次提交
-
-
由 ShenLiang 提交于
* support grad group * fix single card condition
-
- 10 9月, 2021 1 次提交
-
-
由 JZ-LIANG 提交于
* sharding support dp * sharding support mp * sharding support pp
-
- 08 9月, 2021 2 次提交
-
-
由 Yulong Ao 提交于
* add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * add dist * update * update * update * update * update * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update * update * update * update * update * update, test=develop * update, test=develop * update * update * delete unused proto * resotre op_desc * restore type_defs * update var_desc * remove dimss_mapping for proto_pybind * update interface.py * update framework.py * update * update * add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * [WIP] Add the auto completion feature and related codes * [WIP] Improve the auto completion and related codes * [WIP] Make the auto completion to support data-parallel * [WIP] Make the completion support mp and dp+mp * [WIP] Refactor auto completion unit test for MLP * [WIP] Refactor the implementation of DistributedOperatorImpl * [WIP] Improve dims_mapping update rule and fix a bug * [WIP] Support auto completion for one transformer decoder layer * [WIP] Add a minor change * [WIP] Fix a bug within the uint test * Shard XShape tensor, add embedding completion and refactor code * Add the distributed_operators dir to setup.py.in * Improve the completion process and add the unittest for gpt * fix process_mesh ut * fix process_mesh ut * update * update, test=develop * Add support for automatically completing distributed attrs of special ops * update * update * update * fix doc sample codes, test=develop * improve coverage, test=develop * add static_mode check, test=develop * Model the cluster for cost model and physical mapping * update, test=develop * add set_placement, test=develop * Add the check to make sure the candidate tensors' size is great than zero * update doc, test=develop * update doc, test=develop * update doc, test=develop * update doc, test=develop * update, test=develop * Auto mark dist attrs annotated by user * update ndarray to nested list, test=develop * update, test=develop * Add auto-completion module for auto-parallel (based on PR#33804) * Remove unnecessary files * Remove unrelated files for the auto completion pr * Update the unit test to improve the coverage * Modify codes based on reviews * Minor changes for CI * Improve some codes based on new comments * Fix bugs caused by shallow copy in attributes.py * Imporve amend_distributed_attr_for_program in context.py * Other changes for weihang's comments * support shard reader * support shard reader * add parallel mode * update process mesh * add method to compute comm_group * implement dist_embedding forward func * implement dist matmul forward func * implement dist reshape forward func * add transpiler framework * add transpiler forward * implement transpiler forward * implement transpiler backward & update * add process * add unitest * chmod * chmod * chmod * update unitest * add unitest for gpt * remove unused print * rename transpiler --> partitioner * rename transpiler --> partitioner * chmod * chmod * bug fixed * remove amp function * update case for dp mode * update case for dp mode * [Auto Parallel] Integrate all parts with the newest code * Integrate all parts of auto parallel and improve codes * Integrate all parts by AutoParallelizer * Add unit test for AutoParallelizer * Improve auto completion module for pipeline parallel * Add support for matmul_v2 in dist_matmul * Correct the typo "stratergy" to "strategy" * Modify distributed_strategy.proto to conform the main stream * Restore parts of distributed_strategy to conform the develop branch Co-authored-by: Nsandyhouse <lilong12@baidu.com> Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
-
由 Zeng Jinle 提交于
* add fleet api for program pass * turn on apply pass for CI test * fix disable fuse_all_optimizer bug * try to test ci * fix CI * fill unspecified op role * fix fuse_allreduce * add ut to improve coverage * remove useless change * improve c++ coverage * follow some comments * test ir pass pipeline * update doc * reduce ut time again
-
- 30 7月, 2021 1 次提交
-
-
由 wangguanqun 提交于
* add trainer desc config to distributed strategy * code style modified
-
- 27 7月, 2021 1 次提交
-
-
由 Yuang Liu 提交于
-
- 01 7月, 2021 1 次提交
-
-
由 JZ-LIANG 提交于
* dygraph sharding * update unitest hybrid_parallel_communicate_group
-
- 25 6月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 27 5月, 2021 1 次提交
-
-
由 Thunderbrook 提交于
* support ssd in PsCore * remove log * remove bz2 * defalut value * code style * parse table class * code style * add define
-
- 17 5月, 2021 1 次提交
-
-
由 ShenLiang 提交于
* fix precision of mp * fix bug of seed * fix dp * print group
-
- 12 5月, 2021 1 次提交
-
-
由 tangwei12 提交于
* fix cpp lint * fix save/load with unexpected value * fix save and user interface
-
- 06 5月, 2021 1 次提交
-
-
由 zhiboniu 提交于
-
- 25 4月, 2021 1 次提交
-
-
由 lilong12 提交于
* add pp dygraph, test=develop
-
- 22 4月, 2021 2 次提交
- 21 4月, 2021 1 次提交
-
-
由 Yuang Liu 提交于
-
- 19 4月, 2021 1 次提交
-
-
由 ShenLiang 提交于
* support dp & mp
-
- 17 4月, 2021 1 次提交
-
-
由 ShenLiang 提交于
* add model parallel support in dygraph
-
- 07 4月, 2021 1 次提交
-
-
由 zhang wenhui 提交于
* Ascend rc (#30483) * Fix compilcation on CANN20.1 and older (#30494) Fix compilcation on CANN20.1 and older * Add distribution supported (#30578) Add distribution supported * Build praser for Hcom* operators (#30627) Build praser for Hcom* operators * Pass device_ids info from launch to trainer. (#30632) Pass device_ids info from launch to trainer * Add Hccl program group (#30642) Add Hccl program group * Add startup bash files of test_ascend_group. (#30645) Add startup bash files of test_ascend_group * cleanup (#30646) cleanup test_ascend_group.py * [Feature] Build parser to support distributed training (#30658) [Feature] Build parser to support distributed training * fix compilation on ascend-20.1 (#30722) fix compilation on ascend-20.1 * Dev/fix ascend string (#30749) Dev/fix ascend string * code style (#30781) code style * Merge ascend_optimizer and ascend_parser. (#30776) Merge ascend_optimizer and ascend_parser. * Ascendrc add converted op : [range/equal/range/uniform_random/expand/squeeze], fix cast op bug (#30797) Ascendrc add converted op : [range/equal/range/uniform_random/expand/squeeze], fix cast op bug * Add paddle ascend distribution training supported (#30796) Add paddle ascend distribution training supported * pass cxx_flags to gloo cmake (#30857) * Destroy session first. (#30954) Destroy session first. * merge * fix, test=develop * fix, test=develop * fix style, test=develop * fix, test=develop * fix * fix log fatal, test=develop * fix enforce style, test=develop * fix, test=develop * fix, test=develop * fix rccl, test=develop * fix test, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix node_num, test=develop * fix ids str, test=develop * fix ids str, test=develop * fix ids str, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix, test=develop * fix style code, test=develop * fix style code, test=develop * fix style code, test=develop * fix style code, test=develop Co-authored-by: Nhutuxian <hutuxian2011@sina.cn> Co-authored-by: Ngongweibao <weibao.gong@gmail.com> Co-authored-by: NVoid Main <voidmain1313113@gmail.com> Co-authored-by: NLeo Chen <chenqiuliang@baidu.com> Co-authored-by: Ndingsiyu <18369187719@163.com> Co-authored-by: NOleNet <olenet@126.com>
-
- 01 4月, 2021 2 次提交
- 15 3月, 2021 1 次提交
-
-
由 ShenLiang 提交于
-
- 20 2月, 2021 1 次提交
-
-
由 123malin 提交于
* test=develop, save/load, shrink Co-authored-by: NseiriosPlus <tangwei12@baidu.com>
-
- 01 2月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 21 1月, 2021 1 次提交
-
-
由 Zhen Wang 提交于
* Fix the bug in fleet amp_init. * Fix the amp_init unit test.
-
- 20 1月, 2021 1 次提交
-
-
由 huangxu96 提交于
* add fleet amp.init() * add unittest for fleet_amp_init
-
- 12 1月, 2021 1 次提交
-
-
由 Chengmo 提交于
* add save tensor support Co-authored-by: NseiriosPlus <tangwei12@baidu.com>
-
- 22 12月, 2020 1 次提交
-
-
由 ShenLiang 提交于
* fix fleet for multi-stream * fix memcpy for ncclid * use sync to solve move operation
-
- 04 12月, 2020 1 次提交
-
-
由 ShenLiang 提交于
-
- 03 12月, 2020 2 次提交
- 01 12月, 2020 2 次提交
- 27 11月, 2020 1 次提交
-
-
由 ShenLiang 提交于
* add reducer * refine envent for memorycopy * add concat&split for allreduce * apply concat & split for fuse tensor * fix nccl dep * fix the untest, compile problem and ddp initialize problem * fix untest for mac & add some comments & solve the repeated param in sublayers * fix untest for windows & fix document
-
- 19 10月, 2020 1 次提交
-
-
由 MRXLT 提交于
fleet support paddle.optimzier * bug fix * fix fleet_base * bug fix * fix coverage
-
- 16 10月, 2020 1 次提交
-
-
由 WangXi 提交于
-
- 15 10月, 2020 2 次提交
-
-
由 tangwei12 提交于
* add size method for large scale * add large scale UT * add ut for checkpoint
-
由 danleifeng 提交于
* raise error if use multi-cards in fleet non_distributed mode; test=develop
-
- 14 10月, 2020 1 次提交
-
-
由 Chen Weihang 提交于
-