- 22 9月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
-
- 19 9月, 2022 1 次提交
-
-
由 Yulong Ao 提交于
* [AutoParallel] adapt gradient merge pass (#45915) * adapt gradient merge * fix op_role * fix strategy * [Auto Parallel] Gradient Fuse Allreduce (#45643) * bugfix (#45332) * dist embedding support lookup table v1 * add unitest * customize wait_comm * group gradients * bugfix * update program * [Auto Parallel] Improve the APIs (#45776) * [Auto Parallel] Use c++ dist attr in the completion process * [Auto Parallel] Add minor changes * [Auto Parallel] Use c++ dist attr in the completion process * [Auto Parallel] Add minor changes * [Auto Parallel] Add the serialization process for dist attrs * [Auto Parallel] Remove unnecessary comments * [Auto Parallel] Fix some bugs * [Auto Parallel] Fix the code style * [Auto Parallel] Remove unnecessary impls * [Auto Parallel] Fix the importing error * [Auto Parallel] Fix the copy from bugs of op dist attr * [Auto Parallel] Replace the use of constexpr if * [Auto Parallel] Redesign the shard_tensor, shard_op and ProcessMesh * [Auto Parallel] Change API of the completion unittest * [Auto Parallel] Fix the bug when set_attr an int * [Auto Parallel] Add the unittest for the serialization * [Auto Parallel] Add some unit tests * [Auto Paralle] Unify the strategy * [Auto Parallel] Improve the engine api * [Auto Parallel] Reset the changes made to the framework * [Auto Parallel] Change the engine unittest * [Auto Parallel] Update API of the completion and partitioner * [Auto Parallel] Update unit tests using engine api * update shard annotation * [Auto Parallel] Remove the modifications of other modules * [Auto Parallel] Add docs for APIs * add new strategy * [Auto Parallel] Replace the logger * [Auto Parallel] Restore the test_program.py * [Auto Parallel] Change the import rules * [Auto Parallel] Add the examples for Engine * [Auto Parallel] Do some minor changes * [Auto Parallel] Remove yaml dependency * [Auto Parallel] Fix the unittests * add valid after train * bug fix Co-authored-by: Nzhaoyingli <zhaoyingli@baidu.com> Co-authored-by: Ncaozhou <caozhou@radi.ac.cn> Co-authored-by: Ncaozhou <48191911+Caozhou1995@users.noreply.github.com> * [Auto Parallel] Bugfix allreduce fuse for MP (#46086) * bugfix * bugfix * typos fixed * update strategy (#46138) Co-authored-by: Nzhaoyingli <86812880+zhaoyinglia@users.noreply.github.com> Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com> Co-authored-by: Nzhaoyingli <zhaoyingli@baidu.com> Co-authored-by: Ncaozhou <caozhou@radi.ac.cn> Co-authored-by: Ncaozhou <48191911+Caozhou1995@users.noreply.github.com>
-
- 09 9月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* adapt lazy init and fix pass * add unittest * update comment * fix amp and sharding * remove clip_by_norm
-
- 31 8月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* add grad_clip pass * add unittest * add notes * update func * add dist_attr for new op
-
- 23 8月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* add quant pass
-
- 18 8月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* add clip_grad * fix comments * add unittest * update logger
-
- 15 8月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* add collate_fn * fix number of inputs
-
- 12 8月, 2022 1 次提交
-
-
由 JZ-LIANG 提交于
* bugfix * remove scaling * support rescale_grad opt
-
- 13 7月, 2022 1 次提交
-
-
由 JZ-LIANG 提交于
* avoid sync with cpp in partition op * delay eval & predict mode * bugfix for gradient merge pass
-
- 11 7月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* add 'to_static' in engine api * fix cmakelist
-
- 29 6月, 2022 1 次提交
-
-
由 JZ-LIANG 提交于
* fixed bug for pass & engine * fixed bug for benchmark GPT-3
-
- 06 6月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* fix gradient merge * bug fix * update annotation
-
- 05 6月, 2022 1 次提交
-
-
由 Sing_chan 提交于
* use yapf to format all python file * yapf exclude two unittests file for they rely on writing and reading file, and format will break them * disable diff_py_file because too many diff files cause command following failed
-
- 01 6月, 2022 1 次提交
-
-
由 Yulong Ao 提交于
* [Auto Parallel] Add the parallel tuner * [Auto Parallel] Improve the parallel tuner and fix some bugs * upodate cost model * update import Resharder by dist op * update cost model * fix comp cost bug * update cost model * [Auto Parallel] Amend the dist attr for #processses=1 * update cost model and tuner * update cost model and tuner * update cost model and tuner * update cluster * update reshard * [Auto Parallel] Add the estimation from the cost model * [Auto Parallel] Reimplement the backup and restore functions * [Auto Parallel] Fix the bugs of the parallel tuner * [Auto Parallel] Update the engine api and dist context * [Auto Parallel] Work around the high order grad problem * [Auto Parallel] Add some miscellaneous improvements * [Auto Parallel] Add a unittest for DistributedContext Co-authored-by: Ncaozhou <caozhou@radi.ac.cn>
-
- 30 5月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* use original id in dist_op_context.grad_op_id_to_op_id * del assert * remove redundant map
-
- 19 5月, 2022 1 次提交
-
-
由 zhaoyingli 提交于
* slice data in dist_loader & flag to scale grad * bug fix * update unittest * enable static
-
- 10 5月, 2022 1 次提交
-
-
由 Yulong Ao 提交于
* [Auto Parallel] Refactor the engine api and parallelizer * [Auto Parallel] Fix the default dist op for the slice op * [Auto Parallel] Fix the format of planer.py * [Auto Parallel] Fix a bug
-