- 21 1月, 2021 3 次提交
-
-
由 gongweibao 提交于
Pass device_ids info from launch to trainer
-
由 Void Main 提交于
Build praser for Hcom* operators
-
由 gongweibao 提交于
Add distribution supported
-
- 15 1月, 2021 2 次提交
- 14 1月, 2021 2 次提交
-
-
由 Chengmo 提交于
Co-authored-by: NseiriosPlus <tangwei12@baidu.com>
-
由 123malin 提交于
* test=develop, add distributed_infer
-
- 12 1月, 2021 3 次提交
-
-
由 JZ-LIANG 提交于
-
由 tangwei12 提交于
* rename sendrecv.proto to namespace paddle.distributed * split ps with distributed
-
由 Chengmo 提交于
* add save tensor support Co-authored-by: NseiriosPlus <tangwei12@baidu.com>
-
- 08 1月, 2021 2 次提交
-
-
由 Chen Weihang 提交于
-
由 Chengmo 提交于
* add tensor table
-
- 07 1月, 2021 1 次提交
-
-
由 Chen Weihang 提交于
* Simplify the options of spawn based on fleetrun * polish details * polish doc details
-
- 06 1月, 2021 1 次提交
-
-
由 gongweibao 提交于
-
- 05 1月, 2021 3 次提交
-
-
由 WangXi 提交于
-
由 gongweibao 提交于
-
由 Chen Weihang 提交于
* set flags_selectedd_gpus for spawn * add cond for unittest * Delete test_no_single_process_using_multi_gpus_in_spawn.py * Update spawn.py * Update nccl_context.cc
-
- 31 12月, 2020 2 次提交
- 25 12月, 2020 1 次提交
-
-
由 lilong12 提交于
* update, test=develop
-
- 24 12月, 2020 1 次提交
-
-
由 tangwei12 提交于
* oneps (3/4) Co-authored-by: NMrChengmo <cmchengmo@163.com> Co-authored-by: Nmalin10 <malin10@baidu.com> Co-authored-by: Nchengmo <chengmo@baidu.com>
-
- 22 12月, 2020 1 次提交
-
-
由 ShenLiang 提交于
* fix fleet for multi-stream * fix memcpy for ncclid * use sync to solve move operation
-
- 17 12月, 2020 1 次提交
-
-
由 WangXi 提交于
-
- 11 12月, 2020 1 次提交
-
-
由 JZ-LIANG 提交于
* Sharding add hybrid-dp feature * update sharding in distributed_strategy * update sharding unitest * revise code format for sharding
-
- 09 12月, 2020 1 次提交
-
-
由 ShenLiang 提交于
* add tensor_indices in AssignGroupBySize * add rebuild group in reducer
-
- 08 12月, 2020 1 次提交
-
-
由 lilong12 提交于
* update, test=develop
-
- 04 12月, 2020 1 次提交
-
-
由 ShenLiang 提交于
-
- 03 12月, 2020 3 次提交
-
-
由 gongweibao 提交于
-
由 ShenLiang 提交于
-
由 ShenLiang 提交于
* fix doc, test=document_fix
-
- 01 12月, 2020 2 次提交
- 30 11月, 2020 2 次提交
- 27 11月, 2020 4 次提交
-
-
由 ShenLiang 提交于
* add reducer * refine envent for memorycopy * add concat&split for allreduce * apply concat & split for fuse tensor * fix nccl dep * fix the untest, compile problem and ddp initialize problem * fix untest for mac & add some comments & solve the repeated param in sublayers * fix untest for windows & fix document
-
由 Chen Long 提交于
-
由 lilong12 提交于
-
由 lilong12 提交于
-
- 26 11月, 2020 2 次提交
-
-
由 ShenLiang 提交于
* add Inmemorydataset
-
由 JZ-LIANG 提交于
* add lars to fleet meta optimizer * add lamb to proto * add lamb to fleet meta optimizer * fixed syntax bug * fixed syntax bug * fixed syntax error in lamb, add config setter of lamb in distributed_strategy * trigger unitest to rerun * add new unitest func for lamb * revise unitest for lars and lamb * revise dgc meta unitest * revise lars document in distribute_strategy * revise lars lamb document in distributed_strategy.py * revise lars lamb document in distributed_strategy.py * add weight decay exclude logic to lars * restore optimzier.py * restore optimizer.py as develop except lars * add epsilon and exclude fn to distributed_sttrategy * add lars epsilon * revise unitest for fleet lars and lamb * revise lars lamb unitest for CI coverage * revise lars argument api * revise lars argument api * revise lars argument api * revise api doc of lars * fix op role * add sharding save and add_sync_comm_for_test function * add comm_analyse to utlis * revise sharding_utils * add sharding saving unittest * revise sharding utils for unittest * revise sharding en doc * update sharding utils api * add doc for sharding * fixed bug in sharding var size count * update varsize count in sharding * fix sharding num_nccl_comm * Revert "fix sharding num_nccl_comm" This reverts commit d51587c15e9323acf226ddd36154275f0d1daf76.
-