- 27 9月, 2019 1 次提交
-
-
由 tangwei12 提交于
* add a base class for the Communicator * add AsyncCommunicator Impl for async distributed training
-
- 02 9月, 2019 1 次提交
-
-
由 gongweibao 提交于
-
- 28 8月, 2019 1 次提交
-
-
由 tangwei12 提交于
* fix correctness of the communicator * fix a bug in send thread when sending var context is empty, test=develop * add lookup_table_prefetch_op and prefetch optimize, test=develop * remove remote prefetch GPU supported * word2vec force with CPU, test=develop * test dist remote lookup table force with CPU, test=develop
-
- 26 8月, 2019 1 次提交
-
-
由 tangwei12 提交于
* fix sync mode hang in transpiler * remove sync mode in send/recv * replace PADDLE_ENFORCE with PADDLE_ENFORCE_NE
-
- 19 8月, 2019 1 次提交
-
-
由 zhang wenhui 提交于
add fl_listen_and_serv op for Federated_learning and fl_distribute_transpiler add this op to pserver program . This op just listen the endpoint and sum&scale.
-
- 18 8月, 2019 1 次提交
-
-
由 gongweibao 提交于
Unset unittests http_proxy env to avoid timeout.
-
- 14 8月, 2019 1 次提交
-
-
由 Leo Chen 提交于
* remove unused DefaultGradOpDescMaker in REGISTER_OPERATOR(), test=develop * remove SplitIdsOpGradMaker since it is buggy and not tested, update spec file, test=develop
-
- 27 6月, 2019 1 次提交
-
-
由 HaoRen 提交于
* fix prepare context redundant code problem, optimize executor by caching create_varaiables test=develop * supports collective training in executor * make fetch_list runable with variables, add more unittest for use_program_cache test=develop * fix comment test=develop * use unique name for nccl_id * supports output to stream in program_to_code * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code * set op role in collective training * add collective op role * remove orig file * add build optimizer by strategy * add collective strategy * refine collective strategy * add multi-process role maker * refine strategy building factory so that we can easily plugin more strategy * scale loss grad in collective sgd transpiler * add support for distributed fc * code format * revert some features for dist fc * add support for distributed fc training * fix prepare context redundant code problem, optimize executor by caching create_varaiables test=develop * supports collective training in executor * make fetch_list runable with variables, add more unittest for use_program_cache test=develop * use unique name for nccl_id * supports output to stream in program_to_code * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code * set op role in collective training * add collective op role * fix comment test=develop * remove orig file * add build optimizer by strategy * add collective strategy * refine collective strategy * add multi-process role maker * refine strategy building factory so that we can easily plugin more strategy * scale loss grad in collective sgd transpiler * add support for distributed fc * code format * revert some features for dist fc * add support for distributed fc training * test=develop add collective op unittest standard * test=develop remove the test_collective directory * test=develop remove the test_collective directory * remove slicegather test * code format for reducescatter * update attr of shard_index_op * Modify macro nccl_helper * remove test without distribute * macro collective_helper * marcro update * test=develop update support python3.5 * test=develop change gpu memory use to 0.1 when test * test=develop update ut equal func * test=develop set flags to 1.5 * test=develop fix pickle dumple py35 * test=develop fix divide in slice and add sync_comm_stream update atol and rtol to 1e-05 rm shard_index op and test modify read input from file to read from memory remove origin_program in framework and add i/o in c_sync_calc_stream * test=develop update unittest sync operator I/O
-
- 29 5月, 2019 1 次提交
-
-
由 gongweibao 提交于
-
- 27 5月, 2019 1 次提交
-
-
由 gongweibao 提交于
-
- 24 5月, 2019 1 次提交
-
-
由 chengduo 提交于
* This PR adds broadcast for multi-process. And it could be used in dynamic graph to broadcast parameters.
-
- 23 5月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
Async exe support communicator
-
- 17 5月, 2019 1 次提交
-
-
由 Yan Xu 提交于
* add var grad hook test=develop
-
- 25 4月, 2019 1 次提交
-
-
由 Yan Xu 提交于
implement dygraph.parallel.DataParallel to hook reduce op.
-
- 15 4月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
test=develop
-
- 11 4月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 10 4月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 01 4月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 27 3月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 25 3月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 23 3月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 20 3月, 2019 1 次提交
-
-
由 Wu Yi 提交于
* wip allreduce in op * wip * wip * wip * wip adding test * wip for conflict with mp mode * fix tests test=develop * fix cpu build test=develop * fix travis clang format test=develop * fix cpu build test=develop * update api.spec test=develop * delete comment test=develop * fix cpplint test=develop * fix test=develop * follow comment test=develop * add file test=develop * fix build test=develop * update test=develop * to be compatible with sync_bn, and fix mp mode in develop test=develop
-
- 19 3月, 2019 1 次提交
-
-
由 zhhsplendid 提交于
test=develop
-
- 18 3月, 2019 1 次提交
-
-
由 minqiyang 提交于
test=develop
-
- 16 3月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 15 3月, 2019 3 次提交
- 14 3月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 08 3月, 2019 3 次提交
-
-
由 Qiao Longfei 提交于
-
由 Qiao Longfei 提交于
-
由 Qiao Longfei 提交于
-
- 06 3月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 04 3月, 2019 3 次提交
-
-
由 Qiao Longfei 提交于
-
由 Qiao Longfei 提交于
-
由 Qiao Longfei 提交于
-
- 27 2月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 08 2月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 07 2月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-
- 06 2月, 2019 1 次提交
-
-
由 Qiao Longfei 提交于
-