- 26 4月, 2020 1 次提交
-
-
由 Chen Weihang 提交于
* add to_readable_code method, test=develop * polish doc details, test=develop * polish doc note, test=develop * fix unittest error, test=develop * fix coverage, test=develop * add print test, test=develop * add print param, test=develop * hidden to_readable_code api, test=develop * remove original tool methods, test=develop * remove old api using code, test=develop
-
- 18 3月, 2020 1 次提交
-
-
由 tangwei12 提交于
-
- 28 2月, 2020 1 次提交
-
-
由 tianshuo78520a 提交于
-
- 25 2月, 2020 1 次提交
-
-
由 hutuxian 提交于
* Add two types of Metric Calculator: MultiTaskCalculator & CmatchRankCalculator. * Add a config for DynamicAdjustChannelNum function to denote whether we will discard the remaining instances when they are not be distributed evenly. * Remove CPU code in Pull/PushSparse and we will add it back when testing it fully. * Fix some known issues: such as copying persistable vars after one epoch running.
-
- 23 2月, 2020 1 次提交
-
-
由 tianshuo78520a 提交于
-
- 22 2月, 2020 1 次提交
-
-
由 tangwei12 提交于
* add sync communicator and implement
-
- 15 2月, 2020 1 次提交
-
-
由 tangwei12 提交于
* add deprecated for distribute transpiler, will delete it after 2.0.0, test=develop
-
- 17 1月, 2020 1 次提交
-
-
由 tangwei12 提交于
* add half_async in the communicator * fix DistributedStrategy
-
- 13 1月, 2020 1 次提交
-
-
由 123malin 提交于
* test=develop, bug fix for sparse recorder
-
- 07 1月, 2020 2 次提交
- 06 1月, 2020 1 次提交
-
-
由 123malin 提交于
* add distributed_strategy
-
- 12 12月, 2019 1 次提交
-
-
由 tangwei12 提交于
* add fake init for the trainer, fix large memory hold in the trainer * do not merge recv vars from a remote endpoint, test=develop * add recv and save op, merge slice var in one op, save memory * remove hsigmoid with pull sparse, test=develop
-
- 06 12月, 2019 1 次提交
-
-
由 hutuxian 提交于
* Add a single_process_multi_thread transpiler. * Add some UTs. * Fix some API description.
-
- 28 11月, 2019 1 次提交
-
-
由 Kaipeng Deng 提交于
* add Adam beta1/beta2 support Variable. test=develop
-
- 01 11月, 2019 1 次提交
-
-
由 123malin 提交于
* update pserver decay blocks * update distributed notify handler
-
- 17 10月, 2019 1 次提交
-
-
由 tangwei12 提交于
* fix fetch handler error with pslib * fix distributed lookup table op with 1 pserver
-
- 15 10月, 2019 2 次提交
-
-
由 Chengmo 提交于
* test=develop,Fix communicator slow bug * test=develop, delete if() in stop_worker() * test=develop * fix UT, test=develop * fix bug in fetch handler, test=develop * fix bug in fetch handler, test=develop * test=develop, fix fetch barrier bug * test=develop, bug fix * test=develop, bug fix * test=develop, fix bug
-
由 123malin 提交于
* bug fix: invalid learning rate decay in pserver async mode
-
- 11 10月, 2019 1 次提交
-
-
由 tangwei12 提交于
* doc fix, test=develop, test=document_fix
-
- 09 10月, 2019 1 次提交
-
-
由 Chengmo 提交于
* test=develop,test=document_fix,fix transpiler doc,add API.spec
-
- 07 10月, 2019 2 次提交
- 30 9月, 2019 2 次提交
-
-
由 Chengmo 提交于
* refector geo sgd & communicator
-
由 Zeng Jinle 提交于
* add deprecated memory optimize doc, test=develop, test=document_fix * merge develop to solve conflict, test=develop, test=document_fix
-
- 26 9月, 2019 1 次提交
-
-
由 123malin 提交于
* fix DistributeTranspilerConfig document, test=develop
-
- 16 9月, 2019 1 次提交
-
-
由 tangwei12 提交于
fix wrong place with distributed_lookup_table
-
- 06 9月, 2019 1 次提交
-
-
由 123malin 提交于
* fleet api add input check, test=develop
-
- 28 8月, 2019 2 次提交
-
-
由 Yi Liu 提交于
test=develop
-
由 tangwei12 提交于
* fix correctness of the communicator * fix a bug in send thread when sending var context is empty, test=develop * add lookup_table_prefetch_op and prefetch optimize, test=develop * remove remote prefetch GPU supported * word2vec force with CPU, test=develop * test dist remote lookup table force with CPU, test=develop
-
- 26 8月, 2019 1 次提交
-
-
由 tangwei12 提交于
* fix sync mode hang in transpiler * remove sync mode in send/recv * replace PADDLE_ENFORCE with PADDLE_ENFORCE_NE
-
- 16 8月, 2019 1 次提交
-
-
由 Tao Luo 提交于
* remove unused inference_transpiler unit-tests test=develop * remove InferenceTranspiler usage in quantize_transpiler.py test=develop
-
- 12 8月, 2019 1 次提交
-
-
由 gongweibao 提交于
Polish fleet API to support cuda collective mode and nccl2 mode
-
- 10 8月, 2019 1 次提交
-
-
由 Zeng Jinle 提交于
* deprecate python memory optimize, test=develop * remove memory_optimize in unittests, test=develop * add unittests to deprecated interfaces, test=develop
-
- 29 7月, 2019 1 次提交
-
-
由 Zeng Jinle 提交于
* remove legacy memory optimization codes, test=develop * follow huihuang's comments,test=develop * follow luotao's comments, test=develop
-
- 23 7月, 2019 1 次提交
-
-
由 Yi Liu 提交于
* supports distributed classification training * update API.spec * fix evenly division in python3 * change "index_range" to "index_num" in shard_index operator test=document_preview test=develop
-
- 22 7月, 2019 1 次提交
-
-
由 tangwei12 提交于
do some odd jobs, test=develop
-
- 11 7月, 2019 1 次提交
-
-
由 gongweibao 提交于
-
- 02 7月, 2019 1 次提交
-
-
由 Yi Liu 提交于
1. Since allreduce op has 4 reduce types, We split these four reduce types into four ops 2. We also refined the collective op code, e.g. we separated the collective op kernel into CPUKernel and CUDAKernel, and remove the device specified DeviceContext parameter in template as we already knew the target DeviceContext 3. We remove the newly added Collective op role to reduce the complexity of program and graph analysis
-
- 27 6月, 2019 1 次提交
-
-
由 HaoRen 提交于
* fix prepare context redundant code problem, optimize executor by caching create_varaiables test=develop * supports collective training in executor * make fetch_list runable with variables, add more unittest for use_program_cache test=develop * fix comment test=develop * use unique name for nccl_id * supports output to stream in program_to_code * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code * set op role in collective training * add collective op role * remove orig file * add build optimizer by strategy * add collective strategy * refine collective strategy * add multi-process role maker * refine strategy building factory so that we can easily plugin more strategy * scale loss grad in collective sgd transpiler * add support for distributed fc * code format * revert some features for dist fc * add support for distributed fc training * fix prepare context redundant code problem, optimize executor by caching create_varaiables test=develop * supports collective training in executor * make fetch_list runable with variables, add more unittest for use_program_cache test=develop * use unique name for nccl_id * supports output to stream in program_to_code * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code * set op role in collective training * add collective op role * fix comment test=develop * remove orig file * add build optimizer by strategy * add collective strategy * refine collective strategy * add multi-process role maker * refine strategy building factory so that we can easily plugin more strategy * scale loss grad in collective sgd transpiler * add support for distributed fc * code format * revert some features for dist fc * add support for distributed fc training * test=develop add collective op unittest standard * test=develop remove the test_collective directory * test=develop remove the test_collective directory * remove slicegather test * code format for reducescatter * update attr of shard_index_op * Modify macro nccl_helper * remove test without distribute * macro collective_helper * marcro update * test=develop update support python3.5 * test=develop change gpu memory use to 0.1 when test * test=develop update ut equal func * test=develop set flags to 1.5 * test=develop fix pickle dumple py35 * test=develop fix divide in slice and add sync_comm_stream update atol and rtol to 1e-05 rm shard_index op and test modify read input from file to read from memory remove origin_program in framework and add i/o in c_sync_calc_stream * test=develop update unittest sync operator I/O
-