1. 28 10月, 2019 1 次提交
  2. 15 10月, 2019 2 次提交
  3. 07 10月, 2019 1 次提交
  4. 30 9月, 2019 1 次提交
  5. 27 9月, 2019 1 次提交
  6. 02 9月, 2019 1 次提交
  7. 28 8月, 2019 1 次提交
    • T
      Fix the correctness of async mode at distributed training (#18863) · 65c73684
      tangwei12 提交于
      * fix correctness of the communicator
      
      * fix a bug in send thread when sending var context is empty, test=develop
      
      * add lookup_table_prefetch_op and prefetch optimize, test=develop
      
      * remove remote prefetch GPU supported
      
      * word2vec force with CPU, test=develop
      
      * test dist remote lookup table force with CPU, test=develop
      65c73684
  8. 26 8月, 2019 1 次提交
  9. 19 8月, 2019 1 次提交
  10. 18 8月, 2019 1 次提交
  11. 14 8月, 2019 1 次提交
  12. 27 6月, 2019 1 次提交
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
  13. 29 5月, 2019 1 次提交
  14. 27 5月, 2019 1 次提交
  15. 24 5月, 2019 1 次提交
  16. 23 5月, 2019 1 次提交
  17. 17 5月, 2019 1 次提交
  18. 25 4月, 2019 1 次提交
  19. 15 4月, 2019 1 次提交
  20. 11 4月, 2019 1 次提交
  21. 10 4月, 2019 1 次提交
  22. 01 4月, 2019 1 次提交
  23. 27 3月, 2019 1 次提交
  24. 25 3月, 2019 1 次提交
  25. 23 3月, 2019 1 次提交
  26. 20 3月, 2019 1 次提交
    • W
      Collective ops (#15572) · 6382b62f
      Wu Yi 提交于
      * wip allreduce in op
      
      * wip
      
      * wip
      
      * wip
      
      * wip adding test
      
      * wip for conflict with mp mode
      
      * fix tests test=develop
      
      * fix cpu build test=develop
      
      * fix travis clang format test=develop
      
      * fix cpu build test=develop
      
      * update api.spec test=develop
      
      * delete comment test=develop
      
      * fix cpplint test=develop
      
      * fix test=develop
      
      * follow comment test=develop
      
      * add file test=develop
      
      * fix build test=develop
      
      * update test=develop
      
      * to be compatible with sync_bn, and fix mp mode in develop test=develop
      6382b62f
  27. 19 3月, 2019 1 次提交
  28. 18 3月, 2019 1 次提交
  29. 16 3月, 2019 1 次提交
  30. 15 3月, 2019 3 次提交
  31. 14 3月, 2019 1 次提交
  32. 08 3月, 2019 3 次提交
  33. 06 3月, 2019 1 次提交
  34. 04 3月, 2019 2 次提交