1. 17 7月, 2019 1 次提交
  2. 27 6月, 2019 1 次提交
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
  3. 12 6月, 2019 1 次提交
  4. 11 6月, 2019 1 次提交
    • H
      Pipeline Concurrency (#17402) · 969e6378
      hutuxian 提交于
      Add Pipeline Concurrency Train Mode:
      - Cpp: pipeline_trainer & section_worker
      - Python: PipelineOptimizer
      - Add a new data_feed type: PrivateInstantDataFeed
      - Add a test demo of pipeline trainer and the test model is gnn
      - Do not support win32 now
      969e6378
  5. 23 5月, 2019 1 次提交
    • Z
      Fix allocator bug (#16712) · c6189637
      Zeng Jinle 提交于
      * Revert "Revert "Fix allocator bug""
      
      This reverts commit 174d0d0b.
      
      * Revert "fix travis ci"
      
      This reverts commit 5656fa9f.
      
      test=develop
      
      * add inlined_vector.h, test=develop
      
      * add inlined_vector_test,test=develop
      c6189637
  6. 29 3月, 2019 22 次提交
  7. 27 3月, 2019 1 次提交
  8. 26 3月, 2019 2 次提交
  9. 25 3月, 2019 1 次提交
    • S
      split PR · c20db635
      sneaxiy 提交于
      test=develop
      c20db635
  10. 24 3月, 2019 1 次提交
  11. 21 3月, 2019 1 次提交
    • S
      add more unittest · 953214ad
      sneaxiy 提交于
      modify allocator strategy
      remove changes of legacy buddy_allocator
      test=develop
      953214ad
  12. 11 3月, 2019 1 次提交
  13. 05 3月, 2019 1 次提交
  14. 04 3月, 2019 3 次提交
  15. 01 3月, 2019 1 次提交
  16. 11 2月, 2019 1 次提交
    • B
      Fix ngraph compile WITH_DISTRIBUTE=ON (#15636) · f4a0e684
      baojun 提交于
      * fix compile issue with_distribute test=develop
      
      * simplified logic test=develop
      
      * use ngraph dependency test=develop
      
      * set cpu only test=develop
      
      * update test and eliminate fp16 test test=develop
      f4a0e684