1. 16 7月, 2019 1 次提交
  2. 15 7月, 2019 1 次提交
  3. 11 7月, 2019 2 次提交
    • H
      fix cudnn lstm shape bug; test=develop (#18492) · a20b2b43
      Hongyu Liu 提交于
      a20b2b43
    • Z
      Feature/buffer_shared_inplace (#17911) · d3003a16
      Zeng Jinle 提交于
      * feature/buffer_shared_inplace, test=develop
      
      * refine code, test=develop
      
      * fix elementwise_add op cpu inplace and sum inplace bug, test=develop
      
      * add unittest and debug log, test=develop
      
      * fix parallel_executor scope bug, polish code, test=develop
      
      * fix sum op, activation op, single_in_place_inference bug, test=develop
      
      * remove kLocalExecScopeName, test=develop
      
      * fix unittest,test=develop
      
      * fix out_var first version bug, test=develop
      
      * follow comments,test=develop
      d3003a16
  4. 10 7月, 2019 4 次提交
  5. 09 7月, 2019 3 次提交
  6. 08 7月, 2019 1 次提交
  7. 05 7月, 2019 1 次提交
  8. 04 7月, 2019 2 次提交
  9. 03 7月, 2019 6 次提交
  10. 02 7月, 2019 3 次提交
    • L
      rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id() (#18453) · 8f5fffca
      Leo Zhao 提交于
      * rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id()
      
      test=develop
      
      * update session id definition and adjust logic for default behavior
      
      test=develop
      
      * reset logic in mkldnn reuse as most of cases work in default.
      
      test=develop
      8f5fffca
    • Y
      supports collective training with programs (#18392) · a873fa84
      Yi Liu 提交于
      1. Since allreduce op has 4 reduce types, We split these four reduce types into four ops
      2. We also refined the collective op code, e.g. we separated the collective op kernel into CPUKernel and CUDAKernel, and remove the device specified DeviceContext parameter in template as we already knew the target DeviceContext
      3. We remove the newly added Collective op role to reduce the complexity of program and graph analysis
      a873fa84
    • C
      Add find_no_grad_vars in backward.py (#17942) · e0d8c6ac
      chengduo 提交于
      * add not_been_used_vars to no_grad_set
      test=develop
      e0d8c6ac
  11. 01 7月, 2019 2 次提交
  12. 28 6月, 2019 2 次提交
  13. 27 6月, 2019 4 次提交
    • T
      fix communicator with pyreader (#18350) · 999d9a59
      tangwei12 提交于
      * add is_runnning in communicator, test=develop
      999d9a59
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
    • S
      add int8 mkldnn prior_box (#17242) · 9252e8fa
      Sylwester Fraczek 提交于
      add prior_box quantization code
      
      add scale algo rules for prior box
      
      test=develop
      9252e8fa
    • J
      [MKL-DNN] Extending reusing to Elementwise_add_mkldnn op (#18146) · c2efdfd5
      Jacek Czaja 提交于
      * - Reusing of reuder used in elementwise_add_mkldnn
      
      - Added MKL-DNN sum prim reusing
      
      test=develop
      
      - Compilation fixes
      
      test=develop
      
      - Yet another compilation fix
      
      test=develop
      
      - Yet another compilation fix
      
      test=develo
      
      - Yet another linking fix
      
      test=develop
      
      - Final compilation fix
      
      test=develop
      
      - lint fixes
      
      test=develop
      
      - Lint fixes
      
      test=develop
      
      * - Fixes after review
      
      test=develop
      c2efdfd5
  14. 26 6月, 2019 2 次提交
  15. 25 6月, 2019 3 次提交
    • T
      fix softrelu doc (#18324) · 81ec5382
      tensor-tang 提交于
      * fix softrelu doc
      
      test=develop
      
      * update API doc
      
      test=develop
      81ec5382
    • H
      Sequence mask support tensor (#18249) · df2eee71
      Hongyu Liu 提交于
      * sequnce mask support max length tensor input; test=develop
      
      * add rnn_impl.py; test=develop
      
      * add basic gru lstm unittest; test=develop
      
      * fix api spec; test=develop
      
      * fix sequence_mask op bug;
      test=develop
      test=document_preview
      
      * change +-*x to elmentwise_op; test=develop
      
      * add mkl flag; test=develop
      
      * fix rnn impl bug; test=develop
      
      * update api spec; test=develop
      
      * fix doc bug; test=develop
      
      * fix lstm bugs; test=develop
      df2eee71
    • Q
      optimize communicator merge sparse gradient test=develop (#18159) · 0e08e91c
      Qiao Longfei 提交于
      * optimize communicator merge sparse gradient test=develop
      
      * revert multithread selected rows merge add test=develop
      
      * follow comment test=develop
      0e08e91c
  16. 24 6月, 2019 2 次提交
  17. 21 6月, 2019 1 次提交