1. 12 8月, 2019 1 次提交
  2. 09 8月, 2019 1 次提交
  3. 05 8月, 2019 1 次提交
    • L
      fix warpctc.dll not found issue (#18761) · a43a763b
      liuwei1031 提交于
      * fix warpctc.dll not found issue, test=develop
      
      * revert the linux platform change, test=develop
      
      * delete warpctc_lib_path.h.in, test=develop
      
      * add SetPySitePackagePath function
      
      * fix warpctc.dylib not found issue on Mac, test=develop
      
      * improve the paddle lib path setting logic, test=develop
      
      * fix mac ci issue caused by test_warpctc_op unittest, test=develop
      
      * tweak code, test=develop
      a43a763b
  4. 01 8月, 2019 2 次提交
  5. 31 7月, 2019 1 次提交
  6. 30 7月, 2019 1 次提交
  7. 29 7月, 2019 1 次提交
  8. 27 7月, 2019 1 次提交
  9. 25 7月, 2019 1 次提交
  10. 23 7月, 2019 2 次提交
  11. 19 7月, 2019 1 次提交
  12. 18 7月, 2019 2 次提交
  13. 16 7月, 2019 2 次提交
  14. 11 7月, 2019 2 次提交
  15. 10 7月, 2019 2 次提交
  16. 09 7月, 2019 2 次提交
  17. 08 7月, 2019 1 次提交
    • T
      add mkldnn shapeblob cache clear strategy (#18513) · fe32879d
      Tao Luo 提交于
      * add mkldnn shapeblob cache clear strategy
      
      test=develop
      
      * refine with comments
      
      test=develop
      
      * make cache clear strategy more safey
      
      test=develop
      
      * add lock for GetShapeBlobSize
      
      test=develop
      fe32879d
  18. 04 7月, 2019 1 次提交
  19. 03 7月, 2019 1 次提交
  20. 02 7月, 2019 2 次提交
    • L
      rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id() (#18453) · 8f5fffca
      Leo Zhao 提交于
      * rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id()
      
      test=develop
      
      * update session id definition and adjust logic for default behavior
      
      test=develop
      
      * reset logic in mkldnn reuse as most of cases work in default.
      
      test=develop
      8f5fffca
    • Y
      supports collective training with programs (#18392) · a873fa84
      Yi Liu 提交于
      1. Since allreduce op has 4 reduce types, We split these four reduce types into four ops
      2. We also refined the collective op code, e.g. we separated the collective op kernel into CPUKernel and CUDAKernel, and remove the device specified DeviceContext parameter in template as we already knew the target DeviceContext
      3. We remove the newly added Collective op role to reduce the complexity of program and graph analysis
      a873fa84
  21. 01 7月, 2019 1 次提交
  22. 28 6月, 2019 1 次提交
  23. 27 6月, 2019 4 次提交
    • H
      add dependecy of collective_helper (#18365) · 9931bc64
      HaoRen 提交于
      * add dependecy of collective_helper
      
      * test=develop
      fix dependecy of collective_helper
      9931bc64
    • M
      Reset DeviceContext after quantization warmup (#18182) · 84096932
      Michał Gallus 提交于
      test=develop
      84096932
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
    • J
      [MKL-DNN] Extending reusing to Elementwise_add_mkldnn op (#18146) · c2efdfd5
      Jacek Czaja 提交于
      * - Reusing of reuder used in elementwise_add_mkldnn
      
      - Added MKL-DNN sum prim reusing
      
      test=develop
      
      - Compilation fixes
      
      test=develop
      
      - Yet another compilation fix
      
      test=develop
      
      - Yet another compilation fix
      
      test=develo
      
      - Yet another linking fix
      
      test=develop
      
      - Final compilation fix
      
      test=develop
      
      - lint fixes
      
      test=develop
      
      - Lint fixes
      
      test=develop
      
      * - Fixes after review
      
      test=develop
      c2efdfd5
  24. 18 6月, 2019 1 次提交
  25. 14 6月, 2019 1 次提交
  26. 11 6月, 2019 2 次提交
    • J
      [MKL-DNN] Thread-Safety for MKL-DNN reusing Part 1 (#17965) · 84bb45c0
      Jacek Czaja 提交于
      * - removed is_reusing_
      
      * - Added TID to keys for reusing apart from softmax PD
      
      * - compilation fix
      
      * - Yet another compilation fix
      
      * - Batch Norm and Conv adapted
      
      * - Fix to softmax MT
      
      * - Fixes to MT code of MKL-DNN
      
      * - Lint fixes
      
      test=develop
      84bb45c0
    • H
      Pipeline Concurrency (#17402) · 969e6378
      hutuxian 提交于
      Add Pipeline Concurrency Train Mode:
      - Cpp: pipeline_trainer & section_worker
      - Python: PipelineOptimizer
      - Add a new data_feed type: PrivateInstantDataFeed
      - Add a test demo of pipeline trainer and the test model is gnn
      - Do not support win32 now
      969e6378
  27. 10 6月, 2019 1 次提交
  28. 07 6月, 2019 1 次提交