1. 09 7月, 2019 2 次提交
  2. 08 7月, 2019 1 次提交
    • T
      add mkldnn shapeblob cache clear strategy (#18513) · fe32879d
      Tao Luo 提交于
      * add mkldnn shapeblob cache clear strategy
      
      test=develop
      
      * refine with comments
      
      test=develop
      
      * make cache clear strategy more safey
      
      test=develop
      
      * add lock for GetShapeBlobSize
      
      test=develop
      fe32879d
  3. 04 7月, 2019 1 次提交
  4. 03 7月, 2019 1 次提交
  5. 02 7月, 2019 2 次提交
    • L
      rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id() (#18453) · 8f5fffca
      Leo Zhao 提交于
      * rename mkldnn set/get_cur_thread_id() to set/get_cur_mkldnn_session_id()
      
      test=develop
      
      * update session id definition and adjust logic for default behavior
      
      test=develop
      
      * reset logic in mkldnn reuse as most of cases work in default.
      
      test=develop
      8f5fffca
    • Y
      supports collective training with programs (#18392) · a873fa84
      Yi Liu 提交于
      1. Since allreduce op has 4 reduce types, We split these four reduce types into four ops
      2. We also refined the collective op code, e.g. we separated the collective op kernel into CPUKernel and CUDAKernel, and remove the device specified DeviceContext parameter in template as we already knew the target DeviceContext
      3. We remove the newly added Collective op role to reduce the complexity of program and graph analysis
      a873fa84
  6. 01 7月, 2019 1 次提交
  7. 28 6月, 2019 1 次提交
  8. 27 6月, 2019 4 次提交
    • H
      add dependecy of collective_helper (#18365) · 9931bc64
      HaoRen 提交于
      * add dependecy of collective_helper
      
      * test=develop
      fix dependecy of collective_helper
      9931bc64
    • M
      Reset DeviceContext after quantization warmup (#18182) · 84096932
      Michał Gallus 提交于
      test=develop
      84096932
    • H
      supports collective communicated training (#18175) · b7128bac
      HaoRen 提交于
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * fix comment
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * fix prepare context redundant code problem, optimize executor by caching create_varaiables
      test=develop
      
      * supports collective training in executor
      
      * make fetch_list runable with variables, add more unittest for use_program_cache
      test=develop
      
      * use unique name for nccl_id
      
      * supports output to stream in program_to_code
      
      * insert sync_comm_stream before regularization; add skip_op_callstack capability in program_to_code
      
      * set op role in collective training
      
      * add collective op role
      
      * fix comment
      test=develop
      
      * remove orig file
      
      * add build optimizer by strategy
      
      * add collective strategy
      
      * refine collective strategy
      
      * add multi-process role maker
      
      * refine strategy building factory so that we can easily plugin more strategy
      
      * scale loss grad in collective sgd transpiler
      
      * add support for distributed fc
      
      * code format
      
      * revert some features for dist fc
      
      * add support for distributed fc training
      
      * test=develop
      add collective op unittest standard
      
      * test=develop
      remove the test_collective directory
      
      * test=develop
      remove the test_collective directory
      
      * remove slicegather test
      
      * code format for reducescatter
      
      * update attr of shard_index_op
      
      * Modify macro nccl_helper
      
      * remove test without distribute
      
      * macro collective_helper
      
      * marcro update
      
      * test=develop
      update support python3.5
      
      * test=develop change gpu memory use to 0.1 when test
      
      * test=develop
      update ut equal func
      
      * test=develop
      set flags to 1.5
      
      * test=develop fix pickle dumple  py35
      
      * test=develop
      fix divide in slice and add sync_comm_stream
      update atol and rtol to 1e-05
      rm shard_index op and test
      modify read input from file to read from memory
      remove origin_program in framework and add i/o in c_sync_calc_stream
      
      * test=develop update unittest sync operator I/O
      b7128bac
    • J
      [MKL-DNN] Extending reusing to Elementwise_add_mkldnn op (#18146) · c2efdfd5
      Jacek Czaja 提交于
      * - Reusing of reuder used in elementwise_add_mkldnn
      
      - Added MKL-DNN sum prim reusing
      
      test=develop
      
      - Compilation fixes
      
      test=develop
      
      - Yet another compilation fix
      
      test=develop
      
      - Yet another compilation fix
      
      test=develo
      
      - Yet another linking fix
      
      test=develop
      
      - Final compilation fix
      
      test=develop
      
      - lint fixes
      
      test=develop
      
      - Lint fixes
      
      test=develop
      
      * - Fixes after review
      
      test=develop
      c2efdfd5
  9. 18 6月, 2019 1 次提交
  10. 14 6月, 2019 1 次提交
  11. 11 6月, 2019 2 次提交
    • J
      [MKL-DNN] Thread-Safety for MKL-DNN reusing Part 1 (#17965) · 84bb45c0
      Jacek Czaja 提交于
      * - removed is_reusing_
      
      * - Added TID to keys for reusing apart from softmax PD
      
      * - compilation fix
      
      * - Yet another compilation fix
      
      * - Batch Norm and Conv adapted
      
      * - Fix to softmax MT
      
      * - Fixes to MT code of MKL-DNN
      
      * - Lint fixes
      
      test=develop
      84bb45c0
    • H
      Pipeline Concurrency (#17402) · 969e6378
      hutuxian 提交于
      Add Pipeline Concurrency Train Mode:
      - Cpp: pipeline_trainer & section_worker
      - Python: PipelineOptimizer
      - Add a new data_feed type: PrivateInstantDataFeed
      - Add a test demo of pipeline trainer and the test model is gnn
      - Do not support win32 now
      969e6378
  12. 10 6月, 2019 1 次提交
  13. 07 6月, 2019 1 次提交
  14. 05 6月, 2019 1 次提交
  15. 04 6月, 2019 1 次提交
  16. 03 6月, 2019 2 次提交
  17. 29 5月, 2019 1 次提交
  18. 27 5月, 2019 1 次提交
  19. 24 5月, 2019 2 次提交
  20. 23 5月, 2019 2 次提交
  21. 22 5月, 2019 1 次提交
    • G
      Enable the convolution/relu6(bounded_relu) fusion for FP32 on Intel platform. (#17130) · 2281ebf0
      guomingz 提交于
      * Relu6 is the bottleneck op for Mobilenet-v2. As the mkldnn supports the conv/relu6 fusion, we implement it fusion via cpass way. Due to the int8 enabling for this fusion will be supported in MKLDNN v0.20, so this PR is focused on the fp32 optimization.
      
      Below table shows the benchmark(FPS) which measured on skx-8180(28 cores)
      Batch size | with fusion | without fusion
      -- | -- | --
      1 | 214.7 | 53.4
      50 | 1219.727 | 137.280
      
      test=develop
      
      * Fix the format issue
      
      test=develop
      
      * Add the missing nolint comments.
      
      test=develop
      
      * Fix the typos.
      
      test=develop
      
      * Register the conv_brelu_mkldnn_fuse_pass for the MKLDNN engine.
      
      test=develop
      
      * Adjust the indentation.
      
      test=develop
      
      * Add the test_conv_brelu_mkldnn_fuse_pass case.
      
      test=develop
      
      * Slightly update the code per Baidu comments.
      Let the parameter definition embedded into the code.
      That's will make the code easy to understand.
      
      test=develop
      2281ebf0
  22. 20 5月, 2019 1 次提交
  23. 15 5月, 2019 1 次提交
  24. 10 5月, 2019 1 次提交
    • Q
      Double backward of conv2d. (#17211) · e32c9888
      qingqing01 提交于
      * Add conv2d_grad_grad_op
      * Extracte the cuDNN conv algo searching code in conv_cudnn_helper.h.
          - Now use it in conv2d_grad_grad.
          - Will simply the searching code in conv2d and conv2d_grad in next PR.
      * Enhance and fix bug in unit testing of gradient_checker.
      * Support to fetch empty variables,return None in Python.
      e32c9888
  25. 08 5月, 2019 3 次提交
  26. 07 5月, 2019 1 次提交
  27. 30 4月, 2019 1 次提交
  28. 28 4月, 2019 1 次提交
  29. 23 4月, 2019 1 次提交