1. 13 8月, 2019 1 次提交
  2. 26 7月, 2019 1 次提交
  3. 22 7月, 2019 1 次提交
  4. 19 7月, 2019 1 次提交
  5. 09 7月, 2019 1 次提交
  6. 24 5月, 2019 1 次提交
    • M
      [MKL-DNN] Add Fully Connected Op for inference only(#15226) · 0c39b97b
      Michał Gallus 提交于
      * fuse mul and elementwise add to fc
      
      * Reimplement the FC forward operator
      
      * Fix FC MKLDNN integration by transposing weights
      
      * Add FC MKLDNN Pass
      
      test=develop
      
      * FC MKLDNN Pass: change memcpy to std::copy
      
      * Fix MKLDNN FC handling of mismatch input and weights dims
      
      * Lower tolerance for MKL-DNN in resnet50 test
      
      test=develop
      
      * Adjust FC to support MKLDNN Op placement
      
      test=develop
      
      * Adjust Placement Op to set use_mkldnn attribute for graph
      
      test=develop
      
      * MKLDNN FC: fix weights format so that gemm version is called
      
      test=develop
      
      * FC MKLDNN: Remove tolerance decrease from tester_helper
      
      * FC MKL-DNN: Refactor the code, change input reorder to weight reorder
      
      * MKL-DNN FC: Introduce operator caching
      
      test=develop
      
      * FC MKL-DNN: Fix the tensor type in ExpectedKernelType
      
      test=develop
      
      * FC MKL-DNN: fix style changes
      
      test=develop
      
      * FC MKL-DNN: fallback to native on non-supported dim sizes
      
      test=develop
      
      * FC MKLDNN: fix CMake paths
      
      test=develop
      
      * FC MKLDNN: Refine placement pass graph mkldnn attribute
      
      test=develop
      
      * Fix Transpiler error for fuse_conv_eltwise
      
      test=develop
      
      * Fix missing STL includes in files
      
      test=develop
      
      * FC MKL-DNN: Enable new output size computation
      
      Also, refine pass to comply with newest interface.
      test=develop
      
      * FC MKL-DNN: enable only when fc_mkldnn_pass is enabled
      
      * FC MKL-DNN: Allow Weights to use oi or io format
      
      * FC MKL-DNN: Adjust UT to work with correct dims
      
      test=develop
      
      * Enable MKL DEBUG for resnet50 analyzer
      
      test=develop
      
      * FC MKL-DNN: Improve Hashing function
      
      test=develop
      
      * FC MKL-DNN: Fix shape for fc weights in transpiler
      
      * FC MKL-DNN: Update input pointer in re-used fc primitive
      
      * Add log for not handling fc fuse for unsupported dims
      
      test=develop
      
      * FC MKL-DNN: Move transpose from pass to Op Kernel
      
      test=develop
      
      * FC MKL-DNN: Disable transpose in unit test
      
      test=develop
      
      * FC MKL-DNN: Remove fc_mkldnn_pass from default list
      
      * Correct Flag for fake data analyzer tests
      
      test=develop
      
      * FC MKL-DNN: Add comment about fc mkldnn pass disablement
      
      test=develop
      
      * FC MKL-DNN: Disable fc in int8 tests
      
      test=develop
      0c39b97b
  7. 22 5月, 2019 1 次提交
    • G
      Enable the convolution/relu6(bounded_relu) fusion for FP32 on Intel platform. (#17130) · 2281ebf0
      guomingz 提交于
      * Relu6 is the bottleneck op for Mobilenet-v2. As the mkldnn supports the conv/relu6 fusion, we implement it fusion via cpass way. Due to the int8 enabling for this fusion will be supported in MKLDNN v0.20, so this PR is focused on the fp32 optimization.
      
      Below table shows the benchmark(FPS) which measured on skx-8180(28 cores)
      Batch size | with fusion | without fusion
      -- | -- | --
      1 | 214.7 | 53.4
      50 | 1219.727 | 137.280
      
      test=develop
      
      * Fix the format issue
      
      test=develop
      
      * Add the missing nolint comments.
      
      test=develop
      
      * Fix the typos.
      
      test=develop
      
      * Register the conv_brelu_mkldnn_fuse_pass for the MKLDNN engine.
      
      test=develop
      
      * Adjust the indentation.
      
      test=develop
      
      * Add the test_conv_brelu_mkldnn_fuse_pass case.
      
      test=develop
      
      * Slightly update the code per Baidu comments.
      Let the parameter definition embedded into the code.
      That's will make the code easy to understand.
      
      test=develop
      2281ebf0
  8. 24 4月, 2019 1 次提交
  9. 16 4月, 2019 2 次提交
  10. 12 4月, 2019 1 次提交
  11. 26 3月, 2019 1 次提交
  12. 22 3月, 2019 1 次提交
    • X
      Enable MKL-DNN INT8 Concat Kernel. (#16156) · e235882c
      xiaolil1 提交于
      * Enable INT8 Concat Kernel to improve the performance of MobileNet-SSD.
      test=develop
      
      * Optimize UT format.
      test=develop
      
      * Fix UT file address issue.
      test=develop
      
      * Refine the license year.
      test=develop
      
      * Optimize code for new API.
      test=develop
      
      * Restructure INT8 Concat kernel.
      test=develop
      e235882c
  13. 19 3月, 2019 1 次提交
  14. 18 3月, 2019 2 次提交
  15. 06 3月, 2019 2 次提交
    • X
      Add Requantize OP (#15318) · a177d482
      xiaolil1 提交于
      * Enable INT8 ReQuantize OP
      test=develop
      
      * Clean code
      test=develop
      
      * Add comments
      test=develop
      
      * Revert "Clean code"
      test=develop
      
      This reverts commit a7a49b8a.
      
      * Modify requantize op test
      test=develop
      
      * fix requantize UT by moving public function to public test file.
      test=develop
      
      * Fix test fail due to file address change.
      test=develop
      
      * Change file address for requantize op.
      test=develop
      a177d482
    • L
      MKLDNN: Add UT for conv_transpose_mkldnn op. (#16030) · 21156b8d
      lidanqing 提交于
      * MKLDNN: Add UT for conv_transpose_mkldnn op.
      test=develop
      
      * MKLDNN: Add fuse_bias check UT for conv_transpose_mkldnn op.
      test=develop
      21156b8d
  16. 05 3月, 2019 1 次提交
  17. 04 3月, 2019 3 次提交
  18. 01 3月, 2019 1 次提交
  19. 25 2月, 2019 1 次提交
  20. 21 2月, 2019 2 次提交
  21. 29 1月, 2019 1 次提交