1. 24 5月, 2019 1 次提交
    • S
      Conv concat relu quantization (#17466) · 5b2a3c4b
      Sylwester Fraczek 提交于
      * add conv_concat_relu fuse
      
      test=develop
      
      * add test code
      
      test=develop
      
      * added missing include with unordered_map
      
      test=develop
      
      * review fixes for wojtuss
      
      test=develop
      
      * remove 'should (not) be fused' comment statements
      
      one of them was invalid anyway
      
      test=develop
      5b2a3c4b
  2. 22 5月, 2019 1 次提交
    • G
      Enable the convolution/relu6(bounded_relu) fusion for FP32 on Intel platform. (#17130) · 2281ebf0
      guomingz 提交于
      * Relu6 is the bottleneck op for Mobilenet-v2. As the mkldnn supports the conv/relu6 fusion, we implement it fusion via cpass way. Due to the int8 enabling for this fusion will be supported in MKLDNN v0.20, so this PR is focused on the fp32 optimization.
      
      Below table shows the benchmark(FPS) which measured on skx-8180(28 cores)
      Batch size | with fusion | without fusion
      -- | -- | --
      1 | 214.7 | 53.4
      50 | 1219.727 | 137.280
      
      test=develop
      
      * Fix the format issue
      
      test=develop
      
      * Add the missing nolint comments.
      
      test=develop
      
      * Fix the typos.
      
      test=develop
      
      * Register the conv_brelu_mkldnn_fuse_pass for the MKLDNN engine.
      
      test=develop
      
      * Adjust the indentation.
      
      test=develop
      
      * Add the test_conv_brelu_mkldnn_fuse_pass case.
      
      test=develop
      
      * Slightly update the code per Baidu comments.
      Let the parameter definition embedded into the code.
      That's will make the code easy to understand.
      
      test=develop
      2281ebf0
  3. 20 5月, 2019 1 次提交
  4. 08 5月, 2019 1 次提交
  5. 07 5月, 2019 1 次提交
    • Cherry-pick benchmark related changes from release/1.4 (#17156) · a72dbe9a
      石晓伟 提交于
      * cherry-pick commit from 88770542
      
      * cherry-pick commit from 3f0b97df
      
      * cherry-pick from 16691:Anakin subgraph support yolo_v3 and faster-rcnn
      
      (cherry picked from commit 8643dbc2)
      
      * Cherry-Pick from 16662 : Anakin subgraph cpu support
      
      (cherry picked from commit 7ad182e1)
      
      * Cherry-pick from 1662, 16797.. : add anakin int8 support
      
      (cherry picked from commit e14ab180)
      
      * Cherry-pick from 16813 : change singleton to graph RegistBlock
      test=release/1.4
      
      (cherry picked from commit 4b9fa423)
      
      * Cherry Pick : 16837 Support ShuffleNet and MobileNet-v2
      
      Support ShuffleNet and MobileNet-v2, test=release/1.4
      
      (cherry picked from commit a6fb066f)
      
      * Cherry-pick : anakin subgraph add opt config layout argument #16846
      test=release/1.4
      
      (cherry picked from commit 8121b3ec)
      
      * 1. add shuffle_channel_detect
      
      (cherry picked from commit 6efdea89)
      
      * update shuffle_channel op convert, test=release/1.4
      
      (cherry picked from commit e4726a06)
      
      * Modify symbol export rules
      
      test=develop
      a72dbe9a
  6. 28 3月, 2019 1 次提交
    • N
      Anakin ssd support · d065b5bf
      nhzlx 提交于
      refine trt first run
      add quant dequant fuse pass
      omit simplify_anakin_priorbox_detection template
      omit transpose_flatten_concat_fuse template
      test=develop
      d065b5bf
  7. 25 3月, 2019 1 次提交
  8. 21 3月, 2019 1 次提交
  9. 20 3月, 2019 3 次提交
  10. 19 3月, 2019 4 次提交
  11. 18 3月, 2019 1 次提交
    • W
      Add cpu_quantize_pass for C-API quantization (#16127) · 2579ade4
      Wojciech Uss 提交于
      * Add cpu_quantize_pass for C-API quantization
      
      test=develop
      
      * add cpu_quantize_pass test
      
      * fix lint: add include memory unorderd_map and unordered_set
      
      test=develop
      
      * fuse_relu 1
      
      test=develop
      
      * tuned 2 without squash
      
      * fixes
      
      test=develop
      
      * remove unused vars
      
      test=develop
      
      * refactored
      
      test=develop
      
      * fix lint c-style cast -> C++ style cast
      
      test=develop
      
      * remove QuantMax and c style casts
      
      test=develop
      
      * last usage of QuantMax removed
      
      test=develop
      
      * Fix Analysis Predictor UT
      
      Check if memory_optimize_pass has already been added
      to the analysis config before adding a new one, so
      that it is not added multiple times.
      test=develop
      
      * change map to unordered_map
      
      fix the forgotten part of cpu_quantize_pass_tester.cc
      
      test=develop
      
      * removed quantized attribute
      
      * fixed cpu_quantize_pass_tester and op attr comments
      
      test=develop
      
      * removed redundant line
      
      test=debug
      
      * removed gmock
      
      test=develop
      
      * fix after merge
      2579ade4
  12. 16 3月, 2019 1 次提交
  13. 15 3月, 2019 1 次提交
    • Q
      Support sync batch norm. (#16121) · 8ad672a2
      qingqing01 提交于
      * Support Sync Batch Norm.
      * Note, do not enable it in one device.
      
      Usage:
      
      build_strategy = fluid.BuildStrategy()
      build_strategy.sync_batch_norm = True
      binary = fluid.compiler.CompiledProgram(tp).with_data_parallel(
              loss_name=loss_mean.name,
              build_strategy=build_strategy)
      8ad672a2
  14. 14 3月, 2019 1 次提交
    • W
      Add cpu_quantize_squash_pass for C-API quantization (#16128) · b9252f3d
      Wojciech Uss 提交于
      * Add cpu_quantize_squash_pass for C-API quantization
      
      test=develop
      
      * add cpu_quantize_squash_pass teste
      
      * fix lint: add include memory unorderd_map and unordered_set
      
      test=develop
      
      * lint fix 2
      
      * fixes
      
      test=develop
      
      * refactored
      
      test=develop
      
      * fix windows ci
      
      test=develop
      b9252f3d
  15. 13 3月, 2019 1 次提交
  16. 26 2月, 2019 1 次提交
  17. 22 2月, 2019 1 次提交
  18. 31 1月, 2019 1 次提交
  19. 29 1月, 2019 1 次提交
  20. 21 1月, 2019 1 次提交
    • D
      Memory optimization of depthwise conv op and group norm op (#15313) · 9f8f0fc2
      Dun 提交于
      * mem opt
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * test=develop
      
      * refine code  test=develop
      
      * refine code  test=develop
      
      * refine code  test=develop
      
      * refine code  test=develop
      
      * refine with cub test=develop
      
      * fix mkldnn test && remove comments && test=develop
      
      * polish code && test=develop
      
      * add only_forward test && test=develop
      9f8f0fc2
  21. 14 1月, 2019 1 次提交
  22. 13 1月, 2019 1 次提交
  23. 11 1月, 2019 1 次提交
  24. 10 1月, 2019 1 次提交
  25. 08 1月, 2019 1 次提交
  26. 07 1月, 2019 1 次提交
  27. 25 12月, 2018 1 次提交
  28. 16 12月, 2018 1 次提交
  29. 14 12月, 2018 1 次提交
  30. 07 12月, 2018 1 次提交
  31. 03 12月, 2018 1 次提交
  32. 15 11月, 2018 1 次提交
    • S
      add mkldnn prop_kind phase for inference-only case to pooling and activations (#14278) · 8a1eeec5
      Sylwester Fraczek 提交于
      * add is_test to pooling and activations
      
      add prop_kind support for layers activation. conv and pooling
      
      add a pass that sets is_test to true
      
      add transpiler version of is_test pass
      
      test=develop
      
      * patch test and pass
      
      test=develop
      
      * add pass to analyzer.h
      
      test=develop
      
      * add is_test attr description & pass only on mkldnn
      
      in:
      activation_op.cc
      batch_norm_op.cc
      conv_op.cc
      dropout_op.cc
      lrn_op.cc
      pool_op.cc
      sequence_pool_op.cc
      softmax_op.cc
      
      * fix is_test handling for activation pool and conv
      
      * change description of is_test for all layers again
      
      * remove GetAttr(use_mkldnn) from pass
      
      * rename correct_mkldnn_test_phase to is_test
      
      and remove dependency on MKLDNN
      test=develop
      
      * review fix magic number
      
      * two if(..)s into one
      
      * Check is_test once and pass mkldnn forward prop kind
      
      * dereference shared_ptr with * (without get())
      
      test=develop
      
      * add is_test_pass back
      
      test=develop
      8a1eeec5
  33. 14 11月, 2018 1 次提交
  34. 06 11月, 2018 1 次提交
  35. 31 10月, 2018 1 次提交
    • S
      add depthwise conv mkldnn pass · 4e2aaf01
      Sylwester Fraczek 提交于
      added depthwise conv mkldnn pass which for MKLDNN changes depthwise_conv operator to conv operator because for mkldnn this is the same api
      test=develop
      4e2aaf01