1. 09 1月, 2020 1 次提交
  2. 16 12月, 2019 1 次提交
  3. 12 12月, 2019 1 次提交
  4. 27 11月, 2019 1 次提交
    • M
      INT8 Fully-connected (#17641) · 5d7d5482
      Michał Gallus 提交于
      * Implement Int8 FC
      
      * Integrate FC into INT8v2
      
      test=develop
      
      * int8 FC: transpose weights before computing scales
      
      test=develop
      
      * Add support for activation_type string in FC
      
      test=develop
      
      * Disable MKL-DNN's FC in VGG16 and 19
      
      test=develop
      
      * Disable FC quantization when mkldnn FC is disabled
      
      test=develop
      
      * Solve PADDLE_ENFORCES in FC int8
      
      * Fix Paddle enforces and remove const cast
      
      test=develop
      
      * Fix style changes
      
      test=develop
      
      * Fix quantizer_tester test and add fc quantization
      
      test=develop
      
      * Fix FC test fail on CUDA
      
      * Remove unnecessary log from quantize placement pass
      
      test=develop
      
      * Add Thread ID to FC hash key
      
      test=develop
      
      * Add comments to MKL-DNN FC Kernel
      
      test=develop
      
      * Refactor quantizer
      
      test=develop
      
      * Fix linter issues
      
      test=develop
      
      * Fix crash in slim googlenet
      
      test=develop
      
      * Fix PADDLE_ENFORCE messages
      
      test=develop
      5d7d5482
  5. 08 11月, 2019 1 次提交
    • J
      Add transpose2 INT8 for mkl-dnn (#19424) · 77c20835
      joanna.wozna.intel 提交于
      * Add transpose2 INT8 for mkl-dnn
      
      test=develop
      
      * Fix test_transpose_int8_mkldnn
      
      test=develop
      
      * Revert "Merge branch 'develop' into transpose_int8_mkldnn_2"
      
      This reverts commit 34011bdb, reversing
      changes made to 2ce6473f.
      
      * Revert "Revert "Merge branch 'develop' into transpose_int8_mkldnn_2""
      
      This reverts commit 23754dd7.
      
      * Add template to TransposeMKLDNNHandler
      
      test=develop
      
      * Resolve conflict
      
      test=develop
      
      * Restore get_size and refactor
      
      test=develop
      77c20835
  6. 31 10月, 2019 1 次提交
    • H
      GradMaker for dygraph (#19706) · 8c4573a3
      hong 提交于
      * refactor dygraph,test=develop
      
      * fix failed unittest,test=develop
      
      * polish code,test=develop
      
      * check windows ci error,test=develop
      try to fix windows ci error by np.allclose,test=develop
      
      * polish vlog and profiler, test=develop
      
      * try to fix preceding ops order,test=develop
      
      * test transformer in windows ci, test=develop
      
      * use python c-api to speed up tracer.trace,test=develop
      
      * test=develop, fix docker with paddle nccl problem
      
      * test=develop, add ut for debug string and gradient_accumulator
      
      * test=develop, add tests for layer/gradient_accumulator/prepared_op
      
      * test=develop, fix complie error for test_prepared_op
      
      * test=develop, add more ut for dygraph
      
      * test=develop, create API.spec for dygraph api change
      
      * optimize grad maker; test=develop
      
      * optimize grad maker
      
      * test
      
      * grad make optim; test=develop
      
      * fix unittest bugs; test=develop
      
      * add dygraph grad op maker and split_op
      
      * grad op maker refactor; test=develop
      
      * add dygraph grad maker; test=develop
      
      * fix op deformable_conv_v1_op bug; test=develop
      
      * fix deformable_conv prroi pool bugs;
      
      * fix new op grad op maker bug; test=develop
      
      * fix split by ref bug; test=develop
      
      * fix dygraph auto prune bug; test=develop
      
      * fix test_trace bug; test=develop
      
      * fix fused emb seq pool bug; test=develop
      
      * remove useless code in op_desc file; test=develop
      
      * remove useless code, StrVarBaseNode; test=develop
      
      * fix review issues; test=develop
      
      * fix rank_loss grad maker; test=develop
      
      * remove flag in VarBase; test=develop
      
      * fix distributed_notify_op compile bug ; test=develop
      
      * fix reshape op double grad; test=develop
      
      * fix expand as op; test=develop
      
      * add impertive type_defs.h for demo_train; test=develop
      
      * fix inference lib cmake; test=develop
      
      * fix inference lib; test=develop
      
      * fix infernce_lib; test=develop
      
      * fix inference cmake; test=develop
      
      * fix inference lib; test=develop
      
      * fix inference lib; test=develop
      
      * remove condition dygraph grad maker, modify local name; test=develop
      
      * fix split grad maker bug; test=develop
      
      * fix pyramid_op bug; test=develop
      
      * change travis time out limit; test=develop
      
      * restore travis; test=develop
      
      * change timeout limit; test=develop
      8c4573a3
  7. 28 9月, 2019 1 次提交
    • B
      Follow comment of Merged QAT PR 18970 (#19979) · 9de67725
      bingyanghuang 提交于
      * Follow Wangzhen's comment in PR 18970, test=develop
      
      * Review comments, test=develop
      
      * Leave fake quantization around mul
      
      test=develop
      
      * Replace Fake with Real Quantized Mul
      
      test=develop
      
      * Fix bug in quantize placement pass
      
      Nodes in the graph now have checked type instead of node name when they are to be marked for quantization test=develop
      9de67725
  8. 27 9月, 2019 1 次提交
  9. 19 9月, 2019 1 次提交
  10. 03 9月, 2019 1 次提交
    • Y
      A a pass to enable the use of cudnn (#19346) · c5548178
      Yiqun Liu 提交于
      * Add a interface to enable cudnn for inference.
      
      * Add cudnn_placement_pass.
      test=develop
      
      * Set the default value of cudnn_enabled_op_types to null.
      test=develop
      
      * Write the common basic class, placement_pass_base, to refine the codes.
      test=develop
      
      * Call EnableCUDNN in unittest.
      test=develop
      
      * Refine cudnn_placement_pass tester.
      
      * Enable the testing of cudnn_placement_pass in inference's unittest.
      test=develop
      
      * Add the check of op kernels.
      test=develop
      c5548178
  11. 27 8月, 2019 1 次提交
  12. 21 8月, 2019 1 次提交
  13. 15 8月, 2019 1 次提交
  14. 13 8月, 2019 1 次提交
  15. 12 8月, 2019 1 次提交
  16. 01 7月, 2019 1 次提交
    • M
      Fix Pooling output scale (#18186) · 7023a86c
      Michał Gallus 提交于
      * Int8: Fix Pooling output scale
      
      test=develop
      
      * Update scales quantization for certain operators
      
      These include: concat, transpose, pool and reshape. test=develop
      
      * Move concat minimum scale finding to quantizer
      
      test=develop
      7023a86c
  17. 27 6月, 2019 1 次提交
  18. 10 6月, 2019 1 次提交
  19. 28 5月, 2019 1 次提交
    • J
      [MKL-DNN] conv_transpose mkldnn bias pass (#17644) · 6d8075ec
      Jacek Czaja 提交于
      * - changes to graph detector
      
      - Changes to pass
      
      - Added ut for new pass
      
      - use_pass
      
      - Added pass to mkldnn passes
      
      - fix to registration
      
      - improved verbose messaging for conv bias passes
      
      - Lint fixes
      
      test=develop
      
      * - Lint fixes
      
      test=develop
      6d8075ec
  20. 27 5月, 2019 1 次提交
    • S
      add Concat quantization (#17448) · 96845d21
      Sylwester Fraczek 提交于
      * add Concat quantization
      add unit test for quantizing concat
      fix for wrong value when the input is not in map of calculated scales
      add use_quantizer to concat_op.cc
      add scale_algo rules for concat
      
      test=develop
      
      * missing fix for multiple inputs quantize-squash
      
      * wojtuss review fix: adding comment
      
      test=develop
      96845d21
  21. 24 5月, 2019 3 次提交
    • M
      [MKL-DNN] Add Fully Connected Op for inference only(#15226) · 0c39b97b
      Michał Gallus 提交于
      * fuse mul and elementwise add to fc
      
      * Reimplement the FC forward operator
      
      * Fix FC MKLDNN integration by transposing weights
      
      * Add FC MKLDNN Pass
      
      test=develop
      
      * FC MKLDNN Pass: change memcpy to std::copy
      
      * Fix MKLDNN FC handling of mismatch input and weights dims
      
      * Lower tolerance for MKL-DNN in resnet50 test
      
      test=develop
      
      * Adjust FC to support MKLDNN Op placement
      
      test=develop
      
      * Adjust Placement Op to set use_mkldnn attribute for graph
      
      test=develop
      
      * MKLDNN FC: fix weights format so that gemm version is called
      
      test=develop
      
      * FC MKLDNN: Remove tolerance decrease from tester_helper
      
      * FC MKL-DNN: Refactor the code, change input reorder to weight reorder
      
      * MKL-DNN FC: Introduce operator caching
      
      test=develop
      
      * FC MKL-DNN: Fix the tensor type in ExpectedKernelType
      
      test=develop
      
      * FC MKL-DNN: fix style changes
      
      test=develop
      
      * FC MKL-DNN: fallback to native on non-supported dim sizes
      
      test=develop
      
      * FC MKLDNN: fix CMake paths
      
      test=develop
      
      * FC MKLDNN: Refine placement pass graph mkldnn attribute
      
      test=develop
      
      * Fix Transpiler error for fuse_conv_eltwise
      
      test=develop
      
      * Fix missing STL includes in files
      
      test=develop
      
      * FC MKL-DNN: Enable new output size computation
      
      Also, refine pass to comply with newest interface.
      test=develop
      
      * FC MKL-DNN: enable only when fc_mkldnn_pass is enabled
      
      * FC MKL-DNN: Allow Weights to use oi or io format
      
      * FC MKL-DNN: Adjust UT to work with correct dims
      
      test=develop
      
      * Enable MKL DEBUG for resnet50 analyzer
      
      test=develop
      
      * FC MKL-DNN: Improve Hashing function
      
      test=develop
      
      * FC MKL-DNN: Fix shape for fc weights in transpiler
      
      * FC MKL-DNN: Update input pointer in re-used fc primitive
      
      * Add log for not handling fc fuse for unsupported dims
      
      test=develop
      
      * FC MKL-DNN: Move transpose from pass to Op Kernel
      
      test=develop
      
      * FC MKL-DNN: Disable transpose in unit test
      
      test=develop
      
      * FC MKL-DNN: Remove fc_mkldnn_pass from default list
      
      * Correct Flag for fake data analyzer tests
      
      test=develop
      
      * FC MKL-DNN: Add comment about fc mkldnn pass disablement
      
      test=develop
      
      * FC MKL-DNN: Disable fc in int8 tests
      
      test=develop
      0c39b97b
    • S
      Conv concat relu quantization (#17466) · 5b2a3c4b
      Sylwester Fraczek 提交于
      * add conv_concat_relu fuse
      
      test=develop
      
      * add test code
      
      test=develop
      
      * added missing include with unordered_map
      
      test=develop
      
      * review fixes for wojtuss
      
      test=develop
      
      * remove 'should (not) be fused' comment statements
      
      one of them was invalid anyway
      
      test=develop
      5b2a3c4b
    • S
      fix quantize_squash_pass segfault when no tensor linked to Bias (#17292) · bccb0ba4
      Sylwester Fraczek 提交于
      * fix quantize_squash_pass segfault when there is no tensor linked do Bias input
      
      test=develop
      
      * add googlenet test
      
      test=develop
      
      * fix concat CreateKey not using input format
      
      test=develop
      bccb0ba4
  22. 22 5月, 2019 1 次提交
    • G
      Enable the convolution/relu6(bounded_relu) fusion for FP32 on Intel platform. (#17130) · 2281ebf0
      guomingz 提交于
      * Relu6 is the bottleneck op for Mobilenet-v2. As the mkldnn supports the conv/relu6 fusion, we implement it fusion via cpass way. Due to the int8 enabling for this fusion will be supported in MKLDNN v0.20, so this PR is focused on the fp32 optimization.
      
      Below table shows the benchmark(FPS) which measured on skx-8180(28 cores)
      Batch size | with fusion | without fusion
      -- | -- | --
      1 | 214.7 | 53.4
      50 | 1219.727 | 137.280
      
      test=develop
      
      * Fix the format issue
      
      test=develop
      
      * Add the missing nolint comments.
      
      test=develop
      
      * Fix the typos.
      
      test=develop
      
      * Register the conv_brelu_mkldnn_fuse_pass for the MKLDNN engine.
      
      test=develop
      
      * Adjust the indentation.
      
      test=develop
      
      * Add the test_conv_brelu_mkldnn_fuse_pass case.
      
      test=develop
      
      * Slightly update the code per Baidu comments.
      Let the parameter definition embedded into the code.
      That's will make the code easy to understand.
      
      test=develop
      2281ebf0
  23. 16 5月, 2019 1 次提交
  24. 28 3月, 2019 1 次提交
    • C
      Fix the interface of Pass::Apply (#16484) · ed61d67c
      chengduo 提交于
      * modify the interface of Pass::Allay
      test=develop
      
      * Polish code
      test=develop
      
      * Fix Travis CI
      test=develop
      
      * fix Pass::Apply interface
      test=develop
      
      * Fix Travis CI
      test=develop
      ed61d67c
  25. 25 3月, 2019 1 次提交
  26. 21 3月, 2019 1 次提交
  27. 26 2月, 2019 1 次提交
  28. 25 2月, 2019 1 次提交
  29. 22 2月, 2019 1 次提交
  30. 21 2月, 2019 1 次提交
  31. 29 1月, 2019 1 次提交