1. 05 12月, 2022 1 次提交
    • W
      Reverse roll fuse (#46914) · feb68dd1
      Wang Bojun 提交于
      * pass
      
      * pass
      
      * draft version
      
      * share mem opt
      
      * remove sharemem
      
      * add pattern for the case with circle_shift=0
      
      * add UT
      
      * pass opt
      
      * test_fix
      
      * code-commit
      
      * code-style
      
      * code style
      
      * code-style
      
      * ut-fix
      
      * op teller refine
      
      * resolve conflict
      
      * adjust position op_teller list and pass order for swin
      
      * ut code style update
      
      * adjust paddle pass order
      
      * refine pass order
      
      * refine pass order
      
      * refine pass order
      feb68dd1
  2. 01 12月, 2022 1 次提交
  3. 30 11月, 2022 2 次提交
  4. 21 11月, 2022 1 次提交
    • S
      add fc-residual quantization (#46917) · fed0ed34
      Sylwester Fraczek 提交于
      * add fc-residual quantization
      
      * revert removal of check for use_mkldnn
      
      * fix bug
      
      * add disable_logs
      
      * review fix
      
      call twice AreScalesPresntForNodes instead of if-else
      
      * rewrite residual input to output
      
      * revert fc mkldnn taking residual data
      
      * format fix
      
      * fix LoDTensor->DenseTensor
      
      * LoDTensor->DenseTensor
      
      * output->input
      
      * revert changes to unsupported script
      
      revert changes to unsupported script
      
      * remove fc residualdata from output blocklist in cpu_bfloat16_pass.cc
      fed0ed34
  5. 16 11月, 2022 1 次提交
  6. 15 11月, 2022 1 次提交
  7. 14 11月, 2022 1 次提交
  8. 07 11月, 2022 1 次提交
    • H
      suqeeze2 + transpose2 fuse onednn (#47592) · fa874a46
      Hui Zhang 提交于
      * suqeeze2 transpose2 fuse onednn
      
      * format
      
      * fix output shape
      
      * fix conflict
      
      * format
      
      * format
      
      * remove useless
      
      * remove log
      
      * simply pass
      
      * fix comment
      
      * fix
      
      * fix msg
      
      * fix error msg
      
      * format
      fa874a46
  9. 04 11月, 2022 1 次提交
  10. 20 10月, 2022 1 次提交
  11. 18 10月, 2022 1 次提交
    • W
      Merge layernorm trt fuse (#46320) · 5e9f491e
      Wang Bojun 提交于
      * first version, accuracy corrected
      
      * disable debug print
      
      * use blockReduceSum in phi
      
      * add UT
      
      * add opCompat
      
      * code style
      
      * code refine
      
      * bug fix
      
      * code refine
      
      * test fix
      
      * bugfix
      
      * codesytle fix
      
      * code style
      
      * code-style
      
      * code-style
      
      * code-style
      5e9f491e
  12. 17 10月, 2022 2 次提交
  13. 10 10月, 2022 1 次提交
  14. 07 9月, 2022 1 次提交
    • W
      Layernorm shift partition (#45736) · 960109af
      wenbin 提交于
      * first commit
      
      * conver done
      
      * correct format
      
      * layernorm_shift_partition
      
      * correct convert
      
      * redefine plugin
      
      * runable
      
      * bug fix
      
      * modify ShiftPartitionPattern
      
      * correct
      
      * add UT
      
      * modify ut
      
      * compile
      
      * modify enforce
      
      * modify UT
      960109af
  15. 22 8月, 2022 2 次提交
  16. 16 8月, 2022 2 次提交
    • F
      convert multihead to oss (#45019) · f706d95d
      feng_shuai 提交于
      * convert multihead to oss
      
      * fix:bug
      
      * fix:delete const cast
      
      * fix:don't support bias_qk
      
      * add vit pass
      
      * fix:convert bug and add preln_residual_bias
      
      * support length=-1
      
      * add UT for convert
      
      * add no_bias_qk support for gpu_multihead_op
      
      * delete infer_shape depends on bias_qk
      
      * oss just can be used in T4 and A*
      
      * fix:change api for ROCM CI
      f706d95d
    • W
      fix new quant (#45155) · 2fb65e44
      Wangzheee 提交于
      2fb65e44
  17. 04 8月, 2022 1 次提交
    • S
      Matmuls with activation and elementwise_add fuses (#44655) · 0420d514
      Sławomir Siwek 提交于
      * Add unit tests
      
      * matmul_v2 + activation
      
      * matmuls + elementwise_add
      
      * matmul_v2 postops
      
      * transform matmul to v2
      
      * opcompat
      
      * fix fusing matmul with multipe outs
      
      * add shape constraints
      
      * remove unused vars
      
      * change pass order
      
      * - Unit tests to be debugged
      
      - fix
      
      - refactor
      
      - diagnostic
      
      - more diagnostic
      
      - fix
      
      - Fix number two
      
      - fix
      
      - fix
      
      - fix
      
      - alpha added
      
      - more fixes
      
      - compilation fix
      
      - removed diagnostic code
      
      - cosmetic fixes
      
      * lint
      
      * add alpha constraint
      
      * merge matmul refactor
      
      * trigger CI
      
      * - fix
      
      * - another fix
      
      * code style
      
      * add support for matmul+elementwise_add+activation
      
      * code style
      
      * fix bfloat16 bugs
      
      * change append_binary to append_sum
      Co-authored-by: NJacek Czaja <jacek.czaja@intel.com>
      0420d514
  18. 27 7月, 2022 1 次提交
  19. 19 7月, 2022 1 次提交
  20. 12 7月, 2022 1 次提交
  21. 11 7月, 2022 1 次提交
  22. 06 7月, 2022 1 次提交
  23. 05 7月, 2022 1 次提交
  24. 30 6月, 2022 1 次提交
  25. 26 6月, 2022 1 次提交
  26. 23 6月, 2022 1 次提交
    • S
      [external reviewing] Params to int8 pass (#42625) · b8b2d6a9
      Sylwester Fraczek 提交于
      * sylwek
      
      prototype params to int8 pass
      
      * trying to make warmup work
      
      * wip
      
      * wip
      
      * change test to cpp test
      
      * review fixes, refactoring
      
      * more refactoring
      
      * add erasevars
      
      * change test to fixture
      
      * rename pass
      
      and reorder erasevars and graphsaferemovenodes
      
      * fix
      
      * more refactoring and fixed bug
      
      * formatting
      
      * remove scale count
      
      * enfroce message too short
      
      * remove erasevars
      
      erasevars couldbe cauuse of memory issues
      
      some other fixes
      
      * add count of successfull fuses to name of new nodes
      
      * FindVar -> GetVar and use ConvResidual pattern
      
      * use tensor->clear() instead of new variable
      
      * Update paddle/fluid/framework/ir/mkldnn/params_quantization_mkldnn_pass_tester.cc
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * Update paddle/fluid/framework/ir/mkldnn/params_quantization_mkldnn_pass_tester.cc
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * Update paddle/fluid/inference/tests/api/analyzer_lexical_analysis_gru_tester.cc
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * add log (review fix)c
      
      * review fix (2 functions to one)
      
      * code review: Conv->QuantizeConv
      
      * revert
      
      * fix formatting
      
      * remove unused functions
      
      * add paddle enforce
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      b8b2d6a9
  27. 21 6月, 2022 1 次提交
  28. 05 6月, 2022 1 次提交
  29. 02 6月, 2022 1 次提交
  30. 19 5月, 2022 1 次提交
  31. 12 5月, 2022 1 次提交
  32. 28 4月, 2022 1 次提交
    • T
      Bfloat16 refactor (#42238) · 8ad38701
      Tomasz Socha 提交于
      * Refactor Quantization
      
      * Refactor Dequantization
      
      * Classy solution
      
      * Style I
      
      * Style II
      
      * Style III
      
      * Use VLOG(4) for debug info
      
      * Style IV
      8ad38701
  33. 04 4月, 2022 1 次提交
  34. 02 4月, 2022 1 次提交
  35. 16 3月, 2022 1 次提交
  36. 14 3月, 2022 1 次提交
    • T
      Add an elementwise + activation fusion pass. (#36541) · 3f219160
      Tomasz Socha 提交于
      * Add elementwise add and activation fuse pass
      
      * Fix copy ellision
      
      * More flexible pattern detector
      
      * More flexible fusion pass
      
      * Update lists for pass
      
      * Add support for Pow operator
      
      * Add support for more activation types
      
      * Style
      
      * Rename fusion pass
      
      * First version of tests
      
      * Dirty version of pass
      
      * Polished version
      
      * Update pbtxt
      
      * Style
      
      * Update names
      
      * Style
      
      * Use PADDLE_ENFORCE_EQ
      
      * Save error message to variable
      
      * WO for error checks
      
      * CR
      
      * Static style check
      
      * Add missing 'activation_scale' attribute
      
      * Add relu6 and sigmoid activations
      
      * Style
      
      * Fix fuse list formating
      
      * Sync filenames for fuse pass files
      
      * Fix cmake after move
      
      * Fix registration
      
      * Fix pass name in tests
      
      * Add missing activations to checker
      
      * WIPS
      
      * Working mul op
      
      * Working sub
      
      * Working Add
      
      * Remove pten includes
      
      * Remove some forward declarations
      
      * Remove Includes
      
      * Fixes
      
      * Remove default kernels
      
      * Add check if post_ops attributes are avaliable
      
      * Style
      
      * Code adjustment
      
      * Register default kernels
      
      * We have year 2022 not 2021...
      Co-authored-by: Njakpiase <jakpia21@gmail.com>
      Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com>
      
      * Fast review fixes
      Co-authored-by: Njakpiase <jakpia21@gmail.com>
      Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com>
      
      * Review Fix
      
      * Rename one_dnn -> onednn
      
      * Style after review
      
      * Fast and dirty fix for quantization
      
      * Update tests
      
      * Style
      
      * Fix mkldnn_quantizer config
      
      * Add Joanna's suggestion.
      
      * Check if operator is explicitly disables on OneDNN
      
      * Try to use unregistered attributes
      
      * Style
      
      * Test new framework
      
      * FXI
      
      * FXII
      
      * Update test
      
      * Style
      Co-authored-by: Njakpiase <jakpia21@gmail.com>
      Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com>
      3f219160