1. 06 4月, 2023 1 次提交
    • S
      Remove oneDNN-specific attributes from matmul (#49444) · 4d97b25d
      Sławomir Siwek 提交于
      * replace matmul with matmul_v2 in fuse passes
      
      * Remove fusion logic from matmul
      
      * removing fusion methods
      
      * add proper name
      
      * adjust namespaces
      
      * clean attrs in python tests
      
      * delete checkpoint and restore matmul version
      
      * remove unused code
      
      * matmul and reshape/transpose fuses migrated
      
      * split MatmulOneDNN headers
      
      * fuse activation and eltwise_add
      
      * add fuse_activation
      
      * matmul_transpose_reshape/reshape_transpose_matmul
      
      * matmul + elementwise_add (fused)
      
      * activation temporary modifciation
      
      * restore matmul(v1) version 0
      
      * merge newest develop
      
      * remove depedency from other PR
      
      * revert pbtxt
      
      * remove placeholders from matmul_v2
      
      * add description in OPMaker
      
      * remove matmul_v2_op.h and all depedencies
      
      * remove dims changing in base op
      
      * add possibility to fuse already fused_matmul
      
      * restart broken CI
      
      * Empty-Commit
      
      * revert matmul_utils.h
      
      * codestyle
      
      * adjust imports
      
      * add pbtxt file
      
      * 100% matmul unit tests coverage
      
      * trigger CI with minimal changes to develop
      
      * adjust changes to develop
      
      * add fused_matmul op
      
      * inherit base ops
      
      * add "v2"
      
      * move OPMaker
      
      * Gradually add fused_matmul files
      
      * second batch of fused_matmul changes
      
      * split infershapes of matmul_v2 and fused_matmul
      
      * merge code from other PR
      
      * 2023
      
      * inherit fused_matmul from matmul_v2
      
      * Update paddle/phi/backends/onednn/onednn_reuse.h
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * Update paddle/phi/kernels/fusion/onednn/fused_matmul_kernel.cc
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * resolve conflicts
      
      * codestyle
      
      * simplify isgemmlinear
      
      * 2023
      
      * remove import
      
      * reuse methods
      
      * matmul_v2_mkldnn cleanup
      
      * simplify ExecuteMatMulV1Grad
      
      * matmul refactored
      
      * fc
      
      * SetOutMemDescWithLogicalLayoutFusesSupport
      
      * matmul_v2
      
      * alpha support
      
      * group repetetive funcs
      
      * matmul utils
      
      * execute matmul methods
      
      * restore registered kernel names
      
      * split header and impl files
      
      * remove double negatives
      
      * reduce numer of modified files
      
      * adjust ExecuteMatmul
      
      * add scales for ut
      
      * dates
      
      * limit number of modified files
      
      * fluid imports
      
      * remove alpha
      
      * codestyle
      
      ---------
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      4d97b25d
  2. 27 3月, 2023 1 次提交
    • S
      Fused elementwise_(mul/div) (#50428) · 968f7f24
      Sławomir Siwek 提交于
      * extract Op and OPMaker to .h
      
      * extend pattern for fused_op
      
      * set "with_residual" default to false
      
      * adjust fuse passes
      
      * remove fc+eltwise flag
      
      * fused_output_scale
      
      * activation attrs
      
      * remove extra attrs
      
      * fix int8/bf16 unit tests
      
      * simplify RecomputeOutputDims
      
      * remove unused method
      
      * Add description for attributes
      
      * add extra check
      
      * adjust op compats
      
      * update quantize test
      
      * fix protobuf parsing error
      
      * fix int8 performance
      
      * fused elementwises
      
      * merge develop
      
      * remove activation
      
      * restore activation for existing add/sub ops
      968f7f24
  3. 22 3月, 2023 1 次提交
    • S
      Extract fused_transpose op dedicated for oneDNN fuse passes (#50021) · 02296977
      Sławomir Siwek 提交于
      * extract common methods to reuse
      
      * add header for transpose ops
      
      * fused_transpose
      
      * Split big function
      
      * transpose2 tests
      
      * fused_transpose
      
      * Apply extra attributes
      
      * add pbtxt file
      
      * update pbtxt
      
      * Merge develop
      
      * add more strict op compats
      
      * code  style
      
      * remove mkldnn_data_type
      
      * unify SetOutMemDescWithReshape2FuseSupport
      
      * adjust quantize-dequantize for transpose
      
      * remove appendact
      
      * transpose2 quantization
      
      * fix int8 tests
      
      * adjust transpose_op to current develop
      
      * delete fusion code from transpose_kernel
      
      * add fused transpose to NHWC unittest
      
      * change order
      02296977
  4. 13 3月, 2023 1 次提交
    • S
      Fused softplus (#51087) · fdcfa04f
      Sławomir Siwek 提交于
      * mkldnn->onednn
      
      * fused softplus op + kernel
      
      * remove extra attributes
      
      * add missing handler
      
      * change var name
      fdcfa04f
  5. 06 2月, 2023 1 次提交
  6. 03 2月, 2023 1 次提交
    • S
      Replace matmul(v2) with fused_matmul during oneDNN fuse passes (#49515) · 5cfe1645
      Sławomir Siwek 提交于
      * replace matmul with matmul_v2 in fuse passes
      
      * Remove fusion logic from matmul
      
      * removing fusion methods
      
      * add proper name
      
      * adjust namespaces
      
      * clean attrs in python tests
      
      * delete checkpoint and restore matmul version
      
      * remove unused code
      
      * matmul and reshape/transpose fuses migrated
      
      * split MatmulOneDNN headers
      
      * fuse activation and eltwise_add
      
      * add fuse_activation
      
      * matmul_transpose_reshape/reshape_transpose_matmul
      
      * matmul + elementwise_add (fused)
      
      * activation temporary modifciation
      
      * merge newest develop
      
      * remove depedency from other PR
      
      * revert pbtxt
      
      * remove placeholders from matmul_v2
      
      * add description in OPMaker
      
      * remove matmul_v2_op.h and all depedencies
      
      * remove dims changing in base op
      
      * add possibility to fuse already fused_matmul
      
      * restart broken CI
      
      * Empty-Commit
      
      * revert matmul_utils.h
      
      * codestyle
      
      * adjust imports
      
      * add pbtxt file
      
      * 100% matmul unit tests coverage
      
      * trigger CI with minimal changes to develop
      
      * adjust changes to develop
      
      * add fused_matmul op
      
      * inherit base ops
      
      * add "v2"
      
      * move OPMaker
      
      * Gradually add fused_matmul files
      
      * second batch of fused_matmul changes
      
      * split infershapes of matmul_v2 and fused_matmul
      
      * inherit fused_matmul from matmul_v2
      
      * Update paddle/phi/backends/onednn/onednn_reuse.h
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * Update paddle/phi/kernels/fusion/onednn/fused_matmul_kernel.cc
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      ---------
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      5cfe1645
  7. 09 1月, 2023 1 次提交
    • W
      Preln groupnorm (#49463) · 591be3bd
      wenbin 提交于
      * skip_groupnorm
      
      * init
      
      * preln
      
      * add ut
      
      * more assert
      
      * set timeout
      
      * fix windows ci issue
      591be3bd
  8. 21 12月, 2022 1 次提交
    • Z
      Refactor Pass for fused_conv (#48848) · 7f0eb2e3
      zyfncg 提交于
      * refactor conv_activation_mkldnn_fuse_pass
      
      * refactor conv_affine_channel_mkldnn_fuse_pass
      
      * fix conv_activation_mkldnn_fuse_pass
      
      * fix mkldnn unittest
      
      * refactor int8_scale_calculation_mkldnn_pass and params_quantization_mkldnn_pass
      
      * refactor conv_elementwise_add_mkldnn_fuse_pass
      
      * fix quant
      
      * refactor conv_bn_fuse_pass
      
      * fix conv_bn_fuse_pass
      
      * refactor depthwise_conv_bn_fuse_pass
      
      * fix unittest
      
      * fix conv_bn_fuse_pass
      
      * remove redundant conv2d in params_quantization_mkldnn_pass
      
      * fix params_quantization_mkldnn_pass_tester
      7f0eb2e3
  9. 06 12月, 2022 1 次提交
    • Z
      Clear extra input (Bias, ResidualData) in OpMaker of conv2d (#47579) · 0a2dfa38
      zyfncg 提交于
      * delete Bias and ResidualData in OpMaker of conv2d
      
      * delete extra input of conv3d
      
      * refactor pass of conv_bias_fusion
      
      * fix mkldnn dependency
      
      * fix mkldnn compile
      
      * fix test_conv_bias_mkldnn_fuse_pass
      
      * police some code
      
      * remove useless log
      
      * fix analyzer_vit_ocr_tester
      
      * fix conv_activation_mkldnn_fuse_pass
      
      * fix test_analyzer_ocr
      
      * add fused_conv_sig
      
      * fix performence regression
      
      * fix performance regression
      0a2dfa38
  10. 28 11月, 2022 1 次提交
  11. 31 5月, 2022 1 次提交
  12. 10 5月, 2022 1 次提交
  13. 02 4月, 2022 1 次提交
  14. 20 12月, 2021 1 次提交
  15. 14 12月, 2021 1 次提交
  16. 02 11月, 2021 1 次提交
  17. 21 10月, 2021 1 次提交
    • J
      Added matmul_v2+transpose+reshape fuse pass (#36481) · 856cb9c5
      jakpiase 提交于
      * added base changes for matmul_v2+trans+resh fuse pass
      
      * added full matmul_v2+transpose+reshape pass
      
      * removed a file added by mistake
      
      * added reviewers suggestions
      
      * Changed ops type in checking capatibility version
      
      * Deteled one statement
      856cb9c5
  18. 26 8月, 2021 1 次提交
  19. 15 7月, 2021 1 次提交
  20. 14 7月, 2021 1 次提交
  21. 08 7月, 2021 1 次提交
  22. 07 7月, 2021 1 次提交
  23. 06 7月, 2021 1 次提交
  24. 02 7月, 2021 1 次提交
  25. 01 7月, 2021 1 次提交
  26. 28 6月, 2021 1 次提交
  27. 25 6月, 2021 2 次提交
  28. 23 6月, 2021 3 次提交
  29. 22 6月, 2021 1 次提交
  30. 18 6月, 2021 3 次提交
  31. 17 6月, 2021 1 次提交
  32. 16 6月, 2021 3 次提交
  33. 15 6月, 2021 1 次提交