1. 22 5月, 2023 2 次提交
  2. 19 5月, 2023 2 次提交
    • W
      [XPU] fix fallback (#53801) · 4b85e5db
      wz1qqx 提交于
      4b85e5db
    • L
      Add flash attention to speedup fused_gate_attention. (#52731) · d29c1f8e
      limingshu 提交于
      * Reorganize the forward codes of flash-attention.
      
      * Fix forward.
      
      * Remove some noused codes.
      
      * Simplify codes and fix backward.
      
      * Change all LOG(INFO) to VLOG and fix the backward.
      
      * add scale for AF2 flash_attn, much thanks to xreki and shaojie for debug these codes
      
      * decrease the effect of debug print on performance
      
      * Unify the initialize of flashattn arguments.
      
      * Rewirte the reshape of temp_mask and temp_bias.
      
      * API support use_flash_attn.
      
      * Fix compiling error on CI.
      
      * Try to crop the flash-attention lib.
      
      * Correct the condition of whether can use flash-attn.
      
      * Remove the softmax_out argument.
      
      * Remove is_causal.
      
      * Polish codes.
      
      * Fix qkv_transpose_out's shape and scaling of Q * K.
      
      * Update commit of flash-attention.
      
      ---------
      Co-authored-by: NLiu Yiqun <liuyiqun01@baidu.com>
      d29c1f8e
  3. 18 5月, 2023 2 次提交
  4. 15 5月, 2023 2 次提交
  5. 12 5月, 2023 2 次提交
  6. 11 5月, 2023 4 次提交
  7. 10 5月, 2023 1 次提交
  8. 09 5月, 2023 3 次提交
  9. 08 5月, 2023 2 次提交
  10. 06 5月, 2023 1 次提交
  11. 28 4月, 2023 1 次提交
  12. 27 4月, 2023 1 次提交
  13. 26 4月, 2023 1 次提交
  14. 25 4月, 2023 1 次提交
    • Y
      [PHI]Add flags macro for PHI (#52991) · 22e96bde
      YuanRisheng 提交于
      * add flags for phi
      
      * fix compile bugs
      
      * fix ci bugs
      
      * fix inference bugs
      
      * fix cinn' bugs
      
      * fix cinn bugs
      
      * perfect code according comment
      
      * fix ci bugs
      
      * fix ci bugs
      22e96bde
  15. 24 4月, 2023 1 次提交
    • G
      remove some [-Wunused-parameter] (#53185) · 834eb2ba
      Galaxy1458 提交于
      * test,test=develop
      
      * test,test=develop
      
      * test,test=develop
      
      * test,test=develop
      
      * test,test=develop
      
      * test,test=develop
      
      * test,test=develop
      
      * test ,test=develop
      834eb2ba
  16. 20 4月, 2023 2 次提交
  17. 19 4月, 2023 1 次提交
  18. 17 4月, 2023 1 次提交
  19. 14 4月, 2023 2 次提交
  20. 13 4月, 2023 3 次提交
  21. 10 4月, 2023 2 次提交
  22. 09 4月, 2023 1 次提交
  23. 07 4月, 2023 1 次提交
  24. 06 4月, 2023 1 次提交
    • S
      Remove oneDNN-specific attributes from matmul (#49444) · 4d97b25d
      Sławomir Siwek 提交于
      * replace matmul with matmul_v2 in fuse passes
      
      * Remove fusion logic from matmul
      
      * removing fusion methods
      
      * add proper name
      
      * adjust namespaces
      
      * clean attrs in python tests
      
      * delete checkpoint and restore matmul version
      
      * remove unused code
      
      * matmul and reshape/transpose fuses migrated
      
      * split MatmulOneDNN headers
      
      * fuse activation and eltwise_add
      
      * add fuse_activation
      
      * matmul_transpose_reshape/reshape_transpose_matmul
      
      * matmul + elementwise_add (fused)
      
      * activation temporary modifciation
      
      * restore matmul(v1) version 0
      
      * merge newest develop
      
      * remove depedency from other PR
      
      * revert pbtxt
      
      * remove placeholders from matmul_v2
      
      * add description in OPMaker
      
      * remove matmul_v2_op.h and all depedencies
      
      * remove dims changing in base op
      
      * add possibility to fuse already fused_matmul
      
      * restart broken CI
      
      * Empty-Commit
      
      * revert matmul_utils.h
      
      * codestyle
      
      * adjust imports
      
      * add pbtxt file
      
      * 100% matmul unit tests coverage
      
      * trigger CI with minimal changes to develop
      
      * adjust changes to develop
      
      * add fused_matmul op
      
      * inherit base ops
      
      * add "v2"
      
      * move OPMaker
      
      * Gradually add fused_matmul files
      
      * second batch of fused_matmul changes
      
      * split infershapes of matmul_v2 and fused_matmul
      
      * merge code from other PR
      
      * 2023
      
      * inherit fused_matmul from matmul_v2
      
      * Update paddle/phi/backends/onednn/onednn_reuse.h
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * Update paddle/phi/kernels/fusion/onednn/fused_matmul_kernel.cc
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      
      * resolve conflicts
      
      * codestyle
      
      * simplify isgemmlinear
      
      * 2023
      
      * remove import
      
      * reuse methods
      
      * matmul_v2_mkldnn cleanup
      
      * simplify ExecuteMatMulV1Grad
      
      * matmul refactored
      
      * fc
      
      * SetOutMemDescWithLogicalLayoutFusesSupport
      
      * matmul_v2
      
      * alpha support
      
      * group repetetive funcs
      
      * matmul utils
      
      * execute matmul methods
      
      * restore registered kernel names
      
      * split header and impl files
      
      * remove double negatives
      
      * reduce numer of modified files
      
      * adjust ExecuteMatmul
      
      * add scales for ut
      
      * dates
      
      * limit number of modified files
      
      * fluid imports
      
      * remove alpha
      
      * codestyle
      
      ---------
      Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
      4d97b25d