1. 22 8月, 2022 3 次提交
  2. 19 8月, 2022 2 次提交
  3. 18 8月, 2022 2 次提交
  4. 16 8月, 2022 2 次提交
    • F
      convert multihead to oss (#45019) · f706d95d
      feng_shuai 提交于
      * convert multihead to oss
      
      * fix:bug
      
      * fix:delete const cast
      
      * fix:don't support bias_qk
      
      * add vit pass
      
      * fix:convert bug and add preln_residual_bias
      
      * support length=-1
      
      * add UT for convert
      
      * add no_bias_qk support for gpu_multihead_op
      
      * delete infer_shape depends on bias_qk
      
      * oss just can be used in T4 and A*
      
      * fix:change api for ROCM CI
      f706d95d
    • W
      memoptim and fp16 mixed precision (#45132) · fa890092
      Wilber 提交于
      fa890092
  5. 15 8月, 2022 3 次提交
  6. 14 8月, 2022 1 次提交
  7. 12 8月, 2022 1 次提交
  8. 11 8月, 2022 1 次提交
  9. 10 8月, 2022 2 次提交
  10. 09 8月, 2022 1 次提交
  11. 08 8月, 2022 1 次提交
  12. 05 8月, 2022 2 次提交
  13. 04 8月, 2022 4 次提交
    • S
      Matmuls with activation and elementwise_add fuses (#44655) · 0420d514
      Sławomir Siwek 提交于
      * Add unit tests
      
      * matmul_v2 + activation
      
      * matmuls + elementwise_add
      
      * matmul_v2 postops
      
      * transform matmul to v2
      
      * opcompat
      
      * fix fusing matmul with multipe outs
      
      * add shape constraints
      
      * remove unused vars
      
      * change pass order
      
      * - Unit tests to be debugged
      
      - fix
      
      - refactor
      
      - diagnostic
      
      - more diagnostic
      
      - fix
      
      - Fix number two
      
      - fix
      
      - fix
      
      - fix
      
      - alpha added
      
      - more fixes
      
      - compilation fix
      
      - removed diagnostic code
      
      - cosmetic fixes
      
      * lint
      
      * add alpha constraint
      
      * merge matmul refactor
      
      * trigger CI
      
      * - fix
      
      * - another fix
      
      * code style
      
      * add support for matmul+elementwise_add+activation
      
      * code style
      
      * fix bfloat16 bugs
      
      * change append_binary to append_sum
      Co-authored-by: NJacek Czaja <jacek.czaja@intel.com>
      0420d514
    • Z
      [Paddle-TRT] add Rnn (#44678) · ffc8defa
      zhoutianzi666 提交于
      * add rnn
      ffc8defa
    • C
      fix bug (#44875) · c693a027
      ccrrong 提交于
      c693a027
    • W
      convert support multi block. (#44866) · b4a4eef2
      Wilber 提交于
      * convert support multi block.
      
      * update
      b4a4eef2
  14. 03 8月, 2022 1 次提交
  15. 02 8月, 2022 1 次提交
  16. 01 8月, 2022 4 次提交
  17. 29 7月, 2022 2 次提交
  18. 28 7月, 2022 1 次提交
  19. 26 7月, 2022 1 次提交
  20. 25 7月, 2022 1 次提交
  21. 22 7月, 2022 3 次提交
  22. 21 7月, 2022 1 次提交
    • M
      Fc fp16 (#44505) · 3e1280ea
      ming1753 提交于
      * fc support fp16
      
      * add a ‘,’ on paddle_pass_builder.cc
      
      * fc support fp16 on non-cuda.
      3e1280ea