1. 18 8月, 2022 1 次提交
  2. 16 8月, 2022 2 次提交
    • F
      convert multihead to oss (#45019) · f706d95d
      feng_shuai 提交于
      * convert multihead to oss
      
      * fix:bug
      
      * fix:delete const cast
      
      * fix:don't support bias_qk
      
      * add vit pass
      
      * fix:convert bug and add preln_residual_bias
      
      * support length=-1
      
      * add UT for convert
      
      * add no_bias_qk support for gpu_multihead_op
      
      * delete infer_shape depends on bias_qk
      
      * oss just can be used in T4 and A*
      
      * fix:change api for ROCM CI
      f706d95d
    • W
      memoptim and fp16 mixed precision (#45132) · fa890092
      Wilber 提交于
      fa890092
  3. 15 8月, 2022 1 次提交
  4. 14 8月, 2022 1 次提交
  5. 10 8月, 2022 1 次提交
  6. 05 8月, 2022 2 次提交
  7. 04 8月, 2022 3 次提交
    • S
      Matmuls with activation and elementwise_add fuses (#44655) · 0420d514
      Sławomir Siwek 提交于
      * Add unit tests
      
      * matmul_v2 + activation
      
      * matmuls + elementwise_add
      
      * matmul_v2 postops
      
      * transform matmul to v2
      
      * opcompat
      
      * fix fusing matmul with multipe outs
      
      * add shape constraints
      
      * remove unused vars
      
      * change pass order
      
      * - Unit tests to be debugged
      
      - fix
      
      - refactor
      
      - diagnostic
      
      - more diagnostic
      
      - fix
      
      - Fix number two
      
      - fix
      
      - fix
      
      - fix
      
      - alpha added
      
      - more fixes
      
      - compilation fix
      
      - removed diagnostic code
      
      - cosmetic fixes
      
      * lint
      
      * add alpha constraint
      
      * merge matmul refactor
      
      * trigger CI
      
      * - fix
      
      * - another fix
      
      * code style
      
      * add support for matmul+elementwise_add+activation
      
      * code style
      
      * fix bfloat16 bugs
      
      * change append_binary to append_sum
      Co-authored-by: NJacek Czaja <jacek.czaja@intel.com>
      0420d514
    • Z
      [Paddle-TRT] add Rnn (#44678) · ffc8defa
      zhoutianzi666 提交于
      * add rnn
      ffc8defa
    • W
      convert support multi block. (#44866) · b4a4eef2
      Wilber 提交于
      * convert support multi block.
      
      * update
      b4a4eef2
  8. 02 8月, 2022 1 次提交
  9. 01 8月, 2022 3 次提交
  10. 29 7月, 2022 1 次提交
  11. 28 7月, 2022 1 次提交
  12. 26 7月, 2022 1 次提交
  13. 22 7月, 2022 1 次提交
  14. 21 7月, 2022 2 次提交
  15. 19 7月, 2022 3 次提交
  16. 18 7月, 2022 1 次提交
  17. 15 7月, 2022 1 次提交
  18. 13 7月, 2022 1 次提交
  19. 12 7月, 2022 1 次提交
    • S
      matmul+activation fuse pass (#43519) · 3333a439
      Sławomir Siwek 提交于
      * add method for post ops
      
      * format code
      
      * gpd
      
      * format style
      
      * add matmul+act test
      
      * implement matmul+activation
      
      * whitespaces
      
      * code style
      
      * python code format
      
      * Increase UT timeout
      
      * code format
      
      * update style
      
      * generalize activation fuse passes
      
      * change order
      
      * Unify activation GPD
      
      * Revert changes with op_act
      
      * remove softmax mkldnn attrs
      
      * set common name for act attributes
      
      * whitespace
      
      * append postops by helper function
      
      * ut style
      
      * revert changes related to quantization
      
      * Reduce redundancy
      
      * reduce number of parameters
      
      * trigger CI
      
      * validate attribute
      
      * trim unit test
      3333a439
  20. 11 7月, 2022 2 次提交
  21. 08 7月, 2022 1 次提交
  22. 07 7月, 2022 1 次提交
  23. 30 6月, 2022 1 次提交
  24. 29 6月, 2022 2 次提交
  25. 28 6月, 2022 2 次提交
  26. 26 6月, 2022 1 次提交
  27. 24 6月, 2022 1 次提交
  28. 23 6月, 2022 1 次提交