1. 25 9月, 2019 1 次提交
    • B
      add support of matmul with multiple head even different width and height (#19708) · c670058a
      Bob Zhu 提交于
      * add support of matmul with multiple head even different width and height
      
      Original matmul with multiple head supports only the mat_a.width == mat_b.height,
      in that case, mat_b will be horizontally split. In this patch, we extend the
      support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height,
      in this case, mab_b will be vertically split.
      
      One example is A is [3, 8], B is [2, 16], head_number is 4. In this
      case, A will be split as [3, 2], B will be (vertically) split as
      [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16]
      
      test=develop
      
      * add support of matmul with multiple head even different width and height
      
      Original matmul with multiple head supports only the mat_a.width == mat_b.height,
      in that case, mat_b will be horizontally split. In this patch, we extend the
      support when mat_a.width != mat_b.height but mat_a.width/head_number == mat_b.height,
      in this case, mab_b will be vertically split.
      
      One example is A is [3, 8], B is [2, 16], head_number is 4. In this
      case, A will be split as [3, 2], B will be (vertically) split as
      [2, 4]. The final result will be 4 matrix of 4 matrix of [3,4], i.e. [3, 16]
      
      test=develop
      
      * refactor the code of matmul with multiple head even different width and height
      
      test=develop
      c670058a
  2. 20 8月, 2019 1 次提交
  3. 24 7月, 2019 1 次提交
    • B
      Extend Matmul to support matrix multiplication with multiple heads (#18570) · 220eef60
      Bob Zhu 提交于
      * extend matmul op to support multiple head multiplication
      
      With the support of multiple head, the multiplication of two big matrixes is
      split into multiplication of several (head_number) small matrixes. e.g. if
      Mat A is [3, 24] and Mat B is [24, 4], when multiple A and B with head_number
      as 4, Mat A will be split as 4 matrix of [3, 6] and Mat B will be 4 matrix of
      [6, 4]. The result of final matrix will be 4 matrix of [3, 4], i.e. [3, 16].
      220eef60
  4. 04 3月, 2019 1 次提交
  5. 26 2月, 2019 1 次提交
  6. 22 2月, 2019 2 次提交
    • T
      Revert 15770 develop a6910f90 gelu mkl opt (#15872) · ee2321de
      tensor-tang 提交于
      * Revert "Optimze Gelu with MKL Erf function (#15770)"
      
      This reverts commit 676995c8.
      
      * test=develop
      ee2321de
    • Y
      Optimze Gelu with MKL Erf function (#15770) · 676995c8
      Yihua Xu 提交于
      * Optimize for gelu operator
      
      * Set up the low accuracy mode of MKL ERF function.
      
      test=develop
      
      * Only enable MKLML ERF when OS is linux
      
      * Use the speical mklml version included vmsErf function to verify gelu mkl kernel.
      
      test=develop
      
      * Add the CUDA macro to avoid NVCC's compile issue.
      
      test=develop
      
      * Add the TODO comments for mklml library modification.
      
      test=develop
      
      * Clean Code
      
      test=develop
      
      * Add the comment of marco for NVCC compiler.
      
      test=develop
      676995c8
  7. 13 12月, 2018 1 次提交
  8. 27 11月, 2018 1 次提交
  9. 16 11月, 2018 1 次提交
  10. 13 11月, 2018 1 次提交
  11. 22 8月, 2018 5 次提交
  12. 16 8月, 2018 1 次提交
  13. 06 8月, 2018 1 次提交
  14. 03 8月, 2018 2 次提交
  15. 05 7月, 2018 2 次提交
  16. 27 6月, 2018 1 次提交
  17. 20 6月, 2018 1 次提交
  18. 24 5月, 2018 1 次提交
  19. 21 5月, 2018 1 次提交
  20. 08 5月, 2018 2 次提交
  21. 07 5月, 2018 1 次提交
  22. 04 5月, 2018 1 次提交