1. 29 3月, 2023 1 次提交
  2. 24 3月, 2023 1 次提交
    • Z
      Memory Efficient Attention (#51867) · e5ad3859
      ZhangDY-6483 提交于
      * first version, notest
      
      * return final rst, notest
      
      * use infinity() instead of max
      
      * ut structure
      
      * start up of ut
      
      * generate lse
      
      * update
      
      * add depense
      
      * reconstruct cmake
      
      * move file
      
      * add memory efficient attention and fix blasimpl
      
      * update
      
      * update cmake
      
      * add namespace
      
      * update cmake
      
      * use .cu
      
      * update for pad3d
      
      * bug fix
      
      * bug fix
      
      * update
      
      * bug fix
      
      * update enforce
      
      * add test case
      
      * merge the lse pad
      
      * fix kernel_fn of backward
      
      * fix PADDLE_ENFORCE_EQ and phi_api
      
      * fix PADDLE_ENFORCE
      
      * fix PADDLE_ENFORCE
      
      * rerun coverage
      
      * fix memory efficient attention test
      
      * rerun ci
      
      * add cuda version condition
      
      * add cuda version condition
      
      * delete WIP test
      
      * replace PADDLE_ENFORCE
      
      * edit the namespace of datatype in multiple.cc
      
      * rerun
      
      * rerun
      
      ---------
      Co-authored-by: Nliuyuang <liuyuang@baidu.com>
      e5ad3859
  3. 08 3月, 2023 1 次提交
  4. 10 1月, 2023 1 次提交
  5. 27 12月, 2022 1 次提交
  6. 08 11月, 2022 1 次提交
  7. 14 6月, 2022 1 次提交
  8. 04 6月, 2022 1 次提交
  9. 21 4月, 2022 1 次提交
  10. 12 4月, 2022 1 次提交
  11. 22 3月, 2022 1 次提交
  12. 02 3月, 2022 1 次提交
  13. 11 2月, 2022 1 次提交
  14. 31 8月, 2021 1 次提交
    • Z
      New whl release strategy with pruned nv_fatbin (#35239) · 2f3b393d
      Zhanlue Yang 提交于
      [Background]
      Expansion in code size can be irreversible in the long run, leading to huge release packages which
      not only hampers user experience but also exceeds a hard limit of pypi.
      
      In such, NV_FATBIN section takes up 86% of the compiled dylib size, owing to the vast number of GPU
      arches supported.
      
      This PR aims to prune this NV_FATBIN.
      
      [Solution]
      In the new release strategy, two types of whl packages will be involved:
      
      Cubin PIP package:
      PIP package maintains a smaller window for GPU arches support, containing
      sm_60, sm_70, sm_75, sm_80 cubins, covering Pascal - Ampere arches
      
      JIT release package:
      This is a backup for Cubin PIP package, containing compute_35, compute_50, compute_60,
      compute_70, compute_75, compute_80, with best performance and GPU arches coverage.
      
      However, it takes around 10 min to install due to the JIT compilation.
      
      [How to use]
      The new release strategy is disabled by default.
      To compile for Cubin PIP package, add this to cmake: -DCUBIN_RELEASE_PIP
      To compile for JIT release package, add this to cmake: -DJIT_RELEASE_WHL
      2f3b393d
  15. 14 7月, 2021 1 次提交
  16. 06 7月, 2021 1 次提交
    • Z
      Add gpu implementation of shuffle_batch_op (#33938) · c6b6ba1f
      Zeng Jinle 提交于
      * add gpu implementation of shuffle batch
      test=develop
      
      * add thrust cuda patches
      test=develop
      
      * fix macro guard
      
      * fix shuffle batch compile on windows/hip
      
      * fix hip compilation error
      
      * refine CMakeLists.txt
      
      * fix windows compile error
      
      * try to fix windows CI compilation error
      
      * fix windows compilation again
      
      * fix shuffle_batch op test on Windows
      c6b6ba1f
  17. 02 6月, 2021 1 次提交
  18. 26 5月, 2021 1 次提交
  19. 31 3月, 2021 2 次提交
  20. 30 3月, 2021 1 次提交
  21. 17 3月, 2021 1 次提交
  22. 19 2月, 2021 1 次提交
  23. 14 1月, 2021 1 次提交
  24. 27 11月, 2020 1 次提交
    • S
      detect tensorRT plugin fp16 in runtime (#27933) · b9e76a01
      Shang Zhizhou 提交于
      * remove -DSUPPORTS_CUDA_FP16 in cuda.cmake
      
      * comile with cuda9
      
      * add some unittest
      
      * notest;test=coverage
      
      * add unittest for trt plugin swish && split
      
      * update ernie unittest
      
      * fix some error message
      
      * remove repeated judgement of CUDA version in mbEltwiseLayerNormOpConverter
      
      * fix comile errror when CUDA_ARCH_NAME < Pascal"
      
      * fix comile error
      
      * update unittest timeout
      
      * compile with cuda9
      
      * update error msg
      
      * fix code style
      
      * add some comments
      
      * add define IF_CUDA_ARCH_SUPPORT_FP16
      
      * rename IF_CUDA_ARCH_SUPPORT_FP16 to CUDA_ARCH_FP16_SUPPORTED
      b9e76a01
  25. 21 10月, 2020 2 次提交
  26. 18 9月, 2020 1 次提交
  27. 09 9月, 2020 1 次提交
  28. 07 9月, 2020 1 次提交
  29. 20 8月, 2020 1 次提交
  30. 10 8月, 2020 1 次提交
  31. 09 7月, 2020 1 次提交
  32. 16 6月, 2020 1 次提交
  33. 10 6月, 2020 1 次提交
  34. 08 6月, 2020 1 次提交
    • T
      add -DPADDLE_CUDA_BINVER (#24928) · 90d420b1
      T8T9 提交于
      * add -DPADDLE_CUDA_BINVER. test=develop, test=win_gpu
      
      * nvcc will use add_compile_options, avoid using it if you don't want to pass arguments to nvcc. test=develop
      
      * test=develop, test=win_gpu
      90d420b1
  35. 05 6月, 2020 1 次提交
    • T
      Builtin cuda (#24904) · 211ef78c
      T8T9 提交于
      * support CUDA using cmake built-in way (#24395)
      
      * support CUDA using cmake built-in way. test=develop
      
      * test=develop
      
      * cmake_minimum_required 3.10
      
      * test=develop
      211ef78c
  36. 28 5月, 2020 1 次提交
  37. 13 5月, 2020 1 次提交
  38. 12 5月, 2020 1 次提交