1. 14 8月, 2023 1 次提交
    • MarDino's avatar
      Add rmsnorm residual bias add and quant (#55965) · 2ac6a7e4
      MarDino 提交于
      * add rmsnorm residual bias add and quant
      
      * refine python interface
      
      * add rmsnorm unittest
      
      * Add layernorm
      
      * fix layernorm unittest
      
      * refine unittest
      
      * fix example code
      
      * fix review comment
      2ac6a7e4
  2. 27 4月, 2023 1 次提交
    • S
      Move fused feedforward (#53166) · 25b4ba7f
      Sonder 提交于
      * trans fused_feedward Compute function to phi
      
      * add register info
      
      * remove maxfunctor
      
      * move fused feedward to phi
      
      * remove sig file
      
      * remove fliud include
      
      * add include
      
      * add include
      
      * add sig file
      
      * add output register info
      
      * fix sig file
      
      * Update fused_feedforward_sig.cc
      
      * fix grad kernel
      
      * update output register info
      
      * fix
      
      * open fused_feedforward static build
      
      * add optional and fix code style
      
      * fix output info for fused attention
      
      * add optional param
      
      * merge
      25b4ba7f
  3. 13 4月, 2023 1 次提交
    • H
      [enforce.h Decouple logging.h] Delete glog/logging.h from enforce.h (#52651) · 5664ea26
      HongyuJia 提交于
      * [enforce.h Decouple logging.h] Delete glog/logging.h from enforce.h
      
      * Add logging.h for profiler.cc
      
      * Add logging.h for gloo_utils.h
      
      * Add logging.h for addmm_kernel_impl.h
      
      * Add logging.h for addmm_grad_kernel_impl.h
      
      * Add logging.h for p_send_kernel.cu
      
      * Add logging.h for determinant_grad_kernel_impl.h
      
      * Add logging.h for p_recv_kernel.cu
      
      * Add logging.h for elementwise_grad_base.h
      
      * Add logging.h for transfer_layout_kernel.cc
      
      * Add logging.h for eigvals_kernel.cc and index_select_impl.h
      
      * Add logging.h for all files in kernel directory
      
      * Add logging.h for xpu_info.cc
      
      * Add logging.h for xpu
      5664ea26
  4. 06 4月, 2023 1 次提交
    • S
      Move fused_attention op to phi [迁移前向 GPU OpKernel] (#51743) · a7ec8958
      Sonder 提交于
      * add kernel functions
      
      * update kernel functions
      
      * update func parameters' name
      
      * create codes for gpu device
      
      * 调整文件位置
      
      * fix include error
      
      * remove dependent files to phi/
      
      * restore fused_attention_op.cu
      
      * fix dependence errors
      
      * fix dependence errors
      
      * fix include error
      
      * fix all depandence errors[build success]
      
      * remove useless include
      
      * recover useless include
      
      * use phi::ToNCCLDataType
      
      * fix namespace
      
      * update new register code
      
      * fix error in fused_gemm_epilogue_utils
      
      * fix error in FusedAttentionKernel parm
      
      * finish fused_attention registe code[build success]
      
      * add paddle::optional
      
      * add sig file
      
      * fix build error
      
      * fix a include error
      
      * update CMkaeList
      
      * fix parameter sequence
      
      * add include file
      
      * update #if before include
      
      * fix grammly error
      
      * update codes for DropoutParam
      
      * remove const cast
      
      * trans some fluid api to phi api
      
      * add #if
      
      * update test code
      
      * update test codes
      
      * recover test codes
      
      * trans fused_attention to fluid
      
      * move #endif to end
      
      * move #endif
      
      * delete useless files
      
      * use fused attention utils and recover random seed
      
      * remove fluid include in phi
      a7ec8958
  5. 23 2月, 2023 1 次提交
  6. 16 2月, 2023 1 次提交
  7. 29 12月, 2022 1 次提交
  8. 20 12月, 2022 1 次提交
  9. 30 11月, 2022 1 次提交
  10. 28 11月, 2022 1 次提交
  11. 18 11月, 2022 1 次提交
    • MarDino's avatar
      Optimize FusedBiasAddGelu Kernel (#47679) · b0e28540
      MarDino 提交于
      * Add quick gelu and fused bias add kernel
      
      * fix annotation
      
      * remove useless code
      
      * add fast gelu option and set it in multi transformer op
      
      * add flag to restrict if use fast gelu approximate
      
      * fix flags conflict
      
      * fix use tanh function instead
      
      * add cudart version limit
      
      * use phi fast tanh func
      
      * fix comment
      b0e28540
  12. 31 10月, 2022 1 次提交
  13. 28 9月, 2022 1 次提交
    • C
      Remove the declaration of using Tensor in framework/tensor.h (#46432) · e12a905e
      Chen Weihang 提交于
      * remove needless using tensor
      
      * remove needless using tensor
      
      * resolve conflict
      
      * replace tensor using
      
      * fix format error
      
      * revert needless changing
      
      * fix rocm and npu compile error
      
      * fix cinn compile error
      
      * fix format error
      
      * fix mkldnn format error
      
      * fix mkldnn format error
      
      * fix cinn compile error
      
      * fix cinn compile error
      
      * fix cinn compile error
      
      * resolve conflict
      e12a905e
  14. 18 9月, 2022 1 次提交
  15. 01 8月, 2022 1 次提交
    • L
      unify gpu context (#44740) · 86763023
      Leo Chen 提交于
      * remove cudaDeviceContext
      
      * remove more template
      
      * fix rocm compile
      
      * remove alias name CUDADeviceContext
      
      * fix compile
      
      * fix tests
      
      * revert changes
      86763023
  16. 26 6月, 2022 1 次提交
  17. 17 6月, 2022 1 次提交
  18. 05 6月, 2022 1 次提交
  19. 31 5月, 2022 1 次提交
  20. 20 2月, 2022 1 次提交
  21. 15 2月, 2022 1 次提交
  22. 29 1月, 2022 1 次提交
    • L
      Optimize layer norm backward cuda kernel when cols is 1024. (#39247) · 99cfcc09
      Li Min 提交于
      * Add fp16 support for scale/bias for fused_layernnorm_residual_dropout_bias op.
      
      * Remove useless code.
      
      * Remove useless code.
      
      * Optimize layer_norm fwd when cols is 1024.
      
      * Remove useless code.
      
      * Minors.
      
      * Minors.
      
      * Modifications accordding to reviews.
      
      * Minors.
      
      * Optimize layer_norm bwd kernel when cols is 1024.
      
      * Polish layer_norm_bwd_1024 kernel.
      
      * Limit ln_bwd_1024_kernel to paddle_with_cuda.
      
      * Fix double type compile error.
      
      * Add optimization of ln bwd for fused_dropout_add_ln op.
      
      * Polish codes.
      99cfcc09
  23. 17 1月, 2022 1 次提交
    • W
      [Pten] Replace platform::Place to pten::Place. (#38899) · c48a9ad5
      Wilber 提交于
      * add pten::Place data structure.
      
      * update ci problem
      
      * fix ci problem
      
      * update
      
      * using platform::Place=pten::Place
      
      * remove BOOST_GET_CONST for CPUPlace and GPUPlace
      
      * compile pass 25%.
      
      * compile pass 45%
      
      * compile pass 60%
      
      * remove boost_get for xpu npu mlu and ipu
      
      * compile pass on cpu and gpu.
      
      * fix compile problem
      
      * fix compile error.
      
      * update
      
      * fix ci problem
      
      * update
      
      * ci approve
      
      * fix ci problem
      
      * fix ci eager test problem
      
      * remove BOOST_GET_CONST
      
      * fix npu compile
      c48a9ad5
  24. 07 1月, 2022 1 次提交
  25. 28 12月, 2021 1 次提交
  26. 22 10月, 2021 1 次提交
    • L
      Fused attention op forward (#35905) · d4906214
      Li Min 提交于
      功能:本PR的目标是提高attention模块的计算性能。
      为了减少框架层对op的调度开销,本PR通过在C++层手动实现attention模块,对外提供attention 大op;
      为了减少防存开销,本PR采取了两种优化方法:
      (1)在q,k,v计算时通过共享输入X,将该处的gemm,transpose和bias add从三次调用减少为一次;
      (2)使用kernel融合优化技术,在不同cuda kernel之间通过寄存器传输数据;
      d4906214
  27. 29 9月, 2021 1 次提交