1. 07 4月, 2023 1 次提交
  2. 06 4月, 2023 1 次提交
  3. 03 4月, 2023 1 次提交
  4. 24 3月, 2023 1 次提交
    • Y
      [PHI Decoupling]Remove memory header (Part3) (#51288) · 3d78e759
      YuanRisheng 提交于
      * decouple memory copy
      
      * fix ci bugs
      
      * fix ci compile bugs
      
      * fix rocm compile
      
      * fix ci bugs
      
      * decouple memory
      
      * deal with conflict
      
      * fix xpu compile bugs
      
      * fix xpu bugs
      
      * deal with xpu bugs
      
      * fix cmake bugs
      
      * fix windows bugs
      
      * fix ci bugs
      
      * fix ci bugs
      
      * delete redundance code
      
      * add code for pybind
      
      * fix py3 bugs
      
      * fix ci bugs
      3d78e759
  5. 13 3月, 2023 1 次提交
  6. 06 3月, 2023 1 次提交
    • H
      [phi decoupling] decouple dependency to device_context in phi (Part 1) (#50865) · a1006b2b
      Huang Jiyi 提交于
      * move DeviceContextPool to phi
      
      * add EmplaceExternalContextFunc
      
      * update namespace
      
      * update cmake
      
      * fix bugs and create context_pool_impl.h
      
      * replace platform::is_xxx_place
      
      * fix bugs
      
      * update generator
      
      * fix bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix enforce usage
      
      * Revert "fix enforce usage"
      
      This reverts commit 5f521f08a69713cee506e64a00ec6d9fba709e27.
      
      * fix bugs
      
      * rm XPUDeviceContext and CustomDeviceContext
      
      * fix bugs
      
      * fix fix context init bug
      
      * fix bugs after merge
      
      * fix bugs
      
      * fix name
      
      * fix mutable_data
      
      * update and fix bugs
      
      * fix bugs
      
      * update
      
      * fix bugs
      
      * fix name
      
      * fix bugs
      
      * merge
      
      * fix bugs
      
      * create context_pool in phi/backends
      
      * create context_pool in phi/backends
      
      * fix bugs
      
      * fix xpu bugs
      
      * fix rocm bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix bugs
      
      * fix xpu bugs
      
      * update
      
      * update
      
      * fix bugs
      
      * fix bugs
      a1006b2b
  7. 23 2月, 2023 1 次提交
  8. 30 1月, 2023 1 次提交
  9. 23 12月, 2022 1 次提交
  10. 05 12月, 2022 1 次提交
  11. 29 11月, 2022 1 次提交
  12. 18 11月, 2022 1 次提交
  13. 01 11月, 2022 1 次提交
    • C
      Adapting device-specific Extra Attributes for the PHI kernel (#46342) · c923e6c9
      Chen Weihang 提交于
      * add extra attr property set
      
      * add type_info for all context
      
      * add onednn context to all context
      
      * fix context compile error
      
      * simplify conv kernel args
      
      * pass runtime attr into dev_ctx
      
      * fix marco error
      
      * clear conv_grad_kernel extra args
      
      * merge conv_grad_grad into conv_grad
      
      * clear conv2d_grad_grad extra attrs
      
      * clear yaml and eager extra attr
      
      * fix conv1d error
      
      * change to thread local
      
      * fix npu compile failed
      
      * try to fix windows compile failed
      
      * add conv2d onednn phi kernel
      
      * fix ci bugs (#36)
      
      * fix compile bugs (#38)
      
      * fix extra input transform bug (#39)
      
      * support dynamic created attr (#40)
      
      * reset extra info gen code
      
      * rm conv_grad_grad kernel
      
      * reimpl pass attr adapting
      
      * add int attr support
      
      * remove vector inputnames creating
      
      * fix map at error
      
      * Update paddle/phi/kernels/onednn/conv_grad_kernel.cc
      Co-authored-by: NSławomir Siwek <slawomir.siwek@intel.com>
      
      * remove useless extra attrs
      
      * replace mkldnn_engine by onednn_engine
      Co-authored-by: NYuanRisheng <yuanrisheng@baidu.com>
      Co-authored-by: NSławomir Siwek <slawomir.siwek@intel.com>
      c923e6c9
  14. 25 9月, 2022 1 次提交
  15. 09 9月, 2022 1 次提交
  16. 01 9月, 2022 1 次提交
  17. 03 8月, 2022 1 次提交
  18. 01 8月, 2022 1 次提交
    • L
      unify gpu context (#44740) · 86763023
      Leo Chen 提交于
      * remove cudaDeviceContext
      
      * remove more template
      
      * fix rocm compile
      
      * remove alias name CUDADeviceContext
      
      * fix compile
      
      * fix tests
      
      * revert changes
      86763023
  19. 29 7月, 2022 2 次提交
  20. 19 7月, 2022 1 次提交
  21. 14 7月, 2022 1 次提交
    • Y
      [Phi]Improve the mechanism for mkldnn kernel in PHI (#43941) · e9b4d0be
      YuanRisheng 提交于
      * adapt mkldnn kernel in PHI
      
      * fix ci compile bugs
      
      * fix compile bugs
      
      * fix compile bugs
      
      * fix compile bugs
      
      * fix compile bugs
      
      * delete comment
      
      * fix compile bugs in windows-inference
      
      * delete code for converage
      
      * modify code by review
      
      * modify code by review
      
      * add todo
      
      * fix compile bugs
      
      * fix compile bugs
      
      * fix compile bugs
      
      * fix unittest bugsx
      e9b4d0be
  22. 11 7月, 2022 1 次提交
  23. 07 7月, 2022 1 次提交
    • A
      [IPU] support dy2static for IPU merge code (#43770) · 6984fbca
      Allen Guo 提交于
      * feat(): dynamic_to_static support for ipu.
      
      * fix(): format fix.
      
      * fix format
      
      * fix cpplint error
      
      * use phi::errors
      
      * fix format
      
      * fix format
      
      * fix(): add api to restore patched function.
      
      * fix(): identity_loss uses cpu place as expected kernel type.
      
      * doc(): add IPU dy2static related docs.
      
      * fix(): combine test cases.
      
      * fix format
      
      * fix comment
      
      * fix format
      
      * apply comment
      
      * fix compiling
      
      * fix(): align docs.
      
      * fix(): fix identity_loss function docs.
      
      * fix(): adjust mean and sum in identity_loss.
      
      * fix(): minor docs.
      
      * move API to paddle.incubate.identity_loss
      
      * fix UT
      Co-authored-by: Nzhaorui chen <zhaoruic@graphcore.ai>
      6984fbca
  24. 02 7月, 2022 2 次提交
    • L
      unify cpu context, part2 (#44012) · 755438a7
      Leo Chen 提交于
      * fix init()
      
      * delete test_device_context
      
      * replace CPUDeviceContext with CPUContext
      
      * fix test_scalar
      
      * remove dot_op.cc
      
      * fix compile
      755438a7
    • L
      unify cpu context (#43989) · 09096aeb
      Leo Chen 提交于
      * unify cpu context
      
      * fix init()
      
      * delete test_device_context
      
      * fix test_scalar
      09096aeb
  25. 26 6月, 2022 1 次提交
  26. 10 6月, 2022 1 次提交
  27. 08 6月, 2022 1 次提交
  28. 07 6月, 2022 1 次提交
  29. 05 6月, 2022 1 次提交
  30. 23 5月, 2022 2 次提交
  31. 09 4月, 2022 1 次提交
  32. 01 4月, 2022 1 次提交
    • W
      [Eager] Support pinned (#41035) · f3270fc8
      wanghuancoder 提交于
      * support pinned, test=develop
      
      * support async_write, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine,test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      
      * refine, test=develop
      f3270fc8
  33. 31 3月, 2022 1 次提交
    • L
      [new-exec] fit mkldnn op (#41058) · 02cf6764
      Leo Chen 提交于
      * fix bug that some op has no op_role attr
      
      * add mkldnn support for new executor
      
      * fit for mkldnn data_transfer
      
      * fit for mkldnn data_transfer
      02cf6764
  34. 27 3月, 2022 1 次提交
  35. 23 3月, 2022 1 次提交
  36. 07 3月, 2022 1 次提交
    • M
      cuBlasLt Epilogue To Fuse Linear + ReLU|GeLU (#39437) · 2a3d9eca
      Ming-Xu Huang 提交于
      * Added cuBlasLtHandle_t to device context.
      
      * Added fused_gemm_epilogue op.
      
      1. Added fused_gemm_epilogue op to leverage cuBlastLt Epilogue.
      2. Support fusion Act(X*Y + bias), X'dims >=2 and Y'dims shoule be 2.
      2. Act currently only be supported ReLU. (Will add GeLU in the future).
      
      * Added UT to fused_gemm_epilogue op.
      
      * Added LinearAct Pattern
      
      1. Added LinearAct into graph_pattern_detector.* to define (2.)'s
      pattern.
      2. LinearAct is used to detect act(element_add(matmul_v2(x, w), bias)).
      3. act currently only support ReLU (Will support GeLU in the future).
      
      * Added FuseGemmEpiloguePass
      
      1, Added FuseGemmEpiloguePass to handle nn.Linear + Act{ReLU}
      fusion (GeLU will be supported in the future).
      2. Only support matmul_v2 from nn.Linear.
      
      * Added pybind to BuildStrageter.fuse_gemm_epilogue_.
      
      * Added UT for fuse_gemm_epilogue_pass.
      
      * GeLU support and EpilogueSingleton
      
      1. Added GeLU support to fused_gemm_epilogue op.
      2. Added EpilogueSingleton to cache auxiliary pointer.
      3. Added related UTs.
      
      * Rename cublaslt_epilogue_opto gemm_epilogue_op.*.
      
      * Added both train and infer pattern to LinearAct.
      
      1. Added support of fwd graph with grap_ops linking to LinearAct.
      2. Added related changes to fuse_gemm_epilogue_pass for above
      modification.
      
      * Changed CUDA requirement from 11.4 to 11.6 for fuse_gemm_epilogue_pass.
      
      * Added identity activation support to gemm_epilogue_op.
      
      * Added Linear Fusion (matmul_v2 + ele_add)
      
      1. Added matmul_v2 + ele_add pattern to LinearActPattern.
      2. Added matmul_v2 + ele_add support to fuse_gemm_epilogue_pass.
      
      * Rename gemm_epilogue_op.* to fused_gemm_epilogue_op.*
      
      * Add fused_gemm_epilogue_grad op.
      
      1. Added fused_gemm_epilogue_grad to support backward epilogue fusion.
      
      * Add UTs to fused_gemm_epilogue_grad_op.
      
      * Change attribute name in fused_gemm_epilogue_grad_op for clearing.
      
      * Allow DX and DBias be dispensable to fused_gemm_epilogue_grad op.
      
      * Added ElementwiseAdd+Matmul+Act graph pattern detection.
      
      * Fuse backward of Linear( Act(x))
      
      1. Added backward fusion pass to Linear( Act(x)).
      2. Added backward fusion pass to Linear(x).
      
      * Added UTs to backward fusion of Linear(Act(x)).
      
      * Complete document of arguments to fused_gemm_epilogue_op.
      
      * Made arguments of some functions pass by reference.
      
      * Modify code with review comments.
      
      1. Made arguments of some function pass by reference.
      2. Removed redundant code.
      3. Followed Google code style to change code.
      
      * Made 'const' code style be consistent
      
      * Fixed random seed of python UTs.
      
      * Set Compiling constrains to cuBlasLt
      
      1. Require CUDA 11.6+
      2. Remove fuse_gemm_epilogue related tests when CUDA < 11.6.
      
      * Code Reivew from Paddle
      
      1. Changed arguments name is_first_gemm to without_x_gradient for
      clearing.
      2. Applied PADDLE_THROW in fused_gemm_epilogue_op.
      
      * Remove EpilogueSingleton
      
      1. Applied ReserveSpace to replace Epilogue for passing auxiliary
      pointers between FWD and BWD.
      
      * Fix a logical error and enhance UTs.
      
      1. Added act op count checking in UTs.
      2. Fix issue to fuse backward or ReLU(Linear(X)).
      3. TODO: solve GELU fusion issues.
      
      * Fix Linear and GeLU fusion issues.
      
      1. Modified graph_detech_pattern to fit with both linear wiht gelu or
      relu.
      2. Modified data range in Uts to allow negative values.
      
      * Removed fused_gemm_epilogue_op.h.
      
      * Rename namespace pten to phi.
      
      * Rename name of arguments in fused_gemm_epilogue_op
      
      1. bias -> Bias.
      2. out -> Out.
      3. reserve_space -> ReserveSpace.
      
      * Change EpiloguePassActivationCache as local variable.
      
      1. Removed singleton in EpiloguePassActivationCache.
      2. Made EpiloguePassActivationCache as an argument to each pass
      functions.
      2a3d9eca
  37. 03 3月, 2022 1 次提交