1. 29 3月, 2022 1 次提交
  2. 28 3月, 2022 2 次提交
  3. 27 3月, 2022 1 次提交
    • J
      Add StringTensor (#39830) · 0695e1ac
      Jack Zhou 提交于
      * add string tensor and case convert kernels
      
      * Add strings empty kernel; Reorganize the structure of case convert kernel
      
      * Add string infermeta
      
      * Update mutable_data of string tensor
      
      * rename kernel name
      
      * add string copy tmp
      
      * Fix strings copy device bug
      
      * add utf8 gpu converter
      
      * add string tensor c++ api
      
      * Remove mutable_data of string tensor
      
      * update string tensor interface
      
      * remove charcases_flag.h
      
      * remove some fluid headers
      
      * Add make_ddim
      
      * __HIPCC__ -> PADDLE_WITH_HIP
      
      * remove fluid headers
      
      * fix cpu compile
      
      * remove std::hash
      
      * Fix cudaMalloc
      
      * Remove strings/impl directory
      
      * Fix infrt/get_phi_kernel_info.py;Add custom_kernels deps
      
      * Add empty kernel test
      
      * Remove some comments
      
      * Modify lower/upper api encoding type: string->bool
      
      * STRING->PSTRING; Add CreateInferLikeMeta
      
      * Add code gen for C++ String API
      
      * remove strings_api_utils.h
      
      * Add ignore file (strings_api.h, strings_api.cc)
      
      * update strings gen script
      
      * change args order of case convert kernels
      
      * Add comments for pstring, StringTensor
      
      * cpstring_internal.h -> cpstring_impl.h
      
      * Update accordding to comments:
      
      1. Remove fluid headers
      2. paddle::platform::errors -> phi::errors
      3. Use 'place.GetType() == phi::AllocationType::GPU' instead of 'paddle::platform::is_cpu_space()'
      4. Use camel code style
      
      * Remove all singletons in strings kernels
      
      * fix rocm compile
      
      * Fix py3 compile
      
      * Fix c++ coverage
      
      * 1. Add pstring proto type
      2. Add StringTensor debug info
      3. Rename case_convert_kernel to strings_lower_upper
      4. Remove serialize derialize strings kernel
      
      * DataLayout::PSTRING -> DataLayout::PSTRING_UNION
      
      * Register pstring data type
      
      * Fix strings api gen
      
      * Fix dense tensor register pstring dtype
      
      * Fix error messages
      
      * remove line
      
      * add pstring unittest
      
      * remove test string api unitest
      
      * remove empty line
      
      * Remove some headers to decrease the size of executable file
      0695e1ac
  4. 25 3月, 2022 3 次提交
  5. 24 3月, 2022 4 次提交
  6. 23 3月, 2022 2 次提交
  7. 21 3月, 2022 1 次提交
  8. 18 3月, 2022 4 次提交
  9. 17 3月, 2022 2 次提交
  10. 15 3月, 2022 2 次提交
    • C
      Skip infrt when checking log fatal (#40529) · c9f3ad03
      Chen Weihang 提交于
      * skip infrt when checking log fatal, test=document_fix
      
      * remove test=document_fix
      
      * update commit
      c9f3ad03
    • Y
      [Phi]Move Tanh/BRelu/LeakyRelu/ThresholdedRelu Kernels to Phi (#40385) · d7112180
      YuanRisheng 提交于
      * move activation op
      
      * adjust code format
      
      * fix compile bugs
      
      * fix ci bugs
      
      * code format adjust
      
      * code format adjust2
      
      * activate ci status
      
      * modify according to comment
      
      * move activation kernel
      
      * revert relu6
      
      * reduce add code
      
      * perfect use_phi_functor
      
      * completing func name
      
      * fix bugs when run ci
      
      * fix bugs when run infr
      
      * modifpy infrt get kernel signature
      d7112180
  11. 14 3月, 2022 1 次提交
  12. 10 3月, 2022 1 次提交
    • S
      Add trt execute (#40224) · e72ef603
      Shang Zhizhou 提交于
      * add trt.execute
      
      * merge trt.engine type
      
      * update return op
      
      * update comments
      
      * fix style
      
      * fix style
      e72ef603
  13. 09 3月, 2022 2 次提交
    • H
      [Infrt]Update kernel dialect (#40141) · 767647ce
      huzhiqiang 提交于
      767647ce
    • R
      build documents if public apis modified, meanwhile their samplecodes should be tested (#39728) · 041c4bca
      Ren Wei (任卫) 提交于
      * run document_preview when samplecodes be tested
      
      * run document_preview when samplecodes be tested
      
      * sphinx-build symbol link; and build-doc default
      
      * FLUIDDOCDIR typo
      
      * download the required configirations and some other scripts
      
      * install required python packages.
      
      * clone specified branch of docs repo, and if failed, clone the default branch
      
      * clean workspace for docs repo
      
      * use the conf.py imported by https://github.com/PaddlePaddle/docs/pull/4222/
      
      * download and install the boscmd
      
      * Optimaze the code comments.
      
      * specify the pypi index server
      
      * only do doc-build when running in cpu mode
      
      * pull docs pr
      
      git log
      
      paddle_pr_info
      
      * install jq
      
      * force using sphinx-build under py3.7
      
      * using our new domain name for preview
      
      * install python package error
      
      * don't build doc default
      041c4bca
  14. 07 3月, 2022 2 次提交
    • b798fb07
    • M
      cuBlasLt Epilogue To Fuse Linear + ReLU|GeLU (#39437) · 2a3d9eca
      Ming-Xu Huang 提交于
      * Added cuBlasLtHandle_t to device context.
      
      * Added fused_gemm_epilogue op.
      
      1. Added fused_gemm_epilogue op to leverage cuBlastLt Epilogue.
      2. Support fusion Act(X*Y + bias), X'dims >=2 and Y'dims shoule be 2.
      2. Act currently only be supported ReLU. (Will add GeLU in the future).
      
      * Added UT to fused_gemm_epilogue op.
      
      * Added LinearAct Pattern
      
      1. Added LinearAct into graph_pattern_detector.* to define (2.)'s
      pattern.
      2. LinearAct is used to detect act(element_add(matmul_v2(x, w), bias)).
      3. act currently only support ReLU (Will support GeLU in the future).
      
      * Added FuseGemmEpiloguePass
      
      1, Added FuseGemmEpiloguePass to handle nn.Linear + Act{ReLU}
      fusion (GeLU will be supported in the future).
      2. Only support matmul_v2 from nn.Linear.
      
      * Added pybind to BuildStrageter.fuse_gemm_epilogue_.
      
      * Added UT for fuse_gemm_epilogue_pass.
      
      * GeLU support and EpilogueSingleton
      
      1. Added GeLU support to fused_gemm_epilogue op.
      2. Added EpilogueSingleton to cache auxiliary pointer.
      3. Added related UTs.
      
      * Rename cublaslt_epilogue_opto gemm_epilogue_op.*.
      
      * Added both train and infer pattern to LinearAct.
      
      1. Added support of fwd graph with grap_ops linking to LinearAct.
      2. Added related changes to fuse_gemm_epilogue_pass for above
      modification.
      
      * Changed CUDA requirement from 11.4 to 11.6 for fuse_gemm_epilogue_pass.
      
      * Added identity activation support to gemm_epilogue_op.
      
      * Added Linear Fusion (matmul_v2 + ele_add)
      
      1. Added matmul_v2 + ele_add pattern to LinearActPattern.
      2. Added matmul_v2 + ele_add support to fuse_gemm_epilogue_pass.
      
      * Rename gemm_epilogue_op.* to fused_gemm_epilogue_op.*
      
      * Add fused_gemm_epilogue_grad op.
      
      1. Added fused_gemm_epilogue_grad to support backward epilogue fusion.
      
      * Add UTs to fused_gemm_epilogue_grad_op.
      
      * Change attribute name in fused_gemm_epilogue_grad_op for clearing.
      
      * Allow DX and DBias be dispensable to fused_gemm_epilogue_grad op.
      
      * Added ElementwiseAdd+Matmul+Act graph pattern detection.
      
      * Fuse backward of Linear( Act(x))
      
      1. Added backward fusion pass to Linear( Act(x)).
      2. Added backward fusion pass to Linear(x).
      
      * Added UTs to backward fusion of Linear(Act(x)).
      
      * Complete document of arguments to fused_gemm_epilogue_op.
      
      * Made arguments of some functions pass by reference.
      
      * Modify code with review comments.
      
      1. Made arguments of some function pass by reference.
      2. Removed redundant code.
      3. Followed Google code style to change code.
      
      * Made 'const' code style be consistent
      
      * Fixed random seed of python UTs.
      
      * Set Compiling constrains to cuBlasLt
      
      1. Require CUDA 11.6+
      2. Remove fuse_gemm_epilogue related tests when CUDA < 11.6.
      
      * Code Reivew from Paddle
      
      1. Changed arguments name is_first_gemm to without_x_gradient for
      clearing.
      2. Applied PADDLE_THROW in fused_gemm_epilogue_op.
      
      * Remove EpilogueSingleton
      
      1. Applied ReserveSpace to replace Epilogue for passing auxiliary
      pointers between FWD and BWD.
      
      * Fix a logical error and enhance UTs.
      
      1. Added act op count checking in UTs.
      2. Fix issue to fuse backward or ReLU(Linear(X)).
      3. TODO: solve GELU fusion issues.
      
      * Fix Linear and GeLU fusion issues.
      
      1. Modified graph_detech_pattern to fit with both linear wiht gelu or
      relu.
      2. Modified data range in Uts to allow negative values.
      
      * Removed fused_gemm_epilogue_op.h.
      
      * Rename namespace pten to phi.
      
      * Rename name of arguments in fused_gemm_epilogue_op
      
      1. bias -> Bias.
      2. out -> Out.
      3. reserve_space -> ReserveSpace.
      
      * Change EpiloguePassActivationCache as local variable.
      
      1. Removed singleton in EpiloguePassActivationCache.
      2. Made EpiloguePassActivationCache as an argument to each pass
      functions.
      2a3d9eca
  15. 04 3月, 2022 1 次提交
  16. 03 3月, 2022 1 次提交
  17. 02 3月, 2022 3 次提交
  18. 01 3月, 2022 2 次提交
  19. 28 2月, 2022 2 次提交
  20. 22 2月, 2022 3 次提交