1. 27 3月, 2023 1 次提交
    • S
      Fused elementwise_(mul/div) (#50428) · 968f7f24
      Sławomir Siwek 提交于
      * extract Op and OPMaker to .h
      
      * extend pattern for fused_op
      
      * set "with_residual" default to false
      
      * adjust fuse passes
      
      * remove fc+eltwise flag
      
      * fused_output_scale
      
      * activation attrs
      
      * remove extra attrs
      
      * fix int8/bf16 unit tests
      
      * simplify RecomputeOutputDims
      
      * remove unused method
      
      * Add description for attributes
      
      * add extra check
      
      * adjust op compats
      
      * update quantize test
      
      * fix protobuf parsing error
      
      * fix int8 performance
      
      * fused elementwises
      
      * merge develop
      
      * remove activation
      
      * restore activation for existing add/sub ops
      968f7f24
  2. 25 3月, 2023 1 次提交
  3. 24 3月, 2023 2 次提交
    • TaoTao Li's avatar
      add phi operator allreduce/reduce (#51857) · 47f87ad3
      TaoTao Li 提交于
      * add all_reduce, reduce kernel and api
      
      * fix all_reduce reduce ut
      
      fix reduce op maker conflict
      
      fix merge conflicts
      
      * fix conflicts, rename ReduceOp->ReduceBaseOp in reduce_ops
      
      rename allreduce op, to remove
      
      * fix code format
      
      fix comments
      
      * modify test_collective_reduce_api ut timeout
      
      * fix PR-CI-Build
      
      fix comments: format phi operator
      47f87ad3
    • Y
      [PHI Decoupling]Remove memory header (Part3) (#51288) · 3d78e759
      YuanRisheng 提交于
      * decouple memory copy
      
      * fix ci bugs
      
      * fix ci compile bugs
      
      * fix rocm compile
      
      * fix ci bugs
      
      * decouple memory
      
      * deal with conflict
      
      * fix xpu compile bugs
      
      * fix xpu bugs
      
      * deal with xpu bugs
      
      * fix cmake bugs
      
      * fix windows bugs
      
      * fix ci bugs
      
      * fix ci bugs
      
      * delete redundance code
      
      * add code for pybind
      
      * fix py3 bugs
      
      * fix ci bugs
      3d78e759
  4. 23 3月, 2023 5 次提交
  5. 22 3月, 2023 8 次提交
    • HappyHeavyRain's avatar
      Support optimizers operator to be generated (#51767) · 0b008e0c
      HappyHeavyRain 提交于
      * test_get_kernel
      
      * add invoke signature
      
      * change reduce_max
      
      * change frobenius_norm
      
      * reset reduce_max according to composite and change reduce_all
      
      * fix the bug when Scalar(*)
      
      * fix 'scalar when support_tensor'
      
      * change code according to review
      
      * change 'keep_signature' to 'manual_signature' and add some erro info
      
      * support optimizers autogen
      
      * change sgd yaml
      
      * change generate signature
      
      * fix test/cpp/new_executor/CM
      
      * reset signature generated function
      
      * change signature funciton
      
      * change signature funciton
      0b008e0c
    • S
      Extract fused_transpose op dedicated for oneDNN fuse passes (#50021) · 02296977
      Sławomir Siwek 提交于
      * extract common methods to reuse
      
      * add header for transpose ops
      
      * fused_transpose
      
      * Split big function
      
      * transpose2 tests
      
      * fused_transpose
      
      * Apply extra attributes
      
      * add pbtxt file
      
      * update pbtxt
      
      * Merge develop
      
      * add more strict op compats
      
      * code  style
      
      * remove mkldnn_data_type
      
      * unify SetOutMemDescWithReshape2FuseSupport
      
      * adjust quantize-dequantize for transpose
      
      * remove appendact
      
      * transpose2 quantization
      
      * fix int8 tests
      
      * adjust transpose_op to current develop
      
      * delete fusion code from transpose_kernel
      
      * add fused transpose to NHWC unittest
      
      * change order
      02296977
    • W
      add autogen code support for index_add op (#51887) · 3065fa2c
      Wang Xin 提交于
      * add autogen code for index_add op
      
      * bug fixed
      3065fa2c
    • R
      support auto generate for p_norm (#51590) · 2b98993b
      RedContritio 提交于
      * supoort auto generate p_norm
      
      * fix bug in backward
      2b98993b
    • R
      support auto generate for dirichlet (#51601) · ec877d1f
      RedContritio 提交于
      * support auto generate for dirichlet
      
      * use uppercase in args
      
      * use op_compat for name mapping
      ec877d1f
    • W
      Add reduce_max_grad composite rule (#51653) · d04c9cda
      wangxiaoning 提交于
      * max comp
      
      * fix
      
      * add test
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * fix test
      
      * fix api
      d04c9cda
    • S
      Add fused_linear_param_grad_add_kernel (#51805) · f59c5d8b
      sneaxiy 提交于
      * add fused_linear_param_grad_add_kernel
      
      * fix compile error
      
      * remove flag
      
      * fix ci compile error
      
      * fix ci compile error
      
      * revert pylayer revision
      
      * fix ci ut
      
      * improve performance
      f59c5d8b
    • R
      Fix conflict of CppTypeToDataType (#51919) · 535ddd3d
      Ruibiao Chen 提交于
      535ddd3d
  6. 21 3月, 2023 3 次提交
  7. 20 3月, 2023 3 次提交
    • X
      【prim】New layer_norm grad (#51750) · 802a81d0
      xiaoguoguo626807 提交于
      * Add flatten composite rule
      
      * get the right xshape and pass func test
      
      * add cinn unit test
      
      * Remove cinn test, wait for it to be added after repair
      
      * add comp test to test_flatten_contiguous_range_op.py
      
      * remove func test on composite_ops
      
      * Add comments to maybe_wrap_dim func
      
      * remove commented code
      
      * fix the problem with 0D tensor case
      
      * add flatten split rule comment
      
      * fix syntax issues
      
      * block flatten on resnet_prim_cinn
      
      * init change
      
      * tmp commit
      
      * add layer_norm InferMeta check
      
      * cast type modify
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * recover
      
      * big tol
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * Cxx prim custom vjp (#8)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * [dy2static-ci] fix dy2static ci errors.
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [Prim] enable whitelist and blacklist for custom_vjp
      
      * debug log
      
      * clear log
      
      * fix
      
      * nothing
      
      * less memory
      
      * recover utils
      
      * fix
      
      * modify threshold value
      
      * skip layer_norm for test_bert
      
      * back to bert success state
      
      * add epsion
      
      * delete unnecessary compute
      
      * modify amp dtype
      
      * modify * order
      
      * delete sqrt check and fp16
      
      ---------
      Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com>
      Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com>
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      Co-authored-by: Nxiongkun <807377414@qq.com>
      802a81d0
    • Z
      Add LoongArch support (#51109) · 5ed2320c
      Zhang Na 提交于
      5ed2320c
    • L
      Support Linear operation in cuBlaslt and plug into attn_gemm and fusedLinear forward op (#51124) · 2dfc3fa8
      limingshu 提交于
      * optimization for fused linear op
      
      * fix code format
      
      * optimization for linear fused forward
      
      * merge with develop
      
      * fix bugs for gemm_ephilog
      
      * package of cublaslt ephilogue type with enmu
      
      * final fix before code reviewing
      
      * fix missed fusedType typo
      
      * fix code according to review suggestions
      
      * fix windows ci error
      
      * change location of MatmulPlanner
      
      * add some changes for compiler error fix
      
      ---------
      2dfc3fa8
  8. 19 3月, 2023 1 次提交
  9. 18 3月, 2023 1 次提交
  10. 17 3月, 2023 2 次提交
  11. 16 3月, 2023 3 次提交
    • H
      Update from_blob API (#51646) · c07c7712
      Huang Jiyi 提交于
      * remove contexts in tensor_utils
      
      * update from_blob
      
      * update from_blob
      
      * update from_blob
      
      * fix bug
      
      * fix bug
      c07c7712
    • J
      [Auto Parallel Performance] Support BF16 Training (#51285) · 9ded5707
      JZ-LIANG 提交于
      * update env setting
      
      * update pass logic
      
      * dist op support bf16
      
      * backward cast update
      
      * update setting
      
      * update backward
      
      * revert amp pass
      
      * update fp16 backward logic
      
      * register c_embedding bf16
      
      * revert engine
      
      * add unitest
      
      * add unitest
      
      * update unitest
      
      * update cmake
      
      * update math
      
      * update math.py
      
      * update unitest
      
      * update unitest
      
      * revise unitest
      
      * revise unitest
      
      * update unitest
      
      * update unitest
      
      * update unitest
      9ded5707
    • S
      Fix nccl_test_op failure on hopper (#51390) · b5fd7fc1
      Shijie 提交于
      * add sync
      
      * Fix nccl_op_test
      b5fd7fc1
  12. 15 3月, 2023 5 次提交
    • S
      add assign composite backward op (#51430) · 297182f7
      SylarTiaNII 提交于
      * add assign composite backward op
      
      * fix log msg
      
      * code style
      
      * fix comp rule
      
      * replace assign with by_pass
      297182f7
    • J
      【Prim】Custom softmax grad (#51474) · f124c86f
      Jiabin Yang 提交于
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * Cxx prim custom vjp (#8)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * [dy2static-ci] fix dy2static ci errors.
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [Prim] enable whitelist and blacklist for custom_vjp
      
      * support softmax grad
      
      * remove additional code
      
      * add test back
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      Co-authored-by: Nxiongkun <807377414@qq.com>
      f124c86f
    • iSerendipity's avatar
      [PHI] remove operator.h in blas.h (rebase to latest codebase) (#51472) · 427712df
      iSerendipity 提交于
      * Revert "Revert "【Hackathon No.67】remove operator.h in blas.h (#50989)" (#51467)"
      
      This reverts commit b9d91531.
      
      * remove cout
      
      * add header
      
      * fix missing header
      
      * fix refer fluid error
      
      * fix missing header
      
      * 更新 repeat_interleave_grad_kernel_impl.h
      
      Change to phi style datatype.
      
      * 更新 repeat_interleave_grad_kernel_impl.h
      
      Fix missing header
      
      * datatype fluid -> phi
      
      * paddle::experimental -> phi
      
      * fix reference error
      
      * fix reference error
      
      * fix reference error
      
      * fix errors
      
      * fix missing FLAGS
      
      * fix missing headers
      
      * fix missing headers
      
      * fix missing headers
      
      * fix missing headers
      
      * fix missing header
      
      * fix missing header
      
      * fix errors
      427712df
    • R
      support auto generate for nonzero (#51600) · 3734e89a
      RedContritio 提交于
      3734e89a
    • HappyHeavyRain's avatar
      Move the "GetExpectedKernelType" into "get_expected_kernel_func.cc" (#51453) · f0db1f7e
      HappyHeavyRain 提交于
      * test_get_kernel
      
      * add invoke signature
      
      * change reduce_max
      
      * change frobenius_norm
      
      * reset reduce_max according to composite and change reduce_all
      
      * fix the bug when Scalar(*)
      
      * fix 'scalar when support_tensor'
      
      * change code according to review
      
      * change 'keep_signature' to 'manual_signature' and add some erro info
      f0db1f7e
  13. 14 3月, 2023 5 次提交