1. 09 8月, 2023 1 次提交
  2. 08 8月, 2023 1 次提交
  3. 07 8月, 2023 1 次提交
    • N
      cherry-pick fused_rope from develop (#55931) · 8d3a9882
      niuliling123 提交于
      * Add fused_rope forward op (#54351)
      
      * style
      
      * more
      
      * update ctest
      
      * Update legacy_backward.yaml
      
      * Update legacy_ops.yaml
      
      * Update legacy_ops.yaml
      
      * update
      
      * update
      
      * update for move
      
      * Update the rope op according to the comments (#54985)
      
      * Update multiary.cc
      
      * Update __init__.py
      
      * for int64_t and assert
      
      * more
      
      * remove useless assert first
      
      ---------
      Co-authored-by: Nsneaxiy <sneaxiy@126.com>
      8d3a9882
  4. 18 7月, 2023 1 次提交
  5. 13 7月, 2023 1 次提交
  6. 14 6月, 2023 1 次提交
    • P
      support sharding stage1 (#54069) · 974676bc
      pangengzheng 提交于
      * support sharding stage1
      
      * fix unittest
      
      * format
      
      * pass sharded sharding params_and_grads to inner_opt apply_pptimize
      
      * change sharding gradient allreduce to reduce
      
      * support save state_dict adptively and support sharding with mp
      
      * fix sharding test
      
      * test set_state_dict
      
      * add more unit test
      
      * fix global norm of mp case
      
      * polish
      
      * hack to calculate global norm in order to remove diff in calculating global norm values in HybridParallelClipGrad compared to dp
      
      * remove print
      974676bc
  7. 29 5月, 2023 1 次提交
  8. 13 4月, 2023 3 次提交
    • H
      [enforce.h Decouple logging.h] Delete glog/logging.h from enforce.h (#52651) · 5664ea26
      HongyuJia 提交于
      * [enforce.h Decouple logging.h] Delete glog/logging.h from enforce.h
      
      * Add logging.h for profiler.cc
      
      * Add logging.h for gloo_utils.h
      
      * Add logging.h for addmm_kernel_impl.h
      
      * Add logging.h for addmm_grad_kernel_impl.h
      
      * Add logging.h for p_send_kernel.cu
      
      * Add logging.h for determinant_grad_kernel_impl.h
      
      * Add logging.h for p_recv_kernel.cu
      
      * Add logging.h for elementwise_grad_base.h
      
      * Add logging.h for transfer_layout_kernel.cc
      
      * Add logging.h for eigvals_kernel.cc and index_select_impl.h
      
      * Add logging.h for all files in kernel directory
      
      * Add logging.h for xpu_info.cc
      
      * Add logging.h for xpu
      5664ea26
    • Z
      rename_bilinear_tensor_op (#52745) · eb93b5c9
      zhangyuqin1998 提交于
      eb93b5c9
    • L
      Fix bug in cross_entropy in static mode (#52771) · 16c36465
      lzydev 提交于
      * fix bug in cross_entropy in static mode
      
      * fix ci-coverage
      16c36465
  9. 11 4月, 2023 4 次提交
  10. 10 4月, 2023 2 次提交
  11. 09 4月, 2023 1 次提交
  12. 07 4月, 2023 1 次提交
  13. 06 4月, 2023 1 次提交
  14. 04 4月, 2023 2 次提交
  15. 03 4月, 2023 1 次提交
  16. 30 3月, 2023 3 次提交
  17. 29 3月, 2023 1 次提交
    • Y
      Add group_norm composite rule (#51874) · cabf3921
      Yichen Zhang 提交于
      * add group_norm composite rule
      
      * add test for scale_grad and bias_grad
      
      * resolve conflicts
      
      * remove amp in composite_rule.py
      
      * add float16 test
      
      * deal with NHWC format
      
      * keep the composite rule in float16 identical as original kernel
      
      * resolve conflicts
      cabf3921
  18. 27 3月, 2023 2 次提交
  19. 24 3月, 2023 3 次提交
    • TaoTao Li's avatar
      add phi operator allreduce/reduce (#51857) · 47f87ad3
      TaoTao Li 提交于
      * add all_reduce, reduce kernel and api
      
      * fix all_reduce reduce ut
      
      fix reduce op maker conflict
      
      fix merge conflicts
      
      * fix conflicts, rename ReduceOp->ReduceBaseOp in reduce_ops
      
      rename allreduce op, to remove
      
      * fix code format
      
      fix comments
      
      * modify test_collective_reduce_api ut timeout
      
      * fix PR-CI-Build
      
      fix comments: format phi operator
      47f87ad3
    • P
      [PHI]fix momentum dtype infer (#51353) · 648ec795
      PuQing 提交于
      * fix momentum dtype infer
      
      * fix momentum datatype
      
      * fix on cpu
      
      * add momentum
      648ec795
    • Z
      Memory Efficient Attention (#51867) · e5ad3859
      ZhangDY-6483 提交于
      * first version, notest
      
      * return final rst, notest
      
      * use infinity() instead of max
      
      * ut structure
      
      * start up of ut
      
      * generate lse
      
      * update
      
      * add depense
      
      * reconstruct cmake
      
      * move file
      
      * add memory efficient attention and fix blasimpl
      
      * update
      
      * update cmake
      
      * add namespace
      
      * update cmake
      
      * use .cu
      
      * update for pad3d
      
      * bug fix
      
      * bug fix
      
      * update
      
      * bug fix
      
      * update enforce
      
      * add test case
      
      * merge the lse pad
      
      * fix kernel_fn of backward
      
      * fix PADDLE_ENFORCE_EQ and phi_api
      
      * fix PADDLE_ENFORCE
      
      * fix PADDLE_ENFORCE
      
      * rerun coverage
      
      * fix memory efficient attention test
      
      * rerun ci
      
      * add cuda version condition
      
      * add cuda version condition
      
      * delete WIP test
      
      * replace PADDLE_ENFORCE
      
      * edit the namespace of datatype in multiple.cc
      
      * rerun
      
      * rerun
      
      ---------
      Co-authored-by: Nliuyuang <liuyuang@baidu.com>
      e5ad3859
  20. 23 3月, 2023 1 次提交
    • C
      [Prim] add meshgrid composite rule (#51061) · 53bb883d
      chenjian 提交于
      * add meshgrid composite rule
      
      * add meshgrid composite rule
      
      * update
      
      * add into CMakeLists
      
      * fix
      
      * update
      
      * update
      
      * optimize code
      
      * fix meshgrid op
      
      * update test
      53bb883d
  21. 22 3月, 2023 2 次提交
  22. 21 3月, 2023 3 次提交
  23. 20 3月, 2023 3 次提交
    • X
      【prim】New layer_norm grad (#51750) · 802a81d0
      xiaoguoguo626807 提交于
      * Add flatten composite rule
      
      * get the right xshape and pass func test
      
      * add cinn unit test
      
      * Remove cinn test, wait for it to be added after repair
      
      * add comp test to test_flatten_contiguous_range_op.py
      
      * remove func test on composite_ops
      
      * Add comments to maybe_wrap_dim func
      
      * remove commented code
      
      * fix the problem with 0D tensor case
      
      * add flatten split rule comment
      
      * fix syntax issues
      
      * block flatten on resnet_prim_cinn
      
      * init change
      
      * tmp commit
      
      * add layer_norm InferMeta check
      
      * cast type modify
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * recover
      
      * big tol
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * Cxx prim custom vjp (#8)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * [dy2static-ci] fix dy2static ci errors.
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [Prim] enable whitelist and blacklist for custom_vjp
      
      * debug log
      
      * clear log
      
      * fix
      
      * nothing
      
      * less memory
      
      * recover utils
      
      * fix
      
      * modify threshold value
      
      * skip layer_norm for test_bert
      
      * back to bert success state
      
      * add epsion
      
      * delete unnecessary compute
      
      * modify amp dtype
      
      * modify * order
      
      * delete sqrt check and fp16
      
      ---------
      Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com>
      Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com>
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      Co-authored-by: Nxiongkun <807377414@qq.com>
      802a81d0
    • M
      52e1742f
    • zhouweiwei2014's avatar
      Fix unsqueeze with empty axis bug (#51828) · 7a79fd88
      zhouweiwei2014 提交于
      7a79fd88