1. 11 4月, 2023 1 次提交
    • G
      remove -Wimplicit-fallthrough (#52717) · 74542577
      Galaxy1458 提交于
      * delete [-Wno-error=terminate], test=develop
      
      * remove GPUps[-Wterminate],test=develop
      
      * remove some -Wno-, test=develop
      
      * modify ~MatmulDescriptor
      
      * mess
      
      * remove -Wimplicit-fallthrough, test=develop
      
      * remove -Wimplicit-fallthrough, test=develop
      
      * remove -Wimplicit-fallthrough, test=develop
      
      * remove -Wimplicit-fallthrough, test=develop
      
      * remove , test=develop
      74542577
  2. 10 4月, 2023 1 次提交
  3. 06 4月, 2023 2 次提交
    • J
      support more custom vjp (#52533) · 29c28e2f
      Jiabin Yang 提交于
      29c28e2f
    • K
      feat: add composite rule of roll grad (#52532) · 348a36b5
      Kang Zhao 提交于
      * feat: add relu composite rule
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, add python api of relu
      
      * feat: add relu composite rule, commit hook
      
      * fix: maximum type error & ban cinn test
      
      * fix: maximum input sequence bugs
      
      * resolve conflicts
      
      * fix: code style bugs
      
      * add: relu fp16 test
      
      * feat: add rsqrt composite rule
      
      * feat: add rsqrt composite rule
      
      * resolve conflicts of composite rule
      
      * fix: delete check eager
      
      * feat: add roll grad composite rule
      
      * fix minus shift
      
      * fix test roll op
      348a36b5
  4. 03 4月, 2023 1 次提交
  5. 31 3月, 2023 2 次提交
  6. 30 3月, 2023 2 次提交
  7. 29 3月, 2023 1 次提交
  8. 23 3月, 2023 1 次提交
  9. 22 3月, 2023 2 次提交
  10. 21 3月, 2023 2 次提交
  11. 20 3月, 2023 3 次提交
    • X
      【prim】New layer_norm grad (#51750) · 802a81d0
      xiaoguoguo626807 提交于
      * Add flatten composite rule
      
      * get the right xshape and pass func test
      
      * add cinn unit test
      
      * Remove cinn test, wait for it to be added after repair
      
      * add comp test to test_flatten_contiguous_range_op.py
      
      * remove func test on composite_ops
      
      * Add comments to maybe_wrap_dim func
      
      * remove commented code
      
      * fix the problem with 0D tensor case
      
      * add flatten split rule comment
      
      * fix syntax issues
      
      * block flatten on resnet_prim_cinn
      
      * init change
      
      * tmp commit
      
      * add layer_norm InferMeta check
      
      * cast type modify
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * recover
      
      * big tol
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * Cxx prim custom vjp (#8)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * [dy2static-ci] fix dy2static ci errors.
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [Prim] enable whitelist and blacklist for custom_vjp
      
      * debug log
      
      * clear log
      
      * fix
      
      * nothing
      
      * less memory
      
      * recover utils
      
      * fix
      
      * modify threshold value
      
      * skip layer_norm for test_bert
      
      * back to bert success state
      
      * add epsion
      
      * delete unnecessary compute
      
      * modify amp dtype
      
      * modify * order
      
      * delete sqrt check and fp16
      
      ---------
      Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com>
      Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com>
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      Co-authored-by: Nxiongkun <807377414@qq.com>
      802a81d0
    • W
      add sigmoid custom grad for prim (#51768) · ac47d003
      Weilong Wu 提交于
      ac47d003
    • J
      support relue custom vjp (#51742) · 604b7a53
      Jiabin Yang 提交于
      604b7a53
  12. 17 3月, 2023 1 次提交
  13. 15 3月, 2023 3 次提交
  14. 14 3月, 2023 4 次提交
  15. 13 3月, 2023 1 次提交
    • H
      【prim】Maximum grad (#51006) · 4a484973
      heyanru 提交于
      * refresh
      
      * compat
      
      * register
      
      * testop
      
      * fix
      
      * fix
      
      * fox
      
      * cast
      
      * cast
      
      * fix
      
      * type
      
      * fix
      
      * out
      
      * cast
      
      * fix
      
      * fix
      
      * fix
      
      * broad
      
      * broad
      
      * broad
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * broad
      
      * broad
      
      * numel
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * cinn
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      4a484973
  16. 09 3月, 2023 3 次提交
    • G
      add prim erf grad (#50436) · b7e4d974
      GGBond8488 提交于
      * add prim erf grad
      
      * add yaml config for prim erf grad
      
      * add math.h
      
      * add cmath
      
      * add math  defines
      
      * use define math
      
      * use define math
      
      * define M_2_SQRTPI
      
      * M_2_SQRTPI math
      
      * try math.h
      
      * fix typro
      
      * remove pow in erf grad
      
      * use new optest
      
      * add fp16 fp32 test
      
      * remove fp16 test
      b7e4d974
    • S
      add abs composite backward op (#50963) · d0d739ca
      SylarTiaNII 提交于
      * add abs composite backward op
      
      * add missing changes during merge
      
      * modify according to new rules
      
      * local UT OK
      
      * fix typo
      
      * codestyle
      
      * register composite operator
      
      * add fp16 test for abs
      
      * replace experimenta::tensor
      d0d739ca
    • W
      [prim] add elementwise_pow backward (#51230) · d9de3ef6
      wangzhen38 提交于
      * [cinn] add elementwise_pow backward
      
      * [cinn] update unnitest
      
      * [cinn] update by comments
      
      * [cinn] for ci
      
      * [cinn] for ci
      
      * [cinn] for ci
      
      * [cinn] for ci
      
      * [cinn] for ci
      d9de3ef6
  17. 08 3月, 2023 2 次提交
  18. 07 3月, 2023 1 次提交
  19. 06 3月, 2023 1 次提交
  20. 03 3月, 2023 1 次提交
  21. 02 3月, 2023 2 次提交
    • J
      【Prim】Fix slice error and eager comp (#51086) · bbca66f2
      Jiabin Yang 提交于
      * fix attrs copy error
      
      * fix bert by fix slice error
      
      * fix op test
      bbca66f2
    • W
      Add concat grad cinn (#50972) · a4689c90
      wangzhen38 提交于
      * [cinn] concat_grad
      
      * [cinn] concat_grad
      
      * [cinn] concat_grad build success
      
      * [Add PGLBOX] fix unnitest
      
      * [Add PGLBOX] fix unnitest
      
      * [Add PGLBOX] fix codestyle
      
      * [cinn] update by comments
      
      * [cinn] update by comment
      
      * [cinn] add axis check
      a4689c90
  22. 01 3月, 2023 2 次提交
    • J
      【Prim】Fix sqrt grad (#51045) · 5751b7f4
      Jiabin Yang 提交于
      * fix sqrt grad
      
      * fix sqrt grad
      5751b7f4
    • Z
      add topk prim backward (#50679) · 296b3ff0
      zqw_1997 提交于
      * tmp gather vjp
      
      * support gather
      
      * remove useless code
      
      * fix compiling error
      
      * fix ut
      
      * add eager test
      
      * add eager test
      
      * add seed
      
      * small change
      
      * fix cpu error
      
      * fix transpose op compat
      
      * remove tensor index case
      
      * fix prim_cinn
      
      * small commit
      
      * add cumsum prim backward
      
      * small commit
      
      * skip aixs=None test case
      
      * fix op generante eror
      
      * fix static test error
      
      * remove unused code
      
      * fix static test error
      
      * small commit
      
      * skip cpu float16 test case
      
      * skip eager cpu cumsum float16 test case
      
      * add eager and static UT
      
      * fix ut
      
      * add composite backward rule
      
      * fix error
      
      * fix type error and format error
      
      * add try cpu+float16 test
      
      * fix test bugs
      
      * remove test for cpu+float16 and make y[0] be the grad arg
      
      * add cinn test
      
      * fix UT
      
      * fix the wrong dim of v in test cases
      
      * change y[0] to y[1] for grad in UT
      
      * reshape flatten out
      
      * Disable cinn single test
      
      * use scatter_nd_add
      
      * modify the reshape part of topk_grad
      
      * delete useless build file
      
      * to make the syntax right
      
      * modify bug
      
      * try use of put_along_axis
      
      * remove cinn test
      
      * reformat todo
      
      * add silu composite rule
      
      * fix code style.
      
      * add cinn test
      
      * fix composite grad maker code gen
      
      * add prim in cumsum op test
      
      * remove old test
      
      * fix typro
      
      * pass the static test
      
      * fix typro
      
      * modify optest and delete old test files
      
      * remove normal test_top_k_op test
      
      * fix typro
      
      * pass axis=None test case
      
      * buffer comment
      
      * for debug
      
      * add silu fp16 unit test.
      
      * add static guard
      
      * remove forward prim test
      
      * remove same name axis
      
      * modify the test_top_v2_op.py to pass all local tests
      
      * delete the useless testcase
      
      * fix mistake
      
      * add more testcases to test dtype16 and dtype32
      
      ---------
      Co-authored-by: NJiabinYang <360788950@qq.com>
      Co-authored-by: NGGBond8488 <857631483@qq.com>
      Co-authored-by: Nzxcd <228587199@qq.com>
      Co-authored-by: NCharles-hit <wanghao107@baidu.com>
      296b3ff0
  23. 28 2月, 2023 1 次提交
    • X
      【prim】Matmul double grad composite api (#50452) · a0c473f4
      xiaoguoguo626807 提交于
      * modify name
      
      * merge develop
      
      * original code
      
      * build modify
      
      * success 2*2
      
      * fused dim=1 failed
      
      * success
      
      * modify static
      
      * success for static except dim=1
      
      * delete log
      
      * tmp modify
      
      * success
      
      * success
      
      * add fp1664
      
      * delete fp16 cpu test
      
      * stop windows test
      
      * review modify
      
      * modify tanh test
      
      * modify tanh
      
      * fix_conflixt
      
      * modift static prim
      
      * fix_conflict
      
      * Update test_static_prim.cc
      
      * update
      
      * bug fix
      a0c473f4