1. 27 3月, 2023 1 次提交
  2. 23 3月, 2023 2 次提交
  3. 21 3月, 2023 2 次提交
  4. 20 3月, 2023 4 次提交
    • X
      【prim】New layer_norm grad (#51750) · 802a81d0
      xiaoguoguo626807 提交于
      * Add flatten composite rule
      
      * get the right xshape and pass func test
      
      * add cinn unit test
      
      * Remove cinn test, wait for it to be added after repair
      
      * add comp test to test_flatten_contiguous_range_op.py
      
      * remove func test on composite_ops
      
      * Add comments to maybe_wrap_dim func
      
      * remove commented code
      
      * fix the problem with 0D tensor case
      
      * add flatten split rule comment
      
      * fix syntax issues
      
      * block flatten on resnet_prim_cinn
      
      * init change
      
      * tmp commit
      
      * add layer_norm InferMeta check
      
      * cast type modify
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * recover
      
      * big tol
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      * add unittest
      
      * fix typo
      
      * fix typo
      
      * fix map.at
      
      * fix find
      
      * fix test
      
      * fix cinn cache key structure realize
      
      * using ordered map for attributes
      
      * add test by review advice
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * Cxx prim custom vjp (#8)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * Pr 50885 (#7)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557)
      
      * [CINN]Enhance CacheKey hash logic by considering input dtypes
      
      ---------
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix code in a dy2static-friendly way.
      
      * [dystatic] add hooker for prim
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [prim] enable dygraph_to_static to support custom_vjp
      
      * fix cast prim and vjp dtype mapping error bug
      
      * [dy2static-ci] fix dy2static ci errors.
      
      ---------
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      
      * [Prim] enable whitelist and blacklist for custom_vjp
      
      * debug log
      
      * clear log
      
      * fix
      
      * nothing
      
      * less memory
      
      * recover utils
      
      * fix
      
      * modify threshold value
      
      * skip layer_norm for test_bert
      
      * back to bert success state
      
      * add epsion
      
      * delete unnecessary compute
      
      * modify amp dtype
      
      * modify * order
      
      * delete sqrt check and fp16
      
      ---------
      Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com>
      Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com>
      Co-authored-by: NAurelius84 <zhangliujie@baidu.com>
      Co-authored-by: Njiangcheng <thisjiang@qq.com>
      Co-authored-by: Ncxxly <chenxx_id@163.com>
      Co-authored-by: Nxiongkun <807377414@qq.com>
      802a81d0
    • warrentdrew's avatar
      add composite rules for squeeze op (#51539) · 89ff0d59
      warrentdrew 提交于
      * add composite rule for squeeze
      
      * fix pre commit
      
      * fix pre commit
      
      * simplify rules
      
      * arrange code
      
      * fix int axis
      
      * simplify squeeze axis rules
      
      * bugfix
      
      * fix pre commit
      89ff0d59
    • G
      Fluid clean move out fill constant (#49511) · c985b1ac
      GGBond8488 提交于
      * migrate fill_constant to paddle.tensor
      
      * move fill_constant to paddle.tensor and repalce the reference
      
      * add missing fill_constant replacement
      
      * fix typro
      
      * remove unused import fill_constant
      
      * fix zeros import error
      
      * fix circle import
      
      * fix layers.zeros
      
      * fix unitest
      
      * fix unitests
      
      * fix unitest
      
      * use paddle.full replace fill_constant in samplecode
      
      * fix sample code
      
      * recovery xpu test
      
      * recovery xpu test
      
      * fix circle import
      
      * fix utils import error
      
      * fix utils error
      
      * fix circle import
      
      * redo
      
      * fix circle import
      
      * fix prim fill constant import
      
      * fix type error
      
      * fix increase error
      
      * fix test error
      
      * fix fill_constant
      c985b1ac
    • J
      support relue custom vjp (#51742) · 604b7a53
      Jiabin Yang 提交于
      604b7a53
  5. 17 3月, 2023 2 次提交
    • C
      [Prim] support batch_norm vjp (#51283) · ff40a7e5
      cyber-pioneer 提交于
      * add bn vjp
      
      * fix example
      
      * fix code
      
      * fix code
      
      * fix cinn case
      
      * fix code
      
      * fix example
      
      * fix code
      
      * fix example
      
      * fix example
      ff40a7e5
    • M
      Add sqrt composite rule (#51080) · aba9c4d4
      mhy-666 提交于
      * add sqrt composite rule/test
      
      * add sqrt composite rule/test
      
      * fix ops/sqrt, add cinn test
      
      * fix sqrt_comp
      
      * fix sqrt_comp
      
      * fix sqrt_comp
      
      * fix
      
      * fix codestyle
      
      * fix codestyle
      
      * add fp16 test
      
      * add ops/sqrt
      
      * fix
      
      * fix
      
      * fix unitest
      
      * fix
      
      * fix
      
      * fix
      aba9c4d4
  6. 16 3月, 2023 2 次提交
  7. 15 3月, 2023 3 次提交
    • K
      feat: add rsqrt composite rule (#51432) · c9ca7c35
      Kang Zhao 提交于
      * feat: add relu composite rule
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, add python api of relu
      
      * feat: add relu composite rule, commit hook
      
      * fix: maximum type error & ban cinn test
      
      * fix: maximum input sequence bugs
      
      * resolve conflicts
      
      * fix: code style bugs
      
      * add: relu fp16 test
      
      * feat: add rsqrt composite rule
      
      * feat: add rsqrt composite rule
      
      * resolve conflicts of composite rule
      
      * fix: delete check eager
      c9ca7c35
    • J
      【Prim】Support amp logic for layer_norm and softmax (#51473) · 64076727
      Jiabin Yang 提交于
      * support amp logic for layer_norm and softmax
      
      * fix layer_norm amp
      
      * fix layernorm api and dropout fp16
      
      * fix layernorm api and dropout fp16
      
      * fix bn, ln dtype in float16
      
      * fix dropout fp16
      
      * fix comment
      64076727
    • C
      [Prim] add pow composite rule (#51070) · 2d9e103e
      chenjian 提交于
      * add pow composite rule
      
      * fix test
      
      * fix unit test
      
      * update test
      
      * fix test
      
      * update
      2d9e103e
  8. 14 3月, 2023 8 次提交
  9. 13 3月, 2023 2 次提交
  10. 08 3月, 2023 1 次提交
    • K
      feat: add relu composite (#50819) · 079f41c8
      Kang Zhao 提交于
      * feat: add relu composite rule
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, add python api of relu
      
      * feat: add relu composite rule, commit hook
      
      * fix: maximum type error & ban cinn test
      
      * fix: maximum input sequence bugs
      
      * resolve conflicts
      
      * fix: code style bugs
      
      * add: relu fp16 test
      079f41c8
  11. 07 3月, 2023 1 次提交
  12. 03 3月, 2023 1 次提交
    • Z
      add sigmoid composite rule (#50827) · d3352b99
      zxcd 提交于
      * add sigmoid composite rule
      
      * add python api
      
      * fix code style.
      
      * add check_prim=True
      
      * add sigmoid fp16 unit test.
      
      * fix code style.
      
      * rm bf16 check_prim
      
      * fix code style.
      d3352b99
  13. 02 3月, 2023 1 次提交
    • R
      Comp hardswish (#51003) · 51331098
      Roc 提交于
      * add composite op hard swish
      
      * add test grad
      
      * update apis calling
      
      * update date range
      
      * add ut
      
      * tune off cinn for 0-d shape
      
      * skip cinn
      51331098
  14. 01 3月, 2023 2 次提交
    • J
      fix attrs copy error (#51056) · e152e891
      Jiabin Yang 提交于
      e152e891
    • Y
      Add full_like composite rule (#50794) · 7468bab4
      Yichen Zhang 提交于
      * implement composite full_like and simple unit test
      
      * implement op tests for composite full_like op
      
      * some modification as reviewers suggested
      add cinn op test to CMakeLists.txt
      fix code style
      
      * fix code style
      
      * modify input args of prim fill_any_like op
      
      * resolve conflicts
      
      * resolve conflicts
      
      * modify python api and unit tests as suggested
      
      * resolve conflicts
      
      * resolve conflicts
      
      * use framework.dtype to convert dtype in Op test
      7468bab4
  15. 28 2月, 2023 3 次提交
    • I
      Fix some typos (#50914) · 5d8fe822
      iLeGend 提交于
      5d8fe822
    • Z
      add silu composite rule (#50838) · 5d70ba6d
      zxcd 提交于
      * add silu composite rule
      
      * fix code style.
      
      * add silu fp16 unit test.
      5d70ba6d
    • X
      Add flatten composite rule (#50672) · 8220771b
      xysheng-baidu 提交于
      * Add flatten composite rule
      
      * get the right xshape and pass func test
      
      * add cinn unit test
      
      * Remove cinn test, wait for it to be added after repair
      
      * add comp test to test_flatten_contiguous_range_op.py
      
      * remove func test on composite_ops
      
      * Add comments to maybe_wrap_dim func
      
      * remove commented code
      
      * fix the problem with 0D tensor case
      
      * add flatten split rule comment
      
      * fix syntax issues
      
      * block flatten on resnet_prim_cinn
      
      * remove maybe_wrap_dim func
      
      * Use none instead od xshape
      8220771b
  16. 24 2月, 2023 2 次提交
  17. 22 2月, 2023 2 次提交
  18. 21 2月, 2023 1 次提交