1. 15 3月, 2023 3 次提交
    • K
      feat: add rsqrt composite rule (#51432) · c9ca7c35
      Kang Zhao 提交于
      * feat: add relu composite rule
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, add python api of relu
      
      * feat: add relu composite rule, commit hook
      
      * fix: maximum type error & ban cinn test
      
      * fix: maximum input sequence bugs
      
      * resolve conflicts
      
      * fix: code style bugs
      
      * add: relu fp16 test
      
      * feat: add rsqrt composite rule
      
      * feat: add rsqrt composite rule
      
      * resolve conflicts of composite rule
      
      * fix: delete check eager
      c9ca7c35
    • J
      【Prim】Support amp logic for layer_norm and softmax (#51473) · 64076727
      Jiabin Yang 提交于
      * support amp logic for layer_norm and softmax
      
      * fix layer_norm amp
      
      * fix layernorm api and dropout fp16
      
      * fix layernorm api and dropout fp16
      
      * fix bn, ln dtype in float16
      
      * fix dropout fp16
      
      * fix comment
      64076727
    • C
      [Prim] add pow composite rule (#51070) · 2d9e103e
      chenjian 提交于
      * add pow composite rule
      
      * fix test
      
      * fix unit test
      
      * update test
      
      * fix test
      
      * update
      2d9e103e
  2. 14 3月, 2023 8 次提交
  3. 13 3月, 2023 2 次提交
  4. 08 3月, 2023 1 次提交
    • K
      feat: add relu composite (#50819) · 079f41c8
      Kang Zhao 提交于
      * feat: add relu composite rule
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, maximum op
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, polish comments
      
      * feat: add relu composite rule, add python api of relu
      
      * feat: add relu composite rule, commit hook
      
      * fix: maximum type error & ban cinn test
      
      * fix: maximum input sequence bugs
      
      * resolve conflicts
      
      * fix: code style bugs
      
      * add: relu fp16 test
      079f41c8
  5. 07 3月, 2023 1 次提交
  6. 03 3月, 2023 1 次提交
    • Z
      add sigmoid composite rule (#50827) · d3352b99
      zxcd 提交于
      * add sigmoid composite rule
      
      * add python api
      
      * fix code style.
      
      * add check_prim=True
      
      * add sigmoid fp16 unit test.
      
      * fix code style.
      
      * rm bf16 check_prim
      
      * fix code style.
      d3352b99
  7. 02 3月, 2023 1 次提交
    • R
      Comp hardswish (#51003) · 51331098
      Roc 提交于
      * add composite op hard swish
      
      * add test grad
      
      * update apis calling
      
      * update date range
      
      * add ut
      
      * tune off cinn for 0-d shape
      
      * skip cinn
      51331098
  8. 01 3月, 2023 2 次提交
    • J
      fix attrs copy error (#51056) · e152e891
      Jiabin Yang 提交于
      e152e891
    • Y
      Add full_like composite rule (#50794) · 7468bab4
      Yichen Zhang 提交于
      * implement composite full_like and simple unit test
      
      * implement op tests for composite full_like op
      
      * some modification as reviewers suggested
      add cinn op test to CMakeLists.txt
      fix code style
      
      * fix code style
      
      * modify input args of prim fill_any_like op
      
      * resolve conflicts
      
      * resolve conflicts
      
      * modify python api and unit tests as suggested
      
      * resolve conflicts
      
      * resolve conflicts
      
      * use framework.dtype to convert dtype in Op test
      7468bab4
  9. 28 2月, 2023 3 次提交
    • I
      Fix some typos (#50914) · 5d8fe822
      iLeGend 提交于
      5d8fe822
    • Z
      add silu composite rule (#50838) · 5d70ba6d
      zxcd 提交于
      * add silu composite rule
      
      * fix code style.
      
      * add silu fp16 unit test.
      5d70ba6d
    • X
      Add flatten composite rule (#50672) · 8220771b
      xysheng-baidu 提交于
      * Add flatten composite rule
      
      * get the right xshape and pass func test
      
      * add cinn unit test
      
      * Remove cinn test, wait for it to be added after repair
      
      * add comp test to test_flatten_contiguous_range_op.py
      
      * remove func test on composite_ops
      
      * Add comments to maybe_wrap_dim func
      
      * remove commented code
      
      * fix the problem with 0D tensor case
      
      * add flatten split rule comment
      
      * fix syntax issues
      
      * block flatten on resnet_prim_cinn
      
      * remove maybe_wrap_dim func
      
      * Use none instead od xshape
      8220771b
  10. 24 2月, 2023 2 次提交
  11. 22 2月, 2023 2 次提交
  12. 21 2月, 2023 2 次提交
    • C
      [Prim] Add op map (#50673) · 3c7e94d6
      cyber-pioneer 提交于
      * fix flatten op map
      
      * remove prim op all list
      
      * add op map info of full_like
      
      * polish code
      3c7e94d6
    • X
      【prim】Layer norm (#50422) · 2f4763ee
      xiaoguoguo626807 提交于
      * fix composite mean op map
      
      * fix composite check output
      
      * init layer_norm
      
      * init layer_norm
      
      * map output from composite rule to origin op
      
      * add dropout op map
      
      * add input map check
      
      * polish log
      
      * modify rules
      
      * success test_forward
      
      * modify test without cinn
      
      * modify cinn test
      
      * modify cinn test
      
      * except fp64
      
      * except fp64
      
      * delete flatten
      
      * delete unused change
      
      * review
      
      * pass cpu test
      
      * code style
      
      * delete flatten fp16 error
      
      * modify flatten test
      
      ---------
      Co-authored-by: Ncyber-pioneer <chenzhuo@tju.edu.cn>
      2f4763ee
  13. 20 2月, 2023 1 次提交
  14. 16 2月, 2023 1 次提交
    • Z
      Add mean composite rule (#50298) · f7f67b72
      zqw_1997 提交于
      * beta
      
      * small commit
      
      * add batch_norm composite rule
      
      move composite test case
      
      remove unuseful var
      
      add composite op blacklist
      
      * small change v2
      
      * finish the test_composite_mean and test_composite_mean_grad
      
      * add ops assertion to the tests
      
      * add cinn test
      
      * fix the error and inappropriate usage in func: mean_composite
      
      * remove the ref of outer lib in primtives.py
      
      * modify sample code of reduce_sum
      
      * fix composite mean op map
      
      * modify testcases to test more float type
      
      * remove cpu float16 test
      
      * cinn test fix
      
      * remove reduce_max
      
      * change the name sum to sum_x
      
      * change the use of reduce_sum to sum
      
      ---------
      Co-authored-by: Ncyber-pioneer <chenzhuo@tju.edu.cn>
      f7f67b72
  15. 15 2月, 2023 1 次提交
    • C
      fix composite op map (#50397) · ff86aeab
      cyber-pioneer 提交于
      * map output from composite rule to origin op
      
      add mean layer_norm dropout op map
      
      add input map check
      
      composite softmax support input shape []
      
      * composite softmax support shape []
      
      * polish log
      
      * solve conflict
      
      * polish code
      
      * polish op map output
      
      * add check dtype
      ff86aeab
  16. 14 2月, 2023 2 次提交
    • M
      b85af464
    • G
      Add gelu composite rule (#50295) · c364f41d
      GGBond8488 提交于
      * add gelu composite rule
      
      * use full replace fill_constant
      
      * change the form of calculation
      
      * remove float16 test for composite gelu
      
      * reformate code
      
      * remove float16 test case
      
      * add forwad with prim and backward without prim test
      
      * add float16 test for composite gelu and add high dims test
      
      * add float16 test case and high dims test
      
      * shield float16 and cpu test case
      
      * increase train step to 10 in test cinn prim gelu
      
      * replace pow to multiply
      c364f41d
  17. 09 2月, 2023 1 次提交
  18. 07 2月, 2023 1 次提交
  19. 20 1月, 2023 1 次提交
  20. 17 1月, 2023 1 次提交
  21. 13 1月, 2023 1 次提交
  22. 30 12月, 2022 1 次提交
  23. 29 11月, 2022 1 次提交