- 04 4月, 2023 1 次提交
-
-
由 Jiabin Yang 提交于
-
- 03 4月, 2023 2 次提交
-
-
由 cyber-pioneer 提交于
* polish prim arg None check * fix bug
-
由 cyber-pioneer 提交于
* simplify bn vjp code * simplify composite rule * polish name
-
- 31 3月, 2023 2 次提交
-
-
由 Xiaoxu Chen 提交于
-
由 张春乔 提交于
* autofix Co-authored-by: NLiyulingyue <83450930+Liyulingyue@users.noreply.github.com> * revert changes in python/paddle/distributed/fleet/utils/hybrid_parallel_util.py * empty commit, trigger ci * fix test_slice --------- Co-authored-by: NSigureMo <sigure.qaq@gmail.com>
-
- 30 3月, 2023 2 次提交
-
-
由 cyber-pioneer 提交于
* fix_prim * fix bug * add note * fix logic * fix * add note * fix check * fix bug * fix bug * fix bug * add debug * fix check * fix bug * sync print log * fix test case * change default * change test case time
-
由 cyberslack_lee 提交于
[CodeStyle][C416][C417] rewrite unnecessary comprehension with function call and use generator instead of map (#52140) * codestyle c416 c417 * fix error * fix inc * unify all C4 rules into one * fix inc --------- Co-authored-by: NSigureMo <sigure.qaq@gmail.com>
-
- 29 3月, 2023 1 次提交
-
-
由 Yichen Zhang 提交于
* add group_norm composite rule * add test for scale_grad and bias_grad * resolve conflicts * remove amp in composite_rule.py * add float16 test * deal with NHWC format * keep the composite rule in float16 identical as original kernel * resolve conflicts
-
- 28 3月, 2023 2 次提交
-
-
由 张春乔 提交于
-
由 Jiabin Yang 提交于
* optimize composite rule by making scalar shape as []1 * fix shape usage for 0D * fix rules * fix 0D error * fix flatten 0D error * fix bn eval mode * fix bn test * fix flatten
-
- 27 3月, 2023 1 次提交
-
-
由 cyber-pioneer 提交于
-
- 23 3月, 2023 2 次提交
-
-
由 chenjian 提交于
* add meshgrid composite rule * add meshgrid composite rule * update * add into CMakeLists * fix * update * update * optimize code * fix meshgrid op * update test
-
由 Infinity_lee 提交于
-
- 21 3月, 2023 2 次提交
-
-
由 cyber-pioneer 提交于
* simplify batch_norm composite rule * polish code
-
由 zhouweiwei2014 提交于
[Zero-Dim] Support 0D for numel/rank/size/optimizer/create_parameter/create_global_var, fix some usage to adapt 0D (#51566)
-
- 20 3月, 2023 4 次提交
-
-
由 xiaoguoguo626807 提交于
* Add flatten composite rule * get the right xshape and pass func test * add cinn unit test * Remove cinn test, wait for it to be added after repair * add comp test to test_flatten_contiguous_range_op.py * remove func test on composite_ops * Add comments to maybe_wrap_dim func * remove commented code * fix the problem with 0D tensor case * add flatten split rule comment * fix syntax issues * block flatten on resnet_prim_cinn * init change * tmp commit * add layer_norm InferMeta check * cast type modify * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * recover * big tol * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * Cxx prim custom vjp (#8) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [Prim] enable whitelist and blacklist for custom_vjp * debug log * clear log * fix * nothing * less memory * recover utils * fix * modify threshold value * skip layer_norm for test_bert * back to bert success state * add epsion * delete unnecessary compute * modify amp dtype * modify * order * delete sqrt check and fp16 --------- Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com> Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com> Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> Co-authored-by: Nxiongkun <807377414@qq.com>
-
由 warrentdrew 提交于
* add composite rule for squeeze * fix pre commit * fix pre commit * simplify rules * arrange code * fix int axis * simplify squeeze axis rules * bugfix * fix pre commit
-
由 GGBond8488 提交于
* migrate fill_constant to paddle.tensor * move fill_constant to paddle.tensor and repalce the reference * add missing fill_constant replacement * fix typro * remove unused import fill_constant * fix zeros import error * fix circle import * fix layers.zeros * fix unitest * fix unitests * fix unitest * use paddle.full replace fill_constant in samplecode * fix sample code * recovery xpu test * recovery xpu test * fix circle import * fix utils import error * fix utils error * fix circle import * redo * fix circle import * fix prim fill constant import * fix type error * fix increase error * fix test error * fix fill_constant
-
由 Jiabin Yang 提交于
-
- 17 3月, 2023 2 次提交
-
-
由 cyber-pioneer 提交于
* add bn vjp * fix example * fix code * fix code * fix cinn case * fix code * fix example * fix code * fix example * fix example
-
由 mhy-666 提交于
* add sqrt composite rule/test * add sqrt composite rule/test * fix ops/sqrt, add cinn test * fix sqrt_comp * fix sqrt_comp * fix sqrt_comp * fix * fix codestyle * fix codestyle * add fp16 test * add ops/sqrt * fix * fix * fix unitest * fix * fix * fix
-
- 16 3月, 2023 2 次提交
-
-
由 Roc 提交于
-
由 Jiabin Yang 提交于
* support amp logic for layer_norm and softmax * fix layer_norm amp * fix layernorm api and dropout fp16 * fix layernorm api and dropout fp16 * fix bn, ln dtype in float16 * fix dropout fp16 * fix comment * fix cinn dropout amp error
-
- 15 3月, 2023 3 次提交
-
-
由 Kang Zhao 提交于
* feat: add relu composite rule * feat: add relu composite rule, maximum op * feat: add relu composite rule, maximum op * feat: add relu composite rule, polish comments * feat: add relu composite rule, polish comments * feat: add relu composite rule, add python api of relu * feat: add relu composite rule, commit hook * fix: maximum type error & ban cinn test * fix: maximum input sequence bugs * resolve conflicts * fix: code style bugs * add: relu fp16 test * feat: add rsqrt composite rule * feat: add rsqrt composite rule * resolve conflicts of composite rule * fix: delete check eager
-
由 Jiabin Yang 提交于
* support amp logic for layer_norm and softmax * fix layer_norm amp * fix layernorm api and dropout fp16 * fix layernorm api and dropout fp16 * fix bn, ln dtype in float16 * fix dropout fp16 * fix comment
-
由 chenjian 提交于
* add pow composite rule * fix test * fix unit test * update test * fix test * update
-
- 14 3月, 2023 8 次提交
-
-
由 qizhaoaoe 提交于
-
由 xiaoguoguo626807 提交于
* init * modify
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 Aurelius84 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com>
-
由 cyber-pioneer 提交于
-
- 13 3月, 2023 2 次提交
-
-
由 mengziheng 提交于
* first test * add unsqueeze_op
-
由 xysheng-baidu 提交于
* Add expand composite rule * reshape x when dim_in less than dim_out * add tile op for expand * remove rensor shape case when comp prim * enable cinn case * dim_out can't be 0 * update test case for prim type
-
- 08 3月, 2023 1 次提交
-
-
由 Kang Zhao 提交于
* feat: add relu composite rule * feat: add relu composite rule, maximum op * feat: add relu composite rule, maximum op * feat: add relu composite rule, polish comments * feat: add relu composite rule, polish comments * feat: add relu composite rule, add python api of relu * feat: add relu composite rule, commit hook * fix: maximum type error & ban cinn test * fix: maximum input sequence bugs * resolve conflicts * fix: code style bugs * add: relu fp16 test
-
- 07 3月, 2023 1 次提交
-
-
由 ccrrong 提交于
* add stack composite rule * add float16 datatype test
-
- 03 3月, 2023 1 次提交
-
-
由 zxcd 提交于
* add sigmoid composite rule * add python api * fix code style. * add check_prim=True * add sigmoid fp16 unit test. * fix code style. * rm bf16 check_prim * fix code style.
-
- 02 3月, 2023 1 次提交
-
-
由 Roc 提交于
* add composite op hard swish * add test grad * update apis calling * update date range * add ut * tune off cinn for 0-d shape * skip cinn
-