- 23 3月, 2023 2 次提交
-
-
由 Infinity_lee 提交于
-
由 PuQing 提交于
[CodeStyle][C408][C409][C410] Fix unnecessary <dict/list/tuple> call and unnecessary <list/tuple> passed to <list/tupule>() (#51928) * autofix * add select config * autofix C410 * add C410 select
-
- 22 3月, 2023 3 次提交
-
-
由 ShenLiang 提交于
-
由 iSerendipity 提交于
-
由 Ainavo 提交于
* replace assert false with AssertionError * 修改配置文件多余的部分
-
- 21 3月, 2023 3 次提交
-
-
由 Ainavo 提交于
-
由 cyber-pioneer 提交于
* simplify batch_norm composite rule * polish code
-
由 zhouweiwei2014 提交于
[Zero-Dim] Support 0D for numel/rank/size/optimizer/create_parameter/create_global_var, fix some usage to adapt 0D (#51566)
-
- 20 3月, 2023 7 次提交
-
-
由 Ainavo 提交于
* remove super call with parameters * fix bug
-
由 xiaoguoguo626807 提交于
* Add flatten composite rule * get the right xshape and pass func test * add cinn unit test * Remove cinn test, wait for it to be added after repair * add comp test to test_flatten_contiguous_range_op.py * remove func test on composite_ops * Add comments to maybe_wrap_dim func * remove commented code * fix the problem with 0D tensor case * add flatten split rule comment * fix syntax issues * block flatten on resnet_prim_cinn * init change * tmp commit * add layer_norm InferMeta check * cast type modify * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * recover * big tol * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * Cxx prim custom vjp (#8) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [Prim] enable whitelist and blacklist for custom_vjp * debug log * clear log * fix * nothing * less memory * recover utils * fix * modify threshold value * skip layer_norm for test_bert * back to bert success state * add epsion * delete unnecessary compute * modify amp dtype * modify * order * delete sqrt check and fp16 --------- Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com> Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com> Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> Co-authored-by: Nxiongkun <807377414@qq.com>
-
由 zhouweiwei2014 提交于
-
由 GGBond8488 提交于
* remove no used fluid beam_search_decoder * move Layer and related helper to paddle.nn.common * modify Layer references from dygraph.layers.Layer to paddle.nn.common.layers * stash changge * remove fluid layer_object_helper, layers.py * remove fluid layers init * add setip * fix unitest * delete layers in fluid.dygraph * merge paddle.tensor.stat,py * fix circle import * fix curcle import * remove redundant in_dygraph_mode import * revoce paddle.nn.common.* in fluid.__init__ * recovery nn.rnn * paddle.frame use lazy import import paddle.jit to avoid circle import * remove left dygraph.layers ref * merge develop * fix import error * fix test error * fxi merge error * fix test fluid.Layer * fix test error * fix test error * fix import error * fix import error * fix comments * fix circle import * fix rnn import error * fix circle import
-
由 warrentdrew 提交于
* add composite rule for squeeze * fix pre commit * fix pre commit * simplify rules * arrange code * fix int axis * simplify squeeze axis rules * bugfix * fix pre commit
-
由 GGBond8488 提交于
* migrate fill_constant to paddle.tensor * move fill_constant to paddle.tensor and repalce the reference * add missing fill_constant replacement * fix typro * remove unused import fill_constant * fix zeros import error * fix circle import * fix layers.zeros * fix unitest * fix unitests * fix unitest * use paddle.full replace fill_constant in samplecode * fix sample code * recovery xpu test * recovery xpu test * fix circle import * fix utils import error * fix utils error * fix circle import * redo * fix circle import * fix prim fill constant import * fix type error * fix increase error * fix test error * fix fill_constant
-
由 Jiabin Yang 提交于
-
- 17 3月, 2023 4 次提交
-
-
由 qizhaoaoe 提交于
* fluid clean: remove fluid.ir to framework.ir and some funcs form fluid.layer.io to incubate. * delete fluid.ir
-
由 cyber-pioneer 提交于
* add bn vjp * fix example * fix code * fix code * fix cinn case * fix code * fix example * fix code * fix example * fix example
-
由 mhy-666 提交于
* add sqrt composite rule/test * add sqrt composite rule/test * fix ops/sqrt, add cinn test * fix sqrt_comp * fix sqrt_comp * fix sqrt_comp * fix * fix codestyle * fix codestyle * add fp16 test * add ops/sqrt * fix * fix * fix unitest * fix * fix * fix
-
由 Nyakku Shigure 提交于
-
- 16 3月, 2023 2 次提交
-
-
由 Roc 提交于
-
由 Jiabin Yang 提交于
* support amp logic for layer_norm and softmax * fix layer_norm amp * fix layernorm api and dropout fp16 * fix layernorm api and dropout fp16 * fix bn, ln dtype in float16 * fix dropout fp16 * fix comment * fix cinn dropout amp error
-
- 15 3月, 2023 4 次提交
-
-
由 Kang Zhao 提交于
* feat: add relu composite rule * feat: add relu composite rule, maximum op * feat: add relu composite rule, maximum op * feat: add relu composite rule, polish comments * feat: add relu composite rule, polish comments * feat: add relu composite rule, add python api of relu * feat: add relu composite rule, commit hook * fix: maximum type error & ban cinn test * fix: maximum input sequence bugs * resolve conflicts * fix: code style bugs * add: relu fp16 test * feat: add rsqrt composite rule * feat: add rsqrt composite rule * resolve conflicts of composite rule * fix: delete check eager
-
由 Jiabin Yang 提交于
* support amp logic for layer_norm and softmax * fix layer_norm amp * fix layernorm api and dropout fp16 * fix layernorm api and dropout fp16 * fix bn, ln dtype in float16 * fix dropout fp16 * fix comment
-
由 chenjian 提交于
* add pow composite rule * fix test * fix unit test * update test * fix test * update
-
由 wanghuancoder 提交于
* refine _found_inf
-
- 14 3月, 2023 8 次提交
-
-
由 qizhaoaoe 提交于
-
由 xiaoguoguo626807 提交于
* init * modify
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 Aurelius84 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com>
-
由 cyber-pioneer 提交于
-
- 13 3月, 2023 2 次提交
-
-
由 mengziheng 提交于
* first test * add unsqueeze_op
-
由 xysheng-baidu 提交于
* Add expand composite rule * reshape x when dim_in less than dim_out * add tile op for expand * remove rensor shape case when comp prim * enable cinn case * dim_out can't be 0 * update test case for prim type
-
- 10 3月, 2023 2 次提交
-
-
由 pangengzheng 提交于
* support run haokanctr model in heterps-models * polish setup.py * polish JVM_LIB in evn_dict * align infer auc with DistPsArch pre-stable
-
由 sneaxiy 提交于
* add attn_bias.py * add Python interface * add license * add test_attn_bias.py * fix CPU test error * fix ci error
-
- 09 3月, 2023 3 次提交
-
-
由 zqw_1997 提交于
* move fluid.utils to paddle.utils.layers_utils * fix error * delete original fluid layers utils * remove import and old utils * remove more old utils import * change import path of fill_constant in the layers_utils.py * fix mistake * fix error * expose in __init__.py * for comment * when change the ref of func is_sequence, it should change to the root of is_sequence instead * for codecheck
-
由 pangengzheng 提交于
* support run haokanctr model in heterps-models * polish setup.py * polish JVM_LIB in evn_dict
-
由 wangzhen38 提交于
* [clear ps] del some all list * [clear ps] for ci * [clear ps] rm fluid emb
-