- 28 3月, 2023 3 次提交
-
-
由 RedContritio 提交于
* support auto generate static for one_hot_v2 * format
-
由 Wang Xin 提交于
-
由 RedContritio 提交于
* support auto generate for log_softmax * add data_type
-
- 27 3月, 2023 2 次提交
-
-
由 HappyHeavyRain 提交于
* add fused_yaml fused_backward * fix eager_funciton bug * add some comment of fused yaml file * add 'support_dygraph_mode' configuration in fused yaml * delete some 'fused_api.h' in include file * add fused flag in api_gen
-
由 HappyHeavyRain 提交于
* support assign op * support assign infer_var_type * change code according to review * change code according to review * only save 'get_infer_var_type_func' * rest file mode
-
- 24 3月, 2023 2 次提交
-
-
由 TaoTao Li 提交于
* add all_reduce, reduce kernel and api * fix all_reduce reduce ut fix reduce op maker conflict fix merge conflicts * fix conflicts, rename ReduceOp->ReduceBaseOp in reduce_ops rename allreduce op, to remove * fix code format fix comments * modify test_collective_reduce_api ut timeout * fix PR-CI-Build fix comments: format phi operator
-
由 ZhangDY-6483 提交于
* first version, notest * return final rst, notest * use infinity() instead of max * ut structure * start up of ut * generate lse * update * add depense * reconstruct cmake * move file * add memory efficient attention and fix blasimpl * update * update cmake * add namespace * update cmake * use .cu * update for pad3d * bug fix * bug fix * update * bug fix * update enforce * add test case * merge the lse pad * fix kernel_fn of backward * fix PADDLE_ENFORCE_EQ and phi_api * fix PADDLE_ENFORCE * fix PADDLE_ENFORCE * rerun coverage * fix memory efficient attention test * rerun ci * add cuda version condition * add cuda version condition * delete WIP test * replace PADDLE_ENFORCE * edit the namespace of datatype in multiple.cc * rerun * rerun --------- Co-authored-by: Nliuyuang <liuyuang@baidu.com>
-
- 23 3月, 2023 6 次提交
-
-
由 HongyuJia 提交于
[Polish Log] Polish Tensor operants' log: 'OperantsManager reusing XXX mode API {func_name}' (#51991) * [Polish Log] Polish Tensor operants' log: 'OperantsManager reusing XXX mode API {func_name}' * Make API name more precise
-
由 xiaoguoguo626807 提交于
* delete prim flag for matmul_2_grad * delete prim flag for matmul_2_grad * add new setgradoutmeta for matmul_double_grad_node * modify test and delete log * deal with review
-
由 sneaxiy 提交于
* remove fluid deps in fused_linear_param_grad_add_kernel * fix compile error * fix ut error * follow comments
-
由 cxxly 提交于
-
由 Infinity_lee 提交于
-
由 PuQing 提交于
[CodeStyle][C408][C409][C410] Fix unnecessary <dict/list/tuple> call and unnecessary <list/tuple> passed to <list/tupule>() (#51928) * autofix * add select config * autofix C410 * add C410 select
-
- 22 3月, 2023 9 次提交
-
-
由 HappyHeavyRain 提交于
* test_get_kernel * add invoke signature * change reduce_max * change frobenius_norm * reset reduce_max according to composite and change reduce_all * fix the bug when Scalar(*) * fix 'scalar when support_tensor' * change code according to review * change 'keep_signature' to 'manual_signature' and add some erro info * support optimizers autogen * change sgd yaml * change generate signature * fix test/cpp/new_executor/CM * reset signature generated function * change signature funciton * change signature funciton
-
由 ShenLiang 提交于
-
由 Sławomir Siwek 提交于
* extract common methods to reuse * add header for transpose ops * fused_transpose * Split big function * transpose2 tests * fused_transpose * Apply extra attributes * add pbtxt file * update pbtxt * Merge develop * add more strict op compats * code style * remove mkldnn_data_type * unify SetOutMemDescWithReshape2FuseSupport * adjust quantize-dequantize for transpose * remove appendact * transpose2 quantization * fix int8 tests * adjust transpose_op to current develop * delete fusion code from transpose_kernel * add fused transpose to NHWC unittest * change order
-
由 Wang Xin 提交于
* add autogen code for index_add op * bug fixed
-
由 RedContritio 提交于
* supoort auto generate p_norm * fix bug in backward
-
由 RedContritio 提交于
* support auto generate for dirichlet * use uppercase in args * use op_compat for name mapping
-
由 wangxiaoning 提交于
* max comp * fix * add test * fix * fix * fix * fix * fix test * fix api
-
由 sneaxiy 提交于
* add fused_linear_param_grad_add_kernel * fix compile error * remove flag * fix ci compile error * fix ci compile error * revert pylayer revision * fix ci ut * improve performance
-
由 Sylwester Fraczek 提交于
-
- 21 3月, 2023 1 次提交
-
-
由 xiaoguoguo626807 提交于
* init * close cinn * close cinn * add public_python_api
-
- 20 3月, 2023 6 次提交
-
-
由 HappyHeavyRain 提交于
-
由 xiaoguoguo626807 提交于
* Add flatten composite rule * get the right xshape and pass func test * add cinn unit test * Remove cinn test, wait for it to be added after repair * add comp test to test_flatten_contiguous_range_op.py * remove func test on composite_ops * Add comments to maybe_wrap_dim func * remove commented code * fix the problem with 0D tensor case * add flatten split rule comment * fix syntax issues * block flatten on resnet_prim_cinn * init change * tmp commit * add layer_norm InferMeta check * cast type modify * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * recover * big tol * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * Cxx prim custom vjp (#8) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [Prim] enable whitelist and blacklist for custom_vjp * debug log * clear log * fix * nothing * less memory * recover utils * fix * modify threshold value * skip layer_norm for test_bert * back to bert success state * add epsion * delete unnecessary compute * modify amp dtype * modify * order * delete sqrt check and fp16 --------- Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com> Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com> Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> Co-authored-by: Nxiongkun <807377414@qq.com>
-
由 Weilong Wu 提交于
-
由 mayang002 提交于
-
由 Jiabin Yang 提交于
-
由 HongyuJia 提交于
* [Tensor Operants & Prim-Relevant] Tensor supports compare operants * fix dependence of test_comp_static * fix unit test
-
- 17 3月, 2023 1 次提交
-
-
由 cyber-pioneer 提交于
* add bn vjp * fix example * fix code * fix code * fix cinn case * fix code * fix example * fix code * fix example * fix example
-
- 16 3月, 2023 2 次提交
-
-
由 Chitsing KUI 提交于
* rename flash_attn_raw to flash_attn_unpadded * fix static api * fix static return
-
由 Huang Jiyi 提交于
* remove contexts in tensor_utils * update from_blob * update from_blob * update from_blob * fix bug * fix bug
-
- 15 3月, 2023 7 次提交
-
-
由 SylarTiaNII 提交于
* add assign composite backward op * fix log msg * code style * fix comp rule * replace assign with by_pass
-
由 Jiabin Yang 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * Cxx prim custom vjp (#8) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [Prim] enable whitelist and blacklist for custom_vjp * support softmax grad * remove additional code * add test back --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> Co-authored-by: Nxiongkun <807377414@qq.com>
-
由 RedContritio 提交于
-
由 HappyHeavyRain 提交于
* test_get_kernel * add invoke signature * change reduce_max * change frobenius_norm * reset reduce_max according to composite and change reduce_all * fix the bug when Scalar(*) * fix 'scalar when support_tensor' * change code according to review * change 'keep_signature' to 'manual_signature' and add some erro info
-
由 Guoxia Wang 提交于
-
由 zhangyuqin1998 提交于
* Delete hardswish_raw op * fix ut
-
由 xiaoguoguo626807 提交于
* modify_yaml * delete default param * add output for matmul_double_grad
-
- 14 3月, 2023 1 次提交
-
-
由 ccrrong 提交于
* add split_with_num composite rule * add split_with_num composite rule * add split composite rule * update * update test * update test * delete split_with_num_grad
-