- 22 3月, 2023 7 次提交
-
-
由 Sławomir Siwek 提交于
* extract common methods to reuse * add header for transpose ops * fused_transpose * Split big function * transpose2 tests * fused_transpose * Apply extra attributes * add pbtxt file * update pbtxt * Merge develop * add more strict op compats * code style * remove mkldnn_data_type * unify SetOutMemDescWithReshape2FuseSupport * adjust quantize-dequantize for transpose * remove appendact * transpose2 quantization * fix int8 tests * adjust transpose_op to current develop * delete fusion code from transpose_kernel * add fused transpose to NHWC unittest * change order
-
由 Wang Xin 提交于
* add autogen code for index_add op * bug fixed
-
由 RedContritio 提交于
* supoort auto generate p_norm * fix bug in backward
-
由 RedContritio 提交于
* support auto generate for dirichlet * use uppercase in args * use op_compat for name mapping
-
由 wangxiaoning 提交于
* max comp * fix * add test * fix * fix * fix * fix * fix test * fix api
-
由 sneaxiy 提交于
* add fused_linear_param_grad_add_kernel * fix compile error * remove flag * fix ci compile error * fix ci compile error * revert pylayer revision * fix ci ut * improve performance
-
由 Ruibiao Chen 提交于
-
- 21 3月, 2023 3 次提交
-
-
由 iSerendipity 提交于
* move DataType from paddle::experimental to phi * convert namespace * convert namespace * convert namespace * clarify namespace * convert more datatype * Revert "convert more datatype" This reverts commit 083b462959e6a22d4d8767707b628b95b396642e. * convert more in auto_code_generator * fix conflicts for XPU * fix namespace conflicts * fix errors * Revert "fix errors" This reverts commit f9d9958b54ee32141112274c8a5c3c381ab0f876. * fix errors * fix formatting
-
由 Zero Rains 提交于
* fix divide zero bug for softmax_with_cross_entropy * change the single test way * fix the null pointer in graph_khop_sampler * fix the case5
-
由 Difer 提交于
-
- 20 3月, 2023 3 次提交
-
-
由 xiaoguoguo626807 提交于
* Add flatten composite rule * get the right xshape and pass func test * add cinn unit test * Remove cinn test, wait for it to be added after repair * add comp test to test_flatten_contiguous_range_op.py * remove func test on composite_ops * Add comments to maybe_wrap_dim func * remove commented code * fix the problem with 0D tensor case * add flatten split rule comment * fix syntax issues * block flatten on resnet_prim_cinn * init change * tmp commit * add layer_norm InferMeta check * cast type modify * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * recover * big tol * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * Cxx prim custom vjp (#8) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [Prim] enable whitelist and blacklist for custom_vjp * debug log * clear log * fix * nothing * less memory * recover utils * fix * modify threshold value * skip layer_norm for test_bert * back to bert success state * add epsion * delete unnecessary compute * modify amp dtype * modify * order * delete sqrt check and fp16 --------- Co-authored-by: Nxuyongsheng <xuyongsheng@baidu.com> Co-authored-by: Nxysheng-baidu <121540080+xysheng-baidu@users.noreply.github.com> Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> Co-authored-by: Nxiongkun <807377414@qq.com>
-
由 Zhang Na 提交于
-
由 limingshu 提交于
* optimization for fused linear op * fix code format * optimization for linear fused forward * merge with develop * fix bugs for gemm_ephilog * package of cublaslt ephilogue type with enmu * final fix before code reviewing * fix missed fusedType typo * fix code according to review suggestions * fix windows ci error * change location of MatmulPlanner * add some changes for compiler error fix ---------
-
- 19 3月, 2023 1 次提交
-
-
由 Charles-hit 提交于
-
- 18 3月, 2023 1 次提交
-
-
由 Leo Chen 提交于
-
- 17 3月, 2023 2 次提交
-
-
由 Leo Chen 提交于
* support fetch empty tensor on CPUPlace * fix the shape in unittest of empty output
-
由 cyber-pioneer 提交于
* add bn vjp * fix example * fix code * fix code * fix cinn case * fix code * fix example * fix code * fix example * fix example
-
- 16 3月, 2023 3 次提交
-
-
由 Huang Jiyi 提交于
* remove contexts in tensor_utils * update from_blob * update from_blob * update from_blob * fix bug * fix bug
-
由 JZ-LIANG 提交于
* update env setting * update pass logic * dist op support bf16 * backward cast update * update setting * update backward * revert amp pass * update fp16 backward logic * register c_embedding bf16 * revert engine * add unitest * add unitest * update unitest * update cmake * update math * update math.py * update unitest * update unitest * revise unitest * revise unitest * update unitest * update unitest * update unitest
-
由 Shijie 提交于
* add sync * Fix nccl_op_test
-
- 15 3月, 2023 5 次提交
-
-
由 SylarTiaNII 提交于
* add assign composite backward op * fix log msg * code style * fix comp rule * replace assign with by_pass
-
由 Jiabin Yang 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * Cxx prim custom vjp (#8) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [Prim] enable whitelist and blacklist for custom_vjp * support softmax grad * remove additional code * add test back --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> Co-authored-by: Nxiongkun <807377414@qq.com>
-
由 iSerendipity 提交于
* Revert "Revert "【Hackathon No.67】remove operator.h in blas.h (#50989)" (#51467)" This reverts commit b9d91531. * remove cout * add header * fix missing header * fix refer fluid error * fix missing header * 更新 repeat_interleave_grad_kernel_impl.h Change to phi style datatype. * 更新 repeat_interleave_grad_kernel_impl.h Fix missing header * datatype fluid -> phi * paddle::experimental -> phi * fix reference error * fix reference error * fix reference error * fix errors * fix missing FLAGS * fix missing headers * fix missing headers * fix missing headers * fix missing headers * fix missing header * fix missing header * fix errors
-
由 RedContritio 提交于
-
由 HappyHeavyRain 提交于
* test_get_kernel * add invoke signature * change reduce_max * change frobenius_norm * reset reduce_max according to composite and change reduce_all * fix the bug when Scalar(*) * fix 'scalar when support_tensor' * change code according to review * change 'keep_signature' to 'manual_signature' and add some erro info
-
- 14 3月, 2023 5 次提交
-
-
由 engineer1109 提交于
-
由 ccrrong 提交于
* add split_with_num composite rule * add split_with_num composite rule * add split composite rule * update * update test * update test * delete split_with_num_grad
-
由 gouzil 提交于
-
由 cxxly 提交于
-
由 cxxly 提交于
-
- 13 3月, 2023 6 次提交
-
-
由 duanyanhui 提交于
-
由 heyanru 提交于
* refresh * compat * register * testop * fix * fix * fox * cast * cast * fix * type * fix * out * cast * fix * fix * fix * broad * broad * broad * fix * fix * fix * fix * fix * broad * broad * numel * fix * fix * fix * fix * cinn * fix * fix * fix * fix
-
由 shentanyue 提交于
[Lite] Change the source code integration of Paddle Lite to the compilation library integration (#51405)
-
由 YuanRisheng 提交于
* remove transpose infershape * fix ci bugs * fix ci bugs * delete transpose infershape * fix ci bugs * fix ci bugs
-
由 iSerendipity 提交于
* remove fused_matmul from list * add infermeta for fused matmul
-
由 Sławomir Siwek 提交于
* mkldnn->onednn * fused softplus op + kernel * remove extra attributes * add missing handler * change var name
-
- 10 3月, 2023 4 次提交
-
-
由 HappyHeavyRain 提交于
* Add function node in phi_kernel for MKLDNN * fix the bug in 'BuildInferVarKernelContext' * add infer_varkernel_utils.cc * fix the bug:the first two parametes of 'BuildInferVarKernelContext' can't be template variable * change the code according to first review * change the code according to first review * change the mode of paddle_build.sh * change 'infer_var_kernel_fn_' to 'get_kerneltype_forvar_fn_' * add the error information * fix NotFound infomation warning * fix NotFound infomation warning * fix NotFound infomation warning
-
由 Charles-hit 提交于
* support variable parameter in optest * add error msg for use tensor attr in static code gen * fix static code gen * fix prim op test * modify comment * fix op test * fix ci * remove code
-
由 Aurelius84 提交于
* [framework]support pass stop_gradient in static mode * fix control_flow op stop_gradient
-
由 YuanRisheng 提交于
This reverts commit 3f4917f6.
-