- 27 4月, 2023 30 次提交
-
-
由 gouzil 提交于
* [phi] move sequence_pool kernel to phi * mv kernels impl * fix parameter error * clean include * fix compat filename * [phi] move fluid sequence_pool_grad to phi * [phi][compat] sig rm GradVarName * [phi] fix sequence_pool out type * [phi] rm impl, add const string * [phi] fix const str * fix sequence_pooling cmake * [phi] mv sequence_pooling_test * [phi] fix grad sig * [phi] fix sequence_pool is_test error * [phi] fix sequence_pooling gpu include * [phi] mv to impl * [phi] fix SequencePoolFunctor cu include * [phi] modify out max_index int32_t * [phi] add pooltype mapping determine * [phi] fix sequence_pool_sig * [phi] fix sequence_pool_sig sum * [phi] try ci * [phi] fix max_index optional
-
由 Yuanle Liu 提交于
-
由 zhupengyang 提交于
-
由 WangZhen 提交于
[Dy2St]Get grad names when call append backward to fix high order gradient (#53250)
-
由 WangZhen 提交于
* Update slim approve list * Fix id, test=document_fix
-
由 wuhuachaocoding 提交于
-
由 houj04 提交于
-
由 gouzil 提交于
* [static op generation] triangular_solve * [phi] mv triangular_solve_grad to static_backward * [phi] fix import * [phi] mv to ops.yaml、 backward.yaml * fix forward attr * [phi] fix triangular_solve_grad args
-
由 HongyuJia 提交于
* [CINN Support 0D-Tensor] CINN supports 0D-Tensor with trick temporarily * Add unittest
-
由 yangguohao 提交于
-
由 Wang Xin 提交于
-
由 NetPunk 提交于
* support fp16 for maxout op * format code * change api * add test for static float16 * format code * formatting code * atol alignment * experiment—1 * experiment-2 * experiment-3 * format code
-
由 Sonder 提交于
* trans fused_feedward Compute function to phi * add register info * remove maxfunctor * move fused feedward to phi * remove sig file * remove fliud include * add include * add include * add sig file * add output register info * fix sig file * Update fused_feedforward_sig.cc * fix grad kernel * update output register info * fix * open fused_feedforward static build * add optional and fix code style * fix output info for fused attention * add optional param * merge
-
由 Zhang Ting 提交于
* support OD level and skip dynamic loss scaling for bf16
-
由 HongyuJia 提交于
* [Fix CppExtension Unittest] Change CUDAExtension to CppExtension if necessary * Temporarily test cpp_extension under GPU * Split mixed_extension unittest
-
由 HongyuJia 提交于
-
由 HydrogenSulfate 提交于
* add jacobian and hessian in paddle.autograd * disable unitest 'func_multi_input' for bug in high-order gradient of multiply * add dimension checks * add support for 0-D tensor * change return type from Jacobian to Hessian in hessian function * refine Jacobian _flatten function for single xs * refine support for 0-D tensor * 1. add 'func_multi_input' unitest for multiply_grad_kernel bug fixed already. 2. support non-inplace math operation via magical method overwriting. * add unitest for math operation and raise error when 0-D tensor is indexed * add ndim check on ys and xs according to is_batched, and add one unitest * refine docstring of jacobian and hessian * move paddle.incubate.autograd.Jacobian/Hessian to paddle.incubate.autograd.functional.Jacobian/Hessian * remove single_input unitest case because numerical differentiation is wrong * remove 3 unitest for numerical result(reference result) is wrong * 1. rename autodiff.py to autograd.py 2. increase TIMEOUT to 100 * cancel modification for functional Jacobian/Hessian * 1. use tuple as return type instead of list 2. refine docstring * add more unitest case to improve coverage * remove 2 unitest of Hessian for numerical result is wrong * remove 1 unitest of Hessian for numerical result is wrong * remove 1 unitest of Hessian for numerical result is wrong * change unit test to shape check * correct doc and replace incubate API to stable API in _grad
-
由 xiaoguoguo626807 提交于
* modify concat_grad add sum comp rule * modify opcompat
-
由 hua-zi 提交于
* updata Adamw.py out.backward() -> loss.backward() * Update adamw.py
-
由 wanghuancoder 提交于
-
由 JYChen 提交于
-
-
由 engineer1109 提交于
-
由 Galaxy1458 提交于
* test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop
-
由 houj04 提交于
* [XPU] remove scale_loss in parallel.py * [XPU] throw Unimplemented when using Reducer
-
由 huangjiyi 提交于
* update * fix bug
-
由 risemeup1 提交于
* update cmake3.16 to 3.18 * test * Update Dockerfile.ubuntu
-
由 superwinner1 提交于
-
由 cyberslack_lee 提交于
-
由 mengziheng 提交于
* add pad op * add_some_code * modify some code * add some code * add some code * modify some code * add some code * modify some code * Update composite_backward_api.h * modify some code * add some code * add some code * add some code
-
- 26 4月, 2023 10 次提交
-
-
由 zhouweiwei2014 提交于
-
由 zhenhailiu 提交于
* polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish * polish
-
由 huangjiyi 提交于
* rm paddle_install_dir * update test=document_fix * update test=document_fix * update * add test * update * update * update * update * update * update * update * update * update * add todo commont
-
由 mhy-666 提交于
* add scatter_nd_add comp * add scatter_nd_add prim * fix * fix * add public_python_api in TestScatterNdAddSimpleOp setup function * fix composite_backward_api.h * fix composite_backward * add test cases * fix composite_backward_api.h, unittest
-
由 Ruibiao Chen 提交于
* Fix fused_attention_op and fused_feedforward_op bugs in xpu * Fix d_x alloc errors for fused_feedforward_grad_kernel
-
由 xiaoguoguo626807 提交于
* delete prim flag for matmul_2_grad * delete prim flag for matmul_2_grad * recover multiply prune * add rules * add rules * add rules * modify approve rules * fix conflict * recover block
-
由 zqw_1997 提交于
* add test cases, test=allcase * fix test cases, test=allcase * fix test cases, test=allcase * assert_allclose, test=allcase * 1e-5 to 1e-4, test=allcase * change rtol from 1e-4 to 1e-3, test=allcase
-
由 lijialin03 提交于
* modify numel in lbfgs and add a new test case. test=develop * change param 'lr' to 'learning_rate' in lbfgs and its test * add opt LBFGS and change test
-
由 Galaxy1458 提交于
* test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop * test,test=develop
-
由 骑马小猫 提交于
-