- 20 1月, 2021 1 次提交
-
-
由 chentianyu03 提交于
* rewrite abs op * rewrite abs op and remove abs in activation * remove abs register in old codes * fix abs_grad type error * fix abs double_grad output name error * modify abs_grad, abs_grad_grad functor for windows building * format code style * fix the bug of result is nan when the divisor is zero * add missing abs attr and add abs for float16
-
- 22 12月, 2020 1 次提交
-
-
由 whs 提交于
-
- 27 9月, 2020 1 次提交
-
-
由 Jack Zhou 提交于
register log double grad kernel for cpu and cuda
-
- 25 9月, 2020 1 次提交
-
-
由 Zhong Hui 提交于
add abs support double grad for the api 2.0
-
- 06 1月, 2020 1 次提交
-
-
由 Double_V 提交于
* support elu activation double grad,test=develop * delete the code commit in .cc,test=develop * fix relu test unpass, test=develop * add elu double grad kernel and unit test * add caculate dX in elu double grad functor, test=develop * update the commit code,test=develop
-
- 03 1月, 2020 1 次提交
-
-
由 SunAhong1993 提交于
* register int/int64_t/float16 in pow/square kernel,test=develop * add abs/square/exp type,test=develop
-
- 17 9月, 2019 1 次提交
-
-
由 liym27 提交于
improve pow op according to reviews: 1. Delete unnecessary judgement statements in PowGradOpDescMaker; 2. Improve test of test_api; overload GetKernelTypeForVar add stop_gradient=True when attr(factor) is tensor Variable, change examples in API pow. test=develop,test=document_preview
-
- 15 5月, 2019 1 次提交
-
-
由 lvmengsi 提交于
* double backward sqrt * refine unittest. test=develop * refine test. test=develop * remove alpha in unittest. test=develop
-
- 13 5月, 2019 1 次提交
-
-
由 Kaipeng Deng 提交于
* add double grad for square. test=develop * formax code. test=develop * fix for grad sum. test=develop * refine shape. test=develop * refine extract. test=develop
-
- 26 4月, 2019 1 次提交
-
-
由 ceci3 提交于
backward of backward: leaky_relu
-
- 23 4月, 2019 1 次提交
-
-
由 qingqing01 提交于
Support backward of backward for Relu and add a new gradient checker by comparing theoretical and numerical Jacobian. (#16862) * Support backward of backward and a new gradient checker * Rename decorators.py to decorator_helper.py, since Python on Windows CI has decorators package. 1. Add ReluDoubleGradMaker when register relu_grad. 2. Add a new gradient checker by comparing theoretical and numerical Jacobian. Check double gradients by double_grad_check.
-
- 10 4月, 2019 1 次提交
-
-
由 Zeng Jinle 提交于
test=develop
-
- 07 11月, 2018 1 次提交
-
-
由 chengduo 提交于
* add fp16 backward support test=develop * add sum_op fp16 test * disable test_dist_save_load test=develop * add check_grad for sum * add unit test for softmax_grad fp16 test=develop * add scale_op unit test * add mul_grad_op unit test for fp16 * add cross_entropy_grad and eman_grad unit test for fp16 test=develop * fix cross_entropy unit test * add pool2d fp16 unit test * refine conv2d fp16 unit test test=develop * refine activation unit test test=develop * fix ci test=develop * follow zhihong's comment, copy from https://github.com/PaddlePaddle/Paddle/pull/12796 test=develop
-
- 17 8月, 2018 1 次提交
-
- 16 8月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "cherry picked operators changes" * "remove duplicated code" * "add constant setter" * "add get expected kernel" * "fix ci" * "add fill constant"
-
- 17 4月, 2018 1 次提交
-
-
由 dzhwinter 提交于
-
- 10 4月, 2018 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 21 3月, 2018 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 10 2月, 2018 2 次提交
- 26 12月, 2017 1 次提交
-
-
由 Luo Tao 提交于
-
- 12 12月, 2017 1 次提交
-
-
由 QI JUN 提交于
There are mainly following fixes: - take `DeviceContext` as the template parameter of math functors and OpKernel instead of `Place` - remove `eigen_device` interface in base class `DeviceContext` - remove `GetEigenDevice` interface in `ExecutionContext` and base class `DeviceContext` - remove unused `platform::EigenDeviceConverter` - rename `REGISTER_OP_GPU_KERNEL` to `REGISTER_OP_CUDA_KERNEL` - rename `USE_GPU_ONLY_OP` to `USE_CUDA_ONLY_OP`
-
- 27 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Simplize Gradient Check * Stash * Extract apply_backward_pass to backward.py Rename apply_backward_pass to append_backward_ops * Use graph API to check gradient * Fix ci * Fix CI * Fix backward for double precision * Stash * Fix CI * Fix ci * Ignore GRU test * Ignore xe op * Fix CI * Fix softmax with xe gradient The correct equation should be IG = OG * (d_softmax_with_xe()) * Fix typo * Fix merge error * Disable LRN
-
- 09 10月, 2017 1 次提交
-
-
由 Luo Tao 提交于
-
- 06 10月, 2017 2 次提交
- 29 9月, 2017 3 次提交
-
-
由 zhouxiao-coder 提交于
-
由 Abhinav Arora 提交于
-
由 Yu Yang 提交于
-
- 14 9月, 2017 3 次提交
- 13 9月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 08 8月, 2017 1 次提交
-
-
由 dongzhihong 提交于
-
- 07 8月, 2017 1 次提交
-
-
由 dongzhihong 提交于
-
- 04 8月, 2017 1 次提交
-
-
由 liaogang 提交于
-
- 02 8月, 2017 1 次提交
-
-
由 dangqingqing 提交于
-
- 31 7月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 25 7月, 2017 1 次提交
-
-
由 Yu Yang 提交于
Make implement an operator less noisy.
-