- 22 11月, 2021 1 次提交
-
-
由 zhupengyang 提交于
-
- 15 10月, 2021 1 次提交
-
-
由 Jiabin Yang 提交于
* native commit for triple grad of sigmod * Updated unittests files * init functional jacobian api * Updated trible_test func * Updated gradient_checker & test_script * finish test with dtype float32 * add float64 test case * polish code * use atol=1e-5 with dtype float64 * fix for ci * set timeout for test_jacobian * fix dygraph grad to support high differential * polish API docstring * Updated gradient checker and some related files * fix double grad strip error for high differential * fix double grad strip error for high differential * Add Sigmoid triple grad tests * fix dygraph double grad dtype error when calling for high differential senario * Updated triple grad teses func * Use np.random to initialize ddx * Updated triple_grad_check func * add todo for gradient checker and refine some comments * remove additional code * add test for warnging in backward.py * add tanh triple grad * format python code * refine code Co-authored-by: Nveyron95 <veyron_wu@163.com> Co-authored-by: Nlevi131 <limaolin01@baidu.com>
-
- 13 10月, 2021 2 次提交
-
-
由 yujun 提交于
* update * update * update * try make CI pass * doc typo * update doc string
-
由 Jiabin Yang 提交于
* native commit for triple grad of sigmod * Updated unittests files * init functional jacobian api * Updated trible_test func * Updated gradient_checker & test_script * finish test with dtype float32 * add float64 test case * polish code * use atol=1e-5 with dtype float64 * fix for ci * set timeout for test_jacobian * fix dygraph grad to support high differential * polish API docstring * Updated gradient checker and some related files * fix double grad strip error for high differential * fix double grad strip error for high differential * Add Sigmoid triple grad tests * fix dygraph double grad dtype error when calling for high differential senario * Updated triple grad teses func * Use np.random to initialize ddx * Updated triple_grad_check func * add todo for gradient checker and refine some comments * remove additional code * add test for warnging in backward.py * format python code Co-authored-by: Nveyron95 <veyron_wu@163.com> Co-authored-by: Nlevi131 <limaolin01@baidu.com>
-
- 14 9月, 2021 1 次提交
-
-
由 Yiqun Liu 提交于
Implement FunctionTraits to support two kinds of elementwise functor and remove some old codes for broadcast. (#35688)
-
- 13 9月, 2021 2 次提交
- 11 6月, 2021 1 次提交
-
-
由 ronnywang 提交于
-
- 26 5月, 2021 1 次提交
-
-
由 Zhanlue Yang 提交于
Sigmoid: Out = Sigmoid(X) SigmoidGrad: DX = DOut*(1-Out)*Out [This Patch] Out DOut -> SigmoidGradGrad -> DOutNew DDX DDOut DDOut = (1-Out)*Out*DDX DOutNew = (1-2*Out)*DOut*DDX
-
- 20 5月, 2021 1 次提交
-
-
由 limingshu 提交于
-
- 18 5月, 2021 1 次提交
-
-
由 wuhuanzhou 提交于
-
- 10 5月, 2021 1 次提交
-
-
由 Zhang Zheng 提交于
-
- 07 5月, 2021 1 次提交
-
-
由 Zhang Zheng 提交于
-
- 27 4月, 2021 1 次提交
-
-
由 Zhang Zheng 提交于
-
- 15 4月, 2021 1 次提交
-
-
由 Jiabin Yang 提交于
* add IsInitialized * rm additional log and add tanh double grad * rename is_initialized
-
- 02 4月, 2021 1 次提交
-
-
由 niuliling123 提交于
* add leaky_relu forward and backward in activation_op.cu
-
- 29 3月, 2021 1 次提交
-
-
由 niuliling123 提交于
-
- 25 3月, 2021 1 次提交
-
-
由 niuliling123 提交于
-
- 23 3月, 2021 1 次提交
-
-
由 niuliling123 提交于
* add relu forward kernel and backward kernel
-
- 19 2月, 2021 1 次提交
-
-
由 Wojciech Uss 提交于
* Modify relu native implementation * fix GPU performance
-
- 20 1月, 2021 1 次提交
-
-
由 chentianyu03 提交于
* rewrite abs op * rewrite abs op and remove abs in activation * remove abs register in old codes * fix abs_grad type error * fix abs double_grad output name error * modify abs_grad, abs_grad_grad functor for windows building * format code style * fix the bug of result is nan when the divisor is zero * add missing abs attr and add abs for float16
-
- 22 12月, 2020 1 次提交
-
-
由 whs 提交于
-
- 27 9月, 2020 1 次提交
-
-
由 Jack Zhou 提交于
register log double grad kernel for cpu and cuda
-
- 25 9月, 2020 1 次提交
-
-
由 Zhong Hui 提交于
add abs support double grad for the api 2.0
-
- 06 1月, 2020 1 次提交
-
-
由 Double_V 提交于
* support elu activation double grad,test=develop * delete the code commit in .cc,test=develop * fix relu test unpass, test=develop * add elu double grad kernel and unit test * add caculate dX in elu double grad functor, test=develop * update the commit code,test=develop
-
- 03 1月, 2020 1 次提交
-
-
由 SunAhong1993 提交于
* register int/int64_t/float16 in pow/square kernel,test=develop * add abs/square/exp type,test=develop
-
- 17 9月, 2019 1 次提交
-
-
由 liym27 提交于
improve pow op according to reviews: 1. Delete unnecessary judgement statements in PowGradOpDescMaker; 2. Improve test of test_api; overload GetKernelTypeForVar add stop_gradient=True when attr(factor) is tensor Variable, change examples in API pow. test=develop,test=document_preview
-
- 15 5月, 2019 1 次提交
-
-
由 lvmengsi 提交于
* double backward sqrt * refine unittest. test=develop * refine test. test=develop * remove alpha in unittest. test=develop
-
- 13 5月, 2019 1 次提交
-
-
由 Kaipeng Deng 提交于
* add double grad for square. test=develop * formax code. test=develop * fix for grad sum. test=develop * refine shape. test=develop * refine extract. test=develop
-
- 26 4月, 2019 1 次提交
-
-
由 ceci3 提交于
backward of backward: leaky_relu
-
- 23 4月, 2019 1 次提交
-
-
由 qingqing01 提交于
Support backward of backward for Relu and add a new gradient checker by comparing theoretical and numerical Jacobian. (#16862) * Support backward of backward and a new gradient checker * Rename decorators.py to decorator_helper.py, since Python on Windows CI has decorators package. 1. Add ReluDoubleGradMaker when register relu_grad. 2. Add a new gradient checker by comparing theoretical and numerical Jacobian. Check double gradients by double_grad_check.
-
- 10 4月, 2019 1 次提交
-
-
由 Zeng Jinle 提交于
test=develop
-
- 07 11月, 2018 1 次提交
-
-
由 chengduo 提交于
* add fp16 backward support test=develop * add sum_op fp16 test * disable test_dist_save_load test=develop * add check_grad for sum * add unit test for softmax_grad fp16 test=develop * add scale_op unit test * add mul_grad_op unit test for fp16 * add cross_entropy_grad and eman_grad unit test for fp16 test=develop * fix cross_entropy unit test * add pool2d fp16 unit test * refine conv2d fp16 unit test test=develop * refine activation unit test test=develop * fix ci test=develop * follow zhihong's comment, copy from https://github.com/PaddlePaddle/Paddle/pull/12796 test=develop
-
- 17 8月, 2018 1 次提交
-
- 16 8月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "cherry picked operators changes" * "remove duplicated code" * "add constant setter" * "add get expected kernel" * "fix ci" * "add fill constant"
-
- 17 4月, 2018 1 次提交
-
-
由 dzhwinter 提交于
-
- 10 4月, 2018 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 21 3月, 2018 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 10 2月, 2018 1 次提交
-
-
由 Yi Wang 提交于
-