1. 03 12月, 2021 1 次提交
  2. 22 11月, 2021 1 次提交
  3. 15 10月, 2021 1 次提交
    • J
      [New Feature] Support tanh triple grad (#36225) · 808be657
      Jiabin Yang 提交于
      * native commit for triple grad of sigmod
      
      * Updated unittests files
      
      * init functional jacobian api
      
      * Updated trible_test func
      
      * Updated gradient_checker & test_script
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * fix dygraph grad to support high differential
      
      * polish API docstring
      
      * Updated gradient checker and some related files
      
      * fix double grad strip error for high differential
      
      * fix double grad strip error for high differential
      
      * Add Sigmoid triple grad tests
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * Updated triple grad teses func
      
      * Use np.random to initialize ddx
      
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * add tanh triple grad
      
      * format python code
      
      * refine code
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      808be657
  4. 13 10月, 2021 2 次提交
    • Y
      [PaddlePaddle hackathon] + ADD CELU (#36088) · d7064f04
      yujun 提交于
      * update
      
      * update
      
      * update
      
      * try make CI pass
      
      * doc typo
      
      * update doc string
      d7064f04
    • J
      [New Feature] Support triple grad in Paddle (#36187) · 2c44ee7e
      Jiabin Yang 提交于
      * native commit for triple grad of sigmod
      
      * Updated unittests files
      
      * init functional jacobian api
      
      * Updated trible_test func
      
      * Updated gradient_checker & test_script
      
      * finish test with dtype float32
      
      * add float64 test case
      
      * polish code
      
      * use atol=1e-5 with dtype float64
      
      * fix for ci
      
      * set timeout for test_jacobian
      
      * fix dygraph grad to support high differential
      
      * polish API docstring
      
      * Updated gradient checker and some related files
      
      * fix double grad strip error for high differential
      
      * fix double grad strip error for high differential
      
      * Add Sigmoid triple grad tests
      
      * fix dygraph double grad dtype error when calling for high differential senario
      
      * Updated triple grad teses func
      
      * Use np.random to initialize ddx
      
      * Updated triple_grad_check func
      
      * add todo for gradient checker and refine some comments
      
      * remove additional code
      
      * add test for warnging in backward.py
      
      * format python code
      Co-authored-by: Nveyron95 <veyron_wu@163.com>
      Co-authored-by: Nlevi131 <limaolin01@baidu.com>
      2c44ee7e
  5. 14 9月, 2021 1 次提交
  6. 13 9月, 2021 2 次提交
  7. 11 6月, 2021 1 次提交
  8. 26 5月, 2021 1 次提交
  9. 20 5月, 2021 1 次提交
  10. 18 5月, 2021 1 次提交
  11. 10 5月, 2021 1 次提交
  12. 07 5月, 2021 1 次提交
  13. 27 4月, 2021 1 次提交
  14. 15 4月, 2021 1 次提交
  15. 02 4月, 2021 1 次提交
  16. 29 3月, 2021 1 次提交
  17. 25 3月, 2021 1 次提交
  18. 23 3月, 2021 1 次提交
  19. 19 2月, 2021 1 次提交
  20. 20 1月, 2021 1 次提交
    • C
      make abs op support complex types (#30375) · 358106fc
      chentianyu03 提交于
      * rewrite abs op
      
      * rewrite abs op and remove abs in activation
      
      * remove abs register in old codes
      
      * fix abs_grad type error
      
      * fix abs double_grad output name error
      
      * modify abs_grad, abs_grad_grad functor for windows building
      
      * format code style
      
      * fix the bug of result is nan when the divisor is zero
      
      * add missing abs attr and add abs for float16
      358106fc
  21. 22 12月, 2020 1 次提交
  22. 27 9月, 2020 1 次提交
  23. 25 9月, 2020 1 次提交
  24. 06 1月, 2020 1 次提交
    • D
      support elu_op double grad (#21822) · fab4b076
      Double_V 提交于
      * support elu activation double grad,test=develop
      
      * delete the code commit in .cc,test=develop
      
      * fix relu test unpass, test=develop
      
      * add elu double grad kernel and unit test
      
      * add caculate dX in elu double grad functor, test=develop
      
      * update the commit code,test=develop
      fab4b076
  25. 03 1月, 2020 1 次提交
  26. 17 9月, 2019 1 次提交
    • L
      fix pow op, support tensor for agument factor. (#19313) · 677e7144
      liym27 提交于
      improve pow op according to reviews:
      1. Delete unnecessary judgement statements in PowGradOpDescMaker;
      2. Improve test of test_api;
      
      overload GetKernelTypeForVar
      
      add stop_gradient=True when attr(factor) is tensor Variable, change examples in API pow.
      test=develop,test=document_preview
      677e7144
  27. 15 5月, 2019 1 次提交
    • L
      Double backward sqrt (#17387) · 4ef63101
      lvmengsi 提交于
      * double backward sqrt
      
      * refine unittest. test=develop
      
      * refine test. test=develop
      
      * remove alpha in unittest. test=develop
      4ef63101
  28. 13 5月, 2019 1 次提交
    • K
      add double grad for square op (#17173) · 11d3a38f
      Kaipeng Deng 提交于
      * add double grad for square. test=develop
      
      * formax code. test=develop
      
      * fix for grad sum. test=develop
      
      * refine shape. test=develop
      
      * refine extract. test=develop
      11d3a38f
  29. 26 4月, 2019 1 次提交
  30. 23 4月, 2019 1 次提交
    • Q
      Support backward of backward for Relu and add a new gradient checker by... · c1c2633a
      qingqing01 提交于
      Support backward of backward for Relu and add a new gradient checker by comparing theoretical and numerical Jacobian. (#16862)
      
      * Support backward of backward and a new gradient checker
      * Rename decorators.py to decorator_helper.py, since Python on Windows CI has decorators package.
      
      1. Add ReluDoubleGradMaker when register relu_grad.
      2. Add a new gradient checker by comparing theoretical and numerical Jacobian.  Check double gradients by double_grad_check.
      c1c2633a
  31. 10 4月, 2019 1 次提交
  32. 07 11月, 2018 1 次提交
    • C
      Add fp16 backward support (#14202) · a9b5d42d
      chengduo 提交于
      * add fp16 backward support
      test=develop
      
      * add sum_op fp16 test
      
      * disable test_dist_save_load
      test=develop
      
      * add check_grad for sum
      
      * add unit test for softmax_grad fp16
      test=develop
      
      * add scale_op unit test
      
      * add mul_grad_op unit test for fp16
      
      * add cross_entropy_grad and eman_grad unit test for fp16
      test=develop
      
      * fix cross_entropy unit test
      
      * add pool2d fp16 unit test
      
      * refine conv2d fp16 unit test
      test=develop
      
      * refine activation unit test
      test=develop
      
      * fix ci
      test=develop
      
      * follow zhihong's comment, copy from https://github.com/PaddlePaddle/Paddle/pull/12796
      test=develop
      a9b5d42d
  33. 17 8月, 2018 1 次提交
  34. 16 8月, 2018 1 次提交
  35. 17 4月, 2018 1 次提交
  36. 10 4月, 2018 1 次提交
  37. 21 3月, 2018 1 次提交
  38. 12 2月, 2018 1 次提交