1. 09 10月, 2019 1 次提交
  2. 17 9月, 2019 1 次提交
    • L
      fix pow op, support tensor for agument factor. (#19313) · 677e7144
      liym27 提交于
      improve pow op according to reviews:
      1. Delete unnecessary judgement statements in PowGradOpDescMaker;
      2. Improve test of test_api;
      
      overload GetKernelTypeForVar
      
      add stop_gradient=True when attr(factor) is tensor Variable, change examples in API pow.
      test=develop,test=document_preview
      677e7144
  3. 12 8月, 2019 1 次提交
    • H
      Add hard swish op (new op) (#19001) · 20f18930
      huangjun12 提交于
      * add hard_swish activation op (new op)
      test=develop
      
      * remove redundancy files
      
      * modify document content of HardSwish OP
      
      * add API test in test_layers.py
      
      * add dynamic_graph for test_hard_swish
      20f18930
  4. 26 7月, 2019 1 次提交
  5. 22 7月, 2019 1 次提交
  6. 19 7月, 2019 1 次提交
  7. 28 3月, 2019 1 次提交
  8. 10 3月, 2019 1 次提交
  9. 04 3月, 2019 1 次提交
    • D
      polish cudnn related code and fix bug. (#15164) · 4449e855
      dzhwinter 提交于
      * staged.
      
      * polish code
      
      * polish code. test=develop
      
      * polish code. test=develop
      
      * api change. test=develop
      
      * fix default value. test=develop
      
      * fix default value. test=develop
      4449e855
  10. 27 2月, 2019 1 次提交
    • D
      polish cudnn related code and fix bug. (#15164) · 225c11a9
      dzhwinter 提交于
      * staged.
      
      * polish code
      
      * polish code. test=develop
      
      * polish code. test=develop
      
      * api change. test=develop
      
      * fix default value. test=develop
      
      * fix default value. test=develop
      225c11a9
  11. 27 11月, 2018 1 次提交
  12. 07 11月, 2018 1 次提交
    • C
      Add fp16 backward support (#14202) · a9b5d42d
      chengduo 提交于
      * add fp16 backward support
      test=develop
      
      * add sum_op fp16 test
      
      * disable test_dist_save_load
      test=develop
      
      * add check_grad for sum
      
      * add unit test for softmax_grad fp16
      test=develop
      
      * add scale_op unit test
      
      * add mul_grad_op unit test for fp16
      
      * add cross_entropy_grad and eman_grad unit test for fp16
      test=develop
      
      * fix cross_entropy unit test
      
      * add pool2d fp16 unit test
      
      * refine conv2d fp16 unit test
      test=develop
      
      * refine activation unit test
      test=develop
      
      * fix ci
      test=develop
      
      * follow zhihong's comment, copy from https://github.com/PaddlePaddle/Paddle/pull/12796
      test=develop
      a9b5d42d
  13. 15 8月, 2018 1 次提交
  14. 08 8月, 2018 1 次提交
    • C
      Feature/op fusion (#12240) · 7c8b69c7
      chengduo 提交于
      * Add Preface
      
      * Add demo code
      
      * Save file
      
      * Refine code
      
      * seems can work
      
      * use elementwise strategy
      
      * Use ElementwiseComputeEx
      
      * Add comments
      
      * extract functions from operator
      
      * Refine code
      
      * Follow comment
      
      * code refine
      
      * follow comments
      
      * follow comments
      7c8b69c7
  15. 26 7月, 2018 2 次提交
  16. 17 4月, 2018 1 次提交
  17. 16 4月, 2018 1 次提交
  18. 10 4月, 2018 1 次提交
  19. 04 4月, 2018 1 次提交
  20. 29 3月, 2018 1 次提交
  21. 28 3月, 2018 1 次提交
  22. 23 3月, 2018 3 次提交
  23. 21 3月, 2018 2 次提交
  24. 24 2月, 2018 1 次提交
  25. 13 2月, 2018 1 次提交
    • X
      Run Python OP tests in a single Python process to improve test time. (#8362) · cde6241a
      Xin Pan 提交于
      Currently, our tests run with 2 GPUs, the init time is absurdly long:
      about 4s for each process.  Currently, we run each OP test on
      different processes. This PR:
      
      1. create cmake function py_test_modules which will generate the
      Makefile that runs a list of Python unittest module in a single Python
      process.
      
      2. move all "python unittest compatible" (e.g., used the unittest
      package, not just a regular python file). from fluid/tests to
      fluid/tests/unittests.
      
      3. cmake now will run all OP tests in fluid/tests/unittests in a
      single process, except the time-consuming tests, they are separated
      into different processes to utilize parallelism. Please make sure to
      use the unittest package if you put the python test file in
      fluid/tests/unittests
      
      4. remove all exit(0) from fluid/tests/unittests/*.py, exit(0) is used
      to disable unittest, we can not do it when running all tests in a
      single process since it will terminate the process without running the
      other tests. Instead, the test is disabled in
      fluid/tests/unittests/CMakeLists.txt. FIXME is added for each disabled
      item. Please disable the unittest from
      fluid/tests/unittests/CMakeLists.txt, instead of adding exit(0) to the
      Python file, for all Python file in fluid/tests/unittests/.
      
      5. add an option WITH_FAST_BUNDLE_TEST. When OFF, will run the unit
      tests in separate process so that they can be tested individually.
      cde6241a
  26. 12 2月, 2018 1 次提交
  27. 29 1月, 2018 1 次提交
  28. 21 1月, 2018 1 次提交
    • D
      "fix decode bug" (#7711) · e983cc90
      dzhwinter 提交于
      * "fix decode bug"
      
      * "follow commnet"
      
      * "fix error"
      
      * "fix hook bug"
      
      * fix based comment
      
      * fix copyright
      
      * fix based on comment
      e983cc90
  29. 15 1月, 2018 1 次提交
    • D
      Feature/hooks (#7513) · b9b75377
      dzhwinter 提交于
      * add copyright hook
      
      * add copyright hook
      
      * refine copyright hook
      
      * "test copyright hook"
      
      * fix check style
      
      * fix ci
      b9b75377
  30. 26 12月, 2017 1 次提交
  31. 07 12月, 2017 1 次提交
  32. 26 11月, 2017 1 次提交
  33. 14 11月, 2017 1 次提交
  34. 27 10月, 2017 1 次提交
    • Y
      Gradient check use graph (#5027) · be00b0c4
      Yu Yang 提交于
      * Simplize Gradient Check
      
      * Stash
      
      * Extract apply_backward_pass to backward.py
      
      Rename apply_backward_pass to append_backward_ops
      
      * Use graph API to check gradient
      
      * Fix ci
      
      * Fix CI
      
      * Fix backward for double precision
      
      * Stash
      
      * Fix CI
      
      * Fix ci
      
      * Ignore GRU test
      
      * Ignore xe op
      
      * Fix CI
      
      * Fix softmax with xe gradient
      
      The correct equation should be IG = OG * (d_softmax_with_xe())
      
      * Fix typo
      
      * Fix merge error
      
      * Disable LRN
      be00b0c4
  35. 24 10月, 2017 1 次提交
  36. 13 10月, 2017 1 次提交