1. 22 8月, 2019 1 次提交
    • L
      Enhance OpTest to check the consistency of operators when using and not using inplace (#19101) · a9d5fc51
      Leo Chen 提交于
      * add pybind interface to get all inplace ops, test=develop
      
      * enhance OpTest to check whether the consistency of operator when using and not using inplace, test=develop
      
      * handle corner cases in op_test, test=develop
      
      * support outputs without tensor holder_, like XShape in reshape_op, test=develop
      
      * fix bug, some op has GradOpMaker, but actually no grad_op in OpInfoMap, test=develop
      
      * use reshape_grad instead of reshape in FlattenGradOp, test=develop
      
      * fix error debug dims info for variables like XShape, test=develop
      
      * change computational order in sum_op to relieve computation difference using inplace, test=develop
      
      * add inplace_atol to check group_norm, and skip inplace_grad for mkldnn, test=develop
      
      * follow sneaxiy's comments, test=develop
      
      * remove unused DefaultGradOpDescMaker in mkldnn op, test=develop
      a9d5fc51
  2. 11 7月, 2019 1 次提交
    • Z
      Feature/buffer_shared_inplace (#17911) · d3003a16
      Zeng Jinle 提交于
      * feature/buffer_shared_inplace, test=develop
      
      * refine code, test=develop
      
      * fix elementwise_add op cpu inplace and sum inplace bug, test=develop
      
      * add unittest and debug log, test=develop
      
      * fix parallel_executor scope bug, polish code, test=develop
      
      * fix sum op, activation op, single_in_place_inference bug, test=develop
      
      * remove kLocalExecScopeName, test=develop
      
      * fix unittest,test=develop
      
      * fix out_var first version bug, test=develop
      
      * follow comments,test=develop
      d3003a16
  3. 16 6月, 2019 1 次提交
  4. 08 5月, 2019 1 次提交
  5. 07 5月, 2019 1 次提交
  6. 11 12月, 2018 1 次提交
  7. 07 11月, 2018 1 次提交
    • C
      Add fp16 backward support (#14202) · a9b5d42d
      chengduo 提交于
      * add fp16 backward support
      test=develop
      
      * add sum_op fp16 test
      
      * disable test_dist_save_load
      test=develop
      
      * add check_grad for sum
      
      * add unit test for softmax_grad fp16
      test=develop
      
      * add scale_op unit test
      
      * add mul_grad_op unit test for fp16
      
      * add cross_entropy_grad and eman_grad unit test for fp16
      test=develop
      
      * fix cross_entropy unit test
      
      * add pool2d fp16 unit test
      
      * refine conv2d fp16 unit test
      test=develop
      
      * refine activation unit test
      test=develop
      
      * fix ci
      test=develop
      
      * follow zhihong's comment, copy from https://github.com/PaddlePaddle/Paddle/pull/12796
      test=develop
      a9b5d42d
  8. 17 8月, 2018 1 次提交
  9. 16 8月, 2018 1 次提交
  10. 12 2月, 2018 1 次提交
  11. 10 2月, 2018 2 次提交
  12. 12 12月, 2017 1 次提交
    • Q
      Refine device context (#6433) · 61ec0b95
      QI JUN 提交于
      There are mainly following fixes:
      
      - take `DeviceContext` as the template parameter of math functors and OpKernel instead of `Place`
      - remove `eigen_device` interface in base class  `DeviceContext`
      - remove `GetEigenDevice` interface in `ExecutionContext` and base class `DeviceContext`
      - remove unused `platform::EigenDeviceConverter`
      - rename `REGISTER_OP_GPU_KERNEL` to `REGISTER_OP_CUDA_KERNEL`
      - rename `USE_GPU_ONLY_OP` to `USE_CUDA_ONLY_OP`
      61ec0b95
  13. 23 11月, 2017 1 次提交
  14. 27 10月, 2017 1 次提交
    • Y
      Gradient check use graph (#5027) · be00b0c4
      Yu Yang 提交于
      * Simplize Gradient Check
      
      * Stash
      
      * Extract apply_backward_pass to backward.py
      
      Rename apply_backward_pass to append_backward_ops
      
      * Use graph API to check gradient
      
      * Fix ci
      
      * Fix CI
      
      * Fix backward for double precision
      
      * Stash
      
      * Fix CI
      
      * Fix ci
      
      * Ignore GRU test
      
      * Ignore xe op
      
      * Fix CI
      
      * Fix softmax with xe gradient
      
      The correct equation should be IG = OG * (d_softmax_with_xe())
      
      * Fix typo
      
      * Fix merge error
      
      * Disable LRN
      be00b0c4
  15. 03 10月, 2017 2 次提交
  16. 05 9月, 2017 1 次提交
  17. 04 9月, 2017 1 次提交
  18. 26 8月, 2017 1 次提交
  19. 25 8月, 2017 1 次提交
  20. 07 8月, 2017 1 次提交
  21. 04 8月, 2017 1 次提交
  22. 31 7月, 2017 1 次提交
  23. 25 7月, 2017 1 次提交
  24. 19 7月, 2017 1 次提交