- 20 6月, 2019 1 次提交
-
-
由 qingqing01 提交于
* Update backward appending stragety to support double backward and fix some bug. (#18104) * Update backward.py: - If there is no input grad var in all outputs of previous ops, do not append this op into graph. - Only apply this stragety when double backward. * Update some double backward op. * Update sum_op to judge whether a tensor is empty by numel or IsInitialized().
-
- 08 5月, 2019 1 次提交
-
-
由 Yiqun Liu 提交于
* Optimize the cuda implementation of sum_op, which add two lod_tensors inplace. test=develop * Use eigen to add to tensors. test=develop
-
- 07 5月, 2019 1 次提交
-
-
由 zhaoyuchen2018 提交于
* optimize sum op fuse multi eigen kernel calls into one cuda kernel. refine code test=develop. Signed-off-by: Nzhaoyuchen <zhaoyuchen01@baidu.com> * Refine code. test=develop Signed-off-by: Nzhaoyuchen <zhaoyuchen01@baidu.com> * Refine code according to comments. test=develop * refine code delete sum_op_gpu.h test=develop * Fix test error. test=develop Signed-off-by: Nzhaoyuchen <zhaoyuchen01@baidu.com> * refine code in format. test=develop. * refine code test=develop Signed-off-by: Nzhaoyuchen <zhaoyuchen01@baidu.com> * refine code test=develop Signed-off-by: Nzhaoyuchen <zhaoyuchen01@baidu.com>
-
- 11 12月, 2018 1 次提交
-
-
由 Yu Yang 提交于
The macro should be defined by compiler rather than by source. test=develop
-
- 07 11月, 2018 1 次提交
-
-
由 chengduo 提交于
* add fp16 backward support test=develop * add sum_op fp16 test * disable test_dist_save_load test=develop * add check_grad for sum * add unit test for softmax_grad fp16 test=develop * add scale_op unit test * add mul_grad_op unit test for fp16 * add cross_entropy_grad and eman_grad unit test for fp16 test=develop * fix cross_entropy unit test * add pool2d fp16 unit test * refine conv2d fp16 unit test test=develop * refine activation unit test test=develop * fix ci test=develop * follow zhihong's comment, copy from https://github.com/PaddlePaddle/Paddle/pull/12796 test=develop
-
- 17 8月, 2018 1 次提交
-
- 16 8月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "cherry picked operators changes" * "remove duplicated code" * "add constant setter" * "add get expected kernel" * "fix ci" * "add fill constant"
-
- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 10 2月, 2018 2 次提交
- 12 12月, 2017 1 次提交
-
-
由 QI JUN 提交于
There are mainly following fixes: - take `DeviceContext` as the template parameter of math functors and OpKernel instead of `Place` - remove `eigen_device` interface in base class `DeviceContext` - remove `GetEigenDevice` interface in `ExecutionContext` and base class `DeviceContext` - remove unused `platform::EigenDeviceConverter` - rename `REGISTER_OP_GPU_KERNEL` to `REGISTER_OP_CUDA_KERNEL` - rename `USE_GPU_ONLY_OP` to `USE_CUDA_ONLY_OP`
-
- 23 11月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Support int64 for sum op * Refine code
-
- 27 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Simplize Gradient Check * Stash * Extract apply_backward_pass to backward.py Rename apply_backward_pass to append_backward_ops * Use graph API to check gradient * Fix ci * Fix CI * Fix backward for double precision * Stash * Fix CI * Fix ci * Ignore GRU test * Ignore xe op * Fix CI * Fix softmax with xe gradient The correct equation should be IG = OG * (d_softmax_with_xe()) * Fix typo * Fix merge error * Disable LRN
-
- 03 10月, 2017 2 次提交
- 05 9月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 04 9月, 2017 1 次提交
-
-
由 Luo Tao 提交于
-
- 26 8月, 2017 1 次提交
-
-
由 zchen0211 提交于
-
- 25 8月, 2017 1 次提交
-
-
由 zchen0211 提交于
-
- 07 8月, 2017 1 次提交
-
-
由 dongzhihong 提交于
-
- 04 8月, 2017 1 次提交
-
-
由 liaogang 提交于
-
- 31 7月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 25 7月, 2017 1 次提交
-
-
由 Yu Yang 提交于
Make implement an operator less noisy.
-
- 19 7月, 2017 1 次提交
-
-
由 Qiao Longfei 提交于
* a simplest SGD op
-