- 18 1月, 2021 4 次提交
-
-
由 hutuxian 提交于
-
由 QingshuChen 提交于
-
由 taixiurong 提交于
-
由 JZ-LIANG 提交于
-
- 17 1月, 2021 1 次提交
-
-
由 guofei 提交于
* Modify the calculation logic of LambOptimizer
-
- 15 1月, 2021 5 次提交
-
-
由 Zhang Ting 提交于
-
由 Yang Zhang 提交于
built-in `rsqrt` is shadowed
-
由 石晓伟 提交于
-
由 Wojciech Uss 提交于
-
由 wawltor 提交于
* fix the rnn mask memory bug for out of read * update the code for the rnn
-
- 14 1月, 2021 2 次提交
-
-
由 taixiurong 提交于
-
由 ShenLiang 提交于
-
- 13 1月, 2021 1 次提交
-
-
由 GaoWei8 提交于
* softmax backward optimize
-
- 12 1月, 2021 8 次提交
-
-
由 JZ-LIANG 提交于
-
由 lidanqing 提交于
-
由 tangwei12 提交于
* rename sendrecv.proto to namespace paddle.distributed * split ps with distributed
-
由 Double_V 提交于
* fix elugradgrad test fail and error message opt * fix unitest,test=develop * Update prroi_pool_op.h fix error message * opt message,test=develop * fix ci fail,test=develop
-
由 Zhen Wang 提交于
* Fix the accuracy problem of allclose op when using float64 data type in static mode. * Format the code style.
-
由 yaoxuefeng 提交于
-
由 furnace 提交于
-
由 石晓伟 提交于
-
- 11 1月, 2021 9 次提交
-
-
由 Chengmo 提交于
* fix cmake Co-authored-by: NseiriosPlus <tangwei12@baidu.com>
-
由 石晓伟 提交于
-
由 liym27 提交于
Support vector<double> as type of op attribute and op set_value suppport vector<double> as value (#30126)
-
由 wangchaochaohu 提交于
-
由 AshburnLee 提交于
-
由 chentianyu03 提交于
* type promotion for grad * add type promotion for div op
-
由 liym27 提交于
-
由 WeiXin 提交于
* modify error message based on comments * edit code according to review. * Correct spelling according to review.
-
由 wawltor 提交于
add the op error message for the matmul xpu
-
- 10 1月, 2021 2 次提交
-
-
由 GaoWei8 提交于
* optimize softmax forward
-
由 wangchaochaohu 提交于
reduce the occupied size of memory for the fused pattern of elementwise_add Op and activation Op(relu Op for example) (#29885)
-
- 09 1月, 2021 2 次提交
-
-
由 zhang wenhui 提交于
-
由 Jacek Czaja 提交于
* - Added UT for testing elementwise_mul caching * lint fixes
-
- 08 1月, 2021 6 次提交
-
-
由 Zhen Wang 提交于
* add cast ops before and after unsupported fp16 ops. * Keep partial net in FP32 pattern. * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode. * Add fp16 support for adam op. * add multi precision attr for adam. * Fix the bug of test_multi_precision_fp16_train UT. * Code format for CI. * Fix the redefine error about MPTypeTrait on windows. * fix bugs of the _create_accumulators func in Momentum. * fix bug when inserting post cast op. * Add the update_loss_scaling op in allow_set of UnusedVarCheck. * Update for ci coverage. * Add some doc for OptimizerWithMixedPrecision. * Fix the code style. * Imporve the doc of `amp_init`. * Change for fp16 testing if users have the infer program defined in separate way.
-
由 Leo Chen 提交于
-
由 Leo Chen 提交于
* fix dtype of ungenerated grad var * update ut * refine code * set default dtype * fix could_use_cudnn bug * remove debug code * re-implement * fix bug
-
由 Wilber 提交于
-
由 ruri 提交于
-
由 Chengmo 提交于
* add tensor table
-