- 18 11月, 2020 9 次提交
-
-
由 JZ-LIANG 提交于
* add lars to fleet meta optimizer * add lamb to proto * add lamb to fleet meta optimizer * fixed syntax bug * fixed syntax bug * fixed syntax error in lamb, add config setter of lamb in distributed_strategy * trigger unitest to rerun * add new unitest func for lamb * revise unitest for lars and lamb * revise dgc meta unitest * revise lars document in distribute_strategy * revise lars lamb document in distributed_strategy.py * revise lars lamb document in distributed_strategy.py * add weight decay exclude logic to lars * restore optimzier.py * restore optimizer.py as develop except lars * add epsilon and exclude fn to distributed_sttrategy * add lars epsilon * revise unitest for fleet lars and lamb * revise lars lamb unitest for CI coverage * revise lars argument api * revise lars argument api * revise lars argument api * revise api doc of lars * fix op role * add sharding save and add_sync_comm_for_test function * add comm_analyse to utlis * revise sharding_utils * add sharding saving unittest * revise sharding utils for unittest
-
由 xiaoting 提交于
* update interpolate, test=develop * fix coverage, test=develop
-
由 YUNSHEN XIE 提交于
-
由 LielinJiang 提交于
* add arg for vgg
-
由 LielinJiang 提交于
-
由 Shibo Tao 提交于
add two paddle-2.0 apis: paddle.static.io.save_inference_model and paddle.static.io.load_inference_model (#28606) * add two apis: paddle.static.io.save_inference_model and paddle.static.io.load_inference_mode, which are campatible with paddle.fluid.io.save_inference_model and paddle.fluid.io.load_inference_model respectively. * add unittest for new save_inference_model and load_inference_model. test=develop * enhance doc. test=develop * add paddle.enable_static() to test_inference_model_io.py. test=develop
-
由 Chen Weihang 提交于
* add debuging code * change seed & add debug message
-
由 Bai Yifan 提交于
* support user-defined quant and preprocess
-
由 Leo Chen 提交于
* add matmtl_v2 to amp list * support dygraph
-
- 17 11月, 2020 10 次提交
-
-
由 Huihuang Zheng 提交于
Fix a bug that used in PaddleGAN model which used `isinstance` in a for loop
-
由 lilong12 提交于
-
由 Jacek Czaja 提交于
-
由 LielinJiang 提交于
-
由 wangchaochaohu 提交于
-
由 littletomatodonkey 提交于
* adapt pad const * fix comment and rm fluid import * rm stdout * fix note
-
由 Double_V 提交于
* fix pool exclusive and delete disable_static, test=develop * fix pool1d exclusive, test=develop
-
由 Aurelius84 提交于
* fix output dtyp inconsistent with input * refine code
-
由 wangchaochaohu 提交于
-
由 zhupengyang 提交于
-
- 16 11月, 2020 20 次提交
-
-
由 Chen Weihang 提交于
-
由 Zhou Wei 提交于
-
由 cc 提交于
-
由 wawltor 提交于
fix the gradient bug for the topk v2
-
由 wangchaochaohu 提交于
-
由 Kaipeng Deng 提交于
* add WeightedRandomSampler. test=develop
-
由 joanna.wozna.intel 提交于
* Add checkpoint to quantize * Change bfloat16 option
-
由 lidanqing 提交于
* add mkldnn inplace op version test * update mkldnn_inplace fuse pass * update the inplace test
-
由 pangyoki 提交于
* fix test_multinomial * fix test_multinomial add 0 prob
-
由 Kaipeng Deng 提交于
* DataLoader support not auto collate batch. test=develop
-
由 Huihuang Zheng 提交于
PaddleSeg uses tuple as parameter of transpose in dygraph code: https://github.com/PaddlePaddle/PaddleSeg/blob/release/v0.7.0/dygraph/paddleseg/models/danet.py#L152 However, in dy2stat, static code doesn't support the perm as a tuple. This PR fixed it.
-
由 GaoWei8 提交于
-
由 Guo Sheng 提交于
* Fix gradients with ignore_idx in softmax_with_cross_entropy. test=develop * Fix gradients with ignore_idx in softmax_with_cross_entropy on cpu. Remove softmax_with_cross_entropy from op_threshold_white_list. test=develop * Fix test_softmax_cross_entropy_op.py. test=develop
-
由 Guo Sheng 提交于
Fix no_grad setting in AdamW. test=develop
-
由 YUNSHEN XIE 提交于
-
由 qingqing01 提交于
* Support squeezed label as input in paddle.metric.Accuracy * Revert cifar and fix UT
-
由 Wilber 提交于
-
由 LielinJiang 提交于
* fix vgg num classes
-
由 LielinJiang 提交于
* fix summary for rnn gru lstm
-
由 danleifeng 提交于
-
- 13 11月, 2020 1 次提交
-
-
由 Huihuang Zheng 提交于
I found the unittest failed due to batch size. Maybe the reason is that our CI machine has limited memory. I decreased the batch size.
-