- 26 8月, 2022 1 次提交
-
-
由 wanghuancoder 提交于
-
- 15 8月, 2022 1 次提交
-
-
由 Charles-hit 提交于
-
- 05 6月, 2022 1 次提交
-
-
由 Sing_chan 提交于
* use yapf to format all python file * yapf exclude two unittests file for they rely on writing and reading file, and format will break them * disable diff_py_file because too many diff files cause command following failed
-
- 01 6月, 2022 1 次提交
-
-
由 Guoxia Wang 提交于
* fix the bug of adamw which set the attribute in param group not working * fix undefined variable * fix api example typo * add unittest * fix unittest typo
-
- 15 4月, 2022 1 次提交
-
-
由 chentianyu03 提交于
* add adamw yaml * fix test case error * make the name of weight and bias in linear1 and linear2 to be constant
-
- 28 3月, 2022 1 次提交
-
-
由 Jiabin Yang 提交于
-
- 25 3月, 2022 1 次提交
-
-
由 Jiabin Yang 提交于
* refactor eager flags * fix flags error when we switch from eager to dygraph * fix ci problem * fix ci * fix ci * merge develop and fix code style * merge develop and fix code style * fix op test error * fix op test error * fix op test error * fix op test error * fix op test error * merge develop
-
- 27 10月, 2021 1 次提交
-
-
由 zhaoyingli 提交于
-
- 18 10月, 2021 1 次提交
-
-
由 taixiurong 提交于
[XPU AMP] 1. xpu support gradient acc 2. xpu support create tensor in dygraph 3. xpu support update weight params in amp (#36439)
-
- 29 9月, 2021 1 次提交
-
-
由 zhaoyingli 提交于
* update func name * skip cpu * update unittest * update unittest
-
- 26 9月, 2021 1 次提交
-
-
由 zhangbo9674 提交于
* adam to adamw in AdamW * add lr_ratio in adamw * refine logic bug in cpu adamw * delete fix bug for cpu adamw * delete fix bug for cpu adamw
-
- 22 9月, 2021 1 次提交
-
-
由 zhaoyingli 提交于
-
- 17 9月, 2021 1 次提交
-
-
由 zhangbo9674 提交于
* add pure fp16 major function in auto_cast & tracer * support master weight in dygraph for pure fp16 * check mix dtype of fp16&fp32 for check_finite_and_unscale op * change pure fp16 funtion name * refine some bug in auto_cast * refine auto_cast interface logic * add param _casted_by_pure_fp16 for class Layer * support state_dict hook for save model by user appointed dtype in pure_fp16_decorator * refine pure_fp16_decorator as decorator * add unittest * add comment * add comment * support recompute * add comment for auto_cast and decorator * support to_static_state_dict for paddle.jit.save * unlimite models num and optimizers num * add lookup_table in black_list * fix momentum and layer state_dict * fix bug in layer state_dict * fix bug in layer state_dict_helper * refine unittest * refine test_momentun_op * refine interface and some code * refine amp_decorator interface * refine pure fp16 interface * refine master weight interface
-
- 14 9月, 2021 1 次提交
-
-
由 zhaoyingli 提交于
* add layerwise learning rate for adamw * fix format * add unitest * add NotImplementedError * add gpu unitest * update gpuplace
-
- 01 9月, 2021 1 次提交
-
-
由 zhaoyingli 提交于
-
- 23 8月, 2021 1 次提交
-
-
由 zhaoyingli 提交于
* adamw support cuda * adamw support cuda
-
- 17 8月, 2021 1 次提交
-
-
由 Roc 提交于
-
- 02 8月, 2021 1 次提交
-
-
由 sunzhongkai588 提交于
* fix paddle.optimizer test=document_fix * fix paddle.optimizer test=document_fix
-
- 30 7月, 2021 1 次提交
-
-
由 wangguanzhong 提交于
* fix lr in param group * add unittest for adamw
-
- 27 7月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 31 5月, 2021 1 次提交
-
-
由 wangguanzhong 提交于
* support params groups, test=develop * simplify updating opt attr * update according to review
-
- 29 4月, 2021 1 次提交
-
-
由 zhiboniu 提交于
add __all__=[] to python files not in API public list; import * only support in API public list files (#32643)
-
- 27 4月, 2021 1 次提交
-
-
由 xiemoyuan 提交于
* fixed docs. * Fixed docs. test=document_fix code bak. fixed docs. test=document_fix * Revert to previous version of python/paddle/fluid/backward.py * fixed bugs. * test=document_fix. Fixed examples.
-
- 23 4月, 2021 1 次提交
-
-
由 zhiboniu 提交于
-
- 22 4月, 2021 1 次提交
-
-
由 hutuxian 提交于
-
- 05 2月, 2021 1 次提交
-
-
由 Zhen Wang 提交于
* Use correct master weights in AdamW. * Just modify the master weight. * Update for CI Coverage.
-
- 19 1月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 08 1月, 2021 1 次提交
-
-
由 Zhen Wang 提交于
* add cast ops before and after unsupported fp16 ops. * Keep partial net in FP32 pattern. * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode. * Add fp16 support for adam op. * add multi precision attr for adam. * Fix the bug of test_multi_precision_fp16_train UT. * Code format for CI. * Fix the redefine error about MPTypeTrait on windows. * fix bugs of the _create_accumulators func in Momentum. * fix bug when inserting post cast op. * Add the update_loss_scaling op in allow_set of UnusedVarCheck. * Update for ci coverage. * Add some doc for OptimizerWithMixedPrecision. * Fix the code style. * Imporve the doc of `amp_init`. * Change for fp16 testing if users have the infer program defined in separate way.
-
- 07 1月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 29 12月, 2020 1 次提交
-
-
由 Chen Long 提交于
* fix code bugs * fix code bugs test=document_fix * fix code bugs test=document_fix
-
- 28 12月, 2020 1 次提交
-
-
由 Leo Chen 提交于
-
- 27 11月, 2020 1 次提交
-
-
由 ShenLiang 提交于
* add reducer * refine envent for memorycopy * add concat&split for allreduce * apply concat & split for fuse tensor * fix nccl dep * fix the untest, compile problem and ddp initialize problem * fix untest for mac & add some comments & solve the repeated param in sublayers * fix untest for windows & fix document
-
- 24 11月, 2020 1 次提交
-
-
由 Leo Chen 提交于
* upgrade comment string to raw string * fix string in * fix string with ' ' * revert update on comments * upgrade only necessary * fix sample code checker * fix comments with '''
-
- 16 11月, 2020 1 次提交
-
-
由 Guo Sheng 提交于
Fix no_grad setting in AdamW. test=develop
-
- 02 11月, 2020 1 次提交
-
-
由 Guo Sheng 提交于
* Fix lr setting of AdamW when lr is an instance of LRScheduler. test=develop * Fix static graph test mode in test_adamw_op.py. test=develop
-
- 13 10月, 2020 1 次提交
-
-
由 Zhou Wei 提交于
* fix doc and unittest of 2.0 lr_scheduler * fix doc of 2.0 lr_scheduler * fix unittest * fix english doc of lr_scheduler * fix api name of lr scheduler * fix api name of lr scheduler
-
- 12 10月, 2020 1 次提交
-
-
由 MRXLT 提交于
* refine parameters order && doc * update rmsprop doc * refine adam/transpose/unstack/stride_slice * fix bug && doc * fix doc * bug fix * bug fix * fix doc * fix doc * fix doc * fix doc * depercate old strided_slice * update doc * set default value for name * update doc
-
- 29 9月, 2020 1 次提交
-
-
由 Chen Weihang 提交于
* remove data parallel scale loss & apply collective_grads * move apply in minimize * fix failed unittests
-
- 01 9月, 2020 1 次提交
-
-
由 MRXLT 提交于
* update doc * update doc * fix optimizer sample code * add default value for adamw weight_decay * fix adamw * change LearningRateDecay to _LRScheduler * fix adamw;notest * fix load;notest * remove file * bug fix * fix code style * bug fix * add ut * adamw support weight_decay=0 * fix ut * fix set_lr doc * fix doc * change parameters place
-
- 23 8月, 2020 1 次提交
-
-
由 MRXLT 提交于
refine Optimizer/Adam/Admax/RMSProp && add Admw * buf fix * update comment * unify arguments place; notest * fix ut, test=develop * bug fix * fix conflicts, test=develop * add examples code * bug fix * fix comments * fix sample code * add sample code for Optimizer * add adamax ut, test=develop * fix rmsprop ut, test=develop * add ut for optimizer.py and adamw.py * remove TestAdamOptimizerBetaVariable * update api && add ut * update doc && fix ut * add ut Co-authored-by: Nmapingshuo <mps2012@yeah.net>
-