- 23 2月, 2023 1 次提交
-
-
由 Weilong Wu 提交于
This reverts commit 244e7546.
-
- 06 2月, 2023 1 次提交
-
-
由 wanghuancoder 提交于
* refine optimizer create accumulators * refine
-
- 30 1月, 2023 1 次提交
-
-
由 wanghuancoder 提交于
* refine _found_inf
-
- 30 12月, 2022 1 次提交
-
-
由 Sanbu 提交于
* 1219 * temporarily change the num_diff_files limit, test=document_fix * Revert "temporarily change the num_diff_files limit, test=document_fix" This reverts commit 8e70f00ef468d2dad0e38b3da06295ed62990d20. * for codestyle * remove duplicate license * `static mode` -> `static graph mode` * Update hybrid_parallel_inference.py * Update layer_function_generator.py * Update manipulation.py * reset Co-authored-by: NLigoml <39876205+Ligoml@users.noreply.github.com> Co-authored-by: NSigureMo <sigure.qaq@gmail.com>
-
- 25 12月, 2022 1 次提交
-
-
由 wanghuancoder 提交于
-
- 09 12月, 2022 1 次提交
-
-
由 cyber-pioneer 提交于
-
- 29 11月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* isort all files * revert conflicting files * revert conflicting files * revert conflicting files
-
- 17 11月, 2022 1 次提交
-
-
由 Yuang Liu 提交于
Support bfloat16 for adamw and adam optimizer. Fit the lr for pure bf16 training with tensor fusion. (#48041) * add bfloat16 for adamw * set lr not to bfloat16 for pure bf16 training * update the logic * update the adamw optimizer * support bfloat for adam
-
- 10 11月, 2022 1 次提交
-
-
由 WangZhen 提交于
Get grads types from cpp for adam to speed up
-
- 03 11月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* [CodeStyle][py2][U008] remove unnecessary args in `super()` * remove remained args * revert changes in test_pylayer_op * Revert "revert changes in test_pylayer_op" This reverts commit ff185a9ae738afac3b0264f61bde6c6b7f72e7c4. * revert some changes in example code
-
- 26 10月, 2022 1 次提交
-
-
由 sneaxiy 提交于
-
- 23 10月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* update config * re-blacken python code * temporarily disable date and diff_py_file * skip a format
-
- 10 10月, 2022 1 次提交
-
-
由 Shuangchi He 提交于
[CodeStyle][F401] remove unused imports in python_paddle/proto_onnx_optimizer_hapi_autograd_framework. (#46754)
-
- 30 9月, 2022 1 次提交
-
-
由 sneaxiy 提交于
* support pure bfloat16 * support bf16 linear * update PR to pass CI * tiny fix where_grad_kernel.cu * add bfloat16 to selu_grad to pass CI * fix selu grad compilation error
-
- 23 9月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
-
- 14 9月, 2022 1 次提交
-
-
由 Nyakku Shigure 提交于
* trim trailing whitespace * fix `.cmake-format.py` * revert npu ut changes, avoid npu ci error
-
- 29 8月, 2022 1 次提交
-
-
由 HongyuJia 提交于
* add legacy_api.yaml * set merged_momentum inplace only * support inplace optional<vector<tensor>> * add dygraph_mode api * add merged_adam yaml * add merged_adam python api * change testcase of merged_adam and adam * fix import of test_merged_adam_op
-
- 26 8月, 2022 1 次提交
-
-
由 wanghuancoder 提交于
-
- 29 6月, 2022 1 次提交
-
-
由 zyfncg 提交于
* support complexd selected_rows kernel in yaml * support configuring optimizer api in yaml * fix data transform bug
-
- 05 6月, 2022 1 次提交
-
-
由 Sing_chan 提交于
* use yapf to format all python file * yapf exclude two unittests file for they rely on writing and reading file, and format will break them * disable diff_py_file because too many diff files cause command following failed
-
- 13 4月, 2022 1 次提交
-
-
由 chentianyu03 提交于
* add adam yaml * add adam final_state api * add adam_impl
-
- 31 3月, 2022 1 次提交
-
-
由 zhangbo9674 提交于
-
- 25 3月, 2022 1 次提交
-
-
由 Jiabin Yang 提交于
* refactor eager flags * fix flags error when we switch from eager to dygraph * fix ci problem * fix ci * fix ci * merge develop and fix code style * merge develop and fix code style * fix op test error * fix op test error * fix op test error * fix op test error * fix op test error * merge develop
-
- 20 1月, 2022 1 次提交
-
-
由 zhangbo9674 提交于
* fix mp * support merged_momentum for mp
-
- 07 1月, 2022 1 次提交
-
-
由 zhangbo9674 提交于
* add multi tensor for adam * add merged_adam op * refine code * refine adam compute logic
-
- 17 9月, 2021 1 次提交
-
-
由 zhangbo9674 提交于
* add pure fp16 major function in auto_cast & tracer * support master weight in dygraph for pure fp16 * check mix dtype of fp16&fp32 for check_finite_and_unscale op * change pure fp16 funtion name * refine some bug in auto_cast * refine auto_cast interface logic * add param _casted_by_pure_fp16 for class Layer * support state_dict hook for save model by user appointed dtype in pure_fp16_decorator * refine pure_fp16_decorator as decorator * add unittest * add comment * add comment * support recompute * add comment for auto_cast and decorator * support to_static_state_dict for paddle.jit.save * unlimite models num and optimizers num * add lookup_table in black_list * fix momentum and layer state_dict * fix bug in layer state_dict * fix bug in layer state_dict_helper * refine unittest * refine test_momentun_op * refine interface and some code * refine amp_decorator interface * refine pure fp16 interface * refine master weight interface
-
- 02 8月, 2021 1 次提交
-
-
由 sunzhongkai588 提交于
* fix paddle.optimizer test=document_fix * fix paddle.optimizer test=document_fix
-
- 15 7月, 2021 1 次提交
-
-
由 wanghuancoder 提交于
* cache core.ops, test=develop * refine, test=develop
-
- 31 5月, 2021 1 次提交
-
-
由 wangguanzhong 提交于
* support params groups, test=develop * simplify updating opt attr * update according to review
-
- 29 4月, 2021 1 次提交
-
-
由 zhiboniu 提交于
add __all__=[] to python files not in API public list; import * only support in API public list files (#32643)
-
- 28 4月, 2021 1 次提交
-
-
由 Leo Chen 提交于
* add input EpsilonTensor for adam * update python api * add unit test * add npu test * add more ut
-
- 27 4月, 2021 1 次提交
-
-
由 xiemoyuan 提交于
* fixed docs. * Fixed docs. test=document_fix code bak. fixed docs. test=document_fix * Revert to previous version of python/paddle/fluid/backward.py * fixed bugs. * test=document_fix. Fixed examples.
-
- 23 4月, 2021 1 次提交
-
-
由 zhiboniu 提交于
-
- 24 3月, 2021 1 次提交
-
-
由 Chen Weihang 提交于
-
- 01 2月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 08 1月, 2021 1 次提交
-
-
由 Zhen Wang 提交于
* add cast ops before and after unsupported fp16 ops. * Keep partial net in FP32 pattern. * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode. * Add fp16 support for adam op. * add multi precision attr for adam. * Fix the bug of test_multi_precision_fp16_train UT. * Code format for CI. * Fix the redefine error about MPTypeTrait on windows. * fix bugs of the _create_accumulators func in Momentum. * fix bug when inserting post cast op. * Add the update_loss_scaling op in allow_set of UnusedVarCheck. * Update for ci coverage. * Add some doc for OptimizerWithMixedPrecision. * Fix the code style. * Imporve the doc of `amp_init`. * Change for fp16 testing if users have the infer program defined in separate way.
-
- 07 1月, 2021 1 次提交
-
-
由 WangXi 提交于
-
- 28 12月, 2020 1 次提交
-
-
由 Leo Chen 提交于
-
- 27 11月, 2020 1 次提交
-
-
由 ShenLiang 提交于
* add reducer * refine envent for memorycopy * add concat&split for allreduce * apply concat & split for fuse tensor * fix nccl dep * fix the untest, compile problem and ddp initialize problem * fix untest for mac & add some comments & solve the repeated param in sublayers * fix untest for windows & fix document
-
- 24 11月, 2020 1 次提交
-
-
由 Leo Chen 提交于
* upgrade comment string to raw string * fix string in * fix string with ' ' * revert update on comments * upgrade only necessary * fix sample code checker * fix comments with '''
-