- 03 8月, 2020 2 次提交
-
-
由 Dong Daxiang 提交于
* split meta optimizer files * add graph execution in execution, update two properties in DistributedStrategy, unit tests for these features
-
由 gentelyang 提交于
-
- 01 8月, 2020 1 次提交
-
-
由 Yi Liu 提交于
* add localsgd meta optimizer
-
- 31 7月, 2020 7 次提交
-
-
由 Zhang Ting 提交于
-
由 lilong12 提交于
* add pipeline optimizer
-
由 yukavio 提交于
-
由 Bai Yifan 提交于
* Remove slim from paddle framework test=develop Co-authored-by: Nwanghaoshuang <wanghaoshuang@baidu.com>
-
由 ShenLiang 提交于
* fix inverse bug, test=develop * fix the untest, test=develop * add singular checking, test=develop * fix the utest, test=develop * use memory::copy, test=develop * fix bost_get, test=develop * fix position, test=develop
-
由 lilong12 提交于
* hidden the explicit setting of dataset for pipeline training.
-
由 lilong12 提交于
* modify pipeline demo, test=develop
-
- 30 7月, 2020 13 次提交
-
-
由 wawltor 提交于
Update the code for the compare_ops, update the api and doc
-
由 wangchaochaohu 提交于
-
由 wangchaochaohu 提交于
-
由 lilong12 提交于
* fix test_pipeline, test=develop
-
由 Chen Weihang 提交于
* refine error stack format * polish compile traceback format * polish detail format
-
由 mapingshuo 提交于
* add gradient Merge optimizer to meta, test=develop
-
由 tangwei12 提交于
Integrated Trainer of Parameter Server (API add `fluid.contrib.layers.sparse_embedding` only) (#22957) * Integrated Trainer of Parameter Server
-
由 liym27 提交于
-
由 liym27 提交于
[Dy2Stat-ErrorMessage] Add interface:create_origin_info_map and attach_origin_info for AST node (#25627) * Add interface:create_origin_info_map and attach_origin_info for AST node. test=develop * Fix code according to comments from reviewers. test=develop
-
由 hong 提交于
* fix double grad visitid unit; test=develop * change name hash_pair to HashPair; test=develop * follow comment; test=develop
-
由 hong 提交于
* reduce hidden size and loop number; test=develop * change loop number; remove useless code; test=develop
-
由 wawltor 提交于
Fix the argsort and sort op for the api2.0, and update the api
-
- 29 7月, 2020 6 次提交
-
-
由 cc 提交于
* Remove the output for moving_average_abs_max_scale op, test=develop
-
由 Dong Daxiang 提交于
* refine strategy compiler and meta optimizers make async as a_sync
-
由 zhupengyang 提交于
-
由 Chen Weihang 提交于
* remove ProgramTranslator.save_inference_model * adapt save_quantized_model * revert buffer check implemention * remove useless import function
-
由 Chen Weihang 提交于
* simplify buffered reader to improve DataLoader performance * fix 22 failed unittests * fix cuda pinned context condition * fix test_reader_reset failed * fix two failed unittests * change unittest place * polish error messaage * polish cast op GetExpecctedKernelType * remove debug info in unittest
-
由 Huihuang Zheng 提交于
Enhance TracedLayer Error Message Note: this PR uses assert to check type somewhere and check_type somewhere, the reason is that the check_type skips checking when it is under dygraph mode.
-
- 28 7月, 2020 6 次提交
-
-
由 yukavio 提交于
* saving inference model for user defined quantization model * saving inference model for user defined quantization model * fixed ci coverage
-
由 mapingshuo 提交于
-
由 Dong Daxiang 提交于
* add more settings for distributed strategy Basically, DistributedStrategy has several parts of configurations: - BuildStrategy: the same as paddle.fluid.BuildStrategy, but the distributed arguments are moved out of BuildStrategy - ExecutionStrategy: the same as paddle.fluid.ExecutionStrategy - collective communication configs: nccl_comm_num, hierarchical allreduce and so on - distributed algorithms: async_update(mainly used in PS), lars, lamb and so on
-
由 Sylwester Fraczek 提交于
-
由 cc 提交于
-
由 zhupengyang 提交于
-
- 27 7月, 2020 3 次提交
-
-
由 wangchaochaohu 提交于
-
由 Wojciech Uss 提交于
test=develop
-
由 mapingshuo 提交于
-
- 24 7月, 2020 2 次提交
-
-
由 Huihuang Zheng 提交于
Based on the comment here https://github.com/PaddlePaddle/Paddle/blob/b5f8784cab94eae785659787fc529870c87b254c/paddle/fluid/framework/details/build_strategy.h#L49 The unit test which compares Reduce and AllReduce must have diff. The PR_CI_Night runs on P40 machine and it has 8GB GPU, which is smaller than the 16GB normal CI machines. So we decrease the batch size in the past to make it runnable: https://github.com/PaddlePaddle/Paddle/pull/24651/files . Decreasing the batch size makes the difference occurs often. So this PR replace the absolute delta by relative delta. Before this PR, the unit test failure happens with probability about 1/100. After this PR it doesn't happen.
-
由 Zhen Wang 提交于
* fix the double grad bug for the star gan. test=develop * update the retain_graph parameter doc. test=develop * add the unit test for the retain_graph parameter. test=develop
-