- 31 3月, 2021 1 次提交
-
-
由 taixiurong 提交于
-
- 02 3月, 2021 1 次提交
-
-
由 Gradie 提交于
* lamb_op_xpu;test=kunlun * modify lamb_op_xpu.cc;test=kunlun * delete atol lamb_op_xpu; test=kunlun * update xpu.cmake;test=kunlun * test_error 1e-5,lamb_op_xpu;test=kunlun * error1e-5,lamb_op_xpu,test=kunlun * delete atol lamb_xpu;test=kunlun * modify atol,lamb_op_xpy;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu, XPUOptest;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu,modify xpu_cmake; test=kunlun * lamb_op_xpu;test=kunlun * lamb_op_xpu,modify xpucmake;test=kunlun
-
- 19 1月, 2021 1 次提交
-
-
由 ykkk2333 提交于
* add rmsprop_op_xpu test=kunlun * modified rmsprop_op_xpu error code. test=kunlun
-
- 17 1月, 2021 1 次提交
-
-
由 guofei 提交于
* Modify the calculation logic of LambOptimizer
-
- 12 1月, 2021 1 次提交
-
-
由 石晓伟 提交于
-
- 09 1月, 2021 1 次提交
-
-
由 zhang wenhui 提交于
-
- 08 1月, 2021 1 次提交
-
-
由 Zhen Wang 提交于
* add cast ops before and after unsupported fp16 ops. * Keep partial net in FP32 pattern. * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode. * Add fp16 support for adam op. * add multi precision attr for adam. * Fix the bug of test_multi_precision_fp16_train UT. * Code format for CI. * Fix the redefine error about MPTypeTrait on windows. * fix bugs of the _create_accumulators func in Momentum. * fix bug when inserting post cast op. * Add the update_loss_scaling op in allow_set of UnusedVarCheck. * Update for ci coverage. * Add some doc for OptimizerWithMixedPrecision. * Fix the code style. * Imporve the doc of `amp_init`. * Change for fp16 testing if users have the infer program defined in separate way.
-
- 30 12月, 2020 1 次提交
-
-
由 Chengmo 提交于
* fix momentum op register
-
- 21 12月, 2020 1 次提交
-
-
由 LoveAn 提交于
* Test compilation time with less parallel count, notest, test=windows_ci * optimize rules of Unity Build, notest, test=windows_ci, test=windows_op * limit parallel counts used only on GPU, test=develop * remove limit of argument /m:8 on Windows, test=develop
-
- 07 12月, 2020 1 次提交
-
-
由 LoveAn 提交于
* Compiling operator libraries with Unity Build on Windows CPU. * Compiling operator libraries with Unity Build on Windows GPU, no_test, test=windows_ci * Add option in windows ci script, no_test, test=windows_ci * Optimize parallel compiling, test=develop * remove limit of parallel compile and skip some ops in UB, test=develop * remove changes of header file, test=develop * remove changes of header file, test=develop * fix test_eye_op unittest failed, test=develop * Compiling operator libraries with Unity Build on Linux, test=develop * set default WITH_UNITY_BUILD=OFF, test=develop * Move unity build rules into a single file and add comment, test=develop * optimize parallel compilation, test=develop * fix undefined reference error on coverage ci, test=develop
-
- 02 12月, 2020 2 次提交
-
-
由 Zhen Wang 提交于
-
由 Zhen Wang 提交于
* add the weight decay func for the momentum op * Add the multi_precision function in Momentum Optimizer. * Make sure that the initial value of master weights are same with the fp16 weights. * add static loss scaling. * add the rescale_grad function in the pure fp16 training. * use the original momentum updating method. * Polish some codes, such as variable names. * add docstring for apis. * update the var creation details of _create_master_weight. * not modify codes about imperative momentum updating. * Fix the error of test_dist_sparse_tensor_load_momentum UT. * add unit test for multi precision fp16 training. * add more unit tests for CI. * Use lower threshold values for allclose comparing in test_multi_precision_fp16_train UT. * For CI Coverage Checking.
-
- 23 11月, 2020 1 次提交
-
-
由 furnace 提交于
* refactor momentum op to combine weight_decay (scale op and sum op)
-
- 06 11月, 2020 1 次提交
-
-
由 taixiurong 提交于
-
- 19 10月, 2020 2 次提交
-
-
由 yinhaofeng 提交于
* lookup_table_xpu op report errors;test=kunlun * add adam xpu op;test=kunlun * reset lookup * change adam wrong;test=kunlun
-
由 Chengmo 提交于
* fix error message,test=kunlun * fix, test=kunlun
-
- 14 10月, 2020 2 次提交
-
-
由 MRXLT 提交于
* fix adam * fix gpu adam * fix code style * fix ut * update ut add cuda code
-
由 Chen Weihang 提交于
* polish some error message * add white list * revert shell script change
-
- 13 10月, 2020 1 次提交
-
-
由 Chengmo 提交于
* add xpu sgd & momentum
-
- 27 9月, 2020 1 次提交
-
-
由 Chengmo 提交于
* fix sgd/momentum/dpsgd/rmsprop error message
-
- 22 9月, 2020 1 次提交
-
-
由 123malin 提交于
* test=develop, update error message
-
- 21 9月, 2020 1 次提交
-
-
由 MRXLT 提交于
* fix adam * rmsprop support double
-
- 09 9月, 2020 1 次提交
-
-
由 JZ-LIANG 提交于
add lars to fleet meta optimizer
-
- 29 8月, 2020 1 次提交
-
-
由 Jiawei Wang 提交于
* add doc; notest * fix doc; notest * update doc; notest * refine optimizer && adam * refine optimizer; notest * add adam * fix doc * fix doc && add adamw; notest * add error message * bug fix * refine rmsprop && adamax * fix ci * buf fix * update comment * unify arguments place; notest * fix ut, test=develop * bug fix * fix conflicts, test=develop * add examples code * bug fix * fix comments * fix sample code * add sample code for Optimizer * add adamax ut, test=develop * fix rmsprop ut, test=develop * add ut for optimizer.py and adamw.py * first commit of adadelta optimizer * fix learning rate * fix adadelta doc and add sgd momentum * remove unused fluid * fix codestyle * Update test_adam_op.py * Update test_adam_op.py * fix SGD in 2 unittests * fix SGD in 2 unittests * fix ci * fix ut Co-authored-by: NMRXLT <xlt2024@gmail.com> Co-authored-by: Nmapingshuo <mps2012@yeah.net>
-
- 28 8月, 2020 1 次提交
-
-
由 lilong12 提交于
-
- 11 7月, 2020 1 次提交
-
-
由 Chen Weihang 提交于
* fix softmax_with_cross_entropy cuda kernel overflow bug, test=develop * replace old macro & for condition, test=develop * polish details, test=develop
-
- 03 6月, 2020 1 次提交
-
-
由 leesusu 提交于
-
- 13 5月, 2020 3 次提交
-
-
由 gongweibao 提交于
-
由 MRXLT 提交于
-
由 zhang wenhui 提交于
-
- 26 4月, 2020 1 次提交
-
-
由 liuwei1031 提交于
* save InferVarType changes, test=develop * remove code comments, test=develop * tweak code, test=develop * fix compilation warning, update merge_ids_op split_ids_op to new interface, test=develop * modify fused_bn_activation_op, test=develop * fix error of fused_bn_activation_op, test=develop * fix PADDLE_ENFORCE and unittest coverage issue, test=develop * tweak PADDLE_ENFORCE messages, test=develop * improve unittest coverage, test=develop * add StaticGraphInferVarType class, test=develop * rebase develop branch, test=develop * fix unittest error, test=develop * remove comments, test=develop * improve unittest coverage, test=develop * imporve error message and imporve unittest coverage, test=develop * upgrade InferVarType API, test=develop * tweak pyfunc error message, test=develop * fix compilation conflict - save_combine_op, test=develop
-
- 07 4月, 2020 1 次提交
-
-
由 wangchaochaohu 提交于
* add support for value tensor support of fill_constant Op
-
- 04 4月, 2020 1 次提交
-
-
由 Chen Weihang 提交于
* delete invalid check inferface Ref & VectorRef, test=develop * fix vector ref delete error, test=develop * try the new check inferface, test=develop * change all related code with new check macro, test=develop * remove static assert, test=develop * polish detail, test=develop * skip coverage problem, test=develop * add new check macro, test=develop
-
- 27 2月, 2020 1 次提交
-
-
由 zhaoyuchen2018 提交于
* Refine adam op, test=develop * Fuse kernels together to reduce cpu time. * Refine paddle enforce, test=develop * Remove some comments, test=develop * Refine code,test=develop * Refine cuda kernel, test=develop * Refine code according to comments, test=develop
-
- 09 1月, 2020 1 次提交
-
-
由 zhongpu 提交于
* test Optimizer in dygraph, test=develop * add optest for Optimizer in dygraph, test=develop * fix adagrad optimizer, test=develop * fix dpsgd optimizer, test=develop * fix test_optimizer.py, test=develop * fix dpsgd optimizer, this op only support cpu, test=develop * add optest for optimizer, test=develop * add description for dpsgd, test=develop * add rmsprop to white_list in unused_var_check.cc, test=develop * polish code style, test=develop * polish code style, test=develop * delete seed attribute for DpsgdOptimizer, test=develop * change testing to debugging, test=develop
-
- 24 12月, 2019 1 次提交
-
-
由 Aurelius84 提交于
* optimize adam speed by removing _finish_update test=develop * fix SparseAdamFunctor param list test=develop * Remove scale_op in expect_list of adam_op test=develop * fix test optimizer loss assert error test=develop * fix test optimizer loss assert error test=develop * modify PADDLE_ENFORCE usage test=develop * fix op_type in lamb_op.cc test=develop * fix errors ostream format bug test=develop * add betaPowOut in ngraph op test=develop * fix ngraph::op api for gcc8 test=develop * clean code test=develop * modify struct into class test=develop * remove code of beta1Tensor in lamb_op test=develop
-
- 06 12月, 2019 1 次提交
-
-
由 Huihuang Zheng 提交于
Add tests to use dy/dx to make sure the gradient values calculated by the control flow backward is correct. Also fixed bugs detected by those tests. Fix bugs: 1. Unlike sum_op, optimizer ops don't allow uninitialized input tensor. But in conditional_block_grad_op, since the conditional_block may not run, the output gradient tensor may be uninitialized, which will cause the optimizer op error. To fix it, we should let optimizer ops support uninitialized input like sum_op or assign the uninitialized gradient to 0 when the conditional_block_grad_op doesn't run. I found there are about 10+ optimizer ops. **To be simpler, I just assign output gradient of the conditional_block_grad_op to 0 in this PR**. But it can be further explored whether we can make optimizer ops like sum_op to support uninitialized input tensor because theoretically we can speed up without the assigning in conditional_block_grad_op. 2. Infer parameter shapes during append_backward. I didn't know that all our parameters are in global block. When op_desc is inferring shapes at the sub-block, it may not know the shape of gradients of parameters whose shape information is at global block. I fixed it by inferring shapes of gradients from forward var. This PR also did some code clean up: 1. Print the var name when sgd_op catches shape error so that it is easier to debug 2. Fix a typo: dicta -> dict
-
- 29 11月, 2019 2 次提交
-
-
由 Chen Weihang 提交于
* add param & grad shape check for sgd op * add _reshape_inplece interface for dygraph parallel * refine unittest based paddle/models scripts, test=develop * add unittest for parallel grad fuse, test=develop
-
由 hong 提交于
* add_dygraph_execution_context * add dygraph infershape context and execution context; test=develop * fix imperative bug; test=develop * remove inputs outputs interface from execution context, because it have same function with inputNames; test=develop * remove tracer_test ctest; test=develop * fix split op bug; test=develop * fix unitests bug; test=develop * fix distribute test bug; test=develop * fix ngraph compile bug; test=develop * fix grad maker bug; test=develop * fix load op bugs; test=develop * fix operator.cc construct bug; test=develop * remove useless name find in operator; test=develop * add tracer_test; test=develop * fix concat, split bug; test=develop * remove tracer_test unitest; test=develop * fix attribute check bug; test=develop * add test code to fix converage; test=develop * remove useless code, change check backward input in engin; test=develop * unlock var type infer shape;test=develop * add ShareAllLoD api; test=develop * add dygraph infershape context unitest; test=develop * remove increase and decrease lod in dygraph; test=develop * addd override; test=develop * fix increase descrease lod; test=develop * fix paddle_enforce; test=develop * disable lod op dygraph check; test=develop * fix paddle enforce error; test=develop * add comment for op_registry and OperatorBase; test=develop * optimize the comment of op_registry; test=develop * fix format of comment; test=develop * fix format of comment; test=develop * optimize the format of comment; test=develop * optimize the format of the comment; test=develop * optimize comment of op_registry; test=develop
-
- 28 11月, 2019 1 次提交
-
-
由 Kaipeng Deng 提交于
* add Adam beta1/beta2 support Variable. test=develop
-