- 01 12月, 2020 7 次提交
-
-
由 QingshuChen 提交于
* update conv2d & softmax to new xpu api * test=kunlun * remove useless comments * test=kunlun * remote softmax xpu op * test=kunlun * update kunlun softmax * test=kunlun * update xpu unitest * test=kunlun * fix elementwise_grad bug for kunlun *test=kunlun
-
由 Jiawei Wang 提交于
* fix 3 doc * fix 3 doc * Update adadelta.py
-
由 lijianshe02 提交于
-
由 Jiawei Wang 提交于
* add lamb optimizer and unittest * fix momentum resume training * fix momentum acc
-
由 wanghuancoder 提交于
-
由 chentianyu03 提交于
* add complex64 and complex128 type; add +-*/@ and slice opreator for complex types * add test cases for complex elementwise, matmul and getitem unittest * add test cases for complex types * add test cases for complex matmul unittest
-
由 Zhou Wei 提交于
* The leaf tensor concept is exposed and the gradient accumulation of leaf tensor * The leaf tensor concept is exposed and the gradient accumulation of leaf tensor * fix coverage * fix api doc * fix CI unittest * fix CI unittest * fix unitest * empty tensor does’t need inner_var_ * fix some error message
-
- 30 11月, 2020 11 次提交
-
-
由 Chen Weihang 提交于
-
由 hong19860320 提交于
-
由 123malin 提交于
* fix paramete prefetch & device guard Co-authored-by: NMrChengmo <cmchengmo@163.com> Co-authored-by: Nchengmo <chengmo@baidu.com>
-
由 liym27 提交于
* Add a class TensorInplaceVersion to count the inplace version and put it in framework::Tensor instead of Allocation or Variable. * Add a new attribute `_inplace_version` for VarBase. * Raise exception if an inplace operation can result in incorrect gradient computation. * Add a new interface _bump_inplace_version() for VarBase to bump the version whenever the Tensor is modified through an inplace operation. * For api assign, call _bump_inplace_version() when it's an inplace operation inn dynamic mode. * Use original var_wrapper if the inplace_version is not changed. * Replace SnapshotVarWrapperList with SnapshotVarWrapper to optimize performane.
-
由 joejiong 提交于
As the title
-
由 WangXi 提交于
-
由 Chen Weihang 提交于
* fix failed tests in yingchun gived list * add unittests into static_mode_white_list * add enable static * fix dist unittest * skip test_sigmoid_focal_loss_op & add gym * revert no need skip unittests * remove gym
-
由 Thunderbrook 提交于
-
由 liym27 提交于
-
由 WeiXin 提交于
* Changed a variable name error * Add comments * Move member functions of TranslatedLayer out of function * edit code according to review * Edit input argument of '_run_static_graph' * reset due to Segmentation fault * rename variables when stitching graph * modify code according CI * Add comments to '__i_m_p_l__' * remove blanks befor 'Get...' * edit code according to review * Add a comment to '_execution_method_creator' * Edit a comment to '_execution_method_creator'
-
由 wanghuancoder 提交于
* Generate code coverage reports only for incremental files, test=develop * Generate code coverage reports only for incremental files, test=develop * Generate code coverage reports only for incremental files, test=develop * test for diff python file, test=develop * fix no python diff report, test=develop * add cc test file, test=develop * fix bug in generic.cmake, test=develop * for debug no cc report, test=develp * modify compire branch form test_pr to test, test=develop * fix bug, test=develop * test for h file changed, test=develop * debug for redefinition of argument optimize error, test=develop * close -o3 for test, test=develop * remove -o3 for test, test=develop * remove coverage option for nvcc, test=develop * use CMAKE_CXX_FLAGS open coverage option when header file changed, test=develop * reopen -o3, test=develop * remove debug code, test=develop * remove unused code, test=develop
-
- 28 11月, 2020 4 次提交
-
-
由 Huihuang Zheng 提交于
test_mnist failed on CUDA11. We found that it is due to PaddleInference IR Optimization after debugging. We disable it in this PR and we will re-enable it after PaddleInference fixes it.
-
由 liym27 提交于
-
由 liym27 提交于
[Dy2Stat] Fix bug: the return statement should be transformed to an equivalent Paddle/Python if statement, which depends on if conditions of the return stmt. (#29165)
-
由 Huihuang Zheng 提交于
GridGenerator model failed because the output shape of `linspace` is (-1). The reason is that C++ InferShape fixes the shape to (-1): https://github.com/PaddlePaddle/Paddle/blob/5da3d514ebaa6fffd48c4a2e6bb5b16268dae92e/paddle/fluid/operators/linspace_op.cc#L49 We cannot set the shape in C++ infer shape because this Tensor may not be initialized during compile time, but when input `num` of `linspace` is an integer, we know the shape at compiler time. This PR simply set the shape in Python and add GridGenerator as unittest.
-
- 27 11月, 2020 12 次提交
-
-
由 Aurelius84 提交于
-
由 ShenLiang 提交于
* add reducer * refine envent for memorycopy * add concat&split for allreduce * apply concat & split for fuse tensor * fix nccl dep * fix the untest, compile problem and ddp initialize problem * fix untest for mac & add some comments & solve the repeated param in sublayers * fix untest for windows & fix document
-
由 lilong12 提交于
update expand as op to use the shape of the target tensor instead of the target tensor itself. (#29020) * update, test=develop
-
由 Kaipeng Deng 提交于
* alias yolo_loss & decode_yolo_box to paddle.vision. test=develop
-
由 LutaoChu 提交于
add paddle.subtract, optimize paddle.maximum and paddle.minimum
-
由 Jack Zhou 提交于
Add eigen gru and fix the dropout bug in the rnn
-
由 liym27 提交于
[Dynamic-to-Static] Support **kwargs as input of the function which is decorated by `jit.save.to_static` (#29098)
-
由 YUNSHEN XIE 提交于
-
由 Chen Weihang 提交于
-
由 Shang Zhizhou 提交于
* remove -DSUPPORTS_CUDA_FP16 in cuda.cmake * comile with cuda9 * add some unittest * notest;test=coverage * add unittest for trt plugin swish && split * update ernie unittest * fix some error message * remove repeated judgement of CUDA version in mbEltwiseLayerNormOpConverter * fix comile errror when CUDA_ARCH_NAME < Pascal" * fix comile error * update unittest timeout * compile with cuda9 * update error msg * fix code style * add some comments * add define IF_CUDA_ARCH_SUPPORT_FP16 * rename IF_CUDA_ARCH_SUPPORT_FP16 to CUDA_ARCH_FP16_SUPPORTED
-
由 Chen Weihang 提交于
* add symlink force for unittest * open unittest
-
由 Chen Weihang 提交于
-
- 26 11月, 2020 6 次提交
-
-
由 lilong12 提交于
* update, test=develop
-
由 Leo Chen 提交于
* split train_mode and has_grad * fix format * fix ci problems * fix sample code
-
由 YUNSHEN XIE 提交于
-
由 JZ-LIANG 提交于
* add lars to fleet meta optimizer * add lamb to proto * add lamb to fleet meta optimizer * fixed syntax bug * fixed syntax bug * fixed syntax error in lamb, add config setter of lamb in distributed_strategy * trigger unitest to rerun * add new unitest func for lamb * revise unitest for lars and lamb * revise dgc meta unitest * revise lars document in distribute_strategy * revise lars lamb document in distributed_strategy.py * revise lars lamb document in distributed_strategy.py * add weight decay exclude logic to lars * restore optimzier.py * restore optimizer.py as develop except lars * add epsilon and exclude fn to distributed_sttrategy * add lars epsilon * revise unitest for fleet lars and lamb * revise lars lamb unitest for CI coverage * revise lars argument api * revise lars argument api * revise lars argument api * revise api doc of lars * fix op role * add sharding save and add_sync_comm_for_test function * add comm_analyse to utlis * revise sharding_utils * add sharding saving unittest * revise sharding utils for unittest * revise sharding en doc * update sharding utils api * add doc for sharding * fixed bug in sharding var size count * update varsize count in sharding * fix sharding num_nccl_comm * Revert "fix sharding num_nccl_comm" This reverts commit d51587c15e9323acf226ddd36154275f0d1daf76.
-
由 Yanghello 提交于
-
由 WangXi 提交于
-