- 31 8月, 2021 4 次提交
-
-
由 Qi Li 提交于
* [NPU] fix cmake for ascend ci, test=develop * update paddle_build.sh scripts, test=allcase
-
由 Wilber 提交于
-
由 Aganlengzi 提交于
-
由 Aganlengzi 提交于
-
- 30 8月, 2021 3 次提交
-
-
由 xiaoxiaohehe001 提交于
* add_op_unittest
-
由 zhulei 提交于
* [NPU] Add log_loss op * [NPU] Add log_loss op * [NPU] Add log_loss op
-
由 xiongkun 提交于
* tmp * Tile - Assign - Crop * Finish the set value npu kernel and test case in npu * improve the error message * Modify according to zhangliujie * code review
-
- 27 8月, 2021 15 次提交
-
-
由 JYChen 提交于
-
由 xiaoting 提交于
* add maxunppol2d op, test=develop * fix typo, test=develop * fix unpool unitest, test=develop * fix unpool code-example, test=develop * fix for unpool_op_unittest,test=develop * fix example code, test=develop * add noqa:F401, test=develop * fix converage, test=develop * fix unitest for unpool, test=develop * rename unpool2d to unpool, test=develop * rename unpool2d to unpool, test=develop
-
由 Guoxia Wang 提交于
* sparse_momentum_op is used to save w@GRAD memory for gather_op when gather from a large parameter
-
由 WangXi 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 HydrogenSulfate 提交于
-
由 baoachun 提交于
* add elementwise max grad op for npu * add elementwise max grad op for npu * add elementwise max grad op for npu * add elementwise max grad op for npu * add elementwise max grad op for npu
-
由 WeiXin 提交于
* polish the error message of paddle.slice. * polish code.
-
- 26 8月, 2021 6 次提交
-
-
由 WeiXin 提交于
* polish code * polish code. * polish code. * polish code. * polish code.
-
由 Aurelius84 提交于
* Modify into QueueSync QueueAsync * fix complie on MacOS * fix pointer * fix conflict * polish unittest * fix windows fetch error * polish code according reviewer * fix device_guard on CPU place
-
由 Bo Liu 提交于
-
由 zhouweiwei2014 提交于
-
由 shiyutang 提交于
* add_roi_align_npu * update * update * update
-
由 Wilber 提交于
-
- 25 8月, 2021 4 次提交
-
-
由 jakpiase 提交于
* temporary change * fix for expand_v2 * changes after review, activated ppyolov inference test
-
由 Adam Osewski 提交于
* Enable BF16 for creating global tensor and reduce_mean. * Functional test with small model.
-
由 ronnywang 提交于
-
由 taixiurong 提交于
-
- 24 8月, 2021 8 次提交
-
-
由 lilong12 提交于
-
由 Wilber 提交于
-
由 Haohongxiang 提交于
* Add no_sync in data parallel for dynamic graph * modify UT of no_sync * delete test_parallel_dygraph_dataparallel_no_sync.py * add test_parallel_dygraph_no_sync.py * modify run_trainer_with_spawn in UTs * Add UT of complex control flow in no_sync * add specific descriptions and notes for no_sync * check code style * modify UT's TIMEOUT in CMakeLists.txt
-
由 Adam Osewski 提交于
* Small corrections. * Fix lr for bf16. * Revert some changes.
-
由 Jacek Czaja 提交于
* - concat refactoring draft * - cmpilation fixes * - yet another compilation fix * - fix * - compilation fix * - fixes to compilation * - another compilation fix * - fix * - Added overloaded AcquirePrimitiveDesc for concat * - fix * - reserve introduced * - UT fixes * - test concat int8 improved * - fixes * - fix to crash * - lint fixes * - fixes after review * - some other fixes from review
-
由 ronnywang 提交于
* add conv_op_npu and test * add more tests * clean headers & support fp16 * update
-
由 ronnywang 提交于
* add pool2d_op_npu and test * update * update pool2d_backward_navie * clean headers
-
由 Yulong Ao 提交于
* add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * add dist * update * update * update * update * update * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update * update * update * update * update * update, test=develop * update, test=develop * update * update * delete unused proto * resotre op_desc * restore type_defs * update var_desc * remove dimss_mapping for proto_pybind * update interface.py * update framework.py * update * update * add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * [WIP] Add the auto completion feature and related codes * [WIP] Improve the auto completion and related codes * [WIP] Make the auto completion to support data-parallel * [WIP] Make the completion support mp and dp+mp * [WIP] Refactor auto completion unit test for MLP * [WIP] Refactor the implementation of DistributedOperatorImpl * [WIP] Improve dims_mapping update rule and fix a bug * [WIP] Support auto completion for one transformer decoder layer * [WIP] Add a minor change * [WIP] Fix a bug within the uint test * Shard XShape tensor, add embedding completion and refactor code * Add the distributed_operators dir to setup.py.in * Improve the completion process and add the unittest for gpt * fix process_mesh ut * fix process_mesh ut * update * update, test=develop * Add support for automatically completing distributed attrs of special ops * update * update * update * fix doc sample codes, test=develop * improve coverage, test=develop * add static_mode check, test=develop * Model the cluster for cost model and physical mapping * update, test=develop * add set_placement, test=develop * Add the check to make sure the candidate tensors' size is great than zero * update doc, test=develop * update doc, test=develop * update doc, test=develop * update doc, test=develop * update, test=develop * Auto mark dist attrs annotated by user * update ndarray to nested list, test=develop * update, test=develop * Add auto-completion module for auto-parallel (based on PR#33804) * Remove unnecessary files * Remove unrelated files for the auto completion pr * Update the unit test to improve the coverage * Modify codes based on reviews * Minor changes for CI * Improve some codes based on new comments * Fix bugs caused by shallow copy in attributes.py * Imporve amend_distributed_attr_for_program in context.py * Other changes for weihang's comments Co-authored-by: Nsandyhouse <lilong12@baidu.com>
-