- 10 9月, 2021 20 次提交
-
-
由 zhouweiwei2014 提交于
-
由 Zhong Hui 提交于
-
由 zhiboniu 提交于
-
由 Shang Zhizhou 提交于
-
由 hlygit66666 提交于
* add test_cumprod_op * Revert "add test_cumprod_op" This reverts commit c96cf6dff5d09ae7d8cc72c1e8ae4369a153aa19. * recommit * add error message * test input(x) initialize * test use cpu * update test code * add test type * add test case * solve ci problem * add complex case test * add complex case test * fix review problem * fix conflict * fix some docs * change test case * change test case * fix review problems again * fix docs * fix inclusivescan bug
-
由 huangxu96 提交于
This PR supports gradient clip (ClipGradByGlobalNorm) when training with AMP(auto mixed precision).
-
由 chentianyu03 提交于
* add llvm::SmallVector to paddle * rename small vector file * merge paddle small vector to one file * add small_vector_test * modify smallvector test argument type * add string header
-
由 chentianyu03 提交于
* import ska flat_hash_map * add define NOMINMAX macro to fix windows build failed bug * add brackets to std::max in flat_hash_map * move flat_hash_map directions * modify namespace to paddle * modify namespace to paddle * modify namespace to paddle * modify namespace to paddle * rm not used map.h and replace with op_info
-
由 baoachun 提交于
-
由 feng_shuai 提交于
-
由 baoachun 提交于
-
由 ronnywang 提交于
-
由 ShenLiang 提交于
-
由 huangjun12 提交于
* add as_extra to dropout op and lrn op * refine details * fix dropout op maker
-
由 Jiawei Wang 提交于
* fix extra op for expand, expand_as, tile, unstack * Update expand_v2_op.cc
-
由 ceci3 提交于
* fix bn/in/squeeze/syncbn extra * update bn * update * update
-
由 Shang Zhizhou 提交于
* add opdef extra * add reduce mean * update style
-
由 Zeng Jinle 提交于
* fix scatter gather bug: * fix windows ci
-
由 wenbin 提交于
* conv3d * remove const_cast * modify ut * disable dynamic shape for trt6.0 * remove trt5
-
由 pangyoki 提交于
* add asExtra for nce op * fix unittest error in macos * remove asExtra for is_test
-
- 09 9月, 2021 9 次提交
-
-
由 Wei Shengyu 提交于
* mark extra attr for unsqueeze2 * debug for inference
-
由 crystal 提交于
-
由 XGZhang 提交于
-
由 XGZhang 提交于
-
由 whs 提交于
-
由 zhangchunle 提交于
-
由 tianshuo78520a 提交于
-
由 zhangkaihuo 提交于
-
由 0x45f 提交于
* init matrix_rank op, add matrix_rank CPU code and test * add GPU kernel, remove svd_eigen.h * add CPU kernel when tol is tensor * add cpu and gpu code when tol is tensor * fix CI-ROCM error * add matrix_rank API describe, fix PR-CI-Py3 error * fix PR-CI-Windows error, add matrix_rank API test * delete useless comments * fix review * add my code in svd_helper.h * update doc commets * remove spaces
-
- 08 9月, 2021 11 次提交
-
-
由 Leo Chen 提交于
* add backward inplace for dygraph * fix bug * support gradient accumulation
-
由 xiegegege 提交于
-
由 Li Min 提交于
-
由 zhouweiwei2014 提交于
-
由 zhouweiwei2014 提交于
-
由 Yulong Ao 提交于
* add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * add dist * update * update * update * update * update * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update * update * update * update * update * update, test=develop * update, test=develop * update * update * delete unused proto * resotre op_desc * restore type_defs * update var_desc * remove dimss_mapping for proto_pybind * update interface.py * update framework.py * update * update * add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * [WIP] Add the auto completion feature and related codes * [WIP] Improve the auto completion and related codes * [WIP] Make the auto completion to support data-parallel * [WIP] Make the completion support mp and dp+mp * [WIP] Refactor auto completion unit test for MLP * [WIP] Refactor the implementation of DistributedOperatorImpl * [WIP] Improve dims_mapping update rule and fix a bug * [WIP] Support auto completion for one transformer decoder layer * [WIP] Add a minor change * [WIP] Fix a bug within the uint test * Shard XShape tensor, add embedding completion and refactor code * Add the distributed_operators dir to setup.py.in * Improve the completion process and add the unittest for gpt * fix process_mesh ut * fix process_mesh ut * update * update, test=develop * Add support for automatically completing distributed attrs of special ops * update * update * update * fix doc sample codes, test=develop * improve coverage, test=develop * add static_mode check, test=develop * Model the cluster for cost model and physical mapping * update, test=develop * add set_placement, test=develop * Add the check to make sure the candidate tensors' size is great than zero * update doc, test=develop * update doc, test=develop * update doc, test=develop * update doc, test=develop * update, test=develop * Auto mark dist attrs annotated by user * update ndarray to nested list, test=develop * update, test=develop * Add auto-completion module for auto-parallel (based on PR#33804) * Remove unnecessary files * Remove unrelated files for the auto completion pr * Update the unit test to improve the coverage * Modify codes based on reviews * Minor changes for CI * Improve some codes based on new comments * Fix bugs caused by shallow copy in attributes.py * Imporve amend_distributed_attr_for_program in context.py * Other changes for weihang's comments * support shard reader * support shard reader * add parallel mode * update process mesh * add method to compute comm_group * implement dist_embedding forward func * implement dist matmul forward func * implement dist reshape forward func * add transpiler framework * add transpiler forward * implement transpiler forward * implement transpiler backward & update * add process * add unitest * chmod * chmod * chmod * update unitest * add unitest for gpt * remove unused print * rename transpiler --> partitioner * rename transpiler --> partitioner * chmod * chmod * bug fixed * remove amp function * update case for dp mode * update case for dp mode * [Auto Parallel] Integrate all parts with the newest code * Integrate all parts of auto parallel and improve codes * Integrate all parts by AutoParallelizer * Add unit test for AutoParallelizer * Improve auto completion module for pipeline parallel * Add support for matmul_v2 in dist_matmul * Correct the typo "stratergy" to "strategy" * Modify distributed_strategy.proto to conform the main stream * Restore parts of distributed_strategy to conform the develop branch Co-authored-by: Nsandyhouse <lilong12@baidu.com> Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
-
由 will-jl944 提交于
multiply supports bool
-
由 WangXi 提交于
-
由 niuliling123 提交于
-
由 wanghuancoder 提交于
* refactor new executor, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop
-
由 CtfGo 提交于
* mark WhileOp AsExtra attribute * revert kX and kOutputs
-