- 09 9月, 2021 1 次提交
-
-
由 0x45f 提交于
* init matrix_rank op, add matrix_rank CPU code and test * add GPU kernel, remove svd_eigen.h * add CPU kernel when tol is tensor * add cpu and gpu code when tol is tensor * fix CI-ROCM error * add matrix_rank API describe, fix PR-CI-Py3 error * fix PR-CI-Windows error, add matrix_rank API test * delete useless comments * fix review * add my code in svd_helper.h * update doc commets * remove spaces
-
- 08 9月, 2021 31 次提交
-
-
由 Leo Chen 提交于
* add backward inplace for dygraph * fix bug * support gradient accumulation
-
由 xiegegege 提交于
-
由 Li Min 提交于
-
由 zhouweiwei2014 提交于
-
由 zhouweiwei2014 提交于
-
由 Yulong Ao 提交于
* add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * add dist * update * update * update * update * update * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update, test=develop * update * update * update * update * update * update, test=develop * update, test=develop * update * update * delete unused proto * resotre op_desc * restore type_defs * update var_desc * remove dimss_mapping for proto_pybind * update interface.py * update framework.py * update * update * add auto_parallel dir * mv to paddle.distributed * add shard_xx api * add distributed attrs for var * add ut, test=develop * [WIP] Add the auto completion feature and related codes * [WIP] Improve the auto completion and related codes * [WIP] Make the auto completion to support data-parallel * [WIP] Make the completion support mp and dp+mp * [WIP] Refactor auto completion unit test for MLP * [WIP] Refactor the implementation of DistributedOperatorImpl * [WIP] Improve dims_mapping update rule and fix a bug * [WIP] Support auto completion for one transformer decoder layer * [WIP] Add a minor change * [WIP] Fix a bug within the uint test * Shard XShape tensor, add embedding completion and refactor code * Add the distributed_operators dir to setup.py.in * Improve the completion process and add the unittest for gpt * fix process_mesh ut * fix process_mesh ut * update * update, test=develop * Add support for automatically completing distributed attrs of special ops * update * update * update * fix doc sample codes, test=develop * improve coverage, test=develop * add static_mode check, test=develop * Model the cluster for cost model and physical mapping * update, test=develop * add set_placement, test=develop * Add the check to make sure the candidate tensors' size is great than zero * update doc, test=develop * update doc, test=develop * update doc, test=develop * update doc, test=develop * update, test=develop * Auto mark dist attrs annotated by user * update ndarray to nested list, test=develop * update, test=develop * Add auto-completion module for auto-parallel (based on PR#33804) * Remove unnecessary files * Remove unrelated files for the auto completion pr * Update the unit test to improve the coverage * Modify codes based on reviews * Minor changes for CI * Improve some codes based on new comments * Fix bugs caused by shallow copy in attributes.py * Imporve amend_distributed_attr_for_program in context.py * Other changes for weihang's comments * support shard reader * support shard reader * add parallel mode * update process mesh * add method to compute comm_group * implement dist_embedding forward func * implement dist matmul forward func * implement dist reshape forward func * add transpiler framework * add transpiler forward * implement transpiler forward * implement transpiler backward & update * add process * add unitest * chmod * chmod * chmod * update unitest * add unitest for gpt * remove unused print * rename transpiler --> partitioner * rename transpiler --> partitioner * chmod * chmod * bug fixed * remove amp function * update case for dp mode * update case for dp mode * [Auto Parallel] Integrate all parts with the newest code * Integrate all parts of auto parallel and improve codes * Integrate all parts by AutoParallelizer * Add unit test for AutoParallelizer * Improve auto completion module for pipeline parallel * Add support for matmul_v2 in dist_matmul * Correct the typo "stratergy" to "strategy" * Modify distributed_strategy.proto to conform the main stream * Restore parts of distributed_strategy to conform the develop branch Co-authored-by: Nsandyhouse <lilong12@baidu.com> Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
-
由 will-jl944 提交于
multiply supports bool
-
由 WangXi 提交于
-
由 niuliling123 提交于
-
由 wanghuancoder 提交于
* refactor new executor, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop
-
由 CtfGo 提交于
* mark WhileOp AsExtra attribute * revert kX and kOutputs
-
由 niuliling123 提交于
-
由 Leo Chen 提交于
* add clip_by_norm fp16 kernel * add ut
-
由 Shang Zhizhou 提交于
* update slice plugin * add test * fix code style * fix trt6 * update test * fix test * add timeout * update trt version * update cmake
-
由 xiongkun 提交于
* can pass the fake test * add files * modify cmake to pass windows-ci * for ci pass * WITH_GLOO=ON * for pass coverage test * add cpuonly testcase * add * disable nccl when compile with cuda * change python version in cpuonly * add backend argument * add required gpu * add required:gpu
-
由 hong19860320 提交于
-
由 Guoxia Wang 提交于
-
由 Zeng Jinle 提交于
* fix scatter_add_nd and gather bug * fix gather compile error
-
由 Zeng Jinle 提交于
* add fleet api for program pass * turn on apply pass for CI test * fix disable fuse_all_optimizer bug * try to test ci * fix CI * fill unspecified op role * fix fuse_allreduce * add ut to improve coverage * remove useless change * improve c++ coverage * follow some comments * test ir pass pipeline * update doc * reduce ut time again
-
由 zhangkaihuo 提交于
The bug is that access to mean and var is incorrect, and the array will be out of bounds: the shape of mean and var is [batch_size], and the range of thread idx is 0~feature_size, so mean[idx] and var[idx] is incorrect. When batch_size=1, the correct access is mean[0] and var[0], and a unit test with batch_size=1 is added.
-
由 cc 提交于
-
由 lilong12 提交于
* update, test=develop
-
由 lilong12 提交于
* update, test=develop
-
由 feng_shuai 提交于
* merge CMakeList.txt manual * add platform for changethreadnum * repair some bugs according to make error * do nothing just flush CI * forget change thread num * add inplace_atol param for check_output_with_place * Windows * std:min and std::max should be change because of windows
-
由 lilong12 提交于
* support weight sharing
-
由 feng_shuai 提交于
* add some file for jetson-op-test-CE * change diff for some case * rename op_test -> jetson_infer_op * add some case because too large
-
由 Leo Chen 提交于
* release gil before op run * support npu grad test * fix op_test
-
由 Zhong Hui 提交于
-
由 liutiexing 提交于
* Split Tracker and WorkQueue * add WorkQueueGroup * add unittest * fix * update * update * fix compile
-
由 wawltor 提交于
* add the matmul v2 grad kernel * relief the test case time * update the test case for the matmul double grad * remove the unsed code for the matmul double grad * update the test case for the double grad matmul * remove the unused code in dot
-
由 WangXi 提交于
-
- 07 9月, 2021 8 次提交
-
-
由 Zeng Jinle 提交于
* fix scatter_nd_add doc, test=document_fix * update test=document_fix
-
由 xiegegege 提交于
* Add depth while cloning benchmark code,test=document_fix * Add depth while cloning benchmark code, test=document_fix
-
由 yaoxuefeng 提交于
-
由 wangxinxin08 提交于
* add conv op check for illegal input or attributes
-
由 niuliling123 提交于
-
由 Pei Yang 提交于
-
由 Qi Li 提交于
* [NPU] update batch norm op, test=develop * add NHWC support for bn, test=develop
-
由 XiangGao 提交于
Co-authored-by: Nroot <root@bjyz-sys-gpu-kongming9.bjyz.baidu.com>
-