- 20 10月, 2022 7 次提交
-
-
由 zhoutianzi666 提交于
* stride_to_24 * fix CI failing
-
由 JingZhuangzhuang 提交于
-
由 Wang Bojun 提交于
* Enhance the layernorm shift partation fuse op when shift size > 0 (roll shifting) * fix cherry-pick test
-
由 JingZhuangzhuang 提交于
-
由 sneaxiy 提交于
Fix some operators when the tensor.numel() > INT32_MAX
-
由 sneaxiy 提交于
support pure bfloat16 for more ops
-
由 Wen Sun 提交于
* fix: fix incorrect import * fix: fix incorrect usage
-
- 19 10月, 2022 9 次提交
-
-
由 zhaoyingli 提交于
* [Auto Parallel] Make Engine class callable (#46416) * [Auto Parallel] Imporve the user-defined fetches and logging * [Auto Parallel] Make Engine class callable * [Auto Parallel] Update the data loading of tuner * Print IPS in auto parallel Engine (#46554) * [AutoParallel] fix dist_split (#46505) * [AutoParallel] fix dist_split * add unittest * update cmakelist * [AutoParallel] fix sharding (#46572) * [AutoParallel] fix process_mesh (#46583) * [AutoParallel] fix reshard when train with eval (#46605) * [AutoParallel] fix reshard when train with eval * fix mppp * [AutoParallel] fix amp when predict (#46637) * [Auto Parallel]Update comp cost and completion for gpt auto search (#46387) * update comp cost and completion for gpt auto search * add unittest * [Auto Parallel] Fix bugs caused by the inconsistent outputs of Engine API (#46633) * [Auto Parallel] Unify the logger and outputs of Engine API * [Auto Parallel] Fix the bugs of to_static * [Auto Parallel] Adjust the test_to_static.py * [Auto Parallel] Improve the fine-grained APIs (#46552) * [Auto Parallel] Suppport different dataloaders * [Auto Parallel] Add num_shards config for dataset * [Auto Parallel] Unify the logger and outputs of Engine API * [Auto Parallel] Fix the bugs of to_static * [Auto Parallel] Adjust the test_to_static.py * [Auto Parallel] Add the prepare API and replace __call__ with run * [Auto Parallel] Improve the private implementations of Engine * [Auto Parallel] Set capacity of dataloader for opt tuning * [Auto Parallel] [WIP] Change the fine-grained API * [Auto Parallel] Improve APIs to support different user cases * [Auto Parallel] Add removed config * [Auto Parallel] Add imports * [Auto Parallel] Fix bugs for to_static * [Auto Parallel] Remove unnecessary imports * bugfix (#46921) * [Auto Parallel] Fix the bug for None labels (#46987) * [AutoParallel] adapt for gpt-gen (#46771) * for gpt-gen * fix reshard * adapt assign and shape op * add dist_assign & unittest * add conditional block unittest * rename unittest * [Auto Parallel] Fix the bug of completion (#47056) * [Auto Parallel] Fix the bug for None labels * [Auto Parallel] Fix the completion bug * [AutoParallel] add callbacks (#47014) * [AutoParallel] add callbacks * fix unittest * fix dist_context * fix engine * fix cmakelist * fix unittest's returns * fix cmakelist * [Auto Parallel] Add cost interface (#47043) * add cost interface * update inferface and add unittest * update unittest * update inferface * [Auto Parallel]Add parallel tuner (#46189) * add parallel tuner * add unittest * fix unittest * set timeout of unittest * set unittest timeout * fix auto_mode setting * update unittest * sync from develop and update unittest * remove unused import * update unittest * update cmakelist * add unittests Co-authored-by: NYulong Ao <aoyulong@baidu.com> Co-authored-by: NRuibiao Chen <chenruibiao@baidu.com> Co-authored-by: Ncaozhou <48191911+Caozhou1995@users.noreply.github.com> Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
-
由 Zhang Ting 提交于
* strided_slice grad add fp16 support
-
由 yeliang2258 提交于
* Add unsigned int8 propagation * Add or modify unit tests * Correct concat scale checking * Apply review suggestions * Corrections Co-authored-by: Njoanna.wozna.intel <joanna.wozna@intel.com>
-
由 Aurelius84 提交于
Polish @to_static temporary file directory to speed up transformation
-
由 xiongkun 提交于
* [Dy2Static] Support TypeHint for function decorated by @to_static (#47121) * Add TypeHint Transformer * add unittest for typehint transformer * [Dy2Static] Remove GradTransformer (#47063) * [Dy2Static] Remove GradTransformer 1. fix einsum infershape bugs. 2. remove grad_transformer and unify paddle.grad and paddle.static.gradient. 3. add dygraph_and_dy2static_only decorator for dy2static. * fix bugs * rename
-
由 Ghost Screaming 提交于
* Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result is wrong. * Support allow_partial switch, which can be configure in pipeline_configs. If sent tensor are not the same from different hosts, they shouldn't been sent partially and then concated as a whole tensor. * Change name allow_partial to enable_partial_send_recv. * Add global variable _enable_partial_send_recv
-
由 WangZhen 提交于
[CherryPick][Dy2St]Fix recurrent op eager deletion pass error in dy2st
-
由 YangZhou 提交于
* update audio api examples * fix format * format * fix * test api * fix format * fix static check error * fix doc error * fix ci * fix api error * update api.spec * fix ci * fix typo in window gaussian
-
由 Hui Zhang 提交于
Construct exec and ctx only once in cond op to speed up
-
- 18 10月, 2022 7 次提交
-
-
由 Wilber 提交于
-
由 weishengying 提交于
Add symbolic shape deduction function for unfold, scatter_nd_add, p_norm, grid_sampler, pad3d, etc (#46291) (#47003)
-
由 zhouweiwei2014 提交于
新增sparse.is_same_shape、sparse.reshape、sparse.transpose 三个API
-
由 zhoutianzi666 提交于
-
由 Yuang Liu 提交于
* [dygraph sharding] Overlap the reduce and the caculation for sharding stage 2. (#46495) * [dygraph sharding stage 2] sharding broadcast overlap (#46656) * Multi groups for broadcast of sharding stage 2 (#46894)
-
由 Haohongxiang 提交于
* [Dygraph] Fix performance of pp+mp by using send/recv_calc_stream instead of send/recv (#46116) * [Dygraph] Fix Perf of FusedFeedForward and FusedAttention with AllReduce (#46780) * update
-
由 Wang Bojun 提交于
* draft with debug print * remove debug print * bug fix for ci
-
- 17 10月, 2022 9 次提交
-
-
由 Wen Sun 提交于
* Support both use_calc_stream and sync_op in send recv APIs (#46023) * Support both use_calc_stream and sync_op in allgather API (#46295) * Support both use_calc_stream and sync_op in collective communication API (#46761) * Move group and all reduce from collective to communication (#45848) * Completes bfloat16 dtype for collective api in eager mode (#45844) * Fix collective APIs cannot be recognized when building docs (#46962) Co-authored-by: NLiYuRio <63526175+LiYuRio@users.noreply.github.com>
-
由 zhangkaihuo 提交于
cherry-pick : #46322, #46245 Sparse API 支持静态图
-
由 Zhang Zheng 提交于
Optimize performance of depthwise_conv Config: input[2048, 1024, 4, 4], filter[1024, 1, 4, 4], stride=1, pad=0, dilation=1
-
由 Guanghua Yu 提交于
* fix dygraph new format quant * fix unittest * fix conflict
-
由 Allen Guo 提交于
-
由 Allen Guo 提交于
-
由 Allen Guo 提交于
* paddle-inference support custom-ops Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> * fix tolower Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai>
-
由 Allen Guo 提交于
-
由 Zhang Zheng 提交于
为了提升性能,将label的边界检查从python端转移到kernel内,减少额外op的调用,如min、max和同步拷贝等 当前的模板参数IgnoreIndex仅在ignore_index取值范围在[0, dim)时才生效,但是当某个label值超出了边界,ignore_index等于该label,这种情况下是应该仍然能正常计算。虽然当前的计算逻辑在结果上不会出错,但逻辑上仍是有问题的,且模板参数IgnoreIndex是没有必要的
-
- 14 10月, 2022 8 次提交
-
-
由 xiaoxiaohehe001 提交于
-
由 Wilber 提交于
-
由 Guanghua Yu 提交于
-
由 xiaoxiaohehe001 提交于
-
由 Aurelius84 提交于
-
由 Aurelius84 提交于
* [BUG]Fix expand_as_v2 bug while X and Y with different dtype * fix commit
-
由 Zhang Jun 提交于
* fix reshape2 opteller; add elementwise min/max register for tensorrt
-
由 zhoutianzi666 提交于
-