- 23 12月, 2022 1 次提交
-
-
由 haosicheng 提交于
-
- 22 12月, 2022 1 次提交
-
-
由 QingshuChen 提交于
-
- 20 12月, 2022 1 次提交
-
-
由 haosicheng 提交于
-
- 19 12月, 2022 1 次提交
-
-
由 zhangyikun02 提交于
-
- 14 12月, 2022 1 次提交
-
-
由 james 提交于
* nullptr bugfix for XPU pg mode Also a few kernels is added to xpu whitelist * increase error msg length
-
- 08 12月, 2022 1 次提交
-
-
由 haosicheng 提交于
-
- 07 12月, 2022 1 次提交
-
-
由 zhangyikun02 提交于
-
- 06 12月, 2022 2 次提交
- 05 12月, 2022 1 次提交
-
-
由 Ruibiao Chen 提交于
* Replace mutable_data with DeviceContext.Alloc in phi kernels * Fix CI errors * Fix CI errors * Fix CI errors, test=kunlun * Fix CI errors, test=kunlun * Handle rnn_functor * Update approvals
-
- 03 12月, 2022 1 次提交
-
-
由 Yuang Liu 提交于
-
- 02 12月, 2022 3 次提交
-
-
由 Jiabin Yang 提交于
* [Eager] Fix paddle.grad interface * [Eager] Support minimum SubGraph for GeneralGrad * Add needed_nodes to prune grad graph more thoroughly * [Eager] Add grad_node_trans_mapping_ to record which grad_node has been transformed to AccumulationNode * [Eager] Fix paddle.grad interface * Polish code * remove potential_stop_node * Add endding_nodes to enhance genSugraph logic * clear endding_nodes_ * polish code * rename endding_nodes to endding_nades_ * Refactor grad interface * Add register_hook case to fix coverage-ci * Fix code format * Refactor general_grad * Add more code comments * call clear directly to release GradSlotMeta * fix a mistake * fix matmul/ multiply kernel logic and optional input in yaml, fill zeros logic and so on. * fix batch_norm_double_grad yaml optional config * fix tanh_triple_grad yaml and kernels * fix MultiplyTripleGradKernel optional logic * fix merge mistake * fix compile error * remove legacy attr for bn * polish code * fix some kernel * merge develop * fix error * remote log * fix kernel with full like * hide value log behind * hide value log behind * fix matmul_triple grad * fix xpu compile error * fix xpu compile error * fix xpu ut * fix xpu ut * fix_xpu_compile_error Co-authored-by: NWeilong Wu <veyron_wu@163.com>
-
由 Jiabin Yang 提交于
* [Eager] Fix paddle.grad interface * [Eager] Support minimum SubGraph for GeneralGrad * Add needed_nodes to prune grad graph more thoroughly * [Eager] Add grad_node_trans_mapping_ to record which grad_node has been transformed to AccumulationNode * [Eager] Fix paddle.grad interface * Polish code * remove potential_stop_node * Add endding_nodes to enhance genSugraph logic * clear endding_nodes_ * polish code * rename endding_nodes to endding_nades_ * Refactor grad interface * Add register_hook case to fix coverage-ci * Fix code format * Refactor general_grad * Add more code comments * call clear directly to release GradSlotMeta * fix a mistake * fix matmul/ multiply kernel logic and optional input in yaml, fill zeros logic and so on. * fix batch_norm_double_grad yaml optional config * fix tanh_triple_grad yaml and kernels * fix MultiplyTripleGradKernel optional logic * fix merge mistake * fix compile error * remove legacy attr for bn * polish code * fix some kernel * merge develop * fix error * remote log * fix kernel with full like * hide value log behind * hide value log behind * fix matmul_triple grad Co-authored-by: NWeilong Wu <veyron_wu@163.com>
-
由 ykkk2333 提交于
* add stat tool * add roll and roll_grad kernels and strided_slice and strided_slice_grad kernels, test=kunlun * add silu, unfold and their grads,test=kunlun
-
- 01 12月, 2022 2 次提交
-
-
由 zyfncg 提交于
* rename kernel for top_k, slogdeterminant, generate_proposals_v2 * fix bug
-
由 zhangyikun02 提交于
-
- 30 11月, 2022 1 次提交
-
-
由 zhangyikun02 提交于
-
- 29 11月, 2022 1 次提交
-
-
由 haosicheng 提交于
-
- 28 11月, 2022 3 次提交
-
-
由 huangjiyi 提交于
* rm fluid “xpu_header.h” deps in phi * move part of xpu_op_list.h from fluid to phi * add fluid xpu_op_list deps * add glog deps for xpu_op_list in phi * fix PR-CI-Kunlun
-
由 zyfncg 提交于
* add fluid_op_name_map * rename some kernel name * add comments for op-kernel map * refine map name of op to kernel
-
由 haosicheng 提交于
-
- 24 11月, 2022 2 次提交
-
-
由 zhangyikun02 提交于
-
由 zhangyikun02 提交于
-
- 23 11月, 2022 2 次提交
-
-
由 ykkk2333 提交于
* add stat tool * add roll and roll_grad kernels and strided_slice and strided_slice_grad kernels, test=kunlun * add masked_selected_grad kernel,test=kunlun
-
由 zhangyikun02 提交于
-
- 21 11月, 2022 2 次提交
-
-
由 wanghuancoder 提交于
* refine reduce_all
-
由 taixiurong 提交于
-
- 18 11月, 2022 2 次提交
-
-
由 james 提交于
* correct sync behavior for XPU distributed training XPU support event mechanism similar to cuda event, so it is advisable to use an event to sync compute/comm streams for performance. However this mechanism is never fully tested, and inconsistent loss/ending_epochs are reported. Therefore, this PR replaces event sync with stream waiting as a temporary solution. * remove compile warning
-
由 zhangyikun02 提交于
-
- 17 11月, 2022 2 次提交
-
-
由 YuanRisheng 提交于
* standard api * fix xpu bugs
-
由 taixiurong 提交于
-
- 16 11月, 2022 1 次提交
-
-
由 ykkk2333 提交于
* add stat tool * add roll and roll_grad kernels and strided_slice and strided_slice_grad kernels, test=kunlun * embedding and embedding_grad add int32 input, test=kunlun
-
- 15 11月, 2022 1 次提交
-
-
由 zhouweiwei2014 提交于
-
- 11 11月, 2022 1 次提交
-
-
由 zhouweiwei2014 提交于
-
- 10 11月, 2022 6 次提交
-
-
由 zhangyikun02 提交于
-
由 YuanRisheng 提交于
* standard api * fix sparse bugs * fix xpu bugs, test=kunlun * remove hard code for custom unittest * open ci, test=kunlun * deal with conflict
-
由 Wang Xin 提交于
* remove fluid/framework/generator.h from phi * fix PR-CI-Kunlun-KP-Build fail
-
由 huangjiyi 提交于
[PHI Decoupling] remove "paddle/fluid/platform/float16.h" and "paddle/fluid/platform/for_range.h" in phi. (#47817) * rm "paddle/fluid/platform/float16.h" in phi * rm "paddle/fluid/platform/for_range.h" in phi
-
由 zhouweiwei2014 提交于
-
由 james 提交于
* XPU support eager mode * add unittest for XPU eager mode * minor bugfix * minor bugfix, test=kunlun * correct copyright info * 1. remove unsed vars/funcs 2. ProcessGroupBKCL inherit from ProcessGroupStream * bugfix for fp16 in eager mode multi-card, test=kunlun * rebase & fix a few issues * use new processgroup interface, test=kunlun * fix compile issue, test=kunlun
-