- 09 2月, 2023 1 次提交
-
-
由 zhangyikun02 提交于
-
- 01 2月, 2023 1 次提交
-
-
由 zhangyikun02 提交于
-
- 31 1月, 2023 1 次提交
-
-
由 wangshengxiang 提交于
-
- 19 1月, 2023 1 次提交
-
-
由 jameszhang 提交于
* [KUNLUN] add op: maxpool_with_index * use DeviceContext::Alloc() instead of DenseTensor::mutable_data() * fix file format * solve clip unittest failure * minor fix * Revert "solve clip unittest failure" since the issue is fixed in #49535 This reverts commit 1127adc66e79afe35ac3c00bb34e6aaa7cd7d78b. * align with xdnn on the definition of mask in max_pool_with_index * minor
-
- 18 1月, 2023 3 次提交
-
-
由 RuohengMa 提交于
* add reduce_sum_int64 and reduce_sum_int8 xpu kernels * [PHI] add clip grad kernel with support type float32 and int32 * [PHI unittest] add clip_grad unit test * adapt code to clang-format * update xpu api output with clip_grad api * remove int8 support of reduce_sum xpu kernel since it can not pass unit tests * adapt license date, add code for XPUDataType convertion * add int8 support of reduce_sum * add reduce_sum unit tests for dtype int64, int8, and add more test cases * update license date * remove buggy bitwise and, or and xor xpu kernels, refine bitwise not xpu kernel * change license date
-
由 houj04 提交于
-
由 jameszhang 提交于
* revert to use default XPU stream for computing XPUContext now has a null stream by default. If you want to use a separate stream (e.g. in async collective communication), you should create a dedicated XPUContext and invoke its XPUContext::CreateStream() * minor
-
- 16 1月, 2023 1 次提交
-
-
由 QingshuChen 提交于
-
- 13 1月, 2023 4 次提交
-
-
由 jameszhang 提交于
* kunlun add support for c_concat and c_split * replace mutable_data() and ShareDataWith()
-
由 ykkk2333 提交于
-
由 jameszhang 提交于
* fix xpu unittest issue: zero_dim_tensor * deal with leftout issue introduced by #49470
-
由 wangshengxiang 提交于
-
- 12 1月, 2023 3 次提交
-
-
由 YuanRisheng 提交于
-
由 Leo Guo 提交于
xpu2_op_list.cc. test=kunlun
-
由 YuanRisheng 提交于
* rename kernel * delete sig * modify code according comment * fix ci bugs
-
- 09 1月, 2023 2 次提交
-
-
由 QingshuChen 提交于
-
由 ykkk2333 提交于
* migrate shaple sgd, split,sign xpu kernels to phi, test=kunlun * fix dlrm throughput problem, test=kunlun * add xpu einsum, fill_diagonal, and diagonal kernels, test=kunlun
-
- 06 1月, 2023 1 次提交
-
-
由 RuohengMa 提交于
* add bitwise and, bitwise not, bitwise or and bitwise xor * correct typo
-
- 27 12月, 2022 1 次提交
-
-
由 zhangyikun02 提交于
-
- 26 12月, 2022 1 次提交
-
-
由 ykkk2333 提交于
* migrate shaple sgd, split,sign xpu kernels to phi, test=kunlun * fix dlrm throughput problem, test=kunlun
-
- 23 12月, 2022 1 次提交
-
-
由 haosicheng 提交于
-
- 22 12月, 2022 1 次提交
-
-
由 QingshuChen 提交于
-
- 19 12月, 2022 1 次提交
-
-
由 zhangyikun02 提交于
-
- 14 12月, 2022 1 次提交
-
-
由 james 提交于
* nullptr bugfix for XPU pg mode Also a few kernels is added to xpu whitelist * increase error msg length
-
- 09 12月, 2022 3 次提交
-
-
由 james 提交于
This is a PR to catch up with latest xpu white list strategy (https://github.com/PaddlePaddle/Paddle/pull/48606) , since original list only include 'fluid' fashion names, but new list must include 'phi' fashion as well. Refer to paddle/phi/core/kernel_factory.cc for more details.
-
由 haosicheng 提交于
-
由 PuQing 提交于
-
- 08 12月, 2022 3 次提交
-
-
由 haosicheng 提交于
-
由 houj04 提交于
* [XPU] add load op into oplist. * remove test_sampling_id_op_xpu.py
-
由 huangjiyi 提交于
* move cuda_graph from fluid to phi * move device_memory_aligment from fluid to phi * Revert "move device_memory_aligment from fluid to phi" This reverts commit b92fcd39a0a50fdac13278f49be0237a85f3a13f. * update xpu cmake
-
- 07 12月, 2022 2 次提交
-
-
由 QingshuChen 提交于
*test=kunlun
-
由 zhangyikun02 提交于
-
- 06 12月, 2022 1 次提交
-
-
由 QingshuChen 提交于
*test=kunlun
-
- 30 11月, 2022 1 次提交
-
-
由 james 提交于
some legacy code still use xpu_wait() for stream sync -- it only syncs default stream. this PR replaces them with dev_ctx.Wait() to ensure that correct stream is always used
-
- 28 11月, 2022 1 次提交
-
-
由 huangjiyi 提交于
* rm fluid “xpu_header.h” deps in phi * move part of xpu_op_list.h from fluid to phi * add fluid xpu_op_list deps * add glog deps for xpu_op_list in phi * fix PR-CI-Kunlun
-
- 25 11月, 2022 1 次提交
-
-
由 wanghuancoder 提交于
* for xpu multi thread bug test
-
- 24 11月, 2022 1 次提交
-
-
由 PuQing 提交于
-
- 18 11月, 2022 1 次提交
-
-
由 james 提交于
* correct sync behavior for XPU distributed training XPU support event mechanism similar to cuda event, so it is advisable to use an event to sync compute/comm streams for performance. However this mechanism is never fully tested, and inconsistent loss/ending_epochs are reported. Therefore, this PR replaces event sync with stream waiting as a temporary solution. * remove compile warning
-
- 10 11月, 2022 1 次提交
-
-
由 james 提交于
* XPU support eager mode * add unittest for XPU eager mode * minor bugfix * minor bugfix, test=kunlun * correct copyright info * 1. remove unsed vars/funcs 2. ProcessGroupBKCL inherit from ProcessGroupStream * bugfix for fp16 in eager mode multi-card, test=kunlun * rebase & fix a few issues * use new processgroup interface, test=kunlun * fix compile issue, test=kunlun
-
- 01 11月, 2022 1 次提交
-
-
由 Chen Weihang 提交于
* add extra attr property set * add type_info for all context * add onednn context to all context * fix context compile error * simplify conv kernel args * pass runtime attr into dev_ctx * fix marco error * clear conv_grad_kernel extra args * merge conv_grad_grad into conv_grad * clear conv2d_grad_grad extra attrs * clear yaml and eager extra attr * fix conv1d error * change to thread local * fix npu compile failed * try to fix windows compile failed * add conv2d onednn phi kernel * fix ci bugs (#36) * fix compile bugs (#38) * fix extra input transform bug (#39) * support dynamic created attr (#40) * reset extra info gen code * rm conv_grad_grad kernel * reimpl pass attr adapting * add int attr support * remove vector inputnames creating * fix map at error * Update paddle/phi/kernels/onednn/conv_grad_kernel.cc Co-authored-by: NSławomir Siwek <slawomir.siwek@intel.com> * remove useless extra attrs * replace mkldnn_engine by onednn_engine Co-authored-by: NYuanRisheng <yuanrisheng@baidu.com> Co-authored-by: NSławomir Siwek <slawomir.siwek@intel.com>
-