- 21 11月, 2022 1 次提交
-
-
由 PuQing 提交于
-
- 18 11月, 2022 12 次提交
-
-
由 Sławomir Siwek 提交于
* cleanup unused code * unify is_int8 is_bfloat16 * Simplify matmul_v2 FWD kernel * remove RunKernel methods * remove import namespace * remove headers * clean fluid/phi cross imports * remove fluid axpy_handler * delete fluid methods * activations * OneDNNMemDesc * MKLDNNFormatForSize * MatchShapeToLayout * MKLDNNMemoryFormat * MKLDNNFormat * ReorderMKLDNNHandler * to_void_cast * review suggestions * interpolate * remove fluid depedency * init * ExecuteMatMulV2 * rm fluid kernel * matmul_grad * remove mutable_data
-
由 Zuza Gawrysiak 提交于
* Migrate conv_transpose to phi * Move handler to kernel * kernel m * Fix formatting * handler * remove fluid * revert tcp_store * tcp_store * remove unused * Fix declaration * add dnn input * Fix typo Co-authored-by: NSławomir Siwek <slawomir.siwek@intel.com>
-
由 zyfncg 提交于
* fix bug of zero_allocator in host * fix test compile bug * add unittest * update test
-
由 MarDino 提交于
* Add quick gelu and fused bias add kernel * fix annotation * remove useless code * add fast gelu option and set it in multi transformer op * add flag to restrict if use fast gelu approximate * fix flags conflict * fix use tanh function instead * add cudart version limit * use phi fast tanh func * fix comment
-
由 huangjiyi 提交于
* move "paddle/phi/backends/gpu/gpu_device_function.h" to phi * update copyright years * rm "fluid/platform/device/gpu/gpu_device_function.h" in phi * fix rocm-complie bugs
-
由 james 提交于
* correct sync behavior for XPU distributed training XPU support event mechanism similar to cuda event, so it is advisable to use an event to sync compute/comm streams for performance. However this mechanism is never fully tested, and inconsistent loss/ending_epochs are reported. Therefore, this PR replaces event sync with stream waiting as a temporary solution. * remove compile warning
-
由 Tian Zheng 提交于
* Refactor conv_kernel and conv_grad_kernel to provide interface for CUDNNv8 implementation * Fix macro * Add implementation for conv_kernel and conv_grad_kernel * Modification after rebase onto latest develop * Modify plan cache to comply with the API of phi::autotune * Refactor to reduce duplicate code * Review fix: - move functions in conv_kernel_impl_v8.h and conv_grad_kernel_impl_v8.h to conv_kernel.cu and conv_grad_kernelk.cu - add const specifier for input tensor - add logging when plans fail to execute - move CudnnConvBwdFilterV8 and CudnnConvBwdDataV8 to conv_cudnn_frontend.h * - move plan building outside of cache * Fix ROCM build
-
由 Yuang Liu 提交于
-
由 Wang Xin 提交于
* remove "gpu_primitives.h" in fluid namespace * fix PR-CI-GpuPS fail * fix PR-CI-GpuPS fail
-
由 zhangyikun02 提交于
-
由 Sylwester Fraczek 提交于
-
由 huangjiyi 提交于
-
- 17 11月, 2022 13 次提交
-
-
由 Qi Li 提交于
* [NPU] add _npu_identity op and api, test=develop * fix doc * address comments
-
由 xiongkun 提交于
-
由 huangjiyi 提交于
-
由 YuanRisheng 提交于
* standard api * fix xpu bugs
-
由 Mountagha 提交于
-
由 taixiurong 提交于
-
由 Wang Xin 提交于
-
由 Yiqun Liu 提交于
* Implement a common dims simplifier. * Fix the include position error. * Reduce the cpu overhead of broadcast computing.
-
由 huangjiyi 提交于
-
由 huangjiyi 提交于
* rm "paddle/fluid/operators/math.h" in phi * rm "paddle/fluid/operators/math.h" in fluit
-
由 Yuang Liu 提交于
Support bfloat16 for adamw and adam optimizer. Fit the lr for pure bf16 training with tensor fusion. (#48041) * add bfloat16 for adamw * set lr not to bfloat16 for pure bf16 training * update the logic * update the adamw optimizer * support bfloat for adam
-
由 sneaxiy 提交于
* add vectorized bfloat16 atomicAdd * fix compile error * fix compile error again * fix V100 compile error * fix V100 compile again
-
由 zyfncg 提交于
-
- 16 11月, 2022 7 次提交
-
-
由 huangjiyi 提交于
-
由 Qi Li 提交于
* [NPU] update npu prop, test=develop * remove ddim.h * remove diff * update storage prop, test=develop
-
由 HongyuJia 提交于
* simplify depthwise_conv2d phi kernel selection * fix depthwise_conv2d
-
由 Piotr Paturej 提交于
* Enable bf16 in oneDNN bilinear_interp kernel * Fix bilinear_interp_v2 not enabled in models * Remove unnecessary checks
-
由 ykkk2333 提交于
* add stat tool * add roll and roll_grad kernels and strided_slice and strided_slice_grad kernels, test=kunlun * embedding and embedding_grad add int32 input, test=kunlun
-
由 Leo Chen 提交于
-
由 Wang Xin 提交于
-
- 15 11月, 2022 7 次提交
-
-
由 sneaxiy 提交于
-
由 HongyuJia 提交于
* opt error message when selecting kernels under phi * fix for loop * polish error message * polish error message, split into 3 error condition * polish error message
-
由 Yuang Liu 提交于
-
由 zhouweiwei2014 提交于
-
由 Sławomir Siwek 提交于
* cleanup unused code * unify is_int8 is_bfloat16 * Simplify matmul_v2 FWD kernel * remove RunKernel methods * remove import namespace * remove headers * clean fluid/phi cross imports * remove fluid axpy_handler * delete fluid methods * activations * OneDNNMemDesc * MKLDNNFormatForSize * MatchShapeToLayout * MKLDNNMemoryFormat * MKLDNNFormat * ReorderMKLDNNHandler * to_void_cast * review suggestions * interpolate * remove fluid depedency
-
由 huangjiyi 提交于
* rm "paddle/fluid/platform/complex.h" in phi * fix codestyle with pre-commit
-
由 Wang Xin 提交于
-