- 10 4月, 2023 15 次提交
-
-
由 gouzil 提交于
* add autogen code bilinear_tensor_product * [phi] rm cc file
-
由 cyberslack_lee 提交于
-
由 lzydev 提交于
* autogen softmax_with_cross_entropy * fix error in softmax_with_cross_entropy version
-
由 HongyuJia 提交于
* [enforce.h Decouple gflags.h] Move gflags.h from enforce.h to enforce.cc * Add gflags.h for other files * Add gflags.h for other files * Add gflags.h for blas_impl.hip.h * Add gflags.h for miopen_helper.h
-
由 Vvsmile 提交于
* adjust defalut tolerance of output and grad * fix a bug in the grad of OpTest * fix the type of setting defalut value in optest, both forward and backward * add defalut * fix test_sum_op * adjust tolerance * fix the tolerance of eager * add bf16 and fp16 to the activation tests * remove some fixs * fix activation * fix fp16 * fix gelu * fix the activation tests * add bfloat16 specialization to singrad and cosgrad * fix bugs * fix bugs * add unittest * add skip * add fp/bf to rrelu/rrelu_grad * git add rrelu * fix bugs
-
由 qizhaoaoe 提交于
* add fp16 and bf16 support for instance_norm * fix /= operator which not support bf16 * fix instance_norm_grad kernel and unittests. * fix fp32 unittests. * fix instance_norm_kernel and unittests. * fix instance_norm_grad_kernel and unittest threshold. * add fp16/bf16 for instance_norm_grad_grad op. * add bf16 dtype check. * fix conflicts. * fix cpu support for fp32 op and fix type in instance_norm_grad_kernel. * fix type in instance_norm_kernel. * fix bf16 outputs in unittests and refine codes. * fix dx computation. * delete unuseful params and head including. * add fp16/bf16 for static graph. * fix device condiction for instance_norm op. * fix instance_norm_grad_grad and bf16 op tests. * fix op_test to support grad of bf16 can be compared with fp32. * remove updates. * add self-defined grad.
-
由 Wang Xin 提交于
-
由 jjyaoao 提交于
-
由 Zero Rains 提交于
* fix divide zero bug for softmax_with_cross_entropy * change the single test way * can run but slow. the most important is that I do not know why it slow * remove some useless commet * change the copyright to correct * remove some useless change * if repeat_times == 1, we will not use BroadcastKernel
-
由 cyberslack_lee 提交于
-
由 Asthestarsfalll 提交于
* Optimize the performance of logsumexp * Support zero-dim tensor
-
由 lishicheng1996 提交于
-
由 zyfncg 提交于
-
由 Wang Xin 提交于
* add autogen code support for affine_grid op * update op_compat.yaml for affine_grid * update op_compat.yaml for affine_grid * fix AffineGridGradInferMeta * fix CI error * update AffineGridInferMeta
-
由 Galaxy1458 提交于
* delete [-Wno-error=terminate], test=develop * remove GPUps[-Wterminate],test=develop * remove some -Wno-, test=develop * modify ~MatmulDescriptor * mess
-
- 09 4月, 2023 4 次提交
-
-
由 ronnywang 提交于
* [PHI CAPI] support complex dtype kernel * update
-
由 Chitsing KUI 提交于
-
由 shaojie_wang 提交于
-
由 scotty 提交于
* add autogen code support for matrix_nms. * update
-
- 08 4月, 2023 1 次提交
-
-
由 RedContritio 提交于
-
- 07 4月, 2023 7 次提交
-
-
由 lishicheng1996 提交于
-
由 Ruibiao Chen 提交于
* Isolate DenseTensor::set_type from header file * Fix selected_rows
-
由 Zhenghai Zhang 提交于
-
由 TaoTao Li 提交于
fix merge conflicts
-
由 RedContritio 提交于
-
由 Wang Xin 提交于
-
由 engineer1109 提交于
-
- 06 4月, 2023 13 次提交
-
-
由 yuehuayingxueluo 提交于
-
由 Sławomir Siwek 提交于
* replace matmul with matmul_v2 in fuse passes * Remove fusion logic from matmul * removing fusion methods * add proper name * adjust namespaces * clean attrs in python tests * delete checkpoint and restore matmul version * remove unused code * matmul and reshape/transpose fuses migrated * split MatmulOneDNN headers * fuse activation and eltwise_add * add fuse_activation * matmul_transpose_reshape/reshape_transpose_matmul * matmul + elementwise_add (fused) * activation temporary modifciation * restore matmul(v1) version 0 * merge newest develop * remove depedency from other PR * revert pbtxt * remove placeholders from matmul_v2 * add description in OPMaker * remove matmul_v2_op.h and all depedencies * remove dims changing in base op * add possibility to fuse already fused_matmul * restart broken CI * Empty-Commit * revert matmul_utils.h * codestyle * adjust imports * add pbtxt file * 100% matmul unit tests coverage * trigger CI with minimal changes to develop * adjust changes to develop * add fused_matmul op * inherit base ops * add "v2" * move OPMaker * Gradually add fused_matmul files * second batch of fused_matmul changes * split infershapes of matmul_v2 and fused_matmul * merge code from other PR * 2023 * inherit fused_matmul from matmul_v2 * Update paddle/phi/backends/onednn/onednn_reuse.h Co-authored-by: NTomasz Socha <tomasz.socha@intel.com> * Update paddle/phi/kernels/fusion/onednn/fused_matmul_kernel.cc Co-authored-by: NTomasz Socha <tomasz.socha@intel.com> * resolve conflicts * codestyle * simplify isgemmlinear * 2023 * remove import * reuse methods * matmul_v2_mkldnn cleanup * simplify ExecuteMatMulV1Grad * matmul refactored * fc * SetOutMemDescWithLogicalLayoutFusesSupport * matmul_v2 * alpha support * group repetetive funcs * matmul utils * execute matmul methods * restore registered kernel names * split header and impl files * remove double negatives * reduce numer of modified files * adjust ExecuteMatmul * add scales for ut * dates * limit number of modified files * fluid imports * remove alpha * codestyle --------- Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
-
由 Sonder 提交于
* add kernel functions * update kernel functions * update func parameters' name * create codes for gpu device * 调整文件位置 * fix include error * remove dependent files to phi/ * restore fused_attention_op.cu * fix dependence errors * fix dependence errors * fix include error * fix all depandence errors[build success] * remove useless include * recover useless include * use phi::ToNCCLDataType * fix namespace * update new register code * fix error in fused_gemm_epilogue_utils * fix error in FusedAttentionKernel parm * finish fused_attention registe code[build success] * add paddle::optional * add sig file * fix build error * fix a include error * update CMkaeList * fix parameter sequence * add include file * update #if before include * fix grammly error * update codes for DropoutParam * remove const cast * trans some fluid api to phi api * add #if * update test code * update test codes * recover test codes * trans fused_attention to fluid * move #endif to end * move #endif * delete useless files * use fused attention utils and recover random seed * remove fluid include in phi
-
由 scotty 提交于
-
由 RedContritio 提交于
-
由 RedContritio 提交于
-
由 张春乔 提交于
-
由 Jiabin Yang 提交于
-
由 Kang Zhao 提交于
* feat: add relu composite rule * feat: add relu composite rule, maximum op * feat: add relu composite rule, maximum op * feat: add relu composite rule, polish comments * feat: add relu composite rule, polish comments * feat: add relu composite rule, add python api of relu * feat: add relu composite rule, commit hook * fix: maximum type error & ban cinn test * fix: maximum input sequence bugs * resolve conflicts * fix: code style bugs * add: relu fp16 test * feat: add rsqrt composite rule * feat: add rsqrt composite rule * resolve conflicts of composite rule * fix: delete check eager * feat: add roll grad composite rule * fix minus shift * fix test roll op
-
由 zhangyuqin1998 提交于
* Rename conv2d transpose grad grad * fix
-
由 Chitsing KUI 提交于
-
由 sneaxiy 提交于
* fix flash attn * fix another API
-
由 zyfncg 提交于
* update readme * remove unused header file * fix bug * fix onednn * fix onednn * rename header file
-