- 11 4月, 2023 8 次提交
-
-
由 RedContritio 提交于
* support auto generate static for randperm * remove enforce in randperm infermeta
-
由 zhangyuqin1998 提交于
-
由 wangzhen38 提交于
-
由 RedContritio 提交于
* fix error in generator/type_mapping.py * support auto generate for op merged_momentum optimizer
-
由 RedContritio 提交于
* support auto generate for flatten (flatten_contiguous_range) * add data_type for flatten_grad
-
由 Wang Xin 提交于
* add autogen code support for reverse op * bug fixed
-
由 RedContritio 提交于
-
由 RedContritio 提交于
* support auto generate for op momentum optimizer * remove momentum_op.* and update signature * fix dgc momentum op maker error
-
- 10 4月, 2023 9 次提交
-
-
由 lzydev 提交于
* autogen segment_pool * delete legacy_dygraph about segment_pool
-
由 HongyuJia 提交于
* [Opt Performance] Optimize custom operator performance, reconstruct python API auto-gen, add cache and use const inference * opt AutoGradMeta implementation * remove profiler codes * fix unit test * change year, 2021->2023 * fix int64_t parse bug
-
由 gouzil 提交于
* add autogen code bilinear_tensor_product * [phi] rm cc file
-
由 lzydev 提交于
* autogen softmax_with_cross_entropy * fix error in softmax_with_cross_entropy version
-
由 HongyuJia 提交于
* [enforce.h Decouple gflags.h] Move gflags.h from enforce.h to enforce.cc * Add gflags.h for other files * Add gflags.h for other files * Add gflags.h for blas_impl.hip.h * Add gflags.h for miopen_helper.h
-
由 Wang Xin 提交于
-
由 cyberslack_lee 提交于
-
由 zyfncg 提交于
-
由 Wang Xin 提交于
* add autogen code support for affine_grid op * update op_compat.yaml for affine_grid * update op_compat.yaml for affine_grid * fix AffineGridGradInferMeta * fix CI error * update AffineGridInferMeta
-
- 09 4月, 2023 2 次提交
-
-
由 Chitsing KUI 提交于
-
由 scotty 提交于
* add autogen code support for matrix_nms. * update
-
- 08 4月, 2023 1 次提交
-
-
由 RedContritio 提交于
-
- 07 4月, 2023 4 次提交
-
-
由 Zhenghai Zhang 提交于
-
由 TaoTao Li 提交于
fix merge conflicts
-
由 RedContritio 提交于
-
由 engineer1109 提交于
-
- 06 4月, 2023 11 次提交
-
-
由 Sławomir Siwek 提交于
* replace matmul with matmul_v2 in fuse passes * Remove fusion logic from matmul * removing fusion methods * add proper name * adjust namespaces * clean attrs in python tests * delete checkpoint and restore matmul version * remove unused code * matmul and reshape/transpose fuses migrated * split MatmulOneDNN headers * fuse activation and eltwise_add * add fuse_activation * matmul_transpose_reshape/reshape_transpose_matmul * matmul + elementwise_add (fused) * activation temporary modifciation * restore matmul(v1) version 0 * merge newest develop * remove depedency from other PR * revert pbtxt * remove placeholders from matmul_v2 * add description in OPMaker * remove matmul_v2_op.h and all depedencies * remove dims changing in base op * add possibility to fuse already fused_matmul * restart broken CI * Empty-Commit * revert matmul_utils.h * codestyle * adjust imports * add pbtxt file * 100% matmul unit tests coverage * trigger CI with minimal changes to develop * adjust changes to develop * add fused_matmul op * inherit base ops * add "v2" * move OPMaker * Gradually add fused_matmul files * second batch of fused_matmul changes * split infershapes of matmul_v2 and fused_matmul * merge code from other PR * 2023 * inherit fused_matmul from matmul_v2 * Update paddle/phi/backends/onednn/onednn_reuse.h Co-authored-by: NTomasz Socha <tomasz.socha@intel.com> * Update paddle/phi/kernels/fusion/onednn/fused_matmul_kernel.cc Co-authored-by: NTomasz Socha <tomasz.socha@intel.com> * resolve conflicts * codestyle * simplify isgemmlinear * 2023 * remove import * reuse methods * matmul_v2_mkldnn cleanup * simplify ExecuteMatMulV1Grad * matmul refactored * fc * SetOutMemDescWithLogicalLayoutFusesSupport * matmul_v2 * alpha support * group repetetive funcs * matmul utils * execute matmul methods * restore registered kernel names * split header and impl files * remove double negatives * reduce numer of modified files * adjust ExecuteMatmul * add scales for ut * dates * limit number of modified files * fluid imports * remove alpha * codestyle --------- Co-authored-by: NTomasz Socha <tomasz.socha@intel.com>
-
由 scotty 提交于
-
由 RedContritio 提交于
-
由 RedContritio 提交于
-
由 Jiabin Yang 提交于
-
由 Kang Zhao 提交于
* feat: add relu composite rule * feat: add relu composite rule, maximum op * feat: add relu composite rule, maximum op * feat: add relu composite rule, polish comments * feat: add relu composite rule, polish comments * feat: add relu composite rule, add python api of relu * feat: add relu composite rule, commit hook * fix: maximum type error & ban cinn test * fix: maximum input sequence bugs * resolve conflicts * fix: code style bugs * add: relu fp16 test * feat: add rsqrt composite rule * feat: add rsqrt composite rule * resolve conflicts of composite rule * fix: delete check eager * feat: add roll grad composite rule * fix minus shift * fix test roll op
-
由 zhangyuqin1998 提交于
* Rename conv2d transpose grad grad * fix
-
由 sneaxiy 提交于
* fix flash attn * fix another API
-
由 RedContritio 提交于
-
由 RedContritio 提交于
* support auto generate static for randint * move seed from extra to attrs
-
由 RedContritio 提交于
-
- 04 4月, 2023 3 次提交
-
-
由 lzydev 提交于
* autogen embedding * deal * fix bug in CompatMetaTensor::share_lod
-
由 cyberslack_lee 提交于
* bce_loss * fix error * fix * fix * fix * reslove confilict
-
由 zhangyuqin1998 提交于
* rename_bilinear_tensor_product * fix
-
- 03 4月, 2023 2 次提交
-
-
由 HongyuJia 提交于
* [CustomOP Optional Inplace] Custom operator supports inplace optional vector Tensor input * uncomment unittest codes
-
由 RedContritio 提交于
* support auto generate static for gaussian (gaussian_random) * move gaussian_random_batch_size_like Kernels from gaussian_random_op.* to gaussian_random_batch_size_like_op.*
-