- 01 12月, 2022 10 次提交
-
-
由 Wangzheee 提交于
* general optimization for no_varlen multihead
-
由 Wilber 提交于
* update memory_optimize pass
-
由 wanghuancoder 提交于
* do not link python lib in tensor wrapper
-
由 Zhang Jun 提交于
* instance norm support dynamic shape * update unittest
-
由 xiaoxiaohehe001 提交于
-
由 zhoutianzi666 提交于
* remove conv_act_set from graph_pattern_detector.cc
-
由 Zhang Jun 提交于
* Support FP16 in generic TensorRT plugin. * Support FP16 for Pad3D.
-
由 minghaoBD 提交于
* fuse-mt passes compatible with structured pruning
-
由 HongyuJia 提交于
* fix typo error * pass CI-coverage
-
由 zhangyikun02 提交于
-
- 30 11月, 2022 15 次提交
-
-
由 Qi Li 提交于
-
由 zyfncg 提交于
* fix error log for yaml check * remove grad_op of increment
-
由 Netpunk 提交于
* migrate transpose_op.cu.h and gpu_utils.h * format code style * fix some problems * format code * reset tranpose_op.cc * test commit * recover transpose_op.h * delete transpose_op.h * adjust header files order in transpose_op.cc
-
由 ShenLiang 提交于
* fix bug of pylayer * fix bug
-
由 wanghuancoder 提交于
-
由 Aurelius84 提交于
* [Perf]Fix interploate OutSize data transform problem * fix code style * fix grad * fix phi kernel
-
由 MarDino 提交于
* add activation support * fix cublasLt bug * remove useless code and fix test random range
-
由 feng_shuai 提交于
-
由 Yuanle Liu 提交于
-
由 zhangbo9674 提交于
* add fuse act add grad pass * polish code * refine code * add test * refine code
-
由 zyfncg 提交于
* rename some kernel name * fix compile problem
-
由 zyfncg 提交于
* fix bug of eigen_dependency * fix xpu compile
-
由 RichardWooSJTU 提交于
* delete unnecessary shape and slice op Co-authored-by: NYour Name <you@example.com>
-
由 james 提交于
some legacy code still use xpu_wait() for stream sync -- it only syncs default stream. this PR replaces them with dev_ctx.Wait() to ensure that correct stream is always used
-
由 zhangyikun02 提交于
-
- 29 11月, 2022 15 次提交
-
-
由 lzy 提交于
* fix mma_tensorcore (__CUDA_ARCH__) * disable tensorcore by default. disable tensorcore by default, because the judgment of __CUDA_ARCH__ will cause undefined behavior in some environments, can manually enable it on a machine that supports tensorcore.
-
由 HongyuJia 提交于
-
由 Paulina Gacek 提交于
* traspose2 kernel migrated * Got rid of mutable_data * x modification added * ops added in extra info file * Formatting fix * 2 fuse passes with tanpose2 commented * nr of outs changed in 2 passes, passes uncommented * Changes in passes reverted * transpose chnaged in operator.cc * MKLDNN check in operator.cc * Transpose fixes * Fix deleted from operato * template corrected Co-authored-by: NPaulina Gacek <paulinagacek@intel.com>
-
由 张春乔 提交于
* replace LoDTensor with phi::DenseTensor in fluid\operators * replace LoDTensor with phi::DenseTensor in fluid\operators * Update split_lod_tensor_op.cc * Update warpctc_op.cc * Update broadcast_tensors_op.cc * Update crf_decoding_op.cc * Update lstm_op.cc * Update lstm_op.cc * Update lod_reset_op.cc * Update gru_op.cc * Update linear_chain_crf_op.cc * resume 2 files for confilct * Update gru_op.cc * Update linear_chain_crf_op.cc * Update lstm_op.cc
-
由 Nyakku Shigure 提交于
* isort all files * revert conflicting files * revert conflicting files * revert conflicting files
-
由 xiaoxiaohehe001 提交于
-
由 Asthestarsfalll 提交于
* migrate enforce_custom.h from fluid to phi * move to backends/custom/
-
由 Sławomir Siwek 提交于
-
由 Vvsmile 提交于
Optimize the implementation of the argsort operator
-
由 Sławomir Siwek 提交于
* cleanup unused code * unify is_int8 is_bfloat16 * Simplify matmul_v2 FWD kernel * remove RunKernel methods * remove import namespace * remove headers * clean fluid/phi cross imports * remove fluid axpy_handler * delete fluid methods * activations * OneDNNMemDesc * MKLDNNFormatForSize * MatchShapeToLayout * MKLDNNMemoryFormat * MKLDNNFormat * ReorderMKLDNNHandler * to_void_cast * review suggestions * interpolate * remove fluid depedency * init * ExecuteMatMulV2 * rm fluid kernel * matmul_grad * remove mutable_data * mul_grad * matmul fwd * add extra attr * temp disable passes * re-enable passes * workaround for matmul+act * fix for matmul+eltwise_add * fix typo * merge bugfix #48364 * remove merge conflict
-
由 kangguangli 提交于
* fix:add no support for cuda_arch<700 * replace Executor in while op with InterpreterCore * cache InterpreterCore as the member of WhileOp * fix bug: tensor place changed because of assign op in while loop * refine code * refine code * refine code * hot fix * fix compile * merge develop * follow comments * add log for test * remove LoDTensor * set flag control_flow_use_new_executor false Co-authored-by: Nfengshuai <fengshuai03@baidu.com> Co-authored-by: Nzhiqiu <chenqiuliang@baidu.com>
-
由 haosicheng 提交于
-
由 JZ-LIANG 提交于
* get default calc stream from execution ctx instead of global dev ctx pool.
-
由 gem5 提交于
-
由 LiYuRio 提交于
* remove lod_tensor_to_array, array_to_lod_tensor, DynamicRNN * remove less_equal, greater_than, greater_equal, equal, not_equal
-