- 22 12月, 2021 14 次提交
-
-
由 Chen Weihang 提交于
* add pten kernel cmake * add pten kernel cmake function * fix compile error * add enforce include for full kernel * fix compile failed * change cuda to gpu * fix cmake function error
-
由 Zhanlue Yang 提交于
-
由 baoachun 提交于
* add mkldnn reshape_transpose_matmul fuse pass ut and op version check * update reshape_transpose_matmul_mkldnn_fuse_pass ut * update ut
-
由 baoachun 提交于
* update mkldnn batch_norm_activation fuse pass ut * update ut * update mkldnn batch_norm_act_fuse_pass ut * update batch_norm_act_fuse_pass ut * update ut
-
由 王明冬 提交于
-
由 LiYuRio 提交于
-
由 Guanghua Yu 提交于
-
由 Guoxia Wang 提交于
-
由 Chen Weihang 提交于
-
由 YuanRisheng 提交于
* move flatten * fix bugs of test * modify header file * add copy declare * fix compile bugs
-
由 joanna.wozna.intel 提交于
-
由 zyfncg 提交于
* rename full infer_meta * fix merge problem
-
由 wenbin 提交于
* CE fix * more format
-
由 Zhanlue Yang 提交于
-
- 21 12月, 2021 17 次提交
-
-
由 zyfncg 提交于
* add inplace_map for trace_op in pybind * fix inplace problem of setitem * refactor the param format of trace_op Co-authored-by: Npangyoki <pangyoki@126.com>
-
由 baoachun 提交于
* update seqconv_eltadd_relu_fuse_pass ut * update ut * update ut * update ut
-
由 baoachun 提交于
* update squared_mat_sub_fuse_pass ut * update ut * update ut
-
由 Chen Weihang 提交于
* rename cuda to gpu * revert CMake change * resolve conflit * rename other cuda to gpu * poish details
-
由 crystal 提交于
* relu forward opt * add gelu functor * optimize code
-
由 arlesniak 提交于
-
由 Yuang Liu 提交于
-
由 Guoxia Wang 提交于
-
由 baoachun 提交于
* add seqpool_cvm_concat_fuse_pass ut * rename ut name
-
由 chentianyu03 提交于
* fix when out_dtype is same with x.dtype and still transform type error * fix spell error
-
由 sneaxiy 提交于
* mean first version * fix scalar mean * add fp16 dtype for api
-
由 yeliang2258 提交于
* fix timeout bug * update
-
由 baoachun 提交于
* update repeated_fc_relu_fuse_pass ut * update ut
-
由 Haohongxiang 提交于
* update * fix bugs * modify code style * fix bugs of _get_global_group
-
由 heliqi 提交于
* add timeout * add timeout * PassAutoScan base_line use same config * try run base_line * fix dropout Mask of output attr error * fix dropout Mask of output attr error
-
由 石晓伟 提交于
* updates the check_file_diff_approvals for allocation refactor, test=develop * fix a bug, test=develop
-
由 Chen Weihang 提交于
* remove eigen and blas dir * fix declare error
-
- 20 12月, 2021 9 次提交
-
-
由 sneaxiy 提交于
-
由 baoachun 提交于
* add mkldnn conv_transpose_bias fuse pass ut * update conv_transpose_bias_mkldnn_fuse_pass ut * update conv_transpose_bias_mkldnn_fuse_pass ut * update conv_transpose_bias_mkldnn_fuse_pass ut * restrict conv2d data_format in conv_transpose_bias_mkldnn_fuse_pass * update ut timeout setting * update ut
-
由 chentianyu03 提交于
* add pten conj kernel * modify conj_kernel file path * add defined cuda macro to cuda/conj_kernel.h
-
由 baoachun 提交于
-
由 fwenguang 提交于
-
由 From00 提交于
-
由 sneaxiy 提交于
* support FP16 for more ops * add amp list tests * refine reduce_mean_grad * fix OP benchmark ci * fix fp16 reduce_mean * updat ut, but still have some problems * remove mean/reduce_mean fp16 kernel
-
由 Feng Xing 提交于
softmax_with_cross_entropy optimization with soft label. This PR includes optimization of "SoftmaxWithCrossEntropySoftLabel" : compute log_softmax and then compute loss. "CrossEntropySoftLabel" : compute loss with softmax as input. These optimization includes following technics: read data to buffer with vectorization compute max and sum in warp fixed loop size with macro Performance (computation time): softmax_with_cross_entropy_0 (forward) : -40.1% softmax_with_cross_entropy_0 (backward): -41%
-
由 石晓伟 提交于
-