- 25 8月, 2022 4 次提交
-
-
由 ronnywang 提交于
* [NPU] add run_program_op_npu * add run_program_op_npu ut
-
由 hong 提交于
* optimizer conv alog speed * code polish * remove useless code * fix compile error * fix cpu compile error * not use cudnn alog t * add search cache max number * polish code * fix cache test bug * add groups data format to conv args * fix cache test bug * fix cudnn_deterministic bug * fix test switch auto tune bug * fix test swith autotune bug; * fix conv cache bug * fix cache test error * fix cache test bug * fix windows mac compile error * fix workspace search error * update cudnn cache * fix cache test bug; test=develop * fix autotune swith test error * polish code * oplish code
-
由 Rayman 提交于
-
由 USTCKAY 提交于
-
- 24 8月, 2022 6 次提交
-
-
由 Leo Chen 提交于
* make tensor_util contains no cuda code * refine isfinite * revert ut * move isfinite function to its op * fix test * fix compile * std::isnan is not defined for int type on windows * fix windows compile * fix fp16 * fix rocm compile * revert gradient node
-
由 WangZhen 提交于
-
由 mengqingchun02 提交于
* support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support fp16 of adam operator in xpu environment. test=kunlun * support fp16 of adam operator in xpu environment. test=kunlun * support fp16 of adam operator in xpu environment. test=kunlun
-
由 WangZhen 提交于
* Adapt minlength attr for bincount
-
由 wenbin 提交于
* fix * optimize
-
由 zhaoying9105 提交于
-
- 23 8月, 2022 2 次提交
-
-
由 niuliling123 提交于
-
由 YuanRisheng 提交于
* move distribute_fpn_proposals * fix some code * fix yaml bugs * add set dtype * move proposal_impl to funcs * fix compile bugs
-
- 20 8月, 2022 1 次提交
-
-
由 Sing_chan 提交于
* add max_p without test * add test of max_p * make max_p consistent with paddle.maximum
-
- 19 8月, 2022 5 次提交
-
-
由 HongyuJia 提交于
-
由 houj04 提交于
-
由 mengqingchun02 提交于
* support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * make up beam_search_decode operator test cases on xpu and cpu environment. test=kunlun
-
由 dongfangshenzhu 提交于
* add merged_momentum *test=kunlun * add merged_momentum *test=kunlun * add fp16 to merged_momentum,*test=kunlun * change dist_model.cc * add merged_momentum unittest and change momentum,test=kunlun * add merged_momentum unittest and change momentum,test=kunlun * add merged_momentum unittest and change momentum,test=kunlun * add merged_momentum unittest and change momentum,test=kunlun
-
由 mengqingchun02 提交于
* support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun
-
- 18 8月, 2022 2 次提交
-
-
由 pangyoki 提交于
apply buffer_shared_inplace_pass and inplace_addto_op_pass pass to program in Standalone Executor (#45085) * apply inplace addto in python apply_pass * fix * apply inplace pass for program * skip feed and fetch var * fix block_desc.move_from * fix block desc * alltoall remove inplace * fix
-
由 Aurelius84 提交于
* [OpAttr]Squeeze axes support Tensor * add support_tensor * fix unittest * fix coverage
-
- 17 8月, 2022 4 次提交
-
-
由 Aurelius84 提交于
* [OpAttr]Add SupportTensor for OpMaker * fix typo * fix code style * add SupportTensor for concat op * add unittest for register Tensor * add shape checker and split attribute
-
由 Wilber 提交于
* fix multi stream error.
-
由 fwenguang 提交于
-
由 ykkk2333 提交于
* xpu unittest grad compute supports more types, *test=kunlun * add instance norm xpu, *test=kunlun
-
- 16 8月, 2022 5 次提交
-
-
由 Chen Weihang 提交于
* move check finite and unscale kernel into phi * move infershape into phi * move update_loss_scaling kernel into phi * remove original kernels * move update loss scaling infershape into phi * add header for xpu and npu * solve coverage failed * fix npu test failed * remove mutable data in cu file * fix new executor failed * add valid check for meta tensor output
-
由 feng_shuai 提交于
* convert multihead to oss * fix:bug * fix:delete const cast * fix:don't support bias_qk * add vit pass * fix:convert bug and add preln_residual_bias * support length=-1 * add UT for convert * add no_bias_qk support for gpu_multihead_op * delete infer_shape depends on bias_qk * oss just can be used in T4 and A* * fix:change api for ROCM CI
-
由 Aganlengzi 提交于
-
由 feifei-111 提交于
* fix_shape * code style * fix assert * fix to_tensor badreturn
-
由 Sing_chan 提交于
* add select_p * fix bugs * add custom test for select_p; modify select_p primrules * modify according to xiaoxu's comment * add eq_p, select_p, pow_p, use autograd to test grad of high order * add requirement of autograd, modify expected type of eq * modify according to xiaoxu's comment * import primops to use primops.pow
-
- 15 8月, 2022 4 次提交
-
-
由 Yuanle Liu 提交于
-
由 zhangyikun02 提交于
-
由 houj04 提交于
* [XPU] add some collective ops. test=kunlun * use XPUOpTestWrapper. test=kunlun * skip kl1 for collective ops. fix typo: deivce -> device. test=kunlun
-
由 Wilber 提交于
* convert_fp16 support multi block * update * update
-
- 12 8月, 2022 6 次提交
-
-
由 Sławomir Siwek 提交于
* remove v2_transpose_reshape * matmul_transpose_reshape * reshape_transpose_matmul * Add int8 support for matmulV2 * restore ut * adjust old ut * restore parallel UT ruels * remove mkldnn code from base ops * move enforces to pass * remove duplicated functions * delete duplicated enforces * feedback from review * add comments to variables * enable eltwise support * dynamic attribute * remove fusepass tests from op test * remove fuse pass cases from op test * revert introduction of dynamic attributes * style Co-authored-by: Nwozna <joanna.wozna@intel.com>
-
由 kangguangli 提交于
* transfer memcpy_h2d from fluid to phi * use UnchangedInferMeta instead * restore test_standalone_executor * add newline to fix codestyle check * rename pt -> phi * simplify logic and add check * make the comment more clear * remove useless comment * refine code
-
由 Yuanle Liu 提交于
* trt engine input data type should be consistent with trt input bindings type * fix some bugs * fix some bugs * fix some bugs
-
由 duanyanhui 提交于
* enhance grid_sampler to support 3d input
-
由 zyfncg 提交于
-
由 Siming Dai 提交于
* add init file * add op definition and infermeta * add kernel definition funcs * add broadcast infer shape * add gpu forward kernel * delete SUB and DIV * add x_grad * add template * add e_grad for min and max * fix small bug * temp commit * temp commit * add e_grad for sum and mean * fix some compile bug * fix compile bugs * fix compile problem * add sum forward unittest * fix broadcast error, add kernel sig, register e_grad, change unit test * fix grad * add temp grad fix * temp commit * add min max unittest * add max, min unittest, fix mul bug * add cpu forward sum and mean * add forward min max, fix mean unittest * add cpu backward min max * fix code-style * add backward sum mean * fix rocm ci * set uniitest timeout * fix bug of x broadcast to e, gpu grad * fix bug of x broadcast to e, cpu grad * rename BOOST_GET_CONST macro * fix rocm ci * mv graph_send_e_recv to graph_send_ue_recv * move out_size to IntArray * add eager op test * fix max pool type bug, add unittest for api * revise api doc * add fp16 for atomic min and max, add unittest * add unittest * add fp16 support for graph_send_recv * fix unittest fp16 bug * change OutSizeTensor to Out_size * move E to Y * add copyright, fix comment * review code * fix thread block size * fix thread block size * change api attribute name: pool_type to reduce_op, compute_type to message_op * change api attribute name, move pool_type to reduce_op, move compute_type to message_op
-
- 11 8月, 2022 1 次提交
-
-
由 carryyu 提交于
* make affine_grid_op support 5d_input on cpu and gpu
-