- 19 8月, 2022 3 次提交
-
-
由 dongfangshenzhu 提交于
* add merged_momentum *test=kunlun * add merged_momentum *test=kunlun * add fp16 to merged_momentum,*test=kunlun * change dist_model.cc * add merged_momentum unittest and change momentum,test=kunlun * add merged_momentum unittest and change momentum,test=kunlun * add merged_momentum unittest and change momentum,test=kunlun * add merged_momentum unittest and change momentum,test=kunlun
-
由 Charles-hit 提交于
* 修复生成动态图代码时,如果输出没有配置名字,会导致下标越界的问题。 * decide forward_return[0] is not none * 修改反向yaml前向输出只有一个时,未配置名字,那么输出自动生成为out * modify code style
-
由 mengqingchun02 提交于
* support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * support beam_search operator on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * fix beam_search operator bugs on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun * support beam_search_decode operator on xpu. test=kunlun
-
- 18 8月, 2022 7 次提交
-
-
由 heliqi 提交于
* predictor add GetInputType interface * predictor change GetInputType to GetInputTypes * predictor add tester * predictor add tester * predictor change GetInputType to GetInputTypes * predictor change GetInputType to GetInputTypes * predictor add tester
-
由 Weilong Wu 提交于
* [Eager] add get_tensor_from_selected_rows * add PADDLE_ENFORCE to check SelectedRows * use _ prefix in temp
-
由 OccupyMars2025 提交于
-
由 pangyoki 提交于
apply buffer_shared_inplace_pass and inplace_addto_op_pass pass to program in Standalone Executor (#45085) * apply inplace addto in python apply_pass * fix * apply inplace pass for program * skip feed and fetch var * fix block_desc.move_from * fix block desc * alltoall remove inplace * fix
-
由 Aurelius84 提交于
* [OpAttr]Squeeze axes support Tensor * add support_tensor * fix unittest * fix coverage
-
由 zhangxiaoci 提交于
* change to async mode for xpu multi-card training in static graph mode * minor bugfix * irrelevant. move to another pr * move change to other pr * fix stream issue * fix 'stream not meet with current context' error * fix branch diverge, test=kunlun
-
由 JingZhuangzhuang 提交于
* fix infer tans scop * fix infer trans scope * fic infer trans scope * fic infer trans scope Co-authored-by: Ndingjiawei <327396238@qq.com>
-
- 17 8月, 2022 10 次提交
-
-
由 zyfncg 提交于
-
由 Sing_chan 提交于
-
由 Aurelius84 提交于
* [OpAttr]Add SupportTensor for OpMaker * fix typo * fix code style * add SupportTensor for concat op * add unittest for register Tensor * add shape checker and split attribute
-
由 Wilber 提交于
* fix multi stream error.
-
由 Leo Chen 提交于
* use addKernel * fix compile * remove elementwiseAddto * add return * fix custom place
-
由 feng_shuai 提交于
-
由 wanghuancoder 提交于
* fix_stop_gradient
-
由 fwenguang 提交于
-
由 ykkk2333 提交于
* xpu unittest grad compute supports more types, *test=kunlun * add instance norm xpu, *test=kunlun
-
由 sneaxiy 提交于
* fix squared_l2_norm bug * update buffer.h
-
- 16 8月, 2022 11 次提交
-
-
由 Chen Weihang 提交于
* move check finite and unscale kernel into phi * move infershape into phi * move update_loss_scaling kernel into phi * remove original kernels * move update loss scaling infershape into phi * add header for xpu and npu * solve coverage failed * fix npu test failed * remove mutable data in cu file * fix new executor failed * add valid check for meta tensor output
-
由 Weilong Wu 提交于
* [Eager draft] forward_only interface migrate to autograd_api * strings api add dygraph forward function * rm useless comments * draft version for check CI * fix ci * forward-only no need compute_require_grad and pass stop_gradient, rm useless comments * polish yaml and using CPUPlace = phi::CPUPlace * rm useless comments * polish yaml and update some test case * rm useless funcs * polish eager_gen code * polish code
-
由 feng_shuai 提交于
* convert multihead to oss * fix:bug * fix:delete const cast * fix:don't support bias_qk * add vit pass * fix:convert bug and add preln_residual_bias * support length=-1 * add UT for convert * add no_bias_qk support for gpu_multihead_op * delete infer_shape depends on bias_qk * oss just can be used in T4 and A* * fix:change api for ROCM CI
-
由 Aganlengzi 提交于
-
由 feifei-111 提交于
* fix_shape * code style * fix assert * fix to_tensor badreturn
-
由 Wangzheee 提交于
-
由 houj04 提交于
-
由 Wilber 提交于
-
由 Sing_chan 提交于
* add select_p * fix bugs * add custom test for select_p; modify select_p primrules * modify according to xiaoxu's comment * add eq_p, select_p, pow_p, use autograd to test grad of high order * add requirement of autograd, modify expected type of eq * modify according to xiaoxu's comment * import primops to use primops.pow
-
由 Feiyu Chan 提交于
-
由 Yulong Ao 提交于
-
- 15 8月, 2022 7 次提交
-
-
由 Yuanle Liu 提交于
-
由 zlsh80826 提交于
* Reduce pool2d test configuration * Reduce depthwise_conv2d test configuration * Reduce trt_convert_conv2d_fusion test configuration * Reduce trt_convert_conv2d test configuration * Reduce trt_convert_conv2d_transpose test configuration * Reduce trt_convert_hard_swish test configuration * Enhance trt auto scan test error message and mechanism * Increase FP16 trt ut tolerance
-
由 zhangyikun02 提交于
-
由 Hui Zhang 提交于
* rm useless pybind * rm useless ut
-
由 Yulong Ao 提交于
* [Auto Parallel] Move the distributed info from python to c++ * [Auto Parallel] Add dist_attrs for VarDesc and OpDesc * [Auto Parallel] Add the lost file * [Auto Parallel] Make the dist attr be unique_ptr * [Auto Parallel] Add the proto conversion * [Auto Parallel] Improve the proto support * [Auto Parallel] Fix the bugs for adding a device or a link * [Auto Parallel] Add the C++ ProcessMesh and DistributedMapper * [Auto Parallel] Improve the impl of these dist attrs * [Auto Parallel] Pybind11 ProcessMesh and DeviceMesh * [Auto Parallel] Fix the unittest problem * [Auto Parallel] Explicitly add the src file for auto_parallel target * [Auto Parallel] Add the proto depedency explicitly * [Auto Parallel] Fix the cmake bug on windows and mac * [Auto Parallel] Remove the pybind11 header file in process_mesh.h * [Auto Parallel] Remove unused codes * [Auto Parallel] Check whether the dist attr is null * [Auto Parallel] Implement the assign operator for OpDesc explicitly
-
由 houj04 提交于
* [XPU] add some collective ops. test=kunlun * use XPUOpTestWrapper. test=kunlun * skip kl1 for collective ops. fix typo: deivce -> device. test=kunlun
-
由 Wilber 提交于
* convert_fp16 support multi block * update * update
-
- 14 8月, 2022 1 次提交
-
-
由 xiaoxiaohehe001 提交于
This reverts commit 84bf5c31.
-
- 13 8月, 2022 1 次提交
-
-
由 Leo Chen 提交于
* add cached_serialize_str_ * support program hash * add sha * add ut * use hash_str only for new_exe * fix attr order
-