- 15 3月, 2023 7 次提交
-
-
由 Siming Dai 提交于
* add fp16 test for divide, matmul, pnorm * add cumsum fp16 unittest * fix threshold * revert cumsum * fix code-style * fix according to review * fix kernel not found
-
由 Guoxia Wang 提交于
-
由 WangZhen 提交于
-
由 zhangyuqin1998 提交于
* Delete hardswish_raw op * fix ut
-
由 ronnywang 提交于
* [CustomDevice] fix SyncDefaultStream for process_group_custom * update
-
由 wanghuancoder 提交于
* refine _found_inf
-
由 xiaoguoguo626807 提交于
* modify_yaml * delete default param * add output for matmul_double_grad
-
- 14 3月, 2023 33 次提交
-
-
由 zhouweiwei2014 提交于
-
由 Vvsmile 提交于
-
由 pangyoki 提交于
-
由 engineer1109 提交于
-
由 chenxujun 提交于
-
由 zhouweiwei2014 提交于
-
由 ccrrong 提交于
* add split_with_num composite rule * add split_with_num composite rule * add split composite rule * update * update test * update test * delete split_with_num_grad
-
由 engineer1109 提交于
fix abi fix tab
-
由 qizhaoaoe 提交于
-
由 limingshu 提交于
* first commit * fix code bugs in for_loop * fix bugs in cuLoadAddStridedInputs. * optimization for LayerNormBackwardComputeGradInput * add unitest for validating the optimization * fix windows ci error
-
由 gouzil 提交于
-
由 pangyoki 提交于
* cuda graph support multi-stream for new executor * fix windows compile error * delete create_cuda_graph_stream
-
由 zhaoyingli 提交于
-
由 YuhangLi 提交于
* wisemax fp16 support * add bf16 support 4 elementwise_max * append broadcast 4 op 4 fp16 / bf16 * fix elewise_max ut bf16 numeric delta * append fp/bf16 uts * add fp/bf16 uts * change bf16 uts delta * fix some issue * add prim 4 fp16
-
由 wangxiaoning 提交于
-
由 wenbin 提交于
-
由 xiongkun 提交于
-
由 Wang Bojun 提交于
* fix conv2d filter
-
由 xiaoguoguo626807 提交于
* init * modify
-
由 wangxiaoning 提交于
-
由 Infinity_lee 提交于
-
由 zhiboniu 提交于
* add fp16 and bf16 test * update
-
由 Ackeraa 提交于
add register of select Co-authored-by: Nwqgo <1552367872@qq.com>
-
由 Li-fAngyU 提交于
* update empty api to support compex dtype at static mode * code style * code style * 补充注释里的类型描述
-
由 HongyuJia 提交于
-
由 cxxly 提交于
-
由 xiongkun 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * Pr 50885 (#7) * [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com> * [prim] enable dygraph_to_static to support custom_vjp * fix cast prim and vjp dtype mapping error bug * [dy2static-ci] fix dy2static ci errors. --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com>
-
由 cxxly 提交于
-
由 cxxly 提交于
-
由 xiongkun 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes (#50557) * [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com> * [prim] enable dygraph_to_static to support custom_vjp * fix code in a dy2static-friendly way. * [dystatic] add hooker for prim --------- Co-authored-by: NAurelius84 <zhangliujie@baidu.com> Co-authored-by: Njiangcheng <thisjiang@qq.com> Co-authored-by: Ncxxly <chenxx_id@163.com>
-
由 cxxly 提交于
-
由 Aurelius84 提交于
* [CINN]Enhance CacheKey hash logic by considering input dtypes * add unittest * fix typo * fix typo * fix map.at * fix find * fix test * fix cinn cache key structure realize * using ordered map for attributes * add test by review advice --------- Co-authored-by: Njiangcheng <thisjiang@qq.com>
-
由 zhangbo9674 提交于
* add builtin-type DenseTensorType Float16Type Float64Type Int16Type Int64Type * refine comment * refine comment * add classof for Type class * refine test code * add get param func for DenseTensorType * add dyn_cast and refine isa * set default WITH_NEWIR=OFF * refine cast_utils * Refine code by comment * refine code by comment * refine code by comment * refine code by comment * fix bug of dyn_cast * set WITH_NEWIR=OFF * refine code by comment
-