- 11 3月, 2022 1 次提交
-
-
由 xiongkun 提交于
* transfer nll_loss infershape into phi
-
- 10 3月, 2022 10 次提交
-
-
由 xiongkun 提交于
* transfer the infershape of accuracy op into phi * add set_dtype * add setdtype
-
由 Liu-xiandong 提交于
* [phi] move matrix_power op * MatrixInverse fluid -> phi * modify the CMake to fix compile bug * delete useless comment * mutable memory -> phi Alloc * modify the include file * modify the include file * fix bug in CI compiler
-
由 Jeffrey Chen 提交于
* [PHI] Migrate where_index op * [PHI] Fix where_index infermate * [Phi] set where_index out data type
-
由 Liu-xiandong 提交于
* add selected row op and fix bug in ctest * modify the date * fix bug in npu and xpu * modfiy the include file
-
由 Zhang Zheng 提交于
* first commit * cpu kernel * first version * fix compile error * fix compile error * delete v2 * fix * fix * add alias * fix * fix * fix * fix error * fix * fix * fix * fix format
-
由 Zhong Hui 提交于
* move segment_pool to phi. * mark summed ids as optional tensor. * fix as reviews.
-
由 wawltor 提交于
* add the infer shape meta for the graph_send_recv * move the infershape code to another file
-
由 Linjie Chen 提交于
* move argsort to phi * refine files * remove mutable_data * fix unittest * fix unittest * remove infershape * update infershape * fix ci * fix ci * fix ci * fix * fix * fix * fix * fix * fix * fix
-
由 Zhong Hui 提交于
* move arg min max to phi. * move infermeta. * fix as reviews.
-
由 hong 提交于
* move dropout to phi; test=develop * fix xpu, npu compile error; test=develop
-
- 09 3月, 2022 14 次提交
-
-
由 Linjie Chen 提交于
* move viterbi to phi * move infershape to phi * update infershape * fix * resolve conflicts
-
由 YuanRisheng 提交于
-
由 zyfncg 提交于
* save code * fix bug of set_value * add coverage test
-
由 zyfncg 提交于
* support code auto-gene for sparse backward api * fix bug of intermediate api and name of return var
-
由 crystal 提交于
* migrate eigh to phi * optimize code * modify code according to comment * conflict resolution
-
由 wawltor 提交于
* fix the full_like with fill the value of inf * update the test case for the fill_any_like * updae the comments for the full_like
-
由 huangxu96 提交于
* Add wait after Copy * fix wrong place delete
-
由 zhangkaihuo 提交于
Sparse conv3d backward(gpu)
-
由 xiongkun 提交于
* transfer the nll_loss_op and pass the CI * push * fix by self-review * fix by cr * add nll_loss * fix code
-
由 Chen Weihang 提交于
-
由 chenenquan 提交于
* [Phi] Migrate linspace op * [Phi] Migrate linspace op * [Phi] Fix linspace op * [PHI] rename data_tranform to data_type_transform * [PHI] Fix DECLARE and PD
-
由 chentianyu03 提交于
* add reduce_max kernel * add reduce max kernel * update reduce max Argumentmapping * remove reduce_max kernel * remove reduce_max kernel * add reduce max infermeta * rename reduce infermeta
-
由 JingZhuangzhuang 提交于
-
由 Yang 提交于
* fix take_along_axis cuda op register bug * add comma after float Co-authored-by: NChen Weihang <chenwhpro@163.com>
-
- 08 3月, 2022 13 次提交
-
-
由 Yiqun Liu 提交于
-
由 Chen Weihang 提交于
* remove gpudnn suffix & polish cmake * fix typo
-
由 Yang 提交于
* [Phi] move put_along_axis/take_along_axis/maxout * use phi::Copy
-
由 Chen Weihang 提交于
* move matrix inverse into phi * change license year
-
由 YuanRisheng 提交于
[Phi]Move Relu/Cos/Sin/Tan/Acos/Asin/Atan/Sinh/Cosh/Asinh/Acosh/Atanh kernels in Activation to Phi (#40175) * move activation op * adjust code format * fix compile bugs * fix ci bugs * code format adjust * code format adjust2 * activate ci status * modify according to comment
-
由 xiongkun 提交于
* transfer accuracy op and pass the ci * remove header file * fix code * fix code * fix * fix
-
由 WJJ1995 提交于
* support isfinite for phi * mark v2 * fixed bugs * fixed include bugs * deal with comments * decoupling selected_rows * rm bfloat16 * fixed infermeta * fixed code style * rm useless code * replace pt by pd
-
由 zyfncg 提交于
-
由 YuanRisheng 提交于
-
由 Aganlengzi 提交于
* [custom kernel]Upgade support for multi libs * upgrade phi_custom_kernel deps
-
由 wawltor 提交于
* [Phi] transfer old kernel to pten kernel for the graph_send_recv op * update the code for the define of graph_send_recv * fix the gradient problem for graph_send_recv * fix the compile problem * update the enfore message for the windows * update the code for the compiler * update compiler problem for the windows * udpate the code for windows * fix some format problem
-
由 furnace 提交于
* [Phi] move InferShape for truncated_gaussian_random and gaussian_random * [Phi] delete useless codes
-
由 Linjie Chen 提交于
* move infershapes to phi * update code format * update code format
-
- 07 3月, 2022 2 次提交
-
-
由 0x45f 提交于
* move bincount OP to phi * fix dtype * set_dtype by weights or x * fix conflicts
-
由 Ming-Xu Huang 提交于
* Added cuBlasLtHandle_t to device context. * Added fused_gemm_epilogue op. 1. Added fused_gemm_epilogue op to leverage cuBlastLt Epilogue. 2. Support fusion Act(X*Y + bias), X'dims >=2 and Y'dims shoule be 2. 2. Act currently only be supported ReLU. (Will add GeLU in the future). * Added UT to fused_gemm_epilogue op. * Added LinearAct Pattern 1. Added LinearAct into graph_pattern_detector.* to define (2.)'s pattern. 2. LinearAct is used to detect act(element_add(matmul_v2(x, w), bias)). 3. act currently only support ReLU (Will support GeLU in the future). * Added FuseGemmEpiloguePass 1, Added FuseGemmEpiloguePass to handle nn.Linear + Act{ReLU} fusion (GeLU will be supported in the future). 2. Only support matmul_v2 from nn.Linear. * Added pybind to BuildStrageter.fuse_gemm_epilogue_. * Added UT for fuse_gemm_epilogue_pass. * GeLU support and EpilogueSingleton 1. Added GeLU support to fused_gemm_epilogue op. 2. Added EpilogueSingleton to cache auxiliary pointer. 3. Added related UTs. * Rename cublaslt_epilogue_opto gemm_epilogue_op.*. * Added both train and infer pattern to LinearAct. 1. Added support of fwd graph with grap_ops linking to LinearAct. 2. Added related changes to fuse_gemm_epilogue_pass for above modification. * Changed CUDA requirement from 11.4 to 11.6 for fuse_gemm_epilogue_pass. * Added identity activation support to gemm_epilogue_op. * Added Linear Fusion (matmul_v2 + ele_add) 1. Added matmul_v2 + ele_add pattern to LinearActPattern. 2. Added matmul_v2 + ele_add support to fuse_gemm_epilogue_pass. * Rename gemm_epilogue_op.* to fused_gemm_epilogue_op.* * Add fused_gemm_epilogue_grad op. 1. Added fused_gemm_epilogue_grad to support backward epilogue fusion. * Add UTs to fused_gemm_epilogue_grad_op. * Change attribute name in fused_gemm_epilogue_grad_op for clearing. * Allow DX and DBias be dispensable to fused_gemm_epilogue_grad op. * Added ElementwiseAdd+Matmul+Act graph pattern detection. * Fuse backward of Linear( Act(x)) 1. Added backward fusion pass to Linear( Act(x)). 2. Added backward fusion pass to Linear(x). * Added UTs to backward fusion of Linear(Act(x)). * Complete document of arguments to fused_gemm_epilogue_op. * Made arguments of some functions pass by reference. * Modify code with review comments. 1. Made arguments of some function pass by reference. 2. Removed redundant code. 3. Followed Google code style to change code. * Made 'const' code style be consistent * Fixed random seed of python UTs. * Set Compiling constrains to cuBlasLt 1. Require CUDA 11.6+ 2. Remove fuse_gemm_epilogue related tests when CUDA < 11.6. * Code Reivew from Paddle 1. Changed arguments name is_first_gemm to without_x_gradient for clearing. 2. Applied PADDLE_THROW in fused_gemm_epilogue_op. * Remove EpilogueSingleton 1. Applied ReserveSpace to replace Epilogue for passing auxiliary pointers between FWD and BWD. * Fix a logical error and enhance UTs. 1. Added act op count checking in UTs. 2. Fix issue to fuse backward or ReLU(Linear(X)). 3. TODO: solve GELU fusion issues. * Fix Linear and GeLU fusion issues. 1. Modified graph_detech_pattern to fit with both linear wiht gelu or relu. 2. Modified data range in Uts to allow negative values. * Removed fused_gemm_epilogue_op.h. * Rename namespace pten to phi. * Rename name of arguments in fused_gemm_epilogue_op 1. bias -> Bias. 2. out -> Out. 3. reserve_space -> ReserveSpace. * Change EpiloguePassActivationCache as local variable. 1. Removed singleton in EpiloguePassActivationCache. 2. Made EpiloguePassActivationCache as an argument to each pass functions.
-