- 23 3月, 2022 2 次提交
-
-
由 YuanRisheng 提交于
* move activation * fix bugs when run ce
-
由 Zhanlue Yang 提交于
* Removed redundant use of declarations.h * Fixed minor bug
-
- 22 3月, 2022 1 次提交
-
-
由 Leo Chen 提交于
* async prepare deps * fix bug that std::future is not set * add ut * refine code * fix standalone ut * disable prof
-
- 21 3月, 2022 3 次提交
-
-
由 From00 提交于
* Move conv-transpose OPs to phi * Fix CI errors * Fix CI errors
-
由 Siming Dai 提交于
* gpu ps graph engine * remove logs * Add neighbor sampling method * Add actual_sample_size and offset for sampling * Delete Chinese comment * Fix code style Co-authored-by: seemingwang <zsasuke@qq.com>
-
由 Allen Guo 提交于
* sync changes * copy sOpNamescope * fix UTs * add authors Co-authored-by: NXiaobing Wang <xiaobingw@graphcore.ai> Co-authored-by: NAllen Guo <alleng@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: NHan Zhao <hanzhao@graphcore.ai> * fix code-format * fix compile error * add comments for feed_op Co-authored-by: NXiaobing Wang <xiaobingw@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: NHan Zhao <hanzhao@graphcore.ai>
-
- 18 3月, 2022 3 次提交
-
-
由 shentanyue 提交于
* add gelu * fix gelu * add log_softmax * add prelu kernel and prelu/gelu/logsoftmax infershape * fix * fix * fix * fix * fix ci * log_softmax rewrite * fix * fix * fix conflict * fix compile error * fix comment * fix * ci_fix Co-authored-by: NYan Li <liyan665@gmail.com>
-
由 Zhang Zheng 提交于
* first commit * fix compile error * support std::vector<std::srting> * fix * fix op support on GPU by chenweihang * pass test * infershape * add set_dtype * fix order * fix * unify the impl of dt and sr * fix
-
由 chentianyu03 提交于
* move reduce_mean_grad kernel into phi * move reduce_max/min_grad into phi * remove raw max/min grad kernel * fix bug * fix max/min grad error * move all reduce_grad kernel into one file * add prod grad kernel * add infermeta for prod kernel
-
- 17 3月, 2022 6 次提交
-
-
由 Chen Weihang 提交于
* move assign kernel init commit * change vec<tensor> to vec<tensor*> * support tensor array * support api declare * fix test_list failed * fix npu and xpu failed * fix infrt failed * remove assign array size in operator * move assign sr header into sr dir * add infermeta for assign * test op success * fix test_list failed * fix kunlun failed * add set host allocator in tests * support tensor array in arg ctx * open set layout in share_meta * fix meta tensor layout error * fix test failed
-
由 seemingwang 提交于
* extract sub-graph * graph-engine merging * fix * fix * fix heter-ps config
-
由 TeFeng Chen 提交于
-
由 zyfncg 提交于
-
由 YuanRisheng 提交于
-
由 baoachun 提交于
-
- 16 3月, 2022 5 次提交
-
-
由 Zuza 提交于
* Quantize elementwise mul op * Parametrize elementwise functions * Fix code formatting
-
由 zlsh80826 提交于
-
由 Liu-xiandong 提交于
* [kp]fix bug that cannot fallback to CPU normally in XPU KP * fix bug in static graph
-
由 qipengh 提交于
-
由 Yulong Ao 提交于
* [Auto Parallel] Support the auto completion of while_op * [Auto Parallel] Improve the completion algorithms * [Auto Parallel] Fix bugs for ernie inference * [Auto Parallel] Remove attrs which cannot be pickled * [Auto Parallel] make the dims_mappings of LodTensorArray vars empty * [Auto Parallel] Fix bugs for the ernie inference in the pipeline parallel * [Auto Parallel] Remove unncessary comments * [Auto Parallel] Fix a bug of the CMakeLists * [Auto Parallel] Use the newest APIs to write the unit test * [Auto Parallel] Remove unnecessary statements
-
- 15 3月, 2022 4 次提交
-
-
由 Jacek Czaja 提交于
* - Prototype of third solution - fix - compilation fixes - fix - fixe - fix - fix - compilation fix - comment fix - lint update mkldnn conv_elementwise_add_fuse_pass ut - NHWC changes to prelu - alhpa dims - UT fix - fix to UT - lint - Some fixes - added to BWD of prelu NHWC support - reverted removal of resetting cu_layout in clearing of caching * - Small changes * - compilation fix * - fix * - fix * lint * - fixes after internal review * - compilation fix * - lint
-
由 Thunderbrook 提交于
* shard_id * format
-
由 hong 提交于
* move one hot to phi; test=develop * fix bugs; test=develop * fix bugs; test=develop * add infer meta; test=develop * fix bugs; test=develop * resolve confilct * resolve confilct * fix bug; * fix error; test=develop * update; test=develop * polish code; test=develop * add one api in eager mode; test=develop * add one hot test; test=develop * remove use less code; test=develop * fix bug; test=develop * polish code; test=develop * polish code; test=develop
-
由 YuanRisheng 提交于
* move activation op * adjust code format * fix compile bugs * fix ci bugs * code format adjust * code format adjust2 * activate ci status * modify according to comment * move activation kernel * revert relu6 * reduce add code * perfect use_phi_functor * completing func name * fix bugs when run ci * fix bugs when run infr * modifpy infrt get kernel signature
-
- 14 3月, 2022 3 次提交
-
-
由 Tomasz Socha 提交于
* Add elementwise add and activation fuse pass * Fix copy ellision * More flexible pattern detector * More flexible fusion pass * Update lists for pass * Add support for Pow operator * Add support for more activation types * Style * Rename fusion pass * First version of tests * Dirty version of pass * Polished version * Update pbtxt * Style * Update names * Style * Use PADDLE_ENFORCE_EQ * Save error message to variable * WO for error checks * CR * Static style check * Add missing 'activation_scale' attribute * Add relu6 and sigmoid activations * Style * Fix fuse list formating * Sync filenames for fuse pass files * Fix cmake after move * Fix registration * Fix pass name in tests * Add missing activations to checker * WIPS * Working mul op * Working sub * Working Add * Remove pten includes * Remove some forward declarations * Remove Includes * Fixes * Remove default kernels * Add check if post_ops attributes are avaliable * Style * Code adjustment * Register default kernels * We have year 2022 not 2021... Co-authored-by: Njakpiase <jakpia21@gmail.com> Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com> * Fast review fixes Co-authored-by: Njakpiase <jakpia21@gmail.com> Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com> * Review Fix * Rename one_dnn -> onednn * Style after review * Fast and dirty fix for quantization * Update tests * Style * Fix mkldnn_quantizer config * Add Joanna's suggestion. * Check if operator is explicitly disables on OneDNN * Try to use unregistered attributes * Style * Test new framework * FXI * FXII * Update test * Style Co-authored-by: Njakpiase <jakpia21@gmail.com> Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com>
-
由 Jiabin Yang 提交于
* eager, test=develop * fix bug, test=develop * eager, test=develop * merge legacy to fluid * eager, test=develop * eager, test=develop * Refactor TensorAdd func by template and remove gradient_accumulation in eager * Remove needless target name * eager, test=develop * eager, test=develop * Use overload instead of template * Remove legacy code * Remove legacy code * selectedrows, test=develop * Remove DataType test * eager, test=develop * eager, test=develop * support gan, test=develop * Using Tensor directly instead of using EagerTensor * support gradient_accumulation * make test_imperative_lod_tensor_to_selected_rows longer * make test_imperative_lod_tensor_to_selected_rows longer * refine code * ptb, test=develop * Rename all EagerTensor to Tensor * Rename some EagerTensor to Tensor * rename EagerTensor to EagerVariable * eager, test=develop * eager, test=develop * eager, test=develop * eager, test=develop * add more test * eager, test=develop * Support copiable selected rows and merge develop * save load, eager, test=develop * save load, eager, test=develop * refine, test=develop * remove useless _set_value method * refine, test=develop * refine, test=develop * revert static_runner, test=develop * EagerTensor to Tensor, test=develop * refine, test=develop * refine, test=develop * clear grad, test=develop * merge, develop * merge, develop * merge, test=develop * merge, test=develop * Support quant and part of slice * support legacy static save * extend slim tests time * remove imperative on inference * remove imperative on inference * merge develop * fix typo * fix typo * split slice related code into 2 part for imperative and eager * split slice from inference * split slice from inference * fix test_tensor_register_hook * support custom op in eager mode * fix inference deps error * split eager utils from custom operator * fix type match * fix typo Co-authored-by: NWang Huan <wanghuan29@baidu.com> Co-authored-by: NWeilong Wu <veyron_wu@163.com> Co-authored-by: Nwanghuancoder <wanghuancoder@163.com>
-
由 From00 提交于
* Move Pool OPs to phi * Fix CI error * Fix conflicts
-
- 13 3月, 2022 1 次提交
-
-
由 Chen Weihang 提交于
-
- 12 3月, 2022 1 次提交
-
-
由 jiangcheng 提交于
* fix NetBuilder API Name bug in cinn_lib_test * update cinn version to newest
-
- 11 3月, 2022 4 次提交
-
-
由 Sylwester Fraczek 提交于
-
由 Chen Weihang 提交于
* remove needless deps in unittests * add gpu marco * fix other unittests * fix kernel name error * fix test_prepare_op * fix failed dygraph unittests * fix gpu failed tests * fix cinn test failed * fix cinn test failed * fix dropout tests
-
由 chentianyu03 提交于
* add reduce_sum grad kernel * add reduce_grad * modify reduce grad * update reduce grad functions * fix build error * add argument mapping * move cast input after grad * add dims.size=1 cpu reduce_sum grad compute method * update reduce grad GPU * remove raw reduce_sum_grad kernel * modify header files * add namespace funcs for reduce_grad_funcstions
-
由 Aurelius84 提交于
* [Phi] Fix infershape if encounter TensorList and Attr("XXX") * add InferShapeArgumentMappingContext
-
- 10 3月, 2022 1 次提交
-
-
由 wawltor 提交于
* add the infer shape meta for the graph_send_recv * move the infershape code to another file
-
- 09 3月, 2022 2 次提交
-
-
由 YuanRisheng 提交于
-
由 zyfncg 提交于
* save code * fix bug of set_value * add coverage test
-
- 08 3月, 2022 3 次提交
-
-
由 YuanRisheng 提交于
[Phi]Move Relu/Cos/Sin/Tan/Acos/Asin/Atan/Sinh/Cosh/Asinh/Acosh/Atanh kernels in Activation to Phi (#40175) * move activation op * adjust code format * fix compile bugs * fix ci bugs * code format adjust * code format adjust2 * activate ci status * modify according to comment
-
由 YuanRisheng 提交于
-
由 Aganlengzi 提交于
* [custom kernel]Upgade support for multi libs * upgrade phi_custom_kernel deps
-
- 07 3月, 2022 1 次提交
-
-
由 Ming-Xu Huang 提交于
* Added cuBlasLtHandle_t to device context. * Added fused_gemm_epilogue op. 1. Added fused_gemm_epilogue op to leverage cuBlastLt Epilogue. 2. Support fusion Act(X*Y + bias), X'dims >=2 and Y'dims shoule be 2. 2. Act currently only be supported ReLU. (Will add GeLU in the future). * Added UT to fused_gemm_epilogue op. * Added LinearAct Pattern 1. Added LinearAct into graph_pattern_detector.* to define (2.)'s pattern. 2. LinearAct is used to detect act(element_add(matmul_v2(x, w), bias)). 3. act currently only support ReLU (Will support GeLU in the future). * Added FuseGemmEpiloguePass 1, Added FuseGemmEpiloguePass to handle nn.Linear + Act{ReLU} fusion (GeLU will be supported in the future). 2. Only support matmul_v2 from nn.Linear. * Added pybind to BuildStrageter.fuse_gemm_epilogue_. * Added UT for fuse_gemm_epilogue_pass. * GeLU support and EpilogueSingleton 1. Added GeLU support to fused_gemm_epilogue op. 2. Added EpilogueSingleton to cache auxiliary pointer. 3. Added related UTs. * Rename cublaslt_epilogue_opto gemm_epilogue_op.*. * Added both train and infer pattern to LinearAct. 1. Added support of fwd graph with grap_ops linking to LinearAct. 2. Added related changes to fuse_gemm_epilogue_pass for above modification. * Changed CUDA requirement from 11.4 to 11.6 for fuse_gemm_epilogue_pass. * Added identity activation support to gemm_epilogue_op. * Added Linear Fusion (matmul_v2 + ele_add) 1. Added matmul_v2 + ele_add pattern to LinearActPattern. 2. Added matmul_v2 + ele_add support to fuse_gemm_epilogue_pass. * Rename gemm_epilogue_op.* to fused_gemm_epilogue_op.* * Add fused_gemm_epilogue_grad op. 1. Added fused_gemm_epilogue_grad to support backward epilogue fusion. * Add UTs to fused_gemm_epilogue_grad_op. * Change attribute name in fused_gemm_epilogue_grad_op for clearing. * Allow DX and DBias be dispensable to fused_gemm_epilogue_grad op. * Added ElementwiseAdd+Matmul+Act graph pattern detection. * Fuse backward of Linear( Act(x)) 1. Added backward fusion pass to Linear( Act(x)). 2. Added backward fusion pass to Linear(x). * Added UTs to backward fusion of Linear(Act(x)). * Complete document of arguments to fused_gemm_epilogue_op. * Made arguments of some functions pass by reference. * Modify code with review comments. 1. Made arguments of some function pass by reference. 2. Removed redundant code. 3. Followed Google code style to change code. * Made 'const' code style be consistent * Fixed random seed of python UTs. * Set Compiling constrains to cuBlasLt 1. Require CUDA 11.6+ 2. Remove fuse_gemm_epilogue related tests when CUDA < 11.6. * Code Reivew from Paddle 1. Changed arguments name is_first_gemm to without_x_gradient for clearing. 2. Applied PADDLE_THROW in fused_gemm_epilogue_op. * Remove EpilogueSingleton 1. Applied ReserveSpace to replace Epilogue for passing auxiliary pointers between FWD and BWD. * Fix a logical error and enhance UTs. 1. Added act op count checking in UTs. 2. Fix issue to fuse backward or ReLU(Linear(X)). 3. TODO: solve GELU fusion issues. * Fix Linear and GeLU fusion issues. 1. Modified graph_detech_pattern to fit with both linear wiht gelu or relu. 2. Modified data range in Uts to allow negative values. * Removed fused_gemm_epilogue_op.h. * Rename namespace pten to phi. * Rename name of arguments in fused_gemm_epilogue_op 1. bias -> Bias. 2. out -> Out. 3. reserve_space -> ReserveSpace. * Change EpiloguePassActivationCache as local variable. 1. Removed singleton in EpiloguePassActivationCache. 2. Made EpiloguePassActivationCache as an argument to each pass functions.
-