- 08 4月, 2022 2 次提交
-
-
由 Qi Li 提交于
Cherry-pick of #41521
-
由 Zhang Jun 提交于
[cherrypick-2.3] modify infer gpu memory strategy (#41427), remove cudnn_deterministic=True (#41341) (#41491) Co-authored-by: NJingZhuangzhuang <75348594+JZZ-NOTE@users.noreply.github.com>
-
- 06 4月, 2022 1 次提交
-
-
由 Allen Guo 提交于
* remove paddle_ipu shared library * fix unique_name
-
- 03 4月, 2022 1 次提交
-
-
由 FlyingQianMM 提交于
* limit grid dim for index select * mv LimitGridDim into gpu_launch_config.h * fix conflicts * fix conflicts * fix code style * set block to 256 * fix grid setting * set dtype of block_dim to unsigned int
-
- 01 4月, 2022 2 次提交
-
-
由 wanghuancoder 提交于
* support pinned, test=develop * support async_write, test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine,test=develop * refine, test=develop * refine, test=develop * refine, test=develop * refine, test=develop
-
由 z8hanghuan 提交于
* support multi_layer of bilstm,*test=kunlun * support multi_layer of bilstm, *test=kunlun * support multi_layer of bilstm, *test=kunlun * support multi_layer of bilstm, *test=kunlun
-
- 31 3月, 2022 3 次提交
-
-
由 Leo Chen 提交于
* fix bug that some op has no op_role attr * add mkldnn support for new executor * fit for mkldnn data_transfer * fit for mkldnn data_transfer
-
由 chenjian 提交于
* no * maintain old profiler * exclude new python record events for old profiler * maintain old profiler * maintain * maintain old profiler * maintain * fix cmakes
-
由 chenjian 提交于
* no * fix bugs * fix doc according to review * fix api doc format * fix api doc according to review * fix bug and add unit test * fix record event bug * optimize chrome tracing display * fix bug * add comment * add unit test * fix a bug * fix * fix * fix format
-
- 30 3月, 2022 3 次提交
-
-
由 From00 提交于
Add new APIs for GPU memory monitoring (max_memory_allocated, max_memory_reserved, memory_allocated, memory_reserved) (#38657) * Add new API memory_reserved * Add memory_allocated, max_memory_reserved and max_memory_allocater * Fix CI error * Fix CI error * Enhance UT * Add FLAGS_memory_stats_opt * Add STATS macro functions * Add StatAllocator * Fix CI errors * Add UT * Fix CI errors
-
由 ykkk2333 提交于
* add bilinear interpolate v2 to xpu list and unitteset, *test=kunlun * Delete ps_usr_print_log * Delete ps_usr_print_log * Delete xpu_op_test
-
由 houj04 提交于
* swish and pow op for xpu. test=kunlun * fix code style. test=kunlun. * use pow_grad xdnn api. test=kunlun.
-
- 29 3月, 2022 1 次提交
-
-
由 zhangyikun02 提交于
-
- 28 3月, 2022 1 次提交
-
-
由 chenjian 提交于
* no * fix bugs * fix doc according to review * fix api doc format * fix api doc according to review * fix bug and add unit test * fix record event bug
-
- 27 3月, 2022 1 次提交
-
-
由 Leo Chen 提交于
* fit for mkldnn and inplace op * fix compile * refine ut * register op version * fix inplace op * fix transfer_layout
-
- 25 3月, 2022 2 次提交
-
-
由 Aurelius84 提交于
* [Phi] Migrate Adam and Adamw into Phi * fix compile error and unittest ok * fix compile error and unittest ok * fix undefined reference to fLI::FLAGS * test depend on operator * fix cmake * fix xpu compile * fix infrt * fix amp_type_traits * fix amp_type_traits * modify according reviewer * modify according reviewer * fix dtype float16 * fix typo * fix Cmake * fix code style
-
由 FlyingQianMM 提交于
* add maximum limit for grid of reduce, elementwise and gather * add {} after if
-
- 23 3月, 2022 3 次提交
-
-
由 furnace 提交于
* [NPU] add npu support for conv3d and conv3d_grad * [NPU] delete failed unittests due to Ascend not support * [NPU] delete debug codes * [NPU] optimize codes, notest * [NPU] remove const_cast * [NPU] optimize for remove const_cast * [NPU] fix written errors
-
由 From00 提交于
* Performance optimize * Optimize GetAllocator, RWLock and ProcessUnfreedAllocation * Remove test file * Fix CI error * Fix CI errors * Fix CI errors
-
由 chenjian 提交于
* add event record for model profiling * fix format * fix format * fix code example bug * no * add profiler statistic * add profiler feature * fix bug * fix bug * fix bug * fix bug * required: gpu * required: gpu * fix bug * required: gpu * fix ci bug * fix ci error * fix ci error * upgrade document * fix doc * fix ci bug * add doc and fix bug * nothing * fix bug * fix format bug * modify format * add deprecated description for old profiler * fix bug * fix bug * fix * add load_profiler_reuslt doc * add load_profiler_reuslt doc * add load_profiler_reuslt doc * help fix old profiler sample code * add api doc * fix format * fix api doc * fix api doc format * fix api doc format * fix api doc c format * fix api doc format
-
- 21 3月, 2022 4 次提交
-
-
由 Chen Weihang 提交于
* add phi device context pool * change year * fix compile error * fix operator = error * refine init impl * polish details * refine init impl
-
由 zhangyikun02 提交于
-
由 Allen Guo 提交于
* add more ops * add authors Co-authored-by: NXiaobing Wang <xiaobingw@graphcore.ai> Co-authored-by: NAllen Guo <alleng@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: NHan Zhao <hanzhao@graphcore.ai> * rm ipu_strategy.check() * fix UT fail * fix typo Co-authored-by: NXiaobing Wang <xiaobingw@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: NHan Zhao <hanzhao@graphcore.ai>
-
由 Allen Guo 提交于
* sync changes * copy sOpNamescope * fix UTs * add authors Co-authored-by: NXiaobing Wang <xiaobingw@graphcore.ai> Co-authored-by: NAllen Guo <alleng@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: NHan Zhao <hanzhao@graphcore.ai> * fix code-format * fix compile error * add comments for feed_op Co-authored-by: NXiaobing Wang <xiaobingw@graphcore.ai> Co-authored-by: NZhixin Yao <zhixiny@graphcore.ai> Co-authored-by: NZhaorui Chen <zhaoruic@graphcore.ai> Co-authored-by: NHan Zhao <hanzhao@graphcore.ai>
-
- 15 3月, 2022 1 次提交
-
-
由 Jacek Czaja 提交于
* - Prototype of third solution - fix - compilation fixes - fix - fixe - fix - fix - compilation fix - comment fix - lint update mkldnn conv_elementwise_add_fuse_pass ut - NHWC changes to prelu - alhpa dims - UT fix - fix to UT - lint - Some fixes - added to BWD of prelu NHWC support - reverted removal of resetting cu_layout in clearing of caching * - Small changes * - compilation fix * - fix * - fix * lint * - fixes after internal review * - compilation fix * - lint
-
- 14 3月, 2022 3 次提交
-
-
由 Tomasz Socha 提交于
* Add elementwise add and activation fuse pass * Fix copy ellision * More flexible pattern detector * More flexible fusion pass * Update lists for pass * Add support for Pow operator * Add support for more activation types * Style * Rename fusion pass * First version of tests * Dirty version of pass * Polished version * Update pbtxt * Style * Update names * Style * Use PADDLE_ENFORCE_EQ * Save error message to variable * WO for error checks * CR * Static style check * Add missing 'activation_scale' attribute * Add relu6 and sigmoid activations * Style * Fix fuse list formating * Sync filenames for fuse pass files * Fix cmake after move * Fix registration * Fix pass name in tests * Add missing activations to checker * WIPS * Working mul op * Working sub * Working Add * Remove pten includes * Remove some forward declarations * Remove Includes * Fixes * Remove default kernels * Add check if post_ops attributes are avaliable * Style * Code adjustment * Register default kernels * We have year 2022 not 2021... Co-authored-by: Njakpiase <jakpia21@gmail.com> Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com> * Fast review fixes Co-authored-by: Njakpiase <jakpia21@gmail.com> Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com> * Review Fix * Rename one_dnn -> onednn * Style after review * Fast and dirty fix for quantization * Update tests * Style * Fix mkldnn_quantizer config * Add Joanna's suggestion. * Check if operator is explicitly disables on OneDNN * Try to use unregistered attributes * Style * Test new framework * FXI * FXII * Update test * Style Co-authored-by: Njakpiase <jakpia21@gmail.com> Co-authored-by: NSylwester Fraczek <sylwester.fraczek@intel.com>
-
由 Lijunhui 提交于
[KP] Add unittests for brelu,ceil,celu,elu,floor,hard_shrink,hard_sigmoid,log1p,logsigmoid,relu6,silu,soft_relu,softsign,swish (#40448) * solve unexecuted UT * add 24 activation op UT * append swish&thresholded_relu to kpfirst_list * rm thresholded_relu
-
由 liutiexing 提交于
-
- 11 3月, 2022 2 次提交
-
-
由 zhouweiwei2014 提交于
-
由 houj04 提交于
-
- 10 3月, 2022 2 次提交
-
-
由 Lijunhui 提交于
-
由 z8hanghuan 提交于
* add tril_triu for xpu, *test=kunlun * add tril_triu for xpu, *test=kunlun * add tril_triu for xpu, *test=kunlun * add tril_triu for xpu, *test=kunlun * add tril_triu for xpu, *test=kunlun
-
- 09 3月, 2022 1 次提交
-
-
由 fwenguang 提交于
-
- 08 3月, 2022 3 次提交
-
-
由 lilong12 提交于
* add pg_hccl
-
由 Aganlengzi 提交于
* [custom kernel]Upgade support for multi libs * upgrade phi_custom_kernel deps
-
由 chenjian 提交于
* add python profiler package * update according to review * fix bug * fix bug * fix bug * add unit test * Revert "add unit test" This reverts commit 4e69ff71b0645e069afe5dd8fea0d07717852c48. * reduce for pr * add unit test * modify for pr * fix unittest * update for ci coverage * modify according to review * fix bug * improve coverage
-
- 07 3月, 2022 1 次提交
-
-
由 Ming-Xu Huang 提交于
* Added cuBlasLtHandle_t to device context. * Added fused_gemm_epilogue op. 1. Added fused_gemm_epilogue op to leverage cuBlastLt Epilogue. 2. Support fusion Act(X*Y + bias), X'dims >=2 and Y'dims shoule be 2. 2. Act currently only be supported ReLU. (Will add GeLU in the future). * Added UT to fused_gemm_epilogue op. * Added LinearAct Pattern 1. Added LinearAct into graph_pattern_detector.* to define (2.)'s pattern. 2. LinearAct is used to detect act(element_add(matmul_v2(x, w), bias)). 3. act currently only support ReLU (Will support GeLU in the future). * Added FuseGemmEpiloguePass 1, Added FuseGemmEpiloguePass to handle nn.Linear + Act{ReLU} fusion (GeLU will be supported in the future). 2. Only support matmul_v2 from nn.Linear. * Added pybind to BuildStrageter.fuse_gemm_epilogue_. * Added UT for fuse_gemm_epilogue_pass. * GeLU support and EpilogueSingleton 1. Added GeLU support to fused_gemm_epilogue op. 2. Added EpilogueSingleton to cache auxiliary pointer. 3. Added related UTs. * Rename cublaslt_epilogue_opto gemm_epilogue_op.*. * Added both train and infer pattern to LinearAct. 1. Added support of fwd graph with grap_ops linking to LinearAct. 2. Added related changes to fuse_gemm_epilogue_pass for above modification. * Changed CUDA requirement from 11.4 to 11.6 for fuse_gemm_epilogue_pass. * Added identity activation support to gemm_epilogue_op. * Added Linear Fusion (matmul_v2 + ele_add) 1. Added matmul_v2 + ele_add pattern to LinearActPattern. 2. Added matmul_v2 + ele_add support to fuse_gemm_epilogue_pass. * Rename gemm_epilogue_op.* to fused_gemm_epilogue_op.* * Add fused_gemm_epilogue_grad op. 1. Added fused_gemm_epilogue_grad to support backward epilogue fusion. * Add UTs to fused_gemm_epilogue_grad_op. * Change attribute name in fused_gemm_epilogue_grad_op for clearing. * Allow DX and DBias be dispensable to fused_gemm_epilogue_grad op. * Added ElementwiseAdd+Matmul+Act graph pattern detection. * Fuse backward of Linear( Act(x)) 1. Added backward fusion pass to Linear( Act(x)). 2. Added backward fusion pass to Linear(x). * Added UTs to backward fusion of Linear(Act(x)). * Complete document of arguments to fused_gemm_epilogue_op. * Made arguments of some functions pass by reference. * Modify code with review comments. 1. Made arguments of some function pass by reference. 2. Removed redundant code. 3. Followed Google code style to change code. * Made 'const' code style be consistent * Fixed random seed of python UTs. * Set Compiling constrains to cuBlasLt 1. Require CUDA 11.6+ 2. Remove fuse_gemm_epilogue related tests when CUDA < 11.6. * Code Reivew from Paddle 1. Changed arguments name is_first_gemm to without_x_gradient for clearing. 2. Applied PADDLE_THROW in fused_gemm_epilogue_op. * Remove EpilogueSingleton 1. Applied ReserveSpace to replace Epilogue for passing auxiliary pointers between FWD and BWD. * Fix a logical error and enhance UTs. 1. Added act op count checking in UTs. 2. Fix issue to fuse backward or ReLU(Linear(X)). 3. TODO: solve GELU fusion issues. * Fix Linear and GeLU fusion issues. 1. Modified graph_detech_pattern to fit with both linear wiht gelu or relu. 2. Modified data range in Uts to allow negative values. * Removed fused_gemm_epilogue_op.h. * Rename namespace pten to phi. * Rename name of arguments in fused_gemm_epilogue_op 1. bias -> Bias. 2. out -> Out. 3. reserve_space -> ReserveSpace. * Change EpiloguePassActivationCache as local variable. 1. Removed singleton in EpiloguePassActivationCache. 2. Made EpiloguePassActivationCache as an argument to each pass functions.
-
- 04 3月, 2022 2 次提交
- 03 3月, 2022 1 次提交
-
-
由 ronnywang 提交于
-