- 25 4月, 2022 2 次提交
-
-
由 zyfncg 提交于
* optimiaze performance of PreparePhiData (#42093) * Dygraph performance optimization (v2) (#42103) * optimiaze performance of PreparePhiData * dygraph performance optimization * optimize performance of dygraph (#42137)
-
由 Aurelius84 提交于
[Cherry-Pick][Performance]Remove CudaStreamSychornize in ClipGradByGlobalNorm and fix shape op (#42170) * [Performance]Set ShapeKernel with ALL_BACKEND and ALL_LAYOUT (#42138) * [Performance]Set ShapeKernel with ALL_BACKEND and ALL_LAYOUT * [Performance]Set ShapeKernel with ALL_BACKEND and ALL_LAYOUT * [Performance]Remove CudaStreamSychornize in ClipGradByGlobalNorm (#42132)
-
- 24 4月, 2022 2 次提交
-
-
由 tianshuo78520a 提交于
解决编译依赖失败问题
-
由 Weilong Wu 提交于
* [Eager] Fix CastPyArg2scalar for max value of int64 (#42098) * [Eager] Fix CastPyArg2Scalar in Long case * Add more test cases for paddle.clip * Use PyLong_AsLongLong * Fix merge conflicts
-
- 23 4月, 2022 1 次提交
-
-
由 zmxdream 提交于
* add hashtable interface. test=develop * update. test=develop * update. test=develop * fix. test=develop * fix optimizer config for xpups. test=develop * fix. test=develop * fix. test=develop
-
- 22 4月, 2022 7 次提交
-
-
由 Aurelius84 提交于
-
由 pangyoki 提交于
* add _grad_name and _grad_value for eager tensor * fix paddle_enforce * fix paddle_enforce 2 * fix grad_name * _grad_value return lodtensor rather than tensor * fix
-
由 heliqi 提交于
修复ORT在batch变动时,输出shape不对问题
-
由 baoachun 提交于
-
由 Jacek Czaja 提交于
-
由 Baibaifan 提交于
* sharding_for_eager_tensor (#41415) * fix_sharding_copy_right (#41849)
-
由 Allen Guo 提交于
add mixed-precission support for ipu cherry-pick from #41733
-
- 21 4月, 2022 18 次提交
-
-
由 Weilong Wu 提交于
-
由 Zhen Wang 提交于
* Move pass optimizations into CINN.
-
由 zhangyikun02 提交于
-
由 z8hanghuan 提交于
* modify xpu.cmake,*test=kunlun (#41832) * modify xpu.cmake,*test=kunlun * modify xpu.cmake,*test=kunlun * modify xpu.cmake,*test=kunlun * modify xpu.cmake,*test=kunlun * support bilstm,*test=kunlun * [cherry-pick]support multi_layer of bilstm,*test=kunlun
-
由 lilong12 提交于
* fix_nccl_barrier (#41970) * be compatible with the old version of alltoall (#42007) Co-authored-by: NBaibaifan <39549453+Baibaifan@users.noreply.github.com>
-
由 WangXi 提交于
-
由 wangguanqun 提交于
* double accessor and show_scale * double accessor and show_scale * rename * fix bug in pslib config * add unittest
-
由 baoachun 提交于
* update gpu fp16 op blacklist * update blacklist
-
由 baoachun 提交于
-
由 zyfncg 提交于
* [PHI] Support some c++ api in paddle namespace (#41778) * support some c++ api in paddle namespace * change c++ api namespace in custom op * [Phi] Support setting size of vector<Tensor> for out in yaml (#41576) * support setting vector out size in yaml * support setting size of vector<tensor> for out in yaml * add data transform config for shape and size (#41909) * fix api_gen bug
-
由 Chen Weihang 提交于
* [Phi] Support setting size of vector<Tensor> for out in yaml (#41576) * support setting vector out size in yaml * support setting size of vector<tensor> for out in yaml * resolve conflict Co-authored-by: Nzyfncg <zhangyunfei07@baidu.com>
-
由 Aurelius84 提交于
* [Eager]Fix full_like/clip with np.generic type as attribute * support numpy genertic * remove usless code
-
由 TeFeng Chen 提交于
cherry-pick #41795
-
由 baoachun 提交于
-
由 JingZhuangzhuang 提交于
-
由 Jiabin Yang 提交于
* cherry-pick python/paddle/utils/code_gen/backward.yaml * remove unsupported yaml Co-authored-by: NZhanlue Yang <jim19930609@gmail.com>
-
由 Jiabin Yang 提交于
* make fast through to linear * make fast through to linear * add to do for later upgrades * support build once for now
-
由 Chen Weihang 提交于
* polish tensor api details (#41971) * [CustomOp] Fix custom op pinned input error (#41972) * fix custom op pinned input error * fix compile error * fix inference custom op (#41999) * resolve conflict
-
- 20 4月, 2022 10 次提交
-
-
由 lilong12 提交于
-
由 Haohongxiang 提交于
* refactor mp in eager mode * update * update * add uts
-
由 seemingwang 提交于
* gpu_graph engine optimization+ (#41455) * extract sub-graph * graph-engine merging * fix * fix * fix heter-ps config * test performance * test performance * test performance * test * test * update bfs * change cmake * test * test gpu speed * gpu_graph_engine optimization * add ssd layer to graph_engine * fix allocation * fix syntax error * fix syntax error * fix pscore class * fix * recover test * recover test * fix spelling * recover * fix * Cpu gpu graph engine (#41942) * extract sub-graph * graph-engine merging * fix * fix * fix heter-ps config * test performance * test performance * test performance * test * test * update bfs * change cmake * test * test gpu speed * gpu_graph_engine optimization * add ssd layer to graph_engine * fix allocation * fix syntax error * fix syntax error * fix pscore class * fix * recover test * recover test * fix spelling * recover * fix * fix linking problem * remove comment
-
由 Leo Chen 提交于
* [new-exec] shrink downstream map (#41471) * shrink downstream map * shrink last live ops of var * add comment * fix bug * add dependency for send/recv to support pp parallel (#41652) * [new-exec] clear the scope listener after run (#41947) * clear the listener after run * only sync variables in program * refine code * fit for lod_tensor_blocking_queue
-
由 heliqi 提交于
windows编译脚本增加onnxruntime编译选项
-
由 feng_shuai 提交于
-
由 pangyoki 提交于
* support no_need_buffer in eager_fluid state * change no_need_buffer info from fwd_info to bwd_info * fix CI fail, gru_unit donnot use no_need_buffer * fix conflict between no_need_buffer and dispensable * use tensor.define in dispensable * solve conflict * solve conflict
-
由 Jiabin Yang 提交于
Co-authored-by: NZhanlue Yang <jim19930609@gmail.com>
-
由 YuanRisheng 提交于
* support construct scalar using non-cpu tensor * fix bugs when run unittest * fix compile bugs * fix bugs when run ci * fix compile bugs * fix bugs when move copy * perfect unit test * perfect unittest * update according to comment * add target dependency * deal with conflict * fix bugs when run unit test * fix unit test bugs
-
由 Aurelius84 提交于
* update (#41636) * fix bug for eager mode distributed training (#41841) Co-authored-by: Nlilong12 <lilong12@baidu.com>
-