- 16 6月, 2018 1 次提交
-
-
由 qiaolongfei 提交于
-
- 14 6月, 2018 3 次提交
-
-
由 tensor-tang 提交于
-
由 Qiyang Min 提交于
* 1. Create buddy allocator in each places before NcclBcast the variables 2. Check the memory usage of ALL gpus rather than the first one * 1. Make NCCLGroupGuard guards only the ncclBcast part, which avoid ncclGroupEnd blocking the exception throwing 2. NOTE the usage of NCCLGroupGuard * Remove the memory usage check of gpus * Fix code style
-
由 Yang Yang(Tony) 提交于
-
- 13 6月, 2018 3 次提交
-
-
由 qiaolongfei 提交于
-
由 qiaolongfei 提交于
-
由 qiaolongfei 提交于
-
- 12 6月, 2018 1 次提交
-
-
由 Wu Yi 提交于
* wip * use executor.complete to end trainer * fix build * fix build with distribute off * fix typo * fix cmake typo * fix build
-
- 11 6月, 2018 11 次提交
-
-
由 qiaolongfei 提交于
-
由 dzhwinter 提交于
* "add inplace attribute" * "register inplace attribute" * "change se-next model for memory-reuse" * "fix typo" * repick * fix merge conflict * "fix stupid error"
-
由 gongweibao 提交于
-
由 chengduoZH 提交于
replace use_event with use_cuda, because use_event means the program running with CUDA, so use_cuda maybe more intuitive.
-
由 gongweibao 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 qiaolongfei 提交于
-
由 gongweibao 提交于
-
- 10 6月, 2018 5 次提交
-
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
- 08 6月, 2018 5 次提交
-
-
由 Luo Tao 提交于
-
由 fengjiayi 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 Yancey 提交于
* fix dist train error * update by comment
-
- 07 6月, 2018 11 次提交
-
-
由 tensor-tang 提交于
-
由 dzhwinter 提交于
* "split into multiple .ccl" * "refine file structure" * "refine files" * "remove the cmakelist" * "fix typo" * "fix typo" * fix ci
-
由 dzhwinter 提交于
* "init benchmark ops" * "untrack outputs" * "delete some usused code" * "benchmark" * "fix ci" * "fix op test" * "fix uint16 missing" * "fix ci" * "follow comments" * "fix ci" * "follow comments" * "conficts. merge develop branch" * repick * "merge develop branch"
-
由 fengjiayi 提交于
-
由 mozga-intel 提交于
* Add MKLDNN layout support in Paddle Add MKLDNN layout in Paddle so that MKLDNN friendly memory layout can be used in MKLDNN enabled OP kernel. Before this commit, NCHW is hardcode to be used in all MKLDNN op kernels. As a result, non-optimized execution path is selected in MKLDNN primitive which bring worse performance. Besides framework change, three MKLDNN OP kernels were updated for using new MKLDNN layout. They are conv/pool2d/batch_norm. Other MKLDNN OP kernels need be also updated in similar way to achieve best performance. * Add MKLDNN layout support in activation OP * Don't populate layout from input to output when kMKLDNN in * Refine pool mkldnn op kernel * MKLDNN layout * Remove the inferitance from tensor file * MKLDNN layout: refactoring * Remove additional #define to register new operator * Prepare mkldnn tests to work with layout
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 tensor-tang 提交于
-