- 07 5月, 2018 7 次提交
-
-
由 fengjiayi 提交于
Correct tensor copy misuse
-
由 Yancey 提交于
-
由 Kexin Zhao 提交于
-
由 Kexin Zhao 提交于
-
由 Yiqun Liu 提交于
Add the call of DropKids at the end of executor.Run to delete the local scopes created in operators (#10403) * Add the call of DeleteScope to delete the memory of scope created by NewScope. * Call DropKids at the end of executor.Run to delete all local scopes created in operators.
-
由 chengduo 提交于
Feature/update sparse parameter
-
由 Darcy 提交于
* adding operator to blas deps list * use device_context instead to solve cycle deps
-
- 06 5月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "optimizer op support float64" * "fix ci" * "fix ftrl op"
-
- 05 5月, 2018 9 次提交
-
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 fengjiayi 提交于
-
由 Lei Wang 提交于
-
由 Kexin Zhao 提交于
-
由 Siddharth Goyal 提交于
-
由 Jeff Wang 提交于
* Load/save the params from the params_path * Switch to use load_persistables and save_persistables * Instaed of setup the executor to run program and scope. Pass the program to the load_persistables
-
由 Helin Wang 提交于
Fluid new API: dist train without modifying code
-
- 04 5月, 2018 23 次提交
-
-
由 chengduo 提交于
Wrap shfl_x_sync
-
由 chengduoZH 提交于
-
由 Yancey 提交于
* fix remove op * update
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 Yu Yang 提交于
Polish MatMul, clean copy & paste code
-
https://github.com/PaddlePaddle/Paddle由 chengduoZH 提交于
Merge branch 'develop' of https://github.com/PaddlePaddle/Paddle into feature/update_sparse_parameter
-
由 chengduoZH 提交于
-
由 Xin Pan 提交于
Distributed survey
-
由 Tao Luo 提交于
tensorrt convert init
-
由 Qingsheng Li 提交于
* Added auto transform to beam_search_decode_op * Added some comment * Added unittest for beam_search_decode_op on GPU
-
由 Yi Wang 提交于
-
由 Abhinav Arora 提交于
* Fix CPPLint issyes in gru_kernel.h * Fix CPPLint issyes in gru_kernel.h * Fix Compile error
-
由 Yu Yang 提交于
-
由 Lei Wang 提交于
-
由 Kexin Zhao 提交于
-
由 Yancey 提交于
fix delete_ops
-
由 Kexin Zhao 提交于
-
由 Kexin Zhao 提交于
* add test float16 inference accuracy example * complete the test * clean code * add argument parse and refine tests * add shell script * add float16 benchmark code * refine code * prepare for contrib/float16 * put things in contrib float16 folder * update benchmark result * further update benchmark report * add float16 inference report * update report
-
由 Helin Wang 提交于
Works with 1 trainer 1 pserver. 2 trainer 1 pserver will stuck at the end of first step, still investigating. The user only need to set envrionment variables to enable distributed training. run pserver: PADDLE_TRAINING_ROLE=PSERVER PADDLE_PSERVER_IPS=127.0.0.1 PADDLE_TRAINERS=2 PADDLE_CURRENT_IP=127.0.0.1 python no_test_word2vec_new_api.py run trainer: PADDLE_TRAINING_ROLE=TRAINER PADDLE_PSERVER_IPS=127.0.0.1 PADDLE_TRAINERS=2 PADDLE_TRAINER_ID=0 python no_test_word2vec_new_api.py
-
由 Abhinav Arora 提交于
-