提交 a9a228ad 编写于 作者: Y yi.wu

fix dist compile

上级 76ae540f
...@@ -25,6 +25,8 @@ limitations under the License. */ ...@@ -25,6 +25,8 @@ limitations under the License. */
#include "paddle/fluid/operators/detail/grpc_service.h" #include "paddle/fluid/operators/detail/grpc_service.h"
#include "paddle/fluid/operators/detail/send_recv.grpc.pb.h" #include "paddle/fluid/operators/detail/send_recv.grpc.pb.h"
#include "paddle/fluid/operators/detail/send_recv.pb.h" #include "paddle/fluid/operators/detail/send_recv.pb.h"
#include "paddle/fluid/operators/detail/simple_block_queue.h"
#include "paddle/fluid/operators/detail/sendrecvop_utils.h"
namespace paddle { namespace paddle {
namespace operators { namespace operators {
......
...@@ -199,9 +199,9 @@ TEST(LodTensor, Run) { ...@@ -199,9 +199,9 @@ TEST(LodTensor, Run) {
RunTestLodTensor(place); RunTestLodTensor(place);
RunTestLodTensor(place, 1); RunTestLodTensor(place, 1);
#ifdef PADDLE_WITH_CUDA #ifdef PADDLE_WITH_CUDA
platform::CUDAPlace place; platform::CUDAPlace gpu(0);
RunTestLodTensor(place); RunTestLodTensor(gpu);
RunTestLodTensor(place, 1); RunTestLodTensor(gpu, 1);
#endif #endif
} }
...@@ -210,7 +210,7 @@ TEST(SelectedRows, Run) { ...@@ -210,7 +210,7 @@ TEST(SelectedRows, Run) {
RunSerdeTestSelectedRows(place); RunSerdeTestSelectedRows(place);
#ifdef PADDLE_WITH_CUDA #ifdef PADDLE_WITH_CUDA
platform::CUDAPlace place; platform::CUDAPlace gpu;
RunSerdeTestSelectedRows(place); RunSerdeTestSelectedRows(gpu);
#endif #endif
} }
...@@ -93,12 +93,6 @@ class ListenAndServOp : public framework::OperatorBase { ...@@ -93,12 +93,6 @@ class ListenAndServOp : public framework::OperatorBase {
"server program should have at least 2 blocks"); "server program should have at least 2 blocks");
framework::Executor executor(dev_place); framework::Executor executor(dev_place);
std::vector<framework::ExecutorPrepareContext *> blk_ctx_list;
blk_ctx_list.push_back(nullptr); // block0 is not used.
for (int blkid = 1; blkid < num_blocks; ++blkid) {
auto *exe_ctx = executor.Prepare(*program, blkid);
blk_ctx_list.push_back(exe_ctx);
}
// TODO(typhoonzero): change this to a while_op for every cluster-batch. // TODO(typhoonzero): change this to a while_op for every cluster-batch.
bool exit_flag = false; bool exit_flag = false;
...@@ -150,11 +144,11 @@ class ListenAndServOp : public framework::OperatorBase { ...@@ -150,11 +144,11 @@ class ListenAndServOp : public framework::OperatorBase {
// block0 contains only listen_and_serv op, start run from block1. // block0 contains only listen_and_serv op, start run from block1.
for (int blkid = 1; blkid < num_blocks - 1; ++blkid) { for (int blkid = 1; blkid < num_blocks - 1; ++blkid) {
fs.push_back(framework::Async( fs.push_back(framework::Async(
[&executor, &program, &recv_scope, &blk_ctx_list, blkid]() { [&executor, &program, &recv_scope, blkid]() {
int run_block = blkid; // thread local int run_block = blkid; // thread local
try { try {
executor.RunPreparedContext(blk_ctx_list[run_block], executor.Run(*program, &recv_scope, run_block,
&recv_scope, false, false); false, false);
} catch (std::exception &e) { } catch (std::exception &e) {
LOG(ERROR) << "run sub program error " << e.what(); LOG(ERROR) << "run sub program error " << e.what();
} }
...@@ -164,8 +158,8 @@ class ListenAndServOp : public framework::OperatorBase { ...@@ -164,8 +158,8 @@ class ListenAndServOp : public framework::OperatorBase {
// Run global block at final step, or block1 if there are only 2 blocks // Run global block at final step, or block1 if there are only 2 blocks
if (num_blocks >= 2) { if (num_blocks >= 2) {
try { try {
executor.RunPreparedContext(blk_ctx_list[num_blocks - 1], &recv_scope, executor.Run(*program, &recv_scope, num_blocks - 1,
false, false); false, false);
} catch (std::exception &e) { } catch (std::exception &e) {
LOG(ERROR) << "run sub program error " << e.what(); LOG(ERROR) << "run sub program error " << e.what();
} }
...@@ -185,9 +179,9 @@ class ListenAndServOp : public framework::OperatorBase { ...@@ -185,9 +179,9 @@ class ListenAndServOp : public framework::OperatorBase {
sparse_vars.clear(); sparse_vars.clear();
} // while(true) } // while(true)
for (int i = 0; i < num_blocks; ++i) { // for (int i = 0; i < num_blocks; ++i) {
delete blk_ctx_list[i]; // delete blk_ctx_list[i];
} // }
} }
protected: protected:
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册