From f816e116b51a179e3e41e65ac29148994eee1561 Mon Sep 17 00:00:00 2001 From: Yang Yang Date: Tue, 23 Jan 2018 02:24:35 +0000 Subject: [PATCH] clean up --- paddle/operators/parallel_do_op.cc | 6 ------ 1 file changed, 6 deletions(-) diff --git a/paddle/operators/parallel_do_op.cc b/paddle/operators/parallel_do_op.cc index dd8a9e05f9..d977ec43ff 100644 --- a/paddle/operators/parallel_do_op.cc +++ b/paddle/operators/parallel_do_op.cc @@ -174,7 +174,6 @@ class ParallelDoOp : public framework::OperatorBase { lod_tensor_to_be_merged->MergeLoDTensor(lod_tensors, dev_ctx.GetPlace()); } WaitOnPlaces(places); - LOG(INFO) << "End of ParallelGradDo"; } }; @@ -237,7 +236,6 @@ class ParallelDoGradOp : public framework::OperatorBase { WaitOnPlaces(places); AccumulateGrad(scope, place, sub_scopes, places); - LOG(INFO) << "End of ParallelDoGrad"; } void AccumulateGrad(const framework::Scope &scope, @@ -248,15 +246,12 @@ class ParallelDoGradOp : public framework::OperatorBase { std::__cxx11::string tmp_name; auto *tmp = sub_scopes[0]->Var(&tmp_name); - LOG(INFO) << "---" << s; for (size_t i = 1; i < sub_scopes.size(); ++i) { if (!(places[i] == places[0])) { - LOG(INFO) << "---"; CopyOrShare(*sub_scopes[i]->FindVar(s), places[0], tmp); WaitOnPlace(places[0]); } - LOG(INFO) << "---"; auto sum_op = framework::OpRegistry::CreateOp( "sum", {{"X", {s, tmp_name}}}, {{"Out", {s}}}, framework::AttributeMap{}); @@ -264,7 +259,6 @@ class ParallelDoGradOp : public framework::OperatorBase { WaitOnPlace(places[0]); } - LOG(INFO) << "---"; CopyOrShare(*sub_scopes[0]->FindVar(s), place, scope.FindVar(s)); } WaitOnPlaces(places); -- GitLab