diff --git a/paddle/fluid/framework/parallel_executor.cc b/paddle/fluid/framework/parallel_executor.cc index 6e2168a017a56c8541a99c116b251cc34092d48d..977bd7c899833d75b00689b9c1bf54c8fde19f59 100644 --- a/paddle/fluid/framework/parallel_executor.cc +++ b/paddle/fluid/framework/parallel_executor.cc @@ -647,7 +647,7 @@ void ParallelExecutor::FeedAndSplitTensorIntoLocalScopes( "The number(%d) of samples of " "current batch is less than the count(%d) of " "devices(%s), currently, it is not allowed. ", - member_->places_.size(), lod_tensors.size(), + lod_tensors.size(), lod_tensors.size(), (is_cpu_place ? "CPU" : "GPU")); if (is_cpu_place) { error_info += diff --git a/paddle/fluid/pybind/pybind.cc b/paddle/fluid/pybind/pybind.cc index b0030d010f9222e52ee05ee935e917369d91d4d0..f6096fb8ca43b6cac3f1bc03de377c375a1c222d 100644 --- a/paddle/fluid/pybind/pybind.cc +++ b/paddle/fluid/pybind/pybind.cc @@ -1179,7 +1179,8 @@ All parameter, weight, gradient are variables in Paddle. }, R"DOC(The type is BOOL, allow_op_delay represents whether to delay the communication operators to run, it may make the execution faster. - Note that in some models, allow_op_delay may cause program hang. Default False.)DOC") + Note that this option is invalid now, and it will be removed in + next version. Default False.)DOC") .def_property( "num_iteration_per_drop_scope", [](const ExecutionStrategy &self) { @@ -1191,7 +1192,8 @@ All parameter, weight, gradient are variables in Paddle. R"DOC(The type is INT, num_iteration_per_drop_scope indicates how many iterations to clean up the temp variables which is generated during execution. It may make the execution faster, - because the temp variable's shape maybe the same between two iterations. Default 100. + because the temp variable's shape maybe the same between two iterations. + Default 1. NOTES: 1. If you fetch data when calling the 'run', the ParallelExecutor