提交 420ef2a3 编写于 作者: M mindspore-ci-bot 提交者: Gitee

!731 remove extra empty string from log text

Merge pull request !731 from fary86/remove_extra_empty_string_in_log
...@@ -345,7 +345,7 @@ TypePtr StringToNumberType(const std::string &type_name, const std::string &num_ ...@@ -345,7 +345,7 @@ TypePtr StringToNumberType(const std::string &type_name, const std::string &num_
auto bits = std::stoi(type_name.substr(num_type_name.size())); auto bits = std::stoi(type_name.substr(num_type_name.size()));
type = std::make_shared<T>(bits); type = std::make_shared<T>(bits);
} catch (const std::exception &e) { } catch (const std::exception &e) {
MS_LOG(EXCEPTION) << "" << num_type_name << " convert from string error " << e.what(); MS_LOG(EXCEPTION) << num_type_name << " convert from string error " << e.what();
} }
} }
return type; return type;
...@@ -389,7 +389,7 @@ TypePtr TensorStrToType(const std::string &type_name) { ...@@ -389,7 +389,7 @@ TypePtr TensorStrToType(const std::string &type_name) {
} }
type = std::make_shared<TensorType>(element_type); type = std::make_shared<TensorType>(element_type);
} catch (const std::exception &e) { } catch (const std::exception &e) {
MS_LOG(EXCEPTION) << "" << type_name << " convert from string error " << e.what(); MS_LOG(EXCEPTION) << type_name << " convert from string error " << e.what();
} }
} }
...@@ -416,7 +416,7 @@ TypePtr ListStrToType(const std::string &type_name) { ...@@ -416,7 +416,7 @@ TypePtr ListStrToType(const std::string &type_name) {
} }
type = std::make_shared<List>(element_types); type = std::make_shared<List>(element_types);
} catch (const std::exception &e) { } catch (const std::exception &e) {
MS_LOG(EXCEPTION) << "" << type_name << " convert from string error " << e.what(); MS_LOG(EXCEPTION) << type_name << " convert from string error " << e.what();
} }
} }
...@@ -443,7 +443,7 @@ TypePtr TupleStrToType(const std::string &type_name) { ...@@ -443,7 +443,7 @@ TypePtr TupleStrToType(const std::string &type_name) {
} }
type = std::make_shared<Tuple>(element_types); type = std::make_shared<Tuple>(element_types);
} catch (const std::exception &e) { } catch (const std::exception &e) {
MS_LOG(EXCEPTION) << "" << type_name << " convert from string error " << e.what(); MS_LOG(EXCEPTION) << type_name << " convert from string error " << e.what();
} }
} }
return type; return type;
...@@ -484,7 +484,7 @@ TypePtr FunctionStrToType(const std::string &type_name) { ...@@ -484,7 +484,7 @@ TypePtr FunctionStrToType(const std::string &type_name) {
} }
type = std::make_shared<Function>(args_type, retval); type = std::make_shared<Function>(args_type, retval);
} catch (const std::exception &e) { } catch (const std::exception &e) {
MS_LOG(EXCEPTION) << "" << type_name << " convert from string error " << e.what(); MS_LOG(EXCEPTION) << type_name << " convert from string error " << e.what();
} }
} }
return type; return type;
......
...@@ -888,7 +888,7 @@ void FuncGraphUserNodesCollector::OnMoveAllCNode(FuncGraphPtr src, FuncGraphPtr ...@@ -888,7 +888,7 @@ void FuncGraphUserNodesCollector::OnMoveAllCNode(FuncGraphPtr src, FuncGraphPtr
void FuncGraphJDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) { void FuncGraphJDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) {
if (IsValueNode<FuncGraph>(inp) && IsPrimitiveCNode(node, prim::kPrimJ)) { if (IsValueNode<FuncGraph>(inp) && IsPrimitiveCNode(node, prim::kPrimJ)) {
(void)Mod(node->func_graph(), GetValueNode<FuncGraphPtr>(inp), direction); (void)Mod(node->func_graph(), GetValueNode<FuncGraphPtr>(inp), direction);
MS_LOG(DEBUG) << "" << node->func_graph()->ToString() << " users func graph " MS_LOG(DEBUG) << node->func_graph()->ToString() << " users func graph "
<< GetValueNode<FuncGraphPtr>(inp)->ToString() << " which contains J(func_graph), dir: " << direction; << GetValueNode<FuncGraphPtr>(inp)->ToString() << " which contains J(func_graph), dir: " << direction;
} }
} }
...@@ -945,7 +945,7 @@ FuncGraphSetPtr FuncGraphParentsTotalComputer::SeekParents(const FuncGraphPtr &f ...@@ -945,7 +945,7 @@ FuncGraphSetPtr FuncGraphParentsTotalComputer::SeekParents(const FuncGraphPtr &f
void FuncGraphParentsTotalComputer::RealRecompute(FuncGraphPtr fg) { void FuncGraphParentsTotalComputer::RealRecompute(FuncGraphPtr fg) {
MS_EXCEPTION_IF_NULL(fg); MS_EXCEPTION_IF_NULL(fg);
all_parents_direct_ = &(manager_->func_graph_parents_direct()); all_parents_direct_ = &(manager_->func_graph_parents_direct());
MS_LOG(DEBUG) << "" << fg->ToString() << " total func graph dep size:" << (*all_parents_direct_)[fg].size(); MS_LOG(DEBUG) << fg->ToString() << " total func graph dep size:" << (*all_parents_direct_)[fg].size();
func_graph_parents_total_analysis_[fg].update(SeekParents(fg)); func_graph_parents_total_analysis_[fg].update(SeekParents(fg));
MS_LOG(DEBUG) << "FuncGraphParentsTotalComputer end: " << func_graph_parents_total_analysis_[fg].size(); MS_LOG(DEBUG) << "FuncGraphParentsTotalComputer end: " << func_graph_parents_total_analysis_[fg].size();
} }
...@@ -1074,7 +1074,7 @@ void FuncGraphsUsedTotalComputer::RealRecompute(FuncGraphPtr fg) { ...@@ -1074,7 +1074,7 @@ void FuncGraphsUsedTotalComputer::RealRecompute(FuncGraphPtr fg) {
if (func_graph_used_total_analysis_[fg].count(used_fg) == 0) { if (func_graph_used_total_analysis_[fg].count(used_fg) == 0) {
todo_new.push_back(used_fg); todo_new.push_back(used_fg);
} }
MS_LOG(DEBUG) << "" << fg->ToString() << " add func graph " << used_fg->ToString(); MS_LOG(DEBUG) << fg->ToString() << " add func graph " << used_fg->ToString();
func_graph_used_total_analysis_[fg].add(used_fg); func_graph_used_total_analysis_[fg].add(used_fg);
} }
} }
...@@ -1138,7 +1138,7 @@ void RecursiveComputer::CheckRecursiveGraphs(const FuncGraphPtr &fg, std::list<F ...@@ -1138,7 +1138,7 @@ void RecursiveComputer::CheckRecursiveGraphs(const FuncGraphPtr &fg, std::list<F
bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr &fg, const FuncGraphSetPtr &path) { bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr &fg, const FuncGraphSetPtr &path) {
MS_EXCEPTION_IF_NULL(path); MS_EXCEPTION_IF_NULL(path);
if (path->contains(fg)) { if (path->contains(fg)) {
MS_LOG(DEBUG) << "" << fg->ToString() << " had been checked"; MS_LOG(DEBUG) << fg->ToString() << " had been checked";
return false; return false;
} }
MS_EXCEPTION_IF_NULL(manager_); MS_EXCEPTION_IF_NULL(manager_);
...@@ -1149,7 +1149,7 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr &fg, const FuncGraphSetPt ...@@ -1149,7 +1149,7 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr &fg, const FuncGraphSetPt
std::find_if(func_graph_counter_map[fg].begin(), func_graph_counter_map[fg].end(), std::find_if(func_graph_counter_map[fg].begin(), func_graph_counter_map[fg].end(),
[path](const std::pair<FuncGraphPtr, int> iter) { return !path->contains(iter.first); }); [path](const std::pair<FuncGraphPtr, int> iter) { return !path->contains(iter.first); });
if (contains_j != func_graph_counter_map[fg].end()) { if (contains_j != func_graph_counter_map[fg].end()) {
MS_LOG(DEBUG) << "" << fg->ToString() << " contains J(" << contains_j->first->ToString() << ")"; MS_LOG(DEBUG) << fg->ToString() << " contains J(" << contains_j->first->ToString() << ")";
return true; return true;
} }
} }
...@@ -1160,12 +1160,11 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr &fg, const FuncGraphSetPt ...@@ -1160,12 +1160,11 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr &fg, const FuncGraphSetPt
for (auto &item : used[fg]) { for (auto &item : used[fg]) {
auto used_g = item.first; auto used_g = item.first;
if (SeekJ(used_g, path)) { if (SeekJ(used_g, path)) {
MS_LOG(DEBUG) << "" << fg->ToString() << " users func graph " << used_g->ToString() MS_LOG(DEBUG) << fg->ToString() << " users func graph " << used_g->ToString() << " which contains J(func_graph)";
<< " which contains J(func_graph)";
return true; return true;
} }
} }
MS_LOG(DEBUG) << "" << fg->ToString() << " doesn't contain J(func_graph)"; MS_LOG(DEBUG) << fg->ToString() << " doesn't contain J(func_graph)";
return false; return false;
} }
......
...@@ -145,14 +145,14 @@ py::function PrimitivePy::GetComputeFunction() { ...@@ -145,14 +145,14 @@ py::function PrimitivePy::GetComputeFunction() {
static const char *const compute_func_name = "vm_impl"; static const char *const compute_func_name = "vm_impl";
if (py::hasattr(python_obj_, compute_func_name)) { if (py::hasattr(python_obj_, compute_func_name)) {
MS_LOG(INFO) << "" << name() << " compute_func_name"; MS_LOG(INFO) << name() << " compute_func_name";
py::function fn = python_obj_.attr(compute_func_name).cast<py::function>(); py::function fn = python_obj_.attr(compute_func_name).cast<py::function>();
return fn; return fn;
} }
static const std::string vm_module = "mindspore.ops.vm_impl_registry"; static const std::string vm_module = "mindspore.ops.vm_impl_registry";
static const std::string get_vm_impl_fn = "get_vm_impl_fn"; static const std::string get_vm_impl_fn = "get_vm_impl_fn";
MS_LOG(INFO) << "" << name() << ": get_vm_impl_fn"; MS_LOG(INFO) << name() << ": get_vm_impl_fn";
py::function get_fn = parse::python_adapter::GetPyFn(vm_module, get_vm_impl_fn); py::function get_fn = parse::python_adapter::GetPyFn(vm_module, get_vm_impl_fn);
py::function vm_fn = get_fn(python_obj_); py::function vm_fn = get_fn(python_obj_);
......
...@@ -676,7 +676,7 @@ void MultitypeFuncGraph::Register(const std::vector<std::string> &types_name, co ...@@ -676,7 +676,7 @@ void MultitypeFuncGraph::Register(const std::vector<std::string> &types_name, co
for (auto &type_name : types_name) { for (auto &type_name : types_name) {
auto type_ptr = StringToType(type_name); auto type_ptr = StringToType(type_name);
if (type_ptr == nullptr) { if (type_ptr == nullptr) {
MS_LOG(EXCEPTION) << "" << type_name << " convert from string error "; MS_LOG(EXCEPTION) << type_name << " convert from string error ";
} }
types.push_back(type_ptr); types.push_back(type_ptr);
} }
...@@ -955,8 +955,7 @@ int CheckSliceMember(const AbstractBasePtr &member, int default_value, const std ...@@ -955,8 +955,7 @@ int CheckSliceMember(const AbstractBasePtr &member, int default_value, const std
return default_value; return default_value;
} }
MS_LOG(EXCEPTION) << "" << member_name << " should be a AbstractScalar or AbstractNone, but got " MS_LOG(EXCEPTION) << member_name << " should be a AbstractScalar or AbstractNone, but got " << member->ToString();
<< member->ToString();
} }
void GenerateTupleSliceParameter(const AbstractTuplePtr &tuple, const AbstractSlicePtr &slice, int *start_index, void GenerateTupleSliceParameter(const AbstractTuplePtr &tuple, const AbstractSlicePtr &slice, int *start_index,
......
...@@ -246,7 +246,7 @@ AbstractBasePtr InferImplBiasAddGrad(const AnalysisEnginePtr &, const PrimitiveP ...@@ -246,7 +246,7 @@ AbstractBasePtr InferImplBiasAddGrad(const AnalysisEnginePtr &, const PrimitiveP
// Inputs: at least one tensor(y_backprop) // Inputs: at least one tensor(y_backprop)
// Outputs: dbias // Outputs: dbias
if (args_spec_list.empty()) { if (args_spec_list.empty()) {
MS_LOG(EXCEPTION) << "" << primitive->name() << " evaluator at least has 1 parameters, while the input size is " MS_LOG(EXCEPTION) << primitive->name() << " evaluator at least has 1 parameters, while the input size is "
<< args_spec_list.size() << "."; << args_spec_list.size() << ".";
} }
...@@ -255,8 +255,7 @@ AbstractBasePtr InferImplBiasAddGrad(const AnalysisEnginePtr &, const PrimitiveP ...@@ -255,8 +255,7 @@ AbstractBasePtr InferImplBiasAddGrad(const AnalysisEnginePtr &, const PrimitiveP
MS_EXCEPTION_IF_NULL(shape_y); MS_EXCEPTION_IF_NULL(shape_y);
std::vector<int> y_dims = shape_y->shape(); std::vector<int> y_dims = shape_y->shape();
if (y_dims.size() < 2) { if (y_dims.size() < 2) {
MS_LOG(EXCEPTION) << "" << primitive->name() << " input y backprop, dim should >= 2, while " << y_dims.size() MS_LOG(EXCEPTION) << primitive->name() << " input y backprop, dim should >= 2, while " << y_dims.size() << ".";
<< ".";
} }
std::vector<int> bias_dims = {y_dims[1]}; std::vector<int> bias_dims = {y_dims[1]};
ShapePtr ret_shape = std::make_shared<Shape>(bias_dims); ShapePtr ret_shape = std::make_shared<Shape>(bias_dims);
......
...@@ -80,8 +80,7 @@ AbstractBasePtr InferImplDot(const AnalysisEnginePtr &, const PrimitivePtr &prim ...@@ -80,8 +80,7 @@ AbstractBasePtr InferImplDot(const AnalysisEnginePtr &, const PrimitivePtr &prim
auto y_shp_value = y_shp->shape(); auto y_shp_value = y_shp->shape();
// Should be matrix which shape size is 2. // Should be matrix which shape size is 2.
if (x_shp_value.size() != 2 || y_shp_value.size() != 2) { if (x_shp_value.size() != 2 || y_shp_value.size() != 2) {
MS_LOG(EXCEPTION) << "" << op_name MS_LOG(EXCEPTION) << op_name << " evaluator requires input two 2D tensors, while the dimensions of two tensors are "
<< " evaluator requires input two 2D tensors, while the dimensions of two tensors are "
<< x_shp_value.size() << ", " << y_shp_value.size() << " "; << x_shp_value.size() << ", " << y_shp_value.size() << " ";
} }
if (x_shp_value[1] != y_shp_value[0] && x_shp_value[1] != Shape::SHP_ANY && y_shp_value[0] != Shape::SHP_ANY) { if (x_shp_value[1] != y_shp_value[0] && x_shp_value[1] != Shape::SHP_ANY && y_shp_value[0] != Shape::SHP_ANY) {
......
...@@ -171,7 +171,7 @@ class Optimizer : public std::enable_shared_from_this<Optimizer> { ...@@ -171,7 +171,7 @@ class Optimizer : public std::enable_shared_from_this<Optimizer> {
}; };
use_profile ? (WITH(MsProfile::GetProfile()->Step(pass_names_[i])) opt_func) : opt_func(); use_profile ? (WITH(MsProfile::GetProfile()->Step(pass_names_[i])) opt_func) : opt_func();
#ifdef DEBUG #ifdef DEBUG
MS_LOG(DEBUG) << "" << name_ << " round " << counter << " OptPass " << pass_names_[i] << " end."; MS_LOG(DEBUG) << name_ << " round " << counter << " OptPass " << pass_names_[i] << " end.";
auto fg_name = name_ + "_r" + std::to_string(counter) + "_" + std::to_string(i) + "_" + pass_names_[i]; auto fg_name = name_ + "_r" + std::to_string(counter) + "_" + std::to_string(i) + "_" + pass_names_[i];
func_graph->DumpFuncGraph(fg_name); func_graph->DumpFuncGraph(fg_name);
DumpIR(fg_name + ".ir", func_graph); DumpIR(fg_name + ".ir", func_graph);
......
...@@ -37,7 +37,7 @@ void FunctionBlock::AddPrevBlock(const FunctionBlockPtr &block) { prev_blocks_.p ...@@ -37,7 +37,7 @@ void FunctionBlock::AddPrevBlock(const FunctionBlockPtr &block) { prev_blocks_.p
// write variable records the variable name to corresponding node // write variable records the variable name to corresponding node
void FunctionBlock::WriteVariable(const std::string &var_name, const AnfNodePtr &node) { void FunctionBlock::WriteVariable(const std::string &var_name, const AnfNodePtr &node) {
MS_LOG(DEBUG) << "" << func_graph_->ToString() << " write var " << var_name << " with node " << node->DebugString(); MS_LOG(DEBUG) << func_graph_->ToString() << " write var " << var_name << " with node " << node->DebugString();
vars_[var_name] = node; vars_[var_name] = node;
} }
...@@ -71,7 +71,7 @@ AnfNodePtr FunctionBlock::ReadVariable(const std::string &var) { ...@@ -71,7 +71,7 @@ AnfNodePtr FunctionBlock::ReadVariable(const std::string &var) {
TraceManager::DebugTrace(std::make_shared<TracePhi>(debug_info)); TraceManager::DebugTrace(std::make_shared<TracePhi>(debug_info));
ParameterPtr phi_param = std::make_shared<Parameter>(func_graph()); ParameterPtr phi_param = std::make_shared<Parameter>(func_graph());
TraceManager::EndTrace(); TraceManager::EndTrace();
MS_LOG(DEBUG) << "" << func_graph_->ToString() << " generate phi node " << phi_param->ToString() << " for " << var; MS_LOG(DEBUG) << func_graph_->ToString() << " generate phi node " << phi_param->ToString() << " for " << var;
func_graph()->add_parameter(phi_param); func_graph()->add_parameter(phi_param);
phi_nodes_[phi_param] = var; phi_nodes_[phi_param] = var;
WriteVariable(var, phi_param); WriteVariable(var, phi_param);
......
...@@ -333,7 +333,7 @@ void ExecutorPy::GetGeBackendPolicy() const { ...@@ -333,7 +333,7 @@ void ExecutorPy::GetGeBackendPolicy() const {
MS_EXCEPTION_IF_NULL(ms_context); MS_EXCEPTION_IF_NULL(ms_context);
std::string backend = ms_context->backend_policy(); std::string backend = ms_context->backend_policy();
if (backend != "ge") { if (backend != "ge") {
MS_LOG(EXCEPTION) << "" << backend << " backend policy is not supported under ge backend!"; MS_LOG(EXCEPTION) << backend << " backend policy is not supported under ge backend!";
} }
} }
...@@ -491,10 +491,10 @@ void RunPipelineAction(const ActionItem &action, pipeline::ResourcePtr resource, ...@@ -491,10 +491,10 @@ void RunPipelineAction(const ActionItem &action, pipeline::ResourcePtr resource,
// load MindSpore IR from file // load MindSpore IR from file
if (action.first == "symbol_resolve") { if (action.first == "symbol_resolve") {
MS_LOG(DEBUG) << "" << action.first << " read ir file: " << ir_file; MS_LOG(DEBUG) << action.first << " read ir file: " << ir_file;
std::vector<FuncGraphPtr> graphs = ImportIR(ir_file); std::vector<FuncGraphPtr> graphs = ImportIR(ir_file);
if (graphs.size() == 0) { if (graphs.size() == 0) {
MS_LOG(EXCEPTION) << "" << action.first << " read ir file " << ir_file << " failed as no graph found"; MS_LOG(EXCEPTION) << action.first << " read ir file " << ir_file << " failed as no graph found";
} }
auto manager = resource->manager(); auto manager = resource->manager();
MS_EXCEPTION_IF_NULL(manager); MS_EXCEPTION_IF_NULL(manager);
......
...@@ -78,7 +78,7 @@ AnalysisContextPtr AnalysisContext::Filter(const FuncGraphPtr &func_graph) { ...@@ -78,7 +78,7 @@ AnalysisContextPtr AnalysisContext::Filter(const FuncGraphPtr &func_graph) {
oss << ", context: " << iter.second.lock()->ToString() << "]"; oss << ", context: " << iter.second.lock()->ToString() << "]";
} }
oss << "}"; oss << "}";
MS_LOG(EXCEPTION) << "" << oss.str() << " NodeInfo: " << trace::GetDebugInfo(func_graph->debug_info()); MS_LOG(EXCEPTION) << oss.str() << " NodeInfo: " << trace::GetDebugInfo(func_graph->debug_info());
} }
return parent_context; return parent_context;
} }
......
...@@ -33,8 +33,7 @@ void InferEntryLogging(const EvaluatorPtr &evaluator, const AbstractBasePtrList ...@@ -33,8 +33,7 @@ void InferEntryLogging(const EvaluatorPtr &evaluator, const AbstractBasePtrList
MS_LOG(DEBUG) << "Evaluator " << evaluator->ToString() << " run for " << out_conf->node()->scope()->name(); MS_LOG(DEBUG) << "Evaluator " << evaluator->ToString() << " run for " << out_conf->node()->scope()->name();
} }
for (size_t i = 0; i < arg_spec_list.size(); i++) { for (size_t i = 0; i < arg_spec_list.size(); i++) {
MS_LOG(DEBUG) << "" << evaluator->ToString() << " input[" << i MS_LOG(DEBUG) << evaluator->ToString() << " input[" << i << "] abstract value: " << arg_spec_list[i]->ToString();
<< "] abstract value: " << arg_spec_list[i]->ToString();
} }
} }
...@@ -137,7 +136,7 @@ AbstractBasePtrList FuncGraphEvaluator::NormalizeArgs(const AbstractBasePtrList ...@@ -137,7 +136,7 @@ AbstractBasePtrList FuncGraphEvaluator::NormalizeArgs(const AbstractBasePtrList
MS_EXCEPTION_IF_NULL(arg); MS_EXCEPTION_IF_NULL(arg);
return arg->Broaden(); return arg->Broaden();
}); });
MS_LOG(DEBUG) << "" << func_graph_->ToString() << " original: " << mindspore::ToString(args_spec_list) MS_LOG(DEBUG) << func_graph_->ToString() << " original: " << mindspore::ToString(args_spec_list)
<< ", broaded: " << mindspore::ToString(broaded_list); << ", broaded: " << mindspore::ToString(broaded_list);
return broaded_list; return broaded_list;
} }
...@@ -230,20 +229,20 @@ AbstractBasePtr Evaluator::Run(AnalysisEnginePtr engine, const ConfigPtrList &ar ...@@ -230,20 +229,20 @@ AbstractBasePtr Evaluator::Run(AnalysisEnginePtr engine, const ConfigPtrList &ar
MS_EXCEPTION_IF_NULL(cache_); MS_EXCEPTION_IF_NULL(cache_);
auto iter = cache_->find(args_spec_list); auto iter = cache_->find(args_spec_list);
if (iter == cache_->end()) { if (iter == cache_->end()) {
MS_LOG(DEBUG) << "" << evaluator_name << " cache miss, call Infer()."; MS_LOG(DEBUG) << evaluator_name << " cache miss, call Infer().";
AbstractBasePtr ret = Infer(engine, args_spec_list); AbstractBasePtr ret = Infer(engine, args_spec_list);
if (ret == nullptr) { if (ret == nullptr) {
InferFailLogging(shared_from_base<Evaluator>(), args_spec_list, out_conf); InferFailLogging(shared_from_base<Evaluator>(), args_spec_list, out_conf);
MS_LOG(EXCEPTION) << "Evaluator " << evaluator_name << " result is nullptr."; MS_LOG(EXCEPTION) << "Evaluator " << evaluator_name << " result is nullptr.";
} }
MS_EXCEPTION_IF_NULL(ret); MS_EXCEPTION_IF_NULL(ret);
MS_LOG(DEBUG) << "" << evaluator_name << " set cache. return: " << ret->ToString() << "."; MS_LOG(DEBUG) << evaluator_name << " set cache. return: " << ret->ToString() << ".";
(*cache_)[args_spec_list] = ret; (*cache_)[args_spec_list] = ret;
trace::TraceGraphInferLeave(shared_from_base<Evaluator>()); trace::TraceGraphInferLeave(shared_from_base<Evaluator>());
return ret; return ret;
} else { } else {
MS_EXCEPTION_IF_NULL(iter->second); MS_EXCEPTION_IF_NULL(iter->second);
MS_LOG(DEBUG) << "" << evaluator_name << " cache hit. return: " << iter->second->ToString() << "."; MS_LOG(DEBUG) << evaluator_name << " cache hit. return: " << iter->second->ToString() << ".";
trace::TraceGraphInferLeave(shared_from_base<Evaluator>()); trace::TraceGraphInferLeave(shared_from_base<Evaluator>());
return iter->second; return iter->second;
} }
......
...@@ -103,7 +103,7 @@ ShapePtr CheckShapeSame(const std::string &op, const AbstractTensorPtr &tensor_b ...@@ -103,7 +103,7 @@ ShapePtr CheckShapeSame(const std::string &op, const AbstractTensorPtr &tensor_b
ShapePtr shape_base = tensor_base->shape(); ShapePtr shape_base = tensor_base->shape();
ShapePtr shape = tensor->shape(); ShapePtr shape = tensor->shape();
if (*shape != *shape_base) { if (*shape != *shape_base) {
MS_LOG(EXCEPTION) << "" << op << " evaluator first arg shape " << tensor->shape()->ToString() MS_LOG(EXCEPTION) << op << " evaluator first arg shape " << tensor->shape()->ToString()
<< " are not consistent with second arg shape " << tensor_base->shape()->ToString(); << " are not consistent with second arg shape " << tensor_base->shape()->ToString();
} }
return shape_base; return shape_base;
...@@ -113,7 +113,7 @@ TypePtr CheckDtypeSame(const std::string &op, const AbstractTensorPtr &tensor_ba ...@@ -113,7 +113,7 @@ TypePtr CheckDtypeSame(const std::string &op, const AbstractTensorPtr &tensor_ba
TypePtr type_base = tensor_base->element()->BuildType(); TypePtr type_base = tensor_base->element()->BuildType();
TypePtr type = tensor->element()->BuildType(); TypePtr type = tensor->element()->BuildType();
if (*type != *type_base) { if (*type != *type_base) {
MS_LOG(EXCEPTION) << "" << op << " evaluator first arg dtype " << type_base->ToString() MS_LOG(EXCEPTION) << op << " evaluator first arg dtype " << type_base->ToString()
<< " are not consistent with second arg dtype " << type->ToString(); << " are not consistent with second arg dtype " << type->ToString();
} }
return type_base; return type_base;
...@@ -121,14 +121,14 @@ TypePtr CheckDtypeSame(const std::string &op, const AbstractTensorPtr &tensor_ba ...@@ -121,14 +121,14 @@ TypePtr CheckDtypeSame(const std::string &op, const AbstractTensorPtr &tensor_ba
int CheckAxis(const std::string &op, const ValuePtr &axis, int minimum, int max) { int CheckAxis(const std::string &op, const ValuePtr &axis, int minimum, int max) {
if (axis == nullptr) { if (axis == nullptr) {
MS_LOG(EXCEPTION) << "" << op << " evaluator axis is null"; MS_LOG(EXCEPTION) << op << " evaluator axis is null";
} }
if (!axis->isa<Int32Imm>()) { if (!axis->isa<Int32Imm>()) {
MS_LOG(EXCEPTION) << "" << op << " evaluator axis should be int, but got " << axis->type_name(); MS_LOG(EXCEPTION) << op << " evaluator axis should be int, but got " << axis->type_name();
} }
int axis_value = GetValue<int>(axis); int axis_value = GetValue<int>(axis);
if (axis_value > max || axis_value < minimum) { if (axis_value > max || axis_value < minimum) {
MS_LOG(EXCEPTION) << "" << op << " evaluator axis value should be in the range [" << minimum << ", " << max MS_LOG(EXCEPTION) << op << " evaluator axis value should be in the range [" << minimum << ", " << max
<< "], but get " << axis_value; << "], but get " << axis_value;
} }
return axis_value; return axis_value;
...@@ -136,8 +136,7 @@ int CheckAxis(const std::string &op, const ValuePtr &axis, int minimum, int max) ...@@ -136,8 +136,7 @@ int CheckAxis(const std::string &op, const ValuePtr &axis, int minimum, int max)
void CheckArgsSize(const std::string &op, const mindspore::abstract::AbstractBasePtrList &args_spec_list, void CheckArgsSize(const std::string &op, const mindspore::abstract::AbstractBasePtrList &args_spec_list,
size_t size_expect) { size_t size_expect) {
if (args_spec_list.size() != size_expect) { if (args_spec_list.size() != size_expect) {
MS_LOG(EXCEPTION) << "" << op << " input args size should be " << size_expect << ", but got " MS_LOG(EXCEPTION) << op << " input args size should be " << size_expect << ", but got " << args_spec_list.size();
<< args_spec_list.size();
} }
for (size_t i = 0; i < size_expect; i++) { for (size_t i = 0; i < size_expect; i++) {
......
...@@ -70,7 +70,7 @@ ABSTRACT_REPORT_NAME_TRAITS(Class) ...@@ -70,7 +70,7 @@ ABSTRACT_REPORT_NAME_TRAITS(Class)
template <typename T> template <typename T>
std::shared_ptr<T> CheckArg(const std::string &op, const AbstractBasePtrList &args_spec_list, size_t index) { std::shared_ptr<T> CheckArg(const std::string &op, const AbstractBasePtrList &args_spec_list, size_t index) {
if (index >= args_spec_list.size()) { if (index >= args_spec_list.size()) {
MS_EXCEPTION(ValueError) << "" << op << " evaluator args list index out of bound, size " << args_spec_list.size() MS_EXCEPTION(ValueError) << op << " evaluator args list index out of bound, size " << args_spec_list.size()
<< ", index " << index; << ", index " << index;
} }
auto arg = dyn_cast<T>(args_spec_list[index]); auto arg = dyn_cast<T>(args_spec_list[index]);
......
...@@ -122,7 +122,7 @@ AnalysisResult AnalysisEngine::Run(const FuncGraphPtr &func_graph, const Abstrac ...@@ -122,7 +122,7 @@ AnalysisResult AnalysisEngine::Run(const FuncGraphPtr &func_graph, const Abstrac
MS_EXCEPTION_IF_NULL(root_context->func_graph()); MS_EXCEPTION_IF_NULL(root_context->func_graph());
AnfNodeConfigPtr output_conf = MakeConfig(root_context->func_graph()->get_return(), root_context); AnfNodeConfigPtr output_conf = MakeConfig(root_context->func_graph()->get_return(), root_context);
MS_EXCEPTION_IF_NULL(func_graph); MS_EXCEPTION_IF_NULL(func_graph);
MS_LOG(INFO) << "" << func_graph->ToString() << ": Run finished."; MS_LOG(INFO) << func_graph->ToString() << ": Run finished.";
AnalysisResult result; AnalysisResult result;
MS_EXCEPTION_IF_NULL(output_conf); MS_EXCEPTION_IF_NULL(output_conf);
...@@ -167,7 +167,7 @@ AbstractBasePtr AnalysisEngine::Eval(const AnfNodeConfigPtr &conf) { ...@@ -167,7 +167,7 @@ AbstractBasePtr AnalysisEngine::Eval(const AnfNodeConfigPtr &conf) {
for (auto iter : compute_conf_stack_) { for (auto iter : compute_conf_stack_) {
buffer << " -> " << iter->DebugString(); buffer << " -> " << iter->DebugString();
} }
MS_LOG(DEBUG) << "" << buffer.str(); MS_LOG(DEBUG) << buffer.str();
#endif #endif
MS_LOG(DEBUG) << "Begin Eval NodeConfig " << conf->ToString(); MS_LOG(DEBUG) << "Begin Eval NodeConfig " << conf->ToString();
MS_EXCEPTION_IF_NULL(node); MS_EXCEPTION_IF_NULL(node);
......
...@@ -175,7 +175,7 @@ std::vector<int> RealBroadcast(const std::string &op, std::vector<int> x_shape, ...@@ -175,7 +175,7 @@ std::vector<int> RealBroadcast(const std::string &op, std::vector<int> x_shape,
output_i = x_i; output_i = x_i;
} else { } else {
MS_LOG(EXCEPTION) MS_LOG(EXCEPTION)
<< "" << op << op
<< " evaluator the shape of first tensor and the shape of second tensor do not meet the broadcasting " << " evaluator the shape of first tensor and the shape of second tensor do not meet the broadcasting "
"requirements"; "requirements";
} }
......
...@@ -623,7 +623,7 @@ void DfGraphConvertor::InitParamWithData(const TensorOrderMap &tensors) { ...@@ -623,7 +623,7 @@ void DfGraphConvertor::InitParamWithData(const TensorOrderMap &tensors) {
auto node_itor = params_.find(name); auto node_itor = params_.find(name);
// if name not in params_, create a node in graph // if name not in params_, create a node in graph
if (node_itor == params_.end()) { if (node_itor == params_.end()) {
MS_LOG(WARNING) << "" << name << " is not in params, and create a new node."; MS_LOG(WARNING) << name << " is not in params, and create a new node.";
ParameterPtr param = anf_graph_->add_parameter(); ParameterPtr param = anf_graph_->add_parameter();
name = name + "_temp"; name = name + "_temp";
param->set_name(name); param->set_name(name);
......
...@@ -216,8 +216,8 @@ void FinalVM::InstCall(const VectorRef &args) { ...@@ -216,8 +216,8 @@ void FinalVM::InstCall(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 1; const size_t args_size = 1;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
...@@ -232,8 +232,8 @@ void FinalVM::InstTailCall(const VectorRef &args) { ...@@ -232,8 +232,8 @@ void FinalVM::InstTailCall(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 3; const size_t args_size = 3;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
...@@ -261,7 +261,7 @@ void FinalVM::InstTailCall(const VectorRef &args) { ...@@ -261,7 +261,7 @@ void FinalVM::InstTailCall(const VectorRef &args) {
void FinalVM::InstSwitchReturn(const VectorRef &args) { void FinalVM::InstSwitchReturn(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
if (args.size() != 1) { if (args.size() != 1) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires one parameter, while the input size is " << args.size() << "."; MS_LOG(ERROR) << __FUNCTION__ << " requires one parameter, while the input size is " << args.size() << ".";
return; return;
} }
Pop(1); Pop(1);
...@@ -272,8 +272,8 @@ void FinalVM::InstReturn(const VectorRef &args) { ...@@ -272,8 +272,8 @@ void FinalVM::InstReturn(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 2; const size_t args_size = 2;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
...@@ -295,7 +295,7 @@ void FinalVM::InstPartial(const VectorRef &args) { ...@@ -295,7 +295,7 @@ void FinalVM::InstPartial(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 1; const size_t args_size = 1;
if (args.size() < args_size) { if (args.size() < args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " or more parameters, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " or more parameters, while the input size is "
<< args.size() << "."; << args.size() << ".";
return; return;
} }
...@@ -314,8 +314,8 @@ void FinalVM::InstPartial(const VectorRef &args) { ...@@ -314,8 +314,8 @@ void FinalVM::InstPartial(const VectorRef &args) {
void FinalVM::InstSimuSwitch(const VectorRef &args) { void FinalVM::InstSimuSwitch(const VectorRef &args) {
const size_t args_size = 4; const size_t args_size = 4;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
bool cond = utils::cast<bool>(args[0]); bool cond = utils::cast<bool>(args[0]);
...@@ -368,8 +368,8 @@ void FinalVM::InstSimuSwitch(const VectorRef &args) { ...@@ -368,8 +368,8 @@ void FinalVM::InstSimuSwitch(const VectorRef &args) {
void FinalVM::InstRealSwitch(const VectorRef &args) { void FinalVM::InstRealSwitch(const VectorRef &args) {
const size_t args_size = 3; const size_t args_size = 3;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameters, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
...@@ -378,7 +378,7 @@ void FinalVM::InstRealSwitch(const VectorRef &args) { ...@@ -378,7 +378,7 @@ void FinalVM::InstRealSwitch(const VectorRef &args) {
int vfalse = utils::cast<int>(args[2]); int vfalse = utils::cast<int>(args[2]);
BaseRef c = Ref(cond); BaseRef c = Ref(cond);
MS_LOG(DEBUG) << "" << vtrue << " false:" << vfalse << " InstSwitch: " << c.ToString(); MS_LOG(DEBUG) << vtrue << " false:" << vfalse << " InstSwitch: " << c.ToString();
bool bool_value = false; bool bool_value = false;
if (backend_->GetCond(c, &bool_value)) { if (backend_->GetCond(c, &bool_value)) {
MS_LOG(DEBUG) << "Cond:" << bool_value; MS_LOG(DEBUG) << "Cond:" << bool_value;
...@@ -417,8 +417,8 @@ void FinalVM::InstPush(const VectorRef &args) { ...@@ -417,8 +417,8 @@ void FinalVM::InstPush(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 1; const size_t args_size = 1;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
...@@ -431,8 +431,8 @@ void FinalVM::InstInput(const VectorRef &args) { ...@@ -431,8 +431,8 @@ void FinalVM::InstInput(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 1; const size_t args_size = 1;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
...@@ -445,13 +445,13 @@ void FinalVM::InstPadStack(const VectorRef &args) { ...@@ -445,13 +445,13 @@ void FinalVM::InstPadStack(const VectorRef &args) {
MS_LOG(DEBUG) << "Start"; MS_LOG(DEBUG) << "Start";
const size_t args_size = 1; const size_t args_size = 1;
if (args.size() != args_size) { if (args.size() != args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " parameter, while the input size is " << args.size()
<< args.size() << "."; << ".";
return; return;
} }
int sz = utils::cast<int>(args[0]); int sz = utils::cast<int>(args[0]);
MS_LOG(DEBUG) << "" << insts_stack_.size() << " need padstack " << sz << " sp_ " << sp_; MS_LOG(DEBUG) << insts_stack_.size() << " need padstack " << sz << " sp_ " << sp_;
size_t stack_size = insts_stack_.size(); size_t stack_size = insts_stack_.size();
int need = sz - (static_cast<int>(stack_size) - sp_); int need = sz - (static_cast<int>(stack_size) - sp_);
if (need > 0) { if (need > 0) {
...@@ -501,7 +501,7 @@ void FinalVM::InstPushPrim(const VectorRef &args) { ...@@ -501,7 +501,7 @@ void FinalVM::InstPushPrim(const VectorRef &args) {
MS_LOG(DEBUG) << "Start: " << args.size(); MS_LOG(DEBUG) << "Start: " << args.size();
const size_t args_size = 2; const size_t args_size = 2;
if (args.size() < args_size) { if (args.size() < args_size) {
MS_LOG(ERROR) << "" << __FUNCTION__ << " requires " << args_size << " or more parameters, while the input size is " MS_LOG(ERROR) << __FUNCTION__ << " requires " << args_size << " or more parameters, while the input size is "
<< args.size() << "."; << args.size() << ".";
return; return;
} }
......
...@@ -445,7 +445,7 @@ BaseRef RunOperation(const PrimitivePtr &prim, const VectorRef &args) { ...@@ -445,7 +445,7 @@ BaseRef RunOperation(const PrimitivePtr &prim, const VectorRef &args) {
MS_LOG(DEBUG) << "operation start " << prim->name(); MS_LOG(DEBUG) << "operation start " << prim->name();
auto func = operation != nullptr ? operation->GetComputeFunction() : prim->GetComputeFunction(); auto func = operation != nullptr ? operation->GetComputeFunction() : prim->GetComputeFunction();
if (py::isinstance<py::none>(func)) { if (py::isinstance<py::none>(func)) {
MS_LOG(EXCEPTION) << "" << prim->name() << " 's compute function is not implemented"; MS_LOG(EXCEPTION) << prim->name() << " 's compute function is not implemented";
} }
py::tuple py_args = py::tuple(args.size()); py::tuple py_args = py::tuple(args.size());
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册