diff --git a/mindspore/ccsrc/common/trans.cc b/mindspore/ccsrc/common/trans.cc index 4748d5928647d9ead247196a8adab1010728aecb..380c51bcf9fca44647e9428a4937e65dab8b835a 100644 --- a/mindspore/ccsrc/common/trans.cc +++ b/mindspore/ccsrc/common/trans.cc @@ -122,7 +122,7 @@ bool CastKernel(const TypeIdArgs &args, void *dst, const size_t data_size, const TransDataSrc2Dst(args, dst, data_size); break; default: - MS_LOG(ERROR) << "unsupported datatype trans"; + MS_LOG(ERROR) << "Unsupported datatype trans"; return false; } return true; @@ -132,7 +132,7 @@ size_t CubeSizeByType(const TypeId data_type) { const size_t default_error = 0; auto dt_size = TypeIdSize(data_type); if (dt_size < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return default_error; } else if (dt_size == 1) { return kCubeSize * 2; @@ -146,12 +146,12 @@ size_t ShapeSize(const std::vector &shape) { } size_t TypeIdSize(const TypeId data_type) { - const size_t unsupport_type_error = 0; + const size_t unsupported_type_error = 0; auto iter = type_map.find(data_type); if (iter != type_map.end()) { return iter->second; } - return unsupport_type_error; + return unsupported_type_error; } std::vector TransShapeTo4d(const std::vector &shape) { @@ -174,7 +174,7 @@ std::vector TransShapeTo4d(const std::vector &shape) { } break; default: - MS_LOG(EXCEPTION) << "Unexpeted shape size = " << shape.size(); + MS_LOG(EXCEPTION) << "Unexpected shape size = " << shape.size(); } return shape_4d; } @@ -183,7 +183,7 @@ std::vector TransShapeToDevice(const std::vector &shape, const s std::vector device_shape; if (format == kOpFormat_FRAC_NZ) { if (shape.size() < 2) { - MS_EXCEPTION(NotSupportError) << "format " << format << " is not support shape " << shape.size(); + MS_EXCEPTION(NotSupportError) << "Format " << format << " is not support shape " << shape.size(); } if (shape.size() > 2) { (void)std::copy(shape.begin(), shape.end() - 2, std::back_inserter(device_shape)); @@ -231,37 +231,37 @@ std::vector TransShapeToDevice(const std::vector &shape, const s } bool TransDataType(const TypeIdArgs &args, void *result) { - MS_LOG(DEBUG) << "begin trans datatype from " << TypeIdLabel(args.host_data_type) << " to " + MS_LOG(DEBUG) << "Begin trans datatype from " << TypeIdLabel(args.host_data_type) << " to " << TypeIdLabel(args.device_data_type); MS_EXCEPTION_IF_NULL(result); std::pair type_info(args.host_data_type, args.device_data_type); auto iter = mode_map.find(type_info); if (iter == mode_map.end()) { - MS_LOG(ERROR) << "unsupported datatype trans. src_type :" << TypeIdLabel(args.host_data_type) + MS_LOG(ERROR) << "Unsupported datatype trans. src_type :" << TypeIdLabel(args.host_data_type) << ", dst_type:" << TypeIdLabel(args.device_data_type); return false; } auto trans_mode = iter->second; auto type_size = TypeIdSize(args.device_data_type); if (type_size < 1) { - MS_LOG(ERROR) << "invalid host data type."; + MS_LOG(ERROR) << "Invalid host data type."; return false; } if (args.host_shape_size < 1) { - MS_LOG(ERROR) << "invalid host data size."; + MS_LOG(ERROR) << "Invalid host data size."; return false; } if (!CastKernel(args, result, args.host_shape_size, trans_mode)) { - MS_LOG(ERROR) << "failed to trans datatype.."; + MS_LOG(ERROR) << "Failed to trans datatype.."; return false; } return true; } bool TransFormat(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "start trans format."; + MS_LOG(DEBUG) << "Start trans format."; if (TypeIdSize(args.src_data_type) < 1) { - MS_LOG(ERROR) << "invalid datatype.."; + MS_LOG(ERROR) << "Invalid datatype.."; return false; } if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && @@ -276,9 +276,9 @@ bool TransFormat(const FormatArgs &args, void *result) { } bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "start trans format."; + MS_LOG(DEBUG) << "Start trans format."; if (TypeIdSize(args.src_data_type) < 1) { - MS_LOG(ERROR) << "invalid datatype.."; + MS_LOG(ERROR) << "Invalid datatype.."; return false; } if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && @@ -293,15 +293,15 @@ bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { } bool NchwToFracZ(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "trans format from nchw to frac_z"; + MS_LOG(DEBUG) << "Trans format from nchw to frac_z"; MS_EXCEPTION_IF_NULL(result); if (args.host_shape.size() != kNchwDims) { - MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; + MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; return false; } size_t size = TypeIdSize(args.src_data_type); if (size < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return false; } auto n = args.host_shape[0]; @@ -311,7 +311,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { size_t c0 = CubeSizeByType(args.src_data_type); if (c0 < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return false; } size_t c1 = Ceil(c, c0); @@ -327,7 +327,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { size_t dst_size = total_ele_cnt * size; if (dst_size != args.device_size) { - MS_LOG(ERROR) << "illegal total data size." + MS_LOG(ERROR) << "Illegal total data size." << "dst size is :" << dst_size << "device size is :" << args.device_size; return false; } @@ -369,20 +369,20 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { } bool FracZToNchw(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "trans format from frac_z to nchw"; + MS_LOG(DEBUG) << "Trans format from frac_z to nchw"; MS_EXCEPTION_IF_NULL(result); if (args.host_shape.size() != kNchwDims) { - MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; + MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; return false; } size_t size = TypeIdSize(args.src_data_type); if (size < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return false; } size_t total_size = ShapeSize(args.device_shape) * size; if (total_size != args.device_size) { - MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; + MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; return false; } @@ -435,7 +435,7 @@ bool FracZToNchw(const FormatArgs &args, void *result) { bool TransShapeToNz(const std::vector &host_shape, std::vector *hw_shape) { MS_EXCEPTION_IF_NULL(hw_shape); if (host_shape.empty()) { - MS_LOG(ERROR) << "size of vector is 0."; + MS_LOG(ERROR) << "Size of vector is 0."; return false; } switch (host_shape.size()) { @@ -447,7 +447,7 @@ bool TransShapeToNz(const std::vector &host_shape, std::vector * default: auto size = host_shape.size(); if (size < 2) { - MS_LOG(ERROR) << "illegal size."; + MS_LOG(ERROR) << "Illegal size."; return false; } size_t times = 1; @@ -462,26 +462,26 @@ bool TransShapeToNz(const std::vector &host_shape, std::vector * } bool NchwToFracNz(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "trans format from nchw to frac_nz."; + MS_LOG(DEBUG) << "Trans format from nchw to frac_nz."; MS_EXCEPTION_IF_NULL(result); std::vector hw_shape; if (!TransShapeToNz(args.host_shape, &hw_shape)) { - MS_LOG(ERROR) << "trans shape failed.."; + MS_LOG(ERROR) << "Trans shape failed.."; return false; } if (hw_shape.size() < 3 || args.device_shape.size() < 4) { - MS_LOG(ERROR) << "invalid shape size."; + MS_LOG(ERROR) << "Invalid shape size."; return false; } auto size = TypeIdSize(args.src_data_type); if (size < 1) { - MS_LOG(ERROR) << "illegal dtype"; + MS_LOG(ERROR) << "Illegal dtype"; return false; } auto dst_size = ShapeSize(args.device_shape) * size; if (dst_size != args.device_size) { - MS_LOG(ERROR) << "illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; + MS_LOG(ERROR) << "Illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; return false; } auto times = hw_shape.at(0); @@ -538,26 +538,26 @@ bool NchwToFracNz(const FormatArgs &args, void *result) { } bool FracNzToNchw(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "trans format from frac_nz to nchw"; + MS_LOG(DEBUG) << "Trans format from frac_nz to nchw"; MS_EXCEPTION_IF_NULL(result); std::vector hw_shape; if (!TransShapeToNz(args.host_shape, &hw_shape)) { - MS_LOG(ERROR) << "trans shape failed.."; + MS_LOG(ERROR) << "Trans shape failed.."; return false; } if (hw_shape.size() < 3 || args.device_shape.size() < 4) { - MS_LOG(ERROR) << "invalid shape size."; + MS_LOG(ERROR) << "Invalid shape size."; return false; } auto size = TypeIdSize(args.src_data_type); if (size < 1) { - MS_LOG(ERROR) << "illegal dtype"; + MS_LOG(ERROR) << "Illegal dtype"; return false; } auto dst_size = ShapeSize(args.device_shape) * size; if (dst_size != args.device_size) { - MS_LOG(ERROR) << "illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; + MS_LOG(ERROR) << "Illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; return false; } auto times = hw_shape.at(0); @@ -614,20 +614,20 @@ bool FracNzToNchw(const FormatArgs &args, void *result) { } bool NchwToNc1hwc0(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "trans format from nchw to Nc1h1wc0"; + MS_LOG(DEBUG) << "Trans format from nchw to Nc1h1wc0"; MS_EXCEPTION_IF_NULL(result); if (args.host_shape.size() != kNchwDims) { - MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; + MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; return false; } size_t size = TypeIdSize(args.src_data_type); if (size < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return false; } auto total_size = ShapeSize(args.device_shape) * size; if (total_size != args.device_size) { - MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; + MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; return false; } @@ -637,7 +637,7 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) { auto w = args.host_shape[3]; size_t c0 = CubeSizeByType(args.src_data_type); if (c0 < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return false; } size_t c1 = Ceil(c, c0); @@ -687,20 +687,20 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) { } bool Nc1hwc0ToNchw(const FormatArgs &args, void *result) { - MS_LOG(DEBUG) << "trans format from nc1h1wc0 to nchw"; + MS_LOG(DEBUG) << "Trans format from nc1h1wc0 to nchw"; MS_EXCEPTION_IF_NULL(result); if (args.host_shape.size() != kNchwDims) { - MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; + MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; return false; } size_t size = TypeIdSize(args.src_data_type); if (size < 1) { - MS_LOG(ERROR) << "illegal dtype."; + MS_LOG(ERROR) << "Illegal dtype."; return false; } size_t total_size = ShapeSize(args.device_shape) * size; if (total_size != args.device_size) { - MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; + MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; return false; } diff --git a/mindspore/ccsrc/debug/anf_ir_dump.cc b/mindspore/ccsrc/debug/anf_ir_dump.cc index 9e24a17edee751446ea9ad7107492230177a3602..9eb0a376cc41074720f65f76425a5364791134ab 100644 --- a/mindspore/ccsrc/debug/anf_ir_dump.cc +++ b/mindspore/ccsrc/debug/anf_ir_dump.cc @@ -141,7 +141,7 @@ void DumpKernelInfo(const CNodePtr &node, const std::shared_ptr void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMap *para_map) { if (graph == nullptr) { - MS_LOG(INFO) << "param graph is nullptr."; + MS_LOG(INFO) << "Param graph is nullptr."; return; } std::vector parameters = graph->parameters(); @@ -175,17 +175,17 @@ void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMa if (para_map != nullptr) { (*para_map)[p] = para++; } - MS_LOG(DEBUG) << "record param: " << p->ToString() << " graph belong : " << p->func_graph()->ToString(); + MS_LOG(DEBUG) << "Record param: " << p->ToString() << " graph belong : " << p->func_graph()->ToString(); } } void DumpOperator(const AnfNodePtr &op, const std::shared_ptr &gsub) { if (op == nullptr) { - MS_LOG(INFO) << "param op is nullptr"; + MS_LOG(INFO) << "Param op is nullptr"; return; } if (gsub == nullptr) { - MS_LOG(INFO) << "param gsub is nullptr"; + MS_LOG(INFO) << "Param gsub is nullptr"; return; } @@ -338,7 +338,7 @@ void DumpCNode(const CNodePtr &nd, const FuncGraphPtr &sub_graph, OrderedMapinputs().empty()) { - MS_LOG(EXCEPTION) << "input of apply node is empty"; + MS_LOG(EXCEPTION) << "Input of apply node is empty"; } // print operator @@ -376,7 +376,7 @@ void DumpIRInSubgraph(const std::vector &nodes, OrderedMapfunc_graph(); if (sub_graph == nullptr) { - MS_LOG(DEBUG) << "node[" << nd->ToString() << "] belongs to no graph!"; + MS_LOG(DEBUG) << "Node[" << nd->ToString() << "] belongs to no graph!"; continue; } std::shared_ptr gsub = (*sub_graphs)[sub_graph]; @@ -430,12 +430,12 @@ void DumpIR(const std::string &filename, const FuncGraphPtr &graph, bool dump_fu return; } if (filename.size() > PATH_MAX) { - MS_LOG(ERROR) << "file path " << filename << " is too long."; + MS_LOG(ERROR) << "File path " << filename << " is too long."; return; } char real_path[PATH_MAX] = {0}; if (nullptr == realpath(filename.c_str(), real_path)) { - MS_LOG(DEBUG) << "dir " << filename << " does not exit."; + MS_LOG(DEBUG) << "Dir " << filename << " does not exit."; } OrderedMap para_map; diff --git a/mindspore/ccsrc/debug/anf_ir_utils.cc b/mindspore/ccsrc/debug/anf_ir_utils.cc index 98cd2f4b2f52d99acb20f0eac59e292974fd59f2..c25ad862df515fd9c5c146791dce02cec77f02e1 100644 --- a/mindspore/ccsrc/debug/anf_ir_utils.cc +++ b/mindspore/ccsrc/debug/anf_ir_utils.cc @@ -49,7 +49,7 @@ std::string GetMsIrPath(void) { path = path_ptr; char real_path[PATH_MAX] = {0}; if (path.size() > PATH_MAX || nullptr == realpath(path.c_str(), real_path)) { - MS_LOG(EXCEPTION) << "MS IR Path error, " << path_ptr; + MS_LOG(EXCEPTION) << "MS IR path error, " << path_ptr; } path = real_path; } @@ -144,8 +144,8 @@ std::string AnfExporter::GetValueNodeText(const FuncGraphPtr& fg, const ValueNod } std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGraphPtr& mt_func_graph) { - auto py_funs = mt_func_graph->GetPyFunctions(); - if (py_funs.empty()) { + auto py_funcs = mt_func_graph->GetPyFunctions(); + if (py_funcs.empty()) { return ""; } @@ -153,7 +153,7 @@ std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGrap oss << "{"; bool is_first = true; - for (const auto& py_func : py_funs) { + for (const auto& py_func : py_funcs) { if (is_first) { is_first = false; } else { @@ -626,7 +626,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const FuncGraphPt ofs << "\n\n"; (void)func_graph_set.erase(fg); } - ofs << "# num of total funcgraphs: " << exported.size(); + ofs << "# num of total function graphs: " << exported.size(); ofs.close(); } @@ -651,7 +651,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const std::vector ofs << "\n\n"; } - ofs << "# num of total funcgraphs: " << graphs.size(); + ofs << "# num of total function graphs: " << graphs.size(); ofs.close(); } @@ -763,7 +763,7 @@ class Lexer { fin.close(); } } catch (const std::exception& e) { - MS_LOG(ERROR) << "exception when closing file"; + MS_LOG(ERROR) << "Exception when closing file"; } catch (...) { std::string exName(abi::__cxa_current_exception_type()->name()); MS_LOG(ERROR) << "Error occurred when closing file. Exception name: " << exName; @@ -802,7 +802,7 @@ class Lexer { Token token = GetNextTokenInner(); const char* str = token_text[token]; std::string text = (str == nullptr ? GetTokenText() : str); - MS_LOG(DEBUG) << "------parse token] " << text; + MS_LOG(DEBUG) << "------Parse token] " << text; return token; } @@ -1642,7 +1642,7 @@ class IrParser { MS_LOG(EXCEPTION) << "Expect @file at line " << lexer_.GetLineNo(); } - // load prameter default value from serialized file + // load parameter default value from serialized file py::object default_obj = LoadObject(lexer_.GetTokenText()); param->set_default_param(default_obj); @@ -1950,7 +1950,7 @@ class IrParser { return TOK_ERROR; } - // restore python funciton of PrimitivePy from serialized file + // restore python function of PrimitivePy from serialized file py::object py_obj = LoadObject(lexer_.GetTokenText()); PrimitivePyPtr ptr = nullptr; if (py::hasattr(py_obj, "__setattr_flag__") && py::hasattr(py_obj, "_clone")) { @@ -1958,7 +1958,7 @@ class IrParser { py::object new_obj = clone_fn(); ptr = new_obj.cast(); if (ptr == nullptr) { - MS_LOG(EXCEPTION) << "cast to type 'PrimitivePyPtr' error"; + MS_LOG(EXCEPTION) << "Cast to type 'PrimitivePyPtr' error"; } } else { ptr = std::make_shared(id.substr(strlen("PrimitivePy::")), py_obj); @@ -2221,15 +2221,15 @@ class IrParser { }; std::vector ImportIR(const std::string& filename) { - IrParser paser(filename.c_str()); - paser.ParseFile(); - return paser.GetFuncGraphs(); + IrParser parser(filename.c_str()); + parser.ParseFile(); + return parser.GetFuncGraphs(); } #ifdef ENABLE_DUMP_IR void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { if (func_graph == nullptr) { - MS_LOG(ERROR) << "func graph is nullptr"; + MS_LOG(ERROR) << "Func graph is nullptr"; return; } auto ms_context = MsContext::GetInstance(); @@ -2243,16 +2243,16 @@ void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { } std::string file_path = save_graphs_path + "/" + "ms_output_" + suffix + ".pb"; if (file_path.size() > PATH_MAX) { - MS_LOG(ERROR) << "file path " << file_path << " is too long."; + MS_LOG(ERROR) << "File path " << file_path << " is too long."; return; } char real_path[PATH_MAX] = {0}; if (nullptr == realpath(file_path.c_str(), real_path)) { - MS_LOG(DEBUG) << "dir " << file_path << " does not exit."; + MS_LOG(DEBUG) << "Dir " << file_path << " does not exit."; } else { std::string path_string = real_path; if (chmod(common::SafeCStr(path_string), S_IRUSR | S_IWUSR) == -1) { - MS_LOG(ERROR) << "modify file:" << real_path << " to rw fail."; + MS_LOG(ERROR) << "Modify file:" << real_path << " to rw fail."; return; } } diff --git a/mindspore/ccsrc/debug/draw.cc b/mindspore/ccsrc/debug/draw.cc index e0949c04194c11c6507f3dc590dc62dd382cdebf..3e8cbfba1940eeff9104ddb861111805a6404265 100644 --- a/mindspore/ccsrc/debug/draw.cc +++ b/mindspore/ccsrc/debug/draw.cc @@ -362,7 +362,7 @@ Digraph::~Digraph() { fout_.close(); } } catch (const std::exception& e) { - MS_LOG(ERROR) << "exception when closing file " << filename_; + MS_LOG(ERROR) << "Exception when closing file " << filename_; } } diff --git a/mindspore/ccsrc/debug/dump_proto.cc b/mindspore/ccsrc/debug/dump_proto.cc index a12cf8b47f18331d99bd9780c3a8e0796d6e0f3f..a7a1e208a4e7fb30cd7cc2e2ff8484ecf122b01d 100644 --- a/mindspore/ccsrc/debug/dump_proto.cc +++ b/mindspore/ccsrc/debug/dump_proto.cc @@ -208,7 +208,7 @@ void ProtoExporter::SetValueToProto(const ValuePtr& val, irpb::ValueProto* value TypePtr elem_type = dyn_cast(val)->element(); type_proto->mutable_tensor_type()->set_elem_type(GetNumberDataType(elem_type)); } else { - MS_LOG(WARNING) << "Not supported type " << val->type_name(); + MS_LOG(WARNING) << "Unsupported type " << val->type_name(); } } diff --git a/mindspore/ccsrc/debug/e2e_dump.cc b/mindspore/ccsrc/debug/e2e_dump.cc index ba11eafa5f668650f77d754081088422b2609c17..fbe76cdc4744c6eb5e15b98dd1c556cf2fd060e5 100644 --- a/mindspore/ccsrc/debug/e2e_dump.cc +++ b/mindspore/ccsrc/debug/e2e_dump.cc @@ -101,7 +101,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) { auto kernels = dumpSettings.at("kernels"); if (!(enable.is_boolean() && trans_flag.is_boolean() && mode.is_number() && path.is_string() && net_name.is_string() && iteration.is_number() && kernels.is_array())) { - MS_LOG(ERROR) << "element's type in Dump config json is invalid."; + MS_LOG(ERROR) << "Element's type in Dump config json is invalid."; dump_enable_ = false; return false; } @@ -121,7 +121,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) { bool Dump::SetDumpConfFromJsonFile() { const char* config_path_str = std::getenv("MINDSPORE_CONFIG_PATH"); if (config_path_str != nullptr) { - MS_LOG(INFO) << "getenv MINDSPORE_CONFIG_PATH :" << config_path_str; + MS_LOG(INFO) << "Getenv MINDSPORE_CONFIG_PATH :" << config_path_str; } else { MS_LOG(INFO) << "No need E2E Dump. please export MINDSPORE_CONFIG_PATH eg: MINDSPORE_CONFIG_PATH=/etc"; dump_enable_ = false; @@ -132,7 +132,7 @@ bool Dump::SetDumpConfFromJsonFile() { auto id = context_ptr->device_id(); char real_path[PATH_MAX] = {0}; if (nullptr == realpath(config_path_str, real_path)) { - MS_LOG(ERROR) << "env e2e dump path error, " << config_path_str; + MS_LOG(ERROR) << "Env e2e dump path error, " << config_path_str; dump_enable_ = false; return false; } @@ -150,20 +150,20 @@ bool Dump::SetDumpConfFromJsonFile() { bool Dump::DumpToFile(const std::string& filename, const void* data, size_t len) { if (filename.empty() || data == nullptr || len == 0) { - MS_LOG(ERROR) << "incorrect parameter."; + MS_LOG(ERROR) << "Incorrect parameter."; return false; } std::string realpath; bool ret = GetRealPath(filename, &realpath); if (!ret) { - MS_LOG(ERROR) << "get real path failed."; + MS_LOG(ERROR) << "Get real path failed."; return false; } std::ofstream fd; fd.open(realpath, std::ios::binary | std::ios::out); if (!fd.is_open()) { - MS_LOG(ERROR) << "open file " << realpath << " fail."; + MS_LOG(ERROR) << "Open file " << realpath << " fail."; return false; } (void)fd.write(reinterpret_cast(data), SizeToLong(len)); @@ -182,7 +182,7 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) { if (path_split_pos != std::string::npos) { std::string prefix_path = inpath.substr(0, path_split_pos); if (prefix_path.length() >= PATH_MAX) { - MS_LOG(ERROR) << "prefix path is too longer!"; + MS_LOG(ERROR) << "Prefix path is too longer!"; return false; } std::string last_path = inpath.substr(path_split_pos, inpath.length() - path_split_pos); @@ -201,11 +201,11 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) { if (path_split_pos == std::string::npos) { if (inpath.length() >= PATH_MAX) { - MS_LOG(ERROR) << "prefix path is too longer!"; + MS_LOG(ERROR) << "Prefix path is too longer!"; return false; } if (nullptr == realpath(inpath.c_str(), real_path)) { - MS_LOG(ERROR) << "file " << inpath << " does not exit, it will be created."; + MS_LOG(ERROR) << "File " << inpath << " does not exit, it will be created."; } *outpath = std::string(real_path); } @@ -218,7 +218,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { MS_EXCEPTION_IF_NULL(fs); char temp_path[PATH_MAX] = {0}; if (path.length() > PATH_MAX) { - MS_LOG(ERROR) << "path lens is max than " << PATH_MAX; + MS_LOG(ERROR) << "Path lens is max than " << PATH_MAX; return false; } for (uint32_t i = 0; i < path.length(); i++) { @@ -229,7 +229,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { temp_path[i] = '\0'; std::string path_handle(temp_path); if (!fs->FileExist(temp_path)) { - MS_LOG(INFO) << "dir " << path_handle << " does not exit, creating..."; + MS_LOG(INFO) << "Dir " << path_handle << " does not exit, creating..."; if (!fs->CreateDir(temp_path)) { MS_LOG(ERROR) << "Create " << path_handle << " dir error"; return false; @@ -241,7 +241,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { } if (!fs->FileExist(path)) { - MS_LOG(INFO) << "dir " << path << " does not exit, creating..."; + MS_LOG(INFO) << "Dir " << path << " does not exit, creating..."; if (!fs->CreateDir(path)) { MS_LOG(ERROR) << "Create " << path << " dir error"; return false; diff --git a/mindspore/ccsrc/debug/info.cc b/mindspore/ccsrc/debug/info.cc index 6f966e335ca7d6b97c520bb04edda261c976394f..5c1fc372c5fa5d47a454524a30606aad537f6392 100644 --- a/mindspore/ccsrc/debug/info.cc +++ b/mindspore/ccsrc/debug/info.cc @@ -193,7 +193,7 @@ void TraceManager::DebugTrace(const TraceInfoPtr& trace_info) { } TraceContextPtr context = std::make_shared(trace_info); if (trace_info->debug_info() == nullptr) { - MS_LOG(EXCEPTION) << "trace debug info is null"; + MS_LOG(EXCEPTION) << "Trace debug info is null"; } TraceManager::trace_context_stack_.push(context); } @@ -205,7 +205,7 @@ void TraceManager::DebugTrace(const DebugInfoPtr& debug_info, const TraceInfoPtr auto cloned_info = trace_info->clone(); cloned_info->set_debug_info(debug_info); if (cloned_info->debug_info() == nullptr) { - MS_LOG(EXCEPTION) << "trace debug info is null with cloned trace"; + MS_LOG(EXCEPTION) << "Trace debug info is null with cloned trace"; } TraceContextPtr context = std::make_shared(cloned_info); TraceManager::trace_context_stack_.push(context); diff --git a/mindspore/ccsrc/debug/trace.cc b/mindspore/ccsrc/debug/trace.cc index 7ce13052c562e642f953185869f1204c849bc411..51cdd349463f9556b0465d5d330a841f7684cd4a 100644 --- a/mindspore/ccsrc/debug/trace.cc +++ b/mindspore/ccsrc/debug/trace.cc @@ -89,7 +89,7 @@ std::string GetDebugInfo(const DebugInfoPtr& info, SourceLineTip tip) { return ""; } -// a trace info identifys a node transform, so we can trace the node transform through +// a trace info identifies a node transform, so we can trace the node transform through // a link of trace info and debug info std::string GetInfoWithAction(const std::vector& info_vec, SourceLineTip tip) { if (info_vec.size() < 1) { @@ -173,7 +173,7 @@ void DumpInferStack(std::ostringstream& oss) { } auto graph_context = graph_infer->graph_context(); if (graph_context == nullptr) { - MS_LOG(INFO) << "null context continue"; + MS_LOG(INFO) << "Null context continue"; continue; } auto graph = graph_context->func_graph(); @@ -264,7 +264,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename, param_index = 1; auto tagged_func_graphs = CalcTaggedFuncGraphs(); - // first output grapn on the analysis stack + // first output graph on the analysis stack for (const auto& node_cfg : node_cfgs) { auto fg = node_cfg->context()->func_graph(); // the graph is already output, skip it @@ -291,7 +291,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename, ofs << "\n\n"; (void)func_graph_set.erase(fg); } - ofs << "# num of total funcgraphs: " << exported.size(); + ofs << "# num of total function graphs: " << exported.size(); ofs.close(); } @@ -332,7 +332,7 @@ void GetInferStackInfo(std::ostringstream& oss) { MS_LOG(INFO) << "Get graph analysis information *end*"; } -// trace the graph evaluator statck +// trace the graph evaluator stack static std::stack> graph_infer_stack; // trace the cnode infer debug info static std::vector cnode_debug_stack{}; diff --git a/mindspore/ccsrc/debug/trace_info.cc b/mindspore/ccsrc/debug/trace_info.cc index e2da4ffcce925c40805310384f0f69bee4e6dcf8..b01cd1501016210a6e8c72a1c3967f64acdb1fd4 100644 --- a/mindspore/ccsrc/debug/trace_info.cc +++ b/mindspore/ccsrc/debug/trace_info.cc @@ -36,6 +36,6 @@ std::string TraceInfo::GetActionBetweenNode(const DebugInfoPtr& info) { } else if (debug_info()->trace_info() != nullptr) { return act_name + debug_info()->trace_info()->GetActionBetweenNode(info); } - return "not in the traced info"; + return "Not in the traced info"; } } // namespace mindspore diff --git a/mindspore/ccsrc/ir/anf.h b/mindspore/ccsrc/ir/anf.h index e64b1329e96260762548f2f0bf592648c5d9de04..16ccb15c43dcc4c6efccb8839187d5fdf958b352 100644 --- a/mindspore/ccsrc/ir/anf.h +++ b/mindspore/ccsrc/ir/anf.h @@ -83,7 +83,7 @@ class AnfVisitor; // Methods: // func_graph: return FuncGraph that this AnfNode belongs to. // scope: return the scope namespace of this AnfNode. Set it using set_scope. -// abstract: return the cached inferred abstract value. It cantains type, shape +// abstract: return the cached inferred abstract value. It contains type, shape // value. Set New cache using set_abstract. // intermediate_abstract: return the cached inferring abstract value. // Type/Shape: return the related info of this AnfNode. When this AnfNode is an @@ -284,7 +284,7 @@ class Parameter : public ANode { }; using ParameterPtr = std::shared_ptr; -// Value is used to represent the atomic expression metioned in BNF. +// Value is used to represent the atomic expression mentioned in BNF. // It mainly be stored in ValueNode. Value and ValueNode is related definition. class Value : public Base { public: @@ -313,7 +313,7 @@ using ValuePtr = std::shared_ptr; using ValuePtrList = std::vector; // ValueNode is used to hold value. Unlike CNode and Parameter, ValueNode -// do not belong to any particular function graph. +// does not belong to any particular function graph. class ValueNode : public ANode { public: explicit ValueNode(const ValuePtr &value) : value_(value) {} diff --git a/mindspore/ccsrc/ir/dtype/number.cc b/mindspore/ccsrc/ir/dtype/number.cc index 70d0aaeeaa3b7a309d6fecb62d8eeabbbb1be53b..d9ef6bb3bd155f3371f587d9d1d7c94eab86fd8e 100644 --- a/mindspore/ccsrc/ir/dtype/number.cc +++ b/mindspore/ccsrc/ir/dtype/number.cc @@ -34,19 +34,19 @@ bool Number::operator==(const Type& other) const { Int::Int(const int nbits) : Number(IntBitsToTypeId(nbits), nbits, false) { if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { - MS_LOG(EXCEPTION) << "wrong number of bits."; + MS_LOG(EXCEPTION) << "Wrong number of bits."; } } UInt::UInt(const int nbits) : Number(UIntBitsToTypeId(nbits), nbits, false) { if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { - MS_LOG(EXCEPTION) << "wrong number of bits."; + MS_LOG(EXCEPTION) << "Wrong number of bits."; } } Float::Float(const int nbits) : Number(FloatBitsToTypeId(nbits), nbits, false) { if (nbits != 16 && nbits != 32 && nbits != 64) { - MS_LOG(EXCEPTION) << "wrong number of bits."; + MS_LOG(EXCEPTION) << "Wrong number of bits."; } } diff --git a/mindspore/ccsrc/ir/dtype/type.cc b/mindspore/ccsrc/ir/dtype/type.cc index 6169acabbc69773ef2c00aac2e1bc8f1af74b10b..6fbd7f81110443822b0b08c716b5b02ff202efb0 100644 --- a/mindspore/ccsrc/ir/dtype/type.cc +++ b/mindspore/ccsrc/ir/dtype/type.cc @@ -37,7 +37,7 @@ TypeId IntBitsToTypeId(const int nbits) { case 64: return kNumberTypeInt64; default: - MS_LOG(EXCEPTION) << "wrong number of bits."; + MS_LOG(EXCEPTION) << "Wrong number of bits."; } } @@ -52,7 +52,7 @@ TypeId UIntBitsToTypeId(const int nbits) { case 64: return kNumberTypeUInt64; default: - MS_LOG(EXCEPTION) << "wrong number of bits."; + MS_LOG(EXCEPTION) << "Wrong number of bits."; } } @@ -65,7 +65,7 @@ TypeId FloatBitsToTypeId(const int nbits) { case 64: return kNumberTypeFloat64; default: - MS_LOG(EXCEPTION) << "wrong number of bits."; + MS_LOG(EXCEPTION) << "Wrong number of bits."; } } diff --git a/mindspore/ccsrc/ir/func_graph.h b/mindspore/ccsrc/ir/func_graph.h index 13e8c9dc8fca0cf4228ce5009f7436516616fed6..1d58c907559831d80bec1cadefb9e37907cc5f04 100644 --- a/mindspore/ccsrc/ir/func_graph.h +++ b/mindspore/ccsrc/ir/func_graph.h @@ -174,7 +174,7 @@ class FuncGraph : public FuncGraphBase { GraphDebugInfoPtr debug_info(); void set_debug_info(const GraphDebugInfoPtr &info) { if (info == nullptr) { - MS_LOG(EXCEPTION) << "graph set null debug info"; + MS_LOG(EXCEPTION) << "Graph set null debug info"; } this->debug_info_ = info; } diff --git a/mindspore/ccsrc/ir/manager.cc b/mindspore/ccsrc/ir/manager.cc index 09bc7127c1e6e6c2fe46e184670299a80d69adce..c1459014bbce6b20111823414b8adcd5f2d0ff49 100644 --- a/mindspore/ccsrc/ir/manager.cc +++ b/mindspore/ccsrc/ir/manager.cc @@ -817,7 +817,7 @@ void FuncGraphChildDirect::OnMoveAllCNode(FuncGraphPtr src, FuncGraphPtr dst) { void FuncGraphParentsDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) { MS_EXCEPTION_IF_NULL(node); FuncGraphPtr fg1 = node->func_graph(); - // possible chirld parent + // possible child parent if (IsValueNode(inp)) { FuncGraphPtr fg2 = GetValueNode(inp); if (Mod(fg1, ParentProxy(fg2), direction)) { @@ -1181,7 +1181,7 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr& fg, const FuncGraphSetPt } path->add(fg); - // checkg if func graphs used contains J(func_graph); + // check if func graphs used contains J(func_graph); auto& used = this->manager_->func_graphs_used(); for (auto& item : used[fg]) { auto used_g = item.first; diff --git a/mindspore/ccsrc/ir/manager.h b/mindspore/ccsrc/ir/manager.h index 8036bd68c0f401aa2ff96d03e5761912462b6054..dc8f656ae71fc1c034fa150904651bfb77e4baf3 100644 --- a/mindspore/ccsrc/ir/manager.h +++ b/mindspore/ccsrc/ir/manager.h @@ -650,7 +650,7 @@ class FuncGraphTransaction { explicit FuncGraphTransaction(FuncGraphManager* manager) : manager_(manager), changes_() { MS_EXCEPTION_IF_NULL(manager_); if (!manager_->IsManaged()) { - MS_LOG(DEBUG) << "the manager is not managed yet"; + MS_LOG(DEBUG) << "The manager is not managed yet"; } } diff --git a/mindspore/ccsrc/ir/meta_tensor.h b/mindspore/ccsrc/ir/meta_tensor.h index 57c37851b672f2f0b2fe3f4801f269e2ca235d9a..3e28f29f37cef43c2af7023736f3f81d9fa58bc0 100644 --- a/mindspore/ccsrc/ir/meta_tensor.h +++ b/mindspore/ccsrc/ir/meta_tensor.h @@ -148,7 +148,7 @@ class MetaTensor : public Value { // // The constructed MetaTensor object has the same type and shape with meta_tensor. // - // param meta_tensor An exisiting MetaTensor object. + // param meta_tensor An existing MetaTensor object. virtual MetaTensor& operator=(const MetaTensor& meta_tensor); // brief Compares two MetaTensor objects. @@ -166,7 +166,7 @@ class MetaTensor : public Value { TypeId data_type() const { return data_type_; } std::string ToString() const override; std::string DumpText() const override; - // bried Sets the data type of a tensor in its MetaTensor. + // brief Sets the data type of a tensor in its MetaTensor. // // param data_type The data type of the tensor to be set. virtual TypeId set_data_type(const TypeId data_type) { @@ -314,7 +314,7 @@ class Tensor : public MetaTensor { // // The constructed Tensor object has the same type and shape with tensor. // - // param tensor An exisiting Tensor object. + // param tensor An existing Tensor object. Tensor& operator=(const Tensor& tensor); // brief Compares two Tensor objects. @@ -383,7 +383,7 @@ class Tensor : public MetaTensor { // return The [TypeId] of the tensor data. TypeId GetDataType(const py::buffer_info& buf) const; - // bried Sets the data type of a tensor. + // brief Sets the data type of a tensor. // // param data_type The data type of the tensor to be set. // diff --git a/mindspore/ccsrc/ir/visitor.cc b/mindspore/ccsrc/ir/visitor.cc index 70e22b2808544fd80c3753a4561b15232970cb76..efebe3124acce0a649962be553e5a47baf06b11e 100644 --- a/mindspore/ccsrc/ir/visitor.cc +++ b/mindspore/ccsrc/ir/visitor.cc @@ -43,14 +43,13 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vectorcast()->inputs(); - // infact, funcs_size == inps_size - 1 auto funcs_size = funcs.size(); - auto inps_size = inputs.size(); + auto inputs_size = inputs.size(); // check the inputs are matched with the predicate functions if (funcs_size > 0) { // use the predicate function list to check the number of inputs - if (funcs_size != (inps_size - 1)) { + if (funcs_size != (inputs_size - 1)) { return; } @@ -63,7 +62,7 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vectorVisit(inputs[i]); } }; diff --git a/mindspore/ccsrc/pynative/pynative_execute.cc b/mindspore/ccsrc/pynative/pynative_execute.cc index 927e768bbeaa520f6421f0a9ec11bb0956c9b529..e5fb0c69497ae9a3abca07c12eb9851d0bbbb3f0 100644 --- a/mindspore/ccsrc/pynative/pynative_execute.cc +++ b/mindspore/ccsrc/pynative/pynative_execute.cc @@ -36,7 +36,7 @@ #endif const char SINGLE_OP_GRAPH[] = "single_op_graph"; -// primitive unable to infer value for constant input in pynative mode +// primitive unable to infer value for constant input in PyNative mode const std::unordered_set vm_operators = {"partial", "depend"}; namespace mindspore { @@ -45,7 +45,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) { ValuePtr converted_ret = nullptr; bool converted = parse::ConvertData(obj, &converted_ret); if (!converted) { - MS_LOG(EXCEPTION) << "attribute convert error with type:" << std::string(py::str(obj)); + MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj)); } return converted_ret; } @@ -67,7 +67,7 @@ void PynativeInfer(const PrimitivePyPtr& prim, const py::tuple& py_args, OpExecI OpExecInfoPtr GenerateOpExecInfo(const py::args& args) { if (args.size() != PY_ARGS_NUM) { - MS_LOG(ERROR) << "four args are needed by RunOp"; + MS_LOG(ERROR) << "Four args are needed by RunOp"; return nullptr; } auto op_exec_info = std::make_shared(); @@ -145,13 +145,13 @@ py::object RunOpInVM(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat py::object RunOpInMs(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* status) { MS_EXCEPTION_IF_NULL(op_exec_info); - MS_LOG(INFO) << "start run op[" << op_exec_info->op_name << "] with backend policy ms"; + MS_LOG(INFO) << "Start run op[" << op_exec_info->op_name << "] with backend policy ms"; auto ms_context = MsContext::GetInstance(); MS_EXCEPTION_IF_NULL(ms_context); ms_context->set_enable_pynative_infer(true); std::string device_target = ms_context->device_target(); if (device_target != kAscendDevice && device_target != kGPUDevice) { - MS_EXCEPTION(ArgumentError) << "device target [" << device_target << "] is not supported in Pynative mode"; + MS_EXCEPTION(ArgumentError) << "Device target [" << device_target << "] is not supported in Pynative mode"; } std::shared_ptr session = session::SessionFactory::Get().Create(device_target); MS_EXCEPTION_IF_NULL(session); @@ -197,7 +197,7 @@ py::object RunOpWithBackendPolicy(MsBackendPolicy backend_policy, const OpExecIn break; } default: - MS_LOG(ERROR) << "No backend configed for run op"; + MS_LOG(ERROR) << "No backend configured for run op"; } return result; } @@ -240,7 +240,7 @@ py::tuple RunOp(const py::args& args) { } result = RunOpWithBackendPolicy(backend_policy, op_exec_info, &status); if (status != PYNATIVE_SUCCESS) { - MS_LOG(ERROR) << "Fail to run " << op_exec_info->op_name; + MS_LOG(ERROR) << "Failed to run " << op_exec_info->op_name; return err_ret; } diff --git a/mindspore/ccsrc/pynative/pynative_execute_ge.cc b/mindspore/ccsrc/pynative/pynative_execute_ge.cc index 4ed6088494636edfb4f6c0cef90d5a5fc49819d6..180b0006ffb271538a4c63a7d2fc69a824a3580d 100644 --- a/mindspore/ccsrc/pynative/pynative_execute_ge.cc +++ b/mindspore/ccsrc/pynative/pynative_execute_ge.cc @@ -47,7 +47,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) { ValuePtr converted_ret = nullptr; bool converted = parse::ConvertData(obj, &converted_ret); if (!converted) { - MS_LOG(EXCEPTION) << "attribute convert error with type:" << std::string(py::str(obj)); + MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj)); } return converted_ret; } @@ -67,7 +67,7 @@ MeTensorPtr ConvertPyObjToTensor(const py::object& obj) { } else if (py::isinstance(obj)) { me_tensor_ptr = std::make_shared(py::cast(obj), nullptr); } else { - MS_LOG(EXCEPTION) << "run op inputs type is invalid!"; + MS_LOG(EXCEPTION) << "Run op inputs type is invalid!"; } return me_tensor_ptr; } @@ -97,7 +97,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec auto const_op_desc = transform::TransformUtil::GetGeTensorDesc(me_tensor_ptr->shape_c(), me_tensor_ptr->data_type(), kOpFormat_NCHW); if (const_op_desc == nullptr) { - MS_LOG(ERROR) << "Create variable " << op_name << " ouptut descriptor failed!"; + MS_LOG(ERROR) << "Create variable " << op_name << " output descriptor failed!"; return false; } auto pointer_cast_const_op = std::static_pointer_cast(const_op); @@ -108,7 +108,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec continue; } if (adapter->setInput(op, op_input_idx++, const_op)) { - MS_LOG(ERROR) << "fail to set params, index is " << op_input_idx; + MS_LOG(ERROR) << "Failed to set params, index is " << op_input_idx; return false; } graph_input_nodes->push_back(*const_op); @@ -178,7 +178,7 @@ void ToTensorPtr(const OpExecInfoPtr op_exec_info, std::vector* con MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]); auto ge_tensor_ptr = transform::TransformUtil::ConvertTensor(me_tensor_ptr, kOpFormat_NCHW); if (ge_tensor_ptr == nullptr) { - MS_LOG(EXCEPTION) << "convert inputs to GE tensor failed in op " << op_exec_info->op_name << "."; + MS_LOG(EXCEPTION) << "Convert inputs to GE tensor failed in op " << op_exec_info->op_name << "."; } // set inputs for operator to build single node graph inputs->push_back(ge_tensor_ptr); @@ -192,7 +192,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st for (auto& item : op_attrs) { if (!py::isinstance(item.first)) { - MS_LOG(ERROR) << "type error in py dict convert"; + MS_LOG(ERROR) << "Type error in py dict convert"; return PYNATIVE_OP_ATTRS_ERR; } std::string name = py::cast(item.first); @@ -203,7 +203,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st // build graph GeGraphPtr graph = std::make_shared(op_exec_info->op_name); if (BuildSingleOpGraph(op_exec_info, inputs, attrs, graph) == false) { - MS_LOG(ERROR) << "Fail to BuildSingleOpGraph"; + MS_LOG(ERROR) << "Failed to BuildSingleOpGraph"; return PYNATIVE_GRAPH_GE_BUILD_ERR; } @@ -211,7 +211,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st transform::Status ret = transform::DfGraphManager::GetInstance().AddGraph(SINGLE_OP_GRAPH, std::shared_ptr(graph)); if (ret != transform::SUCCESS) { - MS_LOG(ERROR) << "Fail to AddGraph into graph manager"; + MS_LOG(ERROR) << "Failed to AddGraph into graph manager"; return PYNATIVE_GRAPH_MANAGER_ERR; } @@ -289,7 +289,7 @@ py::object RunOpInGE(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat run_ret = graph_runner->RunGraph(run_options, ge_inputs, &ge_outputs); } if (run_ret != transform::Status::SUCCESS) { - MS_LOG(ERROR) << "GraphRunner Fails to Run Graph"; + MS_LOG(ERROR) << "GraphRunner fails to run graph"; *status = PYNATIVE_GRAPH_GE_RUN_ERR; return std::move(err_ret); }