提交 322ffef3 编写于 作者: L leonwanghui

Fix some typo errors in mindspore ir module

Signed-off-by: Nleonwanghui <leon.wanghui@huawei.com>
上级 7a367af9
...@@ -122,7 +122,7 @@ bool CastKernel(const TypeIdArgs &args, void *dst, const size_t data_size, const ...@@ -122,7 +122,7 @@ bool CastKernel(const TypeIdArgs &args, void *dst, const size_t data_size, const
TransDataSrc2Dst<uint16_t, int32_t>(args, dst, data_size); TransDataSrc2Dst<uint16_t, int32_t>(args, dst, data_size);
break; break;
default: default:
MS_LOG(ERROR) << "unsupported datatype trans"; MS_LOG(ERROR) << "Unsupported datatype trans";
return false; return false;
} }
return true; return true;
...@@ -132,7 +132,7 @@ size_t CubeSizeByType(const TypeId data_type) { ...@@ -132,7 +132,7 @@ size_t CubeSizeByType(const TypeId data_type) {
const size_t default_error = 0; const size_t default_error = 0;
auto dt_size = TypeIdSize(data_type); auto dt_size = TypeIdSize(data_type);
if (dt_size < 1) { if (dt_size < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return default_error; return default_error;
} else if (dt_size == 1) { } else if (dt_size == 1) {
return kCubeSize * 2; return kCubeSize * 2;
...@@ -146,12 +146,12 @@ size_t ShapeSize(const std::vector<size_t> &shape) { ...@@ -146,12 +146,12 @@ size_t ShapeSize(const std::vector<size_t> &shape) {
} }
size_t TypeIdSize(const TypeId data_type) { size_t TypeIdSize(const TypeId data_type) {
const size_t unsupport_type_error = 0; const size_t unsupported_type_error = 0;
auto iter = type_map.find(data_type); auto iter = type_map.find(data_type);
if (iter != type_map.end()) { if (iter != type_map.end()) {
return iter->second; return iter->second;
} }
return unsupport_type_error; return unsupported_type_error;
} }
std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) { std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) {
...@@ -174,7 +174,7 @@ std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) { ...@@ -174,7 +174,7 @@ std::vector<size_t> TransShapeTo4d(const std::vector<size_t> &shape) {
} }
break; break;
default: default:
MS_LOG(EXCEPTION) << "Unexpeted shape size = " << shape.size(); MS_LOG(EXCEPTION) << "Unexpected shape size = " << shape.size();
} }
return shape_4d; return shape_4d;
} }
...@@ -183,7 +183,7 @@ std::vector<size_t> TransShapeToDevice(const std::vector<size_t> &shape, const s ...@@ -183,7 +183,7 @@ std::vector<size_t> TransShapeToDevice(const std::vector<size_t> &shape, const s
std::vector<size_t> device_shape; std::vector<size_t> device_shape;
if (format == kOpFormat_FRAC_NZ) { if (format == kOpFormat_FRAC_NZ) {
if (shape.size() < 2) { if (shape.size() < 2) {
MS_EXCEPTION(NotSupportError) << "format " << format << " is not support shape " << shape.size(); MS_EXCEPTION(NotSupportError) << "Format " << format << " is not support shape " << shape.size();
} }
if (shape.size() > 2) { if (shape.size() > 2) {
(void)std::copy(shape.begin(), shape.end() - 2, std::back_inserter(device_shape)); (void)std::copy(shape.begin(), shape.end() - 2, std::back_inserter(device_shape));
...@@ -231,37 +231,37 @@ std::vector<size_t> TransShapeToDevice(const std::vector<size_t> &shape, const s ...@@ -231,37 +231,37 @@ std::vector<size_t> TransShapeToDevice(const std::vector<size_t> &shape, const s
} }
bool TransDataType(const TypeIdArgs &args, void *result) { bool TransDataType(const TypeIdArgs &args, void *result) {
MS_LOG(DEBUG) << "begin trans datatype from " << TypeIdLabel(args.host_data_type) << " to " MS_LOG(DEBUG) << "Begin trans datatype from " << TypeIdLabel(args.host_data_type) << " to "
<< TypeIdLabel(args.device_data_type); << TypeIdLabel(args.device_data_type);
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
std::pair<TypeId, TypeId> type_info(args.host_data_type, args.device_data_type); std::pair<TypeId, TypeId> type_info(args.host_data_type, args.device_data_type);
auto iter = mode_map.find(type_info); auto iter = mode_map.find(type_info);
if (iter == mode_map.end()) { if (iter == mode_map.end()) {
MS_LOG(ERROR) << "unsupported datatype trans. src_type :" << TypeIdLabel(args.host_data_type) MS_LOG(ERROR) << "Unsupported datatype trans. src_type :" << TypeIdLabel(args.host_data_type)
<< ", dst_type:" << TypeIdLabel(args.device_data_type); << ", dst_type:" << TypeIdLabel(args.device_data_type);
return false; return false;
} }
auto trans_mode = iter->second; auto trans_mode = iter->second;
auto type_size = TypeIdSize(args.device_data_type); auto type_size = TypeIdSize(args.device_data_type);
if (type_size < 1) { if (type_size < 1) {
MS_LOG(ERROR) << "invalid host data type."; MS_LOG(ERROR) << "Invalid host data type.";
return false; return false;
} }
if (args.host_shape_size < 1) { if (args.host_shape_size < 1) {
MS_LOG(ERROR) << "invalid host data size."; MS_LOG(ERROR) << "Invalid host data size.";
return false; return false;
} }
if (!CastKernel(args, result, args.host_shape_size, trans_mode)) { if (!CastKernel(args, result, args.host_shape_size, trans_mode)) {
MS_LOG(ERROR) << "failed to trans datatype.."; MS_LOG(ERROR) << "Failed to trans datatype..";
return false; return false;
} }
return true; return true;
} }
bool TransFormat(const FormatArgs &args, void *result) { bool TransFormat(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "start trans format."; MS_LOG(DEBUG) << "Start trans format.";
if (TypeIdSize(args.src_data_type) < 1) { if (TypeIdSize(args.src_data_type) < 1) {
MS_LOG(ERROR) << "invalid datatype.."; MS_LOG(ERROR) << "Invalid datatype..";
return false; return false;
} }
if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) &&
...@@ -276,9 +276,9 @@ bool TransFormat(const FormatArgs &args, void *result) { ...@@ -276,9 +276,9 @@ bool TransFormat(const FormatArgs &args, void *result) {
} }
bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "start trans format."; MS_LOG(DEBUG) << "Start trans format.";
if (TypeIdSize(args.src_data_type) < 1) { if (TypeIdSize(args.src_data_type) < 1) {
MS_LOG(ERROR) << "invalid datatype.."; MS_LOG(ERROR) << "Invalid datatype..";
return false; return false;
} }
if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) && if ((args.host_format == kOpFormat_NCHW || args.host_format == kOpFormat_ND) &&
...@@ -293,15 +293,15 @@ bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) { ...@@ -293,15 +293,15 @@ bool TransFormatFromDeviceToHost(const FormatArgs &args, void *result) {
} }
bool NchwToFracZ(const FormatArgs &args, void *result) { bool NchwToFracZ(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "trans format from nchw to frac_z"; MS_LOG(DEBUG) << "Trans format from nchw to frac_z";
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
if (args.host_shape.size() != kNchwDims) { if (args.host_shape.size() != kNchwDims) {
MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims;
return false; return false;
} }
size_t size = TypeIdSize(args.src_data_type); size_t size = TypeIdSize(args.src_data_type);
if (size < 1) { if (size < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return false; return false;
} }
auto n = args.host_shape[0]; auto n = args.host_shape[0];
...@@ -311,7 +311,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { ...@@ -311,7 +311,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) {
size_t c0 = CubeSizeByType(args.src_data_type); size_t c0 = CubeSizeByType(args.src_data_type);
if (c0 < 1) { if (c0 < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return false; return false;
} }
size_t c1 = Ceil(c, c0); size_t c1 = Ceil(c, c0);
...@@ -327,7 +327,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { ...@@ -327,7 +327,7 @@ bool NchwToFracZ(const FormatArgs &args, void *result) {
size_t dst_size = total_ele_cnt * size; size_t dst_size = total_ele_cnt * size;
if (dst_size != args.device_size) { if (dst_size != args.device_size) {
MS_LOG(ERROR) << "illegal total data size." MS_LOG(ERROR) << "Illegal total data size."
<< "dst size is :" << dst_size << "device size is :" << args.device_size; << "dst size is :" << dst_size << "device size is :" << args.device_size;
return false; return false;
} }
...@@ -369,20 +369,20 @@ bool NchwToFracZ(const FormatArgs &args, void *result) { ...@@ -369,20 +369,20 @@ bool NchwToFracZ(const FormatArgs &args, void *result) {
} }
bool FracZToNchw(const FormatArgs &args, void *result) { bool FracZToNchw(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "trans format from frac_z to nchw"; MS_LOG(DEBUG) << "Trans format from frac_z to nchw";
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
if (args.host_shape.size() != kNchwDims) { if (args.host_shape.size() != kNchwDims) {
MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims;
return false; return false;
} }
size_t size = TypeIdSize(args.src_data_type); size_t size = TypeIdSize(args.src_data_type);
if (size < 1) { if (size < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return false; return false;
} }
size_t total_size = ShapeSize(args.device_shape) * size; size_t total_size = ShapeSize(args.device_shape) * size;
if (total_size != args.device_size) { if (total_size != args.device_size) {
MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size;
return false; return false;
} }
...@@ -435,7 +435,7 @@ bool FracZToNchw(const FormatArgs &args, void *result) { ...@@ -435,7 +435,7 @@ bool FracZToNchw(const FormatArgs &args, void *result) {
bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> *hw_shape) { bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> *hw_shape) {
MS_EXCEPTION_IF_NULL(hw_shape); MS_EXCEPTION_IF_NULL(hw_shape);
if (host_shape.empty()) { if (host_shape.empty()) {
MS_LOG(ERROR) << "size of vector is 0."; MS_LOG(ERROR) << "Size of vector is 0.";
return false; return false;
} }
switch (host_shape.size()) { switch (host_shape.size()) {
...@@ -447,7 +447,7 @@ bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> * ...@@ -447,7 +447,7 @@ bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> *
default: default:
auto size = host_shape.size(); auto size = host_shape.size();
if (size < 2) { if (size < 2) {
MS_LOG(ERROR) << "illegal size."; MS_LOG(ERROR) << "Illegal size.";
return false; return false;
} }
size_t times = 1; size_t times = 1;
...@@ -462,26 +462,26 @@ bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> * ...@@ -462,26 +462,26 @@ bool TransShapeToNz(const std::vector<size_t> &host_shape, std::vector<size_t> *
} }
bool NchwToFracNz(const FormatArgs &args, void *result) { bool NchwToFracNz(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "trans format from nchw to frac_nz."; MS_LOG(DEBUG) << "Trans format from nchw to frac_nz.";
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
std::vector<size_t> hw_shape; std::vector<size_t> hw_shape;
if (!TransShapeToNz(args.host_shape, &hw_shape)) { if (!TransShapeToNz(args.host_shape, &hw_shape)) {
MS_LOG(ERROR) << "trans shape failed.."; MS_LOG(ERROR) << "Trans shape failed..";
return false; return false;
} }
if (hw_shape.size() < 3 || args.device_shape.size() < 4) { if (hw_shape.size() < 3 || args.device_shape.size() < 4) {
MS_LOG(ERROR) << "invalid shape size."; MS_LOG(ERROR) << "Invalid shape size.";
return false; return false;
} }
auto size = TypeIdSize(args.src_data_type); auto size = TypeIdSize(args.src_data_type);
if (size < 1) { if (size < 1) {
MS_LOG(ERROR) << "illegal dtype"; MS_LOG(ERROR) << "Illegal dtype";
return false; return false;
} }
auto dst_size = ShapeSize(args.device_shape) * size; auto dst_size = ShapeSize(args.device_shape) * size;
if (dst_size != args.device_size) { if (dst_size != args.device_size) {
MS_LOG(ERROR) << "illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; MS_LOG(ERROR) << "Illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size;
return false; return false;
} }
auto times = hw_shape.at(0); auto times = hw_shape.at(0);
...@@ -538,26 +538,26 @@ bool NchwToFracNz(const FormatArgs &args, void *result) { ...@@ -538,26 +538,26 @@ bool NchwToFracNz(const FormatArgs &args, void *result) {
} }
bool FracNzToNchw(const FormatArgs &args, void *result) { bool FracNzToNchw(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "trans format from frac_nz to nchw"; MS_LOG(DEBUG) << "Trans format from frac_nz to nchw";
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
std::vector<size_t> hw_shape; std::vector<size_t> hw_shape;
if (!TransShapeToNz(args.host_shape, &hw_shape)) { if (!TransShapeToNz(args.host_shape, &hw_shape)) {
MS_LOG(ERROR) << "trans shape failed.."; MS_LOG(ERROR) << "Trans shape failed..";
return false; return false;
} }
if (hw_shape.size() < 3 || args.device_shape.size() < 4) { if (hw_shape.size() < 3 || args.device_shape.size() < 4) {
MS_LOG(ERROR) << "invalid shape size."; MS_LOG(ERROR) << "Invalid shape size.";
return false; return false;
} }
auto size = TypeIdSize(args.src_data_type); auto size = TypeIdSize(args.src_data_type);
if (size < 1) { if (size < 1) {
MS_LOG(ERROR) << "illegal dtype"; MS_LOG(ERROR) << "Illegal dtype";
return false; return false;
} }
auto dst_size = ShapeSize(args.device_shape) * size; auto dst_size = ShapeSize(args.device_shape) * size;
if (dst_size != args.device_size) { if (dst_size != args.device_size) {
MS_LOG(ERROR) << "illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size; MS_LOG(ERROR) << "Illegal total data size, total_size:" << dst_size << ", device_size:" << args.device_size;
return false; return false;
} }
auto times = hw_shape.at(0); auto times = hw_shape.at(0);
...@@ -614,20 +614,20 @@ bool FracNzToNchw(const FormatArgs &args, void *result) { ...@@ -614,20 +614,20 @@ bool FracNzToNchw(const FormatArgs &args, void *result) {
} }
bool NchwToNc1hwc0(const FormatArgs &args, void *result) { bool NchwToNc1hwc0(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "trans format from nchw to Nc1h1wc0"; MS_LOG(DEBUG) << "Trans format from nchw to Nc1h1wc0";
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
if (args.host_shape.size() != kNchwDims) { if (args.host_shape.size() != kNchwDims) {
MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims;
return false; return false;
} }
size_t size = TypeIdSize(args.src_data_type); size_t size = TypeIdSize(args.src_data_type);
if (size < 1) { if (size < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return false; return false;
} }
auto total_size = ShapeSize(args.device_shape) * size; auto total_size = ShapeSize(args.device_shape) * size;
if (total_size != args.device_size) { if (total_size != args.device_size) {
MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size;
return false; return false;
} }
...@@ -637,7 +637,7 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) { ...@@ -637,7 +637,7 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) {
auto w = args.host_shape[3]; auto w = args.host_shape[3];
size_t c0 = CubeSizeByType(args.src_data_type); size_t c0 = CubeSizeByType(args.src_data_type);
if (c0 < 1) { if (c0 < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return false; return false;
} }
size_t c1 = Ceil(c, c0); size_t c1 = Ceil(c, c0);
...@@ -687,20 +687,20 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) { ...@@ -687,20 +687,20 @@ bool NchwToNc1hwc0(const FormatArgs &args, void *result) {
} }
bool Nc1hwc0ToNchw(const FormatArgs &args, void *result) { bool Nc1hwc0ToNchw(const FormatArgs &args, void *result) {
MS_LOG(DEBUG) << "trans format from nc1h1wc0 to nchw"; MS_LOG(DEBUG) << "Trans format from nc1h1wc0 to nchw";
MS_EXCEPTION_IF_NULL(result); MS_EXCEPTION_IF_NULL(result);
if (args.host_shape.size() != kNchwDims) { if (args.host_shape.size() != kNchwDims) {
MS_LOG(ERROR) << "invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims; MS_LOG(ERROR) << "Invalid host shape, host shape dims:" << args.host_shape.size() << ", expect dims:" << kNchwDims;
return false; return false;
} }
size_t size = TypeIdSize(args.src_data_type); size_t size = TypeIdSize(args.src_data_type);
if (size < 1) { if (size < 1) {
MS_LOG(ERROR) << "illegal dtype."; MS_LOG(ERROR) << "Illegal dtype.";
return false; return false;
} }
size_t total_size = ShapeSize(args.device_shape) * size; size_t total_size = ShapeSize(args.device_shape) * size;
if (total_size != args.device_size) { if (total_size != args.device_size) {
MS_LOG(ERROR) << "illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size; MS_LOG(ERROR) << "Illegal total data size, total_size:" << total_size << ", device_size:" << args.device_size;
return false; return false;
} }
......
...@@ -141,7 +141,7 @@ void DumpKernelInfo(const CNodePtr &node, const std::shared_ptr<SubGraphIRInfo> ...@@ -141,7 +141,7 @@ void DumpKernelInfo(const CNodePtr &node, const std::shared_ptr<SubGraphIRInfo>
void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMap<AnfNodePtr, int32_t> *para_map) { void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMap<AnfNodePtr, int32_t> *para_map) {
if (graph == nullptr) { if (graph == nullptr) {
MS_LOG(INFO) << "param graph is nullptr."; MS_LOG(INFO) << "Param graph is nullptr.";
return; return;
} }
std::vector<AnfNodePtr> parameters = graph->parameters(); std::vector<AnfNodePtr> parameters = graph->parameters();
...@@ -175,17 +175,17 @@ void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMa ...@@ -175,17 +175,17 @@ void DumpParams(const FuncGraphPtr &graph, std::ostringstream &buffer, OrderedMa
if (para_map != nullptr) { if (para_map != nullptr) {
(*para_map)[p] = para++; (*para_map)[p] = para++;
} }
MS_LOG(DEBUG) << "record param: " << p->ToString() << " graph belong : " << p->func_graph()->ToString(); MS_LOG(DEBUG) << "Record param: " << p->ToString() << " graph belong : " << p->func_graph()->ToString();
} }
} }
void DumpOperator(const AnfNodePtr &op, const std::shared_ptr<SubGraphIRInfo> &gsub) { void DumpOperator(const AnfNodePtr &op, const std::shared_ptr<SubGraphIRInfo> &gsub) {
if (op == nullptr) { if (op == nullptr) {
MS_LOG(INFO) << "param op is nullptr"; MS_LOG(INFO) << "Param op is nullptr";
return; return;
} }
if (gsub == nullptr) { if (gsub == nullptr) {
MS_LOG(INFO) << "param gsub is nullptr"; MS_LOG(INFO) << "Param gsub is nullptr";
return; return;
} }
...@@ -338,7 +338,7 @@ void DumpCNode(const CNodePtr &nd, const FuncGraphPtr &sub_graph, OrderedMap<Anf ...@@ -338,7 +338,7 @@ void DumpCNode(const CNodePtr &nd, const FuncGraphPtr &sub_graph, OrderedMap<Anf
} }
if (nd->inputs().empty()) { if (nd->inputs().empty()) {
MS_LOG(EXCEPTION) << "input of apply node is empty"; MS_LOG(EXCEPTION) << "Input of apply node is empty";
} }
// print operator // print operator
...@@ -376,7 +376,7 @@ void DumpIRInSubgraph(const std::vector<AnfNodePtr> &nodes, OrderedMap<AnfNodePt ...@@ -376,7 +376,7 @@ void DumpIRInSubgraph(const std::vector<AnfNodePtr> &nodes, OrderedMap<AnfNodePt
MS_EXCEPTION_IF_NULL(nd); MS_EXCEPTION_IF_NULL(nd);
FuncGraphPtr sub_graph = nd->func_graph(); FuncGraphPtr sub_graph = nd->func_graph();
if (sub_graph == nullptr) { if (sub_graph == nullptr) {
MS_LOG(DEBUG) << "node[" << nd->ToString() << "] belongs to no graph!"; MS_LOG(DEBUG) << "Node[" << nd->ToString() << "] belongs to no graph!";
continue; continue;
} }
std::shared_ptr<SubGraphIRInfo> gsub = (*sub_graphs)[sub_graph]; std::shared_ptr<SubGraphIRInfo> gsub = (*sub_graphs)[sub_graph];
...@@ -430,12 +430,12 @@ void DumpIR(const std::string &filename, const FuncGraphPtr &graph, bool dump_fu ...@@ -430,12 +430,12 @@ void DumpIR(const std::string &filename, const FuncGraphPtr &graph, bool dump_fu
return; return;
} }
if (filename.size() > PATH_MAX) { if (filename.size() > PATH_MAX) {
MS_LOG(ERROR) << "file path " << filename << " is too long."; MS_LOG(ERROR) << "File path " << filename << " is too long.";
return; return;
} }
char real_path[PATH_MAX] = {0}; char real_path[PATH_MAX] = {0};
if (nullptr == realpath(filename.c_str(), real_path)) { if (nullptr == realpath(filename.c_str(), real_path)) {
MS_LOG(DEBUG) << "dir " << filename << " does not exit."; MS_LOG(DEBUG) << "Dir " << filename << " does not exit.";
} }
OrderedMap<AnfNodePtr, int32_t> para_map; OrderedMap<AnfNodePtr, int32_t> para_map;
......
...@@ -49,7 +49,7 @@ std::string GetMsIrPath(void) { ...@@ -49,7 +49,7 @@ std::string GetMsIrPath(void) {
path = path_ptr; path = path_ptr;
char real_path[PATH_MAX] = {0}; char real_path[PATH_MAX] = {0};
if (path.size() > PATH_MAX || nullptr == realpath(path.c_str(), real_path)) { if (path.size() > PATH_MAX || nullptr == realpath(path.c_str(), real_path)) {
MS_LOG(EXCEPTION) << "MS IR Path error, " << path_ptr; MS_LOG(EXCEPTION) << "MS IR path error, " << path_ptr;
} }
path = real_path; path = real_path;
} }
...@@ -144,8 +144,8 @@ std::string AnfExporter::GetValueNodeText(const FuncGraphPtr& fg, const ValueNod ...@@ -144,8 +144,8 @@ std::string AnfExporter::GetValueNodeText(const FuncGraphPtr& fg, const ValueNod
} }
std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGraphPtr& mt_func_graph) { std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGraphPtr& mt_func_graph) {
auto py_funs = mt_func_graph->GetPyFunctions(); auto py_funcs = mt_func_graph->GetPyFunctions();
if (py_funs.empty()) { if (py_funcs.empty()) {
return ""; return "";
} }
...@@ -153,7 +153,7 @@ std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGrap ...@@ -153,7 +153,7 @@ std::string AnfExporter::GetMultitypeFuncGraphText(const prim::MultitypeFuncGrap
oss << "{"; oss << "{";
bool is_first = true; bool is_first = true;
for (const auto& py_func : py_funs) { for (const auto& py_func : py_funcs) {
if (is_first) { if (is_first) {
is_first = false; is_first = false;
} else { } else {
...@@ -626,7 +626,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const FuncGraphPt ...@@ -626,7 +626,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const FuncGraphPt
ofs << "\n\n"; ofs << "\n\n";
(void)func_graph_set.erase(fg); (void)func_graph_set.erase(fg);
} }
ofs << "# num of total funcgraphs: " << exported.size(); ofs << "# num of total function graphs: " << exported.size();
ofs.close(); ofs.close();
} }
...@@ -651,7 +651,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const std::vector ...@@ -651,7 +651,7 @@ void AnfExporter::ExportFuncGraph(const std::string& filename, const std::vector
ofs << "\n\n"; ofs << "\n\n";
} }
ofs << "# num of total funcgraphs: " << graphs.size(); ofs << "# num of total function graphs: " << graphs.size();
ofs.close(); ofs.close();
} }
...@@ -763,7 +763,7 @@ class Lexer { ...@@ -763,7 +763,7 @@ class Lexer {
fin.close(); fin.close();
} }
} catch (const std::exception& e) { } catch (const std::exception& e) {
MS_LOG(ERROR) << "exception when closing file"; MS_LOG(ERROR) << "Exception when closing file";
} catch (...) { } catch (...) {
std::string exName(abi::__cxa_current_exception_type()->name()); std::string exName(abi::__cxa_current_exception_type()->name());
MS_LOG(ERROR) << "Error occurred when closing file. Exception name: " << exName; MS_LOG(ERROR) << "Error occurred when closing file. Exception name: " << exName;
...@@ -802,7 +802,7 @@ class Lexer { ...@@ -802,7 +802,7 @@ class Lexer {
Token token = GetNextTokenInner(); Token token = GetNextTokenInner();
const char* str = token_text[token]; const char* str = token_text[token];
std::string text = (str == nullptr ? GetTokenText() : str); std::string text = (str == nullptr ? GetTokenText() : str);
MS_LOG(DEBUG) << "------parse token] " << text; MS_LOG(DEBUG) << "------Parse token] " << text;
return token; return token;
} }
...@@ -1642,7 +1642,7 @@ class IrParser { ...@@ -1642,7 +1642,7 @@ class IrParser {
MS_LOG(EXCEPTION) << "Expect @file at line " << lexer_.GetLineNo(); MS_LOG(EXCEPTION) << "Expect @file at line " << lexer_.GetLineNo();
} }
// load prameter default value from serialized file // load parameter default value from serialized file
py::object default_obj = LoadObject(lexer_.GetTokenText()); py::object default_obj = LoadObject(lexer_.GetTokenText());
param->set_default_param(default_obj); param->set_default_param(default_obj);
...@@ -1950,7 +1950,7 @@ class IrParser { ...@@ -1950,7 +1950,7 @@ class IrParser {
return TOK_ERROR; return TOK_ERROR;
} }
// restore python funciton of PrimitivePy from serialized file // restore python function of PrimitivePy from serialized file
py::object py_obj = LoadObject(lexer_.GetTokenText()); py::object py_obj = LoadObject(lexer_.GetTokenText());
PrimitivePyPtr ptr = nullptr; PrimitivePyPtr ptr = nullptr;
if (py::hasattr(py_obj, "__setattr_flag__") && py::hasattr(py_obj, "_clone")) { if (py::hasattr(py_obj, "__setattr_flag__") && py::hasattr(py_obj, "_clone")) {
...@@ -1958,7 +1958,7 @@ class IrParser { ...@@ -1958,7 +1958,7 @@ class IrParser {
py::object new_obj = clone_fn(); py::object new_obj = clone_fn();
ptr = new_obj.cast<PrimitivePyPtr>(); ptr = new_obj.cast<PrimitivePyPtr>();
if (ptr == nullptr) { if (ptr == nullptr) {
MS_LOG(EXCEPTION) << "cast to type 'PrimitivePyPtr' error"; MS_LOG(EXCEPTION) << "Cast to type 'PrimitivePyPtr' error";
} }
} else { } else {
ptr = std::make_shared<PrimitivePy>(id.substr(strlen("PrimitivePy::")), py_obj); ptr = std::make_shared<PrimitivePy>(id.substr(strlen("PrimitivePy::")), py_obj);
...@@ -2221,15 +2221,15 @@ class IrParser { ...@@ -2221,15 +2221,15 @@ class IrParser {
}; };
std::vector<FuncGraphPtr> ImportIR(const std::string& filename) { std::vector<FuncGraphPtr> ImportIR(const std::string& filename) {
IrParser paser(filename.c_str()); IrParser parser(filename.c_str());
paser.ParseFile(); parser.ParseFile();
return paser.GetFuncGraphs(); return parser.GetFuncGraphs();
} }
#ifdef ENABLE_DUMP_IR #ifdef ENABLE_DUMP_IR
void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) {
if (func_graph == nullptr) { if (func_graph == nullptr) {
MS_LOG(ERROR) << "func graph is nullptr"; MS_LOG(ERROR) << "Func graph is nullptr";
return; return;
} }
auto ms_context = MsContext::GetInstance(); auto ms_context = MsContext::GetInstance();
...@@ -2243,16 +2243,16 @@ void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) { ...@@ -2243,16 +2243,16 @@ void DumpIRProto(const FuncGraphPtr& func_graph, const std::string& suffix) {
} }
std::string file_path = save_graphs_path + "/" + "ms_output_" + suffix + ".pb"; std::string file_path = save_graphs_path + "/" + "ms_output_" + suffix + ".pb";
if (file_path.size() > PATH_MAX) { if (file_path.size() > PATH_MAX) {
MS_LOG(ERROR) << "file path " << file_path << " is too long."; MS_LOG(ERROR) << "File path " << file_path << " is too long.";
return; return;
} }
char real_path[PATH_MAX] = {0}; char real_path[PATH_MAX] = {0};
if (nullptr == realpath(file_path.c_str(), real_path)) { if (nullptr == realpath(file_path.c_str(), real_path)) {
MS_LOG(DEBUG) << "dir " << file_path << " does not exit."; MS_LOG(DEBUG) << "Dir " << file_path << " does not exit.";
} else { } else {
std::string path_string = real_path; std::string path_string = real_path;
if (chmod(common::SafeCStr(path_string), S_IRUSR | S_IWUSR) == -1) { if (chmod(common::SafeCStr(path_string), S_IRUSR | S_IWUSR) == -1) {
MS_LOG(ERROR) << "modify file:" << real_path << " to rw fail."; MS_LOG(ERROR) << "Modify file:" << real_path << " to rw fail.";
return; return;
} }
} }
......
...@@ -362,7 +362,7 @@ Digraph::~Digraph() { ...@@ -362,7 +362,7 @@ Digraph::~Digraph() {
fout_.close(); fout_.close();
} }
} catch (const std::exception& e) { } catch (const std::exception& e) {
MS_LOG(ERROR) << "exception when closing file " << filename_; MS_LOG(ERROR) << "Exception when closing file " << filename_;
} }
} }
......
...@@ -208,7 +208,7 @@ void ProtoExporter::SetValueToProto(const ValuePtr& val, irpb::ValueProto* value ...@@ -208,7 +208,7 @@ void ProtoExporter::SetValueToProto(const ValuePtr& val, irpb::ValueProto* value
TypePtr elem_type = dyn_cast<TensorType>(val)->element(); TypePtr elem_type = dyn_cast<TensorType>(val)->element();
type_proto->mutable_tensor_type()->set_elem_type(GetNumberDataType(elem_type)); type_proto->mutable_tensor_type()->set_elem_type(GetNumberDataType(elem_type));
} else { } else {
MS_LOG(WARNING) << "Not supported type " << val->type_name(); MS_LOG(WARNING) << "Unsupported type " << val->type_name();
} }
} }
......
...@@ -101,7 +101,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) { ...@@ -101,7 +101,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) {
auto kernels = dumpSettings.at("kernels"); auto kernels = dumpSettings.at("kernels");
if (!(enable.is_boolean() && trans_flag.is_boolean() && mode.is_number() && path.is_string() && if (!(enable.is_boolean() && trans_flag.is_boolean() && mode.is_number() && path.is_string() &&
net_name.is_string() && iteration.is_number() && kernels.is_array())) { net_name.is_string() && iteration.is_number() && kernels.is_array())) {
MS_LOG(ERROR) << "element's type in Dump config json is invalid."; MS_LOG(ERROR) << "Element's type in Dump config json is invalid.";
dump_enable_ = false; dump_enable_ = false;
return false; return false;
} }
...@@ -121,7 +121,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) { ...@@ -121,7 +121,7 @@ bool Dump::IsConfigValid(const nlohmann::json& dumpSettings) {
bool Dump::SetDumpConfFromJsonFile() { bool Dump::SetDumpConfFromJsonFile() {
const char* config_path_str = std::getenv("MINDSPORE_CONFIG_PATH"); const char* config_path_str = std::getenv("MINDSPORE_CONFIG_PATH");
if (config_path_str != nullptr) { if (config_path_str != nullptr) {
MS_LOG(INFO) << "getenv MINDSPORE_CONFIG_PATH :" << config_path_str; MS_LOG(INFO) << "Getenv MINDSPORE_CONFIG_PATH :" << config_path_str;
} else { } else {
MS_LOG(INFO) << "No need E2E Dump. please export MINDSPORE_CONFIG_PATH eg: MINDSPORE_CONFIG_PATH=/etc"; MS_LOG(INFO) << "No need E2E Dump. please export MINDSPORE_CONFIG_PATH eg: MINDSPORE_CONFIG_PATH=/etc";
dump_enable_ = false; dump_enable_ = false;
...@@ -132,7 +132,7 @@ bool Dump::SetDumpConfFromJsonFile() { ...@@ -132,7 +132,7 @@ bool Dump::SetDumpConfFromJsonFile() {
auto id = context_ptr->device_id(); auto id = context_ptr->device_id();
char real_path[PATH_MAX] = {0}; char real_path[PATH_MAX] = {0};
if (nullptr == realpath(config_path_str, real_path)) { if (nullptr == realpath(config_path_str, real_path)) {
MS_LOG(ERROR) << "env e2e dump path error, " << config_path_str; MS_LOG(ERROR) << "Env e2e dump path error, " << config_path_str;
dump_enable_ = false; dump_enable_ = false;
return false; return false;
} }
...@@ -150,20 +150,20 @@ bool Dump::SetDumpConfFromJsonFile() { ...@@ -150,20 +150,20 @@ bool Dump::SetDumpConfFromJsonFile() {
bool Dump::DumpToFile(const std::string& filename, const void* data, size_t len) { bool Dump::DumpToFile(const std::string& filename, const void* data, size_t len) {
if (filename.empty() || data == nullptr || len == 0) { if (filename.empty() || data == nullptr || len == 0) {
MS_LOG(ERROR) << "incorrect parameter."; MS_LOG(ERROR) << "Incorrect parameter.";
return false; return false;
} }
std::string realpath; std::string realpath;
bool ret = GetRealPath(filename, &realpath); bool ret = GetRealPath(filename, &realpath);
if (!ret) { if (!ret) {
MS_LOG(ERROR) << "get real path failed."; MS_LOG(ERROR) << "Get real path failed.";
return false; return false;
} }
std::ofstream fd; std::ofstream fd;
fd.open(realpath, std::ios::binary | std::ios::out); fd.open(realpath, std::ios::binary | std::ios::out);
if (!fd.is_open()) { if (!fd.is_open()) {
MS_LOG(ERROR) << "open file " << realpath << " fail."; MS_LOG(ERROR) << "Open file " << realpath << " fail.";
return false; return false;
} }
(void)fd.write(reinterpret_cast<const char*>(data), SizeToLong(len)); (void)fd.write(reinterpret_cast<const char*>(data), SizeToLong(len));
...@@ -182,7 +182,7 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) { ...@@ -182,7 +182,7 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) {
if (path_split_pos != std::string::npos) { if (path_split_pos != std::string::npos) {
std::string prefix_path = inpath.substr(0, path_split_pos); std::string prefix_path = inpath.substr(0, path_split_pos);
if (prefix_path.length() >= PATH_MAX) { if (prefix_path.length() >= PATH_MAX) {
MS_LOG(ERROR) << "prefix path is too longer!"; MS_LOG(ERROR) << "Prefix path is too longer!";
return false; return false;
} }
std::string last_path = inpath.substr(path_split_pos, inpath.length() - path_split_pos); std::string last_path = inpath.substr(path_split_pos, inpath.length() - path_split_pos);
...@@ -201,11 +201,11 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) { ...@@ -201,11 +201,11 @@ bool Dump::GetRealPath(const std::string& inpath, std::string* outpath) {
if (path_split_pos == std::string::npos) { if (path_split_pos == std::string::npos) {
if (inpath.length() >= PATH_MAX) { if (inpath.length() >= PATH_MAX) {
MS_LOG(ERROR) << "prefix path is too longer!"; MS_LOG(ERROR) << "Prefix path is too longer!";
return false; return false;
} }
if (nullptr == realpath(inpath.c_str(), real_path)) { if (nullptr == realpath(inpath.c_str(), real_path)) {
MS_LOG(ERROR) << "file " << inpath << " does not exit, it will be created."; MS_LOG(ERROR) << "File " << inpath << " does not exit, it will be created.";
} }
*outpath = std::string(real_path); *outpath = std::string(real_path);
} }
...@@ -218,7 +218,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { ...@@ -218,7 +218,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) {
MS_EXCEPTION_IF_NULL(fs); MS_EXCEPTION_IF_NULL(fs);
char temp_path[PATH_MAX] = {0}; char temp_path[PATH_MAX] = {0};
if (path.length() > PATH_MAX) { if (path.length() > PATH_MAX) {
MS_LOG(ERROR) << "path lens is max than " << PATH_MAX; MS_LOG(ERROR) << "Path lens is max than " << PATH_MAX;
return false; return false;
} }
for (uint32_t i = 0; i < path.length(); i++) { for (uint32_t i = 0; i < path.length(); i++) {
...@@ -229,7 +229,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { ...@@ -229,7 +229,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) {
temp_path[i] = '\0'; temp_path[i] = '\0';
std::string path_handle(temp_path); std::string path_handle(temp_path);
if (!fs->FileExist(temp_path)) { if (!fs->FileExist(temp_path)) {
MS_LOG(INFO) << "dir " << path_handle << " does not exit, creating..."; MS_LOG(INFO) << "Dir " << path_handle << " does not exit, creating...";
if (!fs->CreateDir(temp_path)) { if (!fs->CreateDir(temp_path)) {
MS_LOG(ERROR) << "Create " << path_handle << " dir error"; MS_LOG(ERROR) << "Create " << path_handle << " dir error";
return false; return false;
...@@ -241,7 +241,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) { ...@@ -241,7 +241,7 @@ bool Dump::CreateNotExistDirs(const std::string& path) {
} }
if (!fs->FileExist(path)) { if (!fs->FileExist(path)) {
MS_LOG(INFO) << "dir " << path << " does not exit, creating..."; MS_LOG(INFO) << "Dir " << path << " does not exit, creating...";
if (!fs->CreateDir(path)) { if (!fs->CreateDir(path)) {
MS_LOG(ERROR) << "Create " << path << " dir error"; MS_LOG(ERROR) << "Create " << path << " dir error";
return false; return false;
......
...@@ -193,7 +193,7 @@ void TraceManager::DebugTrace(const TraceInfoPtr& trace_info) { ...@@ -193,7 +193,7 @@ void TraceManager::DebugTrace(const TraceInfoPtr& trace_info) {
} }
TraceContextPtr context = std::make_shared<TraceContext>(trace_info); TraceContextPtr context = std::make_shared<TraceContext>(trace_info);
if (trace_info->debug_info() == nullptr) { if (trace_info->debug_info() == nullptr) {
MS_LOG(EXCEPTION) << "trace debug info is null"; MS_LOG(EXCEPTION) << "Trace debug info is null";
} }
TraceManager::trace_context_stack_.push(context); TraceManager::trace_context_stack_.push(context);
} }
...@@ -205,7 +205,7 @@ void TraceManager::DebugTrace(const DebugInfoPtr& debug_info, const TraceInfoPtr ...@@ -205,7 +205,7 @@ void TraceManager::DebugTrace(const DebugInfoPtr& debug_info, const TraceInfoPtr
auto cloned_info = trace_info->clone(); auto cloned_info = trace_info->clone();
cloned_info->set_debug_info(debug_info); cloned_info->set_debug_info(debug_info);
if (cloned_info->debug_info() == nullptr) { if (cloned_info->debug_info() == nullptr) {
MS_LOG(EXCEPTION) << "trace debug info is null with cloned trace"; MS_LOG(EXCEPTION) << "Trace debug info is null with cloned trace";
} }
TraceContextPtr context = std::make_shared<TraceContext>(cloned_info); TraceContextPtr context = std::make_shared<TraceContext>(cloned_info);
TraceManager::trace_context_stack_.push(context); TraceManager::trace_context_stack_.push(context);
......
...@@ -89,7 +89,7 @@ std::string GetDebugInfo(const DebugInfoPtr& info, SourceLineTip tip) { ...@@ -89,7 +89,7 @@ std::string GetDebugInfo(const DebugInfoPtr& info, SourceLineTip tip) {
return ""; return "";
} }
// a trace info identifys a node transform, so we can trace the node transform through // a trace info identifies a node transform, so we can trace the node transform through
// a link of trace info and debug info // a link of trace info and debug info
std::string GetInfoWithAction(const std::vector<DebugInfoPtr>& info_vec, SourceLineTip tip) { std::string GetInfoWithAction(const std::vector<DebugInfoPtr>& info_vec, SourceLineTip tip) {
if (info_vec.size() < 1) { if (info_vec.size() < 1) {
...@@ -173,7 +173,7 @@ void DumpInferStack(std::ostringstream& oss) { ...@@ -173,7 +173,7 @@ void DumpInferStack(std::ostringstream& oss) {
} }
auto graph_context = graph_infer->graph_context(); auto graph_context = graph_infer->graph_context();
if (graph_context == nullptr) { if (graph_context == nullptr) {
MS_LOG(INFO) << "null context continue"; MS_LOG(INFO) << "Null context continue";
continue; continue;
} }
auto graph = graph_context->func_graph(); auto graph = graph_context->func_graph();
...@@ -264,7 +264,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename, ...@@ -264,7 +264,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename,
param_index = 1; param_index = 1;
auto tagged_func_graphs = CalcTaggedFuncGraphs(); auto tagged_func_graphs = CalcTaggedFuncGraphs();
// first output grapn on the analysis stack // first output graph on the analysis stack
for (const auto& node_cfg : node_cfgs) { for (const auto& node_cfg : node_cfgs) {
auto fg = node_cfg->context()->func_graph(); auto fg = node_cfg->context()->func_graph();
// the graph is already output, skip it // the graph is already output, skip it
...@@ -291,7 +291,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename, ...@@ -291,7 +291,7 @@ void AnalyzedFuncGraphExporter::ExportFuncGraph(const std::string& filename,
ofs << "\n\n"; ofs << "\n\n";
(void)func_graph_set.erase(fg); (void)func_graph_set.erase(fg);
} }
ofs << "# num of total funcgraphs: " << exported.size(); ofs << "# num of total function graphs: " << exported.size();
ofs.close(); ofs.close();
} }
...@@ -332,7 +332,7 @@ void GetInferStackInfo(std::ostringstream& oss) { ...@@ -332,7 +332,7 @@ void GetInferStackInfo(std::ostringstream& oss) {
MS_LOG(INFO) << "Get graph analysis information *end*"; MS_LOG(INFO) << "Get graph analysis information *end*";
} }
// trace the graph evaluator statck // trace the graph evaluator stack
static std::stack<std::pair<abstract::EvaluatorPtr, abstract::AnfNodeConfigPtr>> graph_infer_stack; static std::stack<std::pair<abstract::EvaluatorPtr, abstract::AnfNodeConfigPtr>> graph_infer_stack;
// trace the cnode infer debug info // trace the cnode infer debug info
static std::vector<abstract::AnfNodeConfigPtr> cnode_debug_stack{}; static std::vector<abstract::AnfNodeConfigPtr> cnode_debug_stack{};
......
...@@ -36,6 +36,6 @@ std::string TraceInfo::GetActionBetweenNode(const DebugInfoPtr& info) { ...@@ -36,6 +36,6 @@ std::string TraceInfo::GetActionBetweenNode(const DebugInfoPtr& info) {
} else if (debug_info()->trace_info() != nullptr) { } else if (debug_info()->trace_info() != nullptr) {
return act_name + debug_info()->trace_info()->GetActionBetweenNode(info); return act_name + debug_info()->trace_info()->GetActionBetweenNode(info);
} }
return "not in the traced info"; return "Not in the traced info";
} }
} // namespace mindspore } // namespace mindspore
...@@ -83,7 +83,7 @@ class AnfVisitor; ...@@ -83,7 +83,7 @@ class AnfVisitor;
// Methods: // Methods:
// func_graph: return FuncGraph that this AnfNode belongs to. // func_graph: return FuncGraph that this AnfNode belongs to.
// scope: return the scope namespace of this AnfNode. Set it using set_scope. // scope: return the scope namespace of this AnfNode. Set it using set_scope.
// abstract: return the cached inferred abstract value. It cantains type, shape // abstract: return the cached inferred abstract value. It contains type, shape
// value. Set New cache using set_abstract. // value. Set New cache using set_abstract.
// intermediate_abstract: return the cached inferring abstract value. // intermediate_abstract: return the cached inferring abstract value.
// Type/Shape: return the related info of this AnfNode. When this AnfNode is an // Type/Shape: return the related info of this AnfNode. When this AnfNode is an
...@@ -284,7 +284,7 @@ class Parameter : public ANode { ...@@ -284,7 +284,7 @@ class Parameter : public ANode {
}; };
using ParameterPtr = std::shared_ptr<Parameter>; using ParameterPtr = std::shared_ptr<Parameter>;
// Value is used to represent the atomic expression metioned in BNF. // Value is used to represent the atomic expression mentioned in BNF.
// It mainly be stored in ValueNode. Value and ValueNode is related definition. // It mainly be stored in ValueNode. Value and ValueNode is related definition.
class Value : public Base { class Value : public Base {
public: public:
...@@ -313,7 +313,7 @@ using ValuePtr = std::shared_ptr<Value>; ...@@ -313,7 +313,7 @@ using ValuePtr = std::shared_ptr<Value>;
using ValuePtrList = std::vector<ValuePtr>; using ValuePtrList = std::vector<ValuePtr>;
// ValueNode is used to hold value. Unlike CNode and Parameter, ValueNode // ValueNode is used to hold value. Unlike CNode and Parameter, ValueNode
// do not belong to any particular function graph. // does not belong to any particular function graph.
class ValueNode : public ANode { class ValueNode : public ANode {
public: public:
explicit ValueNode(const ValuePtr &value) : value_(value) {} explicit ValueNode(const ValuePtr &value) : value_(value) {}
......
...@@ -34,19 +34,19 @@ bool Number::operator==(const Type& other) const { ...@@ -34,19 +34,19 @@ bool Number::operator==(const Type& other) const {
Int::Int(const int nbits) : Number(IntBitsToTypeId(nbits), nbits, false) { Int::Int(const int nbits) : Number(IntBitsToTypeId(nbits), nbits, false) {
if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) {
MS_LOG(EXCEPTION) << "wrong number of bits."; MS_LOG(EXCEPTION) << "Wrong number of bits.";
} }
} }
UInt::UInt(const int nbits) : Number(UIntBitsToTypeId(nbits), nbits, false) { UInt::UInt(const int nbits) : Number(UIntBitsToTypeId(nbits), nbits, false) {
if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) { if (nbits != 8 && nbits != 16 && nbits != 32 && nbits != 64) {
MS_LOG(EXCEPTION) << "wrong number of bits."; MS_LOG(EXCEPTION) << "Wrong number of bits.";
} }
} }
Float::Float(const int nbits) : Number(FloatBitsToTypeId(nbits), nbits, false) { Float::Float(const int nbits) : Number(FloatBitsToTypeId(nbits), nbits, false) {
if (nbits != 16 && nbits != 32 && nbits != 64) { if (nbits != 16 && nbits != 32 && nbits != 64) {
MS_LOG(EXCEPTION) << "wrong number of bits."; MS_LOG(EXCEPTION) << "Wrong number of bits.";
} }
} }
......
...@@ -37,7 +37,7 @@ TypeId IntBitsToTypeId(const int nbits) { ...@@ -37,7 +37,7 @@ TypeId IntBitsToTypeId(const int nbits) {
case 64: case 64:
return kNumberTypeInt64; return kNumberTypeInt64;
default: default:
MS_LOG(EXCEPTION) << "wrong number of bits."; MS_LOG(EXCEPTION) << "Wrong number of bits.";
} }
} }
...@@ -52,7 +52,7 @@ TypeId UIntBitsToTypeId(const int nbits) { ...@@ -52,7 +52,7 @@ TypeId UIntBitsToTypeId(const int nbits) {
case 64: case 64:
return kNumberTypeUInt64; return kNumberTypeUInt64;
default: default:
MS_LOG(EXCEPTION) << "wrong number of bits."; MS_LOG(EXCEPTION) << "Wrong number of bits.";
} }
} }
...@@ -65,7 +65,7 @@ TypeId FloatBitsToTypeId(const int nbits) { ...@@ -65,7 +65,7 @@ TypeId FloatBitsToTypeId(const int nbits) {
case 64: case 64:
return kNumberTypeFloat64; return kNumberTypeFloat64;
default: default:
MS_LOG(EXCEPTION) << "wrong number of bits."; MS_LOG(EXCEPTION) << "Wrong number of bits.";
} }
} }
......
...@@ -174,7 +174,7 @@ class FuncGraph : public FuncGraphBase { ...@@ -174,7 +174,7 @@ class FuncGraph : public FuncGraphBase {
GraphDebugInfoPtr debug_info(); GraphDebugInfoPtr debug_info();
void set_debug_info(const GraphDebugInfoPtr &info) { void set_debug_info(const GraphDebugInfoPtr &info) {
if (info == nullptr) { if (info == nullptr) {
MS_LOG(EXCEPTION) << "graph set null debug info"; MS_LOG(EXCEPTION) << "Graph set null debug info";
} }
this->debug_info_ = info; this->debug_info_ = info;
} }
......
...@@ -817,7 +817,7 @@ void FuncGraphChildDirect::OnMoveAllCNode(FuncGraphPtr src, FuncGraphPtr dst) { ...@@ -817,7 +817,7 @@ void FuncGraphChildDirect::OnMoveAllCNode(FuncGraphPtr src, FuncGraphPtr dst) {
void FuncGraphParentsDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) { void FuncGraphParentsDirectCollector::OnModEdge(AnfNodePtr node, int, AnfNodePtr inp, EdgeProcessDirection direction) {
MS_EXCEPTION_IF_NULL(node); MS_EXCEPTION_IF_NULL(node);
FuncGraphPtr fg1 = node->func_graph(); FuncGraphPtr fg1 = node->func_graph();
// possible chirld parent // possible child parent
if (IsValueNode<FuncGraph>(inp)) { if (IsValueNode<FuncGraph>(inp)) {
FuncGraphPtr fg2 = GetValueNode<FuncGraphPtr>(inp); FuncGraphPtr fg2 = GetValueNode<FuncGraphPtr>(inp);
if (Mod(fg1, ParentProxy(fg2), direction)) { if (Mod(fg1, ParentProxy(fg2), direction)) {
...@@ -1181,7 +1181,7 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr& fg, const FuncGraphSetPt ...@@ -1181,7 +1181,7 @@ bool FuncGraphJTotalComputer::SeekJ(const FuncGraphPtr& fg, const FuncGraphSetPt
} }
path->add(fg); path->add(fg);
// checkg if func graphs used contains J(func_graph); // check if func graphs used contains J(func_graph);
auto& used = this->manager_->func_graphs_used(); auto& used = this->manager_->func_graphs_used();
for (auto& item : used[fg]) { for (auto& item : used[fg]) {
auto used_g = item.first; auto used_g = item.first;
......
...@@ -650,7 +650,7 @@ class FuncGraphTransaction { ...@@ -650,7 +650,7 @@ class FuncGraphTransaction {
explicit FuncGraphTransaction(FuncGraphManager* manager) : manager_(manager), changes_() { explicit FuncGraphTransaction(FuncGraphManager* manager) : manager_(manager), changes_() {
MS_EXCEPTION_IF_NULL(manager_); MS_EXCEPTION_IF_NULL(manager_);
if (!manager_->IsManaged()) { if (!manager_->IsManaged()) {
MS_LOG(DEBUG) << "the manager is not managed yet"; MS_LOG(DEBUG) << "The manager is not managed yet";
} }
} }
......
...@@ -148,7 +148,7 @@ class MetaTensor : public Value { ...@@ -148,7 +148,7 @@ class MetaTensor : public Value {
// //
// The constructed MetaTensor object has the same type and shape with meta_tensor. // The constructed MetaTensor object has the same type and shape with meta_tensor.
// //
// param meta_tensor An exisiting MetaTensor object. // param meta_tensor An existing MetaTensor object.
virtual MetaTensor& operator=(const MetaTensor& meta_tensor); virtual MetaTensor& operator=(const MetaTensor& meta_tensor);
// brief Compares two MetaTensor objects. // brief Compares two MetaTensor objects.
...@@ -166,7 +166,7 @@ class MetaTensor : public Value { ...@@ -166,7 +166,7 @@ class MetaTensor : public Value {
TypeId data_type() const { return data_type_; } TypeId data_type() const { return data_type_; }
std::string ToString() const override; std::string ToString() const override;
std::string DumpText() const override; std::string DumpText() const override;
// bried Sets the data type of a tensor in its MetaTensor. // brief Sets the data type of a tensor in its MetaTensor.
// //
// param data_type The data type of the tensor to be set. // param data_type The data type of the tensor to be set.
virtual TypeId set_data_type(const TypeId data_type) { virtual TypeId set_data_type(const TypeId data_type) {
...@@ -314,7 +314,7 @@ class Tensor : public MetaTensor { ...@@ -314,7 +314,7 @@ class Tensor : public MetaTensor {
// //
// The constructed Tensor object has the same type and shape with tensor. // The constructed Tensor object has the same type and shape with tensor.
// //
// param tensor An exisiting Tensor object. // param tensor An existing Tensor object.
Tensor& operator=(const Tensor& tensor); Tensor& operator=(const Tensor& tensor);
// brief Compares two Tensor objects. // brief Compares two Tensor objects.
...@@ -383,7 +383,7 @@ class Tensor : public MetaTensor { ...@@ -383,7 +383,7 @@ class Tensor : public MetaTensor {
// return The [TypeId] of the tensor data. // return The [TypeId] of the tensor data.
TypeId GetDataType(const py::buffer_info& buf) const; TypeId GetDataType(const py::buffer_info& buf) const;
// bried Sets the data type of a tensor. // brief Sets the data type of a tensor.
// //
// param data_type The data type of the tensor to be set. // param data_type The data type of the tensor to be set.
// //
......
...@@ -43,14 +43,13 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vector<opt: ...@@ -43,14 +43,13 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vector<opt:
} }
auto &inputs = node->cast<CNodePtr>()->inputs(); auto &inputs = node->cast<CNodePtr>()->inputs();
// infact, funcs_size == inps_size - 1
auto funcs_size = funcs.size(); auto funcs_size = funcs.size();
auto inps_size = inputs.size(); auto inputs_size = inputs.size();
// check the inputs are matched with the predicate functions // check the inputs are matched with the predicate functions
if (funcs_size > 0) { if (funcs_size > 0) {
// use the predicate function list to check the number of inputs // use the predicate function list to check the number of inputs
if (funcs_size != (inps_size - 1)) { if (funcs_size != (inputs_size - 1)) {
return; return;
} }
...@@ -63,7 +62,7 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vector<opt: ...@@ -63,7 +62,7 @@ VisitFuncType AnfVisitor::Match(const PrimitivePtr &prim, const std::vector<opt:
} }
// visit the inputs // visit the inputs
for (size_t i = 1; i < inps_size; i++) { for (size_t i = 1; i < inputs_size; i++) {
this->Visit(inputs[i]); this->Visit(inputs[i]);
} }
}; };
......
...@@ -36,7 +36,7 @@ ...@@ -36,7 +36,7 @@
#endif #endif
const char SINGLE_OP_GRAPH[] = "single_op_graph"; const char SINGLE_OP_GRAPH[] = "single_op_graph";
// primitive unable to infer value for constant input in pynative mode // primitive unable to infer value for constant input in PyNative mode
const std::unordered_set<std::string> vm_operators = {"partial", "depend"}; const std::unordered_set<std::string> vm_operators = {"partial", "depend"};
namespace mindspore { namespace mindspore {
...@@ -45,7 +45,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) { ...@@ -45,7 +45,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) {
ValuePtr converted_ret = nullptr; ValuePtr converted_ret = nullptr;
bool converted = parse::ConvertData(obj, &converted_ret); bool converted = parse::ConvertData(obj, &converted_ret);
if (!converted) { if (!converted) {
MS_LOG(EXCEPTION) << "attribute convert error with type:" << std::string(py::str(obj)); MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj));
} }
return converted_ret; return converted_ret;
} }
...@@ -67,7 +67,7 @@ void PynativeInfer(const PrimitivePyPtr& prim, const py::tuple& py_args, OpExecI ...@@ -67,7 +67,7 @@ void PynativeInfer(const PrimitivePyPtr& prim, const py::tuple& py_args, OpExecI
OpExecInfoPtr GenerateOpExecInfo(const py::args& args) { OpExecInfoPtr GenerateOpExecInfo(const py::args& args) {
if (args.size() != PY_ARGS_NUM) { if (args.size() != PY_ARGS_NUM) {
MS_LOG(ERROR) << "four args are needed by RunOp"; MS_LOG(ERROR) << "Four args are needed by RunOp";
return nullptr; return nullptr;
} }
auto op_exec_info = std::make_shared<OpExecInfo>(); auto op_exec_info = std::make_shared<OpExecInfo>();
...@@ -145,13 +145,13 @@ py::object RunOpInVM(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat ...@@ -145,13 +145,13 @@ py::object RunOpInVM(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat
py::object RunOpInMs(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* status) { py::object RunOpInMs(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* status) {
MS_EXCEPTION_IF_NULL(op_exec_info); MS_EXCEPTION_IF_NULL(op_exec_info);
MS_LOG(INFO) << "start run op[" << op_exec_info->op_name << "] with backend policy ms"; MS_LOG(INFO) << "Start run op[" << op_exec_info->op_name << "] with backend policy ms";
auto ms_context = MsContext::GetInstance(); auto ms_context = MsContext::GetInstance();
MS_EXCEPTION_IF_NULL(ms_context); MS_EXCEPTION_IF_NULL(ms_context);
ms_context->set_enable_pynative_infer(true); ms_context->set_enable_pynative_infer(true);
std::string device_target = ms_context->device_target(); std::string device_target = ms_context->device_target();
if (device_target != kAscendDevice && device_target != kGPUDevice) { if (device_target != kAscendDevice && device_target != kGPUDevice) {
MS_EXCEPTION(ArgumentError) << "device target [" << device_target << "] is not supported in Pynative mode"; MS_EXCEPTION(ArgumentError) << "Device target [" << device_target << "] is not supported in Pynative mode";
} }
std::shared_ptr<session::SessionBasic> session = session::SessionFactory::Get().Create(device_target); std::shared_ptr<session::SessionBasic> session = session::SessionFactory::Get().Create(device_target);
MS_EXCEPTION_IF_NULL(session); MS_EXCEPTION_IF_NULL(session);
...@@ -197,7 +197,7 @@ py::object RunOpWithBackendPolicy(MsBackendPolicy backend_policy, const OpExecIn ...@@ -197,7 +197,7 @@ py::object RunOpWithBackendPolicy(MsBackendPolicy backend_policy, const OpExecIn
break; break;
} }
default: default:
MS_LOG(ERROR) << "No backend configed for run op"; MS_LOG(ERROR) << "No backend configured for run op";
} }
return result; return result;
} }
...@@ -240,7 +240,7 @@ py::tuple RunOp(const py::args& args) { ...@@ -240,7 +240,7 @@ py::tuple RunOp(const py::args& args) {
} }
result = RunOpWithBackendPolicy(backend_policy, op_exec_info, &status); result = RunOpWithBackendPolicy(backend_policy, op_exec_info, &status);
if (status != PYNATIVE_SUCCESS) { if (status != PYNATIVE_SUCCESS) {
MS_LOG(ERROR) << "Fail to run " << op_exec_info->op_name; MS_LOG(ERROR) << "Failed to run " << op_exec_info->op_name;
return err_ret; return err_ret;
} }
......
...@@ -47,7 +47,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) { ...@@ -47,7 +47,7 @@ inline ValuePtr PyAttrValue(const py::object& obj) {
ValuePtr converted_ret = nullptr; ValuePtr converted_ret = nullptr;
bool converted = parse::ConvertData(obj, &converted_ret); bool converted = parse::ConvertData(obj, &converted_ret);
if (!converted) { if (!converted) {
MS_LOG(EXCEPTION) << "attribute convert error with type:" << std::string(py::str(obj)); MS_LOG(EXCEPTION) << "Attribute convert error with type:" << std::string(py::str(obj));
} }
return converted_ret; return converted_ret;
} }
...@@ -67,7 +67,7 @@ MeTensorPtr ConvertPyObjToTensor(const py::object& obj) { ...@@ -67,7 +67,7 @@ MeTensorPtr ConvertPyObjToTensor(const py::object& obj) {
} else if (py::isinstance<py::array>(obj)) { } else if (py::isinstance<py::array>(obj)) {
me_tensor_ptr = std::make_shared<MeTensor>(py::cast<py::array>(obj), nullptr); me_tensor_ptr = std::make_shared<MeTensor>(py::cast<py::array>(obj), nullptr);
} else { } else {
MS_LOG(EXCEPTION) << "run op inputs type is invalid!"; MS_LOG(EXCEPTION) << "Run op inputs type is invalid!";
} }
return me_tensor_ptr; return me_tensor_ptr;
} }
...@@ -97,7 +97,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec ...@@ -97,7 +97,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec
auto const_op_desc = auto const_op_desc =
transform::TransformUtil::GetGeTensorDesc(me_tensor_ptr->shape_c(), me_tensor_ptr->data_type(), kOpFormat_NCHW); transform::TransformUtil::GetGeTensorDesc(me_tensor_ptr->shape_c(), me_tensor_ptr->data_type(), kOpFormat_NCHW);
if (const_op_desc == nullptr) { if (const_op_desc == nullptr) {
MS_LOG(ERROR) << "Create variable " << op_name << " ouptut descriptor failed!"; MS_LOG(ERROR) << "Create variable " << op_name << " output descriptor failed!";
return false; return false;
} }
auto pointer_cast_const_op = std::static_pointer_cast<transform::Constant>(const_op); auto pointer_cast_const_op = std::static_pointer_cast<transform::Constant>(const_op);
...@@ -108,7 +108,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec ...@@ -108,7 +108,7 @@ bool SetInputsForSingleOpGraph(const OpExecInfoPtr& op_exec_info, const std::vec
continue; continue;
} }
if (adapter->setInput(op, op_input_idx++, const_op)) { if (adapter->setInput(op, op_input_idx++, const_op)) {
MS_LOG(ERROR) << "fail to set params, index is " << op_input_idx; MS_LOG(ERROR) << "Failed to set params, index is " << op_input_idx;
return false; return false;
} }
graph_input_nodes->push_back(*const_op); graph_input_nodes->push_back(*const_op);
...@@ -178,7 +178,7 @@ void ToTensorPtr(const OpExecInfoPtr op_exec_info, std::vector<GeTensorPtr>* con ...@@ -178,7 +178,7 @@ void ToTensorPtr(const OpExecInfoPtr op_exec_info, std::vector<GeTensorPtr>* con
MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]); MeTensorPtr me_tensor_ptr = ConvertPyObjToTensor(op_inputs[i]);
auto ge_tensor_ptr = transform::TransformUtil::ConvertTensor(me_tensor_ptr, kOpFormat_NCHW); auto ge_tensor_ptr = transform::TransformUtil::ConvertTensor(me_tensor_ptr, kOpFormat_NCHW);
if (ge_tensor_ptr == nullptr) { if (ge_tensor_ptr == nullptr) {
MS_LOG(EXCEPTION) << "convert inputs to GE tensor failed in op " << op_exec_info->op_name << "."; MS_LOG(EXCEPTION) << "Convert inputs to GE tensor failed in op " << op_exec_info->op_name << ".";
} }
// set inputs for operator to build single node graph // set inputs for operator to build single node graph
inputs->push_back(ge_tensor_ptr); inputs->push_back(ge_tensor_ptr);
...@@ -192,7 +192,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st ...@@ -192,7 +192,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st
for (auto& item : op_attrs) { for (auto& item : op_attrs) {
if (!py::isinstance<py::str>(item.first)) { if (!py::isinstance<py::str>(item.first)) {
MS_LOG(ERROR) << "type error in py dict convert"; MS_LOG(ERROR) << "Type error in py dict convert";
return PYNATIVE_OP_ATTRS_ERR; return PYNATIVE_OP_ATTRS_ERR;
} }
std::string name = py::cast<std::string>(item.first); std::string name = py::cast<std::string>(item.first);
...@@ -203,7 +203,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st ...@@ -203,7 +203,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st
// build graph // build graph
GeGraphPtr graph = std::make_shared<GeGraph>(op_exec_info->op_name); GeGraphPtr graph = std::make_shared<GeGraph>(op_exec_info->op_name);
if (BuildSingleOpGraph(op_exec_info, inputs, attrs, graph) == false) { if (BuildSingleOpGraph(op_exec_info, inputs, attrs, graph) == false) {
MS_LOG(ERROR) << "Fail to BuildSingleOpGraph"; MS_LOG(ERROR) << "Failed to BuildSingleOpGraph";
return PYNATIVE_GRAPH_GE_BUILD_ERR; return PYNATIVE_GRAPH_GE_BUILD_ERR;
} }
...@@ -211,7 +211,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st ...@@ -211,7 +211,7 @@ PynativeStatusCode ConvertAttributes(const OpExecInfoPtr& op_exec_info, const st
transform::Status ret = transform::Status ret =
transform::DfGraphManager::GetInstance().AddGraph(SINGLE_OP_GRAPH, std::shared_ptr<transform::DfGraph>(graph)); transform::DfGraphManager::GetInstance().AddGraph(SINGLE_OP_GRAPH, std::shared_ptr<transform::DfGraph>(graph));
if (ret != transform::SUCCESS) { if (ret != transform::SUCCESS) {
MS_LOG(ERROR) << "Fail to AddGraph into graph manager"; MS_LOG(ERROR) << "Failed to AddGraph into graph manager";
return PYNATIVE_GRAPH_MANAGER_ERR; return PYNATIVE_GRAPH_MANAGER_ERR;
} }
...@@ -289,7 +289,7 @@ py::object RunOpInGE(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat ...@@ -289,7 +289,7 @@ py::object RunOpInGE(const OpExecInfoPtr& op_exec_info, PynativeStatusCode* stat
run_ret = graph_runner->RunGraph(run_options, ge_inputs, &ge_outputs); run_ret = graph_runner->RunGraph(run_options, ge_inputs, &ge_outputs);
} }
if (run_ret != transform::Status::SUCCESS) { if (run_ret != transform::Status::SUCCESS) {
MS_LOG(ERROR) << "GraphRunner Fails to Run Graph"; MS_LOG(ERROR) << "GraphRunner fails to run graph";
*status = PYNATIVE_GRAPH_GE_RUN_ERR; *status = PYNATIVE_GRAPH_GE_RUN_ERR;
return std::move(err_ret); return std::move(err_ret);
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册