// Copyright 2018 The MACE Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include #include #include #include #include #include #include #include #include #include #include #include #include "gflags/gflags.h" #include "mace/port/env.h" #include "mace/port/file_system.h" #include "mace/public/mace.h" #include "mace/utils/logging.h" #include "mace/utils/memory.h" #include "mace/utils/string_util.h" // if convert model to code. #ifdef MODEL_GRAPH_FORMAT_CODE #include "mace/codegen/engine/mace_engine_factory.h" #endif #ifdef MACE_ENABLE_OPENCL namespace mace { const unsigned char *LoadOpenCLBinary(); size_t OpenCLBinarySize(); const unsigned char *LoadOpenCLParameter(); size_t OpenCLParameterSize(); } // namespace mace #endif namespace mace { namespace examples { void ParseShape(const std::string &str, std::vector *shape) { std::string tmp = str; while (!tmp.empty()) { int dim = atoi(tmp.data()); shape->push_back(dim); size_t next_offset = tmp.find(","); if (next_offset == std::string::npos) { break; } else { tmp = tmp.substr(next_offset + 1); } } } std::string FormatName(const std::string input) { std::string res = input; for (size_t i = 0; i < input.size(); ++i) { if (!isalnum(res[i])) res[i] = '_'; } return res; } DeviceType ParseDeviceType(const std::string &device_str) { if (device_str.compare("CPU") == 0) { return DeviceType::CPU; } else if (device_str.compare("GPU") == 0) { return DeviceType::GPU; } else if (device_str.compare("HEXAGON") == 0) { return DeviceType::HEXAGON; } else if (device_str.compare("HTA") == 0) { return DeviceType::HTA; } else { return DeviceType::CPU; } } DataFormat ParseDataFormat(const std::string &data_format_str) { if (data_format_str == "NHWC") { return DataFormat::NHWC; } else if (data_format_str == "NCHW") { return DataFormat::NCHW; } else if (data_format_str == "OIHW") { return DataFormat::OIHW; } else { return DataFormat::NONE; } } DEFINE_string(model_name, "", "model name in model deployment file"); DEFINE_string(input_node, "", "input nodes, separated by comma," "example: input_node0,input_node1"); DEFINE_string(input_shape, "", "input shapes, separated by colon and comma, " "example: 1,224,224,3:1,1,1,10"); DEFINE_string(output_node, "output_node0,output_node1", "output nodes, separated by comma"); DEFINE_string(output_shape, "", "output shapes, separated by colon and comma, " "example: 1,224,224,2:1,1,1,10"); DEFINE_string(input_data_format, "NHWC", "input data formats, NONE|NHWC|NCHW"); DEFINE_string(output_data_format, "NHWC", "output data formats, NONE|NHWC|NCHW"); DEFINE_string(input_file, "", "input file name | input file prefix for multiple inputs."); DEFINE_string(output_file, "", "output file name | output file prefix for multiple outputs"); DEFINE_string(input_dir, "", "input directory name"); DEFINE_string(output_dir, "", "output directory name"); DEFINE_string(opencl_binary_file, "", "compiled opencl binary file path"); DEFINE_string(opencl_parameter_file, "", "tuned OpenCL parameter file path"); DEFINE_string(model_data_file, "", "model data file name, used when model_data_format == file"); DEFINE_string(model_file, "", "model file name, used when load mace model in pb"); DEFINE_string(device, "GPU", "CPU/GPU/HEXAGON"); DEFINE_int32(round, 1, "round"); DEFINE_int32(restart_round, 1, "restart round"); DEFINE_int32(malloc_check_cycle, -1, "malloc debug check cycle, -1 to disable"); DEFINE_int32(gpu_perf_hint, 2, "0:DEFAULT/1:LOW/2:NORMAL/3:HIGH"); DEFINE_int32(gpu_priority_hint, 1, "0:DEFAULT/1:LOW/2:NORMAL/3:HIGH"); DEFINE_int32(omp_num_threads, -1, "num of openmp threads"); DEFINE_int32(cpu_affinity_policy, 1, "0:AFFINITY_NONE/1:AFFINITY_BIG_ONLY/2:AFFINITY_LITTLE_ONLY"); bool RunModel(const std::vector &input_names, const std::vector> &input_shapes, const std::vector &input_data_formats, const std::vector &output_names, const std::vector> &output_shapes, const std::vector &output_data_formats) { // load model DeviceType device_type = ParseDeviceType(FLAGS_device); // configuration // Detailed information please see mace.h MaceStatus status; MaceEngineConfig config(device_type); status = config.SetCPUThreadPolicy( FLAGS_omp_num_threads, static_cast(FLAGS_cpu_affinity_policy)); if (status != MaceStatus::MACE_SUCCESS) { std::cerr << "Set openmp or cpu affinity failed." << std::endl; } #ifdef MACE_ENABLE_OPENCL std::shared_ptr gpu_context; if (device_type == DeviceType::GPU) { // DO NOT USE tmp directory. // Please use APP's own directory and make sure the directory exists. const char *storage_path_ptr = getenv("MACE_INTERNAL_STORAGE_PATH"); const std::string storage_path = std::string(storage_path_ptr == nullptr ? "/data/local/tmp/mace_run/interior" : storage_path_ptr); std::vector opencl_binary_paths = {FLAGS_opencl_binary_file}; gpu_context = GPUContextBuilder() .SetStoragePath(storage_path) .SetOpenCLBinaryPaths(opencl_binary_paths) .SetOpenCLBinary(LoadOpenCLBinary(), OpenCLBinarySize()) .SetOpenCLParameterPath(FLAGS_opencl_parameter_file) .SetOpenCLParameter(LoadOpenCLParameter(), OpenCLParameterSize()) .Finalize(); config.SetGPUContext(gpu_context); config.SetGPUHints( static_cast(FLAGS_gpu_perf_hint), static_cast(FLAGS_gpu_priority_hint)); } #endif // MACE_ENABLE_OPENCL // Create Engine std::shared_ptr engine; MaceStatus create_engine_status; std::unique_ptr model_graph_data = make_unique(); if (FLAGS_model_file != "") { auto fs = GetFileSystem(); auto status = fs->NewReadOnlyMemoryRegionFromFile(FLAGS_model_file.c_str(), &model_graph_data); if (status != MaceStatus::MACE_SUCCESS) { LOG(FATAL) << "Failed to read file: " << FLAGS_model_file; } } std::unique_ptr model_weights_data = make_unique(); if (FLAGS_model_data_file != "") { auto fs = GetFileSystem(); auto status = fs->NewReadOnlyMemoryRegionFromFile( FLAGS_model_data_file.c_str(), &model_weights_data); if (status != MaceStatus::MACE_SUCCESS) { LOG(FATAL) << "Failed to read file: " << FLAGS_model_data_file; } MACE_CHECK(model_weights_data->length() > 0); } // Only choose one of the two type based on the `model_graph_format` // in model deployment file(.yml). #ifdef MODEL_GRAPH_FORMAT_CODE // if model_data_format == code, just pass an empty string("") // to model_data_file parameter. create_engine_status = CreateMaceEngineFromCode( FLAGS_model_name, reinterpret_cast(model_weights_data->data()), model_weights_data->length(), input_names, output_names, config, &engine); #else create_engine_status = CreateMaceEngineFromProto( reinterpret_cast(model_graph_data->data()), model_graph_data->length(), reinterpret_cast(model_weights_data->data()), model_weights_data->length(), input_names, output_names, config, &engine); #endif if (create_engine_status != MaceStatus::MACE_SUCCESS) { std::cerr << "Create engine error, please check the arguments first, " << "if correct, the device may not run the model, " << "please fall back to other strategy." << std::endl; exit(1); } const size_t input_count = input_names.size(); const size_t output_count = output_names.size(); std::map inputs; std::map outputs; std::map inputs_size; for (size_t i = 0; i < input_count; ++i) { int64_t input_size = std::accumulate(input_shapes[i].begin(), input_shapes[i].end(), 1, std::multiplies()); inputs_size[input_names[i]] = input_size; // Only support float and int32 data type auto buffer_in = std::shared_ptr(new float[input_size], std::default_delete()); inputs[input_names[i]] = mace::MaceTensor(input_shapes[i], buffer_in, input_data_formats[i]); } for (size_t i = 0; i < output_count; ++i) { int64_t output_size = std::accumulate(output_shapes[i].begin(), output_shapes[i].end(), 1, std::multiplies()); // Only support float and int32 data type auto buffer_out = std::shared_ptr(new float[output_size], std::default_delete()); outputs[output_names[i]] = mace::MaceTensor(output_shapes[i], buffer_out, output_data_formats[i]); } if (!FLAGS_input_dir.empty()) { DIR *dir_parent; struct dirent *entry; dir_parent = opendir(FLAGS_input_dir.c_str()); if (dir_parent) { while ((entry = readdir(dir_parent))) { std::string file_name = std::string(entry->d_name); std::string prefix = FormatName(input_names[0]); if (file_name.find(prefix) == 0) { std::string suffix = file_name.substr(prefix.size()); for (size_t i = 0; i < input_count; ++i) { file_name = FLAGS_input_dir + "/" + FormatName(input_names[i]) + suffix; std::ifstream in_file(file_name, std::ios::in | std::ios::binary); std::cout << "Read " << file_name << std::endl; if (in_file.is_open()) { in_file.read(reinterpret_cast( inputs[input_names[i]].data().get()), inputs_size[input_names[i]] * sizeof(float)); in_file.close(); } else { std::cerr << "Open input file failed" << std::endl; return -1; } } engine->Run(inputs, &outputs); if (!FLAGS_output_dir.empty()) { for (size_t i = 0; i < output_count; ++i) { std::string output_name = FLAGS_output_dir + "/" + FormatName(output_names[i]) + suffix; std::ofstream out_file(output_name, std::ios::binary); if (out_file.is_open()) { int64_t output_size = std::accumulate(output_shapes[i].begin(), output_shapes[i].end(), 1, std::multiplies()); out_file.write( reinterpret_cast( outputs[output_names[i]].data().get()), output_size * sizeof(float)); out_file.flush(); out_file.close(); } else { std::cerr << "Open output file failed" << std::endl; return -1; } } } } } closedir(dir_parent); } else { std::cerr << "Directory " << FLAGS_input_dir << " does not exist." << std::endl; } } else { for (size_t i = 0; i < input_count; ++i) { std::ifstream in_file(FLAGS_input_file + "_" + FormatName(input_names[i]), std::ios::in | std::ios::binary); if (in_file.is_open()) { in_file.read(reinterpret_cast( inputs[input_names[i]].data().get()), inputs_size[input_names[i]] * sizeof(float)); in_file.close(); } else { std::cerr << "Open input file failed" << std::endl; return -1; } } engine->Run(inputs, &outputs); for (size_t i = 0; i < output_count; ++i) { std::string output_name = FLAGS_output_file + "_" + FormatName(output_names[i]); std::ofstream out_file(output_name, std::ios::binary); int64_t output_size = std::accumulate(output_shapes[i].begin(), output_shapes[i].end(), 1, std::multiplies()); if (out_file.is_open()) { out_file.write( reinterpret_cast(outputs[output_names[i]].data().get()), output_size * sizeof(float)); out_file.flush(); out_file.close(); } else { std::cerr << "Open output file failed" << std::endl; return -1; } } } std::cout << "Finished" << std::endl; return true; } int Main(int argc, char **argv) { std::string usage = "example run\nusage: " + std::string(argv[0]) + " [flags]"; gflags::SetUsageMessage(usage); gflags::ParseCommandLineFlags(&argc, &argv, true); std::cout << "mace version: " << MaceVersion() << std::endl; std::cout << "input node: " << FLAGS_input_node << std::endl; std::cout << "input shape: " << FLAGS_input_shape << std::endl; std::cout << "output node: " << FLAGS_output_node << std::endl; std::cout << "output shape: " << FLAGS_output_shape << std::endl; std::cout << "input_file: " << FLAGS_input_file << std::endl; std::cout << "output_file: " << FLAGS_output_file << std::endl; std::cout << "input_dir: " << FLAGS_input_dir << std::endl; std::cout << "output dir: " << FLAGS_output_dir << std::endl; std::cout << "model_data_file: " << FLAGS_model_data_file << std::endl; std::cout << "model_file: " << FLAGS_model_file << std::endl; std::cout << "device: " << FLAGS_device << std::endl; std::cout << "round: " << FLAGS_round << std::endl; std::cout << "restart_round: " << FLAGS_restart_round << std::endl; std::cout << "gpu_perf_hint: " << FLAGS_gpu_perf_hint << std::endl; std::cout << "gpu_priority_hint: " << FLAGS_gpu_priority_hint << std::endl; std::cout << "omp_num_threads: " << FLAGS_omp_num_threads << std::endl; std::cout << "cpu_affinity_policy: " << FLAGS_cpu_affinity_policy << std::endl; std::vector input_names = Split(FLAGS_input_node, ','); std::vector output_names = Split(FLAGS_output_node, ','); std::vector input_shapes = Split(FLAGS_input_shape, ':'); std::vector output_shapes = Split(FLAGS_output_shape, ':'); const size_t input_count = input_shapes.size(); const size_t output_count = output_shapes.size(); std::vector> input_shape_vec(input_count); std::vector> output_shape_vec(output_count); for (size_t i = 0; i < input_count; ++i) { ParseShape(input_shapes[i], &input_shape_vec[i]); } for (size_t i = 0; i < output_count; ++i) { ParseShape(output_shapes[i], &output_shape_vec[i]); } std::vector raw_input_data_formats = Split(FLAGS_input_data_format, ','); std::vector raw_output_data_formats = Split(FLAGS_output_data_format, ','); std::vector input_data_formats(input_count); std::vector output_data_formats(output_count); for (size_t i = 0; i < input_count; ++i) { input_data_formats[i] = ParseDataFormat(raw_input_data_formats[i]); } for (size_t i = 0; i < output_count; ++i) { output_data_formats[i] = ParseDataFormat(raw_output_data_formats[i]); } bool ret = false; for (int i = 0; i < FLAGS_restart_round; ++i) { std::cout << "restart round " << i << std::endl; ret = RunModel(input_names, input_shape_vec, input_data_formats, output_names, output_shape_vec, output_data_formats); } if (ret) { return 0; } else { return -1; } } } // namespace examples } // namespace mace int main(int argc, char **argv) { mace::examples::Main(argc, argv); }