// Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include // for setprecision #include #include "include/object_detector.h" #ifdef _WIN32 #include #include #else // Linux/Unix #include #endif namespace PaddleDetection { // Load Model and create model predictor void ObjectDetector::LoadModel(const std::string& model_dir, bool use_gpu, const int min_subgraph_size, const int batch_size, const std::string& run_mode, const int gpu_id) { paddle::AnalysisConfig config; std::string prog_file = model_dir + OS_PATH_SEP + "__model__"; std::string params_file = model_dir + OS_PATH_SEP + "__params__"; if (access(prog_file.c_str(), 0) < 0 || access(params_file.c_str(), 0) < 0) { std::cerr << "[ERROR] Model file or parameter file can't be found." << std::endl; success_init_ = false; return; } config.SetModel(prog_file, params_file); if (use_gpu) { config.EnableUseGpu(100, gpu_id); if (run_mode != "fluid") { auto precision = paddle::AnalysisConfig::Precision::kFloat32; if (run_mode == "trt_fp16") { precision = paddle::AnalysisConfig::Precision::kHalf; } else if (run_mode == "trt_int8") { printf("TensorRT int8 mode is not supported now, " "please use 'trt_fp32' or 'trt_fp16' instead"); } else { if (run_mode != "trt_32") { printf("run_mode should be 'fluid', 'trt_fp32' or 'trt_fp16'"); } } config.EnableTensorRtEngine( 1 << 10, batch_size, min_subgraph_size, precision, false, false); } } else { config.DisableGpu(); } config.SwitchUseFeedFetchOps(false); config.SwitchSpecifyInputNames(true); // Memory optimization config.EnableMemoryOptim(); predictor_ = std::move(CreatePaddlePredictor(config)); } // Visualiztion MaskDetector results cv::Mat VisualizeResult(const cv::Mat& img, const std::vector& results, const std::vector& lable_list, const std::vector& colormap) { cv::Mat vis_img = img.clone(); for (int i = 0; i < results.size(); ++i) { int w = results[i].rect[1] - results[i].rect[0]; int h = results[i].rect[3] - results[i].rect[2]; cv::Rect roi = cv::Rect(results[i].rect[0], results[i].rect[2], w, h); // Configure color and text size std::ostringstream oss; oss << std::setiosflags(std::ios::fixed) << std::setprecision(4); oss << lable_list[results[i].class_id] << " "; oss << results[i].confidence; std::string text = oss.str(); int c1 = colormap[3 * results[i].class_id + 0]; int c2 = colormap[3 * results[i].class_id + 1]; int c3 = colormap[3 * results[i].class_id + 2]; cv::Scalar roi_color = cv::Scalar(c1, c2, c3); int font_face = cv::FONT_HERSHEY_COMPLEX_SMALL; double font_scale = 0.5f; float thickness = 0.5; cv::Size text_size = cv::getTextSize(text, font_face, font_scale, thickness, nullptr); cv::Point origin; origin.x = roi.x; origin.y = roi.y; // Configure text background cv::Rect text_back = cv::Rect(results[i].rect[0], results[i].rect[2] - text_size.height, text_size.width, text_size.height); // Draw roi object, text, and background cv::rectangle(vis_img, roi, roi_color, 2); cv::rectangle(vis_img, text_back, roi_color, -1); cv::putText(vis_img, text, origin, font_face, font_scale, cv::Scalar(255, 255, 255), thickness); } return vis_img; } void ObjectDetector::Preprocess(const cv::Mat& ori_im) { // Clone the image : keep the original mat for postprocess cv::Mat im = ori_im.clone(); cv::cvtColor(im, im, cv::COLOR_BGR2RGB); preprocessor_.Run(&im, &inputs_); } void ObjectDetector::Postprocess( const cv::Mat& raw_mat, std::vector* result) { result->clear(); int rh = 1; int rw = 1; if (config_.arch_ == "SSD" || config_.arch_ == "Face") { rh = raw_mat.rows; rw = raw_mat.cols; } int total_size = output_data_.size() / 6; for (int j = 0; j < total_size; ++j) { // Class id int class_id = static_cast(round(output_data_[0 + j * 6])); // Confidence score float score = output_data_[1 + j * 6]; int xmin = (output_data_[2 + j * 6] * rw); int ymin = (output_data_[3 + j * 6] * rh); int xmax = (output_data_[4 + j * 6] * rw); int ymax = (output_data_[5 + j * 6] * rh); int wd = xmax - xmin; int hd = ymax - ymin; if (score > threshold_ && class_id > -1) { ObjectResult result_item; result_item.rect = {xmin, xmax, ymin, ymax}; result_item.class_id = class_id; result_item.confidence = score; result->push_back(result_item); } } } void ObjectDetector::Predict(const cv::Mat& im, std::vector* result) { // Preprocess image Preprocess(im); // Prepare input tensor auto input_names = predictor_->GetInputNames(); for (const auto& tensor_name : input_names) { auto in_tensor = predictor_->GetInputTensor(tensor_name); if (tensor_name == "image") { int rh = inputs_.eval_im_size_f_[0]; int rw = inputs_.eval_im_size_f_[1]; in_tensor->Reshape({1, 3, rh, rw}); in_tensor->copy_from_cpu(inputs_.im_data_.data()); } else if (tensor_name == "im_size") { in_tensor->Reshape({1, 2}); in_tensor->copy_from_cpu(inputs_.ori_im_size_.data()); } else if (tensor_name == "im_info") { in_tensor->Reshape({1, 3}); in_tensor->copy_from_cpu(inputs_.eval_im_size_f_.data()); } else if (tensor_name == "im_shape") { in_tensor->Reshape({1, 3}); in_tensor->copy_from_cpu(inputs_.ori_im_size_f_.data()); } else if (tensor_name == "scale_factor") { in_tensor->Reshape({1, 4}); in_tensor->copy_from_cpu(inputs_.scale_factor_f_.data()); } } // Run predictor predictor_->ZeroCopyRun(); // Get output tensor auto output_names = predictor_->GetOutputNames(); auto out_tensor = predictor_->GetOutputTensor(output_names[0]); std::vector output_shape = out_tensor->shape(); // Calculate output length int output_size = 1; for (int j = 0; j < output_shape.size(); ++j) { output_size *= output_shape[j]; } if (output_size < 6) { std::cerr << "[WARNING] No object detected." << std::endl; } output_data_.resize(output_size); out_tensor->copy_to_cpu(output_data_.data()); // Postprocessing result Postprocess(im, result); } std::vector GenerateColorMap(int num_class) { auto colormap = std::vector(3 * num_class, 0); for (int i = 0; i < num_class; ++i) { int j = 0; int lab = i; while (lab) { colormap[i * 3] |= (((lab >> 0) & 1) << (7 - j)); colormap[i * 3 + 1] |= (((lab >> 1) & 1) << (7 - j)); colormap[i * 3 + 2] |= (((lab >> 2) & 1) << (7 - j)); ++j; lab >>= 3; } } return colormap; } } // namespace PaddleDetection