main.cc 7.6 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
//   Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#include <glog/logging.h>

#include <iostream>
#include <string>
#include <vector>
20 21
#include <sys/types.h>
#include <sys/stat.h>
G
Guanghua Yu 已提交
22
#include <algorithm>
23

C
cnn 已提交
24 25 26 27 28 29 30 31
#ifdef _WIN32
#include <direct.h>
#include <io.h>
#elif LINUX
#include <stdarg.h>
#include <sys/stat.h>
#endif

32
#include "include/object_detector.h"
33
#include <gflags/gflags.h>
34 35 36


DEFINE_string(model_dir, "", "Path of inference model");
37
DEFINE_string(image_file, "", "Path of input image");
38
DEFINE_string(video_path, "", "Path of input video");
G
Guanghua Yu 已提交
39 40
DEFINE_bool(use_gpu, false, "Deprecated, please use `--device` to set the device you want to run.");
DEFINE_string(device, "CPU", "Choose the device you want to run, it can be: CPU/GPU/XPU, default is CPU.");
41
DEFINE_bool(use_camera, false, "Use camera or not");
C
channings 已提交
42 43
DEFINE_string(run_mode, "fluid", "Mode of running(fluid/trt_fp32/trt_fp16)");
DEFINE_int32(gpu_id, 0, "Device id of GPU to execute");
C
channings 已提交
44
DEFINE_int32(camera_id, -1, "Device id of camera to predict");
45 46 47
DEFINE_bool(run_benchmark, false, "Whether to predict a image_file repeatedly for benchmark");
DEFINE_double(threshold, 0.5, "Threshold of score.");
DEFINE_string(output_dir, "output", "Directory of output visualization files.");
48
DEFINE_bool(trt_calib_mode, false, "If the model is produced by TRT offline quantitative calibration, trt_calib_mode need to set True");
49

50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68
static std::string DirName(const std::string &filepath) {
  auto pos = filepath.rfind(OS_PATH_SEP);
  if (pos == std::string::npos) {
    return "";
  }
  return filepath.substr(0, pos);
}

static bool PathExists(const std::string& path){
#ifdef _WIN32
  struct _stat buffer;
  return (_stat(path.c_str(), &buffer) == 0);
#else
  struct stat buffer;
  return (stat(path.c_str(), &buffer) == 0);
#endif  // !_WIN32
}

static void MkDir(const std::string& path) {
69
  if (PathExists(path)) return;
70 71 72 73 74 75 76
  int ret = 0;
#ifdef _WIN32
  ret = _mkdir(path.c_str());
#else
  ret = mkdir(path.c_str(), 0755);
#endif  // !_WIN32
  if (ret != 0) {
77 78
    std::string path_error(path);
    path_error += " mkdir failed!";
79 80 81 82 83 84 85 86 87 88 89 90
    throw std::runtime_error(path_error);
  }
}

static void MkDirs(const std::string& path) {
  if (path.empty()) return;
  if (PathExists(path)) return;

  MkDirs(DirName(path));
  MkDir(path);
}

91 92 93 94
void PredictVideo(const std::string& video_path,
                  PaddleDetection::ObjectDetector* det) {
  // Open video
  cv::VideoCapture capture;
C
channings 已提交
95 96 97 98 99
  if (FLAGS_camera_id != -1){
    capture.open(FLAGS_camera_id);
  }else{
    capture.open(video_path.c_str());
  }
100 101 102 103 104 105 106 107 108 109 110 111
  if (!capture.isOpened()) {
    printf("can not open video : %s\n", video_path.c_str());
    return;
  }

  // Get Video info : resolution, fps
  int video_width = static_cast<int>(capture.get(CV_CAP_PROP_FRAME_WIDTH));
  int video_height = static_cast<int>(capture.get(CV_CAP_PROP_FRAME_HEIGHT));
  int video_fps = static_cast<int>(capture.get(CV_CAP_PROP_FPS));

  // Create VideoWriter for output
  cv::VideoWriter video_out;
C
channings 已提交
112
  std::string video_out_path = "output.mp4";
113
  video_out.open(video_out_path.c_str(),
C
channings 已提交
114
                 0x00000021,
115 116 117 118 119 120 121 122 123 124 125 126 127
                 video_fps,
                 cv::Size(video_width, video_height),
                 true);
  if (!video_out.isOpened()) {
    printf("create video writer failed!\n");
    return;
  }

  std::vector<PaddleDetection::ObjectResult> result;
  auto labels = det->GetLabelList();
  auto colormap = PaddleDetection::GenerateColorMap(labels.size());
  // Capture all frames and do inference
  cv::Mat frame;
C
channings 已提交
128
  int frame_id = 0;
129 130 131 132
  while (capture.read(frame)) {
    if (frame.empty()) {
      break;
    }
133
    det->Predict(frame, 0.5, 0, 1, false, &result);
134 135
    cv::Mat out_im = PaddleDetection::VisualizeResult(
        frame, result, labels, colormap);
C
channings 已提交
136 137 138 139 140 141 142 143 144 145
    for (const auto& item : result) {
      printf("In frame id %d, we detect: class=%d confidence=%.2f rect=[%d %d %d %d]\n",
        frame_id,
        item.class_id,
        item.confidence,
        item.rect[0],
        item.rect[1],
        item.rect[2],
        item.rect[3]);
   }   
146
    video_out.write(out_im);
C
channings 已提交
147
    frame_id += 1;
148 149 150 151 152 153
  }
  capture.release();
  video_out.release();
}

void PredictImage(const std::string& image_path,
154 155 156 157
                  const double threshold,
                  const bool run_benchmark,
                  PaddleDetection::ObjectDetector* det,
                  const std::string& output_dir = "output") {
158 159 160 161
  // Open input image as an opencv cv::Mat object
  cv::Mat im = cv::imread(image_path, 1);
  // Store all detected result
  std::vector<PaddleDetection::ObjectResult> result;
162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184
  if (run_benchmark)
  {
    det->Predict(im, threshold, 100, 100, run_benchmark, &result);
  }else
  {
    det->Predict(im, 0.5, 0, 1, run_benchmark, &result);
    for (const auto& item : result) {
      printf("class=%d confidence=%.4f rect=[%d %d %d %d]\n",
          item.class_id,
          item.confidence,
          item.rect[0],
          item.rect[1],
          item.rect[2],
          item.rect[3]);
    }
    // Visualization result
    auto labels = det->GetLabelList();
    auto colormap = PaddleDetection::GenerateColorMap(labels.size());
    cv::Mat vis_img = PaddleDetection::VisualizeResult(
        im, result, labels, colormap);
    std::vector<int> compression_params;
    compression_params.push_back(CV_IMWRITE_JPEG_QUALITY);
    compression_params.push_back(95);
185 186 187 188 189 190 191
    std::string output_path(output_dir);
    if (output_dir.rfind(OS_PATH_SEP) != output_dir.size() - 1) {
      output_path += OS_PATH_SEP;
    }
    output_path += "output.jpg";
    cv::imwrite(output_path, vis_img, compression_params);
    printf("Visualized output saved as %s\n", output_path.c_str());
192 193 194 195 196 197 198
  }
}

int main(int argc, char** argv) {
  // Parsing command-line
  google::ParseCommandLineFlags(&argc, &argv, true);
  if (FLAGS_model_dir.empty()
199
      || (FLAGS_image_file.empty() && FLAGS_video_path.empty())) {
200
    std::cout << "Usage: ./main --model_dir=/PATH/TO/INFERENCE_MODEL/ "
201
                << "--image_file=/PATH/TO/INPUT/IMAGE/" << std::endl;
202 203 204
    return -1;
  }
  if (!(FLAGS_run_mode == "fluid" || FLAGS_run_mode == "trt_fp32"
205 206
      || FLAGS_run_mode == "trt_fp16" || FLAGS_run_mode == "trt_int8")) {
    std::cout << "run_mode should be 'fluid', 'trt_fp32', 'trt_fp16' or 'trt_int8'.";
207
    return -1;
208
  }
G
Guanghua Yu 已提交
209 210 211 212 213 214 215 216 217
  transform(FLAGS_device.begin(),FLAGS_device.end(),FLAGS_device.begin(),::toupper);
  if (!(FLAGS_device == "CPU" || FLAGS_device == "GPU" || FLAGS_device == "XPU")) {
    std::cout << "device should be 'CPU', 'GPU' or 'XPU'.";
    return -1;
  }
  if (FLAGS_use_gpu) {
    std::cout << "Deprecated, please use `--device` to set the device you want to run.";
    return -1;
  }
218 219

  // Load model and create a object detector
G
Guanghua Yu 已提交
220
  PaddleDetection::ObjectDetector det(FLAGS_model_dir, FLAGS_device,
221
    FLAGS_run_mode, FLAGS_gpu_id, FLAGS_trt_calib_mode);
222
  // Do inference on input video or image
223
  if (!FLAGS_video_path.empty() || FLAGS_use_camera) {
224
    PredictVideo(FLAGS_video_path, &det);
225
  } else if (!FLAGS_image_file.empty()) {
226 227 228
    if (!PathExists(FLAGS_output_dir)) {
      MkDirs(FLAGS_output_dir);
    }
229
    PredictImage(FLAGS_image_file, FLAGS_threshold, FLAGS_run_benchmark, &det, FLAGS_output_dir);
230 231 232
  }
  return 0;
}