diff --git a/paddle/fluid/inference/api/demo_ci/CMakeLists.txt b/paddle/fluid/inference/api/demo_ci/CMakeLists.txt index ae01edb80fcce2f0d1a84086454d5b2a63f611d1..ec8471ef960a2fc44af23c52be09cd678fab3f70 100644 --- a/paddle/fluid/inference/api/demo_ci/CMakeLists.txt +++ b/paddle/fluid/inference/api/demo_ci/CMakeLists.txt @@ -62,10 +62,10 @@ include_directories("${PADDLE_LIB}/third_party/boost") include_directories("${PADDLE_LIB}/third_party/eigen3") if (NOT WIN32) -if (USE_TENSORRT AND WITH_GPU) -include_directories("${TENSORRT_INCLUDE_DIR}") -link_directories("${TENSORRT_LIB_DIR}") -endif() + if (USE_TENSORRT AND WITH_GPU) + include_directories("${TENSORRT_INCLUDE_DIR}") + link_directories("${TENSORRT_LIB_DIR}") + endif() endif(NOT WIN32) if (NOT WIN32) diff --git a/paddle/fluid/inference/api/demo_ci/vis_demo.cc b/paddle/fluid/inference/api/demo_ci/vis_demo.cc index 183f5a86e71014313630543166e931468c5d765d..b9d627b4a510df7901b776de01fc00420c6eec53 100644 --- a/paddle/fluid/inference/api/demo_ci/vis_demo.cc +++ b/paddle/fluid/inference/api/demo_ci/vis_demo.cc @@ -124,9 +124,7 @@ void Main(bool use_gpu, bool use_trt) { config.device = 0; config.max_batch_size = 1; config.fraction_of_gpu_memory = 0.1; // set by yourself - predictor = - CreatePaddlePredictor(config); + predictor = CreatePaddlePredictor(config); } VLOG(3) << "begin to process data"; @@ -166,7 +164,7 @@ int main(int argc, char** argv) { } else if (FLAGS_use_gpu) { paddle::demo::Main(true /*use_gpu*/, false); } else { - paddle::demo::Main(false /* use_gpu*/, false /*use_tensorrt*/); + paddle::demo::Main(false /*use_gpu*/, false /*use_tensorrt*/); } return 0; }