diff --git a/paddle/fluid/inference/api/demo_ci/trt_mobilenet_demo.cc b/paddle/fluid/inference/api/demo_ci/trt_mobilenet_demo.cc index 43776278598681a54e22f8443f728c75a7d43e61..ffb12b5871f088f15e43a1b0ff7e2a8b2f5fd079 100644 --- a/paddle/fluid/inference/api/demo_ci/trt_mobilenet_demo.cc +++ b/paddle/fluid/inference/api/demo_ci/trt_mobilenet_demo.cc @@ -18,15 +18,9 @@ limitations under the License. */ #include #include // use glog instead of CHECK to avoid importing other paddle header files. -#include -#include - -// #include "paddle/fluid/platform/enforce.h" #include "paddle/fluid/inference/demo_ci/utils.h" -#ifdef PADDLE_WITH_CUDA DECLARE_double(fraction_of_gpu_memory_to_use); -#endif DEFINE_string(modeldir, "", "Directory of the inference model."); DEFINE_string(refer, "", "path to reference result for comparison."); DEFINE_string( @@ -38,7 +32,7 @@ namespace paddle { namespace demo { /* - * Use the native fluid engine to inference the demo. + * Use the tensorrt fluid engine to inference the demo. */ void Main() { std::unique_ptr predictor;