From 78c5b9a019a5f51dbc50378c3f5f985078aa5f39 Mon Sep 17 00:00:00 2001 From: MRXLT Date: Thu, 13 Aug 2020 17:27:50 +0800 Subject: [PATCH] bug fix --- .../inferencer-fluid-gpu/include/fluid_gpu_engine.h | 1 + python/paddle_serving_server_gpu/__init__.py | 2 +- python/paddle_serving_server_gpu/serve.py | 2 +- 3 files changed, 3 insertions(+), 2 deletions(-) diff --git a/paddle_inference/inferencer-fluid-gpu/include/fluid_gpu_engine.h b/paddle_inference/inferencer-fluid-gpu/include/fluid_gpu_engine.h index 60d5bcdb..99fb0578 100644 --- a/paddle_inference/inferencer-fluid-gpu/include/fluid_gpu_engine.h +++ b/paddle_inference/inferencer-fluid-gpu/include/fluid_gpu_engine.h @@ -248,6 +248,7 @@ class FluidGpuAnalysisDirCore : public FluidFamilyCore { paddle::AnalysisConfig::Precision::kFloat32, true, true); + LOG(INFO) << "create TensorRT predictor"; } else { if (params.enable_memory_optimization()) { analysis_config.EnableMemoryOptim(); diff --git a/python/paddle_serving_server_gpu/__init__.py b/python/paddle_serving_server_gpu/__init__.py index 0d6936e9..8dda9898 100644 --- a/python/paddle_serving_server_gpu/__init__.py +++ b/python/paddle_serving_server_gpu/__init__.py @@ -248,7 +248,7 @@ class Server(object): def set_gpuid(self, gpuid=0): self.gpuid = gpuid - def use_trt(self): + def set_trt(self): self.use_trt = True def _prepare_engine(self, model_config_paths, device): diff --git a/python/paddle_serving_server_gpu/serve.py b/python/paddle_serving_server_gpu/serve.py index 8f16e0c7..7ed5dc02 100644 --- a/python/paddle_serving_server_gpu/serve.py +++ b/python/paddle_serving_server_gpu/serve.py @@ -65,7 +65,7 @@ def start_gpu_card_model(index, gpuid, args): # pylint: disable=doc-string-miss server.set_ir_optimize(ir_optim) server.set_max_body_size(max_body_size) if args.use_trt: - server.use_trt() + server.set_trt() server.load_model_config(model) server.prepare_server(workdir=workdir, port=port, device=device) -- GitLab