diff --git a/python/paddle_serving_server_gpu/__init__.py b/python/paddle_serving_server_gpu/__init__.py index d8bcd7bbf6aa827ff80fa455726b6f3f356eb810..5ba93becf671bb24186697027824050100300701 100644 --- a/python/paddle_serving_server_gpu/__init__.py +++ b/python/paddle_serving_server_gpu/__init__.py @@ -42,7 +42,7 @@ def serve_args(): "--device", type=str, default="gpu", help="Type of device") parser.add_argument("--gpu_ids", type=str, default="", help="gpu ids") parser.add_argument( - "--name", type=str, default="default", help="Default service name") + "--name", type=str, default="None", help="Default service name") return parser.parse_args() diff --git a/python/paddle_serving_server_gpu/serve.py b/python/paddle_serving_server_gpu/serve.py index cc9b18f6920c46c5d0119e8adfaf8f76ecf2ad26..5d9d96d517d64b21313fda0b44a83b34142b014b 100644 --- a/python/paddle_serving_server_gpu/serve.py +++ b/python/paddle_serving_server_gpu/serve.py @@ -88,4 +88,18 @@ def start_multi_card(args): # pylint: disable=doc-string-missing if __name__ == "__main__": args = serve_args() - start_multi_card(args) + if args.name == "None": + start_multi_card(args) + else: + web_service = WebService(name=args.name) + web_service.load_model_config(args.model) + gpu_ids = [] + if args.gpu_ids == "": + if "CUDA_VISIBLE_DEVICES" in os.environ: + gpu_ids = os.environ["CUDA_VISIBLE_DEVICES"] + if len(gpu_ids) > 0: + gpus = [int(x) for x in gpu_ids.split(",")] + web_service.set_gpus(gpus) + web_service.prepare_server( + workdir=args.workdir, port=args.port, device=args.device) + web_service.run_server() diff --git a/python/paddle_serving_server_gpu/web_serve.py b/python/paddle_serving_server_gpu/web_serve.py deleted file mode 100644 index 734e6d7b93b4f3ad22f330b1545b63c6ac6f2838..0000000000000000000000000000000000000000 --- a/python/paddle_serving_server_gpu/web_serve.py +++ /dev/null @@ -1,39 +0,0 @@ -# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -""" -Usage: - Host a trained paddle model with one line command - Example: - python -m paddle_serving_server.web_serve --model ./serving_server_model --port 9292 -""" -import os -from multiprocessing import Pool, Process -from .web_service import WebService -import paddle_serving_server_gpu as serving -from paddle_serving_server_gpu import serve_args - -if __name__ == "__main__": - args = serve_args() - web_service = WebService(name=args.name) - web_service.load_model_config(args.model) - gpu_ids = [] - if args.gpu_ids == "": - if "CUDA_VISIBLE_DEVICES" in os.environ: - gpu_ids = os.environ["CUDA_VISIBLE_DEVICES"] - if len(gpu_ids) > 0: - gpus = [int(x) for x in gpu_ids.split(",")] - web_service.set_gpus(gpus) - web_service.prepare_server( - workdir=args.workdir, port=args.port, device=args.device) - web_service.run_server()