未验证 提交 333d3814 编写于 作者: B Bin Long 提交者: GitHub

Merge pull request #271 from ShenYuhan/bert_as_service

fix bs doc, fix serving port set
...@@ -68,7 +68,7 @@ $ pip install ujson ...@@ -68,7 +68,7 @@ $ pip install ujson
|模型|网络| |模型|网络|
|:-|:-:| |:-|:-:|
|[ERNIE](https://paddlepaddle.org.cn/hubdetail?name=ERNIE&en_category=SemanticModel)|ERNIE| |[ernie](https://paddlepaddle.org.cn/hubdetail?name=ERNIE&en_category=SemanticModel)|ERNIE|
|[ernie_tiny](https://paddlepaddle.org.cn/hubdetail?name=ernie_tiny&en_category=SemanticModel)|ERNIE| |[ernie_tiny](https://paddlepaddle.org.cn/hubdetail?name=ernie_tiny&en_category=SemanticModel)|ERNIE|
|[ernie_v2_eng_large](https://paddlepaddle.org.cn/hubdetail?name=ernie_v2_eng_large&en_category=SemanticModel)|ERNIE| |[ernie_v2_eng_large](https://paddlepaddle.org.cn/hubdetail?name=ernie_v2_eng_large&en_category=SemanticModel)|ERNIE|
|[ernie_v2_eng_base](https://paddlepaddle.org.cn/hubdetail?name=ernie_v2_eng_base&en_category=SemanticModel)|ERNIE| |[ernie_v2_eng_base](https://paddlepaddle.org.cn/hubdetail?name=ernie_v2_eng_base&en_category=SemanticModel)|ERNIE|
...@@ -225,8 +225,8 @@ Paddle Inference Server exit successfully! ...@@ -225,8 +225,8 @@ Paddle Inference Server exit successfully!
> Q : 如何在一台服务器部署多个模型? > Q : 如何在一台服务器部署多个模型?
> A : 可通过多次启动`Bert Service`,分配不同端口实现。如果使用GPU,需要指定不同的显卡。如同时部署`ernie`和`bert_chinese_L-12_H-768_A-12`,分别执行命令如下: > A : 可通过多次启动`Bert Service`,分配不同端口实现。如果使用GPU,需要指定不同的显卡。如同时部署`ernie`和`bert_chinese_L-12_H-768_A-12`,分别执行命令如下:
> ```shell > ```shell
> $ hub serving start bert_serving -m ernie -p 8866 > $ hub serving start bert_service -m ernie -p 8866
> $ hub serving start bert_serving -m bert_chinese_L-12_H-768_A-12 -p 8867 > $ hub serving start bert_service -m bert_chinese_L-12_H-768_A-12 -p 8867
> ``` > ```
> Q : 启动时显示"Check out http://yq01-gpu-255-129-12-00.epc.baidu.com:8887 in web > Q : 启动时显示"Check out http://yq01-gpu-255-129-12-00.epc.baidu.com:8887 in web
......
...@@ -159,7 +159,7 @@ class ServingCommand(BaseCommand): ...@@ -159,7 +159,7 @@ class ServingCommand(BaseCommand):
module = args.modules module = args.modules
if module is not None: if module is not None:
use_gpu = args.use_gpu use_gpu = args.use_gpu
port = args.port[0] port = args.port
if ServingCommand.is_port_occupied("127.0.0.1", port) is True: if ServingCommand.is_port_occupied("127.0.0.1", port) is True:
print("Port %s is occupied, please change it." % (port)) print("Port %s is occupied, please change it." % (port))
return False return False
...@@ -206,10 +206,12 @@ class ServingCommand(BaseCommand): ...@@ -206,10 +206,12 @@ class ServingCommand(BaseCommand):
if args.sub_command == "start": if args.sub_command == "start":
if args.bert_service == "bert_service": if args.bert_service == "bert_service":
ServingCommand.start_bert_serving(args) ServingCommand.start_bert_serving(args)
else: elif args.bert_service is None:
ServingCommand.start_serving(args) ServingCommand.start_serving(args)
else: else:
ServingCommand.show_help() ServingCommand.show_help()
else:
ServingCommand.show_help()
command = ServingCommand.instance() command = ServingCommand.instance()
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册