提交 0aa84521 编写于 作者: A andyjpaddle

Merge branch 'dygraph' of https://github.com/PaddlePaddle/PaddleOCR into dygraph

......@@ -136,7 +136,7 @@ The recognition model is the same.
2. Run the following command to start the service.
```
# Start the service and save the running log in log.txt
python3 web_service.py &>log.txt &
python3 web_service.py --config=config.yml &>log.txt &
```
After the service is successfully started, a log similar to the following will be printed in log.txt
![](./imgs/start_server.png)
......
......@@ -135,7 +135,7 @@ python3 -m paddle_serving_client.convert --dirname ./ch_PP-OCRv3_rec_infer/ \
2. 启动服务可运行如下命令:
```
# 启动服务,运行日志保存在log.txt
python3 web_service.py &>log.txt &
python3 web_service.py --config=config.yml &>log.txt &
```
成功启动服务后,log.txt中会打印类似如下日志
![](./imgs/start_server.png)
......
......@@ -19,7 +19,7 @@ import copy
import cv2
import base64
# from paddle_serving_app.reader import OCRReader
from ocr_reader import OCRReader, DetResizeForTest
from ocr_reader import OCRReader, DetResizeForTest, ArgsParser
from paddle_serving_app.reader import Sequential, ResizeByFactor
from paddle_serving_app.reader import Div, Normalize, Transpose
from paddle_serving_app.reader import DBPostProcess, FilterBoxes, GetRotateCropImage, SortedBoxes
......@@ -85,7 +85,7 @@ class RecOp(Op):
dt_boxes = copy.deepcopy(self.dt_list)
feed_list = []
img_list = []
max_wh_ratio = 320/48.
max_wh_ratio = 320 / 48.
## Many mini-batchs, the type of feed_data is list.
max_batch_size = 6 # len(dt_boxes)
......@@ -163,5 +163,6 @@ class OcrService(WebService):
uci_service = OcrService(name="ocr")
uci_service.prepare_pipeline_config("config.yml")
FLAGS = ArgsParser().parse_args()
uci_service.prepare_pipeline_config(yml_dict=FLAGS.conf_dict)
uci_service.run_service()
......@@ -86,8 +86,6 @@ function func_serving(){
python_list=(${python_list})
cd ${serving_dir_value}
# cpp serving
unset https_proxy
unset http_proxy
for gpu_id in ${gpu_value[*]}; do
if [ ${gpu_id} = "null" ]; then
web_service_cpp_cmd="${python_list[0]} ${web_service_py} --model ${det_server_value} ${rec_server_value} ${op_key} ${op_value} ${port_key} ${port_value} > serving_log_cpu.log &"
......
......@@ -107,8 +107,6 @@ function func_serving(){
python_list=(${python_list})
cd ${serving_dir_value}
unset https_proxy
unset http_proxy
python=${python_list[0]}
# python serving
......@@ -147,7 +145,7 @@ function func_serving(){
status_check $last_status "${pipeline_cmd}" "${status_log}" "${model_name}"
sleep 2s
done
ps ux | grep -E 'web_service|pipeline' | awk '{print $2}' | xargs kill -s 9
ps ux | grep -E 'web_service' | awk '{print $2}' | xargs kill -s 9
done
done
elif [ ${use_gpu} = "gpu" ]; then
......@@ -197,7 +195,7 @@ function func_serving(){
status_check $last_status "${pipeline_cmd}" "${status_log}" "${model_name}"
sleep 2s
done
ps ux | grep -E 'web_service|pipeline' | awk '{print $2}' | xargs kill -s 9
ps ux | grep -E 'web_service' | awk '{print $2}' | xargs kill -s 9
done
done
else
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册