提交 1c54b6e4 编写于 作者: H HydrogenSulfate

debug

上级 25acd2ea
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/MobileNetV3_large_x1_0_infer/ --dirname:./deploy/paddleserving/MobileNetV3_large_x1_0_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/MobileNetV3_large_x1_0_server/ --serving_server:./deploy/paddleserving/MobileNetV3_large_x1_0_serving/
--serving_client:./deploy/paddleserving/MobileNetV3_large_x1_0_client/ --serving_client:./deploy/paddleserving/MobileNetV3_large_x1_0_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
===========================serving_params===========================
model_name:PPHGNet_small
python:python3.7
inference_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/inference/PPHGNet_small_infer.tar
trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPHGNet_small_infer/
--model_filename:inference.pdmodel
--params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPHGNet_small_serving/
--serving_client:./deploy/paddleserving/PPHGNet_small_client/
serving_dir:./deploy/paddleserving
web_service:classification_web_service.py
--use_gpu:0|null
pipline:pipeline_http_client.py
===========================serving_params===========================
model_name:PPHGNet_tiny
python:python3.7
inference_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/inference/PPHGNet_tiny_infer.tar
trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPHGNet_tiny_infer/
--model_filename:inference.pdmodel
--params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPHGNet_tiny_serving/
--serving_client:./deploy/paddleserving/PPHGNet_tiny_client/
serving_dir:./deploy/paddleserving
web_service:classification_web_service.py
--use_gpu:0|null
pipline:pipeline_http_client.py
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x0_25_infer/ --dirname:./deploy/paddleserving/PPLCNet_x0_25_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x0_25_server/ --serving_server:./deploy/paddleserving/PPLCNet_x0_25_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x0_25_client/ --serving_client:./deploy/paddleserving/PPLCNet_x0_25_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x0_35_infer/ --dirname:./deploy/paddleserving/PPLCNet_x0_35_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x0_35_server/ --serving_server:./deploy/paddleserving/PPLCNet_x0_35_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x0_35_client/ --serving_client:./deploy/paddleserving/PPLCNet_x0_35_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x0_5_infer/ --dirname:./deploy/paddleserving/PPLCNet_x0_5_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x0_5_server/ --serving_server:./deploy/paddleserving/PPLCNet_x0_5_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x0_5_client/ --serving_client:./deploy/paddleserving/PPLCNet_x0_5_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x0_75_infer/ --dirname:./deploy/paddleserving/PPLCNet_x0_75_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x0_75_server/ --serving_server:./deploy/paddleserving/PPLCNet_x0_75_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x0_75_client/ --serving_client:./deploy/paddleserving/PPLCNet_x0_75_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x1_0_infer/ --dirname:./deploy/paddleserving/PPLCNet_x1_0_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x1_0_server/ --serving_server:./deploy/paddleserving/PPLCNet_x1_0_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x1_0_client/ --serving_client:./deploy/paddleserving/PPLCNet_x1_0_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x1_5_infer/ --dirname:./deploy/paddleserving/PPLCNet_x1_5_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x1_5_server/ --serving_server:./deploy/paddleserving/PPLCNet_x1_5_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x1_5_client/ --serving_client:./deploy/paddleserving/PPLCNet_x1_5_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x2_0_infer/ --dirname:./deploy/paddleserving/PPLCNet_x2_0_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x2_0_server/ --serving_server:./deploy/paddleserving/PPLCNet_x2_0_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x2_0_client/ --serving_client:./deploy/paddleserving/PPLCNet_x2_0_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNet_x2_5_infer/ --dirname:./deploy/paddleserving/PPLCNet_x2_5_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNet_x2_5_server/ --serving_server:./deploy/paddleserving/PPLCNet_x2_5_serving/
--serving_client:./deploy/paddleserving/PPLCNet_x2_5_client/ --serving_client:./deploy/paddleserving/PPLCNet_x2_5_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/PPLCNetV2_base_infer/ --dirname:./deploy/paddleserving/PPLCNetV2_base_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/PPLCNetV2_base_server/ --serving_server:./deploy/paddleserving/PPLCNetV2_base_serving/
--serving_client:./deploy/paddleserving/PPLCNetV2_base_client/ --serving_client:./deploy/paddleserving/PPLCNetV2_base_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/ResNet50_infer/ --dirname:./deploy/paddleserving/ResNet50_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/ResNet50_server/ --serving_server:./deploy/paddleserving/ResNet50_serving/
--serving_client:./deploy/paddleserving/ResNet50_client/ --serving_client:./deploy/paddleserving/ResNet50_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/ResNet50_vd_infer/ --dirname:./deploy/paddleserving/ResNet50_vd_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/ResNet50_vd_server/ --serving_server:./deploy/paddleserving/ResNet50_vd_serving/
--serving_client:./deploy/paddleserving/ResNet50_vd_client/ --serving_client:./deploy/paddleserving/ResNet50_vd_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert ...@@ -6,7 +6,7 @@ trans_model:-m paddle_serving_client.convert
--dirname:./deploy/paddleserving/SwinTransformer_tiny_patch4_window7_224_infer/ --dirname:./deploy/paddleserving/SwinTransformer_tiny_patch4_window7_224_infer/
--model_filename:inference.pdmodel --model_filename:inference.pdmodel
--params_filename:inference.pdiparams --params_filename:inference.pdiparams
--serving_server:./deploy/paddleserving/SwinTransformer_tiny_patch4_window7_224_server/ --serving_server:./deploy/paddleserving/SwinTransformer_tiny_patch4_window7_224_serving/
--serving_client:./deploy/paddleserving/SwinTransformer_tiny_patch4_window7_224_client/ --serving_client:./deploy/paddleserving/SwinTransformer_tiny_patch4_window7_224_client/
serving_dir:./deploy/paddleserving serving_dir:./deploy/paddleserving
web_service:classification_web_service.py web_service:classification_web_service.py
......
...@@ -13,15 +13,15 @@ Linux GPU/CPU PYTHON 服务化部署测试的主程序为`test_serving_infer.sh ...@@ -13,15 +13,15 @@ Linux GPU/CPU PYTHON 服务化部署测试的主程序为`test_serving_infer.sh
| PP-ShiTu | PPShiTu_general_rec、PPShiTu_mainbody_det | 支持 | 支持 | | PP-ShiTu | PPShiTu_general_rec、PPShiTu_mainbody_det | 支持 | 支持 |
| PPHGNet | PPHGNet_small | 支持 | 支持 | | PPHGNet | PPHGNet_small | 支持 | 支持 |
| PPHGNet | PPHGNet_tiny | 支持 | 支持 | | PPHGNet | PPHGNet_tiny | 支持 | 支持 |
| PPLCNet | PPLCNet_x0_25 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x0_25 | 支持 | 支持 |
| PPLCNet | PPLCNet_x0_35 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x0_35 | 支持 | 支持 |
| PPLCNet | PPLCNet_x0_5 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x0_5 | 支持 | 支持 |
| PPLCNet | PPLCNet_x0_75 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x0_75 | 支持 | 支持 |
| PPLCNet | PPLCNet_x1_0 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x1_0 | 支持 | 支持 |
| PPLCNet | PPLCNet_x1_5 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x1_5 | 支持 | 支持 |
| PPLCNet | PPLCNet_x2_0 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x2_0 | 支持 | 支持 |
| PPLCNet | PPLCNet_x2_5 | 不支持 | 不支持 | | PPLCNet | PPLCNet_x2_5 | 支持 | 支持 |
| PPLCNetV2 | PPLCNetV2_base | 支持 | 支持 | | PPLCNetV2 | PPLCNetV2_base | 支持 | 支持 |
| ResNet | ResNet50 | 支持 | 支持 | | ResNet | ResNet50 | 支持 | 支持 |
| ResNet | ResNet50_vd | 支持 | 支持 | | ResNet | ResNet50_vd | 支持 | 支持 |
| SwinTransformer | SwinTransformer_tiny_patch4_window7_224 | 支持 | 支持 | | SwinTransformer | SwinTransformer_tiny_patch4_window7_224 | 支持 | 支持 |
...@@ -50,7 +50,7 @@ Linux GPU/CPU PYTHON 服务化部署测试的主程序为`test_serving_infer.sh ...@@ -50,7 +50,7 @@ Linux GPU/CPU PYTHON 服务化部署测试的主程序为`test_serving_infer.sh
```shell ```shell
python3.7 -m pip install -r requirements.txt python3.7 -m pip install -r requirements.txt
``` ```
- 安装 PaddleServing 相关组件,包括serving-server、serving_client、serving-app - 安装 PaddleServing 相关组件,包括serving-server、serving_client、serving-app,自动下载并解压推理模型
```bash ```bash
bash test_tipc/prepare.sh test_tipc/configs/ResNet50/ResNet50_linux_gpu_normal_normal_serving_python_linux_gpu_cpu.txt serving_infer bash test_tipc/prepare.sh test_tipc/configs/ResNet50/ResNet50_linux_gpu_normal_normal_serving_python_linux_gpu_cpu.txt serving_infer
``` ```
......
...@@ -165,24 +165,26 @@ if [ ${MODE} = "serving_infer" ];then ...@@ -165,24 +165,26 @@ if [ ${MODE} = "serving_infer" ];then
${python_name} -m pip install install paddle-serving-server-gpu==0.7.0.post102 ${python_name} -m pip install install paddle-serving-server-gpu==0.7.0.post102
${python_name} -m pip install paddle_serving_client==0.7.0 ${python_name} -m pip install paddle_serving_client==0.7.0
${python_name} -m pip install paddle-serving-app==0.7.0 ${python_name} -m pip install paddle-serving-app==0.7.0
cls_inference_model_url=$(func_parser_value "${lines[3]}") if [[ ${model_name} =~ "ShiTu" ]]; then
cls_tar_name=$(func_get_url_file_name "${cls_inference_model_url}") cls_inference_model_url=$(func_parser_value "${lines[3]}")
det_inference_model_url=$(func_parser_value "${lines[4]}") cls_tar_name=$(func_get_url_file_name "${cls_inference_model_url}")
det_tar_name=$(func_get_url_file_name "${det_inference_model_url}") det_inference_model_url=$(func_parser_value "${lines[4]}")
unset http_proxy det_tar_name=$(func_get_url_file_name "${det_inference_model_url}")
unset https_proxy
if [[ ${det_inference_model_url} -eq null ]]; then
cd ./deploy/paddleserving
wget -nc ${cls_inference_model_url} && tar xf ${cls_tar_name}
cd ../../
else
cd ./deploy cd ./deploy
mkdir models mkdir models
cd models cd models
wget -nc ${cls_inference_model_url} && tar xf ${cls_tar_name} wget -nc ${cls_inference_model_url} && tar xf ${cls_tar_name}
wget -nc ${det_inference_model_url} && tar xf ${det_tar_name} wget -nc ${det_inference_model_url} && tar xf ${det_tar_name}
cd .. cd ..
else
cls_inference_model_url=$(func_parser_value "${lines[3]}")
cls_tar_name=$(func_get_url_file_name "${cls_inference_model_url}")
cd ./deploy/paddleserving
wget -nc ${cls_inference_model_url} && tar xf ${cls_tar_name}
cd ../../
fi fi
unset http_proxy
unset https_proxy
fi fi
if [ ${MODE} = "paddle2onnx_infer" ];then if [ ${MODE} = "paddle2onnx_infer" ];then
......
...@@ -54,15 +54,25 @@ function func_serving_cls(){ ...@@ -54,15 +54,25 @@ function func_serving_cls(){
eval $trans_model_cmd eval $trans_model_cmd
# modify the alias_name of fetch_var to "outputs" # modify the alias_name of fetch_var to "outputs"
server_fetch_var_line_cmd="sed -i '/fetch_var/,/is_lod_tensor/s/alias_name: .*/alias_name: \"prediction\"/' $serving_server_value/serving_server_conf.prototxt" server_fetch_var_line_cmd="sed -i '/fetch_var/,/is_lod_tensor/s/alias_name: .*/alias_name: \"prediction\"/' ${serving_server_value}/serving_server_conf.prototxt"
eval ${server_fetch_var_line_cmd} eval ${server_fetch_var_line_cmd}
client_fetch_var_line_cmd="sed -i '/fetch_var/,/is_lod_tensor/s/alias_name: .*/alias_name: \"prediction\"/' $serving_client_value/serving_client_conf.prototxt" client_fetch_var_line_cmd="sed -i '/fetch_var/,/is_lod_tensor/s/alias_name: .*/alias_name: \"prediction\"/' ${serving_client_value}/serving_client_conf.prototxt"
eval ${client_fetch_var_line_cmd} eval ${client_fetch_var_line_cmd}
prototxt_dataline=$(awk 'NR==1, NR==3{print}' ${serving_server_value}/serving_server_conf.prototxt)
IFS=$'\n'
prototxt_lines=(${prototxt_dataline})
feed_var_name=$(func_parser_value "${prototxt_lines[2]}")
IFS='|'
cd ${serving_dir_value} cd ${serving_dir_value}
unset https_proxy unset https_proxy
unset http_proxy unset http_proxy
# modify the input_name in "classification_web_service.py" to be consistent with feed_var.name in prototxt
set_web_service_feet_var_cmd="sed -i '/preprocess/,/input_imgs}/s/{.*: input_imgs}/{${feed_var_name}: input_imgs}/' ${web_service_py}"
eval ${set_web_service_feet_var_cmd}
model_config=21 model_config=21
serving_server_dir_name=$(func_get_url_file_name "$serving_server_value") serving_server_dir_name=$(func_get_url_file_name "$serving_server_value")
set_model_config_cmd="sed -i '${model_config}s/model_config: .*/model_config: ${serving_server_dir_name}/' config.yml" set_model_config_cmd="sed -i '${model_config}s/model_config: .*/model_config: ${serving_server_dir_name}/' config.yml"
...@@ -215,9 +225,20 @@ function func_serving_rec(){ ...@@ -215,9 +225,20 @@ function func_serving_rec(){
client_fetch_var_line_cmd="sed -i '/fetch_var/,/is_lod_tensor/s/alias_name: .*/alias_name: \"features\"/' $cls_serving_client_value/serving_client_conf.prototxt" client_fetch_var_line_cmd="sed -i '/fetch_var/,/is_lod_tensor/s/alias_name: .*/alias_name: \"features\"/' $cls_serving_client_value/serving_client_conf.prototxt"
eval ${client_fetch_var_line_cmd} eval ${client_fetch_var_line_cmd}
prototxt_dataline=$(awk 'NR==1, NR==3{print}' ${cls_serving_server_value}/serving_server_conf.prototxt)
IFS=$'\n'
prototxt_lines=(${prototxt_dataline})
feed_var_name=$(func_parser_value "${prototxt_lines[2]}")
IFS='|'
cd ${serving_dir_value} cd ${serving_dir_value}
unset https_proxy unset https_proxy
unset http_proxy unset http_proxy
# modify the input_name in "recognition_web_service.py" to be consistent with feed_var.name in prototxt
set_web_service_feet_var_cmd="sed -i '/preprocess/,/input_imgs}/s/{.*: input_imgs}/{${feed_var_name}: input_imgs}/' ${web_service_py}"
eval ${set_web_service_feet_var_cmd}
for python in ${python[*]}; do for python in ${python[*]}; do
if [[ ${python} = "cpp" ]]; then if [[ ${python} = "cpp" ]]; then
for use_gpu in ${web_use_gpu_list[*]}; do for use_gpu in ${web_use_gpu_list[*]}; do
...@@ -257,13 +278,11 @@ function func_serving_rec(){ ...@@ -257,13 +278,11 @@ function func_serving_rec(){
eval $set_devices_cmd eval $set_devices_cmd
web_service_cmd="${python} ${web_service_py} &" web_service_cmd="${python} ${web_service_py} &"
echo $PWD - $web_service_cmd
eval $web_service_cmd eval $web_service_cmd
sleep 5s sleep 5s
for pipeline in ${pipeline_py[*]}; do for pipeline in ${pipeline_py[*]}; do
_save_log_path="${LOG_PATH}/server_infer_cpu_${pipeline%_client*}_batchsize_1.log" _save_log_path="${LOG_PATH}/server_infer_cpu_${pipeline%_client*}_batchsize_1.log"
pipeline_cmd="${python} ${pipeline} > ${_save_log_path} 2>&1 " pipeline_cmd="${python} ${pipeline} > ${_save_log_path} 2>&1 "
echo $PWD - $pipeline_cmd
eval $pipeline_cmd eval $pipeline_cmd
last_status=${PIPESTATUS[0]} last_status=${PIPESTATUS[0]}
eval "cat ${_save_log_path}" eval "cat ${_save_log_path}"
...@@ -291,13 +310,11 @@ function func_serving_rec(){ ...@@ -291,13 +310,11 @@ function func_serving_rec(){
eval $set_devices_cmd eval $set_devices_cmd
web_service_cmd="${python} ${web_service_py} & " web_service_cmd="${python} ${web_service_py} & "
echo $PWD - $web_service_cmd
eval $web_service_cmd eval $web_service_cmd
sleep 10s sleep 10s
for pipeline in ${pipeline_py[*]}; do for pipeline in ${pipeline_py[*]}; do
_save_log_path="${LOG_PATH}/server_infer_gpu_${pipeline%_client*}_batchsize_1.log" _save_log_path="${LOG_PATH}/server_infer_gpu_${pipeline%_client*}_batchsize_1.log"
pipeline_cmd="${python} ${pipeline} > ${_save_log_path} 2>&1" pipeline_cmd="${python} ${pipeline} > ${_save_log_path} 2>&1"
echo $PWD - $pipeline_cmd
eval $pipeline_cmd eval $pipeline_cmd
last_status=${PIPESTATUS[0]} last_status=${PIPESTATUS[0]}
eval "cat ${_save_log_path}" eval "cat ${_save_log_path}"
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册