未验证 提交 db1b2654 编写于 作者: S shangliang Xu 提交者: GitHub

[TIPC] fix serving infer shell (#6206)

上级 ce38daed
......@@ -80,14 +80,6 @@ elif [ ${MODE} = "paddle2onnx_infer" ];then
${python} -m pip install install paddle2onnx
${python} -m pip install onnxruntime==1.10.0
elif [ ${MODE} = "serving_infer" ];then
git clone https://github.com/PaddlePaddle/Serving
cd Serving
bash tools/paddle_env_install.sh
${python} -m pip install -r python/requirements.txt
cd ..
${python} -m pip install paddle-serving-client -i https://pypi.tuna.tsinghua.edu.cn/simple
${python} -m pip install paddle-serving-app -i https://pypi.tuna.tsinghua.edu.cn/simple
${python} -m pip install paddle-serving-server-gpu -i https://pypi.tuna.tsinghua.edu.cn/simple
unset https_proxy http_proxy
else
# download coco lite data
......
......@@ -89,10 +89,6 @@ function func_serving_inference(){
done
}
# build paddle_serving_server
bash deploy/serving/cpp/build_server.sh
echo "################### build finished! ###################"
# run serving infer
Count=0
IFS="|"
......
......@@ -81,9 +81,9 @@ function func_serving_inference(){
}
# set cuda device
GPUID=$2
GPUID=$3
if [ ${#GPUID} -le 0 ];then
env=" "
env="export CUDA_VISIBLE_DEVICES=0"
else
env="export CUDA_VISIBLE_DEVICES=${GPUID}"
fi
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册