提交 ad707bde 编写于 作者: H HydrogenSulfate

slim fleet_train config and fix paddle2onnx for GeneralRecognition model

上级 2e56bec7
===========================paddle2onnx_params=========================== ===========================paddle2onnx_params===========================
model_name:PP-ShiTu_general_rec model_name:GeneralRecognition_PPLCNet_x2_5
python:python3.7 python:python3.7
2onnx: paddle2onnx 2onnx: paddle2onnx
--model_dir:./deploy/models/general_PPLCNet_x2_5_lite_v1.0_infer/ --model_dir:./deploy/models/general_PPLCNet_x2_5_lite_v1.0_infer/
...@@ -9,8 +9,8 @@ python:python3.7 ...@@ -9,8 +9,8 @@ python:python3.7
--opset_version:10 --opset_version:10
--enable_onnx_checker:True --enable_onnx_checker:True
inference_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/rec/models/inference/general_PPLCNet_x2_5_lite_v1.0_infer.tar inference_model_url:https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/rec/models/inference/general_PPLCNet_x2_5_lite_v1.0_infer.tar
inference:./python/predict_cls.py inference:./python/predict_rec.py
Global.use_onnx:True Global.use_onnx:True
Global.inference_model_dir:./models/general_PPLCNet_x2_5_lite_v1.0_infer Global.rec_inference_model_dir:./models/general_PPLCNet_x2_5_lite_v1.0_infer
Global.use_gpu:False Global.use_gpu:False
-c:configs/inference_cls.yaml -c:configs/inference_rec.yaml
\ No newline at end of file \ No newline at end of file
...@@ -85,12 +85,18 @@ if [[ ${MODE} = "cpp_infer" ]]; then ...@@ -85,12 +85,18 @@ if [[ ${MODE} = "cpp_infer" ]]; then
if [[ ! -d "./deploy/cpp/paddle_inference/" ]]; then if [[ ! -d "./deploy/cpp/paddle_inference/" ]]; then
pushd ./deploy/cpp/ pushd ./deploy/cpp/
PADDLEInfer=$3 PADDLEInfer=$3
if [ "" = "$PADDLEInfer" ];then if [ "" = "$PADDLEInfer" ]; then
wget -nc https://paddle-inference-lib.bj.bcebos.com/2.2.2/cxx_c/Linux/GPU/x86-64_gcc8.2_avx_mkl_cuda10.1_cudnn7.6.5_trt6.0.1.5/paddle_inference.tgz --no-check-certificate wget -nc https://paddle-inference-lib.bj.bcebos.com/2.2.2/cxx_c/Linux/GPU/x86-64_gcc8.2_avx_mkl_cuda10.1_cudnn7.6.5_trt6.0.1.5/paddle_inference.tgz --no-check-certificate
tar xf paddle_inference.tgz
else else
wget -nc ${PADDLEInfer} --no-check-certificate wget -nc ${PADDLEInfer} --no-check-certificate
tar_name=$(func_get_url_file_name "$PADDLEInfer")
tar xf ${tar_name}
paddle_inference_install_dir=${tar_name%.*}
if [ ! -d "paddle_inference" ]; then
ln -s ${paddle_inference_install_dir} paddle_inference
fi
fi fi
tar xf paddle_inference.tgz
popd popd
fi fi
if [[ $FILENAME == *infer_cpp_linux_gpu_cpu.txt ]]; then if [[ $FILENAME == *infer_cpp_linux_gpu_cpu.txt ]]; then
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册