From ea83a067caa18e67aac76ab5330ecfc9b08658f3 Mon Sep 17 00:00:00 2001 From: LDOUBLEV Date: Tue, 30 Nov 2021 11:03:47 +0000 Subject: [PATCH] fix bug --- ...rmal_normal_infer_python_linux_gpu_cpu.txt | 2 +- ...rmal_normal_infer_python_linux_gpu_cpu.txt | 4 +- ...rmal_normal_infer_python_linux_gpu_cpu.txt | 2 +- test_tipc/test_inference_python.sh | 65 ++++++++----------- 4 files changed, 31 insertions(+), 42 deletions(-) diff --git a/test_tipc/configs/ch_PP-OCRv2/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt b/test_tipc/configs/ch_PP-OCRv2/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt index 243b2f15..92ba196f 100644 --- a/test_tipc/configs/ch_PP-OCRv2/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt +++ b/test_tipc/configs/ch_PP-OCRv2/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt @@ -1,4 +1,4 @@ -===========================ch_ppocr_mobile_v2.0=========================== +===========================ch_PP-OCRv2=========================== model_name:ch_PP-OCRv2 python:python3.7 null:null diff --git a/test_tipc/configs/ch_ppocr_mobile_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt b/test_tipc/configs/ch_ppocr_mobile_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt index c6aa0fa3..abe5e80e 100644 --- a/test_tipc/configs/ch_ppocr_mobile_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt +++ b/test_tipc/configs/ch_ppocr_mobile_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt @@ -15,7 +15,7 @@ inference:tools/infer/predict_system.py --precision:fp32|fp16 --det_model_dir: --image_dir:./inference/ch_det_data_50/all-sum-510/ -null:null ---benchmark:True --rec_model_dir:./inference/ch_ppocr_mobile_v2.0_rec_infer/ +--benchmark:True +null:null null:null diff --git a/test_tipc/configs/ch_ppocr_server_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt b/test_tipc/configs/ch_ppocr_server_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt index 2a968da0..5493e515 100644 --- a/test_tipc/configs/ch_ppocr_server_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt +++ b/test_tipc/configs/ch_ppocr_server_v2.0/model_linux_gpu_normal_normal_infer_python_linux_gpu_cpu.txt @@ -1,4 +1,4 @@ -===========================ch_ppocr_mobile_v2.0=========================== +===========================ch_ppocr_server_v2.0=========================== model_name:ch_ppocr_server_v2.0 python:python3.7 null:null diff --git a/test_tipc/test_inference_python.sh b/test_tipc/test_inference_python.sh index 6a6cb015..4f45a43f 100644 --- a/test_tipc/test_inference_python.sh +++ b/test_tipc/test_inference_python.sh @@ -6,7 +6,7 @@ FILENAME=$1 # MODE be one of ['whole_infer'] MODE=$2 -dataline=$(awk 'NR==1, NR==17{print}' $FILENAME) +dataline=$(awk 'NR==1, NR==20{print}' $FILENAME) # parser params IFS=$'\n' @@ -15,44 +15,40 @@ lines=(${dataline}) model_name=$(func_parser_value "${lines[1]}") python=$(func_parser_value "${lines[2]}") - -infer_model_dir_list=$(func_parser_value "${lines[3]}") -infer_export_list=$(func_parser_value "${lines[4]}") -infer_is_quant=$(func_parser_value "${lines[5]}") +infer_model_dir_list=$(func_parser_value "${lines[5]}") +infer_export_list=$(func_parser_value "${lines[6]}") +infer_is_quant=$(func_parser_value "${lines[7]}") # parser inference -inference_py=$(func_parser_value "${lines[6]}") -use_gpu_key=$(func_parser_key "${lines[7]}") -use_gpu_list=$(func_parser_value "${lines[7]}") -use_mkldnn_key=$(func_parser_key "${lines[8]}") -use_mkldnn_list=$(func_parser_value "${lines[8]}") -cpu_threads_key=$(func_parser_key "${lines[9]}") -cpu_threads_list=$(func_parser_value "${lines[9]}") -batch_size_key=$(func_parser_key "${lines[10]}") -batch_size_list=$(func_parser_value "${lines[10]}") -use_trt_key=$(func_parser_key "${lines[11]}") -use_trt_list=$(func_parser_value "${lines[11]}") -precision_key=$(func_parser_key "${lines[12]}") -precision_list=$(func_parser_value "${lines[12]}") -infer_model_key=$(func_parser_key "${lines[13]}") -image_dir_key=$(func_parser_key "${lines[14]}") -infer_img_dir=$(func_parser_value "${lines[14]}") -rec_model_key=$(func_parser_key "${lines[15]}") -rec_model_value=$(func_parser_value "${lines[15]}") -benchmark_key=$(func_parser_key "${lines[16]}") -benchmark_value=$(func_parser_value "${lines[16]}") -infer_key1=$(func_parser_key "${lines[17]}") -infer_value1=$(func_parser_value "${lines[17]}") +inference_py=$(func_parser_value "${lines[8]}") +use_gpu_key=$(func_parser_key "${lines[9]}") +use_gpu_list=$(func_parser_value "${lines[9]}") +use_mkldnn_key=$(func_parser_key "${lines[10]}") +use_mkldnn_list=$(func_parser_value "${lines[10]}") +cpu_threads_key=$(func_parser_key "${lines[11]}") +cpu_threads_list=$(func_parser_value "${lines[11]}") +batch_size_key=$(func_parser_key "${lines[12]}") +batch_size_list=$(func_parser_value "${lines[12]}") +use_trt_key=$(func_parser_key "${lines[13]}") +use_trt_list=$(func_parser_value "${lines[13]}") +precision_key=$(func_parser_key "${lines[14]}") +precision_list=$(func_parser_value "${lines[14]}") +infer_model_key=$(func_parser_key "${lines[15]}") +image_dir_key=$(func_parser_key "${lines[16]}") +infer_img_dir=$(func_parser_value "${lines[16]}") +rec_model_key=$(func_parser_key "${lines[17]}") +rec_model_value=$(func_parser_value "${lines[17]}") +benchmark_key=$(func_parser_key "${lines[18]}") +benchmark_value=$(func_parser_value "${lines[18]}") +infer_key1=$(func_parser_key "${lines[19]}") +infer_value1=$(func_parser_value "${lines[19]}") + -echo "DEBUG" -echo "${lines[16]}" -echo "line17" ${lines[17]} LOG_PATH="./test_tipc/output" mkdir -p ${LOG_PATH} status_log="${LOG_PATH}/results_python.log" - function func_inference(){ IFS='|' _python=$1 @@ -61,7 +57,6 @@ function func_inference(){ _log_path=$4 _img_dir=$5 _flag_quant=$6 - echo "FUCNTION " # inference for use_gpu in ${use_gpu_list[*]}; do if [ ${use_gpu} = "False" ] || [ ${use_gpu} = "cpu" ]; then @@ -89,7 +84,6 @@ function func_inference(){ set_infer_params0=$(func_set_params "${rec_model_key}" "${rec_model_value}") set_infer_params1=$(func_set_params "${infer_key1}" "${infer_value1}") command="${_python} ${_script} ${use_gpu_key}=${use_gpu} ${use_mkldnn_key}=${use_mkldnn} ${set_cpu_threads} ${set_model_dir} ${set_batchsize} ${set_infer_params0} ${set_infer_data} ${set_benchmark} ${set_precision} ${set_infer_params1} > ${_save_log_path} 2>&1 " - echo "before cpu command" echo $command eval $command last_status=${PIPESTATUS[0]} @@ -122,7 +116,6 @@ function func_inference(){ set_infer_params0=$(func_set_params "${save_log_key}" "${save_log_value}") set_infer_params1=$(func_set_params "${infer_key1}" "${infer_value1}") command="${_python} ${_script} ${use_gpu_key}=${use_gpu} ${set_tensorrt} ${set_precision} ${set_model_dir} ${set_batchsize} ${set_infer_data} ${set_benchmark} ${set_infer_params1} ${set_infer_params0} > ${_save_log_path} 2>&1 " - echo "before gpu command" eval $command last_status=${PIPESTATUS[0]} eval "cat ${_save_log_path}" @@ -150,9 +143,7 @@ if [ ${MODE} = "whole_infer" ]; then IFS="|" infer_run_exports=(${infer_export_list}) infer_quant_flag=(${infer_is_quant}) - echo "DEBUG2" for infer_model in ${infer_model_dir_list[*]}; do - echo "DEBUG#" # run export if [ ${infer_run_exports[Count]} != "null" ];then save_infer_dir=$(dirname $infer_model) @@ -160,14 +151,12 @@ if [ ${MODE} = "whole_infer" ]; then set_save_infer_key=$(func_set_params "${save_infer_key}" "${save_infer_dir}") export_cmd="${python} ${infer_run_exports[Count]} ${set_export_weight} ${set_save_infer_key}" echo ${infer_run_exports[Count]} - echo $export_cmd eval $export_cmd status_export=$? status_check $status_export "${export_cmd}" "${status_log}" else save_infer_dir=${infer_model} fi - echo "FUCK@" #run inference is_quant=${infer_quant_flag[Count]} if [ ${MODE} = "klquant_infer" ]; then -- GitLab