提交 0cea5504 编写于 作者: G gaotingquan

tipc: fix log path

上级 2700beea
......@@ -323,8 +323,8 @@ if [[ ${MODE} = "paddle2onnx_infer" ]]; then
# prepare paddle2onnx env
python_name=$(func_parser_value "${lines[2]}")
inference_model_url=$(func_parser_value "${lines[10]}")
tar_name=${inference_model_url##*/}
tar_name=$(func_get_url_file_name "$inference_model_url")
${python_name} -m pip install onnx
${python_name} -m pip install paddle2onnx
${python_name} -m pip install onnxruntime
......
......@@ -2,7 +2,7 @@
source test_tipc/common_func.sh
FILENAME=$1
MODE=$2
MODE="paddle2onnx_infer"
# parser params
dataline=$(awk 'NR==1, NR==16{print}' $FILENAME)
......@@ -61,7 +61,7 @@ function func_paddle2onnx(){
# python inference
if [[ ${inference_py} != "null" ]]; then
_save_log_path=".${LOG_PATH}/paddle2onnx_infer_cpu.log"
_save_log_path="${LOG_PATH}/paddle2onnx_infer_cpu.log"
set_model_dir=$(func_set_params "${inference_model_dir_key}" "${inference_model_dir_value}")
set_use_onnx=$(func_set_params "${use_onnx_key}" "${use_onnx_value}")
set_hardware=$(func_set_params "${inference_hardware_key}" "${inference_hardware_value}")
......
......@@ -39,7 +39,7 @@ pipeline_py=$(func_parser_value "${lines[13]}")
function func_serving_cls(){
CLS_ROOT_PATH=$(pwd)
LOG_PATH="${CLS_ROOT_PATH}/test_tipc/output/${model_name}/${MODE}/cpp"
LOG_PATH="${CLS_ROOT_PATH}/test_tipc/output/${model_name}/serving_infer"
mkdir -p ${LOG_PATH}
status_log="${LOG_PATH}/results_cpp_serving.log"
IFS='|'
......@@ -142,7 +142,7 @@ function func_serving_cls(){
function func_serving_rec(){
CLS_ROOT_PATH=$(pwd)
LOG_PATH="${CLS_ROOT_PATH}/test_tipc/output/${model_name}"
LOG_PATH="${CLS_ROOT_PATH}/test_tipc/output/${model_name}/serving_infer"
mkdir -p ${LOG_PATH}
status_log="${LOG_PATH}/results_cpp_serving.log"
......
......@@ -58,7 +58,7 @@ function func_serving_cls(){
for python_ in ${python[*]}; do
if [[ ${python_} =~ "python" ]]; then
trans_log="${LOG_PATH}/python_trans_model.log"
trans_model_cmd="${python_} ${trans_model_py} ${set_dirname} ${set_model_filename} ${set_params_filename} ${set_serving_server} ${set_serving_client} > ${trans_cls_log} 2>&1"
trans_model_cmd="${python_} ${trans_model_py} ${set_dirname} ${set_model_filename} ${set_params_filename} ${set_serving_server} ${set_serving_client} > ${trans_log} 2>&1"
eval ${trans_model_cmd}
last_status=${PIPESTATUS[0]}
status_check $last_status "${trans_model_cmd}" "${status_log}" "${model_name}" "${trans_log}"
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册