未验证 提交 f0b52448 编写于 作者: S shangliang Xu 提交者: GitHub

[TIPC] add tinypose_KL txt, test=document_fix (#6384)

上级 23e9cb95
===========================cpp_infer_params===========================
model_name:tinypose_128x96_KL
python:python3.7
filename:null
##
--output_dir:./output_inference
weights:https://paddledet.bj.bcebos.com/models/keypoint/tinypose_128x96.pdparams
norm_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml -o
quant_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_pact -o
fpgm_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_fpgm -o
distill_export:null
export1:null
export2:null
kl_quant_export:tools/post_quant.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_kl_quant -o
##
opencv_dir:default
infer_mode:null
infer_quant:True
inference:./deploy/cpp/build/main
--device:gpu|cpu
--use_mkldnn:False
--cpu_threads:4
--batch_size:1
--use_tensorrt:null
--run_mode:paddle
--model_dir_keypoint:
--image_dir:./dataset/coco/test2017/
--run_benchmark:False
--model_dir:./output_inference/picodet_s_320_pedestrian
\ No newline at end of file
===========================serving_infer_cpp_params===========================
model_name:tinypose_128x96_KL
python:python3.7
filename:null
##
--output_dir:./output_inference
weights:https://paddledet.bj.bcebos.com/models/keypoint/tinypose_128x96.pdparams
norm_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --export_serving_model True -o
quant_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_pact --export_serving_model True -o
fpgm_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_fpgm --export_serving_model True -o
distill_export:null
export1:null
export2:null
kl_quant_export:tools/post_quant.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config configs/slim/post_quant/tinypose_128x96_ptq.yml --export_serving_model True -o
##
infer_mode:null
infer_quant:True
--model:null
--op:tinypose_128x96
--port:9997
--gpu_ids:null|0
null:null
http_client:deploy/serving/cpp/serving_client.py
--serving_client:null
--image_file:./demo/hrnet_demo.jpg
null:null
\ No newline at end of file
===========================serving_infer_python_params===========================
model_name:tinypose_128x96_KL
python:python3.7
filename:null
##
--output_dir:./output_inference
weights:https://paddledet.bj.bcebos.com/models/keypoint/tinypose_128x96.pdparams
norm_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --export_serving_model True -o
quant_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_pact --export_serving_model True -o
fpgm_export:tools/export_model.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config _template_fpgm --export_serving_model True -o
distill_export:null
export1:null
export2:null
kl_quant_export:tools/post_quant.py -c test_tipc/configs/keypoint/tinypose_128x96.yml --slim_config configs/slim/post_quant/tinypose_128x96_ptq.yml --export_serving_model True -o
##
infer_mode:null
infer_quant:True
web_service:deploy/serving/python/web_service.py --config=deploy/serving/python/config.yml
--model_dir:null
--opt:cpu:op.ppdet.local_service_conf.device_type=0|gpu:op.ppdet.local_service_conf.device_type=1
null:null
http_client:deploy/serving/python/pipeline_http_client.py
--image_file:./demo/hrnet_demo.jpg
null:null
\ No newline at end of file
......@@ -58,6 +58,10 @@ elif [ ${MODE} = "cpp_infer" ];then
wget -nc -P ./output_inference/mask_rcnn_r50_fpn_1x_coco_KL/ https://bj.bcebos.com/v1/paddledet/data/tipc/models/mask_rcnn_r50_fpn_1x_coco_ptq.tar --no-check-certificate
cd ./output_inference/mask_rcnn_r50_fpn_1x_coco_KL/ && tar -xvf mask_rcnn_r50_fpn_1x_coco_ptq.tar && mv -n mask_rcnn_r50_fpn_1x_coco_ptq/* .
cd ../../
elif [[ ${model_name} = "tinypose_128x96_KL" ]]; then
wget -nc -P ./output_inference/tinypose_128x96_KL/ https://bj.bcebos.com/v1/paddledet/data/tipc/models/tinypose_128x96_ptq.tar --no-check-certificate
cd ./output_inference/tinypose_128x96_KL/ && tar -xvf tinypose_128x96_ptq.tar && mv -n tinypose_128x96_ptq/* .
cd ../../
fi
# download mot lite data
wget -nc -P ./dataset/mot/ https://paddledet.bj.bcebos.com/data/tipc/mot_tipc.tar --no-check-certificate
......@@ -124,6 +128,11 @@ elif [ ${MODE} = "serving_infer" ];then
cd ./output_inference/mask_rcnn_r50_fpn_1x_coco_KL/ && tar -xvf mask_rcnn_r50_fpn_1x_coco_ptq.tar && mv -n mask_rcnn_r50_fpn_1x_coco_ptq/* .
cd ../../
eval "${python} -m paddle_serving_client.convert --dirname output_inference/mask_rcnn_r50_fpn_1x_coco_KL/ --model_filename model.pdmodel --params_filename model.pdiparams --serving_server output_inference/mask_rcnn_r50_fpn_1x_coco_KL/serving_server --serving_client output_inference/mask_rcnn_r50_fpn_1x_coco_KL/serving_client"
elif [[ ${model_name} = "tinypose_128x96_KL" ]]; then
wget -nc -P ./output_inference/tinypose_128x96_KL/ https://bj.bcebos.com/v1/paddledet/data/tipc/models/tinypose_128x96_ptq.tar --no-check-certificate
cd ./output_inference/tinypose_128x96_KL/ && tar -xvf tinypose_128x96_ptq.tar && mv -n tinypose_128x96_ptq/* .
cd ../../
eval "${python} -m paddle_serving_client.convert --dirname output_inference/tinypose_128x96_KL/ --model_filename model.pdmodel --params_filename model.pdiparams --serving_server output_inference/tinypose_128x96_KL/serving_server --serving_client output_inference/tinypose_128x96_KL/serving_client"
fi
else
# download coco lite data
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册