From a5873d380a04490be372b78aa079dfccb254af0e Mon Sep 17 00:00:00 2001 From: George Ni Date: Fri, 6 Aug 2021 14:27:04 +0800 Subject: [PATCH] [cherry-pick] fix use_gpu in deploy doc (#3905) * fix use_gpu in deploy doc, test=document_fix * fix deploy doc, test=document_fix --- configs/ppyolo/README.md | 2 +- configs/ppyolo/README_cn.md | 2 +- deploy/TENSOR_RT.md | 4 ++-- deploy/benchmark/benchmark.sh | 4 ++-- deploy/benchmark/benchmark_quant.sh | 2 +- deploy/python/README.md | 2 +- docs/tutorials/GETTING_STARTED_cn.md | 2 +- 7 files changed, 9 insertions(+), 9 deletions(-) diff --git a/configs/ppyolo/README.md b/configs/ppyolo/README.md index 01aa8b170..eca5216cd 100644 --- a/configs/ppyolo/README.md +++ b/configs/ppyolo/README.md @@ -182,7 +182,7 @@ For inference deployment or benchmard, model exported with `tools/export_model.p python tools/export_model.py -c configs/ppyolo/ppyolo_r50vd_dcn_1x_coco.yml -o weights=https://paddledet.bj.bcebos.com/models/ppyolo_r50vd_dcn_1x_coco.pdparams # inference with Paddle Inference library -CUDA_VISIBLE_DEVICES=0 python deploy/python/infer.py --model_dir=output_inference/ppyolo_r50vd_dcn_1x_coco --image_file=demo/000000014439_640x640.jpg --use_gpu=True +CUDA_VISIBLE_DEVICES=0 python deploy/python/infer.py --model_dir=output_inference/ppyolo_r50vd_dcn_1x_coco --image_file=demo/000000014439_640x640.jpg --device=GPU ``` diff --git a/configs/ppyolo/README_cn.md b/configs/ppyolo/README_cn.md index 78bce9756..9eafa8d19 100644 --- a/configs/ppyolo/README_cn.md +++ b/configs/ppyolo/README_cn.md @@ -175,7 +175,7 @@ PP-YOLO模型部署及推理benchmark需要通过`tools/export_model.py`导出 python tools/export_model.py -c configs/ppyolo/ppyolo_r50vd_dcn_1x_coco.yml -o weights=https://paddledet.bj.bcebos.com/models/ppyolo_r50vd_dcn_1x_coco.pdparams # 预测库推理 -CUDA_VISIBLE_DEVICES=0 python deploy/python/infer.py --model_dir=output_inference/ppyolo_r50vd_dcn_1x_coco --image_file=demo/000000014439_640x640.jpg --use_gpu=True +CUDA_VISIBLE_DEVICES=0 python deploy/python/infer.py --model_dir=output_inference/ppyolo_r50vd_dcn_1x_coco --image_file=demo/000000014439_640x640.jpg --device=GPU ``` diff --git a/deploy/TENSOR_RT.md b/deploy/TENSOR_RT.md index d41974327..a9d7677ee 100644 --- a/deploy/TENSOR_RT.md +++ b/deploy/TENSOR_RT.md @@ -53,7 +53,7 @@ TestReader: 以`YOLOv3`为例,使用固定尺寸输入预测: ``` -python python/infer.py --model_dir=../inference_model/yolov3_darknet53_270e_coco/ --image_file=../demo/000000014439.jpg --use_gpu=True --run_mode=trt_fp32 --run_benchmark=True +python python/infer.py --model_dir=./output_inference/yolov3_darknet53_270e_coco/ --image_file=./demo/000000014439.jpg --device=GPU --run_mode=trt_fp32 --run_benchmark=True ``` ### 3.3 TensorRT动态尺寸预测 @@ -73,7 +73,7 @@ Paddle预测库关于动态尺寸输入请查看[Paddle CPP预测](https://www.p 以`Faster RCNN`为例,使用动态尺寸输入预测: ``` -python python/infer.py --model_dir=../inference_model/faster_rcnn_r50_fpn_1x_coco/ --image_file=../demo/000000014439.jpg --use_gpu=True --run_mode=trt_fp16 --run_benchmark=True --trt_max_shape=1280 --trt_min_shape=800 --trt_opt_shape=960 +python python/infer.py --model_dir=./output_inference/faster_rcnn_r50_fpn_1x_coco/ --image_file=./demo/000000014439.jpg --device=GPU --run_mode=trt_fp16 --run_benchmark=True --trt_max_shape=1280 --trt_min_shape=800 --trt_opt_shape=960 ``` ## 4、常见问题QA diff --git a/deploy/benchmark/benchmark.sh b/deploy/benchmark/benchmark.sh index 84840892f..e29aaa884 100644 --- a/deploy/benchmark/benchmark.sh +++ b/deploy/benchmark/benchmark.sh @@ -18,7 +18,7 @@ for use_mkldnn in "True" "False"; do --model_dir=${model_dir} \ --run_benchmark True \ --enable_mkldnn=${use_mkldnn} \ - --use_gpu=False \ + --device=CPU \ --cpu_threads=${threads} \ --image_dir=${img_dir} 2>&1 | tee ${log_path}/${model_name}_cpu_usemkldnn_${use_mkldnn}_cputhreads_${threads}_bs1_infer.log done @@ -29,7 +29,7 @@ for run_mode in "fluid" "trt_fp32" "trt_fp16"; do python deploy/python/infer.py \ --model_dir=${model_dir} \ --run_benchmark=True \ - --use_gpu=True \ + --device=GPU \ --run_mode=${run_mode} \ --image_dir=${img_dir} 2>&1 | tee ${log_path}/${model_name}_gpu_runmode_${run_mode}_bs1_infer.log done diff --git a/deploy/benchmark/benchmark_quant.sh b/deploy/benchmark/benchmark_quant.sh index 1abdca900..a21541dd0 100644 --- a/deploy/benchmark/benchmark_quant.sh +++ b/deploy/benchmark/benchmark_quant.sh @@ -16,7 +16,7 @@ for run_mode in "trt_int8"; do python deploy/python/infer.py \ --model_dir=${model_dir} \ --run_benchmark=True \ - --use_gpu=True \ + --device=GPU \ --run_mode=${run_mode} \ --image_dir=${img_dir} 2>&1 | tee ${log_path}/${model_name}_gpu_runmode_${run_mode}_bs1_infer.log done diff --git a/deploy/python/README.md b/deploy/python/README.md index 73b6d81bf..09294ae95 100644 --- a/deploy/python/README.md +++ b/deploy/python/README.md @@ -21,7 +21,7 @@ PaddleDetection在训练过程包括网络的前向和优化器相关参数, 在终端输入以下命令进行预测: ```bash -python deploy/python/infer.py --model_dir=./inference/yolov3_mobilenet_v1_roadsign --image_file=./demo/road554.png --device=GPU +python deploy/python/infer.py --model_dir=./output_inference/yolov3_mobilenet_v1_roadsign --image_file=./demo/road554.png --device=GPU ``` 参数说明如下: diff --git a/docs/tutorials/GETTING_STARTED_cn.md b/docs/tutorials/GETTING_STARTED_cn.md index b22f054ca..ea94a5e09 100644 --- a/docs/tutorials/GETTING_STARTED_cn.md +++ b/docs/tutorials/GETTING_STARTED_cn.md @@ -243,7 +243,7 @@ python tools/export_model.py -c configs/yolov3/yolov3_mobilenet_v1_roadsign.yml PaddleDetection提供了PaddleInference、PaddleServing、PaddleLite多种部署形式,支持服务端、移动端、嵌入式等多种平台,提供了完善的Python和C++部署方案。 * 在这里,我们以Python为例,说明如何使用PaddleInference进行模型部署 ```bash -python deploy/python/infer.py --model_dir=./inference/yolov3_mobilenet_v1_roadsign --image_file=demo/road554.png --use_gpu=True +python deploy/python/infer.py --model_dir=./output_inference/yolov3_mobilenet_v1_roadsign --image_file=demo/road554.png --device=GPU ``` * 同时`infer.py`提供了丰富的接口,用户进行接入视频文件、摄像头进行预测,更多内容请参考[Python端预测部署](../../deploy/python.md) ### PaddleDetection支持的部署形式说明 -- GitLab