Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleDetection
提交
d88f305d
P
PaddleDetection
项目概览
PaddlePaddle
/
PaddleDetection
大约 1 年 前同步成功
通知
695
Star
11112
Fork
2696
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
184
列表
看板
标记
里程碑
合并请求
40
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
184
Issue
184
列表
看板
标记
里程碑
合并请求
40
合并请求
40
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
d88f305d
编写于
5月 24, 2021
作者:
G
Guanghua Yu
提交者:
GitHub
5月 24, 2021
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Automatically set use_dynamic_shape (#3116)
* Automatically set use_dynamic_shape
上级
d6f36b14
变更
10
显示空白变更内容
内联
并排
Showing
10 changed file
with
37 addition
and
43 deletion
+37
-43
deploy/TENSOR_RT.md
deploy/TENSOR_RT.md
+13
-11
deploy/cpp/include/config_parser.h
deploy/cpp/include/config_parser.h
+9
-0
deploy/cpp/include/object_detector.h
deploy/cpp/include/object_detector.h
+1
-2
deploy/cpp/src/main.cc
deploy/cpp/src/main.cc
+7
-6
deploy/python/infer.py
deploy/python/infer.py
+3
-8
deploy/python/keypoint_det_unite_infer.py
deploy/python/keypoint_det_unite_infer.py
+0
-2
deploy/python/keypoint_infer.py
deploy/python/keypoint_infer.py
+2
-4
deploy/python/topdown_unite_utils.py
deploy/python/topdown_unite_utils.py
+0
-5
deploy/python/utils.py
deploy/python/utils.py
+0
-5
ppdet/engine/export_utils.py
ppdet/engine/export_utils.py
+2
-0
未找到文件。
deploy/TENSOR_RT.md
浏览文件 @
d88f305d
...
@@ -34,31 +34,33 @@ config->EnableTensorRtEngine(1 << 20 /*workspace_size*/,
...
@@ -34,31 +34,33 @@ config->EnableTensorRtEngine(1 << 20 /*workspace_size*/,
### 3.2 TensorRT固定尺寸预测
### 3.2 TensorRT固定尺寸预测
在导出模型时指定模型输入尺寸,设置
`TestReader.inputs_def.image_shape=[3,640,640]`
,具体请参考
[
PaddleDetection模型导出教程
](
../EXPORT_MODEL.md
)
。
例如在模型Reader配置文件中设置:
```
yaml
`TestReader.inputs_def.image_shape`
设置的是输入TensorRT引擎的数据尺寸(在像FasterRCNN中,
`TestReader.inputs_def.image_shape`
指定的是在
`Pad`
操作之前的图像数据尺寸)。
TestReader
:
inputs_def
:
image_shape
:
[
3
,
608
,
608
]
...
```
或者在导出模型时设置
`-o TestReader.inputs_def.image_shape=[3,608,608]`
,模型将会进行固定尺寸预测,具体请参考
[
PaddleDetection模型导出教程
](
../EXPORT_MODEL.md
)
。
可以通过
[
visualdl
](
https://www.paddlepaddle.org.cn/paddle/visualdl/demo/graph
)
打开
`model.pdmodel`
文件,查看输入的第一个Tensor尺寸是否是固定的,如果不指定,尺寸会用
`?`
表示,如下图所示:
可以通过
[
visualdl
](
https://www.paddlepaddle.org.cn/paddle/visualdl/demo/graph
)
打开
`model.pdmodel`
文件,查看输入的第一个Tensor尺寸是否是固定的,如果不指定,尺寸会用
`?`
表示,如下图所示:
![
img
](
../docs/images/input_shape.png
)
![
img
](
../docs/images/input_shape.png
)
同时需要将图像预处理后的尺寸与设置车模型输入尺寸保持一致,需要设置
`infer_cfg.yml`
配置文件中
`Resize OP`
的
`target_size`
参数和
`keep_ratio`
参数。
注意:由于TesnorRT不支持在batch维度进行slice操作,Faster RCNN 和 Mask RCNN
使用固定尺寸输入预测会报错,这两个模型请使用动态尺寸输入
。
注意:由于TesnorRT不支持在batch维度进行slice操作,Faster RCNN 和 Mask RCNN
不能使用固定尺寸输入预测,所以不能设置
`TestReader.inputs_def.image_shape`
字段
。
以
`YOLOv3`
为例,使用
动态
尺寸输入预测:
以
`YOLOv3`
为例,使用
固定
尺寸输入预测:
```
```
python python/infer.py --model_dir=../inference_model/yolov3_darknet53_270e_coco/ --image_file=../demo/000000014439
_640x640
.jpg --use_gpu=True --run_mode=trt_fp32 --run_benchmark=True
python python/infer.py --model_dir=../inference_model/yolov3_darknet53_270e_coco/ --image_file=../demo/000000014439.jpg --use_gpu=True --run_mode=trt_fp32 --run_benchmark=True
```
```
### 3.3 TensorRT动态尺寸预测
### 3.3 TensorRT动态尺寸预测
TensorRT版本>=6时,使用TensorRT预测时,可以支持动态尺寸输入。
TensorRT版本>=6时,使用TensorRT预测时,可以支持动态尺寸输入。
如果模型Reader配置文件中没有设置例如
`TestReader.inputs_def.image_shape=[3,608,608]`
的字段,或者
`image_shape=[3.-1,-1]`
,导出模型将以动态尺寸进行预测。一般RCNN系列模型使用动态图尺寸预测。
Paddle预测库关于动态尺寸输入请查看
[
Paddle CPP预测
](
https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/05_inference_deployment/inference/native_infer.html
)
的
`SetTRTDynamicShapeInfo`
函数说明。
Paddle预测库关于动态尺寸输入请查看
[
Paddle CPP预测
](
https://www.paddlepaddle.org.cn/documentation/docs/zh/guides/05_inference_deployment/inference/native_infer.html
)
的
`SetTRTDynamicShapeInfo`
函数说明。
`python/infer.py`
设置动态尺寸输入参数说明:
`python/infer.py`
设置动态尺寸输入参数说明:
-
use_dynamic_shape 用于设定TensorRT的输入尺寸是否是动态尺寸,默认值:False
-
trt_min_shape 用于设定TensorRT的输入图像height、width中的最小尺寸,默认值:1
-
trt_min_shape 用于设定TensorRT的输入图像height、width中的最小尺寸,默认值:1
-
trt_max_shape 用于设定TensorRT的输入图像height、width中的最大尺寸,默认值:1280
-
trt_max_shape 用于设定TensorRT的输入图像height、width中的最大尺寸,默认值:1280
...
@@ -69,7 +71,7 @@ Paddle预测库关于动态尺寸输入请查看[Paddle CPP预测](https://www.p
...
@@ -69,7 +71,7 @@ Paddle预测库关于动态尺寸输入请查看[Paddle CPP预测](https://www.p
以
`Faster RCNN`
为例,使用动态尺寸输入预测:
以
`Faster RCNN`
为例,使用动态尺寸输入预测:
```
```
python python/infer.py --model_dir=../inference_model/faster_rcnn_r50_fpn_1x_coco/ --image_file=../demo/000000014439.jpg --use_gpu=True --run_mode=trt_fp16 --run_benchmark=True --
use_dynamic_shape=True --
trt_max_shape=1280 --trt_min_shape=800 --trt_opt_shape=960
python python/infer.py --model_dir=../inference_model/faster_rcnn_r50_fpn_1x_coco/ --image_file=../demo/000000014439.jpg --use_gpu=True --run_mode=trt_fp16 --run_benchmark=True --trt_max_shape=1280 --trt_min_shape=800 --trt_opt_shape=960
```
```
## 4、常见问题QA
## 4、常见问题QA
...
...
deploy/cpp/include/config_parser.h
浏览文件 @
d88f305d
...
@@ -91,6 +91,14 @@ class ConfigPaser {
...
@@ -91,6 +91,14 @@ class ConfigPaser {
return
false
;
return
false
;
}
}
// Get use_dynamic_shape for TensorRT
if
(
config
[
"use_dynamic_shape"
].
IsDefined
())
{
use_dynamic_shape_
=
config
[
"use_dynamic_shape"
].
as
<
bool
>
();
}
else
{
std
::
cerr
<<
"Please set use_dynamic_shape."
<<
std
::
endl
;
return
false
;
}
return
true
;
return
true
;
}
}
std
::
string
mode_
;
std
::
string
mode_
;
...
@@ -99,6 +107,7 @@ class ConfigPaser {
...
@@ -99,6 +107,7 @@ class ConfigPaser {
int
min_subgraph_size_
;
int
min_subgraph_size_
;
YAML
::
Node
preprocess_info_
;
YAML
::
Node
preprocess_info_
;
std
::
vector
<
std
::
string
>
label_list_
;
std
::
vector
<
std
::
string
>
label_list_
;
bool
use_dynamic_shape_
;
};
};
}
// namespace PaddleDetection
}
// namespace PaddleDetection
...
...
deploy/cpp/include/object_detector.h
浏览文件 @
d88f305d
...
@@ -64,7 +64,6 @@ class ObjectDetector {
...
@@ -64,7 +64,6 @@ class ObjectDetector {
const
std
::
string
&
run_mode
=
"fluid"
,
const
std
::
string
&
run_mode
=
"fluid"
,
const
int
batch_size
=
1
,
const
int
batch_size
=
1
,
const
int
gpu_id
=
0
,
const
int
gpu_id
=
0
,
bool
use_dynamic_shape
=
false
,
const
int
trt_min_shape
=
1
,
const
int
trt_min_shape
=
1
,
const
int
trt_max_shape
=
1280
,
const
int
trt_max_shape
=
1280
,
const
int
trt_opt_shape
=
640
,
const
int
trt_opt_shape
=
640
,
...
@@ -74,12 +73,12 @@ class ObjectDetector {
...
@@ -74,12 +73,12 @@ class ObjectDetector {
this
->
cpu_math_library_num_threads_
=
cpu_threads
;
this
->
cpu_math_library_num_threads_
=
cpu_threads
;
this
->
use_mkldnn_
=
use_mkldnn
;
this
->
use_mkldnn_
=
use_mkldnn
;
this
->
use_dynamic_shape_
=
use_dynamic_shape
;
this
->
trt_min_shape_
=
trt_min_shape
;
this
->
trt_min_shape_
=
trt_min_shape
;
this
->
trt_max_shape_
=
trt_max_shape
;
this
->
trt_max_shape_
=
trt_max_shape
;
this
->
trt_opt_shape_
=
trt_opt_shape
;
this
->
trt_opt_shape_
=
trt_opt_shape
;
this
->
trt_calib_mode_
=
trt_calib_mode
;
this
->
trt_calib_mode_
=
trt_calib_mode
;
config_
.
load_config
(
model_dir
);
config_
.
load_config
(
model_dir
);
this
->
use_dynamic_shape_
=
config_
.
use_dynamic_shape_
;
this
->
min_subgraph_size_
=
config_
.
min_subgraph_size_
;
this
->
min_subgraph_size_
=
config_
.
min_subgraph_size_
;
threshold_
=
config_
.
draw_threshold_
;
threshold_
=
config_
.
draw_threshold_
;
preprocessor_
.
Init
(
config_
.
preprocess_info_
);
preprocessor_
.
Init
(
config_
.
preprocess_info_
);
...
...
deploy/cpp/src/main.cc
浏览文件 @
d88f305d
...
@@ -49,7 +49,6 @@ DEFINE_int32(gpu_id, 0, "Device id of GPU to execute");
...
@@ -49,7 +49,6 @@ DEFINE_int32(gpu_id, 0, "Device id of GPU to execute");
DEFINE_bool
(
run_benchmark
,
false
,
"Whether to predict a image_file repeatedly for benchmark"
);
DEFINE_bool
(
run_benchmark
,
false
,
"Whether to predict a image_file repeatedly for benchmark"
);
DEFINE_bool
(
use_mkldnn
,
false
,
"Whether use mkldnn with CPU"
);
DEFINE_bool
(
use_mkldnn
,
false
,
"Whether use mkldnn with CPU"
);
DEFINE_int32
(
cpu_threads
,
1
,
"Num of threads with CPU"
);
DEFINE_int32
(
cpu_threads
,
1
,
"Num of threads with CPU"
);
DEFINE_bool
(
use_dynamic_shape
,
false
,
"Trt use dynamic shape or not"
);
DEFINE_int32
(
trt_min_shape
,
1
,
"Min shape of TRT DynamicShapeI"
);
DEFINE_int32
(
trt_min_shape
,
1
,
"Min shape of TRT DynamicShapeI"
);
DEFINE_int32
(
trt_max_shape
,
1280
,
"Max shape of TRT DynamicShapeI"
);
DEFINE_int32
(
trt_max_shape
,
1280
,
"Max shape of TRT DynamicShapeI"
);
DEFINE_int32
(
trt_opt_shape
,
640
,
"Opt shape of TRT DynamicShapeI"
);
DEFINE_int32
(
trt_opt_shape
,
640
,
"Opt shape of TRT DynamicShapeI"
);
...
@@ -361,8 +360,9 @@ int main(int argc, char** argv) {
...
@@ -361,8 +360,9 @@ int main(int argc, char** argv) {
}
}
// Load model and create a object detector
// Load model and create a object detector
PaddleDetection
::
ObjectDetector
det
(
FLAGS_model_dir
,
FLAGS_use_gpu
,
FLAGS_use_mkldnn
,
PaddleDetection
::
ObjectDetector
det
(
FLAGS_model_dir
,
FLAGS_use_gpu
,
FLAGS_use_mkldnn
,
FLAGS_cpu_threads
,
FLAGS_run_mode
,
FLAGS_batch_size
,
FLAGS_gpu_id
,
FLAGS_use_dynamic_shape
,
FLAGS_cpu_threads
,
FLAGS_run_mode
,
FLAGS_batch_size
,
FLAGS_gpu_id
,
FLAGS_trt_min_shape
,
FLAGS_trt_max_shape
,
FLAGS_trt_opt_shape
,
FLAGS_trt_calib_mode
);
FLAGS_trt_min_shape
,
FLAGS_trt_max_shape
,
FLAGS_trt_opt_shape
,
FLAGS_trt_calib_mode
);
// Do inference on input video or image
// Do inference on input video or image
if
(
!
FLAGS_video_file
.
empty
()
||
FLAGS_camera_id
!=
-
1
)
{
if
(
!
FLAGS_video_file
.
empty
()
||
FLAGS_camera_id
!=
-
1
)
{
PredictVideo
(
FLAGS_video_file
,
&
det
);
PredictVideo
(
FLAGS_video_file
,
&
det
);
...
@@ -374,13 +374,14 @@ int main(int argc, char** argv) {
...
@@ -374,13 +374,14 @@ int main(int argc, char** argv) {
if
(
!
FLAGS_image_file
.
empty
())
{
if
(
!
FLAGS_image_file
.
empty
())
{
all_imgs
.
push_back
(
FLAGS_image_file
);
all_imgs
.
push_back
(
FLAGS_image_file
);
if
(
FLAGS_batch_size
>
1
)
{
if
(
FLAGS_batch_size
>
1
)
{
std
::
cout
<<
"batch_size should be 1, when
image_file is not None
"
<<
std
::
endl
;
std
::
cout
<<
"batch_size should be 1, when
set `image_file`.
"
<<
std
::
endl
;
FLAGS_batch_size
=
1
;
return
-
1
;
}
}
}
else
{
}
else
{
GetAllFiles
((
char
*
)
FLAGS_image_dir
.
c_str
(),
all_imgs
);
GetAllFiles
((
char
*
)
FLAGS_image_dir
.
c_str
(),
all_imgs
);
}
}
PredictImage
(
all_imgs
,
FLAGS_batch_size
,
FLAGS_threshold
,
FLAGS_run_benchmark
,
&
det
,
FLAGS_output_dir
);
PredictImage
(
all_imgs
,
FLAGS_batch_size
,
FLAGS_threshold
,
FLAGS_run_benchmark
,
&
det
,
FLAGS_output_dir
);
}
}
return
0
;
return
0
;
}
}
deploy/python/infer.py
浏览文件 @
d88f305d
...
@@ -51,7 +51,6 @@ class Detector(object):
...
@@ -51,7 +51,6 @@ class Detector(object):
use_gpu (bool): whether use gpu
use_gpu (bool): whether use gpu
run_mode (str): mode of running(fluid/trt_fp32/trt_fp16)
run_mode (str): mode of running(fluid/trt_fp32/trt_fp16)
batch_size (int): size of pre batch in inference
batch_size (int): size of pre batch in inference
use_dynamic_shape (bool): use dynamic shape or not
trt_min_shape (int): min shape for dynamic shape in trt
trt_min_shape (int): min shape for dynamic shape in trt
trt_max_shape (int): max shape for dynamic shape in trt
trt_max_shape (int): max shape for dynamic shape in trt
trt_opt_shape (int): opt shape for dynamic shape in trt
trt_opt_shape (int): opt shape for dynamic shape in trt
...
@@ -65,7 +64,6 @@ class Detector(object):
...
@@ -65,7 +64,6 @@ class Detector(object):
use_gpu
=
False
,
use_gpu
=
False
,
run_mode
=
'fluid'
,
run_mode
=
'fluid'
,
batch_size
=
1
,
batch_size
=
1
,
use_dynamic_shape
=
False
,
trt_min_shape
=
1
,
trt_min_shape
=
1
,
trt_max_shape
=
1280
,
trt_max_shape
=
1280
,
trt_opt_shape
=
640
,
trt_opt_shape
=
640
,
...
@@ -79,7 +77,7 @@ class Detector(object):
...
@@ -79,7 +77,7 @@ class Detector(object):
batch_size
=
batch_size
,
batch_size
=
batch_size
,
min_subgraph_size
=
self
.
pred_config
.
min_subgraph_size
,
min_subgraph_size
=
self
.
pred_config
.
min_subgraph_size
,
use_gpu
=
use_gpu
,
use_gpu
=
use_gpu
,
use_dynamic_shape
=
use_dynamic_shape
,
use_dynamic_shape
=
self
.
pred_config
.
use_dynamic_shape
,
trt_min_shape
=
trt_min_shape
,
trt_min_shape
=
trt_min_shape
,
trt_max_shape
=
trt_max_shape
,
trt_max_shape
=
trt_max_shape
,
trt_opt_shape
=
trt_opt_shape
,
trt_opt_shape
=
trt_opt_shape
,
...
@@ -189,7 +187,6 @@ class DetectorSOLOv2(Detector):
...
@@ -189,7 +187,6 @@ class DetectorSOLOv2(Detector):
use_gpu (bool): whether use gpu
use_gpu (bool): whether use gpu
run_mode (str): mode of running(fluid/trt_fp32/trt_fp16)
run_mode (str): mode of running(fluid/trt_fp32/trt_fp16)
batch_size (int): size of pre batch in inference
batch_size (int): size of pre batch in inference
use_dynamic_shape (bool): use dynamic shape or not
trt_min_shape (int): min shape for dynamic shape in trt
trt_min_shape (int): min shape for dynamic shape in trt
trt_max_shape (int): max shape for dynamic shape in trt
trt_max_shape (int): max shape for dynamic shape in trt
trt_opt_shape (int): opt shape for dynamic shape in trt
trt_opt_shape (int): opt shape for dynamic shape in trt
...
@@ -202,7 +199,6 @@ class DetectorSOLOv2(Detector):
...
@@ -202,7 +199,6 @@ class DetectorSOLOv2(Detector):
use_gpu
=
False
,
use_gpu
=
False
,
run_mode
=
'fluid'
,
run_mode
=
'fluid'
,
batch_size
=
1
,
batch_size
=
1
,
use_dynamic_shape
=
False
,
trt_min_shape
=
1
,
trt_min_shape
=
1
,
trt_max_shape
=
1280
,
trt_max_shape
=
1280
,
trt_opt_shape
=
640
,
trt_opt_shape
=
640
,
...
@@ -216,7 +212,7 @@ class DetectorSOLOv2(Detector):
...
@@ -216,7 +212,7 @@ class DetectorSOLOv2(Detector):
batch_size
=
batch_size
,
batch_size
=
batch_size
,
min_subgraph_size
=
self
.
pred_config
.
min_subgraph_size
,
min_subgraph_size
=
self
.
pred_config
.
min_subgraph_size
,
use_gpu
=
use_gpu
,
use_gpu
=
use_gpu
,
use_dynamic_shape
=
use_dynamic_shape
,
use_dynamic_shape
=
self
.
pred_config
.
use_dynamic_shape
,
trt_min_shape
=
trt_min_shape
,
trt_min_shape
=
trt_min_shape
,
trt_max_shape
=
trt_max_shape
,
trt_max_shape
=
trt_max_shape
,
trt_opt_shape
=
trt_opt_shape
,
trt_opt_shape
=
trt_opt_shape
,
...
@@ -328,6 +324,7 @@ class PredictConfig():
...
@@ -328,6 +324,7 @@ class PredictConfig():
self
.
min_subgraph_size
=
yml_conf
[
'min_subgraph_size'
]
self
.
min_subgraph_size
=
yml_conf
[
'min_subgraph_size'
]
self
.
labels
=
yml_conf
[
'label_list'
]
self
.
labels
=
yml_conf
[
'label_list'
]
self
.
mask
=
False
self
.
mask
=
False
self
.
use_dynamic_shape
=
yml_conf
[
'use_dynamic_shape'
]
if
'mask'
in
yml_conf
:
if
'mask'
in
yml_conf
:
self
.
mask
=
yml_conf
[
'mask'
]
self
.
mask
=
yml_conf
[
'mask'
]
self
.
print_config
()
self
.
print_config
()
...
@@ -573,7 +570,6 @@ def main():
...
@@ -573,7 +570,6 @@ def main():
use_gpu
=
FLAGS
.
use_gpu
,
use_gpu
=
FLAGS
.
use_gpu
,
run_mode
=
FLAGS
.
run_mode
,
run_mode
=
FLAGS
.
run_mode
,
batch_size
=
FLAGS
.
batch_size
,
batch_size
=
FLAGS
.
batch_size
,
use_dynamic_shape
=
FLAGS
.
use_dynamic_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
...
@@ -587,7 +583,6 @@ def main():
...
@@ -587,7 +583,6 @@ def main():
use_gpu
=
FLAGS
.
use_gpu
,
use_gpu
=
FLAGS
.
use_gpu
,
run_mode
=
FLAGS
.
run_mode
,
run_mode
=
FLAGS
.
run_mode
,
batch_size
=
FLAGS
.
batch_size
,
batch_size
=
FLAGS
.
batch_size
,
use_dynamic_shape
=
FLAGS
.
use_dynamic_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
...
...
deploy/python/keypoint_det_unite_infer.py
浏览文件 @
d88f305d
...
@@ -158,7 +158,6 @@ def main():
...
@@ -158,7 +158,6 @@ def main():
FLAGS
.
det_model_dir
,
FLAGS
.
det_model_dir
,
use_gpu
=
FLAGS
.
use_gpu
,
use_gpu
=
FLAGS
.
use_gpu
,
run_mode
=
FLAGS
.
run_mode
,
run_mode
=
FLAGS
.
run_mode
,
use_dynamic_shape
=
FLAGS
.
use_dynamic_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
...
@@ -172,7 +171,6 @@ def main():
...
@@ -172,7 +171,6 @@ def main():
FLAGS
.
keypoint_model_dir
,
FLAGS
.
keypoint_model_dir
,
use_gpu
=
FLAGS
.
use_gpu
,
use_gpu
=
FLAGS
.
use_gpu
,
run_mode
=
FLAGS
.
run_mode
,
run_mode
=
FLAGS
.
run_mode
,
use_dynamic_shape
=
FLAGS
.
use_dynamic_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
...
...
deploy/python/keypoint_infer.py
浏览文件 @
d88f305d
...
@@ -46,7 +46,6 @@ class KeyPoint_Detector(object):
...
@@ -46,7 +46,6 @@ class KeyPoint_Detector(object):
model_dir (str): root path of model.pdiparams, model.pdmodel and infer_cfg.yml
model_dir (str): root path of model.pdiparams, model.pdmodel and infer_cfg.yml
use_gpu (bool): whether use gpu
use_gpu (bool): whether use gpu
run_mode (str): mode of running(fluid/trt_fp32/trt_fp16)
run_mode (str): mode of running(fluid/trt_fp32/trt_fp16)
use_dynamic_shape (bool): use dynamic shape or not
trt_min_shape (int): min shape for dynamic shape in trt
trt_min_shape (int): min shape for dynamic shape in trt
trt_max_shape (int): max shape for dynamic shape in trt
trt_max_shape (int): max shape for dynamic shape in trt
trt_opt_shape (int): opt shape for dynamic shape in trt
trt_opt_shape (int): opt shape for dynamic shape in trt
...
@@ -59,7 +58,6 @@ class KeyPoint_Detector(object):
...
@@ -59,7 +58,6 @@ class KeyPoint_Detector(object):
model_dir
,
model_dir
,
use_gpu
=
False
,
use_gpu
=
False
,
run_mode
=
'fluid'
,
run_mode
=
'fluid'
,
use_dynamic_shape
=
False
,
trt_min_shape
=
1
,
trt_min_shape
=
1
,
trt_max_shape
=
1280
,
trt_max_shape
=
1280
,
trt_opt_shape
=
640
,
trt_opt_shape
=
640
,
...
@@ -72,7 +70,7 @@ class KeyPoint_Detector(object):
...
@@ -72,7 +70,7 @@ class KeyPoint_Detector(object):
run_mode
=
run_mode
,
run_mode
=
run_mode
,
min_subgraph_size
=
self
.
pred_config
.
min_subgraph_size
,
min_subgraph_size
=
self
.
pred_config
.
min_subgraph_size
,
use_gpu
=
use_gpu
,
use_gpu
=
use_gpu
,
use_dynamic_shape
=
use_dynamic_shape
,
use_dynamic_shape
=
self
.
pred_config
.
use_dynamic_shape
,
trt_min_shape
=
trt_min_shape
,
trt_min_shape
=
trt_min_shape
,
trt_max_shape
=
trt_max_shape
,
trt_max_shape
=
trt_max_shape
,
trt_opt_shape
=
trt_opt_shape
,
trt_opt_shape
=
trt_opt_shape
,
...
@@ -210,6 +208,7 @@ class PredictConfig_KeyPoint():
...
@@ -210,6 +208,7 @@ class PredictConfig_KeyPoint():
self
.
min_subgraph_size
=
yml_conf
[
'min_subgraph_size'
]
self
.
min_subgraph_size
=
yml_conf
[
'min_subgraph_size'
]
self
.
labels
=
yml_conf
[
'label_list'
]
self
.
labels
=
yml_conf
[
'label_list'
]
self
.
tagmap
=
False
self
.
tagmap
=
False
self
.
use_dynamic_shape
=
yml_conf
[
'use_dynamic_shape'
]
if
'keypoint_bottomup'
==
self
.
archcls
:
if
'keypoint_bottomup'
==
self
.
archcls
:
self
.
tagmap
=
True
self
.
tagmap
=
True
self
.
print_config
()
self
.
print_config
()
...
@@ -384,7 +383,6 @@ def main():
...
@@ -384,7 +383,6 @@ def main():
FLAGS
.
model_dir
,
FLAGS
.
model_dir
,
use_gpu
=
FLAGS
.
use_gpu
,
use_gpu
=
FLAGS
.
use_gpu
,
run_mode
=
FLAGS
.
run_mode
,
run_mode
=
FLAGS
.
run_mode
,
use_dynamic_shape
=
FLAGS
.
use_dynamic_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_min_shape
=
FLAGS
.
trt_min_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_max_shape
=
FLAGS
.
trt_max_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
trt_opt_shape
=
FLAGS
.
trt_opt_shape
,
...
...
deploy/python/topdown_unite_utils.py
浏览文件 @
d88f305d
...
@@ -84,11 +84,6 @@ def argsparser():
...
@@ -84,11 +84,6 @@ def argsparser():
help
=
"Whether use mkldnn with CPU."
)
help
=
"Whether use mkldnn with CPU."
)
parser
.
add_argument
(
parser
.
add_argument
(
"--cpu_threads"
,
type
=
int
,
default
=
1
,
help
=
"Num of threads with CPU."
)
"--cpu_threads"
,
type
=
int
,
default
=
1
,
help
=
"Num of threads with CPU."
)
parser
.
add_argument
(
"--use_dynamic_shape"
,
type
=
ast
.
literal_eval
,
default
=
False
,
help
=
"Dynamic_shape for TensorRT."
)
parser
.
add_argument
(
parser
.
add_argument
(
"--trt_min_shape"
,
type
=
int
,
default
=
1
,
help
=
"min_shape for TensorRT."
)
"--trt_min_shape"
,
type
=
int
,
default
=
1
,
help
=
"min_shape for TensorRT."
)
parser
.
add_argument
(
parser
.
add_argument
(
...
...
deploy/python/utils.py
浏览文件 @
d88f305d
...
@@ -76,11 +76,6 @@ def argsparser():
...
@@ -76,11 +76,6 @@ def argsparser():
help
=
"Whether use mkldnn with CPU."
)
help
=
"Whether use mkldnn with CPU."
)
parser
.
add_argument
(
parser
.
add_argument
(
"--cpu_threads"
,
type
=
int
,
default
=
1
,
help
=
"Num of threads with CPU."
)
"--cpu_threads"
,
type
=
int
,
default
=
1
,
help
=
"Num of threads with CPU."
)
parser
.
add_argument
(
"--use_dynamic_shape"
,
type
=
ast
.
literal_eval
,
default
=
False
,
help
=
"Dynamic_shape for TensorRT."
)
parser
.
add_argument
(
parser
.
add_argument
(
"--trt_min_shape"
,
type
=
int
,
default
=
1
,
help
=
"min_shape for TensorRT."
)
"--trt_min_shape"
,
type
=
int
,
default
=
1
,
help
=
"min_shape for TensorRT."
)
parser
.
add_argument
(
parser
.
add_argument
(
...
...
ppdet/engine/export_utils.py
浏览文件 @
d88f305d
...
@@ -81,10 +81,12 @@ def _dump_infer_config(config, path, image_shape, model):
...
@@ -81,10 +81,12 @@ def _dump_infer_config(config, path, image_shape, model):
arch_state
=
False
arch_state
=
False
from
ppdet.core.config.yaml_helpers
import
setup_orderdict
from
ppdet.core.config.yaml_helpers
import
setup_orderdict
setup_orderdict
()
setup_orderdict
()
use_dynamic_shape
=
True
if
image_shape
[
1
]
==
-
1
else
False
infer_cfg
=
OrderedDict
({
infer_cfg
=
OrderedDict
({
'mode'
:
'fluid'
,
'mode'
:
'fluid'
,
'draw_threshold'
:
0.5
,
'draw_threshold'
:
0.5
,
'metric'
:
config
[
'metric'
],
'metric'
:
config
[
'metric'
],
'use_dynamic_shape'
:
use_dynamic_shape
})
})
infer_arch
=
config
[
'architecture'
]
infer_arch
=
config
[
'architecture'
]
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录