Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
bbdbf3d7
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
bbdbf3d7
编写于
4月 27, 2022
作者:
B
baoachun
提交者:
GitHub
4月 27, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
update test case output threshold (#41242) (#42250)
* update test case output threshold * update testcase
上级
e921cd1a
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
24 addition
and
9 deletion
+24
-9
paddle/fluid/inference/tests/infer_ut/test_det_mv3_db.cc
paddle/fluid/inference/tests/infer_ut/test_det_mv3_db.cc
+20
-5
paddle/fluid/inference/tests/infer_ut/test_ppyolo_mbv3.cc
paddle/fluid/inference/tests/infer_ut/test_ppyolo_mbv3.cc
+1
-1
paddle/fluid/inference/tests/infer_ut/test_resnet50.cc
paddle/fluid/inference/tests/infer_ut/test_resnet50.cc
+3
-3
未找到文件。
paddle/fluid/inference/tests/infer_ut/test_det_mv3_db.cc
浏览文件 @
bbdbf3d7
...
...
@@ -35,11 +35,26 @@ paddle::test::Record PrepareInput(int batch_size, int image_shape = 640) {
void
PrepareDynamicShape
(
paddle_infer
::
Config
*
config
,
int
max_batch_size
=
4
)
{
// set dynamic shape range
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>
min_input_shape
=
{
{
"x"
,
{
1
,
3
,
50
,
50
}}};
{
"x"
,
{
1
,
3
,
224
,
224
}},
{
"conv2d_124.tmp_0"
,
{
1
,
256
,
56
,
56
}},
{
"nearest_interp_v2_2.tmp_0"
,
{
1
,
256
,
56
,
56
}},
{
"nearest_interp_v2_3.tmp_0"
,
{
1
,
64
,
56
,
56
}},
{
"nearest_interp_v2_4.tmp_0"
,
{
1
,
64
,
56
,
56
}},
{
"nearest_interp_v2_5.tmp_0"
,
{
1
,
64
,
56
,
56
}}};
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>
max_input_shape
=
{
{
"x"
,
{
max_batch_size
,
3
,
1600
,
1600
}}};
{
"x"
,
{
max_batch_size
,
3
,
448
,
448
}},
{
"conv2d_124.tmp_0"
,
{
max_batch_size
,
256
,
112
,
112
}},
{
"nearest_interp_v2_2.tmp_0"
,
{
max_batch_size
,
256
,
112
,
112
}},
{
"nearest_interp_v2_3.tmp_0"
,
{
max_batch_size
,
64
,
112
,
112
}},
{
"nearest_interp_v2_4.tmp_0"
,
{
max_batch_size
,
64
,
112
,
112
}},
{
"nearest_interp_v2_5.tmp_0"
,
{
max_batch_size
,
64
,
112
,
112
}}};
std
::
map
<
std
::
string
,
std
::
vector
<
int
>>
opt_input_shape
=
{
{
"x"
,
{
1
,
3
,
640
,
640
}}};
{
"x"
,
{
1
,
3
,
256
,
256
}},
{
"conv2d_124.tmp_0"
,
{
1
,
256
,
64
,
64
}},
{
"nearest_interp_v2_2.tmp_0"
,
{
1
,
256
,
64
,
64
}},
{
"nearest_interp_v2_3.tmp_0"
,
{
1
,
64
,
64
,
64
}},
{
"nearest_interp_v2_4.tmp_0"
,
{
1
,
64
,
64
,
64
}},
{
"nearest_interp_v2_5.tmp_0"
,
{
1
,
64
,
64
,
64
}}};
config
->
SetTRTDynamicShapeInfo
(
min_input_shape
,
max_input_shape
,
opt_input_shape
);
}
...
...
@@ -76,7 +91,7 @@ TEST(tensorrt_tester_det_mv3_db, multi_thread2_trt_fp32_dynamic_shape_bz2) {
int
thread_num
=
2
;
// thread > 2 may OOM
// init input data
std
::
map
<
std
::
string
,
paddle
::
test
::
Record
>
my_input_data_map
;
my_input_data_map
[
"x"
]
=
PrepareInput
(
2
,
640
);
my_input_data_map
[
"x"
]
=
PrepareInput
(
2
,
256
);
// init output data
std
::
map
<
std
::
string
,
paddle
::
test
::
Record
>
infer_output_data
,
truth_output_data
;
...
...
@@ -90,7 +105,7 @@ TEST(tensorrt_tester_det_mv3_db, multi_thread2_trt_fp32_dynamic_shape_bz2) {
FLAGS_modeldir
+
"/inference.pdiparams"
);
config
.
EnableUseGpu
(
100
,
0
);
config
.
EnableTensorRtEngine
(
1
<<
20
,
2
,
3
,
paddle_infer
::
PrecisionType
::
kFloat32
,
false
,
false
);
1
<<
20
,
4
,
3
,
paddle_infer
::
PrecisionType
::
kFloat32
,
false
,
false
);
PrepareDynamicShape
(
&
config
,
4
);
// get groudtruth by disbale ir
paddle_infer
::
services
::
PredictorPool
pred_pool_no_ir
(
config_no_ir
,
1
);
...
...
paddle/fluid/inference/tests/infer_ut/test_ppyolo_mbv3.cc
浏览文件 @
bbdbf3d7
...
...
@@ -93,7 +93,7 @@ TEST(tensorrt_tester_ppyolo_mbv3, multi_thread4_trt_fp32_bz2) {
for
(
int
i
=
0
;
i
<
thread_num
;
++
i
)
{
LOG
(
INFO
)
<<
"join tid : "
<<
i
;
threads
[
i
].
join
();
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
,
1e-2
);
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
,
0.18
);
// TODO(OliverLPH): precision set to 1e-2 since input is fake, change to
// real input later
}
...
...
paddle/fluid/inference/tests/infer_ut/test_resnet50.cc
浏览文件 @
bbdbf3d7
...
...
@@ -87,7 +87,7 @@ TEST(tensorrt_tester_resnet50, trt_fp32_bz2) {
SingleThreadPrediction
(
pred_pool
.
Retrive
(
0
),
&
my_input_data_map
,
&
infer_output_data
);
// check outputs
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
);
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
,
2e-4
);
std
::
cout
<<
"finish test"
<<
std
::
endl
;
}
...
...
@@ -122,7 +122,7 @@ TEST(tensorrt_tester_resnet50, serial_diff_batch_trt_fp32) {
SingleThreadPrediction
(
pred_pool
.
Retrive
(
0
),
&
my_input_data_map
,
&
infer_output_data
);
// check outputs
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
);
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
,
1e-4
);
}
std
::
cout
<<
"finish test"
<<
std
::
endl
;
}
...
...
@@ -164,7 +164,7 @@ TEST(tensorrt_tester_resnet50, multi_thread4_trt_fp32_bz2) {
for
(
int
i
=
0
;
i
<
thread_num
;
++
i
)
{
LOG
(
INFO
)
<<
"join tid : "
<<
i
;
threads
[
i
].
join
();
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
);
CompareRecord
(
&
truth_output_data
,
&
infer_output_data
,
2e-4
);
}
std
::
cout
<<
"finish multi-thread test"
<<
std
::
endl
;
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录