Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Mr.Vain
Mace
提交
3828ac54
Mace
项目概览
Mr.Vain
/
Mace
与 Fork 源项目一致
Fork自
Xiaomi / Mace
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
Mace
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
3828ac54
编写于
3月 31, 2018
作者:
L
liuqi
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Fix benchmark model with new APIs.
上级
e9eaa4d4
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
51 addition
and
60 deletion
+51
-60
mace/benchmark/benchmark_model.cc
mace/benchmark/benchmark_model.cc
+51
-59
tools/benchmark.sh
tools/benchmark.sh
+0
-1
未找到文件。
mace/benchmark/benchmark_model.cc
浏览文件 @
3828ac54
...
...
@@ -12,6 +12,7 @@
#include "gflags/gflags.h"
#include "mace/public/mace.h"
#include "mace/public/mace_runtime.h"
#include "mace/utils/logging.h"
#include "mace/benchmark/stat_summarizer.h"
...
...
@@ -95,9 +96,23 @@ inline int64_t NowMicros() {
return
static_cast
<
int64_t
>
(
tv
.
tv_sec
)
*
1000000
+
tv
.
tv_usec
;
}
DeviceType
ParseDeviceType
(
const
std
::
string
&
device_str
)
{
if
(
device_str
.
compare
(
"CPU"
)
==
0
)
{
return
DeviceType
::
CPU
;
}
else
if
(
device_str
.
compare
(
"NEON"
)
==
0
)
{
return
DeviceType
::
NEON
;
}
else
if
(
device_str
.
compare
(
"OPENCL"
)
==
0
)
{
return
DeviceType
::
OPENCL
;
}
else
if
(
device_str
.
compare
(
"HEXAGON"
)
==
0
)
{
return
DeviceType
::
HEXAGON
;
}
else
{
return
DeviceType
::
CPU
;
}
}
bool
RunInference
(
MaceEngine
*
engine
,
const
std
::
vector
<
mace
::
MaceInputInfo
>
&
input_infos
,
std
::
map
<
std
::
string
,
float
*
>
*
output_infos
,
const
std
::
map
<
std
::
string
,
mace
::
MaceTensor
>
&
input_infos
,
std
::
map
<
std
::
string
,
mace
::
MaceTensor
>
*
output_infos
,
StatSummarizer
*
summarizer
,
int64_t
*
inference_time_us
)
{
MACE_CHECK_NOTNULL
(
output_infos
);
...
...
@@ -106,28 +121,16 @@ bool RunInference(MaceEngine *engine,
if
(
summarizer
)
{
run_metadata_ptr
=
&
run_metadata
;
}
if
(
input_infos
.
size
()
==
1
&&
output_infos
->
size
()
==
1
)
{
const
int64_t
start_time
=
NowMicros
();
bool
s
=
engine
->
Run
(
input_infos
[
0
].
data
,
input_infos
[
0
].
shape
,
output_infos
->
begin
()
->
second
,
run_metadata_ptr
);
const
int64_t
end_time
=
NowMicros
();
if
(
!
s
)
{
LOG
(
ERROR
)
<<
"Error during inference."
;
return
s
;
}
*
inference_time_us
=
end_time
-
start_time
;
}
else
{
const
int64_t
start_time
=
NowMicros
();
bool
s
=
engine
->
Run
(
input_infos
,
*
output_infos
,
run_metadata_ptr
);
const
int64_t
end_time
=
NowMicros
();
const
int64_t
start_time
=
NowMicros
();
mace
::
MaceStatus
s
=
engine
->
Run
(
input_infos
,
output_infos
,
run_metadata_ptr
);
const
int64_t
end_time
=
NowMicros
();
if
(
!
s
)
{
LOG
(
ERROR
)
<<
"Error during inference."
;
return
s
;
}
*
inference_time_us
=
end_time
-
start_time
;
if
(
s
!=
mace
::
MaceStatus
::
MACE_SUCCESS
)
{
LOG
(
ERROR
)
<<
"Error during inference."
;
return
false
;
}
*
inference_time_us
=
end_time
-
start_time
;
if
(
summarizer
!=
nullptr
)
{
summarizer
->
ProcessMetadata
(
run_metadata
);
...
...
@@ -137,8 +140,8 @@ bool RunInference(MaceEngine *engine,
}
bool
Run
(
MaceEngine
*
engine
,
const
std
::
vector
<
mace
::
MaceInputInfo
>
&
input_infos
,
std
::
map
<
std
::
string
,
float
*
>
*
output_infos
,
const
std
::
map
<
std
::
string
,
mace
::
MaceTensor
>
&
input_infos
,
std
::
map
<
std
::
string
,
mace
::
MaceTensor
>
*
output_infos
,
StatSummarizer
*
summarizer
,
int
num_runs
,
double
max_time_sec
,
...
...
@@ -261,12 +264,7 @@ int Main(int argc, char **argv) {
stats_options
.
show_summary
=
FLAGS_show_summary
;
stats
.
reset
(
new
StatSummarizer
(
stats_options
));
DeviceType
device_type
=
CPU
;
if
(
FLAGS_device
==
"OPENCL"
)
{
device_type
=
OPENCL
;
}
else
if
(
FLAGS_device
==
"NEON"
)
{
device_type
=
NEON
;
}
mace
::
DeviceType
device_type
=
ParseDeviceType
(
FLAGS_device
);
// config runtime
mace
::
ConfigOmpThreads
(
FLAGS_omp_num_threads
);
...
...
@@ -302,50 +300,44 @@ int Main(int argc, char **argv) {
mace
::
MACE_MODEL_TAG
::
LoadModelData
(
FLAGS_model_data_file
.
c_str
());
NetDef
net_def
=
mace
::
MACE_MODEL_TAG
::
CreateNet
(
model_data
);
std
::
vector
<
mace
::
MaceInputInfo
>
input_infos
(
input_count
);
std
::
map
<
std
::
string
,
float
*>
output_infos
;
std
::
vector
<
std
::
unique_ptr
<
float
[]
>>
input_datas
(
input_count
);
std
::
vector
<
std
::
unique_ptr
<
float
[]
>>
output_datas
(
output_count
);
std
::
map
<
std
::
string
,
mace
::
MaceTensor
>
inputs
;
std
::
map
<
std
::
string
,
mace
::
MaceTensor
>
outputs
;
for
(
size_t
i
=
0
;
i
<
input_count
;
++
i
)
{
int64_t
input_size
=
std
::
accumulate
(
input_shape_vec
[
i
].
begin
(),
input_shape_vec
[
i
].
end
(),
1
,
std
::
multiplies
<
int64_t
>
());
input_datas
[
i
].
reset
(
new
float
[
input_size
]);
// Allocate input and output
int64_t
input_size
=
std
::
accumulate
(
input_shape_vec
[
i
].
begin
(),
input_shape_vec
[
i
].
end
(),
1
,
std
::
multiplies
<
int64_t
>
());
auto
buffer_in
=
std
::
shared_ptr
<
float
>
(
new
float
[
input_size
],
std
::
default_delete
<
float
[]
>
());
// load input
std
::
ifstream
in_file
(
FLAGS_input_file
+
"_"
+
FormatName
(
input_names
[
i
]),
std
::
ios
::
in
|
std
::
ios
::
binary
);
if
(
in_file
.
is_open
())
{
in_file
.
read
(
reinterpret_cast
<
char
*>
(
input_datas
[
i
]
.
get
()),
in_file
.
read
(
reinterpret_cast
<
char
*>
(
buffer_in
.
get
()),
input_size
*
sizeof
(
float
));
in_file
.
close
();
}
else
{
LOG
(
INFO
)
<<
"Open input file failed"
;
return
-
1
;
}
input_infos
[
i
].
name
=
input_names
[
i
];
input_infos
[
i
].
shape
=
input_shape_vec
[
i
];
input_infos
[
i
].
data
=
input_datas
[
i
].
get
();
inputs
[
input_names
[
i
]]
=
mace
::
MaceTensor
(
input_shape_vec
[
i
],
buffer_in
);
}
for
(
size_t
i
=
0
;
i
<
output_count
;
++
i
)
{
int64_t
output_size
=
std
::
accumulate
(
output_shape_vec
[
i
].
begin
(),
output_shape_vec
[
i
].
end
(),
1
,
std
::
multiplies
<
int64_t
>
());
output_datas
[
i
].
reset
(
new
float
[
output_size
]);
output_infos
[
output_names
[
i
]]
=
output_datas
[
i
].
get
();
int64_t
output_size
=
std
::
accumulate
(
output_shape_vec
[
i
].
begin
(),
output_shape_vec
[
i
].
end
(),
1
,
std
::
multiplies
<
int64_t
>
());
auto
buffer_out
=
std
::
shared_ptr
<
float
>
(
new
float
[
output_size
],
std
::
default_delete
<
float
[]
>
());
outputs
[
output_names
[
i
]]
=
mace
::
MaceTensor
(
output_shape_vec
[
i
],
buffer_out
);
}
// Init model
LOG
(
INFO
)
<<
"Run init"
;
std
::
unique_ptr
<
mace
::
MaceEngine
>
engine_ptr
;
if
(
input_count
==
1
&&
output_count
==
1
)
{
engine_ptr
.
reset
(
new
mace
::
MaceEngine
(
&
net_def
,
device_type
));
}
else
{
engine_ptr
.
reset
(
new
mace
::
MaceEngine
(
&
net_def
,
device_type
,
input_names
,
output_names
));
}
if
(
device_type
==
DeviceType
::
OPENCL
)
{
std
::
unique_ptr
<
mace
::
MaceEngine
>
engine_ptr
(
new
mace
::
MaceEngine
(
&
net_def
,
device_type
,
input_names
,
output_names
));
if
(
device_type
==
DeviceType
::
OPENCL
||
device_type
==
DeviceType
::
HEXAGON
)
{
mace
::
MACE_MODEL_TAG
::
UnloadModelData
(
model_data
);
}
...
...
@@ -355,7 +347,7 @@ int Main(int argc, char **argv) {
int64_t
num_warmup_runs
=
0
;
if
(
FLAGS_warmup_runs
>
0
)
{
bool
status
=
Run
(
engine_ptr
.
get
(),
input
_infos
,
&
output_info
s
,
nullptr
,
Run
(
engine_ptr
.
get
(),
input
s
,
&
output
s
,
nullptr
,
FLAGS_warmup_runs
,
-
1.0
,
inter_inference_sleep_seconds
,
&
warmup_time_us
,
&
num_warmup_runs
);
if
(
!
status
)
{
...
...
@@ -370,7 +362,7 @@ int Main(int argc, char **argv) {
int64_t
no_stat_time_us
=
0
;
int64_t
no_stat_runs
=
0
;
bool
status
=
Run
(
engine_ptr
.
get
(),
input
_infos
,
&
output_info
s
,
Run
(
engine_ptr
.
get
(),
input
s
,
&
output
s
,
nullptr
,
FLAGS_max_num_runs
,
max_benchmark_time_seconds
,
inter_inference_sleep_seconds
,
&
no_stat_time_us
,
&
no_stat_runs
);
if
(
!
status
)
{
...
...
@@ -379,7 +371,7 @@ int Main(int argc, char **argv) {
int64_t
stat_time_us
=
0
;
int64_t
stat_runs
=
0
;
status
=
Run
(
engine_ptr
.
get
(),
input
_infos
,
&
output_info
s
,
status
=
Run
(
engine_ptr
.
get
(),
input
s
,
&
output
s
,
stats
.
get
(),
FLAGS_max_num_runs
,
max_benchmark_time_seconds
,
inter_inference_sleep_seconds
,
&
stat_time_us
,
&
stat_runs
);
if
(
!
status
)
{
...
...
tools/benchmark.sh
浏览文件 @
3828ac54
#!/bin/bash
set
-x
Usage
()
{
echo
"Usage: bash tools/benchmark.sh target_soc model_output_dir option_args"
}
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录