Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Greenplum
Opencv
提交
452fa301
O
Opencv
项目概览
Greenplum
/
Opencv
大约 1 年 前同步成功
通知
7
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
O
Opencv
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
452fa301
编写于
7月 10, 2018
作者:
A
Alexander Alekhin
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
dnn(test): drop CV_ENUM for DNNBackend / DNNTarget
上级
7fe07279
变更
7
隐藏空白更改
内联
并排
Showing
7 changed file
with
85 addition
and
60 deletion
+85
-60
modules/dnn/test/test_backends.cpp
modules/dnn/test/test_backends.cpp
+9
-9
modules/dnn/test/test_caffe_importer.cpp
modules/dnn/test/test_caffe_importer.cpp
+5
-5
modules/dnn/test/test_googlenet.cpp
modules/dnn/test/test_googlenet.cpp
+1
-1
modules/dnn/test/test_halide_layers.cpp
modules/dnn/test/test_halide_layers.cpp
+24
-24
modules/dnn/test/test_precomp.hpp
modules/dnn/test/test_precomp.hpp
+43
-18
modules/dnn/test/test_tf_importer.cpp
modules/dnn/test/test_tf_importer.cpp
+1
-1
modules/dnn/test/test_torch_importer.cpp
modules/dnn/test/test_torch_importer.cpp
+2
-2
未找到文件。
modules/dnn/test/test_backends.cpp
浏览文件 @
452fa301
...
...
@@ -278,19 +278,19 @@ TEST_P(DNNTestNetwork, FastNeuralStyle_eccv16)
processNet
(
"dnn/fast_neural_style_eccv16_starry_night.t7"
,
""
,
inp
,
""
,
""
,
l1
,
lInf
);
}
const
tuple
<
DNNBackend
,
DNN
Target
>
testCases
[]
=
{
const
tuple
<
Backend
,
Target
>
testCases
[]
=
{
#ifdef HAVE_HALIDE
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_CPU
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_CPU
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_OPENCL
),
#endif
#ifdef HAVE_INF_ENGINE
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_CPU
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL_FP16
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_MYRIAD
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_CPU
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL_FP16
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_MYRIAD
),
#endif
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL_FP16
)
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL_FP16
)
};
INSTANTIATE_TEST_CASE_P
(
/*nothing*/
,
DNNTestNetwork
,
testing
::
ValuesIn
(
testCases
));
...
...
modules/dnn/test/test_caffe_importer.cpp
浏览文件 @
452fa301
...
...
@@ -82,7 +82,7 @@ TEST(Test_Caffe, read_googlenet)
ASSERT_FALSE
(
net
.
empty
());
}
typedef
testing
::
TestWithParam
<
tuple
<
bool
,
DNN
Target
>
>
Reproducibility_AlexNet
;
typedef
testing
::
TestWithParam
<
tuple
<
bool
,
Target
>
>
Reproducibility_AlexNet
;
TEST_P
(
Reproducibility_AlexNet
,
Accuracy
)
{
bool
readFromMemory
=
get
<
0
>
(
GetParam
());
...
...
@@ -179,7 +179,7 @@ TEST(Reproducibility_SSD, Accuracy)
normAssertDetections
(
ref
,
out
);
}
typedef
testing
::
TestWithParam
<
DNN
Target
>
Reproducibility_MobileNet_SSD
;
typedef
testing
::
TestWithParam
<
Target
>
Reproducibility_MobileNet_SSD
;
TEST_P
(
Reproducibility_MobileNet_SSD
,
Accuracy
)
{
const
string
proto
=
findDataFile
(
"dnn/MobileNetSSD_deploy.prototxt"
,
false
);
...
...
@@ -234,7 +234,7 @@ TEST_P(Reproducibility_MobileNet_SSD, Accuracy)
INSTANTIATE_TEST_CASE_P
(
/**/
,
Reproducibility_MobileNet_SSD
,
Values
(
DNN_TARGET_CPU
,
DNN_TARGET_OPENCL
,
DNN_TARGET_OPENCL_FP16
));
typedef
testing
::
TestWithParam
<
DNN
Target
>
Reproducibility_ResNet50
;
typedef
testing
::
TestWithParam
<
Target
>
Reproducibility_ResNet50
;
TEST_P
(
Reproducibility_ResNet50
,
Accuracy
)
{
Net
net
=
readNetFromCaffe
(
findDataFile
(
"dnn/ResNet-50-deploy.prototxt"
,
false
),
...
...
@@ -270,7 +270,7 @@ TEST_P(Reproducibility_ResNet50, Accuracy)
INSTANTIATE_TEST_CASE_P
(
/**/
,
Reproducibility_ResNet50
,
Values
(
DNN_TARGET_CPU
,
DNN_TARGET_OPENCL
,
DNN_TARGET_OPENCL_FP16
));
typedef
testing
::
TestWithParam
<
DNN
Target
>
Reproducibility_SqueezeNet_v1_1
;
typedef
testing
::
TestWithParam
<
Target
>
Reproducibility_SqueezeNet_v1_1
;
TEST_P
(
Reproducibility_SqueezeNet_v1_1
,
Accuracy
)
{
Net
net
=
readNetFromCaffe
(
findDataFile
(
"dnn/squeezenet_v1.1.prototxt"
,
false
),
...
...
@@ -413,7 +413,7 @@ TEST(Test_Caffe, multiple_inputs)
normAssert
(
out
,
first_image
+
second_image
);
}
typedef
testing
::
TestWithParam
<
tuple
<
std
::
string
,
DNN
Target
>
>
opencv_face_detector
;
typedef
testing
::
TestWithParam
<
tuple
<
std
::
string
,
Target
>
>
opencv_face_detector
;
TEST_P
(
opencv_face_detector
,
Accuracy
)
{
std
::
string
proto
=
findDataFile
(
"dnn/opencv_face_detector.prototxt"
,
false
);
...
...
modules/dnn/test/test_googlenet.cpp
浏览文件 @
452fa301
...
...
@@ -52,7 +52,7 @@ static std::string _tf(TString filename)
return
(
getOpenCVExtraDir
()
+
"/dnn/"
)
+
filename
;
}
typedef
testing
::
TestWithParam
<
DNN
Target
>
Reproducibility_GoogLeNet
;
typedef
testing
::
TestWithParam
<
Target
>
Reproducibility_GoogLeNet
;
TEST_P
(
Reproducibility_GoogLeNet
,
Batching
)
{
Net
net
=
readNetFromCaffe
(
findDataFile
(
"dnn/bvlc_googlenet.prototxt"
,
false
),
...
...
modules/dnn/test/test_halide_layers.cpp
浏览文件 @
452fa301
...
...
@@ -41,21 +41,21 @@ static void test(LayerParams& params, Mat& input, int backendId, int targetId)
test
(
input
,
net
,
backendId
,
targetId
);
}
static
testing
::
internal
::
ParamGenerator
<
tuple
<
DNNBackend
,
DNN
Target
>
>
dnnBackendsAndTargetsWithHalide
()
static
testing
::
internal
::
ParamGenerator
<
tuple
<
Backend
,
Target
>
>
dnnBackendsAndTargetsWithHalide
()
{
static
const
tuple
<
DNNBackend
,
DNN
Target
>
testCases
[]
=
{
static
const
tuple
<
Backend
,
Target
>
testCases
[]
=
{
#ifdef HAVE_HALIDE
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_CPU
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_CPU
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_HALIDE
,
DNN_TARGET_OPENCL
),
#endif
#ifdef HAVE_INF_ENGINE
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_CPU
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL_FP16
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_MYRIAD
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_CPU
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL_FP16
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_MYRIAD
),
#endif
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL_FP16
)
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL_FP16
)
};
return
testing
::
ValuesIn
(
testCases
);
}
...
...
@@ -89,7 +89,7 @@ TEST_P(Test_Halide_layers, Padding)
////////////////////////////////////////////////////////////////////////////////
// Convolution
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
Vec3i
,
Size
,
Size
,
Size
,
Size
,
Size
,
bool
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
Convolution
;
typedef
TestWithParam
<
tuple
<
Vec3i
,
Size
,
Size
,
Size
,
Size
,
Size
,
bool
,
tuple
<
Backend
,
Target
>
>
>
Convolution
;
TEST_P
(
Convolution
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
())[
0
];
...
...
@@ -154,7 +154,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, Convolution, Combine(
////////////////////////////////////////////////////////////////////////////////
// Deconvolution
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
Vec3i
,
Size
,
Size
,
Size
,
Size
,
Vec4i
,
bool
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
Deconvolution
;
typedef
TestWithParam
<
tuple
<
Vec3i
,
Size
,
Size
,
Size
,
Size
,
Vec4i
,
bool
,
tuple
<
Backend
,
Target
>
>
>
Deconvolution
;
TEST_P
(
Deconvolution
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
())[
0
];
...
...
@@ -220,7 +220,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, Deconvolution, Combine(
////////////////////////////////////////////////////////////////////////////////
// LRN
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
Vec3i
,
int
,
Vec3f
,
bool
,
std
::
string
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
LRN
;
typedef
TestWithParam
<
tuple
<
Vec3i
,
int
,
Vec3f
,
bool
,
std
::
string
,
tuple
<
Backend
,
Target
>
>
>
LRN
;
TEST_P
(
LRN
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
())[
0
];
...
...
@@ -265,7 +265,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, LRN, Combine(
////////////////////////////////////////////////////////////////////////////////
// Average pooling
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
int
,
Size
,
Size
,
Size
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
AvePooling
;
typedef
TestWithParam
<
tuple
<
int
,
Size
,
Size
,
Size
,
tuple
<
Backend
,
Target
>
>
>
AvePooling
;
TEST_P
(
AvePooling
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
());
...
...
@@ -305,7 +305,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, AvePooling, Combine(
////////////////////////////////////////////////////////////////////////////////
// Maximum pooling
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
int
,
Size
,
Size
,
Size
,
Size
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
MaxPooling
;
typedef
TestWithParam
<
tuple
<
int
,
Size
,
Size
,
Size
,
Size
,
tuple
<
Backend
,
Target
>
>
>
MaxPooling
;
TEST_P
(
MaxPooling
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
());
...
...
@@ -344,7 +344,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, MaxPooling, Combine(
////////////////////////////////////////////////////////////////////////////////
// Fully-connected
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
int
,
Size
,
int
,
bool
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
FullyConnected
;
typedef
TestWithParam
<
tuple
<
int
,
Size
,
int
,
bool
,
tuple
<
Backend
,
Target
>
>
>
FullyConnected
;
TEST_P
(
FullyConnected
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
());
...
...
@@ -387,7 +387,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, FullyConnected, Combine(
////////////////////////////////////////////////////////////////////////////////
// SoftMax
////////////////////////////////////////////////////////////////////////////////
typedef
TestWithParam
<
tuple
<
int
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
SoftMax
;
typedef
TestWithParam
<
tuple
<
int
,
tuple
<
Backend
,
Target
>
>
>
SoftMax
;
TEST_P
(
SoftMax
,
Accuracy
)
{
int
inChannels
=
get
<
0
>
(
GetParam
());
...
...
@@ -476,7 +476,7 @@ void testInPlaceActivation(LayerParams& lp, int backendId, int targetId)
test
(
input
,
net
,
backendId
,
targetId
);
}
typedef
TestWithParam
<
tuple
<
bool
,
bool
,
float
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
BatchNorm
;
typedef
TestWithParam
<
tuple
<
bool
,
bool
,
float
,
tuple
<
Backend
,
Target
>
>
>
BatchNorm
;
TEST_P
(
BatchNorm
,
Accuracy
)
{
bool
hasWeights
=
get
<
0
>
(
GetParam
());
...
...
@@ -511,7 +511,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, BatchNorm, Combine(
dnnBackendsAndTargetsWithHalide
()
));
typedef
TestWithParam
<
tuple
<
float
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
ReLU
;
typedef
TestWithParam
<
tuple
<
float
,
tuple
<
Backend
,
Target
>
>
>
ReLU
;
TEST_P
(
ReLU
,
Accuracy
)
{
float
negativeSlope
=
get
<
0
>
(
GetParam
());
...
...
@@ -530,7 +530,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, ReLU, Combine(
dnnBackendsAndTargetsWithHalide
()
));
typedef
TestWithParam
<
tuple
<
std
::
string
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
NoParamActivation
;
typedef
TestWithParam
<
tuple
<
std
::
string
,
tuple
<
Backend
,
Target
>
>
>
NoParamActivation
;
TEST_P
(
NoParamActivation
,
Accuracy
)
{
int
backendId
=
get
<
0
>
(
get
<
1
>
(
GetParam
()));
...
...
@@ -546,7 +546,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, NoParamActivation, Combine(
dnnBackendsAndTargetsWithHalide
()
));
typedef
TestWithParam
<
tuple
<
Vec3f
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
Power
;
typedef
TestWithParam
<
tuple
<
Vec3f
,
tuple
<
Backend
,
Target
>
>
>
Power
;
TEST_P
(
Power
,
Accuracy
)
{
float
power
=
get
<
0
>
(
GetParam
())[
0
];
...
...
@@ -582,7 +582,7 @@ TEST_P(Test_Halide_layers, ChannelsPReLU)
testInPlaceActivation
(
lp
,
backend
,
target
);
}
typedef
TestWithParam
<
tuple
<
bool
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
Scale
;
typedef
TestWithParam
<
tuple
<
bool
,
tuple
<
Backend
,
Target
>
>
>
Scale
;
TEST_P
(
Scale
,
Accuracy
)
{
bool
hasBias
=
get
<
0
>
(
GetParam
());
...
...
@@ -616,7 +616,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, Scale, Combine(
// `--- conv ----^ ^ ^
// `---- ... ------' '
// `-----------------'
typedef
TestWithParam
<
tuple
<
Vec3i
,
Vec3i
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
Concat
;
typedef
TestWithParam
<
tuple
<
Vec3i
,
Vec3i
,
tuple
<
Backend
,
Target
>
>
>
Concat
;
TEST_P
(
Concat
,
Accuracy
)
{
Vec3i
inSize
=
get
<
0
>
(
GetParam
());
...
...
@@ -682,7 +682,7 @@ INSTANTIATE_TEST_CASE_P(Layer_Test_Halide, Concat, Combine(
// `--- conv ----^ ^ ^
// `---- ... ------' '
// `-----------------'
typedef
TestWithParam
<
tuple
<
Vec3i
,
std
::
string
,
int
,
bool
,
tuple
<
DNNBackend
,
DNN
Target
>
>
>
Eltwise
;
typedef
TestWithParam
<
tuple
<
Vec3i
,
std
::
string
,
int
,
bool
,
tuple
<
Backend
,
Target
>
>
>
Eltwise
;
TEST_P
(
Eltwise
,
Accuracy
)
{
Vec3i
inSize
=
get
<
0
>
(
GetParam
());
...
...
modules/dnn/test/test_precomp.hpp
浏览文件 @
452fa301
...
...
@@ -49,15 +49,41 @@
#include "opencv2/dnn.hpp"
#include "test_common.hpp"
namespace
opencv_test
{
namespace
{
using
namespace
cv
::
dnn
;
namespace
cv
{
namespace
dnn
{
CV__DNN_EXPERIMENTAL_NS_BEGIN
static
inline
void
PrintTo
(
const
cv
::
dnn
::
Backend
&
v
,
std
::
ostream
*
os
)
{
switch
(
v
)
{
case
DNN_BACKEND_DEFAULT
:
*
os
<<
"DNN_BACKEND_DEFAULT"
;
return
;
case
DNN_BACKEND_HALIDE
:
*
os
<<
"DNN_BACKEND_HALIDE"
;
return
;
case
DNN_BACKEND_INFERENCE_ENGINE
:
*
os
<<
"DNN_BACKEND_INFERENCE_ENGINE"
;
return
;
case
DNN_BACKEND_OPENCV
:
*
os
<<
"DNN_BACKEND_OPENCV"
;
return
;
}
// don't use "default:" to emit compiler warnings
*
os
<<
"DNN_BACKEND_UNKNOWN("
<<
v
<<
")"
;
}
static
inline
void
PrintTo
(
const
cv
::
dnn
::
Target
&
v
,
std
::
ostream
*
os
)
{
switch
(
v
)
{
case
DNN_TARGET_CPU
:
*
os
<<
"DNN_TARGET_CPU"
;
return
;
case
DNN_TARGET_OPENCL
:
*
os
<<
"DNN_TARGET_OPENCL"
;
return
;
case
DNN_TARGET_OPENCL_FP16
:
*
os
<<
"DNN_TARGET_OPENCL_FP16"
;
return
;
case
DNN_TARGET_MYRIAD
:
*
os
<<
"DNN_TARGET_MYRIAD"
;
return
;
}
// don't use "default:" to emit compiler warnings
*
os
<<
"DNN_TARGET_UNKNOWN("
<<
v
<<
")"
;
}
CV_
ENUM
(
DNNBackend
,
DNN_BACKEND_DEFAULT
,
DNN_BACKEND_HALIDE
,
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_BACKEND_OPENCV
)
CV_ENUM
(
DNNTarget
,
DNN_TARGET_CPU
,
DNN_TARGET_OPENCL
,
DNN_TARGET_OPENCL_FP16
,
DNN_TARGET_MYRIAD
)
CV_
_DNN_EXPERIMENTAL_NS_END
}}
// namespace
static
testing
::
internal
::
ParamGenerator
<
DNNTarget
>
availableDnnTargets
()
namespace
opencv_test
{
using
namespace
cv
::
dnn
;
static
testing
::
internal
::
ParamGenerator
<
Target
>
availableDnnTargets
()
{
static
std
::
vector
<
DNN
Target
>
targets
;
static
std
::
vector
<
Target
>
targets
;
if
(
targets
.
empty
())
{
targets
.
push_back
(
DNN_TARGET_CPU
);
...
...
@@ -69,23 +95,23 @@ static testing::internal::ParamGenerator<DNNTarget> availableDnnTargets()
return
testing
::
ValuesIn
(
targets
);
}
static
testing
::
internal
::
ParamGenerator
<
tuple
<
DNNBackend
,
DNN
Target
>
>
dnnBackendsAndTargets
()
static
testing
::
internal
::
ParamGenerator
<
tuple
<
Backend
,
Target
>
>
dnnBackendsAndTargets
()
{
static
const
tuple
<
DNNBackend
,
DNN
Target
>
testCases
[]
=
{
static
const
tuple
<
Backend
,
Target
>
testCases
[]
=
{
#ifdef HAVE_INF_ENGINE
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_CPU
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL_FP16
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_MYRIAD
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_CPU
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_OPENCL_FP16
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_INFERENCE_ENGINE
,
DNN_TARGET_MYRIAD
),
#endif
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_CPU
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL
),
tuple
<
DNNBackend
,
DNN
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL_FP16
)
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_CPU
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL
),
tuple
<
Backend
,
Target
>
(
DNN_BACKEND_OPENCV
,
DNN_TARGET_OPENCL_FP16
)
};
return
testing
::
ValuesIn
(
testCases
);
}
class
DNNTestLayer
:
public
TestWithParam
<
tuple
<
DNNBackend
,
DNN
Target
>
>
class
DNNTestLayer
:
public
TestWithParam
<
tuple
<
Backend
,
Target
>
>
{
public:
dnn
::
Backend
backend
;
...
...
@@ -156,6 +182,5 @@ protected:
}
};
}}
}
// namespace
#endif
modules/dnn/test/test_tf_importer.cpp
浏览文件 @
452fa301
...
...
@@ -243,7 +243,7 @@ TEST_P(Test_TensorFlow_layers, l2_normalize_3d)
runTensorFlowNet
(
"l2_normalize_3d"
);
}
typedef
testing
::
TestWithParam
<
DNN
Target
>
Test_TensorFlow_nets
;
typedef
testing
::
TestWithParam
<
Target
>
Test_TensorFlow_nets
;
TEST_P
(
Test_TensorFlow_nets
,
MobileNet_SSD
)
{
...
...
modules/dnn/test/test_torch_importer.cpp
浏览文件 @
452fa301
...
...
@@ -100,7 +100,7 @@ static void runTorchNet(String prefix, int targetId = DNN_TARGET_CPU, String out
}
}
typedef
testing
::
TestWithParam
<
DNN
Target
>
Test_Torch_layers
;
typedef
testing
::
TestWithParam
<
Target
>
Test_Torch_layers
;
TEST_P
(
Test_Torch_layers
,
run_convolution
)
{
...
...
@@ -208,7 +208,7 @@ TEST_P(Test_Torch_layers, net_non_spatial)
INSTANTIATE_TEST_CASE_P
(
/**/
,
Test_Torch_layers
,
availableDnnTargets
());
typedef
testing
::
TestWithParam
<
DNN
Target
>
Test_Torch_nets
;
typedef
testing
::
TestWithParam
<
Target
>
Test_Torch_nets
;
TEST_P
(
Test_Torch_nets
,
OpenFace_accuracy
)
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录