Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleDetection
提交
d383fd09
P
PaddleDetection
项目概览
PaddlePaddle
/
PaddleDetection
大约 1 年 前同步成功
通知
695
Star
11112
Fork
2696
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
184
列表
看板
标记
里程碑
合并请求
40
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
184
Issue
184
列表
看板
标记
里程碑
合并请求
40
合并请求
40
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
d383fd09
编写于
8月 27, 2020
作者:
W
wangguanzhong
提交者:
GitHub
8月 27, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
add yolov3_darknet (#1299)
上级
646996f4
变更
16
隐藏空白更改
内联
并排
Showing
16 changed file
with
240 addition
and
241 deletion
+240
-241
configs/mask_rcnn_r50_fpn_1x.yml
configs/mask_rcnn_r50_fpn_1x.yml
+1
-1
configs/yolov3_darknet.yml
configs/yolov3_darknet.yml
+7
-8
configs/yolov3_reader.yml
configs/yolov3_reader.yml
+5
-5
ppdet/data/reader.py
ppdet/data/reader.py
+1
-1
ppdet/modeling/architecture/yolo.py
ppdet/modeling/architecture/yolo.py
+14
-18
ppdet/modeling/backbone/darknet.py
ppdet/modeling/backbone/darknet.py
+72
-35
ppdet/modeling/bbox.py
ppdet/modeling/bbox.py
+13
-29
ppdet/modeling/head/mask_head.py
ppdet/modeling/head/mask_head.py
+1
-1
ppdet/modeling/head/yolo_head.py
ppdet/modeling/head/yolo_head.py
+90
-111
ppdet/modeling/ops.py
ppdet/modeling/ops.py
+10
-18
ppdet/optimizer.py
ppdet/optimizer.py
+1
-1
ppdet/py_op/bbox.py
ppdet/py_op/bbox.py
+1
-1
ppdet/py_op/post_process.py
ppdet/py_op/post_process.py
+2
-1
ppdet/utils/checkpoint.py
ppdet/utils/checkpoint.py
+6
-3
tools/eval.py
tools/eval.py
+13
-5
tools/train.py
tools/train.py
+3
-3
未找到文件。
configs/mask_rcnn_r50_fpn_1x.yml
浏览文件 @
d383fd09
...
@@ -130,7 +130,7 @@ LearningRate:
...
@@ -130,7 +130,7 @@ LearningRate:
gamma
:
0.1
gamma
:
0.1
milestones
:
[
120000
,
160000
]
milestones
:
[
120000
,
160000
]
-
!LinearWarmup
-
!LinearWarmup
start_factor
:
0.3333
start_factor
:
0.3333
333
steps
:
500
steps
:
500
OptimizerBuilder
:
OptimizerBuilder
:
...
...
configs/yolov3_darknet.yml
浏览文件 @
d383fd09
...
@@ -3,13 +3,13 @@ use_gpu: true
...
@@ -3,13 +3,13 @@ use_gpu: true
max_iters
:
500000
max_iters
:
500000
log_smooth_window
:
20
log_smooth_window
:
20
save_dir
:
output
save_dir
:
output
snapshot_iter
:
1
0000
snapshot_iter
:
5
0000
metric
:
COCO
metric
:
COCO
pretrain_weights
:
https://paddle
models.bj.bcebos.com/yolo/darknet53.pdparams
pretrain_weights
:
https://paddle
-imagenet-models-name.bj.bcebos.com/DarkNet53_pretrained.tar
weights
:
output/yolov3_darknet/model_final
weights
:
output/yolov3_darknet/model_final
num_classes
:
80
num_classes
:
80
use_fine_grained_loss
:
false
use_fine_grained_loss
:
false
open_debug
:
Fals
e
load_static_weights
:
Tru
e
YOLOv3
:
YOLOv3
:
anchor
:
AnchorYOLO
anchor
:
AnchorYOLO
...
@@ -18,11 +18,15 @@ YOLOv3:
...
@@ -18,11 +18,15 @@ YOLOv3:
DarkNet
:
DarkNet
:
depth
:
53
depth
:
53
return_idx
:
[
2
,
3
,
4
]
YOLOv3Head
:
YOLOv3Head
:
yolo_feat
:
yolo_feat
:
name
:
YOLOFeat
name
:
YOLOFeat
feat_in_list
:
[
1024
,
768
,
384
]
feat_in_list
:
[
1024
,
768
,
384
]
ignore_thresh
:
0.7
downsample
:
32
label_smooth
:
true
anchor_per_position
:
3
anchor_per_position
:
3
AnchorYOLO
:
AnchorYOLO
:
...
@@ -30,11 +34,6 @@ AnchorYOLO:
...
@@ -30,11 +34,6 @@ AnchorYOLO:
name
:
AnchorGeneratorYOLO
name
:
AnchorGeneratorYOLO
anchors
:
[
10
,
13
,
16
,
30
,
33
,
23
,
30
,
61
,
62
,
45
,
59
,
119
,
116
,
90
,
156
,
198
,
373
,
326
]
anchors
:
[
10
,
13
,
16
,
30
,
33
,
23
,
30
,
61
,
62
,
45
,
59
,
119
,
116
,
90
,
156
,
198
,
373
,
326
]
anchor_masks
:
[[
6
,
7
,
8
],
[
3
,
4
,
5
],
[
0
,
1
,
2
]]
anchor_masks
:
[[
6
,
7
,
8
],
[
3
,
4
,
5
],
[
0
,
1
,
2
]]
anchor_target_generator
:
name
:
AnchorTargetGeneratorYOLO
ignore_thresh
:
0.7
downsample_ratio
:
32
label_smooth
:
true
anchor_post_process
:
anchor_post_process
:
name
:
BBoxPostProcessYOLO
name
:
BBoxPostProcessYOLO
# decode -> clip
# decode -> clip
...
...
configs/yolov3_reader.yml
浏览文件 @
d383fd09
...
@@ -27,8 +27,8 @@ TrainReader:
...
@@ -27,8 +27,8 @@ TrainReader:
-
!BboxXYXY2XYWH
{}
-
!BboxXYXY2XYWH
{}
batch_transforms
:
batch_transforms
:
-
!RandomShape
-
!RandomShape
sizes
:
[
320
,
352
,
384
,
416
,
448
,
480
,
512
,
544
,
576
,
608
]
sizes
:
[
320
,
352
,
384
,
416
,
448
,
480
,
512
,
544
,
576
,
608
]
random_inter
:
True
random_inter
:
True
-
!NormalizeImage
-
!NormalizeImage
mean
:
[
0.485
,
0.456
,
0.406
]
mean
:
[
0.485
,
0.456
,
0.406
]
std
:
[
0.229
,
0.224
,
0.225
]
std
:
[
0.229
,
0.224
,
0.225
]
...
@@ -50,8 +50,8 @@ TrainReader:
...
@@ -50,8 +50,8 @@ TrainReader:
shuffle
:
true
shuffle
:
true
mixup_epoch
:
250
mixup_epoch
:
250
drop_last
:
true
drop_last
:
true
worker_num
:
8
worker_num
:
4
bufsize
:
16
bufsize
:
4
use_process
:
true
use_process
:
true
...
@@ -81,7 +81,7 @@ EvalReader:
...
@@ -81,7 +81,7 @@ EvalReader:
-
!Permute
-
!Permute
to_bgr
:
false
to_bgr
:
false
channel_first
:
True
channel_first
:
True
batch_size
:
8
batch_size
:
1
drop_empty
:
false
drop_empty
:
false
worker_num
:
8
worker_num
:
8
bufsize
:
16
bufsize
:
16
...
...
ppdet/data/reader.py
浏览文件 @
d383fd09
...
@@ -201,7 +201,7 @@ class Reader(object):
...
@@ -201,7 +201,7 @@ class Reader(object):
use_fine_grained_loss
=
False
,
use_fine_grained_loss
=
False
,
num_classes
=
80
,
num_classes
=
80
,
bufsize
=-
1
,
bufsize
=-
1
,
memsize
=
'
3G
'
,
memsize
=
'
500M
'
,
inputs_def
=
None
,
inputs_def
=
None
,
devices_num
=
1
):
devices_num
=
1
):
self
.
_dataset
=
dataset
self
.
_dataset
=
dataset
...
...
ppdet/modeling/architecture/yolo.py
浏览文件 @
d383fd09
...
@@ -17,38 +17,34 @@ class YOLOv3(BaseArch):
...
@@ -17,38 +17,34 @@ class YOLOv3(BaseArch):
'yolo_head'
,
'yolo_head'
,
]
]
def
__init__
(
self
,
anchor
,
backbone
,
yolo_head
,
*
args
,
**
kwargs
):
def
__init__
(
self
,
anchor
,
backbone
,
yolo_head
):
super
(
YOLOv3
,
self
).
__init__
(
*
args
,
**
kwargs
)
super
(
YOLOv3
,
self
).
__init__
()
self
.
anchor
=
anchor
self
.
anchor
=
anchor
self
.
backbone
=
backbone
self
.
backbone
=
backbone
self
.
yolo_head
=
yolo_head
self
.
yolo_head
=
yolo_head
def
model_arch
(
self
,
):
def
model_arch
(
self
,
):
# Backbone
# Backbone
bb_out
=
self
.
backbone
(
self
.
gbd
)
body_feats
=
self
.
backbone
(
self
.
inputs
)
self
.
gbd
.
update
(
bb_out
)
# YOLO Head
# YOLO Head
yolo_head_out
=
self
.
yolo_head
(
self
.
gbd
)
self
.
yolo_head_out
=
self
.
yolo_head
(
body_feats
)
self
.
gbd
.
update
(
yolo_head_out
)
# Anchor
# Anchor
anchor_out
=
self
.
anchor
(
self
.
gbd
)
self
.
anchors
,
self
.
anchor_masks
,
self
.
mask_anchors
=
self
.
anchor
()
self
.
gbd
.
update
(
anchor_out
)
if
self
.
gbd
[
'mode'
]
==
'infer'
:
bbox_out
=
self
.
anchor
.
post_process
(
self
.
gbd
)
self
.
gbd
.
update
(
bbox_out
)
def
loss
(
self
,
):
def
loss
(
self
,
):
yolo_loss
=
self
.
yolo_head
.
loss
(
self
.
gbd
)
yolo_loss
=
self
.
yolo_head
.
loss
(
self
.
inputs
,
self
.
yolo_head_out
,
out
=
{
'loss'
:
yolo_loss
}
self
.
anchors
,
self
.
anchor_masks
,
return
out
self
.
mask_anchors
)
return
yolo_loss
def
infer
(
self
,
):
def
infer
(
self
,
):
bbox
,
bbox_num
=
self
.
anchor
.
post_process
(
self
.
inputs
[
'im_size'
],
self
.
yolo_head_out
,
self
.
mask_anchors
)
outs
=
{
outs
=
{
"bbox"
:
self
.
gbd
[
'predicted_bbox'
]
.
numpy
(),
"bbox"
:
bbox
.
numpy
(),
"bbox_num
s"
:
self
.
gbd
[
'predicted_bbox_nums'
]
,
"bbox_num
"
:
bbox_num
,
'im_id'
:
self
.
gbd
[
'im_id'
].
numpy
()
'im_id'
:
self
.
inputs
[
'im_id'
].
numpy
()
}
}
return
outs
return
outs
ppdet/modeling/backbone/darknet.py
浏览文件 @
d383fd09
...
@@ -16,7 +16,8 @@ class ConvBNLayer(Layer):
...
@@ -16,7 +16,8 @@ class ConvBNLayer(Layer):
stride
=
1
,
stride
=
1
,
groups
=
1
,
groups
=
1
,
padding
=
0
,
padding
=
0
,
act
=
"leaky"
):
act
=
"leaky"
,
name
=
None
):
super
(
ConvBNLayer
,
self
).
__init__
()
super
(
ConvBNLayer
,
self
).
__init__
()
self
.
conv
=
Conv2D
(
self
.
conv
=
Conv2D
(
...
@@ -26,18 +27,18 @@ class ConvBNLayer(Layer):
...
@@ -26,18 +27,18 @@ class ConvBNLayer(Layer):
stride
=
stride
,
stride
=
stride
,
padding
=
padding
,
padding
=
padding
,
groups
=
groups
,
groups
=
groups
,
param_attr
=
ParamAttr
(
param_attr
=
ParamAttr
(
name
=
name
+
'.conv.weights'
),
initializer
=
fluid
.
initializer
.
Normal
(
0.
,
0.02
)),
bias_attr
=
False
,
bias_attr
=
False
,
act
=
None
)
act
=
None
)
bn_name
=
name
+
'.bn'
self
.
batch_norm
=
BatchNorm
(
self
.
batch_norm
=
BatchNorm
(
num_channels
=
ch_out
,
num_channels
=
ch_out
,
param_attr
=
ParamAttr
(
param_attr
=
ParamAttr
(
initializer
=
fluid
.
initializer
.
Normal
(
0.
,
0.02
),
name
=
bn_name
+
'.scale'
,
regularizer
=
L2Decay
(
0.
)),
regularizer
=
L2Decay
(
0.
)),
bias_attr
=
ParamAttr
(
bias_attr
=
ParamAttr
(
initializer
=
fluid
.
initializer
.
Constant
(
0.0
),
name
=
bn_name
+
'.offset'
,
regularizer
=
L2Decay
(
0.
)),
regularizer
=
L2Decay
(
0.
)))
moving_mean_name
=
bn_name
+
'.mean'
,
moving_variance_name
=
bn_name
+
'.var'
)
self
.
act
=
act
self
.
act
=
act
...
@@ -50,7 +51,13 @@ class ConvBNLayer(Layer):
...
@@ -50,7 +51,13 @@ class ConvBNLayer(Layer):
class
DownSample
(
Layer
):
class
DownSample
(
Layer
):
def
__init__
(
self
,
ch_in
,
ch_out
,
filter_size
=
3
,
stride
=
2
,
padding
=
1
):
def
__init__
(
self
,
ch_in
,
ch_out
,
filter_size
=
3
,
stride
=
2
,
padding
=
1
,
name
=
None
):
super
(
DownSample
,
self
).
__init__
()
super
(
DownSample
,
self
).
__init__
()
...
@@ -59,7 +66,8 @@ class DownSample(Layer):
...
@@ -59,7 +66,8 @@ class DownSample(Layer):
ch_out
=
ch_out
,
ch_out
=
ch_out
,
filter_size
=
filter_size
,
filter_size
=
filter_size
,
stride
=
stride
,
stride
=
stride
,
padding
=
padding
)
padding
=
padding
,
name
=
name
)
self
.
ch_out
=
ch_out
self
.
ch_out
=
ch_out
def
forward
(
self
,
inputs
):
def
forward
(
self
,
inputs
):
...
@@ -68,13 +76,23 @@ class DownSample(Layer):
...
@@ -68,13 +76,23 @@ class DownSample(Layer):
class
BasicBlock
(
Layer
):
class
BasicBlock
(
Layer
):
def
__init__
(
self
,
ch_in
,
ch_out
):
def
__init__
(
self
,
ch_in
,
ch_out
,
name
=
None
):
super
(
BasicBlock
,
self
).
__init__
()
super
(
BasicBlock
,
self
).
__init__
()
self
.
conv1
=
ConvBNLayer
(
self
.
conv1
=
ConvBNLayer
(
ch_in
=
ch_in
,
ch_out
=
ch_out
,
filter_size
=
1
,
stride
=
1
,
padding
=
0
)
ch_in
=
ch_in
,
ch_out
=
ch_out
,
filter_size
=
1
,
stride
=
1
,
padding
=
0
,
name
=
name
+
'.0'
)
self
.
conv2
=
ConvBNLayer
(
self
.
conv2
=
ConvBNLayer
(
ch_in
=
ch_out
,
ch_out
=
ch_out
*
2
,
filter_size
=
3
,
stride
=
1
,
padding
=
1
)
ch_in
=
ch_out
,
ch_out
=
ch_out
*
2
,
filter_size
=
3
,
stride
=
1
,
padding
=
1
,
name
=
name
+
'.1'
)
def
forward
(
self
,
inputs
):
def
forward
(
self
,
inputs
):
conv1
=
self
.
conv1
(
inputs
)
conv1
=
self
.
conv1
(
inputs
)
...
@@ -84,14 +102,16 @@ class BasicBlock(Layer):
...
@@ -84,14 +102,16 @@ class BasicBlock(Layer):
class
Blocks
(
Layer
):
class
Blocks
(
Layer
):
def
__init__
(
self
,
ch_in
,
ch_out
,
count
):
def
__init__
(
self
,
ch_in
,
ch_out
,
count
,
name
=
None
):
super
(
Blocks
,
self
).
__init__
()
super
(
Blocks
,
self
).
__init__
()
self
.
basicblock0
=
BasicBlock
(
ch_in
,
ch_out
)
self
.
basicblock0
=
BasicBlock
(
ch_in
,
ch_out
,
name
=
name
+
'.0'
)
self
.
res_out_list
=
[]
self
.
res_out_list
=
[]
for
i
in
range
(
1
,
count
):
for
i
in
range
(
1
,
count
):
res_out
=
self
.
add_sublayer
(
"basic_block_%d"
%
(
i
),
block_name
=
'{}.{}'
.
format
(
name
,
i
)
BasicBlock
(
ch_out
*
2
,
ch_out
))
res_out
=
self
.
add_sublayer
(
block_name
,
BasicBlock
(
ch_out
*
2
,
ch_out
,
name
=
block_name
))
self
.
res_out_list
.
append
(
res_out
)
self
.
res_out_list
.
append
(
res_out
)
self
.
ch_out
=
ch_out
self
.
ch_out
=
ch_out
...
@@ -108,31 +128,46 @@ DarkNet_cfg = {53: ([1, 2, 8, 8, 4])}
...
@@ -108,31 +128,46 @@ DarkNet_cfg = {53: ([1, 2, 8, 8, 4])}
@
register
@
register
@
serializable
@
serializable
class
DarkNet
(
Layer
):
class
DarkNet
(
Layer
):
def
__init__
(
self
,
depth
=
53
,
mode
=
'train'
):
def
__init__
(
self
,
depth
=
53
,
freeze_at
=-
1
,
return_idx
=
[
2
,
3
,
4
],
num_stages
=
5
):
super
(
DarkNet
,
self
).
__init__
()
super
(
DarkNet
,
self
).
__init__
()
self
.
depth
=
depth
self
.
depth
=
depth
self
.
mode
=
mode
self
.
freeze_at
=
freeze_at
self
.
stages
=
DarkNet_cfg
[
self
.
depth
][
0
:
5
]
self
.
return_idx
=
return_idx
self
.
num_stages
=
num_stages
self
.
stages
=
DarkNet_cfg
[
self
.
depth
][
0
:
num_stages
]
self
.
conv0
=
ConvBNLayer
(
self
.
conv0
=
ConvBNLayer
(
ch_in
=
3
,
ch_out
=
32
,
filter_size
=
3
,
stride
=
1
,
padding
=
1
)
ch_in
=
3
,
ch_out
=
32
,
filter_size
=
3
,
stride
=
1
,
padding
=
1
,
name
=
'yolo_input'
)
self
.
downsample0
=
DownSample
(
ch_in
=
32
,
ch_out
=
32
*
2
)
self
.
downsample0
=
DownSample
(
ch_in
=
32
,
ch_out
=
32
*
2
,
name
=
'yolo_input.downsample'
)
self
.
darknet
53
_conv_block_list
=
[]
self
.
darknet_conv_block_list
=
[]
self
.
downsample_list
=
[]
self
.
downsample_list
=
[]
ch_in
=
[
64
,
128
,
256
,
512
,
1024
]
ch_in
=
[
64
,
128
,
256
,
512
,
1024
]
for
i
,
stage
in
enumerate
(
self
.
stages
):
for
i
,
stage
in
enumerate
(
self
.
stages
):
conv_block
=
self
.
add_sublayer
(
"stage_%d"
%
(
i
),
name
=
'stage.{}'
.
format
(
i
)
Blocks
(
conv_block
=
self
.
add_sublayer
(
int
(
ch_in
[
i
]),
32
*
(
2
**
i
),
name
,
Blocks
(
stage
))
int
(
ch_in
[
i
]),
32
*
(
2
**
i
),
stage
,
name
=
name
))
self
.
darknet53_conv_block_list
.
append
(
conv_block
)
self
.
darknet_conv_block_list
.
append
(
conv_block
)
for
i
in
range
(
len
(
self
.
stages
)
-
1
):
for
i
in
range
(
num_stages
-
1
):
down_name
=
'stage.{}.downsample'
.
format
(
i
)
downsample
=
self
.
add_sublayer
(
downsample
=
self
.
add_sublayer
(
"stage_%d_downsample"
%
i
,
down_name
,
DownSample
(
DownSample
(
ch_in
=
32
*
(
2
**
(
i
+
1
)),
ch_out
=
32
*
(
2
**
(
i
+
2
))))
ch_in
=
32
*
(
2
**
(
i
+
1
)),
ch_out
=
32
*
(
2
**
(
i
+
2
)),
name
=
down_name
))
self
.
downsample_list
.
append
(
downsample
)
self
.
downsample_list
.
append
(
downsample
)
def
forward
(
self
,
inputs
):
def
forward
(
self
,
inputs
):
...
@@ -141,10 +176,12 @@ class DarkNet(Layer):
...
@@ -141,10 +176,12 @@ class DarkNet(Layer):
out
=
self
.
conv0
(
x
)
out
=
self
.
conv0
(
x
)
out
=
self
.
downsample0
(
out
)
out
=
self
.
downsample0
(
out
)
blocks
=
[]
blocks
=
[]
for
i
,
conv_block_i
in
enumerate
(
self
.
darknet
53
_conv_block_list
):
for
i
,
conv_block_i
in
enumerate
(
self
.
darknet_conv_block_list
):
out
=
conv_block_i
(
out
)
out
=
conv_block_i
(
out
)
blocks
.
append
(
out
)
if
i
==
self
.
freeze_at
:
if
i
<
len
(
self
.
stages
)
-
1
:
out
.
stop_gradient
=
True
if
i
in
self
.
return_idx
:
blocks
.
append
(
out
)
if
i
<
self
.
num_stages
-
1
:
out
=
self
.
downsample_list
[
i
](
out
)
out
=
self
.
downsample_list
[
i
](
out
)
outs
=
{
'darknet_outs'
:
blocks
[
-
1
:
-
4
:
-
1
]}
return
blocks
return
outs
ppdet/modeling/bbox.py
浏览文件 @
d383fd09
...
@@ -79,27 +79,25 @@ class BBoxPostProcessYOLO(object):
...
@@ -79,27 +79,25 @@ class BBoxPostProcessYOLO(object):
self
.
decode
=
decode
self
.
decode
=
decode
self
.
clip
=
clip
self
.
clip
=
clip
def
__call__
(
self
,
i
nput
s
):
def
__call__
(
self
,
i
m_size
,
yolo_head_out
,
mask_anchor
s
):
# TODO: split yolo_box into 2 steps
# TODO: split yolo_box into 2 steps
# decode
# decode
# clip
# clip
boxes_list
=
[]
boxes_list
=
[]
scores_list
=
[]
scores_list
=
[]
for
i
,
out
in
enumerate
(
inputs
[
'yolo_outs'
]):
for
i
,
head_out
in
enumerate
(
yolo_head_out
):
boxes
,
scores
=
self
.
yolo_box
(
out
,
inputs
[
'im_size'
],
boxes
,
scores
=
self
.
yolo_box
(
head_out
,
im_size
,
mask_anchors
[
i
],
inputs
[
'mask_anchors'
][
i
],
i
,
self
.
num_classes
,
i
)
"yolo_box_"
+
str
(
i
))
boxes_list
.
append
(
boxes
)
boxes_list
.
append
(
boxes
)
scores_list
.
append
(
fluid
.
layers
.
transpose
(
scores
,
perm
=
[
0
,
2
,
1
]))
scores_list
.
append
(
fluid
.
layers
.
transpose
(
scores
,
perm
=
[
0
,
2
,
1
]))
yolo_boxes
=
fluid
.
layers
.
concat
(
boxes_list
,
axis
=
1
)
yolo_boxes
=
fluid
.
layers
.
concat
(
boxes_list
,
axis
=
1
)
yolo_scores
=
fluid
.
layers
.
concat
(
scores_list
,
axis
=
2
)
yolo_scores
=
fluid
.
layers
.
concat
(
scores_list
,
axis
=
2
)
nmsed_
bbox
=
self
.
nms
(
bboxes
=
yolo_boxes
,
scores
=
yolo_scores
)
bbox
=
self
.
nms
(
bboxes
=
yolo_boxes
,
scores
=
yolo_scores
)
# TODO: parse the lod of nmsed_bbox
# TODO: parse the lod of nmsed_bbox
# default batch size is 1
# default batch size is 1
bbox_nums
=
np
.
array
([
0
,
int
(
nmsed_bbox
.
shape
[
0
])],
dtype
=
np
.
int32
)
bbox_num
=
np
.
array
([
int
(
bbox
.
shape
[
0
])],
dtype
=
np
.
int32
)
outs
=
{
"predicted_bbox_nums"
:
bbox_nums
,
"predicted_bbox"
:
nmsed_bbox
}
return
bbox
,
bbox_num
return
outs
@
register
@
register
...
@@ -168,32 +166,18 @@ class AnchorRPN(object):
...
@@ -168,32 +166,18 @@ class AnchorRPN(object):
@
register
@
register
class
AnchorYOLO
(
object
):
class
AnchorYOLO
(
object
):
__inject__
=
[
__inject__
=
[
'anchor_generator'
,
'anchor_post_process'
]
'anchor_generator'
,
'anchor_target_generator'
,
'anchor_post_process'
]
def
__init__
(
self
,
anchor_generator
,
anchor_target_generator
,
def
__init__
(
self
,
anchor_generator
,
anchor_post_process
):
anchor_post_process
):
super
(
AnchorYOLO
,
self
).
__init__
()
super
(
AnchorYOLO
,
self
).
__init__
()
self
.
anchor_generator
=
anchor_generator
self
.
anchor_generator
=
anchor_generator
self
.
anchor_target_generator
=
anchor_target_generator
self
.
anchor_post_process
=
anchor_post_process
self
.
anchor_post_process
=
anchor_post_process
def
__call__
(
self
,
inputs
):
def
__call__
(
self
):
outs
=
self
.
generate_anchors
(
inputs
)
return
self
.
anchor_generator
()
return
outs
def
generate_anchors
(
self
,
inputs
):
outs
=
self
.
anchor_generator
(
inputs
[
'yolo_outs'
])
outs
[
'anchor_module'
]
=
self
return
outs
def
generate_anchors_target
(
self
,
inputs
):
outs
=
self
.
anchor_target_generator
()
return
outs
def
post_process
(
self
,
i
nput
s
):
def
post_process
(
self
,
i
m_size
,
yolo_head_out
,
mask_anchor
s
):
return
self
.
anchor_post_process
(
i
nput
s
)
return
self
.
anchor_post_process
(
i
m_size
,
yolo_head_out
,
mask_anchor
s
)
@
register
@
register
...
...
ppdet/modeling/head/mask_head.py
浏览文件 @
d383fd09
...
@@ -37,7 +37,7 @@ class MaskFeat(Layer):
...
@@ -37,7 +37,7 @@ class MaskFeat(Layer):
mask_conv
.
add_sublayer
(
mask_conv
.
add_sublayer
(
conv_name
,
conv_name
,
Conv2D
(
Conv2D
(
num_channels
=
feat_in
if
j
==
1
else
feat_out
,
num_channels
=
feat_in
if
j
==
0
else
feat_out
,
num_filters
=
feat_out
,
num_filters
=
feat_out
,
filter_size
=
3
,
filter_size
=
3
,
act
=
'relu'
,
act
=
'relu'
,
...
...
ppdet/modeling/head/yolo_head.py
浏览文件 @
d383fd09
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
import
paddle
from
paddle.fluid.dygraph
import
Layer
from
paddle.fluid.dygraph
import
Layer
from
paddle.fluid.param_attr
import
ParamAttr
from
paddle.fluid.param_attr
import
ParamAttr
from
paddle.fluid.initializer
import
Normal
from
paddle.fluid.initializer
import
Normal
from
paddle.fluid.regularizer
import
L2Decay
from
paddle.fluid.regularizer
import
L2Decay
from
paddle.fluid.dygraph.nn
import
Conv2D
,
BatchNorm
from
paddle.fluid.dygraph.nn
import
Conv2D
,
BatchNorm
from
paddle.fluid.dygraph
import
Sequential
from
ppdet.core.workspace
import
register
from
ppdet.core.workspace
import
register
from
..backbone.darknet
import
ConvBNLayer
from
..backbone.darknet
import
ConvBNLayer
class
YoloDetBlock
(
Layer
):
class
YoloDetBlock
(
Layer
):
def
__init__
(
self
,
ch_in
,
channel
):
def
__init__
(
self
,
ch_in
,
channel
,
name
):
super
(
YoloDetBlock
,
self
).
__init__
()
super
(
YoloDetBlock
,
self
).
__init__
()
self
.
ch_in
=
ch_in
self
.
channel
=
channel
assert
channel
%
2
==
0
,
\
assert
channel
%
2
==
0
,
\
"channel {} cannot be divided by 2"
.
format
(
channel
)
"channel {} cannot be divided by 2"
.
format
(
channel
)
conv_def
=
[
self
.
conv0
=
ConvBNLayer
(
[
'conv0'
,
ch_in
,
channel
,
1
,
'.0.0'
],
ch_in
=
ch_in
,
ch_out
=
channel
,
filter_size
=
1
,
stride
=
1
,
padding
=
0
)
[
'conv1'
,
channel
,
channel
*
2
,
3
,
'.0.1'
],
[
'conv2'
,
channel
*
2
,
channel
,
1
,
'.1.0'
],
self
.
conv1
=
ConvBNLayer
(
[
'conv3'
,
channel
,
channel
*
2
,
3
,
'.1.1'
],
ch_in
=
channel
,
[
'route'
,
channel
*
2
,
channel
,
1
,
'.2'
],
ch_out
=
channel
*
2
,
#['tip', channel, channel * 2, 3],
filter_size
=
3
,
]
stride
=
1
,
padding
=
1
)
self
.
conv_module
=
Sequential
()
for
idx
,
(
conv_name
,
ch_in
,
ch_out
,
filter_size
,
self
.
conv2
=
ConvBNLayer
(
post_name
)
in
enumerate
(
conv_def
):
ch_in
=
channel
*
2
,
self
.
conv_module
.
add_sublayer
(
ch_out
=
channel
,
conv_name
,
filter_size
=
1
,
ConvBNLayer
(
stride
=
1
,
ch_in
=
ch_in
,
padding
=
0
)
ch_out
=
ch_out
,
filter_size
=
filter_size
,
self
.
conv3
=
ConvBNLayer
(
padding
=
(
filter_size
-
1
)
//
2
,
ch_in
=
channel
,
name
=
name
+
post_name
))
ch_out
=
channel
*
2
,
filter_size
=
3
,
stride
=
1
,
padding
=
1
)
self
.
route
=
ConvBNLayer
(
ch_in
=
channel
*
2
,
ch_out
=
channel
,
filter_size
=
1
,
stride
=
1
,
padding
=
0
)
self
.
tip
=
ConvBNLayer
(
self
.
tip
=
ConvBNLayer
(
ch_in
=
channel
,
ch_in
=
channel
,
ch_out
=
channel
*
2
,
ch_out
=
channel
*
2
,
filter_size
=
3
,
filter_size
=
3
,
stride
=
1
,
padding
=
1
,
padding
=
1
)
name
=
name
+
'.tip'
)
def
forward
(
self
,
inputs
):
def
forward
(
self
,
inputs
):
out
=
self
.
conv0
(
inputs
)
route
=
self
.
conv_module
(
inputs
)
out
=
self
.
conv1
(
out
)
out
=
self
.
conv2
(
out
)
out
=
self
.
conv3
(
out
)
route
=
self
.
route
(
out
)
tip
=
self
.
tip
(
route
)
tip
=
self
.
tip
(
route
)
return
route
,
tip
return
route
,
tip
class
Upsample
(
Layer
):
def
__init__
(
self
,
scale
=
2
):
super
(
Upsample
,
self
).
__init__
()
self
.
scale
=
scale
def
forward
(
self
,
inputs
):
# get dynamic upsample output shape
shape_nchw
=
fluid
.
layers
.
shape
(
inputs
)
shape_hw
=
fluid
.
layers
.
slice
(
shape_nchw
,
axes
=
[
0
],
starts
=
[
2
],
ends
=
[
4
])
shape_hw
.
stop_gradient
=
True
in_shape
=
fluid
.
layers
.
cast
(
shape_hw
,
dtype
=
'int32'
)
out_shape
=
in_shape
*
self
.
scale
out_shape
.
stop_gradient
=
True
# reisze by actual_shape
out
=
fluid
.
layers
.
resize_nearest
(
input
=
inputs
,
scale
=
self
.
scale
,
actual_shape
=
out_shape
)
return
out
@
register
@
register
class
YOLOFeat
(
Layer
):
class
YOLOFeat
(
Layer
):
def
__init__
(
self
,
feat_in_list
=
[
1024
,
768
,
384
]):
__shared__
=
[
'num_levels'
]
def
__init__
(
self
,
feat_in_list
=
[
1024
,
768
,
384
],
num_levels
=
3
):
super
(
YOLOFeat
,
self
).
__init__
()
super
(
YOLOFeat
,
self
).
__init__
()
self
.
feat_in_list
=
feat_in_list
self
.
feat_in_list
=
feat_in_list
self
.
yolo_blocks
=
[]
self
.
yolo_blocks
=
[]
self
.
route_blocks
=
[]
self
.
route_blocks
=
[]
for
i
in
range
(
3
):
self
.
num_levels
=
num_levels
for
i
in
range
(
self
.
num_levels
):
name
=
'yolo_block.{}'
.
format
(
i
)
yolo_block
=
self
.
add_sublayer
(
yolo_block
=
self
.
add_sublayer
(
"yolo_det_block_%d"
%
(
i
)
,
name
,
YoloDetBlock
(
YoloDetBlock
(
feat_in_list
[
i
],
channel
=
512
//
(
2
**
i
)))
feat_in_list
[
i
],
channel
=
512
//
(
2
**
i
)
,
name
=
name
))
self
.
yolo_blocks
.
append
(
yolo_block
)
self
.
yolo_blocks
.
append
(
yolo_block
)
if
i
<
2
:
if
i
<
self
.
num_levels
-
1
:
name
=
'yolo_transition.{}'
.
format
(
i
)
route
=
self
.
add_sublayer
(
route
=
self
.
add_sublayer
(
"route_%d"
%
i
,
name
,
ConvBNLayer
(
ConvBNLayer
(
ch_in
=
512
//
(
2
**
i
),
ch_in
=
512
//
(
2
**
i
),
ch_out
=
256
//
(
2
**
i
),
ch_out
=
256
//
(
2
**
i
),
filter_size
=
1
,
filter_size
=
1
,
stride
=
1
,
stride
=
1
,
padding
=
0
))
padding
=
0
,
name
=
name
))
self
.
route_blocks
.
append
(
route
)
self
.
route_blocks
.
append
(
route
)
self
.
upsample
=
Upsample
()
def
forward
(
self
,
inputs
):
def
forward
(
self
,
body_feats
):
assert
len
(
body_feats
)
==
self
.
num_levels
body_feats
=
body_feats
[::
-
1
]
yolo_feats
=
[]
yolo_feats
=
[]
for
i
,
block
in
enumerate
(
inputs
[
'darknet_outs'
]
):
for
i
,
block
in
enumerate
(
body_feats
):
if
i
>
0
:
if
i
>
0
:
block
=
fluid
.
layers
.
concat
(
input
=
[
route
,
block
],
axis
=
1
)
block
=
fluid
.
layers
.
concat
(
input
=
[
route
,
block
],
axis
=
1
)
route
,
tip
=
self
.
yolo_blocks
[
i
](
block
)
route
,
tip
=
self
.
yolo_blocks
[
i
](
block
)
yolo_feats
.
append
(
tip
)
yolo_feats
.
append
(
tip
)
if
i
<
2
:
if
i
<
self
.
num_levels
-
1
:
route
=
self
.
route_blocks
[
i
](
route
)
route
=
self
.
route_blocks
[
i
](
route
)
route
=
self
.
upsample
(
route
)
route
=
fluid
.
layers
.
resize_nearest
(
route
,
scale
=
2.
)
outs
=
{
'yolo_feat'
:
yolo_feats
}
return
yolo_feats
return
outs
@
register
@
register
class
YOLOv3Head
(
Layer
):
class
YOLOv3Head
(
Layer
):
__shared__
=
[
'num_classes'
]
__shared__
=
[
'num_classes'
,
'num_levels'
,
'use_fine_grained_loss'
]
__inject__
=
[
'yolo_feat'
]
__inject__
=
[
'yolo_feat'
]
def
__init__
(
self
,
yolo_feat
,
num_classes
=
80
,
anchor_per_position
=
3
):
def
__init__
(
self
,
yolo_feat
,
num_classes
=
80
,
anchor_per_position
=
3
,
num_levels
=
3
,
use_fine_grained_loss
=
False
,
ignore_thresh
=
0.7
,
downsample
=
32
,
label_smooth
=
True
):
super
(
YOLOv3Head
,
self
).
__init__
()
super
(
YOLOv3Head
,
self
).
__init__
()
self
.
num_classes
=
num_classes
self
.
num_classes
=
num_classes
self
.
anchor_per_position
=
anchor_per_position
self
.
anchor_per_position
=
anchor_per_position
self
.
yolo_feat
=
yolo_feat
self
.
yolo_feat
=
yolo_feat
self
.
num_levels
=
num_levels
self
.
yolo_outs
=
[]
self
.
use_fine_grained_loss
=
use_fine_grained_loss
for
i
in
range
(
3
):
self
.
ignore_thresh
=
ignore_thresh
self
.
downsample
=
downsample
self
.
label_smooth
=
label_smooth
self
.
yolo_out_list
=
[]
for
i
in
range
(
num_levels
):
# TODO: optim here
# TODO: optim here
#num_filters = len(cfg.anchor_masks[i]) * (self.num_classes + 5)
#num_filters = len(cfg.anchor_masks[i]) * (self.num_classes + 5)
num_filters
=
self
.
anchor_per_position
*
(
self
.
num_classes
+
5
)
num_filters
=
self
.
anchor_per_position
*
(
self
.
num_classes
+
5
)
name
=
'yolo_output.{}'
.
format
(
i
)
yolo_out
=
self
.
add_sublayer
(
yolo_out
=
self
.
add_sublayer
(
"yolo_out_%d"
%
(
i
)
,
name
,
Conv2D
(
Conv2D
(
num_channels
=
1024
//
(
2
**
i
),
num_channels
=
1024
//
(
2
**
i
),
num_filters
=
num_filters
,
num_filters
=
num_filters
,
...
@@ -152,44 +137,38 @@ class YOLOv3Head(Layer):
...
@@ -152,44 +137,38 @@ class YOLOv3Head(Layer):
stride
=
1
,
stride
=
1
,
padding
=
0
,
padding
=
0
,
act
=
None
,
act
=
None
,
param_attr
=
ParamAttr
(
param_attr
=
ParamAttr
(
name
=
name
+
'.conv.weights'
),
initializer
=
fluid
.
initializer
.
Normal
(
0.
,
0.02
)),
bias_attr
=
ParamAttr
(
bias_attr
=
ParamAttr
(
initializer
=
fluid
.
initializer
.
Constant
(
0.0
),
name
=
name
+
'.conv.bias'
,
regularizer
=
L2Decay
(
0.
))))
regularizer
=
L2Decay
(
0.
))))
self
.
yolo_out_list
.
append
(
yolo_out
)
self
.
yolo_outs
.
append
(
yolo_out
)
def
forward
(
self
,
body_feats
):
assert
len
(
body_feats
)
==
self
.
num_levels
yolo_feats
=
self
.
yolo_feat
(
body_feats
)
yolo_head_out
=
[]
for
i
,
feat
in
enumerate
(
yolo_feats
):
yolo_out
=
self
.
yolo_out_list
[
i
](
feat
)
yolo_head_out
.
append
(
yolo_out
)
return
yolo_head_out
def
loss
(
self
,
inputs
,
head_out
,
anchors
,
anchor_masks
,
mask_anchors
):
if
self
.
use_fine_grained_loss
:
raise
NotImplementedError
def
forward
(
self
,
inputs
):
outs
=
self
.
yolo_feat
(
inputs
)
x
=
outs
[
'yolo_feat'
]
yolo_out_list
=
[]
for
i
,
yolo_f
in
enumerate
(
x
):
yolo_out
=
self
.
yolo_outs
[
i
](
yolo_f
)
yolo_out_list
.
append
(
yolo_out
)
outs
.
update
({
"yolo_outs"
:
yolo_out_list
})
return
outs
def
loss
(
self
,
inputs
):
if
callable
(
inputs
[
'anchor_module'
]):
yolo_targets
=
inputs
[
'anchor_module'
].
generate_anchors_target
(
inputs
)
yolo_losses
=
[]
yolo_losses
=
[]
for
i
,
out
in
enumerate
(
inputs
[
'yolo_outs'
]):
for
i
,
out
in
enumerate
(
head_out
):
# TODO: split yolov3_loss into small ops
# 1. compute target 2. loss
loss
=
fluid
.
layers
.
yolov3_loss
(
loss
=
fluid
.
layers
.
yolov3_loss
(
x
=
out
,
x
=
out
,
gt_box
=
inputs
[
'gt_bbox'
],
gt_box
=
inputs
[
'gt_bbox'
],
gt_label
=
inputs
[
'gt_class'
],
gt_label
=
inputs
[
'gt_class'
],
gt_score
=
inputs
[
'gt_score'
],
gt_score
=
inputs
[
'gt_score'
],
anchors
=
inputs
[
'anchors'
]
,
anchors
=
anchors
,
anchor_mask
=
inputs
[
'anchor_masks'
]
[
i
],
anchor_mask
=
anchor_masks
[
i
],
class_num
=
self
.
num_classes
,
class_num
=
self
.
num_classes
,
ignore_thresh
=
yolo_targets
[
'ignore_thresh'
]
,
ignore_thresh
=
self
.
ignore_thresh
,
downsample_ratio
=
yolo_targets
[
'downsample_ratio'
]
//
2
**
i
,
downsample_ratio
=
self
.
downsample
//
2
**
i
,
use_label_smooth
=
yolo_targets
[
'label_smooth'
]
,
use_label_smooth
=
self
.
label_smooth
,
name
=
'yolo_loss_'
+
str
(
i
))
name
=
'yolo_loss_'
+
str
(
i
))
loss
=
fluid
.
layers
.
reduce_mean
(
loss
)
loss
=
fluid
.
layers
.
reduce_mean
(
loss
)
yolo_losses
.
append
(
loss
)
yolo_losses
.
append
(
loss
)
yolo_loss
=
sum
(
yolo_losses
)
return
{
'loss'
:
sum
(
yolo_losses
)}
return
yolo_loss
ppdet/modeling/ops.py
浏览文件 @
d383fd09
...
@@ -99,20 +99,16 @@ class AnchorGeneratorYOLO(object):
...
@@ -99,20 +99,16 @@ class AnchorGeneratorYOLO(object):
self
.
anchors
=
anchors
self
.
anchors
=
anchors
self
.
anchor_masks
=
anchor_masks
self
.
anchor_masks
=
anchor_masks
def
__call__
(
self
,
yolo_outs
):
def
__call__
(
self
):
anchor_num
=
len
(
self
.
anchors
)
mask_anchors
=
[]
mask_anchors
=
[]
for
i
,
_
in
enumerate
(
yolo_outs
):
for
i
in
range
(
len
(
self
.
anchor_masks
)
):
mask_anchor
=
[]
mask_anchor
=
[]
for
m
in
self
.
anchor_masks
[
i
]:
for
m
in
self
.
anchor_masks
[
i
]:
mask_anchor
.
append
(
self
.
anchors
[
2
*
m
])
assert
m
<
anchor_num
,
"anchor mask index overflow"
mask_anchor
.
append
(
self
.
anchors
[
2
*
m
+
1
])
mask_anchor
.
extend
(
self
.
anchors
[
2
*
m
:
2
*
m
+
2
])
mask_anchors
.
append
(
mask_anchor
)
mask_anchors
.
append
(
mask_anchor
)
outs
=
{
return
self
.
anchors
,
self
.
anchor_masks
,
mask_anchors
"anchors"
:
self
.
anchors
,
"anchor_masks"
:
self
.
anchor_masks
,
"mask_anchors"
:
mask_anchors
}
return
outs
@
register
@
register
...
@@ -305,6 +301,7 @@ class RoIExtractor(object):
...
@@ -305,6 +301,7 @@ class RoIExtractor(object):
self
.
canconical_level
,
self
.
canconical_level
,
self
.
canonical_size
,
self
.
canonical_size
,
rois_num
=
rois_num
)
rois_num
=
rois_num
)
rois_feat_list
=
[]
rois_feat_list
=
[]
for
lvl
in
range
(
self
.
start_level
,
self
.
end_level
+
1
):
for
lvl
in
range
(
self
.
start_level
,
self
.
end_level
+
1
):
roi_feat
=
fluid
.
layers
.
roi_align
(
roi_feat
=
fluid
.
layers
.
roi_align
(
...
@@ -381,24 +378,19 @@ class MultiClassNMS(object):
...
@@ -381,24 +378,19 @@ class MultiClassNMS(object):
@
register
@
register
@
serializable
@
serializable
class
YOLOBox
(
object
):
class
YOLOBox
(
object
):
__shared__
=
[
'num_classes'
]
def
__init__
(
def
__init__
(
self
,
self
,
num_classes
=
80
,
conf_thresh
=
0.005
,
conf_thresh
=
0.005
,
downsample_ratio
=
32
,
downsample_ratio
=
32
,
clip_bbox
=
True
,
):
clip_bbox
=
True
,
):
self
.
num_classes
=
num_classes
self
.
conf_thresh
=
conf_thresh
self
.
conf_thresh
=
conf_thresh
self
.
downsample_ratio
=
downsample_ratio
self
.
downsample_ratio
=
downsample_ratio
self
.
clip_bbox
=
clip_bbox
self
.
clip_bbox
=
clip_bbox
def
__call__
(
self
,
x
,
img_size
,
anchors
,
stage
=
0
,
name
=
None
):
def
__call__
(
self
,
x
,
img_size
,
anchors
,
num_classes
,
stage
=
0
):
outs
=
fluid
.
layers
.
yolo_box
(
x
,
img_size
,
anchors
,
num_classes
,
outs
=
fluid
.
layers
.
yolo_box
(
x
,
img_size
,
anchors
,
self
.
num_classes
,
self
.
conf_thresh
,
self
.
downsample_ratio
//
self
.
conf_thresh
,
self
.
downsample_ratio
//
2
**
stage
,
self
.
clip_bbox
,
name
)
2
**
stage
,
self
.
clip_bbox
)
return
outs
return
outs
...
...
ppdet/optimizer.py
浏览文件 @
d383fd09
...
@@ -82,7 +82,7 @@ class LinearWarmup(object):
...
@@ -82,7 +82,7 @@ class LinearWarmup(object):
def
__call__
(
self
,
base_lr
):
def
__call__
(
self
,
base_lr
):
boundary
=
[]
boundary
=
[]
value
=
[]
value
=
[]
for
i
in
range
(
self
.
steps
):
for
i
in
range
(
self
.
steps
+
1
):
alpha
=
i
/
self
.
steps
alpha
=
i
/
self
.
steps
factor
=
self
.
start_factor
*
(
1
-
alpha
)
+
alpha
factor
=
self
.
start_factor
*
(
1
-
alpha
)
+
alpha
lr
=
base_lr
*
factor
lr
=
base_lr
*
factor
...
...
ppdet/py_op/bbox.py
浏览文件 @
d383fd09
...
@@ -190,7 +190,7 @@ def nms_with_decode(bboxes,
...
@@ -190,7 +190,7 @@ def nms_with_decode(bboxes,
rois_n
=
bboxes_v
[
start
:
end
,
:]
# box
rois_n
=
bboxes_v
[
start
:
end
,
:]
# box
rois_n
=
rois_n
/
im_info
[
i
][
2
]
# scale
rois_n
=
rois_n
/
im_info
[
i
][
2
]
# scale
rois_n
=
delta2bbox
(
bbox_deltas_n
,
rois_n
,
variance_v
)
rois_n
=
delta2bbox
(
bbox_deltas_n
,
rois_n
,
variance_v
)
rois_n
=
clip_bbox
(
rois_n
,
im_info
[
i
][:
2
]
/
im_info
[
i
][
2
]
)
rois_n
=
clip_bbox
(
rois_n
,
np
.
round
(
im_info
[
i
][:
2
]
/
im_info
[
i
][
2
])
)
cls_boxes
=
[[]
for
_
in
range
(
class_nums
)]
cls_boxes
=
[[]
for
_
in
range
(
class_nums
)]
scores_n
=
bbox_probs_v
[
start
:
end
,
:]
scores_n
=
bbox_probs_v
[
start
:
end
,
:]
for
j
in
range
(
1
,
class_nums
):
for
j
in
range
(
1
,
class_nums
):
...
...
ppdet/py_op/post_process.py
浏览文件 @
d383fd09
...
@@ -136,7 +136,8 @@ def get_det_res(bboxes, bbox_nums, image_id, num_id_to_cat_id_map):
...
@@ -136,7 +136,8 @@ def get_det_res(bboxes, bbox_nums, image_id, num_id_to_cat_id_map):
k
=
0
k
=
0
for
i
in
range
(
len
(
bbox_nums
)):
for
i
in
range
(
len
(
bbox_nums
)):
image_id
=
int
(
image_id
[
i
][
0
])
image_id
=
int
(
image_id
[
i
][
0
])
if
bboxes
.
shape
==
(
1
,
1
):
continue
det_nums
=
bbox_nums
[
i
]
det_nums
=
bbox_nums
[
i
]
for
j
in
range
(
det_nums
):
for
j
in
range
(
det_nums
):
dt
=
bboxes
[
k
]
dt
=
bboxes
[
k
]
...
...
ppdet/utils/checkpoint.py
浏览文件 @
d383fd09
...
@@ -3,10 +3,12 @@ from __future__ import division
...
@@ -3,10 +3,12 @@ from __future__ import division
from
__future__
import
print_function
from
__future__
import
print_function
from
__future__
import
unicode_literals
from
__future__
import
unicode_literals
import
errno
import
os
import
os
import
time
import
time
import
re
import
re
import
numpy
as
np
import
numpy
as
np
import
paddle
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
from
.download
import
get_weights_path
from
.download
import
get_weights_path
...
@@ -88,9 +90,10 @@ def load_dygraph_ckpt(model,
...
@@ -88,9 +90,10 @@ def load_dygraph_ckpt(model,
return
model
return
model
def
save_dygraph_ckpt
(
model
,
optimizer
,
save_dir
):
def
save_dygraph_ckpt
(
model
,
optimizer
,
save_dir
,
save_name
):
if
not
os
.
path
.
exists
(
save_dir
):
if
not
os
.
path
.
exists
(
save_dir
):
os
.
makedirs
(
save_dir
)
os
.
makedirs
(
save_dir
)
fluid
.
dygraph
.
save_dygraph
(
model
.
state_dict
(),
save_dir
)
save_path
=
os
.
path
.
join
(
save_dir
,
save_name
)
fluid
.
dygraph
.
save_dygraph
(
optimizer
.
state_dict
(),
save_dir
)
fluid
.
dygraph
.
save_dygraph
(
model
.
state_dict
(),
save_path
)
fluid
.
dygraph
.
save_dygraph
(
optimizer
.
state_dict
(),
save_path
)
print
(
"Save checkpoint:"
,
save_dir
)
print
(
"Save checkpoint:"
,
save_dir
)
tools/eval.py
浏览文件 @
d383fd09
...
@@ -20,6 +20,10 @@ from ppdet.utils.cli import ArgsParser
...
@@ -20,6 +20,10 @@ from ppdet.utils.cli import ArgsParser
from
ppdet.utils.eval_utils
import
coco_eval_results
from
ppdet.utils.eval_utils
import
coco_eval_results
from
ppdet.data.reader
import
create_reader
from
ppdet.data.reader
import
create_reader
from
ppdet.utils.checkpoint
import
load_dygraph_ckpt
,
save_dygraph_ckpt
from
ppdet.utils.checkpoint
import
load_dygraph_ckpt
,
save_dygraph_ckpt
import
logging
FORMAT
=
'%(asctime)s-%(levelname)s: %(message)s'
logging
.
basicConfig
(
level
=
logging
.
INFO
,
format
=
FORMAT
)
logger
=
logging
.
getLogger
(
__name__
)
def
parse_args
():
def
parse_args
():
...
@@ -58,22 +62,26 @@ def run(FLAGS, cfg):
...
@@ -58,22 +62,26 @@ def run(FLAGS, cfg):
# Run Eval
# Run Eval
outs_res
=
[]
outs_res
=
[]
start_time
=
time
.
time
()
sample_num
=
0
for
iter_id
,
data
in
enumerate
(
eval_reader
()):
for
iter_id
,
data
in
enumerate
(
eval_reader
()):
start_time
=
time
.
time
()
# forward
# forward
model
.
eval
()
model
.
eval
()
outs
=
model
(
data
,
cfg
[
'EvalReader'
][
'inputs_def'
][
'fields'
],
'infer'
)
outs
=
model
(
data
,
cfg
[
'EvalReader'
][
'inputs_def'
][
'fields'
],
'infer'
)
outs_res
.
append
(
outs
)
outs_res
.
append
(
outs
)
# log
# log
cost_time
=
time
.
time
()
-
start_time
sample_num
+=
len
(
data
)
print
(
"Eval iter: {}, time: {}"
.
format
(
iter_id
,
cost_time
))
if
iter_id
%
100
==
0
:
logger
.
info
(
"Eval iter: {}"
.
format
(
iter_id
))
cost_time
=
time
.
time
()
-
start_time
logger
.
info
(
'Total sample number: {}, averge FPS: {}'
.
format
(
sample_num
,
sample_num
/
cost_time
))
# Metric
# Metric
coco_eval_results
(
coco_eval_results
(
outs_res
,
outs_res
,
include_mask
=
True
if
'MaskHead'
in
cfg
else
False
,
include_mask
=
True
if
getattr
(
cfg
,
'MaskHead'
,
None
)
else
False
,
dataset
=
cfg
[
'EvalReader'
][
'dataset'
])
dataset
=
cfg
[
'EvalReader'
][
'dataset'
])
...
...
tools/train.py
浏览文件 @
d383fd09
...
@@ -113,7 +113,7 @@ def run(FLAGS, cfg):
...
@@ -113,7 +113,7 @@ def run(FLAGS, cfg):
optimizer
,
optimizer
,
cfg
.
pretrain_weights
,
cfg
.
pretrain_weights
,
ckpt_type
=
FLAGS
.
ckpt_type
,
ckpt_type
=
FLAGS
.
ckpt_type
,
load_static_weights
=
cfg
.
load_static_weights
)
load_static_weights
=
cfg
.
get
(
'load_static_weights'
,
False
)
)
# Parallel Model
# Parallel Model
if
ParallelEnv
().
nranks
>
1
:
if
ParallelEnv
().
nranks
>
1
:
...
@@ -177,8 +177,8 @@ def run(FLAGS, cfg):
...
@@ -177,8 +177,8 @@ def run(FLAGS, cfg):
cfg_name
=
os
.
path
.
basename
(
FLAGS
.
config
).
split
(
'.'
)[
0
]
cfg_name
=
os
.
path
.
basename
(
FLAGS
.
config
).
split
(
'.'
)[
0
]
save_name
=
str
(
save_name
=
str
(
iter_id
)
if
iter_id
!=
cfg
.
max_iters
-
1
else
"model_final"
iter_id
)
if
iter_id
!=
cfg
.
max_iters
-
1
else
"model_final"
save_dir
=
os
.
path
.
join
(
cfg
.
save_dir
,
cfg_name
,
save_name
)
save_dir
=
os
.
path
.
join
(
cfg
.
save_dir
,
cfg_name
)
save_dygraph_ckpt
(
model
,
optimizer
,
save_dir
)
save_dygraph_ckpt
(
model
,
optimizer
,
save_dir
,
save_name
)
def
main
():
def
main
():
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录