Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
models
提交
97574429
M
models
项目概览
PaddlePaddle
/
models
大约 2 年 前同步成功
通知
232
Star
6828
Fork
2962
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
602
列表
看板
标记
里程碑
合并请求
255
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
M
models
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
602
Issue
602
列表
看板
标记
里程碑
合并请求
255
合并请求
255
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
97574429
编写于
9月 18, 2018
作者:
B
Bai Yifan
提交者:
GitHub
9月 18, 2018
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Use pyreader in face detection model (#1183)
* add pyreader
上级
5c47580a
变更
6
隐藏空白更改
内联
并排
Showing
6 changed file
with
268 addition
and
235 deletion
+268
-235
fluid/face_detection/profile.py
fluid/face_detection/profile.py
+71
-82
fluid/face_detection/pyramidbox.py
fluid/face_detection/pyramidbox.py
+11
-10
fluid/face_detection/reader.py
fluid/face_detection/reader.py
+34
-43
fluid/face_detection/train.py
fluid/face_detection/train.py
+147
-97
fluid/face_detection/widerface_eval.py
fluid/face_detection/widerface_eval.py
+3
-1
fluid/object_detection/train.py
fluid/object_detection/train.py
+2
-2
未找到文件。
fluid/face_detection/profile.py
浏览文件 @
97574429
...
@@ -24,13 +24,13 @@ add_arg('parallel', bool, True, "parallel")
...
@@ -24,13 +24,13 @@ add_arg('parallel', bool, True, "parallel")
add_arg
(
'learning_rate'
,
float
,
0.001
,
"Learning rate."
)
add_arg
(
'learning_rate'
,
float
,
0.001
,
"Learning rate."
)
add_arg
(
'batch_size'
,
int
,
20
,
"Minibatch size."
)
add_arg
(
'batch_size'
,
int
,
20
,
"Minibatch size."
)
add_arg
(
'num_iteration'
,
int
,
10
,
"Epoch number."
)
add_arg
(
'num_iteration'
,
int
,
10
,
"Epoch number."
)
add_arg
(
'skip_reader'
,
bool
,
False
,
"Whether to skip data reader."
)
add_arg
(
'use_gpu'
,
bool
,
True
,
"Whether use GPU."
)
add_arg
(
'use_gpu'
,
bool
,
True
,
"Whether use GPU."
)
add_arg
(
'use_pyramidbox'
,
bool
,
True
,
"Whether use PyramidBox model."
)
add_arg
(
'use_pyramidbox'
,
bool
,
True
,
"Whether use PyramidBox model."
)
add_arg
(
'model_save_dir'
,
str
,
'output'
,
"The path to save model."
)
add_arg
(
'model_save_dir'
,
str
,
'output'
,
"The path to save model."
)
add_arg
(
'pretrained_model'
,
str
,
'./vgg_ilsvrc_16_fc_reduced'
,
"The init model path."
)
add_arg
(
'pretrained_model'
,
str
,
'./vgg_ilsvrc_16_fc_reduced'
,
"The init model path."
)
add_arg
(
'resize_h'
,
int
,
640
,
"The resized image height."
)
add_arg
(
'resize_h'
,
int
,
640
,
"The resized image height."
)
add_arg
(
'resize_w'
,
int
,
640
,
"The resized image height."
)
add_arg
(
'resize_w'
,
int
,
640
,
"The resized image height."
)
add_arg
(
'data_dir'
,
str
,
'data'
,
"The base dir of dataset"
)
#yapf: enable
#yapf: enable
...
@@ -43,52 +43,64 @@ def train(args, config, train_file_list, optimizer_method):
...
@@ -43,52 +43,64 @@ def train(args, config, train_file_list, optimizer_method):
use_pyramidbox
=
args
.
use_pyramidbox
use_pyramidbox
=
args
.
use_pyramidbox
model_save_dir
=
args
.
model_save_dir
model_save_dir
=
args
.
model_save_dir
pretrained_model
=
args
.
pretrained_model
pretrained_model
=
args
.
pretrained_model
skip_reader
=
args
.
skip_reader
num_iterations
=
args
.
num_iteration
num_iterations
=
args
.
num_iteration
parallel
=
args
.
parallel
parallel
=
args
.
parallel
num_classes
=
2
num_classes
=
2
image_shape
=
[
3
,
height
,
width
]
image_shape
=
[
3
,
height
,
width
]
devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
)
or
""
startup_prog
=
fluid
.
Program
()
devices_num
=
len
(
devices
.
split
(
","
))
train_prog
=
fluid
.
Program
()
with
fluid
.
program_guard
(
train_prog
,
startup_prog
):
fetches
=
[]
py_reader
=
fluid
.
layers
.
py_reader
(
network
=
PyramidBox
(
image_shape
,
num_classes
,
capacity
=
8
,
sub_network
=
use_pyramidbox
)
shapes
=
[[
-
1
]
+
image_shape
,
[
-
1
,
4
],
[
-
1
,
4
],
[
-
1
,
1
]],
if
use_pyramidbox
:
lod_levels
=
[
0
,
1
,
1
,
1
],
face_loss
,
head_loss
,
loss
=
network
.
train
()
dtypes
=
[
"float32"
,
"float32"
,
"float32"
,
"int32"
],
fetches
=
[
face_loss
,
head_loss
]
use_double_buffer
=
True
)
else
:
with
fluid
.
unique_name
.
guard
():
loss
=
network
.
vgg_ssd_loss
()
image
,
face_box
,
head_box
,
gt_label
=
fluid
.
layers
.
read_file
(
py_reader
)
fetches
=
[
loss
]
fetches
=
[]
network
=
PyramidBox
(
image
=
image
,
epocs
=
12880
//
batch_size
face_box
=
face_box
,
boundaries
=
[
epocs
*
40
,
epocs
*
60
,
epocs
*
80
,
epocs
*
100
]
head_box
=
head_box
,
values
=
[
gt_label
=
gt_label
,
learning_rate
,
learning_rate
*
0.5
,
learning_rate
*
0.25
,
sub_network
=
use_pyramidbox
)
learning_rate
*
0.1
,
learning_rate
*
0.01
if
use_pyramidbox
:
]
face_loss
,
head_loss
,
loss
=
network
.
train
()
fetches
=
[
face_loss
,
head_loss
]
if
optimizer_method
==
"momentum"
:
else
:
optimizer
=
fluid
.
optimizer
.
Momentum
(
loss
=
network
.
vgg_ssd_loss
()
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
fetches
=
[
loss
]
boundaries
=
boundaries
,
values
=
values
),
devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
)
or
""
momentum
=
0.9
,
devices_num
=
len
(
devices
.
split
(
","
))
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
batch_size_per_device
=
batch_size
//
devices_num
)
steps_per_pass
=
12880
//
batch_size
else
:
boundaries
=
[
steps_per_pass
*
50
,
steps_per_pass
*
80
,
optimizer
=
fluid
.
optimizer
.
RMSProp
(
steps_per_pass
*
120
,
steps_per_pass
*
140
]
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
boundaries
,
values
),
values
=
[
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
learning_rate
,
learning_rate
*
0.5
,
learning_rate
*
0.25
,
)
learning_rate
*
0.1
,
learning_rate
*
0.01
]
if
optimizer_method
==
"momentum"
:
optimizer
=
fluid
.
optimizer
.
Momentum
(
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
boundaries
=
boundaries
,
values
=
values
),
momentum
=
0.9
,
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
)
else
:
optimizer
=
fluid
.
optimizer
.
RMSProp
(
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
boundaries
,
values
),
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
)
optimizer
.
minimize
(
loss
)
fluid
.
memory_optimize
(
train_prog
)
optimizer
.
minimize
(
loss
)
fluid
.
memory_optimize
(
fluid
.
default_main_program
())
place
=
fluid
.
CUDAPlace
(
0
)
if
use_gpu
else
fluid
.
CPUPlace
()
place
=
fluid
.
CUDAPlace
(
0
)
if
use_gpu
else
fluid
.
CPUPlace
()
exe
=
fluid
.
Executor
(
place
)
exe
=
fluid
.
Executor
(
place
)
exe
.
run
(
fluid
.
default_startup_program
()
)
exe
.
run
(
startup_prog
)
start_pass
=
0
start_pass
=
0
if
pretrained_model
:
if
pretrained_model
:
...
@@ -106,49 +118,26 @@ def train(args, config, train_file_list, optimizer_method):
...
@@ -106,49 +118,26 @@ def train(args, config, train_file_list, optimizer_method):
if
parallel
:
if
parallel
:
train_exe
=
fluid
.
ParallelExecutor
(
train_exe
=
fluid
.
ParallelExecutor
(
use_cuda
=
use_gpu
,
loss_name
=
loss
.
name
)
use_cuda
=
use_gpu
,
loss_name
=
loss
.
name
,
main_program
=
train_prog
)
train_reader
=
reader
.
train
(
config
,
train_reader
=
reader
.
train_batch_reader
(
config
,
train_file_list
,
batch_size
=
batch_size
)
train_file_list
,
batch_size_per_device
,
def
tensor
(
data
,
place
,
lod
=
None
):
shuffle
=
False
,
t
=
fluid
.
core
.
LoDTensor
()
use_multiprocessing
=
True
,
t
.
set
(
data
,
place
)
num_workers
=
8
,
if
lod
:
max_queue
=
24
)
t
.
set_lod
(
lod
)
py_reader
.
decorate_paddle_reader
(
train_reader
)
return
t
def
run
(
iterations
):
im
,
face_box
,
head_box
,
labels
,
lod
=
next
(
train_reader
)
im_t
=
tensor
(
im
,
place
)
box1
=
tensor
(
face_box
,
place
,
[
lod
])
box2
=
tensor
(
head_box
,
place
,
[
lod
])
lbl_t
=
tensor
(
labels
,
place
,
[
lod
])
feed_data
=
{
'image'
:
im_t
,
'face_box'
:
box1
,
'head_box'
:
box2
,
'gt_label'
:
lbl_t
}
def
run
(
iterations
,
feed_data
):
# global feed_data
# global feed_data
reader_time
=
[]
py_reader
.
start
()
run_time
=
[]
run_time
=
[]
for
batch_id
in
range
(
iterations
):
for
batch_id
in
range
(
iterations
):
start_time
=
time
.
time
()
if
not
skip_reader
:
im
,
face_box
,
head_box
,
labels
,
lod
=
next
(
train_reader
)
im_t
=
tensor
(
im
,
place
)
box1
=
tensor
(
face_box
,
place
,
[
lod
])
box2
=
tensor
(
head_box
,
place
,
[
lod
])
lbl_t
=
tensor
(
labels
,
place
,
[
lod
])
feed_data
=
{
'image'
:
im_t
,
'face_box'
:
box1
,
'head_box'
:
box2
,
'gt_label'
:
lbl_t
}
end_time
=
time
.
time
()
reader_time
.
append
(
end_time
-
start_time
)
start_time
=
time
.
time
()
start_time
=
time
.
time
()
if
parallel
:
if
parallel
:
fetch_vars
=
train_exe
.
run
(
fetch_list
=
[
v
.
name
for
v
in
fetches
],
fetch_vars
=
train_exe
.
run
(
fetch_list
=
[
v
.
name
for
v
in
fetches
])
feed
=
feed_data
)
else
:
else
:
fetch_vars
=
exe
.
run
(
fluid
.
default_main_program
(),
fetch_vars
=
exe
.
run
(
train_prog
,
feed
=
feed_data
,
fetch_list
=
fetches
)
fetch_list
=
fetches
)
end_time
=
time
.
time
()
end_time
=
time
.
time
()
run_time
.
append
(
end_time
-
start_time
)
run_time
.
append
(
end_time
-
start_time
)
...
@@ -158,31 +147,31 @@ def train(args, config, train_file_list, optimizer_method):
...
@@ -158,31 +147,31 @@ def train(args, config, train_file_list, optimizer_method):
else
:
else
:
print
(
"Batch {0}, face loss {1}, head loss {2}"
.
format
(
print
(
"Batch {0}, face loss {1}, head loss {2}"
.
format
(
batch_id
,
fetch_vars
[
0
],
fetch_vars
[
1
]))
batch_id
,
fetch_vars
[
0
],
fetch_vars
[
1
]))
return
run_time
return
reader_time
,
run_time
# start-up
# start-up
run
(
2
,
feed_data
)
run
(
2
)
# profiling
# profiling
start
=
time
.
time
()
start
=
time
.
time
()
if
not
parallel
:
if
not
parallel
:
with
profiler
.
profiler
(
'All'
,
'total'
,
'/tmp/profile_file'
):
with
profiler
.
profiler
(
'All'
,
'total'
,
'/tmp/profile_file'
):
r
eader_time
,
run_time
=
run
(
num_iterations
,
feed_data
)
r
un_time
=
run
(
num_iterations
)
else
:
else
:
r
eader_time
,
run_time
=
run
(
num_iterations
,
feed_data
)
r
un_time
=
run
(
num_iterations
)
end
=
time
.
time
()
end
=
time
.
time
()
total_time
=
end
-
start
total_time
=
end
-
start
print
(
"Total time: {0}, reader time: {1} s, run time: {2} s"
.
format
(
print
(
"Total time: {0}, reader time: {1} s, run time: {2} s"
.
format
(
total_time
,
np
.
sum
(
reader
_time
),
np
.
sum
(
run_time
)))
total_time
,
total_time
-
np
.
sum
(
run
_time
),
np
.
sum
(
run_time
)))
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
print_arguments
(
args
)
print_arguments
(
args
)
data_dir
=
'data/WIDER_train/images/'
data_dir
=
os
.
path
.
join
(
args
.
data_dir
,
'WIDER_train/images/'
)
train_file_list
=
'data/wider_face_split/wider_face_train_bbx_gt.txt'
train_file_list
=
os
.
path
.
join
(
args
.
data_dir
,
'wider_face_split/wider_face_train_bbx_gt.txt'
)
config
=
reader
.
Settings
(
config
=
reader
.
Settings
(
data_dir
=
data_dir
,
data_dir
=
data_dir
,
...
...
fluid/face_detection/pyramidbox.py
浏览文件 @
97574429
...
@@ -63,8 +63,11 @@ def conv_block(input, groups, filters, ksizes, strides=None, with_pool=True):
...
@@ -63,8 +63,11 @@ def conv_block(input, groups, filters, ksizes, strides=None, with_pool=True):
class
PyramidBox
(
object
):
class
PyramidBox
(
object
):
def
__init__
(
self
,
def
__init__
(
self
,
data_shape
,
data_shape
=
None
,
num_classes
=
None
,
image
=
None
,
face_box
=
None
,
head_box
=
None
,
gt_label
=
None
,
use_transposed_conv2d
=
True
,
use_transposed_conv2d
=
True
,
is_infer
=
False
,
is_infer
=
False
,
sub_network
=
False
):
sub_network
=
False
):
...
@@ -74,13 +77,17 @@ class PyramidBox(object):
...
@@ -74,13 +77,17 @@ class PyramidBox(object):
self
.
data_shape
=
data_shape
self
.
data_shape
=
data_shape
self
.
min_sizes
=
[
16.
,
32.
,
64.
,
128.
,
256.
,
512.
]
self
.
min_sizes
=
[
16.
,
32.
,
64.
,
128.
,
256.
,
512.
]
self
.
steps
=
[
4.
,
8.
,
16.
,
32.
,
64.
,
128.
]
self
.
steps
=
[
4.
,
8.
,
16.
,
32.
,
64.
,
128.
]
self
.
num_classes
=
num_classes
self
.
use_transposed_conv2d
=
use_transposed_conv2d
self
.
use_transposed_conv2d
=
use_transposed_conv2d
self
.
is_infer
=
is_infer
self
.
is_infer
=
is_infer
self
.
sub_network
=
sub_network
self
.
sub_network
=
sub_network
self
.
image
=
image
self
.
face_box
=
face_box
self
.
head_box
=
head_box
self
.
gt_label
=
gt_label
# the base network is VGG with atrous layers
# the base network is VGG with atrous layers
self
.
_input
()
if
is_infer
:
self
.
_input
()
self
.
_vgg
()
self
.
_vgg
()
if
sub_network
:
if
sub_network
:
self
.
_low_level_fpn
()
self
.
_low_level_fpn
()
...
@@ -89,12 +96,6 @@ class PyramidBox(object):
...
@@ -89,12 +96,6 @@ class PyramidBox(object):
else
:
else
:
self
.
_vgg_ssd
()
self
.
_vgg_ssd
()
def
feeds
(
self
):
if
self
.
is_infer
:
return
[
self
.
image
]
else
:
return
[
self
.
image
,
self
.
face_box
,
self
.
head_box
,
self
.
gt_label
]
def
_input
(
self
):
def
_input
(
self
):
self
.
image
=
fluid
.
layers
.
data
(
self
.
image
=
fluid
.
layers
.
data
(
name
=
'image'
,
shape
=
self
.
data_shape
,
dtype
=
'float32'
)
name
=
'image'
,
shape
=
self
.
data_shape
,
dtype
=
'float32'
)
...
...
fluid/face_detection/reader.py
浏览文件 @
97574429
...
@@ -231,9 +231,7 @@ def train_generator(settings, file_list, batch_size, shuffle=True):
...
@@ -231,9 +231,7 @@ def train_generator(settings, file_list, batch_size, shuffle=True):
while
True
:
while
True
:
if
shuffle
:
if
shuffle
:
np
.
random
.
shuffle
(
file_dict
)
np
.
random
.
shuffle
(
file_dict
)
images
,
face_boxes
,
head_boxes
,
label_ids
=
[],
[],
[],
[]
batch_out
=
[]
label_offs
=
[
0
]
for
index_image
in
file_dict
.
keys
():
for
index_image
in
file_dict
.
keys
():
image_name
=
file_dict
[
index_image
][
0
]
image_name
=
file_dict
[
index_image
][
0
]
image_path
=
os
.
path
.
join
(
settings
.
data_dir
,
image_name
)
image_path
=
os
.
path
.
join
(
settings
.
data_dir
,
image_name
)
...
@@ -261,7 +259,6 @@ def train_generator(settings, file_list, batch_size, shuffle=True):
...
@@ -261,7 +259,6 @@ def train_generator(settings, file_list, batch_size, shuffle=True):
bbox_sample
.
append
(
float
(
xmax
)
/
im_width
)
bbox_sample
.
append
(
float
(
xmax
)
/
im_width
)
bbox_sample
.
append
(
float
(
ymax
)
/
im_height
)
bbox_sample
.
append
(
float
(
ymax
)
/
im_height
)
bbox_labels
.
append
(
bbox_sample
)
bbox_labels
.
append
(
bbox_sample
)
im
,
sample_labels
=
preprocess
(
im
,
bbox_labels
,
"train"
,
settings
,
im
,
sample_labels
=
preprocess
(
im
,
bbox_labels
,
"train"
,
settings
,
image_path
)
image_path
)
sample_labels
=
np
.
array
(
sample_labels
)
sample_labels
=
np
.
array
(
sample_labels
)
...
@@ -271,46 +268,40 @@ def train_generator(settings, file_list, batch_size, shuffle=True):
...
@@ -271,46 +268,40 @@ def train_generator(settings, file_list, batch_size, shuffle=True):
face_box
=
sample_labels
[:,
1
:
5
]
face_box
=
sample_labels
[:,
1
:
5
]
head_box
=
expand_bboxes
(
face_box
)
head_box
=
expand_bboxes
(
face_box
)
label
=
[
1
]
*
len
(
face_box
)
label
=
[
1
]
*
len
(
face_box
)
batch_out
.
append
((
im
,
face_box
,
head_box
,
label
))
images
.
append
(
im
)
if
len
(
batch_out
)
==
batch_size
:
face_boxes
.
extend
(
face_box
)
yield
batch_out
head_boxes
.
extend
(
head_box
)
batch_out
=
[]
label_ids
.
extend
(
label
)
label_offs
.
append
(
label_offs
[
-
1
]
+
len
(
face_box
))
def
train
(
settings
,
if
len
(
images
)
==
batch_size
:
file_list
,
images
=
np
.
array
(
images
).
astype
(
'float32'
)
batch_size
,
face_boxes
=
np
.
array
(
face_boxes
).
astype
(
'float32'
)
shuffle
=
True
,
head_boxes
=
np
.
array
(
head_boxes
).
astype
(
'float32'
)
use_multiprocessing
=
True
,
label_ids
=
np
.
array
(
label_ids
).
astype
(
'int32'
)
num_workers
=
8
,
yield
images
,
face_boxes
,
head_boxes
,
label_ids
,
label_offs
max_queue
=
24
):
images
,
face_boxes
,
head_boxes
=
[],
[],
[]
def
reader
():
label_ids
,
label_offs
=
[],
[
0
]
try
:
enqueuer
=
GeneratorEnqueuer
(
train_generator
(
settings
,
file_list
,
batch_size
,
shuffle
),
def
train_batch_reader
(
settings
,
use_multiprocessing
=
use_multiprocessing
)
file_list
,
enqueuer
.
start
(
max_queue_size
=
max_queue
,
workers
=
num_workers
)
batch_size
,
shuffle
=
True
,
num_workers
=
8
):
try
:
enqueuer
=
GeneratorEnqueuer
(
train_generator
(
settings
,
file_list
,
batch_size
,
shuffle
),
use_multiprocessing
=
False
)
enqueuer
.
start
(
max_queue_size
=
24
,
workers
=
num_workers
)
generator_output
=
None
while
True
:
while
enqueuer
.
is_running
():
if
not
enqueuer
.
queue
.
empty
():
generator_output
=
enqueuer
.
queue
.
get
()
break
else
:
time
.
sleep
(
0.01
)
yield
generator_output
generator_output
=
None
generator_output
=
None
finally
:
while
True
:
if
enqueuer
is
not
None
:
while
enqueuer
.
is_running
():
enqueuer
.
stop
()
if
not
enqueuer
.
queue
.
empty
():
generator_output
=
enqueuer
.
queue
.
get
()
break
else
:
time
.
sleep
(
0.02
)
yield
generator_output
generator_output
=
None
finally
:
if
enqueuer
is
not
None
:
enqueuer
.
stop
()
return
reader
def
test
(
settings
,
file_list
):
def
test
(
settings
,
file_list
):
...
...
fluid/face_detection/train.py
浏览文件 @
97574429
...
@@ -9,6 +9,7 @@ import time
...
@@ -9,6 +9,7 @@ import time
import
argparse
import
argparse
import
functools
import
functools
import
paddle
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
from
pyramidbox
import
PyramidBox
from
pyramidbox
import
PyramidBox
import
reader
import
reader
...
@@ -21,59 +22,42 @@ add_arg = functools.partial(add_arguments, argparser=parser)
...
@@ -21,59 +22,42 @@ add_arg = functools.partial(add_arguments, argparser=parser)
add_arg
(
'parallel'
,
bool
,
True
,
"Whether use multi-GPU/threads or not."
)
add_arg
(
'parallel'
,
bool
,
True
,
"Whether use multi-GPU/threads or not."
)
add_arg
(
'learning_rate'
,
float
,
0.001
,
"The start learning rate."
)
add_arg
(
'learning_rate'
,
float
,
0.001
,
"The start learning rate."
)
add_arg
(
'batch_size'
,
int
,
16
,
"Minibatch size."
)
add_arg
(
'batch_size'
,
int
,
16
,
"Minibatch size."
)
add_arg
(
'
num_passes'
,
int
,
160
,
"Epoch number."
)
add_arg
(
'
epoc_num'
,
int
,
160
,
"Epoch number."
)
add_arg
(
'use_gpu'
,
bool
,
True
,
"Whether use GPU."
)
add_arg
(
'use_gpu'
,
bool
,
True
,
"Whether use GPU."
)
add_arg
(
'use_pyramidbox'
,
bool
,
True
,
"Whether use PyramidBox model."
)
add_arg
(
'use_pyramidbox'
,
bool
,
True
,
"Whether use PyramidBox model."
)
add_arg
(
'model_save_dir'
,
str
,
'output'
,
"The path to save model."
)
add_arg
(
'model_save_dir'
,
str
,
'output'
,
"The path to save model."
)
add_arg
(
'resize_h'
,
int
,
640
,
"The resized image height."
)
add_arg
(
'resize_h'
,
int
,
640
,
"The resized image height."
)
add_arg
(
'resize_w'
,
int
,
640
,
"The resized image width."
)
add_arg
(
'resize_w'
,
int
,
640
,
"The resized image width."
)
add_arg
(
'mean_BGR'
,
str
,
'104., 117., 123.'
,
"Mean value for B,G,R channel which will be subtracted."
)
add_arg
(
'with_mem_opt'
,
bool
,
True
,
"Whether to use memory optimization or not."
)
add_arg
(
'with_mem_opt'
,
bool
,
True
,
"Whether to use memory optimization or not."
)
add_arg
(
'pretrained_model'
,
str
,
'./vgg_ilsvrc_16_fc_reduced/'
,
"The init model path."
)
add_arg
(
'pretrained_model'
,
str
,
'./vgg_ilsvrc_16_fc_reduced/'
,
"The init model path."
)
add_arg
(
'data_dir'
,
str
,
'data'
,
"The base dir of dataset"
)
add_arg
(
'data_dir'
,
str
,
'data'
,
"The base dir of dataset"
)
#yapf: enable
#yapf: enable
train_parameters
=
{
def
train
(
args
,
config
,
train_file_list
,
optimizer_method
):
"train_images"
:
12880
,
learning_rate
=
args
.
learning_rate
"image_shape"
:
[
3
,
640
,
640
],
batch_size
=
args
.
batch_size
"class_num"
:
2
,
num_passes
=
args
.
num_passes
"batch_size"
:
16
,
height
=
args
.
resize_h
"lr"
:
0.001
,
width
=
args
.
resize_w
"lr_epochs"
:
[
99
,
124
,
149
],
use_gpu
=
args
.
use_gpu
"lr_decay"
:
[
1
,
0.1
,
0.01
,
0.001
],
use_pyramidbox
=
args
.
use_pyramidbox
"epoc_num"
:
160
,
model_save_dir
=
args
.
model_save_dir
"optimizer_method"
:
"momentum"
,
pretrained_model
=
args
.
pretrained_model
"use_pyramidbox"
:
True
with_memory_optimization
=
args
.
with_mem_opt
}
num_classes
=
2
def
optimizer_setting
(
train_params
):
image_shape
=
[
3
,
height
,
width
]
batch_size
=
train_params
[
"batch_size"
]
iters
=
train_params
[
"train_images"
]
//
batch_size
devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
)
or
""
lr
=
train_params
[
"lr"
]
devices_num
=
len
(
devices
.
split
(
","
))
optimizer_method
=
train_params
[
"optimizer_method"
]
boundaries
=
[
i
*
iters
for
i
in
train_params
[
"lr_epochs"
]]
values
=
[
i
*
lr
for
i
in
train_params
[
"lr_decay"
]]
fetches
=
[]
network
=
PyramidBox
(
image_shape
,
num_classes
,
sub_network
=
use_pyramidbox
)
if
use_pyramidbox
:
face_loss
,
head_loss
,
loss
=
network
.
train
()
fetches
=
[
face_loss
,
head_loss
]
else
:
loss
=
network
.
vgg_ssd_loss
()
fetches
=
[
loss
]
steps_per_pass
=
12880
//
batch_size
boundaries
=
[
steps_per_pass
*
99
,
steps_per_pass
*
124
,
steps_per_pass
*
149
]
values
=
[
learning_rate
,
learning_rate
*
0.1
,
learning_rate
*
0.01
,
learning_rate
*
0.001
]
if
optimizer_method
==
"momentum"
:
if
optimizer_method
==
"momentum"
:
optimizer
=
fluid
.
optimizer
.
Momentum
(
optimizer
=
fluid
.
optimizer
.
Momentum
(
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
boundaries
,
values
),
boundaries
=
boundaries
,
values
=
values
),
momentum
=
0.9
,
momentum
=
0.9
,
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
)
)
...
@@ -82,19 +66,76 @@ def train(args, config, train_file_list, optimizer_method):
...
@@ -82,19 +66,76 @@ def train(args, config, train_file_list, optimizer_method):
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
boundaries
,
values
),
learning_rate
=
fluid
.
layers
.
piecewise_decay
(
boundaries
,
values
),
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
regularization
=
fluid
.
regularizer
.
L2Decay
(
0.0005
),
)
)
return
optimizer
def
build_program
(
train_params
,
main_prog
,
startup_prog
,
args
):
use_pyramidbox
=
train_params
[
"use_pyramidbox"
]
image_shape
=
train_params
[
"image_shape"
]
class_num
=
train_params
[
"class_num"
]
with
fluid
.
program_guard
(
main_prog
,
startup_prog
):
py_reader
=
fluid
.
layers
.
py_reader
(
capacity
=
8
,
shapes
=
[[
-
1
]
+
image_shape
,
[
-
1
,
4
],
[
-
1
,
4
],
[
-
1
,
1
]],
lod_levels
=
[
0
,
1
,
1
,
1
],
dtypes
=
[
"float32"
,
"float32"
,
"float32"
,
"int32"
],
use_double_buffer
=
True
)
with
fluid
.
unique_name
.
guard
():
image
,
face_box
,
head_box
,
gt_label
=
fluid
.
layers
.
read_file
(
py_reader
)
fetches
=
[]
network
=
PyramidBox
(
image
=
image
,
face_box
=
face_box
,
head_box
=
head_box
,
gt_label
=
gt_label
,
sub_network
=
use_pyramidbox
)
if
use_pyramidbox
:
face_loss
,
head_loss
,
loss
=
network
.
train
()
fetches
=
[
face_loss
,
head_loss
]
else
:
loss
=
network
.
vgg_ssd_loss
()
fetches
=
[
loss
]
optimizer
=
optimizer_setting
(
train_params
)
optimizer
.
minimize
(
loss
)
return
py_reader
,
fetches
,
loss
def
train
(
args
,
config
,
train_params
,
train_file_list
):
batch_size
=
train_params
[
"batch_size"
]
epoc_num
=
train_params
[
"epoc_num"
]
optimizer_method
=
train_params
[
"optimizer_method"
]
use_pyramidbox
=
train_params
[
"use_pyramidbox"
]
use_gpu
=
args
.
use_gpu
model_save_dir
=
args
.
model_save_dir
pretrained_model
=
args
.
pretrained_model
with_memory_optimization
=
args
.
with_mem_opt
devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
)
or
""
devices_num
=
len
(
devices
.
split
(
","
))
batch_size_per_device
=
batch_size
//
devices_num
iters_per_epoc
=
train_params
[
"train_images"
]
//
batch_size
num_workers
=
8
is_shuffle
=
True
startup_prog
=
fluid
.
Program
()
train_prog
=
fluid
.
Program
()
train_py_reader
,
fetches
,
loss
=
build_program
(
train_params
=
train_params
,
main_prog
=
train_prog
,
startup_prog
=
startup_prog
,
args
=
args
)
optimizer
.
minimize
(
loss
)
if
with_memory_optimization
:
if
with_memory_optimization
:
fluid
.
memory_optimize
(
fluid
.
default_main_program
()
)
fluid
.
memory_optimize
(
train_prog
)
place
=
fluid
.
CUDAPlace
(
0
)
if
use_gpu
else
fluid
.
CPUPlace
()
place
=
fluid
.
CUDAPlace
(
0
)
if
use_gpu
else
fluid
.
CPUPlace
()
exe
=
fluid
.
Executor
(
place
)
exe
=
fluid
.
Executor
(
place
)
exe
.
run
(
fluid
.
default_startup_program
()
)
exe
.
run
(
startup_prog
)
start_
pass
=
0
start_
epoc
=
0
if
pretrained_model
:
if
pretrained_model
:
if
pretrained_model
.
isdigit
():
if
pretrained_model
.
isdigit
():
start_
pass
=
int
(
pretrained_model
)
+
1
start_
epoc
=
int
(
pretrained_model
)
+
1
pretrained_model
=
os
.
path
.
join
(
model_save_dir
,
pretrained_model
)
pretrained_model
=
os
.
path
.
join
(
model_save_dir
,
pretrained_model
)
print
(
"Resume from %s "
%
(
pretrained_model
))
print
(
"Resume from %s "
%
(
pretrained_model
))
...
@@ -103,66 +144,67 @@ def train(args, config, train_file_list, optimizer_method):
...
@@ -103,66 +144,67 @@ def train(args, config, train_file_list, optimizer_method):
(
pretrained_model
))
(
pretrained_model
))
def
if_exist
(
var
):
def
if_exist
(
var
):
return
os
.
path
.
exists
(
os
.
path
.
join
(
pretrained_model
,
var
.
name
))
return
os
.
path
.
exists
(
os
.
path
.
join
(
pretrained_model
,
var
.
name
))
fluid
.
io
.
load_vars
(
exe
,
pretrained_model
,
predicate
=
if_exist
)
fluid
.
io
.
load_vars
(
exe
,
pretrained_model
,
main_program
=
train_prog
,
predicate
=
if_exist
)
train_reader
=
reader
.
train
(
config
,
train_file_list
,
batch_size_per_device
,
shuffle
=
is_shuffle
,
use_multiprocessing
=
True
,
num_workers
=
num_workers
,
max_queue
=
24
)
train_py_reader
.
decorate_paddle_reader
(
train_reader
)
if
args
.
parallel
:
if
args
.
parallel
:
train_exe
=
fluid
.
ParallelExecutor
(
train_exe
=
fluid
.
ParallelExecutor
(
use_cuda
=
use_gpu
,
loss_name
=
loss
.
name
)
main_program
=
train_prog
,
use_cuda
=
use_gpu
,
train_reader
=
reader
.
train_batch_reader
(
config
,
train_file_list
,
batch_size
=
batch_siz
e
)
loss_name
=
loss
.
nam
e
)
def
save_model
(
postfix
):
def
save_model
(
postfix
,
program
):
model_path
=
os
.
path
.
join
(
model_save_dir
,
postfix
)
model_path
=
os
.
path
.
join
(
model_save_dir
,
postfix
)
if
os
.
path
.
isdir
(
model_path
):
if
os
.
path
.
isdir
(
model_path
):
shutil
.
rmtree
(
model_path
)
shutil
.
rmtree
(
model_path
)
print
(
'save models to %s'
%
(
model_path
))
print
(
'save models to %s'
%
(
model_path
))
fluid
.
io
.
save_persistables
(
exe
,
model_path
)
fluid
.
io
.
save_persistables
(
exe
,
model_path
)
def
tensor
(
data
,
place
,
lod
=
None
):
train_py_reader
.
start
()
t
=
fluid
.
core
.
LoDTensor
()
try
:
t
.
set
(
data
,
place
)
for
pass_id
in
range
(
start_epoc
,
epoc_num
):
if
lod
:
t
.
set_lod
(
lod
)
return
t
for
pass_id
in
range
(
start_pass
,
num_passes
):
start_time
=
time
.
time
()
prev_start_time
=
start_time
end_time
=
0
for
batch_id
in
range
(
steps_per_pass
):
im
,
face_box
,
head_box
,
labels
,
lod
=
next
(
train_reader
)
im_t
=
tensor
(
im
,
place
)
box1
=
tensor
(
face_box
,
place
,
[
lod
])
box2
=
tensor
(
head_box
,
place
,
[
lod
])
lbl_t
=
tensor
(
labels
,
place
,
[
lod
])
feeding
=
{
'image'
:
im_t
,
'face_box'
:
box1
,
'head_box'
:
box2
,
'gt_label'
:
lbl_t
}
prev_start_time
=
start_time
start_time
=
time
.
time
()
start_time
=
time
.
time
()
if
args
.
parallel
:
prev_start_time
=
start_time
fetch_vars
=
train_exe
.
run
(
fetch_list
=
[
v
.
name
for
v
in
fetches
],
end_time
=
0
feed
=
feeding
)
batch_id
=
0
else
:
for
batch_id
in
range
(
iters_per_epoc
):
fetch_vars
=
exe
.
run
(
fluid
.
default_main_program
(),
prev_start_time
=
start_time
feed
=
feeding
,
start_time
=
time
.
time
()
fetch_list
=
fetches
)
if
args
.
parallel
:
end_time
=
time
.
time
()
fetch_vars
=
train_exe
.
run
(
fetch_list
=
fetch_vars
=
[
np
.
mean
(
np
.
array
(
v
))
for
v
in
fetch_vars
]
[
v
.
name
for
v
in
fetches
])
if
batch_id
%
10
==
0
:
if
not
args
.
use_pyramidbox
:
print
(
"Pass {0}, batch {1}, loss {2}, time {3}"
.
format
(
pass_id
,
batch_id
,
fetch_vars
[
0
],
start_time
-
prev_start_time
))
else
:
else
:
print
(
"Pass {0}, batch {1}, face loss {2}, head loss {3}, "
\
fetch_vars
=
exe
.
run
(
train_prog
,
"time {4}"
.
format
(
pass_id
,
fetch_list
=
fetches
)
batch_id
,
fetch_vars
[
0
],
fetch_vars
[
1
],
end_time
=
time
.
time
()
start_time
-
prev_start_time
))
fetch_vars
=
[
np
.
mean
(
np
.
array
(
v
))
for
v
in
fetch_vars
]
if
batch_id
%
10
==
0
:
if
pass_id
%
1
==
0
or
pass_id
==
num_passes
-
1
:
if
not
args
.
use_pyramidbox
:
save_model
(
str
(
pass_id
))
print
(
"Pass {0}, batch {1}, loss {2}, time {3}"
.
format
(
pass_id
,
batch_id
,
fetch_vars
[
0
],
start_time
-
prev_start_time
))
else
:
print
(
"Pass {0}, batch {1}, face loss {2}, "
\
"head loss {3}, "
\
"time {4}"
.
format
(
pass_id
,
batch_id
,
fetch_vars
[
0
],
fetch_vars
[
1
],
start_time
-
prev_start_time
))
if
pass_id
%
1
==
0
or
pass_id
==
epoc_num
-
1
:
save_model
(
str
(
pass_id
),
train_prog
)
except
fluid
.
core
.
EOFException
:
train_py_reader
.
reset
()
except
StopIteration
:
train_py_reader
.
reset
()
train_py_reader
.
reset
()
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
args
=
parser
.
parse_args
()
args
=
parser
.
parse_args
()
...
@@ -171,13 +213,21 @@ if __name__ == '__main__':
...
@@ -171,13 +213,21 @@ if __name__ == '__main__':
data_dir
=
os
.
path
.
join
(
args
.
data_dir
,
'WIDER_train/images/'
)
data_dir
=
os
.
path
.
join
(
args
.
data_dir
,
'WIDER_train/images/'
)
train_file_list
=
os
.
path
.
join
(
args
.
data_dir
,
train_file_list
=
os
.
path
.
join
(
args
.
data_dir
,
'wider_face_split/wider_face_train_bbx_gt.txt'
)
'wider_face_split/wider_face_train_bbx_gt.txt'
)
mean_BGR
=
[
float
(
m
)
for
m
in
args
.
mean_BGR
.
split
(
","
)]
image_shape
=
[
3
,
int
(
args
.
resize_h
),
int
(
args
.
resize_w
)]
train_parameters
[
"image_shape"
]
=
image_shape
train_parameters
[
"use_pyramidbox"
]
=
args
.
use_pyramidbox
train_parameters
[
"batch_size"
]
=
args
.
batch_size
train_parameters
[
"lr"
]
=
args
.
learning_rate
train_parameters
[
"epoc_num"
]
=
args
.
epoc_num
config
=
reader
.
Settings
(
config
=
reader
.
Settings
(
data_dir
=
data_dir
,
data_dir
=
data_dir
,
resize_h
=
args
.
resize_h
,
resize_h
=
image_shape
[
1
]
,
resize_w
=
args
.
resize_w
,
resize_w
=
image_shape
[
2
]
,
apply_distort
=
True
,
apply_distort
=
True
,
apply_expand
=
False
,
apply_expand
=
False
,
mean_value
=
[
104.
,
117.
,
123.
]
,
mean_value
=
mean_BGR
,
ap_version
=
'11point'
)
ap_version
=
'11point'
)
train
(
args
,
config
,
train_
file_list
,
optimizer_method
=
"momentum"
)
train
(
args
,
config
,
train_
parameters
,
train_file_list
)
fluid/face_detection/widerface_eval.py
浏览文件 @
97574429
...
@@ -305,7 +305,9 @@ if __name__ == '__main__':
...
@@ -305,7 +305,9 @@ if __name__ == '__main__':
image_shape
=
[
3
,
1024
,
1024
]
image_shape
=
[
3
,
1024
,
1024
]
with
fluid
.
program_guard
(
main_program
,
startup_program
):
with
fluid
.
program_guard
(
main_program
,
startup_program
):
network
=
PyramidBox
(
network
=
PyramidBox
(
image_shape
,
sub_network
=
args
.
use_pyramidbox
,
is_infer
=
True
)
data_shape
=
image_shape
,
sub_network
=
args
.
use_pyramidbox
,
is_infer
=
True
)
infer_program
,
nmsed_out
=
network
.
infer
(
main_program
)
infer_program
,
nmsed_out
=
network
.
infer
(
main_program
)
fetches
=
[
nmsed_out
]
fetches
=
[
nmsed_out
]
fluid
.
io
.
load_persistables
(
fluid
.
io
.
load_persistables
(
...
...
fluid/object_detection/train.py
浏览文件 @
97574429
...
@@ -62,7 +62,7 @@ train_parameters = {
...
@@ -62,7 +62,7 @@ train_parameters = {
def
optimizer_setting
(
train_params
):
def
optimizer_setting
(
train_params
):
batch_size
=
train_params
[
"batch_size"
]
batch_size
=
train_params
[
"batch_size"
]
iters
=
train_params
[
"train_images"
]
/
batch_size
iters
=
train_params
[
"train_images"
]
/
/
batch_size
lr
=
train_params
[
"lr"
]
lr
=
train_params
[
"lr"
]
boundaries
=
[
i
*
iters
for
i
in
train_params
[
"lr_epochs"
]]
boundaries
=
[
i
*
iters
for
i
in
train_params
[
"lr_epochs"
]]
values
=
[
i
*
lr
for
i
in
train_params
[
"lr_decay"
]]
values
=
[
i
*
lr
for
i
in
train_params
[
"lr_decay"
]]
...
@@ -118,7 +118,6 @@ def train(args,
...
@@ -118,7 +118,6 @@ def train(args,
model_save_dir
=
args
.
model_save_dir
model_save_dir
=
args
.
model_save_dir
pretrained_model
=
args
.
pretrained_model
pretrained_model
=
args
.
pretrained_model
epoc_num
=
args
.
epoc_num
use_gpu
=
args
.
use_gpu
use_gpu
=
args
.
use_gpu
parallel
=
args
.
parallel
parallel
=
args
.
parallel
enable_ce
=
args
.
enable_ce
enable_ce
=
args
.
enable_ce
...
@@ -127,6 +126,7 @@ def train(args,
...
@@ -127,6 +126,7 @@ def train(args,
devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
)
or
""
devices
=
os
.
getenv
(
"CUDA_VISIBLE_DEVICES"
)
or
""
devices_num
=
len
(
devices
.
split
(
","
))
devices_num
=
len
(
devices
.
split
(
","
))
batch_size
=
train_params
[
'batch_size'
]
batch_size
=
train_params
[
'batch_size'
]
epoc_num
=
train_params
[
'epoch_num'
]
batch_size_per_device
=
batch_size
//
devices_num
batch_size_per_device
=
batch_size
//
devices_num
iters_per_epoc
=
train_params
[
"train_images"
]
//
batch_size
iters_per_epoc
=
train_params
[
"train_images"
]
//
batch_size
num_workers
=
8
num_workers
=
8
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录