Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleHub
提交
f04836a1
P
PaddleHub
项目概览
PaddlePaddle
/
PaddleHub
大约 1 年 前同步成功
通知
282
Star
12117
Fork
2091
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
200
列表
看板
标记
里程碑
合并请求
4
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleHub
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
200
Issue
200
列表
看板
标记
里程碑
合并请求
4
合并请求
4
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
f04836a1
编写于
3月 28, 2019
作者:
W
wuzewu
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
update finetune method
上级
34855d8e
变更
2
显示空白变更内容
内联
并排
Showing
2 changed file
with
149 addition
and
133 deletion
+149
-133
paddle_hub/finetune/config.py
paddle_hub/finetune/config.py
+3
-1
paddle_hub/finetune/finetune.py
paddle_hub/finetune/finetune.py
+146
-132
未找到文件。
paddle_hub/finetune/config.py
浏览文件 @
f04836a1
...
...
@@ -29,5 +29,7 @@ FinetuneConfig = collections.namedtuple(
'max_seq_len'
,
# for bert
'weight_decay'
,
# for bert
'warmup_proportion'
,
# for bert
'in_tokens'
# for bert
'in_tokens'
,
# for bert
'strategy'
,
'with_memory_optimization'
])
paddle_hub/finetune/finetune.py
浏览文件 @
f04836a1
...
...
@@ -12,146 +12,160 @@
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
absolute_import
from
__future__
import
division
from
__future__
import
print_function
import
os
import
paddle.fluid
as
fluid
import
time
import
numpy
as
np
import
multiprocessing
from
paddle_hub.finetune.optimization
import
bert_optimization
from
paddle_hub.finetune.config
import
FinetuneConfig
import
paddle
import
paddle.fluid
as
fluid
from
paddle_hub.tools.logger
import
logger
def
finetune_and_eval
(
task
,
feed_list
,
data_processor
,
config
=
None
):
# environment setup
if
config
.
use_cuda
:
place
=
fluid
.
CUDAPlace
(
int
(
os
.
getenv
(
'FLAGS_selected_gpus'
,
'0'
)))
dev_count
=
fluid
.
core
.
get_cuda_device_count
()
else
:
place
=
fluid
.
CPUPlace
()
dev_count
=
int
(
os
.
environ
.
get
(
'CPU_NUM'
,
multiprocessing
.
cpu_count
()))
exe
=
fluid
.
Executor
(
place
)
# hub.finetune_and_eval start here
#TODO: to simplify
loss
=
task
.
variable
(
"loss"
)
probs
=
task
.
variable
(
"probs"
)
accuracy
=
task
.
variable
(
"accuracy"
)
num_example
=
task
.
variable
(
"num_example"
)
def
optimizer_config_for_strategy
(
strategy
,
parameters
,
data_processor
,
dev_count
):
# basic configuration
learning_rate
=
1e-4
optimizer
=
fluid
.
optimizer
.
Adam
(
learning_rate
)
regularizer
=
fluid
.
regularizer
.
L2DecayRegularizer
(
regularization_coeff
=
1e-4
)
num_train_examples
=
data_processor
.
get_num_examples
(
phase
=
'train'
)
if
config
.
in_tokens
:
max_train_steps
=
config
.
num_epoch
*
num_train_examples
//
(
config
.
batch_size
//
config
.
max_seq_len
)
//
dev_count
else
:
max_train_steps
=
config
.
num_epoch
*
num_train_examples
//
config
.
batch_size
//
dev_count
return
optimizer
warmup_steps
=
int
(
max_train_steps
*
config
.
warmup_proportion
)
# obtain main program from Task class
train_program
=
task
.
main_program
()
def
_finetune_model
(
task
,
data_processor
,
feed_list
,
config
=
None
,
eval_model
=
False
):
main_program
=
task
.
main_program
()
startup_program
=
task
.
startup_program
()
# clone test program before optimize
test_program
=
train_program
.
clone
(
for_test
=
True
)
loss
=
task
.
variable
(
"loss"
)
accuracy
=
task
.
variable
(
"accuracy"
)
epoch
=
config
.
num_epoch
batch_size
=
config
.
batch_size
learning_rate
=
config
.
learning_rate
use_cuda
=
config
.
use_cuda
batch_size
=
config
.
batch_size
strategy
=
config
.
strategy
with_memory_optimization
=
config
.
with_memory_optimization
checkpoint_dir
=
config
.
checkpoint_dir
bert_optimization
(
loss
,
warmup_steps
,
max_train_steps
,
config
.
learning_rate
,
train_program
,
config
.
weight_decay
)
with
fluid
.
program_guard
(
main_program
,
startup_program
):
# memory optimization
if
use_cuda
:
place
=
fluid
.
CUDAPlace
(
0
)
dev_count
=
fluid
.
core
.
get_cuda_device_count
()
else
:
place
=
fluid
.
CPUPlace
()
dev_count
=
int
(
os
.
environ
.
get
(
'CPU_NUM'
,
multiprocessing
.
cpu_count
()))
optimizer
=
optimizer_config_for_strategy
(
strategy
=
strategy
,
parameters
=
None
,
data_processor
=
data_processor
,
dev_count
=
dev_count
)
data_feeder
=
fluid
.
DataFeeder
(
feed_list
=
feed_list
,
place
=
place
)
exe
=
fluid
.
Executor
(
place
=
place
)
optimizer
.
minimize
(
loss
)
if
with_memory_optimization
:
logger
.
info
(
"Memory optimize start"
)
fluid
.
memory_optimize
(
input_program
=
train_program
,
input_program
=
fluid
.
default_main_program
()
,
skip_opt_set
=
[
# skip task graph variable memory optimization
loss
.
name
,
probs
.
name
,
accuracy
.
name
,
num_example
.
name
accuracy
.
name
])
logger
.
info
(
"Memory optimize end"
)
exe
.
run
(
startup_program
)
feeder
=
fluid
.
DataFeeder
(
feed_list
=
feed_list
,
place
=
place
)
# Traning block
# prepare training dataset
total_loss
,
total_acc
,
total_num_example
=
[],
[],
[]
# initilize all parameters
exe
.
run
(
fluid
.
default_startup_program
())
step
=
0
time_begin
=
time
.
time
()
train_time_used
=
0.0
for
epoch
in
range
(
1
,
config
.
num_epoch
+
1
):
print
(
"Epoch {}"
.
format
(
epoch
))
train_data_generator
=
data_processor
.
data_generator
(
batch_size
=
config
.
batch_size
,
phase
=
'train'
,
shuffle
=
False
)
for
example
in
train_data_generator
():
step
+=
1
logger
.
info
(
"Finetune start"
)
train_time_begin
=
time
.
time
()
np_loss
,
np_acc
,
np_num_example
=
exe
.
run
(
program
=
train_program
,
feed
=
feeder
.
feed
([
example
]),
fetch_list
=
[
loss
,
accuracy
,
num_example
])
train_time_used
+=
time
.
time
()
-
train_time_begin
# Statistic Block
total_loss
.
extend
(
np_loss
*
np_num_example
)
total_acc
.
extend
(
np_acc
*
np_num_example
)
total_num_example
.
extend
(
np_num_example
)
for
index
in
range
(
epoch
):
train_reader
=
paddle
.
batch
(
data_processor
.
data_generator
(
phase
=
'train'
),
batch_size
=
batch_size
)
size
=
accuracy_sum
=
loss_sum
=
0
for
batch
in
train_reader
():
loss_v
,
accuracy_v
=
exe
.
run
(
feed
=
data_feeder
.
feed
(
batch
),
fetch_list
=
[
loss
.
name
,
accuracy
.
name
])
step
+=
1
size
+=
len
(
batch
)
accuracy_sum
+=
accuracy_v
*
len
(
batch
)
loss_sum
+=
loss_v
*
len
(
batch
)
if
step
%
config
.
log_interval
==
0
:
# get training progress
accum_num_example
=
np
.
sum
(
total_num_example
)
print
(
"step {}: loss={:.5f} acc={:.5f} [step/sec: {:.2f}]"
.
format
(
step
,
np
.
sum
(
total_loss
)
/
accum_num_example
,
np
.
sum
(
total_acc
)
/
accum_num_example
,
config
.
log_interval
/
train_time_used
))
# reset statistic variables
total_loss
,
total_acc
,
total_num_example
=
[],
[],
[]
train_time_used
=
0.0
# Evaluation block
if
step
%
config
.
eval_interval
==
0
:
test_data_generator
=
data_processor
.
data_generator
(
batch_size
=
config
.
batch_size
,
phase
=
'test'
,
shuffle
=
False
)
dev_data_generator
=
data_processor
.
data_generator
(
batch_size
=
config
.
batch_size
,
phase
=
'dev'
,
shuffle
=
False
)
evaluate
(
task
,
test_program
,
exe
,
feeder
,
dev_data_generator
)
evaluate
(
task
,
test_program
,
exe
,
feeder
,
test_data_generator
)
# Save model checkpoint
train_time_used
=
time
.
time
()
-
train_time_begin
perf
=
train_time_used
/
config
.
log_interval
train_time_begin
=
time
.
time
()
logger
.
info
(
"step %d: loss=%.5f acc=%.5f [step/sec: %.2f]"
%
(
step
,
loss_sum
/
size
,
accuracy_sum
/
size
,
perf
))
size
=
accuracy_sum
=
loss_sum
=
0
if
step
%
config
.
save_ckpt_interval
==
0
:
save_checkpoint
(
exe
,
train_program
,
step
,
config
.
checkpoint_dir
)
model_save_dir
=
os
.
path
.
join
(
checkpoint_dir
,
"model_parameters_in_step%d"
%
step
)
fluid
.
io
.
save_persistables
(
exe
,
dirname
=
model_save_dir
)
if
eval_model
and
step
%
config
.
eval_interval
==
0
:
eval
(
task
,
data_processor
,
feed_list
,
config
)
# eval before end
if
eval_model
:
eval
(
task
,
data_processor
,
feed_list
,
config
)
logger
.
info
(
"Finetune end"
)
# finish final evaluation on testset
test_data_generator
=
data_processor
.
data_generator
(
batch_size
=
config
.
batch_size
,
phase
=
'test'
,
shuffle
=
False
)
evaluate
(
task
,
test_program
,
exe
,
feeder
,
test_data_generator
)
def
save_model_and_checkpoint
(
task
,
save_dir
):
pass
def
save_checkpoint
(
exe
,
train_program
,
step
,
ckpt_dir
):
#TODO: add global step variable for restore checkpoint like tensorflow
ckpt_step_dir
=
os
.
path
.
join
(
ckpt_dir
,
"step_{}"
.
format
(
step
))
fluid
.
io
.
save_persistables
(
exe
,
ckpt_step_dir
,
train_program
)
def
finetune_and_eval
(
task
,
data_processor
,
feed_list
,
config
=
None
,
):
_finetune_model
(
task
,
data_processor
,
feed_list
,
config
,
eval_model
=
True
)
def
evaluate
(
task
,
test_program
,
exe
,
feeder
,
data_generator
):
def
finetune
(
task
,
data_processor
,
feed_list
,
config
=
None
):
_finetune_model
(
task
,
data_processor
,
feed_list
,
config
,
eval_model
=
False
)
def
eval
(
task
,
data_processor
,
feed_list
,
config
=
None
):
inference_program
=
task
.
inference_program
()
main_program
=
task
.
main_program
()
loss
=
task
.
variable
(
"loss"
)
probs
=
task
.
variable
(
"probs"
)
accuracy
=
task
.
variable
(
"accuracy"
)
num_example
=
task
.
variable
(
"num_example"
)
total_loss
,
total_acc
,
total_num_example
=
[],
[],
[]
eval_step
=
0
use_cuda
=
config
.
use_cuda
batch_size
=
config
.
batch_size
logger
.
info
(
"[Evaluation] start"
)
with
fluid
.
program_guard
(
inference_program
):
place
=
fluid
.
CUDAPlace
(
0
)
if
use_cuda
else
fluid
.
CPUPlace
()
data_feeder
=
fluid
.
DataFeeder
(
feed_list
=
feed_list
,
place
=
place
)
exe
=
fluid
.
Executor
(
place
=
place
)
size
=
accuracy_sum
=
loss_sum
=
0
test_reader
=
paddle
.
batch
(
data_processor
.
data_generator
(
phase
=
'test'
),
batch_size
=
batch_size
)
eval_time_begin
=
time
.
time
()
for
example
in
data_generator
():
eval_step
+=
1
np_loss
,
np_acc
,
np_num_example
=
exe
.
run
(
program
=
test_program
,
feed
=
feeder
.
feed
([
example
]),
fetch_list
=
[
loss
,
accuracy
,
num_example
])
total_loss
.
extend
(
np_loss
*
np_num_example
)
total_acc
.
extend
(
np_acc
*
np_num_example
)
total_num_example
.
extend
(
np_num_example
)
for
index
,
batch
in
enumerate
(
test_reader
()):
loss_v
,
accuracy_v
,
=
exe
.
run
(
feed
=
data_feeder
.
feed
(
batch
),
fetch_list
=
[
loss
,
accuracy
.
name
])
size
+=
len
(
batch
)
accuracy_sum
+=
accuracy_v
*
len
(
batch
)
loss_sum
+=
loss_v
*
len
(
batch
)
eval_time_used
=
time
.
time
()
-
eval_time_begin
accum_num_example
=
np
.
sum
(
total_num_example
)
print
(
"[evaluation] loss={:.5f} acc={:.5f} [step/sec: {:.2f}]"
.
format
(
np
.
sum
(
total_loss
)
/
accum_num_example
,
np
.
sum
(
total_acc
)
/
accum_num_example
,
eval_step
/
eval_time_used
))
perf
=
eval_time_used
/
index
logger
.
info
(
"[Evaluation] loss=%.5f acc=%.5f [step/sec: %.2f]"
%
(
loss_sum
/
size
,
accuracy_sum
/
size
,
perf
))
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录