From 1166014b639e80470152a915c54b479cb8cb7a7c Mon Sep 17 00:00:00 2001 From: Zeyu Chen Date: Sat, 13 Apr 2019 01:38:53 +0800 Subject: [PATCH] migrate BERTFinetune to AdamWeightDecay --- demo/ernie-classification/README.md | 2 +- demo/ernie-classification/ernie_tiny_demo.py | 2 +- demo/ernie-classification/question_answering.py | 2 +- demo/ernie-classification/question_matching.py | 2 +- demo/ernie-classification/sentiment_cls.py | 2 +- demo/ernie-seq-labeling/sequence_labeling.py | 2 +- paddlehub/__init__.py | 2 +- paddlehub/finetune/finetune.py | 6 +++--- paddlehub/finetune/strategy.py | 4 ++-- 9 files changed, 12 insertions(+), 12 deletions(-) diff --git a/demo/ernie-classification/README.md b/demo/ernie-classification/README.md index 45321ae5..32cf3b5c 100644 --- a/demo/ernie-classification/README.md +++ b/demo/ernie-classification/README.md @@ -95,7 +95,7 @@ with fluid.program_guard(program): # NOTE: 必须使用fluid.program_guard接口 ### Step4:选择优化策略并开始Finetune ```python -strategy = hub.BERTFinetuneStrategy( +strategy = hub.AdamWeightDecayStrategy( weight_decay=0.01, learning_rate=5e-5, warmup_strategy="linear_warmup_decay", diff --git a/demo/ernie-classification/ernie_tiny_demo.py b/demo/ernie-classification/ernie_tiny_demo.py index bf0d95d0..a8c24a10 100644 --- a/demo/ernie-classification/ernie_tiny_demo.py +++ b/demo/ernie-classification/ernie_tiny_demo.py @@ -21,7 +21,7 @@ with fluid.program_guard(program): feature=pooled_output, label=label, num_classes=reader.get_num_labels()) # Step4 -strategy = hub.BERTFinetuneStrategy( +strategy = hub.AdamWeightDecayStrategy( learning_rate=5e-5, warmup_proportion=0.1, warmup_strategy="linear_warmup_decay", diff --git a/demo/ernie-classification/question_answering.py b/demo/ernie-classification/question_answering.py index 3429b488..65406c45 100644 --- a/demo/ernie-classification/question_answering.py +++ b/demo/ernie-classification/question_answering.py @@ -62,7 +62,7 @@ if __name__ == '__main__': pooled_output, label, num_classes=num_labels) # Step4: Select finetune strategy, setup config and finetune - strategy = hub.BERTFinetuneStrategy( + strategy = hub.AdamWeightDecayStrategy( weight_decay=args.weight_decay, learning_rate=args.learning_rate, warmup_strategy="linear_warmup_decay", diff --git a/demo/ernie-classification/question_matching.py b/demo/ernie-classification/question_matching.py index af2a0f64..351922c1 100644 --- a/demo/ernie-classification/question_matching.py +++ b/demo/ernie-classification/question_matching.py @@ -62,7 +62,7 @@ if __name__ == '__main__': pooled_output, label, num_classes=num_labels) # Step4: Select finetune strategy, setup config and finetune - strategy = hub.BERTFinetuneStrategy( + strategy = hub.AdamWeightDecayStrategy( weight_decay=args.weight_decay, learning_rate=args.learning_rate, warmup_strategy="linear_warmup_decay", diff --git a/demo/ernie-classification/sentiment_cls.py b/demo/ernie-classification/sentiment_cls.py index ec6b1b4b..fddc8c3a 100644 --- a/demo/ernie-classification/sentiment_cls.py +++ b/demo/ernie-classification/sentiment_cls.py @@ -61,7 +61,7 @@ if __name__ == '__main__': pooled_output, label, num_classes=reader.get_num_labels()) # Step4: Select finetune strategy, setup config and finetune - strategy = hub.BERTFinetuneStrategy( + strategy = hub.AdamWeightDecayStrategy( weight_decay=args.weight_decay, learning_rate=args.learning_rate, warmup_strategy="linear_warmup_decay", diff --git a/demo/ernie-seq-labeling/sequence_labeling.py b/demo/ernie-seq-labeling/sequence_labeling.py index 70280114..cb99b3b9 100644 --- a/demo/ernie-seq-labeling/sequence_labeling.py +++ b/demo/ernie-seq-labeling/sequence_labeling.py @@ -69,7 +69,7 @@ if __name__ == '__main__': num_classes=num_labels) # Select a finetune strategy - strategy = hub.BERTFinetuneStrategy( + strategy = hub.AdamWeightDecayStrategy( weight_decay=args.weight_decay, learning_rate=args.learning_rate, warmup_strategy="linear_warmup_decay", diff --git a/paddlehub/__init__.py b/paddlehub/__init__.py index e9e5b534..183820bf 100644 --- a/paddlehub/__init__.py +++ b/paddlehub/__init__.py @@ -39,5 +39,5 @@ from .finetune.task import create_text_classification_task from .finetune.task import create_img_classification_task from .finetune.finetune import finetune_and_eval from .finetune.config import RunConfig -from .finetune.strategy import BERTFinetuneStrategy +from .finetune.strategy import AdamWeightDecayStrategy from .finetune.strategy import DefaultStrategy diff --git a/paddlehub/finetune/finetune.py b/paddlehub/finetune/finetune.py index 204270ac..f5ab2cd3 100644 --- a/paddlehub/finetune/finetune.py +++ b/paddlehub/finetune/finetune.py @@ -25,7 +25,7 @@ import numpy as np from visualdl import LogWriter from paddlehub.common.logger import logger -from paddlehub.finetune.strategy import BERTFinetuneStrategy, DefaultStrategy +from paddlehub.finetune.strategy import AdamWeightDecayStrategy, DefaultStrategy from paddlehub.finetune.checkpoint import load_checkpoint, save_checkpoint from paddlehub.finetune.evaluate import evaluate_cls_task, evaluate_seq_labeling_task import paddlehub as hub @@ -74,7 +74,7 @@ def _finetune_seq_label_task(task, data_feeder = fluid.DataFeeder(feed_list=feed_list, place=place) # Select strategy - if isinstance(config.strategy, hub.BERTFinetuneStrategy): + if isinstance(config.strategy, hub.AdamWeightDecayStrategy): scheduled_lr = config.strategy.execute(loss, main_program, data_reader, config) elif isinstance(config.strategy, hub.DefaultStrategy): @@ -173,7 +173,7 @@ def _finetune_cls_task(task, data_reader, feed_list, config=None, data_feeder = fluid.DataFeeder(feed_list=feed_list, place=place) # select strategy - if isinstance(config.strategy, hub.BERTFinetuneStrategy): + if isinstance(config.strategy, hub.AdamWeightDecayStrategy): scheduled_lr = config.strategy.execute(loss, main_program, data_reader, config) elif isinstance(config.strategy, hub.DefaultStrategy): diff --git a/paddlehub/finetune/strategy.py b/paddlehub/finetune/strategy.py index 093ae499..f6528810 100644 --- a/paddlehub/finetune/strategy.py +++ b/paddlehub/finetune/strategy.py @@ -61,7 +61,7 @@ class DefaultStrategy(object): return "DefaultStrategy" -class BERTFinetuneStrategy(DefaultStrategy): +class AdamWeightDecayStrategy(DefaultStrategy): def __init__(self, learning_rate=1e-4, warmup_strategy="linear_warmup_decay", @@ -114,7 +114,7 @@ class BERTFinetuneStrategy(DefaultStrategy): # TODO complete __str__() def __str__(self): - return "BERTFintuneStrategy" + return "AdamWeightDecayStrategy" class DefaultFinetuneStrategy(DefaultStrategy): -- GitLab