diff --git a/paddlehub/finetune/strategy.py b/paddlehub/finetune/strategy.py index b06d490600223a831d31b851231ac8bcbac656fd..22fdb61bbc4d5867ab828193d13da339b5fe5527 100644 --- a/paddlehub/finetune/strategy.py +++ b/paddlehub/finetune/strategy.py @@ -91,7 +91,7 @@ class AdamWeightDecayStrategy(DefaultStrategy): lr_scheduler="linear_decay", warmup_proportion=0.0, weight_decay=0.01, - optimizer_name=None): + optimizer_name="adam"): super(AdamWeightDecayStrategy, self).__init__( learning_rate=learning_rate, optimizer_name=optimizer_name) # check strategy correctness