提交 b1f6f568 编写于 作者: W wuzewu

update strategy

上级 a75f8359
......@@ -91,7 +91,7 @@ class AdamWeightDecayStrategy(DefaultStrategy):
lr_scheduler="linear_decay",
warmup_proportion=0.0,
weight_decay=0.01,
optimizer_name=None):
optimizer_name="adam"):
super(AdamWeightDecayStrategy, self).__init__(
learning_rate=learning_rate, optimizer_name=optimizer_name)
# check strategy correctness
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册