提交 da259314 编写于 作者: W weishengyu

update lr

上级 a98bd7a7
......@@ -54,7 +54,7 @@ Optimizer:
momentum: 0.9
lr:
name: Cosine
learning_rate: 0.01
learning_rate: 0.04
regularizer:
name: 'L2'
coeff: 0.0001
......
......@@ -54,7 +54,7 @@ Optimizer:
momentum: 0.9
lr:
name: MultiStepDecay
learning_rate: 0.01
learning_rate: 0.04
milestones: [30, 60, 70, 80, 90, 100]
gamma: 0.5
verbose: False
......
......@@ -53,7 +53,7 @@ Optimizer:
momentum: 0.9
lr:
name: Cosine
learning_rate: 0.01
learning_rate: 0.04
regularizer:
name: 'L2'
coeff: 0.0005
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册