提交 caca3653 编写于 作者: Y Yibing Liu

Change the optimizer from adam to momentum

上级 73722097
...@@ -125,7 +125,8 @@ def profile(args): ...@@ -125,7 +125,8 @@ def profile(args):
class_num=1749, class_num=1749,
parallel=args.parallel) parallel=args.parallel)
adam_optimizer = fluid.optimizer.Adam(learning_rate=args.learning_rate) adam_optimizer = fluid.optimizer.Momentum(
learning_rate=args.learning_rate, momentum=0.9)
adam_optimizer.minimize(avg_cost) adam_optimizer.minimize(avg_cost)
place = fluid.CPUPlace() if args.device == 'CPU' else fluid.CUDAPlace(0) place = fluid.CPUPlace() if args.device == 'CPU' else fluid.CUDAPlace(0)
......
...@@ -143,7 +143,8 @@ def train(args): ...@@ -143,7 +143,8 @@ def train(args):
class_num=1749, class_num=1749,
parallel=args.parallel) parallel=args.parallel)
adam_optimizer = fluid.optimizer.Adam(learning_rate=args.learning_rate) adam_optimizer = fluid.optimizer.Momentum(
learning_rate=args.learning_rate, momentum=0.9)
adam_optimizer.minimize(avg_cost) adam_optimizer.minimize(avg_cost)
# program for test # program for test
...@@ -225,7 +226,7 @@ def train(args): ...@@ -225,7 +226,7 @@ def train(args):
print("\nBatch %d, train cost: %f, train acc: %f" % print("\nBatch %d, train cost: %f, train acc: %f" %
(batch_id, lodtensor_to_ndarray(cost)[0], (batch_id, lodtensor_to_ndarray(cost)[0],
lodtensor_to_ndarray(acc)[0])) lodtensor_to_ndarray(acc)[0]))
# save the latest checkpoints # save the latest checkpoint
if args.checkpoints != '': if args.checkpoints != '':
model_path = os.path.join(args.checkpoints, model_path = os.path.join(args.checkpoints,
"deep_asr.latest.checkpoint") "deep_asr.latest.checkpoint")
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册