未验证 提交 2c4a77ed 编写于 作者: Y Yibing Liu 提交者: GitHub

Enanle gradient clip in dam (#1452)

* Enanle gradient clip in dam

* print memory optimization
上级 6290b4dd
......@@ -206,8 +206,8 @@ def train(args):
loss.persistable = True
logits.persistable = True
# gradient clipping
#fluid.clip.set_gradient_clip(clip=fluid.clip.GradientClipByValue(
# max=1.0, min=-1.0))
fluid.clip.set_gradient_clip(clip=fluid.clip.GradientClipByValue(
max=1.0, min=-1.0))
optimizer = fluid.optimizer.Adam(
learning_rate=fluid.layers.exponential_decay(
......@@ -216,7 +216,9 @@ def train(args):
decay_rate=0.9,
staircase=True))
optimizer.minimize(loss)
print("begin memory optimization ...")
fluid.memory_optimize(train_program)
print("end memory optimization ...")
test_program = fluid.Program()
test_startup = fluid.Program()
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册