From e3db65d5b4726a02ed724e34121fcfd3195d498e Mon Sep 17 00:00:00 2001 From: zhaoyingli <86812880+zhaoyinglia@users.noreply.github.com> Date: Thu, 28 Oct 2021 19:23:11 +0800 Subject: [PATCH] fix dygraph adamw (#36745) (#36794) --- python/paddle/fluid/tests/unittests/test_adamw_op.py | 2 +- python/paddle/optimizer/adamw.py | 5 ++--- 2 files changed, 3 insertions(+), 4 deletions(-) diff --git a/python/paddle/fluid/tests/unittests/test_adamw_op.py b/python/paddle/fluid/tests/unittests/test_adamw_op.py index 0a60f4cba0..dbeb5a4303 100644 --- a/python/paddle/fluid/tests/unittests/test_adamw_op.py +++ b/python/paddle/fluid/tests/unittests/test_adamw_op.py @@ -333,7 +333,7 @@ class TestAdamWOpLayerwiseLR(TestAdamWOp): lr_ratio=simple_lr_fun) loss_ref = np.array( - [4.8383293, 3.0854003, 1.33299, -0.418993, -2.171043]) + [4.8383293, 3.084947, 1.3323904, -0.41943002, -2.1710064]) for i in range(5): a1 = linear1(a) out = linear2(a1) diff --git a/python/paddle/optimizer/adamw.py b/python/paddle/optimizer/adamw.py index f26ee80d0a..9f96b8c9ea 100644 --- a/python/paddle/optimizer/adamw.py +++ b/python/paddle/optimizer/adamw.py @@ -304,9 +304,8 @@ class AdamW(Adam): moment1, moment2, beta1_pow_acc, beta2_pow_acc, master_weight, 'epsilon', self._epsilon, 'lazy_mode', self._lazy_mode, 'min_row_size_to_use_multithread', 1000, 'beta1', _beta1, - 'beta2', _beta2, 'coeff', self._coeff, 'multi_precision', - find_master, 'lr_ratio', lr_ratio_) - + 'beta2', _beta2, "with_decay", with_decay, 'coeff', self._coeff, + 'multi_precision', find_master, 'lr_ratio', lr_ratio_) return None inputs = { -- GitLab