提交 b66ee638 编写于 作者: Y Yang Nie 提交者: Tingquan Gao

fix RMSProp one_dim_param_no_weight_decay

上级 c351dac6
...@@ -232,26 +232,26 @@ class RMSProp(object): ...@@ -232,26 +232,26 @@ class RMSProp(object):
def __call__(self, model_list): def __call__(self, model_list):
# model_list is None in static graph # model_list is None in static graph
parameters = None parameters = None
if len(self.no_weight_decay_name_list) > 0: if model_list:
params_with_decay = [] params_with_decay = []
params_without_decay = [] params_without_decay = []
for m in model_list: for m in model_list:
params = [p for n, p in m.named_parameters() \ for n, p in m.named_parameters():
if not any(nd in n for nd in self.no_weight_decay_name_list)] if any(nd in n for nd in self.no_weight_decay_name_list) \
params_with_decay.extend(params) or (self.one_dim_param_no_weight_decay and len(p.shape) == 1):
params = [p for n, p in m.named_parameters() \ params_without_decay.append(p)
if any(nd in n for nd in self.no_weight_decay_name_list) or (self.one_dim_param_no_weight_decay and len(p.shape) == 1)] else:
params_without_decay.extend(params) params_with_decay.append(p)
parameters = [{ if params_without_decay:
"params": params_with_decay, parameters = [{
"weight_decay": self.weight_decay "params": params_with_decay,
}, { "weight_decay": self.weight_decay
"params": params_without_decay, }, {
"weight_decay": 0.0 "params": params_without_decay,
}] "weight_decay": 0.0
else: }]
parameters = sum([m.parameters() for m in model_list], else:
[]) if model_list else None parameters = params_with_decay
opt = optim.RMSProp( opt = optim.RMSProp(
learning_rate=self.learning_rate, learning_rate=self.learning_rate,
momentum=self.momentum, momentum=self.momentum,
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册