未验证 提交 b1e7b5d5 编写于 作者: F FDInSky 提交者: GitHub

test=master update optimizer doc (#641)

update optimizer doc
上级 6a2cbda5
...@@ -71,9 +71,9 @@ class PolynomialDecay(object): ...@@ -71,9 +71,9 @@ class PolynomialDecay(object):
""" """
Applies polynomial decay to the initial learning rate. Applies polynomial decay to the initial learning rate.
Args: Args:
max_iter (int) The learning rate decay steps. max_iter (int): The learning rate decay steps.
end_lr(float) – End learning rate. end_lr (float): End learning rate.
power (float) Polynomial attenuation coefficient power (float): Polynomial attenuation coefficient
""" """
def __init__(self, max_iter=180000, end_lr=0.0001, power=1.0): def __init__(self, max_iter=180000, end_lr=0.0001, power=1.0):
...@@ -94,8 +94,8 @@ class ExponentialDecay(object): ...@@ -94,8 +94,8 @@ class ExponentialDecay(object):
""" """
Applies exponential decay to the learning rate. Applies exponential decay to the learning rate.
Args: Args:
max_iter (int) The learning rate decay steps. max_iter (int): The learning rate decay steps.
decay_rate (float) The learning rate decay rate. decay_rate (float): The learning rate decay rate.
""" """
def __init__(self, max_iter, decay_rate): def __init__(self, max_iter, decay_rate):
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册