Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
1b83de2e
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
1b83de2e
编写于
4月 23, 2021
作者:
Z
zhiboniu
提交者:
GitHub
4月 23, 2021
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
update 2.0 public api in optimizer (#31944)
上级
7a681f0b
变更
12
隐藏空白更改
内联
并排
Showing
12 changed file
with
36 addition
and
37 deletion
+36
-37
python/paddle/optimizer/__init__.py
python/paddle/optimizer/__init__.py
+23
-15
python/paddle/optimizer/adadelta.py
python/paddle/optimizer/adadelta.py
+0
-2
python/paddle/optimizer/adagrad.py
python/paddle/optimizer/adagrad.py
+0
-2
python/paddle/optimizer/adam.py
python/paddle/optimizer/adam.py
+0
-2
python/paddle/optimizer/adamax.py
python/paddle/optimizer/adamax.py
+0
-2
python/paddle/optimizer/adamw.py
python/paddle/optimizer/adamw.py
+0
-2
python/paddle/optimizer/lamb.py
python/paddle/optimizer/lamb.py
+0
-2
python/paddle/optimizer/lr.py
python/paddle/optimizer/lr.py
+13
-4
python/paddle/optimizer/momentum.py
python/paddle/optimizer/momentum.py
+0
-1
python/paddle/optimizer/optimizer.py
python/paddle/optimizer/optimizer.py
+0
-2
python/paddle/optimizer/rmsprop.py
python/paddle/optimizer/rmsprop.py
+0
-2
python/paddle/optimizer/sgd.py
python/paddle/optimizer/sgd.py
+0
-1
未找到文件。
python/paddle/optimizer/__init__.py
浏览文件 @
1b83de2e
...
@@ -12,19 +12,27 @@
...
@@ -12,19 +12,27 @@
# See the License for the specific language governing permissions and
# See the License for the specific language governing permissions and
# limitations under the License.
# limitations under the License.
__all__
=
[
from
.optimizer
import
Optimizer
# noqa: F401
'Optimizer'
,
'Adagrad'
,
'Adam'
,
'AdamW'
,
'Adamax'
,
'RMSProp'
,
'Adadelta'
,
from
.adagrad
import
Adagrad
# noqa: F401
'SGD'
,
'Momentum'
,
'Lamb'
,
'lr'
from
.adam
import
Adam
# noqa: F401
]
from
.adamw
import
AdamW
# noqa: F401
from
.adamax
import
Adamax
# noqa: F401
from
.rmsprop
import
RMSProp
# noqa: F401
from
.adadelta
import
Adadelta
# noqa: F401
from
.sgd
import
SGD
# noqa: F401
from
.momentum
import
Momentum
# noqa: F401
from
.lamb
import
Lamb
# noqa: F401
from
.
import
lr
# noqa: F401
from
.optimizer
import
Optimizer
__all__
=
[
#noqa
from
.adagrad
import
Adagrad
'Optimizer'
,
from
.adam
import
Adam
'Adagrad'
,
from
.adamw
import
AdamW
'Adam'
,
from
.adamax
import
Adamax
'AdamW'
,
from
.rmsprop
import
RMSProp
'Adamax'
,
from
.adadelta
import
Adadelta
'RMSProp'
,
from
.sgd
import
SGD
'Adadelta'
,
from
.momentum
import
Momentum
'SGD'
,
from
.lamb
import
Lamb
'Momentum'
,
from
.
import
lr
'Lamb'
]
python/paddle/optimizer/adadelta.py
浏览文件 @
1b83de2e
...
@@ -17,8 +17,6 @@ from ..fluid import core
...
@@ -17,8 +17,6 @@ from ..fluid import core
from
..fluid
import
framework
from
..fluid
import
framework
from
..fluid.framework
import
Variable
,
name_scope
from
..fluid.framework
import
Variable
,
name_scope
__all__
=
[
"Adadelta"
]
class
Adadelta
(
Optimizer
):
class
Adadelta
(
Optimizer
):
r
"""
r
"""
...
...
python/paddle/optimizer/adagrad.py
浏览文件 @
1b83de2e
...
@@ -17,8 +17,6 @@ from ..fluid import core
...
@@ -17,8 +17,6 @@ from ..fluid import core
from
..fluid
import
framework
from
..fluid
import
framework
from
..fluid.framework
import
Variable
from
..fluid.framework
import
Variable
__all__
=
[
"Adagrad"
]
class
Adagrad
(
Optimizer
):
class
Adagrad
(
Optimizer
):
r
"""
r
"""
...
...
python/paddle/optimizer/adam.py
浏览文件 @
1b83de2e
...
@@ -24,8 +24,6 @@ from ..fluid.dygraph import base as imperative_base
...
@@ -24,8 +24,6 @@ from ..fluid.dygraph import base as imperative_base
import
paddle
import
paddle
__all__
=
[
"Adam"
]
class
Adam
(
Optimizer
):
class
Adam
(
Optimizer
):
r
"""
r
"""
...
...
python/paddle/optimizer/adamax.py
浏览文件 @
1b83de2e
...
@@ -17,8 +17,6 @@ from ..fluid import core
...
@@ -17,8 +17,6 @@ from ..fluid import core
from
..fluid
import
framework
from
..fluid
import
framework
from
..fluid.framework
import
Variable
,
name_scope
from
..fluid.framework
import
Variable
,
name_scope
__all__
=
[
"Adamax"
]
class
Adamax
(
Optimizer
):
class
Adamax
(
Optimizer
):
r
"""
r
"""
...
...
python/paddle/optimizer/adamw.py
浏览文件 @
1b83de2e
...
@@ -19,8 +19,6 @@ from ..fluid import framework
...
@@ -19,8 +19,6 @@ from ..fluid import framework
from
..fluid.dygraph
import
base
as
imperative_base
from
..fluid.dygraph
import
base
as
imperative_base
import
paddle
import
paddle
__all__
=
[
'AdamW'
]
class
AdamW
(
Adam
):
class
AdamW
(
Adam
):
r
"""
r
"""
...
...
python/paddle/optimizer/lamb.py
浏览文件 @
1b83de2e
...
@@ -17,8 +17,6 @@ from ..fluid import core
...
@@ -17,8 +17,6 @@ from ..fluid import core
from
..fluid
import
framework
from
..fluid
import
framework
from
..fluid.framework
import
Variable
from
..fluid.framework
import
Variable
__all__
=
[
"Lamb"
]
class
Lamb
(
Optimizer
):
class
Lamb
(
Optimizer
):
r
"""
r
"""
...
...
python/paddle/optimizer/lr.py
浏览文件 @
1b83de2e
...
@@ -17,10 +17,19 @@ import numpy
...
@@ -17,10 +17,19 @@ import numpy
import
warnings
import
warnings
from
paddle
import
Tensor
from
paddle
import
Tensor
__all__
=
[
__all__
=
[
#noqa
'LRScheduler'
,
'NoamDecay'
,
'PiecewiseDecay'
,
'NaturalExpDecay'
,
'LRScheduler'
,
'InverseTimeDecay'
,
'PolynomialDecay'
,
'LinearWarmup'
,
'ExponentialDecay'
,
'NoamDecay'
,
'MultiStepDecay'
,
'StepDecay'
,
'LambdaDecay'
,
'ReduceOnPlateau'
,
'PiecewiseDecay'
,
'NaturalExpDecay'
,
'InverseTimeDecay'
,
'PolynomialDecay'
,
'LinearWarmup'
,
'ExponentialDecay'
,
'MultiStepDecay'
,
'StepDecay'
,
'LambdaDecay'
,
'ReduceOnPlateau'
,
'CosineAnnealingDecay'
'CosineAnnealingDecay'
]
]
...
...
python/paddle/optimizer/momentum.py
浏览文件 @
1b83de2e
...
@@ -21,7 +21,6 @@ from ..fluid import unique_name
...
@@ -21,7 +21,6 @@ from ..fluid import unique_name
from
..fluid
import
layers
from
..fluid
import
layers
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
from
paddle.fluid.regularizer
import
L2DecayRegularizer
from
paddle.fluid.regularizer
import
L2DecayRegularizer
__all__
=
[
"Momentum"
]
class
Momentum
(
Optimizer
):
class
Momentum
(
Optimizer
):
...
...
python/paddle/optimizer/optimizer.py
浏览文件 @
1b83de2e
...
@@ -42,8 +42,6 @@ from ..fluid.wrapped_decorator import signature_safe_contextmanager
...
@@ -42,8 +42,6 @@ from ..fluid.wrapped_decorator import signature_safe_contextmanager
from
..
import
compat
as
cpt
from
..
import
compat
as
cpt
from
.lr
import
LRScheduler
from
.lr
import
LRScheduler
__all__
=
[
'Optimizer'
]
class
Optimizer
(
object
):
class
Optimizer
(
object
):
r
"""Optimizer Base class.
r
"""Optimizer Base class.
...
...
python/paddle/optimizer/rmsprop.py
浏览文件 @
1b83de2e
...
@@ -17,8 +17,6 @@ from ..fluid import core
...
@@ -17,8 +17,6 @@ from ..fluid import core
from
..fluid
import
framework
from
..fluid
import
framework
from
..fluid.framework
import
Variable
from
..fluid.framework
import
Variable
__all__
=
[
"RMSProp"
]
class
RMSProp
(
Optimizer
):
class
RMSProp
(
Optimizer
):
r
"""
r
"""
...
...
python/paddle/optimizer/sgd.py
浏览文件 @
1b83de2e
...
@@ -17,7 +17,6 @@ from ..fluid import core
...
@@ -17,7 +17,6 @@ from ..fluid import core
from
..fluid
import
framework
from
..fluid
import
framework
from
..fluid.framework
import
Variable
,
name_scope
from
..fluid.framework
import
Variable
,
name_scope
from
..fluid.dygraph
import
no_grad
from
..fluid.dygraph
import
no_grad
__all__
=
[
"SGD"
]
class
SGD
(
Optimizer
):
class
SGD
(
Optimizer
):
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录