Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
fb641c91
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
fb641c91
编写于
10月 16, 2020
作者:
W
WangXi
提交者:
GitHub
10月 16, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
【paddle.fleet】fleet add _get_applied_meta_list and _get_applied_graph_list (#27952)
上级
fa9d3fa5
变更
8
隐藏空白更改
内联
并排
Showing
8 changed file
with
89 addition
and
6 deletion
+89
-6
python/paddle/distributed/fleet/__init__.py
python/paddle/distributed/fleet/__init__.py
+2
-0
python/paddle/distributed/fleet/base/fleet_base.py
python/paddle/distributed/fleet/base/fleet_base.py
+24
-0
python/paddle/distributed/fleet/base/strategy_compiler.py
python/paddle/distributed/fleet/base/strategy_compiler.py
+10
-4
python/paddle/distributed/fleet/meta_optimizers/amp_optimizer.py
...paddle/distributed/fleet/meta_optimizers/amp_optimizer.py
+1
-1
python/paddle/fluid/tests/unittests/test_fleet_amp_meta_optimizer.py
...le/fluid/tests/unittests/test_fleet_amp_meta_optimizer.py
+6
-0
python/paddle/fluid/tests/unittests/test_fleet_auto.py
python/paddle/fluid/tests/unittests/test_fleet_auto.py
+3
-0
python/paddle/fluid/tests/unittests/test_fleet_base_3.py
python/paddle/fluid/tests/unittests/test_fleet_base_3.py
+40
-0
python/paddle/fluid/tests/unittests/test_fleet_base_4.py
python/paddle/fluid/tests/unittests/test_fleet_base_4.py
+3
-1
未找到文件。
python/paddle/distributed/fleet/__init__.py
浏览文件 @
fb641c91
...
...
@@ -34,6 +34,8 @@ __all__ = [
fleet
=
Fleet
()
_final_strategy
=
fleet
.
_final_strategy
_get_applied_meta_list
=
fleet
.
_get_applied_meta_list
_get_applied_graph_list
=
fleet
.
_get_applied_graph_list
init
=
fleet
.
init
is_first_worker
=
fleet
.
is_first_worker
worker_index
=
fleet
.
worker_index
...
...
python/paddle/distributed/fleet/base/fleet_base.py
浏览文件 @
fb641c91
...
...
@@ -925,6 +925,24 @@ class Fleet(object):
else
:
return
self
.
_context
[
"valid_strategy"
]
def
_get_applied_meta_list
(
self
):
if
"applied_meta_list"
not
in
self
.
_context
:
print
(
"WARNING: You may need to call minimize function before _get_applied_meta_list called"
)
return
[]
else
:
return
self
.
_context
[
"applied_meta_list"
]
def
_get_applied_graph_list
(
self
):
if
"applied_graph_list"
not
in
self
.
_context
:
print
(
"WARNING: You may need to call minimize function before _get_applied_graph_list called"
)
return
[]
else
:
return
self
.
_context
[
"applied_graph_list"
]
def
minimize
(
self
,
loss
,
startup_program
=
None
,
...
...
@@ -1043,6 +1061,12 @@ class Fleet(object):
context
[
"valid_strategy"
]
=
copy
.
deepcopy
(
valid_strategy
)
applied_meta_list
=
self
.
strategy_compiler
.
_get_applied_meta_list
()
applied_graph_list
=
self
.
strategy_compiler
.
_get_applied_graph_list
()
context
[
'applied_meta_list'
]
=
applied_meta_list
context
[
'applied_graph_list'
]
=
applied_graph_list
self
.
_context
=
context
self
.
valid_strategy
=
valid_strategy
...
...
python/paddle/distributed/fleet/base/strategy_compiler.py
浏览文件 @
fb641c91
...
...
@@ -122,13 +122,19 @@ class StrategyCompiler(StrategyCompilerBase):
def
__init__
(
self
):
super
(
StrategyCompiler
,
self
).
__init__
()
self
.
_meta_optimizer
=
None
self
.
_graph_optimizer
=
None
self
.
_meta_optimizer
s
=
[]
self
.
_graph_optimizer
s
=
[]
self
.
_valid_optimizer_list
=
None
self
.
_user_defined_strategy
=
None
self
.
_meta_optimizer_candidates
=
[]
self
.
_graph_optimizer_candidates
=
[]
def
_get_applied_meta_list
(
self
):
return
[
type
(
opt
).
__name__
for
opt
in
self
.
_meta_optimizers
]
def
_get_applied_graph_list
(
self
):
return
[
type
(
opt
).
__name__
for
opt
in
self
.
_graph_optimizers
]
def
_get_valid_strategy
(
self
,
dist_strategy
,
can_not_apply_optimizer_list
):
import
copy
valid_strategy
=
copy
.
deepcopy
(
dist_strategy
)
...
...
@@ -178,8 +184,8 @@ class StrategyCompiler(StrategyCompilerBase):
# and graph_optimizer, the corresponding distributed strategy
# should be updated.
self
.
_meta_optimizers
=
meta_optimizers
self
.
_graph_optimizers
=
graph_optimizers
self
.
_meta_optimizers
=
[]
if
meta_optimizers
is
None
else
meta_optimizers
self
.
_graph_optimizers
=
[]
if
graph_optimizers
is
None
else
graph_optimizers
return_meta
=
None
if
meta_optimizers
==
None
else
meta_optimizers
[
0
]
...
...
python/paddle/distributed/fleet/meta_optimizers/amp_optimizer.py
浏览文件 @
fb641c91
...
...
@@ -72,7 +72,7 @@ class AMPOptimizer(MetaOptimizerBase):
"incr_every_n_steps"
:
1000
,
"decr_every_n_nan_or_inf"
:
2
,
"incr_ratio"
:
2.0
,
"decr_ratio"
:
8.0
,
"decr_ratio"
:
0.8
,
"use_dynamic_loss_scaling"
:
True
}
...
...
python/paddle/fluid/tests/unittests/test_fleet_amp_meta_optimizer.py
浏览文件 @
fb641c91
...
...
@@ -133,8 +133,14 @@ class TestFleetAMPOptimizer(TestFleetMetaOptimizer):
self
.
set_strategy
(
strategy
,
'amp'
)
self
.
set_strategy
(
strategy
,
'recompute'
)
self
.
set_strategy
(
strategy
,
'lamb'
)
self
.
optimizer
(
avg_cost
,
strategy
,
train_prog
,
startup_prog
,
'adam'
)
applied_meta_list
=
fleet
.
_get_applied_meta_list
()
applied_graph_list
=
fleet
.
_get_applied_graph_list
()
print
(
applied_meta_list
,
applied_graph_list
)
self
.
assertEqual
(
len
(
applied_meta_list
),
3
)
ops
=
[
op
.
type
for
op
in
avg_cost
.
block
.
ops
]
outs
=
[
op
.
output
(
'Out'
)[
0
]
for
op
in
avg_cost
.
block
.
ops
if
op
.
type
==
'mul'
...
...
python/paddle/fluid/tests/unittests/test_fleet_auto.py
浏览文件 @
fb641c91
...
...
@@ -48,6 +48,9 @@ class TestDistributedStrategyAuto(unittest.TestCase):
optimizer
=
fleet
.
distributed_optimizer
(
optimizer
,
strategy
=
strategy
)
optimizer
.
minimize
(
avg_cost
)
applied_meta_list
=
fleet
.
_get_applied_meta_list
()
print
(
"applied_meta_list: {}"
.
format
(
applied_meta_list
))
if
__name__
==
"__main__"
:
unittest
.
main
()
python/paddle/fluid/tests/unittests/test_fleet_base_3.py
浏览文件 @
fb641c91
...
...
@@ -18,6 +18,7 @@ import paddle
import
paddle.distributed.fleet
as
fleet
import
paddle.distributed.fleet.base.role_maker
as
role_maker
import
paddle.fluid
as
fluid
paddle
.
enable_static
()
class
TestFleetBase
(
unittest
.
TestCase
):
...
...
@@ -48,5 +49,44 @@ class TestFleetBase(unittest.TestCase):
optimizer
.
minimize
(
avg_cost
)
class
TestFleetBase
(
unittest
.
TestCase
):
def
setUp
(
self
):
os
.
environ
[
"POD_IP"
]
=
"127.0.0.1"
os
.
environ
[
"PADDLE_TRAINER_ENDPOINTS"
]
=
"127.0.0.1:36001"
os
.
environ
[
"PADDLE_TRAINERS_NUM"
]
=
"2"
os
.
environ
[
"PADDLE_PSERVERS_IP_PORT_LIST"
]
=
\
"127.0.0.1:36001,127.0.0.2:36001"
def
test_fleet_get_applied_optimizer
(
self
):
input_x
=
paddle
.
fluid
.
layers
.
data
(
name
=
"x"
,
shape
=
[
32
],
dtype
=
'float32'
)
input_y
=
paddle
.
fluid
.
layers
.
data
(
name
=
"y"
,
shape
=
[
1
],
dtype
=
'int64'
)
fc_1
=
paddle
.
fluid
.
layers
.
fc
(
input
=
input_x
,
size
=
64
,
act
=
'tanh'
)
fc_2
=
paddle
.
fluid
.
layers
.
fc
(
input
=
fc_1
,
size
=
64
,
act
=
'tanh'
)
prediction
=
paddle
.
fluid
.
layers
.
fc
(
input
=
[
fc_2
],
size
=
2
,
act
=
'softmax'
)
cost
=
paddle
.
fluid
.
layers
.
cross_entropy
(
input
=
prediction
,
label
=
input_y
)
avg_cost
=
paddle
.
fluid
.
layers
.
mean
(
x
=
cost
)
fleet
.
init
(
is_collective
=
True
)
meta_list
=
fleet
.
_get_applied_meta_list
()
graph_list
=
fleet
.
_get_applied_graph_list
()
# not called minimize function
self
.
assertEqual
(
len
(
meta_list
),
0
)
self
.
assertEqual
(
len
(
graph_list
),
0
)
strategy
=
fleet
.
DistributedStrategy
()
optimizer
=
paddle
.
fluid
.
optimizer
.
SGD
(
learning_rate
=
0.001
)
optimizer
=
fleet
.
distributed_optimizer
(
optimizer
,
strategy
=
strategy
)
optimizer
.
minimize
(
avg_cost
)
meta_list
=
fleet
.
_get_applied_meta_list
()
graph_list
=
fleet
.
_get_applied_graph_list
()
self
.
assertEqual
(
len
(
meta_list
),
0
)
self
.
assertEqual
(
len
(
graph_list
),
1
)
if
__name__
==
"__main__"
:
unittest
.
main
()
python/paddle/fluid/tests/unittests/test_fleet_base_4.py
浏览文件 @
fb641c91
...
...
@@ -16,6 +16,9 @@ import unittest
import
paddle
import
os
import
paddle.fluid
as
fluid
import
paddle.distributed.fleet
as
fleet
paddle
.
enable_static
()
class
TestFleetBase
(
unittest
.
TestCase
):
...
...
@@ -27,7 +30,6 @@ class TestFleetBase(unittest.TestCase):
"127.0.0.1:36001,127.0.0.2:36001"
def
test_fleet_init
(
self
):
import
paddle.distributed.fleet
as
fleet
os
.
environ
[
"TRAINING_ROLE"
]
=
"PSERVER"
os
.
environ
[
"POD_IP"
]
=
"127.0.0.1"
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录