Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
08524758
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
08524758
编写于
12月 12, 2022
作者:
姜
姜永久
提交者:
GitHub
12月 13, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
remove fleet eager guard tests (#48765)
上级
f3982a9d
变更
13
隐藏空白更改
内联
并排
Showing
13 changed file
with
14 addition
and
46 deletion
+14
-46
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_api.py
...s/unittests/collective/fleet/dygraph_group_sharded_api.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_api_eager.py
...tests/collective/fleet/dygraph_group_sharded_api_eager.py
+1
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage2.py
...nittests/collective/fleet/dygraph_group_sharded_stage2.py
+1
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage2_comm_overlap.py
...ective/fleet/dygraph_group_sharded_stage2_comm_overlap.py
+1
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage2_offload.py
.../collective/fleet/dygraph_group_sharded_stage2_offload.py
+1
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage3.py
...nittests/collective/fleet/dygraph_group_sharded_stage3.py
+1
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage3_offload.py
.../collective/fleet/dygraph_group_sharded_stage3_offload.py
+1
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_optimizer_stage2.py
...sts/collective/fleet/dygraph_sharding_optimizer_stage2.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage2.py
...sts/unittests/collective/fleet/dygraph_sharding_stage2.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage2_offload.py
...tests/collective/fleet/dygraph_sharding_stage2_offload.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage3.py
...sts/unittests/collective/fleet/dygraph_sharding_stage3.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage3_offload.py
...tests/collective/fleet/dygraph_sharding_stage3_offload.py
+0
-3
python/paddle/fluid/tests/unittests/collective/init_process_group.py
...le/fluid/tests/unittests/collective/init_process_group.py
+8
-10
未找到文件。
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_api.py
浏览文件 @
08524758
...
@@ -24,7 +24,6 @@ from paddle.distributed.sharding import (
...
@@ -24,7 +24,6 @@ from paddle.distributed.sharding import (
group_sharded_parallel
,
group_sharded_parallel
,
save_group_sharded_model
,
save_group_sharded_model
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
epoch
=
10
epoch
=
10
...
@@ -196,7 +195,5 @@ def test_sharding_api():
...
@@ -196,7 +195,5 @@ def test_sharding_api():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
pass
fleet
.
init
(
is_collective
=
True
)
fleet
.
init
(
is_collective
=
True
)
test_sharding_api
()
test_sharding_api
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_api_eager.py
浏览文件 @
08524758
...
@@ -22,7 +22,6 @@ from paddle.distributed.sharding import (
...
@@ -22,7 +22,6 @@ from paddle.distributed.sharding import (
group_sharded_parallel
,
group_sharded_parallel
,
save_group_sharded_model
,
save_group_sharded_model
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
epoch
=
10
epoch
=
10
...
@@ -199,5 +198,4 @@ def test_sharding_api():
...
@@ -199,5 +198,4 @@ def test_sharding_api():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
test_sharding_api
()
test_sharding_api
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage2.py
浏览文件 @
08524758
...
@@ -28,7 +28,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_optimizer_sta
...
@@ -28,7 +28,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_optimizer_sta
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage2
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage2
import
(
GroupShardedStage2
,
GroupShardedStage2
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
seed
=
2022
seed
=
2022
...
@@ -246,5 +245,4 @@ def test_dp_stage2():
...
@@ -246,5 +245,4 @@ def test_dp_stage2():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
test_dp_stage2
()
test_dp_stage2
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage2_comm_overlap.py
浏览文件 @
08524758
...
@@ -28,7 +28,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_optimizer_sta
...
@@ -28,7 +28,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_optimizer_sta
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage2
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage2
import
(
GroupShardedStage2
,
GroupShardedStage2
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
seed
=
2022
seed
=
2022
...
@@ -250,5 +249,4 @@ def test_dp_stage2():
...
@@ -250,5 +249,4 @@ def test_dp_stage2():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
test_dp_stage2
()
test_dp_stage2
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage2_offload.py
浏览文件 @
08524758
...
@@ -31,7 +31,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage2 import
...
@@ -31,7 +31,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage2 import
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_utils
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_utils
import
(
GroupShardedScaler
,
GroupShardedScaler
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
seed
=
2021
seed
=
2021
epoch
=
2
epoch
=
2
...
@@ -115,5 +114,4 @@ def test_sharding_stage2_offload():
...
@@ -115,5 +114,4 @@ def test_sharding_stage2_offload():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
test_sharding_stage2_offload
()
test_sharding_stage2_offload
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage3.py
浏览文件 @
08524758
...
@@ -34,7 +34,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage3 import
...
@@ -34,7 +34,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage3 import
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_utils
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_utils
import
(
GroupShardedScaler
,
GroupShardedScaler
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
epoch
=
10
epoch
=
10
...
@@ -320,5 +319,4 @@ def test_stage2_stage3():
...
@@ -320,5 +319,4 @@ def test_stage2_stage3():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
test_stage2_stage3
()
test_stage2_stage3
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_stage3_offload.py
浏览文件 @
08524758
...
@@ -24,7 +24,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage3 import
...
@@ -24,7 +24,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.group_sharded_stage3 import
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_utils
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.group_sharded_utils
import
(
GroupShardedScaler
,
GroupShardedScaler
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
epoch
=
10
epoch
=
10
...
@@ -220,5 +219,4 @@ def test_stage3_offload():
...
@@ -220,5 +219,4 @@ def test_stage3_offload():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
test_stage3_offload
()
test_stage3_offload
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_optimizer_stage2.py
浏览文件 @
08524758
...
@@ -23,7 +23,6 @@ from paddle.distributed.fleet.meta_optimizers.dygraph_optimizer.sharding_optimiz
...
@@ -23,7 +23,6 @@ from paddle.distributed.fleet.meta_optimizers.dygraph_optimizer.sharding_optimiz
ShardingOptimizerStage2
,
ShardingOptimizerStage2
,
)
)
from
paddle.distributed.fleet.utils.internal_storage
import
GradStorage
from
paddle.distributed.fleet.utils.internal_storage
import
GradStorage
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
base_lr
=
0.1
base_lr
=
0.1
...
@@ -142,6 +141,4 @@ def train_mlp():
...
@@ -142,6 +141,4 @@ def train_mlp():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
pass
train_mlp
()
train_mlp
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage2.py
浏览文件 @
08524758
...
@@ -29,7 +29,6 @@ from paddle.distributed.fleet.meta_optimizers.dygraph_optimizer.sharding_optimiz
...
@@ -29,7 +29,6 @@ from paddle.distributed.fleet.meta_optimizers.dygraph_optimizer.sharding_optimiz
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_stage2
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_stage2
import
(
ShardingStage2
,
ShardingStage2
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
seed
=
2022
seed
=
2022
...
@@ -239,7 +238,5 @@ def test_dp_stage2():
...
@@ -239,7 +238,5 @@ def test_dp_stage2():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
pass
fleet
.
init
(
is_collective
=
True
,
strategy
=
strategy
)
fleet
.
init
(
is_collective
=
True
,
strategy
=
strategy
)
test_dp_stage2
()
test_dp_stage2
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage2_offload.py
浏览文件 @
08524758
...
@@ -28,7 +28,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.sharding_stage2 import (
...
@@ -28,7 +28,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.sharding_stage2 import (
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_utils
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_utils
import
(
ShardingScaler
,
ShardingScaler
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
seed
=
2021
seed
=
2021
epoch
=
2
epoch
=
2
...
@@ -119,7 +118,5 @@ def test_sharding_stage2_offload():
...
@@ -119,7 +118,5 @@ def test_sharding_stage2_offload():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
pass
fleet
.
init
(
is_collective
=
True
,
strategy
=
strategy
)
fleet
.
init
(
is_collective
=
True
,
strategy
=
strategy
)
test_sharding_stage2_offload
()
test_sharding_stage2_offload
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage3.py
浏览文件 @
08524758
...
@@ -35,7 +35,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.sharding_stage3 import (
...
@@ -35,7 +35,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.sharding_stage3 import (
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_utils
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_utils
import
(
ShardingScaler
,
ShardingScaler
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
epoch
=
10
epoch
=
10
...
@@ -316,7 +315,5 @@ def test_stage2_stage3():
...
@@ -316,7 +315,5 @@ def test_stage2_stage3():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
pass
fleet
.
init
(
is_collective
=
True
)
fleet
.
init
(
is_collective
=
True
)
test_stage2_stage3
()
test_stage2_stage3
()
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_sharding_stage3_offload.py
浏览文件 @
08524758
...
@@ -25,7 +25,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.sharding_stage3 import (
...
@@ -25,7 +25,6 @@ from paddle.distributed.fleet.meta_parallel.sharding.sharding_stage3 import (
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_utils
import
(
from
paddle.distributed.fleet.meta_parallel.sharding.sharding_utils
import
(
ShardingScaler
,
ShardingScaler
,
)
)
from
paddle.fluid.framework
import
_test_eager_guard
from
paddle.nn
import
Linear
from
paddle.nn
import
Linear
epoch
=
10
epoch
=
10
...
@@ -216,7 +215,5 @@ def test_stage3_offload():
...
@@ -216,7 +215,5 @@ def test_stage3_offload():
if
__name__
==
'__main__'
:
if
__name__
==
'__main__'
:
with
_test_eager_guard
():
pass
fleet
.
init
(
is_collective
=
True
)
fleet
.
init
(
is_collective
=
True
)
test_stage3_offload
()
test_stage3_offload
()
python/paddle/fluid/tests/unittests/collective/init_process_group.py
浏览文件 @
08524758
...
@@ -15,7 +15,6 @@
...
@@ -15,7 +15,6 @@
import
unittest
import
unittest
import
paddle
import
paddle
from
paddle.fluid.framework
import
_test_eager_guard
class
TestProcessGroupFp32
(
unittest
.
TestCase
):
class
TestProcessGroupFp32
(
unittest
.
TestCase
):
...
@@ -26,15 +25,14 @@ class TestProcessGroupFp32(unittest.TestCase):
...
@@ -26,15 +25,14 @@ class TestProcessGroupFp32(unittest.TestCase):
pass
pass
def
test_init_process_group
(
self
):
def
test_init_process_group
(
self
):
with
_test_eager_guard
():
paddle
.
distributed
.
init_parallel_env
()
paddle
.
distributed
.
init_parallel_env
()
paddle
.
distributed
.
new_group
()
paddle
.
distributed
.
new_group
()
group
=
paddle
.
distributed
.
new_group
([
-
1
,
-
2
])
group
=
paddle
.
distributed
.
new_group
([
-
1
,
-
2
])
assert
group
.
process_group
is
None
assert
group
.
process_group
is
None
group
=
paddle
.
distributed
.
collective
.
Group
(
-
1
,
2
,
0
,
[
-
1
,
-
2
])
group
=
paddle
.
distributed
.
collective
.
Group
(
-
1
,
2
,
0
,
[
-
1
,
-
2
])
ret
=
paddle
.
distributed
.
barrier
(
group
)
ret
=
paddle
.
distributed
.
barrier
(
group
)
assert
ret
is
None
assert
ret
is
None
paddle
.
enable_static
()
paddle
.
enable_static
()
in_tensor
=
paddle
.
empty
((
1
,
2
))
in_tensor
=
paddle
.
empty
((
1
,
2
))
in_tensor2
=
paddle
.
empty
((
1
,
2
))
in_tensor2
=
paddle
.
empty
((
1
,
2
))
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录