Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
a1d6e2a8
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
a1d6e2a8
编写于
12月 22, 2022
作者:
W
wanghuancoder
提交者:
GitHub
12月 22, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
delete distribute old dygraph test cast (#49100)
* delete distribute old dygraph test cast
上级
80fe8cbc
变更
11
显示空白变更内容
内联
并排
Showing
11 changed file
with
0 addition
and
67 deletion
+0
-67
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_class_center_sample.py
...sts/collective/fleet/test_parallel_class_center_sample.py
+0
-1
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_mp_layers.py
...tests/collective/fleet/test_parallel_dygraph_mp_layers.py
+0
-1
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_no_sync_gradient_check.py
...ive/fleet/test_parallel_dygraph_no_sync_gradient_check.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_pipeline_parallel.py
...llective/fleet/test_parallel_dygraph_pipeline_parallel.py
+0
-20
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_sharding_parallel.py
...llective/fleet/test_parallel_dygraph_sharding_parallel.py
+0
-3
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_tensor_parallel.py
...collective/fleet/test_parallel_dygraph_tensor_parallel.py
+0
-5
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_margin_cross_entropy.py
...ts/collective/fleet/test_parallel_margin_cross_entropy.py
+0
-3
python/paddle/fluid/tests/unittests/collective/test_collective_global_gather.py
...sts/unittests/collective/test_collective_global_gather.py
+0
-9
python/paddle/fluid/tests/unittests/collective/test_collective_global_scatter.py
...ts/unittests/collective/test_collective_global_scatter.py
+0
-9
python/paddle/fluid/tests/unittests/test_parallel_dygraph_dataparallel.py
...uid/tests/unittests/test_parallel_dygraph_dataparallel.py
+0
-10
python/paddle/tests/test_dist_hapi_model.py
python/paddle/tests/test_dist_hapi_model.py
+0
-3
未找到文件。
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_class_center_sample.py
浏览文件 @
a1d6e2a8
...
...
@@ -20,7 +20,6 @@ from test_parallel_dygraph_dataparallel import TestMultipleGpus
class
TestParallelClassCenterSample
(
TestMultipleGpus
):
def
test_parallel_class_center_sample
(
self
):
self
.
run_mnist_2gpu
(
'parallel_class_center_sample.py'
)
self
.
run_mnist_2gpu
(
'parallel_class_center_sample.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_mp_layers.py
浏览文件 @
a1d6e2a8
...
...
@@ -20,7 +20,6 @@ from test_parallel_dygraph_dataparallel import TestMultipleGpus
class
TestModelParallelLayer
(
TestMultipleGpus
):
def
test_hybrid_parallel_mp_layer
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_layers.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_layers.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_no_sync_gradient_check.py
浏览文件 @
a1d6e2a8
...
...
@@ -20,9 +20,6 @@ from test_parallel_dygraph_dataparallel import TestMultipleGpus
class
TestDataParallelLayer
(
TestMultipleGpus
):
def
test_parallel_dygraph_dataparallel_no_sync
(
self
):
self
.
run_mnist_2gpu
(
'parallel_dygraph_no_sync_gradient_check.py'
)
self
.
run_mnist_2gpu
(
'parallel_dygraph_no_sync_gradient_check.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_pipeline_parallel.py
浏览文件 @
a1d6e2a8
...
...
@@ -23,53 +23,33 @@ class TestHybridPipeParallel(TestMultipleGpus):
self
.
run_mnist_2gpu
(
os
.
path
.
abspath
(
'../../hybrid_parallel_pp_layer.py'
)
)
self
.
run_mnist_2gpu
(
os
.
path
.
abspath
(
'../../hybrid_parallel_pp_layer.py'
),
eager_mode
=
False
,
)
def
test_hybrid_parallel_pp_tuple_inputs
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_embedding.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_embedding.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_shared_weight
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_shared_weight.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_shared_weight.py'
,
eager_mode
=
False
)
def
test_pipeline_parallel_amp
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_amp.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_amp.py'
,
eager_mode
=
False
)
def
test_pipeline_parallel_fp16
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_fp16.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_fp16.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_transformer
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_transformer.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_transformer.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_save_load
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_save_load.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_save_load.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_recompute
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_recompute.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_recompute.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_pp_clip_grad
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_clip_grad.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_clip_grad.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_transformer_unbalanced_data
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_transformer_unbalanced_data.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_pp_transformer_unbalanced_data.py'
,
eager_mode
=
False
,
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_sharding_parallel.py
浏览文件 @
a1d6e2a8
...
...
@@ -22,9 +22,6 @@ class TestHybridParallel(TestMultipleGpus):
# check sharding logic as well as the accuracy with single mode
def
test_hybrid_parallel_sharding_logic
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_sharding_model.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_sharding_model.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_dygraph_tensor_parallel.py
浏览文件 @
a1d6e2a8
...
...
@@ -20,23 +20,18 @@ from test_parallel_dygraph_dataparallel import TestMultipleGpus
class
TestHybridParallel
(
TestMultipleGpus
):
def
test_hybrid_parallel_mp_random
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_random.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_random.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_mp_model
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_model.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_model.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_mp_amp
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_amp.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_amp.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_mp_fp16
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_fp16.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_fp16.py'
,
eager_mode
=
False
)
def
test_hybrid_parallel_mp_clip_grad
(
self
):
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_clip_grad.py'
)
self
.
run_mnist_2gpu
(
'hybrid_parallel_mp_clip_grad.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/fleet/test_parallel_margin_cross_entropy.py
浏览文件 @
a1d6e2a8
...
...
@@ -20,9 +20,6 @@ from test_parallel_dygraph_dataparallel import TestMultipleGpus
class
TestParallelMarginSoftmaxWithCrossEntropy
(
TestMultipleGpus
):
def
test_parallel_margin_cross_entropy
(
self
):
self
.
run_mnist_2gpu
(
'parallel_margin_cross_entropy.py'
)
self
.
run_mnist_2gpu
(
'parallel_margin_cross_entropy.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
python/paddle/fluid/tests/unittests/collective/test_collective_global_gather.py
浏览文件 @
a1d6e2a8
...
...
@@ -29,15 +29,6 @@ class TestCollectiveGlobalGatherAPI(TestDistBase):
"collective_global_gather.py"
,
"global_gather"
,
"nccl"
)
def
test_global_gather_nccl_dygraph
(
self
):
self
.
check_with_place
(
"collective_global_gather_dygraph.py"
,
"global_gather"
,
"nccl"
,
static_mode
=
"0"
,
eager_mode
=
False
,
)
def
test_global_gather_nccl_dygraph_eager
(
self
):
self
.
check_with_place
(
"collective_global_gather_dygraph.py"
,
...
...
python/paddle/fluid/tests/unittests/collective/test_collective_global_scatter.py
浏览文件 @
a1d6e2a8
...
...
@@ -29,15 +29,6 @@ class TestCollectiveSelectScatterAPI(TestDistBase):
"collective_global_scatter.py"
,
"global_scatter"
,
"nccl"
)
def
test_global_scatter_nccl_dygraph
(
self
):
self
.
check_with_place
(
"collective_global_scatter_dygraph.py"
,
"global_scatter"
,
"nccl"
,
static_mode
=
"0"
,
eager_mode
=
False
,
)
def
test_global_scatter_nccl_dygraph_eager
(
self
):
self
.
check_with_place
(
"collective_global_scatter_dygraph.py"
,
...
...
python/paddle/fluid/tests/unittests/test_parallel_dygraph_dataparallel.py
浏览文件 @
a1d6e2a8
...
...
@@ -213,19 +213,9 @@ class TestMultipleWithGloo(unittest.TestCase):
time
.
sleep
(
3
)
class
TestDataParallelGradientCheck
(
TestMultipleGpus
):
def
test_multiple_gpus_dynamic
(
self
):
self
.
run_mnist_2gpu
(
'parallel_dygraph_gradient_check.py'
,
eager_mode
=
False
)
class
TestDataParallelWithPyLayer
(
TestMultipleGpus
):
def
test_parallel_dygraph_dataparallel_with_pylayer
(
self
):
self
.
run_mnist_2gpu
(
'parallel_dygraph_dataparallel_with_pylayer.py'
)
self
.
run_mnist_2gpu
(
'parallel_dygraph_dataparallel_with_pylayer.py'
,
eager_mode
=
False
)
self
.
run_mnist_2gpu
(
'parallel_dygraph_dataparallel_with_pylayer.py'
,
allocator_strategy
=
"naive_best_fit"
,
...
...
python/paddle/tests/test_dist_hapi_model.py
浏览文件 @
a1d6e2a8
...
...
@@ -133,15 +133,12 @@ class TestMultipleGpus(unittest.TestCase):
def
test_hapi_multiple_gpus_static
(
self
):
self
.
run_mnist_2gpu
(
'dist_hapi_mnist_static.py'
)
self
.
run_mnist_2gpu
(
'dist_hapi_mnist_static.py'
,
eager_mode
=
False
)
def
test_hapi_multiple_gpus_dynamic
(
self
):
self
.
run_mnist_2gpu
(
'dist_hapi_mnist_dynamic.py'
)
self
.
run_mnist_2gpu
(
'dist_hapi_mnist_dynamic.py'
,
eager_mode
=
False
)
def
test_hapi_amp_static
(
self
):
self
.
run_mnist_2gpu
(
'dist_hapi_pure_fp16_static.py'
)
self
.
run_mnist_2gpu
(
'dist_hapi_pure_fp16_static.py'
,
eager_mode
=
False
)
if
__name__
==
"__main__"
:
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录