Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
af0ad915
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
af0ad915
编写于
3月 30, 2020
作者:
G
gongweibao
提交者:
GitHub
3月 30, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix testlaunch test=develop (#23304)
上级
2169e6fb
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
7 addition
and
5 deletion
+7
-5
python/paddle/fluid/tests/unittests/CMakeLists.txt
python/paddle/fluid/tests/unittests/CMakeLists.txt
+1
-1
python/paddle/fluid/tests/unittests/test_launch.sh
python/paddle/fluid/tests/unittests/test_launch.sh
+6
-4
未找到文件。
python/paddle/fluid/tests/unittests/CMakeLists.txt
浏览文件 @
af0ad915
...
@@ -300,7 +300,7 @@ if(WITH_DISTRIBUTE)
...
@@ -300,7 +300,7 @@ if(WITH_DISTRIBUTE)
if
(
NOT APPLE
)
if
(
NOT APPLE
)
if
(
WITH_GPU
)
if
(
WITH_GPU
)
# NOTE. test_launch only work in gpu collective mode
# NOTE. test_launch only work in gpu collective mode
bash_test_modules
(
test_launch MODULES test_launch.sh
)
bash_test_modules
(
test_launch MODULES test_launch.sh
ENVS PADDLE_BINARY_DIR=
${
PADDLE_BINARY_DIR
}
)
endif
()
endif
()
bash_test_modules
(
test_launch_ps MODULES test_launch_ps.sh
)
bash_test_modules
(
test_launch_ps MODULES test_launch_ps.sh
)
...
...
python/paddle/fluid/tests/unittests/test_launch.sh
浏览文件 @
af0ad915
#!/bin/bash
#!/bin/bash
set
-e
set
-e
# use default values
# use default values
python
-m
paddle.distributed.launch multi_process.py
# FIXME: random fails on Unknown command lines -c (or -m).
launch_py
=
${
PADDLE_BINARY_DIR
}
/python/paddle/distributed/launch.py
python
${
launch_py
}
multi_process.py
# use paddlecloud
# use paddlecloud
cluster_node_ips
=
"10.0.0.1"
cluster_node_ips
=
"10.0.0.1"
...
@@ -15,7 +17,7 @@ export PADDLE_PORT=35019
...
@@ -15,7 +17,7 @@ export PADDLE_PORT=35019
export
PADDLE_PORTS_NUM
=
2
export
PADDLE_PORTS_NUM
=
2
distributed_args
=
"--use_paddlecloud --cluster_node_ips=
${
cluster_node_ips
}
--node_ip=
${
node_ip
}
--selected_gpus=0,1 --log_dir=testlog"
distributed_args
=
"--use_paddlecloud --cluster_node_ips=
${
cluster_node_ips
}
--node_ip=
${
node_ip
}
--selected_gpus=0,1 --log_dir=testlog"
CUDA_VISIBLE_DEVICES
=
0,1 python
-m
paddle.distributed.launch
${
distributed_args
}
multi_process.py
CUDA_VISIBLE_DEVICES
=
0,1 python
${
launch_py
}
${
distributed_args
}
multi_process.py
str1
=
"selected_gpus:0 worker_endpoints:127.0.0.1:35019,127.0.0.1:35020,127.0.0.2:35019,127.0.0.2:35020 trainers_num:4 current_endpoint:127.0.0.1:35019 trainer_id:0"
str1
=
"selected_gpus:0 worker_endpoints:127.0.0.1:35019,127.0.0.1:35020,127.0.0.2:35019,127.0.0.2:35020 trainers_num:4 current_endpoint:127.0.0.1:35019 trainer_id:0"
str2
=
"selected_gpus:1 worker_endpoints:127.0.0.1:35019,127.0.0.1:35020,127.0.0.2:35019,127.0.0.2:35020 trainers_num:4 current_endpoint:127.0.0.1:35020 trainer_id:1"
str2
=
"selected_gpus:1 worker_endpoints:127.0.0.1:35019,127.0.0.1:35020,127.0.0.2:35019,127.0.0.2:35020 trainers_num:4 current_endpoint:127.0.0.1:35020 trainer_id:1"
...
@@ -50,7 +52,7 @@ unset PADDLE_PORTS_NUM
...
@@ -50,7 +52,7 @@ unset PADDLE_PORTS_NUM
echo
""
echo
""
echo
"paddle.distributed.launch async poll process test"
echo
"paddle.distributed.launch async poll process test"
if
!
CUDA_VISIBLE_DEVICES
=
0,1 python
-m
paddle.distributed.launch
${
distributed_args
}
multi_process.py abort
;
then
if
!
CUDA_VISIBLE_DEVICES
=
0,1 python
${
launch_py
}
${
distributed_args
}
multi_process.py abort
;
then
echo
"train abort as planned"
echo
"train abort as planned"
fi
fi
...
@@ -77,5 +79,5 @@ rm -rf $file_0_0 $file_0_1
...
@@ -77,5 +79,5 @@ rm -rf $file_0_0 $file_0_1
distributed_args
=
"--selected_gpus=0,1 --log_dir=testlog"
distributed_args
=
"--selected_gpus=0,1 --log_dir=testlog"
export
PADDLE_LAUNCH_LOG
=
"test_launch_filelock_0"
export
PADDLE_LAUNCH_LOG
=
"test_launch_filelock_0"
CUDA_VISIBLE_DEVICES
=
0,1 python
-m
paddle.distributed.launch
${
distributed_args
}
find_ports.py
CUDA_VISIBLE_DEVICES
=
0,1 python
${
launch_py
}
${
distributed_args
}
find_ports.py
str_0
=
"worker_endpoints:127.0.0.1:6070,127.0.0.1:6071"
str_0
=
"worker_endpoints:127.0.0.1:6070,127.0.0.1:6071"
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录