From 89890534433f7640e2330d129b148a85fbfbf2b4 Mon Sep 17 00:00:00 2001 From: gongweibao Date: Fri, 4 Dec 2020 14:29:43 +0800 Subject: [PATCH] Fix bug of test_fleet_launch_async.sh (#29332) --- .../paddle/fluid/tests/unittests/test_fleet_launch_async.sh | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh b/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh index 2c0fc0b0629..f50e24f10be 100644 --- a/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh +++ b/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh @@ -27,13 +27,13 @@ export PADDLE_TRAINER_ID=0 export TRAINER_PORTS_NUM=2 -file_0="multi_process_fleetrun.check_0.log" -file_1="multi_process_fleetrun.check_1.log" +file_0="multi_process_fullpath_launch.check_0.log" +file_1="multi_process_fullpath_launch.check_1.log" distributed_args="--ips=${cluster_node_ips} --gpus=0,1 --log_dir=testlog" echo "paddle.distributed.fleet.launch async poll process test" -if ! CUDA_VISIBLE_DEVICES=0,1 python -m paddle.distributed.fleet.launch ${distributed_args} multi_process.py fleetrun abort; then +if ! CUDA_VISIBLE_DEVICES=0,1 python -m paddle.distributed.fleet.launch ${distributed_args} multi_process.py fullpath_launch abort; then echo "train abort as planned" fi -- GitLab