diff --git a/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh b/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh index 2c0fc0b06299dc4aa8099354c4765176accae5ec..f50e24f10beca4d20a2d0b239e390a661ecd19ae 100644 --- a/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh +++ b/python/paddle/fluid/tests/unittests/test_fleet_launch_async.sh @@ -27,13 +27,13 @@ export PADDLE_TRAINER_ID=0 export TRAINER_PORTS_NUM=2 -file_0="multi_process_fleetrun.check_0.log" -file_1="multi_process_fleetrun.check_1.log" +file_0="multi_process_fullpath_launch.check_0.log" +file_1="multi_process_fullpath_launch.check_1.log" distributed_args="--ips=${cluster_node_ips} --gpus=0,1 --log_dir=testlog" echo "paddle.distributed.fleet.launch async poll process test" -if ! CUDA_VISIBLE_DEVICES=0,1 python -m paddle.distributed.fleet.launch ${distributed_args} multi_process.py fleetrun abort; then +if ! CUDA_VISIBLE_DEVICES=0,1 python -m paddle.distributed.fleet.launch ${distributed_args} multi_process.py fullpath_launch abort; then echo "train abort as planned" fi