未验证 提交 36c04102 编写于 作者: L lilong12 提交者: GitHub

Revert "Initialize gloo for low level collective apis (#27356)", test=document_fix (#27665)

上级 99e33373
无相关合并请求
...@@ -21,10 +21,10 @@ void GlooParallelContext::Init() { ...@@ -21,10 +21,10 @@ void GlooParallelContext::Init() {
auto gloo_ptr = paddle::framework::GlooWrapper::GetInstance(); auto gloo_ptr = paddle::framework::GlooWrapper::GetInstance();
gloo_ptr->SetRank(strategy_.rank); gloo_ptr->SetRank(strategy_.rank);
gloo_ptr->SetSize(strategy_.rank_num); gloo_ptr->SetSize(strategy_.rank_num);
gloo_ptr->SetPrefix(strategy_.prefix);
gloo_ptr->SetIface(strategy_.iface); gloo_ptr->SetIface(strategy_.iface);
gloo_ptr->SetTimeoutSeconds(strategy_.init_seconds, strategy_.run_seconds); gloo_ptr->SetTimeoutSeconds(strategy_.init_seconds, strategy_.run_seconds);
gloo_ptr->SetHttpStore(strategy_.ip_address, strategy_.ip_port, gloo_ptr->SetHdfsStore(strategy_.path, strategy_.fs_name, strategy_.fs_ugi);
strategy_.scope);
gloo_ptr->Init(); gloo_ptr->Init();
} }
#endif #endif
......
...@@ -25,11 +25,12 @@ struct GlooParallelStrategy { ...@@ -25,11 +25,12 @@ struct GlooParallelStrategy {
int rank{0}; int rank{0};
int rank_num{1}; int rank_num{1};
std::string iface; std::string iface;
std::string prefix;
int init_seconds{9999999}; int init_seconds{9999999};
int run_seconds{9999999}; int run_seconds{9999999};
std::string ip_address; std::string path;
int ip_port; std::string fs_name;
std::string scope{"worker"}; std::string fs_ugi;
}; };
class GlooParallelContext { class GlooParallelContext {
......
...@@ -62,6 +62,12 @@ void BindGlooContext(py::module *m) { ...@@ -62,6 +62,12 @@ void BindGlooContext(py::module *m) {
[](platform::GlooParallelStrategy &self, const std::string &iface) { [](platform::GlooParallelStrategy &self, const std::string &iface) {
self.iface = iface; self.iface = iface;
}) })
.def_property("prefix",
[](const platform::GlooParallelStrategy &self) {
return self.prefix;
},
[](platform::GlooParallelStrategy &self,
const std::string &prefix) { self.prefix = prefix; })
.def_property("init_seconds", .def_property("init_seconds",
[](const platform::GlooParallelStrategy &self) { [](const platform::GlooParallelStrategy &self) {
return self.init_seconds; return self.init_seconds;
...@@ -77,19 +83,23 @@ void BindGlooContext(py::module *m) { ...@@ -77,19 +83,23 @@ void BindGlooContext(py::module *m) {
self.run_seconds = run_seconds; self.run_seconds = run_seconds;
}) })
.def_property( .def_property(
"ip_address", "path",
[](const platform::GlooParallelStrategy &self) { [](const platform::GlooParallelStrategy &self) { return self.path; },
return self.ip_address; [](platform::GlooParallelStrategy &self, const std::string &path) {
}, self.path = path;
[](platform::GlooParallelStrategy &self, })
const std::string &ip_address) { self.ip_address = ip_address; }) .def_property("fs_name",
.def_property("ip_port", [](const platform::GlooParallelStrategy &self) {
return self.fs_name;
},
[](platform::GlooParallelStrategy &self,
const std::string &fs_name) { self.fs_name = fs_name; })
.def_property("fs_ugi",
[](const platform::GlooParallelStrategy &self) { [](const platform::GlooParallelStrategy &self) {
return self.ip_port; return self.fs_ugi;
}, },
[](platform::GlooParallelStrategy &self, int ip_port) { [](platform::GlooParallelStrategy &self,
self.ip_port = ip_port; const std::string &fs_ugi) { self.fs_ugi = fs_ugi; });
});
py::class_<platform::GlooParallelContext> gloo_ctx(*m, "GlooParallelContext"); py::class_<platform::GlooParallelContext> gloo_ctx(*m, "GlooParallelContext");
gloo_ctx.def(py::init<const platform::GlooParallelStrategy &>()) gloo_ctx.def(py::init<const platform::GlooParallelStrategy &>())
......
...@@ -169,7 +169,8 @@ class Gloo(object): ...@@ -169,7 +169,8 @@ class Gloo(object):
http_server = KVServer(port, size_d) http_server = KVServer(port, size_d)
http_server.start() http_server.start()
wait_seconds = 5 wait_seconds = 5
while http_server_d.get("running", False): while http_server_d.get("running",
False) and not http_server.shoud_stop():
time.sleep(wait_seconds) time.sleep(wait_seconds)
http_server.stop() http_server.stop()
......
...@@ -181,7 +181,7 @@ class KVServer: ...@@ -181,7 +181,7 @@ class KVServer:
self.listen_thread.join() self.listen_thread.join()
self.http_server.server_close() self.http_server.server_close()
def should_stop(self): def shoud_stop(self):
""" """
return whether the server should stop. return whether the server should stop.
......
...@@ -15,10 +15,6 @@ ...@@ -15,10 +15,6 @@
import os import os
import six import six
import warnings import warnings
from multiprocessing import Process, Manager
import netifaces
import time
import sys
from paddle import compat as cpt from paddle import compat as cpt
...@@ -33,39 +29,6 @@ __all__ = ["init_parallel_env"] ...@@ -33,39 +29,6 @@ __all__ = ["init_parallel_env"]
ParallelStrategy = core.ParallelStrategy ParallelStrategy = core.ParallelStrategy
def _start_kv_server(port, http_server_d):
from paddle.distributed.fleet.utils.http_server import KVServer
http_server = KVServer(int(port))
http_server.start()
wait_seconds = 5
while http_server_d.get("running", False):
time.sleep(wait_seconds)
http_server.stop()
def _get_iface_by_ip(ip_address):
"""
Get network interface name by its ip address.
Args:
ip_address (string): ip address
Returns:
Name of the network interface which has the ip address 'ip_address',
or None if the ip_address is not found.
"""
nicfaces = netifaces.interfaces()
for nicface in nicfaces:
message = netifaces.ifaddresses(nicface)
iface_info = message.get(netifaces.AF_INET)
if iface_info:
iface_dict = iface_info[0]
ipaddr = iface_dict.get('addr')
if ipaddr == ip_address:
return nicface
return None
def init_parallel_env(): def init_parallel_env():
""" """
Initialize parallel training environment in dynamic graph mode. Initialize parallel training environment in dynamic graph mode.
...@@ -147,44 +110,7 @@ def init_parallel_env(): ...@@ -147,44 +110,7 @@ def init_parallel_env():
_check_var_exists("PADDLE_TRAINERS_NUM") _check_var_exists("PADDLE_TRAINERS_NUM")
_check_var_exists("PADDLE_TRAINER_ENDPOINTS") _check_var_exists("PADDLE_TRAINER_ENDPOINTS")
if ParallelEnv().world_size < 2: # 3. init NCCL ParallelStrategy
return
# 3: init gloo context
ep_rank_0 = ParallelEnv().trainer_endpoints[0].split(":")
ep_rank = ParallelEnv().trainer_endpoints[ParallelEnv().rank].split(":")
manager = Manager()
# glboal dict to store status
http_server_d = manager.dict()
http_server_d["running"] = False
if ParallelEnv().rank == 0:
http_server = Process(
target=_start_kv_server, args=(int(ep_rank_0[1]), http_server_d))
http_server.daemon = True
http_server_d["running"] = True
http_server.start()
iface = _get_iface_by_ip(ep_rank[0])
if iface is None:
raise ValueError("No network interface associated with the "
"ip address: {}.".format(ep_rank_0[0]))
gloo_strategy = core.GlooParallelStrategy()
gloo_strategy.rank = ParallelEnv().rank
gloo_strategy.rank_num = ParallelEnv().world_size
gloo_strategy.ip_address = ep_rank_0[0]
gloo_strategy.ip_port = int(ep_rank_0[1])
gloo_strategy.iface = iface
default_init_timeout_seconds = 3600
default_run_timeout_seconds = 9999999
gloo_strategy.init_seconds = default_init_timeout_seconds
gloo_strategy.run_seconds = default_run_timeout_seconds
gloo = core.GlooParallelContext(gloo_strategy)
gloo.init()
if ParallelEnv().rank == 0:
http_server_d["running"] = False
http_server.join()
# 4. init NCCL ParallelStrategy
strategy = ParallelStrategy() strategy = ParallelStrategy()
if parallel_helper._is_parallel_ctx_initialized(): if parallel_helper._is_parallel_ctx_initialized():
warnings.warn("The parallel environment has been initialized.") warnings.warn("The parallel environment has been initialized.")
...@@ -192,7 +118,8 @@ def init_parallel_env(): ...@@ -192,7 +118,8 @@ def init_parallel_env():
strategy.local_rank = ParallelEnv().rank strategy.local_rank = ParallelEnv().rank
strategy.trainer_endpoints = ParallelEnv().trainer_endpoints strategy.trainer_endpoints = ParallelEnv().trainer_endpoints
strategy.current_endpoint = ParallelEnv().current_endpoint strategy.current_endpoint = ParallelEnv().current_endpoint
if strategy.nranks < 2:
return
# NOTE(chenweihang): [ why config global place here? ] # NOTE(chenweihang): [ why config global place here? ]
# the dygraph mode will be set to default mode, # the dygraph mode will be set to default mode,
# users will not call `dygraph.guard` or `enable_dygraph` # users will not call `dygraph.guard` or `enable_dygraph`
......
...@@ -366,7 +366,7 @@ def spawn(func, args=(), nprocs=-1, join=True, daemon=False, **options): ...@@ -366,7 +366,7 @@ def spawn(func, args=(), nprocs=-1, join=True, daemon=False, **options):
device = get_device() device = get_device()
if device == 'cpu': if device == 'cpu':
# TODO: not supports cpu parallel now # TODO: not supports cpu parallel now
nprocs = _cpu_num() nprocs = _cpu_num
else: else:
nprocs = core.get_cuda_device_count() nprocs = core.get_cuda_device_count()
......
...@@ -989,7 +989,8 @@ class GeneralRoleMaker(RoleMakerBase): ...@@ -989,7 +989,8 @@ class GeneralRoleMaker(RoleMakerBase):
http_server = KVServer(int(self._http_ip_port[1]), size_d) http_server = KVServer(int(self._http_ip_port[1]), size_d)
http_server.start() http_server.start()
wait_seconds = 5 wait_seconds = 5
while http_server_d.get("running", False): while http_server_d.get("running",
False) and not http_server.shoud_stop():
time.sleep(wait_seconds) time.sleep(wait_seconds)
http_server.stop() http_server.stop()
......
...@@ -173,7 +173,7 @@ class KVServer: ...@@ -173,7 +173,7 @@ class KVServer:
self.listen_thread.join() self.listen_thread.join()
self.http_server.server_close() self.http_server.server_close()
def should_stop(self): def shoud_stop(self):
""" """
return whether the server should stop. return whether the server should stop.
......
...@@ -15,12 +15,6 @@ list(APPEND DIST_TEST_OPS test_parallel_dygraph_sparse_embedding) ...@@ -15,12 +15,6 @@ list(APPEND DIST_TEST_OPS test_parallel_dygraph_sparse_embedding)
list(APPEND DIST_TEST_OPS test_parallel_dygraph_transformer) list(APPEND DIST_TEST_OPS test_parallel_dygraph_transformer)
list(APPEND DIST_TEST_OPS test_listen_and_serv_op) list(APPEND DIST_TEST_OPS test_listen_and_serv_op)
list(APPEND DIST_TEST_OPS test_fleet_graph_execution_meta_optimizer) list(APPEND DIST_TEST_OPS test_fleet_graph_execution_meta_optimizer)
list(APPEND DIST_TEST_OPS test_collective_reduce_api)
list(APPEND DIST_TEST_OPS test_collective_scatter_api)
list(APPEND DIST_TEST_OPS test_collective_barrier_api)
list(APPEND DIST_TEST_OPS test_collective_allreduce_api)
list(APPEND DIST_TEST_OPS test_collective_broadcast_api)
list(APPEND DIST_TEST_OPS test_collective_allgather_api)
set(MIXED_DIST_TEST_OPS ${DIST_TEST_OPS}) set(MIXED_DIST_TEST_OPS ${DIST_TEST_OPS})
#remove distribute unittests. #remove distribute unittests.
list(APPEND MIXED_DIST_TEST_OPS test_dgc_op) list(APPEND MIXED_DIST_TEST_OPS test_dgc_op)
...@@ -68,6 +62,12 @@ if(NOT WITH_GPU OR WIN32) ...@@ -68,6 +62,12 @@ if(NOT WITH_GPU OR WIN32)
LIST(REMOVE_ITEM TEST_OPS test_broadcast) LIST(REMOVE_ITEM TEST_OPS test_broadcast)
LIST(REMOVE_ITEM TEST_OPS test_collective_reduce) LIST(REMOVE_ITEM TEST_OPS test_collective_reduce)
LIST(REMOVE_ITEM TEST_OPS test_collective_scatter) LIST(REMOVE_ITEM TEST_OPS test_collective_scatter)
LIST(REMOVE_ITEM TEST_OPS test_collective_reduce_api)
LIST(REMOVE_ITEM TEST_OPS test_collective_scatter_api)
LIST(REMOVE_ITEM TEST_OPS test_collective_barrier_api)
LIST(REMOVE_ITEM TEST_OPS test_collective_allreduce_api)
LIST(REMOVE_ITEM TEST_OPS test_collective_broadcast_api)
LIST(REMOVE_ITEM TEST_OPS test_collective_allgather_api)
LIST(REMOVE_ITEM TEST_OPS test_reducescatter) LIST(REMOVE_ITEM TEST_OPS test_reducescatter)
LIST(REMOVE_ITEM TEST_OPS test_reducescatter_api) LIST(REMOVE_ITEM TEST_OPS test_reducescatter_api)
endif() endif()
......
...@@ -26,7 +26,6 @@ import functools ...@@ -26,7 +26,6 @@ import functools
import pickle import pickle
from contextlib import closing from contextlib import closing
from six import string_types from six import string_types
import paddle
import paddle.fluid as fluid import paddle.fluid as fluid
import paddle.fluid.unique_name as nameGen import paddle.fluid.unique_name as nameGen
from paddle.fluid import core from paddle.fluid import core
...@@ -61,6 +60,38 @@ class TestCollectiveAPIRunnerBase(object): ...@@ -61,6 +60,38 @@ class TestCollectiveAPIRunnerBase(object):
else: else:
break break
def initCommunicator(self, program, rank, nranks, wait_port,
current_endpoint, endpoints):
other_endpoints = endpoints[:]
other_endpoints.remove(current_endpoint)
if rank == 0 and wait_port:
self.wait_server_ready(other_endpoints)
block = program.global_block()
nccl_id_var = block.create_var(
name=nameGen.generate('nccl_id'),
persistable=True,
type=core.VarDesc.VarType.RAW)
block.append_op(
type='c_gen_nccl_id',
inputs={},
outputs={'Out': nccl_id_var},
attrs={
'rank': rank,
'endpoint': current_endpoint,
'other_endpoints': other_endpoints
})
block.append_op(
type='c_comm_init',
inputs={'X': nccl_id_var},
outputs={},
attrs={
'nranks': nranks,
'rank': rank,
'ring_id': self.global_ring_id
})
def run_trainer(self, args): def run_trainer(self, args):
train_prog = fluid.Program() train_prog = fluid.Program()
startup_prog = fluid.Program() startup_prog = fluid.Program()
...@@ -69,12 +100,23 @@ class TestCollectiveAPIRunnerBase(object): ...@@ -69,12 +100,23 @@ class TestCollectiveAPIRunnerBase(object):
current_endpoint = args["currentendpoint"] current_endpoint = args["currentendpoint"]
nranks = 2 nranks = 2
result = self.get_model(train_prog, startup_prog, rank) result = self.get_model(train_prog, startup_prog, rank)
paddle.distributed.init_parallel_env()
if args['backend'] == 'nccl': if args['backend'] == 'nccl':
self.initCommunicator(startup_prog, rank, nranks, True,
current_endpoint, endpoints)
device_id = int(os.getenv("FLAGS_selected_gpus", "0")) device_id = int(os.getenv("FLAGS_selected_gpus", "0"))
place = fluid.CUDAPlace( place = fluid.CUDAPlace(
device_id) #if args.use_gpu else fluid.CPUPlace() device_id) #if args.use_gpu else fluid.CPUPlace()
else: else:
strategy = fluid.core.GlooParallelStrategy()
strategy.rank = rank
strategy.rank_num = nranks
strategy.prefix = ""
strategy.iface = "lo"
strategy.init_seconds = 999999
strategy.run_seconds = 999999
strategy.path = "/tmp/tmp%d" % args['path_id']
gloo = fluid.core.GlooParallelContext(strategy)
gloo.init()
place = fluid.CPUPlace() place = fluid.CPUPlace()
exe = fluid.Executor(place) exe = fluid.Executor(place)
exe.run(startup_prog) exe.run(startup_prog)
...@@ -157,8 +199,8 @@ class TestDistBase(unittest.TestCase): ...@@ -157,8 +199,8 @@ class TestDistBase(unittest.TestCase):
tr_cmd = "%s %s" tr_cmd = "%s %s"
tr0_cmd = tr_cmd % (self._python_interp, model_file) tr0_cmd = tr_cmd % (self._python_interp, model_file)
tr1_cmd = tr_cmd % (self._python_interp, model_file) tr1_cmd = tr_cmd % (self._python_interp, model_file)
tr0_pipe = open("/tmp/tr0_err.log", "w") tr0_pipe = open("/tmp/tr0_err.log", "wb")
tr1_pipe = open("/tmp/tr1_err.log", "w") tr1_pipe = open("/tmp/tr1_err.log", "wb")
#print(tr0_cmd) #print(tr0_cmd)
tr0_proc = subprocess.Popen( tr0_proc = subprocess.Popen(
tr0_cmd.strip().split(), tr0_cmd.strip().split(),
...@@ -179,10 +221,6 @@ class TestDistBase(unittest.TestCase): ...@@ -179,10 +221,6 @@ class TestDistBase(unittest.TestCase):
# close trainer file # close trainer file
tr0_pipe.close() tr0_pipe.close()
tr1_pipe.close() tr1_pipe.close()
with open("/tmp/tr0_err.log", "r") as f:
sys.stderr.write('trainer 0 stderr file: %s\n' % f.read())
with open("/tmp/tr1_err.log", "r") as f:
sys.stderr.write('trainer 1 stderr file: %s\n' % f.read())
return pickle.loads(tr0_out), pickle.loads( return pickle.loads(tr0_out), pickle.loads(
tr1_out), tr0_proc.pid, tr1_proc.pid tr1_out), tr0_proc.pid, tr1_proc.pid
...@@ -209,7 +247,6 @@ class TestDistBase(unittest.TestCase): ...@@ -209,7 +247,6 @@ class TestDistBase(unittest.TestCase):
if check_error_log: if check_error_log:
required_envs["GLOG_v"] = "3" required_envs["GLOG_v"] = "3"
required_envs["GLOG_logtostderr"] = "1" required_envs["GLOG_logtostderr"] = "1"
required_envs["GLOO_LOG_LEVEL"] = "TRACE"
tr0_out, tr1_out, pid0, pid1 = self._run_cluster(model_file, tr0_out, tr1_out, pid0, pid1 = self._run_cluster(model_file,
required_envs) required_envs)
np.random.seed(pid0) np.random.seed(pid0)
......
...@@ -406,7 +406,7 @@ class TestParallelDyGraphRunnerBase(object): ...@@ -406,7 +406,7 @@ class TestParallelDyGraphRunnerBase(object):
fluid.default_main_program().random_seed = seed fluid.default_main_program().random_seed = seed
np.random.seed(seed) np.random.seed(seed)
import random import random
random.seed(seed) random.seed = seed
model, train_reader, opt = self.get_model() model, train_reader, opt = self.get_model()
nranks = len(args.endpoints.split(",")) if args.endpoints else 1 nranks = len(args.endpoints.split(",")) if args.endpoints else 1
...@@ -456,7 +456,7 @@ class TestParallelDyGraphRunnerBase(object): ...@@ -456,7 +456,7 @@ class TestParallelDyGraphRunnerBase(object):
paddle.static.default_startup_program().random_seed = seed paddle.static.default_startup_program().random_seed = seed
paddle.static.default_main_program().random_seed = seed paddle.static.default_main_program().random_seed = seed
np.random.seed(seed) np.random.seed(seed)
random.seed(seed) random.seed = seed
# get trainer id # get trainer id
args.trainer_id = paddle.distributed.get_rank() args.trainer_id = paddle.distributed.get_rank()
...@@ -499,7 +499,7 @@ class TestParallelDyGraphRunnerBase(object): ...@@ -499,7 +499,7 @@ class TestParallelDyGraphRunnerBase(object):
paddle.static.default_startup_program().random_seed = seed paddle.static.default_startup_program().random_seed = seed
paddle.static.default_main_program().random_seed = seed paddle.static.default_main_program().random_seed = seed
np.random.seed(seed) np.random.seed(seed)
random.seed(seed) random.seed = seed
# get trainer id # get trainer id
args.trainer_id = paddle.distributed.get_rank() args.trainer_id = paddle.distributed.get_rank()
......
...@@ -182,7 +182,7 @@ class TestCloudRoleMaker(unittest.TestCase): ...@@ -182,7 +182,7 @@ class TestCloudRoleMaker(unittest.TestCase):
h.log_message("666") h.log_message("666")
s.get_deleted_size("haha") s.get_deleted_size("haha")
s1 = TmpS() s1 = TmpS()
s1.should_stop() s1.shoud_stop()
if __name__ == "__main__": if __name__ == "__main__":
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册
反馈
建议
客服 返回
顶部