/* Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #include "paddle/fluid/distributed/service/communicator.h" #include #include "paddle/fluid/distributed/table/table.h" #include #include #include #include // NOLINT #include #include // NOLINT #include #include "paddle/fluid/framework/eigen.h" #include "paddle/fluid/framework/selected_rows.h" #include "paddle/fluid/framework/tensor_util.h" #include "paddle/fluid/framework/threadpool.h" #include "paddle/fluid/framework/variable_helper.h" #include "paddle/fluid/platform/profiler.h" #include "paddle/fluid/string/printf.h" #include "paddle/fluid/string/split.h" namespace paddle { namespace distributed { using framework::LoDTensor; using framework::SelectedRows; inline double GetCurrentUS() { struct timeval time; gettimeofday(&time, NULL); return 1e+6 * time.tv_sec + time.tv_usec; } Communicator::Communicator() {} void Communicator::init_gflag(const std::string &gflags) { VLOG(0) << "Init With Gflags:" << gflags; std::vector flags = paddle::string::split_string(gflags); if (flags.size() < 1) { flags.push_back("-max_body_size=314217728"); flags.push_back("-bthread_concurrency=40"); flags.push_back("-socket_max_unwritten_bytes=2048000000"); flags.push_back("-max_connection_pool_size=1950"); } auto it = flags.begin(); flags.insert(it, "exe default"); char *flags_ptr[flags.size()]; for (size_t i = 0; i < flags.size(); ++i) { flags_ptr[i] = (char *)(flags[i].c_str()); } int params_cnt = flags.size(); char **params_ptr = &(flags_ptr[0]); ::google::ParseCommandLineFlags(¶ms_cnt, ¶ms_ptr, true); } std::once_flag Communicator::init_flag_; std::shared_ptr Communicator::communicator_(nullptr); void Communicator::InitBrpcClient( const std::string &dist_desc, const std::vector &host_sign_list) { // not used, just for psclient's init std::map> _dense_pull_regions; for (auto &iter : recv_varname_to_ctx_) { auto tid = iter.first; auto var_names = iter.second; auto ®ions = _dense_pull_regions[tid]; regions.reserve(var_names.size()); for (auto &t : var_names) { Variable *var = recv_scope_->FindVar(t); LoDTensor *tensor = var->GetMutable(); float *w = tensor->data(); paddle::distributed::Region reg(w, tensor->numel()); regions.emplace_back(std::move(reg)); } } if (_worker_ptr.get() == nullptr) { google::protobuf::TextFormat::ParseFromString(dist_desc, &_ps_param); init_gflag(_ps_param.init_gflags()); servers_ = host_sign_list.size(); _ps_env = paddle::distributed::PaddlePSEnvironment(); _ps_env.set_ps_servers(&host_sign_list, servers_); _worker_ptr = std::shared_ptr( paddle::distributed::PSClientFactory::create(_ps_param)); _worker_ptr->configure(_ps_param, _dense_pull_regions, _ps_env, trainer_id_); } return; } void Communicator::RpcRecvDense(const std::vector &varnames, int table_id, Scope *scope) { platform::RecordEvent record_event("Communicator->RpcRecvDense"); std::vector regions; regions.reserve(varnames.size()); for (auto &t : varnames) { Variable *var = scope->Var(t); LoDTensor *tensor = var->GetMutable(); if (platform::is_gpu_place(tensor->place())) { #ifdef PADDLE_WITH_CUDA Variable *temp_var = xpu_temp_scope_->Var(t); LoDTensor *temp_tensor = temp_var->GetMutable(); temp_tensor->Resize(tensor->dims()); float *temp_data = temp_tensor->mutable_data(platform::CPUPlace()); paddle::distributed::Region reg(temp_data, tensor->numel()); regions.emplace_back(std::move(reg)); VLOG(1) << "AsyncCommunicator::RpcRecvDense Var " << t << " table_id " << table_id << " Temp_data[0] " << temp_data[0] << " Temp_data[-1] " << temp_data[tensor->numel() - 1]; #endif } else { float *w = tensor->mutable_data(tensor->place()); paddle::distributed::Region reg(w, tensor->numel()); regions.emplace_back(std::move(reg)); } } auto status = _worker_ptr->pull_dense(regions.data(), regions.size(), table_id); status.wait(); for (auto &t : varnames) { Variable *var = scope->FindVar(t); LoDTensor *tensor = var->GetMutable(); VLOG(1) << "AsyncCommunicator::RecvNoBarrier Var " << t << " On gpu? " << platform::is_gpu_place(tensor->place()); if (platform::is_gpu_place(tensor->place())) { #ifdef PADDLE_WITH_CUDA LoDTensor *temp_tensor = xpu_temp_scope_->FindVar(t)->GetMutable(); framework::TensorCopy(*temp_tensor, tensor->place(), tensor); float *temp_data = temp_tensor->mutable_data(platform::CPUPlace()); VLOG(1) << "AsyncCommunicator::RpcRecvDense Var " << t << " table_id " << table_id << " Temp_data[0] " << temp_data[0] << " Temp_data[-1] " << temp_data[tensor->numel() - 1]; #endif } } return; } void Communicator::RpcSendDenseParam(const std::vector &varnames, int table_id, const Scope &scope) { platform::RecordEvent record_event("Communicator->RpcSendDenseParam"); auto place = platform::CPUPlace(); std::vector regions; for (auto &t : varnames) { Variable *var = scope.FindVar(t); CHECK(var != nullptr) << "var[" << t << "] not found"; LoDTensor *tensor = var->GetMutable(); if (platform::is_gpu_place(tensor->place())) { #ifdef PADDLE_WITH_CUDA Variable *temp_var = xpu_temp_scope_->Var(t); LoDTensor *temp_tensor = temp_var->GetMutable(); temp_tensor->Resize(tensor->dims()); float *temp_data = temp_tensor->mutable_data(platform::CPUPlace()); framework::TensorCopy(*tensor, platform::CPUPlace(), temp_tensor); paddle::distributed::Region reg(temp_data, tensor->numel()); regions.emplace_back(std::move(reg)); VLOG(1) << "AsyncCommunicator::RpcSendDenseParam Var " << t << " table_id " << table_id << " Temp_data[0] " << temp_data[0] << " Temp_data[-1] " << temp_data[tensor->numel() - 1]; #endif } else { float *w = tensor->mutable_data(place); paddle::distributed::Region reg(w, tensor->numel()); regions.emplace_back(std::move(reg)); VLOG(1) << "AsyncCommunicator::RpcSendDenseParam Var " << t << " talbe_id " << table_id << " Temp_data[0] " << w[0] << " Temp_data[-1] " << w[tensor->numel() - 1]; } } auto status = _worker_ptr->push_dense_param(regions.data(), regions.size(), table_id); status.wait(); VLOG(4) << "RPC Send Dense Param " << table_id << " done!"; return; } void Communicator::RpcSendDense(const CommContext &ctx, const Scope &scope) { platform::RecordEvent record_event("Communicator->RpcSendDense"); auto &var_names = ctx.origin_varnames; auto &table_id = ctx.table_id; auto dense_data = std::make_shared>(); size_t request_call_num = _worker_ptr->get_server_nums(); uint32_t num_per_shard = dense_dim_per_shard(ctx.height_sections[0], request_call_num); dense_data->resize(num_per_shard * request_call_num); // accessor->update_dim() = 1 float *data = dense_data->data(); uint32_t pos = 0; for (size_t i = 0; i < var_names.size(); ++i) { const LoDTensor tensor = scope.FindVar(var_names[i])->Get(); size_t count = static_cast(tensor.numel()); const float *g = tensor.data(); CHECK(pos + count <= dense_data->size()) << "invalid dense size, cur pos[" << pos << "]" << " data_num[" << count << "] size[" << dense_data->size() << "]"; memcpy(data + pos, g, count * sizeof(float)); pos += count; } ++_async_call_num; DownpourBrpcClosure *closure = new DownpourBrpcClosure( request_call_num, [this, request_call_num](void *done) { int ret = 0; auto *closure = (DownpourBrpcClosure *)done; for (size_t i = 0; i < request_call_num; ++i) { if (closure->check_response(i, PS_PUSH_DENSE_TABLE) != 0) { ret = -1; break; } } closure->set_promise_value(ret); --_async_call_num; }); auto status = _worker_ptr->push_dense_raw_gradient( table_id, data, dense_data->size(), closure); status.wait(); return; } void Communicator::RpcSendSparseParam(const std::string &varname, int table_id, const Scope &scope) { platform::RecordEvent record_event("Communicator->RpcSendSparseParam"); size_t request_call_num = _worker_ptr->get_server_nums(); std::vector push_g_vec; auto *send_var = scope.FindVar(varname); auto *tensor = send_var->GetMutable(); auto dim = tensor->dims()[1]; uint64_t sparse_num = static_cast(tensor->dims()[0]); std::vector sparse_push_keys(sparse_num); std::iota(sparse_push_keys.begin(), sparse_push_keys.end(), 0); push_g_vec.reserve(sparse_num); for (auto i = 0; i < static_cast(sparse_push_keys.size()); ++i) { push_g_vec.push_back(tensor->data() + i * dim); } DownpourBrpcClosure *closure = new DownpourBrpcClosure( request_call_num, [this, request_call_num](void *done) { int ret = 0; auto *closure = (DownpourBrpcClosure *)done; for (size_t i = 0; i < request_call_num; ++i) { if (closure->check_response(i, PS_PUSH_SPARSE_PARAM) != 0) { ret = -1; break; } } closure->set_promise_value(ret); }); auto status = _worker_ptr->push_sparse_param( table_id, sparse_push_keys.data(), (const float **)push_g_vec.data(), sparse_push_keys.size(), closure); status.wait(); return; } void Communicator::RpcSendSparse(const std::string &var_name, int table_id, const Scope &scope) { platform::RecordEvent record_event("Communicator->RpcSendSparse"); size_t request_call_num = _worker_ptr->get_server_nums(); std::vector sparse_push_keys; std::vector push_g_vec; auto *send_var = scope.FindVar(var_name); auto *tensor = send_var->GetMutable(); auto dim = tensor->value().dims()[1]; std::transform(tensor->rows().begin(), tensor->rows().end(), std::back_inserter(sparse_push_keys), [&](int id) { return static_cast(id); }); for (auto i = 0; i < static_cast(sparse_push_keys.size()); ++i) { push_g_vec.push_back(tensor->mutable_value()->data() + i * dim); } ++_async_call_num; DownpourBrpcClosure *closure = new DownpourBrpcClosure( request_call_num, [this, request_call_num](void *done) { int ret = 0; auto *closure = (DownpourBrpcClosure *)done; for (size_t i = 0; i < request_call_num; ++i) { if (closure->check_response(i, PS_PUSH_SPARSE_TABLE) != 0) { ret = -1; break; } } closure->set_promise_value(ret); --_async_call_num; }); auto status = _worker_ptr->push_sparse_raw_gradient( table_id, sparse_push_keys.data(), (const float **)push_g_vec.data(), sparse_push_keys.size(), closure); status.wait(); return; } void Communicator::RpcRecvSparse(const std::string &varname, int table_id, Scope *scope) { platform::RecordEvent record_event("Communicator->RpcRecvSparse"); auto *send_var = scope->Var(varname); auto *tensor = send_var->GetMutable(); auto dim = tensor->dims()[1]; uint64_t sparse_num = static_cast(tensor->dims()[0]); std::vector sparse_push_keys(sparse_num); std::iota(sparse_push_keys.begin(), sparse_push_keys.end(), 0); std::vector push_g_vec; for (auto i = 0; i < static_cast(sparse_push_keys.size()); ++i) { push_g_vec.push_back(tensor->data() + i * dim); } auto status = _worker_ptr->pull_sparse((float **)push_g_vec.data(), table_id, sparse_push_keys.data(), sparse_push_keys.size()); status.wait(); return; } void Communicator::InitParams(const RecvCtxMap &recv_varname_to_ctx) { if (trainer_id_ == 0) { for (auto &iter : recv_varname_to_ctx) { auto &table_id = iter.first; auto &varnames = iter.second; RpcSendDenseParam(varnames, table_id, *recv_scope_); VLOG(1) << "push dense param to table " << table_id << " from 0' trainer done"; } BarrierWithTable(1); } else { BarrierWithTable(1); for (auto &iter : recv_varname_to_ctx) { auto &table_id = iter.first; auto &varnames = iter.second; RpcRecvDense(varnames, table_id, recv_scope_); VLOG(1) << "pull dense param to table " << table_id << " from 0' trainer done"; } } BarrierWithTable(1); return; } void Communicator::RpcProfilerControl() { if (trainer_id_ == 0) { if (!do_server_profiler_ && platform::IsProfileEnabled()) { // send profiler start flag do_server_profiler_ = true; auto start_status = _worker_ptr->start_profiler(); start_status.wait(); } else if (do_server_profiler_ && !platform::IsProfileEnabled()) { // send profiler end flag auto stop_status = _worker_ptr->stop_profiler(); stop_status.wait(); do_server_profiler_ = false; } } } void AsyncCommunicator::RecvThread() { if (!independent_recv_) return; VLOG(3) << "Independent RecvThread Start and Wait"; while (running_) { int grad_num = grad_num_.load(); if (grad_num > min_send_grad_num_before_recv_) { RecvByCommunicator(); grad_num_.store(0); } else { std::this_thread::sleep_for(std::chrono::milliseconds(10)); } } VLOG(1) << "communicator stopped, independent recv thread exit"; } void AsyncCommunicator::RecvByCommunicator() { if (!running_) return; RecvNoBarrier(); VLOG(3) << "run recv graph end"; } void AsyncCommunicator::RecvNoBarrier() { for (auto &iter : recv_varname_to_ctx_) { auto &table_id = iter.first; auto &varnames = iter.second; RpcRecvDense(varnames, table_id, recv_scope_); } for (auto &iter : recv_varname_to_ctx_) { auto var_names = iter.second; for (auto &t : var_names) { Variable *var = recv_scope_->FindVar(t); LoDTensor *tensor = var->GetMutable(); VLOG(1) << "AsyncCommunicator::RecvNoBarrier Var " << t << " On gpu? " << platform::is_gpu_place(tensor->place()); if (platform::is_gpu_place(tensor->place())) { #ifdef PADDLE_WITH_CUDA LoDTensor *temp_tensor = xpu_temp_scope_->FindVar(t)->GetMutable(); framework::TensorCopy(*temp_tensor, tensor->place(), tensor); #endif } } } return; } void AsyncCommunicator::SendByCommunicator() { std::vector> tasks; tasks.reserve(send_varname_to_ctx_.size()); for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; auto send_recv_task = [this, &ctx] { auto &varnames = ctx.origin_varnames; auto &table_id = ctx.table_id; size_t var_nums = varnames.size(); auto &check_queue = send_varname_to_queue_[varnames[0]]; std::vector>> vars; vars.resize(var_nums); int merged_var_num = 0; int wait_times = 0; while (merged_var_num < max_merge_var_num_) { if (check_queue->Size() == 0) { VLOG(4) << "wait_times -> " << wait_times; if (wait_times >= send_wait_times_) { break; } std::this_thread::sleep_for(std::chrono::milliseconds(10)); wait_times++; continue; } else { wait_times = 0; for (size_t i = 0; i < var_nums; i++) { auto &var_name = varnames[i]; auto &var_queue = send_varname_to_queue_[var_name]; vars[i].push_back(var_queue->Pop()); } merged_var_num++; } } if (merged_var_num == 0) return; for (size_t i = 0; i < var_nums; i++) { auto &var_name = varnames[i]; MergeVars(var_name, vars[i], send_scope_.get(), 1); } if (ctx.is_sparse) { PADDLE_ENFORCE_EQ( varnames.size(), 1, platform::errors::InvalidArgument( "sparse variables can only be merged by one variables")); RpcSendSparse(varnames[0], table_id, *send_scope_); } else { RpcSendDense(ctx, *send_scope_); if (!independent_recv_ && recv_varname_to_ctx_.find(table_id) != recv_varname_to_ctx_.end()) { auto recv_varnames = recv_varname_to_ctx_.at(table_id); RpcRecvDense(recv_varnames, table_id, recv_scope_); } } if (independent_recv_) { grad_num_.fetch_add(1, std::memory_order_relaxed); } }; tasks.emplace_back(send_threadpool_->enqueue(std::move(send_recv_task))); } for (auto &task : tasks) { task.wait(); } return; } void AsyncCommunicator::MainThread() { VLOG(3) << "AsyncCommunicator MainThread start and wait"; while (waiting_ && running_) { std::this_thread::sleep_for(std::chrono::milliseconds(100)); VLOG(3) << "wait for running"; } while (running_) { SendByCommunicator(); RpcProfilerControl(); } VLOG(1) << "communicator stopped, send thread exit"; } void HalfAsyncCommunicator::MainThread() { VLOG(3) << "HalfAsyncCommunicator MainThread start and wait"; while (waiting_ && running_) { std::this_thread::sleep_for(std::chrono::milliseconds(100)); VLOG(3) << "wait for running"; } while (running_) { SendByCommunicator(); BarrierSend(); RecvByCommunicator(); BarrierRecv(); BarrierWeakUp(); } VLOG(1) << "communicator stopped, send thread exit"; } void AsyncCommunicator::InitImpl(const RpcCtxMap &send_varname_to_ctx, const RecvCtxMap &recv_varname_to_ctx, Scope *recv_scope) { send_varname_to_ctx_ = std::move(send_varname_to_ctx); recv_varname_to_ctx_ = std::move(recv_varname_to_ctx); recv_scope_ = std::move(recv_scope); send_scope_.reset(new Scope()); xpu_temp_scope_.reset(new Scope()); for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; auto &varnames = ctx.origin_varnames; for (auto &var_name : varnames) { send_varname_to_queue_[var_name] = std::make_shared>>( send_queue_size_); } } send_threadpool_.reset(new ::ThreadPool(thread_pool_size_)); } AsyncCommunicator::~AsyncCommunicator() { running_ = false; if (main_thread_) main_thread_->join(); if (recv_thread_) recv_thread_->join(); } void AsyncCommunicator::Start() { VLOG(1) << "Communicator start"; if (!communicator_) { VLOG(0) << "Communicator is not inited, do nothing"; } else { VLOG(1) << "start send thread and recv thread"; waiting_ = true; running_ = true; // flushing_ = false; BarrierTriggerReset(max_merge_var_num_); // start send and recv thread main_thread_.reset( new std::thread(std::bind(&AsyncCommunicator::MainThread, this))); if (independent_recv_) { recv_thread_.reset( new std::thread(std::bind(&AsyncCommunicator::RecvThread, this))); } } } void AsyncCommunicator::Stop() { VLOG(1) << "Communicator stop"; running_ = false; if (!communicator_) { VLOG(0) << "Communicator is not inited, do nothing"; } else { if (recv_thread_) { VLOG(1) << "stop recv thread"; recv_thread_->join(); recv_thread_.reset(nullptr); } if (main_thread_) { VLOG(1) << "stop main thread"; main_thread_->join(); main_thread_.reset(nullptr); } } VLOG(1) << "Communicator stop done"; } bool AsyncCommunicator::Check(const std::vector &var_tables) { PADDLE_ENFORCE_EQ( var_tables.size(), 1, platform::errors::InvalidArgument("var_tables.size() == 1 is permitted")); auto table_name = var_tables[0]; if (send_varname_to_ctx_.find(table_name) == send_varname_to_ctx_.end()) return false; return true; } bool AsyncCommunicator::Check(const int table_id) { for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; if (ctx.table_id == table_id) return true; } return false; } void AsyncCommunicator::Send(const std::vector &var_names, const framework::Scope &scope) { waiting_ = false; for (size_t i = 0; i < var_names.size(); i++) { auto *var = scope.FindVar(var_names[i]); auto tmp_grad_var = std::make_shared(); framework::CopyVariable(*var, tmp_grad_var.get()); send_varname_to_queue_[var_names[i]]->Push(tmp_grad_var); } } void HalfAsyncCommunicator::Clean() { for (auto &iter : send_varname_to_queue_) { auto &var_name = iter.first; auto &var_queue = iter.second; while (var_queue->Size() > 0) { var_queue->Pop(); } VLOG(3) << "clean var: " << var_name << " done"; } } void HalfAsyncCommunicator::BarrierTriggerDecrement() { barrier_trigger_--; VLOG(3) << "BarrierTriggerDecrement decrement barrier trigger to " << barrier_trigger_.load(); } void HalfAsyncCommunicator::BarrierTriggerReset(int initial_val) { barrier_trigger_.store(initial_val); VLOG(3) << "BarrierTriggerReset reset barrier trigger to " << barrier_trigger_.load(); } void HalfAsyncCommunicator::Barrier() { barrier_counter_++; if (!running_) { VLOG(3) << "Communicator is not running, release barrier"; return; } { std::unique_lock lk(barrier_mutex_); barrier_cond_.wait(lk, [this] { return (barrier_counter_ == 0); }); } } int HalfAsyncCommunicator::BatchesCounter() { while (running_) { if (barrier_counter_.load() >= barrier_trigger_.load() && barrier_trigger_.load() != 0) { break; } else { std::this_thread::sleep_for(std::chrono::milliseconds(10)); } } return barrier_counter_.load(); } void HalfAsyncCommunicator::SendByCommunicator() { int batches = BatchesCounter(); VLOG(1) << "HalfAsyncCommunicator::BatchesCounter = " << batches; if (batches <= 0) return; std::vector> tasks; tasks.reserve(send_varname_to_ctx_.size()); for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; auto send_recv_task = [this, &ctx, batches] { auto &varnames = ctx.origin_varnames; auto &table_id = ctx.table_id; size_t var_nums = varnames.size(); std::vector>> vars; vars.resize(var_nums); for (size_t i = 0; i < var_nums; i++) { auto &var_name = varnames[i]; auto &var_queue = send_varname_to_queue_[var_name]; for (int j = 0; j < batches; j++) vars[i].push_back(var_queue->Pop()); MergeVars(var_name, vars[i], send_scope_.get(), 1); } if (ctx.is_sparse) { PADDLE_ENFORCE_EQ( varnames.size(), 1, platform::errors::InvalidArgument( "sparse variables can only be merged by one variables")); RpcSendSparse(varnames[0], table_id, *send_scope_); } else { RpcSendDense(ctx, *send_scope_); } }; tasks.emplace_back(send_threadpool_->enqueue(std::move(send_recv_task))); } for (auto &task : tasks) { task.wait(); } return; } void HalfAsyncCommunicator::BarrierWeakUp() { barrier_counter_.store(0); barrier_cond_.notify_all(); } void SyncCommunicator::BarrierSend() { if (!running_) return; BarrierWithTable(0); VLOG(4) << "BarrierSend with SyncCommunicator"; } void SyncCommunicator::BarrierRecv() { if (!running_) return; BarrierWithTable(1); VLOG(4) << "BarrierRecv with SyncCommunicator"; } void GeoCommunicator::Send(const std::vector &var_names, const framework::Scope &scope) { waiting_ = false; auto before_send = GetCurrentUS(); auto table_name = var_names[0]; size_t splited_var_nums = send_varname_to_ctx_[table_name].splited_varnames.size(); std::unordered_map> ids_table; for (size_t j = 0; j < splited_var_nums; j++) { ids_table.insert(std::pair>( send_varname_to_ctx_[table_name].splited_varnames[j], std::unordered_set())); } auto *var = scope.FindVar(table_name); PADDLE_ENFORCE_EQ(var->IsType(), true, platform::errors::InvalidArgument( "Only need to send Sparse Grad in Geo mode.")); auto &rows = var->Get().rows(); // insert ids which has not been record for (size_t j = 0; j < rows.size(); j++) { auto ep_idx = rows[j] % splited_var_nums; ids_table.at(send_varname_to_ctx_[table_name].splited_varnames[ep_idx]) .insert(rows[j]); } for (auto &iter : ids_table) { auto &key = iter.first; auto &sparse_ids_set = iter.second; auto sparse_ids_vec = std::make_shared>(); sparse_ids_vec->assign(sparse_ids_set.begin(), sparse_ids_set.end()); sparse_id_queues_.at(key)->Push(sparse_ids_vec); VLOG(3) << "push " << sparse_ids_vec->size() << " ids to " << key << "'s queue"; } auto after_send = GetCurrentUS(); VLOG(2) << "run send op finish. use time " << (after_send - before_send); } void GeoCommunicator::InitImpl(const RpcCtxMap &send_varname_to_ctx, const RecvCtxMap &recv_varname_to_ctx, Scope *recv_scope) { send_varname_to_ctx_ = std::move(send_varname_to_ctx); recv_varname_to_ctx_ = std::move(recv_varname_to_ctx); recv_scope_ = std::move(recv_scope); PADDLE_ENFORCE_GT( send_varname_to_ctx.size(), 0, platform::errors::InvalidArgument("send var contexts can not be zero")); for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; if (!ctx.is_sparse) continue; auto &varnames = ctx.origin_varnames; PADDLE_ENFORCE_EQ( varnames.size(), 1, platform::errors::InvalidArgument( "sparse variables can only be merged by one variables")); for (auto &splited_var : ctx.splited_varnames) { parallel_task_nums_ += 1; sparse_id_queues_.insert( std::pair>>>>( splited_var, std::make_shared< BlockingQueue>>>( send_queue_size_))); } } send_threadpool_.reset(new ::ThreadPool(thread_pool_size_)); delta_scope_.reset(new Scope()); old_scope_.reset(new Scope()); pserver_scope_.reset(new Scope()); } void GeoCommunicator::InitParams(const RecvCtxMap &recv_varname_to_ctx) { std::vector> tasks; tasks.reserve(recv_varname_to_ctx_.size()); for (auto &iter : recv_varname_to_ctx_) { auto &table_id = iter.first; auto &varnames = iter.second; auto recv_task = [this, &table_id, &varnames] { InitDense(varnames, table_id); }; tasks.emplace_back(send_threadpool_->enqueue(std::move(recv_task))); } for (auto &task : tasks) { task.wait(); } for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; if (!ctx.is_sparse) return; auto &varname = ctx.origin_varnames[0]; auto &table_id = ctx.table_id; auto param = varname.substr(0, varname.size() - 5); InitSparse(param, table_id); } return; } void GeoCommunicator::InitDense(std::vector &varnames, int table_id) { if (trainer_id_ == 0) { RpcSendDenseParam(varnames, table_id, *recv_scope_); BarrierWithTable(1); VLOG(0) << "push dense param to table " << table_id << " from 0' trainer done"; } else { BarrierWithTable(1); RpcRecvDense(varnames, table_id, recv_scope_); VLOG(0) << "push dense param to table " << table_id << " from 0' trainer done"; } // copy to old_scope for (auto &t : varnames) { auto *global_var = recv_scope_->FindVar(t); global_var->GetMutable(); auto *old_var = old_scope_->Var(t); old_var->GetMutable(); framework::CopyVariable(*global_var, old_var); } VLOG(1) << "init dense table " << table_id << " done"; } void GeoCommunicator::SendDense(const CommContext &send_ctx) { platform::RecordEvent record_event("GeoCommunicator->SendDense"); auto &var_names = send_ctx.origin_varnames; auto &table_id = send_ctx.table_id; for (auto &varname : var_names) { auto param_name = GradToParam(varname); auto *var_latest = recv_scope_->FindVar(param_name); auto *var_timestamp = old_scope_->FindVar(param_name); PADDLE_ENFORCE_EQ(var_latest->IsInitialized(), true, platform::errors::Unavailable( "%s is not initialized, please check", param_name)); PADDLE_ENFORCE_EQ(var_timestamp->IsInitialized(), true, platform::errors::Unavailable( "%s is not initialized, please check", param_name)); auto &t_latest = var_latest->Get(); auto t_timestamp = var_timestamp->GetMutable(); auto cpu_ctx = paddle::platform::CPUDeviceContext(); auto *var_delta = delta_scope_->Var(varname); auto *t_delta = var_delta->GetMutable(); t_delta->mutable_data(t_latest.dims(), cpu_ctx.GetPlace()); auto blas = paddle::operators::math::GetBlas( cpu_ctx); blas.VSUB(t_latest.numel(), t_latest.data(), t_timestamp->data(), t_delta->data()); float coefficient = 1.0 / static_cast(trainers_); blas.SCAL(t_latest.numel(), coefficient, t_delta->data()); blas.VADD(t_latest.numel(), t_timestamp->data(), t_delta->data(), t_timestamp->data()); } RpcSendDense(send_ctx, *delta_scope_); VLOG(1) << "Finish Send Dense " << var_names[0] << ", table_id: " << table_id; return; } void GeoCommunicator::RecvDense(const CommContext &send_ctx) { platform::RecordEvent record_event("GeoCommunicator->RecvDense"); auto &table_id = send_ctx.table_id; auto &varnames = recv_varname_to_ctx_.at(table_id); // 1. recv from pserver RpcRecvDense(varnames, table_id, pserver_scope_.get()); // 2.1 pserver - old => delta; 2.2 latest + old => latest 2.3 old => pserver auto cpu_ctx = paddle::platform::CPUDeviceContext(); for (auto &varname : varnames) { auto *var_latest = recv_scope_->FindVar(varname); auto t_latest = var_latest->GetMutable(); auto *var_old = old_scope_->FindVar(varname); auto t_old = var_old->GetMutable(); auto *var_pserver = pserver_scope_->FindVar(varname); auto t_pserver = var_pserver->Get(); auto *var_delta = delta_scope_->Var(varname); auto *t_delta = var_delta->GetMutable(); t_delta->mutable_data(t_latest->dims(), cpu_ctx.GetPlace()); auto blas = paddle::operators::math::GetBlas( cpu_ctx); blas.VSUB(t_latest->numel(), t_pserver.data(), t_old->data(), t_delta->data()); blas.VADD(t_latest->numel(), t_latest->data(), t_delta->data(), t_latest->data()); blas.VCOPY(t_latest->numel(), t_pserver.data(), t_old->data()); } VLOG(1) << "Finish Recv Dense " << varnames[0] << ", table_id: " << table_id; return; } void GeoCommunicator::InitSparse(const std::string &var_name, int table_id) { VLOG(0) << "Init Sparse " << var_name << " : table " << table_id << " begin."; if (trainer_id_ == 0) { RpcSendSparseParam(var_name, table_id, *recv_scope_); BarrierWithTable(1); VLOG(0) << "push sparse param to table " << table_id << " from 0' trainer done"; } else { BarrierWithTable(1); RpcRecvSparse(var_name, table_id, recv_scope_); VLOG(0) << "push dense param to table " << table_id << " from 0' trainer done"; } VLOG(0) << "Init Sparse " << var_name << " : table " << table_id << " done."; auto *global_var = recv_scope_->FindVar(var_name); auto *var = old_scope_->Var(var_name); framework::CopyVariable(*global_var, var); return; } std::vector GeoCommunicator::MergeSparseIds( const std::string &send_varname) { size_t merge_num = 0, wait_times = 0; std::unordered_set sparse_ids; while (merge_num < static_cast(max_merge_var_num_)) { VLOG(3) << "Merge Number of " << send_varname << " = " << merge_num; if (sparse_id_queues_.at(send_varname)->Size() > 0) { wait_times = 0; std::shared_ptr> pop_ids = sparse_id_queues_.at(send_varname)->Pop(); for (size_t j = 0; j < pop_ids->size(); j++) { sparse_ids.insert(pop_ids->at(j)); } merge_num += 1; VLOG(3) << "sparse_id_queues_(" << send_varname << ") pushed"; } else if (sparse_id_queues_.at(send_varname)->Size() == 0) { VLOG(3) << "wait_times -> " << wait_times; if (wait_times >= static_cast(send_wait_times_)) { break; } std::this_thread::sleep_for(std::chrono::milliseconds(10)); wait_times++; continue; } } std::vector res; res.assign(sparse_ids.begin(), sparse_ids.end()); return res; } void GeoCommunicator::SendSparse(const std::string &varname, std::vector &sparse_ids, int table_id, int ep_idx) { platform::RecordEvent record_event("GeoCommunicator->SendSparse"); std::string param_name = SplitedGradToParam(varname); VLOG(1) << "In GeoCommunicator::SendSparse(" << varname << " " << param_name << ", ids.size = " << sparse_ids.size() << ", table_id: " << table_id << ", ep_idx: " << ep_idx; auto *var_latest = recv_scope_->FindVar(param_name); auto *var_old = old_scope_->FindVar(param_name); PADDLE_ENFORCE_EQ(var_latest->IsInitialized(), true, platform::errors::Unavailable( "%s is not initialized, please check", param_name)); PADDLE_ENFORCE_EQ(var_old->IsInitialized(), true, platform::errors::Unavailable( "%s is not initialized, please check", param_name)); auto &t_latest = var_latest->Get(); auto *t_old = var_old->GetMutable(); auto dims1 = t_latest.dims()[1]; auto cpu_ctx = paddle::platform::CPUDeviceContext(); auto *var_delta = delta_scope_->Var(varname); auto *t_delta = var_delta->GetMutable(); auto *var_t_value = t_delta->mutable_value(); var_t_value->Resize({static_cast(sparse_ids.size()), dims1}); auto *t_value = var_t_value->mutable_data(cpu_ctx.GetPlace()); t_delta->set_rows(sparse_ids); t_delta->set_height(t_latest.dims()[0]); auto blas = paddle::operators::math::GetBlas( cpu_ctx); float coefficient = 1.0 / static_cast(trainers_); std::vector push_g_vec; for (auto j = 0; j < static_cast(sparse_ids.size()); ++j) { blas.VSUB(dims1, t_latest.data() + sparse_ids[j] * dims1, t_old->data() + sparse_ids[j] * dims1, t_value + j * dims1); blas.SCAL(dims1, coefficient, t_value + j * dims1); blas.VADD(dims1, t_old->data() + sparse_ids[j] * dims1, t_value + j * dims1, t_old->data() + sparse_ids[j] * dims1); push_g_vec.push_back(t_value + j * dims1); } ++_async_call_num; DownpourBrpcClosure *closure = new DownpourBrpcClosure(1, [this](void *done) { int ret = 0; auto *closure = (DownpourBrpcClosure *)done; if (closure->check_response(0, PS_PUSH_SPARSE_TABLE) != 0) { ret = -1; } closure->set_promise_value(ret); --_async_call_num; }); auto status = _worker_ptr->push_sparse_raw_gradient_partial( table_id, (const uint64_t *)sparse_ids.data(), (const float **)push_g_vec.data(), sparse_ids.size(), closure, ep_idx); status.wait(); VLOG(1) << "Finish Send Sparse " << varname << ", ids.size = " << sparse_ids.size() << ", table_id: " << table_id; return; } void GeoCommunicator::RecvSparse(const std::string &varname, int table_id, int ep_idx) { platform::RecordEvent record_event("GeoCommunicator->RecvSparse"); // 1. recv from pserver std::vector keys; std::vector values; auto status = _worker_ptr->pull_geo_param(table_id, &values, &keys, ep_idx); status.wait(); std::string param = SplitedGradToParam(varname); VLOG(1) << "RecvSparse receive var: " << varname << " " << param << ", " << table_id << "; ids Size: " << keys.size() << "; values size: " << values.size(); auto *var_latest = recv_scope_->FindVar(param); auto *var_old = old_scope_->FindVar(param); auto *t_latest = var_latest->GetMutable(); auto *t_old = var_old->GetMutable(); auto dims1 = t_latest->dims()[1]; auto numel = keys.size() * dims1; std::vector v_delta; v_delta.resize(numel); auto cpu_ctx = paddle::platform::CPUDeviceContext(); auto blas = paddle::operators::math::GetBlas( cpu_ctx); for (auto j = 0; j < static_cast(keys.size()); ++j) { float *latest_data = t_latest->data() + keys[j] * dims1; float *old_data = t_old->data() + keys[j] * dims1; // pserver - old => delta blas.VSUB(dims1, values.data() + j * dims1, old_data, v_delta.data() + j * dims1); // latest + delta => latest blas.VADD(dims1, latest_data, v_delta.data() + j * dims1, latest_data); // pserver => old blas.VCOPY(dims1, values.data() + j * dims1, old_data); } VLOG(1) << "Finish Recv Sparse " << param << ", table_id: " << table_id; } void GeoCommunicator::MainThread() { VLOG(3) << "MainThread start and wait"; while (waiting_ && running_) { std::this_thread::sleep_for(std::chrono::milliseconds(100)); VLOG(3) << "wait for running"; } while (running_) { std::vector> tasks; tasks.reserve(parallel_task_nums_); for (auto &iter : send_varname_to_ctx_) { auto &ctx = iter.second; auto &varnames = ctx.origin_varnames; auto &table_id = ctx.table_id; if (ctx.is_sparse) { PADDLE_ENFORCE_EQ( varnames.size(), 1, platform::errors::InvalidArgument( "sparse variables can only be merged by one variables")); int pserver_num = static_cast(ctx.epmap.size()); for (int ep_idx = 0; ep_idx < pserver_num; ep_idx++) { // varname: emb@GRAD, param_name: emb, splited_varname: emb.delta0 auto send_recv_task = [this, table_id, ep_idx, &ctx] { auto splited_varname = ctx.splited_varnames[ep_idx]; auto sparse_ids = MergeSparseIds(splited_varname); SendSparse(splited_varname, sparse_ids, table_id, ep_idx); RecvSparse(splited_varname, table_id, ep_idx); }; tasks.emplace_back( send_threadpool_->enqueue(std::move(send_recv_task))); } } else { auto send_recv_task = [this, &ctx] { SendDense(ctx); RecvDense(ctx); }; tasks.emplace_back( send_threadpool_->enqueue(std::move(send_recv_task))); } } for (auto &task : tasks) { task.wait(); } } } } // namespace distributed } // namespace paddle