// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. /* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #include "paddle/fluid/framework/fleet/heter_wrapper.h" #if defined(PADDLE_WITH_PSLIB) && !defined(PADDLE_WITH_HETERPS) #include "paddle/fluid/framework/device_worker.h" namespace paddle { namespace framework { std::shared_ptr HeterWrapper::s_instance_ = NULL; bool HeterWrapper::is_initialized_ = false; void HeterWrapper::CreateClient2XpuConnection() { brpc::ChannelOptions options; options.protocol = "baidu_std"; options.connection_type = "single"; options.timeout_ms = 2000000; xpu_channels_.resize(xpu_list_.size()); for (size_t i = 0; i < xpu_list_.size(); ++i) { VLOG(3) << "channel init: " << xpu_list_[i]; xpu_channels_[i].reset(new brpc::Channel()); if (xpu_channels_[i]->Init(xpu_list_[i].c_str(), "", &options) != 0) { VLOG(0) << "server channel init fail"; } } } void HeterWrapper::RegisterServiceHandler(int cmd, HeterServiceHandler func) { service_.RegisterServiceHandler(cmd, func); } void HeterWrapper::SetXpuList(const std::vector& xpu_list) { #ifdef PADDLE_WITH_PSLIB VLOG(3) << "Going to set xpu list"; for (auto& x : xpu_list) { xpu_list_.push_back(x); VLOG(3) << "set xpu list: " << x << " size: " << xpu_list_.size(); } #endif } void HeterWrapper::StartXpuService(const std::string& ip, uint32_t port) { std::string ip_port = ip + ":" + std::to_string(port); VLOG(3) << "xpu server starts at " << ip_port; server_.AddService(&service_, brpc::SERVER_DOESNT_OWN_SERVICE); brpc::ServerOptions options; if (server_.Start(ip_port.c_str(), &options) != 0) { VLOG(0) << "xpu server start fail"; } } // void HeterWrapper::SerializeToReq(const std::string& varname, // Scope* scope, HeterRequest& request) { // auto* req_var = request.mutable_vars(); void HeterWrapper::SerializeToReq(const std::string& varname, Scope* scope, VariableMessage* req_var) { Variable* var = scope->FindVar(varname); if (var == nullptr) { return; } LoDTensor* tensor = var->GetMutable(); req_var->set_varname(varname); req_var->set_type(LOD_TENSOR); req_var->set_data_type(static_cast(tensor->type())); for (auto& dim : framework::vectorize(tensor->dims())) { req_var->add_dims(dim); } const framework::LoD lod = tensor->lod(); if (lod.size() > 0) { req_var->set_lod_level(lod.size()); for (auto& each : lod) { VariableMessage::LodData* lod_inner = req_var->add_lod(); for (auto& d : each) { lod_inner->add_lod_data(d); } } } auto* req_data = req_var->mutable_data(); req_data->clear(); req_data->resize(tensor->numel() * SizeOfType(tensor->type())); char* data_ptr = const_cast(req_data->data()); if (platform::is_cpu_place(tensor->place())) { memcpy(data_ptr, tensor->data(), tensor->numel() * SizeOfType(tensor->type())); } else { #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) memory::Copy(platform::CPUPlace(), data_ptr, tensor->place(), tensor->data(), tensor->numel() * SizeOfType(tensor->type()), nullptr); #endif #ifdef PADDLE_WITH_XPU memory::Copy(platform::CPUPlace(), data_ptr, tensor->place(), tensor->data(), tensor->numel() * SizeOfType(tensor->type())); #endif } } #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) void HeterWrapper::DeSerializeToTensor(Scope* scope, const VariableMessage& req_var, platform::Place place, gpuStream_t stream) { // const VariableMessage& req_var = request->vars(); auto* var = scope->FindVar(req_var.varname()); auto* tensor = var->GetMutable(); std::vector vec_dim; for (auto& x : req_var.dims()) { vec_dim.push_back(x); } tensor->Resize(make_ddim(vec_dim)); LoD lod; for (int i = 0; i < req_var.lod_level(); ++i) { framework::Vector v; for (int j = 0; j < req_var.lod(i).lod_data_size(); ++j) { v.push_back(req_var.lod(i).lod_data(j)); } lod.push_back(v); } tensor->set_lod(lod); void* tensor_data = tensor->mutable_data(place, ToVarType(req_var.data_type())); #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) memory::Copy(place, tensor_data, platform::CPUPlace(), req_var.data().data(), tensor->numel() * SizeOfType(tensor->type()), stream); #else memcpy(tensor_data, req_var.data().data(), tensor->numel() * SizeOfType(tensor->type())); #endif } #endif // void HeterWrapper::DeSerializeToTensor(Scope* scope, // const HeterRequest* request) { void HeterWrapper::DeSerializeToTensor(Scope* scope, const VariableMessage& req_var, platform::Place place) { // const VariableMessage& req_var = request->vars(); auto* var = scope->FindVar(req_var.varname()); auto* tensor = var->GetMutable(); std::vector vec_dim; for (auto& x : req_var.dims()) { vec_dim.push_back(x); } tensor->Resize(make_ddim(vec_dim)); LoD lod; for (int i = 0; i < req_var.lod_level(); ++i) { framework::Vector v; for (int j = 0; j < req_var.lod(i).lod_data_size(); ++j) { v.push_back(req_var.lod(i).lod_data(j)); } lod.push_back(v); } tensor->set_lod(lod); void* tensor_data = tensor->mutable_data(place, ToVarType(req_var.data_type())); #ifdef PADDLE_WITH_XPU memory::Copy(place, tensor_data, platform::CPUPlace(), req_var.data().data(), tensor->numel() * SizeOfType(tensor->type())); #else memcpy(tensor_data, req_var.data().data(), tensor->numel() * SizeOfType(tensor->type())); #endif } framework::proto::VarType::Type HeterWrapper::ToVarType( VariableMessage::Type type) { switch (type) { case VariableMessage::FP32: return framework::proto::VarType::FP32; // NOLINT case VariableMessage::FP64: return framework::proto::VarType::FP64; // NOLINT case VariableMessage::INT32: return framework::proto::VarType::INT32; // NOLINT case VariableMessage::INT64: return framework::proto::VarType::INT64; // NOLINT case VariableMessage::BOOL: return framework::proto::VarType::BOOL; // NOLINT default: PADDLE_THROW(platform::errors::InvalidArgument( "ToVarType:Unsupported type %d", type)); } } void HeterWrapper::StopXpuService(int num) { HeterRequest request; HeterResponse response; brpc::Controller cntl; request.set_cmd(2); // for (size_t i = 0; i < xpu_channels_.size(); ++i) { HeterService_Stub stub(xpu_channels_[num].get()); stub.service(&cntl, &request, &response, NULL); if (cntl.Failed()) { VLOG(0) << "call stop xpu service fail: " << cntl.ErrorText(); } else { VLOG(3) << "call stop xpu service success"; } // } } void HeterWrapper::EndPass(Scope* scope, int num) { HeterRequest request; HeterResponse response; brpc::Controller cntl; request.set_cmd(1); // for (size_t i = 0; i < xpu_channels_.size(); ++i) { HeterService_Stub stub(xpu_channels_[num].get()); stub.service(&cntl, &request, &response, NULL); if (cntl.Failed()) { VLOG(0) << "call end pass fail: " << cntl.ErrorText(); } else { VLOG(3) << "call end pass success"; for (int j = 0; j < response.vars_size(); ++j) { DeSerializeToTensor(scope, response.vars(j), platform::CPUPlace()); } } // } } void HeterWrapper::CallRemoteXpu(std::shared_ptr task, HeterCpuWorker* worker, int mpi_rank, std::vector& send_vars) { HeterRequest request; request.set_cmd(0); request.set_cur_batch(task->cur_batch_); OnHeterRpcDone* done = new OnHeterRpcDone([this, task, worker](void* done) { auto* closure = reinterpret_cast(done); if (closure->cntl.Failed()) { VLOG(0) << "call xpu fail: " << closure->cntl.ErrorText(); } else { VLOG(3) << "call xpu success"; } // DeSerializeToTensor(task->scope_, // closure->response.vars(), platform::CPUPlace()); for (int i = 0; i < closure->response.vars_size(); ++i) { DeSerializeToTensor(task->scope_, closure->response.vars(i), platform::CPUPlace()); } worker->Schedule(task->taskid_); }); // std::vector varnames = {"click", "12345"}; // //varnames.push_back(send_var); // //if (send_var == "_generated_var_412") { // varnames.push_back("filter_by_instag_0.tmp_0"); // varnames.push_back("filter_by_instag_2.tmp_0"); // varnames.push_back("filter_by_instag_0.tmp_1"); // varnames.push_back("concat_1.tmp_0"); // } for (auto& varname : send_vars) { auto* req_var = request.add_vars(); SerializeToReq(varname, task->scope_, req_var); } int num = mpi_rank % xpu_channels_.size(); HeterService_Stub stub(xpu_channels_[num].get()); // stub.service(&cntl, &request, &response, // brpc::NewCallback(&HeterWrapper::RpcCallBack, // response, cntl, worker, task)); stub.service(&done->cntl, &request, &done->response, done); } void HeterWrapper::CallRemoteXpuSync(std::shared_ptr task, HeterCpuWorker* worker, int mpi_rank, std::vector& send_vars) { HeterRequest request; HeterResponse response; brpc::Controller cntl; request.set_cmd(0); request.set_cur_batch(task->cur_batch_); // std::vector varnames = {"concat_1.tmp_0", "click", "12345"}; for (auto& varname : send_vars) { auto* req_var = request.add_vars(); SerializeToReq(varname, task->scope_, req_var); } HeterService_Stub stub(xpu_channels_[0].get()); stub.service(&cntl, &request, &response, NULL); if (cntl.Failed()) { VLOG(0) << "call xpu fail: " << cntl.ErrorText(); } else { VLOG(3) << "call xpu success"; for (int i = 0; i < response.vars_size(); ++i) { DeSerializeToTensor(task->scope_, response.vars(i), platform::CPUPlace()); } } } } // end namespace framework } // end namespace paddle #endif