// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/fluid/framework/details/op_handle_base.h" #include #include namespace paddle { namespace framework { namespace details { std::string OpHandleBase::DebugString() const { std::stringstream ss; ss << Name() << "("; for (auto *var : inputs_) { ss << var->DebugString() << ", "; } ss << ") --> ("; for (auto *var : outputs_) { ss << var->DebugString() << ", "; } ss << ")\n"; return ss.str(); } OpHandleBase::~OpHandleBase() { #ifdef PADDLE_WITH_CUDA for (auto &ev : events_) { PADDLE_ENFORCE(cudaEventDestroy(ev.second)); } #endif } void OpHandleBase::Run(bool use_cuda) { #ifdef PADDLE_WITH_CUDA if (events_.empty() && use_cuda && dev_ctxes_.size() > 0) { for (auto &p : dev_ctxes_) { int dev_id = boost::get(p.first).device; PADDLE_ENFORCE(cudaSetDevice(dev_id)); PADDLE_ENFORCE( cudaEventCreateWithFlags(&events_[dev_id], cudaEventDisableTiming)); } if (IsMultiDeviceTransfer() && dev_ctxes_.size() > 0) { for (auto &out_var : outputs_) { auto *out_var_handle = dynamic_cast(out_var); if (out_var_handle) { int dev_id = boost::get(out_var_handle->place()).device; out_var_handle->SetGenerateEvent(events_.at(dev_id)); } } } else { PADDLE_ENFORCE_EQ(dev_ctxes_.size(), 1UL, "%s should have only one dev_ctx.", Name()); auto &place = dev_ctxes_.begin()->first; int dev_id = boost::get(place).device; for (auto &out_var : outputs_) { auto *out_var_handle = dynamic_cast(out_var); if (out_var_handle) { PADDLE_ENFORCE( platform::is_same_place(place, out_var_handle->place()), "The place of output(%s) is not consistent with the " "place of current op(%s).", out_var_handle->Name(), Name()); out_var_handle->SetGenerateEvent(events_.at(dev_id)); } } } } #else PADDLE_ENFORCE(!use_cuda); #endif RunImpl(); } void OpHandleBase::RecordWaitEventOnCtx(platform::DeviceContext *waited_ctx) { #ifdef PADDLE_WITH_CUDA PADDLE_ENFORCE_NOT_NULL(waited_ctx); if (platform::is_cpu_place(waited_ctx->GetPlace()) || events_.empty()) { for (auto &dev_ctx : dev_ctxes_) { PADDLE_ENFORCE_NOT_NULL(dev_ctx.second); dev_ctx.second->Wait(); } } else { auto stream = static_cast(waited_ctx)->stream(); for (auto &ev : events_) { PADDLE_ENFORCE(cudaStreamWaitEvent(stream, ev.second, 0)); } } #else for (auto &dev_ctx : dev_ctxes_) { dev_ctx.second->Wait(); } #endif } void OpHandleBase::AddInput(VarHandleBase *in) { this->inputs_.emplace_back(in); node_->inputs.push_back(in->Node()); in->AddOutput(this, this->Node()); } void OpHandleBase::AddOutput(VarHandleBase *out) { outputs_.emplace_back(out); node_->outputs.push_back(out->Node()); out->AddInput(this, this->Node()); } void OpHandleBase::WaitInputVarGenerated() { for (auto in_var : inputs_) { if (NeedWait(in_var)) { // Dummy Variable is used to represent dependencies between operators, so // there doesn't add event for it. auto *in_var_handle = dynamic_cast(in_var); if (in_var_handle) { auto &place = in_var_handle->place(); if (platform::is_gpu_place(place)) { #ifdef PADDLE_WITH_CUDA auto stream = static_cast(dev_ctxes_.at(place)) ->stream(); PADDLE_ENFORCE( cudaStreamWaitEvent(stream, in_var_handle->GetEvent(), 0)); #else PADDLE_THROW("Doesn't compile the GPU."); #endif } // There are nothing to do when the place is CPUPlace. } } } } void OpHandleBase::WaitInputVarGenerated(const platform::Place &place) { for (auto in_var : inputs_) { if (NeedWait(in_var)) { // Dummy Variable is used to represent dependencies between operators, so // there doesn't add event for it. auto *in_var_handle = dynamic_cast(in_var); if (in_var_handle) { if (platform::is_gpu_place(in_var_handle->place())) { #ifdef PADDLE_WITH_CUDA auto stream = static_cast( dev_ctxes_.at(in_var_handle->place())) ->stream(); PADDLE_ENFORCE( cudaStreamWaitEvent(stream, in_var_handle->GetEvent(), 0)); #else PADDLE_THROW("Doesn't compile the GPU."); #endif } // There are nothing to do when the place is CPUPlace. } } } } size_t OpHandleBase::NoDummyInputSize() const { size_t cnt = 0; for (auto *in : inputs_) { if (dynamic_cast(in) == nullptr) { ++cnt; } } return cnt; } bool OpHandleBase::NeedWait(VarHandleBase *in_var) { return in_var && in_var->GeneratedOp(); } void OpHandleBase::RunAndRecordEvent(const std::function &callback) { #ifdef PADDLE_WITH_CUDA if (!events_.empty()) { // Use event std::function method = callback; for (auto &p : dev_ctxes_) { method = [method, p, this]() { VLOG(10) << "cudadevicecontext:" << static_cast(p.second) << ", dev_id:" << boost::get(p.first).device; static_cast(p.second)->RecordEvent( events_.at(boost::get(p.first).device), method); }; } method(); } else { #endif callback(); #ifdef PADDLE_WITH_CUDA } #endif } void OpHandleBase::RunAndRecordEvent(platform::Place p, const std::function &callback) { #ifdef PADDLE_WITH_CUDA if (platform::is_cpu_place(p) || events_.empty()) { callback(); } else { auto *ctx = dev_ctxes_.at(p); auto *cuda_ctx = static_cast(ctx); cuda_ctx->RecordEvent(events_.at(boost::get(p).device), callback); } #else callback(); #endif } size_t OpHandleBase::NotReadyInputSize() const { std::unordered_set res; for (auto *var : inputs_) { if (var->GeneratedOp() != nullptr) { res.emplace(var); } } return res.size(); } } // namespace details } // namespace framework } // namespace paddle