// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include #include "paddle/fluid/framework/details/var_handle.h" #include "paddle/fluid/framework/lod_tensor.h" #include "paddle/fluid/framework/selected_rows.h" #include "paddle/fluid/framework/variable.h" #include "paddle/fluid/platform/device_context.h" #include "paddle/fluid/platform/macros.h" namespace paddle { namespace framework { namespace details { // GetTensorFromVar is used in broadcast_op handle and gather_op handle, so it // should be placed in a commonplace. I don't find an appropriate place, so I // temporarily place it in op_handle. Tensor *GetTensorFromVar(Variable *in_var); constexpr char kLocalExecScopeName[] = "@LCOAL_SCOPE@"; class OpHandleBase { private: DISABLE_COPY_AND_ASSIGN(OpHandleBase); public: std::vector inputs_; std::vector outputs_; std::unordered_map dev_ctxes_; #ifdef PADDLE_WITH_CUDA std::unordered_map events_; #endif OpHandleBase() {} std::string DebugString() const; virtual std::string Name() const = 0; virtual ~OpHandleBase(); void Run(bool use_event); virtual void Wait(platform::DeviceContext *waited_dev); void AddInput(VarHandleBase *in); void AddOutput(VarHandleBase *out); // If the Op involves data transfer of multiple devices that // will likely block other computations. virtual bool IsMultiDeviceTransfer() { return false; } protected: virtual void RunImpl() = 0; }; } // namespace details } // namespace framework } // namespace paddle