// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include #include #include #include #include "lite/core/kernel.h" #include "lite/core/op_lite.h" namespace paddle { namespace lite { namespace mir { // Node in a MIR graph. class Node { public: std::list inlinks; std::list outlinks; Node() = default; enum class Role { kArg = 0, kStmt, kNumRoles, /*should be last*/ kUnk, }; class Stmt { // The kernel instances this Statement contains. std::vector> valid_kernels_; // TODO(Superjomn) make this a shared_ptr for resource safety. std::shared_ptr op_; // we hold op to run InferShape public: // Refresh the operator and kernels with the latest OpInfo. void ResetOp(const cpp::OpDesc& op_desc, const std::vector& valid_places, lite::Scope* scope = nullptr); void ResetKernels(const std::vector& valid_places); std::string op_type() const { return op_info()->Type(); } const OpInfo* op_info() const; OpInfo* mutable_op_info(); void SetKernels(std::vector>&& kernels) { valid_kernels_ = std::move(kernels); } std::vector>& kernels() { return valid_kernels_; } void SetOp(const std::shared_ptr& op) { op_ = op; } const std::shared_ptr op() const { return op_; } Place place() const; KernelBase& picked_kernel(); friend std::ostream& operator<<(std::ostream& os, const Stmt& other) { os << "Statement " << other.op_type() << " " << other.place().DebugString(); return os; } // Description. std::string desc; // for cuda multi stream bool need_sync_{false}; int stream_id_{0}; // streams which need to be sync. exclude stream_id_ std::vector sync_streams_{}; }; struct Arg { std::string name; int id{0}; const Type* type{nullptr}; // Weight is a special kind of argument, it is marked as weight explicitly // so that some weight related optimization can take place. bool is_weight{false}; // is_persist indicate that whether is the argument trans from Weight // if the need more than one tool operator(eg. io_copy layout calib), the // argument between them should be persist to make sure it's only run once bool is_persist{false}; int lane{-1}; }; Arg& AsArg(const std::string& name, int id); Arg& AsArg(const std::string& name); Stmt& AsStmt(const std::string& op_type, std::vector>&& kernels, const std::shared_ptr& op) { auto& x = AsStmt(); x.SetOp(op); x.SetKernels(std::move(kernels)); return x; } Stmt* stmt() const { CHECK(IsStmt()); return stmt_.get(); } Arg* arg() const { CHECK(IsArg()); return arg_.get(); } // Set roles. Arg& AsArg() { if (role_ != Role::kUnk) { CHECK(role_ == Role::kArg); return *arg_; } role_ = Role::kArg; arg_.reset(new Arg); return *arg_; } Stmt& AsStmt() { if (role_ != Role::kUnk) { CHECK(role_ == Role::kStmt); return *stmt_; } role_ = Role::kStmt; stmt_.reset(new Stmt); return *stmt_; } friend std::ostream& operator<<(std::ostream& os, Node& other) { os << static_cast(other.role_) << " "; if (!other.IsRoleSet()) { os << "Unk role node"; } if (other.IsArg()) { auto& arg = other.AsArg(); os << "Argument " << arg.name; } if (other.IsStmt()) { auto& arg = other.AsStmt(); os << "Statement " << arg.op_type(); } return os; } // Check roles. bool IsRoleSet() const { return role_ != Role::kUnk; } bool IsStmt() const { return role_ == Role::kStmt; } bool IsArg() const { return role_ == Role::kArg; } private: // Either stmt_ or argument_ is used. std::unique_ptr stmt_; std::unique_ptr arg_; Role role_{Role::kUnk}; }; } // namespace mir } // namespace lite } // namespace paddle