cinn_launch_op.h 6.9 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#pragma once

#include <memory>
#include <string>
#include <unordered_map>
20
#include <unordered_set>
21 22 23
#include "cinn/hlir/framework/graph_compiler.h"
#include "cinn/hlir/framework/scope.h"
#include "cinn/runtime/cinn_runtime.h"
24
#include "cinn/runtime/flags.h"
25
#include "paddle/fluid/framework/data_type.h"
26 27 28
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/operator.h"
#include "paddle/fluid/framework/paddle2cinn/cinn_compiler.h"
29
#include "paddle/fluid/operators/cinn/cinn_launch_context.h"
30 31 32 33

namespace paddle {
namespace operators {

34 35 36
constexpr char kX[] = "X";
constexpr char kOutputs[] = "Out";
constexpr char kCompilationKey[] = "compilation_key";
37 38

using LoDTensor = framework::LoDTensor;
39 40 41 42 43 44 45 46 47 48 49 50 51
using CinnTensor = ::cinn::hlir::framework::Tensor;
using CinnScope = ::cinn::hlir::framework::Scope;
using CinnCompiler = framework::paddle2cinn::CinnCompiler;
using CinnCompiledObject = framework::paddle2cinn::CinnCompiledObject;

namespace details {

// Tranform Paddle place to CINN target
const ::cinn::common::Target& PlaceToCinnTarget(const platform::Place& place);

// Print detailed compilation result of graph for debug
void DebugCinnCompiledResult(const CinnCompiledObject& result);

52 53
// Launch cinn to execute compiled executable program and wait done
void LaunchCinnExecution(const CinnCompiledObject& compiled_obj,
54
                         const CinnLaunchContext& context, void* stream);
55 56 57

// Set cinn FLAGS (such as FLAGS_cinn_cudnn_deterministic) with paddle's FLAGS.
void SetCinnRuntimeFlags();
58 59 60 61 62 63 64 65 66 67 68 69

template <typename DeviceContext>
void* GetStream(const framework::ExecutionContext& ctx) {
  return nullptr;
}

#ifdef PADDLE_WITH_CUDA
template <>
void* GetStream<platform::CUDADeviceContext>(
    const framework::ExecutionContext& ctx);
#endif

70
}  // namespace details
71 72 73 74 75

template <typename DeviceContext, typename T>
class CinnLaunchOpKernel : public framework::OpKernel<T> {
 public:
  void Compute(const framework::ExecutionContext& ctx) const override {
76 77
    const auto& scope = ctx.scope();
    const auto& place = ctx.GetPlace();
78
    void* stream = details::GetStream<DeviceContext>(ctx);
79 80 81 82 83 84 85
    // Step 1. Find graph object and prepare input
    PADDLE_ENFORCE_EQ(ctx.HasAttr(kCompilationKey), true,
                      platform::errors::NotFound(
                          "No Attribute(%s) found for CinnLaunchOp operator.",
                          kCompilationKey));
    const auto& compilation_key =
        ctx.template Attr<std::string>(kCompilationKey);
86
    VLOG(4) << "CinnLaunchOp attribute(" << kCompilationKey << ") "
87 88
            << "value:\n"
            << CinnCompiler::GetInstance()->ReadableKey(compilation_key);
89 90

    auto input_variable_names = ctx.InputNames(kX);
91 92 93 94 95 96 97 98
    const auto& input_tensors = ctx.MultiInput<LoDTensor>(kX);
    std::map<std::string, const LoDTensor*> inputs_name2tensor;
    std::transform(input_variable_names.begin(), input_variable_names.end(),
                   input_tensors.begin(),
                   std::inserter(inputs_name2tensor, inputs_name2tensor.end()),
                   [](const std::string& name, const LoDTensor* tensor) {
                     return std::make_pair(name, tensor);
                   });
99 100

    // Step 2. Get compilation result of the graph
101
    auto target = details::PlaceToCinnTarget(place);
102
    const auto& cinn_compiled_object = CinnCompiler::GetInstance()->Compile(
103
        compilation_key, inputs_name2tensor, target, stream);
104
    details::DebugCinnCompiledResult(cinn_compiled_object);
105

106
    auto* launch_context = cinn_compiled_object.launch_context.get();
107
    // Step 3. Prepare arguments needed for the compiled executable program.
108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124
    launch_context->UpdateCapturedEnv(scope, place);
    if (!launch_context->IsArgumentsInitialized()) {
      VLOG(4) << "CinnLaunchOp prepare arguments";

      // 3.1 Prepare input variables: tensors of input variables have
      //     been initialized before graph compiled, just check the
      //     equiality between tensors of paddle and cinn.
      for (const auto& var_name : input_variable_names) {
        if (!launch_context->IsVariableUsed(var_name)) {
          // some input variables don't need for cinn because they are
          // eliminated by optimized passes or some cinn operators use
          // less variables
          VLOG(4) << "Input variable(" << var_name << ") not used by cinn";
          continue;
        }

        launch_context->AssignExternalVariable(var_name);
125 126
      }

127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142
      // 3.2 Prepare output variables: all output variables should
      //     be initialized and allocated buffer before
      //     the runtime program start execution, the compilation result
      //     includes details of their buffer assginment and we use that to
      //     allocate space in Paddle. For those variables allocated yet,
      //     like persistable parameters, just check the equiality between
      //     Paddle allocation and CINN buffer assginment.
      auto output_variable_names = ctx.OutputNames(kOutputs);
      for (const auto var_name : output_variable_names) {
        PADDLE_ENFORCE_EQ(
            launch_context->IsVariableUsed(var_name), true,
            platform::errors::InvalidArgument(
                "Output variable(%s) not used by cinn", var_name));

        launch_context->AssignExternalVariable(var_name);
      }
143

144 145 146 147 148 149 150 151 152 153
      // 3.3 Prepare internal or temporary variables: Create a temporary
      //     scope to keep internal variables within graph or temporary
      //     variables needed by the compiled runtime program in addition.
      //     Here we directly use the names from CinnScope as Paddle variable
      //     names, because they will not be used outside the graph
      //     and should be destructed after computation finished.
      auto internal_variable_names = launch_context->GetInternalVariableNames();
      for (const auto& var_name : internal_variable_names) {
        launch_context->AssignInternalVariable(var_name);
      }
154
    }
155

156 157 158 159
    // Step 4. Set CINN runtime FLAGS, such as FLAGS_cinn_cudnn_deterministic.
    details::SetCinnRuntimeFlags();

    // Step 5. Launch CINN to execute the compiled executable program
160 161
    VLOG(4) << "Run Cinn compiled executable program with stream: " << stream;
    details::LaunchCinnExecution(cinn_compiled_object, *launch_context, stream);
162
    VLOG(4) << "CinnLaunchOp launch execution done.";
163 164 165 166 167
  }
};

}  // namespace operators
}  // namespace paddle