stream_analyzer.cc 6.4 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#include "paddle/fluid/framework/new_executor/stream_analyzer.h"
16

17 18 19 20 21 22 23 24 25
#include <unordered_set>

namespace paddle {
namespace framework {

/*
 * Parse the var_ids that need to be associated with an event.
 * The caller should guarantee front_op and back_op satisfy the
 * following conditions:
26
 *   1. kQueueSync -> kQueueAsync
27 28 29 30
 *   2. kQueueAsync -> kQueueSync
 *
 * For example: matmul(gpu) -> out_var -> memcpy_d2h
 * out_var should be associated with an event.
L
Leo Chen 已提交
31 32 33 34
 *
 * NOTE(zhiqiu): There are two special case that no event is needed:
 *  1. the variable is marked as NoDataTransformVar
 *  2. the variable is marked as NoNeedDataBuffer
35
 */
L
Leo Chen 已提交
36
std::vector<size_t> StreamAnalyzer::GetNeedEventVarIds(
37 38
    const Instruction& cur_instr, const Instruction& next_instr) {
  std::unordered_set<size_t> unique_var_ids;
39
  for (auto& item : cur_instr.Outputs()) {
40 41 42
    unique_var_ids.insert(item.second.begin(), item.second.end());
  }

L
Leo Chen 已提交
43 44 45 46 47 48 49 50 51 52 53 54
  auto is_no_need_buffer = [&next_instr](std::string name) {
    auto* op = next_instr.OpBase();
    auto& inferer = op->Info().NoNeedBufferVarsInferer();
    if (inferer) {
      auto no_need_buffer_ins =
          inferer(op->Inputs(), op->Outputs(), op->Attrs());
      return no_need_buffer_ins.count(name) != 0;
    }
    return false;
  };

  std::vector<size_t> need_event_var_ids;
55
  for (auto& item : next_instr.Inputs()) {
56
    for (auto var_id : item.second) {
L
Leo Chen 已提交
57 58 59 60 61 62 63 64 65 66 67 68 69 70 71
      if (unique_var_ids.count(var_id) > 0) {
        if (next_instr.NoDataTransformVars().count(var_id)) {
          VLOG(4) << "Skip inserting event at variable " << item.first
                  << " of operator " << next_instr.OpBase()->Type()
                  << " since it is NoDataTransform";
          continue;
        }
        if (is_no_need_buffer(item.first)) {
          VLOG(4) << "Skip inserting event at variable " << item.first
                  << " of operator " << next_instr.OpBase()->Type()
                  << " since it is NoNeedBufferVar";
          continue;
        }

        need_event_var_ids.push_back(var_id);
72 73 74
      }
    }
  }
L
Leo Chen 已提交
75
  return need_event_var_ids;
76 77
}

L
Leo Chen 已提交
78
void StreamAnalyzer::ConstructEventForVar(
79
    const std::vector<size_t>& new_event_var_id, Instruction* next_instr,
L
Leo Chen 已提交
80
    platform::DeviceType waiter_type, const platform::Place& place) {
81 82 83
  for (auto var_id : new_event_var_id) {
    if (var_id2event_.count(var_id) == 0) {
      auto device_event = std::make_shared<platform::DeviceEvent>(
L
Leo Chen 已提交
84
          place, platform::GenerateDeviceEventFlag());
85 86 87
      var_id2event_.emplace(var_id, std::move(device_event));
    }
    // Add events for next_instr.inputs
88
    next_instr->AddInputEvent(var_id, var_id2event_.at(var_id), waiter_type);
89 90 91
  }
}

92 93 94
void StreamAnalyzer::Schedule(const std::vector<size_t>& downstream_ops,
                              std::vector<Instruction>* instructions,
                              size_t op_index) {
95
  auto& cur_instr = instructions->at(op_index);
96
  auto& next_instruction = cur_instr.NextInstructions();
97 98 99 100
  std::vector<size_t> event_var_ids;
  for (auto next_op_id : downstream_ops) {
    auto& next_instr = instructions->at(next_op_id);
    if (IsDirectRun(cur_instr, next_instr)) {
L
Leo Chen 已提交
101 102
      VLOG(4) << "DirectRun: " << cur_instr.OpBase()->Type() << "->"
              << next_instr.OpBase()->Type();
103
      next_instruction.AddDirectRun(next_op_id);
104
    } else {
105
      // Always insert events between different stream
L
Leo Chen 已提交
106 107 108
      auto need_event_var_ids = GetNeedEventVarIds(cur_instr, next_instr);
      event_var_ids.insert(event_var_ids.end(), need_event_var_ids.begin(),
                           need_event_var_ids.end());
109

110
      auto waiter_type = GetWaiterType(next_instr);
L
Leo Chen 已提交
111 112
      ConstructEventForVar(need_event_var_ids, &next_instr, waiter_type,
                           cur_instr.DeviceContext().GetPlace());
113

114
      if (waiter_type == platform::kCPU) {  // GPU -> CPU
L
Leo Chen 已提交
115 116
        VLOG(4) << "SyncRun: " << cur_instr.OpBase()->Type() << "->"
                << next_instr.OpBase()->Type();
117
        next_instruction.AddSyncRun(next_op_id);
118
      } else {  // GPU -> GPU(different stream)
L
Leo Chen 已提交
119 120
        VLOG(4) << "EventRun: " << cur_instr.OpBase()->Type() << "->"
                << next_instr.OpBase()->Type();
121
        next_instruction.ADDEventRun(next_op_id);
122 123
      }
    }
124 125
  }
  // Create events for these cross-stream vars
126
  VLOG(3) << cur_instr.OpBase()->Type()
127 128
          << " event_var_ids.size: " << event_var_ids.size();
  for (auto var_id : event_var_ids) {
129 130
    cur_instr.AddOutputEvent(var_id, var_id2event_.at(var_id),
                             platform::kCUDA /*not used*/);
131 132 133 134
  }
}

platform::DeviceContext* StreamAnalyzer::ParseDeviceContext(
135 136
    const OpFuncNode& op_func_node) {
  auto& op_type = op_func_node.operator_base_->Type();
137
  auto* dev_ctx = op_func_node.dev_ctx_;
A
Aurelius84 已提交
138
  if (op_type == interpreter::kMemcpyD2H) {
139 140
    VLOG(3) << "Get dev_ctx from d2h_context_pool_";
    dev_ctx = d2h_ctx_pool_.Get(place_);
A
Aurelius84 已提交
141
  } else if (op_type == interpreter::kMemcpyH2D) {
142 143 144 145 146 147 148
    VLOG(3) << "Get dev_ctx from h2d_context_pool_";
    dev_ctx = h2d_ctx_pool_.Get(place_);
  }

  return dev_ctx;
}

149 150 151 152
/*
 * NOTE(dev): The following cases are considered as directly run:
 *
 *  1. with same dev_ctx_, such as: CPU -> CPU, GPU -> GPU
L
Leo Chen 已提交
153 154 155 156
 *  2. CPU -> any (it is possible: CPU op->VAR->GPU op, when var is no need
 * buffer or no need data transform)
 *  3. D2H -> CPU
 *  4. CPU -> H2D
157 158 159
 */
bool StreamAnalyzer::IsDirectRun(Instruction& cur_instr,
                                 const Instruction& next_instr) {
160
  return (&cur_instr.DeviceContext() == &next_instr.DeviceContext() ||
L
Leo Chen 已提交
161
          interpreter::IsCpuOp(cur_instr) ||
162 163
          interpreter::IsMemcpyD2H(cur_instr) ||
          interpreter::IsMemcpyH2D(next_instr));
164 165 166
}

platform::DeviceType StreamAnalyzer::GetWaiterType(const Instruction& instr) {
167
  if (instr.KernelType() == OpFuncType::kQueueSync) {
168 169 170 171 172 173
    return platform::kCPU;
  } else {
    return platform::kCUDA;
  }
}

174 175
}  // namespace framework
}  // namespace paddle