tracer.h 7.8 KB
Newer Older
J
Jiabin Yang 已提交
1
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#pragma once

J
Jiabin Yang 已提交
17 18
#include <atomic>
#include <future>  // NOLINT
19
#include <map>
J
Jiabin Yang 已提交
20
#include <memory>
21
#include <string>
22
#include <unordered_map>
23
#include <vector>
24

J
Jiabin Yang 已提交
25
#include "ThreadPool.h"
26
#include "paddle/fluid/framework/garbage_collector.h"
L
Leo Chen 已提交
27
#include "paddle/fluid/imperative/amp_auto_cast.h"
28
#include "paddle/fluid/imperative/basic_engine.h"
29
#include "paddle/fluid/imperative/jit/program_desc_tracer.h"
30
#include "paddle/fluid/imperative/layer.h"
31
#include "paddle/fluid/imperative/layout_autotune.h"
J
Jiabin Yang 已提交
32
#include "paddle/fluid/platform/macros.h"
33
#include "paddle/phi/core/compat/arg_map_context.h"
34 35 36
namespace paddle {
namespace imperative {

L
Leo Chen 已提交
37 38
enum class AmpLevel;

39 40
enum class AmpDtype;

41 42 43 44
using GarbageCollectorMap =
    std::map<platform::Place,
             std::unique_ptr<paddle::framework::GarbageCollector>>;

45 46 47
class UniqueNameGenerator {
 public:
  explicit UniqueNameGenerator(std::string prefix = "") : prefix_(prefix) {}
48
  std::string Generate(std::string key = "dygraph_tmp") {
L
Leo Chen 已提交
49
    return prefix_ + key + "_" + std::to_string(id_++);
50 51 52 53 54 55 56
  }

 private:
  std::atomic<int> id_{0};
  std::string prefix_;
};

57
class Tracer {
J
Jiabin Yang 已提交
58 59
  DISABLE_COPY_AND_ASSIGN(Tracer);

60
 public:
61
  Tracer()
62
      : basic_engine_(new BasicEngine()),
63
        program_desc_tracer_(new jit::ProgramDescTracer()),
64 65 66
        generator_(new UniqueNameGenerator()) {
    expected_place_ = platform::CPUPlace();
  }
67

J
Jiabin Yang 已提交
68
  ~Tracer() = default;
69

J
Jiabin Yang 已提交
70
  template <typename VarType>
71 72 73 74 75 76
  void TraceOp(const std::string& type,
               const NameVarMap<VarType>& ins,
               const NameVarMap<VarType>& outs,
               framework::AttributeMap attrs,
               const platform::Place& place,
               bool trace_backward,
J
Jiabin Yang 已提交
77 78
               const std::map<std::string, std::string>& inplace_map = {},
               paddle::framework::AttributeMap* passed_default_attrs_ = nullptr,
79
               bool use_default_attr_map = true);
J
Jiabin Yang 已提交
80

W
wanghuancoder 已提交
81 82
  template <typename VarType>
  void TraceOpImpl(
83 84
      const std::string& type,
      const NameVarMap<VarType>& ins,
W
wanghuancoder 已提交
85 86
      const NameVarMap<VarType>& outs,
      framework::AttributeMap& attrs,  // NOLINT
87 88
      const platform::Place& place,
      bool trace_backward,
W
wanghuancoder 已提交
89 90 91 92
      const std::map<std::string, std::string>& inplace_map = {},
      paddle::framework::AttributeMap* passed_default_attrs_ = nullptr,
      bool use_default_attr_map = true);

93 94 95 96
  void TraceOp(const std::string& type,
               const NameVarBaseMap& ins,
               const NameVarBaseMap& outs,
               framework::AttributeMap attrs,
97
               const std::map<std::string, std::string>& inplace_map = {});
J
Jiabin Yang 已提交
98

99 100
  void TraceOp(const std::string& type,
               const NameTensorMap& ins,
W
wanghuancoder 已提交
101 102
               const NameTensorMap& outs,
               paddle::framework::AttributeMap& attrs,  // NOLINT
J
Jiabin Yang 已提交
103 104
               const std::map<std::string, std::string>& inplace_map = {});

105 106
  void TraceOp(const std::string& type,
               const NameTensorMap& ins,
W
wanghuancoder 已提交
107 108 109
               const NameTensorMap& outs,
               paddle::framework::AttributeMap attrs);

110 111
  void TraceOp(const std::string& type,
               const NameTensorMap& ins,
W
wanghuancoder 已提交
112 113
               const NameTensorMap& outs,
               paddle::framework::AttributeMap& attrs,  // NOLINT
J
Jiabin Yang 已提交
114 115
               const paddle::platform::Place& place,
               paddle::framework::AttributeMap* default_attrs,
116
               bool use_default_attr_map,
117
               const std::map<std::string, std::string>& inplace_map = {});
118

119
  bool ComputeRequiredGrad(const NameVarBaseMap& ins,
120 121 122 123
                           const NameVarBaseMap& outs,
                           bool trace_backward);
  bool ComputeRequiredGrad(const NameTensorMap& ins,
                           const NameTensorMap& outs,
J
Jiabin Yang 已提交
124
                           bool trace_backward);
J
Jiabin Yang 已提交
125

126 127 128 129 130 131 132 133 134 135 136 137
  void SetEnableProgramDescTracing(bool enabled) {
    enable_program_desc_tracing_ = enabled;
  }

  bool IsProgramDescTracingEnabled() const {
    return enable_program_desc_tracing_;
  }

  jit::ProgramDescTracer* GetProgramDescTracer() {
    return program_desc_tracer_.get();
  }

138
  // Note(Aurelius84): The `tmp` is used as prefix key while naming a temporary
139
  // intermediate var both in imperative and static graph mode. But the
140 141
  // `UniqueNameGenerator` in C++ and `unique_name.py` in Python doesn't share
  // the same auto-increment id. It will create a variable repeatedly with same
142
  // name like `tmp_0` in some cases when transform dygraph into static layers.
143 144
  // So we modify the default prefix key into `eager_tmp` to distinguish with
  // static graph.
145
  std::string GenerateUniqueName(std::string key = "dygraph_tmp") {
146 147 148
    return generator_->Generate(key);
  }

149 150
  BasicEngine* GetEngine() const { return basic_engine_.get(); }

151 152
  platform::Place ExpectedPlace() const { return expected_place_; }

W
WangXi 已提交
153
  void SetExpectedPlace(platform::Place place);
154

155
  bool HasGrad() const { return has_grad_; }
156

157
  void SetHasGrad(bool has_grad) { has_grad_ = has_grad; }
158

L
Leo Chen 已提交
159 160 161 162
  void SetAmpLevel(AmpLevel level) {
    VLOG(4) << "set amp_level to " << static_cast<unsigned int>(level);
    amp_level_ = level;
  }
163

L
Leo Chen 已提交
164
  AmpLevel GetAmpLevel() const { return amp_level_; }
165

166 167 168
  void SetAmpDtype(std::string amp_dtype) {
    VLOG(4) << "set amp_dtype to " << amp_dtype;
    if (amp_dtype == "float16") {
169
      amp_dtype_ = phi::DataType::FLOAT16;
170
    } else if (amp_dtype == "bfloat16") {
171
      amp_dtype_ = phi::DataType::BFLOAT16;
172
    } else {
173
      amp_dtype_ = phi::DataType::FLOAT32;
174 175 176 177
    }
  }

  std::string GetAmpDtype() const {
178
    if (amp_dtype_ == phi::DataType::FLOAT16) {
179
      return std::string("float16");
180
    } else if (amp_dtype_ == phi::DataType::BFLOAT16) {
181 182 183 184 185 186
      return std::string("bfloat16");
    } else {
      return std::string("float32");
    }
  }

Z
Zhang Ting 已提交
187 188
  phi::DataType GetAmpPhiDtype() const { return amp_dtype_; }

189 190 191 192 193 194 195 196 197 198 199 200 201 202
  void DisableLayoutAutoTune() { use_layout_autotune_ = false; }

  void EnableLayoutAutoTune() { use_layout_autotune_ = true; }

  bool UseLayoutAutoTune() {
#if defined(PADDLE_WITH_CUDA)
    if (phi::backends::gpu::TensorCoreAvailable()) {
      return use_layout_autotune_;
    }
#endif
    use_layout_autotune_ = false;
    return false;
  }

203
  phi::KernelSignature GetExpectedKernelSignature(
204 205 206 207
      const std::string& type,
      const NameTensorMap& ins,
      const NameTensorMap& outs,
      framework::AttributeMap attrs) const;
208

209 210 211
  paddle::framework::GarbageCollector* MutableGarbageCollectorIfNotExists(
      const platform::Place& place);

212
 private:
213
  std::unique_ptr<BasicEngine> basic_engine_;
214
  std::unique_ptr<jit::ProgramDescTracer> program_desc_tracer_;
215
  std::unique_ptr<UniqueNameGenerator> generator_;
216
  platform::Place expected_place_;
217
  GarbageCollectorMap gcs_;
218
  static thread_local bool enable_program_desc_tracing_;
219
  static thread_local bool use_layout_autotune_;
Z
Zeng Jinle 已提交
220
  static thread_local bool has_grad_;
221
  static thread_local AmpLevel amp_level_;
222
  static thread_local phi::DataType amp_dtype_;
223 224
};

225 226 227
// To access static variable current_tracer
const std::shared_ptr<Tracer>& GetCurrentTracer();
void SetCurrentTracer(const std::shared_ptr<Tracer>& tracer_);
228 229 230
void IncreaseVarbaseReferenceCountUntilCopyComplete(
    const std::shared_ptr<imperative::VarBase>& var,
    const platform::Place& place);
231

232 233
void PassStopGradient(const NameVarBaseMap& outs, bool generate_grad);

234 235
}  // namespace imperative
}  // namespace paddle