profiler.cc 15.3 KB
Newer Older
1
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
D
dangqingqing 已提交
2 3 4 5

licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
6

D
dangqingqing 已提交
7 8 9 10 11 12 13 14
    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

Y
Yi Wang 已提交
15
#include "paddle/fluid/platform/profiler.h"
X
Xin Pan 已提交
16 17
#include <sys/time.h>
#include <time.h>
18
#include <algorithm>
19
#include <iomanip>
20
#include <map>
21 22
#include <mutex>  // NOLINT
#include <string>
23 24 25
#ifdef PADDLE_WITH_CUDA
#include <cuda.h>
#endif  // PADDLE_WITH_CUDA
26
#include "glog/logging.h"
27 28 29
#include "paddle/fluid/framework/block_desc.h"
#include "paddle/fluid/platform/device_tracer.h"
#include "paddle/fluid/string/printf.h"
D
dangqingqing 已提交
30 31 32 33

namespace paddle {
namespace platform {

34 35
struct EventList;

D
dangqingqing 已提交
36 37 38 39 40 41 42 43 44 45 46 47 48 49 50
// The profiler state, the initial value is ProfilerState::kDisabled
static ProfilerState g_state = ProfilerState::kDisabled;
// The thread local event list only can be accessed by the specific thread
// The thread index of each thread
static thread_local int32_t g_thread_id;
// The g_next_thread_id is a global counter for threads, by the g_thread_id and
// g_next_thread_id, we can know how many threads have created EventList.
static uint32_t g_next_thread_id = 0;
// The global mutex
static std::mutex g_all_event_lists_mutex;
// The total event lists of all threads
static std::list<std::shared_ptr<EventList>> g_all_event_lists;
// The thread local event list only can be accessed by the specific thread
static thread_local std::shared_ptr<EventList> g_event_list;

51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83
struct EventList {
  constexpr static size_t kMB = 1024 * 1024;
  constexpr static size_t kEventBlockSize = 16 * kMB;
  constexpr static size_t kEventSize = sizeof(Event);
  constexpr static size_t kEventAlign = alignof(Event);
  constexpr static size_t kNumBlock =
      kEventBlockSize /
      ((kEventSize + kEventAlign - 1) / kEventAlign * kEventAlign);

  template <typename... Args>
  void Record(Args&&... args) {
    if (event_blocks.empty() || event_blocks.front().size() == kNumBlock) {
      event_blocks.emplace_front();
      event_blocks.front().reserve(kNumBlock);
    }
    event_blocks.front().emplace_back(std::forward<Args>(args)...);
  }

  std::vector<Event> Reduce() {
    std::vector<Event> result;
    for (auto& block : event_blocks) {
      result.insert(result.begin(), std::make_move_iterator(block.begin()),
                    std::make_move_iterator(block.end()));
    }
    event_blocks.clear();
    return result;
  }

  void Clear() { event_blocks.clear(); }

  std::forward_list<std::vector<Event>> event_blocks;
};

D
dangqingqing 已提交
84 85 86 87 88 89 90 91 92
inline uint64_t GetTimeInNsec() {
  using clock = std::conditional<std::chrono::high_resolution_clock::is_steady,
                                 std::chrono::high_resolution_clock,
                                 std::chrono::steady_clock>::type;
  return std::chrono::duration_cast<std::chrono::nanoseconds>(
             clock::now().time_since_epoch())
      .count();
}

X
Xin Pan 已提交
93 94 95 96 97 98
inline uint64_t PosixInNsec() {
  struct timeval tv;
  gettimeofday(&tv, nullptr);
  return 1000 * (static_cast<uint64_t>(tv.tv_sec) * 1000000 + tv.tv_usec);
}

99
Event::Event(EventType type, std::string name, uint32_t thread_id,
D
dangqingqing 已提交
100
             const DeviceContext* dev_ctx)
101
    : type_(type), name_(name), thread_id_(thread_id), has_cuda_(false) {
D
dangqingqing 已提交
102
#ifdef PADDLE_WITH_CUDA
D
dangqingqing 已提交
103 104 105
  has_cuda_ = dev_ctx ? platform::is_gpu_place(dev_ctx->GetPlace()) : false;
  if (has_cuda_) {
    auto* cuda_dev_ctx = static_cast<const CUDADeviceContext*>(dev_ctx);
D
dangqingqing 已提交
106 107 108 109 110 111 112 113 114
    PADDLE_ENFORCE(cudaGetDevice(&device_));
    PADDLE_ENFORCE(cudaEventCreate(&event_));
    auto stream = cuda_dev_ctx->stream();
    PADDLE_ENFORCE(cudaEventRecord(event_, stream));
  }
#endif
  cpu_ns_ = GetTimeInNsec();
}

115
const EventType& Event::type() const { return type_; }
D
dangqingqing 已提交
116

117 118
double Event::CpuElapsedMs(const Event& e) const {
  return (e.cpu_ns_ - cpu_ns_) / (1000000.0);
D
dangqingqing 已提交
119 120
}

121
double Event::CudaElapsedMs(const Event& e) const {
D
dangqingqing 已提交
122 123 124 125 126 127 128
#ifdef PADDLE_WITH_CUDA
  PADDLE_ENFORCE(e.has_cuda() && has_cuda());
  PADDLE_ENFORCE(e.device() == device());
  PADDLE_ENFORCE(cudaEventSynchronize(event_));
  PADDLE_ENFORCE(cudaEventSynchronize(e.event()));
  float ms;
  PADDLE_ENFORCE(cudaEventElapsedTime(&ms, event_, e.event()));
129
  return ms;
D
dangqingqing 已提交
130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156
#else
  PADDLE_THROW("CUDA is not enabled");
#endif
}

#ifdef PADDLE_WITH_CUDA
static void ForEachDevice(std::function<void(int)> func) {
  auto original_device = GetCurrentDeviceId();
  int count = GetCUDADeviceCount();
  for (int i = 0; i < count; i++) {
    SetDeviceId(i);
    func(i);
  }
  SetDeviceId(original_device);
}
#endif

inline EventList& GetEventList() {
  if (!g_event_list) {
    std::lock_guard<std::mutex> guard(g_all_event_lists_mutex);
    g_event_list = std::make_shared<EventList>();
    g_thread_id = g_next_thread_id++;
    g_all_event_lists.emplace_front(g_event_list);
  }
  return *g_event_list;
}

D
dangqingqing 已提交
157
void Mark(const std::string& name, const DeviceContext* dev_ctx) {
158
  GetEventList().Record(EventType::kMark, name, g_thread_id, dev_ctx);
159 160
}

D
dangqingqing 已提交
161
void PushEvent(const std::string& name, const DeviceContext* dev_ctx) {
162
  GetEventList().Record(EventType::kPushRange, name, g_thread_id, dev_ctx);
163 164
}

D
dangqingqing 已提交
165
void PopEvent(const std::string& name, const DeviceContext* dev_ctx) {
166
  GetEventList().Record(EventType::kPopRange, name, g_thread_id, dev_ctx);
D
dangqingqing 已提交
167 168
}

X
Xin Pan 已提交
169 170
RecordEvent::RecordEvent(const std::string& name, const DeviceContext* dev_ctx)
    : start_ns_(PosixInNsec()) {
D
dangqingqing 已提交
171 172
  if (g_state == ProfilerState::kDisabled) return;
  dev_ctx_ = dev_ctx;
Y
Yibing Liu 已提交
173
  name_ = name;
174
  PushEvent(name_, dev_ctx_);
175
  // Maybe need the same push/pop behavior.
X
Xin Pan 已提交
176
  SetCurAnnotation(name_);
D
dangqingqing 已提交
177 178 179 180
}

RecordEvent::~RecordEvent() {
  if (g_state == ProfilerState::kDisabled) return;
X
Xin Pan 已提交
181 182
  DeviceTracer* tracer = GetDeviceTracer();
  if (tracer) {
X
Xin Pan 已提交
183 184
    tracer->AddCPURecords(CurAnnotation(), start_ns_, PosixInNsec(),
                          BlockDepth(), CurThread());
X
Xin Pan 已提交
185
  }
Y
Yibing Liu 已提交
186
  ClearCurAnnotation();
187
  PopEvent(name_, dev_ctx_);
D
dangqingqing 已提交
188
}
D
dangqingqing 已提交
189

X
Xin Pan 已提交
190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217
RecordBlock::RecordBlock(int block_id) : start_ns_(PosixInNsec()) {
  if (g_state == ProfilerState::kDisabled) return;
  SetCurBlock(block_id);
  name_ = string::Sprintf("block_%d", block_id);
}

RecordBlock::~RecordBlock() {
  if (g_state == ProfilerState::kDisabled) return;
  DeviceTracer* tracer = GetDeviceTracer();
  if (tracer) {
    // We try to put all blocks at the same nested depth in the
    // same timeline lane. and distinguish the using thread_id.
    tracer->AddCPURecords(name_, start_ns_, PosixInNsec(), BlockDepth(),
                          CurThread());
  }
  ClearCurBlock();
}

RecordThread::RecordThread(int thread_id) {
  if (g_state == ProfilerState::kDisabled) return;
  SetCurThread(thread_id);
}

RecordThread::~RecordThread() {
  if (g_state == ProfilerState::kDisabled) return;
  ClearCurThread();
}

D
dangqingqing 已提交
218 219 220 221
void EnableProfiler(ProfilerState state) {
  PADDLE_ENFORCE(state != ProfilerState::kDisabled,
                 "Can't enbale profling, since the input state is ",
                 "ProfilerState::kDisabled");
D
dangqingqing 已提交
222
  PADDLE_ENFORCE(g_state == ProfilerState::kDisabled,
D
dangqingqing 已提交
223 224
                 "The profiling state should be disabled when calling ",
                 "EnableProfiler.");
D
dangqingqing 已提交
225
  g_state = state;
226
  if (g_state == ProfilerState::kAll) {
227 228
    GetDeviceTracer()->Enable();
  }
D
dangqingqing 已提交
229
#ifdef PADDLE_WITH_CUDA
D
dangqingqing 已提交
230
  if (g_state == ProfilerState::kCUDA) {
231
    // Generate some dummy events first to reduce the startup overhead.
D
dangqingqing 已提交
232 233
    for (int i = 0; i < 5; i++) {
      ForEachDevice([](int d) {
D
dangqingqing 已提交
234
        DeviceContext* dev_ctx = new CUDADeviceContext(CUDAPlace(d));
D
dangqingqing 已提交
235 236
        Mark("_cuda_startup_", dev_ctx);
        dev_ctx->Wait();
D
dangqingqing 已提交
237
        delete dev_ctx;
D
dangqingqing 已提交
238 239 240 241 242
      });
    }
  }
#endif
  // Mark the profiling start.
D
dangqingqing 已提交
243
  Mark("_start_profiler_", nullptr);
D
dangqingqing 已提交
244 245
}

246
void ResetProfiler() {
D
dangqingqing 已提交
247
  std::lock_guard<std::mutex> guard(g_all_event_lists_mutex);
248 249 250 251 252 253 254 255 256
  for (auto it = g_all_event_lists.begin(); it != g_all_event_lists.end();
       ++it) {
    (*it)->Clear();
  }
}

std::vector<std::vector<Event>> GetAllEvents() {
  std::lock_guard<std::mutex> guard(g_all_event_lists_mutex);
  std::vector<std::vector<Event>> result;
D
dangqingqing 已提交
257 258 259
  for (auto it = g_all_event_lists.begin(); it != g_all_event_lists.end();
       ++it) {
    result.emplace_back((*it)->Reduce());
D
dangqingqing 已提交
260 261 262 263
  }
  return result;
}

264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291
// The information of each event given in the profiling report
struct EventItem {
  std::string name;
  int calls;
  double total_time;
  double min_time;
  double max_time;
  double ave_time;
};

// Print results
void PrintProfiler(const std::vector<std::vector<EventItem>>& events_table,
                   const std::string& sorted_domain, const size_t name_width,
                   const size_t data_width) {
  // Output header information
  std::cout << "\n------------------------->"
            << "     Profiling Report     "
            << "<-------------------------\n\n";
  std::string place;
  if (g_state == ProfilerState::kCPU) {
    place = "CPU";
  } else if (g_state == ProfilerState::kCUDA) {
    place = "CUDA";
  } else if (g_state == ProfilerState::kAll) {
    place = "All";
  } else {
    PADDLE_THROW("Invalid profiler state");
  }
292

293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312
  std::cout << "Place: " << place << std::endl;
  std::cout << "Time unit: ms" << std::endl;
  std::cout << "Sorted by " << sorted_domain
            << " in descending order in the same thread\n\n";
  // Output events table
  std::cout.setf(std::ios::left);
  std::cout << std::setw(name_width) << "Event" << std::setw(data_width)
            << "Calls" << std::setw(data_width) << "Total"
            << std::setw(data_width) << "Min." << std::setw(data_width)
            << "Max." << std::setw(data_width) << "Ave." << std::endl;
  for (size_t i = 0; i < events_table.size(); ++i) {
    for (size_t j = 0; j < events_table[i].size(); ++j) {
      const EventItem& event_item = events_table[i][j];
      std::cout << std::setw(name_width) << event_item.name
                << std::setw(data_width) << event_item.calls
                << std::setw(data_width) << event_item.total_time
                << std::setw(data_width) << event_item.min_time
                << std::setw(data_width) << event_item.max_time
                << std::setw(data_width) << event_item.ave_time << std::endl;
    }
313
  }
314
  std::cout << std::endl;
315 316
}

317 318 319 320
// Parse the event list and output the profiling report
void ParseEvents(const std::vector<std::vector<Event>>& events,
                 EventSortingKey sorted_by = EventSortingKey::kDefault) {
  if (g_state == ProfilerState::kDisabled) return;
321 322

  std::string sorted_domain;
L
Luo Tao 已提交
323
  std::function<bool(const EventItem&, const EventItem&)> sorted_func;
324 325 326
  switch (sorted_by) {
    case EventSortingKey::kCalls:
      sorted_domain = "number of calls";
L
Luo Tao 已提交
327
      sorted_func = [](const EventItem& a, const EventItem& b) {
328 329 330 331 332
        return a.calls > b.calls;
      };
      break;
    case EventSortingKey::kTotal:
      sorted_domain = "total time";
L
Luo Tao 已提交
333
      sorted_func = [](const EventItem& a, const EventItem& b) {
334 335 336 337 338
        return a.total_time > b.total_time;
      };
      break;
    case EventSortingKey::kMin:
      sorted_domain = "minimum time";
L
Luo Tao 已提交
339
      sorted_func = [](const EventItem& a, const EventItem& b) {
340 341 342 343 344
        return a.min_time > b.min_time;
      };
      break;
    case EventSortingKey::kMax:
      sorted_domain = "maximum time";
L
Luo Tao 已提交
345
      sorted_func = [](const EventItem& a, const EventItem& b) {
346 347 348 349 350
        return a.max_time > b.max_time;
      };
      break;
    case EventSortingKey::kAve:
      sorted_domain = "average time";
L
Luo Tao 已提交
351
      sorted_func = [](const EventItem& a, const EventItem& b) {
352 353 354 355
        return a.ave_time > b.ave_time;
      };
      break;
    default:
356
      sorted_domain = "event first end time";
357 358
  }

359
  std::vector<std::vector<EventItem>> events_table;
Y
Yibing Liu 已提交
360
  size_t max_name_width = 0;
361 362
  for (size_t i = 0; i < events.size(); i++) {
    std::list<Event> pushed_events;
363 364 365
    std::vector<EventItem> event_items;
    std::unordered_map<std::string, int> event_idx;

366
    for (size_t j = 0; j < events[i].size(); j++) {
367
      if (events[i][j].type() == EventType::kPushRange) {
368
        pushed_events.push_back(events[i][j]);
369
      } else if (events[i][j].type() == EventType::kPopRange) {
370
        std::list<Event>::reverse_iterator rit = pushed_events.rbegin();
371 372
        while (rit != pushed_events.rend() &&
               rit->name() != events[i][j].name()) {
373 374
          ++rit;
        }
375

376
        if (rit != pushed_events.rend()) {
377 378 379 380
          double event_time = (g_state == ProfilerState::kCUDA ||
                               g_state == ProfilerState::kAll)
                                  ? rit->CudaElapsedMs(events[i][j])
                                  : rit->CpuElapsedMs(events[i][j]);
381

382 383
          std::string event_name =
              "thread" + std::to_string(rit->thread_id()) + "::" + rit->name();
Y
Yibing Liu 已提交
384
          max_name_width = std::max(max_name_width, event_name.size());
385

386 387 388 389 390
          if (event_idx.find(event_name) == event_idx.end()) {
            event_idx[event_name] = event_items.size();
            EventItem event_item = {event_name, 1,          event_time,
                                    event_time, event_time, event_time};
            event_items.push_back(event_item);
391
          } else {
392 393
            int index = event_idx[event_name];
            event_items[index].calls += 1;
394
            // total time
395
            event_items[index].total_time += event_time;
396
            // min time
397 398
            event_items[index].min_time =
                std::min(event_time, event_items[index].min_time);
399
            // max time
400 401
            event_items[index].max_time =
                std::max(event_time, event_items[index].max_time);
402
          }
403

Y
Yibing Liu 已提交
404
          // remove the push marker from the list
405 406
          pushed_events.erase((++rit).base());
        } else {
407 408 409
          LOG(WARNING) << "Cannot find the push marker of event \'"
                       << events[i][j].name()
                       << "\', which will be ignored in profiling report.";
410 411 412
        }
      }
    }
413 414 415 416 417 418
    // average time
    for (auto& item : event_items) {
      item.ave_time = item.total_time / item.calls;
    }
    // sort
    if (sorted_by != EventSortingKey::kDefault) {
419
      std::sort(event_items.begin(), event_items.end(), sorted_func);
420
    }
421

422
    events_table.push_back(event_items);
Y
Yibing Liu 已提交
423
    // log warning if there are events with `push` but without `pop`
424 425
    std::list<Event>::reverse_iterator rit = pushed_events.rbegin();
    while (rit != pushed_events.rend()) {
Y
Yibing Liu 已提交
426 427
      LOG(WARNING) << "Cannot find the pop marker of event \'" << rit->name()
                   << "\', which will be ignored in profiling report.";
428 429
      ++rit;
    }
430
  }
431 432

  // Print report
433
  PrintProfiler(events_table, sorted_domain, max_name_width + 4, 12);
434 435
}

436 437 438 439 440 441 442 443 444 445 446 447 448 449
void DisableProfiler(EventSortingKey sorted_key,
                     const std::string& profile_path) {
  PADDLE_ENFORCE(g_state != ProfilerState::kDisabled,
                 "Can't disable profiling, since it's not starting.");
  // Mark the profiling stop.
  Mark("_stop_profiler_", nullptr);

  std::vector<std::vector<Event>> all_events = GetAllEvents();
  ParseEvents(all_events, sorted_key);
  ResetProfiler();
  DeviceTracer* tracer = GetDeviceTracer();
  if (g_state == ProfilerState::kAll && tracer && tracer->IsEnabled()) {
    tracer->Disable();
    tracer->GenProfile(profile_path);
450
  }
451
  g_state = ProfilerState::kDisabled;
452 453
}

D
dangqingqing 已提交
454 455
}  // namespace platform
}  // namespace paddle