/* Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #pragma once #include #include #include #ifdef PADDLE_WITH_CUDA #include #endif #ifdef PADDLE_WITH_HIP #include #endif #include "paddle/fluid/platform/place.h" #include "paddle/fluid/platform/stream/cuda_stream.h" namespace paddle { namespace platform { enum class EventType { kMark, kPushRange, kPopRange }; enum class EventRole { kOrdinary, // only record op time with op type key kInnerOp, // record op detail time with op type key kUniqueOp, // record op detail time with op unique name key kSpecial, // record event such as PE which is outer of thread local }; class Event { public: // The DeviceContext is used to get the cuda stream. // If CPU profiling mode, can pass nullptr. Event(EventType type, std::string name, uint32_t thread_id, EventRole role = EventRole::kOrdinary, std::string attr = "none"); const EventType& type() const; Event* parent() const { return parent_; } void set_parent(Event* parent) { parent_ = parent; } std::string name() const { return name_; } EventRole role() const { return role_; } uint64_t thread_id() const { return thread_id_; } void set_name(std::string name) { name_ = name; } void set_role(EventRole role) { role_ = role; } std::string attr() const { return attr_; } #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) #ifndef PADDLE_WITH_CUPTI gpuEvent_t event() const { return event_; } int device() const { return device_; } #endif #endif double CpuElapsedMs(const Event& e) const; double CudaElapsedMs(const Event& e) const; private: EventType type_; std::string name_{}; Event* parent_{nullptr}; uint64_t thread_id_; EventRole role_{}; int64_t cpu_ns_; bool visited_status_{false}; std::string attr_; #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) #ifdef PADDLE_WITH_CUPTI int64_t gpu_ns_ = 0; public: void AddCudaElapsedTime(int64_t start_ns, int64_t end_ns) { gpu_ns_ += end_ns - start_ns; } private: #else gpuEvent_t event_ = nullptr; int device_ = -1; #endif #endif }; using EventWithStartNs = std::pair; using ThreadEvents = std::map; class MemEvent { public: MemEvent(EventType type, uint64_t start_ns, uint64_t end_ns, size_t bytes, Place place, int64_t thread_id, const std::string& annotation) : type_(type), start_ns_(start_ns), end_ns_(end_ns), bytes_(bytes), place_(place), thread_id_(thread_id), annotation_(annotation) {} const EventType& type() const { return type_; } uint64_t start_ns() const { return start_ns_; } uint64_t end_ns() const { return end_ns_; } size_t bytes() const { return bytes_; } Place place() const { return place_; } uint64_t thread_id() const { return thread_id_; } const std::string& annotation() const { return annotation_; } private: EventType type_; uint64_t start_ns_ = 0; uint64_t end_ns_ = 0; size_t bytes_; Place place_; uint64_t thread_id_; std::string annotation_; }; class CudaEvent { #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) public: CudaEvent() { #ifdef PADDLE_WITH_HIP hipEventCreateWithFlags(&event_, flags_); #else cudaEventCreateWithFlags(&event_, flags_); #endif } explicit CudaEvent(unsigned int flags) : flags_(flags) { #ifdef PADDLE_WITH_HIP hipEventCreateWithFlags(&event_, flags_); #else cudaEventCreateWithFlags(&event_, flags_); #endif } ~CudaEvent() { #ifdef PADDLE_WITH_HIP hipEventDestroy(event_); #else cudaEventDestroy(event_); #endif } void Record(const paddle::platform::stream::CUDAStream& stream) { #ifdef PADDLE_WITH_HIP PADDLE_ENFORCE_GPU_SUCCESS(hipEventRecord(event_, stream.raw_stream())); #else PADDLE_ENFORCE_GPU_SUCCESS(cudaEventRecord(event_, stream.raw_stream())); #endif } bool Query() { #ifdef PADDLE_WITH_HIP gpuError_t err = hipEventQuery(event_); if (err == hipSuccess) { return true; } if (err == hipErrorNotReady) { return false; } #else gpuError_t err = cudaEventQuery(event_); if (err == cudaSuccess) { return true; } if (err == cudaErrorNotReady) { return false; } #endif PADDLE_ENFORCE_GPU_SUCCESS(err); return false; } void Synchronize() { #ifdef PADDLE_WITH_HIP PADDLE_ENFORCE_GPU_SUCCESS(hipEventSynchronize(event_)); #else PADDLE_ENFORCE_GPU_SUCCESS(cudaEventSynchronize(event_)); #endif } gpuEvent_t GetRawCudaEvent() { return event_; } private: #ifdef PADDLE_WITH_HIP unsigned int flags_ = hipEventDefault; #else unsigned int flags_ = cudaEventDefault; #endif gpuEvent_t event_; #endif }; } // namespace platform } // namespace paddle