/* Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #pragma once #include #include "paddle/phi/api/include/dll_decl.h" #include "paddle/phi/common/data_type.h" #include "paddle/phi/common/place.h" #include "paddle/phi/core/allocator.h" #include "paddle/phi/core/generator.h" namespace phi { class TensorBase; /** * DeviceContext provides device-related interfaces. * * All kernels must access the interfaces provided by the backend through * DeviceContext. */ class PADDLE_API DeviceContext { using DataType = paddle::experimental::DataType; public: /** * @brief Default construct. */ DeviceContext(); /** * @brief Copy construct. */ DeviceContext(const DeviceContext&); /** * @brief Move construct. */ DeviceContext(DeviceContext&&); /** * @brief Move assign operator. */ DeviceContext& operator=(DeviceContext&&); /** * @brief Default destruct. */ virtual ~DeviceContext(); /** * @brief Set the device-related Allocator object. * * @param allocator */ void SetAllocator(const Allocator*); /** * @brief Set the host Allocator object. * * @param allocator */ void SetHostAllocator(const Allocator*); /** * @brief Set the zero-size Allocator object. * * @param allocator */ void SetZeroAllocator(const Allocator*); /** * @brief Set the zero-size Allocator object. * * @param allocator */ void SetPinnedAllocator(const Allocator*); /** * @brief Get the const Allocator object. * * @return Allocator */ const Allocator& GetAllocator() const; /** * @brief Get the const device-related Allocator object. * * @return Allocator */ const Allocator& GetHostAllocator() const; const Allocator& GetZeroAllocator() const; const Allocator& GetPinnedAllocator() const; #ifdef PADDLE_WITH_CUDA /** * @brief Set the CUDA graph Allocator object. * * @param allocator */ void SetCUDAGraphAllocator(const Allocator*); /** * @brief Get the const CUDA graph Allocator object. * * @return Allocator */ const Allocator& GetCUDAGraphAllocator() const; /** * @brief Test whether the CUDA graph allocator is valid * * This method should be called before calling GetCUDAGraphAllocator(). * Other unit can calls GetCUDAGraphAllocator() method, * only when this method returns True! * * @return true if cuda_graph_allocator_ is valid, false otherwise */ bool IsCUDAGraphAllocatorValid() const; #endif /** * @brief Allocate device memory for tensor. */ void* Alloc(TensorBase*, DataType dtype, size_t requested_size = 0, bool pinned = false) const; template T* Alloc(TensorBase* tensor, size_t requested_size = 0, bool pinned = false) const; /** * @brief Allocate host memory for tensor. */ void* HostAlloc(TensorBase* tensor, DataType dtype, size_t requested_size = 0) const; template T* HostAlloc(TensorBase* tensor, size_t requested_size = 0) const; virtual const Place& GetPlace() const = 0; // TODO(wilber): The fluid framework uses wait() in many places, how to delete // this API interface. virtual void Wait() const {} /** * @brief Set the generator for special op. * * @param Generator */ void SetGenerator(Generator*); /** * @brief Get the generator object. * * @return Generator */ Generator* GetGenerator() const; /** * @brief Set the host generator for special op. * * @param Generator */ void SetHostGenerator(Generator*); /** * @brief Get the host generator object. * * @return Generator */ Generator* GetHostGenerator() const; private: struct Impl; std::unique_ptr impl_; }; } // namespace phi