// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include // NOLINT #include #include #include "paddle/fluid/platform/enforce.h" namespace paddle { namespace operators { namespace reader { template class BlockingQueue { // BlockingQueue is for buffered reading and is supposed to use only the // reader package. It is true that we could and we should have been using // framework::Channel, but which has currently a deadlock bug. BlockingQueue // is a workaround and a simplified version of framework::Channel as it // doesn't support GPU and it implements on buffered blocking queue. public: explicit BlockingQueue(size_t capacity, bool speed_test_mode = false) : capacity_(capacity), speed_test_mode_(speed_test_mode), closed_(false) { PADDLE_ENFORCE_GT( capacity_, static_cast(0), "The capacity of a reader::BlockingQueue must be greater than 0."); } bool Send(const T& elem) { std::unique_lock lock(mutex_); send_cv_.wait(lock, [&] { return queue_.size() < capacity_ || closed_; }); if (closed_) { VLOG(5) << "WARNING: Sending an element to a closed reader::BlokcingQueue."; return false; } PADDLE_ENFORCE_LT(queue_.size(), capacity_); queue_.push_back(elem); receive_cv_.notify_one(); return true; } bool Send(T&& elem) { std::unique_lock lock(mutex_); send_cv_.wait(lock, [&] { return queue_.size() < capacity_ || closed_; }); if (closed_) { VLOG(5) << "WARNING: Sending an element to a closed reader::BlokcingQueue."; return false; } PADDLE_ENFORCE_LT(queue_.size(), capacity_); queue_.emplace_back(std::move(elem)); receive_cv_.notify_one(); return true; } bool Receive(T* elem) { std::unique_lock lock(mutex_); receive_cv_.wait(lock, [&] { return !queue_.empty() || closed_; }); if (!queue_.empty()) { PADDLE_ENFORCE_NOT_NULL(elem); *elem = queue_.front(); if (LIKELY(!speed_test_mode_)) { queue_.pop_front(); } send_cv_.notify_one(); return true; } else { PADDLE_ENFORCE(closed_); VLOG(3) << "queue is closed! return nothing."; return false; } } void ReOpen() { std::lock_guard lock(mutex_); VLOG(1) << "reopen queue"; closed_ = false; std::deque new_deque; queue_.swap(new_deque); send_cv_.notify_all(); receive_cv_.notify_all(); } void Close() { std::lock_guard lock(mutex_); VLOG(1) << "close queue"; closed_ = true; send_cv_.notify_all(); receive_cv_.notify_all(); } bool IsClosed() const { std::lock_guard lock(mutex_); return closed_; } size_t Cap() const { std::lock_guard lock(mutex_); return capacity_; } size_t Size() const { std::lock_guard lock(mutex_); return queue_.size(); } private: size_t capacity_; bool speed_test_mode_; bool closed_; std::deque queue_; mutable std::mutex mutex_; mutable std::condition_variable receive_cv_; mutable std::condition_variable send_cv_; }; } // namespace reader } // namespace operators } // namespace paddle