// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include #include #include #include #include #include #include "paddle/fluid/distributed/fleet_executor/interceptor.h" #include "paddle/fluid/distributed/fleet_executor/interceptor_message.pb.h" #include "paddle/fluid/platform/device_context.h" #include "paddle/fluid/platform/enforce.h" #include "paddle/fluid/platform/errors.h" #include "paddle/fluid/platform/macros.h" #include "paddle/fluid/platform/place.h" namespace paddle { namespace framework { class Scope; } namespace distributed { class TaskNode; class InterceptorMessageServiceImpl; class RuntimeGraph; class MessageBus; class Carrier final { public: Carrier() = default; Carrier(int64_t rank, const std::unordered_map& interceptor_id_to_rank) : rank_(rank), interceptor_id_to_rank_(interceptor_id_to_rank) {} ~Carrier(); void Init(int64_t rank, std::shared_ptr runtime_graph, framework::Scope* root_scope, framework::Scope* minibatch_scope, const std::vector& microbatch_scopes, const platform::Place& place); void Release(); void Wait(); // Enqueue a message to corresponding interceptor id bool EnqueueInterceptorMessage(const InterceptorMessage& interceptor_message); // get interceptor based on the interceptor id Interceptor* GetInterceptor(int64_t interceptor_id); // set interceptor with interceptor id Interceptor* SetInterceptor(int64_t interceptor_id, std::unique_ptr); void SetCreatingFlag(bool flag); void SetMsgBus(const std::shared_ptr& msg_bus) { msg_bus_ = msg_bus; } std::condition_variable& GetCondVar(); void Start(); bool IsInit() const; bool Send(const InterceptorMessage& msg); // NOTE: This mutex will be used in interceptor's RunOps function. // This mutex is used for avoiding forward ops and backward ops run // simultaneously, which will lead to a random hang for some sync ops. std::mutex run; private: DISABLE_COPY_AND_ASSIGN(Carrier); // create each Interceptor void CreateInterceptors(); void HandleTmpMessages(); int64_t GetRank(int64_t interceptor_id) const; // interceptor logic id to actually interceptor std::unordered_map> interceptor_idx_to_interceptor_; std::vector source_interceptor_ids_; std::vector message_tmp_{}; std::mutex tmp_message_mutex_; bool creating_interceptors_{true}; std::mutex creating_flag_mutex_; bool is_init_{false}; std::mutex running_mutex_; std::condition_variable cond_var_; std::vector microbatch_scopes_; framework::Scope* root_scope_; framework::Scope* minibatch_scope_; paddle::platform::Place place_; paddle::platform::DeviceContext* dev_ctx_{nullptr}; std::shared_ptr runtime_graph_; std::shared_ptr msg_bus_; int64_t rank_; std::unordered_map interceptor_id_to_rank_; }; } // namespace distributed } // namespace paddle