/* Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #pragma once #include #include #include #include #include #include #include #include "paddle/fluid/framework/program_desc.h" #include "paddle/fluid/framework/scope.h" #include "paddle/fluid/framework/variable_helper.h" #include "paddle/fluid/platform/macros.h" // for DISABLE_COPY_AND_ASSIGN namespace paddle { namespace framework { class NCCLInfo { public: NCCLInfo() {} virtual ~NCCLInfo() {} public: int local_rank_; int global_ranks_; int my_global_rank_; ncclUniqueId nccl_id_; ncclComm_t comm_; cudaStream_t stream_; }; class NCCLWrapper { public: virtual ~NCCLWrapper() {} NCCLWrapper() {} void InitNCCL(); void SetNCCLId(const NCCLInfo& nccl_info); NCCLInfo GetNCCLId(); void SetRankInfo(const int local_rank, const int global_rank, const int ranks); void SyncVar(const int root_rank, const Scope& scope, const std::vector& var_names); static std::shared_ptr GetInstance() { if (NULL == s_instance_) { s_instance_.reset(new paddle::framework::NCCLWrapper()); } return s_instance_; } public: NCCLInfo nccl_info_; private: static std::shared_ptr s_instance_; protected: static bool is_initialized_; DISABLE_COPY_AND_ASSIGN(NCCLWrapper); }; } // end namespace framework } // end namespace paddle