// Copyright (c) 2023 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/phi/kernels/reduce_kernel.h" #include "paddle/phi/backends/all_context.h" #include "paddle/phi/core/kernel_registry.h" #if defined(PADDLE_WITH_GLOO) #include "paddle/phi/core/distributed/gloo_comm_context.h" #endif namespace phi { template void ReduceKernel(const Context& dev_ctx, const DenseTensor& x, int root, int reduce_type, DenseTensor* out) { #if defined(PADDLE_WITH_GLOO) out->Resize(x.dims()); dev_ctx.template Alloc(out); auto comm_ctx = static_cast(dev_ctx.GetCommContext()); PADDLE_ENFORCE_NE( comm_ctx, nullptr, errors::Unavailable("NCCLCommContext is nullptr, collective op should " "has ring_id attr.")); comm_ctx->Reduce(out, x, reduce_type, root); #else PADDLE_THROW( errors::PreconditionNotMet("PaddlePaddle should compile with GPU.")); #endif } } // namespace phi PD_REGISTER_KERNEL(reduce, CPU, ALL_LAYOUT, phi::ReduceKernel, float, double, int, bool, int8_t, uint8_t, int64_t, phi::dtype::float16) {}