// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/phi/kernels/top_k_kernel.h" #include "paddle/phi/backends/xpu/enforce_xpu.h" #include "paddle/phi/core/kernel_registry.h" namespace phi { template void TopkKernel(const Context& dev_ctx, const DenseTensor& x, const Scalar& k_scalar, int axis, bool largest, bool sorted, DenseTensor* out, DenseTensor* indices) { const auto& in_dims = x.dims(); const T* in_data = x.data(); int64_t* indices_data = dev_ctx.template Alloc(indices); T* output_data = dev_ctx.template Alloc(out); const auto& out_dims = out->dims(); PADDLE_ENFORCE_EQ( sorted, true, errors::External( "XPU API does not support unsorted topk operation currently." " Operator will be supported in future update.")); PADDLE_ENFORCE_EQ( largest, true, errors::External( "XPU API does not support smallest topk operation currently." " Operator will be supported in future update.")); if (axis < 0) axis += in_dims.size(); size_t k = k_scalar.to(); if (axis + 1 == in_dims.size()) { xpu::ctx_guard RAII_GUARD(dev_ctx.x_context()); int32_t* indices_int_data = RAII_GUARD.alloc_l3_or_gm(indices->numel()); const size_t row = phi::product(phi::slice_ddim(in_dims, 0, in_dims.size() - 1)); const size_t col = in_dims[in_dims.size() - 1]; int r = xpu::sorted_topk(dev_ctx.x_context(), in_data, output_data, indices_int_data, row, col, k); PADDLE_ENFORCE_XDNN_SUCCESS(r, "sorted_topk"); r = xpu::cast(dev_ctx.x_context(), (const int32_t*)indices_int_data, indices_data, indices->numel()); PADDLE_ENFORCE_XDNN_SUCCESS(r, "cast"); } else { // do transpose if axis is not the last dim of input std::vector trans_axes; for (int i = 0; i < axis; i++) { trans_axes.emplace_back(i); } for (int i = axis + 1; i < in_dims.size(); i++) { trans_axes.emplace_back(i); } trans_axes.emplace_back(axis); // Get input and output dims for transpose DDim trans_dims(in_dims); DDim trans_out_dims(out->dims()); for (size_t i = 0; i < trans_axes.size(); i++) { trans_dims[i] = in_dims[trans_axes[i]]; trans_out_dims[i] = out_dims[trans_axes[i]]; } std::vector x_shape_host(in_dims.size(), 0); for (int i = 0; i < in_dims.size(); ++i) { x_shape_host[i] = in_dims[i]; } xpu::ctx_guard RAII_GUARD(dev_ctx.x_context()); T* trans_in_data = RAII_GUARD.alloc_l3_or_gm(x.numel()); // Transpose and save interval output to trans_in int r = xpu::transpose( dev_ctx.x_context(), in_data, trans_in_data, x_shape_host, trans_axes); PADDLE_ENFORCE_EQ(r, xpu::Error_t::SUCCESS, errors::External("XPU API 1st Transpose kernel" " returns wrong value[%d %s]!", r, XPUAPIErrorMsg[r])); T* trans_out_data = RAII_GUARD.alloc_l3_or_gm(out->numel()); int64_t* trans_idx_data = RAII_GUARD.alloc_l3_or_gm(out->numel()); int32_t* trans_idx_int32_data = RAII_GUARD.alloc_l3_or_gm(out->numel()); const size_t row = phi::product(phi::slice_ddim(trans_dims, 0, trans_dims.size() - 1)); const size_t col = trans_dims[trans_dims.size() - 1]; // Do top k on transposed input r = xpu::sorted_topk(dev_ctx.x_context(), trans_in_data, trans_out_data, trans_idx_int32_data, row, col, k); PADDLE_ENFORCE_XDNN_SUCCESS(r, "sorted_topk"); r = xpu::cast(dev_ctx.x_context(), (const int32_t*)trans_idx_int32_data, trans_idx_data, indices->numel()); PADDLE_ENFORCE_XDNN_SUCCESS(r, "cast"); // Transpose back to original dims std::vector trans_back_axes; for (int i = 0; i < axis; i++) { trans_back_axes.emplace_back(i); } trans_back_axes.emplace_back(trans_out_dims.size() - 1); for (int i = axis; i < trans_out_dims.size() - 1; i++) { trans_back_axes.emplace_back(i); } std::vector trans_out_shape_host(trans_back_axes.size(), 0); for (size_t i = 0; i < trans_back_axes.size(); ++i) { trans_out_shape_host[i] = trans_out_dims[i]; } r = xpu::transpose(dev_ctx.x_context(), trans_out_data, output_data, trans_out_shape_host, trans_back_axes); PADDLE_ENFORCE_EQ(r, xpu::Error_t::SUCCESS, errors::External("XPU API 2nd Transpose kernel" " returns wrong value[%d %s]", r, XPUAPIErrorMsg[r])); r = xpu::transpose(dev_ctx.x_context(), trans_idx_data, indices_data, trans_out_shape_host, trans_back_axes); PADDLE_ENFORCE_EQ(r, xpu::Error_t::SUCCESS, errors::External("XPU API 3rd Transpose kernel" " returns wrong value[%d %s]", r, XPUAPIErrorMsg[r])); } } } // namespace phi PD_REGISTER_KERNEL(topk, XPU, ALL_LAYOUT, phi::TopkKernel, float) {}