// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/phi/kernels/arg_min_max_kernel.h" #include "paddle/phi/backends/gpu/gpu_context.h" #include "paddle/phi/core/kernel_registry.h" #if defined(__NVCC__) || defined(__HIPCC__) #ifdef __NVCC__ #include "cub/cub.cuh" #endif #ifdef __HIPCC__ #include namespace cub = hipcub; #endif #include #include "paddle/fluid/framework/data_type.h" #include "paddle/phi/core/ddim.h" namespace phi { namespace { // NOLINT template using KeyValuePair = cub::KeyValuePair; } // end namespace #define FIXED_BLOCK_DIM_CASE_BASE(log2_block_dim, ...) \ case (1 << (log2_block_dim)): { \ constexpr auto kBlockDim = (1 << (log2_block_dim)); \ __VA_ARGS__; \ } break #define FIXED_BLOCK_DIM_CASE(...) \ FIXED_BLOCK_DIM_CASE_BASE(10, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(9, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(8, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(7, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(6, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(5, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(4, ##__VA_ARGS__); \ FIXED_BLOCK_DIM_CASE_BASE(3, ##__VA_ARGS__); template __global__ void ArgCUDAKernel(const int64_t height, // n * h const int64_t width, // c const int64_t post_size, // h const Reducer reducer, const T init, const T* in, IndType* out) { typedef cub::BlockReduce, BlockDim> BlockReduce; __shared__ typename BlockReduce::TempStorage temp_storage; for (int idx = blockIdx.x; idx < height; idx += gridDim.x) { KeyValuePair kv_pair = {-1, init}; int h = idx / post_size; int w = idx % post_size; for (int k = threadIdx.x; k < width; k += blockDim.x) { kv_pair = reducer({k, in[h * width * post_size + k * post_size + w]}, kv_pair); } kv_pair = BlockReduce(temp_storage).Reduce(kv_pair, reducer); if (threadIdx.x == 0) { out[idx] = static_cast(kv_pair.key); } __syncthreads(); } } template void ComputeFullArg(const phi::GPUContext& dev_ctx, const DenseTensor& input, DenseTensor* indices, const int64_t pre, const int64_t post, const int64_t n) { auto cu_stream = dev_ctx.stream(); auto ComputeBlockSize = [](int64_t col) { auto block_size = 8; if (col > 512) block_size = 1024; else if (col > 256) block_size = 512; else if (col > 128) block_size = 256; else if (col > 64) block_size = 128; else if (col > 32) block_size = 64; else if (col > 16) block_size = 32; else if (col > 8) block_size = 16; #ifdef __HIPCC__ block_size = std::min(block_size, 256); #endif return block_size; }; int64_t max_grid_dimx = dev_ctx.GetCUDAMaxGridDimSize()[0]; int64_t height = pre * post; int64_t width = n; int64_t grid_size = height < max_grid_dimx ? height : max_grid_dimx; const T* in_data = input.data(); IndType* out_data = dev_ctx.template Alloc(indices); if (typeid(Reducer) == typeid(cub::ArgMax)) { switch (ComputeBlockSize(width)) { FIXED_BLOCK_DIM_CASE(ArgCUDAKernel <<>>( height, width, post, Reducer(), std::numeric_limits::lowest(), in_data, out_data)); } } else { switch (ComputeBlockSize(width)) { FIXED_BLOCK_DIM_CASE(ArgCUDAKernel <<>>( height, width, post, Reducer(), std::numeric_limits::max(), in_data, out_data)); } } } template struct VisitDataCudaArgMinMaxFunctor { const Context& dev_ctx; const DenseTensor& x; int64_t axis; bool keepdims; bool flatten; DenseTensor* out; explicit VisitDataCudaArgMinMaxFunctor(const Context& dev_ctx, const DenseTensor& x, int64_t axis, bool keepdims, bool flatten, DenseTensor* out) : dev_ctx(dev_ctx), x(x), axis(axis), keepdims(keepdims), flatten(flatten), out(out) {} template void apply() const { phi::DDim x_dims; int new_axis = axis; if (flatten) { x_dims = phi::make_ddim({x.numel()}); // if flatten, the axis just as 0 new_axis = 0; } else { x_dims = x.dims(); if (axis < 0) new_axis = axis + x.dims().size(); } int64_t numel = x.numel(); int64_t groups = numel / x_dims[new_axis]; int64_t pre = 1; int64_t post = 1; int64_t n = x_dims[new_axis]; for (int i = 0; i < new_axis; i++) { pre *= x_dims[i]; } for (int i = new_axis + 1; i < x_dims.size(); i++) { post *= x_dims[i]; } ComputeFullArg(dev_ctx, x, out, pre, post, n); } }; template void ArgMinMaxOpCUDAKernel(const Context& dev_ctx, const DenseTensor& x, const Scalar& axis, bool keepdims, bool flatten, int dtype, DenseTensor* out) { if (dtype < 0) { paddle::framework::VisitDataTypeTiny( static_cast( paddle::framework::proto::VarType::INT64), VisitDataCudaArgMinMaxFunctor( dev_ctx, x, axis.to(), keepdims, flatten, out)); return; } paddle::framework::VisitDataTypeTiny( static_cast(dtype), VisitDataCudaArgMinMaxFunctor( dev_ctx, x, axis.to(), keepdims, flatten, out)); } template void ArgMinKernel(const Context& dev_ctx, const DenseTensor& x, const Scalar& axis, bool keepdims, bool flatten, int dtype, DenseTensor* out) { ArgMinMaxOpCUDAKernel( dev_ctx, x, axis, keepdims, flatten, dtype, out); } template void ArgMaxKernel(const Context& dev_ctx, const DenseTensor& x, const Scalar& axis, bool keepdims, bool flatten, int dtype, DenseTensor* out) { ArgMinMaxOpCUDAKernel( dev_ctx, x, axis, keepdims, flatten, dtype, out); } #endif } // namespace phi PD_REGISTER_KERNEL(arg_min, GPU, ALL_LAYOUT, phi::ArgMinKernel, phi::dtype::float16, phi::dtype::bfloat16, float, double, int32_t, int64_t, int16_t, uint8_t) {} PD_REGISTER_KERNEL(arg_max, GPU, ALL_LAYOUT, phi::ArgMaxKernel, phi::dtype::float16, phi::dtype::bfloat16, float, double, int32_t, int64_t, int16_t, uint8_t) {}