// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/phi/kernels/matrix_rank_tol_kernel.h" #include "paddle/phi/core/kernel_registry.h" #include "paddle/phi/kernels/abs_kernel.h" #include "paddle/phi/kernels/elementwise_multiply_kernel.h" #include "paddle/phi/kernels/full_kernel.h" #include "paddle/phi/kernels/funcs/compare_functors.h" #include "paddle/phi/kernels/funcs/elementwise_base.h" #include "paddle/phi/kernels/funcs/lapack/lapack_function.h" #include "paddle/phi/kernels/funcs/values_vectors_functor.h" #include "paddle/phi/kernels/impl/matrix_rank_kernel_impl.h" #include "paddle/phi/kernels/reduce_max_kernel.h" #include "paddle/phi/kernels/reduce_sum_kernel.h" #include "paddle/phi/kernels/transpose_kernel.h" namespace phi { template void LapackSVD(const T* x_data, T* eigenvalues_data, int rows, int cols) { char jobz = 'N'; int mx = std::max(rows, cols); int mn = std::min(rows, cols); T* a = const_cast(x_data); int lda = rows; int lwork = 3 * mn + std::max(mx, 7 * mn); std::vector work(lwork); std::vector iwork(8 * mn); int info; phi::funcs::lapackSvd(jobz, rows, cols, a, lda, eigenvalues_data, nullptr, 1, nullptr, 1, work.data(), lwork, iwork.data(), &info); if (info < 0) { PADDLE_THROW(phi::errors::InvalidArgument( "This %s-th argument has an illegal value", info)); } if (info > 0) { PADDLE_THROW(phi::errors::InvalidArgument( "DBDSDC/SBDSDC did not converge, updating process failed. May be you " "passes a invalid matrix.")); } } template void BatchSVD( const T* x_data, T* eigenvalues_data, int batches, int rows, int cols) { int stride = rows * cols; int k = std::min(rows, cols); for (int i = 0; i < batches; ++i) { LapackSVD(x_data + i * stride, eigenvalues_data + i * k, rows, cols); } } template void MatrixRankTolKernel(const Context& dev_ctx, const DenseTensor& x, const DenseTensor& atol_tensor, bool use_default_tol, bool hermitian, DenseTensor* out) { dev_ctx.template Alloc(out); auto dim_x = x.dims(); auto dim_out = out->dims(); int rows = dim_x[dim_x.size() - 2]; int cols = dim_x[dim_x.size() - 1]; int k = std::min(rows, cols); auto numel = x.numel(); int batches = numel / (rows * cols); T rtol_T = 0; if (use_default_tol) { rtol_T = std::numeric_limits::epsilon() * std::max(rows, cols); } DenseTensor eigenvalue_tensor; eigenvalue_tensor.Resize(detail::GetEigenvalueDim(dim_x, k)); auto* eigenvalue_data = dev_ctx.template Alloc(&eigenvalue_tensor); if (hermitian) { phi::funcs::MatrixEighFunctor functor; functor(dev_ctx, x, &eigenvalue_tensor, nullptr, true, false); phi::AbsKernel(dev_ctx, eigenvalue_tensor, &eigenvalue_tensor); } else { DenseTensor trans_x = phi::TransposeLast2Dim(dev_ctx, x); auto* x_data = trans_x.data(); BatchSVD(x_data, eigenvalue_data, batches, rows, cols); } DenseTensor max_eigenvalue_tensor; max_eigenvalue_tensor.Resize(detail::RemoveLastDim(eigenvalue_tensor.dims())); dev_ctx.template Alloc(&max_eigenvalue_tensor); phi::MaxKernel(dev_ctx, eigenvalue_tensor, phi::IntArray({-1}), false, &max_eigenvalue_tensor); DenseTensor temp_rtol_tensor; temp_rtol_tensor = phi::Full(dev_ctx, {1}, static_cast(rtol_T)); DenseTensor rtol_tensor = phi::Multiply(dev_ctx, temp_rtol_tensor, max_eigenvalue_tensor); DenseTensor tol_tensor; tol_tensor.Resize(dim_out); dev_ctx.template Alloc(&tol_tensor); funcs::ElementwiseCompute, T, T>( dev_ctx, atol_tensor, rtol_tensor, -1, GreaterElementFunctor(), &tol_tensor); tol_tensor.Resize(detail::NewAxisDim(tol_tensor.dims(), 1)); DenseTensor compare_result; compare_result.Resize(detail::NewAxisDim(dim_out, k)); dev_ctx.template Alloc(&compare_result); int axis = -1; if (eigenvalue_tensor.dims().size() >= tol_tensor.dims().size()) { funcs::ElementwiseCompute, T, int>( dev_ctx, eigenvalue_tensor, tol_tensor, axis, funcs::GreaterThanFunctor(), &compare_result); } else { funcs::ElementwiseCompute, T, int>( dev_ctx, eigenvalue_tensor, tol_tensor, axis, funcs::LessThanFunctor(), &compare_result); } phi::SumKernel(dev_ctx, compare_result, std::vector{-1}, compare_result.dtype(), false, out); } } // namespace phi PD_REGISTER_KERNEL( matrix_rank_tol, CPU, ALL_LAYOUT, phi::MatrixRankTolKernel, float, double) { }