// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include "paddle/fluid/framework/eigen.h" #include "paddle/fluid/framework/op_registry.h" #include "paddle/fluid/framework/tensor_util.h" #include "paddle/fluid/platform/float16.h" #include "paddle/fluid/platform/transform.h" #include "paddle/phi/kernels/isfinite_kernel.h" #include "paddle/phi/kernels/reduce_all_kernel.h" #include "paddle/phi/kernels/reduce_any_kernel.h" namespace phi { class DenseTensor; } // namespace phi namespace paddle { namespace framework { // store the result bool in gpu tensor, async operation. Faster than above ones. void TensorContainsNAN(const framework::Tensor& tensor, framework::Tensor* out); void TensorContainsInf(const framework::Tensor& tensor, framework::Tensor* out); void TensorIsfinite(const framework::Tensor& tensor, framework::Tensor* out); // copy the result bool to cpu bool TensorContainsNAN(const framework::Tensor& tensor); bool TensorContainsInf(const framework::Tensor& tensor); bool TensorIsfinite(const framework::Tensor& tensor); #define FiniteVisitor(type, reduce_type, device) \ struct type##Visitor##device { \ type##Visitor##device(const phi::DenseTensor& in, phi::DenseTensor* out) \ : in_(in), out_(out) {} \ template \ void apply() const { \ auto place = in_.place(); \ auto* ctx = static_cast( \ platform::DeviceContextPool::Instance().Get(place)); \ Tensor tmp; \ tmp.Resize(in_.dims()); \ out_->Resize({1}); \ std::vector dims(tmp.dims().size()); \ std::iota(dims.begin(), dims.end(), 0); \ phi::type##Kernel(*ctx, in_, &tmp); \ phi::reduce_type##Kernel( \ *ctx, tmp, dims, false, out_); \ } \ const phi::DenseTensor& in_; \ phi::DenseTensor* out_; \ }; FiniteVisitor(Isnan, Any, CPU); FiniteVisitor(Isinf, Any, CPU); FiniteVisitor(Isfinite, All, CPU); #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) FiniteVisitor(Isnan, Any, GPU); FiniteVisitor(Isinf, Any, GPU); FiniteVisitor(Isfinite, All, GPU); #endif // store the result bool in gpu tensor, async operation. Faster than above ones. inline void TensorContainsNAN(const framework::Tensor& tensor, framework::Tensor* out) { auto place = tensor.place(); if (platform::is_cpu_place(tensor.place())) { VisitDataTypeNormal(TransToProtoVarType(tensor.dtype()), IsnanVisitorCPU(tensor, out)); return; } #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) if (platform::is_gpu_place(place)) { VisitDataTypeNormal(TransToProtoVarType(tensor.dtype()), IsnanVisitorGPU(tensor, out)); return; } #endif PADDLE_THROW(platform::errors::Unimplemented("Not supported on %s.", place)); } inline void TensorContainsInf(const framework::Tensor& tensor, framework::Tensor* out) { auto place = tensor.place(); if (platform::is_cpu_place(tensor.place())) { VisitDataTypeNormal(TransToProtoVarType(tensor.dtype()), IsinfVisitorCPU(tensor, out)); return; } #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) if (platform::is_gpu_place(place)) { VisitDataTypeNormal(TransToProtoVarType(tensor.dtype()), IsinfVisitorGPU(tensor, out)); return; } #endif PADDLE_THROW(platform::errors::Unimplemented("Not supported on %s.", place)); } inline void TensorIsfinite(const framework::Tensor& tensor, framework::Tensor* out) { auto place = tensor.place(); if (platform::is_cpu_place(tensor.place())) { VisitDataTypeNormal(TransToProtoVarType(tensor.dtype()), IsfiniteVisitorCPU(tensor, out)); return; } #if defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP) if (platform::is_gpu_place(place)) { VisitDataTypeNormal(TransToProtoVarType(tensor.dtype()), IsfiniteVisitorGPU(tensor, out)); return; } #endif PADDLE_THROW(platform::errors::Unimplemented("Not supported on %s.", place)); } // copy the result bool to cpu inline bool TensorContainsNAN(const framework::Tensor& tensor) { Tensor out; TensorContainsNAN(tensor, &out); return GetValue(&out); } inline bool TensorContainsInf(const framework::Tensor& tensor) { Tensor out; TensorContainsInf(tensor, &out); return GetValue(&out); } inline bool TensorIsfinite(const framework::Tensor& tensor) { Tensor out; TensorIsfinite(tensor, &out); return GetValue(&out); } } // namespace framework namespace operators { struct InfinityFunctor { void operator()(const framework::Tensor& tensor, framework::Tensor* out) { framework::TensorContainsInf(tensor, out); } }; struct NANFunctor { void operator()(const framework::Tensor& tensor, framework::Tensor* out) { framework::TensorContainsNAN(tensor, out); } }; struct IsfiniteFunctor { void operator()(const framework::Tensor& tensor, framework::Tensor* out) { framework::TensorIsfinite(tensor, out); } }; template class OverflowKernel : public framework::OpKernel { public: virtual void Compute(const framework::ExecutionContext& ctx) const { auto* x = ctx.InputVar("X"); auto* out = ctx.Output("Out"); out->mutable_data(ctx.GetPlace()); Functor functor; if (x->IsType()) { auto* in = ctx.Input("X"); functor(*in, out); } else if (x->IsType()) { auto& in = ctx.Input("X")->value(); functor(in, out); } else { PADDLE_ENFORCE_EQ( true, false, platform::errors::InvalidArgument( "The input type mismatch, the type of Input(X) must be Tensor or " "SelectedRows, please check your input.")); } } }; } // namespace operators } // namespace paddle