diff --git a/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.cu b/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.cu index 5a87f39092379246b561cbe20f071652aac1928f..0881b702ec0d80b96fb3e9764d6b81daca97bc27 100644 --- a/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.cu +++ b/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.cu @@ -13,13 +13,13 @@ See the License for the specific language governing permissions and limitations under the License. */ #include "paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.h" -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/fluid/operators/math/softmax_impl.h" #include "paddle/fluid/platform/collective_helper.h" #include "paddle/fluid/platform/device/gpu/nccl_helper.h" #include "paddle/fluid/string/string_helper.h" #include "paddle/phi/backends/gpu/gpu_context.h" #include "paddle/phi/kernels/funcs/axis_utils.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" namespace paddle { namespace operators { @@ -237,9 +237,9 @@ struct CSoftmaxWithCrossEntropyFunctor { auto eigen_predicted_logits = math::EigenMatrix::From(predicted_logits); eigen_loss.device(*dev_ctx.eigen_device()) = - (eigen_sum_exp_logits.log().unaryExpr(math::TolerableValue()) - + (eigen_sum_exp_logits.log().unaryExpr(phi::funcs::TolerableValue()) - eigen_predicted_logits) - .unaryExpr(math::TolerableValue()); + .unaryExpr(phi::funcs::TolerableValue()); eigen_softmax.device(*dev_ctx.eigen_device()) = (eigen_softmax * @@ -372,9 +372,9 @@ struct CSoftmaxWithCrossEntropyProcessGroupFunctor { auto eigen_predicted_logits = math::EigenMatrix::From(predicted_logits); eigen_loss.device(*dev_ctx.eigen_device()) = - (eigen_sum_exp_logits.log().unaryExpr(math::TolerableValue()) - + (eigen_sum_exp_logits.log().unaryExpr(phi::funcs::TolerableValue()) - eigen_predicted_logits) - .unaryExpr(math::TolerableValue()); + .unaryExpr(phi::funcs::TolerableValue()); eigen_softmax.device(*dev_ctx.eigen_device()) = (eigen_softmax * diff --git a/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.h b/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.h index 0336d565de2bf88351714ec9e6d2c38f7b5c8a24..f3a438e729bb142478bae7f800bd60f33beaae93 100644 --- a/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.h +++ b/paddle/fluid/operators/collective/c_softmax_with_cross_entropy_op.h @@ -22,9 +22,9 @@ limitations under the License. */ #include "paddle/fluid/framework/data_type.h" #include "paddle/fluid/framework/lod_tensor.h" #include "paddle/fluid/framework/op_registry.h" -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/fluid/operators/math/softmax.h" #include "paddle/phi/api/include/tensor.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" namespace paddle { namespace operators { diff --git a/paddle/fluid/operators/cross_entropy_op.h b/paddle/fluid/operators/cross_entropy_op.h index 4dcaf7b99f091444ae3a606f2aa28ee9eccd2df0..8ae6f448d24baeccd50e7914423262d7248f4078 100644 --- a/paddle/fluid/operators/cross_entropy_op.h +++ b/paddle/fluid/operators/cross_entropy_op.h @@ -15,8 +15,8 @@ limitations under the License. */ #pragma once #include "paddle/fluid/framework/eigen.h" #include "paddle/fluid/framework/op_registry.h" -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/fluid/platform/for_range.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" #include "paddle/phi/kernels/funcs/math.h" #include "paddle/phi/kernels/funcs/math_function.h" @@ -51,7 +51,7 @@ class CrossEntropyOpKernel : public framework::OpKernel { } int axis_dim = x->dims()[rank - 1]; - math::CrossEntropyFunctor()( + phi::funcs::CrossEntropyFunctor()( ctx.template device_context(), &y_2d, &x_2d, @@ -190,7 +190,7 @@ struct HardLabelCrossEntropyForwardFunctor { label); auto match_x = x_[idx * feature_size_ + label]; - y_[idx] = -math::TolerableValue()(phi::funcs::real_log(match_x)); + y_[idx] = -phi::funcs::TolerableValue()(phi::funcs::real_log(match_x)); match_x_[idx] = match_x; } else { y_[idx] = 0; diff --git a/paddle/fluid/operators/math/CMakeLists.txt b/paddle/fluid/operators/math/CMakeLists.txt index 20933a4162fdc65f80e82b7d4bc81d0546a3d4fe..9bc7473d967cdea10abf6f18e102ae50d227736b 100644 --- a/paddle/fluid/operators/math/CMakeLists.txt +++ b/paddle/fluid/operators/math/CMakeLists.txt @@ -21,7 +21,6 @@ else() math_library(concat_and_split DEPS concat_and_split_functor) endif() math_library(context_project DEPS im2col math_function) -math_library(cross_entropy) math_library(cos_sim_functor) math_library(depthwise_conv) math_library(im2col) diff --git a/paddle/fluid/operators/softmax_with_cross_entropy_op_npu.cc b/paddle/fluid/operators/softmax_with_cross_entropy_op_npu.cc index ddcb07b4d77e479f02e7b58ef6407c894c902453..d42f993f46219b74f86a5d2c663cb29ba6495975 100644 --- a/paddle/fluid/operators/softmax_with_cross_entropy_op_npu.cc +++ b/paddle/fluid/operators/softmax_with_cross_entropy_op_npu.cc @@ -16,10 +16,10 @@ limitations under the License. */ #include #include "paddle/fluid/framework/op_registry.h" -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/fluid/operators/math/softmax.h" #include "paddle/fluid/platform/device/npu/npu_op_runner.h" #include "paddle/phi/kernels/funcs/axis_utils.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" namespace paddle { namespace operators { diff --git a/paddle/phi/kernels/cpu/cross_entropy_kernel.cc b/paddle/phi/kernels/cpu/cross_entropy_kernel.cc index 27675fa8b5a548fdc230aa358afc9015fd1a75ab..cc855a4666043c5ab5d65f57c047fcd44f6324d8 100644 --- a/paddle/phi/kernels/cpu/cross_entropy_kernel.cc +++ b/paddle/phi/kernels/cpu/cross_entropy_kernel.cc @@ -14,11 +14,11 @@ limitations under the License. */ #include "paddle/phi/kernels/cross_entropy_kernel.h" -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/phi/backends/cpu/cpu_context.h" #include "paddle/phi/core/kernel_registry.h" #include "paddle/phi/core/tensor_utils.h" #include "paddle/phi/kernels/funcs/axis_utils.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" #include "paddle/phi/kernels/funcs/math_function.h" #include "paddle/phi/kernels/softmax_kernel.h" @@ -64,7 +64,7 @@ void CrossEntropy(const CPUContext& dev_ctx, DenseTensor out_2d(*out); out_2d.Resize({n, d / axis_dim}); - paddle::operators::math::CrossEntropyFunctor()( + phi::funcs::CrossEntropyFunctor()( dev_ctx, &out_2d, &x_2d, &label_2d, soft_label, ignore_index, axis_dim); } diff --git a/paddle/phi/kernels/funcs/CMakeLists.txt b/paddle/phi/kernels/funcs/CMakeLists.txt index 9c26e33ccebab5a8d5e80cb0fe41c8c19cba7a9e..ac1bd1fd45c7202acd7f21a21d6620aabe9da1c0 100644 --- a/paddle/phi/kernels/funcs/CMakeLists.txt +++ b/paddle/phi/kernels/funcs/CMakeLists.txt @@ -16,6 +16,7 @@ math_library(pooling DEPS dense_tensor) math_library(segment_pooling) math_library(sequence2batch) math_library(matrix_solve DEPS dense_tensor eigen3 blas math_function) +math_library(cross_entropy) cc_library( phi_data_layout_transform diff --git a/paddle/fluid/operators/math/cross_entropy.cc b/paddle/phi/kernels/funcs/cross_entropy.cc similarity index 78% rename from paddle/fluid/operators/math/cross_entropy.cc rename to paddle/phi/kernels/funcs/cross_entropy.cc index f87f5a107e696e57d15833ff79fb4476c0e4e572..b93123fad808e162a48821c4fe400540e8af2833 100644 --- a/paddle/fluid/operators/math/cross_entropy.cc +++ b/paddle/phi/kernels/funcs/cross_entropy.cc @@ -1,4 +1,4 @@ -/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved. +/* Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. @@ -12,20 +12,19 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ -#include "paddle/fluid/operators/math/cross_entropy.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" -#include "paddle/fluid/framework/convert_utils.h" #include "paddle/phi/backends/cpu/cpu_context.h" +#include "paddle/phi/core/utils/data_type.h" -namespace paddle { -namespace operators { -namespace math { +namespace phi { +namespace funcs { using Tensor = phi::DenseTensor; template -using EigenMatrix = framework::EigenMatrix; +using EigenMatrix = phi::EigenMatrix; template struct HardLabelCrossEntropyCPUFunctorImpl { @@ -54,17 +53,17 @@ struct HardLabelCrossEntropyCPUFunctorImpl { for (int j = 0; j < num_remain; j++) { int lbl = static_cast(label_data[i * num_remain + j]); if (lbl != ignore_index_) { - PADDLE_ENFORCE_GE(lbl, - 0, - platform::errors::OutOfRange( - "label value should >= 0 when label " - "value(%f) not equal to ignore_index(%f)", - lbl, - ignore_index_)); + PADDLE_ENFORCE_GE( + lbl, + 0, + phi::errors::OutOfRange("label value should >= 0 when label " + "value(%f) not equal to ignore_index(%f)", + lbl, + ignore_index_)); PADDLE_ENFORCE_LT( lbl, axis_dim_, - platform::errors::OutOfRange( + phi::errors::OutOfRange( "label value should less than the shape of axis dimension " "when label value(%f) not equal to ignore_index(%f), But " "received label value as %ld and shape of axis dimension " @@ -79,7 +78,7 @@ struct HardLabelCrossEntropyCPUFunctorImpl { loss_data[loss_idx] = lbl == ignore_index_ ? 0 - : -math::TolerableValue()(std::log(prob_data[index])); + : -phi::funcs::TolerableValue()(std::log(prob_data[index])); } } } @@ -112,19 +111,18 @@ void CrossEntropyFunctor::operator()( auto loss = EigenMatrix::From(*out); loss.device(*ctx.eigen_device()) = - -((lbl * in.log().unaryExpr(math::TolerableValue())) + -((lbl * in.log().unaryExpr(phi::funcs::TolerableValue())) .reshape(batch_axis_remain) .sum(Eigen::DSizes(1))); } else { HardLabelCrossEntropyCPUFunctorImpl functor_impl( out, prob, labels, ignore_index, axis_dim); - framework::VisitIntDataType(framework::TransToProtoVarType(labels->dtype()), - functor_impl); + phi::VisitDataType(labels->dtype(), functor_impl); } } template class CrossEntropyFunctor; template class CrossEntropyFunctor; -} // namespace math -} // namespace operators -} // namespace paddle + +} // namespace funcs +} // namespace phi diff --git a/paddle/fluid/operators/math/cross_entropy.cu b/paddle/phi/kernels/funcs/cross_entropy.cu similarity index 85% rename from paddle/fluid/operators/math/cross_entropy.cu rename to paddle/phi/kernels/funcs/cross_entropy.cu index 8282f2b8a24f29c610b06835d8f35a647d6c4d45..174c1c1bd934e664108d64d5a97fc4ab2a2e461a 100644 --- a/paddle/fluid/operators/math/cross_entropy.cu +++ b/paddle/phi/kernels/funcs/cross_entropy.cu @@ -1,4 +1,4 @@ -/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved. +/* Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. @@ -12,15 +12,16 @@ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ -#include "paddle/fluid/operators/math/cross_entropy.h" -#include "paddle/fluid/framework/convert_utils.h" -#include "paddle/fluid/platform/device/gpu/gpu_device_function.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" + #include "paddle/phi/backends/gpu/gpu_context.h" +#include "paddle/phi/backends/gpu/gpu_device_function.h" #include "paddle/phi/backends/gpu/gpu_primitives.h" +#include "paddle/phi/core/utils/data_type.h" #include "paddle/phi/kernels/funcs/math.h" -namespace paddle { -namespace operators { -namespace math { + +namespace phi { +namespace funcs { template __global__ void CrossEntropyKernel(T* Y, @@ -38,10 +39,9 @@ __global__ void CrossEntropyKernel(T* Y, D, ignore_index, lbl); - Y[i] = - ignore_index == lbl - ? static_cast(0) - : -math::TolerableValue()(phi::funcs::real_log(X[i * D + lbl])); + Y[i] = ignore_index == lbl ? static_cast(0) + : -phi::funcs::TolerableValue()( + phi::funcs::real_log(X[i * D + lbl])); } } @@ -56,10 +56,11 @@ __global__ void SoftCrossEntropyKernel(T* Y, int idx = blockIdx.x * class_num + tid; int end = blockIdx.x * class_num + class_num; for (; idx < end; idx += blockDim.x) { - val += math::TolerableValue()(phi::funcs::real_log(X[idx])) * label[idx]; + val += phi::funcs::TolerableValue()(phi::funcs::real_log(X[idx])) * + label[idx]; } - val = paddle::platform::reduceSum(val, tid, blockDim.x); + val = phi::backends::gpu::reduceSum(val, tid, blockDim.x); if (threadIdx.x == 0) { Y[blockIdx.x] = -val; } @@ -117,8 +118,8 @@ void CrossEntropyFunctor::operator()( const bool softLabel, const int ignore_index, const int axis_dim) { + T* loss_data = ctx.template Alloc(out); const T* prob_data = prob->data(); - T* loss_data = out->mutable_data(ctx.GetPlace()); int batch_size = prob->dims()[0]; int class_num = prob->dims()[1]; @@ -145,8 +146,7 @@ void CrossEntropyFunctor::operator()( ignore_index, kMaxBlockDim, ctx.stream()); - framework::VisitDataType(framework::TransToProtoVarType(labels->dtype()), - functor); + phi::VisitDataType(labels->dtype(), functor); } } @@ -154,6 +154,5 @@ template class CrossEntropyFunctor; template class CrossEntropyFunctor; template class CrossEntropyFunctor; -} // namespace math -} // namespace operators -} // namespace paddle +} // namespace funcs +} // namespace phi diff --git a/paddle/fluid/operators/math/cross_entropy.h b/paddle/phi/kernels/funcs/cross_entropy.h similarity index 74% rename from paddle/fluid/operators/math/cross_entropy.h rename to paddle/phi/kernels/funcs/cross_entropy.h index fba4c2ebc61c2cd05aaa8e39e67ba87a5f5ffac0..692ba5efef5b734acf16479459cfbdd6b4a91772 100644 --- a/paddle/fluid/operators/math/cross_entropy.h +++ b/paddle/phi/kernels/funcs/cross_entropy.h @@ -1,4 +1,4 @@ -/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved. +/* Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. @@ -15,14 +15,13 @@ limitations under the License. */ #pragma once #include -#include "paddle/fluid/framework/eigen.h" -#include "paddle/fluid/framework/tensor.h" -#include "paddle/fluid/platform/float16.h" +#include "paddle/phi/common/float16.h" +#include "paddle/phi/core/dense_tensor.h" #include "paddle/phi/core/hostdevice.h" +#include "paddle/phi/kernels/funcs/eigen/common.h" -namespace paddle { -namespace operators { -namespace math { +namespace phi { +namespace funcs { template struct TolerableValue { @@ -46,14 +45,15 @@ struct TolerableValue { // Also. In standard implementation of cross entropy, other // framework not has the ValueClipping. template <> -struct TolerableValue { - HOSTDEVICE platform::float16 operator()(const platform::float16& x) const { - if (platform::isfinite(x)) +struct TolerableValue { + HOSTDEVICE phi::dtype::float16 operator()( + const phi::dtype::float16& x) const { + if (phi::dtype::isfinite(x)) return x; - else if (x > static_cast(0)) - return std::numeric_limits::max(); + else if (x > static_cast(0)) + return std::numeric_limits::max(); else - return std::numeric_limits::min(); + return std::numeric_limits::min(); } }; @@ -68,6 +68,5 @@ class CrossEntropyFunctor { const int ignore_index, const int axis_dim); }; -} // namespace math -} // namespace operators -} // namespace paddle +} // namespace funcs +} // namespace phi diff --git a/paddle/phi/kernels/gpu/cross_entropy_grad_kernel.cu b/paddle/phi/kernels/gpu/cross_entropy_grad_kernel.cu index df3e4bd0cf118dbdebc88b585deb4a294e34f4c8..8618f947be45720dcae23de094210d72f025c5b6 100644 --- a/paddle/phi/kernels/gpu/cross_entropy_grad_kernel.cu +++ b/paddle/phi/kernels/gpu/cross_entropy_grad_kernel.cu @@ -22,7 +22,6 @@ limitations under the License. */ namespace cub = hipcub; #endif -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/fluid/operators/math/softmax.h" #include "paddle/fluid/platform/device/gpu/gpu_dnn.h" #include "paddle/phi/backends/gpu/gpu_device_function.h" diff --git a/paddle/phi/kernels/gpu/cross_entropy_kernel.cu b/paddle/phi/kernels/gpu/cross_entropy_kernel.cu index bee9fc801b795f97a7699579b12eef4169610584..93d5f06b665649422eec86b1a79574b08c931965 100644 --- a/paddle/phi/kernels/gpu/cross_entropy_kernel.cu +++ b/paddle/phi/kernels/gpu/cross_entropy_kernel.cu @@ -22,7 +22,6 @@ limitations under the License. */ namespace cub = hipcub; #endif -#include "paddle/fluid/operators/math/cross_entropy.h" #include "paddle/fluid/operators/math/softmax.h" #include "paddle/fluid/platform/device/gpu/gpu_dnn.h" #include "paddle/phi/backends/gpu/gpu_device_function.h" @@ -31,6 +30,7 @@ namespace cub = hipcub; #include "paddle/phi/core/tensor_utils.h" #include "paddle/phi/core/visit_type.h" #include "paddle/phi/kernels/funcs/axis_utils.h" +#include "paddle/phi/kernels/funcs/cross_entropy.h" #include "paddle/phi/kernels/funcs/for_range.h" #include "paddle/phi/kernels/funcs/math_function.h" #include "paddle/phi/kernels/gpudnn/softmax_gpudnn.h" @@ -46,7 +46,7 @@ template static __device__ __forceinline__ T Log(T x) { using AccT = typename dtype::MPTypeTrait::Type; AccT logx = std::log(static_cast(x)); - return paddle::operators::math::TolerableValue()(static_cast(logx)); + return phi::funcs::TolerableValue()(static_cast(logx)); } // Wrapper of exp function. Use exp(float32) for float16 @@ -54,7 +54,7 @@ template static __device__ __forceinline__ T Exp(T x) { using AccT = typename dtype::MPTypeTrait::Type; AccT expx = std::exp(static_cast(x)); - return paddle::operators::math::TolerableValue()(static_cast(expx)); + return phi::funcs::TolerableValue()(static_cast(expx)); } template @@ -1285,16 +1285,15 @@ void CrossEntropyWithSoftmaxCUDAKernel(const GPUContext& dev_ctx, DenseTensor softmax_out_2d(*softmax_out); softmax_out_2d.Resize({n, d}); - // math::CrossEntropyFunctor support axis is the last + // phi::funcs::CrossEntropyFunctor support axis is the last if (axis_v == -1) { - paddle::operators::math::CrossEntropyFunctor()( - dev_ctx, - &loss_2d, - &softmax_2d, - &labels_2d, - soft_label, - ignore_index, - axis_dim); + phi::funcs::CrossEntropyFunctor()(dev_ctx, + &loss_2d, + &softmax_2d, + &labels_2d, + soft_label, + ignore_index, + axis_dim); return; } @@ -1389,14 +1388,13 @@ void CrossEntropyWithSoftmaxCUDAKernel(const GPUContext& dev_ctx, loss_2d.Resize({n, 1}); paddle::operators::math::SoftmaxCUDNNFunctor()( dev_ctx, &logits_2d, &softmax_2d); - paddle::operators::math::CrossEntropyFunctor()( - dev_ctx, - &loss_2d, - &softmax_2d, - &labels_2d, - false, - ignore_index, - axis_dim); + phi::funcs::CrossEntropyFunctor()(dev_ctx, + &loss_2d, + &softmax_2d, + &labels_2d, + false, + ignore_index, + axis_dim); } else { auto* logits_data = logits.data(); auto* labels_data = label.data();