creation.cu 3.7 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#include "paddle/pten/kernels/cuda/creation.h"

#include "paddle/pten/core/kernel_registry.h"
#include "paddle/pten/kernels/functions/eigen/fill.h"

namespace pten {

template <typename T>
void FillAnyLike(const CUDAContext& dev_ctx,
                 const Scalar& val,
                 DenseTensor* out) {
26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50
  auto value = val.to<float>();
  using CommonType = typename std::common_type<
      float,
      typename std::conditional<
          std::is_same<T, paddle::platform::float16>::value,
          float,
          T>::type>::type;

  auto common_type_value = static_cast<CommonType>(value);

  PADDLE_ENFORCE_EQ(
      (common_type_value >=
       static_cast<CommonType>(std::numeric_limits<T>::lowest())) &&
          (common_type_value <=
           static_cast<CommonType>(std::numeric_limits<T>::max())),
      true,
      paddle::platform::errors::InvalidArgument(
          "The filled value is out of range for target type, "
          "current kernel type is %s, the range should between %f "
          "and %f, but now value is %f.",
          typeid(T).name(),
          static_cast<CommonType>(std::numeric_limits<T>::lowest()),
          static_cast<CommonType>(std::numeric_limits<T>::max()),
          static_cast<float>(value)));

51 52 53
  eigen::fill<CUDAContext, T>(dev_ctx, out, val.to<float>());
}

54 55 56 57 58 59 60
template <typename T>
void FillConstant(const CUDAContext& dev_ctx,
                  const Scalar& val,
                  DenseTensor* out) {
  eigen::fill<CUDAContext, T>(dev_ctx, out, val.to<T>());
}

61 62 63 64 65 66 67 68 69
template <typename T>
void FillConstantDynamicShape(const CUDAContext& dev_ctx,
                              const ScalarArray& shape,
                              const Scalar& val,
                              DenseTensor* out) {
  out->Resize(paddle::framework::make_ddim(shape.GetData()));
  eigen::fill<CUDAContext, T>(dev_ctx, out, val.to<T>());
}

70 71 72 73 74 75 76 77 78 79 80 81 82 83
}  // namespace pten

PT_REGISTER_MODULE(CreationCUDA);

PT_REGISTER_KERNEL("fill_any_like",
                   CUDA,
                   ANY,
                   pten::FillAnyLike,
                   float,
                   double,
                   int,
                   int64_t,
                   bool,
                   paddle::platform::float16) {}
84 85 86 87 88 89 90 91 92 93 94 95 96 97 98

PT_REGISTER_KERNEL("fill_constant.scalar",
                   CUDA,
                   ANY,
                   pten::FillConstant,
                   float,
                   double,
                   uint8_t,
                   int16_t,
                   int,
                   int64_t,
                   bool,
                   paddle::platform::float16,
                   paddle::platform::complex<float>,
                   paddle::platform::complex<double>) {}
99 100 101 102 103 104 105 106 107 108 109 110 111 112 113

PT_REGISTER_KERNEL("fill_constant",
                   CUDA,
                   ANY,
                   pten::FillConstantDynamicShape,
                   float,
                   double,
                   uint8_t,
                   int16_t,
                   int,
                   int64_t,
                   bool,
                   paddle::platform::float16,
                   paddle::platform::complex<float>,
                   paddle::platform::complex<double>) {}