utils.h 3.0 KB
Newer Older
J
Jiabin Yang 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21
// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#pragma once
#include <string>
#include <vector>
#include "paddle/fluid/framework/op_proto_maker.h"
#include "paddle/phi/common/data_type.h"
#include "paddle/phi/common/int_array.h"
#include "paddle/phi/common/place.h"
22 23
#include "paddle/phi/core/ddim.h"
using IntArray = paddle::experimental::IntArray;
J
Jiabin Yang 已提交
24 25 26 27 28 29 30 31 32 33 34 35
namespace paddle {
namespace prim {
// We put some api like utils here
template <typename T>
paddle::experimental::Tensor empty(const paddle::experimental::IntArray& shape,
                                   paddle::experimental::DataType dype,
                                   const paddle::Place& place);

template <typename T>
paddle::experimental::Tensor empty_like(const paddle::experimental::Tensor& x,
                                        paddle::experimental::DataType dtype,
                                        const paddle::Place& place);
36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76
template <typename T>
void by_pass(const paddle::experimental::Tensor& x,
             paddle::experimental::Tensor* out);
// These method don't need to be specified
static phi::DDim get_reduce_dims(const phi::DDim& x_dims,
                                 const phi::DDim& y_dims) {
  std::vector<int64_t> result;
  PADDLE_ENFORCE_GE(phi::product(x_dims),
                    phi::product(y_dims),
                    phi::errors::InvalidArgument(
                        "Only x_dims >= y_dims is accepted for "
                        "get_reduce_dims, but we got x_dims: %s, y_dims: %s",
                        x_dims,
                        y_dims));
  int bat = x_dims.size() - y_dims.size();
  for (int i = 0; i < bat; ++i) {
    result.push_back(i);
  }
  for (int i = 0; i < y_dims.size(); ++i) {
    if (y_dims[i] == 1) {
      result.push_back(i + bat);
    } else {
      PADDLE_ENFORCE_EQ(
          y_dims[i],
          x_dims[i + bat],
          platform::errors::InvalidArgument(
              "ReduceDims dimension mismatch. Operands could "
              "not be broadcast together with the shape of x_dims = [%s] and "
              "the shape of y_dims = [%s]. Received [%d] in X is not equal to "
              "[%d] in Y at i:%d.",
              x_dims,
              y_dims,
              x_dims[i + bat],
              y_dims[i],
              i));
    }
  }
  auto res_dims = phi::make_ddim(result);
  VLOG(4) << "Reduce Dims is: " << res_dims;
  return res_dims;
}
J
Jiabin Yang 已提交
77 78
}  // namespace prim
}  // namespace paddle