elementwise_ops.cc 5.0 KB
Newer Older
Y
Yan Chunwei 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14
// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

15
#include "lite/backends/npu/builder.h"
Z
zhupengyang 已提交
16
#include "lite/kernels/npu/bridges/registry.h"
Y
Yan Chunwei 已提交
17 18 19

namespace paddle {
namespace lite {
Z
zhupengyang 已提交
20
namespace kernels {
Y
Yan Chunwei 已提交
21
namespace npu {
Z
zhupengyang 已提交
22
namespace bridges {
Y
Yan Chunwei 已提交
23

24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47
std::vector<int64_t> CvtYShape(const Tensor& x, Tensor* y, int axis) {
  auto x_dims = x.dims();
  CHECK_EQ(x_dims.size(), 4UL) << "[NPU] only support 4-dimension x";
  auto y_dims = y->dims();
  CHECK_GE(x_dims.size(), y_dims.size());

  if (axis < 0) {
    axis += x_dims.size();
  }

  std::vector<int64_t> y_new_shape(y_dims.Vectorize());
  if (y_new_shape.size() == 4UL) {
    return y_new_shape;
  }
  for (int i = 0; i < axis; i++) {
    y_new_shape.insert(y_new_shape.begin(), 1);
  }
  while (y_new_shape.size() < 4) {
    y_new_shape.push_back(1);
  }
  CHECK_EQ(y_new_shape.size(), 4UL);
  return y_new_shape;
}

Y
Yan Chunwei 已提交
48 49 50
node_map_type ElementwiseConverter(
    const std::shared_ptr<lite::OpLite> elementwise_op,
    const node_map_type& inputs_map) {
51 52 53
  auto scope = elementwise_op->scope();
  auto op_info = elementwise_op->op_info();
  auto op_type = op_info->Type();
54
  auto unique_op_type = lite::npu::UniqueName(op_type);
55
  LOG(INFO) << "[NPU] Converting " + op_type + "...";
56

Y
Yan Chunwei 已提交
57 58 59
  auto x_var_name = op_info->Input("X").front();
  auto y_var_name = op_info->Input("Y").front();
  CHECK(inputs_map.find(x_var_name) != inputs_map.end());
60
  auto axis = op_info->GetAttr<int>("axis");
Y
Yan Chunwei 已提交
61

62 63 64
  std::shared_ptr<ge::Operator> elementwise_node = nullptr;
  std::shared_ptr<ge::Operator> x_node = inputs_map.at(x_var_name);
  std::shared_ptr<ge::Operator> y_node = nullptr;
Y
Yan Chunwei 已提交
65
  if (inputs_map.find(y_var_name) != inputs_map.end()) {
66
    y_node = inputs_map.at(y_var_name);
Y
Yan Chunwei 已提交
67
  } else {
68
    auto y_const_node = std::make_shared<ge::op::Const>(y_var_name);
69 70 71 72
    auto x = scope->FindTensor(x_var_name);
    auto y = scope->FindMutableTensor(y_var_name);
    auto y_new_shape = CvtYShape(*x, y, axis);
    y_const_node->set_attr_value(lite::npu::CvtTensor(y, y_new_shape));
73
    y_node = y_const_node;
Y
Yan Chunwei 已提交
74
  }
75 76
  lite::npu::OpList::Global().add(x_node);
  lite::npu::OpList::Global().add(y_node);
Y
Yan Chunwei 已提交
77

78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101
  if (op_type == "elementwise_add" ||
      op_type == "fusion_elementwise_add_activation") {
    auto elt_node = std::make_shared<ge::op::Add>(unique_op_type);
    elt_node->set_input_x1(*x_node);
    elt_node->set_input_x2(*y_node);
    elementwise_node = elt_node;
  } else if (op_type == "elementwise_sub") {
    auto elt_node = std::make_shared<ge::op::Sub>(unique_op_type);
    elt_node->set_input_x1(*x_node);
    elt_node->set_input_x2(*y_node);
    elementwise_node = elt_node;
  } else if (op_type == "elementwise_mul") {
    auto elt_node = std::make_shared<ge::op::Mul>(unique_op_type);
    elt_node->set_input_x(*x_node);
    elt_node->set_input_y(*y_node);
    elementwise_node = elt_node;
  } else if (op_type == "elementwise_div") {
    auto elt_node = std::make_shared<ge::op::RealDiv>(unique_op_type);
    elt_node->set_input_x1(*x_node);
    elt_node->set_input_x2(*y_node);
    elementwise_node = elt_node;
  } else {
    LOG(FATAL) << "unsupported op type: " << op_type;
  }
Y
Yan Chunwei 已提交
102

103
  lite::npu::OpList::Global().add(elementwise_node);
Y
Yan Chunwei 已提交
104 105

  node_map_type outputs_map;
106 107 108 109 110 111 112 113 114 115 116 117 118
  if (op_type == "fusion_elementwise_add_activation") {
    auto act_type = op_info->GetAttr<std::string>("act_type");
    auto act_node =
        std::make_shared<ge::op::Activation>(unique_op_type + "/act");
    act_node->set_input_x(*elementwise_node);
    // TODO(hong19860320) set the coef value for act Ops, such as leaky_relu,
    // clipped_relu etc.
    act_node->set_attr_mode(lite::npu::CvtActMode(act_type));
    lite::npu::OpList::Global().add(act_node);
    outputs_map[op_info->Output("Out").front()] = act_node;
  } else {
    outputs_map[op_info->Output("Out").front()] = elementwise_node;
  }
Y
Yan Chunwei 已提交
119 120 121
  return outputs_map;
}

Z
zhupengyang 已提交
122
}  // namespace bridges
Y
Yan Chunwei 已提交
123
}  // namespace npu
Z
zhupengyang 已提交
124
}  // namespace kernels
Y
Yan Chunwei 已提交
125 126 127 128
}  // namespace lite
}  // namespace paddle

REGISTER_NPU_BRIDGE(elementwise_add,
Z
zhupengyang 已提交
129
                    paddle::lite::kernels::npu::bridges::ElementwiseConverter);
130 131
REGISTER_NPU_BRIDGE(fusion_elementwise_add_activation,
                    paddle::lite::kernels::npu::bridges::ElementwiseConverter);
132 133 134 135 136 137
REGISTER_NPU_BRIDGE(elementwise_sub,
                    paddle::lite::kernels::npu::bridges::ElementwiseConverter);
REGISTER_NPU_BRIDGE(elementwise_mul,
                    paddle::lite::kernels::npu::bridges::ElementwiseConverter);
REGISTER_NPU_BRIDGE(elementwise_div,
                    paddle::lite::kernels::npu::bridges::ElementwiseConverter);