transpose_op.cc 13.5 KB
Newer Older
1
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
X
xzl 已提交
2

L
Luo Tao 已提交
3 4 5
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
X
xzl 已提交
6

L
Luo Tao 已提交
7
    http://www.apache.org/licenses/LICENSE-2.0
X
xzl 已提交
8

L
Luo Tao 已提交
9 10 11 12 13
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
X
xzl 已提交
14

Y
Yi Wang 已提交
15
#include "paddle/fluid/operators/transpose_op.h"
16

17
#include <memory>
18
#include <string>
19
#include <vector>
X
xzl 已提交
20

21 22 23 24
#ifdef PADDLE_WITH_MKLDNN
#include "paddle/fluid/platform/mkldnn_helper.h"
#endif

X
xzl 已提交
25 26 27 28 29 30 31
namespace paddle {
namespace operators {

class TransposeOp : public framework::OperatorWithKernel {
 public:
  using framework::OperatorWithKernel::OperatorWithKernel;

32
  void InferShape(framework::InferShapeContext *ctx) const override {
33 34
    OP_INOUT_CHECK(ctx->HasInput("X"), "Input", "X", "Transpose");
    OP_INOUT_CHECK(ctx->HasOutput("Out"), "Output", "Out", "Transpose");
Q
Qiao Longfei 已提交
35 36
    auto x_dims = ctx->GetInputDim("X");
    std::vector<int> axis = ctx->Attrs().Get<std::vector<int>>("axis");
37

X
xzl 已提交
38
    size_t x_rank = x_dims.size();
X
xzl 已提交
39
    size_t axis_size = axis.size();
X
xzl 已提交
40

41 42 43
    // Note: x_rank > axis_size when fuse squeeze2 + transpose2, else x_rank ==
    // axis_size
    PADDLE_ENFORCE_GE(x_rank,
44
                      axis_size,
45 46
                      platform::errors::InvalidArgument(
                          "The input tensor's dimension "
47
                          "should be equal to or greater than the axis's size. "
48 49
                          "But received input tensor's dimension is %d, "
                          "axis's size is %d",
50 51
                          x_rank,
                          axis_size));
52 53 54

    std::vector<int> count(axis_size, 0);
    for (size_t i = 0; i < axis_size; i++) {
55 56
      PADDLE_ENFORCE_GE(axis[i],
                        0,
57 58 59
                        platform::errors::InvalidArgument(
                            "The axis should be greater than or equal to 0."
                            "But received %d of axis[%d]",
60 61
                            axis[i],
                            i));
62

63
      PADDLE_ENFORCE_EQ(
64 65
          axis[i] < static_cast<int>(axis_size) && ++count[axis[i]] == 1,
          true,
66 67 68 69 70 71 72
          platform::errors::InvalidArgument(
              "Each element of Attribute axis should "
              "be a unique value range from 0 to (dims - 1), "
              "where the dims is the axis's size, "
              "unique value means this axis value can appear only once. "
              "But received axis[%d] is %d, axis_size is %d, "
              "count[axis[%d]] is %d",
73 74 75 76 77
              i,
              axis[i],
              axis_size,
              i,
              count[axis[i]]));
X
xzl 已提交
78
    }
X
xzl 已提交
79

X
xzl 已提交
80
    framework::DDim out_dims(x_dims);
J
Jacek Czaja 已提交
81 82 83
#ifdef PADDLE_WITH_MKLDNN
    // Here we need to match dims to paddle layout
    // as we are producing non-oneDNN result
84
    if (ctx->IsRunMKLDNNKernel() && (x_dims.size() >= 3) &&
J
Jacek Czaja 已提交
85
        (paddle::platform::MKLDNNDeviceContext::tls()
86
             .get_cur_paddle_data_layout() == phi::DataLayout::kNHWC)) {
87
      auto dims = phi::vectorize<int>(x_dims);
J
Jacek Czaja 已提交
88 89 90 91 92 93
      std::rotate(dims.begin() + 1, dims.begin() + 2, dims.end());
      x_dims = x_dims.reshape(dims);
      VLOG(3)
          << "Rotating Shape in Transpose from: kMKLDNN to: kNHWC output_shape";
    }
#endif
94
    for (size_t i = 0; i < axis_size; i++) {
X
xzl 已提交
95
      out_dims[i] = x_dims[axis[i]];
X
xzl 已提交
96
    }
Q
Qiao Longfei 已提交
97
    ctx->SetOutputDim("Out", out_dims);
X
xzl 已提交
98
  }
99 100 101 102

 protected:
  framework::OpKernelType GetExpectedKernelType(
      const framework::ExecutionContext &ctx) const override {
103
    auto data_type = OperatorWithKernel::IndicateVarDataType(ctx, "X");
J
jiahongyu 已提交
104
    auto &data_format = ctx.Attr<std::string>("data_format");
105
    phi::DataLayout layout_ = phi::StringToDataLayout(data_format);
J
jiahongyu 已提交
106
    return framework::OpKernelType(data_type, ctx.GetPlace(), layout_);
107
  }
X
xzl 已提交
108 109 110 111
};

class TransposeOpMaker : public framework::OpProtoAndCheckerMaker {
 public:
Y
Yu Yang 已提交
112
  void Make() override {
113
    AddInput(
X
xzl 已提交
114
        "X",
115 116
        "(Tensor) The input tensor, tensors with rank up to 6 are supported.");
    AddOutput("Out", "(Tensor)The output tensor.");
X
xzl 已提交
117 118
    AddAttr<std::vector<int>>(
        "axis",
119 120 121
        "(vector<int>) A list of values, and the size of the list should be "
        "the same with the input tensor rank. This operator permutes the input "
        "tensor's axes according to the values given.");
122 123
    AddAttr<bool>("use_mkldnn",
                  "(bool, default false) Only used in mkldnn kernel")
124 125
        .SetDefault(false)
        .AsExtra();
126 127 128 129 130 131
    AddAttr<std::string>(
        "data_format",
        "(string, default NCHW) Only used in "
        "An optional string from: \"NHWC\", \"NCHW\". "
        "Defaults to \"NHWC\". Specify the data format of the output data, "
        "the input will be transformed automatically. ")
132 133
        .SetDefault("AnyLayout")
        .AsExtra();
134 135 136 137
    AddAttr<bool>(
        "use_quantizer",
        "(bool, default false) "
        "This parameter is no longer used. Use 'mkldnn_data_type' instead.")
138 139
        .SetDefault(false)
        .AsExtra();
140 141 142 143
    AddAttr<std::string>(
        "mkldnn_data_type",
        "(string, default \"float32\"). Data type of mkldnn kernel")
        .SetDefault("float32")
144 145
        .InEnum({"float32", "int8", "bfloat16"})
        .AsExtra();
146
    /* int8 parameters */
X
xzl 已提交
147
    AddComment(R"DOC(
148 149
Transpose Operator.

150 151
The input tensor will be permuted according to the axes given.
The behavior of this operator is similar to how `numpy.transpose` works.
Y
ying 已提交
152

153 154 155 156 157 158
- suppose the input `X` is a 2-D tensor:
    $$
    X = \begin{pmatrix}
    0 &1 &2 \\
    3 &4 &5
    \end{pmatrix}$$
W
wanghaoshuang 已提交
159

160
    the given `axes` is: $[1, 0]$, and $Y$ = transpose($X$, axis)
W
wanghaoshuang 已提交
161

162
    then the output $Y$ is:
W
wanghaoshuang 已提交
163

164 165 166 167 168 169
    $$
    Y = \begin{pmatrix}
         0 &3 \\
         1 &4  \\
         2 &5
    \end{pmatrix}$$
W
wanghaoshuang 已提交
170

171
- Given a input tensor with shape $(N, C, H, W)$ and the `axes` is
172
$[0, 2, 3, 1]$, then shape of the output tensor will be: $(N, H, W, C)$.
173

X
xzl 已提交
174 175 176 177 178 179 180 181
)DOC");
  }
};

class TransposeOpGrad : public framework::OperatorWithKernel {
 public:
  using framework::OperatorWithKernel::OperatorWithKernel;

182
  void InferShape(framework::InferShapeContext *ctx) const override {
183
    OP_INOUT_CHECK(ctx->HasInput("X"), "Input", "X", "TransposeOpGrad");
184 185 186 187
    OP_INOUT_CHECK(ctx->HasInput(framework::GradVarName("Out")),
                   "Input",
                   framework::GradVarName("Out"),
                   "TransposeOpGrad");
Q
Qiao Longfei 已提交
188 189 190 191 192
    auto x_dims = ctx->GetInputDim("X");
    ctx->SetOutputDim(framework::GradVarName("X"), x_dims);
    if (ctx->HasOutput(framework::GradVarName("X"))) {
      ctx->SetOutputDim(framework::GradVarName("X"), x_dims);
    }
X
xzl 已提交
193
  }
194 195 196 197

 protected:
  framework::OpKernelType GetExpectedKernelType(
      const framework::ExecutionContext &ctx) const override {
198 199
    auto data_type = OperatorWithKernel::IndicateVarDataType(
        ctx, framework::GradVarName("Out"));
J
jiahongyu 已提交
200
    std::string data_format = ctx.Attr<std::string>("data_format");
201
    phi::DataLayout layout_ = phi::StringToDataLayout(data_format);
J
jiahongyu 已提交
202
    return framework::OpKernelType(data_type, ctx.GetPlace(), layout_);
203
  }
X
xzl 已提交
204 205
};

206 207 208 209 210 211 212 213 214 215 216 217 218 219 220
// FIXME(zcd): transpose2 adds an intermediate output(XShape) based on
// transpose, the XShape is used to carry the shape and lod of X which
// will be used in transpose_grad, in this way, the framework can reuse
// the memory of X immediately the transpose2_op is finished.
// Considering compatibility issues, we could not fix transpose2_op
class Transpose2Op : public TransposeOp {
 public:
  Transpose2Op(const std::string &type,
               const framework::VariableNameMap &inputs,
               const framework::VariableNameMap &outputs,
               const framework::AttributeMap &attrs)
      : TransposeOp(type, inputs, outputs, attrs) {}

  void InferShape(framework::InferShapeContext *ctx) const override {
    TransposeOp::InferShape(ctx);
221
    if (!ctx->HasOutput("XShape")) return;
222 223 224 225 226 227
    const auto &in_dims = ctx->GetInputDim("X");
    std::vector<int64_t> x_shape_dim(in_dims.size() + 1);
    x_shape_dim[0] = 0;
    for (int i = 0; i < in_dims.size(); ++i) {
      x_shape_dim[i + 1] = in_dims[i];
    }
228
    ctx->SetOutputDim("XShape", phi::make_ddim(x_shape_dim));
229 230 231 232 233 234
    ctx->ShareLoD("X", /*->*/ "XShape");
  }

 protected:
  framework::OpKernelType GetExpectedKernelType(
      const framework::ExecutionContext &ctx) const override {
235 236
    framework::proto::VarType::Type data_type =
        OperatorWithKernel::IndicateVarDataType(ctx, "X");
J
jiahongyu 已提交
237
    std::string data_format = ctx.Attr<std::string>("data_format");
238
    phi::DataLayout layout_ = phi::StringToDataLayout(data_format);
J
jiahongyu 已提交
239
    return framework::OpKernelType(data_type, ctx.GetPlace(), layout_);
240 241 242
  }
};

243
class Transpose2OpMaker : public framework::OpProtoAndCheckerMaker {
244 245
 public:
  void Make() override {
246 247 248 249 250 251 252 253 254
    AddInput(
        "X",
        "(Tensor) The input tensor, tensors with rank up to 6 are supported.");
    AddOutput("Out", "(Tensor)The output tensor.");
    AddAttr<std::vector<int>>(
        "axis",
        "(vector<int>) A list of values, and the size of the list should be "
        "the same with the input tensor rank. This operator permutes the input "
        "tensor's axes according to the values given.");
255 256 257
    AddOutput("XShape", "(Tensor)The output tensor.")
        .AsIntermediate()
        .AsExtra();
258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285
    AddComment(R"DOC(
Transpose Operator.

The input tensor will be permuted according to the axes given.
The behavior of this operator is similar to how `numpy.transpose` works.

- suppose the input `X` is a 2-D tensor:
    $$
    X = \begin{pmatrix}
    0 &1 &2 \\
    3 &4 &5
    \end{pmatrix}$$

    the given `axes` is: $[1, 0]$, and $Y$ = transpose($X$, axis)

    then the output $Y$ is:

    $$
    Y = \begin{pmatrix}
         0 &3 \\
         1 &4  \\
         2 &5
    \end{pmatrix}$$

- Given a input tensor with shape $(N, C, H, W)$ and the `axes` is
$[0, 2, 3, 1]$, then shape of the output tensor will be: $(N, H, W, C)$.

)DOC");
286 287 288
  }
};

H
hong 已提交
289 290
template <typename T>
class Transpose2GradMaker : public framework::SingleGradOpMaker<T> {
291
 public:
H
hong 已提交
292
  using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
293

294
  void Apply(GradOpPtr<T> grad_op) const override {
295
    grad_op->SetType("transpose2_grad");
H
hong 已提交
296 297 298 299
    grad_op->SetInput("XShape", this->Output("XShape"));
    grad_op->SetInput(framework::GradVarName("Out"), this->OutputGrad("Out"));
    grad_op->SetOutput(framework::GradVarName("X"), this->InputGrad("X"));
    grad_op->SetAttrMap(this->Attrs());
300 301 302
  }
};

303 304 305 306 307 308 309 310 311 312 313 314 315 316
template <typename T>
class Transpose2DoubleGradMaker : public framework::SingleGradOpMaker<T> {
 public:
  using framework::SingleGradOpMaker<T>::SingleGradOpMaker;

  void Apply(GradOpPtr<T> grad_op) const override {
    grad_op->SetType("transpose2");
    grad_op->SetInput("X", this->OutputGrad(framework::GradVarName("X")));
    grad_op->SetOutput("Out", this->InputGrad(framework::GradVarName("Out")));
    grad_op->SetOutput("XShape", this->Input("XShape"));
    grad_op->SetAttrMap(this->Attrs());
  }
};

317 318 319 320 321
class Transpose2OpGrad : public framework::OperatorWithKernel {
 public:
  using framework::OperatorWithKernel::OperatorWithKernel;

  void InferShape(framework::InferShapeContext *ctx) const override {
322 323 324 325 326
    OP_INOUT_CHECK(
        ctx->HasInput("XShape"), "Input", "XShape", "Transpose2OpGrad");
    OP_INOUT_CHECK(ctx->HasInput(framework::GradVarName("Out")),
                   "Input",
                   framework::GradVarName("Out"),
327
                   "Transpose2OpGrad");
328 329
    if (ctx->HasOutput(framework::GradVarName("X"))) {
      auto xshape_dim = ctx->GetInputDim("XShape");
330
      auto x_shape_dim = phi::slice_ddim(xshape_dim, 1, xshape_dim.size());
331 332 333 334 335 336 337 338
      ctx->SetOutputDim(framework::GradVarName("X"), x_shape_dim);
      ctx->ShareLoD("XShape", framework::GradVarName("X"));
    }
  }

 protected:
  framework::OpKernelType GetExpectedKernelType(
      const framework::ExecutionContext &ctx) const override {
339 340 341
    framework::proto::VarType::Type data_type =
        OperatorWithKernel::IndicateVarDataType(ctx,
                                                framework::GradVarName("Out"));
J
jiahongyu 已提交
342
    std::string data_format = ctx.Attr<std::string>("data_format");
343
    phi::DataLayout layout_ = phi::StringToDataLayout(data_format);
J
jiahongyu 已提交
344
    return framework::OpKernelType(data_type, ctx.GetPlace(), layout_);
345 346 347
  }
};

H
hong 已提交
348 349 350 351 352 353 354 355
class TransposeGradInferVarType : public framework::VarTypeInference {
 public:
  void operator()(framework::InferVarTypeContext *ctx) const override {
    ctx->SyncTypeAndDataType(framework::GradVarName("Out"),
                             framework::GradVarName("X"));
  }
};

X
xzl 已提交
356 357 358 359
}  // namespace operators
}  // namespace paddle

namespace ops = paddle::operators;
H
hong 已提交
360
REGISTER_OPERATOR(
361 362 363
    transpose,
    ops::TransposeOp,
    ops::TransposeOpMaker,
H
hong 已提交
364 365
    paddle::framework::DefaultGradOpMaker<paddle::framework::OpDesc, true>,
    paddle::framework::DefaultGradOpMaker<paddle::imperative::OpBase, true>);
366 367
REGISTER_OPERATOR(transpose_grad,
                  ops::TransposeOpGrad,
H
hong 已提交
368
                  ops::TransposeGradInferVarType);
369

370 371 372
REGISTER_OPERATOR(transpose2,
                  ops::Transpose2Op,
                  ops::Transpose2OpMaker,
H
hong 已提交
373 374
                  ops::Transpose2GradMaker<paddle::framework::OpDesc>,
                  ops::Transpose2GradMaker<paddle::imperative::OpBase>);
375 376
REGISTER_OPERATOR(transpose2_grad,
                  ops::Transpose2OpGrad,
H
hong 已提交
377
                  ops::TransposeGradInferVarType,
378 379
                  ops::Transpose2DoubleGradMaker<paddle::framework::OpDesc>,
                  ops::Transpose2DoubleGradMaker<paddle::imperative::OpBase>);