sgd_op.cc 5.5 KB
Newer Older
1
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
Q
Qiao Longfei 已提交
2 3 4 5 6 7 8 9 10 11 12 13 14

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

15
#include <string>
16 17 18

#include "paddle/fluid/operators/optimizers/sgd_op.h"

Q
Qiao Longfei 已提交
19 20 21
namespace paddle {
namespace operators {

D
dongzhihong 已提交
22
class SGDOp : public framework::OperatorWithKernel {
Y
Yu Yang 已提交
23 24 25
 public:
  using framework::OperatorWithKernel::OperatorWithKernel;

C
chengduo 已提交
26
  void InferShape(framework::InferShapeContext *ctx) const override {
C
Chengmo 已提交
27 28 29 30 31 32 33 34 35 36 37 38
    PADDLE_ENFORCE_EQ(ctx->HasInput("Param"), true,
                      platform::errors::NotFound(
                          "Input(Param) of SGDOp should not be null."));
    PADDLE_ENFORCE_EQ(
        ctx->HasInput("Grad"), true,
        platform::errors::NotFound("Input(Grad) of SGDOp should not be null."));
    PADDLE_ENFORCE_EQ(ctx->HasInput("LearningRate"), true,
                      platform::errors::NotFound(
                          "Input(LearningRate) of SGDOp should not be null."));
    PADDLE_ENFORCE_EQ(ctx->HasOutput("ParamOut"), true,
                      platform::errors::NotFound(
                          "Output(ParamOut) of SGDOp should not be null."));
Q
Qiao Longfei 已提交
39

40
    auto lr_dims = ctx->GetInputDim("LearningRate");
41
    PADDLE_ENFORCE_NE(framework::product(lr_dims), 0,
C
Chengmo 已提交
42 43 44 45 46
                      platform::errors::NotFound(
                          "Maybe the Input variable LearningRate has not "
                          "been initialized. You may need to confirm "
                          "if you put exe.run(startup_program) "
                          "after optimizer.minimize function."));
47
    PADDLE_ENFORCE_EQ(framework::product(lr_dims), 1,
C
Chengmo 已提交
48 49 50 51
                      platform::errors::InvalidArgument(
                          "Learning rate should have 1 element. But received "
                          "LearningRate dims [%s]",
                          framework::product(lr_dims)));
52
    auto param_dim = ctx->GetInputDim("Param");
53 54 55 56 57 58
    if (ctx->GetInputsVarType("Grad")[0] ==
        framework::proto::VarType::LOD_TENSOR) {
      PADDLE_ENFORCE_EQ(
          param_dim, ctx->GetInputDim("Grad"),
          platform::errors::InvalidArgument(
              "SGD Operator's input Param and Grad dimensions do not match. "
59 60 61
              "The Param %s shape is [%s], but the Grad %s shape is [%s].",
              ctx->Inputs("Param")[0], param_dim, ctx->Inputs("Grad")[0],
              ctx->GetInputDim("Grad")));
62
    }
63
    ctx->SetOutputDim("ParamOut", param_dim);
Q
Qiao Longfei 已提交
64
  }
65 66 67

 protected:
  framework::OpKernelType GetExpectedKernelType(
C
chengduo 已提交
68
      const framework::ExecutionContext &ctx) const override {
69
    auto data_type = OperatorWithKernel::IndicateVarDataType(ctx, "Param");
Q
qiaolongfei 已提交
70
    return framework::OpKernelType(data_type, ctx.device_context());
71
  }
72 73 74 75 76 77 78 79 80 81 82

  framework::OpKernelType GetKernelTypeForVar(
      const std::string &var_name, const framework::Tensor &tensor,
      const framework::OpKernelType &expected_kernel_type) const {
    if (var_name == "LearningRate") {
      return framework::OpKernelType(tensor.type(), tensor.place(),
                                     tensor.layout());
    }
    return framework::OpKernelType(expected_kernel_type.data_type_,
                                   tensor.place(), tensor.layout());
  }
Q
Qiao Longfei 已提交
83 84
};

Y
Yancey1989 已提交
85 86
class SGDOpInferVarType : public framework::VarTypeInference {
 public:
M
minqiyang 已提交
87
  void operator()(framework::InferVarTypeContext *ctx) const override {
88 89 90 91 92 93 94
    auto in_var_type = ctx->GetInputType("Param");
    PADDLE_ENFORCE_EQ(in_var_type == framework::proto::VarType::SELECTED_ROWS ||
                          in_var_type == framework::proto::VarType::LOD_TENSOR,
                      true, platform::errors::InvalidArgument(
                                "The input Var's type should be LoDtensor or "
                                "SelectedRows, but the received type is %s",
                                in_var_type));
C
chengduo 已提交
95

96
    ctx->SetOutputType("ParamOut", in_var_type, framework::ALL_ELEMENTS);
Y
Yancey1989 已提交
97 98 99
  }
};

D
dongzhihong 已提交
100
class SGDOpMaker : public framework::OpProtoAndCheckerMaker {
101
 public:
Y
Yu Yang 已提交
102
  void Make() override {
103
    AddInput("Param", "(Tensor or SelectedRows) Input parameter");
104
    AddInput("LearningRate", "(Tensor) Learning rate of SGD");
105 106 107
    AddInput("Grad", "(Tensor or SelectedRows) Input gradient");
    AddOutput("ParamOut",
              "(Tensor or SelectedRows, same with Param) "
108
              "Output parameter, should share the same memory with Param");
Q
Qiao Longfei 已提交
109 110
    AddComment(R"DOC(

111
SGD operator
Q
Qiao Longfei 已提交
112

113 114
This operator implements one step of the stochastic gradient descent algorithm.

115
$$param\_out = param - learning\_rate * grad$$
Q
Qiao Longfei 已提交
116 117 118 119

)DOC");
  }
};
Q
qijun 已提交
120

Q
Qiao Longfei 已提交
121 122 123
}  // namespace operators
}  // namespace paddle

D
dongzhihong 已提交
124
namespace ops = paddle::operators;
H
hong 已提交
125 126 127 128 129
REGISTER_OPERATOR(
    sgd, ops::SGDOp, ops::SGDOpMaker,
    paddle::framework::EmptyGradOpMaker<paddle::framework::OpDesc>,
    paddle::framework::EmptyGradOpMaker<paddle::imperative::OpBase>,
    ops::SGDOpInferVarType);
130 131
REGISTER_OP_CPU_KERNEL(
    sgd, ops::SGDOpKernel<paddle::platform::CPUDeviceContext, float>,
132 133
    ops::SGDOpKernel<paddle::platform::CPUDeviceContext,
                     paddle::platform::bfloat16>,
134
    ops::SGDOpKernel<paddle::platform::CPUDeviceContext, double>);