// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/fluid/operators/optimizers/merged_momentum_op.h" namespace paddle { namespace operators { class MergedMomentumOp : public framework::OperatorWithKernel { public: using framework::OperatorWithKernel::OperatorWithKernel; protected: void InferShape(framework::InferShapeContext *ctx) const override {} framework::OpKernelType GetExpectedKernelType( const framework::ExecutionContext &ctx) const override { auto param_dtype = framework::OperatorWithKernel::IndicateVarDataType(ctx, "Param"); return framework::OpKernelType(param_dtype, ctx.GetPlace()); } }; class MergedMomentumOpMaker : public framework::OpProtoAndCheckerMaker { public: void Make() override { AddInput("Param", "(Tensor, default Tensor) " "Input parameter that has to be updated") .AsDuplicable(); AddInput("Grad", "(Tensor, default Tensor) " "Input gradient of the parameter") .AsDuplicable(); AddInput("Velocity", "(Tensor, default Tensor) " "Input velocity (corresponding to the parameter) " "that has to be updated") .AsDuplicable(); AddInput("LearningRate", "(Tensor, default Tensor) " "Input learning rate"); AddInput("MasterParam", "FP32 master weight for AMP.") .AsDispensable() .AsDuplicable(); AddOutput("ParamOut", "(Tensor) This output is updated parameter. " "It shared memory with Input(Param).") .AsDuplicable(); AddOutput("VelocityOut", "(Tensor) This output is updated velocity. " "It shared memory with Input(Velocity).") .AsDuplicable(); AddOutput("MasterParamOut", "The updated FP32 master weight for AMP. " "It shared memory with Input(MasterParam).") .AsDispensable() .AsDuplicable(); AddAttr("mu", "(float) Momentum coefficient"); AddAttr("multi_precision", "(bool, default false) " "Whether to use multi-precision during weight updating.") .SetDefault(false); AddAttr( "rescale_grad", "(float, default 1.0) Multiply the gradient with `rescale_grad`" "before updating. Often choose to be `1.0/batch_size`.") .SetDefault(1.0f); AddComment(R"DOC(Merged Momentum Optimizer.)DOC"); } }; } // namespace operators } // namespace paddle namespace ops = paddle::operators; namespace plat = paddle::platform; REGISTER_OP_WITHOUT_GRADIENT(merged_momentum, ops::MergedMomentumOp, ops::MergedMomentumOpMaker); REGISTER_OP_CPU_KERNEL( merged_momentum, ops::MergedMomentumOpKernel, ops::MergedMomentumOpKernel);