// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "paddle/fluid/framework/details/fuse_adam_op_pass.h" #include #include "paddle/fluid/framework/ir/graph_helper.h" #include "paddle/fluid/framework/op_registry.h" namespace paddle { namespace framework { namespace details { const std::string FuseAdamOpPass::GetOpType() const { return "adam"; } const std::vector FuseAdamOpPass::GetAuxiliaryVarNames() const { return {"Moment1", "Moment2", "Beta1Pow", "Beta2Pow"}; } void FuseAdamOpPass::FuseOptimizerOps( const std::unordered_map> &aux_var_set, const std::unordered_map &fused_vars_name, const std::vector &adam_ops, ir::Graph *graph) const { FuseAdamOps(aux_var_set, fused_vars_name, adam_ops, graph); FuseScaleOps(aux_var_set.at("Beta1Pow"), fused_vars_name.at("Beta1Pow"), adam_ops, graph); FuseScaleOps(aux_var_set.at("Beta2Pow"), fused_vars_name.at("Beta2Pow"), adam_ops, graph); } void FuseAdamOpPass::FuseAdamOps( const std::unordered_map> &vars_set, const std::unordered_map &fused_vars_name, const std::vector &adam_ops, ir::Graph *graph) const { PADDLE_ENFORCE_GT(adam_ops.size(), static_cast(0)); // Check attributions // NOTE: If new attribution is added, the following code maybe need change. int op_role = boost::get( adam_ops[0]->Op()->GetAttr(OpProtoAndCheckerMaker::OpRoleAttrName())); float beta1 = boost::get(adam_ops[0]->Op()->GetAttr("beta1")); float beta2 = boost::get(adam_ops[0]->Op()->GetAttr("beta2")); float epsilon = boost::get(adam_ops[0]->Op()->GetAttr("epsilon")); bool lazy_mode = boost::get(adam_ops[0]->Op()->GetAttr("lazy_mode")); int64_t min_row_size_to_use_multithread = boost::get( adam_ops[0]->Op()->GetAttr("min_row_size_to_use_multithread")); for (auto &adam_op : adam_ops) { PADDLE_ENFORCE_EQ(beta1, boost::get(adam_op->Op()->GetAttr("beta1"))); PADDLE_ENFORCE_EQ(beta2, boost::get(adam_op->Op()->GetAttr("beta2"))); PADDLE_ENFORCE_EQ(epsilon, boost::get(adam_op->Op()->GetAttr("epsilon"))); PADDLE_ENFORCE_EQ(lazy_mode, boost::get(adam_op->Op()->GetAttr("lazy_mode"))); PADDLE_ENFORCE_EQ(min_row_size_to_use_multithread, boost::get(adam_op->Op()->GetAttr( "min_row_size_to_use_multithread"))); PADDLE_ENFORCE_EQ(op_role, boost::get(adam_op->Op()->GetAttr( OpProtoAndCheckerMaker::OpRoleAttrName()))); } // NOTE: fused_var is only exist in scope, so the graph doesn't have fused_var // node. VLOG(10) << "Insert adam to graph "; OpDesc adam_desc(adam_ops[0]->Op()->Block()); adam_desc.SetType("adam"); adam_desc.SetInput(kParam, {fused_vars_name.at(kParam)}); adam_desc.SetInput(kGrad, {fused_vars_name.at(kGrad)}); adam_desc.SetInput("Moment1", {fused_vars_name.at("Moment1")}); adam_desc.SetInput("Moment2", {fused_vars_name.at("Moment2")}); // TODO(zcd): The LearningRate, Beta1Pow, Beta2Pow should be equal. adam_desc.SetInput(kLearningRate, adam_ops[0]->Op()->Input(kLearningRate)); adam_desc.SetInput("Beta1Pow", adam_ops[0]->Op()->Input("Beta1Pow")); adam_desc.SetInput("Beta2Pow", adam_ops[0]->Op()->Input("Beta2Pow")); adam_desc.SetOutput("ParamOut", {fused_vars_name.at(kParam)}); adam_desc.SetOutput("Moment1Out", {fused_vars_name.at("Moment1")}); adam_desc.SetOutput("Moment2Out", {fused_vars_name.at("Moment2")}); adam_desc.SetAttr("beta1", beta1); adam_desc.SetAttr("beta2", beta2); adam_desc.SetAttr("epsilon", epsilon); adam_desc.SetAttr("lazy_mode", lazy_mode); adam_desc.SetAttr("min_row_size_to_use_multithread", min_row_size_to_use_multithread); adam_desc.SetAttr(OpProtoAndCheckerMaker::OpRoleAttrName(), op_role); auto adam_node = graph->CreateOpNode(&adam_desc); InserInputAndOutputForOptOps(adam_ops, adam_node); } void FuseAdamOpPass::FuseScaleOps(const std::vector &beta_name, const std::string &fused_var_name, const std::vector &adam_ops, ir::Graph *graph) const { PADDLE_ENFORCE_EQ(beta_name.size(), adam_ops.size()); const std::string scale_op_name = "scale"; // Get the scale_ops of dealing the adam's beta var. std::vector scale_ops; scale_ops.reserve(beta_name.size()); for (size_t i = 0; i < adam_ops.size(); ++i) { auto &beta_1_pow_name = beta_name[i]; auto beta_pow_iter = std::find_if( adam_ops[i]->inputs.begin(), adam_ops[i]->inputs.end(), [&beta_name, &beta_1_pow_name](ir::Node *var_node) -> bool { return var_node->Var() && var_node->Var()->Name() == beta_1_pow_name; }); PADDLE_ENFORCE(beta_pow_iter != adam_ops[i]->inputs.end()); auto beta_pow_node = *beta_pow_iter; auto scale_op_iter = std::find_if( beta_pow_node->outputs.begin(), beta_pow_node->outputs.end(), [&scale_op_name](ir::Node *op_node) -> bool { return op_node->Op() && op_node->Op()->Type() == scale_op_name; }); PADDLE_ENFORCE(scale_op_iter != beta_pow_node->outputs.end()); scale_ops.emplace_back(*scale_op_iter); } PADDLE_ENFORCE_EQ(scale_ops.size(), beta_name.size()); // Check attributions // NOTE: If new attribution is added, the following code maybe need change. int op_role = boost::get( scale_ops[0]->Op()->GetAttr(OpProtoAndCheckerMaker::OpRoleAttrName())); float scale = boost::get(scale_ops[0]->Op()->GetAttr("scale")); float bias = boost::get(scale_ops[0]->Op()->GetAttr("bias")); bool bias_after_scale = boost::get(scale_ops[0]->Op()->GetAttr("bias_after_scale")); for (auto &scale_op : scale_ops) { PADDLE_ENFORCE_EQ(scale, boost::get(scale_op->Op()->GetAttr("scale"))); PADDLE_ENFORCE_EQ(bias, boost::get(scale_op->Op()->GetAttr("bias"))); PADDLE_ENFORCE_EQ( bias_after_scale, boost::get(scale_op->Op()->GetAttr("bias_after_scale"))); PADDLE_ENFORCE_EQ(op_role, boost::get(scale_op->Op()->GetAttr( OpProtoAndCheckerMaker::OpRoleAttrName()))); } // NOTE: fused_var is only exist in scope, so the graph doesn't have fused_var // node. VLOG(10) << "Insert fused scale to graph."; OpDesc scale_desc(scale_ops[0]->Op()->Block()); scale_desc.SetType("scale"); scale_desc.SetInput("X", {fused_var_name}); scale_desc.SetOutput("Out", {fused_var_name}); scale_desc.SetAttr("scale", scale); scale_desc.SetAttr("bias", bias); scale_desc.SetAttr("bias_after_scale", bias_after_scale); scale_desc.SetAttr(OpProtoAndCheckerMaker::OpRoleAttrName(), op_role); auto scale_node = graph->CreateOpNode(&scale_desc); for (auto scale_op : scale_ops) { // set inputs scale_node->inputs.insert(scale_node->inputs.begin(), scale_op->inputs.begin(), scale_op->inputs.end()); for (auto &input : scale_op->inputs) { std::replace(input->outputs.begin(), input->outputs.end(), scale_op, scale_node); } // set outputs scale_node->outputs.insert(scale_node->outputs.begin(), scale_op->outputs.begin(), scale_op->outputs.end()); for (auto &output : scale_op->outputs) { std::replace(output->inputs.begin(), output->inputs.end(), scale_op, scale_node); } } // Delete scale_ops for (auto &scale_op : scale_ops) { graph->RemoveNode(scale_op); } } } // namespace details } // namespace framework } // namespace paddle REGISTER_PASS(fuse_adam_op_pass, paddle::framework::details::FuseAdamOpPass) .RequirePassAttr(paddle::framework::details::kPlaces) .RequirePassAttr(paddle::framework::details::kLocalScopes);