/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #include #include #include "paddle/fluid/framework/lod_tensor_array.h" #include "paddle/fluid/framework/op_registry.h" #include "paddle/fluid/framework/variable.h" namespace paddle { namespace operators { using framework::Tensor; void LodTensorArray2LodTensorVector(const framework::Scope &scope, const std::string &base_name, const std::string &lod_tensor_array_name, std::vector *res_names) { auto &inx = scope.FindVar(lod_tensor_array_name)->Get(); for (size_t i = 0; i < inx.size(); i++) { std::string var_name = base_name + std::to_string(i); framework::Variable *g_feed_value = const_cast(scope).Var(var_name); auto &feed_input = *(g_feed_value->GetMutable()); feed_input.ShareDataWith(inx[i]); res_names->push_back(var_name); } } void LodTensorVectorResizeFromLodTensorArray( const framework::Scope &scope, const std::string &base_name, const std::string &lod_tensor_array_name, std::vector *res_names) { auto &inx = scope.FindVar(lod_tensor_array_name)->Get(); for (size_t i = 0; i < inx.size(); i++) { std::string var_name = base_name + std::to_string(i); framework::Variable *g_feed_value = const_cast(scope).Var(var_name); auto &feed_input = *(g_feed_value->GetMutable()); auto dims = inx[i].dims(); feed_input.Resize(dims); res_names->push_back(var_name); } } void LodTensorArrayCreateFromLodTensorArray( const framework::Scope &scope, const std::string &input_lod_tensor_array_name, const std::string &output_lod_tensor_array_name) { auto &inx = scope.FindVar(input_lod_tensor_array_name) ->Get(); auto &grad_inx = *scope.FindVar(output_lod_tensor_array_name) ->GetMutable(); for (size_t i = 0; i < inx.size(); i++) { std::string var_name = output_lod_tensor_array_name + std::to_string(i); framework::Variable *g_feed_value = const_cast(scope).Var(var_name); auto &feed_input = *(g_feed_value->GetMutable()); grad_inx.push_back(feed_input); } } class LoDTensorArray2TensorOp : public framework::OperatorBase { public: using OperatorBase::OperatorBase; private: void RunImpl(const framework::Scope &scope, const platform::Place &place) const override { auto axis = Attr("axis"); framework::AttributeMap attrs; attrs["axis"] = axis; auto &inx = scope.FindVar(Input("X"))->Get(); auto &out = *scope.FindVar(Output("Out"))->GetMutable(); auto &out_inx = *scope.FindVar(Output("OutIndex"))->GetMutable(); const size_t n = inx.size(); PADDLE_ENFORCE_GT(n, 0, "Input tensorarray size should > 0."); std::string base_name = Inputs("X")[0]; std::vector names; // get the input tensorarray items' dim in out_inx auto out_inx_dim = out_inx.dims(); out_inx_dim[0] = inx.size(); out_inx.Resize(out_inx_dim); auto &local_scope = scope.NewScope(); std::string var_name = "out_index"; framework::Variable *tmp_index_var = local_scope.Var(var_name); auto &tmp_index_tensor = *(tmp_index_var->GetMutable()); tmp_index_tensor.Resize(out_inx_dim); int *tmp_index_data = tmp_index_tensor.mutable_data(platform::CPUPlace()); auto out_dims = inx[0].dims(); size_t out_dim_sum = 0; for (size_t index = 0; index < inx.size(); index++) { auto inx_dims = inx[index].dims(); out_dim_sum += inx_dims[axis]; tmp_index_data[index] = inx_dims[axis]; } out_inx.ShareDataWith(tmp_index_tensor); // get input array items' dims out_dims[axis] = out_dim_sum; out.Resize(out_dims); LodTensorArray2LodTensorVector(local_scope, base_name, Input("X"), &names); // Invoke concat Op auto concat_op = framework::OpRegistry::CreateOp( "concat", {{"X", names}}, {{"Out", {Output("Out")}}}, attrs); concat_op->Run(local_scope, place); } }; class LoDTensorArray2TensorOpMaker : public framework::OpProtoAndCheckerMaker { public: void Make() override { AddInput("X", "Input LoDTensorArray of tensor_array_to_tensor operator."); AddOutput("Out", "Output tensor of tensor_array_to_tensor operator."); AddOutput("OutIndex", "Output input LoDTensorArray items' dims of " "tensor_array_to_tensor operator."); AddAttr("axis", "The axis along which the input tensors will be concatenated.") .SetDefault(0); AddComment(R"DOC( tensor_array_to_tensor Operator. Concatenate the input LoDTensorArray along dimension axis to the output Tensor. Examples: Input = {[1,2], [3,4], [5,6]} axis = 0 Output = [[1,2], [3,4], [5,6]] OutputIndex = [1,1,1] )DOC"); } }; class LoDTensorArray2TensorOpInferShape : public framework::InferShapeBase { public: void operator()(framework::InferShapeContext *ctx) const override {} }; class LoDTensorArray2TensorGradInferShape : public framework::InferShapeBase { public: void operator()(framework::InferShapeContext *context) const override {} }; class LoDTensorArray2TensorGradInferVarType : public framework::VarTypeInference { public: void operator()(const framework::OpDesc &op_desc, framework::BlockDesc *block) const override { for (auto &out_var : op_desc.Output(framework::GradVarName("X"))) { block->Var(out_var)->SetType(framework::proto::VarType::LOD_TENSOR_ARRAY); } } }; class LoDTensorArray2TensorGradOp : public framework::OperatorBase { public: using OperatorBase::OperatorBase; private: void RunImpl(const framework::Scope &scope, const platform::Place &place) const override { auto axis = Attr("axis"); framework::AttributeMap attrs; attrs["axis"] = axis; auto &inx = scope.FindVar(Input("X"))->Get(); const size_t n = inx.size(); PADDLE_ENFORCE_GT(n, 0, "Input tensorarray size should > 0."); std::string base_name = Inputs("X")[0]; std::vector names; LodTensorArray2LodTensorVector(scope, base_name, Input("X"), &names); // grad auto dx_name = Output(framework::GradVarName("X")); auto dout_name = Input(framework::GradVarName("Out")); std::vector grad_names; LodTensorVectorResizeFromLodTensorArray(scope, "grad_name", Input("X"), &grad_names); auto concat_grad_op = framework::OpRegistry::CreateOp( "concat_grad", {{"X", names}, {"Out@GRAD", {dout_name}}}, {{"X@GRAD", grad_names}}, attrs); concat_grad_op->Run(scope, place); LodTensorArrayCreateFromLodTensorArray(scope, Input("X"), dx_name); auto &grad_inx = *scope.FindVar(dx_name)->GetMutable(); for (size_t i = 0; i < grad_names.size(); i++) { std::string var_name = grad_names[i]; auto &feed_input = scope.FindVar(var_name)->Get(); grad_inx[i].ShareDataWith(feed_input); } } }; } // namespace operators } // namespace paddle USE_OP(concat); namespace ops = paddle::operators; REGISTER_OPERATOR(tensor_array_to_tensor, ops::LoDTensorArray2TensorOp, ops::LoDTensorArray2TensorOpMaker, ops::LoDTensorArray2TensorOpInferShape, paddle::framework::DefaultGradOpDescMaker); REGISTER_OPERATOR(tensor_array_to_tensor_grad, ops::LoDTensorArray2TensorGradOp, ops::LoDTensorArray2TensorGradInferShape, ops::LoDTensorArray2TensorGradInferVarType);