read_op.cc 5.2 KB
Newer Older
F
fengjiayi 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14
//   Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

Y
Yi Wang 已提交
15 16
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/reader.h"
Y
Yu Yang 已提交
17
#include "paddle/fluid/operators/detail/safe_ref.h"
W
Wu Yi 已提交
18
#include "paddle/fluid/platform/profiler.h"
F
fengjiayi 已提交
19 20 21 22 23 24 25 26 27 28 29

namespace paddle {
namespace operators {

class ReadInferShape : public framework::InferShapeBase {
 public:
  void operator()(framework::InferShapeContext* ctx) const override {
    PADDLE_ENFORCE(ctx->HasInput("Reader"),
                   "The ReadOp must take a reader as input.");
    PADDLE_ENFORCE(ctx->HasOutputs("Out"),
                   "The ReadOp should be assigned with output.");
F
fengjiayi 已提交
30
    std::vector<framework::DDim> reader_dims = ctx->GetReaderDims("Reader");
F
fengjiayi 已提交
31 32 33 34 35
    std::vector<std::string> out_names = ctx->Outputs("Out");
    PADDLE_ENFORCE_EQ(
        reader_dims.size(), out_names.size(),
        "The reader's dim number doesn't match the output number.");
    ctx->SetOutputsDim("Out", reader_dims);
S
sneaxiy 已提交
36 37 38 39 40 41 42 43 44 45 46 47 48
    if (!ctx->IsRuntime()) {
      auto in_desc =
          boost::get<framework::VarDesc*>(ctx->GetInputVarPtrs("Reader")[0]);
      auto in_lod_levels = in_desc->GetLoDLevels();
      auto out_var_ptrs = ctx->GetOutputVarPtrs("Out");
      PADDLE_ENFORCE_EQ(in_lod_levels.size(), out_var_ptrs.size(),
                        "LoDLevels of Input(Reader) must be the same as the "
                        "number of Outputs(Out).");
      for (size_t i = 0; i < out_var_ptrs.size(); ++i) {
        auto* out_desc = boost::get<framework::VarDesc*>(out_var_ptrs[i]);
        out_desc->SetLoDLevel(in_lod_levels[i]);
      }
    }
F
fengjiayi 已提交
49 50 51 52 53 54 55 56 57
  }
};

class ReadInferVarType : public framework::VarTypeInference {
 public:
  void operator()(const framework::OpDesc& op_desc,
                  framework::BlockDesc* block) const override {
    std::string reader_name = op_desc.Input("Reader")[0];
    std::vector<std::string> out_names = op_desc.Output("Out");
F
fengjiayi 已提交
58 59
    framework::VarDesc* reader = block->FindVarRecursive(reader_name);
    auto dtypes = reader->GetDataTypes();
F
fengjiayi 已提交
60 61
    PADDLE_ENFORCE_EQ(dtypes.size(), out_names.size());
    for (size_t i = 0; i < dtypes.size(); ++i) {
F
fengjiayi 已提交
62
      framework::VarDesc& out = block->FindRecursiveOrCreateVar(out_names[i]);
63
      out.SetType(framework::proto::VarType::LOD_TENSOR);
F
fengjiayi 已提交
64 65 66 67 68 69 70 71
      out.SetDataType(dtypes[i]);
    }
  }
};

class ReadOp : public framework::OperatorBase {
 public:
  using framework::OperatorBase::OperatorBase;
72 73 74 75

 private:
  void RunImpl(const framework::Scope& scope,
               const platform::Place& dev_place) const override {
F
fengjiayi 已提交
76
    framework::ReaderHolder* reader =
Y
Yu Yang 已提交
77 78 79
        detail::Ref(scope.FindVar(Input("Reader")),
                    "Cannot find reader variable %s", Input("Reader"))
            .GetMutable<framework::ReaderHolder>();
F
fengjiayi 已提交
80 81
    std::vector<std::string> out_arg_names = Outputs("Out");
    std::vector<framework::LoDTensor> ins;
W
Wu Yi 已提交
82 83 84 85 86 87

    // For profiling
    platform::DeviceContextPool& pool = platform::DeviceContextPool::Instance();
    auto& ctx = *pool.Get(dev_place);
    platform::RecordEvent record_event(Type(), &ctx);

F
fengjiayi 已提交
88
    reader->ReadNext(&ins);
89
    if (ins.empty()) {
F
fengjiayi 已提交
90
      if (Attr<bool>("throw_eof_exp")) {
91
        PADDLE_THROW_EOF();
F
fengjiayi 已提交
92 93 94 95 96 97
      } else {
        ins.resize(out_arg_names.size());
        for (auto& tensor : ins) {
          // data type is not important for subsequent DataBalanceOpHandle
          tensor.mutable_data<float>(framework::make_ddim({0}), dev_place);
        }
98 99
      }
    }
F
fengjiayi 已提交
100
    PADDLE_ENFORCE_EQ(ins.size(), out_arg_names.size());
101
    for (size_t i = 0; i < out_arg_names.size(); ++i) {
F
fengjiayi 已提交
102 103 104 105 106 107 108 109 110 111
      auto* out =
          scope.FindVar(out_arg_names[i])->GetMutable<framework::LoDTensor>();
      out->ShareDataWith(ins[i]);
      out->set_lod(ins[i].lod());
    }
  }
};

class ReadOpMaker : public framework::OpProtoAndCheckerMaker {
 public:
Y
Yu Yang 已提交
112
  void Make() override {
F
fengjiayi 已提交
113 114
    AddInput("Reader", "(ReaderHolder) The executed reader.");
    AddOutput("Out", "(LoDTensor) The output data.").AsDuplicable();
Y
yuyang18 已提交
115 116 117 118 119 120 121
    AddAttr<bool>(
        "throw_eof_exp",
        "If set true, an exception will be thrown when the Reader "
        "yields empty (which means there is no next data).\n"
        "NOTES: This flag must be true always. It will be set to false"
        " only when the data-balance is enabled in ParallelExecutor"
        " and it is set by ParallelExecutor instance, not users.")
F
fengjiayi 已提交
122
        .SetDefault(true);
F
fengjiayi 已提交
123 124 125 126
    AddComment(R"DOC(
      Read Operator

      Execute a given reader once and output data.
F
fengjiayi 已提交
127
    )DOC");
F
fengjiayi 已提交
128 129 130 131
  }
};

}  // namespace operators
F
fengjiayi 已提交
132 133 134 135 136
}  // namespace paddle

namespace ops = paddle::operators;
REGISTER_OPERATOR(read, ops::ReadOp, ops::ReadInferShape, ops::ReadOpMaker,
                  paddle::framework::EmptyGradOpMaker, ops::ReadInferVarType);