executor.cc 5.3 KB
Newer Older
Q
qijun 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#include "paddle/framework/executor.h"
Y
Yang Yang 已提交
16

Y
Yang Yang 已提交
17
#include <set>
Y
Yang Yang 已提交
18

Y
Yang Yu 已提交
19
#include "gflags/gflags.h"
Q
QI JUN 已提交
20
#include "paddle/framework/feed_fetch_type.h"
Y
Yu Yang 已提交
21
#include "paddle/framework/lod_rank_table.h"
Y
Yu Yang 已提交
22
#include "paddle/framework/lod_tensor_array.h"
Q
qijun 已提交
23
#include "paddle/framework/op_registry.h"
Y
Yang Yu 已提交
24
#include "paddle/platform/place.h"
Y
Yang Yu 已提交
25

26
DECLARE_bool(do_memory_benchmark);
Y
Yang Yu 已提交
27 28 29
DEFINE_bool(check_nan_inf, false,
            "Checking whether operator produce NAN/INF or not. It will be "
            "extremely slow so please use this flag wisely.");
Q
qijun 已提交
30 31 32 33

namespace paddle {
namespace framework {

Y
Yang Yang 已提交
34 35 36
const std::string kFeedOpType = "feed";
const std::string kFetchOpType = "fetch";

D
dzhwinter 已提交
37
Executor::Executor(const platform::Place& place) : place_(place) {}
Q
qijun 已提交
38

Y
Yancey 已提交
39
static void CreateTensor(Variable* var, proto::VarDesc::VarType var_type) {
40
  if (var_type == proto::VarDesc::LOD_TENSOR) {
Q
QI JUN 已提交
41
    var->GetMutable<LoDTensor>();
42
  } else if (var_type == proto::VarDesc::SELECTED_ROWS) {
Q
QI JUN 已提交
43
    var->GetMutable<SelectedRows>();
44
  } else if (var_type == proto::VarDesc::FEED_MINIBATCH) {
Q
QI JUN 已提交
45
    var->GetMutable<FeedFetchList>();
46
  } else if (var_type == proto::VarDesc::FETCH_LIST) {
Q
QI JUN 已提交
47
    var->GetMutable<FeedFetchList>();
48
  } else if (var_type == proto::VarDesc::STEP_SCOPES) {
Y
Yu Yang 已提交
49
    var->GetMutable<std::vector<framework::Scope>>();
50
  } else if (var_type == proto::VarDesc::LOD_RANK_TABLE) {
Y
Yu Yang 已提交
51
    var->GetMutable<LoDRankTable>();
52
  } else if (var_type == proto::VarDesc::LOD_TENSOR_ARRAY) {
Y
Yu Yang 已提交
53
    var->GetMutable<LoDTensorArray>();
Y
Yang Yu 已提交
54 55
  } else if (var_type == proto::VarDesc::PLACE_LIST) {
    var->GetMutable<platform::PlaceList>();
Q
QI JUN 已提交
56 57
  } else {
    PADDLE_THROW(
Y
Yu Yang 已提交
58
        "Variable type %d is not in "
Y
Yang Yu 已提交
59 60
        "[LoDTensor, SelectedRows, FEED_MINIBATCH, FETCH_LIST, LOD_RANK_TABLE,"
        " PLACE_LIST]",
Y
Yu Yang 已提交
61
        var_type);
Q
QI JUN 已提交
62 63 64
  }
}

Y
Yang Yu 已提交
65 66
static void CheckTensorNANOrInf(const std::string& name,
                                const framework::Tensor& tensor) {
Y
Yang Yu 已提交
67
  if (tensor.memory_size() == 0) {
Y
Yang Yu 已提交
68 69
    return;
  }
Y
Yang Yu 已提交
70 71
  if (tensor.type().hash_code() != typeid(float).hash_code() &&
      tensor.type().hash_code() != typeid(double).hash_code()) {
Y
Yang Yu 已提交
72 73 74 75 76 77
    return;
  }
  PADDLE_ENFORCE(!framework::HasInf(tensor), "Tensor %s has Inf", name);
  PADDLE_ENFORCE(!framework::HasNAN(tensor), "Tensor %s has NAN", name);
}

Y
Yu Yang 已提交
78
void Executor::Run(const ProgramDesc& pdesc, Scope* scope, int block_id,
T
typhoonzero 已提交
79
                   bool create_local_scope, bool create_vars) {
Y
Yang Yang 已提交
80
  // TODO(tonyyang-svail):
Y
Yang Yang 已提交
81
  //    - only runs on the first device (i.e. no interdevice communication)
Y
Yang Yang 已提交
82
  //    - will change to use multiple blocks for RNN op and Cond Op
83
  PADDLE_ENFORCE_LT(static_cast<size_t>(block_id), pdesc.Size());
84
  auto& block = pdesc.Block(block_id);
Y
Yang Yang 已提交
85

Y
Yu Yang 已提交
86
  Scope* local_scope = scope;
T
typhoonzero 已提交
87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105
  if (create_vars) {
    if (create_local_scope) {
      local_scope = &scope->NewScope();
      for (auto& var : block.AllVars()) {
        if (var->Name() == framework::kEmptyVarName) {
          continue;
        }

        if (var->Persistable()) {
          auto* ptr = scope->Var(var->Name());
          CreateTensor(ptr, var->GetType());
          VLOG(3) << "Create Variable " << var->Name()
                  << " global, which pointer is " << ptr;
        } else {
          auto* ptr = local_scope->Var(var->Name());
          CreateTensor(ptr, var->GetType());
          VLOG(3) << "Create Variable " << var->Name()
                  << " locally, which pointer is " << ptr;
        }
106
      }
T
typhoonzero 已提交
107 108
    } else {
      for (auto& var : block.AllVars()) {
Y
Yu Yang 已提交
109 110
        auto* ptr = local_scope->Var(var->Name());
        CreateTensor(ptr, var->GetType());
T
typhoonzero 已提交
111 112
        VLOG(3) << "Create variable " << var->Name() << ", which pointer is "
                << ptr;
Y
Yu Yang 已提交
113
      }
T
typhoonzero 已提交
114 115
    }  // if (create_local_scope)
  }    // if (create_vars)
Y
Yang Yang 已提交
116

117 118
  for (auto& op_desc : block.AllOps()) {
    auto op = paddle::framework::OpRegistry::CreateOp(*op_desc);
119
    VLOG(3) << op->DebugStringEx(local_scope);
D
dzhwinter 已提交
120
    op->Run(*local_scope, place_);
121 122 123 124
    if (FLAGS_do_memory_benchmark) {
      VLOG(2) << "Memory used after operator " + op->Type() + " running: "
              << memory::memory_usage(place_);
    }
Y
Yang Yu 已提交
125 126 127 128 129 130 131 132 133
    if (FLAGS_check_nan_inf) {
      for (auto& vname : op->OutputVars(true)) {
        auto* var = local_scope->FindVar(vname);
        if (var == nullptr) continue;
        if (var->IsType<framework::LoDTensor>()) {
          CheckTensorNANOrInf(vname, var->Get<framework::LoDTensor>());
        }
      }
    }
Y
Yu Yang 已提交
134
  }
G
gongweibao 已提交
135
  if (create_vars && create_local_scope) {
Y
Yu Yang 已提交
136
    scope->DeleteScope(local_scope);
Q
qijun 已提交
137
  }
138 139 140 141 142 143
  if (FLAGS_do_memory_benchmark) {
    VLOG(2) << "-------------------------------------------------------";
    VLOG(2) << "Memory used after deleting local scope: "
            << memory::memory_usage(place_);
    VLOG(2) << "-------------------------------------------------------";
  }
Q
qijun 已提交
144 145 146 147
}

}  // namespace framework
}  // namespace paddle