executor.h 4.6 KB
Newer Older
1
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
Q
qijun 已提交
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#pragma once

L
Liu Yiqun 已提交
17 18 19
#include <map>
#include <string>
#include <vector>
S
sneaxiy 已提交
20
#include "paddle/fluid/framework/garbage_collector.h"
Y
Yi Wang 已提交
21 22 23 24 25
#include "paddle/fluid/framework/op_info.h"
#include "paddle/fluid/framework/program_desc.h"
#include "paddle/fluid/framework/scope.h"
#include "paddle/fluid/framework/tensor.h"
#include "paddle/fluid/platform/device_context.h"
Q
qijun 已提交
26 27 28

namespace paddle {
namespace framework {
29
extern void InitializeVariable(Variable* var, proto::VarType::Type var_type);
Q
Qiao Longfei 已提交
30

S
sneaxiy 已提交
31 32 33 34 35 36
template <typename T>
std::unordered_map<std::string, T> GetNonPersistableReferenceCount(
    const ProgramDesc& prog, size_t block_id) {
  auto& block = prog.Block(block_id);
  std::unordered_map<std::string, T> ref_cnts;

S
sneaxiy 已提交
37 38 39 40 41 42 43 44 45
  auto update_ref_cnts = [&](OpDesc* op_desc, const VariableNameMap& name_map) {
    for (auto& name_pair : name_map) {
      for (auto& name : name_pair.second) {
        auto* var_desc = block.FindVar(name);
        if (var_desc == nullptr || var_desc->Persistable()) continue;
        auto type = var_desc->Proto()->type().type();
        if (type != proto::VarType::LOD_TENSOR &&
            type != proto::VarType::SELECTED_ROWS) {
          continue;
S
sneaxiy 已提交
46 47
        }

S
sneaxiy 已提交
48 49 50 51 52
        auto it = ref_cnts.find(name);
        if (it != ref_cnts.end()) {
          ++it->second;
        } else {
          ref_cnts[name] = 1;
S
sneaxiy 已提交
53 54 55
        }
      }
    }
S
sneaxiy 已提交
56 57 58 59 60
  };

  for (auto op_desc : block.AllOps()) {
    update_ref_cnts(op_desc, op_desc->Inputs());
    update_ref_cnts(op_desc, op_desc->Outputs());
S
sneaxiy 已提交
61 62 63 64
  }
  return ref_cnts;
}

Q
Qiao Longfei 已提交
65 66 67 68
struct ExecutorPrepareContext {
  ExecutorPrepareContext(const framework::ProgramDesc& prog, size_t block_id);
  ~ExecutorPrepareContext();

S
sneaxiy 已提交
69 70
  void ResetReferenceCount() { cur_ref_cnts_ = ref_cnts_; }

Q
Qiao Longfei 已提交
71 72 73
  const framework::ProgramDesc& prog_;
  size_t block_id_;
  std::vector<std::unique_ptr<OperatorBase>> ops_;
S
sneaxiy 已提交
74 75

  std::unordered_map<std::string, int> ref_cnts_;
S
sneaxiy 已提交
76
  std::unordered_map<std::string, int> cur_ref_cnts_;
Q
Qiao Longfei 已提交
77 78
};

Q
qijun 已提交
79 80
class Executor {
 public:
D
dzhwinter 已提交
81 82
  // TODO(dzhwinter) : Do not rely on this function, it will be removed
  explicit Executor(const platform::DeviceContext& device)
D
dzhwinter 已提交
83
      : Executor(device.GetPlace()) {}
D
dzhwinter 已提交
84

D
dzhwinter 已提交
85
  explicit Executor(const platform::Place& place);
Y
Yang Yang 已提交
86

W
Wu Yi 已提交
87
  /*
Y
Yancey1989 已提交
88 89
   * Close this Executor.
   * Calling this method will send complete messages to all pserver instances.
W
Wu Yi 已提交
90
   */
Y
Yancey1989 已提交
91
  void Close();
W
Wu Yi 已提交
92

Y
Yang Yang 已提交
93 94 95 96 97 98 99
  /* @Brief
   * Runtime evaluation of the given ProgramDesc under certain Scope
   *
   * @param
   *  ProgramDesc
   *  Scope
   */
Y
Yu Yang 已提交
100 101
  void Run(const ProgramDesc& prog, Scope* scope, int block_id,
           bool create_local_scope = true, bool create_vars = true);
Q
qijun 已提交
102

X
fix  
Xin Pan 已提交
103
  // This API is very slow.
104
  void Run(const ProgramDesc& program, Scope* scope,
105 106
           std::map<std::string, const LoDTensor*>* feed_targets,
           std::map<std::string, LoDTensor*>* fetch_targets,
W
Wu Yi 已提交
107
           bool create_local_scope = true, bool create_vars = true,
108
           const std::string& feed_holder_name = "feed",
109
           const std::string& fetch_holder_name = "fetch");
110

Q
Qiao Longfei 已提交
111 112
  static std::unique_ptr<ExecutorPrepareContext> Prepare(
      const ProgramDesc& program, int block_id);
Y
Yu Yang 已提交
113

T
typhoonzero 已提交
114 115 116
  static std::vector<std::shared_ptr<ExecutorPrepareContext>> Prepare(
      const ProgramDesc& program, const std::vector<int>& block_ids);

L
Liu Yiqun 已提交
117
  void CreateVariables(const ProgramDesc& pdesc, Scope* scope, int block_id);
118

Y
Yu Yang 已提交
119 120
  void RunPreparedContext(ExecutorPrepareContext* ctx, Scope* scope,
                          bool create_local_scope = true,
Q
qiaolongfei 已提交
121
                          bool create_vars = true, bool keep_kids = false);
Y
Yu Yang 已提交
122

X
fix  
Xin Pan 已提交
123
  // This API is very slow.
124
  void RunPreparedContext(ExecutorPrepareContext* ctx, Scope* scope,
125 126
                          std::map<std::string, const LoDTensor*>* feed_targets,
                          std::map<std::string, LoDTensor*>* fetch_targets,
W
Wu Yi 已提交
127
                          bool create_local_scope = true,
L
Liu Yiqun 已提交
128
                          bool create_vars = true,
129
                          const std::string& feed_holder_name = "feed",
L
Liu Yiqun 已提交
130
                          const std::string& fetch_holder_name = "fetch");
131

132 133
  void EnableMKLDNN(const ProgramDesc& program);

Q
qijun 已提交
134
 private:
D
dzhwinter 已提交
135
  const platform::Place place_;
Q
qijun 已提交
136
};
Q
qijun 已提交
137 138 139

}  // namespace framework
}  // namespace paddle