multi_devices_helper.h 4.6 KB
Newer Older
Y
Yu Yang 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
//   Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#pragma once

Y
Yu Yang 已提交
17 18
#include <memory>
#include <string>
C
chengduo 已提交
19 20 21
#include <unordered_map>
#include <unordered_set>
#include <utility>
X
Xin Pan 已提交
22
#include <vector>
W
wanghuancoder 已提交
23

X
clean  
Xin Pan 已提交
24
#include "paddle/fluid/framework/details/op_handle_base.h"
25
#include "paddle/fluid/framework/details/scope_buffered_ssa_graph_executor.h"
X
clean  
Xin Pan 已提交
26
#include "paddle/fluid/framework/details/var_handle.h"
W
wanghuancoder 已提交
27 28
#include "paddle/fluid/framework/ir/graph.h"
#include "paddle/fluid/framework/ir/pass.h"
Z
Zeng Jinle 已提交
29 30
#include "paddle/fluid/framework/op_desc.h"
#include "paddle/fluid/framework/op_proto_maker.h"
Y
Yu Yang 已提交
31 32 33
#include "paddle/fluid/framework/program_desc.h"
#include "paddle/fluid/platform/place.h"

W
wanghuancoder 已提交
34 35 36 37 38
namespace paddle {
namespace framework {
class OpDesc;
}  // namespace framework
}  // namespace paddle
X
Xin Pan 已提交
39

Y
Yu Yang 已提交
40 41 42 43
namespace paddle {
namespace framework {
namespace details {

X
clean  
Xin Pan 已提交
44 45 46
// all variable in each devices.
// The outside vector is the device vector. Each element of this vector is a
// map from variable name to variables. The variables, who have the same name,
C
chenxujun 已提交
47 48
// will have a different version. The offset in the
// `std::vector<VarHandle*>` is the version of variables.
Y
Yancey1989 已提交
49
typedef std::vector<std::unordered_map<std::string, std::vector<VarHandle *>>>
X
Xin Pan 已提交
50
    GraphVars;
C
chengduo 已提交
51
constexpr char kGraphVars[] = "vars";
Y
Yancey1989 已提交
52

53 54
constexpr char kNRanks[] = "nranks";

C
chengduo 已提交
55
constexpr char kPlaces[] = "places";
56
constexpr char kGlobalScope[] = "global_scope";
C
chengduo 已提交
57
constexpr char kLocalScopes[] = "local_scopes";
C
chengduo 已提交
58
constexpr char kNCCLCtxs[] = "nccl_ctxs";
59
constexpr char kBKCLCtxs[] = "bkcl_ctxs";
60
constexpr char kUseHierarchicalAllReduce[] = "use_hierarchical_allreduce";
C
chengduo 已提交
61 62 63 64

// aux variables to represent dependency. Useful to resolve data hazard.
typedef std::unordered_set<VarHandleBase *> GraphDepVars;
constexpr char kGraphDepVars[] = "dep_vars";
C
chengduo 已提交
65

66
typedef std::unordered_map<std::string, details::VariableInfo> FusedVars;
C
chengduo 已提交
67
constexpr char kFusedVars[] = "fused_vars";
C
chengduo 已提交
68 69 70 71 72
constexpr char kFusedVarNamePrefix[] = "@FUSEDVAR@";

typedef std::string FusedOptType;
constexpr char kFusedOptType[] = "fused_opt_type";

73
typedef std::vector<std::string> FusedGrads;
C
chengduo 已提交
74
constexpr char kFusedGrads[] = "fused_gradients";
C
chengduo 已提交
75 76

typedef std::vector<std::pair<std::string, std::string>> ParamsAndGrads;
77 78 79
constexpr char kParamsAndDenseGrads[] = "params_and_dense_grads";
constexpr char kParamsAndSparseGrads[] = "params_and_sparse_grads";

80 81 82
typedef std::unordered_set<std::string> PinnedVars;
constexpr char kPinnedVars[] = "pinned_vars";

C
chengduo 已提交
83
typedef std::vector<std::vector<std::pair<std::string, std::string>>>
84
    GroupParamsAndGrads;
85
constexpr char kGroupParamsAndDenseGrads[] = "group_params_dense_grads";
C
chengduo 已提交
86

Z
Zeng Jinle 已提交
87 88 89 90
inline bool IsOpRole(const OpDesc &op, OpRole role) {
  const auto &attrs = op.GetAttrMap();
  auto iter = attrs.find(OpProtoAndCheckerMaker::OpRoleAttrName());
  if (iter == attrs.end()) return false;
R
Ruibiao Chen 已提交
91
  return static_cast<bool>(PADDLE_GET_CONST(int, iter->second) &
Z
Zeng Jinle 已提交
92 93 94 95 96 97 98
                           static_cast<int>(role));
}

inline std::vector<std::string> GetOpRoleVarsOrEmpty(const OpDesc &op) {
  const auto &attrs = op.GetAttrMap();
  auto iter = attrs.find(OpProtoAndCheckerMaker::OpRoleVarAttrName());
  if (iter == attrs.end()) return {};
R
Ruibiao Chen 已提交
99
  auto &ret = PADDLE_GET_CONST(std::vector<std::string>, iter->second);
Z
Zeng Jinle 已提交
100
  PADDLE_ENFORCE_EQ(
101 102
      ret.size() % 2,
      0,
Z
Zeng Jinle 已提交
103 104
      platform::errors::InvalidArgument(
          "The size of attribute %s must be an even number, but got %d",
105 106
          OpProtoAndCheckerMaker::OpRoleVarAttrName(),
          ret.size()));
R
Ruibiao Chen 已提交
107
  return PADDLE_GET_CONST(std::vector<std::string>, iter->second);
Z
Zeng Jinle 已提交
108 109
}

110 111
bool IsDataParallelInferenceGraph(const ir::Graph &graph);

112 113 114 115 116 117 118
std::vector<std::unique_ptr<ir::Graph>> TrySeparateToMultipleSingleDeviceGraphs(
    ir::Graph *graph);

bool HasDropLastReadOp(const ir::Graph &graph);

bool HasKeepLastReadOp(const ir::Graph &graph);

119
template <typename T>
120 121
void CopyGraphAttrIfExists(const ir::Graph &src,
                           ir::Graph *dst,
122 123 124 125 126 127 128
                           const std::string &name) {
  if (src.Has(name)) {
    auto &attr = src.Get<T>(name);
    dst->Set(name, new T(attr));
  }
}

Y
Yu Yang 已提交
129 130 131
}  // namespace details
}  // namespace framework
}  // namespace paddle