build_strategy.cc 7.1 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#include "paddle/fluid/framework/details/build_strategy.h"

#include "paddle/fluid/framework/details/multi_devices_graph_check_pass.h"
#include "paddle/fluid/framework/details/multi_devices_graph_print_pass.h"
19
#include "paddle/fluid/framework/details/reduce_op_handle.h"
S
sneaxiy 已提交
20
#include "paddle/fluid/framework/details/sequential_execution_pass.h"
21 22 23 24 25 26 27
#include "paddle/fluid/framework/ir/graph.h"
#include "paddle/fluid/framework/ir/graph_viz_pass.h"

namespace paddle {
namespace framework {
namespace details {

28
static inline bool SeqOnlyAllReduceOps(const BuildStrategy &strategy) {
Y
Yancey1989 已提交
29 30 31
  return (!strategy.enable_sequential_execution_ &&
          strategy.num_trainers_ > 1) ||
         strategy.enable_parallel_graph_;
32 33
}

34 35 36 37
class ParallelExecutorPassBuilder : public ir::PassBuilder {
 public:
  explicit ParallelExecutorPassBuilder(const BuildStrategy &strategy)
      : ir::PassBuilder(), strategy_(strategy) {
S
sneaxiy 已提交
38 39 40 41
    if (strategy_.enable_sequential_execution_) {
      AppendPass("sequential_execution_pass");
    }

X
Xin Pan 已提交
42
    // Add a graph viz pass to record a graph.
43 44 45 46 47 48 49
    if (!strategy_.debug_graphviz_path_.empty()) {
      auto viz_pass = AppendPass("graph_viz_pass");
      const std::string graph_path = string::Sprintf(
          "%s%s", strategy_.debug_graphviz_path_.c_str(), "_original_graph");
      viz_pass->Set<std::string>("graph_viz_path", new std::string(graph_path));
    }

X
Xin Pan 已提交
50
    // Add op fusion.
51
    if (strategy.fuse_elewise_add_act_ops_) {
X
Xin Pan 已提交
52
      auto fuse_elewise_add_act_pass = AppendPass("fuse_elewise_add_act_pass");
X
Xin Pan 已提交
53
      // Add a graph viz pass to record a graph.
54
      if (!strategy.debug_graphviz_path_.empty()) {
X
Xin Pan 已提交
55
        auto viz_pass = AppendPass("graph_viz_pass");
56 57
        const std::string graph_path = string::Sprintf(
            "%s%s", strategy.debug_graphviz_path_.c_str(), "_fused_graph");
X
Xin Pan 已提交
58 59
        viz_pass->Set<std::string>("graph_viz_path",
                                   new std::string(graph_path));
60 61 62
      }
    }

63 64 65 66 67 68 69 70 71 72 73
    CollectiveContext *context = CollectiveContext::GetInstance();
    context->endpoints_ = strategy_.trainers_endpoints_;
    context->trainer_id_ = strategy_.trainer_id_;
    PADDLE_ENFORCE(strategy_.trainer_id_ >= 0, "trainer_id_ >= 0");
    if (strategy_.trainer_id_ > 0) {
      PADDLE_ENFORCE((unsigned)(strategy_.trainer_id_) <
                         strategy_.trainers_endpoints_.size(),
                     "trainer_id_ < endpoints_ size");
    }
    VLOG(1) << "CollectiveContext:" << context->String();

74 75 76 77
    // Convert graph to run on multi-devices.
    auto multi_devices_pass = AppendPass("multi_devices_pass");
    multi_devices_pass->SetNotOwned<const BuildStrategy>("strategy",
                                                         &strategy_);
78 79
    multi_devices_pass->Set<int>("num_trainers",
                                 new int(strategy_.num_trainers_));
80

X
Xin Pan 已提交
81
    // Add a graph print pass to record a graph with device info.
82 83 84 85 86 87 88 89 90 91
    if (!strategy_.debug_graphviz_path_.empty()) {
      auto multi_devices_print_pass = AppendPass("multi_devices_print_pass");
      multi_devices_print_pass->SetNotOwned<const std::string>(
          "debug_graphviz_path", &strategy_.debug_graphviz_path_);
      multi_devices_print_pass->Set<details::GraphvizSSAGraphPrinter>(
          "graph_printer", new details::GraphvizSSAGraphPrinter);
    }

    // Verify that the graph is correct for multi-device executor.
    AppendPass("multi_devices_check_pass");
S
sneaxiy 已提交
92

93 94 95 96
    if (SeqOnlyAllReduceOps(strategy)) {
      AppendPass("all_reduce_deps_pass");
    }

S
sneaxiy 已提交
97 98 99
    if (strategy_.remove_unnecessary_lock_) {
      AppendPass("modify_op_lock_and_record_event_pass");
    }
100 101 102 103 104 105
  }

 private:
  BuildStrategy strategy_;
};

106
std::shared_ptr<ir::PassBuilder> BuildStrategy::CreatePassesFromStrategy(
X
Xin Pan 已提交
107 108
    bool finalize_strategy) const {
  if (is_finalized_) {
109 110
    return pass_builder_;
  }
111
  pass_builder_.reset(new ParallelExecutorPassBuilder(*this));
X
Xin Pan 已提交
112 113
  if (finalize_strategy) {
    is_finalized_ = true;
114
  }
X
fix  
Xin Pan 已提交
115
  return pass_builder_;
116 117 118 119 120 121 122
}

std::unique_ptr<ir::Graph> BuildStrategy::Apply(
    const ProgramDesc &main_program, const std::vector<platform::Place> &places,
    const std::string &loss_var_name,
    const std::unordered_set<std::string> &param_names,
    const std::vector<Scope *> &local_scopes,
P
peizhilin 已提交
123
#if defined(PADDLE_WITH_CUDA) && !defined(_WIN32)
124 125 126 127
    const bool use_cuda, platform::NCCLContextMap *nccl_ctxs) const {
#else
    const bool use_cuda) const {
#endif
128 129
  // Create a default one if not finalized by user.
  CreatePassesFromStrategy(false);
X
fix  
Xin Pan 已提交
130 131 132 133 134 135 136 137 138 139 140 141 142 143 144

  std::unique_ptr<ir::Graph> graph(new ir::Graph(main_program));

  for (std::shared_ptr<ir::Pass> &pass : pass_builder_->AllPasses()) {
    if (pass->Type() == "multi_devices_pass") {
      pass->Erase("places");
      pass->SetNotOwned<const std::vector<platform::Place>>("places", &places);
      pass->Erase("loss_var_name");
      pass->SetNotOwned<const std::string>("loss_var_name", &loss_var_name);
      pass->Erase("params");
      pass->SetNotOwned<const std::unordered_set<std::string>>("params",
                                                               &param_names);
      pass->Erase("local_scopes");
      pass->SetNotOwned<const std::vector<Scope *>>("local_scopes",
                                                    &local_scopes);
P
peizhilin 已提交
145
#if defined(PADDLE_WITH_CUDA) && !defined(_WIN32)
X
fix  
Xin Pan 已提交
146 147 148
      platform::NCCLContextMap *nctx = use_cuda ? nccl_ctxs : nullptr;
      pass->Erase("nccl_ctxs");
      pass->SetNotOwned<platform::NCCLContextMap>("nccl_ctxs", nctx);
149
#endif
S
sneaxiy 已提交
150
    } else if (pass->Type() == "sequential_execution_pass") {
151 152
      LOG(INFO) << "set enable_sequential_execution:"
                << enable_sequential_execution_;
153 154 155 156 157 158

      pass->Erase(kAllOpDescs);
      pass->Set<const std::vector<OpDesc *>>(
          kAllOpDescs,
          new std::vector<OpDesc *>(main_program.Block(0).AllOps()));
    } else if (pass->Type() == "all_reduce_deps_pass") {
159 160
      LOG(INFO) << "SeqOnlyAllReduceOps:" << SeqOnlyAllReduceOps(*this)
                << ", num_trainers:" << num_trainers_;
161

S
sneaxiy 已提交
162 163 164 165
      pass->Erase(kAllOpDescs);
      pass->Set<const std::vector<OpDesc *>>(
          kAllOpDescs,
          new std::vector<OpDesc *>(main_program.Block(0).AllOps()));
X
fix  
Xin Pan 已提交
166 167 168
    }
    graph = pass->Apply(std::move(graph));
  }
169 170 171 172 173 174 175 176
  return graph;
}
}  // namespace details
}  // namespace framework
}  // namespace paddle

USE_PASS(fuse_elewise_add_act_pass);
USE_PASS(graph_viz_pass);
177
USE_PASS(multi_batch_merge_pass);
178 179 180
USE_PASS(multi_devices_pass);
USE_PASS(multi_devices_check_pass);
USE_PASS(multi_devices_print_pass);
S
sneaxiy 已提交
181
USE_PASS(sequential_execution_pass);
182
USE_PASS(all_reduce_deps_pass);
S
sneaxiy 已提交
183
USE_PASS(modify_op_lock_and_record_event_pass);