build_strategy.cc 6.5 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#include "paddle/fluid/framework/details/build_strategy.h"

#include "paddle/fluid/framework/details/multi_devices_graph_check_pass.h"
#include "paddle/fluid/framework/details/multi_devices_graph_print_pass.h"
19
#include "paddle/fluid/framework/details/reduce_op_handle.h"
S
sneaxiy 已提交
20
#include "paddle/fluid/framework/details/sequential_execution_pass.h"
21 22 23 24 25 26 27
#include "paddle/fluid/framework/ir/graph.h"
#include "paddle/fluid/framework/ir/graph_viz_pass.h"

namespace paddle {
namespace framework {
namespace details {

28 29 30 31
static inline bool SeqOnlyAllReduceOps(const BuildStrategy &strategy) {
  return (!strategy.enable_sequential_execution_ && strategy.num_trainers_ > 1);
}

32 33 34 35
class ParallelExecutorPassBuilder : public ir::PassBuilder {
 public:
  explicit ParallelExecutorPassBuilder(const BuildStrategy &strategy)
      : ir::PassBuilder(), strategy_(strategy) {
S
sneaxiy 已提交
36 37 38 39
    if (strategy_.enable_sequential_execution_) {
      AppendPass("sequential_execution_pass");
    }

X
Xin Pan 已提交
40
    // Add a graph viz pass to record a graph.
41 42 43 44 45 46 47
    if (!strategy_.debug_graphviz_path_.empty()) {
      auto viz_pass = AppendPass("graph_viz_pass");
      const std::string graph_path = string::Sprintf(
          "%s%s", strategy_.debug_graphviz_path_.c_str(), "_original_graph");
      viz_pass->Set<std::string>("graph_viz_path", new std::string(graph_path));
    }

X
Xin Pan 已提交
48
    // Add op fusion.
49
    if (strategy.fuse_elewise_add_act_ops_) {
X
Xin Pan 已提交
50
      auto fuse_elewise_add_act_pass = AppendPass("fuse_elewise_add_act_pass");
X
Xin Pan 已提交
51
      // Add a graph viz pass to record a graph.
52
      if (!strategy.debug_graphviz_path_.empty()) {
X
Xin Pan 已提交
53
        auto viz_pass = AppendPass("graph_viz_pass");
54 55
        const std::string graph_path = string::Sprintf(
            "%s%s", strategy.debug_graphviz_path_.c_str(), "_fused_graph");
X
Xin Pan 已提交
56 57
        viz_pass->Set<std::string>("graph_viz_path",
                                   new std::string(graph_path));
58 59 60 61 62 63 64
      }
    }

    // Convert graph to run on multi-devices.
    auto multi_devices_pass = AppendPass("multi_devices_pass");
    multi_devices_pass->SetNotOwned<const BuildStrategy>("strategy",
                                                         &strategy_);
65 66
    multi_devices_pass->Set<int>("num_trainers",
                                 new int(strategy_.num_trainers_));
67

X
Xin Pan 已提交
68
    // Add a graph print pass to record a graph with device info.
69 70 71 72 73 74 75 76 77 78
    if (!strategy_.debug_graphviz_path_.empty()) {
      auto multi_devices_print_pass = AppendPass("multi_devices_print_pass");
      multi_devices_print_pass->SetNotOwned<const std::string>(
          "debug_graphviz_path", &strategy_.debug_graphviz_path_);
      multi_devices_print_pass->Set<details::GraphvizSSAGraphPrinter>(
          "graph_printer", new details::GraphvizSSAGraphPrinter);
    }

    // Verify that the graph is correct for multi-device executor.
    AppendPass("multi_devices_check_pass");
S
sneaxiy 已提交
79

80 81 82 83
    if (SeqOnlyAllReduceOps(strategy)) {
      AppendPass("all_reduce_deps_pass");
    }

S
sneaxiy 已提交
84 85 86
    if (strategy_.remove_unnecessary_lock_) {
      AppendPass("modify_op_lock_and_record_event_pass");
    }
87 88 89 90 91 92
  }

 private:
  BuildStrategy strategy_;
};

93
std::shared_ptr<ir::PassBuilder> BuildStrategy::CreatePassesFromStrategy(
X
Xin Pan 已提交
94 95
    bool finalize_strategy) const {
  if (is_finalized_) {
96 97
    return pass_builder_;
  }
98
  pass_builder_.reset(new ParallelExecutorPassBuilder(*this));
X
Xin Pan 已提交
99 100
  if (finalize_strategy) {
    is_finalized_ = true;
101
  }
X
fix  
Xin Pan 已提交
102
  return pass_builder_;
103 104 105 106 107 108 109
}

std::unique_ptr<ir::Graph> BuildStrategy::Apply(
    const ProgramDesc &main_program, const std::vector<platform::Place> &places,
    const std::string &loss_var_name,
    const std::unordered_set<std::string> &param_names,
    const std::vector<Scope *> &local_scopes,
P
peizhilin 已提交
110
#if defined(PADDLE_WITH_CUDA) && !defined(_WIN32)
111 112 113 114
    const bool use_cuda, platform::NCCLContextMap *nccl_ctxs) const {
#else
    const bool use_cuda) const {
#endif
115 116
  // Create a default one if not finalized by user.
  CreatePassesFromStrategy(false);
X
fix  
Xin Pan 已提交
117 118 119 120 121 122 123 124 125 126 127 128 129 130 131

  std::unique_ptr<ir::Graph> graph(new ir::Graph(main_program));

  for (std::shared_ptr<ir::Pass> &pass : pass_builder_->AllPasses()) {
    if (pass->Type() == "multi_devices_pass") {
      pass->Erase("places");
      pass->SetNotOwned<const std::vector<platform::Place>>("places", &places);
      pass->Erase("loss_var_name");
      pass->SetNotOwned<const std::string>("loss_var_name", &loss_var_name);
      pass->Erase("params");
      pass->SetNotOwned<const std::unordered_set<std::string>>("params",
                                                               &param_names);
      pass->Erase("local_scopes");
      pass->SetNotOwned<const std::vector<Scope *>>("local_scopes",
                                                    &local_scopes);
P
peizhilin 已提交
132
#if defined(PADDLE_WITH_CUDA) && !defined(_WIN32)
X
fix  
Xin Pan 已提交
133 134 135
      platform::NCCLContextMap *nctx = use_cuda ? nccl_ctxs : nullptr;
      pass->Erase("nccl_ctxs");
      pass->SetNotOwned<platform::NCCLContextMap>("nccl_ctxs", nctx);
136
#endif
S
sneaxiy 已提交
137
    } else if (pass->Type() == "sequential_execution_pass") {
138 139 140 141 142 143 144 145 146 147 148
      VLOG(1) << "set enable_sequential_execution:"
              << enable_sequential_execution_;

      pass->Erase(kAllOpDescs);
      pass->Set<const std::vector<OpDesc *>>(
          kAllOpDescs,
          new std::vector<OpDesc *>(main_program.Block(0).AllOps()));
    } else if (pass->Type() == "all_reduce_deps_pass") {
      VLOG(1) << "SeqOnlyAllReduceOps:" << SeqOnlyAllReduceOps(*this)
              << ", num_trainers:" << num_trainers_;

S
sneaxiy 已提交
149 150 151 152
      pass->Erase(kAllOpDescs);
      pass->Set<const std::vector<OpDesc *>>(
          kAllOpDescs,
          new std::vector<OpDesc *>(main_program.Block(0).AllOps()));
X
fix  
Xin Pan 已提交
153 154 155
    }
    graph = pass->Apply(std::move(graph));
  }
156 157 158 159 160 161 162 163
  return graph;
}
}  // namespace details
}  // namespace framework
}  // namespace paddle

USE_PASS(fuse_elewise_add_act_pass);
USE_PASS(graph_viz_pass);
164
USE_PASS(multi_batch_merge_pass);
165 166 167
USE_PASS(multi_devices_pass);
USE_PASS(multi_devices_check_pass);
USE_PASS(multi_devices_print_pass);
S
sneaxiy 已提交
168
USE_PASS(sequential_execution_pass);
169
USE_PASS(all_reduce_deps_pass);
S
sneaxiy 已提交
170
USE_PASS(modify_op_lock_and_record_event_pass);