build_strategy.cc 5.0 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */

#include "paddle/fluid/framework/details/build_strategy.h"

#include "paddle/fluid/framework/details/multi_devices_graph_check_pass.h"
#include "paddle/fluid/framework/details/multi_devices_graph_print_pass.h"
#include "paddle/fluid/framework/ir/graph.h"
#include "paddle/fluid/framework/ir/graph_viz_pass.h"

namespace paddle {
namespace framework {
namespace details {

class ParallelExecutorPassBuilder : public ir::PassBuilder {
 public:
  explicit ParallelExecutorPassBuilder(const BuildStrategy &strategy)
      : ir::PassBuilder(), strategy_(strategy) {
X
Xin Pan 已提交
30
    // Add a graph viz pass to record a graph.
31 32 33 34 35 36 37
    if (!strategy_.debug_graphviz_path_.empty()) {
      auto viz_pass = AppendPass("graph_viz_pass");
      const std::string graph_path = string::Sprintf(
          "%s%s", strategy_.debug_graphviz_path_.c_str(), "_original_graph");
      viz_pass->Set<std::string>("graph_viz_path", new std::string(graph_path));
    }

X
Xin Pan 已提交
38
    // Add op fusion.
39
    if (strategy.fuse_elewise_add_act_ops_) {
X
Xin Pan 已提交
40
      auto fuse_elewise_add_act_pass = AppendPass("fuse_elewise_add_act_pass");
X
Xin Pan 已提交
41
      // Add a graph viz pass to record a graph.
42
      if (!strategy.debug_graphviz_path_.empty()) {
X
Xin Pan 已提交
43
        auto viz_pass = AppendPass("graph_viz_pass");
44 45
        const std::string graph_path = string::Sprintf(
            "%s%s", strategy.debug_graphviz_path_.c_str(), "_fused_graph");
X
Xin Pan 已提交
46 47
        viz_pass->Set<std::string>("graph_viz_path",
                                   new std::string(graph_path));
48 49 50 51 52 53 54 55
      }
    }

    // Convert graph to run on multi-devices.
    auto multi_devices_pass = AppendPass("multi_devices_pass");
    multi_devices_pass->SetNotOwned<const BuildStrategy>("strategy",
                                                         &strategy_);

X
Xin Pan 已提交
56
    // Add a graph print pass to record a graph with device info.
57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72
    if (!strategy_.debug_graphviz_path_.empty()) {
      auto multi_devices_print_pass = AppendPass("multi_devices_print_pass");
      multi_devices_print_pass->SetNotOwned<const std::string>(
          "debug_graphviz_path", &strategy_.debug_graphviz_path_);
      multi_devices_print_pass->Set<details::GraphvizSSAGraphPrinter>(
          "graph_printer", new details::GraphvizSSAGraphPrinter);
    }

    // Verify that the graph is correct for multi-device executor.
    AppendPass("multi_devices_check_pass");
  }

 private:
  BuildStrategy strategy_;
};

X
fix  
Xin Pan 已提交
73 74
std::shared_ptr<ir::PassBuilder> BuildStrategy::CreatePassesFromStrategy()
    const {
75
  pass_builder_.reset(new ParallelExecutorPassBuilder(*this));
X
fix  
Xin Pan 已提交
76
  return pass_builder_;
77 78 79 80 81 82 83 84 85 86 87 88
}

std::unique_ptr<ir::Graph> BuildStrategy::Apply(
    const ProgramDesc &main_program, const std::vector<platform::Place> &places,
    const std::string &loss_var_name,
    const std::unordered_set<std::string> &param_names,
    const std::vector<Scope *> &local_scopes,
#ifdef PADDLE_WITH_CUDA
    const bool use_cuda, platform::NCCLContextMap *nccl_ctxs) const {
#else
    const bool use_cuda) const {
#endif
X
Xin Pan 已提交
89
  // Create a default one if not initialized by user.
90
  if (!pass_builder_) {
X
fix  
Xin Pan 已提交
91
    CreatePassesFromStrategy();
92
  }
X
fix  
Xin Pan 已提交
93 94 95 96 97

  std::unique_ptr<ir::Graph> graph(new ir::Graph(main_program));

  for (std::shared_ptr<ir::Pass> &pass : pass_builder_->AllPasses()) {
    if (pass->Type() == "multi_devices_pass") {
S
sneaxiy 已提交
98 99 100 101 102
      pass->Erase("enable_sequence_execution");
      if (enable_sequence_execution_) {
        pass->Set("enable_sequence_execution", new bool(true));
      }

X
fix  
Xin Pan 已提交
103 104 105 106 107 108 109 110 111 112
      pass->Erase("places");
      pass->SetNotOwned<const std::vector<platform::Place>>("places", &places);
      pass->Erase("loss_var_name");
      pass->SetNotOwned<const std::string>("loss_var_name", &loss_var_name);
      pass->Erase("params");
      pass->SetNotOwned<const std::unordered_set<std::string>>("params",
                                                               &param_names);
      pass->Erase("local_scopes");
      pass->SetNotOwned<const std::vector<Scope *>>("local_scopes",
                                                    &local_scopes);
113
#ifdef PADDLE_WITH_CUDA
X
fix  
Xin Pan 已提交
114 115 116
      platform::NCCLContextMap *nctx = use_cuda ? nccl_ctxs : nullptr;
      pass->Erase("nccl_ctxs");
      pass->SetNotOwned<platform::NCCLContextMap>("nccl_ctxs", nctx);
117
#endif
X
fix  
Xin Pan 已提交
118 119 120
    }
    graph = pass->Apply(std::move(graph));
  }
121 122 123 124 125 126 127 128 129 130 131
  return graph;
}
}  // namespace details
}  // namespace framework
}  // namespace paddle

USE_PASS(fuse_elewise_add_act_pass);
USE_PASS(graph_viz_pass);
USE_PASS(multi_devices_pass);
USE_PASS(multi_devices_check_pass);
USE_PASS(multi_devices_print_pass);