graph_execution_optimizer.py 8.7 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
#   Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and

import paddle
from paddle.fluid.framework import core
from paddle.fluid import compiler
from .meta_optimizer_base import MetaOptimizerBase
from ..base.private_helper_function import wait_server_ready
D
Dong Daxiang 已提交
19
import logging
20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35


class GraphExecutionOptimizer(MetaOptimizerBase):
    def __init__(self, optimizer):
        super(GraphExecutionOptimizer, self).__init__(optimizer)
        self.inner_opt = optimizer
        # we do not allow meta optimizer to be inner optimizer currently
        self.meta_optimizers_white_list = []

    def _is_graph_out(self):
        return True

    def _can_apply(self):
        """
        Basically, this is PE, and almost all programs can be executed here
        """
D
Dong Daxiang 已提交
36 37 38 39
        if not self.role_maker._is_collective:
            # update me. currently, if parameter server is used
            # graph execution optimizer can not be applied
            return False
40 41 42 43 44 45 46 47 48 49 50
        return True

    def backward(self,
                 loss,
                 startup_program=None,
                 parameter_list=None,
                 no_grad_set=None,
                 callbacks=None):
        pass

    # should fix the variable 
51
    def _setup_nccl_op(self, startup_program, main_program, build_strategy):
52
        trainer_endpoints = self.role_maker.get_trainer_endpoints()
53
        trainers = trainer_endpoints
54 55 56 57 58
        trainer_id = self.role_maker.worker_index()
        current_endpoint = self.role_maker.get_trainer_endpoints()[trainer_id]
        trainer_endpoints_env = ",".join(trainer_endpoints)
        trainers_num = self.role_maker.worker_num()
        if trainer_id == 0:
D
Dong Daxiang 已提交
59 60 61
            other_trainer_endpoints = trainer_endpoints[:]
            other_trainer_endpoints.remove(current_endpoint)
            wait_server_ready(other_trainer_endpoints)
62 63
        nccl_id_var = startup_program.global_block().create_var(
            name="NCCLID", persistable=True, type=core.VarDesc.VarType.RAW)
64
        for i in range(1, build_strategy.nccl_comm_num):
65 66 67 68 69
            startup_program.global_block().create_var(
                name="NCCLID_{}".format(i),
                persistable=True,
                type=core.VarDesc.VarType.RAW)

70 71
        if build_strategy.use_hierarchical_allreduce:
            for i in range(0, build_strategy.nccl_comm_num):
72 73 74 75 76 77 78 79 80 81 82 83 84 85
                startup_program.global_block().create_var(
                    name="Hierarchical_inter_NCCLID_{}".format(i),
                    persistable=True,
                    type=core.VarDesc.VarType.RAW)
                startup_program.global_block().create_var(
                    name="Hierarchical_exter_NCCLID_{}".format(i),
                    persistable=True,
                    type=core.VarDesc.VarType.RAW)

        startup_program.global_block().append_op(
            type="gen_nccl_id",
            inputs={},
            outputs={"NCCLID": nccl_id_var},
            attrs={
D
Dong Daxiang 已提交
86
                "trainers": trainer_endpoints,
87
                "trainer_id": trainer_id,
88
                "nccl_comm_num": build_strategy.nccl_comm_num,
89
                "use_hierarchical_allreduce":
90
                build_strategy.use_hierarchical_allreduce,
91
                "hierarchical_allreduce_inter_ranks":
92
                build_strategy.hierarchical_allreduce_inter_nranks
93 94 95
            })

    def _try_to_compile(self, startup_program, main_program, loss):
96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126
        import copy
        dist_strategy = self.user_defined_strategy
        local_build_strategy = paddle.fluid.BuildStrategy()
        local_build_strategy.enable_sequential_execution = \
                    dist_strategy.build_strategy.enable_sequential_execution
        local_build_strategy.fuse_elewise_add_act_ops = \
                    dist_strategy.build_strategy.fuse_elewise_add_act_ops
        local_build_strategy.fuse_bn_act_ops = \
                    dist_strategy.build_strategy.fuse_bn_act_ops
        local_build_strategy.enable_auto_fusion = \
                    dist_strategy.build_strategy.enable_auto_fusion
        local_build_strategy.fuse_relu_depthwise_conv = \
                    dist_strategy.build_strategy.fuse_relu_depthwise_conv
        local_build_strategy.fuse_broadcast_ops = \
                    dist_strategy.build_strategy.fuse_broadcast_ops
        local_build_strategy.fuse_all_optimizer_ops = \
                    dist_strategy.build_strategy.fuse_all_optimizer_ops
        local_build_strategy.enable_inplace = \
                    dist_strategy.build_strategy.enable_inplace
        local_build_strategy.use_hierarchical_allreduce = \
                    dist_strategy.use_hierarchical_allreduce
        local_build_strategy.hierarchical_allreduce_inter_nranks = \
                    dist_strategy.hierarchical_allreduce_inter_nranks
        local_build_strategy.sync_batch_norm = \
                    dist_strategy.sync_batch_norm
        local_build_strategy.fuse_all_reduce_ops = \
                    dist_strategy.fuse_all_reduce_ops
        local_build_strategy.nccl_comm_num = \
                    dist_strategy.nccl_comm_num

        exe_strategy = self.user_defined_strategy.execution_strategy
127
        node_num = self.role_maker.worker_num()
128

129 130 131 132 133
        if self.role_maker._is_collective:
            assert node_num >= 1, "nccl2 node_num must >= 1, now:{}" % node_num

        if node_num <= 1:
            # local mode
134
            if local_build_strategy.nccl_comm_num > 1:
135
                logging.warn("set nccl_comm_num=1 since you only have 1 node.")
136
            local_build_strategy.nccl_comm_num = 1
137

138
            if local_build_strategy.use_hierarchical_allreduce:
139 140 141
                logging.warn(
                    "set hierachical_allreduce=False since you only have 1 node."
                )
142
            local_build_strategy.use_hierarchical_allreduce = False
143

144
        sync_allreduce = dist_strategy.sync_nccl_allreduce
145
        if sync_allreduce:
146 147 148 149 150 151
            paddle.fluid.framework.set_flags({
                "FLAGS_sync_nccl_allreduce": True
            })
            exe_strategy.num_threads = local_build_strategy.nccl_comm_num + 1
            if local_build_strategy.use_hierarchical_allreduce:
                exe_strategy.num_threads = 2 * local_build_strategy.nccl_comm_num + 1
152 153 154
            if exe_strategy.num_threads > 4:
                logging.warn(
                    "if you use hierachical_allreduce or "
155
                    "with multi nccl comm, please set distributed_strategy.sync_nccl_allreduce=False"
156 157
                )

158
        sync_batch_norm = local_build_strategy.sync_batch_norm
159
        if sync_batch_norm:
160 161
            local_build_strategy.nccl_comm_num = 1
            local_build_strategy.use_hierarchical_allreduce = False
162 163 164 165 166 167 168
            exe_strategy.num_threads = 1
            logging.warn(
                "use sync_batch_norm will hang when set num_threads > 1, so "
                "set num_threads=1, nccl_comm_num=1, hierachical_allreduce=False."
            )

        # TODO(guru4elephant): should be an independent optimizer
169
        self._setup_nccl_op(startup_program, main_program, local_build_strategy)
170

171 172 173
        local_build_strategy.num_trainers = self.role_maker.worker_num()
        local_build_strategy.trainer_id = self.role_maker.worker_index()
        local_build_strategy.trainers_endpoints = self.role_maker.get_trainer_endpoints(
174
        )
175
        local_build_strategy.enable_backward_optimizer_op_deps = True
176 177 178 179 180

        self._compiled_program = compiler.CompiledProgram(main_program)

        self._compiled_program.with_data_parallel(
            loss_name=loss.name,
181
            build_strategy=local_build_strategy,
182 183 184 185 186
            exec_strategy=exe_strategy,
            share_vars_from=None)

        return self._compiled_program

D
Dong Daxiang 已提交
187 188 189 190
    def _disable_strategy(self, dist_strategy):
        # TODO(guru4elephant): should close all PE related flags here
        pass

191 192 193 194 195 196 197 198 199
    def minimize(self,
                 loss,
                 startup_program=None,
                 parameter_list=None,
                 no_grad_set=None):
        if startup_program == None:
            startup_program = paddle.default_startup_program()
        compiled_program = self._try_to_compile(startup_program,
                                                loss.block.program, loss)
200
        loss.block.program._graph = compiled_program
201 202 203

        # just return self.optimizer_ops and self.param_grads
        return None, None