rpc_server.cc 3.2 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#include <fstream>
#include <iostream>
#include <limits>
#include <string>

20
#include "paddle/fluid/operators/distributed/rpc_server.h"
21 22 23

namespace paddle {
namespace operators {
24
namespace distributed {
25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45

void RPCServer::ShutDown() {
  LOG(INFO) << "RPCServer ShutDown ";
  ShutDownImpl();

  exit_flag_ = true;
  barrier_cond_.notify_all();
  rpc_cond_.notify_all();
}

void RPCServer::SavePort() const {
  auto file_path = string::Sprintf("/tmp/paddle.%d.port", ::getpid());
  std::ofstream port_file;
  port_file.open(file_path);
  port_file << selected_port_;
  port_file.close();
  VLOG(4) << "selected port written to " << file_path;
}

void RPCServer::WaitBarrier(const std::string& rpc_name) {
  std::unique_lock<std::mutex> lock(this->mutex_);
W
Wu Yi 已提交
46
  barrier_cond_.wait(lock, [this, &rpc_name] {
47 48 49
    return (barrier_counter_[rpc_name] >= client_num_ || exit_flag_.load());
  });

W
Wu Yi 已提交
50 51
  VLOG(3) << "batch_barrier_: " << rpc_name << " "
          << barrier_counter_[rpc_name];
52 53 54
}

void RPCServer::IncreaseBatchBarrier(const std::string rpc_name) {
W
Wu Yi 已提交
55
  VLOG(4) << "RPCServer begin IncreaseBatchBarrier " << rpc_name;
56
  int b = 0;
W
Wu Yi 已提交
57 58
  std::unique_lock<std::mutex> lock(mutex_);
  b = ++barrier_counter_[rpc_name];
59
  if (b >= client_num_) {
W
Wu Yi 已提交
60
    lock.unlock();
61
    barrier_cond_.notify_all();
W
Wu Yi 已提交
62
    lock.lock();
63 64 65
  }
}

W
Wu Yi 已提交
66 67 68 69 70 71 72 73
void RPCServer::DecreaseClientNum() {
  {
    std::unique_lock<std::mutex> lock(mutex_);
    client_num_--;
  }
  barrier_cond_.notify_all();
}

74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103
void RPCServer::ResetBarrierCounter() {
  VLOG(3) << "RPCServer ResetBarrierCounter ";
  std::unique_lock<std::mutex> lock(mutex_);
  for (auto& t : barrier_counter_) {
    t.second = 0;
  }
}

void RPCServer::RegisterRPC(const std::string& rpc_name,
                            RequestHandler* handler, int thread_num) {
  rpc_call_map_[rpc_name] = handler;
  rpc_thread_num_[rpc_name] = thread_num;

  static int cond = -1;
  rpc_cond_map_[rpc_name] = ++cond;
  VLOG(4) << "RegisterRPC rpc_name:" << rpc_name << ", handler:" << handler
          << ", cond:" << rpc_cond_map_[rpc_name];
}

void RPCServer::SetCond(const std::string& rpc_name) {
  VLOG(3) << "RPCServer SetCond " << rpc_name;
  {
    std::unique_lock<std::mutex> lock(mutex_);
    cur_cond_ = rpc_cond_map_[rpc_name];
  }

  rpc_cond_.notify_all();
}

void RPCServer::WaitCond(const std::string& rpc_name) {
W
Wu Yi 已提交
104
  VLOG(4) << "RPCServer WaitCond " << rpc_name;
105 106 107 108 109 110 111 112 113 114 115
  int cond = 0;
  {
    std::unique_lock<std::mutex> lock(mutex_);
    cond = rpc_cond_map_[rpc_name];
  }

  std::unique_lock<std::mutex> lock(mutex_);
  rpc_cond_.wait(
      lock, [=] { return (cur_cond_.load() == cond || exit_flag_.load()); });
}

116
}  // namespace distributed
117 118
}  // namespace operators
}  // namespace paddle