large_scale_kv.h 7.6 KB
Newer Older
T
tangwei12 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26
// Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#pragma once

#include <ThreadPool.h>
#include <functional>
#include <future>  // NOLINT
#include <memory>
#include <string>
#include <thread>  // NOLINT
#include <unordered_map>
#include <unordered_set>
#include <utility>
#include <vector>
27
#include "gflags/gflags.h"
T
tangwei12 已提交
28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50

#include "paddle/fluid/distributed/common/utils.h"
#include "paddle/fluid/distributed/table/depends/initializers.h"
#include "paddle/fluid/framework/generator.h"
#include "paddle/fluid/framework/lod_tensor.h"
#include "paddle/fluid/framework/rw_lock.h"
#include "paddle/fluid/framework/selected_rows.h"
#include "paddle/fluid/framework/tensor.h"
#include "paddle/fluid/framework/threadpool.h"
#include "paddle/fluid/framework/variable.h"
#include "paddle/fluid/platform/device_context.h"
#include "paddle/fluid/platform/enforce.h"
#include "paddle/fluid/platform/place.h"
#include "paddle/fluid/platform/port.h"
#include "paddle/fluid/string/printf.h"
#include "paddle/fluid/string/string_helper.h"

namespace paddle {
namespace distributed {

enum Mode { training, infer };

struct VALUE {
T
tangwei12 已提交
51 52
  explicit VALUE(size_t length)
      : length_(length),
53
        count_(0),
T
tangwei12 已提交
54
        unseen_days_(0),
55 56
        need_save_(false),
        is_entry_(false) {
T
tangwei12 已提交
57
    data_.resize(length);
58
    memset(data_.data(), 0, sizeof(float) * length);
T
tangwei12 已提交
59 60
  }

T
tangwei12 已提交
61 62
  size_t length_;
  std::vector<float> data_;
T
tangwei12 已提交
63
  int count_;
64 65 66
  int unseen_days_;  // use to check knock-out
  bool need_save_;   // whether need to save
  bool is_entry_;    // whether knock-in
T
tangwei12 已提交
67 68
};

69 70 71 72 73
inline bool count_entry(std::shared_ptr<VALUE> value, int threshold) {
  return value->count_ >= threshold;
}

inline bool probility_entry(std::shared_ptr<VALUE> value, float threshold) {
T
tangwei12 已提交
74
  UniformInitializer uniform = UniformInitializer({"uniform", "0", "0", "1"});
75 76 77
  return uniform.GetValue() >= threshold;
}

T
tangwei12 已提交
78 79
class ValueBlock {
 public:
T
tangwei12 已提交
80 81 82 83 84 85 86 87 88 89 90 91
  explicit ValueBlock(const std::vector<std::string> &value_names,
                      const std::vector<int> &value_dims,
                      const std::vector<int> &value_offsets,
                      const std::unordered_map<std::string, int> &value_idx,
                      const std::vector<std::string> &init_attrs,
                      const std::string &entry_attr)
      : value_names_(value_names),
        value_dims_(value_dims),
        value_offsets_(value_offsets),
        value_idx_(value_idx) {
    for (int x = 0; x < value_dims.size(); ++x) {
      value_length_ += value_dims[x];
T
tangwei12 已提交
92 93
    }

T
tangwei12 已提交
94 95
    // for Entry
    {
T
tangwei12 已提交
96
      auto slices = string::split_string<std::string>(entry_attr, ":");
97 98
      if (slices[0] == "none") {
        entry_func_ = std::bind(&count_entry, std::placeholders::_1, 0);
T
tangwei12 已提交
99
      } else if (slices[0] == "count_filter_entry") {
100 101
        int threshold = std::stoi(slices[1]);
        entry_func_ = std::bind(&count_entry, std::placeholders::_1, threshold);
T
tangwei12 已提交
102
      } else if (slices[0] == "probability_entry") {
103
        float threshold = std::stof(slices[1]);
T
tangwei12 已提交
104
        entry_func_ =
105
            std::bind(&probility_entry, std::placeholders::_1, threshold);
T
tangwei12 已提交
106
      } else {
107
        PADDLE_THROW(platform::errors::InvalidArgument(
T
tangwei12 已提交
108 109
            "Not supported Entry Type : %s, Only support [CountFilterEntry, "
            "ProbabilityEntry]",
110
            slices[0]));
T
tangwei12 已提交
111 112
      }
    }
T
tangwei12 已提交
113 114 115 116 117 118 119 120 121 122 123 124 125 126 127

    // for Initializer
    {
      for (auto &attr : init_attrs) {
        auto slices = string::split_string<std::string>(attr, "&");

        if (slices[0] == "gaussian_random") {
          initializers_.emplace_back(
              std::make_shared<GaussianInitializer>(slices));
        } else if (slices[0] == "fill_constant") {
          initializers_.emplace_back(
              std::make_shared<FillConstantInitializer>(slices));
        } else if (slices[0] == "uniform_random") {
          initializers_.emplace_back(
              std::make_shared<UniformInitializer>(slices));
C
Chengmo 已提交
128 129 130
        } else if (slices[0] == "truncated_gaussian_random") {
          initializers_.emplace_back(
              std::make_shared<TruncatedGaussianInitializer>(slices));
T
tangwei12 已提交
131 132 133 134 135 136
        } else {
          PADDLE_THROW(platform::errors::InvalidArgument(
              "%s can not be supported", attr));
        }
      }
    }
T
tangwei12 已提交
137 138 139 140
  }

  ~ValueBlock() {}

T
tangwei12 已提交
141
  std::vector<float *> Get(const uint64_t &id,
142 143
                           const std::vector<std::string> &value_names,
                           const std::vector<int> &value_dims) {
T
tangwei12 已提交
144 145 146 147
    auto pts = std::vector<float *>();
    pts.reserve(value_names.size());
    auto &values = values_.at(id);
    for (int i = 0; i < static_cast<int>(value_names.size()); i++) {
148 149 150
      PADDLE_ENFORCE_EQ(
          value_dims[i], value_dims_[i],
          platform::errors::InvalidArgument("value dims is not match"));
T
tangwei12 已提交
151 152
      pts.push_back(values->data_.data() +
                    value_offsets_.at(value_idx_.at(value_names[i])));
T
tangwei12 已提交
153
    }
T
tangwei12 已提交
154
    return pts;
T
tangwei12 已提交
155 156
  }

157
  // pull
158 159
  float *Init(const uint64_t &id, const bool with_update = true,
              const int counter = 1) {
160 161 162 163 164
    if (!Has(id)) {
      values_[id] = std::make_shared<VALUE>(value_length_);
    }

    auto &value = values_.at(id);
T
tangwei12 已提交
165

166
    if (with_update) {
167
      AttrUpdate(value, counter);
168 169 170
    }

    return value->data_.data();
T
tangwei12 已提交
171 172
  }

173
  void AttrUpdate(std::shared_ptr<VALUE> value, const int counter) {
174 175
    // update state
    value->unseen_days_ = 0;
176
    value->count_ += counter;
177 178 179 180 181 182 183 184 185

    if (!value->is_entry_) {
      value->is_entry_ = entry_func_(value);
      if (value->is_entry_) {
        // initialize
        for (int x = 0; x < value_names_.size(); ++x) {
          initializers_[x]->GetValue(value->data_.data() + value_offsets_[x],
                                     value_dims_[x]);
        }
T
tangwei12 已提交
186
        value->need_save_ = true;
T
tangwei12 已提交
187
      }
T
tangwei12 已提交
188 189
    } else {
      value->need_save_ = true;
T
tangwei12 已提交
190
    }
191 192

    return;
T
tangwei12 已提交
193 194
  }

195 196 197 198 199 200 201 202 203
  // dont jude if (has(id))
  float *Get(const uint64_t &id) {
    auto &value = values_.at(id);
    return value->data_.data();
  }

  // for load, to reset count, unseen_days
  std::shared_ptr<VALUE> GetValue(const uint64_t &id) { return values_.at(id); }

T
tangwei12 已提交
204
  bool GetEntry(const uint64_t &id) {
205
    auto &value = values_.at(id);
T
tangwei12 已提交
206
    return value->is_entry_;
T
tangwei12 已提交
207 208
  }

209 210 211 212
  void SetEntry(const uint64_t &id, const bool state) {
    auto &value = values_.at(id);
    value->is_entry_ = state;
  }
T
tangwei12 已提交
213

214 215 216 217 218 219 220 221 222
  void Shrink(const int threshold) {
    for (auto iter = values_.begin(); iter != values_.end();) {
      auto &value = iter->second;
      value->unseen_days_++;
      if (value->unseen_days_ >= threshold) {
        iter = values_.erase(iter);
      } else {
        ++iter;
      }
T
tangwei12 已提交
223
    }
224
    return;
T
tangwei12 已提交
225 226 227 228 229 230 231 232 233 234 235 236 237
  }

 private:
  bool Has(const uint64_t id) {
    auto got = values_.find(id);
    if (got == values_.end()) {
      return false;
    } else {
      return true;
    }
  }

 public:
T
tangwei12 已提交
238 239
  std::unordered_map<uint64_t, std::shared_ptr<VALUE>> values_;
  size_t value_length_ = 0;
T
tangwei12 已提交
240 241

 private:
T
tangwei12 已提交
242 243 244 245 246
  const std::vector<std::string> &value_names_;
  const std::vector<int> &value_dims_;
  const std::vector<int> &value_offsets_;
  const std::unordered_map<std::string, int> &value_idx_;

247
  std::function<bool(std::shared_ptr<VALUE>)> entry_func_;
T
tangwei12 已提交
248
  std::vector<std::shared_ptr<Initializer>> initializers_;
T
tangwei12 已提交
249 250 251 252
};

}  // namespace distributed
}  // namespace paddle