common_graph_table.cc 88.7 KB
Newer Older
S
seemingwang 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14
// Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

15
#include "paddle/fluid/distributed/ps/table/common_graph_table.h"
16

S
seemingwang 已提交
17
#include <time.h>
18

S
seemingwang 已提交
19
#include <algorithm>
20
#include <chrono>
S
seemingwang 已提交
21 22
#include <set>
#include <sstream>
23

D
danleifeng 已提交
24
#include "gflags/gflags.h"
S
seemingwang 已提交
25
#include "paddle/fluid/distributed/common/utils.h"
26
#include "paddle/fluid/distributed/ps/table/graph/graph_node.h"
L
lxsbupt 已提交
27 28
#include "paddle/fluid/framework/fleet/fleet_wrapper.h"
#include "paddle/fluid/framework/fleet/heter_ps/graph_gpu_wrapper.h"
29
#include "paddle/fluid/framework/generator.h"
D
danleifeng 已提交
30 31
#include "paddle/fluid/framework/io/fs.h"
#include "paddle/fluid/platform/timer.h"
S
seemingwang 已提交
32 33
#include "paddle/fluid/string/printf.h"
#include "paddle/fluid/string/string_helper.h"
34

D
danleifeng 已提交
35
DECLARE_bool(graph_load_in_parallel);
L
lxsbupt 已提交
36 37 38
DECLARE_bool(graph_get_neighbor_id);
DECLARE_int32(gpugraph_storage_mode);
DECLARE_uint64(gpugraph_slot_feasign_max_num);
D
danleifeng 已提交
39

S
seemingwang 已提交
40 41 42
namespace paddle {
namespace distributed {

43
#ifdef PADDLE_WITH_HETERPS
44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59
int32_t GraphTable::Load_to_ssd(const std::string &path,
                                const std::string &param) {
  bool load_edge = (param[0] == 'e');
  bool load_node = (param[0] == 'n');
  if (load_edge) {
    bool reverse_edge = (param[1] == '<');
    std::string edge_type = param.substr(2);
    return this->load_edges_to_ssd(path, reverse_edge, edge_type);
  }
  if (load_node) {
    std::string node_type = param.substr(1);
    return this->load_nodes(path, node_type);
  }
  return 0;
}

D
danleifeng 已提交
60
paddle::framework::GpuPsCommGraphFea GraphTable::make_gpu_ps_graph_fea(
L
lxsbupt 已提交
61 62 63 64 65 66 67 68 69
    int gpu_id, std::vector<uint64_t> &node_ids, int slot_num) {
  size_t shard_num = 64;
  std::vector<std::vector<uint64_t>> bags(shard_num);
  std::vector<uint64_t> feature_array[shard_num];
  std::vector<uint8_t> slot_id_array[shard_num];
  std::vector<uint64_t> node_id_array[shard_num];
  std::vector<paddle::framework::GpuPsFeaInfo> node_fea_info_array[shard_num];
  for (size_t i = 0; i < shard_num; i++) {
    auto predsize = node_ids.size() / shard_num;
D
danleifeng 已提交
70
    bags[i].reserve(predsize * 1.2);
L
lxsbupt 已提交
71 72 73 74
    feature_array[i].reserve(predsize * 1.2 * slot_num);
    slot_id_array[i].reserve(predsize * 1.2 * slot_num);
    node_id_array[i].reserve(predsize * 1.2);
    node_fea_info_array[i].reserve(predsize * 1.2);
D
danleifeng 已提交
75 76 77
  }

  for (auto x : node_ids) {
L
lxsbupt 已提交
78
    int location = x % shard_num;
79 80
    bags[location].push_back(x);
  }
D
danleifeng 已提交
81

82
  std::vector<std::future<int>> tasks;
L
lxsbupt 已提交
83 84 85 86 87 88 89
  if (slot_feature_num_map_.size() == 0) {
    slot_feature_num_map_.resize(slot_num);
    for (int k = 0; k < slot_num; ++k) {
      slot_feature_num_map_[k] = 0;
    }
  }

90
  for (size_t i = 0; i < bags.size(); i++) {
91
    if (bags[i].size() > 0) {
L
lxsbupt 已提交
92
      tasks.push_back(_cpu_worker_pool[gpu_id]->enqueue([&, i, this]() -> int {
D
danleifeng 已提交
93 94 95
        uint64_t node_id;
        paddle::framework::GpuPsFeaInfo x;
        std::vector<uint64_t> feature_ids;
96
        for (size_t j = 0; j < bags[i].size(); j++) {
97
          // TODO(danleifeng): use FEATURE_TABLE instead
D
danleifeng 已提交
98 99
          Node *v = find_node(1, bags[i][j]);
          node_id = bags[i][j];
100
          if (v == NULL) {
D
danleifeng 已提交
101 102 103
            x.feature_size = 0;
            x.feature_offset = 0;
            node_fea_info_array[i].push_back(x);
104
          } else {
D
danleifeng 已提交
105 106 107 108
            // x <- v
            x.feature_offset = feature_array[i].size();
            int total_feature_size = 0;
            for (int k = 0; k < slot_num; ++k) {
L
lxsbupt 已提交
109 110 111 112
              auto feature_ids_size =
                  v->get_feature_ids(k, feature_array[i], slot_id_array[i]);
              if (slot_feature_num_map_[k] < feature_ids_size) {
                slot_feature_num_map_[k] = feature_ids_size;
D
danleifeng 已提交
113
              }
L
lxsbupt 已提交
114
              total_feature_size += feature_ids_size;
D
danleifeng 已提交
115 116 117 118 119 120 121 122 123 124
            }
            x.feature_size = total_feature_size;
            node_fea_info_array[i].push_back(x);
          }
          node_id_array[i].push_back(node_id);
        }
        return 0;
      }));
    }
  }
125
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
L
lxsbupt 已提交
126 127 128 129 130 131 132 133 134

  std::stringstream ss;
  for (int k = 0; k < slot_num; ++k) {
    ss << slot_feature_num_map_[k] << " ";
  }
  VLOG(0) << "slot_feature_num_map: " << ss.str();

  tasks.clear();

D
danleifeng 已提交
135 136
  paddle::framework::GpuPsCommGraphFea res;
  uint64_t tot_len = 0;
L
lxsbupt 已提交
137
  for (size_t i = 0; i < shard_num; i++) {
D
danleifeng 已提交
138 139 140 141 142 143
    tot_len += feature_array[i].size();
  }
  VLOG(0) << "Loaded feature table on cpu, feature_list_size[" << tot_len
          << "] node_ids_size[" << node_ids.size() << "]";
  res.init_on_cpu(tot_len, (unsigned int)node_ids.size(), slot_num);
  unsigned int offset = 0, ind = 0;
L
lxsbupt 已提交
144 145 146 147 148 149 150 151 152 153 154 155 156 157 158
  for (size_t i = 0; i < shard_num; i++) {
    tasks.push_back(
        _cpu_worker_pool[gpu_id]->enqueue([&, i, ind, offset, this]() -> int {
          auto start = ind;
          for (size_t j = 0; j < node_id_array[i].size(); j++) {
            res.node_list[start] = node_id_array[i][j];
            res.fea_info_list[start] = node_fea_info_array[i][j];
            res.fea_info_list[start++].feature_offset += offset;
          }
          for (size_t j = 0; j < feature_array[i].size(); j++) {
            res.feature_list[offset + j] = feature_array[i][j];
            res.slot_id_list[offset + j] = slot_id_array[i][j];
          }
          return 0;
        }));
D
danleifeng 已提交
159
    offset += feature_array[i].size();
L
lxsbupt 已提交
160
    ind += node_id_array[i].size();
D
danleifeng 已提交
161
  }
L
lxsbupt 已提交
162
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
D
danleifeng 已提交
163 164 165 166
  return res;
}

paddle::framework::GpuPsCommGraph GraphTable::make_gpu_ps_graph(
L
lxsbupt 已提交
167
    int idx, const std::vector<uint64_t> &ids) {
D
danleifeng 已提交
168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197
  std::vector<std::vector<uint64_t>> bags(task_pool_size_);
  for (int i = 0; i < task_pool_size_; i++) {
    auto predsize = ids.size() / task_pool_size_;
    bags[i].reserve(predsize * 1.2);
  }
  for (auto x : ids) {
    int location = x % shard_num % task_pool_size_;
    bags[location].push_back(x);
  }

  std::vector<std::future<int>> tasks;
  std::vector<uint64_t> node_array[task_pool_size_];  // node id list
  std::vector<paddle::framework::GpuPsNodeInfo> info_array[task_pool_size_];
  std::vector<uint64_t> edge_array[task_pool_size_];  // edge id list

  for (size_t i = 0; i < bags.size(); i++) {
    if (bags[i].size() > 0) {
      tasks.push_back(_shards_task_pool[i]->enqueue([&, i, this]() -> int {
        node_array[i].resize(bags[i].size());
        info_array[i].resize(bags[i].size());
        edge_array[i].reserve(bags[i].size());

        for (size_t j = 0; j < bags[i].size(); j++) {
          auto node_id = bags[i][j];
          node_array[i][j] = node_id;
          Node *v = find_node(0, idx, node_id);
          if (v != nullptr) {
            info_array[i][j].neighbor_offset = edge_array[i].size();
            info_array[i][j].neighbor_size = v->get_neighbor_size();
            for (size_t k = 0; k < v->get_neighbor_size(); k++) {
198 199
              edge_array[i].push_back(v->get_neighbor_id(k));
            }
D
danleifeng 已提交
200 201 202
          } else {
            info_array[i][j].neighbor_offset = 0;
            info_array[i][j].neighbor_size = 0;
203 204 205 206 207 208
          }
        }
        return 0;
      }));
    }
  }
209
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
D
danleifeng 已提交
210

S
seemingwang 已提交
211
  int64_t tot_len = 0;
212
  for (int i = 0; i < task_pool_size_; i++) {
213 214
    tot_len += edge_array[i].size();
  }
D
danleifeng 已提交
215 216

  paddle::framework::GpuPsCommGraph res;
S
seemingwang 已提交
217 218
  res.init_on_cpu(tot_len, ids.size());
  int64_t offset = 0, ind = 0;
219
  for (int i = 0; i < task_pool_size_; i++) {
220
    for (size_t j = 0; j < node_array[i].size(); j++) {
221
      res.node_list[ind] = node_array[i][j];
D
danleifeng 已提交
222 223
      res.node_info_list[ind] = info_array[i][j];
      res.node_info_list[ind++].neighbor_offset += offset;
224
    }
225
    for (size_t j = 0; j < edge_array[i].size(); j++) {
226 227 228 229 230 231
      res.neighbor_list[offset + j] = edge_array[i][j];
    }
    offset += edge_array[i].size();
  }
  return res;
}
232

233
int32_t GraphTable::add_node_to_ssd(
D
danleifeng 已提交
234
    int type_id, int idx, uint64_t src_id, char *data, int len) {
235
  if (_db != NULL) {
D
danleifeng 已提交
236
    char ch[sizeof(int) * 2 + sizeof(uint64_t)];
237 238
    memcpy(ch, &type_id, sizeof(int));
    memcpy(ch + sizeof(int), &idx, sizeof(int));
D
danleifeng 已提交
239
    memcpy(ch + sizeof(int) * 2, &src_id, sizeof(uint64_t));
240
    std::string str;
241 242
    if (_db->get(src_id % shard_num % task_pool_size_,
                 ch,
D
danleifeng 已提交
243
                 sizeof(int) * 2 + sizeof(uint64_t),
244
                 str) == 0) {
L
lxsbupt 已提交
245 246
      const uint64_t *stored_data =
          reinterpret_cast<const uint64_t *>(str.c_str());  // NOLINT
D
danleifeng 已提交
247 248 249 250
      int n = str.size() / sizeof(uint64_t);
      char *new_data = new char[n * sizeof(uint64_t) + len];
      memcpy(new_data, stored_data, n * sizeof(uint64_t));
      memcpy(new_data + n * sizeof(uint64_t), data, len);
251 252
      _db->put(src_id % shard_num % task_pool_size_,
               ch,
D
danleifeng 已提交
253
               sizeof(int) * 2 + sizeof(uint64_t),
L
lxsbupt 已提交
254
               reinterpret_cast<char *>(new_data),
D
danleifeng 已提交
255
               n * sizeof(uint64_t) + len);
256 257
      delete[] new_data;
    } else {
258 259
      _db->put(src_id % shard_num % task_pool_size_,
               ch,
D
danleifeng 已提交
260
               sizeof(int) * 2 + sizeof(uint64_t),
L
lxsbupt 已提交
261
               reinterpret_cast<char *>(data),
262
               len);
263
    }
264
  }
265 266 267
  return 0;
}
char *GraphTable::random_sample_neighbor_from_ssd(
268
    int idx,
D
danleifeng 已提交
269
    uint64_t id,
270 271 272
    int sample_size,
    const std::shared_ptr<std::mt19937_64> rng,
    int &actual_size) {
273 274 275 276 277
  if (_db == NULL) {
    actual_size = 0;
    return NULL;
  }
  std::string str;
S
seemingwang 已提交
278
  VLOG(2) << "sample ssd for key " << id;
D
danleifeng 已提交
279
  char ch[sizeof(int) * 2 + sizeof(uint64_t)];
280 281
  memset(ch, 0, sizeof(int));
  memcpy(ch + sizeof(int), &idx, sizeof(int));
D
danleifeng 已提交
282
  memcpy(ch + sizeof(int) * 2, &id, sizeof(uint64_t));
283 284
  if (_db->get(id % shard_num % task_pool_size_,
               ch,
D
danleifeng 已提交
285
               sizeof(int) * 2 + sizeof(uint64_t),
286
               str) == 0) {
L
lxsbupt 已提交
287
    const uint64_t *data = reinterpret_cast<const uint64_t *>(str.c_str());
D
danleifeng 已提交
288
    int n = str.size() / sizeof(uint64_t);
289
    std::unordered_map<int, int> m;
D
danleifeng 已提交
290
    // std::vector<uint64_t> res;
291 292 293 294 295 296 297 298 299 300 301 302 303 304 305
    int sm_size = std::min(n, sample_size);
    actual_size = sm_size * Node::id_size;
    char *buff = new char[actual_size];
    for (int i = 0; i < sm_size; i++) {
      std::uniform_int_distribution<int> distrib(0, n - i - 1);
      int t = distrib(*rng);
      // int t = rand() % (n-i);
      int pos = 0;
      auto iter = m.find(t);
      if (iter != m.end()) {
        pos = iter->second;
      } else {
        pos = t;
      }
      auto iter2 = m.find(n - i - 1);
306

307 308 309 310 311 312
      int key2 = iter2 == m.end() ? n - i - 1 : iter2->second;
      m[t] = key2;
      m.erase(n - i - 1);
      memcpy(buff + i * Node::id_size, &data[pos], Node::id_size);
      // res.push_back(data[pos]);
    }
S
seemingwang 已提交
313
    for (int i = 0; i < actual_size; i += 8) {
L
lxsbupt 已提交
314 315
      VLOG(2) << "sampled an neighbor "
              << *reinterpret_cast<uint64_t *>(&buff[i]);
S
seemingwang 已提交
316
    }
317 318 319 320 321
    return buff;
  }
  actual_size = 0;
  return NULL;
}
322 323

int64_t GraphTable::load_graph_to_memory_from_ssd(int idx,
D
danleifeng 已提交
324 325
                                                  std::vector<uint64_t> &ids) {
  std::vector<std::vector<uint64_t>> bags(task_pool_size_);
326 327 328 329 330 331 332 333 334
  for (auto x : ids) {
    int location = x % shard_num % task_pool_size_;
    bags[location].push_back(x);
  }
  std::vector<std::future<int>> tasks;
  std::vector<int64_t> count(task_pool_size_, 0);
  for (size_t i = 0; i < bags.size(); i++) {
    if (bags[i].size() > 0) {
      tasks.push_back(_shards_task_pool[i]->enqueue([&, i, idx, this]() -> int {
D
danleifeng 已提交
335
        char ch[sizeof(int) * 2 + sizeof(uint64_t)];
336 337 338 339
        memset(ch, 0, sizeof(int));
        memcpy(ch + sizeof(int), &idx, sizeof(int));
        for (size_t k = 0; k < bags[i].size(); k++) {
          auto v = bags[i][k];
D
danleifeng 已提交
340
          memcpy(ch + sizeof(int) * 2, &v, sizeof(uint64_t));
341
          std::string str;
D
danleifeng 已提交
342
          if (_db->get(i, ch, sizeof(int) * 2 + sizeof(uint64_t), str) == 0) {
343
            count[i] += (int64_t)str.size();
344
            for (size_t j = 0; j < str.size(); j += sizeof(uint64_t)) {
L
lxsbupt 已提交
345 346
              uint64_t id =
                  *reinterpret_cast<const uint64_t *>(str.c_str() + j);
347 348 349 350 351 352 353 354 355
              add_comm_edge(idx, v, id);
            }
          }
        }
        return 0;
      }));
    }
  }

356
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
357 358 359 360 361 362 363 364 365 366 367 368
  int64_t tot = 0;
  for (auto x : count) tot += x;
  return tot;
}

void GraphTable::make_partitions(int idx, int64_t byte_size, int device_len) {
  VLOG(2) << "start to make graph partitions , byte_size = " << byte_size
          << " total memory cost = " << total_memory_cost;
  if (total_memory_cost == 0) {
    VLOG(0) << "no edges are detected,make partitions exits";
    return;
  }
369 370
  auto &weight_map = node_weight[0][idx];
  const double a = 2.0, y = 1.25, weight_param = 1.0;
371 372 373 374 375 376 377 378 379
  int64_t gb_size_by_discount = byte_size * 0.8 * device_len;
  if (gb_size_by_discount <= 0) gb_size_by_discount = 1;
  int part_len = total_memory_cost / gb_size_by_discount;
  if (part_len == 0) part_len = 1;

  VLOG(2) << "part_len = " << part_len
          << " byte size = " << gb_size_by_discount;
  partitions[idx].clear();
  partitions[idx].resize(part_len);
380
  std::vector<double> weight_cost(part_len, 0);
381
  std::vector<int64_t> memory_remaining(part_len, gb_size_by_discount);
382
  std::vector<double> score(part_len, 0);
D
danleifeng 已提交
383
  std::unordered_map<uint64_t, int> id_map;
384 385 386 387 388
  std::vector<rocksdb::Iterator *> iters;
  for (int i = 0; i < task_pool_size_; i++) {
    iters.push_back(_db->get_iterator(i));
    iters[i]->SeekToFirst();
  }
389
  size_t next = 0;
390
  while (iters.size()) {
391
    if (next >= iters.size()) {
392 393 394 395 396 397 398
      next = 0;
    }
    if (!iters[next]->Valid()) {
      iters.erase(iters.begin() + next);
      continue;
    }
    std::string key = iters[next]->key().ToString();
L
lxsbupt 已提交
399 400
    int type_idx = *(reinterpret_cast<const int *>(key.c_str()));
    int temp_idx = *(reinterpret_cast<const int *>(key.c_str() + sizeof(int)));
401
    if (type_idx != 0 || temp_idx != idx) {
402 403 404 405 406
      iters[next]->Next();
      next++;
      continue;
    }
    std::string value = iters[next]->value().ToString();
407
    std::uint64_t i_key =
L
lxsbupt 已提交
408
        *reinterpret_cast<const uint64_t *>(key.c_str() + sizeof(int) * 2);
409 410 411 412 413 414 415
    for (int i = 0; i < part_len; i++) {
      if (memory_remaining[i] < (int64_t)value.size()) {
        score[i] = -100000.0;
      } else {
        score[i] = 0;
      }
    }
416
    for (size_t j = 0; j < value.size(); j += sizeof(uint64_t)) {
L
lxsbupt 已提交
417
      uint64_t v = *(reinterpret_cast<const uint64_t *>(value.c_str() + j));
418 419 420 421 422 423
      int index = -1;
      if (id_map.find(v) != id_map.end()) {
        index = id_map[v];
        score[index]++;
      }
    }
424 425 426 427 428 429 430
    double base, weight_base = 0;
    double w = 0;
    bool has_weight = false;
    if (weight_map.find(i_key) != weight_map.end()) {
      w = weight_map[i_key];
      has_weight = true;
    }
431 432
    int index = 0;
    for (int i = 0; i < part_len; i++) {
433
      base = gb_size_by_discount - memory_remaining[i] + value.size();
434
      if (has_weight) {
435
        weight_base = weight_cost[i] + w * weight_param;
436
      } else {
437 438 439
        weight_base = 0;
      }
      score[i] -= a * y * std::pow(1.0 * base, y - 1) + weight_base;
440 441 442 443 444 445 446
      if (score[i] > score[index]) index = i;
      VLOG(2) << "score" << i << " = " << score[i] << " memory left "
              << memory_remaining[i];
    }
    id_map[i_key] = index;
    partitions[idx][index].push_back(i_key);
    memory_remaining[index] -= (int64_t)value.size();
447
    if (has_weight) weight_cost[index] += w;
448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465
    iters[next]->Next();
    next++;
  }
  for (int i = 0; i < part_len; i++) {
    if (partitions[idx][i].size() == 0) {
      partitions[idx].erase(partitions[idx].begin() + i);
      i--;
      part_len--;
      continue;
    }
    VLOG(2) << " partition " << i << " size = " << partitions[idx][i].size();
    for (auto x : partitions[idx][i]) {
      VLOG(2) << "find a id " << x;
    }
  }
  next_partition = 0;
}

466 467 468 469
void GraphTable::export_partition_files(int idx, std::string file_path) {
  int part_len = partitions[idx].size();
  if (part_len == 0) return;
  if (file_path == "") file_path = ".";
470
  if (file_path[file_path.size() - 1] != '/') {
471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494
    file_path += "/";
  }
  std::vector<std::future<int>> tasks;
  for (int i = 0; i < part_len; i++) {
    tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
        [&, i, idx, this]() -> int {
          std::string output_path =
              file_path + "partition_" + std::to_string(i);

          std::ofstream ofs(output_path);
          if (ofs.fail()) {
            VLOG(0) << "creating " << output_path << " failed";
            return 0;
          }
          for (auto x : partitions[idx][i]) {
            auto str = std::to_string(x);
            ofs.write(str.c_str(), str.size());
            ofs.write("\n", 1);
          }
          ofs.close();
          return 0;
        }));
  }

495
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
496
}
497 498
void GraphTable::clear_graph(int idx) {
  for (auto p : edge_shards[idx]) {
L
lxsbupt 已提交
499
    p->clear();
500 501 502 503 504 505 506 507
    delete p;
  }

  edge_shards[idx].clear();
  for (size_t i = 0; i < shard_num_per_server; i++) {
    edge_shards[idx].push_back(new GraphShard());
  }
}
L
lxsbupt 已提交
508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628

void GraphTable::release_graph() {
  // Before releasing graph, prepare for sampling ids and embedding keys.
  build_graph_type_keys();

  if (FLAGS_gpugraph_storage_mode ==
      paddle::framework::GpuGraphStorageMode::WHOLE_HBM) {
    build_graph_total_keys();
  }
  // clear graph
  if (FLAGS_gpugraph_storage_mode == paddle::framework::GpuGraphStorageMode::
                                         MEM_EMB_FEATURE_AND_GPU_GRAPH ||
      FLAGS_gpugraph_storage_mode == paddle::framework::GpuGraphStorageMode::
                                         SSD_EMB_AND_MEM_FEATURE_GPU_GRAPH) {
    clear_edge_shard();
  } else {
    clear_graph();
  }
}

void GraphTable::release_graph_edge() {
  if (FLAGS_gpugraph_storage_mode ==
      paddle::framework::GpuGraphStorageMode::WHOLE_HBM) {
    build_graph_total_keys();
  }
  clear_edge_shard();
}

void GraphTable::release_graph_node() {
  build_graph_type_keys();
  if (FLAGS_gpugraph_storage_mode != paddle::framework::GpuGraphStorageMode::
                                         MEM_EMB_FEATURE_AND_GPU_GRAPH &&
      FLAGS_gpugraph_storage_mode != paddle::framework::GpuGraphStorageMode::
                                         SSD_EMB_AND_MEM_FEATURE_GPU_GRAPH) {
    clear_feature_shard();
  } else {
    merge_feature_shard();
    feature_shrink_to_fit();
  }
}

void GraphTable::clear_edge_shard() {
  VLOG(0) << "begin clear edge shard";
  std::vector<std::future<int>> tasks;
  for (auto &type_shards : edge_shards) {
    for (auto &shard : type_shards) {
      tasks.push_back(
          load_node_edge_task_pool->enqueue([&shard, this]() -> int {
            delete shard;
            return 0;
          }));
    }
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  for (auto &shards : edge_shards) {
    shards.clear();
    for (size_t i = 0; i < shard_num_per_server; i++) {
      shards.push_back(new GraphShard());
    }
  }
  VLOG(0) << "finish clear edge shard";
}

void GraphTable::clear_feature_shard() {
  VLOG(0) << "begin clear feature shard";
  std::vector<std::future<int>> tasks;
  for (auto &type_shards : feature_shards) {
    for (auto &shard : type_shards) {
      tasks.push_back(
          load_node_edge_task_pool->enqueue([&shard, this]() -> int {
            delete shard;
            return 0;
          }));
    }
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  for (auto &shards : feature_shards) {
    shards.clear();
    for (size_t i = 0; i < shard_num_per_server; i++) {
      shards.push_back(new GraphShard());
    }
  }
  VLOG(0) << "finish clear feature shard";
}

void GraphTable::feature_shrink_to_fit() {
  std::vector<std::future<int>> tasks;
  for (auto &type_shards : feature_shards) {
    for (auto &shard : type_shards) {
      tasks.push_back(
          load_node_edge_task_pool->enqueue([&shard, this]() -> int {
            shard->shrink_to_fit();
            return 0;
          }));
    }
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
}

void GraphTable::merge_feature_shard() {
  VLOG(0) << "begin merge_feature_shard";
  std::vector<std::future<int>> tasks;
  for (size_t i = 0; i < feature_shards[0].size(); i++) {
    tasks.push_back(load_node_edge_task_pool->enqueue([i, this]() -> int {
      for (size_t j = 1; j < feature_shards.size(); j++) {
        feature_shards[0][i]->merge_shard(feature_shards[j][i]);
      }
      return 0;
    }));
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  feature_shards.resize(1);
}

void GraphTable::clear_graph() {
  VLOG(0) << "begin clear_graph";
  clear_edge_shard();
  clear_feature_shard();
  VLOG(0) << "finish clear_graph";
}

629
int32_t GraphTable::load_next_partition(int idx) {
630
  if (next_partition >= static_cast<int>(partitions[idx].size())) {
631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667
    VLOG(0) << "partition iteration is done";
    return -1;
  }
  clear_graph(idx);
  load_graph_to_memory_from_ssd(idx, partitions[idx][next_partition]);
  next_partition++;
  return 0;
}
int32_t GraphTable::load_edges_to_ssd(const std::string &path,
                                      bool reverse_edge,
                                      const std::string &edge_type) {
  int idx = 0;
  if (edge_type == "") {
    VLOG(0) << "edge_type not specified, loading edges to " << id_to_edge[0]
            << " part";
  } else {
    if (edge_to_id.find(edge_type) == edge_to_id.end()) {
      VLOG(0) << "edge_type " << edge_type
              << " is not defined, nothing will be loaded";
      return 0;
    }
    idx = edge_to_id[edge_type];
  }
  total_memory_cost = 0;
  auto paths = paddle::string::split_string<std::string>(path, ";");
  int64_t count = 0;
  std::string sample_type = "random";
  for (auto path : paths) {
    std::ifstream file(path);
    std::string line;
    while (std::getline(file, line)) {
      VLOG(0) << "get a line from file " << line;
      auto values = paddle::string::split_string<std::string>(line, "\t");
      count++;
      if (values.size() < 2) continue;
      auto src_id = std::stoll(values[0]);
      auto dist_ids = paddle::string::split_string<std::string>(values[1], ";");
D
danleifeng 已提交
668
      std::vector<uint64_t> dist_data;
669 670
      for (auto x : dist_ids) {
        dist_data.push_back(std::stoll(x));
D
danleifeng 已提交
671
        total_memory_cost += sizeof(uint64_t);
672
      }
673 674 675
      add_node_to_ssd(0,
                      idx,
                      src_id,
L
lxsbupt 已提交
676
                      reinterpret_cast<char *>(dist_data.data()),
677
                      static_cast<int>(dist_data.size() * sizeof(uint64_t)));
678 679 680 681 682 683 684
    }
  }
  VLOG(0) << "total memory cost = " << total_memory_cost << " bytes";
  return 0;
}

int32_t GraphTable::dump_edges_to_ssd(int idx) {
S
seemingwang 已提交
685
  VLOG(2) << "calling dump edges to ssd";
686 687 688 689 690 691 692 693
  std::vector<std::future<int64_t>> tasks;
  auto &shards = edge_shards[idx];
  for (size_t i = 0; i < shards.size(); ++i) {
    tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
        [&, i, this]() -> int64_t {
          int64_t cost = 0;
          std::vector<Node *> &v = shards[i]->get_bucket();
          for (size_t j = 0; j < v.size(); j++) {
D
danleifeng 已提交
694
            std::vector<uint64_t> s;
695
            for (size_t k = 0; k < v[j]->get_neighbor_size(); k++) {
696 697
              s.push_back(v[j]->get_neighbor_id(k));
            }
D
danleifeng 已提交
698
            cost += v[j]->get_neighbor_size() * sizeof(uint64_t);
699 700 701
            add_node_to_ssd(0,
                            idx,
                            v[j]->get_id(),
L
lxsbupt 已提交
702
                            (char *)(s.data()),  // NOLINT
D
danleifeng 已提交
703
                            s.size() * sizeof(uint64_t));
704 705 706 707 708 709 710 711 712
          }
          return cost;
        }));
  }
  for (size_t i = 0; i < tasks.size(); i++) total_memory_cost += tasks[i].get();
  return 0;
}
int32_t GraphTable::make_complementary_graph(int idx, int64_t byte_size) {
  VLOG(0) << "make_complementary_graph";
713
  const size_t fixed_size = byte_size / 8;
D
danleifeng 已提交
714
  std::vector<std::unordered_map<uint64_t, int>> count(task_pool_size_);
715 716 717 718 719 720 721 722
  std::vector<std::future<int>> tasks;
  auto &shards = edge_shards[idx];
  for (size_t i = 0; i < shards.size(); ++i) {
    tasks.push_back(
        _shards_task_pool[i % task_pool_size_]->enqueue([&, i, this]() -> int {
          std::vector<Node *> &v = shards[i]->get_bucket();
          size_t ind = i % this->task_pool_size_;
          for (size_t j = 0; j < v.size(); j++) {
S
seemingwang 已提交
723
            // size_t location = v[j]->get_id();
D
danleifeng 已提交
724
            for (size_t k = 0; k < v[j]->get_neighbor_size(); k++) {
725 726 727 728 729 730
              count[ind][v[j]->get_neighbor_id(k)]++;
            }
          }
          return 0;
        }));
  }
S
seemingwang 已提交
731
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
D
danleifeng 已提交
732 733 734
  std::unordered_map<uint64_t, int> final_count;
  std::map<int, std::vector<uint64_t>> count_to_id;
  std::vector<uint64_t> buffer;
S
seemingwang 已提交
735
  clear_graph(idx);
736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752

  for (int i = 0; i < task_pool_size_; i++) {
    for (auto &p : count[i]) {
      final_count[p.first] = final_count[p.first] + p.second;
    }
    count[i].clear();
  }
  for (auto &p : final_count) {
    count_to_id[p.second].push_back(p.first);
    VLOG(2) << p.first << " appear " << p.second << " times";
  }
  auto iter = count_to_id.rbegin();
  while (iter != count_to_id.rend() && byte_size > 0) {
    for (auto x : iter->second) {
      buffer.push_back(x);
      if (buffer.size() >= fixed_size) {
        int64_t res = load_graph_to_memory_from_ssd(idx, buffer);
S
seemingwang 已提交
753
        buffer.clear();
754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770
        byte_size -= res;
      }
      if (byte_size <= 0) break;
    }
    iter++;
  }
  if (byte_size > 0 && buffer.size() > 0) {
    int64_t res = load_graph_to_memory_from_ssd(idx, buffer);
    byte_size -= res;
  }
  std::string sample_type = "random";
  for (auto &shard : edge_shards[idx]) {
    auto bucket = shard->get_bucket();
    for (size_t i = 0; i < bucket.size(); i++) {
      bucket[i]->build_sampler(sample_type);
    }
  }
D
danleifeng 已提交
771

772 773
  return 0;
}
774
#endif
775

776
/*
777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809
int CompleteGraphSampler::run_graph_sampling() {
  pthread_rwlock_t *rw_lock = graph_table->rw_lock.get();
  pthread_rwlock_rdlock(rw_lock);
  std::cout << "in graph sampling" << std::endl;
  sample_nodes.clear();
  sample_neighbors.clear();
  sample_res.clear();
  sample_nodes.resize(gpu_num);
  sample_neighbors.resize(gpu_num);
  sample_res.resize(gpu_num);
  std::vector<std::vector<std::vector<paddle::framework::GpuPsGraphNode>>>
      sample_nodes_ex(graph_table->task_pool_size_);
  std::vector<std::vector<std::vector<int64_t>>> sample_neighbors_ex(
      graph_table->task_pool_size_);
  for (int i = 0; i < graph_table->task_pool_size_; i++) {
    sample_nodes_ex[i].resize(gpu_num);
    sample_neighbors_ex[i].resize(gpu_num);
  }
  std::vector<std::future<int>> tasks;
  for (size_t i = 0; i < graph_table->shards.size(); ++i) {
    tasks.push_back(
        graph_table->_shards_task_pool[i % graph_table->task_pool_size_]
            ->enqueue([&, i, this]() -> int {
              if (this->status == GraphSamplerStatus::terminating) return 0;
              paddle::framework::GpuPsGraphNode node;
              std::vector<Node *> &v =
                  this->graph_table->shards[i]->get_bucket();
              size_t ind = i % this->graph_table->task_pool_size_;
              for (size_t j = 0; j < v.size(); j++) {
                size_t location = v[j]->get_id() % this->gpu_num;
                node.node_id = v[j]->get_id();
                node.neighbor_size = v[j]->get_neighbor_size();
                node.neighbor_offset =
L
lxsbupt 已提交
810
                    static_cast<int>sample_neighbors_ex[ind][location].size();
811 812 813 814 815 816 817 818 819 820
                sample_nodes_ex[ind][location].emplace_back(node);
                for (int k = 0; k < node.neighbor_size; k++)
                  sample_neighbors_ex[ind][location].push_back(
                      v[j]->get_neighbor_id(k));
              }
              return 0;
            }));
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  tasks.clear();
821
  for (int i = 0; i < gpu_num; i++) {
822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848
    tasks.push_back(
        graph_table->_shards_task_pool[i % graph_table->task_pool_size_]
            ->enqueue([&, i, this]() -> int {
              if (this->status == GraphSamplerStatus::terminating) return 0;
              int total_offset = 0;
              size_t ind = i % this->graph_table->task_pool_size_;
              for (int j = 0; j < this->graph_table->task_pool_size_; j++) {
                for (size_t k = 0; k < sample_nodes_ex[j][ind].size(); k++) {
                  sample_nodes[ind].push_back(sample_nodes_ex[j][ind][k]);
                  sample_nodes[ind].back().neighbor_offset += total_offset;
                }
                size_t neighbor_size = sample_neighbors_ex[j][ind].size();
                total_offset += neighbor_size;
                for (size_t k = 0; k < neighbor_size; k++) {
                  sample_neighbors[ind].push_back(
                      sample_neighbors_ex[j][ind][k]);
                }
              }
              return 0;
            }));
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();

  if (this->status == GraphSamplerStatus::terminating) {
    pthread_rwlock_unlock(rw_lock);
    return 0;
  }
849
  for (int i = 0; i < gpu_num; i++) {
850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883
    sample_res[i].node_list = sample_nodes[i].data();
    sample_res[i].neighbor_list = sample_neighbors[i].data();
    sample_res[i].node_size = sample_nodes[i].size();
    sample_res[i].neighbor_size = sample_neighbors[i].size();
  }
  pthread_rwlock_unlock(rw_lock);
  if (this->status == GraphSamplerStatus::terminating) {
    return 0;
  }
  callback(sample_res);
  return 0;
}
void CompleteGraphSampler::init(size_t gpu_num, GraphTable *graph_table,
                                std::vector<std::string> args) {
  this->gpu_num = gpu_num;
  this->graph_table = graph_table;
}

int BasicBfsGraphSampler::run_graph_sampling() {
  pthread_rwlock_t *rw_lock = graph_table->rw_lock.get();
  pthread_rwlock_rdlock(rw_lock);
  while (rounds > 0 && status == GraphSamplerStatus::running) {
    for (size_t i = 0; i < sample_neighbors_map.size(); i++) {
      sample_neighbors_map[i].clear();
    }
    sample_neighbors_map.clear();
    std::vector<int> nodes_left(graph_table->shards.size(),
                                node_num_for_each_shard);
    std::promise<int> prom;
    std::future<int> fut = prom.get_future();
    sample_neighbors_map.resize(graph_table->task_pool_size_);
    int task_size = 0;
    std::vector<std::future<int>> tasks;
    int init_size = 0;
884 885
    //__sync_fetch_and_add
    std::function<int(int, int64_t)> bfs = [&, this](int i, int id) -> int {
886 887 888 889 890 891 892 893 894 895 896 897 898
      if (this->status == GraphSamplerStatus::terminating) {
        int task_left = __sync_sub_and_fetch(&task_size, 1);
        if (task_left == 0) {
          prom.set_value(0);
        }
        return 0;
      }
      size_t ind = i % this->graph_table->task_pool_size_;
      if (nodes_left[i] > 0) {
        auto iter = sample_neighbors_map[ind].find(id);
        if (iter == sample_neighbors_map[ind].end()) {
          Node *node = graph_table->shards[i]->find_node(id);
          if (node != NULL) {
899 900 901
            nodes_left[i]--;
            sample_neighbors_map[ind][id] = std::vector<int64_t>();
            iter = sample_neighbors_map[ind].find(id);
902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925
            size_t edge_fetch_size =
                std::min((size_t) this->edge_num_for_each_node,
                         node->get_neighbor_size());
            for (size_t k = 0; k < edge_fetch_size; k++) {
              int64_t neighbor_id = node->get_neighbor_id(k);
              int node_location = neighbor_id % this->graph_table->shard_num %
                                  this->graph_table->task_pool_size_;
              __sync_add_and_fetch(&task_size, 1);
              graph_table->_shards_task_pool[node_location]->enqueue(
                  bfs, neighbor_id % this->graph_table->shard_num, neighbor_id);
              iter->second.push_back(neighbor_id);
            }
          }
        }
      }
      int task_left = __sync_sub_and_fetch(&task_size, 1);
      if (task_left == 0) {
        prom.set_value(0);
      }
      return 0;
    };
    for (size_t i = 0; i < graph_table->shards.size(); ++i) {
      std::vector<Node *> &v = graph_table->shards[i]->get_bucket();
      if (v.size() > 0) {
L
lxsbupt 已提交
926
        int search_size = std::min(init_search_size, static_cast<int>v.size());
927 928 929 930 931 932 933
        for (int k = 0; k < search_size; k++) {
          init_size++;
          __sync_add_and_fetch(&task_size, 1);
          int64_t id = v[k]->get_id();
          graph_table->_shards_task_pool[i % graph_table->task_pool_size_]
              ->enqueue(bfs, i, id);
        }
934 935 936 937 938 939 940 941 942 943
      }  // if
    }
    if (init_size == 0) {
      prom.set_value(0);
    }
    fut.get();
    if (this->status == GraphSamplerStatus::terminating) {
      pthread_rwlock_unlock(rw_lock);
      return 0;
    }
944
    VLOG(0) << "BasicBfsGraphSampler finishes the graph searching task";
945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973
    sample_nodes.clear();
    sample_neighbors.clear();
    sample_res.clear();
    sample_nodes.resize(gpu_num);
    sample_neighbors.resize(gpu_num);
    sample_res.resize(gpu_num);
    std::vector<std::vector<std::vector<paddle::framework::GpuPsGraphNode>>>
        sample_nodes_ex(graph_table->task_pool_size_);
    std::vector<std::vector<std::vector<int64_t>>> sample_neighbors_ex(
        graph_table->task_pool_size_);
    for (int i = 0; i < graph_table->task_pool_size_; i++) {
      sample_nodes_ex[i].resize(gpu_num);
      sample_neighbors_ex[i].resize(gpu_num);
    }
    tasks.clear();
    for (size_t i = 0; i < (size_t)graph_table->task_pool_size_; ++i) {
      tasks.push_back(
          graph_table->_shards_task_pool[i]->enqueue([&, i, this]() -> int {
            if (this->status == GraphSamplerStatus::terminating) {
              return 0;
            }
            paddle::framework::GpuPsGraphNode node;
            auto iter = sample_neighbors_map[i].begin();
            size_t ind = i;
            for (; iter != sample_neighbors_map[i].end(); iter++) {
              size_t location = iter->first % this->gpu_num;
              node.node_id = iter->first;
              node.neighbor_size = iter->second.size();
              node.neighbor_offset =
L
lxsbupt 已提交
974
                  static_cast<int>sample_neighbors_ex[ind][location].size();
975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991
              sample_nodes_ex[ind][location].emplace_back(node);
              for (auto k : iter->second)
                sample_neighbors_ex[ind][location].push_back(k);
            }
            return 0;
          }));
    }

    for (size_t i = 0; i < tasks.size(); i++) {
      tasks[i].get();
      sample_neighbors_map[i].clear();
    }
    tasks.clear();
    if (this->status == GraphSamplerStatus::terminating) {
      pthread_rwlock_unlock(rw_lock);
      return 0;
    }
992
    for (size_t i = 0; i < (size_t)gpu_num; i++) {
993 994 995 996 997 998 999 1000 1001
      tasks.push_back(
          graph_table->_shards_task_pool[i % graph_table->task_pool_size_]
              ->enqueue([&, i, this]() -> int {
                if (this->status == GraphSamplerStatus::terminating) {
                  pthread_rwlock_unlock(rw_lock);
                  return 0;
                }
                int total_offset = 0;
                for (int j = 0; j < this->graph_table->task_pool_size_; j++) {
1002 1003
                  for (size_t k = 0; k < sample_nodes_ex[j][i].size(); k++) {
                    sample_nodes[i].push_back(sample_nodes_ex[j][i][k]);
1004 1005
                    sample_nodes[i].back().neighbor_offset += total_offset;
                  }
1006
                  size_t neighbor_size = sample_neighbors_ex[j][i].size();
1007 1008
                  total_offset += neighbor_size;
                  for (size_t k = 0; k < neighbor_size; k++) {
1009
                    sample_neighbors[i].push_back(sample_neighbors_ex[j][i][k]);
1010 1011 1012 1013 1014 1015 1016 1017 1018 1019
                  }
                }
                return 0;
              }));
    }
    for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
    if (this->status == GraphSamplerStatus::terminating) {
      pthread_rwlock_unlock(rw_lock);
      return 0;
    }
1020
    for (int i = 0; i < gpu_num; i++) {
1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037
      sample_res[i].node_list = sample_nodes[i].data();
      sample_res[i].neighbor_list = sample_neighbors[i].data();
      sample_res[i].node_size = sample_nodes[i].size();
      sample_res[i].neighbor_size = sample_neighbors[i].size();
    }
    pthread_rwlock_unlock(rw_lock);
    if (this->status == GraphSamplerStatus::terminating) {
      return 0;
    }
    callback(sample_res);
    rounds--;
    if (rounds > 0) {
      for (int i = 0;
           i < interval && this->status == GraphSamplerStatus::running; i++) {
        std::this_thread::sleep_for(std::chrono::seconds(1));
      }
    }
1038
    VLOG(0)<<"bfs returning";
1039 1040 1041 1042 1043 1044 1045
  }
  return 0;
}
void BasicBfsGraphSampler::init(size_t gpu_num, GraphTable *graph_table,
                                std::vector<std::string> args) {
  this->gpu_num = gpu_num;
  this->graph_table = graph_table;
1046 1047 1048 1049 1050
  init_search_size = args.size() > 0 ? std::stoi(args[0]) : 10;
  node_num_for_each_shard = args.size() > 1 ? std::stoi(args[1]) : 10;
  edge_num_for_each_node = args.size() > 2 ? std::stoi(args[2]) : 10;
  rounds = args.size() > 3 ? std::stoi(args[3]) : 1;
  interval = args.size() > 4 ? std::stoi(args[4]) : 60;
1051 1052 1053
}

#endif
1054
*/
S
seemingwang 已提交
1055 1056 1057
std::vector<Node *> GraphShard::get_batch(int start, int end, int step) {
  if (start < 0) start = 0;
  std::vector<Node *> res;
L
lxsbupt 已提交
1058
  for (int pos = start; pos < std::min(end, static_cast<int>(bucket.size()));
1059
       pos += step) {
S
seemingwang 已提交
1060 1061 1062 1063 1064 1065 1066
    res.push_back(bucket[pos]);
  }
  return res;
}

size_t GraphShard::get_size() { return bucket.size(); }

D
danleifeng 已提交
1067
int32_t GraphTable::add_comm_edge(int idx, uint64_t src_id, uint64_t dst_id) {
1068 1069 1070 1071 1072 1073
  size_t src_shard_id = src_id % shard_num;

  if (src_shard_id >= shard_end || src_shard_id < shard_start) {
    return -1;
  }
  size_t index = src_shard_id - shard_start;
1074 1075
  edge_shards[idx][index]->add_graph_node(src_id)->build_edges(false);
  edge_shards[idx][index]->add_neighbor(src_id, dst_id, 1.0);
1076 1077
  return 0;
}
1078
int32_t GraphTable::add_graph_node(int idx,
D
danleifeng 已提交
1079
                                   std::vector<uint64_t> &id_list,
1080
                                   std::vector<bool> &is_weight_list) {
1081
  auto &shards = edge_shards[idx];
1082
  size_t node_size = id_list.size();
D
danleifeng 已提交
1083
  std::vector<std::vector<std::pair<uint64_t, bool>>> batch(task_pool_size_);
1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094
  for (size_t i = 0; i < node_size; i++) {
    size_t shard_id = id_list[i] % shard_num;
    if (shard_id >= shard_end || shard_id < shard_start) {
      continue;
    }
    batch[get_thread_pool_index(id_list[i])].push_back(
        {id_list[i], i < is_weight_list.size() ? is_weight_list[i] : false});
  }
  std::vector<std::future<int>> tasks;
  for (size_t i = 0; i < batch.size(); ++i) {
    if (!batch[i].size()) continue;
1095 1096 1097 1098 1099 1100 1101 1102
    tasks.push_back(
        _shards_task_pool[i]->enqueue([&shards, &batch, i, this]() -> int {
          for (auto &p : batch[i]) {
            size_t index = p.first % this->shard_num - this->shard_start;
            shards[index]->add_graph_node(p.first)->build_edges(p.second);
          }
          return 0;
        }));
1103 1104 1105 1106 1107
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  return 0;
}

D
danleifeng 已提交
1108
int32_t GraphTable::remove_graph_node(int idx, std::vector<uint64_t> &id_list) {
1109
  size_t node_size = id_list.size();
D
danleifeng 已提交
1110
  std::vector<std::vector<uint64_t>> batch(task_pool_size_);
1111 1112 1113 1114 1115
  for (size_t i = 0; i < node_size; i++) {
    size_t shard_id = id_list[i] % shard_num;
    if (shard_id >= shard_end || shard_id < shard_start) continue;
    batch[get_thread_pool_index(id_list[i])].push_back(id_list[i]);
  }
1116
  auto &shards = edge_shards[idx];
1117 1118 1119
  std::vector<std::future<int>> tasks;
  for (size_t i = 0; i < batch.size(); ++i) {
    if (!batch[i].size()) continue;
1120 1121 1122 1123 1124 1125 1126 1127
    tasks.push_back(
        _shards_task_pool[i]->enqueue([&shards, &batch, i, this]() -> int {
          for (auto &p : batch[i]) {
            size_t index = p % this->shard_num - this->shard_start;
            shards[index]->delete_node(p);
          }
          return 0;
        }));
1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  return 0;
}

void GraphShard::clear() {
  for (size_t i = 0; i < bucket.size(); i++) {
    delete bucket[i];
  }
  bucket.clear();
  node_location.clear();
}

GraphShard::~GraphShard() { clear(); }
1142

D
danleifeng 已提交
1143
void GraphShard::delete_node(uint64_t id) {
1144 1145 1146 1147
  auto iter = node_location.find(id);
  if (iter == node_location.end()) return;
  int pos = iter->second;
  delete bucket[pos];
1148
  if (pos != static_cast<int>(bucket.size()) - 1) {
1149 1150 1151 1152 1153 1154
    bucket[pos] = bucket.back();
    node_location[bucket.back()->get_id()] = pos;
  }
  node_location.erase(id);
  bucket.pop_back();
}
D
danleifeng 已提交
1155
GraphNode *GraphShard::add_graph_node(uint64_t id) {
S
seemingwang 已提交
1156 1157 1158 1159
  if (node_location.find(id) == node_location.end()) {
    node_location[id] = bucket.size();
    bucket.push_back(new GraphNode(id));
  }
L
lxsbupt 已提交
1160
  return reinterpret_cast<GraphNode *>(bucket[node_location[id]]);
S
seemingwang 已提交
1161 1162
}

1163 1164 1165 1166 1167 1168
GraphNode *GraphShard::add_graph_node(Node *node) {
  auto id = node->get_id();
  if (node_location.find(id) == node_location.end()) {
    node_location[id] = bucket.size();
    bucket.push_back(node);
  }
L
lxsbupt 已提交
1169
  return reinterpret_cast<GraphNode *>(bucket[node_location[id]]);
1170
}
D
danleifeng 已提交
1171 1172

FeatureNode *GraphShard::add_feature_node(uint64_t id, bool is_overlap) {
S
seemingwang 已提交
1173 1174 1175
  if (node_location.find(id) == node_location.end()) {
    node_location[id] = bucket.size();
    bucket.push_back(new FeatureNode(id));
L
lxsbupt 已提交
1176
    return reinterpret_cast<FeatureNode *>(bucket[node_location[id]]);
D
danleifeng 已提交
1177 1178
  }
  if (is_overlap) {
L
lxsbupt 已提交
1179
    return reinterpret_cast<FeatureNode *>(bucket[node_location[id]]);
S
seemingwang 已提交
1180
  }
D
danleifeng 已提交
1181 1182

  return NULL;
S
seemingwang 已提交
1183 1184
}

D
danleifeng 已提交
1185
void GraphShard::add_neighbor(uint64_t id, uint64_t dst_id, float weight) {
S
seemingwang 已提交
1186 1187 1188
  find_node(id)->add_edge(dst_id, weight);
}

D
danleifeng 已提交
1189
Node *GraphShard::find_node(uint64_t id) {
S
seemingwang 已提交
1190 1191 1192 1193
  auto iter = node_location.find(id);
  return iter == node_location.end() ? nullptr : bucket[iter->second];
}

1194
GraphTable::~GraphTable() {
L
lxsbupt 已提交
1195 1196 1197
#ifdef PADDLE_WITH_GPU_GRAPH
  clear_graph();
#endif
1198 1199
}

Z
zhaocaibei123 已提交
1200
int32_t GraphTable::Load(const std::string &path, const std::string &param) {
S
seemingwang 已提交
1201 1202 1203 1204
  bool load_edge = (param[0] == 'e');
  bool load_node = (param[0] == 'n');
  if (load_edge) {
    bool reverse_edge = (param[1] == '<');
1205 1206
    std::string edge_type = param.substr(2);
    return this->load_edges(path, reverse_edge, edge_type);
S
seemingwang 已提交
1207 1208 1209 1210 1211 1212 1213 1214
  }
  if (load_node) {
    std::string node_type = param.substr(1);
    return this->load_nodes(path, node_type);
  }
  return 0;
}

D
danleifeng 已提交
1215 1216 1217
std::string GraphTable::get_inverse_etype(std::string &etype) {
  auto etype_split = paddle::string::split_string<std::string>(etype, "2");
  std::string res;
1218
  if (etype_split.size() == 3) {
D
danleifeng 已提交
1219 1220 1221 1222 1223 1224 1225
    res = etype_split[2] + "2" + etype_split[1] + "2" + etype_split[0];
  } else {
    res = etype_split[1] + "2" + etype_split[0];
  }
  return res;
}

L
lxsbupt 已提交
1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333
int32_t GraphTable::parse_type_to_typepath(
    std::string &type2files,
    std::string graph_data_local_path,
    std::vector<std::string> &res_type,
    std::unordered_map<std::string, std::string> &res_type2path) {
  auto type2files_split =
      paddle::string::split_string<std::string>(type2files, ",");
  if (type2files_split.size() == 0) {
    return -1;
  }
  for (auto one_type2file : type2files_split) {
    auto one_type2file_split =
        paddle::string::split_string<std::string>(one_type2file, ":");
    auto type = one_type2file_split[0];
    auto type_dir = one_type2file_split[1];
    res_type.push_back(type);
    res_type2path[type] = graph_data_local_path + "/" + type_dir;
  }
  return 0;
}

int32_t GraphTable::parse_edge_and_load(std::string etype2files,
                                        std::string graph_data_local_path,
                                        int part_num,
                                        bool reverse) {
  std::vector<std::string> etypes;
  std::unordered_map<std::string, std::string> edge_to_edgedir;
  int res = parse_type_to_typepath(
      etype2files, graph_data_local_path, etypes, edge_to_edgedir);
  if (res != 0) {
    VLOG(0) << "parse edge type and edgedir failed!";
    return -1;
  }
  VLOG(0) << "etypes size: " << etypes.size();
  VLOG(0) << "whether reverse: " << reverse;
  is_load_reverse_edge = reverse;
  std::string delim = ";";
  size_t total_len = etypes.size();

  std::vector<std::future<int>> tasks;
  for (size_t i = 0; i < total_len; i++) {
    tasks.push_back(
        _shards_task_pool[i % task_pool_size_]->enqueue([&, i, this]() -> int {
          std::string etype_path = edge_to_edgedir[etypes[i]];
          auto etype_path_list = paddle::framework::localfs_list(etype_path);
          std::string etype_path_str;
          if (part_num > 0 &&
              part_num < static_cast<int>(etype_path_list.size())) {
            std::vector<std::string> sub_etype_path_list(
                etype_path_list.begin(), etype_path_list.begin() + part_num);
            etype_path_str =
                paddle::string::join_strings(sub_etype_path_list, delim);
          } else {
            etype_path_str =
                paddle::string::join_strings(etype_path_list, delim);
          }
          this->load_edges(etype_path_str, false, etypes[i]);
          if (reverse) {
            std::string r_etype = get_inverse_etype(etypes[i]);
            this->load_edges(etype_path_str, true, r_etype);
          }
          return 0;
        }));
  }
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
  return 0;
}

int32_t GraphTable::parse_node_and_load(std::string ntype2files,
                                        std::string graph_data_local_path,
                                        int part_num) {
  std::vector<std::string> ntypes;
  std::unordered_map<std::string, std::string> node_to_nodedir;
  int res = parse_type_to_typepath(
      ntype2files, graph_data_local_path, ntypes, node_to_nodedir);
  if (res != 0) {
    VLOG(0) << "parse node type and nodedir failed!";
    return -1;
  }
  std::string delim = ";";
  std::string npath = node_to_nodedir[ntypes[0]];
  auto npath_list = paddle::framework::localfs_list(npath);
  std::string npath_str;
  if (part_num > 0 && part_num < static_cast<int>(npath_list.size())) {
    std::vector<std::string> sub_npath_list(npath_list.begin(),
                                            npath_list.begin() + part_num);
    npath_str = paddle::string::join_strings(sub_npath_list, delim);
  } else {
    npath_str = paddle::string::join_strings(npath_list, delim);
  }

  if (ntypes.size() == 0) {
    VLOG(0) << "node_type not specified, nothing will be loaded ";
    return 0;
  }
  if (FLAGS_graph_load_in_parallel) {
    this->load_nodes(npath_str, "");
  } else {
    for (size_t j = 0; j < ntypes.size(); j++) {
      this->load_nodes(npath_str, ntypes[j]);
    }
  }
  return 0;
}

int32_t GraphTable::load_node_and_edge_file(std::string etype2files,
                                            std::string ntype2files,
                                            std::string graph_data_local_path,
D
danleifeng 已提交
1334 1335
                                            int part_num,
                                            bool reverse) {
L
lxsbupt 已提交
1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352
  std::vector<std::string> etypes;
  std::unordered_map<std::string, std::string> edge_to_edgedir;
  int res = parse_type_to_typepath(
      etype2files, graph_data_local_path, etypes, edge_to_edgedir);
  if (res != 0) {
    VLOG(0) << "parse edge type and edgedir failed!";
    return -1;
  }
  std::vector<std::string> ntypes;
  std::unordered_map<std::string, std::string> node_to_nodedir;
  res = parse_type_to_typepath(
      ntype2files, graph_data_local_path, ntypes, node_to_nodedir);
  if (res != 0) {
    VLOG(0) << "parse node type and nodedir failed!";
    return -1;
  }

D
danleifeng 已提交
1353 1354
  VLOG(0) << "etypes size: " << etypes.size();
  VLOG(0) << "whether reverse: " << reverse;
L
lxsbupt 已提交
1355
  is_load_reverse_edge = reverse;
D
danleifeng 已提交
1356 1357 1358 1359 1360 1361 1362 1363
  std::string delim = ";";
  size_t total_len = etypes.size() + 1;  // 1 is for node

  std::vector<std::future<int>> tasks;
  for (size_t i = 0; i < total_len; i++) {
    tasks.push_back(
        _shards_task_pool[i % task_pool_size_]->enqueue([&, i, this]() -> int {
          if (i < etypes.size()) {
L
lxsbupt 已提交
1364
            std::string etype_path = edge_to_edgedir[etypes[i]];
D
danleifeng 已提交
1365 1366
            auto etype_path_list = paddle::framework::localfs_list(etype_path);
            std::string etype_path_str;
1367
            if (part_num > 0 &&
L
lxsbupt 已提交
1368
                part_num < static_cast<int>(etype_path_list.size())) {
D
danleifeng 已提交
1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382
              std::vector<std::string> sub_etype_path_list(
                  etype_path_list.begin(), etype_path_list.begin() + part_num);
              etype_path_str =
                  paddle::string::join_strings(sub_etype_path_list, delim);
            } else {
              etype_path_str =
                  paddle::string::join_strings(etype_path_list, delim);
            }
            this->load_edges(etype_path_str, false, etypes[i]);
            if (reverse) {
              std::string r_etype = get_inverse_etype(etypes[i]);
              this->load_edges(etype_path_str, true, r_etype);
            }
          } else {
L
lxsbupt 已提交
1383
            std::string npath = node_to_nodedir[ntypes[0]];
D
danleifeng 已提交
1384 1385
            auto npath_list = paddle::framework::localfs_list(npath);
            std::string npath_str;
L
lxsbupt 已提交
1386 1387
            if (part_num > 0 &&
                part_num < static_cast<int>(npath_list.size())) {
D
danleifeng 已提交
1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409
              std::vector<std::string> sub_npath_list(
                  npath_list.begin(), npath_list.begin() + part_num);
              npath_str = paddle::string::join_strings(sub_npath_list, delim);
            } else {
              npath_str = paddle::string::join_strings(npath_list, delim);
            }

            if (ntypes.size() == 0) {
              VLOG(0) << "node_type not specified, nothing will be loaded ";
              return 0;
            }
            if (FLAGS_graph_load_in_parallel) {
              this->load_nodes(npath_str, "");
            } else {
              for (size_t j = 0; j < ntypes.size(); j++) {
                this->load_nodes(npath_str, ntypes[j]);
              }
            }
          }
          return 0;
        }));
  }
1410
  for (size_t i = 0; i < tasks.size(); i++) tasks[i].get();
D
danleifeng 已提交
1411 1412 1413
  return 0;
}

S
seemingwang 已提交
1414
int32_t GraphTable::get_nodes_ids_by_ranges(
1415 1416 1417
    int type_id,
    int idx,
    std::vector<std::pair<int, int>> ranges,
D
danleifeng 已提交
1418 1419
    std::vector<uint64_t> &res) {
  std::mutex mutex;
S
seemingwang 已提交
1420 1421
  int start = 0, end, index = 0, total_size = 0;
  res.clear();
1422
  auto &shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
D
danleifeng 已提交
1423
  std::vector<std::future<size_t>> tasks;
L
lxsbupt 已提交
1424 1425
  for (size_t i = 0;
       i < shards.size() && index < static_cast<int>(ranges.size());
1426
       i++) {
1427
    end = total_size + shards[i]->get_size();
S
seemingwang 已提交
1428
    start = total_size;
1429 1430
    while (start < end && index < static_cast<int>(ranges.size())) {
      if (ranges[index].second <= start) {
S
seemingwang 已提交
1431
        index++;
1432
      } else if (ranges[index].first >= end) {
S
seemingwang 已提交
1433 1434 1435 1436 1437 1438 1439 1440
        break;
      } else {
        int first = std::max(ranges[index].first, start);
        int second = std::min(ranges[index].second, end);
        start = second;
        first -= total_size;
        second -= total_size;
        tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
D
danleifeng 已提交
1441 1442 1443 1444 1445 1446 1447 1448 1449
            [&shards, this, first, second, i, &res, &mutex]() -> size_t {
              std::vector<uint64_t> keys;
              shards[i]->get_ids_by_range(first, second, &keys);

              size_t num = keys.size();
              mutex.lock();
              res.reserve(res.size() + num);
              for (auto &id : keys) {
                res.push_back(id);
1450
                std::swap(res[rand() % res.size()],
L
lxsbupt 已提交
1451
                          res[static_cast<int>(res.size()) - 1]);
D
danleifeng 已提交
1452 1453 1454 1455
              }
              mutex.unlock();

              return num;
S
seemingwang 已提交
1456 1457 1458
            }));
      }
    }
1459
    total_size += shards[i]->get_size();
S
seemingwang 已提交
1460
  }
1461
  for (size_t i = 0; i < tasks.size(); i++) {
D
danleifeng 已提交
1462
    tasks[i].get();
S
seemingwang 已提交
1463 1464 1465 1466
  }
  return 0;
}

D
danleifeng 已提交
1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479
std::pair<uint64_t, uint64_t> GraphTable::parse_node_file(
    const std::string &path, const std::string &node_type, int idx) {
  std::ifstream file(path);
  std::string line;
  uint64_t local_count = 0;
  uint64_t local_valid_count = 0;

  int num = 0;
  std::vector<paddle::string::str_ptr> vals;
  size_t n = node_type.length();
  while (std::getline(file, line)) {
    if (strncmp(line.c_str(), node_type.c_str(), n) != 0) {
      continue;
1480
    }
D
danleifeng 已提交
1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494
    vals.clear();
    num = paddle::string::split_string_ptr(
        line.c_str() + n + 1, line.length() - n - 1, '\t', &vals);
    if (num == 0) {
      continue;
    }
    uint64_t id = std::strtoul(vals[0].ptr, NULL, 10);
    size_t shard_id = id % shard_num;
    if (shard_id >= shard_end || shard_id < shard_start) {
      VLOG(4) << "will not load " << id << " from " << path
              << ", please check id distribution";
      continue;
    }
    local_count++;
S
seemingwang 已提交
1495

D
danleifeng 已提交
1496 1497 1498 1499 1500 1501 1502
    size_t index = shard_id - shard_start;
    auto node = feature_shards[idx][index]->add_feature_node(id, false);
    if (node != NULL) {
      node->set_feature_size(feat_name[idx].size());
      for (int i = 1; i < num; ++i) {
        auto &v = vals[i];
        parse_feature(idx, v.ptr, v.len, node);
S
seemingwang 已提交
1503
      }
D
danleifeng 已提交
1504 1505 1506 1507 1508 1509 1510
    }
    local_valid_count++;
  }
  VLOG(2) << "node_type[" << node_type << "] loads " << local_count
          << " nodes from filepath->" << path;
  return {local_count, local_valid_count};
}
S
seemingwang 已提交
1511

D
danleifeng 已提交
1512 1513 1514 1515 1516 1517 1518
std::pair<uint64_t, uint64_t> GraphTable::parse_node_file(
    const std::string &path) {
  std::ifstream file(path);
  std::string line;
  uint64_t local_count = 0;
  uint64_t local_valid_count = 0;
  int idx = 0;
S
seemingwang 已提交
1519

D
danleifeng 已提交
1520 1521
  auto path_split = paddle::string::split_string<std::string>(path, "/");
  auto path_name = path_split[path_split.size() - 1];
S
seemingwang 已提交
1522

D
danleifeng 已提交
1523 1524
  int num = 0;
  std::vector<paddle::string::str_ptr> vals;
S
seemingwang 已提交
1525

D
danleifeng 已提交
1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562
  while (std::getline(file, line)) {
    vals.clear();
    num = paddle::string::split_string_ptr(
        line.c_str(), line.length(), '\t', &vals);
    if (vals.empty()) {
      continue;
    }
    std::string parse_node_type = vals[0].to_string();
    auto it = feature_to_id.find(parse_node_type);
    if (it == feature_to_id.end()) {
      VLOG(0) << parse_node_type << "type error, please check";
      continue;
    }
    idx = it->second;
    uint64_t id = std::strtoul(vals[1].ptr, NULL, 10);
    size_t shard_id = id % shard_num;
    if (shard_id >= shard_end || shard_id < shard_start) {
      VLOG(4) << "will not load " << id << " from " << path
              << ", please check id distribution";
      continue;
    }
    local_count++;

    size_t index = shard_id - shard_start;
    auto node = feature_shards[idx][index]->add_feature_node(id, false);
    if (node != NULL) {
      for (int i = 2; i < num; ++i) {
        auto &v = vals[i];
        parse_feature(idx, v.ptr, v.len, node);
      }
    }
    local_valid_count++;
  }
  VLOG(2) << local_valid_count << "/" << local_count << " nodes from filepath->"
          << path;
  return {local_count, local_valid_count};
}
S
seemingwang 已提交
1563

1564
// // TODO(danleifeng): opt load all node_types in once reading
D
danleifeng 已提交
1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580
int32_t GraphTable::load_nodes(const std::string &path, std::string node_type) {
  auto paths = paddle::string::split_string<std::string>(path, ";");
  uint64_t count = 0;
  uint64_t valid_count = 0;
  int idx = 0;
  if (FLAGS_graph_load_in_parallel) {
    if (node_type == "") {
      VLOG(0) << "Begin GraphTable::load_nodes(), will load all node_type once";
    }
    std::vector<std::future<std::pair<uint64_t, uint64_t>>> tasks;
    for (size_t i = 0; i < paths.size(); i++) {
      tasks.push_back(load_node_edge_task_pool->enqueue(
          [&, i, this]() -> std::pair<uint64_t, uint64_t> {
            return parse_node_file(paths[i]);
          }));
    }
1581
    for (size_t i = 0; i < tasks.size(); i++) {
D
danleifeng 已提交
1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595
      auto res = tasks[i].get();
      count += res.first;
      valid_count += res.second;
    }
  } else {
    VLOG(0) << "Begin GraphTable::load_nodes() node_type[" << node_type << "]";
    if (node_type == "") {
      VLOG(0) << "node_type not specified, loading edges to "
              << id_to_feature[0] << " part";
    } else {
      if (feature_to_id.find(node_type) == feature_to_id.end()) {
        VLOG(0) << "node_type " << node_type
                << " is not defined, nothing will be loaded";
        return 0;
S
seemingwang 已提交
1596
      }
D
danleifeng 已提交
1597 1598 1599 1600 1601 1602 1603
      idx = feature_to_id[node_type];
    }
    for (auto path : paths) {
      VLOG(2) << "Begin GraphTable::load_nodes(), path[" << path << "]";
      auto res = parse_node_file(path, node_type, idx);
      count += res.first;
      valid_count += res.second;
S
seemingwang 已提交
1604 1605 1606
    }
  }

D
danleifeng 已提交
1607 1608
  VLOG(0) << valid_count << "/" << count << " nodes in node_type[ " << node_type
          << "] are loaded successfully!";
S
seemingwang 已提交
1609 1610 1611
  return 0;
}

1612 1613 1614 1615 1616 1617 1618 1619 1620
int32_t GraphTable::build_sampler(int idx, std::string sample_type) {
  for (auto &shard : edge_shards[idx]) {
    auto bucket = shard->get_bucket();
    for (size_t i = 0; i < bucket.size(); i++) {
      bucket[i]->build_sampler(sample_type);
    }
  }
  return 0;
}
D
danleifeng 已提交
1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679

std::pair<uint64_t, uint64_t> GraphTable::parse_edge_file(
    const std::string &path, int idx, bool reverse) {
  std::string sample_type = "random";
  bool is_weighted = false;
  std::ifstream file(path);
  std::string line;
  uint64_t local_count = 0;
  uint64_t local_valid_count = 0;
  uint64_t part_num = 0;
  if (FLAGS_graph_load_in_parallel) {
    auto path_split = paddle::string::split_string<std::string>(path, "/");
    auto part_name_split = paddle::string::split_string<std::string>(
        path_split[path_split.size() - 1], "-");
    part_num = std::stoull(part_name_split[part_name_split.size() - 1]);
  }

  while (std::getline(file, line)) {
    size_t start = line.find_first_of('\t');
    if (start == std::string::npos) continue;
    local_count++;
    uint64_t src_id = std::stoull(&line[0]);
    uint64_t dst_id = std::stoull(&line[start + 1]);
    if (reverse) {
      std::swap(src_id, dst_id);
    }
    size_t src_shard_id = src_id % shard_num;
    if (FLAGS_graph_load_in_parallel) {
      if (src_shard_id != (part_num % shard_num)) {
        continue;
      }
    }

    float weight = 1;
    size_t last = line.find_last_of('\t');
    if (start != last) {
      weight = std::stof(&line[last + 1]);
      sample_type = "weighted";
      is_weighted = true;
    }

    if (src_shard_id >= shard_end || src_shard_id < shard_start) {
      VLOG(4) << "will not load " << src_id << " from " << path
              << ", please check id distribution";
      continue;
    }
    size_t index = src_shard_id - shard_start;
    auto node = edge_shards[idx][index]->add_graph_node(src_id);
    if (node != NULL) {
      node->build_edges(is_weighted);
      node->add_edge(dst_id, weight);
    }

    local_valid_count++;
  }
  VLOG(2) << local_count << " edges are loaded from filepath->" << path;
  return {local_count, local_valid_count};
}

1680 1681
int32_t GraphTable::load_edges(const std::string &path,
                               bool reverse_edge,
1682
                               const std::string &edge_type) {
1683 1684 1685
#ifdef PADDLE_WITH_HETERPS
  if (search_level == 2) total_memory_cost = 0;
#endif
1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697
  int idx = 0;
  if (edge_type == "") {
    VLOG(0) << "edge_type not specified, loading edges to " << id_to_edge[0]
            << " part";
  } else {
    if (edge_to_id.find(edge_type) == edge_to_id.end()) {
      VLOG(0) << "edge_type " << edge_type
              << " is not defined, nothing will be loaded";
      return 0;
    }
    idx = edge_to_id[edge_type];
  }
1698

S
seemingwang 已提交
1699
  auto paths = paddle::string::split_string<std::string>(path, ";");
D
danleifeng 已提交
1700 1701 1702 1703 1704 1705
  uint64_t count = 0;
  uint64_t valid_count = 0;

  VLOG(0) << "Begin GraphTable::load_edges() edge_type[" << edge_type << "]";
  if (FLAGS_graph_load_in_parallel) {
    std::vector<std::future<std::pair<uint64_t, uint64_t>>> tasks;
1706
    for (size_t i = 0; i < paths.size(); i++) {
D
danleifeng 已提交
1707 1708 1709 1710 1711
      tasks.push_back(load_node_edge_task_pool->enqueue(
          [&, i, idx, this]() -> std::pair<uint64_t, uint64_t> {
            return parse_edge_file(paths[i], idx, reverse_edge);
          }));
    }
1712
    for (size_t j = 0; j < tasks.size(); j++) {
D
danleifeng 已提交
1713 1714 1715 1716 1717 1718 1719 1720 1721
      auto res = tasks[j].get();
      count += res.first;
      valid_count += res.second;
    }
  } else {
    for (auto path : paths) {
      auto res = parse_edge_file(path, idx, reverse_edge);
      count += res.first;
      valid_count += res.second;
S
seemingwang 已提交
1722 1723
    }
  }
D
danleifeng 已提交
1724 1725
  VLOG(0) << valid_count << "/" << count << " edge_type[" << edge_type
          << "] edges are loaded successfully";
L
lxsbupt 已提交
1726 1727
  std::string edge_size = edge_type + ":" + std::to_string(valid_count);
  edge_type_size.push_back(edge_size);
S
seemingwang 已提交
1728

1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739
#ifdef PADDLE_WITH_HETERPS
  if (search_level == 2) {
    if (count > 0) {
      dump_edges_to_ssd(idx);
      VLOG(0) << "dumping edges to ssd, edge count is reset to 0";
      clear_graph(idx);
      count = 0;
    }
    return 0;
  }
#endif
D
danleifeng 已提交
1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753

  if (!build_sampler_on_cpu) {
    // To reduce memory overhead, CPU samplers won't be created in gpugraph.
    // In order not to affect the sampler function of other scenario,
    // this optimization is only performed in load_edges function.
    VLOG(0) << "run in gpugraph mode!";
  } else {
    std::string sample_type = "random";
    VLOG(0) << "build sampler ... ";
    for (auto &shard : edge_shards[idx]) {
      auto bucket = shard->get_bucket();
      for (size_t i = 0; i < bucket.size(); i++) {
        bucket[i]->build_sampler(sample_type);
      }
1754 1755 1756
    }
  }

S
seemingwang 已提交
1757 1758 1759
  return 0;
}

D
danleifeng 已提交
1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780
Node *GraphTable::find_node(int type_id, uint64_t id) {
  size_t shard_id = id % shard_num;
  if (shard_id >= shard_end || shard_id < shard_start) {
    return nullptr;
  }
  Node *node = nullptr;
  size_t index = shard_id - shard_start;
  auto &search_shards = type_id == 0 ? edge_shards : feature_shards;
  for (auto &search_shard : search_shards) {
    PADDLE_ENFORCE_NOT_NULL(search_shard[index],
                            paddle::platform::errors::InvalidArgument(
                                "search_shard[%d] should not be null.", index));
    node = search_shard[index]->find_node(id);
    if (node != nullptr) {
      break;
    }
  }
  return node;
}

Node *GraphTable::find_node(int type_id, int idx, uint64_t id) {
S
seemingwang 已提交
1781 1782
  size_t shard_id = id % shard_num;
  if (shard_id >= shard_end || shard_id < shard_start) {
1783
    return nullptr;
S
seemingwang 已提交
1784 1785
  }
  size_t index = shard_id - shard_start;
1786
  auto &search_shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
D
danleifeng 已提交
1787 1788 1789
  PADDLE_ENFORCE_NOT_NULL(search_shards[index],
                          paddle::platform::errors::InvalidArgument(
                              "search_shard[%d] should not be null.", index));
1790
  Node *node = search_shards[index]->find_node(id);
S
seemingwang 已提交
1791 1792
  return node;
}
D
danleifeng 已提交
1793
uint32_t GraphTable::get_thread_pool_index(uint64_t node_id) {
1794
  return node_id % shard_num % shard_num_per_server % task_pool_size_;
S
seemingwang 已提交
1795
}
1796

D
danleifeng 已提交
1797 1798
uint32_t GraphTable::get_thread_pool_index_by_shard_index(
    uint64_t shard_index) {
S
seemingwang 已提交
1799
  return shard_index % shard_num_per_server % task_pool_size_;
1800 1801
}

1802 1803
int32_t GraphTable::clear_nodes(int type_id, int idx) {
  auto &search_shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
Z
zhangchunle 已提交
1804
  for (size_t i = 0; i < search_shards.size(); i++) {
1805
    search_shards[i]->clear();
1806 1807 1808 1809
  }
  return 0;
}

1810 1811 1812
int32_t GraphTable::random_sample_nodes(int type_id,
                                        int idx,
                                        int sample_size,
S
seemingwang 已提交
1813 1814 1815
                                        std::unique_ptr<char[]> &buffer,
                                        int &actual_size) {
  int total_size = 0;
1816
  auto &shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
1817
  for (size_t i = 0; i < shards.size(); i++) {
1818
    total_size += shards[i]->get_size();
S
seemingwang 已提交
1819 1820 1821 1822 1823 1824 1825 1826 1827
  }
  if (sample_size > total_size) sample_size = total_size;
  int range_num = random_sample_nodes_ranges;
  if (range_num > sample_size) range_num = sample_size;
  if (sample_size == 0 || range_num == 0) return 0;
  std::vector<int> ranges_len, ranges_pos;
  int remain = sample_size, last_pos = -1, num;
  std::set<int> separator_set;
  for (int i = 0; i < range_num - 1; i++) {
L
lxsbupt 已提交
1828
    while (separator_set.find(num = rand() % (sample_size - 1)) !=  // NOLINT
1829 1830
           separator_set.end()) {
    }
S
seemingwang 已提交
1831 1832 1833 1834 1835 1836 1837 1838 1839 1840
    separator_set.insert(num);
  }
  for (auto p : separator_set) {
    ranges_len.push_back(p - last_pos);
    last_pos = p;
  }
  ranges_len.push_back(sample_size - 1 - last_pos);
  remain = total_size - sample_size + range_num;
  separator_set.clear();
  for (int i = 0; i < range_num; i++) {
L
lxsbupt 已提交
1841
    while (separator_set.find(num = rand() % remain) !=  // NOLINT
1842 1843
           separator_set.end()) {
    }
S
seemingwang 已提交
1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854
    separator_set.insert(num);
  }
  int used = 0, index = 0;
  last_pos = -1;
  for (auto p : separator_set) {
    used += p - last_pos - 1;
    last_pos = p;
    ranges_pos.push_back(used);
    used += ranges_len[index++];
  }
  std::vector<std::pair<int, int>> first_half, second_half;
L
lxsbupt 已提交
1855
  int start_index = rand() % total_size;  // NOLINT
1856
  for (size_t i = 0; i < ranges_len.size() && i < ranges_pos.size(); i++) {
1857
    if (ranges_pos[i] + ranges_len[i] - 1 + start_index < total_size) {
S
seemingwang 已提交
1858 1859
      first_half.push_back({ranges_pos[i] + start_index,
                            ranges_pos[i] + ranges_len[i] + start_index});
L
lxsbupt 已提交
1860
    } else if ((ranges_pos[i] + start_index) >= total_size) {
S
seemingwang 已提交
1861 1862 1863 1864 1865 1866 1867 1868 1869 1870
      second_half.push_back(
          {ranges_pos[i] + start_index - total_size,
           ranges_pos[i] + ranges_len[i] + start_index - total_size});
    } else {
      first_half.push_back({ranges_pos[i] + start_index, total_size});
      second_half.push_back(
          {0, ranges_pos[i] + ranges_len[i] + start_index - total_size});
    }
  }
  for (auto &pair : first_half) second_half.push_back(pair);
D
danleifeng 已提交
1871
  std::vector<uint64_t> res;
1872
  get_nodes_ids_by_ranges(type_id, idx, second_half, res);
D
danleifeng 已提交
1873
  actual_size = res.size() * sizeof(uint64_t);
S
seemingwang 已提交
1874 1875 1876 1877 1878
  buffer.reset(new char[actual_size]);
  char *pointer = buffer.get();
  memcpy(pointer, res.data(), actual_size);
  return 0;
}
1879
int32_t GraphTable::random_sample_neighbors(
1880
    int idx,
D
danleifeng 已提交
1881
    uint64_t *node_ids,
1882 1883 1884
    int sample_size,
    std::vector<std::shared_ptr<char>> &buffers,
    std::vector<int> &actual_sizes,
1885
    bool need_weight) {
S
seemingwang 已提交
1886
  size_t node_num = buffers.size();
1887
  std::function<void(char *)> char_del = [](char *c) { delete[] c; };
S
seemingwang 已提交
1888
  std::vector<std::future<int>> tasks;
1889 1890
  std::vector<std::vector<uint32_t>> seq_id(task_pool_size_);
  std::vector<std::vector<SampleKey>> id_list(task_pool_size_);
S
seemingwang 已提交
1891
  size_t index;
1892 1893 1894 1895
  for (size_t idy = 0; idy < node_num; ++idy) {
    index = get_thread_pool_index(node_ids[idy]);
    seq_id[index].emplace_back(idy);
    id_list[index].emplace_back(idx, node_ids[idy], sample_size, need_weight);
S
seemingwang 已提交
1896
  }
1897

1898
  for (size_t i = 0; i < seq_id.size(); i++) {
S
seemingwang 已提交
1899 1900
    if (seq_id[i].size() == 0) continue;
    tasks.push_back(_shards_task_pool[i]->enqueue([&, i, this]() -> int {
D
danleifeng 已提交
1901
      uint64_t node_id;
S
seemingwang 已提交
1902 1903 1904 1905
      std::vector<std::pair<SampleKey, SampleResult>> r;
      LRUResponse response = LRUResponse::blocked;
      if (use_cache) {
        response =
1906
            scaled_lru->query(i, id_list[i].data(), id_list[i].size(), r);
S
seemingwang 已提交
1907
      }
1908
      size_t index = 0;
S
seemingwang 已提交
1909 1910 1911 1912
      std::vector<SampleResult> sample_res;
      std::vector<SampleKey> sample_keys;
      auto &rng = _shards_task_rng_pool[i];
      for (size_t k = 0; k < id_list[i].size(); k++) {
1913
        if (index < r.size() &&
S
seemingwang 已提交
1914
            r[index].first.node_key == id_list[i][k].node_key) {
1915 1916 1917
          int idy = seq_id[i][k];
          actual_sizes[idy] = r[index].second.actual_size;
          buffers[idy] = r[index].second.buffer;
S
seemingwang 已提交
1918 1919 1920
          index++;
        } else {
          node_id = id_list[i][k].node_key;
1921 1922 1923
          Node *node = find_node(0, idx, node_id);
          int idy = seq_id[i][k];
          int &actual_size = actual_sizes[idy];
S
seemingwang 已提交
1924
          if (node == nullptr) {
1925 1926
#ifdef PADDLE_WITH_HETERPS
            if (search_level == 2) {
S
seemingwang 已提交
1927
              VLOG(2) << "enter sample from ssd for node_id " << node_id;
1928
              char *buffer_addr = random_sample_neighbor_from_ssd(
1929
                  idx, node_id, sample_size, rng, actual_size);
1930
              if (actual_size != 0) {
S
seemingwang 已提交
1931
                std::shared_ptr<char> &buffer = buffers[idy];
1932 1933
                buffer.reset(buffer_addr, char_del);
              }
S
seemingwang 已提交
1934
              VLOG(2) << "actual sampled size from ssd = " << actual_sizes[idy];
1935 1936 1937
              continue;
            }
#endif
S
seemingwang 已提交
1938 1939 1940
            actual_size = 0;
            continue;
          }
1941
          std::shared_ptr<char> &buffer = buffers[idy];
S
seemingwang 已提交
1942
          std::vector<int> res = node->sample_k(sample_size, rng);
1943 1944 1945
          actual_size =
              res.size() * (need_weight ? (Node::id_size + Node::weight_size)
                                        : Node::id_size);
S
seemingwang 已提交
1946
          int offset = 0;
D
danleifeng 已提交
1947
          uint64_t id;
S
seemingwang 已提交
1948 1949 1950
          float weight;
          char *buffer_addr = new char[actual_size];
          if (response == LRUResponse::ok) {
1951
            sample_keys.emplace_back(idx, node_id, sample_size, need_weight);
S
seemingwang 已提交
1952 1953
            sample_res.emplace_back(actual_size, buffer_addr);
            buffer = sample_res.back().buffer;
1954
          } else {
S
seemingwang 已提交
1955 1956 1957 1958 1959 1960
            buffer.reset(buffer_addr, char_del);
          }
          for (int &x : res) {
            id = node->get_neighbor_id(x);
            memcpy(buffer_addr + offset, &id, Node::id_size);
            offset += Node::id_size;
1961 1962 1963 1964 1965
            if (need_weight) {
              weight = node->get_neighbor_weight(x);
              memcpy(buffer_addr + offset, &weight, Node::weight_size);
              offset += Node::weight_size;
            }
1966 1967
          }
        }
1968
      }
S
seemingwang 已提交
1969
      if (sample_res.size()) {
1970 1971
        scaled_lru->insert(
            i, sample_keys.data(), sample_res.data(), sample_keys.size());
1972 1973 1974
      }
      return 0;
    }));
S
seemingwang 已提交
1975
  }
S
seemingwang 已提交
1976 1977
  for (auto &t : tasks) {
    t.get();
S
seemingwang 已提交
1978 1979 1980 1981
  }
  return 0;
}

1982
int32_t GraphTable::get_node_feat(int idx,
D
danleifeng 已提交
1983
                                  const std::vector<uint64_t> &node_ids,
S
seemingwang 已提交
1984 1985 1986 1987
                                  const std::vector<std::string> &feature_names,
                                  std::vector<std::vector<std::string>> &res) {
  size_t node_num = node_ids.size();
  std::vector<std::future<int>> tasks;
1988
  for (size_t idy = 0; idy < node_num; ++idy) {
D
danleifeng 已提交
1989
    uint64_t node_id = node_ids[idy];
S
seemingwang 已提交
1990
    tasks.push_back(_shards_task_pool[get_thread_pool_index(node_id)]->enqueue(
1991 1992
        [&, idx, idy, node_id]() -> int {
          Node *node = find_node(1, idx, node_id);
S
seemingwang 已提交
1993 1994 1995 1996

          if (node == nullptr) {
            return 0;
          }
1997
          for (size_t feat_idx = 0; feat_idx < feature_names.size();
1998
               ++feat_idx) {
S
seemingwang 已提交
1999
            const std::string &feature_name = feature_names[feat_idx];
2000
            if (feat_id_map[idx].find(feature_name) != feat_id_map[idx].end()) {
S
seemingwang 已提交
2001 2002
              // res[feat_idx][idx] =
              // node->get_feature(feat_id_map[feature_name]);
2003 2004
              auto feat = node->get_feature(feat_id_map[idx][feature_name]);
              res[feat_idx][idy] = feat;
S
seemingwang 已提交
2005 2006 2007 2008
            }
          }
          return 0;
        }));
S
seemingwang 已提交
2009
  }
2010 2011
  for (size_t idy = 0; idy < node_num; ++idy) {
    tasks[idy].get();
S
seemingwang 已提交
2012 2013 2014 2015 2016
  }
  return 0;
}

int32_t GraphTable::set_node_feat(
2017
    int idx,
D
danleifeng 已提交
2018
    const std::vector<uint64_t> &node_ids,
S
seemingwang 已提交
2019 2020 2021 2022
    const std::vector<std::string> &feature_names,
    const std::vector<std::vector<std::string>> &res) {
  size_t node_num = node_ids.size();
  std::vector<std::future<int>> tasks;
2023
  for (size_t idy = 0; idy < node_num; ++idy) {
D
danleifeng 已提交
2024
    uint64_t node_id = node_ids[idy];
S
seemingwang 已提交
2025
    tasks.push_back(_shards_task_pool[get_thread_pool_index(node_id)]->enqueue(
2026
        [&, idx, idy, node_id]() -> int {
S
seemingwang 已提交
2027
          size_t index = node_id % this->shard_num - this->shard_start;
2028 2029
          auto node = feature_shards[idx][index]->add_feature_node(node_id);
          node->set_feature_size(this->feat_name[idx].size());
2030
          for (size_t feat_idx = 0; feat_idx < feature_names.size();
2031
               ++feat_idx) {
S
seemingwang 已提交
2032
            const std::string &feature_name = feature_names[feat_idx];
2033 2034 2035
            if (feat_id_map[idx].find(feature_name) != feat_id_map[idx].end()) {
              node->set_feature(feat_id_map[idx][feature_name],
                                res[feat_idx][idy]);
S
seemingwang 已提交
2036 2037 2038 2039
            }
          }
          return 0;
        }));
S
seemingwang 已提交
2040
  }
2041 2042
  for (size_t idy = 0; idy < node_num; ++idy) {
    tasks[idy].get();
S
seemingwang 已提交
2043 2044 2045 2046
  }
  return 0;
}

D
danleifeng 已提交
2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081
void string_vector_2_string(std::vector<std::string>::iterator strs_begin,
                            std::vector<std::string>::iterator strs_end,
                            char delim,
                            std::string *output) {
  size_t i = 0;
  for (std::vector<std::string>::iterator iter = strs_begin; iter != strs_end;
       ++iter) {
    if (i > 0) {
      *output += delim;
    }

    *output += *iter;
    ++i;
  }
}

void string_vector_2_string(
    std::vector<paddle::string::str_ptr>::iterator strs_begin,
    std::vector<paddle::string::str_ptr>::iterator strs_end,
    char delim,
    std::string *output) {
  size_t i = 0;
  for (auto iter = strs_begin; iter != strs_end; ++iter) {
    if (i > 0) {
      output->append(&delim, 1);
    }
    output->append((*iter).ptr, (*iter).len);
    ++i;
  }
}

int GraphTable::parse_feature(int idx,
                              const char *feat_str,
                              size_t len,
                              FeatureNode *node) {
S
seemingwang 已提交
2082 2083
  // Return (feat_id, btyes) if name are in this->feat_name, else return (-1,
  // "")
D
danleifeng 已提交
2084 2085
  thread_local std::vector<paddle::string::str_ptr> fields;
  fields.clear();
L
lxsbupt 已提交
2086
  char c = slot_feature_separator_.at(0);
D
danleifeng 已提交
2087 2088
  paddle::string::split_string_ptr(feat_str, len, c, &fields);

L
lxsbupt 已提交
2089 2090 2091 2092 2093 2094 2095 2096
  thread_local std::vector<paddle::string::str_ptr> fea_fields;
  fea_fields.clear();
  c = feature_separator_.at(0);
  paddle::string::split_string_ptr(fields[1].ptr,
                                   fields[1].len,
                                   c,
                                   &fea_fields,
                                   FLAGS_gpugraph_slot_feasign_max_num);
D
danleifeng 已提交
2097 2098 2099 2100 2101
  std::string name = fields[0].to_string();
  auto it = feat_id_map[idx].find(name);
  if (it != feat_id_map[idx].end()) {
    int32_t id = it->second;
    std::string *fea_ptr = node->mutable_feature(id);
2102
    std::string dtype = this->feat_dtype[idx][id];
S
seemingwang 已提交
2103
    if (dtype == "feasign") {
D
danleifeng 已提交
2104 2105 2106
      //      string_vector_2_string(fields.begin() + 1, fields.end(), ' ',
      //      fea_ptr);
      FeatureNode::parse_value_to_bytes<uint64_t>(
L
lxsbupt 已提交
2107
          fea_fields.begin(), fea_fields.end(), fea_ptr);
D
danleifeng 已提交
2108
      return 0;
S
seemingwang 已提交
2109
    } else if (dtype == "string") {
L
lxsbupt 已提交
2110 2111
      string_vector_2_string(
          fea_fields.begin(), fea_fields.end(), ' ', fea_ptr);
D
danleifeng 已提交
2112
      return 0;
S
seemingwang 已提交
2113
    } else if (dtype == "float32") {
D
danleifeng 已提交
2114
      FeatureNode::parse_value_to_bytes<float>(
L
lxsbupt 已提交
2115
          fea_fields.begin(), fea_fields.end(), fea_ptr);
D
danleifeng 已提交
2116
      return 0;
S
seemingwang 已提交
2117
    } else if (dtype == "float64") {
D
danleifeng 已提交
2118
      FeatureNode::parse_value_to_bytes<double>(
L
lxsbupt 已提交
2119
          fea_fields.begin(), fea_fields.end(), fea_ptr);
D
danleifeng 已提交
2120
      return 0;
S
seemingwang 已提交
2121
    } else if (dtype == "int32") {
D
danleifeng 已提交
2122
      FeatureNode::parse_value_to_bytes<int32_t>(
L
lxsbupt 已提交
2123
          fea_fields.begin(), fea_fields.end(), fea_ptr);
D
danleifeng 已提交
2124
      return 0;
S
seemingwang 已提交
2125
    } else if (dtype == "int64") {
D
danleifeng 已提交
2126
      FeatureNode::parse_value_to_bytes<uint64_t>(
L
lxsbupt 已提交
2127
          fea_fields.begin(), fea_fields.end(), fea_ptr);
D
danleifeng 已提交
2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176
      return 0;
    }
  } else {
    VLOG(2) << "feature_name[" << name << "] is not in feat_id_map, ntype_id["
            << idx << "] feat_id_map_size[" << feat_id_map.size() << "]";
  }

  return -1;
}
// thread safe shard vector merge
class MergeShardVector {
 public:
  MergeShardVector(std::vector<std::vector<uint64_t>> *output, int slice_num) {
    _slice_num = slice_num;
    _shard_keys = output;
    _shard_keys->resize(slice_num);
    _mutexs = new std::mutex[slice_num];
  }
  ~MergeShardVector() {
    if (_mutexs != nullptr) {
      delete[] _mutexs;
      _mutexs = nullptr;
    }
  }
  // merge shard keys
  void merge(const std::vector<std::vector<uint64_t>> &shard_keys) {
    // add to shard
    for (int shard_id = 0; shard_id < _slice_num; ++shard_id) {
      auto &dest = (*_shard_keys)[shard_id];
      auto &src = shard_keys[shard_id];

      _mutexs[shard_id].lock();
      dest.insert(dest.end(), src.begin(), src.end());
      _mutexs[shard_id].unlock();
    }
  }

 private:
  int _slice_num = 0;
  std::mutex *_mutexs = nullptr;
  std::vector<std::vector<uint64_t>> *_shard_keys;
};

int GraphTable::get_all_id(int type_id,
                           int slice_num,
                           std::vector<std::vector<uint64_t>> *output) {
  MergeShardVector shard_merge(output, slice_num);
  auto &search_shards = type_id == 0 ? edge_shards : feature_shards;
  std::vector<std::future<size_t>> tasks;
2177 2178
  for (size_t idx = 0; idx < search_shards.size(); idx++) {
    for (size_t j = 0; j < search_shards[idx].size(); j++) {
D
danleifeng 已提交
2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200
      tasks.push_back(_shards_task_pool[j % task_pool_size_]->enqueue(
          [&search_shards, idx, j, slice_num, &shard_merge]() -> size_t {
            std::vector<std::vector<uint64_t>> shard_keys;
            size_t num =
                search_shards[idx][j]->get_all_id(&shard_keys, slice_num);
            // add to shard
            shard_merge.merge(shard_keys);
            return num;
          }));
    }
  }
  for (size_t i = 0; i < tasks.size(); ++i) {
    tasks[i].wait();
  }
  return 0;
}

int GraphTable::get_all_neighbor_id(
    int type_id, int slice_num, std::vector<std::vector<uint64_t>> *output) {
  MergeShardVector shard_merge(output, slice_num);
  auto &search_shards = type_id == 0 ? edge_shards : feature_shards;
  std::vector<std::future<size_t>> tasks;
2201 2202
  for (size_t idx = 0; idx < search_shards.size(); idx++) {
    for (size_t j = 0; j < search_shards[idx].size(); j++) {
D
danleifeng 已提交
2203 2204 2205 2206 2207 2208 2209 2210 2211
      tasks.push_back(_shards_task_pool[j % task_pool_size_]->enqueue(
          [&search_shards, idx, j, slice_num, &shard_merge]() -> size_t {
            std::vector<std::vector<uint64_t>> shard_keys;
            size_t num = search_shards[idx][j]->get_all_neighbor_id(&shard_keys,
                                                                    slice_num);
            // add to shard
            shard_merge.merge(shard_keys);
            return num;
          }));
S
seemingwang 已提交
2212 2213
    }
  }
D
danleifeng 已提交
2214 2215 2216 2217
  for (size_t i = 0; i < tasks.size(); ++i) {
    tasks[i].wait();
  }
  return 0;
S
seemingwang 已提交
2218 2219
}

D
danleifeng 已提交
2220 2221 2222 2223 2224
int GraphTable::get_all_id(int type_id,
                           int idx,
                           int slice_num,
                           std::vector<std::vector<uint64_t>> *output) {
  MergeShardVector shard_merge(output, slice_num);
2225
  auto &search_shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
D
danleifeng 已提交
2226 2227
  std::vector<std::future<size_t>> tasks;
  VLOG(3) << "begin task, task_pool_size_[" << task_pool_size_ << "]";
Z
zhangchunle 已提交
2228
  for (size_t i = 0; i < search_shards.size(); i++) {
2229
    tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
D
danleifeng 已提交
2230 2231 2232 2233 2234 2235
        [&search_shards, i, slice_num, &shard_merge]() -> size_t {
          std::vector<std::vector<uint64_t>> shard_keys;
          size_t num = search_shards[i]->get_all_id(&shard_keys, slice_num);
          // add to shard
          shard_merge.merge(shard_keys);
          return num;
2236 2237 2238 2239 2240
        }));
  }
  for (size_t i = 0; i < tasks.size(); ++i) {
    tasks[i].wait();
  }
D
danleifeng 已提交
2241 2242 2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253
  VLOG(3) << "end task, task_pool_size_[" << task_pool_size_ << "]";
  return 0;
}

int GraphTable::get_all_neighbor_id(
    int type_id,
    int idx,
    int slice_num,
    std::vector<std::vector<uint64_t>> *output) {
  MergeShardVector shard_merge(output, slice_num);
  auto &search_shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
  std::vector<std::future<size_t>> tasks;
  VLOG(3) << "begin task, task_pool_size_[" << task_pool_size_ << "]";
2254
  for (size_t i = 0; i < search_shards.size(); i++) {
D
danleifeng 已提交
2255 2256 2257 2258 2259 2260 2261 2262 2263
    tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
        [&search_shards, i, slice_num, &shard_merge]() -> size_t {
          std::vector<std::vector<uint64_t>> shard_keys;
          size_t num =
              search_shards[i]->get_all_neighbor_id(&shard_keys, slice_num);
          // add to shard
          shard_merge.merge(shard_keys);
          return num;
        }));
2264
  }
D
danleifeng 已提交
2265 2266 2267 2268 2269
  for (size_t i = 0; i < tasks.size(); ++i) {
    tasks[i].wait();
  }
  VLOG(3) << "end task, task_pool_size_[" << task_pool_size_ << "]";
  return 0;
2270
}
D
danleifeng 已提交
2271 2272 2273 2274 2275 2276 2277 2278 2279

int GraphTable::get_all_feature_ids(
    int type_id,
    int idx,
    int slice_num,
    std::vector<std::vector<uint64_t>> *output) {
  MergeShardVector shard_merge(output, slice_num);
  auto &search_shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
  std::vector<std::future<size_t>> tasks;
2280
  for (size_t i = 0; i < search_shards.size(); i++) {
D
danleifeng 已提交
2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296
    tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
        [&search_shards, i, slice_num, &shard_merge]() -> size_t {
          std::vector<std::vector<uint64_t>> shard_keys;
          size_t num =
              search_shards[i]->get_all_feature_ids(&shard_keys, slice_num);
          // add to shard
          shard_merge.merge(shard_keys);
          return num;
        }));
  }
  for (size_t i = 0; i < tasks.size(); ++i) {
    tasks[i].wait();
  }
  return 0;
}

L
lxsbupt 已提交
2297 2298 2299 2300 2301 2302 2303 2304 2305 2306
int GraphTable::get_node_embedding_ids(
    int slice_num, std::vector<std::vector<uint64_t>> *output) {
  if (is_load_reverse_edge && !FLAGS_graph_get_neighbor_id) {
    return get_all_id(0, slice_num, output);
  } else {
    get_all_id(0, slice_num, output);
    return get_all_neighbor_id(0, slice_num, output);
  }
}

2307 2308 2309
int32_t GraphTable::pull_graph_list(int type_id,
                                    int idx,
                                    int start,
2310
                                    int total_size,
S
seemingwang 已提交
2311
                                    std::unique_ptr<char[]> &buffer,
2312 2313
                                    int &actual_size,
                                    bool need_feature,
S
seemingwang 已提交
2314 2315 2316
                                    int step) {
  if (start < 0) start = 0;
  int size = 0, cur_size;
2317
  auto &search_shards = type_id == 0 ? edge_shards[idx] : feature_shards[idx];
S
seemingwang 已提交
2318
  std::vector<std::future<std::vector<Node *>>> tasks;
2319 2320
  for (size_t i = 0; i < search_shards.size() && total_size > 0; i++) {
    cur_size = search_shards[i]->get_size();
S
seemingwang 已提交
2321 2322 2323 2324 2325 2326 2327
    if (size + cur_size <= start) {
      size += cur_size;
      continue;
    }
    int count = std::min(1 + (size + cur_size - start - 1) / step, total_size);
    int end = start + (count - 1) * step + 1;
    tasks.push_back(_shards_task_pool[i % task_pool_size_]->enqueue(
2328 2329
        [&search_shards, this, i, start, end, step, size]()
            -> std::vector<Node *> {
2330
          return search_shards[i]->get_batch(start - size, end - size, step);
S
seemingwang 已提交
2331 2332 2333 2334 2335 2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358
        }));
    start += count * step;
    total_size -= count;
    size += cur_size;
  }
  for (size_t i = 0; i < tasks.size(); ++i) {
    tasks[i].wait();
  }
  size = 0;
  std::vector<std::vector<Node *>> res;
  for (size_t i = 0; i < tasks.size(); i++) {
    res.push_back(tasks[i].get());
    for (size_t j = 0; j < res.back().size(); j++) {
      size += res.back()[j]->get_size(need_feature);
    }
  }
  char *buffer_addr = new char[size];
  buffer.reset(buffer_addr);
  int index = 0;
  for (size_t i = 0; i < res.size(); i++) {
    for (size_t j = 0; j < res[i].size(); j++) {
      res[i][j]->to_buffer(buffer_addr + index, need_feature);
      index += res[i][j]->get_size(need_feature);
    }
  }
  actual_size = size;
  return 0;
}
S
seemingwang 已提交
2359

D
danleifeng 已提交
2360 2361 2362 2363
void GraphTable::set_feature_separator(const std::string &ch) {
  feature_separator_ = ch;
}

L
lxsbupt 已提交
2364 2365 2366 2367
void GraphTable::set_slot_feature_separator(const std::string &ch) {
  slot_feature_separator_ = ch;
}

D
danleifeng 已提交
2368
int32_t GraphTable::get_server_index_by_id(uint64_t id) {
S
seemingwang 已提交
2369 2370
  return id % shard_num / shard_num_per_server;
}
Z
zhaocaibei123 已提交
2371
int32_t GraphTable::Initialize(const TableParameter &config,
2372 2373 2374
                               const FsClientParameter &fs_config) {
  LOG(INFO) << "in graphTable initialize";
  _config = config;
Z
zhaocaibei123 已提交
2375
  if (InitializeAccessor() != 0) {
2376 2377 2378
    LOG(WARNING) << "Table accessor initialize failed";
    return -1;
  }
S
seemingwang 已提交
2379

2380 2381 2382 2383 2384 2385 2386
  if (_afs_client.initialize(fs_config) != 0) {
    LOG(WARNING) << "Table fs_client initialize failed";
    // return -1;
  }
  auto graph = config.graph_parameter();
  shard_num = _config.shard_num();
  LOG(INFO) << "in graphTable initialize over";
Z
zhaocaibei123 已提交
2387
  return Initialize(graph);
2388
}
2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399 2400 2401 2402 2403 2404 2405 2406

void GraphTable::load_node_weight(int type_id, int idx, std::string path) {
  auto paths = paddle::string::split_string<std::string>(path, ";");
  int64_t count = 0;
  auto &weight_map = node_weight[type_id][idx];
  for (auto path : paths) {
    std::ifstream file(path);
    std::string line;
    while (std::getline(file, line)) {
      auto values = paddle::string::split_string<std::string>(line, "\t");
      count++;
      if (values.size() < 2) continue;
      auto src_id = std::stoull(values[0]);
      double weight = std::stod(values[1]);
      weight_map[src_id] = weight;
    }
  }
}
Z
zhaocaibei123 已提交
2407
int32_t GraphTable::Initialize(const GraphParameter &graph) {
2408
  task_pool_size_ = graph.task_pool_size();
D
danleifeng 已提交
2409
  build_sampler_on_cpu = graph.build_sampler_on_cpu();
2410

2411
#ifdef PADDLE_WITH_HETERPS
2412 2413 2414 2415 2416
  _db = NULL;
  search_level = graph.search_level();
  if (search_level >= 2) {
    _db = paddle::distributed::RocksDBHandler::GetInstance();
    _db->initialize("./temp_gpups_db", task_pool_size_);
2417
  }
2418 2419 2420 2421 2422 2423 2424 2425 2426
// gpups_mode = true;
// auto *sampler =
//     CREATE_PSCORE_CLASS(GraphSampler, graph.gpups_graph_sample_class());
// auto slices =
//     string::split_string<std::string>(graph.gpups_graph_sample_args(), ",");
// std::cout << "slices" << std::endl;
// for (auto x : slices) std::cout << x << std::endl;
// sampler->init(graph.gpu_num(), this, slices);
// graph_sampler.reset(sampler);
2427
#endif
2428 2429 2430 2431 2432 2433 2434 2435 2436
  if (shard_num == 0) {
    server_num = 1;
    _shard_idx = 0;
    shard_num = graph.shard_num();
  }
  use_cache = graph.use_cache();
  if (use_cache) {
    cache_size_limit = graph.cache_size_limit();
    cache_ttl = graph.cache_ttl();
L
lxsbupt 已提交
2437
    make_neighbor_sample_cache(cache_size_limit, cache_ttl);
2438
  }
S
seemingwang 已提交
2439 2440 2441
  _shards_task_pool.resize(task_pool_size_);
  for (size_t i = 0; i < _shards_task_pool.size(); ++i) {
    _shards_task_pool[i].reset(new ::ThreadPool(1));
2442
    _shards_task_rng_pool.push_back(paddle::framework::GetCPURandomEngine(0));
S
seemingwang 已提交
2443
  }
D
danleifeng 已提交
2444 2445
  load_node_edge_task_pool.reset(new ::ThreadPool(load_thread_num));

2446
  auto graph_feature = graph.graph_feature();
2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458
  auto node_types = graph.node_types();
  auto edge_types = graph.edge_types();
  VLOG(0) << "got " << edge_types.size() << "edge types in total";
  feat_id_map.resize(node_types.size());
  for (int k = 0; k < edge_types.size(); k++) {
    VLOG(0) << "in initialize: get a edge_type " << edge_types[k];
    edge_to_id[edge_types[k]] = k;
    id_to_edge.push_back(edge_types[k]);
  }
  feat_name.resize(node_types.size());
  feat_shape.resize(node_types.size());
  feat_dtype.resize(node_types.size());
L
lxsbupt 已提交
2459
  VLOG(0) << "got " << node_types.size() << " node types in total";
2460 2461 2462 2463 2464 2465 2466 2467 2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481
  for (int k = 0; k < node_types.size(); k++) {
    feature_to_id[node_types[k]] = k;
    auto node_type = node_types[k];
    auto feature = graph_feature[k];
    id_to_feature.push_back(node_type);
    int feat_conf_size = static_cast<int>(feature.name().size());

    for (int i = 0; i < feat_conf_size; i++) {
      // auto &f_name = common.attributes()[i];
      // auto &f_shape = common.dims()[i];
      // auto &f_dtype = common.params()[i];
      auto &f_name = feature.name()[i];
      auto &f_shape = feature.shape()[i];
      auto &f_dtype = feature.dtype()[i];
      feat_name[k].push_back(f_name);
      feat_shape[k].push_back(f_shape);
      feat_dtype[k].push_back(f_dtype);
      feat_id_map[k][f_name] = i;
      VLOG(0) << "init graph table feat conf name:" << f_name
              << " shape:" << f_shape << " dtype:" << f_dtype;
    }
  }
2482 2483 2484 2485
  // this->table_name = common.table_name();
  // this->table_type = common.name();
  this->table_name = graph.table_name();
  this->table_type = graph.table_type();
S
seemingwang 已提交
2486 2487
  VLOG(0) << " init graph table type " << this->table_type << " table name "
          << this->table_name;
2488
  // int feat_conf_size = static_cast<int>(common.attributes().size());
2489
  // int feat_conf_size = static_cast<int>(graph_feature.name().size());
S
seemingwang 已提交
2490 2491
  VLOG(0) << "in init graph table shard num = " << shard_num << " shard_idx"
          << _shard_idx;
S
seemingwang 已提交
2492 2493 2494
  shard_num_per_server = sparse_local_shard_num(shard_num, server_num);
  shard_start = _shard_idx * shard_num_per_server;
  shard_end = shard_start + shard_num_per_server;
S
seemingwang 已提交
2495 2496
  VLOG(0) << "in init graph table shard idx = " << _shard_idx << " shard_start "
          << shard_start << " shard_end " << shard_end;
2497
  edge_shards.resize(id_to_edge.size());
2498 2499
  node_weight.resize(2);
  node_weight[0].resize(id_to_edge.size());
2500 2501 2502
#ifdef PADDLE_WITH_HETERPS
  partitions.resize(id_to_edge.size());
#endif
2503
  for (size_t k = 0; k < edge_shards.size(); k++) {
2504 2505 2506
    for (size_t i = 0; i < shard_num_per_server; i++) {
      edge_shards[k].push_back(new GraphShard());
    }
2507
  }
2508
  node_weight[1].resize(id_to_feature.size());
2509
  feature_shards.resize(id_to_feature.size());
2510
  for (size_t k = 0; k < feature_shards.size(); k++) {
2511 2512 2513
    for (size_t i = 0; i < shard_num_per_server; i++) {
      feature_shards[k].push_back(new GraphShard());
    }
2514 2515
  }

S
seemingwang 已提交
2516 2517
  return 0;
}
2518

L
lxsbupt 已提交
2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537 2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555 2556 2557 2558 2559 2560 2561 2562 2563
void GraphTable::init_worker_poll(int gpu_num) {
  _cpu_worker_pool.resize(gpu_num);
  for (int i = 0; i < gpu_num; i++) {
    _cpu_worker_pool[i].reset(new ::ThreadPool(16));
  }
}

void GraphTable::build_graph_total_keys() {
  VLOG(0) << "begin insert edge to graph_total_keys";
  // build node embedding id
  std::vector<std::vector<uint64_t>> keys;
  this->get_node_embedding_ids(1, &keys);
  graph_total_keys_.insert(
      graph_total_keys_.end(), keys[0].begin(), keys[0].end());

  VLOG(0) << "finish insert edge to graph_total_keys";
}

void GraphTable::build_graph_type_keys() {
  VLOG(0) << "begin build_graph_type_keys";
  graph_type_keys_.clear();
  graph_type_keys_.resize(this->feature_to_id.size());

  int cnt = 0;
  for (auto &it : this->feature_to_id) {
    auto node_idx = it.second;
    std::vector<std::vector<uint64_t>> keys;
    this->get_all_id(1, node_idx, 1, &keys);
    type_to_index_[node_idx] = cnt;
    graph_type_keys_[cnt++] = std::move(keys[0]);
  }
  VLOG(0) << "finish build_graph_type_keys";

  VLOG(0) << "begin insert feature into graph_total_keys";
  // build feature embedding id
  for (auto &it : this->feature_to_id) {
    auto node_idx = it.second;
    std::vector<std::vector<uint64_t>> keys;
    this->get_all_feature_ids(1, node_idx, 1, &keys);
    graph_total_keys_.insert(
        graph_total_keys_.end(), keys[0].begin(), keys[0].end());
  }
  VLOG(0) << "finish insert feature into graph_total_keys";
}

2564 2565
}  // namespace distributed
};  // namespace paddle