db_impl.h 40.2 KB
Newer Older
1
//  Copyright (c) 2011-present, Facebook, Inc.  All rights reserved.
2 3 4 5
//  This source code is licensed under the BSD-style license found in the
//  LICENSE file in the root directory of this source tree. An additional grant
//  of patent rights can be found in the PATENTS file in the same directory.
//
J
jorlow@chromium.org 已提交
6 7 8
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file. See the AUTHORS file for names of contributors.
9
#pragma once
K
Kai Liu 已提交
10

H
Haobo Xu 已提交
11
#include <atomic>
12
#include <deque>
13
#include <functional>
14
#include <limits>
I
Igor Canadi 已提交
15
#include <list>
16
#include <queue>
17
#include <set>
I
Igor Canadi 已提交
18
#include <string>
19 20
#include <utility>
#include <vector>
K
kailiu 已提交
21

22
#include "db/column_family.h"
23
#include "db/compaction_job.h"
24
#include "db/dbformat.h"
25
#include "db/flush_job.h"
26 27
#include "db/flush_scheduler.h"
#include "db/internal_stats.h"
A
agiardullo 已提交
28 29
#include "db/log_writer.h"
#include "db/snapshot_impl.h"
30
#include "db/version_edit.h"
I
Igor Canadi 已提交
31
#include "db/wal_manager.h"
32 33
#include "db/write_controller.h"
#include "db/write_thread.h"
34
#include "db/writebuffer.h"
K
Kai Liu 已提交
35 36
#include "memtable_list.h"
#include "port/port.h"
37 38 39 40
#include "rocksdb/db.h"
#include "rocksdb/env.h"
#include "rocksdb/memtablerep.h"
#include "rocksdb/transaction_log.h"
S
sdong 已提交
41
#include "table/scoped_arena_iterator.h"
42
#include "util/autovector.h"
I
Igor Canadi 已提交
43 44
#include "util/event_logger.h"
#include "util/hash.h"
45
#include "util/instrumented_mutex.h"
46 47
#include "util/stop_watch.h"
#include "util/thread_local.h"
48

49
namespace rocksdb {
J
jorlow@chromium.org 已提交
50 51 52 53 54 55

class MemTable;
class TableCache;
class Version;
class VersionEdit;
class VersionSet;
56
class Arena;
A
agiardullo 已提交
57
class WriteCallback;
I
Igor Canadi 已提交
58
struct JobContext;
59
struct ExternalSstFileInfo;
J
jorlow@chromium.org 已提交
60 61 62

class DBImpl : public DB {
 public:
I
Igor Canadi 已提交
63
  DBImpl(const DBOptions& options, const std::string& dbname);
J
jorlow@chromium.org 已提交
64 65 66
  virtual ~DBImpl();

  // Implementations of the DB interface
67 68
  using DB::Put;
  virtual Status Put(const WriteOptions& options,
69
                     ColumnFamilyHandle* column_family, const Slice& key,
I
Igor Sugak 已提交
70
                     const Slice& value) override;
71 72
  using DB::Merge;
  virtual Status Merge(const WriteOptions& options,
73
                       ColumnFamilyHandle* column_family, const Slice& key,
I
Igor Sugak 已提交
74
                       const Slice& value) override;
75 76
  using DB::Delete;
  virtual Status Delete(const WriteOptions& options,
I
Igor Sugak 已提交
77 78
                        ColumnFamilyHandle* column_family,
                        const Slice& key) override;
A
Andres Noetzli 已提交
79 80 81 82
  using DB::SingleDelete;
  virtual Status SingleDelete(const WriteOptions& options,
                              ColumnFamilyHandle* column_family,
                              const Slice& key) override;
83
  using DB::Write;
I
Igor Sugak 已提交
84 85
  virtual Status Write(const WriteOptions& options,
                       WriteBatch* updates) override;
A
agiardullo 已提交
86

87
  using DB::Get;
J
jorlow@chromium.org 已提交
88
  virtual Status Get(const ReadOptions& options,
89
                     ColumnFamilyHandle* column_family, const Slice& key,
I
Igor Sugak 已提交
90
                     std::string* value) override;
91 92 93
  using DB::MultiGet;
  virtual std::vector<Status> MultiGet(
      const ReadOptions& options,
94
      const std::vector<ColumnFamilyHandle*>& column_family,
I
Igor Sugak 已提交
95 96
      const std::vector<Slice>& keys,
      std::vector<std::string>* values) override;
97

98
  virtual Status CreateColumnFamily(const ColumnFamilyOptions& options,
99
                                    const std::string& column_family,
I
Igor Sugak 已提交
100 101
                                    ColumnFamilyHandle** handle) override;
  virtual Status DropColumnFamily(ColumnFamilyHandle* column_family) override;
102

103 104 105 106
  // Returns false if key doesn't exist in the database and true if it may.
  // If value_found is not passed in as null, then return the value if found in
  // memory. On return, if value was found, then value_found will be set to true
  // , otherwise false.
107
  using DB::KeyMayExist;
108
  virtual bool KeyMayExist(const ReadOptions& options,
109
                           ColumnFamilyHandle* column_family, const Slice& key,
I
Igor Sugak 已提交
110 111
                           std::string* value,
                           bool* value_found = nullptr) override;
112 113
  using DB::NewIterator;
  virtual Iterator* NewIterator(const ReadOptions& options,
I
Igor Sugak 已提交
114
                                ColumnFamilyHandle* column_family) override;
115 116
  virtual Status NewIterators(
      const ReadOptions& options,
I
Igor Canadi 已提交
117
      const std::vector<ColumnFamilyHandle*>& column_families,
I
Igor Sugak 已提交
118 119 120
      std::vector<Iterator*>* iterators) override;
  virtual const Snapshot* GetSnapshot() override;
  virtual void ReleaseSnapshot(const Snapshot* snapshot) override;
121
  using DB::GetProperty;
122
  virtual bool GetProperty(ColumnFamilyHandle* column_family,
I
Igor Sugak 已提交
123
                           const Slice& property, std::string* value) override;
124 125 126
  using DB::GetIntProperty;
  virtual bool GetIntProperty(ColumnFamilyHandle* column_family,
                              const Slice& property, uint64_t* value) override;
127 128 129
  using DB::GetAggregatedIntProperty;
  virtual bool GetAggregatedIntProperty(const Slice& property,
                                        uint64_t* aggregated_value) override;
130
  using DB::GetApproximateSizes;
131
  virtual void GetApproximateSizes(ColumnFamilyHandle* column_family,
132 133
                                   const Range* range, int n, uint64_t* sizes,
                                   bool include_memtable = false) override;
134
  using DB::CompactRange;
135 136 137
  virtual Status CompactRange(const CompactRangeOptions& options,
                              ColumnFamilyHandle* column_family,
                              const Slice* begin, const Slice* end) override;
138

139
  using DB::CompactFiles;
I
Igor Sugak 已提交
140 141 142 143 144
  virtual Status CompactFiles(const CompactionOptions& compact_options,
                              ColumnFamilyHandle* column_family,
                              const std::vector<std::string>& input_file_names,
                              const int output_level,
                              const int output_path_id = -1) override;
145

146 147 148
  virtual Status PauseBackgroundWork() override;
  virtual Status ContinueBackgroundWork() override;

149 150 151
  virtual Status EnableAutoCompaction(
      const std::vector<ColumnFamilyHandle*>& column_family_handles) override;

152
  using DB::SetOptions;
I
Igor Sugak 已提交
153 154 155
  Status SetOptions(
      ColumnFamilyHandle* column_family,
      const std::unordered_map<std::string, std::string>& options_map) override;
156

157
  using DB::NumberLevels;
I
Igor Sugak 已提交
158
  virtual int NumberLevels(ColumnFamilyHandle* column_family) override;
159
  using DB::MaxMemCompactionLevel;
I
Igor Sugak 已提交
160
  virtual int MaxMemCompactionLevel(ColumnFamilyHandle* column_family) override;
161
  using DB::Level0StopWriteTrigger;
I
Igor Sugak 已提交
162 163 164 165
  virtual int Level0StopWriteTrigger(
      ColumnFamilyHandle* column_family) override;
  virtual const std::string& GetName() const override;
  virtual Env* GetEnv() const override;
166
  using DB::GetOptions;
I
Igor Sugak 已提交
167 168
  virtual const Options& GetOptions(
      ColumnFamilyHandle* column_family) const override;
169 170
  using DB::GetDBOptions;
  virtual const DBOptions& GetDBOptions() const override;
171 172
  using DB::Flush;
  virtual Status Flush(const FlushOptions& options,
I
Igor Sugak 已提交
173
                       ColumnFamilyHandle* column_family) override;
174
  virtual Status SyncWAL() override;
I
Igor Canadi 已提交
175

I
Igor Sugak 已提交
176
  virtual SequenceNumber GetLatestSequenceNumber() const override;
I
Igor Canadi 已提交
177 178

#ifndef ROCKSDB_LITE
I
Igor Sugak 已提交
179 180
  virtual Status DisableFileDeletions() override;
  virtual Status EnableFileDeletions(bool force) override;
181
  virtual int IsFileDeletionsEnabled() const;
I
Igor Canadi 已提交
182
  // All the returned filenames start with "/"
183
  virtual Status GetLiveFiles(std::vector<std::string>&,
184
                              uint64_t* manifest_file_size,
I
Igor Sugak 已提交
185 186
                              bool flush_memtable = true) override;
  virtual Status GetSortedWalFiles(VectorLogPtr& files) override;
I
Igor Canadi 已提交
187

188 189 190
  virtual Status GetUpdatesSince(
      SequenceNumber seq_number, unique_ptr<TransactionLogIterator>* iter,
      const TransactionLogIterator::ReadOptions&
I
Igor Sugak 已提交
191 192
          read_options = TransactionLogIterator::ReadOptions()) override;
  virtual Status DeleteFile(std::string name) override;
193 194
  Status DeleteFilesInRange(ColumnFamilyHandle* column_family,
                            const Slice* begin, const Slice* end);
195

I
Igor Sugak 已提交
196 197
  virtual void GetLiveFilesMetaData(
      std::vector<LiveFileMetaData>* metadata) override;
198 199 200 201 202 203 204 205 206

  // Obtains the meta data of the specified column family of the DB.
  // Status::NotFound() will be returned if the current DB does not have
  // any column family match the specified name.
  // TODO(yhchiang): output parameter is placed in the end in this codebase.
  virtual void GetColumnFamilyMetaData(
      ColumnFamilyHandle* column_family,
      ColumnFamilyMetaData* metadata) override;

207 208 209 210
  // experimental API
  Status SuggestCompactRange(ColumnFamilyHandle* column_family,
                             const Slice* begin, const Slice* end);

211 212
  Status PromoteL0(ColumnFamilyHandle* column_family, int target_level);

A
agiardullo 已提交
213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232
  // Similar to Write() but will call the callback once on the single write
  // thread to determine whether it is safe to perform the write.
  virtual Status WriteWithCallback(const WriteOptions& write_options,
                                   WriteBatch* my_batch,
                                   WriteCallback* callback);

  // Returns the sequence number that is guaranteed to be smaller than or equal
  // to the sequence number of any key that could be inserted into the current
  // memtables. It can then be assumed that any write with a larger(or equal)
  // sequence number will be present in this memtable or a later memtable.
  //
  // If the earliest sequence number could not be determined,
  // kMaxSequenceNumber will be returned.
  //
  // If include_history=true, will also search Memtables in MemTableList
  // History.
  SequenceNumber GetEarliestMemTableSequenceNumber(SuperVersion* sv,
                                                   bool include_history);

  // For a given key, check to see if there are any records for this key
233 234 235 236
  // in the memtables, including memtable history.  If cache_only is false,
  // SST files will also be checked.
  //
  // If a key is found, *found_record_for_key will be set to true and
237
  // *seq will be set to the stored sequence number for the latest
238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256
  // operation on this key or kMaxSequenceNumber if unknown.
  // If no key is found, *found_record_for_key will be set to false.
  //
  // Note: If cache_only=false, it is possible for *seq to be set to 0 if
  // the sequence number has been cleared from the record.  If the caller is
  // holding an active db snapshot, we know the missing sequence must be less
  // than the snapshot's sequence number (sequence numbers are only cleared
  // when there are no earlier active snapshots).
  //
  // If NotFound is returned and found_record_for_key is set to false, then no
  // record for this key was found.  If the caller is holding an active db
  // snapshot, we know that no key could have existing after this snapshot
  // (since we do not compact keys that have an earlier snapshot).
  //
  // Returns OK or NotFound on success,
  // other status on unexpected error.
  Status GetLatestSequenceForKey(SuperVersion* sv, const Slice& key,
                                 bool cache_only, SequenceNumber* seq,
                                 bool* found_record_for_key);
A
agiardullo 已提交
257

258 259 260 261 262 263 264
  using DB::AddFile;
  virtual Status AddFile(ColumnFamilyHandle* column_family,
                         const ExternalSstFileInfo* file_info,
                         bool move_file) override;
  virtual Status AddFile(ColumnFamilyHandle* column_family,
                         const std::string& file_path, bool move_file) override;

I
Igor Canadi 已提交
265
#endif  // ROCKSDB_LITE
266

267 268 269 270 271 272
  // Similar to GetSnapshot(), but also lets the db know that this snapshot
  // will be used for transaction write-conflict checking.  The DB can then
  // make sure not to compact any keys that would prevent a write-conflict from
  // being detected.
  const Snapshot* GetSnapshotForWriteConflictBoundary();

I
Igor Canadi 已提交
273 274 275 276
  // checks if all live files exist on file system and that their file sizes
  // match to our in-memory records
  virtual Status CheckConsistency();

277
  virtual Status GetDbIdentity(std::string& identity) const override;
278

I
Igor Canadi 已提交
279
  Status RunManualCompaction(ColumnFamilyData* cfd, int input_level,
280
                             int output_level, uint32_t output_path_id,
281
                             const Slice* begin, const Slice* end,
282
                             bool exclusive,
283
                             bool disallow_trivial_move = false);
284

285 286 287
  // Return an internal iterator over the current state of the database.
  // The keys of this iterator are internal keys (see format.h).
  // The returned iterator should be deleted when no longer needed.
S
sdong 已提交
288 289
  InternalIterator* NewInternalIterator(
      Arena* arena, ColumnFamilyHandle* column_family = nullptr);
290

291
#ifndef NDEBUG
J
jorlow@chromium.org 已提交
292
  // Extra methods (for testing) that are not in the public DB interface
I
Igor Canadi 已提交
293
  // Implemented in db_impl_debug.cc
J
jorlow@chromium.org 已提交
294

295
  // Compact any files in the named level that overlap [*begin, *end]
296
  Status TEST_CompactRange(int level, const Slice* begin, const Slice* end,
297 298
                           ColumnFamilyHandle* column_family = nullptr,
                           bool disallow_trivial_move = false);
J
jorlow@chromium.org 已提交
299

300
  // Force current memtable contents to be flushed.
301 302
  Status TEST_FlushMemTable(bool wait = true,
                            ColumnFamilyHandle* cfh = nullptr);
J
jorlow@chromium.org 已提交
303

304
  // Wait for memtable compaction
305
  Status TEST_WaitForFlushMemTable(ColumnFamilyHandle* column_family = nullptr);
306 307 308 309

  // Wait for any compaction
  Status TEST_WaitForCompact();

310 311
  // Return the maximum overlapping data (in bytes) at next level for any
  // file at a level >= 1.
312 313
  int64_t TEST_MaxNextLevelOverlappingBytes(ColumnFamilyHandle* column_family =
                                                nullptr);
314

A
Abhishek Kona 已提交
315 316
  // Return the current manifest file no.
  uint64_t TEST_Current_Manifest_FileNo();
317

318
  // get total level0 file size. Only for testing.
319
  uint64_t TEST_GetLevel0TotalSize();
320

321 322
  void TEST_GetFilesMetaData(ColumnFamilyHandle* column_family,
                             std::vector<std::vector<FileMetaData>>* metadata);
323

324 325 326 327 328 329 330 331 332 333
  void TEST_LockMutex();

  void TEST_UnlockMutex();

  // REQUIRES: mutex locked
  void* TEST_BeginWrite();

  // REQUIRES: mutex locked
  // pass the pointer that you got from TEST_BeginWrite()
  void TEST_EndWrite(void* w);
334

335
  uint64_t TEST_MaxTotalInMemoryState() const {
336 337
    return max_total_in_memory_state_;
  }
338

339 340
  size_t TEST_LogsToFreeSize();

341 342
  uint64_t TEST_LogfileNumber();

343 344 345 346 347 348
  // Returns column family name to ImmutableCFOptions map.
  Status TEST_GetAllImmutableCFOptions(
      std::unordered_map<std::string, const ImmutableCFOptions*>* iopts_map);

  Cache* TEST_table_cache() { return table_cache_.get(); }

349
  WriteController& TEST_write_controler() { return write_controller_; }
350

351 352 353
  uint64_t TEST_FindMinLogContainingOutstandingPrep();
  uint64_t TEST_FindMinPrepLogReferencedByMemTable();

354
#endif  // NDEBUG
I
Igor Canadi 已提交
355

356 357 358 359
  // Return maximum background compaction alowed to be scheduled based on
  // compaction status.
  int BGCompactionsAllowed() const;

I
Igor Canadi 已提交
360
  // Returns the list of live files in 'live' and the list
K
kailiu 已提交
361
  // of all files in the filesystem in 'candidate_files'.
I
Igor Canadi 已提交
362
  // If force == false and the last call was less than
363
  // db_options_.delete_obsolete_files_period_micros microseconds ago,
I
Igor Canadi 已提交
364 365
  // it will not fill up the job_context
  void FindObsoleteFiles(JobContext* job_context, bool force,
I
Igor Canadi 已提交
366 367 368 369 370 371
                         bool no_full_scan = false);

  // Diffs the files listed in filenames and those that do not
  // belong to live files are posibly removed. Also, removes all the
  // files in sst_delete_files and log_delete_files.
  // It is not necessary to hold the mutex when invoking this method.
I
Igor Canadi 已提交
372
  void PurgeObsoleteFiles(const JobContext& background_contet);
I
Igor Canadi 已提交
373

I
Igor Sugak 已提交
374
  ColumnFamilyHandle* DefaultColumnFamily() const override;
375

376 377
  const SnapshotList& snapshots() const { return snapshots_; }

378
  void CancelAllBackgroundWork(bool wait);
379

A
agiardullo 已提交
380 381 382 383 384 385 386 387 388 389 390
  // Find Super version and reference it. Based on options, it might return
  // the thread local cached one.
  // Call ReturnAndCleanupSuperVersion() when it is no longer needed.
  SuperVersion* GetAndRefSuperVersion(ColumnFamilyData* cfd);

  // Similar to the previous function but looks up based on a column family id.
  // nullptr will be returned if this column family no longer exists.
  // REQUIRED: this function should only be called on the write thread or if the
  // mutex is held.
  SuperVersion* GetAndRefSuperVersion(uint32_t column_family_id);

A
agiardullo 已提交
391 392 393
  // Same as above, should called without mutex held and not on write thread.
  SuperVersion* GetAndRefSuperVersionUnlocked(uint32_t column_family_id);

A
agiardullo 已提交
394 395 396 397 398 399 400 401 402 403
  // Un-reference the super version and return it to thread local cache if
  // needed. If it is the last reference of the super version. Clean it up
  // after un-referencing it.
  void ReturnAndCleanupSuperVersion(ColumnFamilyData* cfd, SuperVersion* sv);

  // Similar to the previous function but looks up based on a column family id.
  // nullptr will be returned if this column family no longer exists.
  // REQUIRED: this function should only be called on the write thread.
  void ReturnAndCleanupSuperVersion(uint32_t colun_family_id, SuperVersion* sv);

A
agiardullo 已提交
404 405 406 407
  // Same as above, should called without mutex held and not on write thread.
  void ReturnAndCleanupSuperVersionUnlocked(uint32_t colun_family_id,
                                            SuperVersion* sv);

A
agiardullo 已提交
408 409 410 411 412
  // REQUIRED: this function should only be called on the write thread or if the
  // mutex is held.  Return value only valid until next call to this function or
  // mutex is released.
  ColumnFamilyHandle* GetColumnFamilyHandle(uint32_t column_family_id);

A
agiardullo 已提交
413 414 415
  // Same as above, should called without mutex held and not on write thread.
  ColumnFamilyHandle* GetColumnFamilyHandleUnlocked(uint32_t column_family_id);

416 417 418 419 420 421 422 423 424 425 426 427 428 429
  // Returns the number of currently running flushes.
  // REQUIREMENT: mutex_ must be held when calling this function.
  int num_running_flushes() {
    mutex_.AssertHeld();
    return num_running_flushes_;
  }

  // Returns the number of currently running compactions.
  // REQUIREMENT: mutex_ must be held when calling this function.
  int num_running_compactions() {
    mutex_.AssertHeld();
    return num_running_compactions_;
  }

430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445
  // hollow transactions shell used for recovery.
  // these will then be passed to TransactionDB so that
  // locks can be reacquired before writing can resume.
  struct RecoveredTransaction {
    uint64_t log_number_;
    std::string name_;
    WriteBatch* batch_;
    explicit RecoveredTransaction(const uint64_t log, const std::string& name,
                                  WriteBatch* batch)
        : log_number_(log), name_(name), batch_(batch) {}

    ~RecoveredTransaction() { delete batch_; }
  };

  bool allow_2pc() const { return db_options_.allow_2pc; }

R
Reid Horuff 已提交
446 447 448 449 450
  std::unordered_map<std::string, RecoveredTransaction*>
  recovered_transactions() {
    return recovered_transactions_;
  }

451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477
  RecoveredTransaction* GetRecoveredTransaction(const std::string& name) {
    auto it = recovered_transactions_.find(name);
    if (it == recovered_transactions_.end()) {
      return nullptr;
    } else {
      return it->second;
    }
  }

  void InsertRecoveredTransaction(const uint64_t log, const std::string& name,
                                  WriteBatch* batch) {
    recovered_transactions_[name] = new RecoveredTransaction(log, name, batch);
    MarkLogAsContainingPrepSection(log);
  }

  void DeleteRecoveredTransaction(const std::string& name) {
    auto it = recovered_transactions_.find(name);
    assert(it != recovered_transactions_.end());
    auto* trx = it->second;
    recovered_transactions_.erase(it);
    MarkLogAsHavingPrepSectionFlushed(trx->log_number_);
    delete trx;
  }

  void MarkLogAsHavingPrepSectionFlushed(uint64_t log);
  void MarkLogAsContainingPrepSection(uint64_t log);

478
 protected:
H
heyongqiang 已提交
479 480
  Env* const env_;
  const std::string dbname_;
481
  unique_ptr<VersionSet> versions_;
482
  const DBOptions db_options_;
L
Lei Jin 已提交
483
  Statistics* stats_;
484 485
  std::unordered_map<std::string, RecoveredTransaction*>
      recovered_transactions_;
H
heyongqiang 已提交
486

S
sdong 已提交
487 488 489 490
  InternalIterator* NewInternalIterator(const ReadOptions&,
                                        ColumnFamilyData* cfd,
                                        SuperVersion* super_version,
                                        Arena* arena);
491

492 493 494
  // Except in DB::Open(), WriteOptionsFile can only be called when:
  // 1. WriteThread::Writer::EnterUnbatched() is used.
  // 2. db_mutex is held
495
  Status WriteOptionsFile();
496 497 498 499

  // The following two functions can only be called when:
  // 1. WriteThread::Writer::EnterUnbatched() is used.
  // 2. db_mutex is NOT held
500 501 502
  Status RenameTempFileToOptionsFile(const std::string& file_name);
  Status DeleteObsoleteOptionsFiles();

503
  void NotifyOnFlushCompleted(ColumnFamilyData* cfd, FileMetaData* file_meta,
504
                              const MutableCFOptions& mutable_cf_options,
505
                              int job_id, TableProperties prop);
506

O
Ori Bernstein 已提交
507
  void NotifyOnCompactionCompleted(ColumnFamilyData* cfd,
508 509
                                   Compaction *c, const Status &st,
                                   const CompactionJobStats& job_stats,
510
                                   int job_id);
O
Ori Bernstein 已提交
511

Y
Yueh-Hsuan Chiang 已提交
512 513 514 515 516 517
  void NewThreadStatusCfInfo(ColumnFamilyData* cfd) const;

  void EraseThreadStatusCfInfo(ColumnFamilyData* cfd) const;

  void EraseThreadStatusDbInfo() const;

A
agiardullo 已提交
518
  Status WriteImpl(const WriteOptions& options, WriteBatch* updates,
519 520 521 522 523 524
                   WriteCallback* callback = nullptr,
                   uint64_t* log_used = nullptr, uint64_t log_ref = 0,
                   bool disable_memtable = false);

  uint64_t FindMinLogContainingOutstandingPrep();
  uint64_t FindMinPrepLogReferencedByMemTable();
A
agiardullo 已提交
525

J
jorlow@chromium.org 已提交
526 527
 private:
  friend class DB;
528
  friend class InternalStats;
R
Reid Horuff 已提交
529
  friend class TransactionImpl;
I
Igor Canadi 已提交
530
#ifndef ROCKSDB_LITE
L
Lei Jin 已提交
531
  friend class ForwardIterator;
I
Igor Canadi 已提交
532
#endif
533
  friend struct SuperVersion;
L
Lei Jin 已提交
534
  friend class CompactedDBImpl;
A
agiardullo 已提交
535 536 537
#ifndef NDEBUG
  friend class XFTransactionWriteHandler;
#endif
538
  struct CompactionState;
539

S
Stanislau Hlebik 已提交
540
  struct WriteContext;
J
jorlow@chromium.org 已提交
541 542 543 544 545 546

  Status NewDB();

  // Recover the descriptor from persistent storage.  May do a significant
  // amount of work to recover recently logged updates.  Any changes to
  // be made to the descriptor are added to *edit.
547
  Status Recover(const std::vector<ColumnFamilyDescriptor>& column_families,
548 549
                 bool read_only = false, bool error_if_log_file_exist = false,
                 bool error_if_data_exists_in_logs = false);
J
jorlow@chromium.org 已提交
550 551 552

  void MaybeIgnoreError(Status* s) const;

553 554
  const Status CreateArchivalDirectory();

J
jorlow@chromium.org 已提交
555 556 557
  // Delete any unneeded files and stale in-memory entries.
  void DeleteObsoleteFiles();

I
Igor Canadi 已提交
558 559
  // Background process needs to call
  //     auto x = CaptureCurrentFileNumberInPendingOutputs()
560
  //     auto file_num = versions_->NewFileNumber();
I
Igor Canadi 已提交
561 562
  //     <do something>
  //     ReleaseFileNumberFromPendingOutputs(x)
563 564
  // This will protect any file with number `file_num` or greater from being
  // deleted while <do something> is running.
I
Igor Canadi 已提交
565 566 567 568 569 570 571 572 573 574 575 576
  // -----------
  // This function will capture current file number and append it to
  // pending_outputs_. This will prevent any background process to delete any
  // file created after this point.
  std::list<uint64_t>::iterator CaptureCurrentFileNumberInPendingOutputs();
  // This function should be called with the result of
  // CaptureCurrentFileNumberInPendingOutputs(). It then marks that any file
  // created between the calls CaptureCurrentFileNumberInPendingOutputs() and
  // ReleaseFileNumberFromPendingOutputs() can now be deleted (if it's not live
  // and blocked by any other pending_outputs_ calls)
  void ReleaseFileNumberFromPendingOutputs(std::list<uint64_t>::iterator v);

577
  // Flush the in-memory write buffer to storage.  Switches to a new
J
jorlow@chromium.org 已提交
578
  // log-file/memtable and writes a new descriptor iff successful.
I
Igor Canadi 已提交
579 580 581 582
  Status FlushMemTableToOutputFile(ColumnFamilyData* cfd,
                                   const MutableCFOptions& mutable_cf_options,
                                   bool* madeProgress, JobContext* job_context,
                                   LogBuffer* log_buffer);
J
jorlow@chromium.org 已提交
583

S
Stanislau Hlebik 已提交
584 585 586
  // REQUIRES: log_numbers are sorted in ascending order
  Status RecoverLogFiles(const std::vector<uint64_t>& log_numbers,
                         SequenceNumber* max_sequence, bool read_only);
J
jorlow@chromium.org 已提交
587

588
  // The following two methods are used to flush a memtable to
Y
Yueh-Hsuan Chiang 已提交
589
  // storage. The first one is used at database RecoveryTime (when the
590 591 592
  // database is opened) and is heavyweight because it holds the mutex
  // for the entire period. The second method WriteLevel0Table supports
  // concurrent flush memtables to storage.
593 594
  Status WriteLevel0TableForRecovery(int job_id, ColumnFamilyData* cfd,
                                     MemTable* mem, VersionEdit* edit);
S
sdong 已提交
595 596 597

  // num_bytes: for slowdown case, delay time is calculated based on
  //            `num_bytes` going through.
598
  Status DelayWrite(uint64_t num_bytes);
599

I
Igor Canadi 已提交
600
  Status ScheduleFlushes(WriteContext* context);
601

I
Igor Canadi 已提交
602
  Status SwitchMemtable(ColumnFamilyData* cfd, WriteContext* context);
S
Stanislau Hlebik 已提交
603

H
heyongqiang 已提交
604
  // Force current memtable contents to be flushed.
605
  Status FlushMemTable(ColumnFamilyData* cfd, const FlushOptions& options);
H
heyongqiang 已提交
606

607
  // Wait for memtable flushed
608
  Status WaitForFlushMemTable(ColumnFamilyData* cfd);
H
heyongqiang 已提交
609

I
Igor Canadi 已提交
610
#ifndef ROCKSDB_LITE
611 612 613
  Status CompactFilesImpl(
      const CompactionOptions& compact_options, ColumnFamilyData* cfd,
      Version* version, const std::vector<std::string>& input_file_names,
614 615
      const int output_level, int output_path_id, JobContext* job_context,
      LogBuffer* log_buffer);
I
Igor Canadi 已提交
616
#endif  // ROCKSDB_LITE
617 618 619

  ColumnFamilyData* GetColumnFamilyDataByName(const std::string& cf_name);

620
  void MaybeScheduleFlushOrCompaction();
621 622
  void SchedulePendingFlush(ColumnFamilyData* cfd);
  void SchedulePendingCompaction(ColumnFamilyData* cfd);
623
  static void BGWorkCompaction(void* arg);
624
  static void BGWorkFlush(void* db);
625 626
  static void UnscheduleCallback(void* arg);
  void BackgroundCallCompaction(void* arg);
627
  void BackgroundCallFlush();
I
Igor Canadi 已提交
628
  Status BackgroundCompaction(bool* madeProgress, JobContext* job_context,
629
                              LogBuffer* log_buffer, void* m = 0);
I
Igor Canadi 已提交
630
  Status BackgroundFlush(bool* madeProgress, JobContext* job_context,
H
Haobo Xu 已提交
631
                         LogBuffer* log_buffer);
J
jorlow@chromium.org 已提交
632

633 634
  void PrintStatistics();

635
  // dump rocksdb.stats to LOG
636 637
  void MaybeDumpStats();

638 639
  // Return the minimum empty level that could hold the total data in the
  // input level. Return the input level, if such level could not be found.
640 641
  int FindMinimumEmptyLevelFitting(ColumnFamilyData* cfd,
      const MutableCFOptions& mutable_cf_options, int level);
642

643 644 645
  // Move the files in the input level to the target level.
  // If target_level < 0, automatically calculate the minimum level that could
  // hold the data set.
I
Igor Canadi 已提交
646
  Status ReFitLevel(ColumnFamilyData* cfd, int level, int target_level = -1);
647

648 649 650 651 652 653
  // helper functions for adding and removing from flush & compaction queues
  void AddToCompactionQueue(ColumnFamilyData* cfd);
  ColumnFamilyData* PopFirstFromCompactionQueue();
  void AddToFlushQueue(ColumnFamilyData* cfd);
  ColumnFamilyData* PopFirstFromFlushQueue();

654 655 656
  // helper function to call after some of the logs_ were synced
  void MarkLogsSynced(uint64_t up_to, bool synced_dir, const Status& status);

657 658
  const Snapshot* GetSnapshotImpl(bool is_write_conflict_boundary);

J
jorlow@chromium.org 已提交
659
  // table_cache_ provides its own synchronization
I
Igor Canadi 已提交
660
  std::shared_ptr<Cache> table_cache_;
J
jorlow@chromium.org 已提交
661

662
  // Lock over the persistent DB state.  Non-nullptr iff successfully acquired.
J
jorlow@chromium.org 已提交
663 664
  FileLock* db_lock_;

665 666 667 668
  // The mutex for options file related operations.
  // NOTE: should never acquire options_file_mutex_ and mutex_ at the
  //       same time.
  InstrumentedMutex options_files_mutex_;
J
jorlow@chromium.org 已提交
669
  // State below is protected by mutex_
670
  InstrumentedMutex mutex_;
671

I
Igor Canadi 已提交
672
  std::atomic<bool> shutting_down_;
673 674
  // This condition variable is signaled on these conditions:
  // * whenever bg_compaction_scheduled_ goes down to 0
675
  // * if AnyManualCompaction, whenever a compaction finishes, even if it hasn't
676 677 678 679 680
  // made any progress
  // * whenever a compaction made any progress
  // * whenever bg_flush_scheduled_ value decreases (i.e. whenever a flush is
  // done, even if it didn't make any progress)
  // * whenever there is an error in background flush or compaction
681
  InstrumentedCondVar bg_cv_;
682
  uint64_t logfile_number_;
S
Sage Weil 已提交
683 684
  std::deque<uint64_t>
      log_recycle_files;  // a list of log files that we can recycle
685
  bool log_dir_synced_;
I
Igor Canadi 已提交
686
  bool log_empty_;
687
  ColumnFamilyHandleImpl* default_cf_handle_;
688
  InternalStats* default_cf_internal_stats_;
689
  unique_ptr<ColumnFamilyMemTablesImpl> column_family_memtables_;
I
Igor Canadi 已提交
690 691
  struct LogFileNumberSize {
    explicit LogFileNumberSize(uint64_t _number)
692
        : number(_number) {}
I
Igor Canadi 已提交
693 694
    void AddSize(uint64_t new_size) { size += new_size; }
    uint64_t number;
695 696 697 698
    uint64_t size = 0;
    bool getting_flushed = false;
  };
  struct LogWriterNumber {
699 700 701 702 703 704 705 706 707 708 709 710 711 712
    // pass ownership of _writer
    LogWriterNumber(uint64_t _number, log::Writer* _writer)
        : number(_number), writer(_writer) {}

    log::Writer* ReleaseWriter() {
      auto* w = writer;
      writer = nullptr;
      return w;
    }
    void ClearWriter() {
      delete writer;
      writer = nullptr;
    }

713
    uint64_t number;
714 715 716
    // Visual Studio doesn't support deque's member to be noncopyable because
    // of a unique_ptr as a member.
    log::Writer* writer;  // own
717 718
    // true for some prefix of logs_
    bool getting_synced = false;
I
Igor Canadi 已提交
719 720
  };
  std::deque<LogFileNumberSize> alive_log_files_;
721 722 723 724 725 726 727 728 729 730
  // Log files that aren't fully synced, and the current log file.
  // Synchronization:
  //  - push_back() is done from write thread with locked mutex_,
  //  - pop_front() is done from any thread with locked mutex_,
  //  - back() and items with getting_synced=true are not popped,
  //  - it follows that write thread with unlocked mutex_ can safely access
  //    back() and items with getting_synced=true.
  std::deque<LogWriterNumber> logs_;
  // Signaled when getting_synced becomes false for some of the logs_.
  InstrumentedCondVar log_sync_cv_;
I
Igor Canadi 已提交
731 732 733 734
  uint64_t total_log_size_;
  // only used for dynamically adjusting max_total_wal_size. it is a sum of
  // [write_buffer_size * max_write_buffer_number] over all column families
  uint64_t max_total_in_memory_state_;
735 736 737
  // If true, we have only one (default) column family. We use this to optimize
  // some code-paths
  bool single_column_family_mode_;
738 739 740
  // If this is non-empty, we need to delete these log files in background
  // threads. Protected by db mutex.
  autovector<log::Writer*> logs_to_free_;
I
Igor Canadi 已提交
741

S
sdong 已提交
742 743
  bool is_snapshot_supported_;

744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771
  // Class to maintain directories for all database paths other than main one.
  class Directories {
   public:
    Status SetDirectories(Env* env, const std::string& dbname,
                          const std::string& wal_dir,
                          const std::vector<DbPath>& data_paths);

    Directory* GetDataDir(size_t path_id);

    Directory* GetWalDir() {
      if (wal_dir_) {
        return wal_dir_.get();
      }
      return db_dir_.get();
    }

    Directory* GetDbDir() { return db_dir_.get(); }

   private:
    std::unique_ptr<Directory> db_dir_;
    std::vector<std::unique_ptr<Directory>> data_dirs_;
    std::unique_ptr<Directory> wal_dir_;

    Status CreateAndNewDirectory(Env* env, const std::string& dirname,
                                 std::unique_ptr<Directory>* directory) const;
  };

  Directories directories_;
772

773 774
  WriteBuffer write_buffer_;

I
Igor Canadi 已提交
775 776
  WriteThread write_thread_;

777
  WriteBatch tmp_batch_;
778

779
  WriteController write_controller_;
S
sdong 已提交
780 781 782 783 784

  // Size of the last batch group. In slowdown mode, next write needs to
  // sleep if it uses up the quota.
  uint64_t last_batch_group_size_;

I
Igor Canadi 已提交
785
  FlushScheduler flush_scheduler_;
786

J
jorlow@chromium.org 已提交
787 788
  SnapshotList snapshots_;

I
Igor Canadi 已提交
789 790 791 792 793 794 795 796 797 798
  // For each background job, pending_outputs_ keeps the current file number at
  // the time that background job started.
  // FindObsoleteFiles()/PurgeObsoleteFiles() never deletes any file that has
  // number bigger than any of the file number in pending_outputs_. Since file
  // numbers grow monotonically, this also means that pending_outputs_ is always
  // sorted. After a background job is done executing, its file number is
  // deleted from pending_outputs_, which allows PurgeObsoleteFiles() to clean
  // it up.
  // State is protected with db mutex.
  std::list<uint64_t> pending_outputs_;
J
jorlow@chromium.org 已提交
799

800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825
  // flush_queue_ and compaction_queue_ hold column families that we need to
  // flush and compact, respectively.
  // A column family is inserted into flush_queue_ when it satisfies condition
  // cfd->imm()->IsFlushPending()
  // A column family is inserted into compaction_queue_ when it satisfied
  // condition cfd->NeedsCompaction()
  // Column families in this list are all Ref()-erenced
  // TODO(icanadi) Provide some kind of ReferencedColumnFamily class that will
  // do RAII on ColumnFamilyData
  // Column families are in this queue when they need to be flushed or
  // compacted. Consumers of these queues are flush and compaction threads. When
  // column family is put on this queue, we increase unscheduled_flushes_ and
  // unscheduled_compactions_. When these variables are bigger than zero, that
  // means we need to schedule background threads for compaction and thread.
  // Once the background threads are scheduled, we decrease unscheduled_flushes_
  // and unscheduled_compactions_. That way we keep track of number of
  // compaction and flush threads we need to schedule. This scheduling is done
  // in MaybeScheduleFlushOrCompaction()
  // invariant(column family present in flush_queue_ <==>
  // ColumnFamilyData::pending_flush_ == true)
  std::deque<ColumnFamilyData*> flush_queue_;
  // invariant(column family present in compaction_queue_ <==>
  // ColumnFamilyData::pending_compaction_ == true)
  std::deque<ColumnFamilyData*> compaction_queue_;
  int unscheduled_flushes_;
  int unscheduled_compactions_;
826

827
  // count how many background compactions are running or have been scheduled
828
  int bg_compaction_scheduled_;
J
jorlow@chromium.org 已提交
829

830 831 832
  // stores the number of compactions are currently running
  int num_running_compactions_;

833 834 835
  // number of background memtable flush jobs, submitted to the HIGH pool
  int bg_flush_scheduled_;

836 837 838
  // stores the number of flushes are currently running
  int num_running_flushes_;

H
hans@chromium.org 已提交
839 840
  // Information for a manual compaction
  struct ManualCompaction {
I
Igor Canadi 已提交
841
    ColumnFamilyData* cfd;
842 843
    int input_level;
    int output_level;
844
    uint32_t output_path_id;
L
Lei Jin 已提交
845
    Status status;
846
    bool done;
847
    bool in_progress;             // compaction request being processed?
848 849 850
    bool incomplete;              // only part of requested range compacted
    bool exclusive;               // current behavior of only one manual
    bool disallow_trivial_move;   // Force actual compaction to run
851 852
    const InternalKey* begin;     // nullptr means beginning of key range
    const InternalKey* end;       // nullptr means end of key range
853
    InternalKey* manual_end;      // how far we are compacting
854
    InternalKey tmp_storage;      // Used to keep track of compaction progress
855 856 857 858 859 860 861 862
    InternalKey tmp_storage1;     // Used to keep track of compaction progress
    Compaction* compaction;
  };
  std::deque<ManualCompaction*> manual_compaction_dequeue_;

  struct CompactionArg {
    DBImpl* db;
    ManualCompaction* m;
H
hans@chromium.org 已提交
863
  };
J
jorlow@chromium.org 已提交
864 865 866 867

  // Have we encountered a background error in paranoid mode?
  Status bg_error_;

868
  // shall we disable deletion of obsolete files
869 870 871 872 873 874
  // if 0 the deletion is enabled.
  // if non-zero, files will not be getting deleted
  // This enables two different threads to call
  // EnableFileDeletions() and DisableFileDeletions()
  // without any synchronization
  int disable_delete_obsolete_files_;
875

I
Igor Canadi 已提交
876 877
  // next time when we should run DeleteObsoleteFiles with full scan
  uint64_t delete_obsolete_files_next_run_;
878

879
  // last time stats were dumped to LOG
H
Haobo Xu 已提交
880
  std::atomic<uint64_t> last_stats_dump_time_microsec_;
881

882 883 884 885
  // Each flush or compaction gets its own job id. this counter makes sure
  // they're unique
  std::atomic<int> next_job_id_;

886 887 888 889
  // A flag indicating whether the current rocksdb database has any
  // data that is not yet persisted into either WAL or SST file.
  // Used when disableWAL is true.
  bool has_unpersisted_data_;
890

H
heyongqiang 已提交
891
  static const int KEEP_LOG_FILE_NUM = 1000;
D
Dmitri Smirnov 已提交
892
  // MSVC version 1800 still does not have constexpr for ::max()
893
  static const uint64_t kNoTimeOut = port::kMaxUint64;
D
Dmitri Smirnov 已提交
894

H
heyongqiang 已提交
895
  std::string db_absolute_path_;
H
heyongqiang 已提交
896

897
  // The options to access storage files
L
Lei Jin 已提交
898
  const EnvOptions env_options_;
899

I
Igor Canadi 已提交
900 901 902 903
#ifndef ROCKSDB_LITE
  WalManager wal_manager_;
#endif  // ROCKSDB_LITE

I
Igor Canadi 已提交
904 905 906
  // Unified interface for logging events
  EventLogger event_logger_;

907
  // A value of > 0 temporarily disables scheduling of background work
908
  int bg_work_paused_;
909

910 911 912
  // A value of > 0 temporarily disables scheduling of background compaction
  int bg_compaction_paused_;

913 914 915
  // Guard against multiple concurrent refitting
  bool refitting_level_;

916 917 918
  // Indicate DB was opened successfully
  bool opened_successfully_;

919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940
  // minmum log number still containing prepared data.
  // this is used by FindObsoleteFiles to determine which
  // flushed logs we must keep around because they still
  // contain prepared data which has not been flushed or rolled back
  std::priority_queue<uint64_t, std::vector<uint64_t>, std::greater<uint64_t>>
      min_log_with_prep_;

  // to be used in conjunction with min_log_with_prep_.
  // once a transaction with data in log L is committed or rolled back
  // rather than removing the value from the heap we add that value
  // to prepared_section_completed_ which maps LOG -> instance_count
  // since a log could contain multiple prepared sections
  //
  // when trying to determine the minmum log still active we first
  // consult min_log_with_prep_. while that root value maps to
  // a value > 0 in prepared_section_completed_ we decrement the
  // instance_count for that log and pop the root value in
  // min_log_with_prep_. This will work the same as a min_heap
  // where we are deleteing arbitrary elements and the up heaping.
  std::unordered_map<uint64_t, uint64_t> prepared_section_completed_;
  std::mutex prep_heap_mutex_;

J
jorlow@chromium.org 已提交
941 942 943 944
  // No copying allowed
  DBImpl(const DBImpl&);
  void operator=(const DBImpl&);

945 946 947 948 949 950
  // Return the earliest snapshot where seqno is visible.
  // Store the snapshot right before that, if any, in prev_snapshot
  inline SequenceNumber findEarliestVisibleSnapshot(
    SequenceNumber in,
    std::vector<SequenceNumber>& snapshots,
    SequenceNumber* prev_snapshot);
951

I
Igor Canadi 已提交
952
  // Background threads call this function, which is just a wrapper around
I
Igor Canadi 已提交
953 954 955
  // the InstallSuperVersion() function. Background threads carry
  // job_context which can have new_superversion already
  // allocated.
I
Igor Canadi 已提交
956
  void InstallSuperVersionAndScheduleWorkWrapper(
I
Igor Canadi 已提交
957 958
      ColumnFamilyData* cfd, JobContext* job_context,
      const MutableCFOptions& mutable_cf_options);
L
Lei Jin 已提交
959

960 961 962
  // All ColumnFamily state changes go through this function. Here we analyze
  // the new state and we schedule background work if we detect that the new
  // state needs flush or compaction.
I
Igor Canadi 已提交
963 964 965
  SuperVersion* InstallSuperVersionAndScheduleWork(
      ColumnFamilyData* cfd, SuperVersion* new_sv,
      const MutableCFOptions& mutable_cf_options);
I
Igor Canadi 已提交
966

I
Igor Canadi 已提交
967
#ifndef ROCKSDB_LITE
I
Igor Canadi 已提交
968 969 970
  using DB::GetPropertiesOfAllTables;
  virtual Status GetPropertiesOfAllTables(ColumnFamilyHandle* column_family,
                                          TablePropertiesCollection* props)
971
      override;
972
  virtual Status GetPropertiesOfTablesInRange(
973
      ColumnFamilyHandle* column_family, const Range* range, std::size_t n,
974 975
      TablePropertiesCollection* props) override;

I
Igor Canadi 已提交
976
#endif  // ROCKSDB_LITE
977

978 979
  // Function that Get and KeyMayExist call with no_io true or false
  // Note: 'value_found' from KeyMayExist propagates here
980 981 982
  Status GetImpl(const ReadOptions& options, ColumnFamilyHandle* column_family,
                 const Slice& key, std::string* value,
                 bool* value_found = nullptr);
983

984
  bool GetIntPropertyInternal(ColumnFamilyData* cfd,
985 986
                              const DBPropertyInfo& property_info,
                              bool is_locked, uint64_t* value);
987 988 989 990 991

  bool HasPendingManualCompaction();
  bool HasExclusiveManualCompaction();
  void AddManualCompaction(ManualCompaction* m);
  void RemoveManualCompaction(ManualCompaction* m);
992
  bool ShouldntRunManualCompaction(ManualCompaction* m);
993 994
  bool HaveManualCompaction(ColumnFamilyData* cfd);
  bool MCOverlap(ManualCompaction* m, ManualCompaction* m1);
J
jorlow@chromium.org 已提交
995 996 997 998 999 1000 1001
};

// Sanitize db options.  The caller should delete result.info_log if
// it is not equal to src.info_log.
extern Options SanitizeOptions(const std::string& db,
                               const InternalKeyComparator* icmp,
                               const Options& src);
1002
extern DBOptions SanitizeOptions(const std::string& db, const DBOptions& src);
S
Siying Dong 已提交
1003

M
miguelportilla 已提交
1004 1005 1006 1007 1008 1009 1010
// Fix user-supplied options to be reasonable
template <class T, class V>
static void ClipToRange(T* ptr, V minvalue, V maxvalue) {
  if (static_cast<V>(*ptr) > maxvalue) *ptr = maxvalue;
  if (static_cast<V>(*ptr) < minvalue) *ptr = minvalue;
}

1011
}  // namespace rocksdb