1. 10 4月, 2021 1 次提交
  2. 08 4月, 2021 4 次提交
    • G
      Fix flush reason attribution (#8150) · 48cd7a3a
      Giuseppe Ottaviano 提交于
      Summary:
      Current flush reason attribution is misleading or incorrect (depending on what the original intention was):
      
      - Flush due to WAL reaching its maximum size is attributed to `kWriteBufferManager`
      - Flushes due to full write buffer and write buffer manager are not distinguishable, both are attributed to `kWriteBufferFull`
      
      This changes the first to a new flush reason `kWALFull`, and splits the second between `kWriteBufferManager` and `kWriteBufferFull`.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8150
      
      Reviewed By: zhichao-cao
      
      Differential Revision: D27569645
      
      Pulled By: ot
      
      fbshipit-source-id: 7e3c8ca186a6e71976e6b8e937297eebd4b769cc
      48cd7a3a
    • A
      Enable backup/restore for Integrated BlobDB in stress and crash tests (#8165) · 0be89e87
      Akanksha Mahajan 提交于
      Summary:
      Enable backup/restore functionality with Integrated BlobDB in
      db_stress and crash test.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8165
      
      Test Plan:
      Ran python3 -u tools/db_crashtest.py --simple whitebox along
      with :
        1. decreased "backup_in_one" value for backups to be more frequent and
        2. manually changed code for "enable_blob_file" to be always true and
           apply blobdb params 100% for testing purpose.
      
      Reviewed By: ltamasi
      
      Differential Revision: D27636025
      
      Pulled By: akankshamahajan15
      
      fbshipit-source-id: 0d0e0d1479ced163f992872dc998e79c581bfc99
      0be89e87
    • A
      Integrated BlobDB for backup/restore support (#8129) · d52b520d
      Akanksha Mahajan 提交于
      Summary:
      Add support for blob files for backup/restore like table files.
          Since DB session ID is currently not supported for blob files (there is no place to store it in
          the header), so for blob files uses the
          kLegacyCrc32cAndFileSize naming scheme even if
          share_files_with_checksum_naming is set to kUseDbSessionId.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8129
      
      Test Plan: Add new test units
      
      Reviewed By: ltamasi
      
      Differential Revision: D27408510
      
      Pulled By: akankshamahajan15
      
      fbshipit-source-id: b27434d189a639ef3e6ad165c61a143a2daaf06e
      d52b520d
    • P
      Fix read-only DB writing to filesystem with write_dbid_to_manifest (#8164) · a4e82a3c
      Peter Dillinger 提交于
      Summary:
      Fixing another crash test failure in the case of
      write_dbid_to_manifest=true and reading a backup as read-only DB.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8164
      
      Test Plan:
      enhanced unit test for backup as read-only DB, ran
      blackbox_crash_test more with elevated backup_one_in
      
      Reviewed By: zhichao-cao
      
      Differential Revision: D27622237
      
      Pulled By: pdillinger
      
      fbshipit-source-id: 680d0f99ddb465a601737f2e3f2c80efd47384fb
      a4e82a3c
  3. 07 4月, 2021 7 次提交
    • P
      Fix crash test with backup as read-only DB (#8161) · 35af0433
      Peter Dillinger 提交于
      Summary:
      Forgot to re-test crash test after adding read-only filesystem
      enforcement to https://github.com/facebook/rocksdb/issues/8142. The problem is ReadOnlyFileSystem would reject
      CreateDirIfMissing whenever DBOptions::create_if_missing=true. The fix
      that is better for users is to allow CreateDirIfMissing in
      ReadOnlyFileSystem if the directory exists, so that they don't cause a
      failure on using create_if_missing with opening backups as read-only
      DBs. Added this option test to the unit test (in addition to being in the
      crash test).
      
      Also fixed a couple of lints.
      
      And some better messaging from 'make format' so that when you run it
      with uncommitted changes, it's clear that it's only checking the
      uncommitted changes.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8161
      
      Test Plan: local blackbox_crash_test with amplified backup_one_in
      
      Reviewed By: ajkr
      
      Differential Revision: D27614409
      
      Pulled By: pdillinger
      
      fbshipit-source-id: 63ccb626c7e34c200d61c6bca2a8f60da9015179
      35af0433
    • S
      Update installation instructions (#8158) · 6db3af11
      Sahir Hoda 提交于
      Summary:
      Updated instructions for installing zstd on CentOS and note about clang-format dependency
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8158
      
      Reviewed By: riversand963
      
      Differential Revision: D27598665
      
      Pulled By: ajkr
      
      fbshipit-source-id: e349eeb91147f3163e170cc29c8460b06d739b5b
      6db3af11
    • P
      Make backups openable as read-only DBs (#8142) · 879357fd
      Peter Dillinger 提交于
      Summary:
      A current limitation of backups is that you don't know the
      exact database state of when the backup was taken. With this new
      feature, you can at least inspect the backup's DB state without
      restoring it by opening it as a read-only DB.
      
      Rather than add something like OpenAsReadOnlyDB to the BackupEngine API,
      which would inhibit opening stackable DB implementations read-only
      (if/when their APIs support it), we instead provide a DB name and Env
      that can be used to open as a read-only DB.
      
      Possible follow-up work:
      
      * Add a version of GetBackupInfo for a single backup.
      * Let CreateNewBackup return the BackupID of the newly-created backup.
      
      Implementation details:
      
      Refactored ChrootFileSystem to split off new base class RemapFileSystem,
      which allows more general remapping of files. We use this base class to
      implement BackupEngineImpl::RemapSharedFileSystem.
      
      To minimize API impact, I decided to just add these fields `name_for_open`
      and `env_for_open` to those set by GetBackupInfo when
      include_file_details=true. Creating the RemapSharedFileSystem adds a bit
      to the memory consumption, perhaps unnecessarily in some cases, but this
      has been mitigated by (a) only initialize the RemapSharedFileSystem
      lazily when GetBackupInfo with include_file_details=true is called, and
      (b) using the existing `shared_ptr<FileInfo>` objects to hold most of the
      mapping data.
      
      To enhance API safety, RemapSharedFileSystem is wrapped by new
      ReadOnlyFileSystem which rejects any attempts to write. This uncovered a
      couple of places in which DB::OpenForReadOnly would write to the
      filesystem, so I fixed these. Added a release note because this affects
      logging.
      
      Additional minor refactoring in backupable_db.cc to support the new
      functionality.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8142
      
      Test Plan:
      new test (run with ASAN and UBSAN), added to stress test and
      ran it for a while with amplified backup_one_in
      
      Reviewed By: ajkr
      
      Differential Revision: D27535408
      
      Pulled By: pdillinger
      
      fbshipit-source-id: 04666d310aa0261ef6b2385c43ca793ce1dfd148
      879357fd
    • Y
      Fix a bug for SeekForPrev with partitioned filter and prefix (#8137) · 09528f9f
      Yanqin Jin 提交于
      Summary:
      According to https://github.com/facebook/rocksdb/issues/5907, each filter partition "should include the bloom of the prefix of the last
      key in the previous partition" so that SeekForPrev() in prefix mode can return correct result.
      The prefix of the last key in the previous partition does not necessarily have the same prefix
      as the first key in the current partition. Regardless of the first key in current partition, the
      prefix of the last key in the previous partition should be added. The existing code, however,
      does not follow this. Furthermore, there is another issue: when finishing current filter partition,
      `FullFilterBlockBuilder::AddPrefix()` is called for the first key in next filter partition, which effectively
      overwrites `last_prefix_str_` prematurely. Consequently, when the filter block builder proceeds
      to the next partition, `last_prefix_str_` will be the prefix of its first key, leaving no way of adding
      the bloom of the prefix of the last key of the previous partition.
      
      Prefix extractor is FixedLength.2.
      ```
      [  filter part 1   ]    [  filter part 2    ]
                        abc    d
      ```
      When SeekForPrev("abcd"), checking the filter partition will land on filter part 2 because "abcd" > "abc"
      but smaller than "d".
      If the filter in filter part 2 happens to return false for the test for "ab", then SeekForPrev("abcd") will build
      incorrect iterator tree in non-total-order mode.
      
      Also fix a unit test which starts to fail following this PR. `InDomain` should not fail due to assertion
      error when checking on an arbitrary key.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8137
      
      Test Plan:
      ```
      make check
      ```
      
      Without this fix, the following command will fail pretty soon.
      ```
      ./db_stress --acquire_snapshot_one_in=10000 --avoid_flush_during_recovery=0 \
      --avoid_unnecessary_blocking_io=0 --backup_max_size=104857600 --backup_one_in=0 \
      --batch_protection_bytes_per_key=0 --block_size=16384 --bloom_bits=17 \
      --bottommost_compression_type=disable --cache_index_and_filter_blocks=1 --cache_size=1048576 \
      --checkpoint_one_in=0 --checksum_type=kxxHash64 --clear_column_family_one_in=0 \
      --compact_files_one_in=1000000 --compact_range_one_in=1000000 --compaction_ttl=0 \
      --compression_max_dict_buffer_bytes=0 --compression_max_dict_bytes=0 \
      --compression_parallel_threads=1 --compression_type=zstd --compression_zstd_max_train_bytes=0 \
      --continuous_verification_interval=0 --db=/dev/shm/rocksdb/rocksdb_crashtest_whitebox \
      --db_write_buffer_size=8388608 --delpercent=5 --delrangepercent=0 --destroy_db_initially=0 --enable_blob_files=0 \
      --enable_compaction_filter=0 --enable_pipelined_write=1 --file_checksum_impl=big --flush_one_in=1000000 \
      --format_version=5 --get_current_wal_file_one_in=0 --get_live_files_one_in=1000000 --get_property_one_in=1000000 \
      --get_sorted_wal_files_one_in=0 --index_block_restart_interval=4 --index_type=2 --ingest_external_file_one_in=0 \
      --iterpercent=10 --key_len_percent_dist=1,30,69 --level_compaction_dynamic_level_bytes=True \
      --log2_keys_per_lock=10 --long_running_snapshots=1 --mark_for_compaction_one_file_in=0 \
      --max_background_compactions=20 --max_bytes_for_level_base=10485760 --max_key=100000000 --max_key_len=3 \
      --max_manifest_file_size=1073741824 --max_write_batch_group_size_bytes=16777216 --max_write_buffer_number=3 \
      --max_write_buffer_size_to_maintain=8388608 --memtablerep=skip_list --mmap_read=1 --mock_direct_io=False \
      --nooverwritepercent=0 --open_files=500000 --ops_per_thread=20000000 --optimize_filters_for_memory=0 --paranoid_file_checks=1 --partition_filters=1 --partition_pinning=0 --pause_background_one_in=1000000 \
      --periodic_compaction_seconds=0 --prefixpercent=5 --progress_reports=0 --read_fault_one_in=0 --read_only=0 \
      --readpercent=45 --recycle_log_file_num=0 --reopen=20 --secondary_catch_up_one_in=0 \
      --snapshot_hold_ops=100000 --sst_file_manager_bytes_per_sec=104857600 \
      --sst_file_manager_bytes_per_truncate=0 --subcompactions=2 --sync=0 --sync_fault_injection=False \
      --target_file_size_base=2097152 --target_file_size_multiplier=2 --test_batches_snapshots=0 --test_cf_consistency=0 \
      --top_level_index_pinning=0 --unpartitioned_pinning=1 --use_blob_db=0 --use_block_based_filter=0 \
      --use_direct_io_for_flush_and_compaction=0 --use_direct_reads=0 --use_full_merge_v1=0 --use_merge=0 \
      --use_multiget=0 --use_ribbon_filter=0 --use_txn=0 --user_timestamp_size=8 --verify_checksum=1 \
      --verify_checksum_one_in=1000000 --verify_db_one_in=100000 --write_buffer_size=4194304 \
      --write_dbid_to_manifest=1 --writepercent=35
      ```
      
      Reviewed By: pdillinger
      
      Differential Revision: D27553054
      
      Pulled By: riversand963
      
      fbshipit-source-id: 60e391e4a2d8d98a9a3172ec5d6176b90ec3de98
      09528f9f
    • S
      Remove check for status returned by `InvalidatePageCache` (#8156) · c4d0e66d
      sunby 提交于
      Summary:
      Failures in `InvalidatePageCache` will change the API contract. So we remove the status check for `InvalidatePageCache` in `SstFileWriter::Add()`, `SstFileWriter::Finish` and `Rep::DeleteRange`
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8156
      
      Reviewed By: riversand963
      
      Differential Revision: D27597012
      
      Pulled By: ajkr
      
      fbshipit-source-id: 2872051695d50cc47ed0f2848dc582464c00076f
      c4d0e66d
    • Y
      Reset pinnable slice before using it in Get() (#8154) · 2d8518f5
      Yanqin Jin 提交于
      Summary:
      Fixes https://github.com/facebook/rocksdb/issues/6548.
      If we do not reset the pinnable slice before calling get, we will see the following assertion failure
      while running the test with multiple column families.
      ```
      db_bench: ./include/rocksdb/slice.h:168: void rocksdb::PinnableSlice::PinSlice(const rocksdb::Slice&, rocksdb::Cleanable*): Assertion `!pinned_' failed.
      ```
      This happens in `BlockBasedTable::Get()`.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8154
      
      Test Plan:
      ./db_bench --benchmarks=fillseq -num_column_families=3
      ./db_bench --benchmarks=readrandom -use_existing_db=1 -num_column_families=3
      
      Reviewed By: ajkr
      
      Differential Revision: D27587589
      
      Pulled By: riversand963
      
      fbshipit-source-id: 7379e7649ba40f046d6a4014c9ad629cb3f9a786
      2d8518f5
    • A
      Update ZStd. Fixes an issue with Make 3.82 (#8155) · ffd3f493
      Adam Retter 提交于
      Summary:
      The previous version of ZStd doesn't build correctly with Make 3.82. Updating it resolves the issue.
      
      jay-zhuang This also needs to be cherry-picked to:
      1. 6.17.fb
      2. 6.18.fb
      3. 6.19.fb
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8155
      
      Reviewed By: riversand963
      
      Differential Revision: D27596460
      
      Pulled By: ajkr
      
      fbshipit-source-id: ac8492245e6273f54efcc1587346a797a91c9441
      ffd3f493
  4. 06 4月, 2021 3 次提交
  5. 05 4月, 2021 1 次提交
    • P
      Make tests "parallel" and "passing ASC" by default (#8146) · bd7ddf58
      Peter Dillinger 提交于
      Summary:
      New tests should by default be expected to be parallelizeable
      and passing with ASSERT_STATUS_CHECKED. Thus, I'm changing those two
      lists to exclusions rather than inclusions.
      
      For the set of exclusions, I only listed things that currently failed
      for me when attempting not to exclude, or had some other documented
      reason. This marks many more tests as "parallel," which will potentially
      cause some failures from self-interference, but we can address those as
      they are discovered.
      
      Also changed CircleCI ASC test to be parallelized; the easy way to do
      that is to exclude building tests that don't pass ASC, which is now a
      small set.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8146
      
      Test Plan: Watch CI, etc.
      
      Reviewed By: riversand963
      
      Differential Revision: D27542782
      
      Pulled By: pdillinger
      
      fbshipit-source-id: bdd74bcd912a963ee33f3fc0d2cad2567dc7740f
      bd7ddf58
  6. 04 4月, 2021 2 次提交
  7. 03 4月, 2021 1 次提交
  8. 02 4月, 2021 3 次提交
    • A
      Add request_id in IODebugContext. (#8045) · 689b13e6
      Akanksha Mahajan 提交于
      Summary:
      Add request_id in IODebugContext which will be populated by
          underlying FileSystem for IOTracing purposes. Update IOTracer to trace
          request_id in the tracing records. Provided API
          IODebugContext::SetRequestId which will set the request_id and enable
          tracing for request_id. The API hides the implementation and underlying
          file system needs to call this API directly.
      
      Update DB::StartIOTrace API and remove redundant Env* from the
          argument as its not used and DB already has Env that is passed down to
          IOTracer.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8045
      
      Test Plan: Update unit test.
      
      Differential Revision: D26899871
      
      Pulled By: akankshamahajan15
      
      fbshipit-source-id: 56adef52ee5af0fb3060b607c3af1ec01635fa2b
      689b13e6
    • R
      version_set_test.cc: remove a redundent obj copy (#7880) · 5025c7ec
      rockeet 提交于
      Summary:
      Remove redundant obj copy
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/7880
      
      Reviewed By: akankshamahajan15
      
      Differential Revision: D26921119
      
      Pulled By: riversand963
      
      fbshipit-source-id: f227da688b067870a069e728a67799a8a95fee99
      5025c7ec
    • Z
      Replace Status with IOStatus for block fetcher IO function (#8130) · 17002365
      Zhichao Cao 提交于
      Summary:
      To propagate the IOStatus from file reads to RocksDB read logic, some of the existing status needs to be replaced by IOStatus.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8130
      
      Test Plan: make check
      
      Reviewed By: anand1976
      
      Differential Revision: D27440188
      
      Pulled By: zhichao-cao
      
      fbshipit-source-id: bbe7622c2106fe4e46871d60f7c26944e5030d78
      17002365
  9. 01 4月, 2021 4 次提交
    • A
      Fix compression dictionary sampling with dedicated range tombstone SSTs (#8141) · c43a37a9
      Andrew Kryczka 提交于
      Summary:
      Return early in case there are zero data blocks when
      `BlockBasedTableBuilder::EnterUnbuffered()` is called. This crash can
      only be triggered by applying dictionary compression to SST files that
      contain only range tombstones. It cannot be triggered by a low buffer
      limit alone since we only consider entering unbuffered mode after
      buffering a data block causing the limit to be breached, or `Finish()`ing the file. It also cannot
      be triggered by a totally empty file because those go through
      `Abandon()` rather than `Finish()` so unbuffered mode is never entered.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8141
      
      Test Plan: added a unit test that repro'd the "Floating point exception"
      
      Reviewed By: riversand963
      
      Differential Revision: D27495640
      
      Pulled By: ajkr
      
      fbshipit-source-id: a463cfba476919dc5c5c380800a75a86c31ffa23
      c43a37a9
    • shadowlux's avatar
      Merge checks into one (#8138) · a3a943bf
      shadowlux 提交于
      Summary: Pull Request resolved: https://github.com/facebook/rocksdb/pull/8138
      
      Reviewed By: zhichao-cao
      
      Differential Revision: D27475616
      
      Pulled By: riversand963
      
      fbshipit-source-id: d2815eed578a90c53d6a4e0dc4aaa232516eb4f8
      a3a943bf
    • A
      Add sample_for_compression results to table properties (#8139) · 1ba2b8a5
      Andrew Kryczka 提交于
      Summary:
      Added `TableProperties::{fast,slow}_compression_estimated_data_size`.
      These properties are present in block-based tables when
      `ColumnFamilyOptions::sample_for_compression > 0` and the necessary
      compression library is supported when the file is generated. They
      contain estimates of what `TableProperties::data_size` would be if the
      "fast"/"slow" compression library had been used instead. One
      limitation is we do not record exactly which "fast" (ZSTD or Zlib)
      or "slow" (LZ4 or Snappy) compression library produced the result.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8139
      
      Test Plan:
      - new unit test
      - ran `db_bench` with `sample_for_compression=1`; verified the `data_size` property matches the `{slow,fast}_compression_estimated_data_size` when the same compression type is used for the output file compression and the sampled compression
      
      Reviewed By: riversand963
      
      Differential Revision: D27454338
      
      Pulled By: ajkr
      
      fbshipit-source-id: 9529293de93ddac7f03b2e149d746e9f634abac4
      1ba2b8a5
    • J
      Fix getApproximateMemTableStats() return type (#8098) · a781b103
      Jay Zhuang 提交于
      Summary:
      Which should return 2 long instead of an array.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8098
      
      Reviewed By: mrambacher
      
      Differential Revision: D27308741
      
      Pulled By: jay-zhuang
      
      fbshipit-source-id: 44beea2bd28cf6779b048bebc98f2426fe95e25c
      a781b103
  10. 31 3月, 2021 5 次提交
    • M
      Pass PLATFORM_FLAGS in build_detect_platform (#8111) · 493a4e28
      mrambacher 提交于
      Summary:
      At least under MacOS, some things were excluded from the build (like Snappy) because the compilation flags were not passed in correctly.  This PR does a few things:
      - Passes the EXTRA_CXX/LDFLAGS into build_detect_platform.  This means that if some tool (like TBB for example) is not installed in a standard place, it could still be detected by build_detect_platform.  In this case, the developer would invoke: "EXTRA_CXXFLAGS=<path to TBB include> EXTRA_LDFLAGS=<path to TBB library> make", and the build script would find the tools in the extra location.
      - Changes the compilation tests to use PLATFORM_CXXFLAGS.  This change causes the EXTRA_FLAGS passed in to the script to be included in the compilation check.  Additionally, flags set by the script itself (like --std=c++11) will be used during the checks.
      
      Validated that the make_platform.mk file generated on Linux does not change with this change.  On my MacOS machine, the SNAPPY libraries are now available (they were not before as they required --std=c++11 to build).
      
      I also verified that I can build against TBB installed on my Mac by passing in the EXTRA CXX and LD FLAGS to the location in which TBB is installed.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8111
      
      Reviewed By: jay-zhuang
      
      Differential Revision: D27353516
      
      Pulled By: mrambacher
      
      fbshipit-source-id: b6b378c96dbf678bab1479556dcbcb49c47e807d
      493a4e28
    • Z
      Fix error_handler_fs_test failure due to statistics (#8136) · 335c5a6b
      Zhichao Cao 提交于
      Summary:
      Fix error_handler_fs_test failure due to statistics, it will fails due to multi-thread running and resume is different.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8136
      
      Test Plan: make check
      
      Reviewed By: akankshamahajan15
      
      Differential Revision: D27448828
      
      Pulled By: zhichao-cao
      
      fbshipit-source-id: b94255c45e9e66e93334b5ca2e4e1bfcba23fc20
      335c5a6b
    • S
      Fix possible hang issue in ~DBImpl() when flush is scheduled in LOW pool (#8125) · e6534900
      sherriiiliu 提交于
      Summary:
      In DBImpl::CloseHelper, we wait for bg_compaction_scheduled_
      and bg_flush_scheduled_ to drop to 0. Unschedule is called prior
      to cancel any unscheduled flushes/compactions. It is assumed that
      anything in the high priority is a flush, and anything in the low
      priority pool is a compaction. This assumption, however, is broken when
      the high-pri pool is full.
      As a result, bg_compaction_scheduled_ can go < 0 and bg_flush_scheduled_
      will remain > 0 and DB can be in hang state.
      The fix is, we decrement the `bg_{flush,compaction,bottom_compaction}_scheduled_`
      inside the `Unschedule{Flush,Compaction,BottomCompaction}Callback()`s. DB
      `mutex_` will make the counts atomic in `Unschedule`.
      Related discussion: https://github.com/facebook/rocksdb/issues/7928
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8125
      
      Test Plan: Added new test case which hangs without the fix.
      
      Reviewed By: jay-zhuang
      
      Differential Revision: D27390043
      
      Pulled By: ajkr
      
      fbshipit-source-id: 78a367fba9a59ac5607ad24bd1c46dc16d5ec110
      e6534900
    • J
      Unittest uses unique test db name (#8124) · 9418403c
      Jay Zhuang 提交于
      Summary:
      thread_id is only unique within a process. If we run the same test-set with multiple processes, it could cause db path collision between 2 runs, error message will be like:
      ```
      ...
      IO error: While lock file: /tmp/rocksdbtest-501//deletefile_test_8093137327721791717/LOCK: Resource temporarily unavailable
      ...
      ```
      This is could be likely reproduced by:
      ```
      gtest-parallel ./deletefile_test --gtest_filter=DeleteFileTest.BackgroundPurgeCFDropTest -r 1000 -w 1000
      ```
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8124
      
      Reviewed By: ajkr
      
      Differential Revision: D27435195
      
      Pulled By: jay-zhuang
      
      fbshipit-source-id: 850fc72cdb660edf93be9a1ca9327008c16dd720
      9418403c
    • A
      Vulnerability issue in kramdown dependency (#8131) · f03606cd
      Akanksha Mahajan 提交于
      Summary:
      GitHub has detected that a package defined in the
      docs/Gemfile.lock file of the facebook/rocksdb repository contains a
      security vulnerability.
      This patch fixes it by upgrading the version of kramdown to 2.3.1
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8131
      
      Reviewed By: jay-zhuang
      
      Differential Revision: D27418776
      
      Pulled By: akankshamahajan15
      
      fbshipit-source-id: 0a4b0b85922b9958afcbc44560584701b1c6c82d
      f03606cd
  11. 30 3月, 2021 7 次提交
    • P
      Add thread safety to BackupEngine, explain more (#8115) · ec11c23c
      Peter Dillinger 提交于
      Summary:
      BackupEngine previously had unclear but strict concurrency
      requirements that the API user must follow for safe use. Now we make
      that clear, by separating operations into "Read," "Append," and "Write"
      operations, and specifying which combinations are safe across threads on
      the same BackupEngine object (previously none; now all, using a
      read-write lock), and which are safe across different BackupEngine
      instances open on the same backup_dir.
      
      The changes to backupable_db.h should be backward compatible. It is
      mostly about eliminating copies of what should be the same function and
      (unsurprisingly) useful documentation comments were often placed on
      only one of the two copies. With the re-organization, we are also
      grouping different categories of operations. In the future we might add
      BackupEngineReadAppendOnly, but that didn't seem necessary.
      
      To mark API Read operations 'const', I had to mark some implementation
      functions 'const' and some fields mutable.
      
      Functional changes:
      * Added RWMutex locking around public API functions to implement thread
      safety on a single object. To avoid future bugs, this is another
      internal class layered on top (removing many "override" in
      BackupEngineImpl). It would be possible to allow more concurrency
      between operations, rather than mutual exclusion, but IMHO not worth the
      work.
      * Fixed a race between Open() (Initialize()) and CreateNewBackup() for
      different objects on the same backup_dir, where Initialize() could
      delete the temporary meta file created during CreateNewBackup().
      (This was found by the new test.)
      
      Also cleaned up a couple of "status checked" TODOs, and improved a
      checksum mismatch error message to include involved files.
      
      Potential follow-up work:
      * CreateNewBackup has an API wart because it doesn't tell you the
      BackupID it just created, which makes it of limited use in a multithreaded
      setting.
      * We could also consider a Refresh() function to catch up to
      changes made from another BackupEngine object to the same dir.
      * Use a lock file to prevent multiple writer BackupEngines, but this
      won't work on remote filesystems not supporting lock files.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8115
      
      Test Plan:
      new mini-stress test in backup unit tests, run with gcc,
      clang, ASC, TSAN, and UBSAN, 100 iterations each.
      
      Reviewed By: ajkr
      
      Differential Revision: D27347589
      
      Pulled By: pdillinger
      
      fbshipit-source-id: 28d82ed2ac672e44085a739ddb19d297dad14b15
      ec11c23c
    • I
      Adding safer permissions to PosixFilesystem::NewLogger (#8106) · 04191e1c
      Imanol-Mikel Barba Sabariego 提交于
      Summary:
      We have observed rocksdb databases creating info log files with world-writeable permissions.
      
      The reason why the file is created like so is because stdio streams opened with fopen calls use mode 0666, and while normally most systems have a umask of 022, in some occasions (for instance, while running daemons), you may find that the application is running with a less restrictive umask. The result is that when opening the DB, the LOG file would be created with world-writeable perms:
      
      ```
      $ ls -lh db/
      total 6.4M
      -rw-r--r-- 1 ibarba users  115 Mar 24 17:41 000004.log
      -rw-r--r-- 1 ibarba users   16 Mar 24 17:41 CURRENT
      -rw-r--r-- 1 ibarba users   37 Mar 24 17:41 IDENTITY
      -rw-r--r-- 1 ibarba users    0 Mar 24 17:41 LOCK
      -rw-rw-r-- 1 ibarba users 114K Mar 24 17:41 LOG
      -rw-r--r-- 1 ibarba users  514 Mar 24 17:41 MANIFEST-000003
      -rw-r--r-- 1 ibarba users  31K Mar 24 17:41 OPTIONS-000018
      -rw-r--r-- 1 ibarba users  31K Mar 24 17:41 OPTIONS-000020
      ```
      
      This diff replaces the fopen call with a regular open() call restricting mode, and then using fdopen to associate an stdio stream with that file descriptor. Resulting in the following files being created:
      
      ```
      -rw-r--r-- 1 ibarba users   58 Mar 24 18:16 000004.log
      -rw-r--r-- 1 ibarba users   16 Mar 24 18:16 CURRENT
      -rw-r--r-- 1 ibarba users   37 Mar 24 18:16 IDENTITY
      -rw-r--r-- 1 ibarba users    0 Mar 24 18:16 LOCK
      -rw-r--r-- 1 ibarba users 111K Mar 24 18:16 LOG
      -rw-r--r-- 1 ibarba users  514 Mar 24 18:16 MANIFEST-000003
      -rw-r--r-- 1 ibarba users  31K Mar 24 18:16 OPTIONS-000018
      -rw-r--r-- 1 ibarba users  31K Mar 24 18:16 OPTIONS-000020
      ```
      
      With the correct permissions
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8106
      
      Reviewed By: akankshamahajan15
      
      Differential Revision: D27415377
      
      Pulled By: mrambacher
      
      fbshipit-source-id: 97ac6c215700a7ea306f4a1fdf9fcf64a3cbb202
      04191e1c
    • J
      Compaction should not move data to up level (#8116) · a037bb35
      Jay Zhuang 提交于
      Summary: Pull Request resolved: https://github.com/facebook/rocksdb/pull/8116
      
      Reviewed By: ajkr, mrambacher
      
      Differential Revision: D27353828
      
      Pulled By: jay-zhuang
      
      fbshipit-source-id: 42703fb01b04d92cc097d7979e64798448852e88
      a037bb35
    • A
      range_tree requires GNU libc on ppc64 (#8070) · 24b7ebee
      Adam Retter 提交于
      Summary:
      If the platform is ppc64 and the libc is not GNU libc, then we exclude the range_tree from compilation.
      
      See https://jira.percona.com/browse/PS-7559
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8070
      
      Reviewed By: jay-zhuang
      
      Differential Revision: D27246004
      
      Pulled By: mrambacher
      
      fbshipit-source-id: 59d8433242ce7ce608988341becb4f83312445f5
      24b7ebee
    • K
      Fix comment spelling (#7960) · 25ae3807
      kshair 提交于
      Summary:
      terated -> treated
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/7960
      
      Reviewed By: ajkr
      
      Differential Revision: D26677005
      
      Pulled By: zhichao-cao
      
      fbshipit-source-id: 6221305afb263aa60f674a4113aa30cb8f3914e6
      25ae3807
    • M
      Fix check in db_bench for num shard bits to match check in LRUCache (#8110) · 1be38676
      mrambacher 提交于
      Summary:
      The check in db_bench for table_cache_numshardbits was 0 < bits <= 20, whereas the check in LRUCache was 0 < bits < 20.  Changed the two values to match to avoid a crash in db_bench on a null cache.
      
      Fixes https://github.com/facebook/rocksdb/issues/7393
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8110
      
      Reviewed By: zhichao-cao
      
      Differential Revision: D27353522
      
      Pulled By: mrambacher
      
      fbshipit-source-id: a414bd23b5bde1f071146b34cfca5e35c02de869
      1be38676
    • Y
      fix typo (#8118) · 70e80c91
      yaphet 提交于
      Summary: Pull Request resolved: https://github.com/facebook/rocksdb/pull/8118
      
      Reviewed By: ajkr
      
      Differential Revision: D27367488
      
      Pulled By: zhichao-cao
      
      fbshipit-source-id: 6ed598c74ab9232f2e56326b3a30476d473699d7
      70e80c91
  12. 29 3月, 2021 2 次提交
    • M
      Fix spelling in comments in include/rocksdb/ (#8120) · 524b10bd
      mrambacher 提交于
      Summary:
      Ran a spell check over the comments in the include/rocksdb directory and fixed any mis-spellings.
      
      There are still some variable names that are spelled incorrectly (like SizeApproximationOptions::include_memtabtles, SstFileMetaData::oldest_ancester_time) that were not fixed, as those would break compilation.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8120
      
      Reviewed By: zhichao-cao
      
      Differential Revision: D27366034
      
      Pulled By: mrambacher
      
      fbshipit-source-id: 6a3f3674890bb6acc751e9c5887a8fbb6adca5df
      524b10bd
    • Y
      Disable partitioned filters in ts stress test (#8127) · ae7a7956
      Yanqin Jin 提交于
      Summary:
      Currently, partitioned filter does not support user-defined timestamp. Disable it for now in ts stress test so that
      the contrun jobs can proceed.
      
      Pull Request resolved: https://github.com/facebook/rocksdb/pull/8127
      
      Test Plan: make crash_test_with_ts
      
      Reviewed By: ajkr
      
      Differential Revision: D27388488
      
      Pulled By: riversand963
      
      fbshipit-source-id: 5ccff18121cb537bd82f2ac072cd25efb625c666
      ae7a7956