• P
    New Bloom filter implementation for full and partitioned filters (#6007) · f059c7d9
    Peter Dillinger 提交于
    Summary:
    Adds an improved, replacement Bloom filter implementation (FastLocalBloom) for full and partitioned filters in the block-based table. This replacement is faster and more accurate, especially for high bits per key or millions of keys in a single filter.
    
    Speed
    
    The improved speed, at least on recent x86_64, comes from
    * Using fastrange instead of modulo (%)
    * Using our new hash function (XXH3 preview, added in a previous commit), which is much faster for large keys and only *slightly* slower on keys around 12 bytes if hashing the same size many thousands of times in a row.
    * Optimizing the Bloom filter queries with AVX2 SIMD operations. (Added AVX2 to the USE_SSE=1 build.) Careful design was required to support (a) SIMD-optimized queries, (b) compatible non-SIMD code that's simple and efficient, (c) flexible choice of number of probes, and (d) essentially maximized accuracy for a cache-local Bloom filter. Probes are made eight at a time, so any number of probes up to 8 is the same speed, then up to 16, etc.
    * Prefetching cache lines when building the filter. Although this optimization could be applied to the old structure as well, it seems to balance out the small added cost of accumulating 64 bit hashes for adding to the filter rather than 32 bit hashes.
    
    Here's nominal speed data from filter_bench (200MB in filters, about 10k keys each, 10 bits filter data / key, 6 probes, avg key size 24 bytes, includes hashing time) on Skylake DE (relatively low clock speed):
    
    $ ./filter_bench -quick -impl=2 -net_includes_hashing # New Bloom filter
    Build avg ns/key: 47.7135
    Mixed inside/outside queries...
      Single filter net ns/op: 26.2825
      Random filter net ns/op: 150.459
        Average FP rate %: 0.954651
    $ ./filter_bench -quick -impl=0 -net_includes_hashing # Old Bloom filter
    Build avg ns/key: 47.2245
    Mixed inside/outside queries...
      Single filter net ns/op: 63.2978
      Random filter net ns/op: 188.038
        Average FP rate %: 1.13823
    
    Similar build time but dramatically faster query times on hot data (63 ns to 26 ns), and somewhat faster on stale data (188 ns to 150 ns). Performance differences on batched and skewed query loads are between these extremes as expected.
    
    The only other interesting thing about speed is "inside" (query key was added to filter) vs. "outside" (query key was not added to filter) query times. The non-SIMD implementations are substantially slower when most queries are "outside" vs. "inside". This goes against what one might expect or would have observed years ago, as "outside" queries only need about two probes on average, due to short-circuiting, while "inside" always have num_probes (say 6). The problem is probably the nastily unpredictable branch. The SIMD implementation has few branches (very predictable) and has pretty consistent running time regardless of query outcome.
    
    Accuracy
    
    The generally improved accuracy (re: Issue https://github.com/facebook/rocksdb/issues/5857) comes from a better design for probing indices
    within a cache line (re: Issue https://github.com/facebook/rocksdb/issues/4120) and improved accuracy for millions of keys in a single filter from using a 64-bit hash function (XXH3p). Design details in code comments.
    
    Accuracy data (generalizes, except old impl gets worse with millions of keys):
    Memory bits per key: FP rate percent old impl -> FP rate percent new impl
    6: 5.70953 -> 5.69888
    8: 2.45766 -> 2.29709
    10: 1.13977 -> 0.959254
    12: 0.662498 -> 0.411593
    16: 0.353023 -> 0.0873754
    24: 0.261552 -> 0.0060971
    50: 0.225453 -> ~0.00003 (less than 1 in a million queries are FP)
    
    Fixes https://github.com/facebook/rocksdb/issues/5857
    Fixes https://github.com/facebook/rocksdb/issues/4120
    
    Unlike the old implementation, this implementation has a fixed cache line size (64 bytes). At 10 bits per key, the accuracy of this new implementation is very close to the old implementation with 128-byte cache line size. If there's sufficient demand, this implementation could be generalized.
    
    Compatibility
    
    Although old releases would see the new structure as corrupt filter data and read the table as if there's no filter, we've decided only to enable the new Bloom filter with new format_version=5. This provides a smooth path for automatic adoption over time, with an option for early opt-in.
    Pull Request resolved: https://github.com/facebook/rocksdb/pull/6007
    
    Test Plan: filter_bench has been used thoroughly to validate speed, accuracy, and correctness. Unit tests have been carefully updated to exercise new and old implementations, as well as the logic to select an implementation based on context (format_version).
    
    Differential Revision: D18294749
    
    Pulled By: pdillinger
    
    fbshipit-source-id: d44c9db3696e4d0a17caaec47075b7755c262c5f
    f059c7d9
bloom_impl.h 14.5 KB