Bloom Filters

(eli.thegreenplace.net)

199 points | by mfrw a day ago ago

60 comments

  • jeffparsons 17 hours ago

    I recently discovered "compact approximators", which can be seen as a generalisation of Bloom filters. Instead of telling you "element X is probably in the set", they tell you "a lower bound on the value associated with key X".

    If there are no collisions, you get the true value. The data structure also doesn't "fill up" the way a Bloom filter does — the older data just gets probabilistically worse, so in some use cases you can keep using the one structure continuously.

    My particular use case (which led me to "invent" the structure and then discover that of course it's already been described in the literature, haha) is implicit cache invalidation. The approximator structure(s) store keys like "latest time member X was removed from group Y". I can then use that to lazily decide which cached values I can still use, instead of doing a potentially expensive search to see what I need to invalidate at the time the group member was removed.

    I'm new to using them, so I keep getting excited about new use cases — much like when I was first introduced to Bloom filters.

    • thomasmg 11 hours ago

      This is a bit similar to count-min sketch, which can be used to calculate frequencies in a stream.

      Also interesting are "Invertible Bloom Lookup Tables" (IBLT), https://arxiv.org/pdf/1101.2245 , which allow to add and remove data, and allow to retrieve the remaining data if "little" data remains. That means, it can be used for error correction: all all the _correct_ data (let's say 1 GB of correct data) into a 1 MB IBLT. Let's say a downloader finds that the checksum doesn't match: he can download that 1 MB IBLT, and remove the data from his (invalid) download. What remains is the delta: the error correction data. I know, there are other ways you could do error correction, but this is very fast, and quite interesting technically.

    • nyrikki 12 hours ago

      While there are real use cases for the above, if you are looking at bloom filters make sure you check the above works for your need.

      First-order with least fixed point FO(LFP) == P As P=co-P but we think NP!=co-NP, bloom filters often have value for access to a small and fast path to co-NP

      In that case the collisions don't matter because often excluding membership is more valuable than proving membership and you have to choose one.

      That is why outlook used it to reduce address completion, even if a hit requires the expensive call to the server, the set of email addresses not in your address book is far larger.

      But it is all horses for courses.

      If your need does fit in P you have a lot of options, while the options for co-NP is far more rarified.

    • abetusk 14 hours ago

      Can you provide a link to a paper or reference?

    • taneq 17 hours ago

      Ooh, this seems like it could be useful for collision avoidance (where you need to know a lower bound to your time to impact.)

  • Const-me 13 hours ago

    There’s a useful but non-obvious property of these filters.

    If you have two Bloom filters which use same set of possible keys and same filter setup, you can compute intersection of their represented sets with a bitwise AND operation over the bitsets in these filters.

    If you store bitsets aligned and padded by SIMD vector size (for example, an array of __m128i in C++ as opposed to the array of uint64 values in the golang example), computing such intersection is very efficient on modern computers.

    I once used that technique to filter 3D objects from a large set using the intersection of 3 range queries, one per coordinate. Compute two Bloom filters filled with the objects returned by X and Y queries, intersect the filters, then iterate over the objects returned by the remaining Z query testing against the intersected Bloom filter. Due to probabilistic nature of the Bloom filter you still need to test for X and Y queries for the objects tested positive by the filter, however with proper setup the filter should reject vast majority of them.

    • econ 11 minutes ago

      From what I understand bloom filters have a hash per item but when i invented (hah) them I used a bit array for each property where each bit describes an item at the same offset. When searching for some properties one can do an AND on those entire arrays and eliminate candidates really fast.

    • ChuckMcM 10 hours ago

      That's pretty clever. In my 3D experiments I always struggled with minimizing the number of objects in the world I had to consider for rendering on each frame based on the view/fog(range). This seems like it would help with that.

  • ozgrakkurt 14 hours ago

    Would recommend reading rocksdb implementation of bloom filter and ribbon filter to anyone wanting learn more about the production level implementation side. It is extremely well explained in the comments and is the state of the art implementation as far as I know.

    https://github.com/facebook/rocksdb/blob/main/util/bloom_imp...

    https://github.com/facebook/rocksdb/blob/main/util/ribbon_al...

    • almostgotcaught 2 hours ago

      These are the kinds of comments that I write when I work really hard (and very long) on a PR and I know no one will really dig into it (kind of like "well at least I committed the findings to posterity").

      • econ 23 minutes ago

        If you keep at it someone someday will be blown away.

  • celeritascelery 16 hours ago

    The author says that with a 1.2GB bloom filter and 7 hash functions lookup is only 30ns. I have to assume this is because the cache has loaded all the benchmark values. My guess is that the benchmark is only testing a few elements. In real world scenarios with a bloom filter this big most of those 7 hash lookups will be into cold cache since 1.2 GB is too large. That means lookups are much longer than 30ns. Probably still faster than going to network or database though.

    • eliben 15 hours ago

      Updated the result to 80ns - thanks for flagging this. This grows with the size of the data (because more cache misses), and running the benchmark on the full billion takes a while.

      [That said, on a hot production bloom filter, much can be loaded into caches anyway so it's not an entirely un-realistic scenario that some of these are cache hits]

    • returningfory2 16 hours ago

      This is the benchmark they wrote: https://github.com/eliben/code-for-blog/blob/7278526923168d2...

      The benchmark alternates between ~1 million different keys to check in the filter, explicitly to account for cache effects.

      • Tuna-Fish 15 hours ago

        A single lookup is going to take more than 30ns, the reason they only see that is that the OoO machinery of their CPU is good enough to run those lookups in parallel.

    • Tuna-Fish 15 hours ago

      Yes, 30ns means it's in cache. But bloom filters are surprisingly fast for the amount of lookups they do, because they all happen in parallel and there is a lot of parallelism in modern memory subsystems, so that you essentially only pay the cost of a single random read for the entire lookup. If using 1GB pages, you can still realistically talk about <100ns lookups.

      • bjornsing 4 hours ago

        > so that you essentially only pay the cost of a single random read for the entire lookup

        Why would you ever pay more than that for a bloom filter lookup? I mean, I don’t see how that has anything to do with parallelism in memory subsystems. But I may be missing something.

  • aranw 14 hours ago

    Sam Rose has also written a great article on Bloom Filters and has some great animations to help illustrate it too. Definitely worth checking out https://samwho.dev/bloom-filters/

  • noelwelsh 20 hours ago

    Hash functions are kinda magic. Bloom filters are one of the fun tricks you can play with them (though Bloom filters themselves are quite old at this point; check the literature for various alternatives that are better in specific situations.) The field of data sketches and streaming algorithms has many others. k-Minimum values is a nice example of a set cardinality estimator that is very easy to understand.

    • f_devd 20 hours ago

      > check the literature for various alternatives that are better in specific situations

      For direct replacements, see Xor(+)-Filters: https://arxiv.org/abs/1912.08258 and the subsequent evolution into Binary Fuse Filters: https://arxiv.org/abs/2201.01174

      • sparkie 18 hours ago

        Also quite recent are Ribbon filters: https://arxiv.org/abs/2103.02515

        But these, and xor/binary-fuse-filters are only direct replacements for Bloom filters in some problems. There are still problems for which Bloom filters are a better solution.

        Bloom filters have the advantage that the filter produced by adding two elements, is the same as the bitwise-or of a pair of filters made by adding each element separately.

            bloom([x, y]) == bloom([x]) | bloom([y])
        
        Also, the filter produced by bitwise-and, `bloom([x]) & bloom([y])`, cannot have any bits set in it which are not also set by `bloom([x, y])`. We can assert that

            bloom([x]) & bloom([x, y]) == bloom([x])
            bloom([y]) & bloom([x, y]) == bloom([y])
            (bloom([x]) & bloom([y])) | bloom([x, y]) == bloom([x, y])
        
        There are practical applications that this can provide which are not satisfied by the "optimized" variants. The bitwise-or does set union (or least upper bound), and the bitwise-and does a set intersection (or greatest lower bound), though the filter produced by `bloom([x, y])` has a higher false positive rate than the filter produced by `bloom([x]) & bloom([y])`.

                   bloom([x]) | bloom([y])              == bloom([x, y])
                  🡥                       🡤
            bloom([x])                   bloom([y])
                  🡤                       🡥
                   bloom([x]) & bloom([y])
      • samuel 20 hours ago

        Interesting, but not quite the same:

        Xor and binary fuse filters require access to the full set of keys at construction time. In this sense, they are immutable. Alternatives have typically a fixed memory usage and a maximal capacity, but they also allow more flexibility such as progressive construction (adding keys one by one).

      • virexene 20 hours ago

        I may have missed something when skimming the paper, but it sounds like Xor filters are constructed offline and can't be modified efficiently afterwards, whereas Bloom filters can be inserted into efficiently. So they don't seem to be an exact replacement

        • f_devd 18 hours ago

          No, you are correct. I misremembered with cuckoofilters.

    • Snawoot 19 hours ago

      Cuckoo filters are more efficient alternative. The algorithm of displacement is also very notable: how we can use random swaps to place data in cells optimally.

  • redbell 12 hours ago

    Related: https://news.ycombinator.com/item?id=42293832

    Also, this -somehow- related Ask HN is a true classic, at least for me: https://news.ycombinator.com/item?id=32186203

  • klaussilveira 12 hours ago

    You can even use Bloom Filters for rate limiting with counting BF:

    https://github.com/FastFilter/fastfilter_cpp/blob/f27873fac4...

    • thomasmg 11 hours ago

      Yes, this is the succinct counting blocked Bloom filter I wrote. I wanted to write a paper about it. It requires twice the space of a blocked Bloom filter (which is about half the space of a regular counting Bloom filter). Reads are very fast, and updates are okish.

      • klaussilveira 8 hours ago

        Well, thank you for that! That library has taught me a lot. :)

  • gitroom 15 hours ago

    pretty cool seeing how stuff like this keeps getting new uses - i always wanna tinker with these things but never have the right problem yet tbh

  • fooker 14 hours ago

    Here's my use case for it: I want really small filters (say 64 bits or 128 bits), and use these to represent 50-100 element subsets of a ~1000-2000 element set.

    I know with a traditional bloom filter this would give me a lot of false positives. Is there an alternative that would fare better ?

    • thomasmg 11 hours ago

      Well, what problem do you want to solve? What you describe is not a use case but a possible solution... this smells like an xy problem...

      • fooker 6 hours ago

        The problem is:

        Represent (possibly overlapping, hence trees are tricky) subsets of size 50-100 out of a set of size of size 1000-2000.

        Some amount of false positives is acceptable but not like 50%.

  • esafak 15 hours ago
  • bobosha 18 hours ago

    what are some real-world usecases that people use it for?

    • mfrw 16 hours ago

      I like to think of it as a magical checklist, it helps us to quickly check if something _might_ be there, without actually looking through everything.

      A few non-exhaustive real world use-cases that come to mind:

      - Databases: To quickly check if a record might exist before doing a costly disk lookup.

      - Spell Checkers: To check if a word might be in the dictionary.

      - Spam Filters: To check if an email sender is on a list of known spammers.

      - Browser Security: Chrome uses Bloom filters to check if a site might be malicious.

      - Password Checker: To check if a password is known to be leaked.

      - Web Caches: To check if a URL or resource is definitely not in the cache.

      - Distributed Systems: To avoid sending data that another system definitely doesn’t need.

    • thesz 14 hours ago

      https://news.ycombinator.com/item?id=42486610

      Discussion of how Bloom filters made SQLite much faster.

    • gww 18 hours ago

      They are used somewhat commonly in bioinformatics where lookup tables can have long keys and many entries [1].

      1. https://en.m.wikipedia.org/wiki/Bloom_filters_in_bioinformat...

    • withinboredom 17 hours ago

      If you have a whole cluster of machines that have data on them and you need to ask: "does this machine probably have or not have this data?"; a bloom filter will tell you an answer. It can save a ton of time since a bloom filter's answer is "probably yes" and "definitely no."

    • nsteel 18 hours ago
    • la_fayette 13 hours ago

      Initially, Bitcoin light wallets were built with bloom filters. So a full node would only propagate transactions, which satisfy a bloom filter, given by light client to that light client. The bloom filter seems not to be privacy preserving, that was one reason why this was abondend by some wallets. However, bitcoinj and wallets built on top of it, might still use this...

    • andrewstuart 17 hours ago

      Bloom filters can be extremely fast to tell you if something is not in a dataset.

      They can give false positives incorrectly indicating an element might be in the set when it's not, but never false negatives

      Knowing (fast) if something is not in a dataset can be very useful.

    • leeoniya 18 hours ago

      permission checks (allow/denylists)

      • sparkie 17 hours ago

        Using them for whitelists is probably not a great idea because they can give false positives. An attacker could potentially flood the filter with fake accounts and increase the rate of false positives, increasing the chance they're granted access.

        For blacklists, potentially more suitable, but since it can also give false positives, it could deny permission to people who should have it. An attacker might also attack this - by flooding the filter with accounts that deliberately get blacklisted, they could lock out people who should have access.

        Obviously this is very use-case specific - it's probably not the best approach to doing permissions if security is paramount.

        • withinboredom 17 hours ago

          No, but they can tell you a user is definitely not in an allowlist or blocklist. That is useful, especially if it can save a database lookup on every check.

          • sparkie 16 hours ago

            That may work, but there are potential issues with that regarding timing attacks. If an attacker could make many attempts to access a resource, they may be able to figure out who (probably) has access with a brute-force timing test, and narrow down an attack target.

            • withinboredom 16 hours ago

              I'm not sure I understand. Generally, an allow-list/block-list is for authorized resources? By the time you are doing this check, the user is already authenticated and this is part of authorization. So, the user shouldn't be able to authenticate as arbitrary users to do a timing attack. If they can, you have bigger problems.

  • thrance 18 hours ago

    I've known about Bloom Filters for a while now, and I like the idea of them, but I've never had a need for them yet.

    • speed_spread 17 hours ago

      Good for you. Their main utility is in adtech and surveillance.

      • okr 17 hours ago

        Or, in general, all data intensive applications.

  • andrewstuart 17 hours ago

    Bloom filters - when I eventually learned about them - were nothing at all like what I had assumed from the name. And very interesting too.

    • eimrine 17 hours ago

      "Bloom's filter" that's a more correct name which doesn't let to make any assumptions.

      "Hashmap" - maybe a less correct name but it lets to make more correct assumption.

  • londons_explore 19 hours ago

    A whole load of data structures, like bloom filters, are 'efficiency tricks'.

    For example, "check if we might have Mr Smith's data cached by looking in this bloom filter". As Long as the answer is usually right, that's good enough.

    I do wonder if in the future we'll use a mini online-trained AI to achieve the same thing. "Ask the AI if we have MR Smiths data cached".

    Rather like you might ask an office clerk "Is Mr smith that troublesome customer you were dealing with last week? Do you still have his file on your desk?"

    • nkrisc 18 hours ago

      Why would that be better than a bloom filter (or similar)?

      • chupy 14 hours ago

        Because it has AI in it's name /s

    • ozgrakkurt 14 hours ago

      There are actually “learned” bloom filters if anyone is interested in machine learning/bloom filter relation. but it is not related to chatbots

    • esafak 15 hours ago

      The way that would work is that the LLM would translate that sentence into a tool call, which would query a data store that does the heavy lifting. Also, there is an ML task called "Learning to hash", which is about optimizing the hash for the task: https://learning2hash.github.io/