260

https://xkcd.com/2934

Alt text:

Sometimes, you can tell Bloom filters are the wrong tool for the job, but when they're the right one you can never be sure.

top 17 comments
sorted by: hot top controversial new old
[-] hydroptic@sopuli.xyz 58 points 6 months ago

Do you want to learn about probabilistic data structures?

  • maybe
  • no
[-] LPThinker@lemmy.world 31 points 6 months ago

For anyone interested in learning more about bloom filters, this is a technical but extremely accessible and easy to follow introduction to them, including some excellent interactive visualizations: https://samwho.dev/bloom-filters/

[-] gbuttersnaps@programming.dev 6 points 6 months ago

This was a great read, thanks for sharing!

[-] randomaccount43543@lemmy.world 22 points 6 months ago
[-] Xeroxchasechase@lemmy.world 16 points 6 months ago

Is'nt bloom filter a shader that makes the picture look hazy and bright?

[-] Ephera@lemmy.ml 30 points 6 months ago

That's the bloom shader effect: https://en.wikipedia.org/wiki/Bloom_(shader_effect)

But yeah, some people might refer to that as "bloom filter", although it's not what's meant here.

[-] rockSlayer@lemmy.world 6 points 6 months ago

A bloom filter is a data structure that is most useful in creating a spell check algorithm

[-] hydroptic@sopuli.xyz 13 points 6 months ago

That's definitely not what they're most useful for. I mean, you probably can use a bloom filter for implementing spell check, but saying that's where they're most useful severely misses the point of probabilistic set membership queries.

Bloom filters and their relatives are great when you have a huge set of values – eg. 100s of millions of user IDs in some database – and you want to have a very fast way of checking whether some value might be in that set, without having to query the database. Naturally this assumes that you've prepopulated a bloom filter with whatever values you need to be checking.

If the result of the bloom filter query is "nope", you know that the value's definitely not in the set, but if the result is "maybe" then you can go ahead and double-check by querying the database. This means that the vast majority of checks don't have to hit that slow DB at all, and even though you'll get some false positives this'll still be much much much faster than having to go through that DB every time.

[-] jaybone@lemmy.world 1 points 6 months ago

In this example, what would you use to prepopulate the filter?

[-] hydroptic@sopuli.xyz 1 points 6 months ago

Which example do you mean?

If you meant my user ID example, you'd prepopulate the bloom filter with existing user IDs on eg. service startup or whatever, and then update the filter every time a new user ID is added – keeping in mind that the false positive rate will grow as more are added, and that at some point you may need to create a new filter with a bigger backing bit array

[-] jaybone@lemmy.world 2 points 6 months ago

So you’re just putting a bunch of values in memory that you can access quickly, like similar to a hash set contains(), maybe hoping for O(1) time. But other than that there’s no trick to it?

[-] hydroptic@sopuli.xyz 2 points 6 months ago* (last edited 6 months ago)

Well, yes and no. With a straight-up hash set, you're keeping set_size * bits_per_element bits plus whatever the overhead of the hash table is in memory, which might not be tenable for very large sets, but with a Bloom filter that has eg. ~1% false positive rate and an ideal k parameter (number of hash functions, see eg. the Bloom filter wiki article) you're only keeping ~10 bits per element completely regardless of element size because they don't store the elements themselves or even their full hashes – they only tell you whether some element is probably in the set or not, but you can't eg. enumerate the elements in the set. As an example of memory usage, a Bloom filter that has a false positive rate of ~1% for 500 million elements would need 571 MiB (noting that although the size of the filter doesn't grow when you insert elements, the false positive rate goes up once you go past that 500 million element count.)

Lookup and insertion time complexity for a Bloom filter is O(k) where k is the parameter I mentioned and a constant – ie. effectively O(1).

Probabilistic set membership queries are mainly useful when you're dealing with ginormous sets of elements that you can't shove into a regular in-memory hash set. A good example in the wiki article is CDN cache filtering:

Nearly three-quarters of the URLs accessed from a typical web cache are "one-hit-wonders" that are accessed by users only once and never again. It is clearly wasteful of disk resources to store one-hit-wonders in a web cache, since they will never be accessed again. To prevent caching one-hit-wonders, a Bloom filter is used to keep track of all URLs that are accessed by users. A web object is cached only when it has been accessed at least once before, i.e., the object is cached on its second request.

[-] Audalin@lemmy.world 10 points 6 months ago

There's a recent algorithm using somewhat similar ideas for approximate counting of unique objects in a stream with constant memory:

https://www.quantamagazine.org/computer-scientists-invent-an-efficient-new-way-to-count-20240516/

[-] Mubelotix@jlai.lu 2 points 6 months ago

I think I like hash-based probabilistic counting better, but this is interesting

[-] Venat0r@lemmy.world 1 points 5 months ago* (last edited 5 months ago)

Not to be confused with Bloom shader effect.

this post was submitted on 18 May 2024
260 points (98.5% liked)

xkcd

8836 readers
246 users here now

A community for a webcomic of romance, sarcasm, math, and language.

founded 1 year ago
MODERATORS