The Privacy Theater of Hashed Pii
Key topics
The article discusses how hashing personally identifiable information (PII) like phone numbers is often used as a superficial measure to claim privacy compliance, while the discussion revolves around the effectiveness and limitations of such practices.
Snapshot generated from the HN discussion
Discussion Activity
Active discussionFirst comment
2h
Peak period
17
Day 7
Avg / period
6.5
Based on 26 loaded comments
Key moments
- 01Story posted
Oct 19, 2025 at 11:07 PM EDT
3 months ago
Step 01 - 02First comment
Oct 20, 2025 at 12:51 AM EDT
2h after posting
Step 02 - 03Peak activity
17 comments in Day 7
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 28, 2025 at 4:01 PM EDT
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
If you added a salt, this would still allow you to reverse some particular hashed phone number in about 4 hours, it just wouldn't allow you to do all of them at the same time.
If it's not stored alongside the hash it's not a salt, it's something else.
https://en.wikipedia.org/wiki/Salt_(cryptography)
That is not even true. The definition in the article does not substantiate it. There is no requirement for the salt to be stored alongside the hash.
The definition in the article is sufficiently clear. This is all that a salt is:
> a salt is random data fed as an additional input to a one-way function that hashes data
With regard to effective anonymization, the salt is stored by the generator, but not in the exported dataset.
This isn't perfect, but there hasn't been a single customer (bank) that pushed back against it yet.
Salting does mostly solve the problem from an information theory standpoint. Correlation analysis is a borderline paranoia thing if you are practicing reasonable hygiene elsewhere.
This year it's advertised as confidential, rather than anonymous, so I suppose that is an improvement.
There is no honour amongst data thieves.
While of course not realising that GDPR implementation is partially on them and that some of those metrics are literally impossible to implement without breaching into GDPR territory. Any company saying that they are "fully GDPR compliant" but also giving you retention and attribution metrics by default is probably confusing you in this way.
https://www.microsoft.com/en-us/research/blog/password-monit...
If anything, this article understates the problem. A single Nvidia RTX4090 can calculate 164 billion MD5 hashes per second running hashcat software:
https://gist.github.com/Chick3nman/32e662a5bb63bc4f51b847bb4...
That said, surprisingly few people are aware of this fact, even senior technical leadership at Big Tech companies, so I'm not surprised dodgy Ad-Tech companies are not either, and it might be an illustration of Hanlon's Razor: do not ascribe to malice what can be better explained by incompetence (even if ad-tech companies long ago forfeited the benefit of doubt).
All the laws were passed so that companies don't not compare their customer lists without asking the customer first.
I hope some government agency picks that up and strikes such BS with might.
If you are BambooHR customer having people in your HR system - you have to ask person if you can check if they are up in BambooHR, guess what if they say no or yes you already have half of the job done.
Putting it into a hash and seeing if you have it in your database is still sharing that requires consent. Fuckers.