The End of Anonymous: How AI Stripped Away Online Privacy for Just $4

A groundbreaking study reveals that artificial intelligence can now unmask anonymous social media users with alarming accuracy and at a cost that makes mass surveillance economically viable for the first time.
The $4 Identity Crisis
The age of online anonymity may be over. A team of researchers from ETH Zurich and Anthropic has demonstrated that large language models can strip away digital pseudonyms with unprecedented efficiency, correctly identifying anonymous users for as little as one to four dollars per person.
The implications are staggering. What once required teams of human investigators working for weeks can now be accomplished by AI in minutes. Simon Lermen, an AI engineer at MATS Research and lead author of the study, puts it bluntly: ‘Ask yourself: could a team of smart investigators figure out who you are from your posts? If yes, LLM agents can likely do the same, and the cost of doing so is only going down.’
In their most striking experiment, the researchers successfully matched 226 out of 338 anonymous Hacker News users to their real LinkedIn profiles with 90% precision. The entire operation cost less than $2,000 – roughly the price of a mid-range laptop.
How the Digital Detective Works
The AI system operates like a hyper-efficient private investigator, but one that never sleeps and processes information at superhuman speed. The researchers developed a three-step pipeline that extracts identity-relevant features from user posts, searches for candidate matches using semantic embeddings, and then reasons over the top candidates to verify matches.
Unlike previous deanonymization methods that required structured data, this approach works directly on raw, unstructured text – the kind of casual comments and posts that fill social media platforms. The AI looks for seemingly innocuous details: a mention of walking a dog named ‘Biscuit’ through Dolores Park, complaints about school in Seattle, or discussions about coding in Python.
What makes this particularly insidious is that each component of the attack appears benign. Summarizing text, generating embeddings, ranking candidates – none of these steps seem inherently malicious, making the system difficult to detect or restrict through conventional safeguards.
The Economics of Exposure
The true revolution isn’t in the technique – skilled human investigators could theoretically connect these dots. It’s in the economics. What was once prohibitively expensive is now accessible to anyone with an internet connection and access to commercial AI models.
Daniel Paleka, the study’s co-author from ETH Zurich, told researchers he was surprised by ‘how little information it takes to connect two accounts.’ The system doesn’t need access to specialized databases or custom training – it works with standard commercial models like ChatGPT and Claude.
This cost reduction transforms the threat landscape entirely. A motivated intelligence service could always unmask a specific dissident with enough analyst hours, but running that process against every pseudonymous account in a protest movement was prohibitively expensive. Now, mass surveillance becomes economically viable for the first time.
Real-World Implications
The researchers didn’t stop at theoretical demonstrations. They successfully identified nine out of 125 scientists from anonymized interview transcripts in Anthropic’s research dataset, using only descriptions of their research projects. On Reddit, users who commented on ten or more films in movie discussion communities could be traced in 48% of cases.
The study reveals that traditional anonymization methods are failing in the AI era. A separate 2025 paper found that state-of-the-art identifier-removal techniques still leave significant personal information recoverable from surrounding text. The ‘practical obscurity’ that once protected pseudonymous users – the assumption that linking scattered posts was too labor-intensive – no longer holds.
The most vulnerable groups are predictable: activists, whistleblowers, abuse survivors, and anyone who depends on anonymity for their safety. For these individuals, the stakes aren’t just privacy – they’re potentially life and death.
Fighting Back in the Age of AI
The researchers propose several countermeasures, though they’re honest about their limitations. Platforms could enforce stricter rate limits on API access to user data, detect automated scraping, and restrict bulk data exports. These measures would increase the cost of attacks but wouldn’t prevent them entirely.
For individual users, the most effective protection remains the most inconvenient: share less information online. Much less. The researchers recommend compartmentalizing digital identities – using different platforms for different interests, adopting different writing styles, and never using the same pseudonym across multiple services.
As Lermen notes, every comment is now a potential clue, and AI has learned to read them all at once. The old spy rule applies more than ever: never use the same cover name in two countries. In the digital age, every platform is a different country, and the borders are more porous than ever.
The Future of Digital Privacy
This research represents a fundamental shift in how we must think about online privacy. The assumption that pseudonymous posts are safe because linking them requires disproportionate effort has been shattered. As AI models become more capable and costs continue to fall, the situation will only worsen.
The trajectory is clear: deanonymization capability scales predictably with model improvements. In experiments matching Reddit users across movie discussion communities, switching from low to high reasoning effort roughly doubled the correct identification rate. Models keep getting better at reasoning, costs per token keep falling, and both curves are accelerating.
What we’re witnessing isn’t just a technical advancement – it’s the end of an era. The internet was built on the assumption that practical obscurity would protect ordinary users. That assumption is now obsolete. The question isn’t whether AI will continue to erode online anonymity, but how quickly, and whether society can adapt its privacy expectations and protections in time.









