Deepfakes for Code and the Asymmetric Internet
I recently came across a GitHub repo that I found fascinating. It’s called ruvnet/RuView; it has ~29k stars and is the #1 trending repo for this month as of this writing. It claims to turn commodity WiFi signals into real-time human pose estimation and vital sign monitoring (the idea is real and based on actual research). It made the rounds on social media. People on Reddit are even asking how to protect themselves against it.
RuView is the #1 trending repo on GitHub this month even though it doesn't do anything useful.This particular repo, however, does nothing useful. The project’s “pose estimation” is a hardcoded skeleton template wiggled by sine waves, and with a hard-coded walking animation. Before the author rewrote (obscured?) everything in Rust, this was even clearer: The code that was supposed to return data from the WiFi sensor returned random numbers.
It’s essentially a deepfake, but for code. There’s plausible-looking Rust and Python code, a convincing README, all the right buzzwords. It looks real enough to fool thousands of developers into starring it.
This isn’t an isolated incident; the author has over a hundred similar repos on their GitHub, all of them seemingly AI-generated.1
These repos are symptoms of something more fundamental. The internet always made it cheap to distribute noise—spam and SEO clickbait have been around for decades. But AI has made it cheap to produce convincing noise too. A single person can now mass-produce plausible-looking repos, articles, and images at near-zero marginal cost. And every receiver pays the verification cost—cross-referencing, fact-checking, checking if code actually works—independently.
But AI doesn’t just increase the noise. It simultaneously makes it possible to extract signal at scale. The same technology creates the noise and powers the filter—but only for those who can afford it.
Consider Meta. They’ve invested billions in AI infrastructure for ad targeting. Apple’s App Tracking Transparency (ATT) made it significantly harder to track users across apps, degrading the signal that advertisers rely on. In response Meta built models to infer user intent and behavior from noisier data—and it worked: by Q4 2023 they were reporting record revenue of $40B, up 25% year-over-year, and completely recovered from a brutal 2022.
On the Q3 2025 earnings call, Zuckerberg said:
But any compute that we don’t need for [AI research], we feel pretty good that we’re going to be able to absorb a very large amount of that to just convert into more intelligence and better recommendations in our family of apps and ads in a profitable way.
Meta has essentially built industrial-grade signal extraction for their advertising channel—and they think it will keep getting better the more compute they can throw at it.
ATT, it turns out, was a gift in disguise for Meta. If ad targeting becomes harder due to noise, the companies that are best at signal extraction gain a more defensible moat. The increasing noise actually helps incumbents by raising the bar beyond what smaller players can afford. This is happening for Meta and I expect the same dynamic to play out in financial markets, intelligence, and anywhere else that extracting signal from noisy data confers an advantage.
This asymmetry is structural—it follows directly from the economics of AI deployment. Building and running sophisticated models for signal extraction is expensive while the cost of noise generation is approaching zero. So we end up with a world where only well-resourced actors—large tech companies, governments, sophisticated financial firms—can afford verification at scale. For the average user, the utility of the internet degrades. For smaller companies, it becomes increasingly difficult to compete. This should worry anybody who wants an open and egalitarian internet.
Footnotes
-
I don’t know why these were produced. Maybe it’s a marketing scheme—the author offers their consultancy services for $1,500/hour. ↩︎