Austin Weeks Built a 124-Star Bot Trap That Wastes AI Companies' Millions
I watched a colleague spend three days debugging why their scraper was running forever, burning through AWS credits on what seemed like infinite pagination. Turns out they'd hit a tarpit—a deliberate trap designed to waste bot resources. That memory came flooding back when I saw Austin Weeks' Miasma trending on GitHub.
Released on March 18, 2024, this Rust-based tool does something beautifully vindictive: it traps AI web scrapers in an endless maze of poisoned content. 124 stars in 11 days. 323 Hacker News points. The developer community is clearly hungry for revenge.
The Technical Middle Finger
Miasma works with elegant simplicity. Website owners deploy it as a reverse proxy, then scatter hidden HTML links throughout their pages using display: none and aria-hidden attributes. When scrapers follow these invisible breadcrumbs, they're redirected to /bots or similar trap paths.
What happens next is poetic justice:
- Self-referential pages that link to more trap pages
- Misleading, low-quality "slop" data designed to poison training datasets
- Infinite loops that burn through scraper compute resources
- Legitimate crawlers stay safe via robots.txt whitelisting
<> "It's a technical middle finger to scrapers," one Hacker News commenter called it. The phrase captures something deeper—developer frustration boiling over into direct action./>
The Hypocrisy Problem
Here's what really stings: companies like OpenAI aggressively scrape the entire web while simultaneously blocking scrapers from their own sites. They build billion-dollar models on "stolen" content, then cry foul when others try to peek at their data.
Weeks addresses this head-on by auto-rejecting AI-generated pull requests. Even his contribution policy is a statement about human versus machine labor.
The timing isn't coincidental. This tool emerged as polite requests and robots.txt files prove toothless against determined AI companies. Why ask nicely when you can fight back?
Smart Detection, Smarter Defense
Beyond resource exhaustion, Miasma serves as sophisticated bot detection. Any browser visit to those hidden poison links immediately flags automation—perfect for instant bans. Spanish tech coverage highlighted how startups can protect their IP through simple proxy integration.
The low-memory, resource-efficient design makes economic sense. Why should small sites subsidize training data for companies worth hundreds of billions?
Hacker News discussions revealed another layer: shared IP blocking across verified domains. Imagine a distributed defense network where caught scrapers face coordinated bans. The arms race is escalating.
Poisoning the Well
Critics worry about "degrading shared web data quality" and "harming legitimate AI efforts." But whose playground are we really protecting? AI companies already cherry-pick the highest-quality content while ignoring creator preferences.
The ethical debate misses the power imbalance. Individual developers and small sites have zero leverage in data licensing negotiations. Tools like Miasma level the playing field by making unauthorized scraping uneconomical.
Some coverage bizarrely frames this as "advancing AI" or opening "new frontiers." That misses the point entirely. This isn't innovation—it's digital self-defense.
The Escalation Continues
Miasma joins a growing ecosystem of anti-scraping tools as developers shift from hoping for legislative protection to building their own. The 232 Hacker News comments reveal pent-up frustration finally finding an outlet.
Sophisticated scrapers will adapt. They'll improve bot detection, respect hidden links, maybe even parse CSS to avoid display:none traps. But each defensive move forces them to spend more resources and develop more complex systems.
My Bet: Within six months, we'll see Miasma-style traps deployed across thousands of sites, forcing AI companies into expensive cat-and-mouse games or legitimate data partnerships. The age of free scraping is ending, one poison pit at a time.

