Skip to content
Go back

Dealing With AI Web Crawlers: The Carrot and The Stick

Published: Sep 2, 2025
Vancouver, Canada

The open web is under siege. As detailed in a recent article by The Register, a new generation of AI web crawlers is hammering websites with traffic spikes ‘up to ten or even twenty times normal levels within minutes.’ This isn’t the polite, well-behaved crawling of the past; this is a brute-force data grab that is knocking small sites offline and threatening passion projects with existential bandwidth bills.

The stories from the trenches are grim. One sysadmin on Hacker News described the onslaught as being ‘close to having a site get Slashdotted every single day.’ He noted the real problem isn’t just bandwidth, but the database and compute load from endless page requests. Another developer, the creator of the popular Linux gaming resource ProtonDB, shared how a single scraper was set to drive his monthly hosting bill up by $500—for a free service he runs for the community.

The old social contract of the web, governed by the humble robots.txt file, has broken down. Many AI companies, backed by billions in funding, either can’t or won’t build well-behaved crawlers. The incentives are broken. They have no reason to be good citizens, and website owners are left bearing the full cost.

To survive, we need a new strategy. Passive defense is no longer enough. We need to actively create a new set of incentives and disincentives. My proposed strategy has two parts: a carrot to encourage good behavior and a stick to punish the bad.

The Carrot: llms.txt — The Welcome Mat

The first part of the strategy is cooperation. The llms.txt standard is an open, consent-based protocol designed to help well-behaved bots. Instead of forcing a crawler to parse complex HTML, navigation, and ads, you provide a simple markdown file that points them directly to clean, LLM-friendly content.

Think of it as a welcome mat. You are politely showing ‘good’ bots the most efficient way to get the data they need without breaking the furniture.

  • For the Bot Operator: Their job becomes easier, cheaper, and more efficient.
  • For the Site Owner: Your server load is drastically reduced, as you’re serving simple, static content.

It’s a decentralized, ‘Bazaar’-style solution that empowers individual creators to cooperate with bots on their own terms.

The Stick: AI Labyrinth — The Trapdoor

The second part of the strategy is confrontation. For every bot that ignores your robots.txt file and your shiny new llms.txt welcome mat, you need a stick. Cloudflare’s AI Labyrinth is a perfect example. It’s a weapon of entrapment designed to punish misbehaving bots.

Think of it as a trapdoor. When a ‘bad’ bot is detected, it’s sent into a resource-wasting maze of fake, AI-generated content. The bot wastes its time, compute, and money processing useless data, while your origin server remains untouched.

  • For the Bot Operator: Their job becomes harder, more expensive, and less effective. A strong financial incentive is created to either fix their broken crawler or go elsewhere.
  • For the Site Owner: You actively punish trespassers, offload their traffic, and make your site a far less attractive target for abuse.

This is a centralized, ‘Cathedral’-style solution, where a powerful gatekeeper (Cloudflare) provides protection to the masses.

At a Glance: Two Sides of a Modern Defense

These two approaches are not competitors. They are complementary tools designed to handle two different classes of bots.

Featurellms.txtAI Labyrinth
PhilosophyCooperative (Consent)Adversarial (Entrapment)
MechanismProvides a clean, static, easy-to-parse data path.Serves fake, AI-generated content to waste bot resources.
Target BotsWell-behaved crawlers that follow the standard.Misbehaving crawlers that ignore rules.
ImplementationDecentralized: An open standard for any server.Centralized: A proprietary service (e.g., via Cloudflare).
Primary GoalReduce server load and provide clean data.Increase the cost and reduce the efficiency of bad scraping.
Impact on BotsMakes their job easier and cheaper.Makes their job harder and more expensive.

The Pragmatic Path Forward

This dual approach directly addresses the nuances of the problem. llms.txt is a diplomatic offer to the competent and well-intentioned. However, it does nothing to deter bots that are either malicious or simply incompetent.

AI Labyrinth is brutally effective against both. It doesn’t care about intent, only behavior. A poorly coded bot that gets stuck in a loop and a malicious bot that intentionally ignores rules are treated the same: they both fall through the trapdoor. This is the necessary stick for when the carrot is ignored.

This does, however, raise a philosophical question about the web’s future. As one commenter noted, ‘The Cathedral won. Full stop.’ By relying on a service like Cloudflare, we reinforce the web’s centralization. While llms.txt champions the decentralized ‘Bazaar,’ AI Labyrinth is a pragmatic admission that the Bazaar may need a Cathedral’s walls to protect it from being pillaged. In the face of an existential threat, pragmatism must often win out.

The Ideal Setup: A United Defense

The most robust strategy is not to choose one, but to use both.

  1. Lay the Welcome Mat: Implement llms.txt on your site. This is your good-faith offer of cooperation. It handles the ‘good’ bots, rewarding their citizenship and protecting your server.
  2. Set the Trapdoor: Enable a service like AI Labyrinth. This is your security system. It catches and neutralizes every scraper that ignores your clearly posted rules of engagement.

This combination creates a powerful and clear incentive structure: cooperate, and we will help you. Trespass, and you will pay the price.

The era of passive web administration is over. The old norms have been shattered by a gold rush for data. By combining the carrot of llms.txt with the stick of AI Labyrinth, creators and hobbyists can reclaim control, protect their resources, and build a more defensible web for the age of AI.