Something broke the social contract of the internet, and most site owners haven’t noticed yet.

For 30 years, web crawling worked on a simple handshake: bots index your content, search engines send you traffic, everyone wins. robots.txt was a gentleman’s agreement, and it held because the incentives were aligned. Google needed your content. You needed Google’s traffic. The deal made sense.

AI crawlers operate under no such bargain. They take your content to train models or power retrieval systems, and they send nothing back. No attribution. No link. No traffic. Just a line in your server logs and a slightly higher infrastructure bill.

This is the AI scraping war. And it is quietly reshaping how the open web works.

The Scale Nobody Talks About

The conversation around AI scraping tends to focus on the ethics. Should AI companies be allowed to train on public web data? That debate matters. But it misses a more immediate, more concrete problem: the raw volume of requests these crawlers generate is causing real operational damage.

In late 2025, several mid-size publishers reported that AI crawlers had surpassed human visitors in total bandwidth consumption. Not bot traffic in general. Just the AI crawlers. One technical blog documented GPTBot alone accounting for 35% of their total requests over a 30-day window.

This isn’t a rounding error. When your CDN bill goes up because OpenAI’s crawler is hitting every page on your site three times a week, that’s a real cost with no corresponding revenue. For a company with a large content library, we’re talking thousands of dollars a month in extra infrastructure spend that didn’t exist two years ago.

And the problem compounds. Unlike Googlebot, which crawls strategically based on PageRank signals and crawl budgets, many AI crawlers take the brute force approach. They want everything. Every page, every revision, as frequently as possible. They don’t care about your crawl-delay directive because most of them don’t support it.

robots.txt Is a Polite Fiction

When the AI scraping conversation first gained traction in 2023, the immediate response from site owners was to update robots.txt. Block GPTBot, block CCBot, block ClaudeBot, problem solved.

Except it wasn’t.

robots.txt has always been advisory. There’s no technical enforcement. A well-behaved crawler checks the file and respects it. A less scrupulous one simply ignores it. And in the AI scraping ecosystem, the incentive to ignore it is enormous. Training data is the moat. The more data you can feed your model, the better it performs. That creates a direct financial incentive to crawl aggressively and ask forgiveness later.

The known, branded crawlers from major AI companies mostly comply. OpenAI’s GPTBot, Anthropic’s ClaudeBot, and Google’s extended crawlers generally honor robots.txt directives. But they represent a fraction of the total AI crawling activity. The long tail of smaller AI companies, research groups, and outright data brokers often operates with far less regard for publisher preferences.

The result is a two-tier system. Blocking the big names is easy. Blocking the scrapers that don’t announce themselves is a fundamentally harder problem that robots.txt was never designed to solve.

The User Agent Shell Game

AI scrapers that want to avoid detection have a trivially simple option: lie about who they are.

Changing a user agent string takes one line of code. A scraper that identifies itself as GPTBot can be blocked instantly. A scraper that identifies itself as a regular Chrome browser on Windows 10 blends into normal traffic seamlessly. There is no verification mechanism built into HTTP that prevents this.

This is where the arms race gets interesting. The behavioral signatures of an AI crawler are subtly different from a real browser. Crawlers tend to request pages in systematic patterns. They rarely load images, CSS, or JavaScript. They don’t generate mouse movements or scroll events. Their TLS handshake fingerprints don’t match the browsers they claim to be.

But detection is getting harder, not easier. Modern scraping infrastructure runs on headless browsers like Playwright and Puppeteer that execute JavaScript, render pages, and produce TLS fingerprints nearly identical to real Chrome. Some scraping services even route through residential proxy networks, making IP-based detection unreliable.

We’re watching a classic security arms race. Every new detection technique gets a countermeasure within months. Every countermeasure spawns a more sophisticated detection method. The cycle accelerates, and the cost of keeping up falls entirely on site owners.

The Infrastructure Tax

Let’s talk numbers, because the cost structure of AI scraping is asymmetric in a way that should alarm anyone running a content-heavy website.

An AI scraper fetching your content costs you money in three ways:

1. Bandwidth and compute. Every request hits your servers or CDN. If your pages are server-rendered, each request also consumes CPU cycles. A crawler hitting 10,000 pages generates real load.

2. Cache pollution. Aggressive crawling can push real user content out of edge caches, degrading performance for actual visitors. When your CDN sees a flood of requests for deep archive pages, it allocates cache space to content nobody is actually reading.

3. Log noise and analytics contamination. Bot traffic that slips past your filters shows up in your analytics as visits with 100% bounce rates and zero engagement. This pollutes your data and makes it harder to understand what real users are doing on your site.

None of these costs show up on the AI company’s balance sheet. They externalize the expense of data acquisition onto every website they crawl. It’s a subsidy that site owners never agreed to provide.

What Actually Works (And What Doesn’t)

After watching this play out for two years, some patterns have emerged around what actually reduces unwanted AI crawling.

robots.txt alone: minimal impact. It stops the compliant crawlers, which are the ones causing the least damage. The aggressive scrapers either ignore it or never check.

Rate limiting: helpful but blunt. Throttling requests from suspicious IPs or user agents can reduce load, but aggressive rate limiting risks blocking legitimate users behind shared IPs or VPNs. It’s a balancing act with no perfect setting.

Behavioral analysis: the most effective approach. Identifying crawlers by how they behave rather than what they claim to be produces the best results. Real humans don’t request 500 pages in 60 seconds. Real browsers load CSS and execute JavaScript. Real sessions have mouse movements and scroll patterns. Behavioral detection catches scrapers regardless of their user agent or IP address.

JavaScript challenges: surprisingly effective. Many AI scrapers still don’t execute JavaScript. Serving a lightweight JS challenge before delivering content filters out a significant portion of scraping traffic. The tradeoff is a slight increase in page load time for real users.

Honeypot pages: high signal, low noise. Creating pages that are invisible to human visitors but discoverable by crawlers provides a clean detection signal. When a bot requests a page that no real user would ever find, you know exactly what you’re dealing with.

The common thread across these techniques is that they rely on behavior, not identity. You can’t trust what a bot tells you about itself. You can trust what it does.

The Bigger Question: What Happens to the Open Web?

Here’s what keeps me up at night about all of this. The AI scraping war has a second-order effect that nobody in the LLM space seems willing to address.

The open web has always run on a value exchange. People publish content freely because they get something back, whether that’s search traffic, ad revenue, reputation, or community engagement. That exchange is what made the web worth building on.

AI crawling disrupts this exchange in a fundamental way. Your content trains a model that answers questions without ever sending users to your site. The information gets extracted, the source gets nothing. Taken to its logical conclusion, this creates a massive free-rider problem. Why publish anything publicly if the primary beneficiary is an AI company’s training pipeline?

We’re already seeing the early signs. Some publishers have moved content behind paywalls specifically because of AI scraping. Others have started serving degraded content to suspected bots. A few have abandoned the open web entirely, moving to newsletters and gated communities where they control distribution.

This isn’t healthy. The open web got us here precisely because information flowed freely. Walling everything off makes the internet worse for everyone, including the AI companies that depend on fresh public content to keep their models current.

Where This Is Heading

The AI scraping war won’t be resolved by technology alone. robots.txt clearly isn’t sufficient, but neither is an endless arms race of detection and evasion. Some combination of legal frameworks, technical standards, and market pressure will eventually produce a new equilibrium.

A few things seem likely:

Authenticated crawling will become standard. Rather than relying on user agent strings, expect to see cryptographic verification of crawler identity. If a bot claims to be GPTBot, it should be able to prove it. Early versions of this already exist in proposals from the W3C and independent standards groups.

Revenue sharing will emerge for training data. The current arrangement where AI companies extract value without compensation is unstable. Licensing deals between AI companies and major publishers are already common. Standardized micropayment or licensing frameworks for smaller publishers will follow, though probably not as quickly as they should.

Behavioral detection becomes table stakes. Any serious bot management strategy in 2026 already includes behavioral analysis. Within a few years, it will be as basic and expected as having an SSL certificate. Sites that rely solely on user agent filtering will be the ones getting scraped hardest.

The legal landscape will catch up. The EU’s AI Act already includes provisions around training data provenance. Copyright cases are working through US courts. The legal uncertainty that currently enables aggressive scraping won’t last forever.

What You Should Do Right Now

If you run a website with content worth protecting, here’s the practical minimum:

  1. Audit your bot traffic. Look at your server logs and identify how much of your traffic comes from AI crawlers, both the ones announcing themselves and the ones pretending to be browsers.

  2. Deploy behavioral detection. User agent filtering is necessary but not sufficient. You need to identify crawlers by what they do, not what they say they are.

  3. Implement JavaScript challenges. Even a lightweight challenge filters out the majority of unsophisticated scrapers.

  4. Set up honeypots. Hidden links and invisible pages give you a clean signal when something is crawling your site without following normal navigation paths.

  5. Monitor your infrastructure costs. If your bandwidth usage has climbed steadily without a corresponding increase in real traffic, AI crawlers are probably the reason.

The open web is worth fighting for. But protecting it requires understanding the new threat model, and right now, that model is a fleet of AI crawlers consuming content at a pace the internet’s original social contract never anticipated.


Ready to see what’s actually crawling your site? WebDecoy’s behavioral detection identifies AI scrapers regardless of their user agent, and our honeypot system catches the ones trying to hide.

Want to see WebDecoy in action?

Get a personalized demo from our team.

Request Demo