Many sites are struggling to keep up with ai data scrapers – Breaking News & Latest Updates 2026
Skip to main content
E
External Link
Many sites are struggling to keep up with AI data scrapers.

A report from 404 Media reveals how AI companies like Anthropic are bypassing a website’s robots.txt file by deploying new web crawlers with different names. This makes it more difficult for websites to block crawlers, as they constantly need to update their files to include the new bots:

Anthropic’s current and active crawler is called “CLAUDEBOT.” Neither Reuters nor Condé Nast, for example, blocks CLAUDEBOT. This means that these websites—and hundreds of others who have copy pasted old blocker lists—are not actually blocking Anthropic.

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.
Comments
Loading comments
Getting the conversation ready...