The New York Times has blocked OpenAI’s web crawler, meaning that OpenAI can’t use content from the publication to train its AI models. If you check the NYT’s robots.txt page, you can see that the NYT disallows GPTBot, the crawler that OpenAI introduced earlier this month. Based on the Internet Archive’s Wayback Machine, it appears NYT blocked the crawler as early as August 17th.
The New York Times blocks OpenAI’s web crawler
The NYT’s robot.txt page that controls how it appears to automated bots built to index the internet now specifically disallows OpenAI’s GPTBot.
The NYT’s robot.txt page that controls how it appears to automated bots built to index the internet now specifically disallows OpenAI’s GPTBot.


The change comes after the NYT updated its terms of service at the beginning of this month to prohibit the use of its content to train AI models. New York Times spokesperson Charlie Stadtlander spokesperson declined to comment. OpenAI didn’t immediately reply to a request for comment.
The NYT is also considering legal action against OpenAI for intellectual property rights violations, NPR reported last week. If it did sue, the Times would be joining others like Sarah Silverman and two other authors who sued the company in July over its use of Books3, a dataset used to train ChatGPT that may have thousands of copyrighted works, as well as Matthew Butterick, a programmer and lawyer who alleges the company’s data scraping practices amount to software piracy.
Update August 21st, 7:55PM ET: The New York Times declined to comment.
Most Popular
- Sony’s PlayStation 5 is $200 off for the first time since December
- Anthropic’s most dangerous AI model just fell into the wrong hands
- Elon Musk admits that millions of Tesla vehicles won’t get unsupervised FSD
- You’re about to feel the AI money squeeze
- Microsoft launches ‘vibe working’ in Word, Excel, and PowerPoint












