/admin/, /private/.Each major AI company operates its own crawler that visits websites to collect content. This content is used for two main purposes: training AI models and powering real-time retrieval in AI search features. Here is who they are:
User-agent: * with broad Disallow rules to block anything that isn't a known search engine. This was reasonable practice before AI crawlers existed but now it silently blocks GPTBot, ClaudeBot, and every other AI crawler that wasn't on the original allowlist.Anthropic-AI without realizing that ClaudeBot is a separate crawler or vice versa. The result is unintended partial blocking.Run a free LLM visibility analysis the Analyzer checks your robots.txt live and tells you exactly which AI crawlers are blocked.