robots.txt
Text file placed at the root of a website to indicate to indexing robots which pages to explore or avoid.
What is robots.txt?
The robots.txt file is a web standard that allows site owners to communicate with indexing robots (crawlers). It indicates which parts of the site can or cannot be explored. In the context of GEO, it's crucial to configure robots.txt to allow AI crawlers (GPTBot, ClaudeBot, etc.) to access your content, unless you have legal or strategic reasons to block them.
How Qwairy Makes This Actionable
Qwairy analyzes your robots.txt file to verify AI crawler accessibility. Our crawlability analysis checks if GPTBot, ClaudeBot, Google-Extended, and other AI crawlers can access your content, identifying any blocking rules that might prevent AI visibility.
Frequently Asked Questions
Related Terms
AI Crawler
Indexing robot used by AI companies to collect data intended to train or feed their models.
llms.txt
Standardized file proposing a structured summary of a site's content to optimize its understanding by LLMs.
GPTBot
OpenAI's web crawler used to collect data intended to train GPT models.
ClaudeBot
Anthropic's indexing robot used to train and improve Claude models.
Crawlability
Ability of a website to be explored efficiently by indexing robots.