Understanding Your Score
Crawlability
Whether AI systems can discover, access, and index your website's content.
Overview
What Crawlability Measures
Crawlability is the foundation of your AURA score. It evaluates whether AI crawlers can physically reach and index your pages. If AI systems can't crawl your site, nothing else matters — they can't recommend content they can't see.
This dimension runs on every audit, including free ones, and checks a series of technical signals that determine your site's accessibility to AI bots.
What We Check
Technical Checks
AURA runs the following technical checks to evaluate whether AI systems can access your site.
- Robots.txt Configuration — Does your robots.txt file allow AI crawlers like GPTBot, ClaudeBot, and Google-Extended? Sites that block these bots are invisible to the AI systems that could recommend them.
- AI Bot Directives — Beyond robots.txt, we check whether your site explicitly allows or blocks specific AI training and search bots. This includes directives for GPTBot, ClaudeBot, Anthropic, Google-Extended, and others.
- Meta Robots Tags — Are your pages using noindex or nofollow meta tags that prevent AI crawlers from indexing or following links? These tags can accidentally hide important content.
- XML Sitemap — Does your site have a valid, discoverable XML sitemap? Sitemaps help AI crawlers find all your important pages efficiently rather than relying on link discovery alone.
- Schema.org Presence — Does your site include structured data markup? Even basic Schema.org (Organization, WebSite) helps AI systems understand your site's identity.
- Server Response — How quickly does your server respond? Slow response times can cause AI crawlers to time out or deprioritize your site.
- Page Structure — Does your homepage have a proper H1 heading and a clear page title? These help AI systems understand what your site is about at first glance.
- Content Rendering — Can AI crawlers read your content from the raw HTML, or does it require JavaScript execution? Content that only appears after JS rendering may be invisible to some AI crawlers.
- llms.txt — Does your site include an llms.txt file? This emerging standard helps AI systems understand your site's purpose, key pages, and how to interact with your content.
Common Problems
Common Issues
These are the most frequent reasons for a low Crawlability score.
- AI bots blocked in robots.txt — The most common issue. Many sites inherited robots.txt rules that block GPTBot or ClaudeBot without realizing it.
- WAF or bot protection blocking crawlers — Cloudflare, AWS WAF, and other security tools may block our crawler even when robots.txt allows AI bots. The report distinguishes between intentional blocking (robots.txt) and infrastructure blocking (WAF/firewall).
- Missing XML sitemap — Without a sitemap, AI crawlers rely on link discovery, which may miss important pages buried deep in your site structure.
- Slow server response — Response times over 3 seconds can cause crawler timeouts and reduce how frequently AI systems revisit your site.
Taking Action
How to Improve
Follow these steps to improve your Crawlability score, ordered by typical impact.
- Audit your robots.txt — Ensure GPTBot, ClaudeBot, Google-Extended, and Anthropic-AI are not disallowed. If in doubt, remove blanket User-agent: * Disallow rules and whitelist specific bots instead.
- Check your WAF settings — If your report shows crawler blocking despite correct robots.txt, your security infrastructure is likely the cause. Whitelist known AI bot user agents in your WAF/CDN settings.
- Submit an XML sitemap — Create a sitemap at /sitemap.xml listing all important pages. Submit it to Google Search Console — AI systems often discover sitemaps through the same channels.
- Add an llms.txt file — Create a /llms.txt file describing your site's purpose, key pages, and content organization. This is a simple text file that takes minutes to create.
- Ensure server-side rendering — If your site uses a JavaScript framework (React, Vue, Angular), configure server-side rendering so the HTML contains your content without requiring JS execution.
FAQ
Frequently Asked Questions
-
Why does my Crawlability score differ from my SEO crawl score?Traditional SEO tools check whether Googlebot can crawl your site. AURA's Crawlability specifically checks AI-related bots like GPTBot, ClaudeBot, and Google-Extended, which may have different access rules in your robots.txt.
-
My robots.txt allows all bots but I still got a low score. Why?Your security infrastructure (Cloudflare, WAF, or cookie walls) may be blocking crawlers at the network level before they reach your robots.txt. The report's "crawler blocked" indicator shows when this happens.
-
Is Crawlability included in the free plan?Yes. Crawlability and Authority are the two dimensions that run on every audit, including free Starter plan audits.