If you've searched for your brand or industry on ChatGPT, Perplexity, or Claude and your website never appears in the answers — you're not alone. I've scanned 1,000+ websites and found that most sites fail at least one critical AI search readiness check.
Here's the diagnostic checklist I use:
# Check your robots.txt for AI crawler blocks
curl -s https://yoursite.com/robots.txt | grep -i -E "gptbot|claudebot|perplexitybot|google-extended|bytespider"If you see Disallow: / for any of these, that's your problem. ~30% of websites block AI crawlers without knowing it.
# Add to your robots.txt
User-agent: GPTBot
Allow: /
User-agent: ClaudeBot
Allow: /
User-agent: PerplexityBot
Allow: /
User-agent: Google-Extended
Allow: /
# See what AI crawlers actually see (no JavaScript execution)
curl -s https://yoursite.com | head -100If your main content isn't in the raw HTML output, AI crawlers probably see a blank page. You need server-side rendering (SSR).
curl -s -o /dev/null -w "%{http_code}" https://yoursite.com/llms.txtIf this returns 404, you're missing an easy win. llms.txt tells AI systems what your site is about and where to find key content.
# Your Site Name
> One sentence describing what your site does.
## Key Pages
- [Product](https://yoursite.com/product): What you offer
- [Pricing](https://yoursite.com/pricing): Plans and pricing
- [Docs](https://yoursite.com/docs): Documentation and guides
- [Blog](https://yoursite.com/blog): Latest articles and insights
# Check for JSON-LD structured data
curl -s https://yoursite.com | grep -c "application/ld+json"If the count is 0, you have no structured data. AI search engines use this to understand your content type (article, FAQ, how-to, product, etc.).
Instead of checking each signal manually, you can run a free scan at GEOScore AI — it checks all 9 AI search readiness signals and gives you a 0-100 score with specific fix recommendations. Takes ~60 seconds, no signup.
| Site Type | Typical Score | Main Issues |
|---|---|---|
| WordPress blogs | 40-55 | Missing llms.txt, weak structured data |
| React/Next.js SPAs | 20-40 | Client-side rendering, no SSR |
| Shopify stores | 35-50 | Limited robots.txt control, thin content |
| Static sites (Hugo, Jekyll) | 50-70 | Usually good! Fast, server-rendered, clean HTML |
| Enterprise sites | 25-45 | Complex JS, aggressive security headers |
Created by William Wang, founder of GEOScore AI