Readiness scorecard

Is your site AI agent ready?

See how AI-friendly your public site is: one checklist, one score, clear next steps.

How it works

Three steps from URL to an actionable readiness report.

  1. 1

    Paste your URL

    We only use your public site address and a short checklist of common entry points—not your whole catalog.

  2. 2

    We check safely

    Tight limits on time and data per run, so you get a fast snapshot without handing us a wide crawl.

  3. 3

    Get a score and fixes

    Each check explains what it is, why it matters, and what to change.

What we check

Nine weighted signals. Same list you will see on every scan report.

  • robots.txt

    8 pts

    Standard rules file crawlers and AI agents read first.

    Discovery

  • sitemap.xml

    8 pts

    Machine-readable list of important URLs on your site.

    Discovery

  • Security headers

    6 pts

    Baseline HTTP security headers on your HTML homepage response.

    Policy & bots

  • Canonical & Open Graph

    5 pts

    Canonical URL, Open Graph tags, and robots meta for the homepage.

    Discovery

  • Structured data (JSON-LD)

    5 pts

    JSON-LD structured data for Organization, Product, or FAQ content.

    Discovery

  • Content signals

    6 pts

    Content-Signal lines in robots.txt declare AI usage preferences.

    Policy & bots

  • AI bot rules

    8 pts

    Structured Allow/Disallow coverage for major AI crawlers in robots.txt.

    Policy & bots

  • AI context files

    5 pts

    ai.txt plus machine-readable /.well-known/ai-context.json.

    Discovery

  • llms.txt

    10 pts

    Plain-text file aimed at LLMs: what your site is and how to use it.

    Discovery

  • Markdown negotiation

    12 pts

    Homepage returns markdown when the client asks for it.

    Content negotiation

  • MIME profile

    3 pts

    Expected vs actual Content-Type across probed endpoints.

    Diagnostics

  • Fetch performance

    3 pts

    Latency buckets for key fetches (fast / moderate / slow).

    Diagnostics

  • Agent skills index

    8 pts

    JSON index of agent skills your site exposes.

    Agent capabilities

  • MCP server card

    8 pts

    MCP server card JSON under /.well-known.

    Agent capabilities

  • API catalog

    4 pts

    Well-known API catalog with linkset metadata.

    Agent capabilities

  • Well-known discovery

    3 pts

    Optional well-known endpoints (security, OAuth, ai-context).

    Agent capabilities

Example report

Every run includes a ring score, pass counts, and cards with plain-English context.

Score72/ 100

On track

example.com

Sample · not a live scan

Passing

robots.txt, sitemap, llms.txt, …

Needs work

Markdown negotiation, MCP card, …

FAQ

What are AI agents?

AI agents are software systems that can take actions toward a goal with some autonomy—often by reading documentation, calling tools or APIs, or browsing the web. They range from chat assistants to crawlers and workflow bots. This site focuses on how clearly your public website signals rules and capabilities to those systems.

What are common kinds of AI agents?

Rough buckets include: conversational assistants (chat over your help content), crawlers and indexers (search and AI training pipelines), and tool-using agents (MCP, plugins, or APIs that fetch pages on your behalf). Different agents respect different signals—robots.txt, llms.txt, and structured metadata all help reduce ambiguity.

What is AI agent readiness?

It measures how well autonomous tools can discover your rules, preferred content formats (like markdown), and machine-readable capability files—without brittle full-site scraping.

How do I prepare my public website for AI agents?

Publish clear, machine-readable signals: a valid robots.txt (and optional Content-Signal lines), a sitemap, llms.txt if you want a curated LLM entry point, honest Content-Type headers, and well-known files such as MCP server cards or agent skill indexes where relevant. Fix broken redirects and avoid blocking helpful paths for legitimate automated clients.

What does an AI agent typically do when it visits my site?

It usually fetches a small set of URLs—homepage, robots.txt, sitemap, maybe llms.txt or /.well-known files—then decides what it may quote, summarize, or link to. It does not magically understand intent; it follows HTTP responses, headers, and the text you expose.

Is my website “agent ready”?

There is no universal certificate. Readiness means your public endpoints clearly state policy and capabilities and return consistent, honest responses. This scanner runs lightweight server checks against those endpoints and scores common gaps so you can prioritize fixes.

Why do AI or automation projects sometimes disappoint?

Common issues include unclear data or policies, brittle integrations, missing observability, and expectations that outpace what models can safely do. For websites specifically, vague robots rules, missing llms.txt, or misleading Content-Types make automated clients guess—and guesses are often wrong.

Why does llms.txt matter?

It is a simple, curated file models can read first so they route traffic respectfully and surface the URLs and policies you care about.

What are Content Signals?

Optional directives in robots.txt that declare preferences for how AI systems use your content for training, snippets, and summarization.

AI Agent Readiness Scanner