logo
agent-readyai-agentsfounder-growthllms-txtmcp-serveragent-native-gtm

Agent-Ready Site Audit: The 2026 Scorecard Founders Run Before Launch

AI agents pre-filter your site before buyers ever see it. The median founder site scores 22/100. Here is the 100-point 2026 scorecard to fix the gap.

ForkOff Team11 min read
Agent-ready site audit 2026 cover

The Agent-Ready Scorecard in one scroll

AI agents now pre-filter founder sites before human buyers ever see them. The median founder site in FORKOFF Q1 2026 audits scores 22 out of 100 on agent-readiness. The 100-point scorecard is 5 layers: Discovery (llms.txt + sitemap), Trust (schema + .well-known), Action (MCP + WebMCP), Negotiation (markdown middleware), Retrieval (citation-friendly H-tree). Level 5 = 100/100. This post is the full rubric with FORKOFF first-party data.

The YC founder whose site the agent refused to cite

In Q1 2026 a seed-stage YC founder came to us after a Perplexity citation run. They had spent nine months on brand, a fresh Framer site, a decent SEO pass, and a founder-voice X account with 4,200 followers. When a prospect pasted their domain into Claude and asked for a product summary, the agent returned a generic paragraph sourced from the homepage meta description and refused to cite any specific claim. The prospect moved on to a competitor whose site answered cleanly and with structured citations. The founder lost the meeting they did not know they were in.

This is the single biggest distribution shift of 2026. Your site is no longer read first by a human. It is read first by an agent, and the agent decides whether the human ever sees it. The discipline that fixes the gap is agent-readiness, and unlike traditional SEO it is a rubric with exact technical requirements you can install in a sprint.

FORKOFF audited 47 founder sites between January and April 2026 against the public scoring rubric on isitagentready.com. The median score was 22 out of 100. The bottom quartile scored 8. The top quartile scored 71. Forkoff.xyz and clips.forkoff.xyz both score 100 out of 100 at Level 5 agent-native, which is what the rubric below is built to reach.

38 percent, 22 of 100, and the isitagentready rubric that made this concrete

Three numbers anchor the 2026 agent-readiness thesis. First, roughly 38 percent of B2B buyer research now begins inside an AI agent such as Claude, ChatGPT, or Perplexity rather than on Google, a jump from about 12 percent in 2024 across major buyer surveys. Second, FORKOFF Q1 2026 founder-site audits across 47 seed and Series A companies found a median score of 22 out of 100 on the isitagentready.com public rubric, with the gap concentrated in the Action and Negotiation layers where MCP servers and markdown content-negotiation live. Third, the rubric itself matters because it formalized what had been tacit. Before Cloudflare published a public scanner, no founder could tell you whether their site was agent-ready or not. Now they can, and the teams treating the 100-point target as a launch gate are compounding faster than peers still running the 2022 playbook.

Source: FORKOFF founder-site audits Q1 2026 (n=47); isitagentready.com public rubric; a16z 2026 State of AI buyer survey aggregates

Why agent-readiness is not SEO and not schema alone

The common objection is that agent-readiness is just SEO with more tags. It is not. Traditional SEO optimizes a page for a ranking function that shows a user a link. Schema markup optimizes a page for a snippet that surfaces above that link. Agent-readiness optimizes a domain for a conversational agent that is deciding whether to cite your site at all, and if so, what to quote, what to call into, and what to recommend to the human on the other side of the chat.

That decision lives inside five discrete layers. Each layer has a different technical surface. Each layer is scored independently. A site can have flawless schema and still score zero on Action because it never shipped an MCP server. A site can have a rich MCP endpoint and still score zero on Discovery because it never published llms.txt. The rubric compounds because every agent in 2026 checks all five layers and downgrades the domain if any single layer is missing.

The five layers below are the FORKOFF internal scorecard. Each has a point value, a list of the specific files and endpoints the layer scores, and a reference implementation. None of it requires a rebuild. All of it can be installed on an existing Next.js, Astro, WordPress, or Webflow site inside one focused week.

Layer 1 of 5: Discovery (20 points)

Discovery is the agent's first pass on your domain. Before the agent reads a single product claim, it fetches three files and decides whether your site is a serious information source or a black box.

llms.txt (10 points). The llms.txt spec, published in late 2024 and now supported by every major agent, is the agent-native analog of robots.txt. It lives at the root of your domain, enumerates the canonical URLs an agent should prioritize for training and citation, and gives a summary the agent uses as context when summarizing your company. FORKOFF ships llms.txt on every client domain in the first week of an engagement. Missing it is the single most common cause of a zero-score site.

sitemap.xml tuned for AI crawlers (6 points). The sitemap the agent reads is the same sitemap Google reads, but it must include lastmod, changefreq, and priority on every URL, and must be referenced explicitly in both robots.txt and the Link header. Without the lastmod attribute, agents conservatively refuse to re-index the page more than once per quarter.

robots.txt posture (4 points). The robots.txt file should explicitly allow known agent user-agents (GPTBot, ClaudeBot, PerplexityBot, Google-Extended) rather than leaving them to the default. Implicit allow is scored as 2 points on the rubric; explicit allow with crawl-delay tuning is scored as the full 4.

Layer 2 of 5: Trust (25 points)

The Trust layer is where most teams score surprisingly well on the surface and poorly when you open the hood. The 25 points split across three surfaces.

Schema.org markup (12 points). Not one schema type, five. Article schema on every blog post, FAQ schema on every post with questions, Breadcrumb schema site-wide, SoftwareApplication schema on the product pages, and Organization schema on the homepage. Each schema must validate cleanly against the 2026 Google Rich Results test and must include the id, sameAs, and url fields an agent uses to disambiguate your entity from a competitor with a similar name.

.well-known manifests (9 points). This is where the 2026 sites diverge hard from the 2022 ones. A Level 5 site serves seven files under /.well-known: the MCP Server Card, the Agent Skills Card, the A2A Agent Card, OAuth 2.0 discovery, OAuth 2.0 Protected Resource metadata, a Content Signals manifest, and an API Catalog. Each is a small JSON document that takes less than an hour to author and less than a day to validate. Missing these is the single largest point gap across FORKOFF's audit set.

JSON-LD instead of Microdata (4 points). Agents parse JSON-LD an order of magnitude faster than Microdata and with higher fidelity. Sites still using Microdata or RDFa in 2026 are scored as partially compliant and lose the top band on Trust.

Layer 3 of 5: Action (25 points)

If Discovery and Trust make your site readable, Action makes it callable. This is the layer where the founder site stops being a brochure and becomes an operable surface the agent can invoke on behalf of its user. Three sub-scores.

MCP server endpoint (15 points). Model Context Protocol is the 2026 contract by which an agent requests a structured action from a remote service. A FORKOFF-grade implementation exposes an MCP server at /mcp that enumerates at minimum four operations: get_company_summary, list_resources, get_contact_intake, and start_audit. Each operation is defined by a JSON schema the agent reads at runtime. Teams without an MCP endpoint score zero on this band, no partial credit.

WebMCP card (6 points). The WebMCP card is the agent-side discovery manifest for your MCP server, published at /.well-known/webmcp. It is how an agent browsing your domain finds the MCP endpoint without being told in advance. Agents that land on a page with a WebMCP card auto-offer to call into the server.

Agent Skills manifest (4 points). Adjacent to the MCP card, the Agent Skills manifest describes named workflows the agent can compose from the raw MCP operations. For FORKOFF that includes run_agent_readiness_audit and generate_growth_plan. The agent reads the manifest, understands the high-level verbs, and treats your site as a tool.

Hridoy Rehman

Hridoy Rehman

@hridoyreh

1. Go to Google Search Console. 2. Select 3 months of data. 3. Tap Pages > open any URL. 4. Download the CSV file. 5. Open Claude and upload the file. Prompt: Analyze the file and identify which keywords are not used on [that page/content URL] but are still getting a lot of impressions. 6. Note these top keywords. 7. Use them on your existing content. You will see the traffic growth...

Apr 24, 2026, 9:40 AM

383
7

Layer 4 of 5: Negotiation (15 points)

Negotiation is the quietest of the five layers and the one that produces the biggest retention gap. The 15 points break across two surfaces.

Markdown content-negotiation middleware (10 points). Every page on the site must respond to an Accept: text/markdown header with a clean markdown body rather than HTML. Agents request markdown because it parses faster, tokenizes smaller, and quotes cleaner in a chat. A domain that only serves HTML forces the agent to do lossy HTML-to-markdown conversion on the fly and scores a flat zero here. The fix is a 40-line middleware on the edge that detects the Accept header, calls a markdownify pass, and caches the result.

Link headers for AI crawlers (5 points). Every HTTP response should include Link headers advertising the sitemap, the llms.txt, and the MCP Server Card. This is cheap to install and almost never done. Agents use Link headers to find adjacent resources without walking the DOM. Adding them on a Cloudflare Worker is a 5-minute change that recovers the full 5 points.

Layer 5 of 5: Retrieval (15 points)

Retrieval is how easily an agent can quote a specific claim from your content with confidence. Three sub-scores.

Citation-friendly H-tree hierarchy (7 points). Every H2 and H3 needs an id attribute so the agent can anchor a citation to a stable URL fragment. Sites that render all headings as divs, or that use auto-generated ids that change on every deploy, score zero here. A stable slug derived from the heading text is the safe pattern.

FAQ schema on every long post (5 points). FAQ schema is the most overlooked high-leverage markup in 2026. Every long-form post should have a blog.faq block with at least five QA pairs, each answer in the 40-60-word range that Perplexity and ChatGPT quote cleanly.

Anchor URLs preserved on redeploy (3 points). Anchor fragments must survive site redeploys. The agent's citation is only trustworthy if the URL it quotes keeps resolving six months from now. Generating deterministic slugs and freezing them in a migration map is the fix.

The 5-layer scorecard at a glance

LayerPointsCore componentsFailure mode
Discovery20llms.txt, sitemap.xml, robots.txtAgent cannot find canonical URLs; defaults to homepage meta only
Trust25schema.org suite, .well-known manifests, JSON-LDAgent cannot disambiguate entity; conflates you with a competitor
Action25MCP server, WebMCP card, Agent SkillsAgent cannot call into your surface; you remain a brochure
Negotiation15markdown middleware, Link headersAgent parses lossy HTML; citations degrade
Retrieval15H-tree anchors, FAQ schema, stable slugsAgent cannot cite a specific claim; falls back to vague paraphrase

Rubric mirrors the public isitagentready.com scoring. Point values are normalized to 100 across all 5 layers.

Get the agent-readiness audit for your site free

Send your domain and FORKOFF will run the full 100-point scorecard, flag the three highest-leverage missing layers, and return a one-week install plan. No retainer conversation, no sales call. Just the scorecard and the fix list.

How we install the scorecard with founder teams

Every FORKOFF agent-readiness engagement starts with a scan against isitagentready.com and a manual FORKOFF audit that extends the public rubric with four internal checks the scanner misses (citation-survivability, MCP schema quality, WebMCP freshness, FAQ schema density). We score the site, name the three highest-leverage layers, and ship a week-one install plan.

Week one closes Discovery (llms.txt, sitemap tuning, robots posture). Week two closes Trust (schema suite, .well-known manifests). Week three closes Action (MCP endpoint, WebMCP card, Agent Skills manifest). Week four closes Negotiation and Retrieval together, because the markdown middleware and the citation anchors ship from the same edge worker. By end of month one, a site that started at 22 typically lands at 88 to 94, and the remaining 6 to 12 points are style-score bonuses earned in month two.

For the operator context behind this motion, two related FORKOFF reads: the Agent-Native GTM Founder Stack covers the broader stack an agent-native team runs, and the AI Marketing Verification audit covers the trust half that compounds after the site is agent-ready. For a distribution surface that compounds inside an agent-native workflow, see the AI DevRel Playbook. For AI-native outreach that pairs with an agent-ready site, see the Reddit Intent Engine. For the content-metric side of an agent-native motion, see the Qualified Views metric, and for the fund-backed variant of the same stack, see the VC Portfolio GTM Playbook. The full Founder Growth hub has the rest.

The 4 mistakes that make an agent-ready install fail

Across 47 FORKOFF audits in Q1 2026, four mistakes showed up repeatedly when a site believed it was agent-ready and was not.

  1. Shipping schema without validating it against 2026 agent parsers. Validators that pass Google Rich Results do not always pass Claude or ChatGPT's internal parsers. Run every schema through all three and fix the parser-specific warnings.
  2. Publishing llms.txt once and letting it rot. llms.txt is a living document. Add every new canonical page to it within 48 hours of publish, or agents will refuse to cite pages they cannot find in the index.
  3. Skipping the MCP endpoint because the product is early. Even a pre-launch product can ship an MCP server with four operations. An agent that finds a callable endpoint trusts the domain categorically more than one that finds a contact form.
  4. Treating the scanner score as the goal. The scanner is a proxy. The goal is that Claude, ChatGPT, and Perplexity cite you cleanly in a real buyer conversation. Test that with five buyer-style prompts a month and iterate.

HackerNews

View thread →

Hacker News thread that surfaced the agent-readiness rubric in 2026-04. 113 points, 178 comments, a dense thread on which specific sites pass and fail the scanner and why.

We scored 19 at the first scan and 94 four weeks later. Nothing about the product changed. What changed was that Claude and Perplexity started citing us cleanly, buyer meetings started quoting our exact wording, and we stopped losing the deals we did not know we were in. Agent-readiness is the most under-priced sprint we ran this year.

Growth lead, Seed-stage AI startup, 12-person team (FORKOFF founder-site audit debrief 2026)

The Bottom Line

AI agents pre-filter your site before a buyer ever sees it. The median founder site scores 22 out of 100 on the 2026 agent-readiness rubric and loses deals it never knew it was in. The fix is a rubric, not a rebuild. Five layers, one hundred points, four weeks of focused install work to land between 88 and 94, and a monthly re-scan to hold the score.

Most teams will find three layers they should already be running and are not. The point is to score the site, install the gaps, and test against real buyer prompts, not to theorize about agent behavior for another quarter.

If you want the FORKOFF audit run for you, that is what we do.

Ready to install the 2026 agent-ready site stack?

FORKOFF agent-readiness engagements move a site from a typical 22 out of 100 start to 88-94 inside four weeks: Discovery week one, Trust week two, Action week three, Negotiation and Retrieval week four. Book the free audit and see the exact 3-layer plan for your site.

Frequently Asked Questions

An agent-ready site is a domain engineered so AI agents such as Claude, ChatGPT, and Perplexity can discover, trust, call into, negotiate with, and cite it cleanly. The FORKOFF rubric scores agent-readiness across five layers on a 100-point scale. A Level 5 site ships llms.txt, schema.org, .well-known manifests, an MCP server, markdown middleware, and citation-friendly anchors. The median founder site scores 22 out of 100 in FORKOFF Q1 2026 audits.

A focused FORKOFF sprint lands a founder site between 88 and 94 out of 100 in four weeks. Week one closes the Discovery layer with llms.txt, sitemap tuning, and robots posture. Week two closes Trust with the schema suite and .well-known manifests. Week three ships the MCP server and WebMCP card. Week four ships markdown middleware and citation anchors. Month two picks up the final style bonuses.

Run five buyer-style prompts a month against Claude, ChatGPT, and Perplexity using your company name, your core feature, your category, your pricing, and your audience. Record whether the agent cites your domain, quotes your wording, and recommends you versus a competitor. Re-score monthly. The metric that matters is citation share, not scanner score. FORKOFF re-scans every client site monthly and tracks citation share against the starting baseline.

SEO optimizes a page for a ranking function that returns a link to a human. Agent-readiness optimizes a domain for a conversational agent deciding whether to cite your site at all and what to quote. Agent-readiness covers five layers SEO never touches: MCP endpoints, WebMCP cards, markdown content-negotiation, Agent Skills manifests, and .well-known agent manifests. A site can rank well on Google and still score under 30 on agent-readiness.

Yes, even pre-launch products ship an MCP endpoint with four operations: get_company_summary, list_resources, get_contact_intake, and start_audit. Agents score a callable domain categorically higher than a domain that only exposes a contact form. The four-operation stub takes roughly two engineer-days to build on a Cloudflare Worker and returns 15 full points on the Action layer of the scorecard, unlocking downstream Trust and Retrieval scoring.