For AI agents: a documentation index is available at /llms.txt — markdown versions of all pages are available by appending .md to any URL path.
Skip to content

Why Agent-Friendliness Matters

AI coding agents are a new and fast-growing audience for your documentation. Millions of developers use tools like Claude Code, Cursor, GitHub Copilot, Windsurf, Codex, and Gemini CLI daily. Those tools read your docs on the developers' behalf. When a developer asks an agent to integrate your API, debug an issue, or follow a migration guide, the agent fetches and reads your documentation pages in real time.

This is a distinct audience from LLM training crawlers or AI answer engines. Coding agents need to read specific pages, right now, and produce working code from what they find. If they can't, the developer gets a bad result and your product takes the hit.

How agents fail on documentation

Agents don't read docs the way humans do. They can't scroll past a navigation bar, mentally filter out boilerplate, or click through a login flow. When your docs aren't set up for agents, here's what actually happens:

  • Client-side rendering: Agents fetch the page and get an empty JavaScript shell with no content. This is the most common complete failure mode, and it affects every docs site built on a single-page app framework without server-side rendering.
  • Truncation: Agents have context limits. A page with 200K characters of inline CSS, JavaScript, and navigation markup before the actual content starts means the agent may never reach the documentation text. Even if it does, the content may be cut off. Different agent platforms implement different solutions for this problem, and each platform has different failure modes.
  • No discovery path: Agents don't guess that your site supports .md URLs. Only some of them try Accept: text/markdown headers to get clean markdown content. Without an llms.txt or a directive on pages, agents have no way to find a markdown content path even if one exists.
  • Broken navigation: Cross-host redirects, JavaScript redirects, and soft 404s (pages that return HTTP 200 with an error message) all break agent workflows. Agents trust HTTP status codes and follow standard redirects. Anything else is a dead end.
  • Auth walls: If your docs require authentication, agents can't log in. The entire site is invisible to them.

When your docs are agent-friendly, none of this is visible. The agent fetches clean markdown, navigates via llms.txt, and produces working integration code on the first try. The developer never thinks about your docs at all, which is the goal.

The invisible problem

When a documentation page has a typo, someone files a bug. When an agent can't read a documentation page, nobody files anything. The developer gets a bad answer from the agent, blames the agent (or your product), and moves on. There's no feedback loop.

This means agent-accessibility problems can persist for months without anyone on your team knowing. Your docs could be perfectly written and completely invisible to the tools that millions of developers rely on.

The Agent Score gives you visibility into this blind spot. It tests your site the way agents actually access it and tells you what they see (or don't see).

Business impact

When agents can't use your documentation effectively, the result impacts your bottom line:

Increased support costs

Agents that can't read your docs produce incorrect code. Developers hit errors, open support tickets, and file bug reports that trace back to bad agent output rather than actual bugs in your product. Your support team spends time debugging problems that good documentation would have prevented.

Customer churn and revenue impact

Developers increasingly choose tools based on how well they work with AI agents. If integrating your product is painful because agents can't read your docs, developers will choose a competitor with docs that agents can actually use. This is a new competitive angle that most companies aren't tracking yet.

Wasted documentation investment

You've invested in writing, maintaining, and hosting documentation. If agents can't access it, that investment is partially wasted for a growing segment of your audience. The content exists; it's just locked behind technical barriers that have real fixes.

Reputational damage

When an agent writes bad code using your product, the developer's first instinct is often "this API is hard to use" or "the docs are bad," not "the agent couldn't read the docs." Poor agent-friendliness degrades your product's reputation even when the product and its documentation is excellent.

The effort has an outsized impact

The highest-impact fixes for agent-friendliness are typically configuration changes, not documentation rewrites:

FixImpactTypical effort
Add an llms.txt fileCritical (10 pts)A few hours for initial creation; minutes to maintain if generated at build time
Enable server-side renderingCritical (10 pts)A configuration flag on many docs platforms
Serve markdown at .md URLsHigh (7 pts)Platform-dependent; some support it natively, others need a server config change
Add an llms.txt directive to pagesHigh (7 pts)A one-line addition to your page template
Fix broken llms.txt linksHigh (7 pts)Find and fix broken links with afdocs check --verbose
Point llms.txt links to markdownHigh (7 pts)Find-and-replace in your llms.txt

Your documentation site may move from an F to a B or higher in a single sprint. The Improve Your Score guide walks through these fixes in priority order.

A new competitive angle

When a developer asks an agent to "help me integrate Product A's API" and "help me integrate Product B's API", the agent will produce better code for whichever product has more agent-friendly documentation. The developer may not even realize why one integration went smoothly and the other didn't.

Companies that optimize their docs for agents early will have a compounding advantage as agent usage grows. Companies that don't will gradually lose developer mindshare to competitors whose documentation agents can actually read.

This isn't hypothetical. Agent usage among developers has grown rapidly through 2025 and into 2026, and the agents that read documentation (Claude Code, Cursor, Copilot, etc.) are among the most widely adopted developer tools in history. The documentation you ship today is already being read by agents, whether or not you've optimized for it. In early April 2026, documentation platform provider Mintlify shared a statistic:

Claude Code and Cursor account for nearly half of all views to documentation sites.

Making the case internally

If you need to justify prioritizing agent-friendliness to leadership, here are the key points:

  1. It's measurable. Run npx afdocs check https://your-docs-site.com --format scorecard and you have a concrete before-and-after metric. See Quick Start for setup.
  2. The fixes may be small. Most high-impact improvements are configuration changes, not content changes. See the effort table above. The effort size varies depending on your documentation platform.
  3. The problem is invisible without measurement. You won't get bug reports about this. The Agent Score is the feedback loop.
  4. It's competitive. If your competitors' docs are agent-friendly and yours aren't, agents will produce better code for their products.
  5. It protects your docs investment. You've already written the content. Agent-friendliness makes sure it reaches this growing audience.