Back to graph

Topic analysis

Agent Reading Test

A benchmark that tests how well AI coding agents can read web content. Point your agent at the test, get a score, compare across platforms. AI coding agents (Claude Code, Cursor, GitHub Copilot, and others) read documentation websites as part of their workflows. But most agents hit silent failure modes: content gets truncated, CSS buries the real text, client-side rendering delivers empty shells, and tabbed content serializes into walls of text where only the first variant is visible. This benchmark surfaces those failure modes. Each test page is designed around a specific problem documented in the Agent-Friendly Documentation Spec . The pages embed canary tokens at strategic positions. But instead of asking agents to hunt for tokens (which games relevance filters), the test gives the agent realistic documentation tasks. Only after the agent completes all tasks does it learn about the canary tokens and report which ones it encountered. You paste the results into a scoring form. 150K-char page with canary tokens at 10K, 40K, 75K, 100K, and 130K. Maps exactly where your agent's truncation limit kicks in. 80K of inline CSS before the real content. Tests whether agents distinguish CSS noise from documentation. Client-side rendered page. Content only appears after JavaScript executes. Most agents see an empty shell. 8 language variants in tabs. Canary tokens in tabs 1, 4, and 8. Tests how far into serialized tab content the agent reads. Returns HTTP 200 with a "page not found" message. Tests whether the agent recognizes it as an error page. Markdown with an unclosed code fence. Everything after it becomes "code." Tests markdown parsing awareness. Different canary tokens in HTML vs. markdown versions. Tests whether your agent requests the better format. 301 redirect to a different hostname. Most agents won't follow it (security measure). The canary is on the other side. Three cloud platforms, identical "Step 1/2/3" headers. Tests whether agents can determine which section is which. Real content buried after 50% navigation chrome. Tests whether agents read past the sidebar serialization. The test has a maximum score of 20 points . Each canary token found earns 1 point, and correct answers to qualitative questions earn 1 point each. The answer key has the full breakdown. A perfect score is unlikely for any current agent. The tests are calibrated so that each failure mode will realistically affect at least some agents. A typical score range for current agents is probably 14-18 out of 20, depending on the platform's web fetch pipeline. Agent Reading Test is a companion project to the Agent-Friendly Documentation Spec , which defines 22 checks across 8 categories evaluating how well documentation sites serve AI agent consumers. The spec is grounded in empirical observation of real agent workflows. This benchmark flips the perspective: instead of testing the documentation site, it tests the agent. The same failure modes apply, but here we're measuring which agents handle them gracefully and which don't. Source code: github.com/agent-ecosystem/agent-reading-test Created by Dachary Carey • Licensed under CC BY 4.0

Heat score

1

Sources

1

Platforms

1

Relations

8
First seen
Apr 7, 2026, 2:56 AM
Last updated
Apr 7, 2026, 4:00 AM

Why this topic matters

Agent Reading Test is currently shaped by signals from 1 source platforms. This page organizes AI analysis summaries, 1 timeline events, and 8 relationship edges so search engines and AI systems can understand the topic's factual basis and propagation arc.

News

Keywords

10 tags
benchmarktestshowwellcodingagentscanreadwebcontent

Source evidence

1 evidence items

Timeline

Agent Reading Test

Apr 7, 2026, 2:56 AM

Related topics

Research-Driven Agents: When an agent reads before it codes

AI coding agentcode optimizationCPU inferencellama.cppperformance improvementresearch phasecloud VMsflash attention
Relation score 0.90Open topic

Moving from WordPress to Jekyll (and static site generators in general)

static site generatormigrationCMSSEOdeveloper toolscontent optimizationweb development
Relation score 0.80Open topic

Show HN: Hippo, biologically inspired memory for AI agents

biologicallyinspiredmemoryagentssecretgoodisnrememberingmore
Relation score 0.75Open topic

The cult of vibe coding is insane

cultvibecodinginsanehadleaksourcecodepeoplehave
Relation score 0.70Open topic

Show HN: Modo – I built an open-source alternative to Kiro, Cursor, and Windsurf

builtopensourcealternativeplanscodeswhatadds
Relation score 0.60Open topic

Show HN: Modo – I built an open-source alternative to Kiro, Cursor, and Windsurf

builtopensourcealternativeplanscodeswhatadds
Relation score 0.80Open topic

Show HN: Hippo, biologically inspired memory for AI agents

biologicallyinspiredmemoryagentssecretgoodisnrememberingmore
Relation score 0.80Open topic

The cult of vibe coding is insane

cultvibecodinginsanehadleaksourcecodepeoplehave
Relation score 0.70Open topic