canaisee

Rubric v1.3

Every scan is scored against the published rubric below. Rubric changes require two approvals — EAC engineering (measurement correctness) and Evangent leadership (grading integrity) — and every change bumps the version string. Historical scans retain the rubric version that was active when they ran.

The two-tier composite

v1.3 replaces the v1.2 four-categories-weighted model with a two-tier Hygiene + Frontier composite. Hygiene measures whether agents can read the site at all — sensible robots.txt, clean HTML, content reachable without JavaScript, semantic structure, decent token efficiency. Frontier measures genuinely ahead-of-the-curve agent-native features — Accept: text/markdown,/.well-known/mcp.json, WebMCP tool declarations, Content-Signal headers, A2A agent cards. Most well-built 2026 sites land in the B range on hygiene alone; Frontier is what lifts a site to A and A+.

Composite formula

// Content site
composite = hygiene * 0.85 + frontier * 0.15

// Interactive site
composite = hygiene * 0.75 + frontier * 0.25

Interactive sites get more frontier credit because WebMCP tools and MCP endpoints genuinely matter more there — an e-commerce site with agent tools is materially more usable than one without.

Site-type classification

Sites are classified automatically as Content (blog, church, news, docs, nonprofit) or Interactive (SaaS, e-commerce, booking, dashboards). Classification uses schema.org types, form patterns, and URL hints, and can be overridden by the user.

Hygiene checks

Hygiene is a weighted average of three sub-categories. The sub-category weights roll up into the Hygiene composite.

Sub-categoryWeight in Hygiene
Crawler accessibility30%
Content readability35%
Semantic structure35%
CheckSub-categoryWithin-sub-category weight
robots.txt AI-agent policycrawler45%
JavaScript dependencycrawler40%
Login / paywall wallcrawler15%
Accept: text/markdown negotiationreadability15%
Content-to-chrome ratioreadability45%
Token efficiencyreadability40%
JSON-LD / schema.orgsemantic30%
Open Graph & Twitter Cardssemantic20%
Heading hierarchysemantic20%
Language declarationsemantic10%
sitemap.xmlsemantic20%

Frontier checks

Frontier is generous at the low end and demanding at the high end. A site with any two of the top three (content negotiation, MCP, WebMCP or rich structured data) scores 50+. A site with all three plus decent minor signals scores 90+. The top scores (95+) are reserved for sites that have genuinely built comprehensive agentic-web support.

CheckWeight in FrontierApplies to
Accept: text/markdown negotiation25%both
MCP server discovery25%both
WebMCP tool declarations20%interactive
Rich structured data20%content
Content-Signal response header10%both
A2A agent card10%both
llms.txt / ai.txt intent signal5%both
.md mirror5%both

Letter grades

CompositeGradeMeaning
95–100A+Agentic-web leadership. Genuinely exceptional.
88–94AAhead of the curve on AI readability.
78–87BFundamentals solid; frontier features would push to A.
65–77CReadable, but real gaps. Clear next steps.
50–64DSerious issues; agents struggle here.
0–49FUnreadable to most agents.

Human-only ceiling (§4.6)

A site that scores below 15 on Frontier — meaning essentially no agent-native features — has its composite capped at 85 (top of B) regardless of how strong its Hygiene score is. This is a softer replacement for the v1.2 D+ penalty: pristine HTML and perfect semantic structure still top out at B until the site adds some agent-native surface. The cap is disclosed on every affected scorecard — it’s aspirational, not punishing: B is a respectable grade that says “you’re doing the basics right” and the cap preserves a meaningful difference between sites that have invested in the agentic web and sites that have only invested in the traditional web.

Where we're opinionated

The rubric tracks live agentic-web protocols — Accept: text/markdown, /.well-known/mcp.json, WebMCP, Content-Signal — because empirical data shows these are what agents actually use in 2026. llms.txt and ai.txt are kept as minor signals of intent, not load-bearing checks; independent log studies show ~0.1% of AI crawler requests fetch them.

Governance

Rubric changes are tracked in the lib/rubric/ directory of the open-source repo, one file per version. Breaking changes live alongside the previous version so historical scans continue to read their own version's scoring logic. The rubric is reviewed quarterly against the current agentic-web landscape.