DEV Community

Cover image for New Standards for the Agent-Readable Website
Kelvin
Kelvin

Posted on

New Standards for the Agent-Readable Website

Saw Garry Tan’s post.

Then ended up here:

https://x.com/flynnjamm/status/2023465136204419096?s=46

And I couldn’t stop thinking about one thing:

How do you sell to agents?

Not humans.

Not people clicking blue links.

Agents.


The shift is real

Search clicks are dropping.

  • Gartner says traditional search volume could fall ~25% by 2026.
  • Ahrefs saw ~34% lower CTR when Google AI Overviews show up.
  • Publisher referrals from Google are down ~30%+ globally.

People are getting answers without visiting sites.

SEO was built for clicks.

Agents are built for extraction.


Bots are everywhere

AI bot traffic is up 300%+ (Akamai).

More machines reading your site than humans.

But your site?

Still written for humans.


The actual problem

It’s never been easier to launch a site.

It’s also never been easier to launch one agents can’t understand.

We have pieces:

  • robots.txt
  • schema.org
  • OpenAPI
  • llms.txt

But no unified “AI-ready” standard.

No Lighthouse for agents.


The idea: ART

Agent Readiness Toolkit (ART)

Think Lighthouse, but for AI discoverability.

It would:

  • Audit how legible your site is to agents
  • Standardize a clean “agent contract”
  • Simulate what an LLM would extract
  • Give you an Agent Readiness Score

Because soon:

Your customer might never visit your site.

Your pricing page might be summarized somewhere else.

Your product might be recommended by a model.

In that world, ranking doesn’t matter.

Being extractable does.

SEO shaped the last 20 years.

Agent readiness might shape the next 20.

Top comments (0)