DEV Community

Cover image for AI coding agents nail the code. They break on the scaffolding.
Adam
Adam

Posted on • Originally published at weeklyfoo.com

AI coding agents nail the code. They break on the scaffolding.

Signup here for the newsletter to get the weekly digest right into your inbox.

When AI coding agents fall short, the failure is almost never "it generated bad code." The failure is everything else — the environment state, the implicit decisions baked into a config three years ago, the reason that dependency is pinned to that version.

Joel Andrews published one of the more grounded take-downs of AI coding agents this week. Not a hype piece in either direction — just a practitioner's account of where these tools consistently fall apart in real production environments. The pattern he describes maps to something Daniel Miessler flagged in his April AI synthesis: most of what developers call "work" was actually maintaining an elaborate, fragile state required for work to happen. The AI exposed it by stumbling directly into it.

This is what "scaffolding" means in practice. Your agent can write a correct function. It cannot know why your test suite is configured with that specific mock, or why a particular environment variable gets overridden in staging but not local. That context lives in heads and in Slack threads and in the institutional memory of whoever set it up in 2022. The model doesn't have it.

Drew Breunig's framing of the second phase of agentic development makes more sense through this lens. Phase one was: can the AI generate code? (Yes, mostly.) Phase two is: can the agent operate within the scaffolding well enough to do something useful end-to-end? The answer right now is: inconsistently, and the variance is high. Spec-driven development isn't just a prompt engineering trick — it's an attempt to externalize the scaffolding into something the agent can read.

What's encouraging is that we're getting honest data now. The honeymoon period where everything was demos and benchmarks is ending. Practitioners are writing up where it works and where it doesn't, and the failure modes are getting specific enough to be actionable.

Some uncomfortable truths about AI coding agents — grounded production observations from Joel Andrews on where agents consistently fall short, with specifics that let you actually plan around them

The 2nd Phase of Agentic Development — Drew Breunig's model for the phase shift from "can it code?" to spec-driven idiosyncratic tooling

The Most Important Ideas in AI Right Now (April 2026) — Daniel Miessler's quarterly synthesis, worth reading for the self-improvement and transparency threads that contextualize where the capability ceiling actually sits

Inside the Claude Code source — Haseeb Qureshi's reverse-engineering of the Claude Code CLI from the full TypeScript leak; seeing the internals clarifies how the scaffolding problem is being handled at the tool level

The next useful thing the AI tooling community could build isn't a better code generator. It's better scaffolding capture — structured context files, agent instructions, decision logs that encode why the codebase is the way it is. The teams building that layer now will get substantially more out of agents than teams that keep treating them as fast autocomplete.


Want to read more? Check out the full issue here.

To sign up for the weekly newsletter, visit weeklyfoo.com.

Top comments (0)