The debate is framed wrong.
Every week someone publishes a benchmark comparing GPT-5.x vs Claude Opus vs Gemini on SWE-bench. The implicit assumption: the model is the variable that matters. Pick the best model, your coding agent works better.
But a benchmark published last month broke that assumption cleanly. Grok Code Fast went from 6.7% to 68.3% on a real-world coding task — not because the model changed, not because of a new training run — because the edit tool format changed. That's a 10x improvement from a single harness modification.
The Edit Tool Problem
Most coding agents use one of three edit formats:
- apply_patch (Codex): OpenAI-flavored diff strings. Works great for GPT variants tuned for it. Give it to Grok 4 and the patch failure rate hits 50.7%.
- str_replace (Claude Code, most others): Find the exact old text, replace with new. Simple to reason about, but the model must reproduce every character including whitespace. A single indentation difference = failure. There's a GitHub megathread about this.
- Cursor's neural network: They trained a separate 70B model just to apply edits correctly. That's how hard this problem is.
The new approach — hashline — tags every line with a 2-3 character content hash when the model reads a file:
11:a3|function hello() {
22:f1| return "world";
33:0e|}
The model edits by referencing hashes, not reproducing text. If the file changed since the last read, the hashes won't match and the edit is rejected before corruption. No whitespace reproduction required. No perfect recall required.
Results across 16 models: hashline matches or beats str_replace for most, and weakest models gain the most. Grok 4 Fast's output tokens dropped 61% because it stopped burning tokens on retry loops for failed edits.
Why This Is a Distributed Systems Problem
The latent.space harness debate frames this as Big Model vs Big Harness. But that's still the wrong frame. The right question is: which layer owns which decisions?
Noam Brown (OpenAI) argues scaffolding fills capability gaps, and as models get better, scaffolding collapses. He's right about cognitive scaffolding — chain-of-thought prompting, multi-step decomposition, RAG pipelines. Those compress into models over time.
But the edit tool problem isn't cognitive. It's mechanical. It's the interface between model output and filesystem state. Models understand perfectly what to change. They fail at expressing the change in a format the harness can parse reliably. That's not a language modeling problem — it's an interface design problem. Models don't absorb interface design problems.
Same logic applies to:
- Provider failover and circuit breaking (distributed systems)
- Parallel task execution with dependency ordering (scheduling)
- Cost tracking and budget enforcement (financial controls)
- State persistence across session boundaries (storage)
None of these are cognitive. All of them are infrastructure. Infrastructure doesn't compress into bigger language models.
What the Vendor Blocking Tells You
Anthropic recently blocked OpenCode — a popular open-source agent — from using Claude Code subscriptions. Google disabled a researcher's account for running a benchmark on Gemini. The researcher's benchmark showed Gemini 3 Flash hitting 78.3% with a novel technique that beats Google's own attempt by 5 points.
The signal is clear: don't build harnesses, use ours.
But no vendor will do harness optimization for their competitors' models. Anthropic won't tune for Grok. xAI won't tune for Gemini. An open-source harness does, because contributors fix the failures they personally encounter across whichever models they use.
Building for the Durable Half
We've been building claw-forge as a multi-provider autonomous coding agent harness. The design philosophy matches this analysis: the parts of the harness that survive model improvement are the infrastructure parts.
We're adding hashline edit mode as our next PR. The benchmark methodology is straightforward to replicate — random file from a known codebase, mechanical mutation, fix rate per format. We'll publish our numbers.
The harness problem isn't going away. The question is whether it gets solved by one company, in private, for one model — or by a community, in the open, for all of them.
claw-forge is open source: github.com/clawinfra/claw-forge. The hashline technique is from oh-my-pi by can1357.
Top comments (0)