<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Andrey Kucherenko</title>
    <description>The latest articles on DEV Community by Andrey Kucherenko (@kucherenko).</description>
    <link>https://dev.to/kucherenko</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F203305%2F7d767a58-54d2-4bc8-946a-13a1482f9414.png</url>
      <title>DEV Community: Andrey Kucherenko</title>
      <link>https://dev.to/kucherenko</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/kucherenko"/>
    <language>en</language>
    <item>
      <title>16 frameworks. One Blind Spot</title>
      <dc:creator>Andrey Kucherenko</dc:creator>
      <pubDate>Wed, 06 May 2026 13:12:36 +0000</pubDate>
      <link>https://dev.to/kucherenko/16-frameworks-one-blind-spot-20cg</link>
      <guid>https://dev.to/kucherenko/16-frameworks-one-blind-spot-20cg</guid>
      <description>&lt;p&gt;Every AI agent framework on the market today has the same fatal flaw.&lt;/p&gt;

&lt;p&gt;They will take your half-baked, 3 a.m., "wouldn't it be cool if..." idea and dutifully - &lt;strong&gt;expensively&lt;/strong&gt; - turn it into a beautifully specified, well-architected, properly tested pile of garbage.&lt;/p&gt;

&lt;p&gt;Spec-driven? Sure. &lt;br&gt;
Multi-agent? Of course. &lt;br&gt;
Self-reviewing? Absolutely. &lt;br&gt;
Adversarial? They'll claim it.&lt;/p&gt;

&lt;p&gt;But ask any of them a single question:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Did you challenge the idea before you built it?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;Silence.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;This is Part 1 of a two-part series. Part 1 maps the landscape and identifies the structural gap. Part 2 shows what filling that gap actually looks like - at the level of agents, prompts, and termination conditions.&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;The garbage-in industrial complex&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Here's the workflow every AI framework optimizes for:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;User has an idea.&lt;/li&gt;
&lt;li&gt;Framework generates a spec.&lt;/li&gt;
&lt;li&gt;Framework writes the code.&lt;/li&gt;
&lt;li&gt;Framework tests the code.&lt;/li&gt;
&lt;li&gt;User gets a working implementation of a bad idea.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Sixteen frameworks.&lt;/strong&gt; The space exploded after GitHub released Spec-Kit in September 2025 - every framework I'll show you below either predates that by months or appeared in its wake. All of them are sophisticated machinery for steps 2 through 4:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Planning agents. &lt;/li&gt;
&lt;li&gt;Reviewer agents. &lt;/li&gt;
&lt;li&gt;Test-writing agents. &lt;/li&gt;
&lt;li&gt;Constitution enforcers. &lt;/li&gt;
&lt;li&gt;Traceability auditors. &lt;/li&gt;
&lt;li&gt;Agents that argue with each other about implementation details.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Nobody is arguing about whether the thing should exist.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Concrete example: &lt;strong&gt;Spec-Kit,&lt;/strong&gt; one of the most-installed SDD tools, will dutifully run &lt;code&gt;specify init&lt;/code&gt; and create a &lt;code&gt;.specify/&lt;/code&gt; directory with a nine-article constitution, a spec template, a plan template, and a tasks template - &lt;strong&gt;for an idea that should never have been written down.&lt;/strong&gt; &lt;/p&gt;

&lt;p&gt;The constitution will then block your merge if the implementation violates Article III's testing standards. It will not block the merge if the feature itself was &lt;strong&gt;a bad idea&lt;/strong&gt;. That's not in scope.&lt;/p&gt;

&lt;p&gt;The dirty secret of the AI agent space is that most ideas users bring to these systems are bad. Not evil-bad. Just unexamined. Half-thought. &lt;strong&gt;Solving the wrong problem.&lt;/strong&gt; Built on assumptions that fall apart under thirty seconds of pressure.&lt;/p&gt;

&lt;p&gt;And the frameworks &lt;strong&gt;happily burn tokens&lt;/strong&gt; turning that into an artifact.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;"But we have reviewers"&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;No, you don't. Not really.&lt;/p&gt;

&lt;p&gt;I want to be specific here, because "frameworks have reviewers" is technically true and substantively misleading. Let me walk through what current reviewer agents actually do - and what they don't.&lt;/p&gt;

&lt;p&gt;Take &lt;strong&gt;Superpowers&lt;/strong&gt;, one of the most-installed Claude Code plugins right now. After implementation, two reviewer subagents spawn in fresh contexts: a spec compliance reviewer and a code quality reviewer. Each returns a PASS or FAIL verdict. If either fails, the task isn't marked complete. &lt;/p&gt;

&lt;p&gt;This is genuinely good engineering - it eliminates the most common AI coding failure mode, which is the agent declaring victory on code that doesn't actually work. But notice the framing. The spec compliance reviewer asks: &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Does this code match the spec?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The code quality reviewer asks: &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Is this code well-written? &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Neither one asks &lt;strong&gt;whether the spec made sense in the first place.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Take &lt;strong&gt;MUSUBI&lt;/strong&gt;, the most formally rigorous open-source SDD tool I've seen. It ships with a @traceability-auditor skill that blocks merges if requirement-to-test coverage drops below 100%, and a @constitution-enforcer skill that checks every file against nine constitutional articles before merging. &lt;/p&gt;

&lt;p&gt;The traceability matrix maps every requirement to a design section, a code file, and a test file. Beautiful enforcement. But the requirements themselves - the things being traced - were written before the auditor ever ran. They were never on trial. &lt;strong&gt;If the requirement was wrong, the matrix just makes sure the wrong thing is fully implemented and fully tested.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Take &lt;strong&gt;CSDD&lt;/strong&gt;, the academic methodology from Marri. It embeds security constraints directly into the project's constitution — &lt;em&gt;SEC-001 (CWE-89) - SQL Injection Prevention,&lt;/em&gt; with MUST-level enforcement, semgrep integration, and rejection of any code that violates the rule. The reported result is a 73% reduction in security defects. Genuinely impressive - for security defects [1].  &lt;strong&gt;The constitution still doesn't ask whether the feature should exist.&lt;/strong&gt; It just makes sure that &lt;strong&gt;if&lt;/strong&gt; you build the feature, you don't build it with SQL injection in it.&lt;/p&gt;

&lt;p&gt;Or look at &lt;strong&gt;Intent&lt;/strong&gt;, the commercial L2 framework with a $252M war chest behind it. Its Verifier agent atomically updates the spec when the implementing agent discovers an API change or a missing requirement. &lt;/p&gt;

&lt;p&gt;This is genuinely innovative - bidirectional Living Spec is a real advance. But the Verifier verifies the &lt;em&gt;spec-to-code consistency&lt;/em&gt;, not the &lt;em&gt;idea-to-need consistency&lt;/em&gt;. &lt;strong&gt;The spec is still gospel.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I could keep going. Every reviewer in every framework reviews &lt;strong&gt;the implementation against the spec.&lt;/strong&gt; &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;They check whether the code matches what was planned. &lt;br&gt;
They catch bugs. &lt;br&gt;
They suggest refactors. &lt;br&gt;
They block bad merges.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;None of them ask:&lt;/strong&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Should this have been planned in the first place?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The spec is treated as gospel the moment it's written. Every downstream agent assumes the upstream decision was correct. The whole pipeline is a confidence-laundering machine — your vague idea goes in one end, a fully-specced deliverable comes out the other, and at no point did anyone say: &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Wait, this is dumb.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;That's not a reviewer problem. That's a design problem. Reviewers can only check what they're pointed at, and they're all pointed downstream.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Sixteen frameworks. One empty column.&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;I went through the most popular spec-driven frameworks I could find - all of them open-source and commercial,  production and research. Sixteen of them. &lt;/p&gt;

&lt;p&gt;For each one I cataloged: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;what quality mechanisms it actually ships;&lt;/li&gt;
&lt;li&gt;what gates it enforces;&lt;/li&gt;
&lt;li&gt;what it blocks;&lt;/li&gt;
&lt;li&gt;what it merely suggests.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Here's the result.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2mma04tud7s2riw3xw61.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2mma04tud7s2riw3xw61.jpg" alt="Sixteen frameworks. One empty column." width="800" height="729"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The frameworks group into three rigor levels. The taxonomy comes from Piskala's 2026 paper [2], and I find it useful precisely because it cuts across popularity - the most-starred tool isn't necessarily the most rigorous, and the most rigorous tool isn't necessarily the most useful. Three different things.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;L1 - Spec-First.&lt;/strong&gt; &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;You write a spec before code. The spec lives in the feature branch. After merge, it becomes historical documentation. The promise is flexibility; the cost is low. &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Most frameworks live here: Spec-Kit, Superpowers, BMAD-METHOD, GSD, cc-sdd, Don Cheli SDD, Agent OS v3, Shotgun, SpecSwarm, OWASP Skill, WordPress SDD. &lt;/p&gt;

&lt;p&gt;Eleven of the sixteen. This is where almost all real-world adoption happens, because the overhead is bearable and the discipline is real.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;L2 - Spec-Anchored.&lt;/strong&gt; &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The spec lives the full project lifecycle. Every change requires an addendum. Every addendum propagates to design, code, and tests. The promise is auditability and traceability; the cost is process overhead. &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Three frameworks here: MUSUBI (orthodox traceability with EARS format), Intent (modern bidirectional Living Spec), OpenSpec (brownfield-first with delta specs).&lt;/p&gt;

&lt;p&gt;This level pays off when an auditor is going to ask hard questions and you need to answer them with receipts.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;L3 - Spec-as-Source.&lt;/strong&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The spec is the only edited artifact. Code is generated and carries a literal // GENERATED FROM SPEC - DO NOT EDIT comment. &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Two examples: Tessl (commercial, currently in beta) and CSDD (academic). The promise is zero spec drift by construction. &lt;/p&gt;

&lt;p&gt;The cost is that this approach is still aspirational outside narrow contexts - generated-code workflows, security-critical microservices, certain OSS libraries.&lt;/p&gt;

&lt;p&gt;Now look across all three levels at the guardrails each framework offers. Every kind you can imagine is present somewhere.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Spec gates&lt;/strong&gt; that block implementation without an approved design - the &lt;code&gt;HARD-GATE&lt;/code&gt; in Superpowers' brainstorming skill, the prerequisite chain in Spec-Kit's CLI, the Contract signature in MUSUBI.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;TDD enforcement&lt;/strong&gt; that refuses production code without a failing test first - Superpowers' iron law, MUSUBI's blocking Article III, Don Cheli SDD's TypeScript orchestrator gate at 85% coverage.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Constitutions&lt;/strong&gt; that block merges when code violates project principles - Spec-Kit's nine articles, MUSUBI's nine articles, CSDD's CWE-mapped security clauses, WordPress SDD's domain-specific constitution.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Traceability matrices&lt;/strong&gt; linking requirement to design to code to test - MUSUBI's 100% coverage requirement, CSDD's Compliance Traceability Matrix, Tessl's Spec Registry, Intent's bidirectional sync.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Code review subagents&lt;/strong&gt; with fresh contexts and PASS/FAIL verdicts - Superpowers' two reviewers, MUSUBI's traceability-auditor, Intent's Verifier.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Five columns of guardrails.&lt;/strong&gt; All of them point in the same direction. Every single one is asking a variant of the same question: &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Did the implementation follow the rules?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The last two columns of the matrix - &lt;strong&gt;Adversarial debate&lt;/strong&gt; and &lt;strong&gt;Nash equilibrium&lt;/strong&gt; - &lt;strong&gt;are empty for fifteen&lt;/strong&gt; of the sixteen frameworks.&lt;/p&gt;

&lt;p&gt;Nobody, &lt;strong&gt;except one&lt;/strong&gt;, checks the spec against attack before it gets written. That is the hole.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;The phase that's missing&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;There's a phase that has to happen &lt;strong&gt;before&lt;/strong&gt; specification, and it's the most important one. Call it whatever you want. I call it &lt;strong&gt;The Grilling.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Grilling is what happens when you take an idea to a room full of people who don't care about your feelings and they tear it apart for an hour.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;You leave that room with one of three outcomes:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;The idea is dead.&lt;/strong&gt; &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Good. You just saved a week.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;The idea is different now.&lt;/strong&gt; &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Better. The thing you build is the thing that should exist.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;The idea survives intact.&lt;/strong&gt; &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Rare. But now it's &lt;em&gt;load-bearing,&lt;/em&gt; not vibes.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Every good engineering org runs some version of this. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;RFCs with adversarial review. &lt;/li&gt;
&lt;li&gt;Design docs with explicit Alternatives Considered and Drawbacks sections. &lt;/li&gt;
&lt;li&gt;Pre-mortems. &lt;/li&gt;
&lt;li&gt;Red teams. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The whole &lt;strong&gt;point&lt;/strong&gt; is to find out the idea is bad &lt;strong&gt;before&lt;/strong&gt; you've built it, not after.&lt;/p&gt;

&lt;p&gt;AI frameworks skipped this entirely. They went straight from &lt;em&gt;"user said something"&lt;/em&gt; to &lt;em&gt;"let's spec it."&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;What's missing isn't another reviewer&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;It would be tempting to look at this gap and say: just add another reviewer agent, one that reviews the spec instead of the code. Problem solved.&lt;/p&gt;

&lt;p&gt;It wouldn't work, and the reasons it wouldn't work are &lt;strong&gt;exactly what makes the missing phase interesting.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;A reviewer agent operates on a single artifact and checks it against rules. That works for &lt;em&gt;code-against-spec&lt;/em&gt; or &lt;em&gt;spec-against-constitution&lt;/em&gt;. &lt;/p&gt;

&lt;p&gt;It doesn't work for &lt;em&gt;idea-against-reality&lt;/em&gt;, because there's &lt;strong&gt;no static&lt;/strong&gt; rulebook to check against. The check has to be &lt;strong&gt;dynamic&lt;/strong&gt; - you have to actually &lt;strong&gt;attack the idea&lt;/strong&gt; from multiple angles, see what breaks, see what holds, see where the assumptions are fragile. You can't write that as a checklist.&lt;/p&gt;

&lt;p&gt;It also has to run &lt;em&gt;with grounding&lt;/em&gt;. An attacker that doesn't know:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;what's already in the codebase;&lt;/li&gt;
&lt;li&gt;what's already been tried;&lt;/li&gt;
&lt;li&gt;what failed before;&lt;/li&gt;
&lt;li&gt;what constraints apply. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That attacker just &lt;strong&gt;hallucinates&lt;/strong&gt; objections that don't apply to the real system. &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;The Devil's Advocate role only has weight when it's standing on verified ground truth.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And it has to &lt;em&gt;terminate&lt;/em&gt; on a substantive condition, not a procedural one. &lt;em&gt;"We've talked enough"&lt;/em&gt; is not the same as &lt;em&gt;"The idea has been adequately tested&lt;/em&gt;." &lt;/p&gt;

&lt;p&gt;Frameworks that have multi-agent debate often &lt;strong&gt;stop&lt;/strong&gt; after a fixed number of rounds, or when a synthesizer declares the discussion mature. &lt;strong&gt;That's exhaustion, not equilibrium.&lt;/strong&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;So what would a framework with this phase actually look like?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Once you take the question seriously, the constraints multiply quickly. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;It can't be a single prompt. &lt;/li&gt;
&lt;li&gt;It can't happen before reconnaissance of the actual codebase. &lt;/li&gt;
&lt;li&gt;It needs structural termination.&lt;/li&gt;
&lt;li&gt;It has to live inside a larger pipeline with downstream gates, otherwise the verdict gets quietly ignored.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I spent a lot of time trying to add this on top of existing frameworks. None of them quite fit. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The L1 tools are too downstream — by the time their gates fire, the idea is already a spec, and the spec is already being implemented. &lt;/li&gt;
&lt;li&gt;The L2 tools add traceability but still treat the spec as input, not as the thing being challenged. &lt;/li&gt;
&lt;li&gt;The L3 tools generate code from spec - but that just makes the spec even more sacred, not less.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The mismatch was structural. The missing phase had to sit upstream of where every existing framework starts.&lt;/p&gt;

&lt;p&gt;So I built &lt;strong&gt;&lt;a href="https://github.com/kucherenko/gangsta" rel="noopener noreferrer"&gt;Gangsta Agents&lt;/a&gt;&lt;/strong&gt; - a new open-source agentic framework &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;where adversarial scrutiny is a phase, not a prompt; &lt;/li&gt;
&lt;li&gt;where it sits before the spec, not after it; &lt;/li&gt;
&lt;li&gt;and where the verdict it produces flows through a chain of gates that the team can't quietly skip.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The framework has six phases: &lt;br&gt;
&lt;strong&gt;Reconnaissance → Grilling → Sit-Down → Resource Development → The Hit → Laundering.&lt;/strong&gt; &lt;/p&gt;

&lt;p&gt;&lt;em&gt;The mafia metaphor is intentional&lt;/em&gt; - every phase has a role, every artifact has a name, every transition has a signature from the Don (user). It's a young project (first stable release in April 2026), and the rest of the pipeline is worth its own treatment.&lt;/p&gt;

&lt;p&gt;But the phase that's genuinely new — the phase that doesn't exist anywhere else - is Phase 2.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Coming in Part 2&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Part 2 is the deep dive: how Grilling actually runs as a phase, what the three subagents do (Proposer, Devil's Advocate, Synthesizer), why the user sits at the table as the Don, what Nash Equilibrium termination looks like in practice, and when not to use Grilling at all. With diagrams, decision rules, and a concrete session example.&lt;/p&gt;

&lt;p&gt;If the chart in this article changed how you see the SDD landscape, the next one is where the alternative gets specified.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/kucherenko/gangsta" rel="noopener noreferrer"&gt;github.com/kucherenko/gangsta&lt;/a&gt; &lt;br&gt;
&lt;a href="https://gangsta.page" rel="noopener noreferrer"&gt;gangsta.page&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;References&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;[1] Marri, S. R. Constitutional Spec-Driven Development: Enforcing Security by Construction in AI-Assisted Code Generation. arXiv:2602.02584, January 2026. &lt;a href="https://arxiv.org/abs/2602.02584" rel="noopener noreferrer"&gt;https://arxiv.org/abs/2602.02584&lt;/a&gt;. The 73% figure is from a single banking microservices case study (n=1) — replication needed, but the methodology is solid.&lt;/p&gt;

&lt;p&gt;[2] Piskala, D. B. Spec-Driven Development: From Code to Contract in the Age of AI Coding Assistants. arXiv:2602.00180, January 2026. &lt;a href="https://arxiv.org/abs/2602.00180" rel="noopener noreferrer"&gt;https://arxiv.org/abs/2602.00180&lt;/a&gt;. Source of the L1/L2/L3 rigor taxonomy used in the matrix.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>llm</category>
      <category>opensource</category>
    </item>
  </channel>
</rss>
