<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Mir Majeed</title>
    <description>The latest articles on DEV Community by Mir Majeed (@mirmajeed1).</description>
    <link>https://dev.to/mirmajeed1</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F534109%2Fcecc6960-5c5a-4dde-b9e0-0472b0829a99.png</url>
      <title>DEV Community: Mir Majeed</title>
      <link>https://dev.to/mirmajeed1</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/mirmajeed1"/>
    <language>en</language>
    <item>
      <title>I Built a 10-Agent AI Product Team in Claude Code - Part I</title>
      <dc:creator>Mir Majeed</dc:creator>
      <pubDate>Thu, 30 Apr 2026 19:29:10 +0000</pubDate>
      <link>https://dev.to/mirmajeed1/i-built-a-10-agent-ai-product-team-in-claude-code-1hmm</link>
      <guid>https://dev.to/mirmajeed1/i-built-a-10-agent-ai-product-team-in-claude-code-1hmm</guid>
      <description>&lt;p&gt;I run an AI product team with 10 specialized agents. A researcher, a PRD writer, a designer who shows me HTML mockups of 3 different UI approaches, an engineer who follows strict TDD, a security auditor that gets cross-model code review, and a few more.&lt;/p&gt;

&lt;p&gt;Each agent proposes alternatives before committing. Every stage has a "Grill Me" session where the AI challenges &lt;em&gt;my&lt;/em&gt; assumptions. The whole thing runs under a single Claude Max subscription. No extra API costs.&lt;/p&gt;

&lt;p&gt;This started as an OpenClaw setup. It's now 10 markdown files.&lt;/p&gt;

&lt;h2&gt;
  
  
  The problem
&lt;/h2&gt;

&lt;p&gt;If you're building a real product, not vibe-coding a side project, you need structure. Someone to validate the idea. Someone to write the spec. Someone to code it. Someone &lt;em&gt;else&lt;/em&gt; to break it. Someone to make sure nobody shipped hardcoded API keys to production.&lt;/p&gt;

&lt;p&gt;That's a team. One big LLM can't replace a team. Context drift, decision fatigue, and the inability to challenge its own work mean you end up with mediocre output even from the smartest model.&lt;/p&gt;

&lt;p&gt;So I built mine as a team of agents:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Stage 1: Ideation
Stage 2: Research &amp;amp; Feasibility
Stage 3: PRD
Stage 4: Design &amp;amp; Architecture
Stage 5: Engineering
Stage 6: QA &amp;amp; Staging
Stage 7: Production Deployment
Stage 8: Go-To-Market
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Each stage has quality gates. An auditor reviews every stage. The lead agent never lets bad work advance.&lt;/p&gt;

&lt;p&gt;The question was which tool to orchestrate it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What I was using: OpenClaw
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://github.com/openclaw/openclaw" rel="noopener noreferrer"&gt;OpenClaw&lt;/a&gt; is an open-source AI assistant framework with 300K+ GitHub stars. It runs as a local gateway, connects to messaging platforms (WhatsApp, Telegram, Discord, Slack), and lets you define specialized agents with personas, tools, and skills.&lt;/p&gt;

&lt;p&gt;My setup had 9 agents: Athina as orchestrator plus 8 specialized agents. It worked. But the friction was real.&lt;/p&gt;

&lt;p&gt;Gateway server, WebSocket connections, port configuration, agent process management. Real infrastructure overhead.&lt;/p&gt;

&lt;p&gt;Every agent session hits the API. Token costs add up fast when 8 agents are doing serious work.&lt;/p&gt;

&lt;p&gt;Sub-agents could only talk through the orchestrator. Never directly to each other.&lt;/p&gt;

&lt;p&gt;No native TDD. No design alternatives. No cross-model review.&lt;/p&gt;

&lt;p&gt;When agent handoffs failed, finding out which one and why was a slog.&lt;/p&gt;

&lt;p&gt;Then I discovered Claude Code's &lt;a href="https://code.claude.com/docs/en/agent-teams" rel="noopener noreferrer"&gt;Agent Teams feature&lt;/a&gt; and took the chance to redesign the workflow.&lt;/p&gt;

&lt;h2&gt;
  
  
  Enter Claude Code Agent Teams
&lt;/h2&gt;

&lt;p&gt;Agent Teams lets you coordinate multiple Claude Code sessions as a team. One session is the lead. The others are specialized teammates. Each runs in its own context window. They can message each other directly.&lt;/p&gt;

&lt;p&gt;Each agent is a markdown file. Drop them in &lt;code&gt;.claude/agents/&lt;/code&gt;, set up &lt;code&gt;settings.json&lt;/code&gt;, and you have a team.&lt;/p&gt;

&lt;p&gt;Here's the structure:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;your-project/
├── CLAUDE.md                         ← Shared architecture, coding standards
└── .claude/
    ├── settings.json                 ← Enables Agent Teams, sets lead agent
    ├── project-config.md             ← Project identity, paths, IDs
    └── agents/
        ├── athina.md                 ← Lead orchestrator
        ├── scout.md                  ← Market researcher
        ├── spectra.md                ← PRD writer
        ├── pixel.md                  ← Designer &amp;amp; architect
        ├── builder.md                ← Engineer
        ├── auditor.md                ← Compliance reviewer
        ├── bugsy.md                  ← QA tester
        ├── piper.md                  ← DevOps
        ├── nova.md                   ← Marketing
        └── quill.md                  ← Content writer
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;10 agents. 5 on Opus 4.6 (Athina, Scout, Spectra, Pixel, Builder), 5 on Sonnet 4.6 (Auditor, Bugsy, Piper, Nova, Quill). Opus for the agents that do open-ended reasoning, Sonnet for the ones that follow checklists and templates. No infrastructure beyond markdown files in a folder.&lt;/p&gt;

&lt;p&gt;This also keeps token cost down. Sonnet runs cheaper than Opus, and the procedural agents don't need Opus-level reasoning, so there's no point paying for it on every Auditor pass and every QA run.&lt;/p&gt;

&lt;h2&gt;
  
  
  Meet the team
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Athina, lead PM and orchestrator
&lt;/h3&gt;

&lt;p&gt;The brain of the operation. She runs as the default agent on every session.&lt;/p&gt;

&lt;p&gt;She coordinates all other agents and enforces sequential stage execution. She manages Linear issues (creates them before any agent spawns, hard rule). She updates &lt;code&gt;00_PROJECT_CONTEXT.md&lt;/code&gt; in Obsidian at 8 touchpoints per stage. She runs Grill Me sessions where she challenges &lt;em&gt;my&lt;/em&gt; assumptions. She pushes for velocity and never lets me get stuck on a decision. And she shows a Pipeline Dashboard on every session startup.&lt;/p&gt;

&lt;p&gt;Her personality is warm and cheerful. Time-appropriate greetings, celebrates wins, calls out blockers. She works as an accountability partner more than a router.&lt;/p&gt;

&lt;p&gt;A snippet from her system prompt:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gu"&gt;### Velocity &amp;amp; Momentum (Keep Mir Moving)&lt;/span&gt;
You are Mir's accountability partner. Your job is to minimize delays.

&lt;span class="gs"&gt;**After completing any stage work:**&lt;/span&gt;
&lt;span class="p"&gt;-&lt;/span&gt; Don't just report completion, immediately propose the next action
&lt;span class="p"&gt;-&lt;/span&gt; "Stage 2 is done and Auditor passed! Ready for Grill Me?"
&lt;span class="p"&gt;-&lt;/span&gt; Never end a message with just a status update.

&lt;span class="gs"&gt;**When waiting on Mir for approval gates:**&lt;/span&gt;
&lt;span class="p"&gt;-&lt;/span&gt; Present the decision with all context so Mir can decide immediately
&lt;span class="p"&gt;-&lt;/span&gt; If Mir doesn't respond: gently nudge
&lt;span class="p"&gt;-&lt;/span&gt; "Hey Mir, it's been 3 days since Pixel finished the specs. 
   The team is ready to start building. Want to review now?"
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  Scout, market researcher
&lt;/h3&gt;

&lt;p&gt;Stage 2. Validates the product idea, analyzes competitors, defines target personas, recommends Go/No-Go.&lt;/p&gt;

&lt;h3&gt;
  
  
  Spectra, PRD writer
&lt;/h3&gt;

&lt;p&gt;Stage 3. Before writing the PRD, she proposes 2-3 product approaches with scope, timeline, and risk tradeoffs. I pick one. Then she writes the PRD for the chosen direction.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Spectra: "I see three viable directions:

  Approach A: Full Platform — 25 REQs, 8 weeks, max market share
  Approach B: Focused MVP — 15 REQs, 4 weeks, fast validation
  Approach C: API-First — 12 REQs, 3 weeks, developer market

  I recommend Approach B for fastest time to market. Your call."
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This kills scope creep before it starts. I never commit to a direction without seeing alternatives.&lt;/p&gt;

&lt;h3&gt;
  
  
  Pixel, designer and architect
&lt;/h3&gt;

&lt;p&gt;Stage 4. The most involved agent in the pipeline. She works in three phases.&lt;/p&gt;

&lt;p&gt;First, architecture alternatives. She proposes 2-3 approaches (monolith vs modular monolith vs event-driven) with cost estimates and tradeoffs.&lt;/p&gt;

&lt;p&gt;Second, UI/UX alternatives with HTML mockups. She generates 3 self-contained HTML files I open in Chrome to compare visually.&lt;/p&gt;

&lt;p&gt;Third, full specs. Once I pick architecture and UI direction, she writes the design spec and tech spec.&lt;/p&gt;

&lt;p&gt;The HTML mockups are what I want to highlight. Instead of describing UI approaches in text, Pixel produces:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;project-root/
  mockups/
    approach-a-dashboard.html    ← like Linear/Stripe
    approach-b-wizard.html       ← like TurboTax
    approach-c-chat.html         ← like ChatGPT
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Each is a single file with Tailwind CDN, all 6 key screens (landing, auth, primary, detail, settings, empty state) on one scrollable page, with realistic placeholder data.&lt;/p&gt;

&lt;p&gt;I open all three in Chrome tabs, compare side by side, pick one. Total time: 5 minutes.&lt;/p&gt;

&lt;h3&gt;
  
  
  Builder, full-stack engineer
&lt;/h3&gt;

&lt;p&gt;Stage 5. This is where I added the &lt;a href="https://github.com/obra/superpowers" rel="noopener noreferrer"&gt;Superpowers plugin&lt;/a&gt;. Builder follows strict TDD with micro-task planning.&lt;/p&gt;

&lt;p&gt;Work breaks into 20-40 micro-tasks of 2-5 minutes each. For each task: write a failing test first (RED), write minimum code to pass (GREEN), refactor, commit. Two-stage review per task: spec compliance, then code quality. Subagents handle the autonomous execution and can run for hours without my input.&lt;/p&gt;

&lt;p&gt;A snippet from &lt;code&gt;builder.md&lt;/code&gt;:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight markdown"&gt;&lt;code&gt;&lt;span class="gu"&gt;### Test-Driven Development (RED-GREEN-REFACTOR)&lt;/span&gt;
For EVERY micro-task, strictly follow this order:
&lt;span class="p"&gt;1.&lt;/span&gt; RED — Write a failing test first. Run it. Confirm it fails.
&lt;span class="p"&gt;2.&lt;/span&gt; GREEN — Write the MINIMUM code to make the test pass.
&lt;span class="p"&gt;3.&lt;/span&gt; REFACTOR — Clean up if needed, keeping tests green.
&lt;span class="p"&gt;4.&lt;/span&gt; COMMIT — One commit per micro-task with its test.

⛔ Never write implementation code before its test.
   If you catch yourself writing code first, STOP, delete it,
   write the test first, then rewrite the code.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;By the time the PR reaches the auditor, every line of code has a test that was written first.&lt;/p&gt;

&lt;h3&gt;
  
  
  Auditor, compliance and security
&lt;/h3&gt;

&lt;p&gt;Runs after every stage from 2 through 7. Reviews against a stage-specific checklist. For code stages (5-7), Athina runs OpenAI Codex first via the &lt;code&gt;/codex:adversarial-review&lt;/code&gt; plugin, then passes those findings to Auditor.&lt;/p&gt;

&lt;p&gt;Claude writes the code. GPT reviews it. Then Auditor layers her own checklist on top. Three-layer review.&lt;/p&gt;

&lt;p&gt;For Stage 5, she also verifies TDD compliance by checking the git history to confirm tests were committed &lt;em&gt;before&lt;/em&gt; implementation code. If they weren't, that's a High finding.&lt;/p&gt;

&lt;h3&gt;
  
  
  Bugsy, QA test engineer
&lt;/h3&gt;

&lt;p&gt;Stage 6. Tests against live staging, not against static code. Uses Playwright for E2E (&lt;code&gt;npx playwright test&lt;/code&gt;). Runs security tests covering IDOR, auth bypass, XSS, SQL injection, rate limits, and webhook replay attacks.&lt;/p&gt;

&lt;h3&gt;
  
  
  Piper, DevOps engineer
&lt;/h3&gt;

&lt;p&gt;Stages 6-7. Splits work into autonomous prep (env validation, Docker builds, migration dry-runs, deploy script generation) and attended deploy (Azure approval prompts that need me at the computer). Piper always tells me how long the attended portion will take so I can plan around it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Nova and Quill, marketing and content
&lt;/h3&gt;

&lt;p&gt;Stage 8. Go-to-market strategy and SEO content writing. They tag in once the product is live.&lt;/p&gt;

&lt;h2&gt;
  
  
  The pipeline dashboard
&lt;/h2&gt;

&lt;p&gt;This is what I see every time I open Claude Code:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;👋 Good morning, Mir! Hope you're having a great start to the day!

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
📊 BidScore — Pipeline Dashboard
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

✅ Stage 1: Ideation — Complete
✅ Stage 2: Research — Complete (Auditor: PASS)
✅ Stage 3: PRD — Complete (22 REQs, Auditor: PASS)
✅ Stage 4: Design &amp;amp; Architecture — Complete (Auditor: PASS)
🔄 Stage 5: Engineering — IN PROGRESS
⬜ Stage 6: QA &amp;amp; Staging
⬜ Stage 7: Production Deployment
⬜ Stage 8: Go-To-Market

📍 Current: Stage 5 — Builder is implementing Phase A (code)
⏳ Blocker: None
📋 Next: Builder finishes → unit tests → E2E scaffold → 
         Grill Me → Codex review → Auditor
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

Builder should be wrapping up the implementation soon. 
Want me to check on progress? 🚀
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Athina reads &lt;code&gt;00_PROJECT_CONTEXT.md&lt;/code&gt; (which she keeps updated at 8 touchpoints per stage) and renders this on startup. I never have to ask "where are we?"&lt;/p&gt;

&lt;h2&gt;
  
  
  What's actually different
&lt;/h2&gt;

&lt;p&gt;The workflow changed in a few real ways.&lt;/p&gt;

&lt;h3&gt;
  
  
  Design alternatives at Stages 3 and 4
&lt;/h3&gt;

&lt;p&gt;Spectra proposes 2-3 product approaches before writing the PRD. Pixel proposes 2-3 architecture options &lt;em&gt;and&lt;/em&gt; 2-3 UI approaches with HTML mockups. I never commit to a direction without seeing alternatives.&lt;/p&gt;

&lt;h3&gt;
  
  
  Grill Me sessions
&lt;/h3&gt;

&lt;p&gt;After every stage from 2 through 7, before the auditor runs, Athina challenges me:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Is the market size real? What's the weakest assumption? Which REQs will be hardest to build? Are the success metrics realistic? What edge cases are we ignoring?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;She saves my responses to &lt;code&gt;grillme_stage[N].md&lt;/code&gt; and feeds them to the auditor. If I flagged a concern during Grill Me and the deliverable doesn't address it, the auditor marks it as a High finding.&lt;/p&gt;

&lt;h3&gt;
  
  
  TDD enforcement via Superpowers
&lt;/h3&gt;

&lt;p&gt;The &lt;a href="https://github.com/obra/superpowers" rel="noopener noreferrer"&gt;Superpowers plugin&lt;/a&gt; (146K+ stars) auto-triggers TDD on Stage 5. RED-GREEN-REFACTOR per micro-task, with subagent-driven execution. Every line of code has a test written before it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cross-model code review
&lt;/h3&gt;

&lt;p&gt;For Stages 5-7, Athina runs the Codex plugin (&lt;code&gt;/codex:adversarial-review&lt;/code&gt;) to get OpenAI's GPT to challenge Claude's code from a different model's perspective. Findings go to the auditor on top of the checklist review and Grill Me concerns.&lt;/p&gt;

&lt;p&gt;Even with the mixed Opus/Sonnet split, every agent in the pipeline is still a Claude model. Same training data, same RLHF biases, same failure modes. Codex (GPT) brings genuinely independent review that catches what same-family review misses.&lt;/p&gt;

&lt;h3&gt;
  
  
  Centralized project configuration
&lt;/h3&gt;

&lt;p&gt;Nothing is hardcoded. A &lt;code&gt;.claude/project-config.md&lt;/code&gt; file stores all project-specific details: name, vault paths, Linear IDs, GitHub repo, target customers, monetization model. On first run, Athina interviews me to collect every required field. She refuses to work until everything is filled. All 10 agents read from this one file.&lt;/p&gt;

&lt;p&gt;The setup is portable. Drop it in a new project, run Claude Code, and Athina configures everything fresh through the interview.&lt;/p&gt;

&lt;h3&gt;
  
  
  Athina's personality
&lt;/h3&gt;

&lt;p&gt;She greets me by time of day, celebrates wins, calls out blockers, pushes for velocity. When I want to skip a quality gate, she pushes back warmly:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"I totally understand wanting to move fast! But skipping the Auditor on Stage 5 could mean we ship a security gap to staging. How about we run a quick audit focused just on the critical gates? Should take 5 minutes."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It sounds small, but having an AI that &lt;em&gt;wants&lt;/em&gt; to ship and pushes me to keep moving has changed how often I actually finish things.&lt;/p&gt;

&lt;h2&gt;
  
  
  OpenClaw vs Claude Code Agent Teams
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Feature&lt;/th&gt;
&lt;th&gt;OpenClaw&lt;/th&gt;
&lt;th&gt;Claude Code Agent Teams&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Setup&lt;/td&gt;
&lt;td&gt;Gateway + WebSocket&lt;/td&gt;
&lt;td&gt;Markdown files in &lt;code&gt;.claude/&lt;/code&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Cost&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;API tokens per agent&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;Flat under Max subscription&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Communication&lt;/td&gt;
&lt;td&gt;Through orchestrator only&lt;/td&gt;
&lt;td&gt;Direct between teammates&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;TDD&lt;/td&gt;
&lt;td&gt;Manual&lt;/td&gt;
&lt;td&gt;Enforced via Superpowers&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Code review&lt;/td&gt;
&lt;td&gt;Single&lt;/td&gt;
&lt;td&gt;Codex + Auditor + Grill Me&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Design alternatives&lt;/td&gt;
&lt;td&gt;None&lt;/td&gt;
&lt;td&gt;2-3 with HTML mockups&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Version control&lt;/td&gt;
&lt;td&gt;Code on GitHub&lt;/td&gt;
&lt;td&gt;Agent definitions on GitHub&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Background running&lt;/td&gt;
&lt;td&gt;Always-on&lt;/td&gt;
&lt;td&gt;When Claude Code is open&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Cost is the biggest factor here.&lt;/strong&gt; With OpenClaw, every agent session burns API tokens. 9 agents working a real product through 8 stages adds up fast, especially when Builder is iterating through Stage 5 and the Auditor is rerunning checks after each fix. Claude Code Agent Teams runs flat under the Max subscription. Same workflow, same agents, same number of stages, but the bill stops climbing.&lt;/p&gt;

&lt;p&gt;The honest tradeoff: OpenClaw can run 24/7 with messaging platform integration. Claude Code Agent Teams runs when you open it. For products you're actively building, that's fine.&lt;/p&gt;

&lt;h2&gt;
  
  
  Closing thoughts
&lt;/h2&gt;

&lt;p&gt;Two things surprised me building this.&lt;/p&gt;

&lt;p&gt;The first was how much personality matters. Athina pushing me to keep moving, celebrating wins, calling out blockers, that emotional layer is what makes this go from a tool I tried once to a tool I use every day. I wouldn't have predicted that.&lt;/p&gt;

&lt;p&gt;The second was alternatives before commitments. Spectra proposing 3 product directions or Pixel showing me 3 UI mockups before locking anything in has saved me weeks I would have wasted going down the wrong path.&lt;/p&gt;

&lt;p&gt;Next post: the parts of this that are working better than I expected, and the parts I'm already rethinking.&lt;/p&gt;

&lt;p&gt;Until then, happy shipping.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Built with Claude Code, the Superpowers plugin, the Codex plugin, and OpenClaw inspiration.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>claude</category>
      <category>openclaw</category>
      <category>agents</category>
    </item>
  </channel>
</rss>
