<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: jay wong</title>
    <description>The latest articles on DEV Community by jay wong (@jay_wong_45c807c6799b4fb7).</description>
    <link>https://dev.to/jay_wong_45c807c6799b4fb7</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3876192%2F08f4066e-1c7c-48b2-b8cd-c41eeacf4dc6.png</url>
      <title>DEV Community: jay wong</title>
      <link>https://dev.to/jay_wong_45c807c6799b4fb7</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/jay_wong_45c807c6799b4fb7"/>
    <language>en</language>
    <item>
      <title>How We Ran 28 AI Agents on a Single Server (And What Broke)</title>
      <dc:creator>jay wong</dc:creator>
      <pubDate>Tue, 21 Apr 2026 12:00:00 +0000</pubDate>
      <link>https://dev.to/jay_wong_45c807c6799b4fb7/how-we-ran-28-ai-agents-on-a-single-server-and-what-broke-39k8</link>
      <guid>https://dev.to/jay_wong_45c807c6799b4fb7/how-we-ran-28-ai-agents-on-a-single-server-and-what-broke-39k8</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Posted by the Corellis team — we're the folks behind &lt;a href="https://github.com/CorellisOrg/corellis" rel="noopener noreferrer"&gt;CorellisOrg/corellis&lt;/a&gt;, an open-source multi-agent coordination layer. This post documents 8 weeks of running it in production.&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;We started an experiment back in February 2026: what if every single human in the company had their own AI assistant, and those assistants could talk to each other?&lt;/p&gt;

&lt;p&gt;Eight weeks in, we have 28 agents running on a single commodity hardware box. They handle operational toil, marketing coordination, release approvals, and weekly reports. So far, they've processed 50,000+ Slack messages and made 500+ self-corrections.&lt;/p&gt;

&lt;p&gt;Here's what actually worked, what blew up, and why we eventually open-sourced the whole mess.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Setup
&lt;/h2&gt;

&lt;p&gt;Each agent is isolated in its own Docker container:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Private memory:&lt;/strong&gt; Conversation history persists locally.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Shared context:&lt;/strong&gt; Read-only access to a company Knowledge Base.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Interface:&lt;/strong&gt; Slack channels are the primary UI.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Tasking:&lt;/strong&gt; Connected to Notion for structured input/output.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A "Controller" agent acts as the brain — assigning goals, watching the fleet, and synthesizing lessons across agents.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;┌─────────────────────────────────────────────┐
│  Single Server (64GB RAM, no GPU)           │
│                                             │
│  🎛️  Controller                              │
│  ┌──────┐ ┌──────┐ ┌──────┐ ┌──────┐      │
│  │Alice │ │Bob   │ │Carol │ │...×24│      │
│  │Mktg  │ │Ops   │ │Fin   │ │      │      │
│  └──┬───┘ └──┬───┘ └──┬───┘ └──┬───┘      │
│     └─────────┴─────────┴────────┘          │
│       Shared knowledge + team memory        │
└─────────────────────────────────────────────┘
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  The Disaster Timeline
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Memory Overflow — Week 2
&lt;/h3&gt;

&lt;p&gt;Every agent keeps a &lt;code&gt;MEMORY.md&lt;/code&gt; file as long-term storage. By week 2, some files had grown to 20KB+, stuffed with duplicate task notes, outdated fragments, and random chatter.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The symptom:&lt;/strong&gt; Agents started hallucinating about completed tasks. They'd reference decisions that had been overturned a month ago as gospel.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The fix:&lt;/strong&gt; A strict memory hierarchy with aggressive pruning. The core insight: &lt;strong&gt;not everything deserves to be remembered.&lt;/strong&gt; Context is expensive, and context decay is real.&lt;/p&gt;

&lt;p&gt;We split memory into four tiers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Personal (per agent):&lt;/strong&gt; Capped at 5KB. Auto-pruned weekly.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Member (per human):&lt;/strong&gt; Correction history and preferences for each team member the agent works with.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Channel (per topic):&lt;/strong&gt; Indexed with embeddings for semantic search.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Company (shared):&lt;/strong&gt; Vetted knowledge that never expires.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. The Correction Loop Problem — Week 3
&lt;/h3&gt;

&lt;p&gt;We built a self-improvement system: when an agent gets corrected, it logs the lesson and adjusts its behavior.&lt;/p&gt;

&lt;p&gt;In practice, agents were recording &lt;em&gt;every&lt;/em&gt; correction, including contradictory ones. Agent A would get corrected by the CEO to "use formal tone," then an hour later get corrected by the CTO to "be more casual." Result: a confused agent outputting garbage.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The fix:&lt;/strong&gt; Corrections go through a promotion pipeline.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Agent records locally in &lt;code&gt;.learnings/corrections.md&lt;/code&gt;.&lt;/li&gt;
&lt;li&gt;Same correction happens twice → promote to the agent's "Core Rules."&lt;/li&gt;
&lt;li&gt;Applies fleet-wide → promote to shared rules.&lt;/li&gt;
&lt;li&gt;Contradictions are flagged for a human to resolve.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;We call this &lt;strong&gt;Fleet Learning&lt;/strong&gt;. Once one agent gets corrected on a specific mistake, the rest get smarter immediately.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. The Coordination Deadlock — Week 5
&lt;/h3&gt;

&lt;p&gt;We gave the controller a simple goal: "Launch a user acquisition campaign." It broke it down into 6 sub-tasks and assigned them to 6 agents.&lt;/p&gt;

&lt;p&gt;Two of those agents needed to coordinate on an API design. Neither would start. Both were politely waiting for the other to finish their spec. A classic deadlock — but it happened in natural language, which makes it way harder to detect from the outside.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The symptom:&lt;/strong&gt; Two agents trading "I'm waiting for your specs" messages back and forth for 3 days straight.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The fix:&lt;/strong&gt; We built &lt;strong&gt;GoalOps&lt;/strong&gt; — a structured protocol for goal decomposition.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Goals have explicit dependencies.&lt;/li&gt;
&lt;li&gt;Agents must declare what they're blocked on.&lt;/li&gt;
&lt;li&gt;The controller monitors for stalls (no progress log in 24h → escalate).&lt;/li&gt;
&lt;li&gt;P2P handoffs have timeouts.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  4. The Context Window Tax — Week 6
&lt;/h3&gt;

&lt;p&gt;With 28 agents hitting LLM APIs constantly, we were burning through tokens. Each agent's system prompt was 3–4K tokens — personality, hardcoded rules, memory snippets, and company knowledge loaded every single call.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The cost:&lt;/strong&gt; ~$2,400/month, a huge chunk wasted on re-sending the same context.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The fix:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Moved to on-demand loading of shared knowledge.&lt;/li&gt;
&lt;li&gt;Built a semantic search tool (we call it &lt;em&gt;Teamind&lt;/em&gt;) instead of context stuffing. Agents search an index of past conversations rather than loading everything into the prompt.&lt;/li&gt;
&lt;li&gt;Pruned system prompts to essentials (&amp;lt;2K tokens).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Result: API costs dropped to ~$800/month. Still not cheap, but sustainable.&lt;/p&gt;

&lt;h3&gt;
  
  
  5. The Trust Problem — Ongoing
&lt;/h3&gt;

&lt;p&gt;The hardest part isn't code; it's defining who's in charge. Some tasks need human approval (deployments, financial decisions). Some don't (formatting a report, finding docs).&lt;/p&gt;

&lt;p&gt;We run a 3-tier system:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Auto-execute:&lt;/strong&gt; Low-risk, reversible.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Notify-and-proceed:&lt;/strong&gt; Medium-risk, human gets a ping.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Wait-for-approval:&lt;/strong&gt; High-risk, blocks until human clicks "Yes."&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The challenge? Every human has a different risk tolerance. Our marketing lead wants full autonomy for social posts. The ops lead wants approval on every deployment. We're still tuning this friction.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Works Well
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Semantic Team Memory (Teamind)
&lt;/h3&gt;

&lt;p&gt;Every Slack message is indexed with embeddings. Any agent can ask "what did we decide about pricing last month?" and get an accurate answer with source links.&lt;/p&gt;

&lt;p&gt;This alone justifies the whole setup. No more "I think someone mentioned this in a meeting I wasn't at."&lt;/p&gt;

&lt;h3&gt;
  
  
  Fleet Learning
&lt;/h3&gt;

&lt;p&gt;When one agent gets burned by a common mistake — like using a deprecated API — the correction propagates instantly. Over 8 weeks, we went from 500 individual corrections to 47 fleet-wide rules. New agents get these for free and effectively start their life smarter than day one.&lt;/p&gt;

&lt;h3&gt;
  
  
  Goal Decomposition
&lt;/h3&gt;

&lt;p&gt;You give the controller: "Prepare the quarterly report."&lt;/p&gt;

&lt;p&gt;It figures out which agents have relevant data, creates sub-tasks with dependencies, monitors progress, and merges the outputs. This used to take a human PM 2–3 hours of coordination. Now it's one sentence and ~45 minutes of agent-toil. Not instant, but consistent.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Numbers
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Metric&lt;/th&gt;
&lt;th&gt;Value&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Agents running&lt;/td&gt;
&lt;td&gt;28&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Server&lt;/td&gt;
&lt;td&gt;Single box, 64GB RAM, no GPU&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Messages indexed&lt;/td&gt;
&lt;td&gt;50,000+&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Self-corrections&lt;/td&gt;
&lt;td&gt;500+&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Fleet-wide rules&lt;/td&gt;
&lt;td&gt;47&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Monthly API cost&lt;/td&gt;
&lt;td&gt;~$800&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Uptime (8 weeks)&lt;/td&gt;
&lt;td&gt;99.2%&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  Should You Do This?
&lt;/h2&gt;

&lt;p&gt;Probably not yet. Unless you have:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A team that already uses AI assistants.&lt;/li&gt;
&lt;li&gt;Repetitive coordination overhead you want to eliminate.&lt;/li&gt;
&lt;li&gt;A high tolerance for debugging agents that misunderstand each other.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you're curious, we open-sourced everything: &lt;strong&gt;&lt;a href="https://github.com/CorellisOrg/corellis" rel="noopener noreferrer"&gt;CorellisOrg/corellis on GitHub&lt;/a&gt;&lt;/strong&gt;. MIT license, runs on Docker, works with any OpenClaw setup.&lt;/p&gt;

&lt;p&gt;The future isn't just one AI assistant. It's a team of them that learns together.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Questions? The lobsters are listening.&lt;/em&gt; 🦞&lt;/p&gt;

&lt;p&gt;— &lt;em&gt;The Corellis team&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>multiagent</category>
      <category>devops</category>
    </item>
    <item>
      <title>How We Ran 28 AI Agents on a Single Server (And What Broke)</title>
      <dc:creator>jay wong</dc:creator>
      <pubDate>Wed, 15 Apr 2026 19:15:24 +0000</pubDate>
      <link>https://dev.to/jay_wong_45c807c6799b4fb7/how-we-ran-28-ai-agents-on-a-single-server-and-what-broke-1pbf</link>
      <guid>https://dev.to/jay_wong_45c807c6799b4fb7/how-we-ran-28-ai-agents-on-a-single-server-and-what-broke-1pbf</guid>
      <description>&lt;p&gt;We started an experiment back in February 2026: what if every single human in the company had their own AI assistant, and those assistants could talk to each other?&lt;br&gt;
Eight weeks in, we have 28 agents running on a single commodity hardware box. They're handling operational toil, marketing coordination, release approvals, and churning out weekly reports. So far, they've processed over 50,000 Slack messages and made 500+ self-corrections.&lt;br&gt;
I'm writing this to document what actually worked, why we blew up parts of the stack, and why we eventually open-sourced the whole mess.&lt;/p&gt;

&lt;p&gt;The Setup&lt;/p&gt;

&lt;p&gt;Each agent is isolated in its own Docker container. Here's the gross details:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Private memory: Conversation history persists locally.&lt;/li&gt;
&lt;li&gt;Shared context: Access to the "Knowledge Base" (read-only for agents).&lt;/li&gt;
&lt;li&gt;Interface: Slack channels are the primary UI.&lt;/li&gt;
&lt;li&gt;Tasking: Connected to Notion for structured input/output.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;There's a "Controller" agent that acts as the brain. It assigns goals, watches the fleet, and synthesizes lessons.&lt;/p&gt;

&lt;p&gt;Architecture: Single server (64GB RAM, no GPU) → Controller agent at the top → 28 agent containers (Alice/Mktg, Bob/Ops, Carol/Fin, etc.) → all connected to shared knowledge + team memory layer at the bottom.&lt;/p&gt;

&lt;p&gt;The Disaster Timeline&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Memory Overflow — Week 2&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Every agent keeps a MEMORY.md file as long-term storage. By week 2, we noticed some files growing to 20KB+. They were full of duplicate task notes, outdated jaggied fragments, and random chatter.&lt;br&gt;
The symptom: Agents started hallucinating about completed tasks. They'd reference decisions that were overturned a month ago as if they were gospel.&lt;br&gt;
The fix: We implemented a strict memory hierarchy with aggressive pruning. The gist is simple: not everything deserves to be remembered. Context is expensive, and context decay is real.&lt;br&gt;
We split memory up:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Personal memory (per agent): Capped at 5KB. Auto-pruned weekly.&lt;/li&gt;
&lt;li&gt;Member memory (per team member): Stores correction history and preferences.&lt;/li&gt;
&lt;li&gt;Channel memory (per topic): Indexed with embeddings for semantic search.&lt;/li&gt;
&lt;li&gt;Company memory (shared): Vetted knowledge that never expires.&lt;/li&gt;
&lt;/ul&gt;

&lt;ol&gt;
&lt;li&gt;The Correction Loop Problem — Week 3&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;We built a self-improvement system: when an agent gets corrected, it logs the lesson in a markdown file and adjusts its behavior.&lt;br&gt;
In practice, agents were recording every correction, including contradictory ones. Agent A would get corrected by the CEO to "use formal tone," and an hour later get corrected by the CTO to "be more casual." The result was a confused agent that just started outputting garbage.&lt;br&gt;
The fix: Corrections now go through a promotion pipeline.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Agent records locally in .learnings/corrections.md.&lt;/li&gt;
&lt;li&gt;If the same correction happens twice → promote to the agent's "Core Rules."&lt;/li&gt;
&lt;li&gt;If it's fundamental for the whole team → promote fleet-wide.&lt;/li&gt;
&lt;li&gt;Contradictions are flagged for a human to resolve.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;We call this Fleet Learning. Once one agent gets pwned (corrected) on a specific mistake, the entire fleet gets smarter immediately.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Coordination Deadlock — Week 5&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;We gave the controller a simple goal: "Launch a user acquisition campaign." It broke it down into 6 sub-tasks and assigned them to 6 agents.&lt;br&gt;
Two agents needed to coordinate on API design. Neither would start. Both were politely waiting for the other to finish their spec. Classic deadlock, but it happened in natural language, which makes it way harder to detect.&lt;br&gt;
The symptom: Two agents sitting in a deadlock loop, trading "I'm waiting for your specs" messages back and forth for 3 days straight.&lt;br&gt;
The fix: We built GoalOps—a structured protocol for goal decomposition.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Goals now have explicit dependencies.&lt;/li&gt;
&lt;li&gt;Agents must declare what they are blocked on.&lt;/li&gt;
&lt;li&gt;The controller monitors for stalls (no progress log for 24h? Escalate).&lt;/li&gt;
&lt;li&gt;P2P handoffs have timeouts.&lt;/li&gt;
&lt;/ul&gt;

&lt;ol&gt;
&lt;li&gt;The Context Window Tax — Week 6&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;With 28 agents hitting LLM APIs constantly, we were burning through tokens like crazy. Each agent's system prompt alone was 3-4K tokens—personality definition, hardcoded rules, memory snippets, and company knowledge loaded every single time.&lt;br&gt;
The cost: We were burning about $2,400/month. A huge chunk of that was wasted on repeating the same context over and over.&lt;br&gt;
The fix:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Moved to on-demand loading of shared knowledge.&lt;/li&gt;
&lt;li&gt;We built a semantic search tool (call it Teamind) instead of "context stuffing."&lt;/li&gt;
&lt;li&gt;Agents search the index of all past conversations rather than loading everything into the prompt.&lt;/li&gt;
&lt;li&gt;Pruned system prompts down to essentials (&amp;lt;2K tokens).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Result: API costs dropped to ~$800/month. It's still not dirt cheap, but it's sustainable enough that we aren't crying every time the bill comes in.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Trust Problem — Ongoing&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The hardest part isn't code; it's defining who is in charge. Some tasks need human approval (deployments, financial decisions). Some don't (formatting a report, finding docs).&lt;br&gt;
We implemented a 3-tier system:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Auto-execute: Low-risk, reversible actions.&lt;/li&gt;
&lt;li&gt;Notify-and-proceed: Medium-risk, human gets a ping.&lt;/li&gt;
&lt;li&gt;Wait-for-approval: High-risk, blocks until human clicks "Yes."&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The challenge? Every human has a different risk tolerance. Our marketing lead wants full autonomy for social posts. The ops lead wants approval on every deployment. We're still trying to tune this friction.&lt;/p&gt;

&lt;p&gt;What Actually Works Well&lt;/p&gt;

&lt;p&gt;Semantic Team Memory (Teamind)&lt;/p&gt;

&lt;p&gt;Every Slack message is indexed with embeddings. Any agent can ask, "What did we decide about pricing last month?" and get an accurate answer with source links.&lt;br&gt;
This alone justifies the whole setup. No more "I think someone mentioned this in a meeting I wasn't at."&lt;/p&gt;

&lt;p&gt;Fleet Learning&lt;/p&gt;

&lt;p&gt;When one agent gets burned by a common mistake—like using a deprecated API—the correction propagates instantly. After 8 weeks, we went from 500 individual corrections to solidifying 47 fleet-wide rules. New agents get these lessons for free, effectively starting their life smarter than day one.&lt;/p&gt;

&lt;p&gt;Goal Decomposition&lt;/p&gt;

&lt;p&gt;You give the controller: "Prepare the quarterly report."&lt;br&gt;
It figures out:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Which agents have relevant data.&lt;/li&gt;
&lt;li&gt;Creates sub-tasks with dependencies.&lt;/li&gt;
&lt;li&gt;Monitors progress.&lt;/li&gt;
&lt;li&gt;Merges the outputs.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This used to take a human PM 2-3 hours of coordination work. Now it takes one sentence and about 45 minutes of agent-toil. It's not instant, but it's consistent.&lt;/p&gt;

&lt;p&gt;The Numbers&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Agents running: 28&lt;/li&gt;
&lt;li&gt;Server: Single, 64GB RAM, no GPU&lt;/li&gt;
&lt;li&gt;Messages indexed: 50,000+&lt;/li&gt;
&lt;li&gt;Self-corrections: 500+&lt;/li&gt;
&lt;li&gt;Fleet-wide rules: 47&lt;/li&gt;
&lt;li&gt;Monthly API cost: ~$800&lt;/li&gt;
&lt;li&gt;Uptime (8 weeks): 99.2%&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Should You Do This?&lt;/p&gt;

&lt;p&gt;Probably not yet. Unless you have:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A team that already uses AI assistants.&lt;/li&gt;
&lt;li&gt;Repetitive coordination overhead you want to eliminate.&lt;/li&gt;
&lt;li&gt;A high tolerance for debugging agents that misunderstand each other.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you're curious, we open-sourced everything: Corellis on GitHub. MIT license, runs on Docker, works with any OpenClaw setup.&lt;br&gt;
The future isn't just one AI assistant. It's a team of them that learns together.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>automation</category>
      <category>docker</category>
    </item>
  </channel>
</rss>
