<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: cyw</title>
    <description>The latest articles on DEV Community by cyw (@cyw95).</description>
    <link>https://dev.to/cyw95</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3395587%2F11e2d787-300b-4a37-aa2e-6a1cbf27dc86.jpeg</url>
      <title>DEV Community: cyw</title>
      <link>https://dev.to/cyw95</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/cyw95"/>
    <language>en</language>
    <item>
      <title>I built an open-source "Vercel for AI Agents" and could save AWS bill by 40%</title>
      <dc:creator>cyw</dc:creator>
      <pubDate>Tue, 29 Jul 2025 01:37:53 +0000</pubDate>
      <link>https://dev.to/cyw95/i-built-an-open-source-vercel-for-ai-agents-and-could-save-aws-bill-by-40-2bo1</link>
      <guid>https://dev.to/cyw95/i-built-an-open-source-vercel-for-ai-agents-and-could-save-aws-bill-by-40-2bo1</guid>
      <description>&lt;p&gt;&lt;strong&gt;TL;DR:&lt;/strong&gt;&lt;br&gt;
Like Vercel, but for stateful AI agents. Deploy your container and instantly get an agent with persistent memory, auto-recovery, and a live API endpoint — no infrastructure work required.&lt;/p&gt;



&lt;p&gt;Hey Dev.to 👋 &lt;/p&gt;

&lt;p&gt;I’m Cyw, I recently launched an open-source project called Agentainer — a local-first platform for deploying long-running AI agents with zero DevOps. You can check it out here:&lt;/p&gt;

&lt;p&gt;GitHub: &lt;a href="https://github.com/oso95/Agentainer-lab" rel="noopener noreferrer"&gt;Agentainer-Lab(GitHub)&lt;/a&gt;&lt;br&gt;
Website: &lt;a href="https://agentainer.io" rel="noopener noreferrer"&gt;Agentainer&lt;/a&gt;&lt;/p&gt;



&lt;p&gt;&lt;strong&gt;Why I built this&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Most infrastructure today is optimized for stateless, short-lived workloads — think Lambda, Cloud Run, or even Kubernetes pods. But when I started working with LLM agents, I realized these tools weren’t a good fit.&lt;/p&gt;

&lt;p&gt;Agents need to persist memory. They run for hours. They crash and need to pick up where they left off. Getting that to work in production often means duct-taping together:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Docker + EFS for persistent volumes&lt;/li&gt;
&lt;li&gt;Retry queues for failed requests&lt;/li&gt;
&lt;li&gt;Auto-restart logic&lt;/li&gt;
&lt;li&gt;Gateways for safe public access&lt;/li&gt;
&lt;li&gt;And probably some bash scripts you hope never break&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Honestly? I got tired of it. So I built Agentainer to simplify the stack.&lt;/p&gt;



&lt;p&gt;🔍 How It Compares&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;tr&gt;
&lt;th&gt;Feature&lt;/th&gt;
&lt;th&gt;Agentainer&lt;/th&gt;
&lt;th&gt;Raw Docker&lt;/th&gt;
&lt;th&gt;Kubernetes&lt;/th&gt;
&lt;th&gt;Serverless&lt;/th&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;b&gt;Deployment Speed&lt;/b&gt;&lt;/td&gt;
&lt;td&gt;✅ &amp;lt; 30 seconds&lt;/td&gt;
&lt;td&gt;⚠️ Manual setup&lt;/td&gt;
&lt;td&gt;❌ Complex YAML&lt;/td&gt;
&lt;td&gt;✅ Fast&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;b&gt;State Management&lt;/b&gt;&lt;/td&gt;
&lt;td&gt;✅ Built-in Redis&lt;/td&gt;
&lt;td&gt;❌ DIY&lt;/td&gt;
&lt;td&gt;⚠️ External&lt;/td&gt;
&lt;td&gt;❌ Stateless&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;b&gt;Request Persistence&lt;/b&gt;&lt;/td&gt;
&lt;td&gt;✅ Automatic&lt;/td&gt;
&lt;td&gt;❌ Not included&lt;/td&gt;
&lt;td&gt;❌ Not included&lt;/td&gt;
&lt;td&gt;❌ Lost on timeout&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;b&gt;Crash Recovery&lt;/b&gt;&lt;/td&gt;
&lt;td&gt;✅ With replay&lt;/td&gt;
&lt;td&gt;⚠️ Restart only&lt;/td&gt;
&lt;td&gt;⚠️ Restart only&lt;/td&gt;
&lt;td&gt;✅ Auto-retry&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;b&gt;Local Development&lt;/b&gt;&lt;/td&gt;
&lt;td&gt;✅ Optimized&lt;/td&gt;
&lt;td&gt;✅ Native&lt;/td&gt;
&lt;td&gt;❌ Heavy&lt;/td&gt;
&lt;td&gt;❌ Cloud only&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;b&gt;LLM-Specific&lt;/b&gt;&lt;/td&gt;
&lt;td&gt;✅ Purpose-built&lt;/td&gt;
&lt;td&gt;❌ Generic&lt;/td&gt;
&lt;td&gt;❌ Generic&lt;/td&gt;
&lt;td&gt;❌ Generic&lt;/td&gt;
&lt;/tr&gt;
&lt;/table&gt;&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;What it does&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Agentainer runs each agent in an isolated container with:&lt;/li&gt;
&lt;li&gt;Persistent memory (via volume mounts)&lt;/li&gt;
&lt;li&gt;Crash recovery (auto-restarts with state checkpointing)&lt;/li&gt;
&lt;li&gt;Request replay (if the agent dies mid-task, the request isn’t lost)&lt;/li&gt;
&lt;li&gt;Clean networking (each agent gets a dedicated API endpoint with proxying built-in)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All you do is:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;agentainer deploy &lt;span class="nt"&gt;--name&lt;/span&gt; my-agent &lt;span class="nt"&gt;--image&lt;/span&gt; my-docker-image
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;And your agent is live. No exposed ports, no infra glue, no YAML.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Built for agents (but could work for anything)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Originally, I made this so my own coding agent could deploy and manage infrastructure on its own — without wasting tokens repeating shell commands. Everything is CLI-first, and also accessible by API. So if you want your own agents to spin up services, restart themselves, or shut down gracefully — they can.&lt;/p&gt;

&lt;p&gt;And even though I built it for agents, technically you could use this for any long-running containerized app. It just happens to be extra helpful when you're dealing with memory/stateful logic.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;About the “save 40%” part&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I did some napkin math, and depending on your traffic and use case, replacing per-request billing with a flat-cost EC2 or VPS instance can easily save money. One dev I talked to migrated a chunk of their Lambda-based backend to Agentainer and saw their AWS bill drop significantly. YMMV, but the economics are real — especially if you're running:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;High-throughput agents&lt;/li&gt;
&lt;li&gt;Memory-heavy processes&lt;/li&gt;
&lt;li&gt;Repeated calls that don’t need instant scale-to-zero behavior&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Even for smaller projects, you might get better UX (no cold starts!) and spend less time wiring things up.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;What’s next?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Right now, Agentainer Lab is focused on local development and self-hosting. But the Agentainer.io roadmap includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Observability / logging&lt;/li&gt;
&lt;li&gt;Built-in backups and restore&lt;/li&gt;
&lt;li&gt;Audit trails&lt;/li&gt;
&lt;li&gt;Auto-scaling clusters for production loads&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you’re interested, you can sign up for early access on the site. We’ll send you one email when production is live — then delete your email from our database.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Try it out&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;GitHub&lt;/strong&gt;: &lt;a href="https://github.com/oso95/Agentainer-lab" rel="noopener noreferrer"&gt;github.com/oso95/Agentainer-lab&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Platform&lt;/strong&gt;: &lt;a href="https://agentainer.io" rel="noopener noreferrer"&gt;agentainer.io&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;Would love to hear your thoughts. Are you building agents? Running persistent workloads? Using too much YAML? Drop your current setup or ideas in the comments — I’m all ears!&lt;/p&gt;

</description>
      <category>aws</category>
      <category>ai</category>
      <category>devops</category>
      <category>webdev</category>
    </item>
  </channel>
</rss>
