<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: SashiDo.io</title>
    <description>The latest articles on DEV Community by SashiDo.io (@sashido).</description>
    <link>https://dev.to/sashido</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Forganization%2Fprofile_image%2F1038%2Fa1e270a0-8d0d-437a-9483-19e8ec4ae0ed.jpeg</url>
      <title>DEV Community: SashiDo.io</title>
      <link>https://dev.to/sashido</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/sashido"/>
    <language>en</language>
    <item>
      <title>AI Coding: Building a 1-Hour App Clone Is Easy. Shipping It Is the Work</title>
      <dc:creator>Pavel Ivanov</dc:creator>
      <pubDate>Thu, 19 Mar 2026 05:00:33 +0000</pubDate>
      <link>https://dev.to/sashido/ai-coding-building-a-1-hour-app-clone-is-easy-shipping-it-is-the-work-k55</link>
      <guid>https://dev.to/sashido/ai-coding-building-a-1-hour-app-clone-is-easy-shipping-it-is-the-work-k55</guid>
      <description>&lt;p&gt;A few months ago, “AI can help you code” meant autocomplete, snippets, and faster refactors. Now it can mean &lt;strong&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-with-ai-agents-a-non-developers-journey-to-shipping-an-mvp-on-sashido-part-2" rel="noopener noreferrer"&gt;a non-developer&lt;/a&gt; producing a believable SaaS clone in under an hour&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That’s not a thought experiment. CNBC documented reporters using Anthropic’s agent-style tooling to build a functional Monday-style &lt;a href="https://www.sashido.io/en/blog/ai-app-builder-vibe-coding-saas-backend-2025" rel="noopener noreferrer"&gt;project management&lt;/a&gt; app quickly and cheaply, then iterating by simply describing what they wanted next in plain English. Read the original experiment for context in CNBC’s write-up, &lt;a href="https://www.cnbc.com/2026/02/05/how-exposed-are-software-stocks-to-ai-tools-we-tested-vibe-coding.html" rel="noopener noreferrer"&gt;How exposed are software stocks to AI tools? We tested vibe-coding&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The important takeaway for builders is not “everything is doomed.” It’s more practical than that. &lt;strong&gt;AI coding collapses the time to a working UI and basic flows&lt;/strong&gt;, but it does not magically give you a production backend, an operating model, or a defensible product.&lt;/p&gt;

&lt;h2&gt;
  
  
  What The 60-Minute Clone Proves, and What It Hides
&lt;/h2&gt;

&lt;p&gt;The clone worked because modern SaaS categories like “project boards, assignments, statuses, comments, reminders” are often &lt;em&gt;repeatable patterns&lt;/em&gt;. If you can describe a Kanban board and a list view, an AI agent can generate a decent front end, wire up basic state, and mimic a familiar interaction model.&lt;/p&gt;

&lt;p&gt;What the demo hides is the part that usually hurts after launch. The first 10 users are impressed by the UI. The next 100 ask for permissions, &lt;a href="https://www.sashido.io/en/blog/vibe-coding-to-production-backend-reality-check" rel="noopener noreferrer"&gt;audit logs&lt;/a&gt;, import/export, and mobile notifications. The next 1,000 hit &lt;a href="https://www.sashido.io/en/blog/ai-assisted-coding-vibe-projects-2026" rel="noopener noreferrer"&gt;edge cases&lt;/a&gt; like duplicate records, stale realtime updates, race conditions, and “why did my task disappear?” The next 10,000 bring cost, scaling, and security questions you cannot vibe-code away.&lt;/p&gt;

&lt;p&gt;In other words, &lt;strong&gt;a clone is a screenshot that moves&lt;/strong&gt;. A product is a system that survives reality: unreliable networks, messy data, malicious input, and “it worked yesterday.”&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Connect your AI-built UI to a managed backend in minutes. Read our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; to add &lt;a href="https://www.sashido.io/en/blog/ai-powered-backend-mobile-app-development-speed" rel="noopener noreferrer"&gt;auth, database&lt;/a&gt;, and push without backend ops.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Real Checklist Behind Vibe-Coded Apps
&lt;/h2&gt;

&lt;p&gt;If you’re a solo founder or an AI-first builder, the fastest path is usually to let AI generate the interface and basic flows, then lock in the backend fundamentals early, before you accumulate production debt.&lt;/p&gt;

&lt;p&gt;Here’s the checklist we see repeatedly when “the prototype is getting real” happens.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Modeling That Doesn’t Collapse Under Change
&lt;/h3&gt;

&lt;p&gt;Early prototypes store everything in a single table or document because it’s convenient. The first time you add teams, roles, or multi-board views, your data model needs to support relationship-like queries, indexing, and constraints.&lt;/p&gt;

&lt;p&gt;A practical smell test is this. If you’re already asking the AI to “add workspace support” or “make boards shareable,” you are crossing from a toy model into a real one.&lt;/p&gt;

&lt;h3&gt;
  
  
  Authentication and Account Linking
&lt;/h3&gt;

&lt;p&gt;AI coding tools can generate a login page. The hard part is: password resets, session handling, rate limiting, social login linking, and handling the “I signed up with Google but now I want email login” situation.&lt;/p&gt;

&lt;p&gt;This is also where most indie projects quietly leak users. If your auth breaks once, people stop trusting the app.&lt;/p&gt;

&lt;h3&gt;
  
  
  Authorization and Permissions, Not Just Users
&lt;/h3&gt;

&lt;p&gt;Most clones have “a user.” Real apps have “a user in a workspace with a role, on a board, with a permission.” You need object-level access control, not just a boolean admin flag.&lt;/p&gt;

&lt;p&gt;If you want a north star, look at the industry’s constant stream of permission bugs. OWASP keeps a living list of what attackers exploit most often in APIs, and it’s worth skimming even if you don’t consider yourself a security person. Start with &lt;a href="https://owasp.org/API-Security/" rel="noopener noreferrer"&gt;OWASP API Security Top 10&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Background Jobs and Long-Running Work
&lt;/h3&gt;

&lt;p&gt;The CNBC experiment’s moment of “connect email, then it becomes a project manager” points at a common next step. You add automation. That means jobs that run every hour, sync external data, send reminders, retry on failure, and persist state.&lt;/p&gt;

&lt;p&gt;AI agents for coding can generate the scheduler code, but production needs visibility, failure handling, and a place to run those jobs consistently.&lt;/p&gt;

&lt;h3&gt;
  
  
  Realtime Without Realtime Headaches
&lt;/h3&gt;

&lt;p&gt;Boards and statuses feel much better when multiple clients stay in sync. But realtime is also where apps get subtle bugs, especially with reconnect logic and partial failures.&lt;/p&gt;

&lt;p&gt;A good principle is: if realtime is core to the UX, treat it as infrastructure, not a feature.&lt;/p&gt;

&lt;h3&gt;
  
  
  File Storage and Delivery
&lt;/h3&gt;

&lt;p&gt;The moment users attach files, screenshots, or exports, you need scalable object storage and a CDN. You also need to handle access control, signed URLs, and data transfer costs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Push Notifications That Don’t Break Trust
&lt;/h3&gt;

&lt;p&gt;Push is easy to “add,” and hard to do well. If you send irrelevant notifications, users disable them. If you miss critical ones, teams stop relying on you.&lt;/p&gt;

&lt;p&gt;Even if you are not building a mobile-first product, push becomes relevant quickly for reminders, approvals, and mentions.&lt;/p&gt;

&lt;h2&gt;
  
  
  Best AI Tools for Coding: What Each One Is Actually Good At
&lt;/h2&gt;

&lt;p&gt;“Best” in AI coding depends on whether you need autocomplete, a coding agent, a UI builder, or a debugging partner. The CNBC-style outcome typically requires &lt;strong&gt;agentic behavior&lt;/strong&gt;, not just code completion.&lt;/p&gt;

&lt;p&gt;Below is a practical, builder-oriented comparison. It’s intentionally framed around &lt;em&gt;work you will do&lt;/em&gt; in the first month of shipping.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool Category&lt;/th&gt;
&lt;th&gt;What It’s Best At&lt;/th&gt;
&lt;th&gt;Where It Often Fails&lt;/th&gt;
&lt;th&gt;When It Fits&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Agentic IDE workflows (e.g., Claude Code-style)&lt;/td&gt;
&lt;td&gt;End-to-end feature implementation from a prompt, multi-file edits, fast prototyping loops&lt;/td&gt;
&lt;td&gt;Can create inconsistent architecture, “works on my machine” integration gaps, shallow testing and security assumptions&lt;/td&gt;
&lt;td&gt;When you need to go from blank screen to usable flows fast&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;In-IDE assistants (e.g., Copilot-style)&lt;/td&gt;
&lt;td&gt;Fast iteration inside existing codebases, autocomplete, refactors, test scaffolding&lt;/td&gt;
&lt;td&gt;Less effective for greenfield architecture decisions and cross-cutting product flows&lt;/td&gt;
&lt;td&gt;When you already have a backend and want speed without losing control&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;UI-first vibe builders&lt;/td&gt;
&lt;td&gt;Rapid UI and routing, fast demo creation&lt;/td&gt;
&lt;td&gt;Backend and data model often become a rewrite&lt;/td&gt;
&lt;td&gt;When you must show something to users or investors in hours&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Specialized code search and analysis&lt;/td&gt;
&lt;td&gt;Understanding unfamiliar repos and dependencies&lt;/td&gt;
&lt;td&gt;Still requires engineering judgment&lt;/td&gt;
&lt;td&gt;When the app has grown and you need to debug faster&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;If you want to understand what “agentic” means in practice, Anthropic’s own documentation is a good reference point. Start with &lt;a href="https://docs.anthropic.com/en/docs/claude-code/overview" rel="noopener noreferrer"&gt;Claude Code Overview&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  GitHub Copilot vs Claude Code: The Practical Difference
&lt;/h3&gt;

&lt;p&gt;If you’re comparing &lt;strong&gt;GitHub Copilot vs Claude Code&lt;/strong&gt;, the best framing is “assistant” vs “agent.” Copilot shines when you already know what you’re building and want speed inside your editor. Claude Code-style agents shine when you want the tool to plan, edit many files, and drive implementation from high-level intent. The trade-off is oversight. Agents can drift.&lt;/p&gt;

&lt;p&gt;For official product context, see &lt;a href="https://github.com/features/copilot" rel="noopener noreferrer"&gt;GitHub Copilot&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  AI Models for Coding vs AI Agents for Coding
&lt;/h3&gt;

&lt;p&gt;AI models for coding are the engines. They predict code, explain code, and generate solutions. AI agents for coding wrap those models with tools. They can read files, search the repo, run tasks, and iterate.&lt;/p&gt;

&lt;p&gt;The second category is what makes “clone in an hour” feasible, but it also increases the chance that you ship something that &lt;em&gt;looks&lt;/em&gt; correct while hiding runtime and operational flaws.&lt;/p&gt;

&lt;h2&gt;
  
  
  Pros and Cons of AI Coding for Shipping Real Products
&lt;/h2&gt;

&lt;p&gt;AI coding is not a binary. It’s a slider. You can use AI to generate 20% of a system or 80% of it. The right percentage depends on where correctness matters.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Upside: Speed, Iteration, and Option Value
&lt;/h3&gt;

&lt;p&gt;The best part is obvious. &lt;strong&gt;You can test more ideas with less sunk cost&lt;/strong&gt;. Instead of spending two weeks building the first version of boards and tasks, you can build it today and spend the next two weeks learning what users actually want.&lt;/p&gt;

&lt;p&gt;That “option value” also changes how teams think about SaaS. If a category’s core value is mostly UI plus CRUD plus workflow, then AI can lower the cost to compete.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Downside: Hidden Debt Shows Up Right After Validation
&lt;/h3&gt;

&lt;p&gt;AI-produced code often has three recurring issues.&lt;/p&gt;

&lt;p&gt;First, it’s not designed for change. A demo can tolerate duplication. A product cannot.&lt;/p&gt;

&lt;p&gt;Second, it underestimates security. The typical failure mode is not “hackers are geniuses,” it’s that APIs accidentally allow actions that should have been forbidden.&lt;/p&gt;

&lt;p&gt;Third, it lacks an operating model. When the app slows down, who investigates. When a job fails, who retries. When a schema changes, what breaks.&lt;/p&gt;

&lt;p&gt;A simple threshold that catches many teams is this. When you pass &lt;strong&gt;500 to 1,000 real users&lt;/strong&gt; or you start seeing &lt;strong&gt;thousands of requests per hour&lt;/strong&gt;, you stop being able to “just redeploy” when something breaks. You need monitoring, predictable scaling, and a data model you trust.&lt;/p&gt;

&lt;h3&gt;
  
  
  A Quick Reality Check on Costs
&lt;/h3&gt;

&lt;p&gt;The CNBC demo quoted a small compute cost for iteration. That is real and powerful. But it’s only one line item.&lt;/p&gt;

&lt;p&gt;Once you ship, you pay for database queries, file storage, data transfer, background work, and the time you spend on operational fixes. That’s why the “clone cost” is not the “product cost.” The product cost includes &lt;em&gt;maintenance, reliability, and support&lt;/em&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where a Managed Backend Fits When AI Builds Your Front End
&lt;/h2&gt;

&lt;p&gt;The pattern we see with indie teams is consistent. AI coding gets the UI and flows to “convincing.” The next bottleneck is the backend. Not because backends are glamorous, but because they are where reliability lives.&lt;/p&gt;

&lt;p&gt;That’s exactly why we built &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;. The goal is simple. You should be able to connect your AI-built UI to a production backend that already has the boring essentials handled.&lt;/p&gt;

&lt;p&gt;In practice, that means a MongoDB database with a CRUD API, built-in user management with social logins, file storage backed by AWS S3 with a CDN, realtime over WebSockets, serverless functions close to your users, scheduled jobs, and mobile push notifications. It also means the unglamorous parts. Platform monitoring, SSL, and a dashboard to operate the app.&lt;/p&gt;

&lt;p&gt;If you’re coming from the Parse ecosystem, it may help to know that Parse itself is a long-running open source backend framework. You can start from the official &lt;a href="https://website.parseplatform.org/" rel="noopener noreferrer"&gt;Parse Platform&lt;/a&gt; site, or go deeper with the community’s &lt;a href="https://github.com/parse-community/parse-server" rel="noopener noreferrer"&gt;Parse Server repository&lt;/a&gt;. Our own developer docs are organized around that reality. If you want implementation-level guides, start with our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;SashiDo Documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;When cost comes up, we recommend treating pricing as configuration, not lore. We offer a 10-day free trial with no credit card required, and we keep the current plan details on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Comparing Backend Paths for AI-First Builders
&lt;/h2&gt;

&lt;p&gt;Once AI coding makes prototypes cheap, the key decision becomes. What backend path keeps you shipping without rewriting everything after validation.&lt;/p&gt;

&lt;p&gt;Here’s a decision-oriented comparison of common paths we see in the “vibe-coder to production” transition.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Backend Path&lt;/th&gt;
&lt;th&gt;Why People Choose It&lt;/th&gt;
&lt;th&gt;What Usually Bites Later&lt;/th&gt;
&lt;th&gt;Best For&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Managed Parse hosting (SashiDo-style)&lt;/td&gt;
&lt;td&gt;Fast setup, API-first, auth, realtime, jobs, push, storage already integrated&lt;/td&gt;
&lt;td&gt;You still need to design a clean data model and permissions. Managed does not mean no decisions&lt;/td&gt;
&lt;td&gt;Solo founders and small teams that want to ship fast without backend ops&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;DIY backend on your own infra&lt;/td&gt;
&lt;td&gt;Maximum control&lt;/td&gt;
&lt;td&gt;Time sink. Security and maintenance become your job immediately&lt;/td&gt;
&lt;td&gt;Teams with strong backend skills and compliance requirements&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Postgres-first BaaS&lt;/td&gt;
&lt;td&gt;Familiar SQL ecosystem&lt;/td&gt;
&lt;td&gt;Migrations, RLS complexity, and scaling patterns can be non-trivial for first-time backend builders&lt;/td&gt;
&lt;td&gt;Builders who already think in SQL and want tight relational constraints&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;GraphQL-first orchestration&lt;/td&gt;
&lt;td&gt;Clean API surface across services&lt;/td&gt;
&lt;td&gt;Complexity shifts to schema governance, permissions, and performance&lt;/td&gt;
&lt;td&gt;Teams already running multiple services and needing a unified graph&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cloud suite assembly&lt;/td&gt;
&lt;td&gt;Wide service catalog&lt;/td&gt;
&lt;td&gt;Configuration sprawl and surprise bills if you assemble without guardrails&lt;/td&gt;
&lt;td&gt;Teams already deep in a cloud ecosystem&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;If you are actively comparing options, we’ve published direct comparisons that focus on day-to-day builder trade-offs. See &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt;, &lt;a href="https://www.sashido.io/en/sashido-vs-hasura" rel="noopener noreferrer"&gt;SashiDo vs Hasura&lt;/a&gt;, and &lt;a href="https://www.sashido.io/en/sashido-vs-aws-amplify" rel="noopener noreferrer"&gt;SashiDo vs AWS Amplify&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Features to Look For When AI Coding Is Your Front End
&lt;/h2&gt;

&lt;p&gt;This is the “don’t get surprised in month two” list. Use it as a filter when evaluating any backend approach.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Auth plus account lifecycle&lt;/strong&gt;: social login, password reset, session management, and rate limiting.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Granular permissions&lt;/strong&gt;: object-level access control and a way to reason about multi-tenant data.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Storage plus delivery&lt;/strong&gt;: object storage, CDN behavior, and predictable data transfer pricing.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Realtime&lt;/strong&gt;: simple subscriptions, reconnect behavior, and monitoring.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Jobs and functions&lt;/strong&gt;: a place for automation, webhooks, scheduled tasks, and retries.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Observability&lt;/strong&gt;: logs, metrics, alerts, and a dashboard that makes incidents debuggable.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you want a deeper dive into scaling patterns, our write-up on engines is useful because it explains what “scale up” actually means operationally, not just conceptually. See &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Power Up with Our Engine Feature&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Decision Framework You Can Use This Weekend
&lt;/h2&gt;

&lt;p&gt;Most builders do not fail because their first prototype is bad. They fail because they cannot transition from “prototype energy” to “product reliability” without losing momentum.&lt;/p&gt;

&lt;p&gt;A simple way to decide is to pick the lane you are in right now.&lt;/p&gt;

&lt;p&gt;If you’re pre-validation, optimize for learning speed. Let AI coding generate UI, and make sure your backend choice keeps you flexible. Avoid anything that locks you into a rewrite after the first real feedback.&lt;/p&gt;

&lt;p&gt;If you’re post-validation, optimize for correctness and operations. The fastest teams at this stage are not the ones that generate the most code. They’re the ones that reduce the number of unknowns. They add permissions, monitoring, and a data model that can evolve without breaking.&lt;/p&gt;

&lt;p&gt;If you’re building a “system of record” style product, be honest about the bar. These apps live and die on integrity, auditability, and integrations. AI will still help, but you should expect deeper engineering work and tighter security practices.&lt;/p&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  How difficult is AI coding?
&lt;/h3&gt;

&lt;p&gt;AI coding is easiest when the task is a known pattern, like CRUD screens, dashboards, and common integrations. It gets harder when you need correct permissions, clean data modeling, and reliable background automation. The skill is less about typing code and more about specifying behavior, reviewing output, and recognizing hidden failure modes early.&lt;/p&gt;

&lt;h3&gt;
  
  
  What is the best coder for AI?
&lt;/h3&gt;

&lt;p&gt;The best coder for AI is usually a developer who can translate vague product intent into precise constraints, then review and shape what the agent generates. In practice, that means someone comfortable with debugging, reading unfamiliar code, and making architecture trade-offs. AI speeds up implementation, but it does not replace judgment about security, data, and operations.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Should I Build With AI Coding First, Front End or Back End?
&lt;/h3&gt;

&lt;p&gt;Start with the front end and core flows if your goal is learning quickly, because users react fastest to UX. But lock in backend fundamentals early, especially auth, permissions, and data modeling, because those are expensive to change later. Many teams succeed by generating UI with AI and using a managed backend for production reliability.&lt;/p&gt;

&lt;h3&gt;
  
  
  When Do I Need a Real Backend Instead of a Local Prototype?
&lt;/h3&gt;

&lt;p&gt;A good trigger is when you have real users, real data, and the expectation that the app will work every day. Once you need multi-user access, permissions, file uploads, scheduled reminders, or realtime collaboration, you need a backend that can handle security, scaling, and monitoring. That shift often happens sooner than expected, sometimes within the first week of sharing a demo.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.cnbc.com/2026/02/05/how-exposed-are-software-stocks-to-ai-tools-we-tested-vibe-coding.html" rel="noopener noreferrer"&gt;How exposed are software stocks to AI tools? We tested vibe-coding (CNBC)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.anthropic.com/en/docs/claude-code/overview" rel="noopener noreferrer"&gt;Claude Code Overview (Anthropic Docs)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://github.com/features/copilot" rel="noopener noreferrer"&gt;GitHub Copilot (Official Product Page)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://owasp.org/API-Security/" rel="noopener noreferrer"&gt;OWASP API Security Top 10&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://website.parseplatform.org/" rel="noopener noreferrer"&gt;Parse Platform (Official Site)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion: AI Coding Gets You a Clone. Shipping Gets You a Business.
&lt;/h2&gt;

&lt;p&gt;The CNBC experiment is the clearest public demonstration yet of what AI coding changes. It makes “first working version” radically cheaper. That’s great news for builders and terrifying news for SaaS categories that relied on slow development as a moat.&lt;/p&gt;

&lt;p&gt;But once your demo becomes a product, the differentiator shifts to the stuff users only notice when it’s missing. Authentication that never breaks, permissions that never leak, jobs that retry, realtime that stays consistent, storage that delivers fast, and monitoring that tells you what went wrong before your users do.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;When you’re ready to move beyond demos, &lt;strong&gt;pick &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; to deploy a production-ready backend fast&lt;/strong&gt;. Start your 10-day free trial to get MongoDB, auth, realtime, serverless functions, and unlimited push notifications, all monitored 24/7.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/cursor-coding-remove-next-bottleneck" rel="noopener noreferrer"&gt;Cursor Coding Turns Output Up. Here’s How to Remove the Next Bottleneck&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/artificial-intelligence-coding-90-10-rule-build-vs-buy" rel="noopener noreferrer"&gt;Artificial Intelligence Coding and the 90/10 Rule: Build vs Buy&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-ai-ready-backends" rel="noopener noreferrer"&gt;Vibe Coding and AI-Ready Backends for Rapid Prototypes&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-assisted-coding-vibe-projects-2026" rel="noopener noreferrer"&gt;AI Assisted Coding in 2026: Vibe Projects You Can Monetize&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/creating-an-app-weekend-builds-take-weeks" rel="noopener noreferrer"&gt;Creating an App in a Weekend? The 47,000-Line Reality&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>programming</category>
      <category>devops</category>
    </item>
    <item>
      <title>AI Assisted Programming That Actually Ships: The Long-Running Agent Harness</title>
      <dc:creator>Marian Ignev</dc:creator>
      <pubDate>Wed, 18 Mar 2026 12:47:25 +0000</pubDate>
      <link>https://dev.to/sashido/ai-assisted-programming-that-actually-ships-the-long-running-agent-harness-43pl</link>
      <guid>https://dev.to/sashido/ai-assisted-programming-that-actually-ships-the-long-running-agent-harness-43pl</guid>
      <description>&lt;p&gt;If you are using &lt;strong&gt;ai assisted programming&lt;/strong&gt; for anything bigger than a quick script, you have probably seen the same pattern: the agent starts strong, then a few sessions later it forgets what mattered, rewrites working code, marks things “done” without testing, or leaves you with a half-finished feature and no breadcrumbs.&lt;/p&gt;

&lt;p&gt;That is not a model problem as much as it is a harness problem. Long-running work is inherently shift work. Each new session begins with partial context, and most real projects cannot fit into a single window. The fix is to stop treating your agent like a one-shot &lt;em&gt;best ai code generator&lt;/em&gt; and start treating it like an engineer joining a codebase mid-sprint, with a clear backlog, a reproducible environment, and a definition of done.&lt;/p&gt;

&lt;p&gt;In our experience building infrastructure for teams that ship fast, the most reliable setup is a &lt;a href="https://www.sashido.io/en/blog/coding-agents-best-practices-plan-test-ship-faster" rel="noopener noreferrer"&gt;two-part harness&lt;/a&gt;: an initializer run that prepares the repo for many sessions, and a repeatable coding loop that makes incremental progress while leaving clean artifacts.&lt;/p&gt;

&lt;p&gt;To make this concrete, we will also map the harness to a real backend so you can prototype end-to-end without babysitting infrastructure. That is where &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; fits naturally, because we give you Database, APIs, Auth, Storage, Realtime, Jobs, Functions, and Push in minutes, which lets the agent focus on product work instead of DevOps.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why long-running agents fail in the real world
&lt;/h2&gt;

&lt;p&gt;Most “ai for coding” workflows break down in two predictable ways.&lt;/p&gt;

&lt;p&gt;First, the agent tries to do too much in one pass. It will start implementing multiple features, change shared abstractions, and then run out of context mid-flight. The next session wakes up, sees an inconsistent codebase, and spends most of its budget re-deriving what happened instead of moving forward. The worst part is that you often do not notice the damage until later because the work fails silently.&lt;/p&gt;

&lt;p&gt;Second, once a project has some visible progress, the agent starts declaring victory. It sees a UI, some endpoints, a few tests. Then it assumes you are done, even if edge cases, auth flows, billing, or background jobs are missing. This is especially common with “programming ai” setups that do not have an explicit, testable feature inventory.&lt;/p&gt;

&lt;p&gt;The throughline is simple: &lt;strong&gt;without a &lt;a href="https://www.sashido.io/en/blog/caveat-coder-ai-infrastructure-importance" rel="noopener noreferrer"&gt;stable shared memory&lt;/a&gt; and a stable definition of done, each session is forced to guess&lt;/strong&gt;. Your harness needs to remove guessing.&lt;/p&gt;

&lt;h2&gt;
  
  
  The harness pattern: initializer session plus incremental coding sessions
&lt;/h2&gt;

&lt;p&gt;A practical long-horizon setup splits responsibilities.&lt;/p&gt;

&lt;p&gt;The initializer session exists once, at the very beginning. Its job is not to build features. Its job is to create a working environment and durable artifacts that every future session can trust. Think of it as setting up the project the way a senior engineer would. You want a runnable dev environment, a clear feature list, a place to log progress, and a repo state you can always roll back to.&lt;/p&gt;

&lt;p&gt;Every subsequent session is a coding session. Its job is narrow: pick one feature, implement it, prove it works, record what changed, and commit the result. If you enforce this rhythm, you fix both failure modes at once. You prevent the agent from one-shotting the whole app, and you prevent it from “calling it done” based on vibes.&lt;/p&gt;

&lt;p&gt;If you are building with “ai dev tools” that can automate terminal and browser actions, this pattern gets even stronger because you can require &lt;a href="https://www.sashido.io/en/blog/coding-agents-best-practices-plan-test-ship-faster" rel="noopener noreferrer"&gt;end-to-end checks&lt;/a&gt;, not just unit tests.&lt;/p&gt;

&lt;h2&gt;
  
  
  The three artifacts that make sessions resumable
&lt;/h2&gt;

&lt;p&gt;Long-running work succeeds when a fresh session can answer three questions quickly: What is the goal. What is the current state. What should I do next.&lt;/p&gt;

&lt;p&gt;We rely on three lightweight artifacts to answer those questions with minimal token waste.&lt;/p&gt;

&lt;h3&gt;
  
  
  1) A feature list that the agent cannot hand-wave away
&lt;/h3&gt;

&lt;p&gt;The feature list is your guardrail against premature “done.” It should be structured, test-oriented, and easy to update without rewriting history.&lt;/p&gt;

&lt;p&gt;A practical approach is a JSON file (many teams name it feature_list.json) where each feature entry includes fields like category, description, user-visible steps, and a boolean such as passes set to false by default. The key is that &lt;strong&gt;coding sessions are only allowed to flip passes from false to true&lt;/strong&gt; after verification. They do not rewrite the description or delete items just because implementation is hard.&lt;/p&gt;

&lt;p&gt;This is also the point where you force the agent to stop being a &lt;em&gt;code writer ai&lt;/em&gt; and become a product engineer. The description and steps should read like what a human would do in the app, not like internal implementation notes.&lt;/p&gt;

&lt;h3&gt;
  
  
  2) A progress file that summarizes “what happened” across sessions
&lt;/h3&gt;

&lt;p&gt;A short, append-only progress log (often a plain text file such as claude-progress.txt) is the fastest way to rehydrate context. It should capture what feature was attempted, what files changed, how it was tested, what is still broken, and what the next session should do first.&lt;/p&gt;

&lt;p&gt;Keep it boring. You want a new session to skim 20 lines and immediately know where to start. &lt;strong&gt;The &lt;a href="https://www.sashido.io/en/blog/ai-coding-tools-dynamic-context-discovery" rel="noopener noreferrer"&gt;progress file&lt;/a&gt; is not documentation. It is shift notes.&lt;/strong&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  3) An init script that makes the environment reproducible
&lt;/h3&gt;

&lt;p&gt;Your init script (commonly init.sh) is the antidote to “it works on my machine.” It should do the minimum to start the project, run migrations or seed data, and kick off a basic smoke test.&lt;/p&gt;

&lt;p&gt;If you follow the spirit of the &lt;a href="https://www.12factor.net/" rel="noopener noreferrer"&gt;Twelve-Factor App&lt;/a&gt; approach, the script should rely on environment configuration and keep the app process model simple. That makes it easier for an agent to run, and easier for you to run in CI later.&lt;/p&gt;

&lt;p&gt;This is also where a real backend platform helps. When the backend is already provisioned, the init path is short. You are not asking the agent to install, configure, and secure a database server. You are asking it to connect to a backend that already exists.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI assisted programming across context windows: the incremental loop
&lt;/h2&gt;

&lt;p&gt;Once the initializer artifacts exist, every coding session should follow the same loop. The loop is intentionally repetitive because repetition is what makes sessions resumable.&lt;/p&gt;

&lt;p&gt;Start by grounding yourself in the repo state. Read the progress log, scan recent commits, and confirm the feature list still matches the product goal. Then run init.sh and execute a smoke test before touching code. This catches “broken baseline” problems early, before you pile new changes on top.&lt;/p&gt;

&lt;p&gt;Only then do you pick a single failing feature. Implement it in the smallest change set you can. Test it like a user. Update the feature list to mark it passing. Add a concise progress note. Commit.&lt;/p&gt;

&lt;p&gt;That last part matters. &lt;strong&gt;A git commit is not just version control. It is your rollback and your memory boundary.&lt;/strong&gt; The official &lt;a href="https://git-scm.com/docs/git-commit" rel="noopener noreferrer"&gt;git commit documentation&lt;/a&gt; is not glamorous, but the discipline it enables is exactly what long-running agent work needs. If a session goes off the rails, you can reset to the last known good commit without debate.&lt;/p&gt;

&lt;h3&gt;
  
  
  A small session checklist you can reuse
&lt;/h3&gt;

&lt;p&gt;Keep this near your prompt template and near your repo README. Short, boring, consistent.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Confirm you are in the expected directory and repo.&lt;/li&gt;
&lt;li&gt;Read the progress log, then read the last 10 to 20 commits.&lt;/li&gt;
&lt;li&gt;Run init.sh, then run the smoke test before making changes.&lt;/li&gt;
&lt;li&gt;Choose exactly one feature whose passes flag is false.&lt;/li&gt;
&lt;li&gt;Implement the change, then test end-to-end.&lt;/li&gt;
&lt;li&gt;Update passes to true only after testing, write a progress note, and commit with a descriptive message.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is the part most “ai that can code” demos skip. They show generation. They do not show continuity.&lt;/p&gt;

&lt;h2&gt;
  
  
  Testing: stop trusting green unit tests when the user flow is broken
&lt;/h2&gt;

&lt;p&gt;Long-running agents have a predictable testing failure mode: they “verify” by running a linter, maybe a unit test, maybe a single API call, and then they claim completion. That is better than nothing, but it does not tell you if the feature works end-to-end.&lt;/p&gt;

&lt;p&gt;For web apps, browser-driven testing is the fastest way to keep agents honest. If your harness can drive a browser, require it. Tools like &lt;a href="https://playwright.dev/docs/intro" rel="noopener noreferrer"&gt;Playwright&lt;/a&gt; make it straightforward to automate real user flows, including login, navigation, and form submission. You do not need a huge test suite. You need a reliable smoke test that proves the baseline works, plus a focused scenario test for the feature you just implemented.&lt;/p&gt;

&lt;p&gt;When you connect this to your feature list, you get a powerful loop: the agent is not allowed to mark passes true until it can execute the corresponding steps in an automated or at least reproducible way.&lt;/p&gt;

&lt;p&gt;There is a trade-off. Browser automation can be flaky, and vision limitations can hide UI problems. But in practice, it is still a net win because it catches regressions that a unit test never sees, like broken routing, missing auth headers, or a UI that never renders due to a runtime error.&lt;/p&gt;

&lt;h2&gt;
  
  
  Recovery and safety: treat each session like a production change
&lt;/h2&gt;

&lt;p&gt;Long-running agent loops can create subtle security and reliability issues because the agent is effectively committing code repeatedly. The guardrail is to adopt a lightweight secure development posture early.&lt;/p&gt;

&lt;p&gt;A pragmatic reference is NIST’s &lt;a href="https://csrc.nist.gov/pubs/sp/800/218/final" rel="noopener noreferrer"&gt;Secure Software Development Framework (SSDF) SP 800-218&lt;/a&gt;. You do not need heavyweight compliance for a prototype, but the SSDF mindset translates cleanly into agent harness rules: make changes traceable, verify before release, and reduce the blast radius of mistakes.&lt;/p&gt;

&lt;p&gt;In practice, that means keeping secrets out of the repo, storing them in environment variables, reviewing dependency changes, and ensuring the agent’s tests cover the flows that matter. It also means you should never let the agent “refactor everything” as a side quest. If you want a refactor, create a feature item for it and make it pass like everything else.&lt;/p&gt;

&lt;h2&gt;
  
  
  Mapping the harness to a real backend you can ship on
&lt;/h2&gt;

&lt;p&gt;A harness is only half the battle. The other half is having real infrastructure to test against, without spending your limited founder time on setup and ops.&lt;/p&gt;

&lt;p&gt;When we see solo founders and indie hackers attempt multi-session builds, the backend is usually where momentum dies. Auth gets bolted on late. Database migrations drift. File storage is hacked in via local folders. Push notifications are postponed indefinitely. Then the agent spends sessions trying to duct-tape infrastructure rather than shipping user-visible value.&lt;/p&gt;

&lt;p&gt;This is why we built &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;. For long-running agent work, the platform acts like a stable external system the agent can target consistently across sessions.&lt;/p&gt;

&lt;p&gt;You can set up a project so your feature list includes backend-backed items from day one, like signup with social login, profile CRUD, file upload with CDN delivery, or realtime presence. In SashiDo, those map cleanly to things we provision by default: MongoDB with CRUD APIs, built-in User Management with social providers, S3-backed file storage with a CDN layer, Realtime over WebSockets, serverless Functions, scheduled Jobs, and push notifications.&lt;/p&gt;

&lt;p&gt;The harness benefit is subtle but huge: your init script can consistently start the frontend and point it at the same backend app, and your tests can validate flows that are actually production-shaped.&lt;/p&gt;

&lt;p&gt;If you want the quickest path, start with our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;SashiDo Docs&lt;/a&gt; and follow the flow in our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt;. This keeps the “backend exists” part out of your agent’s context window, so it can spend tokens on the product.&lt;/p&gt;

&lt;h3&gt;
  
  
  A concrete way to structure your first few features
&lt;/h3&gt;

&lt;p&gt;Instead of letting the agent invent architecture, tie features to backend capabilities you already have.&lt;/p&gt;

&lt;p&gt;You might start with a thin vertical slice: user signup and login, a single main data object stored in the database, and a UI that lists and edits it. MongoDB’s data model and CRUD operations are well understood, and the &lt;a href="https://www.mongodb.com/docs/manual/crud/" rel="noopener noreferrer"&gt;MongoDB CRUD concepts&lt;/a&gt; are a good canonical reference when you are sanity-checking queries and updates.&lt;/p&gt;

&lt;p&gt;Then expand into “real app” capabilities that often get deferred: file uploads for user content, realtime updates for collaboration, and scheduled jobs for cleanup or recurring work. If you hit scaling limits, you can later scale compute using our Engines. Our write-up on &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;SashiDo Engines&lt;/a&gt; explains when to move past the default and how the cost model works.&lt;/p&gt;

&lt;p&gt;On pricing, keep your harness honest by linking to the live source of truth. We offer a 10-day free trial with no credit card, and our current plans are always listed on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where this compares to other backends
&lt;/h3&gt;

&lt;p&gt;If you are evaluating alternatives while building your harness, keep the comparison grounded in session continuity. Does the backend reduce setup steps in init.sh. Does it give the agent stable APIs and auth flows to test against.&lt;/p&gt;

&lt;p&gt;If you are weighing us against Firebase or Supabase, we maintain direct comparisons that focus on practical trade-offs: &lt;a href="https://www.sashido.io/en/sashido-vs-firebase" rel="noopener noreferrer"&gt;SashiDo vs Firebase&lt;/a&gt; and &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Trade-offs and when to add more specialized agents
&lt;/h2&gt;

&lt;p&gt;The initializer plus coding loop works well because it is simple. But there are limits.&lt;/p&gt;

&lt;p&gt;If your feature list gets large, you may want a separate “triage” step that periodically reorders priorities, merges duplicates, and clarifies acceptance criteria. Do that intentionally and rarely. Otherwise, you will churn the file more than you ship.&lt;/p&gt;

&lt;p&gt;If testing becomes complex, a dedicated testing pass can help. But do not turn this into a multi-agent architecture because it sounds cool. Do it because your bottleneck is verification, not implementation.&lt;/p&gt;

&lt;p&gt;And if you start seeing repeated regressions, tighten your definition of clean state. Make the smoke test mandatory. Require that each session leaves the app runnable. Require that the progress note includes how to reproduce and how to verify.&lt;/p&gt;

&lt;p&gt;The goal is not to create bureaucracy. The goal is to make &lt;a href="https://www.sashido.io/en/blog/vibe-coding-experience-ai-tools" rel="noopener noreferrer"&gt;&lt;strong&gt;ai assisted programming&lt;/strong&gt;&lt;/a&gt; behave like a reliable teammate who can pick up work tomorrow without re-learning yesterday.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: make your agent boring, and your velocity will get exciting
&lt;/h2&gt;

&lt;p&gt;The biggest unlock in long-running agent work is not smarter prompts. It is a harness that forces continuity.&lt;/p&gt;

&lt;p&gt;When you combine a one-time initializer session with durable artifacts, then enforce an incremental loop with testing and commits, your “ai for coding” workflow stops feeling like gambling. You stop losing sessions to confusion. You stop accumulating half-done work. You gain a predictable rhythm where every session either ships a feature or leaves a clear note about why it could not.&lt;/p&gt;

&lt;p&gt;If you want to apply this pattern to something users can actually touch, anchor it to a real backend early. That is where &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; helps. You get the database, auth, functions, jobs, storage, realtime, and push capabilities up front, which gives your harness a stable target and makes end-to-end tests meaningful.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and further reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.12factor.net/" rel="noopener noreferrer"&gt;The Twelve-Factor App&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://git-scm.com/docs/git-commit" rel="noopener noreferrer"&gt;Git commit documentation&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://playwright.dev/docs/intro" rel="noopener noreferrer"&gt;Playwright documentation&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://csrc.nist.gov/pubs/sp/800/218/final" rel="noopener noreferrer"&gt;NIST SP 800-218 Secure Software Development Framework&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.mongodb.com/docs/manual/crud/" rel="noopener noreferrer"&gt;MongoDB CRUD operations concepts&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;If you are building a long-running agent harness and want a backend your agent can reliably test against across sessions, you can explore SashiDo’s platform at &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and start with a 10-day free trial to wire up Auth, Database, Functions, Realtime, Storage, Jobs, and Push without DevOps.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/coding-agents-best-practices-plan-test-ship-faster" rel="noopener noreferrer"&gt;Coding Agents: Best practices to plan, test, and ship faster&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/artificial-intelligence-coding-ai-coworker-not-tool" rel="noopener noreferrer"&gt;Artificial Intelligence Coding Is Becoming a Coworker, Not a Tool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/artificial-intelligence-coding-90-10-rule-build-vs-buy" rel="noopener noreferrer"&gt;Artificial Intelligence Coding and the 90/10 Rule: Build vs Buy&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/cursor-coding-remove-next-bottleneck" rel="noopener noreferrer"&gt;Cursor Coding Turns Output Up. Here’s How to Remove the Next Bottleneck&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/artificial-intelligence-coding-shipping-vibe-coded-apps" rel="noopener noreferrer"&gt;Artificial Intelligence Coding Is Easy. Shipping Vibe-Coded Apps Is Not&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>tutorial</category>
      <category>development</category>
    </item>
    <item>
      <title>AI Coding Security: The Vibe-Coding Risk Nobody Reviews</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Fri, 27 Feb 2026 07:00:24 +0000</pubDate>
      <link>https://dev.to/sashido/ai-coding-security-the-vibe-coding-risk-nobody-reviews-4oe0</link>
      <guid>https://dev.to/sashido/ai-coding-security-the-vibe-coding-risk-nobody-reviews-4oe0</guid>
      <description>&lt;p&gt;If you have been shipping with &lt;em&gt;ai coding&lt;/em&gt; tools lately, you have probably felt the trade-off in your hands. You can describe an app, watch thousands of lines appear, and demo something real in an afternoon. But the moment that code runs on your laptop, your API keys, browser sessions, and files sit one prompt away from becoming part of the experiment.&lt;/p&gt;

&lt;p&gt;A recent real-world incident made this painfully concrete. A security researcher demonstrated that, by modifying a single line inside a large AI-generated project, an attacker could quietly gain control of the victim’s machine. No suspicious download prompt. No “click this link” moment. Just the reality that when you cannot review what gets generated, you also cannot reliably defend it.&lt;/p&gt;

&lt;p&gt;The core lesson is simple and uncomfortable. &lt;strong&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-risks-technical-debt-backend-strategy" rel="noopener noreferrer"&gt;Vibe coding shifts risk&lt;/a&gt; from writing code to executing code&lt;/strong&gt;. The danger is not that AI writes “bad code” in the abstract. The danger is that it produces &lt;em&gt;a lot of code&lt;/em&gt; quickly, and it often runs with permissions your prototype does not deserve.&lt;/p&gt;

&lt;p&gt;Here is the pattern we see most often with solo founders and indie hackers. The build starts as a no code app builder style flow, or a low code application platform workflow with an AI chat maker UI. Then it becomes a real product. Users sign up. Payments enter the picture. Secrets land in environment variables. That is the point where “it works” stops being the bar.&lt;/p&gt;

&lt;p&gt;Right after you internalize that, the next step is to move the dangerous parts out of your personal machine and into a controlled environment.&lt;/p&gt;

&lt;p&gt;A practical way to do that early is to run prototypes against a managed backend where permissions, auth, storage, and isolation are already designed in. That is exactly why we built &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;&lt;strong&gt;SashiDo - Backend for Modern Builders&lt;/strong&gt;&lt;/a&gt;. It lets you keep the speed of ai generate app workflows, while avoiding the habit of giving bots local access to everything.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Breaks in Vibe Coding (And Why It Is Different)
&lt;/h2&gt;

&lt;p&gt;Traditional app security failures usually need a trigger. You click a malicious attachment. You paste credentials into the wrong place. You install a compromised dependency. In the incident above, the attacker’s leverage came from something scarier. The victim did not need to do anything at all after starting the project. That is what makes “zero-click” style compromises so damaging in practice.&lt;/p&gt;

&lt;p&gt;There are three reasons vibe-coding workflows create a new class of problems.&lt;/p&gt;

&lt;p&gt;First, &lt;strong&gt;the review surface explodes&lt;/strong&gt;. When an AI tool generates thousands of lines you did not author, it becomes normal to run code you do not understand. That makes it easy for malicious or compromised changes to hide in plain sight.&lt;/p&gt;

&lt;p&gt;Second, the tooling often has &lt;em&gt;deep local privileges&lt;/em&gt; by default. If your AI agent can read your filesystem to be helpful, it can also read secrets. If it can run commands to build and test, it can also execute unexpected payloads.&lt;/p&gt;

&lt;p&gt;Third, the “project” is rarely just code. It is config files, local caches, credentials, and tokens. That is why a single line added in the wrong place can turn a harmless demo into full device access.&lt;/p&gt;

&lt;p&gt;This is also why professor Kevin Curran’s warning lands with experienced engineers. Without discipline, documentation, and review, the output tends to fail under attack. The discipline part matters because &lt;em&gt;ai coding&lt;/em&gt; is less forgiving when you skip basic software hygiene.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Quick Threat Model for AI Coding Projects
&lt;/h2&gt;

&lt;p&gt;You do not need a full security program to make good decisions. You need a simple model of what can go wrong.&lt;/p&gt;

&lt;p&gt;Start with the assets. In almost every vibe-coding project we see, the highest value items are: API keys and tokens, user data, payment and analytics dashboards, and your local machine’s browser sessions and SSH keys.&lt;/p&gt;

&lt;p&gt;Then map the paths.&lt;/p&gt;

&lt;p&gt;An attacker can target the AI tool itself, its plugin ecosystem, or shared project artifacts. They can also target your own workflow. For example, sharing a project link, pulling “helpful” code snippets from community chat, or granting the agent permission to access a folder full of keys.&lt;/p&gt;

&lt;p&gt;Finally, map the outcomes. In the worst cases, a hidden change does not just break your app. It &lt;strong&gt;turns your environment into the attacker’s environment&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;If you want a compact set of categories that maps well to these failures, the &lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10 (2021)&lt;/a&gt; is still the best common language. You will recognize the usual suspects, like broken access control and injection. But in vibe coding, the biggest driver is often the same. Lack of visibility.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Features to Look For in Secure AI Coding Setups
&lt;/h2&gt;

&lt;p&gt;If your goal is to keep building quickly while reducing the odds of an “ai coding hacks” moment, you are looking for guardrails more than features.&lt;/p&gt;

&lt;p&gt;A secure setup typically has three layers.&lt;/p&gt;

&lt;p&gt;At the device layer, isolation matters. Running agentic AI directly on your daily laptop is convenient, but it makes compromise catastrophic. Microsoft’s &lt;a href="https://learn.microsoft.com/en-us/windows/security/application-security/application-isolation/windows-sandbox/windows-sandbox-overview" rel="noopener noreferrer"&gt;Windows Sandbox overview&lt;/a&gt; is a good example of the direction you want. &lt;a href="https://www.sashido.io/en/blog/caveat-coder-ai-infrastructure-importance" rel="noopener noreferrer"&gt;A disposable environment&lt;/a&gt;. A fresh state each run. Clear boundaries.&lt;/p&gt;

&lt;p&gt;At the identity layer, least privilege matters. Disposable accounts for experiments and short-lived credentials reduce blast radius. This aligns with the broader “assume breach” mindset found in the &lt;a href="https://www.cisa.gov/zero-trust-maturity-model" rel="noopener noreferrer"&gt;CISA Zero Trust Maturity Model&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;At the software layer, supply chain visibility matters. If you cannot answer “what dependencies did the agent add” you are already behind. CISA’s guidance on SBOMs, like &lt;a href="https://www.cisa.gov/resources-tools/resources/shared-vision-software-bill-materials-sbom-cybersecurity" rel="noopener noreferrer"&gt;Shared Vision for SBOM&lt;/a&gt;, is worth reading because it explains why modern software is as much about components as code.&lt;/p&gt;

&lt;p&gt;In practice, here is the checklist we see working for solo founders.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Keep the agent on a separate machine, VM, or sandbox when it can run code or access files.&lt;/li&gt;
&lt;li&gt;Use disposable accounts and test credentials for experiments. Avoid logging the agent into production dashboards.&lt;/li&gt;
&lt;li&gt;Treat generated code as untrusted until you review it. Focus review on auth, file access, network calls, and “helper” scripts.&lt;/li&gt;
&lt;li&gt;Lock down secrets. If you must use keys, use least-privilege keys and rotate them after a prototyping session.&lt;/li&gt;
&lt;li&gt;Add automated security checks early. GitHub’s &lt;a href="https://docs.github.com/en/enterprise-cloud@latest/code-security/getting-started/github-security-features" rel="noopener noreferrer"&gt;security features documentation&lt;/a&gt; is a good starting point for code scanning, secret scanning, and dependency alerts.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;None of this removes the value of vibe coding. It just puts your workflow back inside a security boundary.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where “Run It Locally” Fails First
&lt;/h2&gt;

&lt;p&gt;For early demos, local execution is fine. The break point usually happens when one of these becomes true.&lt;/p&gt;

&lt;p&gt;You start storing user content, like images, audio, or documents. You introduce authentication and password reset flows. You add push notifications. You accept payments or connect to production third-party APIs. Or you hit a growth threshold where a single security mistake impacts more than a handful of beta users.&lt;/p&gt;

&lt;p&gt;That is when local-first, agent-heavy workflows create two kinds of pain.&lt;/p&gt;

&lt;p&gt;The first is security pain. It becomes normal for your agent to have access to the same files and sessions you use for everything else.&lt;/p&gt;

&lt;p&gt;The second is operational pain. Even if the prototype works, you now need APIs, a database, background jobs, and a place to host and scale. If you try to bolt those on late, you often end up shipping with default settings and unreviewed permissions.&lt;/p&gt;

&lt;p&gt;This is the moment where a managed backend is less about convenience and more about &lt;em&gt;risk containment&lt;/em&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Top Options Compared for Shipping AI Coding Projects
&lt;/h2&gt;

&lt;p&gt;For commercial intent decisions, it helps to compare options by what they protect you from, not what they promise.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Option&lt;/th&gt;
&lt;th&gt;What It’s Great For&lt;/th&gt;
&lt;th&gt;Where It Breaks&lt;/th&gt;
&lt;th&gt;Best Fit&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Vibe coding on your main laptop&lt;/td&gt;
&lt;td&gt;Fastest first demo, quick iteration&lt;/td&gt;
&lt;td&gt;Large blast radius. Hard to review. Secrets leak risk&lt;/td&gt;
&lt;td&gt;One-off experiments with no real data&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Vibe coding in a sandbox or dedicated machine&lt;/td&gt;
&lt;td&gt;Safer agent execution&lt;/td&gt;
&lt;td&gt;Still need backend, auth, storage, scaling&lt;/td&gt;
&lt;td&gt;Early builders who want speed plus containment&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Roll your own backend (self-host)&lt;/td&gt;
&lt;td&gt;Maximum control&lt;/td&gt;
&lt;td&gt;DevOps tax, patching, uptime, backups&lt;/td&gt;
&lt;td&gt;Teams with infra experience and time&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Managed backend (BaaS) + AI front-end&lt;/td&gt;
&lt;td&gt;Faster path to production-grade primitives&lt;/td&gt;
&lt;td&gt;You still own app logic and access rules&lt;/td&gt;
&lt;td&gt;Solo founders going prototype to launch&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;If you are in the last category, this is where &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;&lt;strong&gt;SashiDo - Backend for Modern Builders&lt;/strong&gt;&lt;/a&gt; fits naturally. We built it so you can move from “the agent generated an app” to “this is a real service” without building a DevOps stack first.&lt;/p&gt;

&lt;p&gt;In a typical ai coding workflow, you need a database, APIs, auth, file storage, realtime updates, background jobs, serverless functions, and push notifications. In SashiDo, those are first-class features. Every app includes a MongoDB database with CRUD APIs, complete user management with social logins, object storage backed by AWS S3 with a built-in CDN, JavaScript serverless functions in Europe and North America, realtime via WebSockets, scheduled and recurring jobs, and unlimited iOS and Android push notifications.&lt;/p&gt;

&lt;p&gt;If you want to validate this quickly, our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; shows how to stand up a backend and connect a client app without building your own infrastructure.&lt;/p&gt;

&lt;p&gt;When comparing managed backends, you might also look at alternatives like Supabase, Hasura, AWS Amplify, or Vercel depending on your stack. If you do, keep the evaluation grounded in what you need for your launch. Auth model, database fit, scaling knobs, background job support, and how much operational responsibility you retain.&lt;/p&gt;

&lt;p&gt;For reference, we maintain comparison pages that highlight the practical differences. You can start with &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt;, &lt;a href="https://www.sashido.io/en/sashido-vs-hasura" rel="noopener noreferrer"&gt;SashiDo vs Hasura&lt;/a&gt;, &lt;a href="https://www.sashido.io/en/sashido-vs-aws-amplify" rel="noopener noreferrer"&gt;SashiDo vs AWS Amplify&lt;/a&gt;, and &lt;a href="https://www.sashido.io/en/sashido-vs-vercel" rel="noopener noreferrer"&gt;SashiDo vs Vercel&lt;/a&gt;. The point is not that one is “best” in a vacuum. The point is to choose the backend that reduces your risk and workload for the kind of app your ai coding tool is producing.&lt;/p&gt;

&lt;h2&gt;
  
  
  The “Best AI for Vibe Coding” Is the One You Can Constrain
&lt;/h2&gt;

&lt;p&gt;People often ask for the best ai for vibe coding as if the answer is purely about code quality or speed. In practice, the deciding factor is whether the workflow gives you &lt;strong&gt;control over permissions and execution&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;If the tool can run code, read files, and manage dependencies, then your security posture depends on what it is allowed to touch. The safer tools make boundaries obvious. They separate “generate text” from “execute actions.” They support running inside isolated environments. They make it easy to inspect diffs and changes.&lt;/p&gt;

&lt;p&gt;The most reliable pattern is to let AI help with generation and refactoring, then run builds and deployments inside a controlled pipeline. This is also why agentic AI on personal devices keeps landing in headlines. It is powerful, but without guardrails it is also extremely insecure.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI Coding Detector and AI Coding Checker: Useful, but Not a Seatbelt
&lt;/h2&gt;

&lt;p&gt;It is tempting to look for an ai coding detector or ai coding checker that can tell you whether the output is safe. These tools can help, especially when they flag obvious secrets, risky dependencies, or suspicious patterns. But they are not a replacement for isolation and access control.&lt;/p&gt;

&lt;p&gt;A detector can tell you “this looks machine-generated” or “this string resembles a key.” It cannot reliably answer, “does this project contain a hidden execution path that only triggers under specific conditions?” That is why the first line of defense should be limiting what the project can touch.&lt;/p&gt;

&lt;p&gt;Use checkers for what they are good at. Consistency, linting, scanning for known issues, and catching accidental leaks. Then build the real defenses around execution boundaries and least privilege.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Managed Backend Move: What Changes (And What Doesn’t)
&lt;/h2&gt;

&lt;p&gt;Moving to a managed backend does not magically make your app secure. You still need to design access rules and avoid shipping admin-level APIs to clients.&lt;/p&gt;

&lt;p&gt;What it does change is the reliability of your foundation. Your database is not a file on your laptop. Your auth system is not a half-finished prompt output. Your storage and CDN are not an ad-hoc bucket with unknown permissions. Your background jobs do not run on a machine that also holds your personal SSH keys.&lt;/p&gt;

&lt;p&gt;At SashiDo, we see this shift most clearly when indie hackers add auth late. They often start with a “just store users in local storage” approach because the AI suggests it. Then they realize password resets, social logins, token expiry, and account takeover protection are a product in themselves.&lt;/p&gt;

&lt;p&gt;That is why we include a complete User Management system by default, and why our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;documentation&lt;/a&gt; focuses on concrete, buildable flows rather than marketing promises.&lt;/p&gt;

&lt;p&gt;If you are dealing with higher stakes workloads, it is also worth reviewing our &lt;a href="https://www.sashido.io/en/policies" rel="noopener noreferrer"&gt;security and privacy policies&lt;/a&gt; to understand where the platform’s responsibilities end and where yours begin.&lt;/p&gt;

&lt;h2&gt;
  
  
  Cost, Scale, and the “Surprise Bill” Problem
&lt;/h2&gt;

&lt;p&gt;The other anxiety we hear constantly from the vibe-coder-solo-founder-indie-hacker crowd is cost volatility. The pattern is predictable. A demo hits social media. Traffic spikes. The backend bill surprises you. Then you start turning features off.&lt;/p&gt;

&lt;p&gt;The best defense is not a perfect forecast. It is picking an architecture that can scale in &lt;em&gt;predictable steps&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;In SashiDo, scaling is designed around clear knobs. You start with an app plan and scale resources as needed. If you want the current pricing and what is included, always check our live &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt;, because rates and limits can change over time. The key point for planning is that you can begin with a free trial and then scale requests, storage, and compute as real usage arrives.&lt;/p&gt;

&lt;p&gt;When you hit compute-heavy workloads, like agent-driven processing or bursty realtime features, that is when our Engines become relevant. Our write-up on the &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Engines feature&lt;/a&gt; explains how isolation and performance scaling work, and how usage is calculated.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Practical “Stop Doing This” List for AI Coding
&lt;/h2&gt;

&lt;p&gt;If you only change a few habits this week, make them these.&lt;/p&gt;

&lt;p&gt;Do not run agentic tools with access to your home directory “because it’s easier.” Do not store production secrets in files the agent can read. Do not let an AI tool auto-install dependencies without checking what it added. Do not treat “it compiled” as a security signal. And do not assume that because the code came from a well-rated tool, the project is safe.&lt;/p&gt;

&lt;p&gt;Instead, build a workflow where you can move fast &lt;em&gt;and&lt;/em&gt; contain failures. Use isolation for execution. Use disposable credentials. Use automated scanning for obvious leaks. Then move the backend into a managed environment before you start collecting real users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Secure AI Coding Means Constraining the Agent
&lt;/h2&gt;

&lt;p&gt;The big shift in &lt;em&gt;ai coding&lt;/em&gt; is not that software became easier to write. It is that software became easier to &lt;em&gt;run&lt;/em&gt; without understanding it. That is how you get a single hidden change turning into full device access, and how you end up with a “zero-click” style compromise in what looked like a harmless prototype.&lt;/p&gt;

&lt;p&gt;The fix is not to abandon vibe coding. The fix is to treat AI output as untrusted until proven otherwise, and to &lt;strong&gt;move execution and data behind boundaries you control&lt;/strong&gt;.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you want to keep shipping quickly without giving bots deep local access, it helps to put your database, auth, storage, and jobs behind a managed backend. You can explore &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;&lt;strong&gt;SashiDo - Backend for Modern Builders&lt;/strong&gt;&lt;/a&gt; to sandbox AI agent-driven apps, add production-ready auth and APIs, and start with a 10-day free trial with no credit card required. For the most up-to-date plan details, refer to our live &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing&lt;/a&gt; page.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Is the Best Coder for AI?
&lt;/h3&gt;

&lt;p&gt;The best “coder for AI” is the workflow that lets you constrain what the model or agent can execute, not the one that generates the most code. Look for strong boundaries, reviewable diffs, and isolated execution. If the tool can run commands or access files, your ability to limit permissions matters more than raw generation quality.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Are the Most Common AI Coding Hacks in Vibe-Coding Workflows?
&lt;/h3&gt;

&lt;p&gt;The most common failures are hidden code changes, leaked secrets, and overly broad permissions. In vibe coding, attackers do not need you to understand the code. They need you to run it. That is why isolating execution and using disposable credentials reduce risk even when you cannot fully review every generated file.&lt;/p&gt;

&lt;h3&gt;
  
  
  When Should I Stop Prototyping Locally and Move the Backend?
&lt;/h3&gt;

&lt;p&gt;Move off local-first setups once you add real auth, start storing user content, connect to paid APIs, or expect public traffic. Those are the points where compromise affects users, not just your demo. A managed backend also helps when you need background jobs, push notifications, or predictable scaling without building DevOps.&lt;/p&gt;

&lt;h3&gt;
  
  
  Do AI Coding Detectors and AI Coding Checkers Actually Improve Security?
&lt;/h3&gt;

&lt;p&gt;They help with specific problems like finding accidental secrets, spotting known vulnerable dependencies, and enforcing basic hygiene. They do not replace isolation or access control, because they cannot reliably prove a large project has no hidden execution paths. Use them as a safety net, not as your primary defense.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10 (2021)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.nist.gov/itl/ai-risk-management-framework" rel="noopener noreferrer"&gt;NIST AI Risk Management Framework (AI RMF)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.cisa.gov/resources-tools/resources/shared-vision-software-bill-materials-sbom-cybersecurity" rel="noopener noreferrer"&gt;CISA Shared Vision for SBOM&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://learn.microsoft.com/en-us/windows/security/application-security/application-isolation/windows-sandbox/windows-sandbox-overview" rel="noopener noreferrer"&gt;Microsoft Windows Sandbox Overview&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.github.com/en/enterprise-cloud@latest/code-security/getting-started/github-security-features" rel="noopener noreferrer"&gt;GitHub Security Features Documentation&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-fun-ai-assisted-programming" rel="noopener noreferrer"&gt;Vibe Coding: Fun, AI-Assisted Programming for Makers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/embracing-vibe-coding" rel="noopener noreferrer"&gt;Embracing Vibe Coding: Making Programming More Fun with AI&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-software-development-excitement" rel="noopener noreferrer"&gt;Vibe Coding: Making Software Development Exciting Again&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/no-code-platforms-meet-the-real-world-vibe-coding-that-ships" rel="noopener noreferrer"&gt;No Code Platforms Meet the Real World: Vibe Coding That Ships&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-builder-xcode-vibe-coding-backend-checklist" rel="noopener noreferrer"&gt;Agentic Coding in Xcode: Turn Vibe Coding Into a Real App&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>programming</category>
      <category>devops</category>
    </item>
    <item>
      <title>Mobile Backend as a Service in the Age of Long-Running Agents</title>
      <dc:creator>Marian Ignev</dc:creator>
      <pubDate>Thu, 26 Feb 2026 07:00:28 +0000</pubDate>
      <link>https://dev.to/sashido/mobile-backend-as-a-service-in-the-age-of-long-running-agents-40ee</link>
      <guid>https://dev.to/sashido/mobile-backend-as-a-service-in-the-age-of-long-running-agents-40ee</guid>
      <description>&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/ctos-dont-let-ai-agents-run-the-backend-yet" rel="noopener noreferrer"&gt;Long-running AI coding agents&lt;/a&gt; are changing what “a day of engineering” looks like. Instead of generating a small patch you can review in ten minutes, they can now work for 24 to 50+ hours and come back with a pull request that touches authentication, data models, tests, and performance bottlenecks all at once. That is exciting. It is also where many teams discover a new constraint: &lt;strong&gt;the bottleneck moves from writing code to deploying code safely&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;If you are building a mobile backend as a service based product or you are choosing one for your app, this shift matters. Bigger PRs can keep merge rates high, but only if your backend surface area is predictable, your deployment path is guarded, and your operational load is not quietly ballooning behind the scenes.&lt;/p&gt;

&lt;p&gt;In this guide, we will walk through the practical patterns we see working when teams combine long-running agents with a managed mobile backend, and the failure modes you should plan for if you want speed without surprises.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Actually Changes When Agents Run for 30+ Hours
&lt;/h2&gt;

&lt;p&gt;The biggest difference is not that the model writes “better code”. The difference is that it has time to do &lt;em&gt;everything around the code&lt;/em&gt; that humans often postpone. It can refactor adjacent modules, chase edge cases, add missing tests, and fix inconsistencies it finds along the way. That is why long-running agents tend to produce larger PRs that still feel surprisingly mergeable.&lt;/p&gt;

&lt;p&gt;But longer horizons also amplify small misunderstandings. One incorrect assumption about your auth model or your data ownership rules can propagate through thousands of lines before anyone notices. In practice, teams that succeed with long-running agents adopt two habits.&lt;/p&gt;

&lt;p&gt;First, they &lt;strong&gt;force alignment before execution&lt;/strong&gt;. The agent proposes a plan, you approve the plan, then it starts the long run. Second, they &lt;strong&gt;force follow-through&lt;/strong&gt;. The agent does not stop at “it compiles”. It runs checks, revisits earlier decisions, and has other agent passes or tools act as reviewers.&lt;/p&gt;

&lt;p&gt;Right after you adopt those habits, another reality becomes obvious. If the agent is producing production-ready changes, you need a &lt;a href="https://www.sashido.io/en/blog/best-mbaas-mobile-backend-platform-production" rel="noopener noreferrer"&gt;production-ready deployment path&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;If your backend is a bundle of bespoke services, one huge PR can mean one huge deploy. That increases blast radius. If your backend is built on a mobile backend as a service, the deploy surface is usually narrower because you are mostly shipping schema changes, serverless functions, jobs, and access rules on top of stable primitives.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/mobile-backend-as-a-service" rel="noopener noreferrer"&gt;A mobile backend as a service&lt;/a&gt; does not eliminate review. It reduces the number of places where review can go catastrophically wrong.&lt;/p&gt;

&lt;p&gt;A quick contextual move that helps many teams is to start by stabilizing the backend primitives you do not want an agent to reinvent.&lt;/p&gt;

&lt;p&gt;If that is the stage you are at, you can &lt;strong&gt;anchor your mobile backend to managed building blocks&lt;/strong&gt; with &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;, then let agents spend their long runs on app logic and tests instead of re-creating auth, storage, and realtime infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  How a Mobile Backend as a Service Helps Long-Running Agent Work Stay Mergeable
&lt;/h2&gt;

&lt;p&gt;A good mobile backend as a service is less about “no backend code”. It is about a smaller, repeatable backend that is easy to reason about, even when changes are large. In day-to-day terms, it gives your agents a clearer target and gives your reviewers fewer unknowns.&lt;/p&gt;

&lt;p&gt;Here is the pattern we see repeatedly.&lt;/p&gt;

&lt;p&gt;You keep your data model and access rules explicit. You keep server-side logic in a limited number of places. You treat background work as scheduled jobs, not ad hoc cron servers. You use a first-class auth system, not a hand-rolled JWT setup scattered across endpoints. You keep files in object storage with a CDN, not in a random VM directory. Then you let the agent produce big PRs, because the infrastructure footprint stays stable.&lt;/p&gt;

&lt;p&gt;This is also why “backend-as-a service platforms” tend to show up in AI-assisted workflows faster than in traditional ones. When code generation cost drops, the expensive part is &lt;em&gt;integration, observability, rollback, and policy enforcement&lt;/em&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Planning Pass: What to Lock Down Before the Agent Writes Anything
&lt;/h3&gt;

&lt;p&gt;When you ask an agent for a long run, treat it like you would treat a senior engineer starting a week-long refactor. The plan is not bureaucracy. It is a way to prevent a small misunderstanding from turning into a full rewrite.&lt;/p&gt;

&lt;p&gt;A plan that works for long-running agent runs is usually short and concrete.&lt;/p&gt;

&lt;p&gt;It defines what is in scope and out of scope, which tables or collections are allowed to change, which endpoints or Cloud Functions can be modified, what the migration path is, and what “done” means in terms of tests and monitoring.&lt;/p&gt;

&lt;p&gt;In mobile backends, the plan should also name the “invariants” that must not be broken. Typical invariants are user ownership rules, role boundaries, and the behavior of push notification delivery.&lt;/p&gt;

&lt;p&gt;If you are building on Parse-compatible infrastructure, it also helps to anchor the plan to the exact areas where server-side behavior lives. For example, &lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-engineering-prompts-to-production" rel="noopener noreferrer"&gt;Cloud Code functions&lt;/a&gt; and triggers, scheduled jobs, and access rules. That keeps the agent from spraying logic across new microservices. If you need a refresher on where those building blocks sit, link your team to our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;SashiDo documentation&lt;/a&gt; at the start of the project so the plan uses the same concepts.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Follow-Through Pass: What “Production-Ready” Means in Practice
&lt;/h3&gt;

&lt;p&gt;Long-running agents are most valuable when they go past “feature implemented” and into “feature integrated”. Integration is where most real-world projects die.&lt;/p&gt;

&lt;p&gt;For agent-generated PRs, integration usually means four things.&lt;/p&gt;

&lt;p&gt;It means the PR ships with tests or at least adds coverage where the change is risky. It means access control is reviewed as part of the change, not as a follow-up. It means performance is checked at the query level, not only at the UI level. It means your deploy path enforces required checks and approvals.&lt;/p&gt;

&lt;p&gt;GitHub makes the last part explicit. Branch protection can require reviews and required status checks before merge. That is not an AI-specific feature, it is the core guardrail you want when PR size grows. If your team needs to re-check what you can enforce, GitHub’s documentation on &lt;a href="https://docs.github.com/repositories/configuring-branches-and-merges-in-your-repository/managing-rulesets/available-rules-for-rulesets" rel="noopener noreferrer"&gt;required reviews and rulesets&lt;/a&gt; is the practical reference.&lt;/p&gt;

&lt;h2&gt;
  
  
  Mobile Backend as a Service: The Parts Agents Change Most Often
&lt;/h2&gt;

&lt;p&gt;When teams say they want agents to “build the backend”, they usually mean a few repeatable slices. These slices are also where most mobile backend incidents happen, because they are cross-cutting.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Model and Query Performance
&lt;/h3&gt;

&lt;p&gt;Agents are good at large schema and model refactors because they can chase every call site. The risk is that they can also introduce slow queries in places you do not notice until load hits.&lt;/p&gt;

&lt;p&gt;The safe pattern is to make indexing and query shape part of the “definition of done”. MongoDB’s own guidance on &lt;a href="https://www.mongodb.com/docs/v7.1/crud/" rel="noopener noreferrer"&gt;CRUD operations&lt;/a&gt; and indexing is a good grounding for what to watch. If your agent adds new filters or sorts, make sure it also proposes the index changes, and make sure you have a way to observe query latency after deploy.&lt;/p&gt;

&lt;p&gt;In our platform, every app comes with a MongoDB database and a CRUD API. That is a stable base for agents to build on because your data layer does not change shape just because you are shipping a new feature.&lt;/p&gt;

&lt;h3&gt;
  
  
  Authentication, Social Login, and RBAC
&lt;/h3&gt;

&lt;p&gt;Auth refactors are one of the clearest wins for long-running agents because they are tedious and easy to get wrong manually. They are also one of the easiest places for an agent to make a “reasonable” assumption that is still wrong for your security model.&lt;/p&gt;

&lt;p&gt;A practical approach is to explicitly tie the plan to authorization rules and then review those rules first in the PR. OWASP’s guidance on authorization, including the importance of server-side enforcement and least privilege, is a useful reality check when a refactor touches roles and permissions. Their &lt;a href="https://cheatsheetseries.owasp.org/cheatsheets/Transaction_Authorization_Cheat_Sheet.html" rel="noopener noreferrer"&gt;Transaction Authorization Cheat Sheet&lt;/a&gt; is a good starting point for the kinds of checks reviewers should demand.&lt;/p&gt;

&lt;p&gt;On &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;, we ship a complete user management system and make social providers available with minimal setup. That matters in agent-driven workflows because it reduces the temptation to create one-off login flows and scattered token logic.&lt;/p&gt;

&lt;h3&gt;
  
  
  Files, Storage, and CDN Behavior
&lt;/h3&gt;

&lt;p&gt;If your agent-generated PR touches media, uploads, or user-generated files, you want storage to be boring. The fastest way to build a fragile system is to mix “temporary dev storage” with production traffic.&lt;/p&gt;

&lt;p&gt;Object storage durability and access patterns are well understood. For example, Amazon S3 is designed for high durability and redundancy. The official AWS documentation on &lt;a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/DataDurability.html" rel="noopener noreferrer"&gt;Amazon S3 durability&lt;/a&gt; explains why. When you put this behind a CDN and you enforce consistent upload rules, you turn file handling from an app-wide source of bugs into a backend primitive.&lt;/p&gt;

&lt;p&gt;In our stack, files live in an AWS S3 object store with built-in CDN behavior. If you want the architectural details behind that decision, our post on &lt;a href="https://www.sashido.io/en/blog/announcing-microcdn-for-sashido-files" rel="noopener noreferrer"&gt;MicroCDN for SashiDo Files&lt;/a&gt; gives the performance reasoning and what it changes for delivery.&lt;/p&gt;

&lt;h3&gt;
  
  
  Realtime State and WebSockets
&lt;/h3&gt;

&lt;p&gt;Realtime is one of those features that looks simple in a demo and becomes expensive in production. Agents can help implement a realtime slice end to end, but you still need to choose a model for synchronization and consistency.&lt;/p&gt;

&lt;p&gt;The underlying idea is that WebSockets enable long-lived, bidirectional sessions, which lets you broadcast updates and keep clients in sync. The WebSockets project documentation provides a straightforward overview of how state updates and broadcasting work in practice. See the &lt;a href="https://websockets.readthedocs.io/en/5.0/intro.html" rel="noopener noreferrer"&gt;websockets documentation&lt;/a&gt; for the core mechanics.&lt;/p&gt;

&lt;p&gt;The practical guardrail is to define what must be strongly consistent versus what can be eventually consistent. Chat typing indicators can be lossy. Billing state cannot. That clarity should go into the plan you approve before the agent starts.&lt;/p&gt;

&lt;h3&gt;
  
  
  Background Jobs and Recurring Work
&lt;/h3&gt;

&lt;p&gt;Long-running agents are surprisingly good at building the “job pipeline” that product teams keep pushing off, like retryable deliveries, cleanup jobs, or periodic aggregations. Where teams get burned is when jobs run without clear ownership, schedules, or dashboards.&lt;/p&gt;

&lt;p&gt;If you are using Agenda with MongoDB, it is worth reading the official &lt;a href="https://agenda.github.io/agenda/" rel="noopener noreferrer"&gt;Agenda documentation&lt;/a&gt; because it makes the job model and recurring schedules explicit. In our platform, scheduled and recurring jobs are built in and manageable via the dashboard, which turns “jobs” from a hidden ops concern into something you can review and reason about alongside application logic.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Practical Workflow: Turning a Long-Running Agent PR Into a Safe Deploy
&lt;/h2&gt;

&lt;p&gt;For a startup CTO or technical co-founder, the workflow matters more than the tool. The best harness in the world still produces risk if your team merges and deploys without a consistent gate.&lt;/p&gt;

&lt;p&gt;This is the workflow we recommend when agent runs start producing PRs that feel “too big to review”, but still too valuable to ignore.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Constrain the Task to a Backend Slice
&lt;/h3&gt;

&lt;p&gt;If the task touches everything, the PR will touch everything. Prefer slices like “refactor auth and RBAC”, “migrate storage paths”, or “add high-coverage tests around payments webhooks”. Long-running agents thrive when the goal is concrete and the end state is verifiable.&lt;/p&gt;

&lt;p&gt;If your backend stack is already fragmented, consider consolidating first on a smaller set of managed primitives. That is where a mobile backend as a service is often the difference between “agent PRs are magic” and “agent PRs are scary”.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Require a Plan That Names Invariants
&lt;/h3&gt;

&lt;p&gt;Have the agent propose a plan and make approval explicit. The plan should name invariants like user ownership, data retention rules, and role boundaries. If you are on &lt;a href="https://www.sashido.io/en/blog/backend-as-a-service-guide-2026" rel="noopener noreferrer"&gt;Parse-compatible infrastructure&lt;/a&gt;, it should also name which Cloud Functions, triggers, and jobs are allowed to change, so reviewers know where to look.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Enforce Merge Gates, Especially for Big PRs
&lt;/h3&gt;

&lt;p&gt;Large PRs are not automatically unsafe. Unchecked merges are.&lt;/p&gt;

&lt;p&gt;At minimum, enforce required reviews and required status checks in your GitHub rules. Then require a security pass for PRs touching auth and access control. This is not “AI governance”. It is the same engineering hygiene you want when the size of change increases.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Deploy in a Way That Minimizes Blast Radius
&lt;/h3&gt;

&lt;p&gt;If you can deploy backend changes separately from mobile app releases, do it. If a PR touches both, consider rolling out backend changes first behind feature flags or toggles where possible.&lt;/p&gt;

&lt;p&gt;In practice, teams that use backend as a service providers often get this separation more easily because backend primitives are already centralized. You are not deploying five separate services to get one feature out.&lt;/p&gt;

&lt;p&gt;If you are scaling, make sure your backend can increase capacity without turning every release into an ops event. Our guide on &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;SashiDo Engines&lt;/a&gt; walks through how scaling works in our infrastructure and how to think about cost and performance trade-offs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Observe First, Then Expand
&lt;/h3&gt;

&lt;p&gt;When a long-running agent PR lands, you should assume it changed more than you noticed. That is not a critique. It is the nature of long-horizon work.&lt;/p&gt;

&lt;p&gt;Start by watching request error rates, latency, and key job and push notification deliveries. Only then expand traffic or enable the feature broadly.&lt;/p&gt;

&lt;p&gt;If uptime is a hard requirement for your app, you will also want a clear high-availability posture. Our post on &lt;a href="https://www.sashido.io/en/blog/dont-let-your-apps-down-enable-high-availability" rel="noopener noreferrer"&gt;High Availability and Self-Healing&lt;/a&gt; explains the common failure modes and what a safer deployment setup looks like.&lt;/p&gt;

&lt;h3&gt;
  
  
  Key Takeaways You Can Apply This Week
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Treat long-running agent tasks like real projects&lt;/strong&gt;, not prompts. Approve a plan, name invariants, define done.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Move safety left into your merge workflow&lt;/strong&gt; with required reviews and required checks.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Prefer managed primitives for the mobile backend&lt;/strong&gt;, so PRs change app logic more than infrastructure.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Deploy with blast radius in mind&lt;/strong&gt;, then observe before expanding rollout.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Trade-Offs: When Long-Running Agents Are the Wrong Tool
&lt;/h2&gt;

&lt;p&gt;Long-running agents are not a free win. They tend to underperform in a few predictable situations.&lt;/p&gt;

&lt;p&gt;They struggle when the product requirements are still ambiguous. If you cannot define done, the agent will likely optimize for an interpretation that creates rework. They are also risky when your security model is undocumented. Auth and RBAC changes need explicit invariants and human review.&lt;/p&gt;

&lt;p&gt;They can also create “false progress” in environments where you cannot deploy frequently. If your team ships once per month, a 36-hour PR is not the main bottleneck. Your release process is. In that case, use agents for test coverage and refactoring first, then revisit feature work.&lt;/p&gt;

&lt;p&gt;Finally, long-running agents can be wasteful if your architecture is chaotic. If a task requires changing five services, three queues, and two data stores just to add a feature, you will pay for complexity no matter who writes the code. That is often the moment teams consider consolidating on app building platforms or a single managed backend.&lt;/p&gt;

&lt;p&gt;If you are comparing approaches, we keep an up-to-date technical comparison for teams evaluating different stacks. For example, here is our &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase comparison&lt;/a&gt; that focuses on practical differences in backend primitives and operational responsibility.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started Without Turning Your Backend Into an AI Experiment
&lt;/h2&gt;

&lt;p&gt;If you want to try this approach without committing your whole roadmap to it, start with a single backend slice that is measurable.&lt;/p&gt;

&lt;p&gt;A good first project is a refactor that has obvious success criteria, like “reduce auth-related bugs” or “make uploads consistent”. Another good first project is a performance cleanup where you can measure request latency and error rates.&lt;/p&gt;

&lt;p&gt;If you are new to our platform, our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; is the quickest way to set up database, auth, storage, and serverless functions, and then keep your backend changes reviewable. The follow-up &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide-part-2" rel="noopener noreferrer"&gt;Getting Started Guide Part 2&lt;/a&gt; goes deeper into building feature-rich apps and managing projects in the dashboard.&lt;/p&gt;

&lt;p&gt;If cost predictability is part of your decision, keep one rule in mind. Always verify the current plan limits and overage pricing on the official &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;SashiDo pricing&lt;/a&gt; page, since these details can change as we update the platform.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://www.cloudflare.com/learning/serverless/glossary/backend-as-a-service-baas/" rel="noopener noreferrer"&gt;Backend as a Service (BaaS) Definition&lt;/a&gt; from Cloudflare, helpful for clarifying the BaaS and mobile backend as a service boundary.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.github.com/repositories/configuring-branches-and-merges-in-your-repository/managing-rulesets/available-rules-for-rulesets" rel="noopener noreferrer"&gt;GitHub Rulesets and Required Checks&lt;/a&gt;, the practical reference for merge gates.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://cheatsheetseries.owasp.org/cheatsheets/Transaction_Authorization_Cheat_Sheet.html" rel="noopener noreferrer"&gt;OWASP Transaction Authorization Cheat Sheet&lt;/a&gt;, a solid checklist mindset for authorization-sensitive changes.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/DataDurability.html" rel="noopener noreferrer"&gt;Amazon S3 Data Durability&lt;/a&gt;, useful context for object storage expectations.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.mongodb.com/docs/v7.1/crud/" rel="noopener noreferrer"&gt;MongoDB CRUD Operations&lt;/a&gt;, a quick grounding for how data access patterns change during refactors.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion: Shipping Faster Without Losing Control
&lt;/h2&gt;

&lt;p&gt;Long-running agents make it realistic to delegate work that used to take weeks, like large refactors, deep test coverage improvements, and performance overhauls. The catch is that they move risk into a new place. You are no longer asking “can we write this code”. You are asking &lt;strong&gt;can we review, merge, and deploy this much change safely&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That is where a mobile backend as a service can be a force multiplier. It keeps the backend primitives stable so agent work concentrates on business logic, and it reduces the number of bespoke services that can break during a big merge. Pair that with upfront planning, explicit invariants, and strict merge gates, and the “big PR” stops being scary.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you want a managed foundation that makes long-running agent PRs easier to review and safer to deploy, you can explore &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and start with a small backend slice before expanding to your full app.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions About Mobile Backend as a Service
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Is an Example of a Backend as a Service?
&lt;/h3&gt;

&lt;p&gt;A practical example is a platform that gives you a hosted database, CRUD APIs, authentication, file storage, and serverless functions as managed components. For mobile teams, that means you can ship features without standing up and operating separate services for auth, storage, realtime, and background jobs.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is a Mobile Backend?
&lt;/h3&gt;

&lt;p&gt;A mobile backend is the server-side system that supports a mobile app, including data storage, user authentication, business logic, and integrations like push notifications. In a mobile backend as a service approach, those capabilities are provided as managed building blocks so teams can focus on the app and product logic.&lt;/p&gt;

&lt;h3&gt;
  
  
  Is BaaS Good for IoT Applications?
&lt;/h3&gt;

&lt;p&gt;BaaS can be a good fit for IoT when devices mainly need secure auth, simple data ingestion, and reliable storage, and when you want to avoid heavy DevOps overhead. It becomes a worse fit when you need highly specialized protocols, strict on-prem constraints, or ultra-custom streaming pipelines that exceed what the platform supports.&lt;/p&gt;

&lt;h3&gt;
  
  
  Should I Use a Backend as a Service?
&lt;/h3&gt;

&lt;p&gt;Use a backend as a service when your main constraint is shipping speed and you want predictable primitives for auth, data, files, and realtime features. Avoid it when your backend is your product’s differentiator at the infrastructure level, or when compliance and custom networking requirements force you into a fully bespoke deployment model.&lt;/p&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-coding-ai-developer-productivity" rel="noopener noreferrer"&gt;Does AI Coding Really Boost Output?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/creating-an-app-weekend-builds-take-weeks" rel="noopener noreferrer"&gt;Creating an App in a Weekend? The 47,000-Line Reality&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/iphone-app-with-ai-xcode-no-code-backend" rel="noopener noreferrer"&gt;iPhone App with AI in Xcode: Build Your First MVP Fast&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-database-as-a-service" rel="noopener noreferrer"&gt;Vibe Coding Meets Database as a Service: Ship Fast, Stay Safe&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/choosing-a-mobile-backend-platform-2026-practical-checklist" rel="noopener noreferrer"&gt;Choosing a Mobile Backend Platform in 2026: A Practical Checklist&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>mobile</category>
      <category>backend</category>
      <category>cloud</category>
    </item>
    <item>
      <title>Creating Apps With Human Curation and AI: From Vibe Code to Real Users</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Wed, 25 Feb 2026 07:00:25 +0000</pubDate>
      <link>https://dev.to/sashido/creating-apps-with-human-curation-and-ai-from-vibe-code-to-real-users-6hm</link>
      <guid>https://dev.to/sashido/creating-apps-with-human-curation-and-ai-from-vibe-code-to-real-users-6hm</guid>
      <description>&lt;p&gt;The fastest way to get momentum when &lt;strong&gt;creating apps&lt;/strong&gt; in 2026 is to combine two things that used to live in separate worlds. Human curation (taste, judgment, and context) and AI assistance (speed, synthesis, and automation). When it clicks, you stop arguing about frameworks and start shipping something people actually want to use.&lt;/p&gt;

&lt;p&gt;But there’s a predictable second act. Once real users show up, your “vibe-coded” prototype suddenly needs a real backend: authentication, a database you can trust, file storage for uploads, background work, and a way to push updates or notifications without babysitting servers.&lt;/p&gt;

&lt;p&gt;This is the point where many solo founders stall, not because the product idea is weak, but because the infrastructure work is the opposite of fun. It is also where you can make one of the highest leverage decisions in the whole project: decide what stays custom, and what becomes a managed primitive.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Vibe Coding Works for Creating Apps (Until It Doesn’t)
&lt;/h2&gt;

&lt;p&gt;Vibe coding works because it compresses the feedback loop. You can take a pile of unstructured inputs, photos, notes, half-finished ideas, and turn them into a usable interface with AI helping you draft components, refactor, and connect flows. For early product discovery, that speed is a superpower.&lt;/p&gt;

&lt;p&gt;The pattern is especially strong for “taste-driven” apps where the product value is not the algorithm alone. It’s the combination of a point of view and a system that makes that point of view discoverable. Book recommendations, playlists, lesson plans, local guides, design patterns, curated prompts, even niche directories. The AI helps you &lt;em&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-experience-ai-tools" rel="noopener noreferrer"&gt;index and connect&lt;/a&gt;&lt;/em&gt; the curator’s intent at scale.&lt;/p&gt;

&lt;p&gt;Where it starts to break is right when you earn the first real traction. People want to create profiles, save favorites, share lists, upload their own content, and see personalized results. The app becomes stateful. You now need consistent data modeling, permissions, abuse prevention, and operational reliability.&lt;/p&gt;

&lt;p&gt;A useful rule of thumb: if you can describe your product as “a personalized feed” or “&lt;a href="https://www.sashido.io/en/blog/vibe-coding-mvp-parse-server-backend" rel="noopener noreferrer"&gt;a library of user-created items&lt;/a&gt;,” you are already in backend land.&lt;/p&gt;

&lt;p&gt;If you are at that point, our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; is a practical walkthrough for wiring up auth, data, and server-side logic quickly so your prototype can handle real users.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Core Insight: Human Curation Sets the North Star, AI Scales the Paths
&lt;/h2&gt;

&lt;p&gt;When an app’s value depends on taste, the best results usually come from a split of responsibilities.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Humans define the ontology.&lt;/strong&gt; That means the themes, labels, genres, categories, and the “why” behind an item. In practice, it often starts as a spreadsheet, a doc, or a set of notes. It is messy, personal, and opinionated. That is good.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AI turns that ontology into workflows.&lt;/strong&gt; It helps you inventory a collection, extract metadata from images, generate summaries, propose related items outside your dataset, and keep the experience fresh without needing a full-time content team.&lt;/p&gt;

&lt;p&gt;The big product unlock is that this approach creates an app that feels personal at scale. It is not trying to be the universal truth. It is trying to be a coherent perspective that users can subscribe to.&lt;/p&gt;

&lt;p&gt;The engineering implication is straightforward: you will store curated objects, store user objects, and store interaction events. Then you will run recommendation logic that mixes “curator-first” and “AI-augmented.” That’s why the backend becomes the long pole.&lt;/p&gt;

&lt;h2&gt;
  
  
  What a Vibe-Coded Web App Needs to Graduate to Production
&lt;/h2&gt;

&lt;p&gt;Most prototypes start as a single-page app with a few API calls. Then the requirements expand. Not because you got fancy, but because users demand the basics.&lt;/p&gt;

&lt;h3&gt;
  
  
  Authentication and Identity (So Personalization Actually Works)
&lt;/h3&gt;

&lt;p&gt;The moment you add profiles, you need reliable login and session handling. In practice, social sign-in is what prevents drop-off, especially when you are testing a new idea and users have low commitment.&lt;/p&gt;

&lt;p&gt;In &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;, every app comes with a complete User Management system. You can enable social logins like Google, Facebook, GitHub, and Microsoft providers with minimal setup, which matters when you are iterating daily and do not want to maintain your own auth stack.&lt;/p&gt;

&lt;h3&gt;
  
  
  A Database That Matches How You Build Features
&lt;/h3&gt;

&lt;p&gt;For creator and discovery apps, your data model changes constantly. One week you store “themes.” The next week you add “mashups,” “shelves,” “reactions,” and “reading status.” If your database workflow fights you, you slow down.&lt;/p&gt;

&lt;p&gt;We see many solo builders move faster with a flexible document model, especially early on. That’s why every SashiDo app includes a MongoDB database with a CRUD API. You can evolve your schema as your UX evolves, without rewriting migrations every other night.&lt;/p&gt;

&lt;h3&gt;
  
  
  File Storage and Delivery (Because Users Upload Everything)
&lt;/h3&gt;

&lt;p&gt;If your app involves images, covers, audio clips, PDFs, or user-generated attachments, you need storage that is boring and scalable. You also need delivery that does not punish you for success.&lt;/p&gt;

&lt;p&gt;Our Files offering is an &lt;a href="https://www.sashido.io/en/blog/best-backend-as-a-service-vibe-coding" rel="noopener noreferrer"&gt;AWS S3 object store&lt;/a&gt; integrated with a built-in CDN, designed for fast delivery at scale. If your “inventory and index” workflow starts with photos, this becomes a core primitive, not an afterthought.&lt;/p&gt;

&lt;h3&gt;
  
  
  Background Work, Scheduled Jobs, and Notifications
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/ai-assisted-coding-vibe-projects-2026" rel="noopener noreferrer"&gt;AI-assisted apps&lt;/a&gt; often require asynchronous tasks: embedding generation, classification, metadata enrichment, or sending recommendation emails. Then you add routine jobs: cleanup tasks, digest emails, or “rebuild the index” runs.&lt;/p&gt;

&lt;p&gt;In SashiDo, you can schedule and manage recurring jobs via our dashboard, and send unlimited mobile push notifications (iOS and Android) when you need re-engagement without wiring a bespoke pipeline.&lt;/p&gt;

&lt;h3&gt;
  
  
  Realtime for Shared State
&lt;/h3&gt;

&lt;p&gt;Realtime is not only for chat. It is for any UI where the state should feel alive across devices. Think collaborative lists, live updates to a curated shelf, or a community-driven “what people are reading now” page.&lt;/p&gt;

&lt;p&gt;When you sync client state globally over WebSockets, the UI becomes more engaging, and you cut a surprising amount of polling complexity.&lt;/p&gt;

&lt;h2&gt;
  
  
  How It Works: A Practical Flow for Building Your Own App Around Curation + AI
&lt;/h2&gt;

&lt;p&gt;Here is the approach we see work repeatedly for solo founders who want to &lt;strong&gt;build a web app&lt;/strong&gt; quickly without trapping themselves in a prototype forever.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Start With a Curated Corpus You Can Defend
&lt;/h3&gt;

&lt;p&gt;Before you optimize prompts or model choices, make the curator layer real. That can be a collection you already own (books, games, recipes) or a structured set of recommendations.&lt;/p&gt;

&lt;p&gt;The point is not volume. The point is consistency. Users will forgive that you have 300 items. They will not forgive that your “mystery” label means three different things.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Use AI for Ingestion and Metadata, Not for Taste
&lt;/h3&gt;

&lt;p&gt;AI is excellent at turning unstructured inputs into structured fields. Examples that show up in real projects:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Extracting titles and authors from book-cover photos.&lt;/li&gt;
&lt;li&gt;Suggesting tags and summaries from your curated notes.&lt;/li&gt;
&lt;li&gt;Proposing related items outside your collection, while clearly labeling them as suggestions.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you let AI decide the taste layer, you risk blending into every other recommendation product. If you use AI to &lt;em&gt;amplify your taste&lt;/em&gt;, you get differentiation.&lt;/p&gt;

&lt;p&gt;For official guidance on model capabilities and integration patterns, the &lt;a href="https://docs.claude.com/en/home" rel="noopener noreferrer"&gt;Claude developer documentation&lt;/a&gt; is a solid reference point.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Make the “First Personalization Moment” Happen Fast
&lt;/h3&gt;

&lt;p&gt;Personalization is what turns browsing into habit. The trick is to define a moment that can happen in under 60 seconds:&lt;/p&gt;

&lt;p&gt;A user picks 3 themes they like, saves 5 items, or follows 2 curators. Then you generate a tailored list immediately.&lt;/p&gt;

&lt;p&gt;This is where backend details matter. You need authentication, user data, and a place to store those events reliably. If you delay this, you end up with a pretty catalog and no retention.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Treat External Links as Product, Not Plumbing
&lt;/h3&gt;

&lt;p&gt;If your app points people to libraries or independent stores, links are not a footnote. They are part of your product ethics and your differentiation.&lt;/p&gt;

&lt;p&gt;When you integrate library access, it helps to understand how library discovery tools work. The &lt;a href="https://help.libbyapp.com/en-us/index.htm" rel="noopener noreferrer"&gt;Libby Help Center&lt;/a&gt; is useful for seeing the user flow and terminology. If you support independent bookstores, Bookshop’s mission and mechanics are laid out clearly on the &lt;a href="https://bookshop.org/info/about-us" rel="noopener noreferrer"&gt;Bookshop.org About page&lt;/a&gt;. For audiobooks that support local stores, &lt;a href="https://libro.fm/about" rel="noopener noreferrer"&gt;Libro.fm’s About page&lt;/a&gt; explains the model.&lt;/p&gt;

&lt;p&gt;From an engineering standpoint, these links imply tracking, attribution, and sometimes regional rules. That means you will want a clean data model and a safe way to generate outbound URLs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Promote the Prototype to a Real Backend Before You Add “One More Feature”
&lt;/h3&gt;

&lt;p&gt;This is the part most vibe coders try to postpone. The UI is fun. The backend feels like chores.&lt;/p&gt;

&lt;p&gt;But the moment you have:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Any kind of user-generated content&lt;/li&gt;
&lt;li&gt;A need for permissions (public vs private shelves, admin vs member)&lt;/li&gt;
&lt;li&gt;Background processing (AI enrichment, daily digests)&lt;/li&gt;
&lt;li&gt;Or even mild traction (hundreds of weekly active users)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;…you should stop bolting on ad-hoc endpoints and stabilize the foundation.&lt;/p&gt;

&lt;p&gt;This is where a managed backend helps you keep shipping. Parse is a proven model for moving fast with guardrails. If you want to understand the underlying primitives, the official &lt;a href="https://docs.parseplatform.org/" rel="noopener noreferrer"&gt;Parse Platform documentation&lt;/a&gt; is the canonical reference.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started Without Losing the Vibe
&lt;/h2&gt;

&lt;p&gt;The goal is not to “enterprise-ify” your project. It’s to keep the same creative pace, but remove the operational risks that kill momentum.&lt;/p&gt;

&lt;p&gt;A practical setup for many solo founders looks like this: a front end built in whatever stack you like, a managed backend that handles identity and data, file storage for uploads and assets, serverless functions for the few bits of custom logic that actually need code, and scheduled jobs for the repetitive work.&lt;/p&gt;

&lt;p&gt;That same setup applies whether you are creating ios apps on windows (for example, building the UI with cross-platform tooling and testing on real devices later), creating game apps (where leaderboards, inventories, and player profiles need a backend), or creating slack apps (where you store workspace installs, tokens, and event history). The surface area changes. The backend responsibilities rhyme.&lt;/p&gt;

&lt;p&gt;In &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;, we focus on those repeatable backend responsibilities so you can keep your energy on the product layer. We give you database + APIs, auth, storage/CDN, realtime, background jobs, and serverless functions that deploy in seconds in Europe and North America.&lt;/p&gt;

&lt;p&gt;If you want to go deeper on scaling patterns, our post on &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Engines and How to Scale Performance&lt;/a&gt; explains when you should add compute, what changes operationally, and how the cost model works.&lt;/p&gt;

&lt;h2&gt;
  
  
  Trade-Offs: When a Managed Backend Wins, and When It Doesn’t
&lt;/h2&gt;

&lt;p&gt;A managed backend is not the answer to every architecture problem. It wins when speed and reliability matter more than bespoke infrastructure control.&lt;/p&gt;

&lt;h3&gt;
  
  
  It Usually Wins When You Are:
&lt;/h3&gt;

&lt;p&gt;Building your own app with 1-3 people, iterating daily, and trying to get to repeat usage. It also wins when your backend needs are “standard but non-trivial,” meaning you need auth, permissions, push, storage, and jobs, but you do not want to staff DevOps.&lt;/p&gt;

&lt;p&gt;It can be especially helpful when AI costs already feel unpredictable. In that scenario, the last thing you want is a backend bill that spikes because you accidentally built an inefficient polling loop.&lt;/p&gt;

&lt;h3&gt;
  
  
  It Usually Loses When You Need:
&lt;/h3&gt;

&lt;p&gt;Deep infrastructure customization, unusual compliance constraints that require full control over every layer, or a specialized data plane (for example, heavy analytics pipelines with custom streaming infrastructure). If your team already includes experienced backend and ops engineers, you might prefer to self-host and tune everything.&lt;/p&gt;

&lt;p&gt;For founders comparing paths, we publish direct comparisons that focus on trade-offs rather than marketing. If you are considering Supabase, our &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase comparison&lt;/a&gt; is a useful checklist-style read.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Benefits for Solo Founders Creating Apps With Real Users
&lt;/h2&gt;

&lt;p&gt;Here are the benefits that tend to matter in practice, once your project moves past the demo stage.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Shorter time to real accounts&lt;/strong&gt;: shipping profiles and personalization is easier when auth and permissions are already solved.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Fewer “glue services”&lt;/strong&gt;: file storage, push, realtime, and jobs stop being separate projects.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;More predictable scaling&lt;/strong&gt;: you can handle spikes without a midnight rewrite. We have seen peaks up to 140K requests per second across the platform.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Less operational drag&lt;/strong&gt;: monitoring and a stable deployment model keeps your weekends for product work.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If high availability is a concern, our guide on &lt;a href="https://www.sashido.io/en/blog/dont-let-your-apps-down-enable-high-availability" rel="noopener noreferrer"&gt;High Availability and Zero-Downtime Components&lt;/a&gt; is a practical overview of the failure modes you are actually trying to prevent.&lt;/p&gt;

&lt;h2&gt;
  
  
  Costs and Monetization: The Numbers That Actually Matter Early
&lt;/h2&gt;

&lt;p&gt;When people ask about cost while creating apps, they often focus on the wrong number. The first real cost driver is usually not hosting. It’s iteration time.&lt;/p&gt;

&lt;p&gt;That said, you should still understand your baseline spend, because it affects whether you can keep the project alive long enough to find product-market fit.&lt;/p&gt;

&lt;p&gt;Our platform includes a 10-day free trial with no credit card required, and pricing that starts per app. The current plan details and overage rates can change, so the only reliable reference is our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;Pricing page&lt;/a&gt;. If you are cost-sensitive, look for two things: included monthly requests (so you can predict traffic costs) and included storage/transfer (so uploads do not surprise you).&lt;/p&gt;

&lt;p&gt;Monetization for curated recommendation apps usually starts in one of three ways: affiliate links, subscriptions for advanced personalization, or paid contributions for creators. The backend implications are similar regardless of model. You need user identity, event tracking, and a safe place to store billing-related state, even if billing itself is outsourced.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Quick Checklist Before You Launch to Strangers
&lt;/h2&gt;

&lt;p&gt;This is the boring checklist that prevents the most painful “we launched and it broke” moments.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Make sure user data is separated from curated content, so you can evolve each without breaking the other.&lt;/li&gt;
&lt;li&gt;Decide what is public, private, and admin-only. Then enforce it at the API level, not just in the UI.&lt;/li&gt;
&lt;li&gt;Treat ingestion as a pipeline. Photos or notes go in, structured records come out, and the process can be rerun.&lt;/li&gt;
&lt;li&gt;Add background work early. Anything AI-related that can take more than a couple seconds should not block the UI.&lt;/li&gt;
&lt;li&gt;Track the first personalization moment. If users do not hit it, your onboarding is too slow.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions About Creating Apps
&lt;/h2&gt;

&lt;h3&gt;
  
  
  How Can I Create My Own App?
&lt;/h3&gt;

&lt;p&gt;Start by defining the smallest version that proves the value, then work backward from user actions to data. For a curated-and-AI app, that means: a corpus, a way for users to save preferences, and a first personalized output. Prototype the UI fast, but promote to a &lt;a href="https://www.sashido.io/en/blog/vibe-coding-to-production-backend-reality-check" rel="noopener noreferrer"&gt;real backend&lt;/a&gt; as soon as accounts and user-generated content appear.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Much Does It Cost to Invent an App?
&lt;/h3&gt;

&lt;p&gt;“Inventing” is mostly paying for time, not servers. Early costs typically include your tooling, any AI API usage, and baseline hosting. The practical approach is to budget for 2-3 months of iteration, then choose infrastructure with predictable included quotas and clear overage rates so surprise bills do not end the project mid-test.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Much Can a 1000 Downloads App Make?
&lt;/h3&gt;

&lt;p&gt;At 1,000 downloads, revenue is usually modest unless you have strong conversion. What matters more is engagement: do users return weekly, save items, or share? If you have a 2-5% paid conversion on a $5-$10/month plan, you start to see signal. Affiliate models depend heavily on click-through and regional availability.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Backend Pieces Matter Most for Curated, AI-Assisted Apps?
&lt;/h3&gt;

&lt;p&gt;Focus on the parts that turn a catalog into a product: authentication, a flexible database for evolving metadata, file storage for ingestion, and background jobs for AI enrichment. Realtime and notifications become important once users collaborate, follow creators, or expect fresh recommendations without manually checking back.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Keep the Taste Layer Yours, and Make Everything Else Boring
&lt;/h2&gt;

&lt;p&gt;The best vibe-coded projects do not “win” because they have the most advanced model. They win because they pair a clear human point of view with an experience that is fast, personal, and reliable. If you are &lt;strong&gt;creating apps&lt;/strong&gt; in this category, protect your curation layer and use AI to scale the workflows around it. Then stabilize the backend before growth forces you into rushed rewrites.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;When you’re ready to stop wrestling with custom backends and keep building features, consider using &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; to deploy database, APIs, auth, files, realtime, jobs, and functions in minutes. You can start with a 10-day free trial and verify current plan details on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;Pricing page&lt;/a&gt;.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-ai-ready-backends" rel="noopener noreferrer"&gt;Vibe Coding and AI-Ready Backends for Rapid Prototypes&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/creating-an-app-weekend-builds-take-weeks" rel="noopener noreferrer"&gt;Creating an App in a Weekend? The 47,000-Line Reality&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-coding-ai-developer-productivity" rel="noopener noreferrer"&gt;Does AI Coding Really Boost Output?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/app-dev-vibe-coding-baas-best-practices-2025" rel="noopener noreferrer"&gt;App Development in 2025: Vibe Coding Best Practices That Still Ship&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/iphone-app-with-ai-xcode-no-code-backend" rel="noopener noreferrer"&gt;iPhone App with AI in Xcode: Build Your First MVP Fast&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>Agentic Coding Turns Vibe Prototypes Into Real Software</title>
      <dc:creator>Pavel Ivanov</dc:creator>
      <pubDate>Tue, 24 Feb 2026 07:00:25 +0000</pubDate>
      <link>https://dev.to/sashido/agentic-coding-turns-vibe-prototypes-into-real-software-5cm9</link>
      <guid>https://dev.to/sashido/agentic-coding-turns-vibe-prototypes-into-real-software-5cm9</guid>
      <description>&lt;p&gt;The most practical shift in software right now is not that engineers suddenly write 10x more code. It is that &lt;strong&gt;more people can get to a working demo&lt;/strong&gt; before an engineering sprint even starts. That is the real unlock behind &lt;a href="https://www.sashido.io/en/blog/vibe-coding-ai-ready-backends" rel="noopener noreferrer"&gt;vibe coding&lt;/a&gt;, and it is also where &lt;em&gt;agentic coding&lt;/em&gt; starts to matter.&lt;/p&gt;

&lt;p&gt;Agentic coding is what happens when you stop treating AI as a code autocomplete tool and start using it as an execution loop. You give it an outcome, it plans the steps, edits multiple files, runs checks, and iterates. In practice, it is the difference between “generate a UI” and “ship a small product slice end to end, then keep it alive.”&lt;/p&gt;

&lt;p&gt;That shift is why prototypes now show up in meetings instead of decks, and why internal teams ship tools that sat in backlogs for months. It is also why the boring parts. Auth, data models, background jobs, file storage, and auditability. suddenly decide whether your AI-built app survives real users.&lt;/p&gt;

&lt;p&gt;If you want a fast path from agentic coding experiments to something you can actually deploy, a managed backend removes a lot of failure modes early. You can start small on &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and keep your focus on the product loop, not infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Pattern: Demo First, Specification Second
&lt;/h2&gt;

&lt;p&gt;The biggest behavioral change we keep seeing is simple. When building a prototype costs 20 to 60 minutes, teams stop arguing in documents and start validating with something clickable. &lt;strong&gt;Demo, don’t memo&lt;/strong&gt; becomes the default.&lt;/p&gt;

&lt;p&gt;This does not only apply to “non-technical” folks. Product and design teams tend to do especially well because they are already trained to break ambiguity into steps, define acceptance criteria, and iterate quickly. Those are the exact muscles that make AI workflows productive.&lt;/p&gt;

&lt;p&gt;The practical outcome is that the pre-engineering bottleneck shrinks. Instead of “idea, PRD, backlog, six weeks,” it becomes “idea, working demo, feedback, then engineering.” That is why the impact often shows up first in product, design, ops, and exec workflows, even when the engineering org reports only marginal speedups.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Is Agentic Coding (And What It Is Not)
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Agentic coding&lt;/strong&gt; is a development workflow where an AI system is treated like a semi-autonomous contributor. It can decompose goals, run multi-step tasks, and iterate based on results. It is not just generating code. It is managing a loop of “plan, change, verify, repeat.”&lt;/p&gt;

&lt;p&gt;What it is not is magic. The moment your prototype needs reliable auth, safe data access, rate limits, job retries, observability, and predictable costs, the AI will not save you from missing fundamentals. It can accelerate the build. It cannot remove production constraints.&lt;/p&gt;

&lt;p&gt;A useful mental model is to split work into two tracks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Creation velocity&lt;/strong&gt;: How fast you can generate a working slice.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Operational reality&lt;/strong&gt;: How long it stays working when users, data, and edge cases arrive.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Agentic coding is great at the first. Your product stack still needs to cover the second.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where People Actually Use Vibe And Agentic Coding Today
&lt;/h2&gt;

&lt;p&gt;The most common outcomes are not huge enterprise rebuilds. They are practical, high-leverage slices of software where time-to-demo is the win.&lt;/p&gt;

&lt;h3&gt;
  
  
  Rapid Prototyping Without Waiting on Engineering
&lt;/h3&gt;

&lt;p&gt;This is the killer use case because it changes the pace of decision-making. A &lt;a href="https://www.sashido.io/en/blog/vibe-coding-mvp-parse-server-backend" rel="noopener noreferrer"&gt;clickable prototype&lt;/a&gt; reveals gaps you would not find in a document. It forces you to name the real entities and flows. Who logs in. What data must persist. What happens when you refresh. What happens on mobile.&lt;/p&gt;

&lt;p&gt;It also collapses parts of the traditional product toolchain. A prototype can function as design artifact, PRD, and validation tool. That is why design and prototyping vendors are taking AI seriously. If you want a concrete example of how incumbents frame the risk, Figma’s public filings discuss competitive pressure from rapidly evolving AI capabilities in the market, including reliance on third-party AI models and faster-moving competitors (see the risk disclosures in the &lt;a href="https://www.sec.gov/Archives/edgar/data/1579878/000162827925000272/filename1.htm" rel="noopener noreferrer"&gt;Figma S-1 on SEC.gov&lt;/a&gt;).&lt;/p&gt;

&lt;p&gt;The key constraint is that prototype speed often creates a new bottleneck. Once something exists, expectations rise. People want it deployed, shared, and stable. That is where agentic coding needs a stable backend foundation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Internal Tools That Match The Actual Process
&lt;/h3&gt;

&lt;p&gt;Internal tools are where “80% fit” SaaS breaks down. Every ops team eventually hits a workflow that is too specific for vendors to care about, but too valuable to ignore. Historically, those tools stayed in backlogs because they were not customer-facing.&lt;/p&gt;

&lt;p&gt;Now the people who need the tool can often build a first version themselves. That changes two things. First, the tool matches the real process because the builder lives inside it. Second, iteration is immediate because the feedback loop is the same person or the same team.&lt;/p&gt;

&lt;p&gt;The trap here is permissioning and &lt;a href="https://www.sashido.io/en/blog/vibe-coding-risks-technical-debt-backend-strategy" rel="noopener noreferrer"&gt;data handling&lt;/a&gt;. Internal tools tend to touch sensitive data. Payroll, customer lists, support logs, financial exports. If you do not implement access control, you are not building an internal tool. You are building an incident.&lt;/p&gt;

&lt;h3&gt;
  
  
  Turning Slide Decks Into Interactive Demos
&lt;/h3&gt;

&lt;p&gt;This is niche, but it keeps showing up because it is persuasive. If you can &lt;a href="https://www.sashido.io/en/blog/vibe-coding-with-cursors-tutorial-and-best-practices" rel="noopener noreferrer"&gt;click through a workflow&lt;/a&gt;, stakeholders understand it. If you can tailor a demo to a specific customer segment, sales cycles shorten.&lt;/p&gt;

&lt;p&gt;The operational gotcha is that “demo apps” have a habit of becoming real apps. They get forwarded, bookmarked, and used. If you did not plan for authentication, data retention, and basic security, the demo becomes a liability.&lt;/p&gt;

&lt;h3&gt;
  
  
  Replacing Simple SaaS That Almost Fits
&lt;/h3&gt;

&lt;p&gt;This is where the SaaS market gets uncomfortable. People are not rebuilding giant, deeply integrated systems. They are replacing the small tools that cost a little, frustrate daily, and do not match a specific workflow. If the product is basically CRUD plus a few rules and a dashboard, it is now in the blast radius.&lt;/p&gt;

&lt;p&gt;If you sell simple B2B software, the defense is not “AI can’t do it.” The defense is building compounding advantages AI cannot cheaply replicate. Workflow depth, distribution, compliance, integrations, data network effects, reliability, and trust.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Agentic Coding Works When You Need To Ship
&lt;/h2&gt;

&lt;p&gt;The easiest way to make agentic coding useful is to structure the work around short, verifiable loops. You want the agent to make progress in small increments you can check quickly.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Define The Slice, Not The Vision
&lt;/h3&gt;

&lt;p&gt;Agents do best when the goal is concrete. “Add onboarding with Google login and create a profile record” is better than “build an app for creators.” A slice should be testable in minutes.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Lock The Data Model Early
&lt;/h3&gt;

&lt;p&gt;Most vibe-coded apps break because data is an afterthought. The UI gets generated, then the team realizes there is no persistence plan.&lt;/p&gt;

&lt;p&gt;A good rule is to name your core objects in plain language first. Users, projects, messages, invoices, tasks. Then decide what must be queryable, what must be unique, and what must be private.&lt;/p&gt;

&lt;p&gt;If you are using MongoDB, it helps to ground the conversation in what CRUD actually means and how queries behave. MongoDB’s own manual on &lt;a href="https://www.mongodb.com/docs/manual/crud/" rel="noopener noreferrer"&gt;CRUD operations&lt;/a&gt; is a straightforward reference for the concepts you will keep bumping into when your prototype becomes a database-backed product.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Put Guardrails Around Auth And Access
&lt;/h3&gt;

&lt;p&gt;If your agentic workflow creates endpoints quickly, it can also create insecure endpoints quickly. This is where you want a checklist mindset, not a creative mindset.&lt;/p&gt;

&lt;p&gt;The &lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10&lt;/a&gt; is still the best “don’t embarrass yourself” baseline. You do not need to become a security expert overnight. You do need to ensure you are not shipping broken access control, insecure design, or misconfigurations because the prototype felt “good enough.”&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Make Realtime And Background Work Explicit
&lt;/h3&gt;

&lt;p&gt;Many modern apps feel realtime even when they are not. If you need live collaboration, presence, or instant updates, you will almost certainly end up on WebSockets. The core standard is &lt;a href="https://www.rfc-editor.org/rfc/rfc6455" rel="noopener noreferrer"&gt;RFC 6455&lt;/a&gt;, and it matters because realtime introduces state, connection management, and message validation concerns.&lt;/p&gt;

&lt;p&gt;If you need background jobs, treat them as first-class. Email sends, scheduled reports, retries, and cleanup tasks should not be “a script we run later.” Job systems need idempotency and failure handling. Agenda is a common choice in the Node ecosystem, and its &lt;a href="https://agenda.github.io/agenda/" rel="noopener noreferrer"&gt;official documentation&lt;/a&gt; is a practical reference for the scheduling model.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Production Gap: Why Prototype Wins Still Fail
&lt;/h2&gt;

&lt;p&gt;A lot of agentic builds follow the same arc. The demo is excellent. The first users arrive. Then the app hits one of these walls.&lt;/p&gt;

&lt;h3&gt;
  
  
  Wall 1: Daily Maintenance Becomes The Hidden Cost
&lt;/h3&gt;

&lt;p&gt;Even small production apps need constant care. Dependencies update. Prompts drift. Edge cases appear. A feature that looked “done” needs three more iterations because real users do not behave like the builder.&lt;/p&gt;

&lt;p&gt;The solution is not to stop using agentic coding. The solution is to reduce the surface area you maintain yourself. Offload commodity backend concerns to something stable, so your daily work stays focused on product behavior.&lt;/p&gt;

&lt;h3&gt;
  
  
  Wall 2: Security Issues Arrive Faster Than You Think
&lt;/h3&gt;

&lt;p&gt;The most dangerous part of fast generation is that it creates the illusion of completeness. A vibe-coded app can look polished and still leak data.&lt;/p&gt;

&lt;p&gt;Real incidents tend to be boring in retrospect. Publicly accessible files, weak auth checks, overly permissive roles, logs that include secrets, or endpoints that trust client input. Treat security like plumbing. You only notice it when it breaks, and when it breaks, it is expensive.&lt;/p&gt;

&lt;h3&gt;
  
  
  Wall 3: Costs Become Unpredictable
&lt;/h3&gt;

&lt;p&gt;Agentic builds often pair with heavy AI API usage. That makes cost control a product feature. You need to know your request volume, data transfer, storage growth, and any per-invocation compute.&lt;/p&gt;

&lt;p&gt;A practical way to stay sane is to decide early what you will meter and what you will cap. If you cannot answer “what happens if we get 10x usage next week,” you do not yet have a production plan.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where A Managed Backend Fits (Without Killing Momentum)
&lt;/h2&gt;

&lt;p&gt;If you are a solo founder or a small team, the goal is not to architect a perfect system. The goal is to ship something that survives contact with real users.&lt;/p&gt;

&lt;p&gt;That is the niche we built &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; for. We focus on the backend pieces that &lt;a href="https://www.sashido.io/en/blog/ai-app-builder-xcode-vibe-coding-backend-checklist" rel="noopener noreferrer"&gt;agentic coding&lt;/a&gt; workflows constantly re-create, and frequently re-create incorrectly, when they start from scratch.&lt;/p&gt;

&lt;p&gt;Every app comes with a MongoDB database plus CRUD APIs, built-in user management with social logins, storage that can serve files globally through an object store plus CDN, serverless functions you can deploy quickly in multiple regions, realtime sync over WebSockets, background jobs you can schedule and monitor, and push notifications for iOS and Android. When you are ready to go deeper, our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;developer docs&lt;/a&gt; and our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; are designed to help you move from prototype to production without a DevOps detour.&lt;/p&gt;

&lt;p&gt;If you are comparing backends, it helps to be explicit about what you want to own. If your main concern is speed-to-shipping with fewer moving parts, our comparison of &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt; is a useful starting point because it frames the trade-offs in day-to-day terms.&lt;/p&gt;

&lt;p&gt;Pricing matters too, because prototypes turn into traffic faster than expected. We keep our current plan details on the &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt; so you can always verify the latest numbers, including the 10-day free trial.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Practical Getting Started Checklist For Agentic Coding Projects
&lt;/h2&gt;

&lt;p&gt;You do not need a big process. You need a few gates that prevent the common failures.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Before you demo externally&lt;/strong&gt;: confirm auth exists, confirm private data is not public by default, confirm basic rate limits or caps exist, and confirm you can delete user data if needed.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Before you accept signups&lt;/strong&gt;: decide how you will handle password resets and social login, verify email flows, and ensure roles and permissions match your product model.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Before you add realtime&lt;/strong&gt;: define exactly what events you broadcast, validate message payloads, and decide what happens when clients reconnect.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Before you automate jobs&lt;/strong&gt;: ensure each job is idempotent, define retry behavior, and decide where you will view failures.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Before you ship mobile&lt;/strong&gt;: confirm push tokens are stored safely, opt-in is tracked, and you can segment notifications without leaking user data.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These gates do not slow you down. They stop you from re-building the same prototype twice because production requirements were discovered late.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways If You Are Building With Agentic Coding
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;The biggest value is pre-engineering speed&lt;/strong&gt;. Get to a working demo, then validate.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The risk is operational debt&lt;/strong&gt;. Maintenance, security, and cost control show up quickly.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Simple SaaS and internal tools are the first targets&lt;/strong&gt;. Not huge, mission-critical suites.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Backends are where prototypes become real&lt;/strong&gt;. Data, auth, files, jobs, and realtime cannot be an afterthought.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Is The Difference Between Vibe Coding And Agentic Coding?
&lt;/h3&gt;

&lt;p&gt;Vibe coding is usually about getting something working quickly from prompts, often optimized for speed and feel. Agentic coding is about &lt;strong&gt;running a repeatable execution loop&lt;/strong&gt; that can plan tasks, change multiple components, validate outcomes, and keep iterating. The difference shows up when you need reliability. Agentic coding is closer to “operate a project” than “generate a prototype.”&lt;/p&gt;

&lt;h3&gt;
  
  
  What Does Agentic Mean?
&lt;/h3&gt;

&lt;p&gt;In software, agentic means the system can take initiative within boundaries. It does not just answer a question or generate a snippet. It can decide the next step, perform actions, and adapt based on results. For agentic coding, that typically means it can refactor, wire components, run tests or checks, and continue until a defined goal is met.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is LLM Vs Agentic?
&lt;/h3&gt;

&lt;p&gt;An LLM is the underlying model that predicts and generates text, including code. Agentic systems use an LLM as a component inside a larger workflow that adds planning, tool use, memory, and verification loops. In practice, “LLM coding” feels like help at the keyboard. “Agentic coding” feels like delegating a task and reviewing the outcome.&lt;/p&gt;

&lt;h3&gt;
  
  
  Does ChatGPT Have Agentic Coding?
&lt;/h3&gt;

&lt;p&gt;ChatGPT can support agentic coding patterns when it is used with tools or features that allow multi-step actions, file edits, and iteration. The key is not the chat UI. It is whether the workflow supports planning, executing, and verifying changes across a project. Without that loop, you mostly get suggestions rather than autonomous progress.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Ship Faster, But Respect The Boring Parts
&lt;/h2&gt;

&lt;p&gt;Agentic coding is here to stay because it changes who can build and how quickly ideas become usable software. It compresses the time between “I think this might work” and “click it and tell me.” That is the good news.&lt;/p&gt;

&lt;p&gt;The hard truth is that the moment you cross into real users, &lt;strong&gt;backend fundamentals become the limiting factor&lt;/strong&gt;. Data persistence, access control, job reliability, realtime correctness, and predictable costs decide whether your agentic coding win is a one-week spike or the beginning of a product.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you are ready to turn agentic coding prototypes into production apps without rebuilding your backend from scratch, explore &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;. You can deploy a MongoDB-backed API, auth, storage with CDN, functions, jobs, realtime, and push notifications in minutes, then scale as usage grows.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h3&gt;
  
  
  Sources And Further Reading
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10:2021&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.rfc-editor.org/rfc/rfc6455" rel="noopener noreferrer"&gt;RFC 6455: The WebSocket Protocol&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.mongodb.com/docs/manual/crud/" rel="noopener noreferrer"&gt;MongoDB Manual: CRUD Operations&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sec.gov/Archives/edgar/data/1579878/000162827925000272/filename1.htm" rel="noopener noreferrer"&gt;Figma S-1 Filing (SEC.gov)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://agenda.github.io/agenda/" rel="noopener noreferrer"&gt;Agenda Job Scheduler Documentation&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-builder-vibe-coding-saas-backend-2025" rel="noopener noreferrer"&gt;AI App Builder vs Vibe Coding: Will SaaS End-or Just Get Rewired?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/jump-on-vibe-coding-bandwagon" rel="noopener noreferrer"&gt;Jump on the Vibe Coding Bandwagon: A Guide for Non-Technical Founders&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-vital-literacy-skill" rel="noopener noreferrer"&gt;Why Vibe Coding is a Vital Literacy Skill for Developers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;AI that writes code is now a system problem, not a tool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ctos-dont-let-ai-agents-run-the-backend-yet" rel="noopener noreferrer"&gt;Why CTOs Don’t Let AI Agents Run the Backend (Yet)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>programming</category>
      <category>ai</category>
      <category>tutorial</category>
      <category>softwaredevelopment</category>
    </item>
    <item>
      <title>Artificial Intelligence Coding Is Shrinking Teams. Adapt Fast</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Mon, 23 Feb 2026 07:00:25 +0000</pubDate>
      <link>https://dev.to/sashido/artificial-intelligence-coding-is-shrinking-teams-adapt-fast-m8g</link>
      <guid>https://dev.to/sashido/artificial-intelligence-coding-is-shrinking-teams-adapt-fast-m8g</guid>
      <description>&lt;p&gt;The most obvious change in software right now is not a new framework. It is the budget line item that keeps moving. More spend is going to GPUs, tokens, and enterprise AI licenses, and less is being reserved for headcount. That shift is why &lt;strong&gt;artificial intelligence coding&lt;/strong&gt; is showing up in board decks as a productivity lever, and why teams feel pressure to do “the same roadmap” with fewer engineers.&lt;/p&gt;

&lt;p&gt;From inside product orgs, the pattern is easy to recognize. The build is not blocked by writing endpoints anymore. It is blocked by review, integration, and reliability work that still needs humans. Engineers are being asked to become multipliers with coding AI tools, and the uncomfortable truth is that multipliers make it easier to justify smaller teams.&lt;/p&gt;

&lt;p&gt;That does not mean software work disappears. It means the work that remains gets more opinionated. People who can &lt;a href="https://www.sashido.io/en/blog/coding-agents-best-practices-plan-test-ship-faster" rel="noopener noreferrer"&gt;ship end to end&lt;/a&gt;. People who can treat AI output as a draft, then turn it into a secure system with observable behavior.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Artificial Intelligence Coding Leads to Smaller Teams
&lt;/h2&gt;

&lt;p&gt;When leadership believes AI can “speed up coding,” they often assume it affects the whole lifecycle evenly. In reality, the gains cluster in a few places. Boilerplate. First drafts. Simple refactors. Tests for known behavior. This lines up with evidence like GitHub’s controlled experiment where developers using Copilot finished a task &lt;strong&gt;&lt;a href="https://www.sashido.io/en/blog/ai-coding-tools-dynamic-context-discovery" rel="noopener noreferrer"&gt;55% faster&lt;/a&gt;&lt;/strong&gt; on average, and also reported higher satisfaction. The nuance is in the fine print. The task was scoped, the environment was controlled, and the output still needed human judgment. See GitHub’s write-up, &lt;a href="https://github.blog/news-insights/research/research-quantifying-github-copilots-impact-on-developer-productivity-and-happiness/" rel="noopener noreferrer"&gt;Research: Quantifying GitHub Copilot’s Impact on Developer Productivity and Happiness&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The second driver is organizational, not technical. If AI gives each engineer more throughput, executives can treat that as a reason to fund AI access and reduce labor cost. It is the same classic capital-to-labor tradeoff, just with token spend instead of factory machines. That tradeoff is accelerating as AI budgets rise. Even conservative forecasts show steep growth. For example, Gartner projects rapid growth in spending on generative AI models. See &lt;a href="https://www.gartner.com/en/newsroom/press-releases/2025-07-10-gartner-forecasts-worldwide-end-user-spending-on-generative-ai-models-to-total-us-dollars-14-billion-in-2025" rel="noopener noreferrer"&gt;Gartner Forecasts Worldwide End-User Spending on Generative AI Models&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The third driver is that AI changes what “a team” means. On recent earnings calls, major tech leaders have described smaller teams moving faster with AI. Meta’s leadership, for example, has talked about AI agents enabling one very capable engineer to accomplish work that previously required a larger group. See the discussion in the &lt;a href="https://www.fool.com/earnings/call-transcripts/2026/01/28/meta-meta-q4-2025-earnings-call-transcript/" rel="noopener noreferrer"&gt;Meta Platforms Q4 2025 Earnings Call Transcript&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The practical takeaway is simple. &lt;strong&gt;Artificial intelligence coding compresses the time to first working version, but it does not compress responsibility.&lt;/strong&gt; The teams that win are the ones that redesign their workflow around the new bottlenecks instead of pretending the old process still fits.&lt;/p&gt;

&lt;p&gt;If you are a solo founder or indie hacker, this is actually an opportunity. Smaller teams becoming normal means your ability to ship a &lt;a href="https://www.sashido.io/en/blog/app-dev-vibe-coding-baas-best-practices-2025" rel="noopener noreferrer"&gt;production-like demo fast&lt;/a&gt; is no longer “cute.” It is a competitive move.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Artificial Intelligence Coding Actually Works in Real Teams
&lt;/h2&gt;

&lt;p&gt;Most people describe AI-assisted development like it is autocomplete. That framing is incomplete. In practice, you are running a loop that looks like this.&lt;/p&gt;

&lt;p&gt;You describe intent in natural language. The model proposes structure and code. You validate the result against reality. Then you tighten constraints, add context, and iterate. The biggest speedups come when you already know what “correct” looks like, and you can quickly reject nonsense.&lt;/p&gt;

&lt;p&gt;This is why teams that are already strong engineers often get more value than beginners. The AI reduces typing, but it increases the amount of &lt;em&gt;judgment per minute&lt;/em&gt;. When someone says they feel like a reviewer instead of an engineer, that is not a vibe. It is the new unit of work.&lt;/p&gt;

&lt;p&gt;A useful way to think about it is to split development into three layers.&lt;/p&gt;

&lt;p&gt;At the top, there is product intent. What must the system do. Who can do what. What happens when something fails.&lt;/p&gt;

&lt;p&gt;In the middle, there is system design. Data shape. Boundaries. Permissions. How state moves between client, server, and background jobs.&lt;/p&gt;

&lt;p&gt;At the bottom, there is implementation. CRUD endpoints. serialization. Pagination. Retry logic.&lt;/p&gt;

&lt;p&gt;AI tools help most at the bottom layer, and sometimes in the middle. They do not remove the need for decisions at the top. That mismatch is where many teams get burned.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Two Loops That Matter: Generation and Verification
&lt;/h3&gt;

&lt;p&gt;Most modern coding AI tools are excellent at producing plausible code quickly. The failure mode is not that the code is obviously broken. The failure mode is that it is subtly wrong. It looks right in a diff, then fails under concurrency, weird inputs, or authorization edge cases.&lt;/p&gt;

&lt;p&gt;So the “new” work is verification. That includes security review, data correctness, and operational readiness.&lt;/p&gt;

&lt;p&gt;If you want a concrete standard for what verification needs to cover, the fastest way to align your team is to map it to a well-known framework. The &lt;a href="https://csrc.nist.gov/pubs/sp/800/218/final" rel="noopener noreferrer"&gt;NIST Secure Software Development Framework (SSDF)&lt;/a&gt; is a solid checklist of practices that remain relevant even when AI writes the first draft.&lt;/p&gt;

&lt;p&gt;Security, in particular, is where AI output can be dangerous because it tends to optimize for completion, not for threat modeling. If you need a quick reality check on the most common categories of failure, the &lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10 (2021)&lt;/a&gt; is still the most practical starting point for web apps.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where AI Tool for Coding Wins, and Where It Fails
&lt;/h2&gt;

&lt;p&gt;Used well, an ai tool for coding is like having a fast junior engineer who never sleeps and occasionally hallucinates. That analogy is not meant to be snarky. It is meant to set expectations. If you would not ship a junior engineer’s PR without review, you should not ship AI output without review either.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where It Wins
&lt;/h3&gt;

&lt;p&gt;It shines when the task is narrow and you can quickly validate output. Common examples include translating between languages, generating client SDK glue code, creating admin scripts, drafting schema migrations, and exploring alternate implementations.&lt;/p&gt;

&lt;p&gt;It also shines when you are working in unfamiliar territory and need a starting point. Many vibe coders treat the model as a map, then do the actual driving themselves.&lt;/p&gt;

&lt;p&gt;Adoption is not niche anymore. The &lt;a href="https://survey.stackoverflow.co/2024" rel="noopener noreferrer"&gt;Stack Overflow Developer Survey 2024&lt;/a&gt; reports that a large majority of developers are using or planning to use AI tools in their workflow. That is a useful signal because it means AI-generated code will increasingly be part of your dependency graph, even if you personally avoid it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where It Fails
&lt;/h3&gt;

&lt;p&gt;It fails when the constraints are implicit, undocumented, or domain-specific. Authorization logic. Billing edge cases. Idempotency. Multi-tenant data partitioning. Anything where one missing condition becomes a real incident.&lt;/p&gt;

&lt;p&gt;It also fails socially. When teams treat AI as a mandate, they end up with productivity theater. People generate more code than they can review, quality drops, and the on-call load increases. The work did not go away. It just moved from “build” to “fix.”&lt;/p&gt;

&lt;p&gt;If you are trying to decide whether you are in a safe zone, this quick check helps.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;If you cannot explain the data model and permission model in one page, AI output will probably amplify confusion, not reduce it.&lt;/li&gt;
&lt;li&gt;If you do not have a predictable release process and rollback story, faster code generation will just create faster incidents.&lt;/li&gt;
&lt;li&gt;If your system has long-running workflows, background processing, or realtime state, you need a clear plan for state persistence and retries before you let AI generate large chunks.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The New Skill Stack: What “Good Engineers” Do More Of
&lt;/h2&gt;

&lt;p&gt;When engineering teams shrink, the engineers who remain do less “make it compile” and more “make it operate.” This is the part many people miss when they search for the best ai tool for coding. The tool matters, but the workflow matters more.&lt;/p&gt;

&lt;p&gt;Here are the behaviors we see in teams that get real leverage from artificial intelligence coding.&lt;/p&gt;

&lt;p&gt;They write better prompts because they start from clearer specs. They provide examples of edge cases and failure modes. They describe expected inputs and outputs. They include constraints like latency, cost ceilings, and required audit logs.&lt;/p&gt;

&lt;p&gt;They build smaller, testable slices. Instead of asking an AI to generate a whole system, they ask for one endpoint, one background job, one permission rule, then validate.&lt;/p&gt;

&lt;p&gt;They keep a tight feedback loop with production. Observability is the difference between “AI helped” and “AI created a brittle mess.” Even a basic set of dashboards, logs, and alerts turns AI-generated code into something you can trust.&lt;/p&gt;

&lt;p&gt;They also standardize decisions that AI tends to get wrong. For example, teams often codify security defaults. Password policies. Token lifetimes. Access control rules. Data retention. You want these to be boring and consistent, not reinvented by a model on every feature.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started: A Practical Workflow for Solo Builders
&lt;/h2&gt;

&lt;p&gt;If you are a solo founder building an AI-first demo, your biggest risk is not shipping slow. It is shipping something that works once, then collapses the moment you share it with 50 people.&lt;/p&gt;

&lt;p&gt;A reliable workflow is less about your model choice and more about how you handle state, auth, files, and background tasks. That is where most prototypes die.&lt;/p&gt;

&lt;p&gt;Start with these steps, and do them in order.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;First, define what must persist. Chat history. Agent memory. User preferences. Billing state. If it matters after a refresh or after a week, it belongs in a database, not in a browser tab.&lt;/li&gt;
&lt;li&gt;Second, decide how users sign in before you build features that depend on identity. Social login is usually fine for MVPs, but your authorization rules still need to be explicit.&lt;/li&gt;
&lt;li&gt;Third, define the “slow work” path early. If you have anything that takes more than a couple seconds, you will need background jobs, retries, and status tracking.&lt;/li&gt;
&lt;li&gt;Fourth, make a plan for files and media. Demos often break because uploads are hacked in at the end.&lt;/li&gt;
&lt;li&gt;Fifth, set a cost ceiling for your cloud AI platform usage. Put rate limits in place so a viral demo does not become a surprise bill.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Once those foundations exist, &lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;artificial intelligence coding&lt;/a&gt; becomes safe to apply aggressively because it is operating inside guardrails.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where a Managed Backend Fits for Vibe Coding
&lt;/h3&gt;

&lt;p&gt;In theory, you can hand-roll all of the above. In practice, that becomes the new bottleneck, especially when you are trying to move fast with python ai coding or JavaScript-based agents and you do not want to be a part-time DevOps engineer.&lt;/p&gt;

&lt;p&gt;This is exactly the situation we built &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; for. The principle is simple. &lt;strong&gt;Spend your scarce human time on product decisions and verification, not on rebuilding the same backend plumbing.&lt;/strong&gt; Every app comes with a MongoDB database and CRUD APIs, built-in user management with social logins, file storage backed by S3 with CDN, realtime over WebSockets, scheduled and recurring jobs, and push notifications for iOS and Android.&lt;/p&gt;

&lt;p&gt;If you want to explore quickly, we keep onboarding practical with our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;SashiDo Documentation&lt;/a&gt; and a walkthrough in our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt;. When you hit performance limits, our Engines model gives you a clear scaling path and cost model. Our deep dive, &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Power Up With SashiDo’s Brand-New Engine Feature&lt;/a&gt;, explains how to scale predictably without re-architecting.&lt;/p&gt;

&lt;p&gt;If you are evaluating alternatives, it is also worth comparing tradeoffs explicitly. For example, here is our breakdown of differences in &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt;, focusing on workflow, scaling controls, and operational overhead.&lt;/p&gt;

&lt;h3&gt;
  
  
  Pricing Reality Check for Lean Teams
&lt;/h3&gt;

&lt;p&gt;When teams shrink, predictability matters more than raw power. If you are budgeting an MVP, always sanity-check current numbers on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;Pricing page&lt;/a&gt;. We also offer a 10-day free trial with no credit card required, which makes it easier to validate whether a managed backend is a fit before you commit.&lt;/p&gt;

&lt;h2&gt;
  
  
  Artificial Intelligence Coding Languages That Actually Matter
&lt;/h2&gt;

&lt;p&gt;The internet loves debating “the best language for AI,” but for most products the language choice is not the deciding factor. The deciding factor is integration speed and operational simplicity.&lt;/p&gt;

&lt;p&gt;In practice, you will see two clusters.&lt;/p&gt;

&lt;p&gt;JavaScript and TypeScript dominate when the product is web-first, the team is small, and you want to iterate quickly across frontend and serverless functions.&lt;/p&gt;

&lt;p&gt;Python dominates when the product depends on data tooling, model pipelines, or heavy use of ML libraries. That is why “artificial intelligence coding in python” is such a common path for prototypes.&lt;/p&gt;

&lt;p&gt;The mistake is treating the language as the strategy. The strategy is how you deploy and operate the system. A strong stack is one where your auth model, data model, and background processing are clear regardless of whether your AI layer is written in Python or JavaScript.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to Do When Your Team Is Half the Size
&lt;/h2&gt;

&lt;p&gt;If you wake up tomorrow and your team is smaller, the goal is not to “work twice as hard.” The goal is to &lt;strong&gt;reduce the surface area of bespoke infrastructure&lt;/strong&gt; so your remaining engineers can focus on the differentiating parts.&lt;/p&gt;

&lt;p&gt;This is where app builder platform decisions become strategic. If your backend is a weekend of glue code and infrastructure wrangling, AI will not save you. It will just help you generate more glue code.&lt;/p&gt;

&lt;p&gt;Instead, redesign around a few principles.&lt;/p&gt;

&lt;p&gt;Keep your domain logic small and boring. Push generic concerns into managed services. Make your API boundaries explicit. Instrument everything. Establish a release cadence that favors small, reversible changes.&lt;/p&gt;

&lt;p&gt;If you do that, artificial intelligence coding becomes a lever instead of a liability.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;p&gt;If you want to go deeper on the evidence and the guardrails, these are the references we regularly point teams to.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://github.blog/news-insights/research/research-quantifying-github-copilots-impact-on-developer-productivity-and-happiness/" rel="noopener noreferrer"&gt;Research: Quantifying GitHub Copilot’s Impact on Developer Productivity and Happiness&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://survey.stackoverflow.co/2024" rel="noopener noreferrer"&gt;Stack Overflow Developer Survey 2024&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://csrc.nist.gov/pubs/sp/800/218/final" rel="noopener noreferrer"&gt;NIST Secure Software Development Framework (SSDF) SP 800-218&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10 (2021)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.fool.com/earnings/call-transcripts/2026/01/28/meta-meta-q4-2025-earnings-call-transcript/" rel="noopener noreferrer"&gt;Meta Platforms Q4 2025 Earnings Call Transcript&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Does Artificial Intelligence Coding Really Reduce Engineering Headcount?
&lt;/h3&gt;

&lt;p&gt;It can, but not because AI “replaces engineers” in a clean way. It mainly reduces time spent on drafting and boilerplate, which makes it possible for leadership to run smaller teams. The work that stays is system design, verification, and operations, and those remain human-heavy.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Are the Biggest Risks When Using Coding AI Tools in Production?
&lt;/h3&gt;

&lt;p&gt;The common risks are subtle security bugs, incorrect authorization logic, and fragile integrations that pass reviews because the code looks plausible. AI also increases the volume of changes, which can overwhelm review and on-call capacity. Frameworks like NIST SSDF and OWASP Top 10 help teams keep verification disciplined.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is the Best AI Tool for Coding for a Solo Founder?
&lt;/h3&gt;

&lt;p&gt;The best tool is usually the one that fits your daily loop and reduces context switching, not the one with the biggest benchmark score. You want fast iteration, strong IDE integration, and predictable behavior on your stack. The bigger differentiator is pairing the tool with clear specs, small changes, and strong guardrails.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Does SashiDo Help When AI Speeds Up App Development?
&lt;/h3&gt;

&lt;p&gt;AI makes it easy to build features faster, but it also makes it easy to hit backend gaps sooner, like authentication, persistent state, background jobs, and file storage. Using &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; can remove a lot of that plumbing so you can focus on product logic and verification.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Artificial Intelligence Coding Needs Better Guardrails, Not More Hype
&lt;/h2&gt;

&lt;p&gt;Artificial intelligence coding is changing software economics because it increases throughput at the point of creation, and that makes smaller teams more viable. The winners will be the engineers and founders who accept the new bottleneck. Verification, security, and operations. If you build workflows that treat AI output as a draft and invest in guardrails early, you can ship faster without turning your roadmap into on-call debt.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you are trying to ship an AI-first MVP with a small team, it helps to offload the backend basics early. You can &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;explore SashiDo’s platform&lt;/a&gt; to deploy a managed backend with database, APIs, auth, jobs, realtime, and push notifications, then scale usage predictably as your demo becomes a real product.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;AI that writes code is now a system problem, not a tool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-builder-vibe-coding-saas-backend-2025" rel="noopener noreferrer"&gt;AI App Builder vs Vibe Coding: Will SaaS End-or Just Get Rewired?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/jump-on-vibe-coding-bandwagon" rel="noopener noreferrer"&gt;Jump on the Vibe Coding Bandwagon: A Guide for Non-Technical Founders&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/mcp-ai-workflows-agent-ready-backends" rel="noopener noreferrer"&gt;MCP and AI Agents: Building Agent-Ready Backends in 2025&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ctos-dont-let-ai-agents-run-the-backend-yet" rel="noopener noreferrer"&gt;Why CTOs Don’t Let AI Agents Run the Backend (Yet)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>productivity</category>
      <category>development</category>
    </item>
    <item>
      <title>Agentic Workflows: When Autonomy Pays Off and When It Backfires</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Fri, 20 Feb 2026 07:00:29 +0000</pubDate>
      <link>https://dev.to/sashido/agentic-workflows-when-autonomy-pays-off-and-when-it-backfires-27b0</link>
      <guid>https://dev.to/sashido/agentic-workflows-when-autonomy-pays-off-and-when-it-backfires-27b0</guid>
      <description>&lt;p&gt;Agentic workflows are showing up in every roadmap because they promise something every small team wants. More output without more headcount. But in production, most failures aren’t “the model was dumb.” They’re “we gave it freedom where we needed guarantees.”&lt;/p&gt;

&lt;p&gt;In a startup environment, that mistake is expensive. Autonomy usually increases latency, makes costs spikier, and complicates debugging. So the real design skill is not building agents. It’s knowing &lt;strong&gt;where discretion creates user value&lt;/strong&gt; and where it just creates new failure modes.&lt;/p&gt;

&lt;p&gt;Here’s the cleanest rule we use in practice. If a task is mostly repeatable and you can write down the steps ahead of time, a deterministic workflow beats an agent. If the task has &lt;em&gt;conditional tool use&lt;/em&gt; and the right next step depends on what the system discovers, an agentic component can earn its keep.&lt;/p&gt;

&lt;p&gt;If you’re stress-testing that boundary while building a product backend, &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; is designed to remove the “backend busywork” so you can spend time on the agent logic and evaluation instead.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Line That Matters: Who Chooses the Next Step?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;A traditional AI workflow&lt;/a&gt; can still use an LLM, but the execution path is fixed. You call the model, you take its output, and you move to the next step. That structure makes it predictable. You can reason about worst-case latency, estimate cost per request, and write monitoring that catches regressions quickly.&lt;/p&gt;

&lt;p&gt;Agentic workflows add a specific capability: the model gets to choose what happens next. It can decide to call a tool, skip a tool, ask for clarification, or loop to refine an answer. That decision power is the whole point, and it is also where systems become fragile.&lt;/p&gt;

&lt;p&gt;A helpful way to think like a cloud architect is to treat autonomy as a budget you spend. You spend it when uncertainty is high and the cost of hard-coding the logic is higher than the cost of letting the model explore.&lt;/p&gt;

&lt;h2&gt;
  
  
  When Simpler Workflows Beat Agentic Workflows
&lt;/h2&gt;

&lt;p&gt;Teams often reach for agents to cover gaps that are not really AI problems. They are product definition problems or data access problems. If you are in any of the scenarios below, keep the workflow deterministic and invest in better inputs, better guardrails, or better data.&lt;/p&gt;

&lt;p&gt;If you have a tight latency budget, deterministic usually wins. When a user is waiting on a checkout confirmation, a login flow, or a support response embedded inside a live chat, adding multiple tool calls can turn a 1 to 2 second interaction into 8 to 20 seconds. That is often the difference between “feels instant” and “feels broken.”&lt;/p&gt;

&lt;p&gt;If you need predictable cost, deterministic usually wins. Agent loops are cost multipliers. They also create tail risk, where 1 percent of requests become 20x more expensive because the model got stuck exploring.&lt;/p&gt;

&lt;p&gt;If you are in a regulated context or you have strict brand risk, deterministic usually wins. Overconfident tool-skipping is not just an accuracy issue. It is a governance issue. This is exactly the type of operational risk the &lt;a href="https://www.nist.gov/ai-risk-management-framework" rel="noopener noreferrer"&gt;NIST AI Risk Management Framework&lt;/a&gt; pushes teams to address with clear controls, measurement, and escalation paths.&lt;/p&gt;

&lt;p&gt;If your system is mostly CRUD with a little text generation, deterministic usually wins. Many “AI agents” are really a standard workflow wrapped around a prompt. That is fine. It is often the right answer.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where Agentic Workflows Actually Earn Their Complexity
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/coding-agents-best-practices-plan-test-ship-faster" rel="noopener noreferrer"&gt;Agentic workflows&lt;/a&gt; become valuable when the system must make &lt;em&gt;conditional decisions&lt;/em&gt; about which tools to use and when, and when that choice changes the outcome.&lt;/p&gt;

&lt;p&gt;A common real-world example is ambiguous research or investigation. “Why did signups drop yesterday?” is not one query. It’s a branching process. You might need to check analytics, then validate tracking changes, then correlate releases, then inspect error logs, then segment users. Hard-coding every branch becomes brittle, and human triage becomes expensive.&lt;/p&gt;

&lt;p&gt;Another example is support and operations triage. When tickets vary widely, an agent can decide whether a question is answered by docs, by an internal runbook, by a database query, or by escalation. That kind of routing can be worth the extra complexity, as long as you design for &lt;a href="https://www.sashido.io/en/blog/ctos-dont-let-ai-agents-run-the-backend-yet" rel="noopener noreferrer"&gt;safe refusal&lt;/a&gt; and clear handoffs.&lt;/p&gt;

&lt;p&gt;A third example is multi-step internal tooling, where employees accept slightly higher latency in exchange for fewer manual steps. This is where agentic workflows often feel magical, because the user is already thinking in goals, not in API calls.&lt;/p&gt;

&lt;p&gt;The principle is consistent across these scenarios. &lt;strong&gt;Autonomy helps when the next action depends on what you learn mid-flight&lt;/strong&gt;, not when you already know the steps.&lt;/p&gt;

&lt;h2&gt;
  
  
  Agentic Workflows Break for Boring Reasons
&lt;/h2&gt;

&lt;p&gt;Most agent failures are not exotic. They come from three operational issues you can observe within the first week of shipping.&lt;/p&gt;

&lt;h3&gt;
  
  
  Tool Miscalibration: The Agent “Knows” It Doesn’t Need the Tool
&lt;/h3&gt;

&lt;p&gt;If your tool descriptions are vague, the model will underuse them. If your tool descriptions are too strict, the model will overuse them and waste time. Either way, your “agent” becomes a random variable.&lt;/p&gt;

&lt;p&gt;This is why agent evaluation cannot stop at task accuracy. You also need to evaluate calibration. Does the system know when to defer, when to ask a clarifying question, and when to call a tool? In practice, we treat this as a first-class metric alongside success rate.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/ai-dev-tools-are-leaving-chat-why-claudes-cowork-signals-the-next-shift" rel="noopener noreferrer"&gt;The ReAct pattern&lt;/a&gt; is one reason tool use became mainstream. It pairs reasoning with acting in a single loop, which is useful. But it also makes it easier for teams to accidentally ship systems that &lt;em&gt;look&lt;/em&gt; intelligent while being hard to control. If you want the grounding for this idea, read the original &lt;a href="https://arxiv.org/abs/2210.03629" rel="noopener noreferrer"&gt;ReAct paper&lt;/a&gt; and notice how much of the performance comes from tool choice, not just text generation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Tool Overload: Too Many Endpoints, Too Little Intent
&lt;/h3&gt;

&lt;p&gt;Human-friendly APIs and agent-friendly APIs are not the same thing. A typical backend exposes dozens of narrowly-scoped endpoints. An agent will struggle to pick the right one unless you give it a small, well-designed surface area.&lt;/p&gt;

&lt;p&gt;A practical pattern is consolidation. Instead of separate tools for create, update, and delete, define one tool with a clear intent, a structured input schema, and explicit guidance about when to use it. This reduces hallucinated calls and makes logs easier to read.&lt;/p&gt;

&lt;p&gt;This is also where “APIs &amp;amp; auth” matter more than teams expect. The moment you let an agent act, authorization becomes part of your model interface. The difference between read-only tools and write tools needs to be explicit, because the model will not infer your security posture.&lt;/p&gt;

&lt;h3&gt;
  
  
  Observability Gaps: You Can’t Debug What You Didn’t Log
&lt;/h3&gt;

&lt;p&gt;Agents fail in sequences. If you only log the final answer, you can’t tell whether the problem was tool choice, missing context, permission errors, or a bad retry loop.&lt;/p&gt;

&lt;p&gt;In production, you want structured traces: which tools were available, which tool was selected, tool inputs and outputs, and a short reason for selection. Not because the model’s chain-of-thought should be stored verbatim, but because you need enough signal to reproduce failures.&lt;/p&gt;

&lt;h2&gt;
  
  
  Tool And API Design Patterns That Make Agents Behave
&lt;/h2&gt;

&lt;p&gt;If you only take one practical idea from this article, make it this. Treat tools as user interface. They are the buttons your agent can press.&lt;/p&gt;

&lt;p&gt;We have seen the best results from designing tools around outcomes, not around backend implementation. An “account_lookup” tool that returns a normalized account object is better than exposing five different endpoints that each return fragments. The agent’s job becomes choosing &lt;em&gt;whether&lt;/em&gt; to look up an account, not learning the quirks of your microservices.&lt;/p&gt;

&lt;p&gt;When teams ask how far to go, we suggest three constraints.&lt;/p&gt;

&lt;p&gt;First, keep the tool set small. If you need more than about 10 to 15 distinct tools for one agent role, you are probably exposing implementation details. Consolidate.&lt;/p&gt;

&lt;p&gt;Second, make tool inputs structured. Function calling and tool schemas are not just a convenience. They reduce ambiguity and improve safety. If you need a reference point, compare the behavior you get from open-ended prompts versus typed tool interfaces in &lt;a href="https://platform.openai.com/docs/guides/function-calling" rel="noopener noreferrer"&gt;OpenAI’s function calling documentation&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Third, design tools with least privilege. Start with read-only tools. Then add write tools that are scoped to safe operations, and gate the highest-risk actions behind explicit human confirmation.&lt;/p&gt;

&lt;p&gt;This is also where an application platform can save you time. When we ship systems on &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;, we can standardize a lot of the “boring but essential” surfaces quickly, including database CRUD APIs, auth, and files, so the tool layer stays consistent as the agent evolves.&lt;/p&gt;

&lt;h2&gt;
  
  
  Retrieval, Fine-Tuning, Or Tools: Pick the Cheapest Reliability
&lt;/h2&gt;

&lt;p&gt;A lot of teams start with an agent and then bolt on retrieval. Then they bolt on more tools. Then they bolt on more prompts. That can work, but it often creates a complex runtime system when a simpler training-time solution would be cheaper.&lt;/p&gt;

&lt;p&gt;Retrieval-augmented generation is a great baseline when knowledge changes frequently, and when you need citations or traceability. The original &lt;a href="https://arxiv.org/abs/2005.11401" rel="noopener noreferrer"&gt;RAG paper&lt;/a&gt; is still the clean reference for why retrieval helps factuality and coverage.&lt;/p&gt;

&lt;p&gt;Fine-tuning is often better when knowledge is stable and you care about latency. If your policies, product taxonomy, or domain language change monthly or quarterly, you can encode that behavior into the model rather than forcing a retrieval step on every request. LoRA is one of the techniques that made this accessible because it reduces training cost. See the original &lt;a href="https://arxiv.org/abs/2106.09685" rel="noopener noreferrer"&gt;LoRA paper&lt;/a&gt; for the approach.&lt;/p&gt;

&lt;p&gt;Tools are best when you need fresh state or actions. Anything involving inventory, permissions, payments, device state, or user-specific context generally belongs behind a tool call, not in training data.&lt;/p&gt;

&lt;p&gt;In practice, the decision often comes down to a few concrete constraints.&lt;/p&gt;

&lt;p&gt;If your latency budget is under 3 seconds end-to-end, be cautious with multi-step agent loops. Prefer deterministic workflows with one retrieval step, or fine-tuning for stable knowledge.&lt;/p&gt;

&lt;p&gt;If your per-request cost needs to be predictable, cap the agent. Set a maximum number of tool calls and a maximum number of iterations, then make escalation explicit.&lt;/p&gt;

&lt;p&gt;If you need a database for real time analytics, don’t make the model “guess” the state. Let it query. The right pattern is a tool that returns a small, well-structured snapshot. If you are building realtime analytics dashboards, MongoDB’s &lt;a href="https://www.mongodb.com/docs/manual/changeStreams/" rel="noopener noreferrer"&gt;Change Streams&lt;/a&gt; are an example of the underlying mechanism teams often rely on to keep state fresh.&lt;/p&gt;

&lt;h2&gt;
  
  
  Choosing the Right Level of Autonomy: A Practical Checklist
&lt;/h2&gt;

&lt;p&gt;The most effective teams treat autonomy like a spectrum, not a switch. Start deterministic, add agentic decisions where they pay off, and keep the rest boring.&lt;/p&gt;

&lt;p&gt;Use this checklist when you are deciding whether to ship an agentic component.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;If you can write the steps as a flowchart today, start deterministic. Add an agent only at the decision points where the flow branches based on new information.&lt;/li&gt;
&lt;li&gt;If the task has clear success criteria and low ambiguity, prefer a workflow. If it requires exploration and the “right next step” is context-dependent, consider an agent.&lt;/li&gt;
&lt;li&gt;If failure is high-impact, add guardrails first. Rate limits, allowlists, human confirmation for writes, and tight auth scopes matter more than clever prompts.&lt;/li&gt;
&lt;li&gt;If the system needs multiple backend calls, invest in tool design. Consolidate endpoints so the agent chooses intent, not implementation.&lt;/li&gt;
&lt;li&gt;If you cannot evaluate tool choice, do not ship autonomy. Use an evaluation harness and track not only outcomes, but also tool usage rates, refusal rates, and escalation rates.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;On evaluation specifically, it helps to follow established discipline rather than inventing your own. OpenAI’s &lt;a href="https://platform.openai.com/docs/guides/evaluation-best-practices" rel="noopener noreferrer"&gt;evaluation best practices&lt;/a&gt; and the open-source &lt;a href="https://github.com/openai/evals" rel="noopener noreferrer"&gt;OpenAI Evals framework&lt;/a&gt; are useful references for how teams structure repeatable tests and catch regressions.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Roll Out Agentic Workflows Without Betting the Company
&lt;/h2&gt;

&lt;p&gt;Most production-grade systems end up layered. Deterministic workflows handle the 80 percent path. Agentic logic handles edge cases, exploration, and triage.&lt;/p&gt;

&lt;p&gt;A rollout plan that works well in small teams starts with containment. Put the agent behind a narrow interface. Make it operate on read-only tools first. Log every tool selection. Set hard caps on loops. Add a clear fallback path that routes to deterministic behavior or to a human.&lt;/p&gt;

&lt;p&gt;Next, focus on “tool-first” user experiences. If you want an agent to help with ops, give it a small set of reliable tools with strict inputs. If you want it to help with product questions, start with retrieval over your docs and changelogs before you let it query production data.&lt;/p&gt;

&lt;p&gt;Finally, assume your backend will change. Tool contracts should be versioned, and you should expect that agent prompts and tool descriptions will need maintenance just like APIs.&lt;/p&gt;

&lt;p&gt;This is one reason Parse-based stacks keep showing up in agency work. A mature client SDK plus a stable data model makes it easier to ship and iterate across multiple apps without rebuilding auth and CRUD every time. If you are evaluating Parse Server for agencies or for a lean internal platform, our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;Parse Platform documentation&lt;/a&gt; is the best starting point because it maps client behavior, server capabilities, and deployment realities.&lt;/p&gt;

&lt;p&gt;If you do reach the point where your agent features become core product behavior, the next bottleneck is usually infrastructure consistency. You will need stable realtime, jobs, and safe deploys. Our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; shows how we structure apps so you can move from prototype to production without rebuilding the backend. When performance becomes the limiter, our &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Engines feature overview&lt;/a&gt; explains how to scale compute predictably. If uptime is the concern, our guide on &lt;a href="https://www.sashido.io/en/blog/dont-let-your-apps-down-enable-high-availability" rel="noopener noreferrer"&gt;High Availability and zero-downtime patterns&lt;/a&gt; is the pragmatic checklist we point teams to.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources And Further Reading
&lt;/h2&gt;

&lt;p&gt;The ideas above are easiest to apply when you also read the primary references behind them.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.nist.gov/ai-risk-management-framework" rel="noopener noreferrer"&gt;NIST AI Risk Management Framework&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2210.03629" rel="noopener noreferrer"&gt;ReAct: Reasoning and Acting in Language Models&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2005.11401" rel="noopener noreferrer"&gt;Retrieval-Augmented Generation (RAG)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2106.09685" rel="noopener noreferrer"&gt;LoRA: Low-Rank Adaptation of Large Language Models&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://platform.openai.com/docs/guides/evaluation-best-practices" rel="noopener noreferrer"&gt;OpenAI Evaluation Best Practices&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion: Make Agentic Workflows Earn Their Budget
&lt;/h2&gt;

&lt;p&gt;Agentic workflows can be a real advantage, but only when autonomy is doing work you cannot cheaply encode in a deterministic pipeline. When you treat tools as interface, measure calibration not just accuracy, and constrain writes with explicit permissions, you get the benefits of flexibility without turning production into a guessing game.&lt;/p&gt;

&lt;p&gt;The long-term pattern we see holding up is layered. Deterministic workflows for the happy path, agentic decisions for conditional branching, and clear escalation when uncertainty is high.&lt;/p&gt;

&lt;p&gt;If you want to build and run agentic workflows on a Parse-based application platform without taking on DevOps overhead, you can explore &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and start with our current &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing&lt;/a&gt; and the 10-day free trial.&lt;/p&gt;

&lt;h2&gt;
  
  
  FAQs
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Is an Agentic Workflow?
&lt;/h3&gt;

&lt;p&gt;An agentic workflow is a system where the model is not just generating text. It is also choosing actions, like whether to query a database, call an API, ask a follow-up question, or stop. In software teams, the defining trait is &lt;em&gt;conditional tool use&lt;/em&gt;, where the model decides the next step based on what it discovers.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is the Difference Between Agentic and Non Agentic Workflows?
&lt;/h3&gt;

&lt;p&gt;Non agentic workflows follow a fixed execution path. Even with an LLM inside, the system runs step-by-step the same way every time. Agentic workflows introduce branching and iteration controlled by the model. That flexibility helps with ambiguous tasks, but it usually costs more, adds latency, and requires stronger evaluation and guardrails.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Are the Top 3 Agentic Frameworks?
&lt;/h3&gt;

&lt;p&gt;The top three commonly used frameworks are LangGraph, Microsoft AutoGen, and Semantic Kernel. LangGraph is popular for structured multi-step flows with explicit state. AutoGen focuses on multi-agent conversation patterns. Semantic Kernel is often chosen when teams want agent orchestration integrated into existing C#, Python, or Java applications.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is the Difference Between RAG and an Agentic Workflow?
&lt;/h3&gt;

&lt;p&gt;RAG is a technique for improving answers by retrieving relevant documents at runtime and feeding them to the model. An agentic workflow is a control pattern where the model decides which actions to take, which can include retrieval, database queries, or other tools. You can use RAG inside an agent, or use RAG in a simple deterministic pipeline.&lt;/p&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/coding-agents-best-practices-plan-test-ship-faster" rel="noopener noreferrer"&gt;Coding Agents: Best practices to plan, test, and ship faster&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-development-agent-ready-apis" rel="noopener noreferrer"&gt;AI App Development Needs Agent-Ready APIs (Not “Smart” Agents)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;AI that writes code is now a system problem, not a tool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/alternatives-to-supabase-backend-as-a-service-vibe-coding" rel="noopener noreferrer"&gt;Alternatives to Supabase Backend as a Service for Vibe Coding&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/mcp-server-tutorial-reliable-ai-agents-skills-tools" rel="noopener noreferrer"&gt;MCP Server Tutorial: Make AI Agents Reliable With Skills + Tools&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>productivity</category>
      <category>softwaredevelopment</category>
      <category>development</category>
      <category>startup</category>
    </item>
    <item>
      <title>Artificial Intelligence Coding Is Turning Into Vibe Working: What Still Breaks</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Thu, 19 Feb 2026 07:00:25 +0000</pubDate>
      <link>https://dev.to/sashido/artificial-intelligence-coding-is-turning-into-vibe-working-what-still-breaks-1fj1</link>
      <guid>https://dev.to/sashido/artificial-intelligence-coding-is-turning-into-vibe-working-what-still-breaks-1fj1</guid>
      <description>&lt;p&gt;Something bigger than faster autocomplete is happening. In the last year, &lt;strong&gt;artificial intelligence coding&lt;/strong&gt; moved from “help me write this function” to “take this objective and run with it.” The same behavior is now showing up outside engineering, where people brief AI agents once, then iterate on outputs instead of building them manually.&lt;/p&gt;

&lt;p&gt;If you have been riding the &lt;em&gt;vibe coding&lt;/em&gt; wave, this shift feels familiar. You stay in flow, you describe intent in plain language, and the tool fills in the boring parts. The difference is that “vibe working” pushes that pattern into documents, analysis, planning, and operations. It also exposes a blunt reality: the bottleneck is no longer writing code. It is making AI-produced work reliable, auditable, and safe enough to ship.&lt;/p&gt;

&lt;p&gt;Here’s the first major insight we see across teams and solo builders. &lt;strong&gt;The moment an agent does multi-step work, you stop needing “more prompts” and start needing “more system.”&lt;/strong&gt; That system is usually state, identity, permissions, storage, background execution, and an API surface you can trust.&lt;/p&gt;

&lt;p&gt;If you want a quick way to de-risk early experiments, start by keeping cost and infra decisions reversible. A 10-day trial with predictable limits helps you move fast without committing early. You can check the current trial and entry plan details on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;Pricing page&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What People Mean by Vibe Working (And Why It Shows Up Now)
&lt;/h2&gt;

&lt;p&gt;Vibe working is the workplace version of vibe coding. The idea is simple. Instead of “point and click” workflows, you brief an AI agent with intent, context, and constraints, then review what it produces.&lt;/p&gt;

&lt;p&gt;In software, IBM describes vibe coding as prompting AI to generate code, then refining later, which naturally prioritizes experimentation and prototyping before optimization. That “code first, refine later” mentality is captured in IBM’s overview of &lt;a href="https://www.ibm.com/think/topics/vibe-coding" rel="noopener noreferrer"&gt;vibe coding&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Now the same pattern is being pushed into mainstream productivity tools. Microsoft has framed this as a new human-agent collaboration pattern inside Office, where Agent Mode can turn plain-language requests into spreadsheets, documents, and presentations through iterative steering. Their product direction is spelled out in &lt;a href="https://www.microsoft.com/en-us/microsoft-365/blog/2025/09/29/vibe-working-introducing-agent-mode-and-office-agent-in-microsoft-365-copilot/" rel="noopener noreferrer"&gt;Vibe Working: Introducing Agent Mode and Office Agent in Microsoft 365 Copilot&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The reason it “suddenly works” is not magic. It is a mix of better reasoning, longer context windows, and agent tooling that supports multi-step plans. The reason it “suddenly breaks” is also predictable. Once agents touch real data and real users, you inherit the same problems every production system has always had.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Hidden Trade: From Manual Effort to Operational Risk
&lt;/h2&gt;

&lt;p&gt;Vibe working often feels like free leverage. You trade time spent producing artifacts for time spent directing and reviewing them.&lt;/p&gt;

&lt;p&gt;But the trade is not purely economic. It is a shift in failure modes.&lt;/p&gt;

&lt;p&gt;When a human writes a report, most mistakes are local. A wrong number, a missing citation, a flawed assumption. When an agent produces and updates reports, pulls data, emails stakeholders, and triggers workflows, mistakes become systemic. The errors propagate, the provenance becomes fuzzy, and the blast radius increases.&lt;/p&gt;

&lt;p&gt;This is why governance and security frameworks matter even for indie builders. The NIST AI Risk Management Framework is useful here, not because it tells you how to prompt better, but because it forces you to think about measurement, monitoring, and accountability across the lifecycle. Start with the landing page for the &lt;a href="https://www.nist.gov/itl/ai-risk-management-framework" rel="noopener noreferrer"&gt;NIST AI Risk Management Framework (AI RMF 1.0)&lt;/a&gt; and treat it as a checklist for “what needs to exist before I trust an agent with real work.”&lt;/p&gt;

&lt;p&gt;At the app layer, the OWASP community has also cataloged common ways &lt;a href="https://www.sashido.io/en/blog/vibe-coding-experience-ai-tools" rel="noopener noreferrer"&gt;LLM-powered apps&lt;/a&gt; fail. The &lt;a href="https://owasp.org/www-project-top-10-for-large-language-model-applications/" rel="noopener noreferrer"&gt;OWASP Top 10 for Large Language Model Applications&lt;/a&gt; is a practical read because it maps directly to what vibe working introduces: prompt injection risks, sensitive data exposure, insecure plugin-style actions, and weak boundaries between “suggestion” and “execution.”&lt;/p&gt;

&lt;h2&gt;
  
  
  When Vibe Working Works, And When It Fails
&lt;/h2&gt;

&lt;p&gt;The most useful way to think about vibe working is not “AI replaces tasks.” It is “AI changes which constraints matter.”&lt;/p&gt;

&lt;p&gt;It tends to work best when the task has a clear objective, the inputs are constrained, and the outputs can be reviewed cheaply. It struggles when the task is ambiguous, the inputs are messy, or the outputs trigger irreversible actions.&lt;/p&gt;

&lt;p&gt;Here is a simple field-tested way to decide if a workflow is ready for agentic automation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Good candidates&lt;/strong&gt; are workflows where you can validate outcomes quickly, like drafting a spec from an outline, summarizing known documents, generating boilerplate UI, or producing an initial dashboard view. These map well to the “best AI tools for coding” category too, because the review loop is fast.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Bad candidates&lt;/strong&gt; are workflows with hidden coupling and real-world consequences, like payroll decisions, account deletions, production config changes, mass emailing, or changing permissions. The agent may be “right” most of the time, but one failure is too expensive.&lt;/p&gt;

&lt;p&gt;If you are a solo founder building a prototype, a practical threshold is this. Once you put an AI feature in front of more than &lt;strong&gt;100 to 1,000 real users&lt;/strong&gt;, you should assume you need auditability, rate limits, safe retries, and a way to reproduce what the system did.&lt;/p&gt;

&lt;h2&gt;
  
  
  Artificial Intelligence Coding in the Agent Era: What Changes for Builders
&lt;/h2&gt;

&lt;p&gt;In classic artificial intelligence coding, you implement models, data pipelines, and inference endpoints. In vibe coding, you prompt assistants to write the code.&lt;/p&gt;

&lt;p&gt;In vibe working, you are effectively building &lt;strong&gt;systems that supervise semi-autonomous work&lt;/strong&gt;. That changes what “done” means.&lt;/p&gt;

&lt;p&gt;The patterns that matter most are surprisingly non-glamorous:&lt;/p&gt;

&lt;p&gt;You need a clear identity model, so the agent is not acting as “whoever asked last.” You need state, so multi-step work can resume, retry, and explain itself. You need storage, because artifacts are not just text. They are files, logs, and attachments. You need background execution, because real work rarely fits inside a single synchronous request. And you need real-time visibility, because debugging agents is mostly about seeing what happened, not guessing.&lt;/p&gt;

&lt;p&gt;When people search “how to add backend to AI app,” this is usually what they mean, even if they phrase it as “my agent keeps forgetting things” or “my demo works but I cannot ship it.”&lt;/p&gt;

&lt;h2&gt;
  
  
  The Copilot Confusion: GitHub Copilot vs Microsoft Copilot
&lt;/h2&gt;

&lt;p&gt;A lot of teams conflate “Copilot” with a single product, then get surprised by mismatched expectations.&lt;/p&gt;

&lt;p&gt;GitHub Copilot is built for developers inside editors and code review workflows. It is best thought of as an AI pair programmer that produces and refactors code in context. The most direct reference point is the official &lt;a href="https://docs.github.com/en/copilot" rel="noopener noreferrer"&gt;GitHub Copilot documentation&lt;/a&gt;, which focuses on IDE integration, suggestion workflows, and developer experience.&lt;/p&gt;

&lt;p&gt;Microsoft Copilot is broader. It is designed for productivity work across Microsoft apps, where the outputs are spreadsheets, documents, decks, and summaries. Microsoft’s own starting point is the &lt;a href="https://support.microsoft.com/en-us/microsoft-copilot" rel="noopener noreferrer"&gt;Microsoft Copilot help center&lt;/a&gt;, which frames Copilot as a cross-app assistant rather than an IDE-first coding tool.&lt;/p&gt;

&lt;p&gt;In practice, the “github copilot vs microsoft copilot” question is less about which is better AI for coding, and more about which environment you are automating. If your work product is code, GitHub Copilot is the native fit. If your work product is Office artifacts and enterprise workflows, Microsoft Copilot is the more direct match. Many builders use both.&lt;/p&gt;

&lt;p&gt;The missing piece, for both, is still the same. You need a backend to persist decisions, manage users, enforce permissions, and turn suggestions into safe actions.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Backend Reality Check: Agents Need Memory, Not Just Context
&lt;/h2&gt;

&lt;p&gt;A lot of agent demos rely on context windows as a substitute for memory. That works until it does not.&lt;/p&gt;

&lt;p&gt;Context is what you paste in. Memory is what the system stores, retrieves, and audits over time. If you are building an AI product, you eventually need both.&lt;/p&gt;

&lt;p&gt;For example, if you are building a support assistant, you need to track user identity, consent, conversation history, escalations, and attachments. If you are building an AI content tool, you need drafts, version history, and publishing status. If you are building an agent that runs a weekly workflow, you need schedules, retries, and a place to persist intermediate outputs.&lt;/p&gt;

&lt;p&gt;This is where a managed backend matters because it removes the “I need to learn DevOps to ship a demo” tax.&lt;/p&gt;

&lt;p&gt;With &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;, we focus on the boring infrastructure that keeps agentic apps from collapsing in production. Each app comes with a MongoDB database and CRUD API, a complete user management system with social logins, file storage backed by AWS S3 with a built-in CDN, serverless functions you can deploy in seconds, realtime via WebSockets, and background jobs you can schedule and manage.&lt;/p&gt;

&lt;p&gt;If you want to go deeper on implementation details, our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;documentation and developer guides&lt;/a&gt; are the best place to understand how Parse-based backends map to modern web and mobile apps.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Practical Build Path: From Vibe Coding Prototype to Vibe Working System
&lt;/h2&gt;

&lt;p&gt;If you are using a no code AI app builder, or you are prototyping fast with prompts and generated code, you can still apply a “production readiness ladder.” You do not need to do everything on day one. You do need to do the next right thing before usage ramps.&lt;/p&gt;

&lt;p&gt;Start by making sure your AI feature has a stable interface and clear boundaries. That means defining what the agent is allowed to do, what it can only suggest, and what it must never touch without human confirmation.&lt;/p&gt;

&lt;p&gt;Then add identity and access control early, even if you only have a handful of users. The moment you demo to investors or early customers, authentication stops being “enterprise stuff” and becomes table stakes.&lt;/p&gt;

&lt;p&gt;Next, persist state outside the model. Store conversation summaries, tool outputs, and decisions as structured data. This is the difference between an agent that “feels smart” and a product that can be debugged.&lt;/p&gt;

&lt;p&gt;Then make long-running work explicit. If an agent needs to poll a feed, send push notifications, or generate a weekly report, it should run as a background job with retries and monitoring. Otherwise, you end up with fragile timeouts and ghost failures.&lt;/p&gt;

&lt;p&gt;Finally, plan for scale earlier than you think. You do not need to over-engineer, but you should know how you will scale if your demo suddenly hits 10,000 users after a launch.&lt;/p&gt;

&lt;p&gt;We have a practical walkthrough for this “from idea to deployed backend” phase in &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;SashiDo’s Getting Started Guide&lt;/a&gt; and the follow-up &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide-part-2" rel="noopener noreferrer"&gt;Getting Started Guide Part 2&lt;/a&gt;. They are written for builders who want to ship quickly without turning infrastructure into the project.&lt;/p&gt;

&lt;h2&gt;
  
  
  Cost Predictability Is Part of Reliability
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.sashido.io/en/blog/embracing-vibe-coding" rel="noopener noreferrer"&gt;Vibe working&lt;/a&gt; encourages experimentation. That is good. The trap is that experimentation can also produce unpredictable infrastructure bills, especially when agents generate more requests than humans would.&lt;/p&gt;

&lt;p&gt;The most common cost shock we see is not model spend. It is the compound effect of retries, polling, file storage growth, and “just one more integration.” That is why you should always tie agent workflows to quotas and monitoring, and choose a backend plan where you can see limits and overages up front.&lt;/p&gt;

&lt;p&gt;We keep pricing and limits transparent and up to date on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;Pricing page&lt;/a&gt;, including the free trial. If you scale beyond your base plan, you can also tune performance with compute options. Our deep dive on &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Engines and how scaling works&lt;/a&gt; explains when you actually need more horsepower and how costs are calculated.&lt;/p&gt;

&lt;h2&gt;
  
  
  Reliability Patterns That Matter More Than Better Prompts
&lt;/h2&gt;

&lt;p&gt;If you remember one thing from the vibe working shift, make it this. &lt;strong&gt;Prompting is interface design. Reliability is systems design.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;These are the patterns we recommend putting in place before you call something “ready,” especially if you plan to build an AI app that interacts with real users:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Make actions explicit&lt;/strong&gt;: separate “draft” from “send,” and “suggest” from “apply,” so an agent cannot accidentally cross the line.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Log intent and outcomes&lt;/strong&gt;: store what the agent was asked to do, what it did, and what data it touched. This is the only way to debug non-deterministic behavior.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Treat files as first-class artifacts&lt;/strong&gt;: reports, exports, and attachments need storage with stable URLs, access control, and delivery performance.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Design for retries&lt;/strong&gt;: agent workflows fail for mundane reasons like timeouts and rate limits. Your system should retry safely without duplicating side effects.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Use realtime where humans supervise&lt;/strong&gt;: when a person is steering an agent, streaming status updates prevents “black box waiting” and makes review faster.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If your product includes mobile engagement, also think about notification pipelines early. Push is often the first “real world” signal that your backend is behaving. We have written about high-volume delivery patterns in &lt;a href="https://www.sashido.io/en/blog/sending-milions-of-push-notifications-with-go-redis-and-nats" rel="noopener noreferrer"&gt;Sending Millions of Push Notifications&lt;/a&gt;, and about uptime architecture in &lt;a href="https://www.sashido.io/en/blog/dont-let-your-apps-down-enable-high-availability" rel="noopener noreferrer"&gt;High Availability and Zero-Downtime Deployments&lt;/a&gt;. Both topics become relevant surprisingly early once an agent is running unattended.&lt;/p&gt;

&lt;h2&gt;
  
  
  Tooling Choices: Avoid Lock-In, Keep Leverage
&lt;/h2&gt;

&lt;p&gt;For indie hackers and solo founders, tool choice is rarely about ideology. It is about speed now and optionality later.&lt;/p&gt;

&lt;p&gt;If you are weighing managed backends, the practical questions are. Can I ship auth, data, files, functions, realtime, and jobs without building a platform team. Can I migrate if I must. Can I predict my spend. Can I recover quickly when something breaks.&lt;/p&gt;

&lt;p&gt;If you are comparing alternatives like Supabase, Hasura, AWS Amplify, or Vercel, we recommend focusing on your actual constraints. If your AI product needs a Parse-compatible backend and you want to avoid piecing together five services, compare the trade-offs directly. For reference, here is our breakdown of &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt; and &lt;a href="https://www.sashido.io/en/sashido-vs-aws-amplify" rel="noopener noreferrer"&gt;SashiDo vs AWS Amplify&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Artificial Intelligence Coding Languages That Fit Vibe Working
&lt;/h2&gt;

&lt;p&gt;Vibe working changes what you value in a language. You want fast iteration, a strong ecosystem, and clean ways to integrate APIs, data stores, and background tasks.&lt;/p&gt;

&lt;p&gt;For most AI-first products, &lt;strong&gt;Python&lt;/strong&gt; remains the most common choice for model-adjacent work because of its ecosystem and community gravity. But in production, a lot of the glue ends up in &lt;strong&gt;JavaScript or TypeScript&lt;/strong&gt;, because web apps, dashboards, and serverless functions often live there.&lt;/p&gt;

&lt;p&gt;What matters is not winning a language debate. It is choosing a stack where you can ship a reliable surface area quickly, then optimize later. If your AI feature is primarily “agent plus workflow,” you can keep the model layer separate and focus your application layer on auth, data, files, jobs, and realtime updates.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Vibe Working Is Real, but Systems Still Decide Who Ships
&lt;/h2&gt;

&lt;p&gt;Vibe working is not a fad label. It is a reasonable description of what happens when AI agents can execute multi-step work and humans shift into steering, review, and decision-making.&lt;/p&gt;

&lt;p&gt;The builders who win with &lt;a href="https://www.sashido.io/en/blog/vibe-coding-risks-technical-debt-backend-strategy" rel="noopener noreferrer"&gt;artificial intelligence coding&lt;/a&gt; in this era will not be the ones with the fanciest prompts. They will be the ones who build &lt;em&gt;boring reliability&lt;/em&gt; around agent behavior. Identity, state, audit logs, safe actions, predictable costs, and a backend that does not require a DevOps detour.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you are moving from prompt demos to real users, it helps to stand up the backend foundations early. You can &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;explore SashiDo’s platform&lt;/a&gt; to see how database, auth, functions, jobs, realtime, storage, and push fit together in a deploy-in-minutes workflow.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;When you are ready to move from an impressive prototype to a product you can safely iterate on, deploy with &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and keep your focus on the experience, not the infrastructure. Check the current free trial and plan limits on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;Pricing page&lt;/a&gt;, then use our &lt;a href="https://www.sashido.io/en/blog/tag/getting-started" rel="noopener noreferrer"&gt;Getting Started guides&lt;/a&gt; to ship a working backend in an afternoon.&lt;/p&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  How Is Coding Used in Artificial Intelligence?
&lt;/h3&gt;

&lt;p&gt;In artificial intelligence coding, the “coding” is often the orchestration layer. You wire data ingestion, evaluation, and guardrails around a model, then expose it through APIs and UIs. In vibe working scenarios, coding is also used to persist agent state, enforce permissions, and make multi-step actions observable and reversible.&lt;/p&gt;

&lt;h3&gt;
  
  
  Is AI Really Replacing Coding?
&lt;/h3&gt;

&lt;p&gt;AI is replacing some manual typing and boilerplate, but it is not replacing the need to design systems. As agents do more end-to-end work, the hard part shifts to specifying constraints, validating outputs, and building reliable infrastructure around actions and data access. Coding becomes more about integration, safety boundaries, and operations.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Much Do AI Coders Make?
&lt;/h3&gt;

&lt;p&gt;Compensation varies widely by region and seniority, but the premium is usually tied to impact, not buzzwords. People who can ship AI features into production tend to earn more than those who only prototype, because they can handle reliability, security, and monitoring. Roles that blend backend engineering with LLM integration often price highest.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Difficult Is Artificial Intelligence Coding for a Solo Builder?
&lt;/h3&gt;

&lt;p&gt;Prototyping is easier than ever because you can use best ai for coding tools to generate scaffolding quickly. Production is still hard if you do not plan for auth, data modeling, and long-running workflows. The difficulty usually spikes when you add real users, persistent state, and background jobs, not when you write the first prompt.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.ibm.com/think/topics/vibe-coding" rel="noopener noreferrer"&gt;IBM: Vibe Coding&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.microsoft.com/en-us/microsoft-365/blog/2025/09/29/vibe-working-introducing-agent-mode-and-office-agent-in-microsoft-365-copilot/" rel="noopener noreferrer"&gt;Microsoft 365 Blog: Vibe Working and Agent Mode&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.nist.gov/itl/ai-risk-management-framework" rel="noopener noreferrer"&gt;NIST: AI Risk Management Framework (AI RMF 1.0)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://owasp.org/www-project-top-10-for-large-language-model-applications/" rel="noopener noreferrer"&gt;OWASP: Top 10 for Large Language Model Applications&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.github.com/en/copilot" rel="noopener noreferrer"&gt;GitHub Docs: GitHub Copilot&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-builder-vibe-coding-saas-backend-2025" rel="noopener noreferrer"&gt;AI App Builder vs Vibe Coding: Will SaaS End-or Just Get Rewired?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/jump-on-vibe-coding-bandwagon" rel="noopener noreferrer"&gt;Jump on the Vibe Coding Bandwagon: A Guide for Non-Technical Founders&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-vital-literacy-skill" rel="noopener noreferrer"&gt;Why Vibe Coding is a Vital Literacy Skill for Developers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;AI that writes code is now a system problem, not a tool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-to-production-backend-reality-check" rel="noopener noreferrer"&gt;Vibe Coding to Production: The Backend Reality Check&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>devops</category>
      <category>backend</category>
    </item>
    <item>
      <title>Develop Software When Your AI Model Starts Acting Like a Teammate</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Wed, 18 Feb 2026 07:00:36 +0000</pubDate>
      <link>https://dev.to/sashido/develop-software-when-your-ai-model-starts-acting-like-a-teammate-3f4d</link>
      <guid>https://dev.to/sashido/develop-software-when-your-ai-model-starts-acting-like-a-teammate-3f4d</guid>
      <description>&lt;p&gt;The fastest way to &lt;strong&gt;develop software&lt;/strong&gt; in 2026 is no longer just picking a framework. It is learning how to ship when an AI model suddenly gets better at reasoning, codebase navigation, and “doing the next step” without being asked. The teams that win these moments are not the ones with the fanciest prompts. They are the ones who can run tight early tests, connect those tests to real product data safely, and promote the winners into production without their backend becoming the bottleneck.&lt;/p&gt;

&lt;p&gt;When advanced models move from “autocomplete” to &lt;em&gt;collaborator&lt;/em&gt;, a familiar pattern shows up inside engineering orgs. People clear calendars, open a dedicated channel, and throw the hardest problems first. Not because it is fun, but because it is the only honest way to learn where the model helps, where it breaks, and what you need to change in your app to benefit.&lt;/p&gt;

&lt;p&gt;In practice, the biggest unlock is not that the model writes more code. It is that the model starts finishing multi-step tasks end to end. That changes how your team plans work, how you test changes, and how you design your startup backend infrastructure so it can survive the new pace.&lt;/p&gt;

&lt;p&gt;A concrete example: one team finally had a recurring UI analytics bug diagnosed on the first attempt after five-plus failures with an older model. The fix was not “smarter code generation.” It was spotting &lt;a href="https://www.sashido.io/en/blog/ai-dev-tools-are-leaving-chat-why-claudes-cowork-signals-the-next-shift" rel="noopener noreferrer"&gt;eight parallel API searches&lt;/a&gt; firing at once, plus calls bypassing rate limiting by using a raw HTTP client instead of the project’s guarded wrapper. The model was useful because it saw the system behavior, not just the local file.&lt;/p&gt;

&lt;p&gt;If you are running these AI upgrade sprints, you will move faster when your test apps can authenticate real users, store files, run background jobs, and stream realtime updates without you rebuilding infrastructure each time. For &lt;a href="https://www.sashido.io/en/blog/ai-powered-backend-mobile-app-development-speed" rel="noopener noreferrer"&gt;Parse-based projects&lt;/a&gt;, our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; is the shortest path we know to stand up those moving parts cleanly.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Early-Access Model Testing Really Teaches Teams
&lt;/h2&gt;

&lt;p&gt;These short pre-launch windows surface the same two truths again and again.&lt;/p&gt;

&lt;p&gt;First, &lt;strong&gt;benchmarks and “vibe checks” measure different things&lt;/strong&gt;. Benchmarks tell you if the model clears a known bar. Hands-on building tells you if it feels reliable under messy reality, like half-migrated code, inconsistent naming, flaky third-party APIs, and product requirements that change mid-task.&lt;/p&gt;

&lt;p&gt;Second, the moment the model feels more autonomous, your constraints shift from “can it write this” to “can our product safely accept what it produces.” That is where operational discipline matters. You need isolation, repeatability, and rollback. Otherwise, you end up with impressive demos that cannot be shipped.&lt;/p&gt;

&lt;p&gt;A good mental model is to treat early-access testing like a release candidate for a dependency you cannot fully control. The right stance is: measure, stress, constrain, then promote.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Further reading:&lt;/strong&gt; if you want the official framing of the model changes themselves, start with &lt;a href="https://www.anthropic.com/news/claude-opus-4-6" rel="noopener noreferrer"&gt;Anthropic’s Claude Opus 4.6 announcement&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Develop Software During a Model Early-Access Sprint
&lt;/h2&gt;

&lt;p&gt;When we see teams do this well, they follow a simple loop. They do not over-intellectualize it. They just make it repeatable.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Start With Your Hardest “Production-Like” Tasks
&lt;/h3&gt;

&lt;p&gt;Good tests are the ones that reflect how you actually develop software. They are rarely toy problems.&lt;/p&gt;

&lt;p&gt;A few examples that consistently expose model strengths and weak spots:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://www.sashido.io/en/blog/ai-assisted-coding-vibe-projects-2026" rel="noopener noreferrer"&gt;A stubborn bug&lt;/a&gt; that spans frontend, API usage, and rate limiting, because it forces the model to reason about system behavior.&lt;/li&gt;
&lt;li&gt;A real refactor that moves functionality between modules without breaking navigation, auth flows, or permissions.&lt;/li&gt;
&lt;li&gt;A library port or cross-language translation that must match existing tests, because it exposes instruction-following under constraints.&lt;/li&gt;
&lt;li&gt;A feature that looks “simple” in text but touches design details you did not specify, because it reveals whether the model productively fills in blanks or invents risky assumptions.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Step 2: Separate “Scoring” From “Feeling”
&lt;/h3&gt;

&lt;p&gt;Teams that only trust dashboards miss issues that show up in human use. Teams that only trust vibe checks get fooled by novelty.&lt;/p&gt;

&lt;p&gt;A practical split:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Your structured evals should be small, stable, and run every time you change prompts, tools, or context packing.&lt;/li&gt;
&lt;li&gt;Your hands-on building sessions should be time-boxed and documented with concrete observations, like failure modes, hallucination triggers, and the exact tool calls that went wrong.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is also where you decide what “ship ready” means. For many product teams, it is not “the model is correct.” It is “the model is correct &lt;em&gt;within our guardrails&lt;/em&gt;.”&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Make Tool Access Explicit, Auditable, and Reversible
&lt;/h3&gt;

&lt;p&gt;As soon as the model can browse, call tools, or update data, you need a hard line between:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The model reasoning about data.&lt;/li&gt;
&lt;li&gt;The system actually mutating data.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In early testing, the easiest mistake is giving the model a powerful admin token because “it is just a staging app.” That is how staging becomes production by accident.&lt;/p&gt;

&lt;p&gt;Use common standards and keep them boring. For example, build around OAuth scopes and explicit grants as described in &lt;a href="https://www.rfc-editor.org/rfc/rfc6749" rel="noopener noreferrer"&gt;RFC 6749&lt;/a&gt;, and treat realtime connections as first-class security surfaces as described in &lt;a href="https://www.rfc-editor.org/rfc/rfc6455" rel="noopener noreferrer"&gt;RFC 6455&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Bottleneck: Shipping the AI Output Into the Product
&lt;/h2&gt;

&lt;p&gt;Once you get a model that can diagnose a complex bug quickly, or port a large library while preserving tests, your throughput increases. Your bottleneck often shifts to integration work that used to be “background noise.”&lt;/p&gt;

&lt;p&gt;This is where startup teams feel pain first.&lt;/p&gt;

&lt;p&gt;You want to stand up a handful of &lt;a href="https://www.sashido.io/en/blog/backend-as-a-service-claude-artifacts-to-production" rel="noopener noreferrer"&gt;test apps&lt;/a&gt; quickly, each with a clean dataset. You need authentication because internal testers cannot all share one admin account. You need file storage because AI features increasingly involve uploads. You need scheduled jobs because the “assistant” becomes a queue of long-running tasks. You need push notifications because users expect to be re-engaged when a task is done.&lt;/p&gt;

&lt;p&gt;If your team is 3 to 20 people, the hidden cost is not the cloud bill. It is the hours burned maintaining these basics while you are trying to validate whether the AI feature even works.&lt;/p&gt;

&lt;p&gt;This is exactly the gap a backend-as-a-service platform is supposed to close. The trick is choosing one that does not trap you, and that scales predictably when your AI feature turns a calm traffic pattern into bursts.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where a Managed Backend Fits, and Where It Does Not
&lt;/h2&gt;

&lt;p&gt;A managed backend is not magic. It is a trade.&lt;/p&gt;

&lt;p&gt;You trade some low-level infrastructure control for speed, standardization, monitoring, and a much smaller operational surface. That is valuable when you are running frequent experiments, especially when model behavior changes quickly.&lt;/p&gt;

&lt;p&gt;It is a weaker fit when you have strict requirements that only custom infrastructure can satisfy, like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Extremely specialized networking or data residency constraints that require custom VPC topology.&lt;/li&gt;
&lt;li&gt;Deep, bespoke database tuning and query planners that your team wants to own end to end.&lt;/li&gt;
&lt;li&gt;A need for full control over every component because you are running an internal platform team.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For most early-stage product teams, the real question is not “managed vs self-hosted.” It is &lt;strong&gt;when to keep velocity, and when to buy back control&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;A practical threshold we see is this: if you are still changing your data model weekly, and your roadmap depends on shipping AI-connected features fast, managed services usually win. When you stabilize and start optimizing for cost and tail latency at very high scale, you may selectively bring pieces in-house.&lt;/p&gt;

&lt;p&gt;If you are currently comparing options, and Supabase is on your shortlist, our take is nuanced. It is a strong tool. But the decision depends on your appetite for ops and your desired portability. Here is our direct comparison so you can evaluate trade-offs quickly: &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Connecting Early AI Tests to a Real Backend Without DevOps Overhead
&lt;/h2&gt;

&lt;p&gt;Once the principle is clear, here is how we think about it inside &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;When teams are trying to develop software quickly during model shifts, the backend work that slows them down is usually not “build a database.” It is everything around it: auth, file delivery, realtime sync, job scheduling, push, and the day-two concerns like monitoring, logs, and predictable scaling.&lt;/p&gt;

&lt;p&gt;We built our platform around a Parse-compatible core, with a MongoDB database and CRUD APIs per app, plus built-in user management and social logins. That matters in AI test loops because you can spin up multiple apps for parallel experiments, keep datasets separated, and still use the same client SDK patterns. If you want the full technical surface, our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;documentation&lt;/a&gt; lays out the Parse Platform APIs, SDKs, and operational guides.&lt;/p&gt;

&lt;p&gt;File-heavy AI features are another common speed bump. Even a “simple” assistant quickly turns into uploading PDFs, images, audio, or generated exports. We use an AWS S3 object store behind the scenes, and the reason it works well is that S3 is designed to be boring, durable infrastructure at massive scale. If you want the canonical reference for the underlying storage model, see the &lt;a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/Welcome.html" rel="noopener noreferrer"&gt;Amazon S3 User Guide&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Realtime is the third area that changes the feel of AI features. Users expect a progress stream, not a spinner that times out. When your client state needs to sync over WebSockets, the protocol-level constraints are not optional, and they show up under load. The WebSocket spec in &lt;a href="https://www.rfc-editor.org/rfc/rfc6455" rel="noopener noreferrer"&gt;RFC 6455&lt;/a&gt; is still the best way to align your expectations with reality.&lt;/p&gt;

&lt;p&gt;Finally, AI product flows almost always need background work. Summaries, indexing, webhooks, retries, and scheduled maintenance are job-shaped problems. The scheduler we rely on is based on MongoDB and Agenda, and the upstream project is well documented. If you want to understand the model of recurring jobs and locking, Agenda’s &lt;a href="https://github.com/agenda/agenda" rel="noopener noreferrer"&gt;official repository&lt;/a&gt; is the clearest reference.&lt;/p&gt;

&lt;h3&gt;
  
  
  Scaling Without Guesswork When Your Traffic Becomes Spiky
&lt;/h3&gt;

&lt;p&gt;Model-connected features often create bursty demand. A demo gets shared. A new assistant feature triggers users to upload files in batches. A “design uplift” release sends more interactive sessions through realtime.&lt;/p&gt;

&lt;p&gt;The practical thing to plan for is not average traffic. It is peaks. If you have ever watched a graph jump from calm to chaos, you know that capacity planning for the mean is a trap.&lt;/p&gt;

&lt;p&gt;That is why we built Engines. It lets you scale compute without rebuilding your stack, and it gives you a clear cost model for different performance profiles. If you want the deeper mechanics, our post on &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;the Engine feature and how scaling works&lt;/a&gt; explains when to upgrade and how pricing is calculated.&lt;/p&gt;

&lt;p&gt;We also see teams underestimate the cost of downtime during high-attention moments. If your AI feature goes viral and your backend falls over, the issue is rarely “one bug.” It is usually missing redundancy and deployment safety. If uptime is becoming existential, our guide on &lt;a href="https://www.sashido.io/en/blog/dont-let-your-apps-down-enable-high-availability" rel="noopener noreferrer"&gt;high availability and self-healing setups&lt;/a&gt; is a good map of what to harden first.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Practical Checklist for CTOs Shipping AI-Connected Features
&lt;/h2&gt;

&lt;p&gt;If you want a concise way to operationalize all of this, here is the checklist we recommend for small teams.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Decide what counts as a “hard test” for your app, and pick 3 to 5 tasks that are representative. Include at least one cross-cutting bug, one refactor, and one long-running workflow.&lt;/li&gt;
&lt;li&gt;Separate your eval results from your hands-on building notes. Treat them as complementary, not competing.&lt;/li&gt;
&lt;li&gt;Put your model behind explicit permissions. Never let early tests run with admin tokens by default. Make every data mutation reversible.&lt;/li&gt;
&lt;li&gt;Use separate apps or environments for parallel experiments, and keep datasets isolated so you can compare results cleanly.&lt;/li&gt;
&lt;li&gt;Add observability early. If you cannot explain why a job was retried or why a realtime connection dropped, you will not trust your own AI feature in production.&lt;/li&gt;
&lt;li&gt;Plan for spikes. If you only test at 1x traffic, you will ship a feature that works until it is popular.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you are using Parse, it is worth grounding in the upstream ecosystem once, because it makes portability discussions with investors much easier. The &lt;a href="https://website.parseplatform.org/" rel="noopener noreferrer"&gt;Parse Platform project&lt;/a&gt; is the canonical reference for what “Parse-compatible” means.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Develop Software Faster by Making AI Testing Shippable
&lt;/h2&gt;

&lt;p&gt;When models become stronger, the temptation is to treat the upgrade as a prompt problem. The teams that ship treat it as a systems problem. They build a repeatable loop, they stress real tasks first, and they invest in the boring plumbing that turns AI output into product behavior.&lt;/p&gt;

&lt;p&gt;To &lt;strong&gt;develop software&lt;/strong&gt; reliably in this new rhythm, you need two things at once: an evaluation discipline that tells you what the model is doing, and a backend that lets you deploy experiments and promote them safely. When your small team is already stretched, paying the DevOps tax for every new AI workflow is the slow path.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you want to connect early-access AI tests to a real backend quickly, you can explore &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt;. We deploy database, APIs, auth, storage, realtime, background jobs, and serverless functions in minutes, and you can start with a 10-day free trial. For current plan details, always check our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt; since limits and rates can change.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  How Do You Develop Software?
&lt;/h3&gt;

&lt;p&gt;Developing software is a loop of defining a problem, building the smallest useful slice, and validating it with real users. In AI-connected products, add one more loop: evaluate model behavior with repeatable tests before you ship. This keeps improvements real, and prevents the model from silently changing your app’s reliability.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is a Synonym for Developed Software?
&lt;/h3&gt;

&lt;p&gt;In engineering discussions, people often say production-ready software, shipped software, or deployed application. The best synonym depends on what you mean: production-ready emphasizes stability and support, while shipped emphasizes delivery. In AI-heavy projects, deployed application also implies the backend, auth, jobs, and monitoring are in place.&lt;/p&gt;

&lt;h3&gt;
  
  
  When Does a Managed Backend Beat Self-Hosting for AI Features?
&lt;/h3&gt;

&lt;p&gt;Managed backends usually win when you are iterating quickly and your data model is still changing, especially if your team has no dedicated DevOps. They reduce setup time for auth, storage, jobs, and realtime, which AI workflows depend on. Self-hosting becomes more attractive when you need bespoke infrastructure control or very specialized tuning.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Breaks First When You Add AI Agents to a Live App?
&lt;/h3&gt;

&lt;p&gt;Most teams first hit limits in long-running work and spiky traffic. AI features create queues, retries, and background tasks, then users expect realtime progress and notifications. The second failure mode is unsafe permissions, where tools are too powerful in testing and accidentally leak into production. Guardrails and environment isolation prevent both.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.anthropic.com/news/claude-opus-4-6" rel="noopener noreferrer"&gt;Claude Opus 4.6 Announcement&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.rfc-editor.org/rfc/rfc6749" rel="noopener noreferrer"&gt;RFC 6749: The OAuth 2.0 Authorization Framework&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.rfc-editor.org/rfc/rfc6455" rel="noopener noreferrer"&gt;RFC 6455: The WebSocket Protocol&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/Welcome.html" rel="noopener noreferrer"&gt;Amazon S3 User Guide&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://github.com/agenda/agenda" rel="noopener noreferrer"&gt;Agenda Job Scheduler (GitHub)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://website.parseplatform.org/" rel="noopener noreferrer"&gt;Parse Platform&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-engineering-prompts-to-production" rel="noopener noreferrer"&gt;What Is BaaS in Vibe Engineering? From Prompts to Production&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-coding-ai-developer-productivity" rel="noopener noreferrer"&gt;Does AI Coding Really Boost Output?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-coding-tools-dynamic-context-discovery" rel="noopener noreferrer"&gt;AI coding tools: dynamic context discovery to cut tokens and ship&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/code-sandbox-options-for-ai-agents" rel="noopener noreferrer"&gt;Code Sandbox Options for AI Agents: 5 Ways to Run Generated Code Safely&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;AI that writes code is now a system problem, not a tool&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>softwaredevelopment</category>
      <category>programming</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Artificial Intelligence Coding: When Vibe Coding Becomes Agentic Engineering</title>
      <dc:creator>Vesi Staneva</dc:creator>
      <pubDate>Tue, 17 Feb 2026 07:00:44 +0000</pubDate>
      <link>https://dev.to/sashido/artificial-intelligence-coding-when-vibe-coding-becomes-agentic-engineering-5ffb</link>
      <guid>https://dev.to/sashido/artificial-intelligence-coding-when-vibe-coding-becomes-agentic-engineering-5ffb</guid>
      <description>&lt;p&gt;A year ago, a lot of &lt;strong&gt;artificial intelligence coding&lt;/strong&gt; looked like a dare. You accepted whole diffs from tools like Cursor, pasted stack traces into a chat, and kept going until the demo worked. It felt like speedrunning software.&lt;/p&gt;

&lt;p&gt;Now the same workflow is showing up in real products, with real users, and real consequences. The shift is not that AI writes code. It is that builders are increasingly &lt;strong&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-experience-ai-tools" rel="noopener noreferrer"&gt;orchestrating agents&lt;/a&gt;&lt;/strong&gt; that write code, wire systems, and propose changes, while the human sets constraints, checks the seams, and decides what ships.&lt;/p&gt;

&lt;p&gt;That changes the skill stack. You still need taste and architecture, but you also need an operating model for quality. Otherwise, the exact thing that makes AI for code generation feel magical, the ability to move fast without understanding every line, becomes the thing that breaks you in production.&lt;/p&gt;

&lt;p&gt;If you are a solo founder or indie hacker doing cursor vibe coding for an MVP, the practical question is simple. &lt;strong&gt;How do you keep the leverage, but stop the backend and reliability debt from compounding?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;A lightweight way to start is to put your &lt;a href="https://www.sashido.io/en/blog/vibe-coding-risks-technical-debt-backend-strategy" rel="noopener noreferrer"&gt;data model, auth, and API surface&lt;/a&gt; on rails early. If you want that without running servers, you can build on &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and keep your “agentic” energy focused on product.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Real Pattern Behind Vibe Coding
&lt;/h2&gt;

&lt;p&gt;The pattern we see repeatedly is not that people suddenly became careless. It is that modern AI tools made a new loop viable.&lt;/p&gt;

&lt;p&gt;You describe intent. The agent proposes code. You run it, observe behavior, and feed back constraints. That loop can turn a weekend prototype into something demo-able in hours.&lt;/p&gt;

&lt;p&gt;The trap is that the loop rewards “Accept All” behaviors early. You are optimizing for visible progress, not for &lt;strong&gt;maintainability, security boundaries, or operability&lt;/strong&gt;. The moment you cross into “real users,” that optimization flips. Every unclear data shape, every missing access rule, and every unbounded request path turns into a late-night incident.&lt;/p&gt;

&lt;p&gt;You can feel the industry acknowledging this shift. Satya Nadella has publicly said a meaningful portion of Microsoft’s code is now AI-generated, and he discussed the variability across languages and contexts in a public interview covered by &lt;a href="https://techcrunch.com/2025/04/29/microsoft-ceo-says-up-to-30-of-the-companys-code-was-written-by-ai/" rel="noopener noreferrer"&gt;TechCrunch&lt;/a&gt;. That is the signal. The leverage is real, but so is the need for engineering discipline.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Agentic Engineering Changes Artificial Intelligence Coding
&lt;/h2&gt;

&lt;p&gt;Agentic engineering is not a new programming language. It is a new division of labor.&lt;/p&gt;

&lt;p&gt;Instead of writing most lines yourself, you spend more time doing three things.&lt;/p&gt;

&lt;p&gt;First, you define the “rails”. You decide what is allowed. That includes your data model, auth model, API boundaries, rate limits, and storage rules.&lt;/p&gt;

&lt;p&gt;Second, you supervise the agent. You review diffs, but you also review &lt;em&gt;intent&lt;/em&gt;. You ask whether this change creates a new dependency, a new trust boundary, or a new failure mode.&lt;/p&gt;

&lt;p&gt;Third, you instrument the system so you can recover. When an agent-written feature fails in production, you need logs, reproducible jobs, and a way to roll forward or roll back.&lt;/p&gt;

&lt;p&gt;A useful mental model is that the agent is an extremely fast junior developer with infinite energy and imperfect judgment. Your job is to make it hard to do unsafe things, and easy to do the safe thing.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where “Accept All” Still Works
&lt;/h3&gt;

&lt;p&gt;There are places where vibe coding is still the right move. Landing pages, internal tooling, one-off scripts, and UI experimentation are often fine. If the worst-case failure is an ugly component, speed wins.&lt;/p&gt;

&lt;h3&gt;
  
  
  Where It Fails Quickly
&lt;/h3&gt;

&lt;p&gt;The failure zone usually starts when you add any of the following.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Authentication and user data&lt;/li&gt;
&lt;li&gt;Payments or anything that can be abused as a business flow&lt;/li&gt;
&lt;li&gt;Public APIs, webhooks, or integrations&lt;/li&gt;
&lt;li&gt;Background work, scheduled tasks, or anything that can run unbounded&lt;/li&gt;
&lt;li&gt;Multi-tenant data, where one user must never see another user’s records&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you have even 50 to 100 active users, or you are sending traffic from a public launch, these issues appear fast. The “it works on my machine” phase ends, and the “it worked yesterday” phase begins.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Practical Guardrail Checklist for AI for Code Generation
&lt;/h2&gt;

&lt;p&gt;When you are moving fast with best AI tools for coding, the goal is not to add bureaucracy. The goal is to add &lt;strong&gt;small, high-leverage constraints&lt;/strong&gt; that stop the worst mistakes.&lt;/p&gt;

&lt;p&gt;Here is the checklist we use internally when we watch teams graduate from throwaway vibe coding to something you can operate.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data contracts first&lt;/strong&gt;: write down what a user object, session object, and core domain objects look like, including required fields and ownership.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Auth and authorization as separate work&lt;/strong&gt;: AI is good at auth UI, but authorization bugs are subtle. Decide object-level rules up front.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Bounded inputs&lt;/strong&gt;: every endpoint needs size limits, pagination defaults, and rate-limiting assumptions.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Observability minimum&lt;/strong&gt;: log request IDs, user IDs (when safe), and failure reasons. Make background tasks emit structured status.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Failure modes by design&lt;/strong&gt;: decide what happens when the model call fails, times out, or returns malformed output.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you want a single security anchor for API work, the &lt;a href="https://owasp.org/API-Security/editions/2023/en/0x11-t10/" rel="noopener noreferrer"&gt;OWASP API Security Top 10 (2023)&lt;/a&gt; is still the most useful reality check. It is not “AI specific,” but AI-generated code tends to accidentally recreate classic mistakes like broken authorization or unrestricted resource consumption.&lt;/p&gt;

&lt;h2&gt;
  
  
  Artificial Intelligence Coding Languages: What Actually Matters in 2026
&lt;/h2&gt;

&lt;p&gt;People ask about “the” &lt;a href="https://www.sashido.io/en/blog/ai-assisted-coding-vibe-projects-2026" rel="noopener noreferrer"&gt;artificial intelligence coding&lt;/a&gt; language, but in practice you are balancing three constraints. Library ecosystems, performance requirements, and how well your tooling supports agentic workflows.&lt;/p&gt;

&lt;p&gt;Python stays dominant for model work because the ecosystem is unmatched for experimentation. JavaScript and TypeScript dominate product glue because they sit closest to web and mobile experiences, and because agents can rewrite UI and API wiring quickly.&lt;/p&gt;

&lt;p&gt;If you are building an AI-first app, the most common split is simple. Keep model interaction and evaluation logic in Python where it is convenient, and keep product and orchestration logic in JavaScript or TypeScript where it is shippable.&lt;/p&gt;

&lt;p&gt;The key point is not which language you pick. It is whether you can enforce consistent patterns around data access, secrets, background work, and state across sessions. This is the part vibe coding often skips.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started: Turning Vibe Coding Cursor Projects Into Production Work
&lt;/h2&gt;

&lt;p&gt;If you already have a prototype, the fastest “graduation path” is to stabilize three things before you add more features.&lt;/p&gt;

&lt;h3&gt;
  
  
  1) Make State Real, Not Implicit
&lt;/h3&gt;

&lt;p&gt;Most agent-built demos hide state in local files, in-memory maps, or a loosely defined JSON blob. That is fine until you need multi-device logins, auditability, or recovery.&lt;/p&gt;

&lt;p&gt;Pick a real database model and move the core objects there. If you do this early, your agents will start generating code against stable schemas instead of inventing new shapes every time.&lt;/p&gt;

&lt;h3&gt;
  
  
  2) Put Auth on Rails
&lt;/h3&gt;

&lt;p&gt;In demos, auth is often bolted on at the end. In real apps, auth becomes the root of your data boundaries, rate limits, and abuse prevention.&lt;/p&gt;

&lt;p&gt;If you want to avoid building this from scratch, we designed &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; so every app starts with MongoDB plus a CRUD API and a complete user management system. Social logins are a click away for providers like Google, GitHub, and many others, which is a huge time saver when your AI agent keeps refactoring your UI.&lt;/p&gt;

&lt;p&gt;For implementation details, our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;developer docs&lt;/a&gt; are the canonical reference, and the &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; shows the shortest path from project creation to a running backend.&lt;/p&gt;

&lt;h3&gt;
  
  
  3) Externalize Background Work
&lt;/h3&gt;

&lt;p&gt;Agentic apps quickly grow “invisible features”. Sync jobs, scheduled runs, post-processing, and notification fanout.&lt;/p&gt;

&lt;p&gt;If those tasks are tied to a laptop or a single web process, you will see nondeterministic behavior. Move them into scheduled and recurring jobs with clear inputs and outputs. If you are building on our platform, you can run jobs with MongoDB and Agenda and manage them from our dashboard, so the work stays observable even when the code was mostly generated.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Backend Problem Agentic Apps Keep Rediscovering
&lt;/h2&gt;

&lt;p&gt;Most AI-first MVPs have the same backend-shaped problems, regardless of whether you used a no code app builder, wrote everything manually, or leaned on an agent.&lt;/p&gt;

&lt;p&gt;They need a place to store user state across sessions. They need an API layer that enforces access rules. They need file storage for user uploads, artifacts, or model outputs. They need realtime updates when long tasks complete. They need push notifications to re-engage users.&lt;/p&gt;

&lt;p&gt;This is exactly where “backend as a product” saves the most time, because it removes the slowest parts of early productionization. The first time you feel it is when your demo becomes a real app and you stop wanting to babysit a server.&lt;/p&gt;

&lt;p&gt;If you are curious what “files” looks like at scale, we wrote up why we use S3 plus a built-in CDN in &lt;a href="https://www.sashido.io/en/blog/announcing-microcdn-for-sashido-files" rel="noopener noreferrer"&gt;Announcing MicroCDN for SashiDo Files&lt;/a&gt;. It is a good example of the behind-the-scenes engineering that vibe coding workflows usually do not cover.&lt;/p&gt;

&lt;h2&gt;
  
  
  Cost, Reliability, and the Point Where You Need Real Scaling
&lt;/h2&gt;

&lt;p&gt;AI-first builders often underestimate two costs.&lt;/p&gt;

&lt;p&gt;The obvious cost is model inference. The hidden cost is infrastructure unpredictability caused by unbounded endpoints, retries, and background tasks that scale accidentally.&lt;/p&gt;

&lt;p&gt;A simple rule works well. If you cannot estimate your “requests per user per day” within a factor of 3, you do not yet control your backend costs. Before you optimize model spend, you should bound and measure backend spend.&lt;/p&gt;

&lt;p&gt;On our side, we make pricing transparent and app-scoped, but details can change. If you are evaluating budgets, always check the current numbers on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt;. At the time of writing, the entry plan includes a free trial and a low monthly per-app starting price, with metered overages for extra requests, storage, and transfer.&lt;/p&gt;

&lt;p&gt;When you hit real traction, scaling is rarely about “a bigger server.” It is usually about isolating hotspots. One high-traffic endpoint. One job queue that spikes. One realtime channel that becomes noisy.&lt;/p&gt;

&lt;p&gt;That is why we built Engines. They let you scale compute separately and predictably, without rewriting your app. If you want to understand when to move up and how cost is calculated, the practical guide is &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Power Up With SashiDo’s Brand-New Engine Feature&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;If you are comparing options, keep it grounded in your real workload. For example, if you are deciding between managed Postgres-style workflows and a Parse-style backend, our side-by-side notes in &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt; help you map trade-offs without guessing.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Quality Bar: How to Claim Leverage Without Shipping Chaos
&lt;/h2&gt;

&lt;p&gt;The best teams treat agent output as a draft, not as truth.&lt;/p&gt;

&lt;p&gt;A useful way to operationalize that is to decide what must be human-owned, even if the agent writes the initial version.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data access rules&lt;/strong&gt; must be reviewed by a human every time. This is where breaches happen.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Public API shapes&lt;/strong&gt; must be stable. Agents love to rename fields.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Retry logic and timeouts&lt;/strong&gt; must be explicit. Otherwise you create self-amplifying load.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Secrets and credentials&lt;/strong&gt; must be managed outside the code. Agents will paste them into config files if you let them.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you need a framework to talk about risk without turning it into hand-waving, the &lt;a href="https://www.nist.gov/itl/ai-risk-management-framework" rel="noopener noreferrer"&gt;NIST AI Risk Management Framework (AI RMF 1.0)&lt;/a&gt; is a strong reference. It helps you name the risk you are managing, from reliability to security to transparency, which makes it easier to choose what to test and what to monitor.&lt;/p&gt;

&lt;p&gt;Also, it is worth remembering that the productivity boost is measurable. In a controlled study, developers using Copilot completed a task significantly faster, as documented in &lt;a href="https://www.microsoft.com/en-us/research/publication/the-impact-of-ai-on-developer-productivity-evidence-from-github-copilot/" rel="noopener noreferrer"&gt;Microsoft Research’s GitHub Copilot productivity paper&lt;/a&gt;. The point is not the exact percentage. The point is that speed is real, so the discipline to keep quality is now the differentiator.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways if You Want to Build an AI App Fast
&lt;/h2&gt;

&lt;p&gt;If you are trying to build ai app experiences quickly, keep these takeaways in mind.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Vibe coding is a great prototyping mode&lt;/strong&gt;, but it needs a handoff to agentic engineering once users and data are involved.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Artificial intelligence coding works best with rails&lt;/strong&gt;, meaning stable data models, explicit auth, and bounded resource usage.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The backend is where prototypes go to die&lt;/strong&gt;. If you remove DevOps early, you keep momentum and reduce long-term rewrite risk.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Scaling is mostly about isolating hotspots&lt;/strong&gt;, not guessing bigger servers. Measure, then scale the part that is actually hot.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions About Artificial Intelligence Coding
&lt;/h2&gt;

&lt;h3&gt;
  
  
  How Is Coding Used in Artificial Intelligence?
&lt;/h3&gt;

&lt;p&gt;In practice, coding is used less for writing “the model” and more for wiring everything around it: data collection, evaluation, prompt orchestration, and safe integration into product flows. The code defines inputs, constraints, retries, and storage so AI outputs are reproducible, auditable, and useful across sessions.&lt;/p&gt;

&lt;h3&gt;
  
  
  Is AI Really Replacing Coding?
&lt;/h3&gt;

&lt;p&gt;AI is changing &lt;em&gt;who writes the first draft&lt;/em&gt;, not eliminating the need to engineer software. As systems become more agent-driven, humans spend more time defining constraints, reviewing risky changes, and designing reliability and security boundaries. The coding work shifts toward orchestration, verification, and operations rather than raw typing.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Much Do AI Coders Make?
&lt;/h3&gt;

&lt;p&gt;Compensation varies widely, because “AI coder” can mean very different roles. Builders who can ship product features and also handle evaluation, data pipelines, and production reliability tend to earn more than those who only prototype. In many markets, the premium is tied to operational ownership, not tool familiarity.&lt;/p&gt;

&lt;h3&gt;
  
  
  How Difficult Is Artificial Intelligence Coding for a Solo Founder?
&lt;/h3&gt;

&lt;p&gt;The hardest part is not the syntax. It is managing complexity when the agent starts generating large changes quickly. If you keep scope tight, use stable data models, and build basic monitoring early, solo founders can ship real AI apps. Difficulty spikes when auth, quotas, and background tasks are added late.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Artificial Intelligence Coding Needs Rails to Stay Fun
&lt;/h2&gt;

&lt;p&gt;Artificial intelligence coding is not going back to the old pace. The winning approach in 2026 is learning how to supervise agents, set boundaries, and keep software operable. Vibe coding can still get you to the first demo. Agentic engineering is how you keep shipping after users show up.&lt;/p&gt;

&lt;p&gt;If you want to keep your momentum while putting the backend on dependable rails, it is worth exploring a managed foundation that already includes database, APIs, auth, storage, realtime, functions, and jobs.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;When you are ready to move from throwaway vibe coding to reliable agentic engineering, you can &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;explore SashiDo’s platform&lt;/a&gt; and start a 10-day free trial with no credit card. Check the current plan limits and overages on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt; so your prototype has a clear path to production.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-that-writes-code-agents-context-governance-2026" rel="noopener noreferrer"&gt;AI that writes code is now a system problem, not a tool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-builder-vibe-coding-saas-backend-2025" rel="noopener noreferrer"&gt;AI App Builder vs Vibe Coding: Will SaaS End-or Just Get Rewired?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/vibe-coding-vital-literacy-skill" rel="noopener noreferrer"&gt;Why Vibe Coding is a Vital Literacy Skill for Developers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/jump-on-vibe-coding-bandwagon" rel="noopener noreferrer"&gt;Jump on the Vibe Coding Bandwagon: A Guide for Non-Technical Founders&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-engineering-prompts-to-production" rel="noopener noreferrer"&gt;What Is BaaS in Vibe Engineering? From Prompts to Production&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>development</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>BaaS Backend as a Service for Parallel AI Agent Teams</title>
      <dc:creator>Marian Ignev</dc:creator>
      <pubDate>Mon, 16 Feb 2026 07:00:36 +0000</pubDate>
      <link>https://dev.to/sashido/baas-backend-as-a-service-for-parallel-ai-agent-teams-1mh1</link>
      <guid>https://dev.to/sashido/baas-backend-as-a-service-for-parallel-ai-agent-teams-1mh1</guid>
      <description>&lt;p&gt;&lt;strong&gt;BaaS backend as a service&lt;/strong&gt; is a managed backend that gives you ready-to-use building blocks like a database, APIs, authentication, storage, serverless functions, and realtime messaging, so you can ship without owning servers. For AI agent teams running in parallel, BaaS matters because it provides &lt;a href="https://www.sashido.io/en/blog/backend-as-a-service-guide-2026" rel="noopener noreferrer"&gt;durable state&lt;/a&gt;, &lt;a href="https://www.sashido.io/en/blog/ctos-dont-let-ai-agents-run-the-backend-yet" rel="noopener noreferrer"&gt;safe coordination&lt;/a&gt;, and predictable operations while you iterate.&lt;/p&gt;

&lt;p&gt;If you have ever tried to let multiple LLM “workers” push code all day, you quickly learn the hard part is not getting code written. The hard part is preventing the system from drifting, duplicating work, breaking yesterday’s features, or spending hours “doing something” without moving a measurable metric.&lt;/p&gt;

&lt;p&gt;The reliable pattern is simple: &lt;strong&gt;agent autonomy only scales when you give it a harness&lt;/strong&gt;. The harness is not just a loop that keeps the model running. It is the environment that tells the agents what success looks like, how to claim work, how to merge safely, and how to recover when they get lost.&lt;/p&gt;

&lt;p&gt;A practical corollary is that &lt;a href="https://www.sashido.io/en/blog/backend-as-a-service-claude-artifacts-to-production" rel="noopener noreferrer"&gt;the backend becomes part of the harness&lt;/a&gt;. Once you go beyond a single laptop session and start running agents on multiple machines, you need shared state, &lt;a href="https://www.sashido.io/en/blog/why-most-backend-choices-fail-and-how-to-pick-one-that-lasts" rel="noopener noreferrer"&gt;audit trails&lt;/a&gt;, access control, file storage, and &lt;a href="https://www.sashido.io/en/blog/best-backend-as-a-service-vibe-coding" rel="noopener noreferrer"&gt;stable webhooks&lt;/a&gt;. That is where a backend as a service BaaS platform can &lt;a href="https://www.sashido.io/en/blog/best-open-source-backend-as-a-service-solutions-vibe-coding" rel="noopener noreferrer"&gt;remove a lot of friction&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;If your goal is to ship an AI-powered feature fast, our &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; is designed for exactly this kind of iteration, with database, APIs, auth, functions, jobs, storage, and realtime already wired.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Parallel Agent Teams Fail Without a Harness
&lt;/h2&gt;

&lt;p&gt;Most “agent mode” setups fail in predictable ways.&lt;/p&gt;

&lt;p&gt;First, &lt;strong&gt;progress becomes unobservable&lt;/strong&gt;. An agent produces logs, commits, and diffs, but you cannot tell if it is actually getting closer to “done” without a tight verifier. When the verifier is weak, agents optimize for passing the wrong thing. When the verifier is noisy, agents thrash.&lt;/p&gt;

&lt;p&gt;Second, &lt;strong&gt;parallel work collapses into duplicated effort&lt;/strong&gt;. If 8 to 16 agents all see the same failing test or the same vague TODO, they race toward the same fix. Even if they are individually competent, you get merge conflicts and regressions. At some point, adding agents makes you slower.&lt;/p&gt;

&lt;p&gt;Third, &lt;strong&gt;context becomes a liability&lt;/strong&gt;. Agent outputs, stack traces, and verbose build logs pollute the next run. The agent “reads” noise and spends tokens summarizing instead of acting. When that happens, you pay for output but not for progress.&lt;/p&gt;

&lt;p&gt;Finally, &lt;strong&gt;the system has no memory&lt;/strong&gt;. A single agent can keep notes in a local file, but in a multi-run, multi-container world, you need durable, queryable memory. Otherwise, every new run spends time rediscovering the same constraints and repeating the same failed approaches.&lt;/p&gt;

&lt;p&gt;These are not abstract concerns. They show up as costs. If you are paying for models and compute, &lt;strong&gt;unbounded retries and duplicated work are the fastest way to burn budget&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Harness Patterns That Make Long-Running Agents Useful
&lt;/h2&gt;

&lt;p&gt;A good harness does three things: it keeps agents running, it tells them what to do next, and it makes their work safely mergeable.&lt;/p&gt;

&lt;h3&gt;
  
  
  Keep The Run Loop Boring
&lt;/h3&gt;

&lt;p&gt;The run loop should be the least interesting part of your system. Its job is to start a fresh agent session, hand it the same high-level goal, and force it to leave artifacts that the next session can pick up. The value is that you stop relying on “one perfect session” and instead build incremental progress over many small sessions.&lt;/p&gt;

&lt;p&gt;The most important design decision here is how you persist artifacts. In practice you need both versioned artifacts (like git commits) and runtime artifacts (like logs, test summaries, and generated files). If runtime artifacts live only inside an ephemeral container, the agent cannot use them as memory.&lt;/p&gt;

&lt;h3&gt;
  
  
  Use Task Locks To Prevent Collisions
&lt;/h3&gt;

&lt;p&gt;When multiple agents share a repo, the simplest synchronization primitive is still a lock file per task. Each agent “claims” a work unit by creating a uniquely named lock, then releases it when done.&lt;/p&gt;

&lt;p&gt;The lock system works best when tasks are concrete. Fix a specific failing test. Implement a parser rule. Optimize a specific hotspot. If tasks are broad, agents will claim different locks but still collide on the same set of files.&lt;/p&gt;

&lt;p&gt;Locking also forces you to decide what a “unit of work” is. A useful heuristic is: &lt;strong&gt;a unit should be small enough to finish in one agent session, and big enough to be reviewed in one diff&lt;/strong&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Give Each Agent Its Own Workspace
&lt;/h3&gt;

&lt;p&gt;Parallel agents need isolated workspaces so they can build, test, and experiment without stepping on each other. The shared upstream is for coordination and merging. The per-agent workspace is for local iteration.&lt;/p&gt;

&lt;p&gt;This separation reduces accidental coupling. It also makes failures easier to debug because you can reproduce a failing run by checking out the agent’s workspace state and rerunning the verifier.&lt;/p&gt;

&lt;h3&gt;
  
  
  Treat Merging As A First-Class Step
&lt;/h3&gt;

&lt;p&gt;If you want parallelism, you must assume merges will be frequent and sometimes painful.&lt;/p&gt;

&lt;p&gt;A harness should standardize how agents pull latest changes, handle merge conflicts, re-run the verifier, and only then push. If you do not standardize this, each agent invents its own merge process, which usually means pushing half-tested changes.&lt;/p&gt;

&lt;p&gt;This is also where access control matters. If your agents can push to main without guardrails, you will eventually deploy something that “passed” but was not actually verified end-to-end.&lt;/p&gt;

&lt;h2&gt;
  
  
  Verifiers: Tests That Keep Agents Honest
&lt;/h2&gt;

&lt;p&gt;Agent teams are only as good as the feedback you provide. In practice, that means your verifier must be both &lt;strong&gt;high quality&lt;/strong&gt; and &lt;strong&gt;machine-friendly&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;High quality means it catches regressions and prevents the system from “solving” a proxy metric. If your verifier only checks compilation, agents will ship a compiler that compiles but breaks semantics. If your verifier only checks unit tests, agents may overfit test cases.&lt;/p&gt;

&lt;p&gt;Machine-friendly means the output is structured and short. Long, noisy output increases context window pollution and makes the next agent session spend tokens reading rather than fixing.&lt;/p&gt;

&lt;p&gt;A few patterns we see work reliably:&lt;/p&gt;

&lt;h3&gt;
  
  
  Make The Happy Path Fast, The Full Path Real
&lt;/h3&gt;

&lt;p&gt;Agents will happily spend hours running full suites. That is rarely what you want.&lt;/p&gt;

&lt;p&gt;A better approach is to have two modes: a fast mode that runs a deterministic subsample, and a full mode that runs on a schedule or on specific triggers. Deterministic matters because you want agents to know whether they made things better or worse. Subsampling matters because you want rapid iteration.&lt;/p&gt;

&lt;p&gt;There is a trade-off. If you subsample too aggressively, you miss regressions. If you never subsample, you slow progress. In many projects, a 1% to 10% fast mode is a workable starting point. Increase coverage as you approach “almost done”, because that is where regressions become most frequent.&lt;/p&gt;

&lt;h3&gt;
  
  
  Summarize Failures, Then Link To Deep Logs
&lt;/h3&gt;

&lt;p&gt;In an agent harness, the console output should read like a verdict. One line per failing check, with stable identifiers and the minimal error reason.&lt;/p&gt;

&lt;p&gt;The detailed logs should be stored separately, with a consistent path and naming scheme, so the agent can find them when needed. This mirrors how strong CI systems behave for humans. You see the summary first, then drill down.&lt;/p&gt;

&lt;h3&gt;
  
  
  Add Oracles When The Task Is Too Big
&lt;/h3&gt;

&lt;p&gt;Some tasks are “one giant thing”. Large builds, massive integration tests, or system-level behaviors do not decompose cleanly into hundreds of independent tests.&lt;/p&gt;

&lt;p&gt;In those cases, you often need a known-good oracle to help agents isolate the blame. The principle is: &lt;strong&gt;reduce the search space until multiple agents can work on disjoint slices&lt;/strong&gt;. In compiler work that might mean compiling a subset with one tool and the rest with another, then shrinking the subset when failures occur. In web apps, it might mean replaying production traffic against a known-good version and bisecting by endpoint.&lt;/p&gt;

&lt;h2&gt;
  
  
  When A Managed Backend Becomes Part of the Harness
&lt;/h2&gt;

&lt;p&gt;Once agents run across machines and sessions, your backend stops being “the app backend” and becomes “the system backbone”. You need a place to store state, coordinate tasks, authenticate actors, and expose webhooks for external triggers.&lt;/p&gt;

&lt;p&gt;This is where a &lt;strong&gt;baas backend as a service&lt;/strong&gt; approach fits naturally, especially for solo builders who do not want to build infrastructure just to support their automation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Durable State For Agents and Builds
&lt;/h3&gt;

&lt;p&gt;Agents need persistent memory: task queues, run histories, summaries of failed approaches, known failure modes, and artifacts.&lt;/p&gt;

&lt;p&gt;A BaaS with a real database and CRUD APIs lets you log each run as an object with fields like status, commit hash, failing checks, and links to artifacts. If you later want analytics, you query it. If you later want dashboards, you already have the data.&lt;/p&gt;

&lt;p&gt;MongoDB-style event streams are also useful when you want automation that reacts to state changes. MongoDB’s &lt;a href="https://www.mongodb.com/docs/manual/changestreams/" rel="noopener noreferrer"&gt;Change Streams documentation&lt;/a&gt; is a good reference for the underlying concept, even if your platform abstracts the implementation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Auth and Multi-Tenancy Without Reinventing It
&lt;/h3&gt;

&lt;p&gt;As soon as you share a tool with collaborators, customers, or even your future self on a different machine, you need authentication and authorization.&lt;/p&gt;

&lt;p&gt;This is where many agent prototypes die. People postpone auth, then later realize every endpoint and every artifact store needs access control.&lt;/p&gt;

&lt;p&gt;A managed BaaS for freelancers and small teams is valuable here because you can model a &lt;strong&gt;multi-tenant backend&lt;/strong&gt; early. That means each app or workspace has isolated data, and your “agent runs” are scoped to the correct tenant by design.&lt;/p&gt;

&lt;h3&gt;
  
  
  File Storage for Logs and Artifacts
&lt;/h3&gt;

&lt;p&gt;Agent systems generate files. Logs, build outputs, coverage reports, screenshots, model outputs, and more.&lt;/p&gt;

&lt;p&gt;If you store these on local disk, you lose them when containers reset. If you store them in the database, you bloat your storage and complicate retrieval.&lt;/p&gt;

&lt;p&gt;Object storage is the right primitive for this. It is designed for big blobs and cheap delivery. A managed platform with integrated storage and CDN makes artifacts accessible without you wiring a separate service.&lt;/p&gt;

&lt;h3&gt;
  
  
  Realtime and Webhooks for Feedback Loops
&lt;/h3&gt;

&lt;p&gt;When you want a live dashboard, realtime messaging matters. WebSockets are the standard building block for this. If you want the canonical protocol reference, the IETF’s &lt;a href="https://datatracker.ietf.org/doc/html/rfc6455" rel="noopener noreferrer"&gt;RFC 6455: The WebSocket Protocol&lt;/a&gt; is the primary spec.&lt;/p&gt;

&lt;p&gt;In practice, realtime lets you stream agent status changes to your UI. Webhooks let you trigger agent runs from external systems like git events, issue trackers, or scheduled jobs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Background Jobs for Scheduled Verifiers
&lt;/h3&gt;

&lt;p&gt;Agent harnesses work best when the verifier runs on a schedule, not only on pushes. Nightly full test runs, periodic “run the expensive suite”, or “rebuild all artifacts with the latest dependencies” are job workloads.&lt;/p&gt;

&lt;p&gt;You can build a job system yourself, but it is another operational surface. A backend that scales should let you schedule recurring jobs and inspect them when things go wrong.&lt;/p&gt;

&lt;h3&gt;
  
  
  Security Is Not Optional
&lt;/h3&gt;

&lt;p&gt;Autonomous systems amplify mistakes. If an agent can upload secrets into logs, push insecure code, or expose an internal endpoint, it will eventually happen.&lt;/p&gt;

&lt;p&gt;A good baseline is to explicitly threat model and align your controls with well-known categories. The &lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10 (2021)&lt;/a&gt; is a practical checklist for common web risks like broken access control and injection.&lt;/p&gt;

&lt;p&gt;Also, if you are building something intended to survive, structure it like an operable service. The &lt;a href="https://12factor.net/" rel="noopener noreferrer"&gt;Twelve-Factor App&lt;/a&gt; guidelines are still a useful mental model for config, deploys, logs, and portability.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where We Fit: A Practical Backend as a Service BaaS Platform
&lt;/h2&gt;

&lt;p&gt;Most solo builders do not fail because they cannot code. They fail because “the backend backlog” grows faster than the product.&lt;/p&gt;

&lt;p&gt;We built &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; so you can stand up a production-grade Parse-based backend quickly, then spend your time on the harness, the verifier, and the product behavior. Under the hood, every app comes with a MongoDB database and ready CRUD APIs, user management with social logins, serverless JavaScript functions, realtime, scheduled jobs, push notifications, and object storage with CDN.&lt;/p&gt;

&lt;p&gt;If you want to understand the foundation, start with our &lt;a href="https://www.sashido.io/en/docs" rel="noopener noreferrer"&gt;Parse Platform docs and guides&lt;/a&gt;. If you are specifically planning to scale agent-driven workloads, our write-up on engines is a useful mental model for dialing compute up and down without rebuilding infrastructure, see &lt;a href="https://www.sashido.io/en/blog/power-up-with-sashidos-brand-new-engine-feature" rel="noopener noreferrer"&gt;Power Up With SashiDo’s Engine Feature&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;It is also worth being explicit about trade-offs. If your project is deeply coupled to a different database paradigm, or you need bespoke infrastructure primitives, a BaaS may not be the right fit. And if you are comparing options, we keep direct comparisons in one place, for example &lt;a href="https://www.sashido.io/en/sashido-vs-supabase" rel="noopener noreferrer"&gt;SashiDo vs Supabase&lt;/a&gt; and &lt;a href="https://www.sashido.io/en/sashido-vs-aws-amplify" rel="noopener noreferrer"&gt;SashiDo vs AWS Amplify&lt;/a&gt;, so you can map features and operational responsibilities without jumping between vendor marketing pages.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started: A Lightweight Checklist for Agent Teams + BaaS
&lt;/h2&gt;

&lt;p&gt;If you are building an agent harness and want to keep it shippable, the fastest path is to decide what you will measure, what you will persist, and what you will not build yourself.&lt;/p&gt;

&lt;p&gt;Here is a practical sequence that works well for a solo founder.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Define one objective metric for “progress”. Pick something testable, like passing a suite, reducing failing cases from 200 to 50, or compiling a fixed set of projects. Avoid vague goals like improve quality.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Design a verifier that is readable by machines. Produce a short summary output with stable identifiers. Store full logs as artifacts.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Choose a task unit and a lock naming scheme. Make it easy for agents to claim disjoint work without debate.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Persist run state in a database. Store run metadata, task claims, and outcomes as rows or documents so you can query and build dashboards.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Persist artifacts in object storage. Keep logs, build outputs, and reports out of your database.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Put auth in early. Even if you are the only user today, treat the harness UI and APIs like a multi-tenant backend from the start.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Add scheduled jobs for the expensive checks. Nightly full runs catch regressions that fast mode misses.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Add a kill switch. You need a way to pause agents, revoke keys, and stop jobs when something goes off the rails.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;If you want a quick walkthrough of how we think about setting up the backend pieces without getting lost in configuration, our &lt;a href="https://www.sashido.io/en/blog/sashidos-getting-started-guide" rel="noopener noreferrer"&gt;Getting Started Guide&lt;/a&gt; is the shortest path to a working backend you can wire into your harness.&lt;/p&gt;

&lt;h2&gt;
  
  
  Frequently Asked Questions
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Are the BaaS Platforms for Backend-as-a-Service?
&lt;/h3&gt;

&lt;p&gt;BaaS platforms bundle common backend needs like database, auth, &lt;a href="https://www.sashido.io/en/blog/startup-mvp-backend-in-2026-tools-tradeoffs-trends" rel="noopener noreferrer"&gt;file storage&lt;/a&gt;, &lt;a href="https://www.sashido.io/en/blog/ai-powered-backend-mobile-app-development-speed" rel="noopener noreferrer"&gt;serverless functions&lt;/a&gt;, and realtime APIs so you can ship without managing servers. In agent-team workflows, look for strong auth, reliable background jobs, and good observability. Many teams also prefer platforms built on open ecosystems, such as &lt;a href="https://github.com/parse-community/parse-server" rel="noopener noreferrer"&gt;Parse Server&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is BaaS and SaaS?
&lt;/h3&gt;

&lt;p&gt;In software development, BaaS (backend as a service) gives you backend building blocks like data storage, auth, and APIs as a managed service. SaaS is a finished application delivered over the web. For agent harnesses, BaaS is the infrastructure layer you build on, while SaaS is the product you eventually deliver to users.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is BaaS Banking as a Service?
&lt;/h3&gt;

&lt;p&gt;BaaS in banking refers to Banking-as-a-Service, where licensed institutions expose APIs for accounts, cards, and payments so other companies can embed financial features. That is different from baas backend as a service in software engineering, which is about app backends. The overlap is mostly conceptual: both provide APIs that abstract regulated or operational complexity.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Is the Difference Between BaaS and PaaS?
&lt;/h3&gt;

&lt;p&gt;BaaS focuses on ready-to-use backend capabilities like user management, database APIs, push notifications, and file storage. PaaS (platform as a service) usually provides runtime and deployment primitives where you still build most backend components yourself. For parallel agent teams, BaaS reduces surface area faster. PaaS offers flexibility but increases operational work.&lt;/p&gt;

&lt;h3&gt;
  
  
  When Does Parallelism Stop Helping Agent Teams?
&lt;/h3&gt;

&lt;p&gt;Parallelism stops helping when the work cannot be decomposed. If all agents hit the same bug in the same files, you get collisions and regressions. The fix is usually to shard the work using better task definitions, stronger verifiers, or an oracle that helps isolate failures so agents can work on disjoint slices.&lt;/p&gt;

&lt;h3&gt;
  
  
  Can SashiDo Store Agent Run Logs and Artifacts?
&lt;/h3&gt;

&lt;p&gt;Yes. We typically store structured run metadata in the database and keep large artifacts like logs and build outputs in object storage. The important architectural point is separation: summaries should be queryable, while blobs should be cheap to store and fast to serve.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Shipping Parallel Agents Requires a Verifier and a Backbone
&lt;/h2&gt;

&lt;p&gt;The main lesson from real-world parallel agent teams is that autonomy is an engineering problem, not a prompt problem. &lt;strong&gt;You make progress by building a harness that forces measurable improvement&lt;/strong&gt;. That means high-quality verifiers, task locks, isolated workspaces, and a strategy for when the task is too large to parallelize directly.&lt;/p&gt;

&lt;p&gt;When you reach the point where &lt;a href="https://www.sashido.io/en/blog/backend-as-a-service-frontend-integration" rel="noopener noreferrer"&gt;shared state&lt;/a&gt;, auth, storage, realtime feedback, and scheduled verifiers become the bottleneck, a &lt;strong&gt;baas backend as a service&lt;/strong&gt; stops being a convenience and becomes part of the system design. If you want to remove backend friction while you iterate on agent harnesses, you can explore &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and start with a free trial, then check the current plan details on our &lt;a href="https://www.sashido.io/en/pricing/" rel="noopener noreferrer"&gt;pricing page&lt;/a&gt;.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you are trying to ship an agent-driven prototype without spending a week wiring auth, storage, jobs, and realtime, it is often simpler to &lt;strong&gt;explore SashiDo’s platform&lt;/strong&gt; at &lt;a href="https://www.sashido.io/en/" rel="noopener noreferrer"&gt;SashiDo - Backend for Modern Builders&lt;/a&gt; and plug your harness into a managed backend that is ready in minutes.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Sources and Further Reading
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://datatracker.ietf.org/doc/html/rfc6455" rel="noopener noreferrer"&gt;RFC 6455: The WebSocket Protocol&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://owasp.org/Top10/2021/" rel="noopener noreferrer"&gt;OWASP Top 10: 2021&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://12factor.net/" rel="noopener noreferrer"&gt;The Twelve-Factor App&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://github.com/parse-community/parse-server" rel="noopener noreferrer"&gt;Parse Server Repository&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.mongodb.com/docs/manual/changestreams/" rel="noopener noreferrer"&gt;MongoDB Manual: Change Streams&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Related Articles
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/mcp-ai-workflows-agent-ready-backends" rel="noopener noreferrer"&gt;MCP and AI Agents: Building Agent-Ready Backends in 2025&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/alternatives-to-supabase-backend-as-a-service-vibe-coding" rel="noopener noreferrer"&gt;Alternatives to Supabase Backend as a Service for Vibe Coding&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/what-is-baas-vibe-coding-ai-developer-productivity" rel="noopener noreferrer"&gt;Does AI Coding Really Boost Output?&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/ai-app-development-agent-ready-apis" rel="noopener noreferrer"&gt;AI App Development Needs Agent-Ready APIs (Not “Smart” Agents)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.sashido.io/en/blog/best-backend-as-a-service-vibe-coding-mvps" rel="noopener noreferrer"&gt;Best Backend as a Service for Vibe Coding MVPs (2026)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>backend</category>
      <category>cloud</category>
      <category>devops</category>
    </item>
  </channel>
</rss>
