<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Cauã Ferraz</title>
    <description>The latest articles on DEV Community by Cauã Ferraz (@cauaferraz).</description>
    <link>https://dev.to/cauaferraz</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3853793%2F1bf108fe-cf1b-4cd4-a9ee-91879d87fd17.jpeg</url>
      <title>DEV Community: Cauã Ferraz</title>
      <link>https://dev.to/cauaferraz</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/cauaferraz"/>
    <language>en</language>
    <item>
      <title>Why AI agent teams are just hoping their agents behave</title>
      <dc:creator>Cauã Ferraz</dc:creator>
      <pubDate>Tue, 31 Mar 2026 15:45:43 +0000</pubDate>
      <link>https://dev.to/cauaferraz/why-ai-agent-teams-are-just-hoping-their-agents-behave-2loa</link>
      <guid>https://dev.to/cauaferraz/why-ai-agent-teams-are-just-hoping-their-agents-behave-2loa</guid>
      <description>&lt;p&gt;I'm 19, studying computer engineering in Brazil. A few weeks ago I was testing an AI agent with no restrictions. Just to see what it would do.&lt;/p&gt;

&lt;p&gt;It was destructive.&lt;/p&gt;

&lt;p&gt;Nothing permanent, I caught it. But it was the kind of moment where you sit back and think: what if I hadn't been watching? What if this was running in production? What if someone else's agent is doing this right now and nobody is watching?&lt;/p&gt;

&lt;p&gt;That's when I realized the problem. Everyone is racing to give agents more tools, more autonomy, more access. But nobody is building the layer that controls what they can actually do with it. The assumption is that a good prompt is enough. It isn't.&lt;/p&gt;

&lt;h2&gt;
  
  
  The gap nobody is talking about
&lt;/h2&gt;

&lt;p&gt;The AI agent space has exploded. LangChain, CrewAI, browser-use, OpenAI Agents SDK, the tooling for building agents has never been better. You can have an agent browsing the web, writing code, calling APIs, and moving files in an afternoon.&lt;/p&gt;

&lt;p&gt;But here's what I couldn't find: a serious answer to "how do I control what my agent can actually do at runtime?"&lt;/p&gt;

&lt;p&gt;The common answers I got:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;"Write a good system prompt"&lt;/li&gt;
&lt;li&gt;"Add some input validation"&lt;/li&gt;
&lt;li&gt;"Just don't give it dangerous tools"&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These are not answers. These are hopes dressed up as engineering.&lt;/p&gt;

&lt;p&gt;A good system prompt doesn't stop an agent from being manipulated through prompt injection. Input validation doesn't catch an agent that decides &lt;code&gt;rm -rf ./old_stuff&lt;/code&gt; is a reasonable interpretation of "clean up." And "don't give it dangerous tools" directly contradicts the reason you're using agents in the first place.&lt;/p&gt;

&lt;h2&gt;
  
  
  What actually needs to exist
&lt;/h2&gt;

&lt;p&gt;The thing missing is embarrassingly simple: a policy layer that sits between your agent and the world.&lt;/p&gt;

&lt;p&gt;Not prompt engineering. Not vibes. An actual enforcement layer that says:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt; This agent can read from &lt;code&gt;./workspace&lt;/code&gt; but cannot delete anything&lt;/li&gt;
&lt;li&gt; This agent can call the OpenAI API but not your production database&lt;/li&gt;
&lt;li&gt; This command requires a human to approve it before it executes&lt;/li&gt;
&lt;li&gt; Everything gets logged, always&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The goal isn't to babysit every action manually, that defeats the purpose of automation. The goal is to define the boundaries once, enforce them automatically, and only surface the genuinely ambiguous decisions to a human.&lt;/p&gt;

&lt;p&gt;This is what firewalls do for networks. This is what WAFs do for web apps. Agents need the same thing, and almost nobody is building it.&lt;/p&gt;

&lt;h2&gt;
  
  
  So I built it
&lt;/h2&gt;

&lt;p&gt;I built AgentGuard, an open source runtime firewall for AI agents.&lt;/p&gt;

&lt;p&gt;It's a Go proxy that sits between your agent and its tools. You define policies in YAML. The proxy enforces them in real time, blocking, holding for approval, logging everything. It has adapters for LangChain, CrewAI, browser-use, and MCP. There's a dashboard that shows you live what your agents are doing and lets you approve or deny actions with one click.&lt;/p&gt;

&lt;p&gt;It's not finished. The SQLite audit backend isn't done. Some adapters are still rough. But the core works, and I think the core is the right idea.&lt;/p&gt;


&lt;div class="ltag-github-readme-tag"&gt;
  &lt;div class="readme-overview"&gt;
    &lt;h2&gt;
      &lt;img src="https://assets.dev.to/assets/github-logo-5a155e1f9a670af7944dd5e12375bc76ed542ea80224905ecaf878b9157cdefc.svg" alt="GitHub logo"&gt;
      &lt;a href="https://github.com/Caua-ferraz" rel="noopener noreferrer"&gt;
        Caua-ferraz
      &lt;/a&gt; / &lt;a href="https://github.com/Caua-ferraz/AgentGuard" rel="noopener noreferrer"&gt;
        AgentGuard
      &lt;/a&gt;
    &lt;/h2&gt;
    &lt;h3&gt;
      AgentGuard is a firewall for AI agents, preventing that any unwanted surprises go without supervision by your agent
    &lt;/h3&gt;
  &lt;/div&gt;
  &lt;div class="ltag-github-body"&gt;
    
&lt;div id="readme" class="md"&gt;
&lt;p&gt;
  &lt;a rel="noopener noreferrer" href="https://github.com/Caua-ferraz/AgentGuard/docs/assets/banner.svg"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fraw.githubusercontent.com%2FCaua-ferraz%2FAgentGuard%2FHEAD%2Fdocs%2Fassets%2Fbanner.svg" alt="AgentGuard" width="720"&gt;&lt;/a&gt;
&lt;/p&gt;

&lt;p&gt;
  &lt;strong&gt;The firewall for AI agents.&lt;/strong&gt;&lt;br&gt;
  Policy enforcement, real-time oversight, and full audit logging for autonomous AI systems
&lt;/p&gt;

&lt;p&gt;
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard#quickstart" rel="noopener noreferrer"&gt;Quickstart&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard#why-agentguard" rel="noopener noreferrer"&gt;Why AgentGuard&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard#architecture" rel="noopener noreferrer"&gt;Architecture&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard#policy-engine" rel="noopener noreferrer"&gt;Policy Engine&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard#dashboard" rel="noopener noreferrer"&gt;Dashboard&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard#adapters" rel="noopener noreferrer"&gt;Adapters&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard/docs/SETUP.md" rel="noopener noreferrer"&gt;Setup Guide&lt;/a&gt; •
  &lt;a href="https://github.com/Caua-ferraz/AgentGuard/docs/CONTRIBUTING.md" rel="noopener noreferrer"&gt;Contributing&lt;/a&gt;
&lt;/p&gt;

&lt;div class="markdown-heading"&gt;
&lt;h2 class="heading-element"&gt;The Problem&lt;/h2&gt;
&lt;/div&gt;

&lt;p&gt;Every trending AI project is giving agents more autonomy — running shell commands, browsing the web, calling APIs, moving money, even performing penetration tests. But &lt;strong&gt;nobody is building the guardrails.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Right now, most teams deploying AI agents are just... hoping they behave.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;AgentGuard&lt;/strong&gt; fixes that.&lt;/p&gt;

&lt;div class="markdown-heading"&gt;
&lt;h2 class="heading-element"&gt;Why AgentGuard&lt;/h2&gt;
&lt;/div&gt;

&lt;p&gt;&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;br&gt;
&lt;thead&gt;
&lt;br&gt;
&lt;tr&gt;
&lt;br&gt;
&lt;th&gt;Without AgentGuard&lt;/th&gt;
&lt;br&gt;
&lt;th&gt;With AgentGuard&lt;/th&gt;
&lt;br&gt;
&lt;/tr&gt;
&lt;br&gt;
&lt;/thead&gt;
&lt;br&gt;
&lt;tbody&gt;
&lt;br&gt;
&lt;tr&gt;
&lt;br&gt;
&lt;td&gt;Agent runs &lt;code&gt;rm -rf /&lt;/code&gt; — you find out later&lt;/td&gt;
&lt;br&gt;
&lt;td&gt;Policy blocks destructive commands before execution&lt;/td&gt;
&lt;br&gt;
&lt;/tr&gt;
&lt;br&gt;
&lt;tr&gt;
&lt;br&gt;
&lt;td&gt;Agent calls production API with no oversight&lt;/td&gt;
&lt;br&gt;
&lt;td&gt;Action paused, you get a Slack/webhook notification to approve&lt;/td&gt;
&lt;br&gt;
&lt;/tr&gt;
&lt;br&gt;
&lt;tr&gt;
&lt;br&gt;
&lt;td&gt;No record of what the agent did or why&lt;/td&gt;
&lt;br&gt;
&lt;td&gt;Full audit trail with timestamps, reasoning, and decisions&lt;/td&gt;
&lt;br&gt;
&lt;/tr&gt;
&lt;br&gt;
&lt;tr&gt;
&lt;br&gt;
&lt;td&gt;"It worked on my machine" debugging&lt;/td&gt;
&lt;br&gt;
&lt;td&gt;Query any agent session from the audit&lt;/td&gt;
&lt;br&gt;
&lt;/tr&gt;
&lt;br&gt;
&lt;/tbody&gt;
&lt;br&gt;
&lt;/table&gt;&lt;/div&gt;…&lt;/p&gt;
&lt;/div&gt;
&lt;br&gt;
  &lt;/div&gt;
&lt;br&gt;
  &lt;div class="gh-btn-container"&gt;&lt;a class="gh-btn" href="https://github.com/Caua-ferraz/AgentGuard" rel="noopener noreferrer"&gt;View on GitHub&lt;/a&gt;&lt;/div&gt;
&lt;br&gt;
&lt;/div&gt;
&lt;br&gt;


&lt;p&gt;In 5 days it's been cloned by 165 unique developers with almost no active distribution. I think that says something about how real this problem is.&lt;/p&gt;

&lt;h2&gt;
  
  
  The thing I keep thinking about
&lt;/h2&gt;

&lt;p&gt;Only 14.4% of organizations send AI agents to production with full security approval. 88% reported confirmed or suspected AI agent security incidents last year.&lt;/p&gt;

&lt;p&gt;Everyone is moving fast. Nobody is building the guardrails.&lt;/p&gt;

&lt;p&gt;I don't know if AgentGuard is the right answer. But I'm pretty confident "hope" isn't.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>opensource</category>
      <category>agents</category>
    </item>
  </channel>
</rss>
