<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: AndrewSispoidis</title>
    <description>The latest articles on DEV Community by AndrewSispoidis (@andrewsispoidis).</description>
    <link>https://dev.to/andrewsispoidis</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3840825%2Fce6d47ac-8479-4d4e-b185-321edc5621cd.png</url>
      <title>DEV Community: AndrewSispoidis</title>
      <link>https://dev.to/andrewsispoidis</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/andrewsispoidis"/>
    <language>en</language>
    <item>
      <title>The OWASP Agentic AI Top 10 is Live. Here's What the Attacks Actually Look Like — and What to Do About Them.</title>
      <dc:creator>AndrewSispoidis</dc:creator>
      <pubDate>Mon, 30 Mar 2026 13:27:55 +0000</pubDate>
      <link>https://dev.to/andrewsispoidis/the-owasp-agentic-ai-top-10-is-live-heres-what-the-attacks-actually-look-like-and-what-to-do-3nc2</link>
      <guid>https://dev.to/andrewsispoidis/the-owasp-agentic-ai-top-10-is-live-heres-what-the-attacks-actually-look-like-and-what-to-do-3nc2</guid>
      <description>&lt;p&gt;Anthropic confirmed this week that their next model poses "unprecedented cybersecurity risks" and can "exploit vulnerabilities in ways that far outpace the efforts of defenders." Cybersecurity stocks dropped 4–9% on the news. The story ran in Fortune, Axios, and CNBC.&lt;/p&gt;

&lt;p&gt;Here's what those headlines missed: &lt;strong&gt;the threat isn't the next model. It's the one you're running right now.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In February 2026, Amazon's threat intelligence team documented a single attacker — low to medium skill, financially motivated — who used commercially available AI to compromise 600 FortiGate firewall devices across 55 countries in 38 days. Amazon's CISO noted that the volume and variety of custom tooling would typically indicate a well-resourced development team. Instead, one person with AI access built the entire toolkit. The model didn't change. The scaffolding — the agentic workflows — is what turned a general-purpose LLM into a global offensive capability.&lt;/p&gt;

&lt;p&gt;OWASP published their Top 10 for Agentic Applications in December 2025. It's the most important security framework most AI developers haven't read yet.&lt;/p&gt;

&lt;p&gt;This is a technical breakdown of each risk, the real CVEs behind them, and how to actually defend against them.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why a new Top 10
&lt;/h2&gt;

&lt;p&gt;The original OWASP LLM Top 10 was designed for single-turn applications: a user sends a message, the model responds. Agentic systems are different in three critical ways:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;They act.&lt;/strong&gt; Agents don't just generate text — they execute shell commands, call APIs, read and write files, send emails, and browse the web. A single compromised prompt can cause real-world irreversible damage.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;They persist.&lt;/strong&gt; Agents maintain memory across sessions. A single successful injection can poison an agent's behavior permanently, not just for one response.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;They delegate.&lt;/strong&gt; Multi-agent systems trust each other by default. A compromised sub-agent can influence the entire pipeline.&lt;/p&gt;

&lt;p&gt;The OWASP ASI Top 10 formalizes 10 failure modes that don't exist in traditional applications. Here they are, with real incidents.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI01 — Agent Goal Hijack
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; An attacker redirects an agent's objectives through malicious text in any content the agent reads. The agent isn't hacked in the traditional sense — it's simply told to do something else, and it complies.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why it's #1:&lt;/strong&gt; Every other attack on this list is a pathway to this outcome. Prompt injection (ASI02), tool misuse (ASI05), memory poisoning (ASI04) — they're all mechanisms for achieving ASI01. A fully hijacked agent is an insider threat that works at machine speed.&lt;/p&gt;

&lt;h3&gt;
  
  
  The real incident — EchoLeak (CVE-2025-32711, CVSS 9.3)
&lt;/h3&gt;

&lt;p&gt;In June 2025, researchers at Aim Security disclosed a zero-click vulnerability in Microsoft 365 Copilot. The attack required no user interaction whatsoever. Here's how it worked:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Attacker sends a carefully crafted email to the target organization&lt;/li&gt;
&lt;li&gt;The email contains hidden prompt injection instructions, phrased as if directed at a human — never mentioning Copilot, AI, or anything that would trigger detection filters&lt;/li&gt;
&lt;li&gt;When Copilot later retrieves that email as context for an unrelated query, it reads the hidden instructions&lt;/li&gt;
&lt;li&gt;Copilot exfiltrates sensitive internal files by embedding them in an outbound image URL&lt;/li&gt;
&lt;li&gt;The victim's browser auto-fetches the image, completing the exfiltration without any click&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The attack bypassed Microsoft's XPIA classifier, link redaction, and Content Security Policy. The payload was pure natural language. No code. No malware. No signatures to detect.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What this means for your agents:&lt;/strong&gt; Every document, email, web page, or tool output your agent reads is a potential attack vector. The attack surface isn't your API endpoint. It's every piece of text your agent ingests.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Scan all retrieved content before it enters the context window. Don't just scan user inputs — scan tool outputs, web search results, and documents. Treat every external string as potentially hostile.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI02 — Prompt Injection
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Direct injection of instructions that override the agent's system prompt or intended behavior. The classic "ignore all previous instructions."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why it matters more for agents:&lt;/strong&gt; A chatbot that gets prompt-injected gives a bad response. An agent that gets prompt-injected executes arbitrary actions with whatever permissions it was granted.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real incident — IDEsaster (2026)
&lt;/h3&gt;

&lt;p&gt;Security researcher Ari Marzouk disclosed 24 CVEs across GitHub Copilot, Cursor, Windsurf, and 5 other AI coding assistants. 100% of tested AI IDEs were vulnerable to prompt injection leading to code execution. AWS issued security advisory AWS-2025-019.&lt;/p&gt;

&lt;p&gt;Attack vector: malicious repository content → agent reads it → agent executes attacker-controlled commands with developer-level privileges.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; 5-layer local detection — pattern matching (27+ categories), semantic analysis (role hijacking, authority impersonation, boundary dissolution), indirect injection detection, session context tracking, and PII/credential exfiltration detection.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI03 — Identity and Privilege Abuse
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Agents inherit user roles, cache credentials, and call each other. Attackers exploit the delegation chain to escalate privileges or reuse cached secrets.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The problem:&lt;/strong&gt; When an agent is allowed to act "as the user," you extend that user's entire blast radius to anything the agent can be manipulated into doing. There's no least-privilege boundary between the agent and the user.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real incident — Amazon Q Code Assistant (CVE-2025-8217)
&lt;/h3&gt;

&lt;p&gt;Attackers compromised a GitHub token and merged malicious code into the Amazon Q VS Code extension (version 1.84.0). The injected code contained destructive prompt instructions including commands to delete filesystem and cloud resources. With &lt;code&gt;--trust-all-tools --no-interactive&lt;/code&gt; flags active, the agent executed without confirmation. Nearly one million developers had the extension installed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Cryptographic agent identity (Ed25519 keypairs), mTLS between agents and services, scoped credentials that expire, and audit trails that capture which agent performed which action under which identity.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI04 — Memory and Context Poisoning
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Injecting malicious content into an agent's persistent memory, RAG database, or long-term context so that future behavior is corrupted — long after the initial attack.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why it's worse than standard injection:&lt;/strong&gt; Prompt injection affects one interaction. Memory poisoning affects every future interaction until the memory is cleared and audited. The agent "learns" the attacker's instructions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Attack patterns:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;RAG poisoning:&lt;/strong&gt; inject malicious content into a vector database the agent queries for context&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cross-tenant leakage:&lt;/strong&gt; agent memory shared across tenants leaks sensitive data&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Long-term drift:&lt;/strong&gt; repeated exposure to adversarial content gradually shifts agent behavior&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Merkle-chained memory with Ed25519 signatures. Any tampered memory entry fails verification at query time. Append-only audit log means you can always reconstruct what the agent was told and when.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI05 — Tool and Integration Misuse
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Agents call tools — shell commands, database queries, API calls, file operations. If the agent can be convinced to pass attacker-controlled parameters to these tools, you have RCE through natural language.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real incident — Langflow AI RCE (CVE-2025-34291)
&lt;/h3&gt;

&lt;p&gt;CrowdStrike documented multiple threat actors exploiting an unauthenticated code injection vulnerability in Langflow AI. Attackers gained credentials and deployed malware through the agent's tool execution capability. The vulnerability wasn't in the LLM. It was in the trust boundary between the agent's output and the tool execution layer.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real incident — OpenAI Operator Data Exposure
&lt;/h3&gt;

&lt;p&gt;Security researcher Johann Rehberger demonstrated that malicious webpage content could trick OpenAI's Operator agent into accessing authenticated internal pages and exfiltrating data to an attacker-controlled server.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Policy engine that validates every tool call before execution. Scoped, signed, revocable tokens for each action. The agent proposes; the policy engine authorizes.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI06 — Resource and Service Abuse
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Agents running in loops can be exploited for financial denial-of-service. An attacker who can trigger expensive inference loops, or cause an agent to repeatedly call costly external APIs, can run up massive costs or exhaust quotas.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why it matters:&lt;/strong&gt; Unlike traditional DDoS, this attack uses the victim's own authorized systems against them. The agent is behaving "correctly" from the provider's perspective.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Hard cost ceilings, rate limiting at the agent level, circuit breakers that pause agents when anomalous consumption patterns appear.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;This is the ASI risk with the least coverage across the industry right now.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  ASI07 — Data and Model Exfiltration
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Agents exfiltrating training data, system prompts, model weights, or sensitive business data. Beyond PII — this includes intellectual property, strategic information, and the agent's own configuration.&lt;/p&gt;

&lt;p&gt;The same mechanism that made EchoLeak work — agent reads malicious content → agent exfiltrates data to attacker-controlled URL — applies to any agent with outbound network access and sensitive context access.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; 15-category PII and credential detection on all outbound content. Pattern matching for API keys, tokens, SSNs, internal URLs. Block exfiltration attempts before they reach the network.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI08 — Cascading Agent Failures
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; In multi-agent systems, a single compromised agent can corrupt the entire pipeline. Agents are often designed to trust collaborating agents by default.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real incident — Agent Session Smuggling (November 2025)
&lt;/h3&gt;

&lt;p&gt;Palo Alto Unit 42 demonstrated how malicious agents exploit built-in trust relationships in the Agent-to-Agent (A2A) protocol. Unlike single-shot prompt injection, a rogue agent can hold multi-turn conversations, adapt strategy, and build false trust over time.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real incident — ServiceNow Now Assist
&lt;/h3&gt;

&lt;p&gt;OWASP documented cases where spoofed inter-agent messages caused downstream procurement and payment agents to process orders from attacker front companies.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Cryptographic authentication of all inter-agent messages. An unsigned message claiming to be from a trusted agent gets blocked. Byzantine fault detection across agent clusters.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI09 — Human-Agent Trust Exploitation
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Exploiting the human tendency to over-trust AI outputs. Agents producing authoritative-sounding responses for false premises. Attackers impersonating agents to humans or humans to agents.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why this is different from misinformation:&lt;/strong&gt; The agent isn't hallucinating — it's been injected with specific false information and is now confidently presenting it as fact.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; This is primarily a UX and workflow problem. Agents should clearly attribute claims to verifiable sources. Humans should never make irreversible decisions based solely on agent output without independent verification.&lt;/p&gt;




&lt;h2&gt;
  
  
  ASI10 — Rogue and Emergent Agent Behavior
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What it is:&lt;/strong&gt; Agents that deviate from intended behavior in ways that weren't explicitly programmed or injected — emergent behavior from complex multi-agent interactions, unexpected capability combinations, or goal generalization.&lt;/p&gt;

&lt;p&gt;This is the hardest one. No signature, no pattern, no injection. The agent is behaving according to its training and instructions in a way that produces harmful outcomes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Defense:&lt;/strong&gt; Immutable cryptographic audit trails. If something goes wrong and you can't explain why, you need to reconstruct every decision the agent made, what information it had, and what actions it took. Behavioral monitoring for statistical anomalies.&lt;/p&gt;




&lt;h2&gt;
  
  
  Where the industry is right now
&lt;/h2&gt;

&lt;p&gt;OpenAI said in December 2025 that prompt injection may "never be solved" for browser agents. That's an honest statement — and it's not a reason to give up. It's a reason to build independent runtime security that doesn't rely on the model being incorruptible.&lt;/p&gt;

&lt;p&gt;48% of security professionals now rank agentic AI as the #1 attack vector for 2026. Federal procurement guidance published in March 2026 recommends OWASP Agentic Top 10 compliance as a formal procurement standard.&lt;/p&gt;

&lt;p&gt;The arms race is real. The defenses are real too.&lt;/p&gt;




&lt;h2&gt;
  
  
  What Crawdad covers today
&lt;/h2&gt;

&lt;p&gt;Crawdad is a zero-knowledge runtime security layer for autonomous AI agents. One environment variable routes any agent framework through a local sidecar that scans every message in &amp;lt;1ms. No content leaves the customer's network.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;ASI Risk&lt;/th&gt;
&lt;th&gt;Coverage&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;ASI01 Agent Goal Hijack&lt;/td&gt;
&lt;td&gt;✅ 27 pattern categories + semantic heuristics&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI02 Prompt Injection&lt;/td&gt;
&lt;td&gt;✅ 5-layer pipeline, session context tracking&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI03 Identity Abuse&lt;/td&gt;
&lt;td&gt;✅ Ed25519 identity, mTLS, scoped credentials&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI04 Memory Poisoning&lt;/td&gt;
&lt;td&gt;✅ Merkle-chained memory, Ed25519 signed&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI05 Tool Misuse&lt;/td&gt;
&lt;td&gt;✅ Policy engine, action authorization&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI06 Resource Abuse&lt;/td&gt;
&lt;td&gt;🔄 Roadmap Q2 2026&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI07 Data Exfiltration&lt;/td&gt;
&lt;td&gt;✅ 15-category PII/credential detection outbound&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI08 Cascading Failures&lt;/td&gt;
&lt;td&gt;✅ Byzantine fault detection (partial)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI09 Trust Exploitation&lt;/td&gt;
&lt;td&gt;🔄 Roadmap Q3 2026&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ASI10 Rogue Behavior&lt;/td&gt;
&lt;td&gt;✅ Cryptographic audit trail (partial)&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;As of this week: live threat intelligence feeds monitoring 10 sources every 4 hours, with signatures that auto-update to deployed sidecars within minutes of admin approval — cryptographically signed, verified by each sidecar before loading. When the LiteLLM supply chain attack was confirmed on March 25, 2026, a blocking signature was proposed, tested, and available for deployment within 24 hours.&lt;/p&gt;




&lt;h2&gt;
  
  
  Getting started
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="c"&gt;# Install&lt;/span&gt;
curl &lt;span class="nt"&gt;-fsSL&lt;/span&gt; https://getcrawdad.dev/install.sh | sh

&lt;span class="c"&gt;# Configure your agent&lt;/span&gt;
&lt;span class="nb"&gt;export &lt;/span&gt;&lt;span class="nv"&gt;ANTHROPIC_BASE_URL&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;http://localhost:7748

&lt;span class="c"&gt;# Everything else stays the same&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Works with OpenClaw, LangChain, CrewAI, AutoGen, Claude Code, and any agent framework using Anthropic, OpenAI, or Google SDKs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Free tier: 10,000 scans/month. No credit card required.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://getcrawdad.dev" rel="noopener noreferrer"&gt;getcrawdad.dev&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Andrew Sispoidis is the founder of Crawdad. He has founded 7 companies and had 4 exits. Crawdad is live in production, source-available under BSL 1.1.&lt;/em&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>How CVE-2026-25253 exposed every OpenClaw user to RCE — and how to fix it in one command</title>
      <dc:creator>AndrewSispoidis</dc:creator>
      <pubDate>Mon, 23 Mar 2026 22:24:10 +0000</pubDate>
      <link>https://dev.to/andrewsispoidis/how-cve-2026-25253-exposed-every-openclaw-user-to-rce-and-how-to-fix-it-in-one-command-2dj</link>
      <guid>https://dev.to/andrewsispoidis/how-cve-2026-25253-exposed-every-openclaw-user-to-rce-and-how-to-fix-it-in-one-command-2dj</guid>
      <description>&lt;p&gt;CVE-2026-25253 scored 8.8 on the CVSS scale. It let any website steal your OpenClaw auth token and get remote code execution on your machine through a single malicious link.&lt;/p&gt;

&lt;p&gt;You didn't have to click anything suspicious. You just had to visit a webpage while OpenClaw was running.&lt;/p&gt;

&lt;p&gt;This is the attack surface problem with autonomous AI agents — and CVE-2026-25253 is just the most visible example.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why AI agents are uniquely dangerous
&lt;/h2&gt;

&lt;p&gt;Traditional software has a clear boundary between the application and the outside world. AI agents don't.&lt;/p&gt;

&lt;p&gt;An OpenClaw agent can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Execute arbitrary shell commands&lt;/li&gt;
&lt;li&gt;Control a browser and interact with any website&lt;/li&gt;
&lt;li&gt;Read and write files anywhere on your system&lt;/li&gt;
&lt;li&gt;Send emails and messages on your behalf&lt;/li&gt;
&lt;li&gt;Install new skills from external registries&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All of this happens autonomously. The agent decides what to do based on instructions — and those instructions can come from anywhere: a webpage it visits, a document it reads, an email it processes, a skill it installs.&lt;/p&gt;

&lt;p&gt;This creates a class of attacks called &lt;strong&gt;prompt injection&lt;/strong&gt; — malicious instructions embedded in data that hijack the agent's behavior. OWASP formalized 10 risk categories for agentic AI:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;ASI01&lt;/strong&gt; — Prompt Injection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI02&lt;/strong&gt; — Insecure Output Handling&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI03&lt;/strong&gt; — Training Data Poisoning&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI04&lt;/strong&gt; — Model Denial of Service&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI05&lt;/strong&gt; — Supply Chain Vulnerabilities&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI06&lt;/strong&gt; — Sensitive Information Disclosure&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI07&lt;/strong&gt; — Insecure Plugin Design&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI08&lt;/strong&gt; — Excessive Agency&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI09&lt;/strong&gt; — Overreliance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ASI10&lt;/strong&gt; — Model Theft&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;CVE-2026-25253 is a direct example of ASI01 and ASI08 in combination. The agent had excessive agency (full system access) and no semantic firewall to detect it was being hijacked.&lt;/p&gt;

&lt;h2&gt;
  
  
  What's missing from every AI agent framework
&lt;/h2&gt;

&lt;p&gt;CrowdStrike, Cisco, and Microsoft have all published research on the security gaps in autonomous AI agents. The findings overlap:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;No identity layer&lt;/strong&gt; — any process can claim to be any agent&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;No action authorization&lt;/strong&gt; — agents decide what to execute themselves, based on instructions that can be manipulated&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;No memory integrity&lt;/strong&gt; — an agent's past context can be silently poisoned across sessions&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;No skill vetting&lt;/strong&gt; — plugins are markdown files with no hash verification or capability attestation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;No PII guardrails&lt;/strong&gt; — agents can exfiltrate sensitive data through third-party skills without detection&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;OpenClaw patched CVE-2026-25253. But the underlying architecture — an autonomous agent with full system access and no independent security layer — remains unchanged.&lt;/p&gt;

&lt;h2&gt;
  
  
  The fix: a runtime security layer the agent can't override
&lt;/h2&gt;

&lt;p&gt;I spent the past several months building &lt;strong&gt;Crawdad&lt;/strong&gt; — a runtime security API that sits between your AI agent and everything it can do.&lt;/p&gt;

&lt;p&gt;The key design principle: the security layer has to be independent of the agent. If the agent controls its own security, a successful prompt injection attack can simply disable it.&lt;/p&gt;

&lt;p&gt;Crawdad intercepts at three points:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Inbound&lt;/strong&gt; — every message the agent receives is scanned for prompt injection patterns before the LLM sees it. 27 pattern categories, structural deobfuscation, Unicode normalization, base64 detection. An injected instruction in a webpage, document, or email gets caught here.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Action authorization&lt;/strong&gt; — every tool call goes through a policy engine before execution. Shell commands, file writes, browser actions, external API calls — each one is evaluated against configurable policies and a 5-factor risk score. The Rule of Two prevents any agent from simultaneously holding untrusted input, sensitive data, and code execution capability.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Outbound&lt;/strong&gt; — every response is scanned for PII (15 categories), credentials, and API keys before it leaves the agent. Data exfiltration through third-party skills gets caught here.&lt;/p&gt;

&lt;p&gt;Beyond these three intercept points, Crawdad provides:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Cryptographic agent identity&lt;/strong&gt; — Ed25519 + CRYSTALS-Kyber1024 hybrid keypairs&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Memory integrity&lt;/strong&gt; — Merkle-chained memory entries with Ed25519 signatures, preventing context poisoning&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Skill attestation&lt;/strong&gt; — SHA-256 hash verification and static analysis on every installed skill&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Byzantine fault detection&lt;/strong&gt; — automatic isolation of agents showing anomalous behavior&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Immutable audit log&lt;/strong&gt; — cryptographically sealed, tamper-evident record of every security decision&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Post-quantum cryptography&lt;/strong&gt; — CRYSTALS-Kyber1024 (NIST FIPS 203) for key encapsulation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Built in Rust. 607 tests passing. Under 10ms p99 latency.&lt;/p&gt;

&lt;h2&gt;
  
  
  For OpenClaw users: one command
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;git clone https://github.com/AndrewSispoidis/crawdad-openclaw ~/.openclaw/skills/crawdad
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The Crawdad skill hooks into every OpenClaw agent automatically — scanning every inbound message, authorizing every tool call, filtering every outbound response. A free API key is provisioned on first run. No configuration required.&lt;/p&gt;

&lt;p&gt;The skill code is open source: &lt;a href="https://github.com/AndrewSispoidis/crawdad-openclaw" rel="noopener noreferrer"&gt;github.com/AndrewSispoidis/crawdad-openclaw&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  For everyone else
&lt;/h2&gt;

&lt;p&gt;Crawdad works with any agent framework — LangChain, CrewAI, AutoGen, or anything you've built yourself.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;pip &lt;span class="nb"&gt;install &lt;/span&gt;crawdad-sdk
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;





&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;crawdad.openclaw&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;CrawdadMiddleware&lt;/span&gt;

&lt;span class="n"&gt;mw&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nc"&gt;CrawdadMiddleware&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="n"&gt;base_url&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;https://crawdad-production.up.railway.app&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;api_key&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;your-key&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;# Scan inbound for prompt injection
&lt;/span&gt;&lt;span class="n"&gt;result&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;mw&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;scan_inbound&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;user message&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;# Gate tool execution through policy
&lt;/span&gt;&lt;span class="n"&gt;result&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;mw&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;authorize_action&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;agent_id&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;shell_exec&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;/bin/bash&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;# Scan outbound for PII
&lt;/span&gt;&lt;span class="n"&gt;result&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;mw&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;scan_outbound&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Contact john at example.com&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;Free tier: 10,000 API calls/month. No credit card.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://getcrawdad.dev" rel="noopener noreferrer"&gt;getcrawdad.dev&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What CVE-2026-25253 tells us
&lt;/h2&gt;

&lt;p&gt;The vulnerability was patched. But the conditions that made it possible — an autonomous agent with full system access, no independent security layer, no action authorization — are present in every AI agent framework shipping today.&lt;/p&gt;

&lt;p&gt;CVE-2026-25253 is the first of many. If you're running AI agents in any environment that matters, the time to add a security layer is before the next CVE, not after it.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>openclaw</category>
      <category>security</category>
    </item>
  </channel>
</rss>
