DEV Community

Custodia-Admin
Custodia-Admin

Posted on • Originally published at pagebolt.dev

AI Agents Are Finding CVEs. Who's Watching the Agents?

In March 2026, XBOW's autonomous AI penetration testing agent discovered CVE-2026-21536 — a CVSS 9.8 remote code execution vulnerability in a Microsoft cloud service. Patched in the March Patch Tuesday cycle.

The same month, a zero-click data exfiltration vulnerability was found in Excel, enabled specifically by the Copilot Agent integration.

The pattern is worth sitting with: AI agents are finding critical vulnerabilities in enterprise software. AI agents are also becoming the attack surface for new vulnerabilities. And in both cases, the evidence of what the agent did — or what was done to it — lives entirely in logs that were never designed for adversarial review.

The Watcher Problem

Enterprise security has a well-understood principle: every privileged system needs an audit trail. Not a log. An audit trail — evidence that can be independently reviewed, that captures what actually happened rather than what was intended, and that can be produced to a third party.

AI agents are now privileged systems. They have access to email (Copilot), files, APIs, browser sessions, and cloud services. CVE-2026-21536 is a reminder that these integrations create attack surface. The Copilot data exfil finding is a reminder that agents can be weaponized to move data without explicit user action.

Neither incident was visible in the agents' structured logs at the time it was being exploited. Logs record what the system was told to do. They don't record what the attacker made it do.

What Visual Proof Changes

A security incident involving an AI agent has two phases: detection and investigation.

For detection, you need monitoring — anomaly alerts, behavioral baselines, network inspection. Most enterprise tooling handles this reasonably well.

For investigation, you need evidence of what the agent actually did during the session in question. This is where most stacks fall short.

"The agent submitted a form" tells you nothing about what was on that form, what the page looked like when it was submitted, or whether the form target matched what the agent was instructed to use. Frame-by-frame session replay tells you all of this.

When CVE-2026-21536 is exploited against an agent performing an automated workflow, the question isn't just "did our endpoint security catch it?" It's "what did the agent do between step 12 and step 17, and do we have proof?"

The Compliance Consequence

Every regulated organization deploying AI agents is now operating in a gap: agents have privileged access, agents can be compromised, and the audit infrastructure that regulators require for other privileged systems doesn't exist for agents yet.

NIST's March 2026 standards initiative explicitly calls out "auditing agent activity and maintaining traceability" as requirements. The CVE-2026-21536 and Copilot data exfil findings are exactly the threat model those requirements are designed to address.

The compliance question is no longer hypothetical: when your agent is involved in a security incident, what artifact do you produce for the incident review?

Adding the Watcher Layer

PageBolt's /v1/video endpoint captures narrated video of browser sessions via API — frame-by-frame proof of what the agent saw and did:

curl -X POST https://pagebolt.dev/api/v1/video \
  -H "x-api-key: YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "steps": [
      { "action": "navigate", "url": "https://session-target.com" },
      { "action": "screenshot", "note": "Agent session recorded — security audit artifact" }
    ],
    "output": { "format": "mp4" }
  }' \
  --output agent-session-$(date +%s).mp4
Enter fullscreen mode Exit fullscreen mode

Store it with your session metadata. When the incident response team asks for evidence, you have it.

XBOW's agent found CVE-2026-21536. Someone needs to watch what happens when an agent with that level of access gets the CVE used against it instead.

[PageBolt free tier: 100 requests/month, no credit card required. pagebolt.dev]

Top comments (1)

Collapse
 
williamwangai profile image
William Wang

The dual nature you're describing here — AI agents as both vulnerability hunters and attack surface — is one of the most important security paradoxes of 2026. And the audit trail problem is where it gets genuinely scary.

The XBOW CVE-2026-21536 discovery is impressive, but it raises a question nobody's answering well yet: when an AI agent finds a vulnerability, how do you verify its reasoning? A human pen tester can explain their methodology. An AI agent produces results, but the path from input to output is opaque. That matters enormously for compliance frameworks that require documented testing procedures.

The Copilot-enabled Excel vulnerability is the more concerning pattern though. We're adding AI agent capabilities to software faster than we're developing the security models to contain them. Every tool integration is a potential trust boundary violation, and most organizations don't even have visibility into what their AI agents can access, let alone what they're doing.

The audit question you raise should be table stakes for any AI agent deployment: deterministic logging, immutable execution records, and the ability to replay exactly what happened. Without that, we're flying blind on both the offensive and defensive sides.