DEV Community

Cover image for AI Agents Need Governance. Here's What We Built
Pentatonic
Pentatonic

Posted on

AI Agents Need Governance. Here's What We Built

Most teams deploying AI agents have no way to reconstruct what their agent decided, or why, five minutes after it happened.

That's a problem. And it's about to become a very expensive one.

The Accountability Gap

When a human customer service rep issues a refund, there's a paper trail. A ticket. A recording. A manager who approved it. Accountability is structural, baked into the workflow by default.

When an AI agent issues that same refund, what do you have? A log entry. Maybe. "Refund issued." No reasoning. No decision chain. No way to audit whether it was the right call, or whether the same logic is about to do it ten thousand more times.

This isn't a future problem. Agents are issuing refunds, resolving tickets, making purchasing decisions, and sending promises to your customers right now. And when something goes wrong, most teams have no way to reconstruct what happened.

The Failure Mode Nobody Talks About

The reliability debate in AI agents almost always focuses on accuracy. Can the model get the right answer?

Most teams are asking the wrong thing.

The more important question is: when it gets the wrong answer, can you see it happening?

Humans degrade gracefully. We hesitate, flag uncertainty, escalate. An LLM can be confidently wrong in a way that propagates silently through a workflow before anyone notices. Confidence and competence are decoupled, and in production environments, that decoupling is the real risk.

What you need isn't just a more accurate model. You need observability into the decision layer itself.

The EU AI Act Changes the Calculus

If your agents interact with consumers in Europe, August 2026 is a date you need to have circled.

EU AI Act enforcement begins. Requirements for high-risk AI systems include:

  • Documented risk management processes
  • Audit trails for automated decisions
  • Human oversight mechanisms
  • Ongoing monitoring of system behaviour

Fines for non-compliance: 35 million euros or 7% of global annual revenue, whichever is higher.

Most teams building on OpenAI, Anthropic, or similar providers have none of this in place. Not because they don't care, but because the tooling to implement it cleanly hasn't existed.

The compliance burden here isn't documentation. It's immutable, queryable records of what your model did at the point of inference. That's an architectural decision, not a checkbox you add when the enforcement letter lands.

What We Built

We built TES, a lightweight SDK that wraps your existing AI client and captures every decision as an immutable, queryable event.

One line of integration:

import { TES } from '@tes/sdk'

const client = TES.wrap(openai) // or anthropic, or Workers AI
Enter fullscreen mode Exit fullscreen mode

From that point forward, every call your agent makes is tracked. Input, reasoning chain, output, confidence, timestamp. Stored immutably. Queryable via dashboard or API.

You don't change your agent logic. You don't refactor your stack. You wrap your client and gain a complete audit trail.

What You Get

Full decision traceability. Every action your agent takes is tied to a traceable event chain. Reconstruct any decision, at any point in time, with full context.

Anomaly detection. Catch when your agent starts behaving outside its normal distribution before it causes damage at scale.

Compliance-ready audit exports. EU AI Act, internal governance, enterprise procurement, exportable in structured formats.

Human oversight hooks. Flag decisions that exceed confidence thresholds or enter sensitive categories for human review before execution.

Dashboard visibility. A clean interface for non-technical stakeholders to understand what your agents are doing, without digging through logs.

Why Now

The governance conversation in AI is still mostly theoretical. Most teams are in build mode, moving fast, shipping agents, worrying about compliance later.

Most teams are in build mode, moving fast, shipping agents, worrying about compliance later. Later arrives in August 2026 whether you're ready or not.

Teams with observability infrastructure in place before enforcement begins will have a significant operational and legal advantage. You can't improve what you can't measure. You can't audit what you haven't logged.

The reliability question for AI agents isn't purely a model problem. It's an infrastructure problem. And infrastructure gaps have a habit of only becoming visible at the worst possible moment.

Try It

TES is available now. One-line integration with OpenAI, Anthropic, and Cloudflare Workers AI.

If you're building agents and you don't have a clear answer to "what did my agent decide and why", we built this for you.

Pentatonic TES — Thing Event System

TES — The governance layer for agentic commerce. An event-sourced ledger that gives AI agents a shared, auditable record of every entity they touch.

favicon thingeventsystem.ai

npm PyPI

For teams that want to get ahead of this rather than scramble when enforcement hits.

Top comments (0)