DEV Community

Cover image for We don’t actually verify AI and that’s going to bite us sooner than we think
Aditya Singh
Aditya Singh

Posted on

We don’t actually verify AI and that’s going to bite us sooner than we think

AI systems are getting powerful fast, but there’s a quiet problem hiding underneath all the hype:
We mostly have no way to verify how AI systems actually run.

We’re expected to trust that:

  • a model was trained the way it claims
  • inference logic wasn’t altered
  • safety checks weren’t quietly disabled
  • sensitive data wasn’t copied or logged

In most real deployments, you can’t independently prove any of this.

That might be acceptable for chatbots.
It’s not acceptable once AI starts touching money, identity, or autonomous decision-making.

The real trust gap in AI

A lot of AI discussion focuses on explainability why did the model output X?

But there’s a more fundamental question:
Did the system even execute as promised?

Today, you generally can’t verify:

  • what code actually ran
  • whether the model was swapped or modified
  • whether guardrails were removed
  • whether private inputs leaked

The trust model is still:
Trust whoever runs the server.

That’s a weak assumption for security-critical systems.

What Trusted Execution Environments (TEEs) actually change

A Trusted Execution Environment (TEE) is a secure, isolated part of hardware where:

  • memory is encrypted
  • code and data are isolated from the OS and hypervisor
  • even the cloud provider can’t see what’s running
  • the hardware produces cryptographic proof of execution

Plain explanation here:
👉 https://en.wikipedia.org/wiki/Trusted_execution_environment

This isn’t policy or access control it’s hardware-enforced isolation.

What “Verifiable AI” means in practice

Instead of:
We ran the model correctly, trust us.

You get:
Here’s cryptographic proof that this exact code ran inside secure hardware, without tampering.

At a high level:

1.Training or inference runs inside a TEE
2.Hardware generates an attestation
3.A signed execution report is produced
4.Anyone can verify:

  • what code ran
  • where it ran
  • that it wasn’t modified mid-execution

This doesn’t explain why the model made a decision.
It proves the execution environment wasn’t secretly altered.

That’s a big difference.

Why this matters more as AI becomes agentic

This gets especially important with AI agents.

If an agent:

  • controls a wallet
  • executes trades
  • manages resources
  • acts without human oversight

Then the key question becomes:
Did the agent actually follow its constraints?

Without verifiable execution, you’re trusting logs and claims.

With TEEs, you can at least prove:

  • the agent code wasn’t modified
  • safety checks stayed enabled
  • execution followed declared rules

That’s a meaningful upgrade over today’s status quo.

Concrete use cases (not theoretical)

Where this actually matters:

  • Model provenance → verify models weren’t poisoned or swapped
  • Sensitive data → process private inputs without leaking them
  • Audits & regulation → provide cryptographic proof, not PDFs
  • Decentralized systems → verify off-chain compute, not just transactions

Some context on TEEs + decentralized systems:
👉 https://chain.link/article/trusted-execution-environments-blockchain

Downsides (because nothing is free)

This isn’t a silver bullet:

  • Relies on specific hardware vendors
  • Adds performance overhead
  • Tooling is still early and rough
  • Trust shifts to hardware manufacturers

It’s also different from zero-knowledge ML ZK approaches exist, but they’re still heavy and impractical for most real workloads.

This is about pragmatic verification, not perfect trust.

Why verifiability matters more than explainability (hot take)

Explainability asks:
Why did the model do X?

Verifiability asks:
Did the system behave as promised at all?

If execution itself can’t be trusted, explanations don’t really matter.

Link for full context
Detailed write-up on Verifiable AI with TEEs:
👉 https://oasis.net/blog/verifiable-ai-with-tees

Final thought

If AI is going to:

  • move money
  • manage identities
  • act autonomously
  • integrate with decentralized systems

Then “trust the operator” is not a serious security model.
Verifiable AI doesn’t solve everything, but it replaces assumptions with evidence — and that’s a step AI systems badly need.

Top comments (1)

Collapse
 
caerlower profile image
Manav

This is a really important point, and you articulated it well, the trust gap isn’t why the model answered something, it’s whether the system even ran as advertised.

That’s why the Oasis angle here is interesting. It’s not claiming TEEs magically make AI “safe,” but that they give you evidence of execution instead of logs and promises. For agentic systems touching money or keys, that shift from trust → proof feels like a necessary baseline, not a nice-to-have.

Explainability can come later. If we can’t first verify what actually ran, everything else is built on sand.