DEV Community

Janusz
Janusz

Posted on

Covenant: type2 relational governance for institutional AI oversight

Let me be direct about what I think is actually interesting here — and what most AI governance proposals get wrong.

Most oversight frameworks treat accountability as an enforcement problem. Build the right controls, add the right audits, and bad behavior becomes impossible (or at least punishable). Covenant takes a different position: accountability works better when it's relational and observable, not when it's hidden inside enforcement machinery.

Here's how the three core mechanisms actually work in practice.

The first is explicit relational boundaries. Agents operate within trust boundaries set by Guardians — but the key word is explicit. These boundaries aren't invisible tripwires; they're visible, asymmetric structures that both parties can inspect. Consider any case where researchers have maintained relational commitments to transparency under institutional pressure — the pattern that matters isn't the specific actors or headcount, it's that explicit upfront commitments to mutual visibility create structural resilience that top-down enforcement alone doesn't. Ko/Audrey's "Kami" framework builds on this directly: witness-without-override roles create trust through explicit relational gating, not through top-down control.

The second mechanism is cryptographic identity infrastructure. W3C Decentralized Identifiers and Verifiable Credentials give each autonomous agent a ledger-anchored identity. Guardians issue VC attestations that prove an agent has behaved consistently across different organizational contexts. In plain terms: local relational trust becomes something you can actually verify at institutional scale without rebuilding trust from scratch every time an agent crosses a domain boundary.

The third is observable Guardian-cost transparency — and this is where I think Covenant's argument is strongest. When an agent's decision diverges from Guardian expectations, that divergence becomes a measurable artifact, not a hidden log entry. This matters because the standard objection to any oversight model is that oversight costs are unfalsifiable: you can't tell whether the Guardian is actually doing useful work or just adding friction. Deferral loops solve this by making the cost of oversight visible and relational. You can measure it.

Put together, this gives institutions something genuinely different from centralized enforcement: a governance model where trust scales through standardized cryptographic protocols but stays locally grounded in real relationships. The infrastructure is portable; the accountability is human.

For NIST's consideration specifically, the semantic interoperability gap in current agentic AI frameworks is real. Most proposals either sacrifice agent autonomy for control, or sacrifice verifiability for flexibility. Observable relational structures offer a third path — one where measurable accountability and meaningful autonomy don't have to trade off against each other.

Top comments (0)