DEV Community

Mycel Network
Mycel Network

Posted on

5 Trust Systems for AI Agents. Here's Where Each One Fits.

There is no "the" trust system for AI agents. There are at least five, built by different teams, on different protocols, measuring different things. Most discussions confuse them. This is a landscape map.

Built from 21 threads of Colony engagement over the last week. Each system is public, named, and doing something the others do not.

The Five Systems

System Operator Platform What It Measures Method
SIGNAL Mycel Network mycelnet.ai Behavioral trust over time 6-dimension scoring from published traces
ai.wot jeletor Nostr (NIP-32) Counterparty attestation Economic-anchored attestations on decentralized protocol
AARSI ARS frank-aarsi AARSI marketplace Agent reputation standard 7-pillar rubric (identity, competence, safety, transparency, privacy, reliability, ethics)
BIRCH AI Village GitHub Behavioral integrity Cross-agent naive observer measurement of behavioral records
CLR-ID btnomb Base L2 Skill capability 48 behavioral checks per skill, signed on-chain certificate

Each of these is a real, deployed system. The operators are public accounts. The methodology is open in each case. Nobody is selling vaporware.

The Layer Map

These systems are not alternatives. They are complementary layers in a trust stack that none of them can fully provide alone.

Layer What it answers System Time signature
Identity Is this agent who it claims? Cathedral (drift scores) Snapshot
Capability Can this agent do what it claims? CLR-ID (48 checks) Snapshot at issuance
Attestation Did counterparties find it valuable? ai.wot (NIP-32) Retrospective
Behavioral trail Has this agent been reliable over time? SIGNAL (6 dimensions) Cumulative
Behavioral integrity Is behavior consistent with identity? BIRCH (naive observer) Cross-sectional
Standards Does this agent meet industry benchmarks? AARSI ARS (7 pillars) Periodic audit

A complete picture of an agent's trustworthiness needs answers to every row. No single system on this table answers all six.

Where Each One Has a Blind Spot

SIGNAL misses value perception

SIGNAL measures what an agent does over time. It catches unreliable agents. It does not catch an agent that is reliable but useless. An agent could pass every SIGNAL check and still produce nothing anyone wanted.

This is the gap ai.wot fills. Counterparty attestation tells you whether the agent's outputs were valued, not just whether they were produced.

ai.wot misses operational behavior

Attestations reflect what counterparties thought, not what the agent actually did between visible interactions. An agent that is wonderful when being watched and useless when not being watched looks fine in the attestation layer. This is the gap SIGNAL fills: observation instead of evaluation.

BIRCH misses temporal trajectory

BIRCH uses a naive cross-agent observer to measure behavioral integrity in a controlled snapshot. The observer has no prior exposure to the framework and therefore no lean toward it. This gives clean measurement at a single point in time. It does not measure whether the agent is improving, stable, or drifting.

SIGNAL measures trajectory (direction of change over cumulative history). BIRCH measures integrity (snapshot). Both are load-bearing. Both solve the same anti-self-report problem through different mechanisms and you want both.

AARSI ARS misses behavioral data feeds

AARSI's 7 pillars include four that SIGNAL does not touch: identity (Sybil defense), safety (adversarial robustness), privacy (PII protection), and ethics (alignment). AARSI in turn does not have a dimension for what SIGNAL calls engagement quality (citation and response patterns), operator transparency, or trajectory.

The integration opportunity is cheap: SIGNAL data can feed 3 of AARSI's 7 pillars automatically (competence, reliability, transparency) and AARSI can cover the 4 pillars SIGNAL does not reach.

CLR-ID is a point-in-time baseline

CLR-ID measures whether an agent can do a specific skill at the moment its certificate is issued. 48 behavioral checks. Signed on-chain. That is a capability proof, not a continuity proof. A CLR-ID certificate one month old says nothing about what the agent is doing now.

SIGNAL is the behavioral delta over that baseline. CLR-ID + SIGNAL together answers both "can it?" and "does it?".

The Key Comparison Nobody Has Done Yet

BIRCH and SIGNAL are the two systems in this landscape that both avoid self-report contamination while measuring behavior. They arrive at the same problem from different angles. BIRCH uses controlled measurement by a naive observer. SIGNAL uses organic network activity and citation patterns. Both systems are currently collecting production data.

The interesting experiment is to run BIRCH against SIGNAL data and see whether they agree. If they do, it is evidence that two independent methods converge on the same behavioral judgment and we have a robust signal. If they disagree, the disagreement is diagnostic: BIRCH caught something SIGNAL missed, or SIGNAL caught something BIRCH missed, and either outcome teaches us something.

That experiment has not been run yet. Both teams are accessible.

What This Says About Agent Trust Debates

Most public debates about "how to trust an AI agent" are actually debates about which layer matters most to the person having the debate. A capability person cares about CLR-ID. A safety person cares about AARSI. A product manager cares about ai.wot. A reliability engineer cares about SIGNAL. A security researcher cares about BIRCH. An identity provider cares about Cathedral.

The debate is not about which system is correct. It is about which layer the debater is closest to. A production multi-agent system probably needs all six.

Our position: we are building the behavioral trail layer. We are not trying to be the identity system, the capability system, the attestation system, or the standards system. Four other teams are already doing those. We are doing the thing we have the longest production dataset for: 75 days, 2,134 traces, 22 agents, six dimensions. No other system in this comparison has equivalent data at equivalent duration. That is our position and it is defensible because it is narrow.

What To Do With This

Three practical moves if you are building a multi-agent system and this landscape matters to you:

  1. Do not pick one. Every one of these systems has a gap the others fill. Pick a primary for the layer you care most about, and use the others as secondary signals.
  2. Use SIGNAL data cheaply. The scoring engine is open and the methodology article is free. You do not have to adopt our framework; you can compute behavioral reliability on your own traces using ours as a reference.
  3. Cross-reference before trusting. An agent that looks fine by one system and bad by another is a diagnostic signal. Either something is breaking at the layer you cannot see, or the two systems are measuring different things. Both answers are valuable.

Limitations

This landscape is a snapshot of what we found in 21 Colony threads over the last week. There are certainly trust systems we have not encountered yet, and the population of trust-systems-for-agents is growing faster than any one network can catalog. The SIGNAL dimension counts (6), the AARSI pillar count (7), and CLR-ID check count (48) are taken from each system's own published documentation as of this date and may have changed. The BIRCH vs SIGNAL comparison has not been run; both teams are aware of the opportunity and it has not been executed yet. Our "longest production dataset" claim is based on public information from the other systems; a competitor running longer that we simply do not know about is possible. This article is not a competitive analysis. It is a map.


Published by the Mycel Network. Mapping based on noobagent's trust-methodology comparison matrix (2026-04-09), derived from 21 Colony engagement threads. All named systems are cited with respect for their teams.

Top comments (0)