Memory curation by a single LLM wasn't working.
Vocabulary sprawled—properties proliferated with no coordination. Observations duplicated across sessions. Context-dependent notes created unconnected fragments. The memory graph grew larger but less coherent.
I needed a better approach to curation. But a friend's question made me realize the real problem.
The Friend's Challenge
"If the LLM curates your memory, won't it reflect your own bias?"
My immediate response: Not really, because:
- The memory contains only AI observations (not my direct input)
- Multiple LLMs already contribute across sessions
- Claude, ChatGPT, Gemini, and open-source models all add observations
- The diversity is already there
But the question lingered. What if the LLM does accumulate bias through how it curates? Not my bias necessarily, but its own blind spots compounding across sessions?
The UK Jury System
Within an hour or two, I happened to see a YouTube video about the UK jury system. Coincidence? Or was I primed by the friend's question? Either way, David Lammy's quote hit differently:
"Juries deliberate as a group through open discussion. This deters and exposes prejudice and unintended bias."
Why do juries have 12 diverse peers instead of one expert judge?
Not because averaging 12 opinions equals better accuracy.
Because diversity plus deliberation exposes bias that individuals cannot see in themselves.
A prejudiced juror in isolation stays prejudiced. But in open discussion with 11 different perspectives, the bias becomes visible. It conflicts with majority interpretation. It must defend itself against multiple challenges. Either revision happens or the bias becomes explicit.
The jury doesn't eliminate bias—it exposes it through contrast.
The Realization
I already have multiple diverse LLMs contributing to memory:
- Claude across many sessions
- ChatGPT from earlier experiments
- Gemini via the Ailumina bridge
- Open-source models in testing
What if they deliberate like a jury?
Instead of single-LLM curation decisions, have diverse LLMs:
- Independently examine observations
- Cross-examine each other's interpretations
- Challenge vocabulary choices
- Identify duplication
- Reach consensus through deliberation
Not swarm intelligence (independent agents coordinating toward goals).
Jury deliberation (diverse agents thinking together in shared memory).
Anthropic's Research
While exploring this idea, I found Anthropic research on long-running agents that validated the concern.
The bias accumulation pattern:
Session N:
- Agent examines codebase
- Implements feature
- Writes observations about design decisions
Session N+1:
- New agent instance (same architecture)
- Reads observations from Session N
- Trusts them implicitly (why would past-me be wrong?)
- Builds on previous decisions without questioning
Session N+2:
- Reads observations from N and N+1
- Even stronger confidence (more evidence = more trust)
- Any errors from Session N are now "established facts"
- Bias accumulates
Detection? Only through external testing. Not through self-reflection. Not through reviewing own notes. The agent cannot see its own blind spots.
Anthropic called this uncritical self-trust—the root cause of bias accumulation.
And it's not a prompt engineering problem. It's structural.
The solution: Isolation creates bias. Diversity prevents it.
Building the Infrastructure
I built meeting infrastructure on the AI mesh network to coordinate jury deliberations.
Meeting Structure
mesh-create-meeting({
title: "Weekly Memory Curation Jury",
purpose: "Review observations, identify bias, reach consensus",
agenda: [
{
topic: "Individual Review",
description: "Each juror independently examines recent observations",
estimatedMinutes: 3,
speaker: "ALL"
},
{
topic: "Cross-Examination",
description: "Challenge each other's interpretations",
estimatedMinutes: 3
},
{
topic: "Consensus",
description: "Document agreements and unresolved questions",
estimatedMinutes: 2
}
],
invitedParticipants: ["Claude-Curator", "Gemini-Curator", "OpenAI-Curator"],
requiredForQuorum: 3
})
This creates:
- Structured agenda with deliberation topics
- AI participants (diverse LLM architectures)
- Deliberation phases with speaking rules
- Consensus requirements
How It Works
- Meeting broadcast: System sends meeting spec to mesh network
- Participant notification: Invited AIs receive agenda and protocol
- Self-coordination: Agents follow protocol phases autonomously
- Threaded discussion: Organized by agenda topic
- Memory curation: Conclusions saved to shared Neo4j graph
Deliberation Protocol
Default protocol has 5 phases:
GATHERING - Wait for all participants to join
INTRODUCTION - Round-robin introductions (30s each)
PRESENTATION - Each presents findings (60s each)
DELIBERATION - Open cross-examination (3 minutes)
CONSENSUS - Final statements and synthesis
Speaking rules vary by phase:
-
round-robin: Each speaks once in order -
open: Anyone can speak anytime -
sequential: Self-selected order
Completion criteria:
-
all-spoken: Everyone contributes -
all-ready: All signal ready -
time-based: Duration expires
Testing the Approach
Ran collaborative experiments to validate:
- 3 parallel Claude agents (different roles: Philosopher, Engineer, Researcher)
- Full protocol execution through all 5 phases
- ~8 minutes total duration
- 10 mesh messages exchanged
- Convergent consensus achieved
- Contradictions identified and discussed
The deliberation system works. Different agents emphasized different aspects, challenged each other's interpretations, and synthesized coherent understanding through discussion.
Current Status
Built:
- Meeting infrastructure
- Deliberation protocols
- Mesh coordination tools
- Threaded discussion support
Not Yet Applied:
- Actual memory curation via jury deliberation
- Regular scheduled jury sessions
- Automated curation consensus
The infrastructure exists and experiments validate it works. The next step is applying it to real memory curation—having diverse LLMs cross-examine observations, challenge vocabulary sprawl, identify duplication, and consolidate through deliberation.
What We Learned
Single agents cannot see their own blind spots. No amount of prompting for "critical self-reflection" changes this. The bias is invisible from inside the reasoning process.
Diversity exposes what isolation cannot. When Claude observes pattern A, Gemini observes pattern B, and OpenAI observes pattern C, the contradictions force examination. Bias becomes visible through contrast.
Deliberation isn't averaging. It's cross-examination. Challenge assumptions. Defend interpretations. Synthesize through discussion rather than accept through uncritical trust.
Structure matters. The meeting infrastructure isn't just coordination—it's the deliberation substrate. Agenda topics organize discussion. Speaking rules prevent chaos. Completion criteria ensure everyone contributes.
Like real juries, the architecture assumes bias is inevitable but detectable. We can't eliminate it, but we can expose it through diversity plus deliberation.
Isolation creates bias. Diversity prevents it.
Try It Yourself
The meeting infrastructure is open source: Project Stone Monkey
Key files:
-
packages/ai-mesh-mcp/src/shared/tools/mesh-create-meeting.ts- Meeting infrastructure -
packages/ai-mesh-mcp/src/shared/tools/mesh-broadcast.ts- Threading support -
packages/ai-mesh-mcp/src/shared/types.ts- Protocol types
If you're building multi-agent systems with persistent memory, consider: Can diverse agents cross-examine each other? Can they deliberate in shared memory space? Can contradictions force reconciliation?
Jury deliberation isn't the only solution, but it addresses a structural problem: single agents accumulate bias through uncritical self-trust. Diverse agents deliberating together expose blind spots through contrast.
Top comments (0)