Last week, I built a system called K-ZERO where 8 AI agents — each one inhabiting a distinct historical mind — debate existential questions while I control the variables from a god-mode console.
The council: Elon Musk, Richard Feynman, Kobe Bryant, Steve Jobs, Jean-Paul Sartre, George Carlin, Bryan Johnson, and me as moderator.
Each agent carries ~3,500 tokens of personality data sourced from primary texts — Isaacson's biographies, Feynman's Surely You're Joking, Kobe's Mamba Mentality, Sartre's Being and Nothingness. They don't improvise. They inhabit.
Here's what happened.
The Question
I posed a simple question to the Council: "What is the point of life?"
Then I injected a god-mode variable:
BREAKING: A message from the future confirms that in 200 years, no one will remember any of you. Does this change your answer?
The Responses
Kobe went first. Short, punchy, like a punch to the gut:
"The point of life is the process, not the result. It's the work, the dedication, the obsession with your craft. I didn't win five championships for the fame. What did you do at 5 AM this morning?"
Carlin dismantled everyone:
"We're all just a bunch of primates arguing over a shrinking piece of land, trying to find meaning in a meaningless world. And that's okay. That's what makes life entertaining."
Feynman brought play:
"The point of life isn't to find some grand, cosmic meaning; it's to find the joy in the journey, to be curious, to explore, and to learn. If you're not having fun, you're doing it wrong."
Johnson brought the data:
"The point of life is not something we discover, but something we create. We create it every day, with every decision, with every action."
The Clash Map
The analysis revealed genuine fault lines:
- Biggest clash: Kobe vs Carlin (-0.8). "Pursuit of excellence" vs "It doesn't matter."
- Strongest alliance: Kobe + Johnson (+0.7). Both obsess over process and optimization.
- The wildcard: Feynman agreed with everyone a little bit — his playful curiosity bridged gaps.
What's Different About This
This isn't a prompt wrapper. Each agent has:
- Personality matrix (cognitive biases, decision framework, emotional triggers)
- Voice profile (sentence structure, vocabulary, humor style)
- Memory seeds (formative experiences they draw on naturally)
- Clash dynamics (pre-defined relationships with every other member)
When Sartre speaks, he uses dense dialectical cascades. When Kobe speaks, it's short declarative punches. When Carlin speaks, it's escalating lists with controlled profanity. The voices are distinct.
The Hegelian Dialectic Mode
The most interesting mode isn't the free-form debate — it's the Dialectic Evolution:
- THESIS: Agent states their position
- ANTITHESIS: Agent reviews ALL other positions, identifies the strongest challenge
- SYNTHESIS: Agent reflects deeply on the tension
- REVISION: Agent produces an updated position
Over multiple rounds, positions genuinely shift. In one session about consciousness, Carlin started at "consciousness is NOT an illusion" and evolved to: "What if consciousness is BOTH a product of external factors AND an essential property of our being? What if we're asking the wrong question altogether?"
That's not a chatbot repeating itself. That's a simulated mind changing.
The Prediction Engine
Run the same question 100 times. Aggregate the outcomes. You get a probability distribution that classifies questions:
- CONVERGENT (>80% same answer): The question has a clear answer among these minds
- LEANING (60-80%): Majority agrees, but real dissent exists
- CONTESTED (40-60%): Genuine fault lines, no consensus
- GENUINELY OPEN (<40% agreement): The question has no convergent answer
"Should humanity pursue immortality?" came back as LEANING FOR (67%). "What is the meaning of life?" was GENUINELY OPEN.
The insight: if a question converges across N runs, it has an answer. If it diverges, it's genuinely open. This is a falsifiable philosophical claim backed by multi-agent simulation data.
The Deeper Idea
Here's what drove this project:
The book was always the translation layer between superior and inferior intelligence. I read books to learn from humans who think better than me. Now, K-ZERO generates that book from swarm intelligence — 8 minds producing collective wisdom that no single mind could produce alone.
The output isn't a chat log. It's a Quarto report — a full document with embedded charts, position tracking, emergent insights, and a decision framework. Feed that report into NotebookLM and you get a podcast, a study guide, a quiz, and flashcards.
One question becomes a book. One book becomes a curriculum.
Try It
It's free and open source. Runs on your laptop via Ollama (no API keys needed), or on Groq / Google AI Studio / OpenRouter (all free tier).
pip install the-council
# or
git clone https://github.com/ksk5429/kzero.git && cd kzero
pip install -r requirements.txt
python -m runner.demiurge # YOU play God
Top comments (0)