Abstract (The "Elevator Pitch"):
Most AI interfaces treat LLMs like chat-boxes. We believe they are a Society of Minds. In 48 hours, we are building OourMind.io, a multi-sensory interpretability lab that visualizes how Mistral Large 3 selects and shifts between latent personas (Social Agents) to answer a prompt.
A standard "wrapper app" won't win this hackathon. We need to visualize the geometry of thought.
π οΈ The Architecture: The "Body" and the "Brain"
To execute this on a budget and a strict deadline, we are ruthlessly separating the Static Visual Theater (Body) from the Live Metadata Inference (Brain).
Phase 1: The Visual Stage (frontend/oourmind.io)
The Core: React + Three.js/Spline. We aren't building a chat interface. We're building a geometry viewer.
The State Engine: A simple JavaScript function that maps Mistral's metadata (e.g., Tone: 0.8, Structure: Grid) to Spline "States" and ElevenLabs audio files.
Phase 2: The "Social Agent" Interrogation (Backend/Jupyter)
Instead of a live, fragile API connection, we are running 3 High-Impact Case Studies in a stable backend:
The Moral Dilemma: (Tests Ethicist vs. Utilitarian bias).
The Creative Abstract: (Tests Fluidity).
*The Logical Paradox: *(Tests Structure).
We force Mistral 3 to output a JSON object containing its metadata:
** π
The 48-Hour Execution Sprint
**
Day 1: The Extraction (The Science)
H0-H6: Finalize the Spline geometries and ElevenLabs voice profiles for the 3 target personas (Analytical, Creative, Technical).
H7-H12: Backend Interrogation. Running the "Case Study" prompts on Mistral Large 3 to extract the raw activation metadata and response content.
Day 2: The Exhibition (The Demo)
H13-H18: Hardcoding the extracted JSON from Day 1 into the oourmind.io frontend. If the user clicks "Moral Dilemma," the site "plays back" the pre-recorded 3D and audio state.
H19-H22: Record the Vision Video. This is 90% of the judging score. The video shows the vision, not just the code.
H23-H24: Polish documentation and submit.
π Future Evolution (Post-Hackathon)
The MVP is just a snapshot. Here is the Production Pipeline we will build next:
1. Neuron-Activation Heatmaps
Move beyond simple geometry to a live visualization of actual neurons firing within Mistralβs latent layers as it generates text. This is true interpretability.
2. The "Persona Switchboard"
An interactive dashboard where a human auditor can manually force Mistral to switch personas mid-sentence (e.g., from "Aggressive Lawyer" to "Helpful Mediator").
3. Verification of Trust (Governance Dashboard)
We will integrate a "Verifiable Persona Signature" (e.g., using a protocol like hummiin.io). This provides a decentralized, auditable receipt that the model was in a safe "persona state" during critical inference. This is the governance layer for corporate AI deployment.
Closing Statement:
We aren't building another tool to generate content. We're building a tool to understand the character of the content generator.
Top comments (0)