DEV Community

Cover image for I Built an AGI Benchmark — And Tested It Against Top AI Models
felipe muniz
felipe muniz

Posted on

I Built an AGI Benchmark — And Tested It Against Top AI Models

Radial Comparison

Most AI benchmarks today measure accuracy.

But here’s the problem:

Accuracy ≠ Intelligence.

So I built something different.

An experimental evaluation suite designed to measure cognitive behavior — not just outputs.

And the results were… surprising.

🧠 What This Benchmark Measures

Instead of one score, the system evaluates multiple cognitive dimensions:

Reasoning

Planning

Memory

Metacognition

Agency

Self-correction

Epistemic calibration

Contradiction awareness

Grounding fidelity

Task adaptation

Citation integrity

Each model gets a cognitive profile — like a brain scan.

🧪 The Experiment

I tested multiple models including:

ATIC (my architecture)

GPT

Claude

Gemini

Each was evaluated across controlled tasks with:

identical prompts

multiple seeds

automated scoring

judge validation

📊 What Happened

Grounding changed everything.

When grounding was enabled:

epistemic calibration improved

contradiction detection improved

reasoning stability improved

In other words:

grounding didn’t just make answers better
it made thinking better

🏆 Composite Results
Model Score
Claude 0.875
ATIC (grounded) 0.844
GPT 0.812
ATIC (no grounding) 0.781
Gemini 0.708
⚠️ Important Takeaway

We’re benchmarking AI wrong.

Current leaderboards reward:

memorization

pattern matching

dataset familiarity

But real intelligence requires:

self-correction

uncertainty awareness

causal consistency

adaptive reasoning

Benchmarks that ignore these are measuring performance, not cognition.

🔬 Why This Matters

If AGI is the goal, we need metrics that evaluate:

the structure of thought — not just the correctness of answers.

That’s what this project tries to do.

📂 Full Open-Source Repo

Includes:

benchmark engine

scoring system

evaluation datasets

visualization scripts

reproducible results

👉 https://github.com/AletheionAGI/benchmark_agi_suite

🚀 Challenge to the Community

Run your model on it.

Break it.
Improve it.
Fork it.
Benchmark it.

Because AGI won’t be built by one lab.

It’ll be built by people who measure intelligence correctly.

If you're working on reasoning systems, cognitive architectures, or evaluation science — I want to hear your thoughts.

Top comments (0)