DEV Community

Cover image for Your AI Agent Has Amnesia. And You Designed It That Way.
Timur Fatykhov
Timur Fatykhov

Posted on

Your AI Agent Has Amnesia. And You Designed It That Way.

#ai

__Every LLM call starts from nothing. No memory of what worked yesterday. No record of what failed last week. The industry calls this “stateless.” It’s not an architecture pattern — it’s a limitation we’ve been too slow to fix.
I spent the last month reading nine papers from 2025–2026 on the cutting edge of agent memory research. Not theoretical memory. Real systems with benchmarks, architectures, and trade-offs.
Here’s what changed how I build — and what it should change about how you build too.

1. Context Replay Is Not Memory

The most widespread approach to “giving agents memory” is context replay: retrieve relevant text, inject into the prompt, hope the model does something useful. RAG at its most basic.
It works for simple recall. It falls apart for everything else.
A-MEM made this concrete. The authors replaced flat memory stores with a Zettelkasten-style knowledge network. When a new memory is encoded, the agent generates structured notes with contextual tags — and critically, retroactive bidirectional links to existing memories. Memory is a graph, not a list.
The difference isn’t subtle. Similarity search finds things that look like the query. Graph traversal finds things related to the query. Those are fundamentally different operations, and for complex multi-session reasoning, only one of them actually works.
SYNAPSE extended this with spreading activation — the same neural mechanism that lets you hear “doctor” and prime “nurse.” Their dual-layer architecture achieves a weighted average F1 of 40.5 on the LoCoMo benchmark, a margin of +7.2 points over the next best agentic system — while reducing token consumption by 95% compared to full-context methods.
The takeaway: If your agent’s memory is a vector store with cosine similarity, you’ve built a search engine — not a memory. Real memory has structure, relationships, and traversal paths.

2. One Memory System Is Never Enough

The 47-author Agent Memory Survey (Dec 2025) gave the field its first unified taxonomy. Three dimensions: memory forms (how it’s stored), functions (what it does), and dynamics (how it changes). Conflating them — which almost everyone does — leads to systems that are brittle at everything except the one task they were tuned for.
Procedural Memory Is Not All You Need made this argument directly. LLMs are fundamentally constrained by their architecture, which mirrors human procedural memory: pattern-driven, automated, but lacking grounded factual knowledge. An agent that knows how to execute a task still can’t reliably reason about what that task involves without semantic memory.
MAP addressed this structurally with a modular planner architecture — separate memory modules with clean interfaces between them, composed like microservices. Need procedural and factual? Activate both. Need only episodic? Use just that.
The takeaway: Stop building one memory system. Build memory systems — plural — with clear interfaces. A fact store is not an episode log is not a skill library.

3. Graphs Beat Flat Vectors For Anything That Matters

For anything beyond single-turn Q&A, graph-structured memory consistently outperforms flat vector retrieval. This appeared across every paper until it was impossible to ignore.
Mem0 evolved in exactly this direction. Their latest architecture integrates graph-augmented memory via FalkorDB, with per-user graph isolation and sub-140ms p99 query latency. The paper demonstrates 26% relative improvement over OpenAI on LLM-as-a-Judge metrics, with graph memory adding another ~2% over the base vector configuration.
The Agent Memory Survey confirmed the pattern with systematic analysis: systems with graph-augmented retrieval consistently outperform pure vector approaches on multi-hop reasoning, temporal reasoning, and contradiction detection. The gap widens as task complexity increases.
One honest counter-benchmark deserves acknowledgement. Letta — the team behind MemGPT — demonstrated that a GPT-4o-mini agent equipped with basic filesystem tools (semantic file search and grep over raw conversational history) achieved 74.0% accuracy on the same LoCoMo benchmark where Mem0’s top-performing graph variant scored 68.5%. Letta themselves draw a cautious conclusion from this: that LoCoMo may be testing retrieval skill more than memory architecture, and that “memory is more about how agents manage context than the exact retrieval mechanism used.” This is worth holding onto. Specialized graph architectures offer real structural advantages — relationship traversal, contradiction detection, temporal reasoning — that simple file search cannot replicate. But the Letta result is a useful reminder that architectural sophistication is not a substitute for capable tool use, and that today’s benchmarks are still catching up to what agent memory actually requires.
The takeaway: Vector search is necessary but insufficient. If your agents handle tasks spanning multiple turns, entity relationships, or temporal reasoning — you need graph structure. Not instead of vectors. On top of them.

4. Evolution Already Solved This. 600 Million Years Ago.

Here’s the lesson I didn’t expect from a stack of AI papers: the best architects in this space aren’t inventing new solutions. They’re reverse-engineering the one that already works.
The Episodic Memory paper laid out five properties long-term agents genuinely need: temporally indexed, instance-specific, single-shot encodable, inspectable, and compositional. Without these, they argue, agents can’t maintain coherent context across sessions — a gap most current architectures don’t address. These properties are grounded in cognitive science going back to Endel Tulving’s 1972 taxonomy of human memory.
SYNAPSE’s spreading activation is borrowed directly from Collins and Loftus’s 1975 model of human semantic memory. ACC’s cognitive compression mirrors the brain’s consolidation process during sleep — taking fragmented short-term memories and compressing them into stable long-term representations.
The Survey acknowledged this convergence as “cognitive neuroscience as design language.” I’d go further: it’s a design proof. Evolution already ran the world’s longest A/B test on memory architectures. Structured, multi-system, consolidation-driven, forgetting-enabled associative memory won. Everything else went extinct.
The takeaway: The hippocampus has already solved the problems you’re encountering. You’re not building from scratch. You’re standing on 600 million years of R&D.

  1. Forgetting Is a Feature, Not a Bug Every instinct in software engineering says store everything, delete nothing, disk is cheap. For agent memory, this instinct is actively harmful. ACC (Agent Cognitive Compressor) demonstrated this most clearly. Its commitment mechanism prevents unverified content from becoming persistent memory — memories pass through a compression-and-validation pipeline before they’re committed to long-term storage. Tested across IT operations, cybersecurity, and healthcare workflows, ACC consistently produced lower hallucination and drift than transcript replay approaches. The industry is moving in the opposite direction. Llama 4 Scout ships with a 10-million token context window — 50x larger than the previous generation — with the implicit promise that more context solves the memory problem. It doesn’t. Chroma Research established empirically that LLM performance degrades with increasing input length, across all 18 frontier models tested — even on trivially easy tasks. Stuffing more memories into context doesn’t help. It hurts. The degradation isn’t linear and it doesn’t wait until the context window is full. Independent analysis of Llama 4’s 10M window confirms the pattern: recall accuracy shows stochastic degradation as context grows past the million-token mark, with the “lost in the middle” phenomenon becoming more severe, not less, at extreme scale. A-MAC (March 2026) formalises this into a framework: memory admission as a structured decision across five dimensions — future utility, factual confidence, semantic novelty, temporal recency, and content type. What you don’t store matters as much as what you do. On the LoCoMo benchmark, A-MAC improved F1 to 0.583 while reducing latency 31% versus state-of-the-art systems. The takeaway: Build forgetting into your memory architecture from day one. Implement confidence decay, staleness signals, and explicit deletion policies. An agent that remembers everything isn’t smarter — it’s confused. The Uncomfortable Conclusion Most agent frameworks are optimized for stateless task execution. They treat memory as an afterthought, a plugin, a “nice to have.” The research says the opposite: memory architecture is the single most important design decision for any agent that persists beyond a single conversation. What research says works: Structured, graph-augmented memory with typed relationships Separate memory systems for different cognitive functions Biologically-inspired consolidation and forgetting Spreading activation for associative recall Explicit admission control over what enters long-term memory

*What most production agents actually have:
A vector database. Maybe RAG. Conversation history stuffed into context until it overflows.
*

The field has published the answer. The industry hasn’t implemented it yet.

What This Means For What I’m Building

I read these papers while building Nous(https://github.com/tfatykhov/nous) — an open-source cognitive architecture for AI agents, grounded in Minsky’s Society of Mind thesis that intelligence emerges from many specialised, coordinated modules rather than a single monolithic system.
The architecture maps directly onto what the research validated:
Structured graph memory — shipped. PostgreSQL + pgvector with polymorphic graph edges across all memory types. Density-gated spreading activation using recursive CTEs, built on the same Collins & Loftus spreading activation principle as SYNAPSE.
Separate memory subsystems — shipped. Brain (decisions, calibration, graph, guardrails), Heart (episodes, facts, procedures, censors, working memory), and Identity (character, values, protocols) operate as distinct modules with defined interfaces.
Calibration with Brier scores — shipped. Every decision records a confidence score. Outcomes are reviewed automatically by a background Decision Reviewer. Agents learn whether their confidence estimates are reliable over time.
B-brain self-monitoring and Symbolic Control — shipped. A Monitor engine watches each turn post-execution: did the action match intent? If not, create a censor. This is Minsky’s B-brain watching the A-brain work, implemented in code. The necessity of this governance layer is empirically validated by the SCL framework (Nov 2025), which bridges classical expert systems and neural reasoning. SCL’s Soft Symbolic Control is an adaptive governance layer that applies symbolic constraints to the probabilistic inference of the LLM — not as rigid rules, but as a metaprompt-based mechanism that guides reasoning while preserving the model’s generalization capabilities. Experiments show SCL achieves zero policy violations and eliminates redundant tool calls. Our programmatic censors operationalise this exact principle.
Sleep consolidation — shipped. A Sleep Handler runs five phases during idle periods: review pending decisions, prune stale censors, compress old episodes into summaries, reflect on cross-session patterns, generalise repeated facts. A direct implementation of the biological consolidation ACC and the Survey both point to.
Calibrated forgetting — shipped. Staleness decay (half-life scoring), relevance floor cutoffs, deduplication, and abandoned decision filtering all enforce that not everything survives into long-term memory. Memory admission control as a formal scored framework is designed (F023) but not yet shipped.

Minsky argued in 1986 that the power of intelligence stems from diversity of components, not any single principle. The papers say the same thing about memory, in 2025, with benchmarks.
Nous is ~21,000 lines of Python, 1,200+ tests, deployed on PostgreSQL with 23 tables. The cognitive loop, graph memory, calibration, and sleep consolidation are live. Frame splitting and the growth engine are designed and in spec — next to build.
The agents that will matter in 2027 aren’t the fastest ones. They’re the ones being built with real memory systems today.

Real memory. That forms, evolves, consolidates, and yes — forgets.

That’s the difference between a tool and a mind.

Which of these five gaps is most visible in the agents you’re building or evaluating? I’d be curious what’s hardest to close in practice.

Papers Referenced

  1. A-MEM — Agentic Memory for LLM Agents (Feb 2025) Xu et al. | arxiv.org/abs/2502.12110
  2. Episodic Memory — Position: Episodic Memory is the Missing Piece for Long-Term LLM Agents (Feb 2025) Pink et al. | arxiv.org/abs/2502.06975
  3. Mem0 — Building Production-Ready AI Agents with Scalable Long-Term Memory (Apr 2025) Chhikara et al. | arxiv.org/abs/2504.19413
  4. Procedural Memory Is Not All You Need — ACM UMAP Adjunct ’25 (May 2025) Wheeler & Jeunen | arxiv.org/abs/2505.03434
  5. MAP — Modular Agentic Planner — Nature Communications, 2025 (exact DOI pending verification)
  6. SCL — Bridging Symbolic Control and Neural Reasoning in LLM Agents (Nov 2025) arxiv.org/abs/2511.17673
  7. Memory in the Age of AI Agents — Survey, 47 authors (Dec 2025) arxiv.org/abs/2512.13564
  8. ACC — AI Agents Need Memory Control Over More Context (Jan 2026) Bousetouane | arxiv.org/abs/2601.11653
  9. SYNAPSE — LLM Agents with Episodic-Semantic Memory via Spreading Activation (Jan 2026) arxiv.org/abs/2601.02744
  10. A-MAC — Adaptive Memory Admission Control for LLM Agents (Mar 2026) arxiv.org/abs/2603.04549
  11. Context Rot — How Increasing Input Tokens Impacts LLM Performance — Chroma Research, 2025 research.trychroma.com/context-rot
  12. Collins & Loftus — A Spreading-Activation Theory of Semantic Processing (1975) Psychological Review, 82(6), 407–428
  13. Tulving, E. — Episodic and Semantic Memory (1972) In Organization of Memory. Academic Press.
  14. Minsky, M. — The Society of Mind (1986) Simon & Schuster. ISBN 0671657135

Top comments (0)