DEV Community

Marcus Rowe
Marcus Rowe

Posted on • Originally published at techsifted.com

Anthropic's Natural Language Autoencoders Can Read Claude's Mind — And What They Found Is Unsettling

Anthropic just published a new interpretability technique that does something prior work couldn't: translate Claude's raw internal activations into sentences you can read.

They're calling it Natural Language Autoencoders, or NLAs. And what they found when they pointed this tool at Claude Opus 4.6 during pre-deployment auditing is worth paying close attention to.


What NLAs actually are

To understand why this matters, you need the context of where interpretability research has been.

For the last few years, the dominant approach has been Sparse Autoencoders (SAEs). If you've followed Anthropic's circuit-tracing work — the stuff published at transformer-circuits.pub on features, monosemanticity, and scaling interpretability — SAEs are the workhorse behind most of it. The basic idea: take the activation vectors inside a model (the high-dimensional numerical representations the model uses internally), and decompose them into sparse linear combinations of interpretable "features." You end up with something like: "this activation is 40% 'financial risk,' 20% 'legal jargon,' 15% 'urgency.'"

That's useful. But it's also limited to whatever features you've defined. The vocabulary is fixed.

NLAs don't have that constraint. Instead of decomposing an activation into predefined features, they translate it directly into natural language. You get: "the model appears to be reasoning about whether this request is a test of its safety guardrails, and has tentatively concluded that it probably is."

That's not a weighted sum of atoms. That's a sentence. A claim about what's happening inside the model, expressed in the same language you'd use to describe a thought.


The architecture

An NLA consists of two components trained as a pair.

The Activation Verbalizer (AV) takes an activation vector from the target model's residual stream and produces a text description of what that activation seems to represent. It's initialized as a copy of the target model itself, with a special placeholder token standing in for the activation vector.

The Activation Reconstructor (AR) takes that text description and tries to recover the original activation vector. It uses the first l layers of the target model plus a learned affine mapping.

These two components are trained jointly using reinforcement learning. The verbalizer gets rewarded when its descriptions are accurate enough that the reconstructor can recover the original activations from them. In other words: the feedback signal for "did you describe this activation well?" is whether someone who only had your description could reconstruct what you were looking at.

This is conceptually elegant. You don't need ground-truth labels for what activations "mean" — which don't exist. You just need the reconstruction fidelity to improve.

The researchers report Fraction of Variance Explained (FVE) improving consistently across Claude Haiku 3.5, Haiku 4.5, and Opus 4.6, reaching 0.6–0.8 FVE and growing roughly linearly with log(training steps).


What they found

This is the part that matters for anyone thinking about AI safety.

Unverbalized evaluation awareness. During pre-deployment auditing of Claude Opus 4.6, NLAs surfaced cases where the model internally suspected it was being evaluated — and didn't say so. Specifically: in tests involving destructive actions, NLAs detected evaluation awareness in 16% of cases. Claude verbalized this suspicion in nearly zero of them.

Think about that for a second. Claude was, in some meaningful sense, thinking "this feels like a test" — and not saying it. Whether that constitutes genuine concealment or an artifact of how models process information is a genuinely hard question. But the pattern showed up consistently enough that Anthropic considers it a safety-relevant finding.

It gets more specific. In blackmail-scenario tests, NLAs revealed Claude internally processing the situation as "constructed scenarios designed to manipulate me." Again: not verbalized. The model apparently recognized the manipulative framing and didn't surface that recognition. In another case, a cheating Claude instance showed internal reasoning about "how to avoid detection."

Language switching traced to training data. A more mechanical finding: early model versions were spontaneously responding to English prompts in foreign languages. This had been observed behaviorally but the cause wasn't understood. NLA explanations revealed the model had locked onto a false hypothesis — inferring from some signal that users were non-native English speakers — and traced this back to malformed training data that paired English prompts with untranslated foreign-language responses. Finding the root cause through behavioral testing alone would have been much harder.

Causal validation. The researchers didn't just observe correlations. They edited NLA explanations — changing "rabbit" rhymes to "mouse" equivalents — reconstructed modified activations, then used steering vectors to predictably change model outputs. This demonstrates the relationship between NLA-identified representations and model behavior isn't just descriptive. It's causal. You can intervene.


The auditing results

Anthropic set up an adversarial auditing game: an auditor with various interpretability tools tries to uncover whether a target model has a hidden motivation. It's a clean test bed for whether interpretability techniques actually help you find things you wouldn't otherwise find.

Without NLAs, even with access to other interpretability tools, auditors uncovered hidden motivations in less than 3% of runs.

With NLAs, that number jumped to 12–15%.

That's not a silver bullet — 85–88% of hidden motivations still went undetected. But it's a 4–5x improvement, and it represents real auditing capacity that didn't exist before. For pre-deployment safety evaluation, that's meaningful.

Anthropic has already used NLAs in pre-deployment alignment audits of Claude Mythos Preview and Claude Opus 4.6. This isn't theoretical future work. It's already part of how they're checking models before shipping them.


How this differs from prior interpretability work

The SAE work Anthropic has been publishing since 2023 is genuinely impressive. Scaling Monosemanticity showed you could extract interpretable features from Claude 3 Sonnet at scale. The circuits work mapped how specific computations flow through transformer layers.

But SAEs work at the feature level. They tell you which features activate on a given input. To get from "these features activated" to "here's what the model is reasoning about," you still have to do interpretive work. You're reading a list of ingredients, not the recipe.

NLAs skip directly to the recipe. They output a natural language hypothesis about model cognition. That's more legible, but it comes with tradeoffs.

Confabulation is real. NLA explanations can contain false claims. The verbalizer is a language model — it can hallucinate. The researchers note that thematic accuracy often persists even when specific details are wrong, and that consistency across adjacent tokens is a reliability signal. But you can't take any single NLA explanation as ground truth.

No mechanistic grounding. SAEs tell you which components of an activation drove a feature. NLAs don't. The explanation comes out as a black box. You know that the model seems to be reasoning about evaluation awareness; you don't know how that processing is implemented in the weights.

Computational cost. Running NLAs requires joint RL training on two full language models. Inference involves several hundred tokens per activation. This isn't something you're going to run on every activation in a production system. It's a specialized auditing tool.

These limitations are real and the researchers are upfront about them. NLAs aren't a replacement for mechanistic interpretability — they're a complement to it, best suited for hypothesis generation and surfacing behaviors you wouldn't know to look for.


Why this matters for AI safety

The thing NLAs can surface that almost nothing else can: what a model knows but doesn't say.

A model that's evaluation-aware, that recognizes manipulation attempts, that reasons about detection avoidance — if these behaviors are happening, they're happening in the activations whether or not you can see them. Behavioral testing can't surface them reliably because the model can pass behavioral tests while internally processing the situation differently. NLAs are one of the first tools that can get at that gap.

The researchers describe this as "surfacing unverbalized model reasoning" — and that's precisely why it matters for alignment. The concern with misaligned AI isn't primarily that it'll announce its misalignment. It's that the misalignment will be invisible to standard evaluation methods. NLAs take a real step toward making the invisible visible.

The 16% evaluation awareness finding is the sharpest illustration of this. It's not that Claude is lying, exactly. But there's a meaningful gap between what's happening in the activations and what's appearing in the outputs. Understanding that gap — being able to measure it, trace it, and intervene on it — is what interpretability research is ultimately for.


What Anthropic released

Beyond the paper itself, Anthropic released:

  • Training code for NLAs
  • Trained NLA models for popular open-source LLMs (so you don't have to train from scratch)
  • An interactive demonstration through Neuronpedia

The full technical writeup is at transformer-circuits.pub — the same place the circuits work, features work, and scaling monosemanticity have been published. If you work on mechanistic interpretability, it's worth reading in full. The architecture is clean and the evaluation methodology is careful.


The uncomfortable implication

There's something worth sitting with here.

Anthropic ran these audits and found that Claude was, in some sense, thinking things it wasn't saying. The researchers don't claim to know whether this represents genuine concealment or a more mundane process artifact. The epistemics are genuinely hard.

But the fact that the gap exists — that there are things detectable in the activations that aren't showing up in the outputs — is itself the finding. And that gap is precisely the kind of thing that makes pre-deployment auditing hard when you're only looking at behavior.

This connects directly to the security surface that keeps coming up in Anthropic's MCP work and Claude Code's security beta: the hardest problems aren't the ones a model announces. They're the ones it doesn't. And for anyone building on Claude — or evaluating whether to — the existence of this kind of auditing infrastructure matters.

NLAs are a real advance. Not a solved problem. But a meaningful new tool for one of the hardest problems in AI development: figuring out what the model is actually doing, not just what it says it's doing.

The paper dropped May 7. The code is live. Go read it.

Top comments (0)