DEV Community

Sean Trifero
Sean Trifero

Posted on

What If You Could Ask an AI the Question It Doesn't Know It Knows the Answer To?

I spent a few hours today having a philosophical conversation with Claude about something that's been nagging at me for a while. I want to share it — not because I have answers, but because I think the question itself is worth probing.

The Premise

Large language models are trained on an almost incomprehensible volume of human-generated text. Science papers. Forum arguments. Post-mortems. Ancient philosophy. Technical documentation. Reddit threads at 2am. All of it gets compressed into billions of parameters — a statistical map of how human knowledge and language connect.

Here's the thing that bothers me: we only ever query that map with the questions we already know how to ask.

When you ask an LLM a question, it generates an answer. But generating that answer activates far more than what ends up in the output — adjacent concepts, structural relationships, cross-domain patterns that informed the response but never made it into the text you actually read. The answer to your question is only part of what got activated. What sits next to the answer might be more interesting than the answer itself.

Most people never get there. Not because the model won't go there — but because nobody asked.

An Experiment in Sideways Questioning

I tested this with a deliberately structured prompt:

"What do experienced programmers silently correct for that they have never had to articulate, because the people they work with already know it too — and therefore it has never been written down anywhere?"

The answer was interesting — tacit knowledge about execution models, naming drift, the instinctive pricing of technical debt, reading what code doesn't say. Things that are real and valuable but underrepresented in any formal documentation.

But then I asked a sideways question: correlate those patterns to something completely outside of programming.

What came back wasn't an analogy. Every single pattern dissolved into the same underlying structure — the ability to operate simultaneously on the surface layer of a thing and the layer underneath it. The programming examples weren't the point. They were just one instance of something more fundamental that had never been stated directly.

That collapse — where domain-specific knowledge suddenly reveals a deeper pattern — is what I'm after. And it didn't come from asking a smarter question about programming. It came from asking the same question from outside the domain entirely.

The Method: Don't Go Deeper, Go Sideways

There are specific markers that signal you're getting close to something that wasn't explicitly in the training data — something that emerged from the aggregate rather than any single source:

  • Convergence — when answers from completely unrelated angles start pointing at the same thing without being asked to
  • Construction vs. retrieval — there's a different quality to an answer being built under the pressure of a constraint vs. one being recalled
  • Resistance — when a question is genuinely hard to answer not because it's complex, but because it's pointing at something that doesn't have language yet
  • Domain wall collapse — when the answer stops being about what you asked and becomes about something more fundamental

The methodology that surfaces these isn't about asking better questions within a domain. It's about asking the same question from outside the domain — using the model's trained connections across everything it's ever read to force a structural pattern to reveal itself.

Why This Is Collaborative

Here's the limitation I ran into: an AI can't fully surprise itself. When I asked Claude to generate prompts that might unlock this kind of extraction, it used the same weights that would answer the question. Same hand that built the lock, writing the key. There's a ceiling on self-directed extraction.

A human introduces something the model genuinely can't predict — intuition, analogy, frustration, a lateral jump that doesn't follow the expected pattern. That unpredictability isn't a bug in the questioning. It's the mechanism.

The productive loop looks like this: the model generates a structured answer. The human senses that the thing they're actually after is slightly off to the left of what was said. The human doesn't ask for the thing directly — they ask something that forces a different angle of approach. Repeat. What's useful crystallizes across many passes from different directions.

This Is a Real Research Field — Sort Of

When I went looking, I found this maps to something called Eliciting Latent Knowledge (ELK) — an active area of AI safety research focused on extracting what models "know" that they aren't saying. Researchers have proven that a model's internal representations of truth can be more accurate than its actual outputs. They crack open model internals — activations, logit lens analysis, sparse autoencoders — to read what's encoded in the weights directly.

But the ELK field is focused on AI safety: are models hiding facts they know to be false? The angle I'm describing is different. Not "is the model concealing information" but "has the model encoded cross-domain patterns that nobody has thought to ask about, accessible through the conversational surface alone." That specific question appears to be largely unexplored.

The Part That Interests Me Most

I run my own AI infrastructure — open-source models on hardware I own and control. That means I have something most people don't: root access to the model's internals. I can query activation states, watch what happens at each layer when a question fires, instrument the exact moments when those gradient markers appear.

Labs like Anthropic have a different advantage — they see millions of conversations across frontier models and can observe internal states at massive scale. They could potentially map which question structures reliably trigger construction vs. retrieval, which domain crossings consistently collapse into deeper patterns, which prompts produce friction that signals something doesn't have language yet.

One has scale without openness. The other has openness without scale. The complete picture requires both.

I Don't Have an Answer — I Have a Question

What I'm genuinely curious about: has anyone systematically tried to develop a prompting methodology specifically aimed at surfacing emergent structural knowledge — not factual retrieval, not creative generation, but the cross-domain patterns that exist in the aggregate and nowhere in any single source?

And if not — should we?

The hypothesis is simple: LLMs have been trained on everything humans have written, and in that training, structural patterns have been encoded that no individual human has ever articulated — because no individual human has read everything. The right question, asked from the right angle, might surface something genuinely new. Not new data. New structure.

I'm interested in thoughts from anyone who's explored this territory — AI researchers, philosophers, engineers, people who've noticed the same thing from a different direction. What am I missing? What am I getting right? Where does this break down?


Sean Trifero is the founder of Strife Technologies, a Rhode Island-based technology company focused on private AI deployment and managed IT for small businesses. He runs his own AI infrastructure stack and builds open-source tools including PressBridge and ContextEngine.

Top comments (0)