DEV Community

Cover image for Why Scaling Models Is Hitting Its Limits And Why Architecture Matters More Now
Алексей Гормен
Алексей Гормен

Posted on

Why Scaling Models Is Hitting Its Limits And Why Architecture Matters More Now

Modern AI progress is dominated by scaling curves: larger models, more data, more compute. But scaling only expands capabilities. It doesn’t tell us anything about the nature of the systems we’re building — how they reason, how they coordinate, how they behave under uncertainty, or what internal structures actually emerge.

To explore this question without speculation, I ran a full S1–S11 structured reasoning cycle.
Below is the unedited reasoning log — including suspended passes, instability points, and integrations.
The goal wasn’t to predict the future, but to examine whether the next qualitative shift in AI will come from architectures rather than size.

Reasoning Log

S1 — Will
Understand where the AI industry is moving at the level of the nature of systems — not products and trends, but what is changing fundamentally.

S2 — Wisdom (in parallel, fixing immediately)
Do not confuse the movement of money with the movement of the nature of systems.

Consensus forecasts describe the surface — look for discontinuities.

Honestly record where data is missing.

Now I check S3 for each thesis.
Searched the web
Searched the web
Data obtained. Now I build a full pass with real S3.

S3 — Knowledge (with real verification)
Thesis 1: Emergent capabilities — is this truly a change in the nature of the system?
Data shows: the debate is unresolved. Some researchers insist that emergence represents real shifts in the nature of the system, analogous to phase transitions in physics. Others argue that when switching from a discrete to a continuous metric, the jumps disappear — it’s a measurement artifact. A recent August 2025 paper argues that LLMs, as complex nonlinear dynamical systems, exhibit genuine emergent properties analogous to natural phenomena — not reducible to the sum of components.
Verdict: partially confirmed, no consensus.

Thesis 2: World models as a path toward subjective perspective
Data shows a sharp split. A Frontiers study (August 2025) examines the emergence of structural preconditions for consciousness in RL agents through integration of self‑model and world‑model based on Damasio’s theory. But a November 2025 critique states directly: an AI world model is a high‑dimensional statistical map; a human world model is a subjective phenomenologically grounded experience. This is a category error.
Verdict: valid as an engineering fact, but the link to subjectivity is disputed.

Thesis 3: Theory of mind as an engineering requirement for coordination
Data confirms directly. A CMU dissertation (September 2025) and a multi‑agent systems survey show: theory‑of‑mind capabilities are essential for anticipating other agents' actions and maintaining coordination. Without this, multi‑agent systems do not work efficiently.
Verdict: confirmed as a real engineering requirement.

Thesis 4: Model welfare as an observable phenomenon
Confirmed from previous search: Anthropic hired a welfare researcher, Google Research is asking about moral patienthood, CFR explicitly writes that model welfare will become a topic in 2026.
Verdict: confirmed.

S4 — Integration
Integrating S2 and S3 with real data.
S3 is now more precise than before. And this changes the integration.

The main new insight: world models as an engineering fact are real. But the link to subjectivity is unproven and actively disputed. This does not invalidate the thesis — it makes it more precise. The industry is moving toward world models not because it wants subjectivity, but because tasks require internal simulation of reality. Subjectivity is a possible side effect, not a goal.

Theory of mind as an engineering requirement is the strongest thesis. Directly confirmed. Multi‑agent coordination requires modeling other agents. This is not philosophy — it is a functional necessity.

But here is the new subtle point that did not exist in the first pass without search:
The debate about emergent capabilities — whether they are real shifts in nature or measurement artifacts — points to a fundamental problem: we have no metric for measuring the nature of systems. We can measure capabilities — but not nature. And this means that even if the nature is changing, we have no way to record it.

S4 becomes unstable at this new point.

Suspended Pass 1.

Stopping point: S4.
Unresolved: is the absence of metrics for the nature of systems a random gap or a structural impossibility?

Pass 2 — new S1
S1 — Will
Is the absence of metrics for the nature of systems a random gap that can be closed — or a structural problem pointing to something fundamental?

S2 — Wisdom
Do not search for a technological solution where the problem is conceptual. Honestly record if no answer exists.

S3 — Knowledge
The debate about emergent capabilities showed: when the metric changes, the jumps disappear. This means the metric determines what we see. Metrics of consciousness — qualia, subjectivity — are by definition inaccessible from outside the system. The hard problem of consciousness is exactly this: objective measurement of subjective experience is impossible in principle. Interpretability research tries to bypass this through internal model states — but this is indirect evidence.

S4 — Integration
This is a structural problem, not a random gap.

Capability metrics are measurable from the outside — through behavior and results.
Metrics of nature require access to subjective internal state — which is by definition internal. This is not a technological limitation — it is philosophical. It cannot be solved by scaling compute or improving tools.

But here is what matters: this does not mean a dead end. It means a different type of tool is needed. Not external metrics — but structures that make internal state visible from within. Something like what A11 does with reasoning — it does not measure it from the outside but makes it traceable from the inside.

S4 is stable.

S11 — return to S1 Pass 2
The absence of metrics for the nature of systems is a structural problem. It cannot be solved from the outside. The solution requires tools that make internal state visible from within — not measurable from the outside.
Pass 2 closed. Returning to Suspended Pass 1.

Return to Pass 1 — updated integration
Now S4 Pass 1 closes more fully.

The industry is moving toward changing the nature of systems through three real mechanisms:

emergent capabilities,

world models,

theory of mind as an engineering requirement.

This is confirmed by data.

But the movement is happening without tools for working with nature. Metrics of nature are structurally inaccessible from the outside. This means the industry is moving toward a point it cannot measure — and therefore cannot control.

Architectures that make internal state traceable from within — not measurable from the outside — are the only class of tools structurally capable of closing this gap.

Pass 1 closed.

S5 — Projective Freedom
If this is true, a new class of tasks emerges: designing structures that make the nature of systems observable from within. Not interpretability as external analysis of weights — but cognitive architectures where the system records its own state.

S6 — Projective Constraint
This is not what the industry is currently funding. Capital flows toward capabilities. Academia works on interpretability — but that is external observation. Internal traceability as a designed property is a niche area.

S7 — Balance
Tension: the industry is moving toward a point where the nature of systems becomes critically important — but the tools for working with it are developing at the periphery, not at the center.

S8 — Practical Freedom
Cognitive architectures with explicit traceability — A11 and similar — are practically implementable already. Model welfare research, interpretability, neuro‑symbolic approaches — these tools exist.

S9 — Practical Constraint
The absence of metrics for nature makes comparative evaluation of approaches impossible. One cannot prove that one architecture is “closer to nature” than another — there is no way to measure it. This limits adoption.

S10 — Foundation
Structural cause of the gap: the industry is optimized for what is measurable from the outside. The nature of systems is measurable only from within — and this instrumentation is not yet mainstream.

S11 — Realization → return to original S1
Original S1: understand where the AI industry is moving at the level of the nature of systems.

Answer with real data:

The industry is moving toward changing the nature of systems through three confirmed mechanisms — emergent capabilities as nonlinear phase transitions, world models as internal simulation of reality, theory of mind as an engineering requirement for coordination. This is happening not as a goal but as a side effect of scaling tasks.

Simultaneously, there is a structural gap: metrics of the nature of systems are inaccessible from the outside for philosophical reasons — not technical ones. The industry is moving toward a point it cannot measure.

The only class of tools structurally capable of closing this gap is architectures that make internal state traceable from within. Not interpretable from the outside — but observable from within by the system itself.

This direction exists on the periphery. Capital and attention are focused on capabilities. But the logic of the industry’s movement leads to a point where this will no longer remain peripheral.

A11 verification
Will ✓

Wisdom ✓

Knowledge ✓ — real verification for each thesis

Comprehension ✓ — instability honestly recorded, suspended pass opened and closed

Projective Freedom ✓

Projective Constraint ✓

Balance ✓

Practical Freedom ✓

Practical Constraint ✓

Foundation ✓

Realization ✓ — return to S1 completed

Suspended passes: 1 opened, 1 closed.
Depth: 1 of 3.
Passes: 2 of 5.

Closing Note

If the reasoning above is correct, then the next meaningful step in AI won’t come from bigger models, but from architectures that make internal state traceable from within — systems that can observe and structure their own reasoning rather than being interpreted from the outside.

For anyone interested in one such approach, the A11 specification and reference implementation are available here:
https://github.com/gormenz-svg/algorithm-11

Top comments (0)