<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: oleg kholin</title>
    <description>The latest articles on DEV Community by oleg kholin (@oleg_kholin_551a551b).</description>
    <link>https://dev.to/oleg_kholin_551a551b</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3807609%2F2fb9874d-192f-4bb5-8740-44ec7ecff512.png</url>
      <title>DEV Community: oleg kholin</title>
      <link>https://dev.to/oleg_kholin_551a551b</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/oleg_kholin_551a551b"/>
    <language>en</language>
    <item>
      <title>The Impact of AI Agent Development on Smartphone Screen Size: An Analysis of Trends, Paradoxes, and Architectural Shifts</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Mon, 04 May 2026 13:44:50 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/the-impact-of-ai-agent-development-on-smartphone-screen-size-an-analysis-of-trends-paradoxes-and-374h</link>
      <guid>https://dev.to/oleg_kholin_551a551b/the-impact-of-ai-agent-development-on-smartphone-screen-size-an-analysis-of-trends-paradoxes-and-374h</guid>
      <description>&lt;p&gt;Introduction&lt;br&gt;
The rapid growth and development of AI agents often leads to what seems, at first glance, an obvious thought — smartphone screen size must inevitably shrink. Indeed, if an intelligent agent can perform tasks by voice, work in the background, and deliver brief summaries instead of long feeds — why do we need a six-inch display? The logic appears flawless. However, a deeper analysis reveals that behind this assumption lies a series of paradoxes, and the real trends point in an entirely different direction.&lt;br&gt;
Moreover, the question of screen size turns out to be merely the tip of the iceberg. Behind it stand far larger processes: a shift in the architecture of human interaction with computing, the emergence of a new class of market players, and — perhaps most unexpectedly — a revolution that may begin not with complex tasks, but with the simplest "remind me not to miss the turn."&lt;br&gt;
In this work, we will examine the arguments on both sides, conduct their critical analysis, identify fundamental trends, and show that the question of screen size, for the first time in the history of smartphones, may be decided not by the manufacturer, but by the consumer — and that the answer lies deeper than it appears.&lt;br&gt;
Arguments in Favor of Screen Reduction&lt;br&gt;
Proponents of screen reduction put forward a number of arguments that appear compelling at first glance. Among them:&lt;br&gt;
• Voice interaction is becoming the primary channel, making tactile input on a large screen redundant.&lt;br&gt;
• AI agents are capable of generating brief summaries of texts, emails, and notifications, eliminating the need for a large display area.&lt;br&gt;
• Agents perform tasks autonomously in the background, reducing screen usage time.&lt;br&gt;
• The development of AR glasses transfers visual information from the smartphone screen to a wearable device.&lt;br&gt;
• The growing ecosystem of wearable devices (smartwatches, AI pins, AI-enabled earbuds) distributes functions across multiple devices.&lt;br&gt;
• The development of neural interfaces may, in the long term, eliminate the need for a visual channel altogether.&lt;br&gt;
At a surface level, these arguments form a coherent picture: AI takes over tasks — the screen becomes less essential — the device shrinks. However, critical analysis of each of these arguments reveals significant weaknesses.&lt;br&gt;
Critical Analysis: Why the Arguments "For" Don't Hold Up&lt;br&gt;
Voice assistants have existed since 2011 (Apple Siri), yet over thirteen years of their presence on the market, the average smartphone screen size has only grown. Voice remains a niche scenario — timers, music, simple queries — while for complex tasks such as comparing products, navigating a document, or reading, the visual interface remains indispensable.&lt;br&gt;
Wearable devices do indeed take over certain functions from the smartphone: earbuds have assumed audio, watches — notifications and fitness tracking. But in practice, users are not prepared to carry five devices instead of one, and the smartphone remains a universal tool — a "Swiss army knife" of digital life.&lt;br&gt;
AR/VR technologies are perhaps the most promising direction, yet mass adoption of lightweight and affordable AR glasses is a matter of at least five to ten years. Current solutions (Apple Vision Pro at 600g and $3,500) are far from the mass market. Added to this is the social stigma, well known from the Google Glass experience.&lt;br&gt;
Neural interfaces, with all due respect to the Neuralink project, remain in the realm of science fiction for the mass consumer — the horizon of their practical application is measured in decades.&lt;br&gt;
Finally, the trend toward UX simplification has historically not led to smaller screens. On the contrary — more whitespace in the interface, larger typography, greater visual comfort. The iPhone grew in size precisely when Apple was simplifying iOS.&lt;br&gt;
Arguments Against Screen Reduction&lt;br&gt;
The arguments on the opposing side rest on fundamental rather than circumstantial factors:&lt;br&gt;
Growth of video consumption. TikTok, YouTube Shorts, Reels, streaming platforms — video content is growing exponentially. The average user spends more than three hours per day watching video on a smartphone. AI amplifies this trend through personalized recommendations and AI-generated video content. No one will watch video on a screen smaller than the current one.&lt;br&gt;
Visual verification. The more tasks an AI agent performs autonomously, the greater the user's need to verify the result before confirmation. A booked hotel, a sent payment, a composed letter to a supervisor — all of this requires visual review. An agent's error can cost real money.&lt;br&gt;
Privacy. In the office, on public transport, in a café, voice interaction is impossible or socially unacceptable. This is not a technological limitation that can be overcome by an engineering solution — it is a fundamental property of human coexistence. As long as we live among other people, the screen remains a private channel of interaction.&lt;br&gt;
Visual communication. Memes, stickers, video messages, stories — modern communication is approximately 70% visual. AI amplifies this trend by generating stickers, filters, and AI avatars. People communicate through images, and for that, a screen is needed.&lt;br&gt;
Identified Trends&lt;br&gt;
A deep analysis of both groups of arguments allows us to identify six fundamental trends, which we propose to divide into two categories.&lt;br&gt;
Trends From the Device&lt;br&gt;
The arguments "for" screen reduction, upon closer examination, point not to a smaller display, but to three technological trends:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Distributed computing. The smartphone is ceasing to be the sole device — computation is being distributed among glasses, watches, earbuds, and other devices. The screen is not shrinking — the smartphone is losing its monopoly.&lt;/li&gt;
&lt;li&gt; Multimodal interaction. The number of ways to interact with a device is increasing — voice, gestures, gaze, touch. The user chooses a channel depending on context: at home — voice, in the subway — screen, while driving — voice. The screen does not disappear; it becomes one of many channels.&lt;/li&gt;
&lt;li&gt; Transition from executor to supervisor. The user is doing less themselves and increasingly reviewing the results of AI's work. This does not reduce the need for a screen — it changes the nature of its use.
Trends From the Human
The arguments "against" reveal three trends rooted not in technology, but in human nature:&lt;/li&gt;
&lt;li&gt; Explosive growth of generated content. AI endlessly creates visual content — images, video, charts, tables. The volume grows exponentially, while consumption remains a visual process that requires a screen.&lt;/li&gt;
&lt;li&gt; Deficit of trust in AI. The more autonomy agents receive, the greater the need for transparency and verification of their actions. Verification is a visual task, and it requires screen space.&lt;/li&gt;
&lt;li&gt; Privacy as a permanent barrier. Social norms and the need for confidentiality limit the spread of alternative interfaces — voice-based, AR glasses with cameras, neural interfaces. This barrier is not technological and cannot be overcome through engineering.
Interconnection of Trends
The most significant finding of this analysis is the discovery of systemic interconnections between the two groups of trends. Trends from the device and trends from the human do not contradict each other — they complement and mutually reinforce one another.
The tendency toward the user's transition into the role of supervisor is amplified by the deficit of trust in AI. The more autonomous the agent, the greater the need for a screen to visually verify its actions. Multimodality of interaction collides with the barrier of privacy: new channels — voice, AR — are constrained by social norms, and the screen remains the primary private interface. The distribution of computing across devices cannot keep pace with the explosive growth of content: the volume of AI-generated visual material grows faster than the device ecosystem can distribute it.
Who Determines Screen Size: The Manufacturer or the Consumer?
At this stage, however, it is necessary to ask a question that calls into doubt all the preceding logic: does consumer behavior actually determine screen size?
Historical practice suggests otherwise. Before 2007, any consumer survey would have shown absolute loyalty to physical keyboards: BlackBerry was iconic precisely for its tactile feedback, the Nokia E-series sold in the millions. The consumer did not ask for a virtual keyboard — Steve Jobs imposed it, and within a few years, physical buttons on smartphones disappeared as a class. The same story played out with the headphone jack, the removable battery, the SD card slot — all things the consumer "wanted," until the manufacturer decided otherwise.
TikTok did not make screens large — Apple and Samsung made screens large, and TikTok emerged as a product optimized for the already existing vertical six-inch display. First the hardware — then the content to fit it. Apple killed the mini lineup not because the consumer "didn't want a compact smartphone" — the consumer wanted it and bought it — but because the margins were lower.
In this logic, screen size is determined not by user needs, but by the manufacturer's product strategy: OLED panel costs, patent wars, camera-driven chassis thickness requirements, supplier agreements. And any analysis of user patterns — how many hours they watch video, how they verify AI's work — turns out to be methodologically fragile.
The Turning Point: The AI Agent as a New Player
And here we arrive at what is perhaps the most important conclusion of the first part of this study.
The manufacturer of an AI agent is not Samsung competing with Apple over display brightness. It is a player from a different industry altogether, one that changes the very object of consumption. OpenAI, Anthropic, Google with Gemini — they sell not a device, but the ability to perform a task. And if that ability is accessible through any carrier — a smartphone, glasses, a speaker, an AI pin — then the hardware manufacturer's monopoly on shaping demand collapses.
Previously, "needs" were shaped by the smartphone itself: Apple defined vertical video, and the industry followed. Now the AI agent is an independent product that the user selects separately from the device. For the first time, a reverse movement emerges: a person chooses an agent for their task, and then selects a carrier for the chosen agent. The hardware manufacturer finds itself in the position of follower, not dictator.
This explains the failures of the Humane AI Pin and Rabbit R1 not as "the consumer wasn't ready," but as "the consumer was given a real choice for the first time" — and chose. Previously, such a choice did not exist: you bought a BlackBerry — you used the keyboard; you bought an iPhone — you used the glass. When a real choice of form factor for the same AI function appeared, it turned out that AI Pin and Rabbit were not what people needed — they needed a screen. This is a market vote that did not exist for the keyboard in 2007: back then, no alternative was offered.
Beyond the Screen: AI as an Operating System
However, the analysis would be incomplete if we stopped at the question of screen size. Behind it, an architectural shift of far greater magnitude comes into view.
From Apps to the Agent Layer
The history of computing has seen several fundamental transitions: DOS gave way to Windows, the desktop web to mobile operating systems, web search to app ecosystems. Each time, what changed was not merely the technology, but the foundational model of human access to computation.
The next possible transition is from an app-centric OS to an agent-centric OS. The user interacts not with a set of applications, but with a unified agent layer, where "apps" become invisible backend tools. The precursors are already visible: AI browsers are partially replacing search and navigation, intent-first UX allows the user to articulate a task instead of opening a specific application, and cross-app orchestration promises to become a superstructure over the fragmented app economy.
In this scenario, the smartphone ceases to be a "container for applications" and becomes a terminal for accessing the agent. The screen, microphone, camera, sensors — all remain, but the value shifts from iOS/Android to the agent system. AI is potentially capable of not merely weakening device manufacturers, but of creating a post-OS paradigm, where the traditional mobile operating system becomes the same kind of "invisible layer" that BIOS became for most users.
Control of the market in this case may pass to whoever builds the dominant agent OS — be it OpenAI, Meta, Google, or a yet-unknown player. Zuckerberg was premature with the Facebook Phone, attempting to turn a social network into a device shell. But the bet back then was on the social graph, whereas today's AI agent operates on a cognitive graph — it can simultaneously become the shell, the interface, the coordinator, the search engine, and the workflow. This is potentially far more powerful.
The Main Barrier Is Not Hardware
However, the primary barrier on the path to an agent OS is not the hardware implementation. Creating an "AI phone" is technically possible today. The real barrier is orchestration, trust, and ecosystem depth. The agent must reliably execute actions and have access to payments, identity, messaging, APIs, and security systems. The winner will not be whoever makes a "smartphone with AI," but whoever creates a new computational environment of trust.
The Revolution Begins With "Remind Me"
And here we arrive at what may be the most unexpected turn of this entire study. Virtually all futuristic models overestimate complex scenarios — "organize a vacation," "manage my finances," "replace the OS entirely" — and systematically underestimate micro-mundane attention management.
Cognitive Scaffolding Instead of Superintelligence
The real mass AI-native experience may begin not with the automation of complex tasks, but with the simplest requests:
• "Remind me not to miss the turn."
• "Remind me when my grandson gets home."
• "Remind me when it's 7 o'clock."
Yes, even that. Not an alarm — but "remind me." The difference is fundamental: an alarm is a tool that needs to be configured. "Remind me" is the delegation of an intention to an agent that will figure out the method of execution on its own.
This is a shift from task execution to cognitive scaffolding — not "do something complex for me," but "hold my context better than I can myself." Not a command executor, but a keeper of unfinished intentions — what can most precisely be called an ambient guardian of intention.
Why "Remind Me" Is More Powerful Than "Organize"
Micro-mundane scenarios possess three critical advantages over complex agent tasks.
First, frequency: such requests arise hundreds of times per week, not once a month. Second, a low threshold of trust: "remind me to turn" carries no financial risk, unlike "send $3,000," making delegation psychologically comfortable. Third, habit formation: if a system reliably maintains everyday context, it becomes a cognitive prosthesis that is difficult to abandon.
Historically, technologies win not through maximum complexity, but through the minimization of minor frustrations: autocomplete, GPS, push notifications, autosave. AI may win the mass market through anticipatory reminders — proactive nudges that connect geolocation, time, family graph, habits, calendar, and behavior into a unified contextual memory.
Agent OS as a Replacement of Forgetting
From this perspective, the agent operating system begins not as a replacement of applications, but as a replacement of forgetting. The first true AI revolution may turn out to be not in the automation of labor, but in the automation of memory and attention. And if this happens, then the simplest "remind me…" scenarios may become for the agent era what the alarm clock was for the early mobile phone: not a spectacular feature, but an everyday point of dependency.
What This Means for the Screen
At the same time, the revolution is first logical, then form-factor-driven. An AI OS will more likely first change the structure of interaction — kill app navigation, remove part of the UI complexity — than shrink the physical display. In the "reminder" scenario, the screen is needed less as a workspace, but more as a point of confirmation and trust calibration: "You asked to be reminded before the turn — now," "Grandson is home," "7:00." Brief, contextual, minimal messages.
Paradoxically, this brings us back to the original question — but on a different level. The screen does not shrink because of AI agents as such. But if the agent OS wins through micro-mundane scenarios, the nature of screen usage will change so radically that the question of its size may be reformulated anew — no longer by the manufacturer or today's consumer, but by a new model of interaction in which the screen becomes not a workspace, but a window of confirmation.
Conclusion
The initial assumption that the development of AI agents will lead to a reduction in smartphone screen size finds no confirmation upon deep analysis. AI creates more visual content than ever before. Humans need to verify the agent's work more and more. Alternative interfaces run up against social and cultural barriers.
But the main conclusions lie deeper than display size.
First. The AI agent, existing above devices and platforms, breaks for the first time in twenty years the hardware manufacturers' monopoly on shaping the user experience. The consumer receives a real choice of form factor for the first time — and the early results of this vote (the failure of screenless AI devices) speak in favor of the screen.
Second. Behind the question of screen size stands an architectural shift on the scale of DOS→Windows: a transition from an app-centric to an agent-centric operating system, where AI may become not a feature within the smartphone, but a new level of operational logic, calling into question for the first time in the mobile era the centrality of iOS and Android.
Third. Mass adoption of the agent paradigm will most likely begin not with complex automation scenarios, but with the simplest cognitive scaffolding — "remind me," "warn me," "don't let me forget." The first true AI revolution may turn out to be a revolution not of labor, but of memory and attention.
The smartphone screen will likely not shrink. But the world in which we look at it will change beyond recognition.&lt;/li&gt;
&lt;/ol&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>mobile</category>
      <category>ux</category>
    </item>
    <item>
      <title>Adaptive Company: A CSS-like Language for Describing Organizational Structure Dynamics in Crisis</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Wed, 29 Apr 2026 10:42:06 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/adaptive-company-a-css-like-language-for-describing-organizational-structure-dynamics-in-crisis-1gfk</link>
      <guid>https://dev.to/oleg_kholin_551a551b/adaptive-company-a-css-like-language-for-describing-organizational-structure-dynamics-in-crisis-1gfk</guid>
      <description>&lt;p&gt;Three Phrases Every Consultant Hears&lt;br&gt;
"I can't see how the crisis is affecting the company"&lt;br&gt;
"I can't see how the company behaves in a crisis"&lt;br&gt;
"I can't see how we can change during a crisis"&lt;br&gt;
The key word here is can't see. Not "don't know," not "don't understand." Specifically — can't see. The problem isn't a lack of data. The problem is the absence of a language capable of describing and showing what happens to a company's structure under pressure.&lt;/p&gt;




&lt;p&gt;The Problem: Company Structure Is Described as a Photograph&lt;br&gt;
Classical tools — org charts, UML diagrams, process descriptions — capture a single state. A snapshot. Here are the departments, here are the connections, here are the functions.&lt;br&gt;
But in a crisis, the structure moves. People become overloaded, roles blur, departments contract, business directions die off. A static diagram won't show any of this. It becomes outdated the moment it's created.&lt;br&gt;
What's needed is not a snapshot, but rules for how the snapshot changes. Not a description of the structure, but a description of how the structure transforms under pressure.&lt;/p&gt;




&lt;p&gt;Theoretical Framework: Taleb and Three Types of Systems&lt;br&gt;
Nassim Taleb identifies three levels of system response to stress:&lt;br&gt;
• Fragile — breaks under pressure&lt;br&gt;
• Robust — withstands pressure, maintaining functionality&lt;br&gt;
• Antifragile — grows stronger from pressure&lt;br&gt;
What is described in this article is not antifragility. A company doesn't become stronger from a crisis on its own. This is about robustness through managed structural reorganization: the ability to maintain functionality by reorganizing from within.&lt;/p&gt;




&lt;p&gt;An Analogy from Frontend: Adaptive Layout, Not Fluid&lt;br&gt;
In web development, there are two approaches to how a website responds to screen size changes:&lt;br&gt;
Fluid layout — everything changes smoothly, continuously, totally. Blocks stretch, compress, overflow. There are no fixed modes. If the rules are poorly defined — elements break out of bounds, the interface falls apart.&lt;br&gt;
Adaptive layout — the system operates in clearly defined modes with specific boundaries. When a threshold is reached — the layout restructures: the composition changes, blocks appear or disappear, the placement logic becomes different.&lt;br&gt;
A company that responds to crisis in a "fluid" manner — without clearly defined modes — becomes unreadable to external players. Clients, partners, suppliers, regulators don't understand: who is responsible for what right now? What commitments are in effect? Where are the boundaries?&lt;br&gt;
A company built on the principle of adaptive layout presents clear states to the outside world. Flexibility lives inside. A readable operating mode is what's shown outside.&lt;/p&gt;




&lt;p&gt;Two Levels of Change&lt;br&gt;
A company's structure in crisis doesn't change in just one way. There are two fundamentally different mechanisms:&lt;br&gt;
Continuous Level (Within the Structure)&lt;br&gt;
Roles, functions, employee workload, task distribution — all of this changes smoothly, within ranges. As long as the structure holds — the system compensates for pressure by redistributing the load.&lt;br&gt;
Example: a department of 10 people, each with one professional profile. The crisis reduces headcount — the number of profiles per person grows. People take on more.&lt;br&gt;
Discrete Level (The Structure Itself)&lt;br&gt;
Departments, business directions, organizational units, connections between them — these change abruptly. This is reassembly: merging departments, removing management layers, shutting down business directions.&lt;br&gt;
Example: employee profiles are a reflection of business directions. When a department shrinks below a critical threshold — it's no longer about overloading people. The budget can't sustain it, adjacent departments can't cope. This means the business directions themselves must be cut.&lt;br&gt;
The key point: first, the system compensates internally. When the internal resource is exhausted — structural reassembly occurs.&lt;/p&gt;




&lt;p&gt;Crisis Is Not a Single Parameter&lt;br&gt;
A single cause of crisis generates multiple parallel consequences for a company:&lt;br&gt;
• Financial pressure (cash flow, budgets)&lt;br&gt;
• Staffing shortages (layoffs, overload)&lt;br&gt;
• Operational overload (processes can't keep up)&lt;br&gt;
• External environment pressure (regulators, market, clients)&lt;br&gt;
A crisis doesn't strike along a single axis. It hits the entire system simultaneously.&lt;/p&gt;




&lt;p&gt;Cascading Threats: Not a Gradation, but a Screen Rotation&lt;br&gt;
When multiple threats coincide — the company's response is not sequential, not "step by step." It's not a gradual deterioration with a transition to the next level.&lt;br&gt;
It's an instant mode switch. An analogy from the same frontend world: not a window resize, but a screen orientation change — portrait → landscape. Everything restructures at once and entirely.&lt;br&gt;
A company that lacks pre-defined rules for such a switch loses controllability at that very moment.&lt;/p&gt;




&lt;p&gt;Language Architecture: Two Layers&lt;br&gt;
Layer A — Structure (UML Level)&lt;br&gt;
The base description: departments, people, roles, functions, connections, business directions. This is the company's "skeleton" at a given moment. Classical UML handles this well.&lt;br&gt;
Layer B — Deformation Rules (CSS Level)&lt;br&gt;
A description of how the structure changes as pressure shifts:&lt;br&gt;
• At certain environmental parameters → headcount changes&lt;br&gt;
• At certain parameters → departments are reorganized&lt;br&gt;
• At certain parameters → new functions and responsibilities are introduced&lt;br&gt;
• At certain parameters → business directions are cut&lt;br&gt;
This is not a static description, but a set of cascading transformation rules — analogous to how CSS defines rules for changing the display when conditions change.&lt;/p&gt;




&lt;p&gt;Why the Visual Layer Matters&lt;br&gt;
An important note: the company's business model doesn't change through this process. What changes is the internal organization. But to manage these changes, they need to be visible.&lt;br&gt;
When the structure and its dynamics are visualized:&lt;br&gt;
• Assessment criteria can be assigned to every element's state&lt;br&gt;
• UI dashboards can be built for monitoring and analysis&lt;br&gt;
• It becomes possible to see in real time: where the overload is, where the gaps are, where the structure is at its limit, where reassembly is already needed&lt;br&gt;
Visualization is not decoration. It is a management instrument. Without it, the executive returns to those same three phrases: "I can't see."&lt;/p&gt;




&lt;p&gt;Conclusion&lt;br&gt;
A company is a system with two layers of dynamics:&lt;br&gt;
• Internal → continuous redistribution of roles, functions, workload&lt;br&gt;
• Structural → discrete reassembly of departments, connections, business directions&lt;br&gt;
What's needed is a language that describes not a single state of the company, but the rules of transition between modes. A language in which:&lt;br&gt;
• The UML level defines the structure&lt;br&gt;
• The CSS level defines the rules of its transformation&lt;br&gt;
• The visual layer turns this into a manageable, observable system&lt;br&gt;
Then the executive stops "not seeing" — and begins to manage not by intuition, but by architecture.&lt;/p&gt;

</description>
      <category>css</category>
      <category>design</category>
      <category>leadership</category>
      <category>management</category>
    </item>
    <item>
      <title>The Evolution of the 3D Printing Problem: From Technological Optimism to Structural Deadlock</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Sat, 11 Apr 2026 11:08:43 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/the-evolution-of-the-3d-printing-problem-from-technological-optimism-to-structural-deadlock-5g0k</link>
      <guid>https://dev.to/oleg_kholin_551a551b/the-evolution-of-the-3d-printing-problem-from-technological-optimism-to-structural-deadlock-5g0k</guid>
      <description>&lt;p&gt;The development of 3D printing over the past decades has been accompanied by a persistent expectation of its inevitable mass adoption. The logic appeared straightforward: the technology matured, hardware became cheaper, materials became widely available, and software gradually improved in usability. Within this framework, it was assumed that further cost reduction and simplification would eventually make the 3D printer as common a household device as a paper printer or a microwave oven.&lt;/p&gt;

&lt;p&gt;However, the actual trajectory has been different. Despite technological maturity and accessibility, 3D printing has not become part of everyday domestic life. This discrepancy between expectation and reality is often explained through familiar arguments: the lack of a “killer use case,” high barriers to entry, poor economic competitiveness compared to mass-produced goods, and inferior product quality. Yet these explanations remain superficial and fail to address deeper structural causes.&lt;/p&gt;

&lt;p&gt;The Initial Misframing: False Universality&lt;/p&gt;

&lt;p&gt;The core issue begins with how the question itself is framed. The assumption that any mature technology must become mass-market ignores a fundamental distinction between classes of tasks. Some technologies serve daily or regularly recurring needs, while others address rare, highly variable, and context-specific problems.&lt;/p&gt;

&lt;p&gt;Low cost and accessibility are not sufficient conditions for mass adoption. There are many examples of inexpensive, highly capable devices that never become household standards because they do not correspond to everyday needs. The ability to use a tool does not imply the necessity of using it.&lt;/p&gt;

&lt;p&gt;In this context, 3D printing was incorrectly positioned from the outset. It was treated as a potential mass household technology, whereas by its nature it belongs to the category of specialized tools—similar to equipment used in workshops or production environments.&lt;/p&gt;

&lt;p&gt;Reframing the Context: From “Every Home” to “Every Workshop”&lt;/p&gt;

&lt;p&gt;Correcting the framing leads to a different interpretation. A 3D printer is not a household appliance in the conventional sense. It is a tool designed for solving problems that arise irregularly but require a high degree of customization.&lt;/p&gt;

&lt;p&gt;From this perspective, it becomes clear that the technology has already found stable domains of application. Jewelry production, custom components for technical devices, advertising and promotional items, and educational construction kits all demonstrate effective use of 3D printing. These domains share a common characteristic: small batch sizes, high variability, and the absence of economic justification for traditional industrial manufacturing.&lt;/p&gt;

&lt;p&gt;Thus, the issue is not the absence of demand, but its nature. The demand is not mass-market—it is niche, yet stable and reproducible.&lt;/p&gt;

&lt;p&gt;The Illusion of Technological Limitations&lt;/p&gt;

&lt;p&gt;Many arguments against broader adoption of 3D printing rely on outdated assumptions. Claims about insufficient precision, strength, or functionality increasingly fail to reflect current reality. Modern desktop systems are capable of producing working mechanical components suitable for practical use without additional finishing.&lt;/p&gt;

&lt;p&gt;Other limitations, such as water resistance or consistency of output, are often interpreted as inherent to the technology. In practice, however, these depend heavily on process parameters. Their resolution lies in standardization and reproducibility of settings, not in altering the underlying physics of the process.&lt;/p&gt;

&lt;p&gt;Thus, many perceived “limitations” are not technological but infrastructural.&lt;/p&gt;

&lt;p&gt;The Ecosystem as a Consequence of Task Structure&lt;/p&gt;

&lt;p&gt;Another commonly cited issue is the lack of a developed ecosystem—unified model libraries, standardized print profiles, and user-friendly tools. However, a deeper analysis shows that an ecosystem cannot emerge independently of a structured understanding of tasks.&lt;/p&gt;

&lt;p&gt;In mature engineering and software systems, the primary layer is not the toolset but the ontology of objects and operations. Users work not with abstract geometry, but with entities that have parameters and behavior. This allows systems to scale through extensions, reuse, and accumulation of knowledge.&lt;/p&gt;

&lt;p&gt;In 3D printing, the situation is reversed: tools exist, but there is no shared understanding of what tasks are being solved or how. As a result, each user constructs an individual workflow, and accumulated experience does not scale across the system.&lt;/p&gt;

&lt;p&gt;Under these conditions, an ecosystem cannot be built directly. It can only emerge as a byproduct of task systematization.&lt;/p&gt;

&lt;p&gt;The Representation Problem: From Geometry to Parameters&lt;/p&gt;

&lt;p&gt;The dominant model exchange format—static geometric files—limits reuse and adaptability. These models contain no information about purpose, constraints, or functional parameters.&lt;/p&gt;

&lt;p&gt;A parametric approach, by contrast, defines objects through relationships and constraints. This enables adaptation to specific conditions without breaking functionality. However, adoption of this approach is constrained by the lack of accessible tools aligned with real-world workflows.&lt;/p&gt;

&lt;p&gt;The gap between existing CAD systems and practical user behavior remains one of the central barriers.&lt;/p&gt;

&lt;p&gt;The Role of Adjacent Technologies&lt;/p&gt;

&lt;p&gt;The evolution of 3D printing is closely tied to the maturity of adjacent technologies. One of the most critical missing components is affordable, accurate 3D scanning. The ability to quickly capture the geometry of existing objects would significantly simplify many practical workflows, particularly those involving replication or repair.&lt;/p&gt;

&lt;p&gt;The absence of such tools increases labor costs and reduces accessibility, further limiting adoption. In this sense, 3D printing remains partially constrained by the immaturity of its technological ecosystem.&lt;/p&gt;

&lt;p&gt;The Limits of Generative Solutions&lt;/p&gt;

&lt;p&gt;Attempts to compensate for the lack of models through generative approaches encounter a fundamental limitation. Generative systems are oriented toward creating new forms, while many real-world tasks require accurate reproduction of existing objects under functional constraints.&lt;/p&gt;

&lt;p&gt;Without embedded engineering logic, generated models may appear plausible but fail in practical use. This highlights the distinction between form synthesis and engineering design. The former may assist the latter, but cannot replace it.&lt;/p&gt;

&lt;p&gt;The Absence of a Dominant Use Scenario&lt;/p&gt;

&lt;p&gt;Another defining feature of 3D printing is the absence of a dominant, unifying application scenario. In successful technological domains, development is typically organized around a small number of clearly defined use cases, which drive standardization and infrastructure.&lt;/p&gt;

&lt;p&gt;In contrast, 3D printing is characterized by a wide range of fragmented applications without consolidation. This fragmentation hinders standardization and slows ecosystem development.&lt;/p&gt;

&lt;p&gt;The Non-Obvious Cause: The Absence of a Risk-Bearing Actor&lt;/p&gt;

&lt;p&gt;The deepest layer of the problem lies in the distribution of risk. Building a fully functional ecosystem requires long-term investment, coordination across multiple layers, and acceptance of uncertainty. Yet the benefits of such an ecosystem are distributed across many participants, while the costs are concentrated on whoever initiates it.&lt;/p&gt;

&lt;p&gt;Hardware manufacturers are incentivized to protect proprietary advantages rather than standardize. Software companies focus on high-margin enterprise markets. Open-source communities lack the resources to deliver robust, production-grade systems. Investors are reluctant to engage with long-term, uncertain, and weakly monetizable opportunities.&lt;/p&gt;

&lt;p&gt;As a result, no actor emerges for whom building the ecosystem is a rational decision. This creates a structural deadlock: the technology exists, demand exists in niches, partial solutions exist—but integration does not occur.&lt;/p&gt;

&lt;p&gt;This distinguishes 3D printing from cases of successful technological scaling. In those cases, there is always an actor for whom the cost of inaction exceeds the cost of building the system. That actor may be a company, a consortium, or a public institution—but it exists.&lt;/p&gt;

&lt;p&gt;In 3D printing, such an actor has not yet emerged. Moreover, the current distribution of incentives actively discourages their appearance. Benefits are diffuse, while risks are concentrated.&lt;/p&gt;

&lt;p&gt;Therefore, the absence of an ecosystem is not the root cause, but a consequence. The root cause lies in the economics of risk. As long as the cost of integration exceeds its expected return for any individual participant, systemic solutions will remain unrealized.&lt;/p&gt;

&lt;p&gt;The Resulting Picture&lt;/p&gt;

&lt;p&gt;3D printing is not a failed mass-market technology. It is a mature tool for a specific class of problems that do not align with everyday consumer use.&lt;/p&gt;

&lt;p&gt;Its limitations are not primarily technological, but structural:&lt;/p&gt;

&lt;p&gt;incorrect framing of mass adoption as a goal;&lt;br&gt;
absence of a formalized task space;&lt;br&gt;
inadequate model representation formats;&lt;br&gt;
mismatch between tools and real workflows;&lt;br&gt;
immaturity of adjacent technologies;&lt;br&gt;
lack of dominant application scenarios;&lt;br&gt;
absence of an actor willing to bear integration risk.&lt;br&gt;
Future Directions&lt;/p&gt;

&lt;p&gt;The future of 3D printing depends less on improving hardware and more on advancing the organization of knowledge and systems around it:&lt;/p&gt;

&lt;p&gt;developing a clear taxonomy of tasks and use cases;&lt;br&gt;
transitioning from geometric to parametric models;&lt;br&gt;
creating tools aligned with actual workflows;&lt;br&gt;
standardizing print profiles by object type rather than hardware;&lt;br&gt;
advancing accessible methods for geometry acquisition;&lt;br&gt;
identifying a limited number of scalable application domains.&lt;/p&gt;

&lt;p&gt;Until such developments occur, 3D printing will remain an effective but localized tool—widely used in professional and semi-professional contexts, yet lacking a mechanism for broader systemic adoption.&lt;/p&gt;

</description>
      <category>design</category>
      <category>discuss</category>
      <category>product</category>
      <category>science</category>
    </item>
    <item>
      <title>Rising on the Shoulders of Giants: Top 10 Technologies That Succeeded by "Hijacking" Infrastructure</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Fri, 10 Apr 2026 04:58:47 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/rising-on-the-shoulders-of-giants-top-10-technologies-that-succeeded-by-hijacking-infrastructure-2p0l</link>
      <guid>https://dev.to/oleg_kholin_551a551b/rising-on-the-shoulders-of-giants-top-10-technologies-that-succeeded-by-hijacking-infrastructure-2p0l</guid>
      <description>&lt;p&gt;Introduction: Why Do Some Technologies "Fly" While Others Stagnate?&lt;br&gt;
According to the "Four Trees" framework, the most expensive and time-consuming stage of progress is growing the "Roots"—Trees 3 &amp;amp; 4 (the auxiliary tools and the infrastructure for mass production). Most brilliant ideas perish because they attempt to grow these roots from scratch.&lt;br&gt;
However, there exists a strategy of "Engineering Refactoring." Instead of growing its own tree, a Strategic Subject performs a scan of the global market and identifies a mature Tree 4 in a completely different industry. By "hijacking" this existing infrastructure and adapting it to a new task, they achieve a lightning-fast leap to Tree 2 (the mass product). This approach allows them to collapse the cascade of costs by 80–90% and seize the market.&lt;br&gt;
Below are 10 examples of how identifying and utilizing "foreign" auxiliary trees created new technological empires.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Consumer Drones
Tree 1 (The Idea): An autonomous flying robot.
The Donor (Tree 4): The smartphone industry.
The Refactoring: DJI and others performed a "reconnaissance from the small," assembling drones from "spare parts" intended for phones. This collapsed the cost of such devices from $100,000 to $1,000.&lt;/li&gt;
&lt;li&gt;Artificial Intelligence (AI/Neural Networks)
Tree 1 (The Idea): Machine learning via neural networks.
The Donor (Tree 4): The gaming industry (GPUs).
The Refactoring: AI stagnated for decades until developers "recognized" in NVIDIA’s gaming video cards the perfect tool for matrix calculations. AI "parasitized" the global demand for video games, gaining a ready-made computational base.&lt;/li&gt;
&lt;li&gt;Electric Vehicles (Tesla)
Tree 1 (The Idea): A mass-market, high-performance electric car.
The Donor (Tree 4): The consumer electronics industry (laptops).
The Refactoring: Elon Musk used standard 18650 lithium-ion cells already mass-produced for laptops. Tesla rose on the "roots" planted by Panasonic and Sony, which were already mature and cheap.&lt;/li&gt;
&lt;li&gt;Uber and Modern Logistics
Tree 1 (The Idea): Real-time global management of movement.
The Donor (Tree 4): Military satellite navigation (GPS) + 4G networks.
The Refactoring: Uber utilized the existing Tree 4 created by the Pentagon and telecommunication giants. By changing the mechanism—using an app on a "borrowed" Tree 4—they annihilated the old taxi industry.&lt;/li&gt;
&lt;li&gt;Medical Express Tests (PCR/COVID Tests)
Tree 1 (The Idea): Instant molecular diagnostics at a scale of billions.
The Donor (Tree 4): The food and beverage industry (PET bottle production).
The Refactoring: The industry "hijacked" the production of PET preforms (the blanks used for soda bottles). The same mass-production lines of Tree 4 allowed the world to be flooded with cheap tests.&lt;/li&gt;
&lt;li&gt;"Digital Twins" in Architecture
Tree 1 (The Idea): Photo-realistic modeling of cities and factories.
The Donor (Tree 4): Game Engines (Unreal Engine / Unity).
The Refactoring: Hijacking game-engine technology allowed architects to design factories with a speed and fidelity that traditional CAD systems could never match.&lt;/li&gt;
&lt;li&gt;Warehouse and Domestic Robots (LiDARs and Sensors)
Tree 1 (The Idea): Robots that navigate space autonomously.
The Donor (Tree 4): The automotive industry (ADAS systems).
The Refactoring: Robotics "harvested the fruit" from trees planted by Toyota and Mercedes, which collapsed the price of LiDARs and cameras by implementing them in mass-market cars.&lt;/li&gt;
&lt;li&gt;Cloud Computing (AWS)
Tree 1 (The Idea): Selling computational power as a utility.
The Donor (Tree 4): Amazon’s retail infrastructure.
The Refactoring: Today, half the internet runs on the "roots" Amazon originally grew to support its own online bookstore.&lt;/li&gt;
&lt;li&gt;Vertical Farming (AgroTech)
Tree 1 (The Idea): Year-round food production within urban centers.
The Donor (Tree 4): The LED display industry (TVs and Smartphones).
The Refactoring: Vertical farming became viable only when mass production of screens collapsed the price of LEDs. Agriculture "plugged into" the auxiliary tree of the electronics industry.&lt;/li&gt;
&lt;li&gt;3D Bioprinting
Tree 1 (The Idea): Printing human organs and tissues.
The Donor (Tree 4): Office inkjet printing.
The Refactoring: The first bioprinters were created by hacking standard HP and Epson printers. The mature Tree 4 mechanics of droplet delivery were hijacked to print living cells.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Investment Radar: Identifying Future Unicorns through Tree Analysis&lt;br&gt;
For an investor, the "Four Trees" model serves as a precision instrument to identify hidden market leaders at an early stage. A true "unicorn" is often born not in a lab, but at the intersection of a mature, external infrastructure and a novel business objective.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;"Donor Signal" Detection
The beginning of an explosive growth phase for a new player can often be predicted by analyzing the financial reporting of entirely different companies—component suppliers.
The Core Concept: A sudden, unexplained spike in sales of specific components among "shoulder companies" (producers of chips, sensors, batteries) often signals the emergence of a hidden giant in an adjacent industry.
Example: An abnormal growth in demand for miniature gyroscopes and Li-Po batteries between 2008 and 2010 was the signal for the birth of the consumer drone industry, well before brands like DJI became household names. Investors should watch those who suddenly begin buying "smartphone parts" for non-smartphone purposes.&lt;/li&gt;
&lt;li&gt;Investment Arbitrage on the "Broken Cascade"
The most profitable companies are those that reach the mass-market product stage (Tree 2) without bearing the colossal costs of building their own fundamental base (Trees 3–4).
The Strategy: Look for companies with abnormally high margins in their early stages and a low level of Capital Expenditure (CapEx) compared to industry leaders.
The Indicator: If a company shows results comparable to industry giants while having a 10x smaller R&amp;amp;D budget, it is a sure sign they have successfully performed an Engineering Refactoring and are using someone else's infrastructure as a free resource.&lt;/li&gt;
&lt;li&gt;Searching for "Universal Roots" (Predicting the Next Wave)
An investor can predict the next wave of technological breakthroughs simply by evaluating the maturity of specific "donor markets."
The Forecast: As soon as a technology in one industry (e.g., satellite internet, laser scanners, or solid-state batteries) reaches the Tree 4 stage—meaning it has become cheap, reliable, and mass-produced—it automatically opens a "window of opportunity" for new Tree 1 ideas in adjacent fields.
The Action: Analyze which mature "roots" are currently sitting on the market, unused. The entity that is first to "graft" these roots onto a new market niche will become the next global leader.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Conclusion: The Lesson of the Strategic Subject&lt;br&gt;
All these examples share a single pattern: Success came not through the invention of new "roots," but through the ability to recognize them in other industries.&lt;br&gt;
The winner is not the one who tries to build everything from scratch, but the one who performs Engineering Refactoring and "grafts" their idea onto the strongest and most massive Tree 4 available on the market. Do not just follow science journals; watch the supply chains of mass-market components. Real business subjectivity is the ability to recognize someone else's mature Tree 4 faster than the competition and rebuild your architecture to exploit it, collapsing the cost structures of everyone else in the process!&lt;/p&gt;

</description>
      <category>architecture</category>
      <category>product</category>
      <category>startup</category>
      <category>systemdesign</category>
    </item>
    <item>
      <title>The Invisible Roots of Progress: Top 10 Supermaterials Stuck in the Laboratory</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Wed, 08 Apr 2026 20:20:26 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/the-invisible-roots-of-progress-top-10-supermaterials-stuck-in-the-laboratory-3do2</link>
      <guid>https://dev.to/oleg_kholin_551a551b/the-invisible-roots-of-progress-top-10-supermaterials-stuck-in-the-laboratory-3do2</guid>
      <description>&lt;p&gt;The popular essay &lt;strong&gt;"The Four Trees"&lt;/strong&gt; offers an original lens through which to view technological progress. According to this concept, the development of any technology rests upon four metaphorical "trees":&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 (The Idea):&lt;/strong&gt; The fundamental concept or laboratory proof-of-concept. The principle is proven; the physics works.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Tree 2 (The Mass Product):&lt;/strong&gt; The stage of mass production and widespread infrastructure. What we produce at scale and use in daily life.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Trees 3 &amp;amp; 4 (The Auxiliary Roots):&lt;/strong&gt; Auxiliary tools and the secondary technologies used to produce them. These are the "hidden" roots—the lithography machines, the specialized furnaces, the methods of purification, and the precise manipulation of matter.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  The Great Barrier: From Assembly to Integration
&lt;/h3&gt;

&lt;p&gt;The reason many supermaterials fail to go mainstream is deeper than mere "cost." We are currently stuck in the trap of &lt;strong&gt;Miniaturization&lt;/strong&gt;. This is the stage where we simply shrink individual components and attempt to connect them (similar to how vacuum tubes were replaced by discrete transistors).&lt;/p&gt;

&lt;p&gt;The true revolutionary leap is &lt;strong&gt;Micro-miniaturization&lt;/strong&gt; (Integration). This is the transition from "assembling discrete parts" to "forming a structure." In microelectronics, we don't solder millions of transistors together; we grow them simultaneously as a single, integrated structure on a silicon wafer through deposition and etching.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The tragedy of modern supermaterials:&lt;/strong&gt; We still treat them as "discrete parts" (we try to "cut" graphene or "glue" a nanotube). We are still thinking in the category of "assembly," whereas we desperately need a "lithography for materials." Until we learn to form the structure of a device directly &lt;em&gt;out of&lt;/em&gt; the material itself, we will remain in the era of "expensive transistors," never reaching the era of "cheap integrated circuits."&lt;/p&gt;

&lt;p&gt;Below are the Top 10 Tree 1 materials waiting for their "integrated revolution."&lt;/p&gt;




&lt;h3&gt;
  
  
  1. Graphene: Two-Dimensional Carbon
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Proven in 2004. A single-atom-thick layer of carbon. The strongest and most conductive material in the universe.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; We are still trying to "transfer" it like a delicate film. This is the era of assembly. For graphene to reach Tree 2, it must be grown directly into the specific regions of a chip as part of a unified integrated circuit, rather than being "pasted" onto existing ones.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Nitinol and Shape Memory Alloys
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Alloys (e.g., Titanium-Nickel) that return to a complex original shape when heated.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; We currently produce them as discrete "parts" (stents, wires). We lack the technology to integrate "shape memory" directly into the 3D structure of a product during the fabrication stage, allowing the material itself to act as the mechanism.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Carbon Nanotubes (CNTs)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Cylindrical carbon structures, 100 times stronger than steel and lighter than aluminum.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; We can produce "nanopowder" (a discrete additive), but we cannot yet form a continuous macro-structure (like a thread or a sheet) without losing their unique properties at the molecular boundaries. We need a method of "weaving" the structure at the moment of formation, rather than attempting to assemble billions of individual fibers.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  4. Metallic Glasses (Amorphous Metals)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Metals with a disordered, liquid-like atomic structure. Extremely strong and immune to corrosion.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; We are limited by the requirement of "extreme cooling rates," which restricts us to making only thin ribbons or small parts. We lack the Tree 4 technology to form bulk structures where the amorphous state is preserved during the casting of large masses.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  5. Aerogels: "Frozen Smoke"
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; A crystalline lattice consisting of 99% air. The lightest solid and the world’s best thermal insulator.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; Production requires supercritical drying in high-pressure autoclaves—a boutique, "batch-assembly" method. To become Tree 2, aerogels must evolve into a material that can be deposited like a spray-on foam directly at a construction site, forming its structure in-situ.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  6. MXenes: 2D Metals
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Two-dimensional crystals made of metals and carbon, capable of ultra-fast battery charging.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; Currently obtained through "etching" (a subtractive and dirty chemical process). This is a discrete, wasteful method. We need the technology to "grow" MXenes directly as electrodes within the pre-integrated structure of a battery.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  7. Borophene: Single-Atom Boron
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; A 2D layer of boron, even stronger and more flexible than graphene.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; It can only survive in an ultra-high vacuum. We lack the technology for "integrated encapsulation"—where the material is grown and immediately sealed with a protective atomic layer in a single, continuous process.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  8. Perovskites: "Printable" Solar Power
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Crystals that convert light to electricity more efficiently than silicon.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; They degrade rapidly in the presence of moisture. The solution is not just "better chemistry" but a breakthrough in "integrated sandwich-structure" fabrication, where the active perovskite and its transparent protection are formed as a unified, airtight structure during the printing process.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  9. High-Entropy Alloys (HEAs)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Alloys made of 5 or more metals in equal proportions, offering extreme heat and radiation resistance.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; The problem of homogeneity. We need "atomic mixing" tools (such as high-speed laser deposition) so that the alloy is formed directly as the final part, rather than being smelted into a bulk ingot that requires further, less precise processing.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  10. Synthetic Spider Silk: Bio-Kevlar
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tree 1 Status:&lt;/strong&gt; Stronger than Kevlar and more elastic than Nylon.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Trees 3–4 Bottleneck:&lt;/strong&gt; We can produce the protein (the raw material), but we cannot yet "form the thread" (the structure) with the same molecular grace as a spider. This is the transition from "brewing a soup" in a bioreactor to "molecular-level weaving."&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  Final Conclusion
&lt;/h3&gt;

&lt;p&gt;Today’s supermaterials are stuck at the &lt;strong&gt;transistor level of the 1950s&lt;/strong&gt;. We have learned how to create them, but we have not yet learned how to unite them into "Integrated Circuits of Matter."&lt;/p&gt;

&lt;p&gt;The problem is not that these technologies are inherently too expensive; the problem is that we are still trying to &lt;strong&gt;assemble&lt;/strong&gt; the future by hand, piece by piece, instead of &lt;strong&gt;forming&lt;/strong&gt; its structure as a unified whole. The entity that first creates a "lithography for materials"—moving from the assembly of parts to the integrated growth of systems—will become the new technological leader, controlling real progress!&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Decontextualization of Anthropomorphism in Robotics: An Epistemological and Physical Analysis of a Paradigm Shift</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Sat, 28 Mar 2026 19:38:24 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/decontextualization-of-anthropomorphism-in-robotics-an-epistemological-and-physical-analysis-of-a-41oc</link>
      <guid>https://dev.to/oleg_kholin_551a551b/decontextualization-of-anthropomorphism-in-robotics-an-epistemological-and-physical-analysis-of-a-41oc</guid>
      <description>&lt;p&gt;Abstract&lt;br&gt;
Contemporary robotics relies heavily on anthropomorphic morphology as a universal standard for interaction with the human environment. This paper proposes a reconsideration of this assumption. It is shown that anthropomorphism is often not an engineering necessity, but a historical and cognitive legacy that emerged within a specific infrastructural context. A method of decontextualizing anthropomorphic assumptions is proposed through physical stress-tests of the environment, demonstrating the limitations of humanoid architectures. The concept of separating social and physical compatibility of agents is introduced, and a program of adaptive tests is proposed aimed at identifying optimal morphologies for extreme and unstable environments. The paper formulates a framework for transitioning from anthropocentric design to physically conditioned morphological optimization.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Introduction
The majority of contemporary robotics is developed within an implicit assumption:
a robot must be compatible with infrastructure created for humans.
From this assumption, a second is often derived:
a robot must be anthropomorphic.
However, this assumption is rarely subjected to systematic analysis at the level of the physics of interaction with the real environment. In many scenarios — rescue operations, unstable surfaces, chemically active zones, destroyed infrastructures — humanoid morphology proves to be not optimal, and sometimes physically unstable.
This leads to the key research question:
can anthropomorphism be not a universal engineering solution, but a contextually limited historical design strategy?
This paper proposes viewing the current situation as an epistemological shift in robotics, in which a change in the physical context of tasks naturally leads to the loss of applicability of old paradigms — without the need for their direct refutation.&lt;/li&gt;
&lt;li&gt;The Physical Impossibility of Old Paradigms
Scientific paradigms sometimes cease to function not because they are logically refuted, but because the physical context in which they were effective disappears.
In engineering systems this manifests particularly clearly:
a change in environment automatically changes the optimal morphology of the agent.
When a robot must function under conditions of:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;unstable surfaces,&lt;br&gt;
degrading objects,&lt;br&gt;
chemically active environments,&lt;br&gt;
phase transitions of materials,&lt;/p&gt;

&lt;p&gt;anthropomorphic architecture begins to encounter fundamental limitations of mechanics and stability.&lt;br&gt;
In this sense, the paradigm shift occurs not through argumentation, but through the physics of tasks.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Historical Contextuality of Infrastructure
Human infrastructure is often perceived as a universal environment to which all artificial agents must adapt. However, infrastructures were formed historically and adapted to human biomechanics.
The history of technology shows that:
environment and agents evolve jointly.
Examples:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;automobiles led to the appearance of asphalt roads and traffic signals,&lt;br&gt;
wheelchairs — to ramps and changes in architectural standards,&lt;br&gt;
electric vehicles — to the infrastructure of charging stations.&lt;/p&gt;

&lt;p&gt;Thus, the requirement for robots to fully adapt to existing infrastructure ignores the historical precedent of technological co-evolution.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Biomechanics as a Variable
Human biomechanics is often regarded as a fixed model of interaction with the environment. However, real conditions demonstrate the opposite.
When the environment changes, the movement strategy changes as well:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;on slippery surfaces, additional points of support are used,&lt;br&gt;
on steep inclines, the center of mass is shifted,&lt;br&gt;
in unstable environments, the contact area is increased.&lt;/p&gt;

&lt;p&gt;Consequently, the requirement to universally preserve anthropomorphic form means fixing one point in the space of possible morphologies, ignoring the adaptability characteristic of engineering systems.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Social Acceptance Does Not Require Anthropomorphism
One of the common arguments in favor of humanoid robots is the convenience of interaction.
However, empirical reality demonstrates the opposite.
Many technological agents have been successfully integrated into the social environment without anthropomorphic form:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;ATMs,&lt;br&gt;
order terminals,&lt;br&gt;
self-checkout machines,&lt;br&gt;
information kiosks.&lt;/p&gt;

&lt;p&gt;People interact with them without discomfort, because the key factor turns out to be not resemblance to a human, but:&lt;/p&gt;

&lt;p&gt;predictability of behavior&lt;br&gt;
and clarity of functional role.&lt;/p&gt;

&lt;p&gt;This allows two independent parameters to be identified:&lt;/p&gt;

&lt;p&gt;social compatibility&lt;br&gt;
and physical compatibility.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Physical Limitations of Anthropomorphic Morphology
Humanoid architecture possesses a number of characteristics that in certain environments become limitations:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;high center of mass&lt;br&gt;
limited contact area&lt;br&gt;
discrete step kinematics&lt;br&gt;
high sensitivity to loss of traction&lt;br&gt;
complexity of load redistribution&lt;/p&gt;

&lt;p&gt;A thought experiment involving a humanoid robot moving across thin ice with a load demonstrates these limitations.&lt;br&gt;
Narrow contact points create high pressure on the surface, increasing the probability of structural failure of the support. A high center of mass reduces stability during sliding, and the discrete structure of the step limits adaptation to a continuously changing surface.&lt;br&gt;
Under such conditions, anthropomorphism becomes not merely a suboptimal solution, but a potential engineering error.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Method of Decontextualization of Anthropomorphism
This paper proposes a method that can be described as the decontextualization of engineering myths.
The idea of the method is as follows:
an engineering hypothesis is tested not only in a standard environment, but also under conditions of changing physical context.
If the model loses its operability when the environment changes, its universality proves to be illusory.
The method includes three stages:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;changing the physical context of the task&lt;br&gt;
observing the degradation of the original morphology&lt;br&gt;
searching for an alternative architecture&lt;/p&gt;

&lt;p&gt;In this way, the paradigm is dismantled naturally — through incompatibility with new conditions.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Evolution of Adaptivity Tests
To identify the limitations of anthropomorphic systems, a sequence of tests of increasing complexity is proposed.
Test 1 — Retention of a Degrading Object
The system must accompany an object that is gradually losing its shape and structure.
This tests the ability to adapt to continuous changes in geometry.
Test 2 — Evacuation of a Human During Morphological Collapse
A human in an extreme environment may lose the structural stability of the body. The robot must complete the rescue before the point of irreversible damage.
This introduces a temporal and dynamic component to the task.
Test 3 — Extraction of Materials Before Phase Transition
The task includes chemical and physical instability of the object, which may become dangerous.
Here the ability of the system to operate under conditions of changing matter is tested.
These tests shift the focus from the form of the robot to the physics of interaction.&lt;/li&gt;
&lt;li&gt;The Precedents Approach
Instead of directly asserting a new paradigm theoretically, development through a chain of engineering precedents is possible.
First, tasks appear in which old architectures systematically fail. Then different research groups independently find new solutions.
As a result, a new practice is formed that gradually becomes the standard.
This path is slow, but sustainable:
it creates change through the accumulation of facts, not through declarations.&lt;/li&gt;
&lt;li&gt;Separation of Social and Physical Architecture
One of the key conclusions of the paper is the necessity of separating two functions of the robot:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;the social interface&lt;br&gt;
and the physical executive body.&lt;/p&gt;

&lt;p&gt;Anthropomorphism may be useful as an interface:&lt;/p&gt;

&lt;p&gt;for communication,&lt;br&gt;
for recognition of intentions,&lt;br&gt;
for reducing the cognitive load on the human.&lt;/p&gt;

&lt;p&gt;However, physical morphology must be determined exclusively by the physics of the environment.&lt;br&gt;
Separating these levels opens up new architectures of robotic systems.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Co-evolution of Environment and Agents
The history of technology shows that environments change under the influence of new agents.
In the long term, the emergence of new types of robots may lead to changes in infrastructure:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;adaptive surfaces,&lt;br&gt;
dynamic transport environments,&lt;br&gt;
hybrid architectural systems.&lt;/p&gt;

&lt;p&gt;Thus, the future of robotics may lie not in copying human form, but in the joint design of environment and agents.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Conclusion
The paper proposes an analytical framework in which anthropomorphism is viewed not as a universal standard of robotics, but as a historically conditioned strategy, effective only in certain contexts.
It is shown that a change in the physical environment of tasks naturally leads to the loss of applicability of old morphologies. Under these conditions, a transition becomes necessary — from anthropocentric design to physically conditioned morphological optimization.
Such a transition is not the victory of one idea over another, but rather a change in the plane on which questions are posed in robotics.
On the new plane, the key criterion is no longer the robot's resemblance to a human, but the correspondence of its form to the physics of the environment.&lt;/li&gt;
&lt;/ol&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>computerscience</category>
      <category>science</category>
    </item>
    <item>
      <title>AI and Creativity: Making Sense of a Real Shift</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Wed, 25 Mar 2026 10:25:02 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/ai-and-creativity-making-sense-of-a-real-shift-2pj4</link>
      <guid>https://dev.to/oleg_kholin_551a551b/ai-and-creativity-making-sense-of-a-real-shift-2pj4</guid>
      <description>&lt;h3&gt;
  
  
  Our perspective on Tim Green's article &lt;a href="https://dev.to/rawveg/no-consent-no-credit-no-pay-23p5"&gt;"No Consent, No Credit, No Pay"&lt;/a&gt;
&lt;/h3&gt;




&lt;h2&gt;
  
  
  What Is Actually Happening
&lt;/h2&gt;

&lt;p&gt;The public debate around generative AI and artists' rights is focused on legal details — datasets, lawsuits, licensing models. But behind all this noise, the main point escapes notice: we are witnessing not a technological improvement of existing tools, but the elimination of the very need for intermediaries between an idea and its realisation.&lt;/p&gt;

&lt;p&gt;The chain used to look like this: &lt;strong&gt;idea → specialist → tool → result&lt;/strong&gt;. Now it looks like this: &lt;strong&gt;idea → result&lt;/strong&gt;. AI did not replace the designer with a better Photoshop. It made the designer an unnecessary link in the chain. And this is permanent.&lt;/p&gt;

&lt;p&gt;Filters and plugins never cancelled the tools themselves. 3ds Max, After Effects, Photoshop — they remained necessary, which meant the people who knew how to use them remained necessary too. AI became a thin client that replaced both the tools and the specialists in one move: designers, layout artists, retouchers, illustrators, pattern makers. You no longer need them — and you no longer need what they used either.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Analogy Everyone Misses
&lt;/h2&gt;

&lt;p&gt;The authors of such articles and the participants in court proceedings compare what is happening to piracy or copyright infringement on the internet. That is an imprecise analogy.&lt;/p&gt;

&lt;p&gt;The precise one is &lt;strong&gt;Gutenberg's printing press&lt;/strong&gt;. It did not give scribes a better writing tool. It made the scribe an unnecessary link in the distribution of text. The profession did not disappear overnight — new niches emerged, new forms of craft. But the economic foundation was undermined irreversibly.&lt;/p&gt;

&lt;p&gt;Even closer is &lt;strong&gt;the history of photography&lt;/strong&gt;. The arrival of smartphones with quality cameras did not destroy photographers in direct competition. It simply turned out that the vast majority of consumers were satisfied with the level of photography available on their phones. Professionals survived in niches: artistic reportage, film photography, studio work. Enthusiasts of film still exist — just as enthusiasts of valve amplifiers do. But the monopoly on quality imagery collapsed forever.&lt;/p&gt;

&lt;p&gt;AI is doing exactly the same thing to illustration and design.&lt;/p&gt;




&lt;h2&gt;
  
  
  What Really Worries Artists — and What They Are Confusing
&lt;/h2&gt;

&lt;p&gt;The article lists grievances: lack of consent, attribution, compensation, style copying. Let us examine each honestly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Style copying&lt;/strong&gt; is painful, but it is not new. On DeviantArt, an interesting new style gets copied the very next day — without any AI involved. Style has never been a subject of legal protection in any jurisdiction. AI has merely accelerated and scaled what was already happening.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Legal lawsuits&lt;/strong&gt; are not built on style but on fact: specific images were physically present in the LAION-5B dataset, downloaded and used for commercial purposes without the authors' permission. This is closer to a real violation — but even here the boundary is blurred. If you train a model on photographs of interiors where paintings hang on walls, purchased by the homeowners — who is the infringer? The law does not yet have an answer.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The demand for attribution&lt;/strong&gt; from AI seems strange when we never demanded it from human artists inspired by each other's work. This is not an argument — it is a symptom of disorientation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Compensation&lt;/strong&gt; is the only genuinely strong argument. Companies earned billions not from specific images but from models for whose creation those images were indispensable. The Spotify analogy works here: the platform pays authors not because it reproduces a specific track, but because it uses the entire catalogue as the foundation of its business. The logic of Getty Images and Sweden's STIM is exactly this — and it is convincing.&lt;/p&gt;

&lt;p&gt;But even winning every lawsuit and receiving royalties will not bring back the corporate illustration market. It is gone — just as photographers lost the family portrait market.&lt;/p&gt;




&lt;h2&gt;
  
  
  Where the Real Protection Lies
&lt;/h2&gt;

&lt;p&gt;Practice shows that those who survive are not those who litigate, but those who keep moving.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Unique technique&lt;/strong&gt; is real protection. An authorial system built on strict geometric curves, non-trivial mathematical foundations, rare combinations — this is opaque to AI. It averages what appeared frequently in the dataset. What is rare and original it cannot reproduce — it struggles even to describe it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fusion&lt;/strong&gt; — combining two or more styles in non-trivial combinations — produces works that AI cannot decompose into base layers if that particular combination was absent from its training. Papercut art plus liquid art: AI sees the result but cannot see the structure.&lt;/p&gt;

&lt;p&gt;However, there is a crucial practical nuance here. "Idea → result" is not yet a straight arrow. AI in its current mass form does not know your visual language. It averages. It drifts. It draws things you did not ask for. Ask it for a shadow from a fountain — and it will try to draw the fountain too. This is not a flaw that will be patched in the next update. It is a fundamental property of a model trained on averaged mass data: it does not understand local logic, a part without the whole.&lt;/p&gt;

&lt;p&gt;This means the gap between intention and realisation is still real. And as long as that gap exists, craft does not disappear — it simply changes its form.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Selling process, not just result&lt;/strong&gt; — the most sustainable model. An artist who sells not only paintings but brush profiles, techniques, and tools sells something AI does not produce. AI uses brushes but does not sell them.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Live contact with the audience&lt;/strong&gt; creates attachment to the author as a person, not to the genre. This is what economists call switching cost — viewers may go to AI for an image, but they come back to the artist for the human being.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Deficit That Was Always There
&lt;/h2&gt;

&lt;p&gt;The deepest problem is neither technical nor legal. An idea is never protected by copyright anywhere — only its specific realisation. DeviantArt demonstrates this daily: a new idea survives one day before the first imitator appears. AI merely accelerates an already existing process.&lt;/p&gt;

&lt;p&gt;The real deficit — of original ideas — existed long before generative AI. Where did the calligraphers go? They still exist, but the bulk of calligraphy is now produced by electronic and software means. This is not a tragedy — it is a civilisational shift.&lt;/p&gt;




&lt;h2&gt;
  
  
  What Comes Next — and Why It Cannot Be Stopped Either
&lt;/h2&gt;

&lt;p&gt;There is one more dimension that changes the entire picture.&lt;/p&gt;

&lt;p&gt;Right now AI is a mass tool with averaged models. This is like a typewriter: everyone gets the same font. But the moment is approaching when an artist will be able to train a model on materials they have personally selected — including their own paintings, their own visual language, their own logic of form. And this cannot be stopped either.&lt;/p&gt;

&lt;p&gt;A personally trained model closes the loop. It knows your visual language. It understands your local logic. It realises your specific intention rather than an averaged one. The gap between intention and realisation — the shadow without the fountain — becomes your problem to solve, not a limitation imposed by someone else's model.&lt;/p&gt;

&lt;p&gt;This is no longer "a thin client replaced the specialist." This is the specialist acquiring a personal thin client. A qualitatively different situation. And it destroys the linear picture of "AI displacing the artist." The real trajectory is more complex: mass AI displaces the mass market, but personal AI amplifies the unique author.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Real Scale of What Is Happening
&lt;/h2&gt;

&lt;p&gt;The authors of articles like this one and the participants in court proceedings are discussing compensation for the past. That is understandable and humanly just. But the future is already structured differently.&lt;/p&gt;

&lt;p&gt;The real question is not legal. It is civilisational: what to do with the mass release of creative professions — just as the industrial revolution released manual labour, and digital photography released portrait photographers.&lt;/p&gt;

&lt;p&gt;History gives no grounds for panic — every such shift generated new niches, new forms of mastery, new markets. But it gives no grounds for illusion either. What is gone is gone forever.&lt;/p&gt;

&lt;p&gt;The artists filing lawsuits are fighting for compensation for the past. Those who will thrive are the ones who understand that the instrument has changed, build a personal relationship with their audience, develop techniques that cannot be averaged, and — when the moment comes — train their own model on their own material.&lt;/p&gt;

&lt;p&gt;That is not the end of craft. That is craft in its new form.&lt;/p&gt;

&lt;p&gt;And we have to live with that.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>career</category>
      <category>design</category>
      <category>discuss</category>
    </item>
    <item>
      <title>Four Trees of Technological Progress. What Usually Goes Unseen.</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Mon, 23 Mar 2026 09:24:31 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/four-trees-of-technological-progress-what-usually-goes-unseen-3akn</link>
      <guid>https://dev.to/oleg_kholin_551a551b/four-trees-of-technological-progress-what-usually-goes-unseen-3akn</guid>
      <description>&lt;p&gt;Everyone knows the game Civilization. It has a technology tree — you research writing, mathematics unlocks, then astronomy, then navigation. Clean and logical.&lt;br&gt;
But the real technological tree is more complex. Behind every node of the visible tree hide three invisible trees. And without them the main tree simply cannot grow.&lt;br&gt;
Let us go through this with concrete examples.&lt;/p&gt;

&lt;p&gt;Tree 1 — The Main Tree of Technologies&lt;br&gt;
This is what everyone sees. A new idea or scientific discovery gives birth to a new technology.&lt;br&gt;
Einstein in 1917 described the theory of stimulated emission of light. In 1960 Theodore Maiman built the first working laser. In 1903 the Wright brothers understood how to control the lift of a wing. In 1938 Otto Hahn split the uranium atom. These are all nodes of the main tree — ideas that change the world.&lt;/p&gt;

&lt;p&gt;Tree 2 — The Tree of Instruments&lt;br&gt;
Every new technology requires a new instrument for its realisation. This is the first hidden tree — the direct manifestation of the main tree.&lt;br&gt;
The laser required a ruby crystal of special purity, mirrors accurate to the nanometre, a flash lamp of strictly defined power. The Wright brothers' aircraft required a light petrol engine — a steam engine was too heavy. The atomic bomb required centrifuges for separating uranium isotopes — without them uranium enrichment is physically impossible. The transistor required monocrystalline silicon of extraordinary purity — 99.9999999%.&lt;br&gt;
It would seem that is all. There is an idea, there is an instrument for its realisation. But this is where things get truly interesting.&lt;/p&gt;

&lt;p&gt;Tree 3 — The Shadow Tree of Instruments&lt;br&gt;
The new instrument from Tree 2 cannot be created with old instruments. To produce it, entirely new auxiliary instruments are needed — not for the final product, but specifically for the creation of the main instrument.&lt;br&gt;
To build the ASML lithographic machine for chip production — special machines for polishing lenses to atomic-level precision were needed. These machines did not exist before the task arose. To build the engine for the Wright brothers' aircraft — Charles Taylor, their mechanic, constructed a special lathe for turning aluminium cylinders. The existing machines of that time did not provide the necessary precision. To build the centrifuges for uranium enrichment in the Manhattan Project — special balancing machines had to be built, because rotor imbalance at 50,000 revolutions per minute destroyed the centrifuge within seconds. To grow monocrystalline silicon for transistors — Gordon Teal at Bell Labs in 1950 built a special Czochralski apparatus with temperature control accurate to fractions of a degree. Nothing like it existed in industry at that time.&lt;/p&gt;

&lt;p&gt;Tree 4 — The Shadow Tree of Technologies&lt;br&gt;
But that is still not the bottom. To create the auxiliary instruments from Tree 3 — new technologies are needed specifically for their production. This is the fourth tree, the most invisible of all.&lt;br&gt;
To polish the lenses for ASML — a new technology of ion-beam etching of glass was required, which simply did not exist before. It was developed by the company Zeiss specifically for this task. Zeiss has existed since 1846 and today is the sole manufacturer of these lenses in the world. To balance the rotors of centrifuges in the Manhattan Project — a new technology of dynamic balancing of rotating bodies at ultra-high speeds was required. Before this, balancing had only been applied to steam turbines at speeds hundreds of times lower. To control the temperature in the monocrystalline silicon growing apparatus — new thermocouple and temperature regulator technology was required, with precision unachievable by industrial equipment of that era. To build Taylor's lathe for aluminium cylinders — a new technology of aluminium cutting was required. Aluminium was then a new metal and would "smear" ordinary cutting tools — special geometry and cutting speeds were needed.&lt;/p&gt;

&lt;p&gt;From Electronics to AI — The Same Four Trees&lt;br&gt;
Let us apply the same system to artificial intelligence — one of the most complex technological chains in the history of humanity.&lt;br&gt;
Tree 1 — The Main Tree of AI Technologies&lt;br&gt;
In 1943 mathematician Warren McCulloch and neurophysiologist Walter Pitts described a mathematical model of a neuron — a binary element that either fires or does not. This was pure theory, with no practical application.&lt;br&gt;
In 1957 Frank Rosenblatt created the perceptron — the first learning neural network. But it could only solve linearly separable problems and hit a dead end.&lt;br&gt;
In 1974 Paul Werbos described the backpropagation algorithm. The idea allowed a network to learn from its errors, adjusting weights at each layer. In 1986 Geoffrey Hinton, Rumelhart and Williams republished it and made it practical. This thawed AI research after years of "winter".&lt;br&gt;
In 1997 LSTMs appeared — Long Short-Term Memory networks that could work with sequences and retain context. But they processed text strictly one word at a time — parallel processing was impossible.&lt;br&gt;
In 2017 researchers at Google published the paper "Attention Is All You Need". The transformer architecture made it possible to process all text simultaneously rather than sequentially — each token is compared against all others through the attention mechanism. From this grew GPT, BERT, Claude and all of modern AI.&lt;br&gt;
Tree 2 — The Tree of AI Instruments&lt;br&gt;
The transformer architecture requires processing enormous matrices of numbers simultaneously — billions of matrix multiplication operations per second. A CPU is physically incapable of this — it is sequential. What was needed was a GPU — a graphics processing unit, architecturally designed for parallel computation.&lt;br&gt;
Training modern large language models requires not just GPUs, but thousands of GPUs working simultaneously. By 2024 Meta had 600,000 H100 GPUs dedicated to AI research and development. Elon Musk's startup xAI built the Colossus supercomputer with 200,000 NVIDIA H100/H200 GPUs.&lt;br&gt;
Beyond GPUs, special chips with high-speed HBM — High Bandwidth Memory — were needed, which ordinary GPUs did not have. And specialised server platforms with hundreds of gigabytes of RAM and high-speed inter-processor NVLink connections.&lt;br&gt;
Tree 3 — The Shadow Tree of AI Instruments&lt;br&gt;
GPUs for AI contain billions of transistors on an area the size of a fingernail. To manufacture them, lithographic machines with extreme ultraviolet radiation are needed — EUV lithography machines from ASML. These are machines the height of a double-decker bus, costing 150 to 200 million dollars each. Before the 2010s such machines simply did not exist — they had to be created specifically for this purpose.&lt;br&gt;
To manufacture GPU chips, ISO Class 1 cleanrooms are required — spaces in which no more than 10 particles of size 0.1 micron are present per cubic metre of air. This is millions of times cleaner than ordinary air. Creating such spaces required new filtration systems, new construction materials, new working protocols.&lt;br&gt;
To connect thousands of GPUs together into clusters for AI training, specialised high-speed InfiniBand and NVLink switches were needed — equipment that before the AI era existed only in narrowly specialised supercomputers and was not commercially available at the required scale.&lt;br&gt;
Tree 4 — The Shadow Tree of AI Technologies&lt;br&gt;
To build the ASML EUV lithography machine, a new technology for generating extreme ultraviolet light was required — a laser beam strikes a droplet of tin in a vacuum 50,000 times per second, creating a plasma at the temperature of the surface of the Sun, which emits EUV light. This technology was developed from scratch over more than 20 years through the joint efforts of ASML, Zeiss and dozens of other companies.&lt;br&gt;
To create HBM memory for GPUs, a new technology of three-dimensional chip packaging was needed — TSV, Through-Silicon Vias. This is a method of vertically connecting silicon layers through microscopic holes. Before the AI era, such packaging density was needed by no one.&lt;br&gt;
To train transformers on thousands of GPUs simultaneously, a new GPU programming technology was needed — CUDA. It was released by NVIDIA in 2007 and became central to NVIDIA's strategy of positioning the GPU as a universal tool for scientific applications. By 2015 CUDA development was increasingly focused on accelerating machine learning workloads. NVIDIA developed specialised libraries — cuDNN for deep learning and cuBLAS for linear algebra. CUDA is not simply a library. It is an entirely new technology for programming parallel computation, which did not exist before it.&lt;/p&gt;

&lt;p&gt;The Conclusion — The Correct Order of Growth:&lt;br&gt;
A step forward in the main tree of technologies actually looks like this:&lt;br&gt;
First a new idea or technology emerges. To realise it, a new main instrument is needed. But that instrument cannot be created with old instruments — new auxiliary instruments are needed. And to create the auxiliary instruments, new auxiliary technologies are needed.&lt;br&gt;
That is, the real direction of movement is reverse — first Tree 4 grows, then Tree 3, then Tree 2, and only then does Tree 1 become possible.&lt;br&gt;
For AI this looked as follows: first in 2007 CUDA appeared — a new GPU programming technology (Tree 4). This made possible the creation of next-generation GPU clusters — H100 and A100 (Tree 3). Those in turn provided the instruments for practically training transformers at industrial scale (Tree 2). And only then did the theoretical idea from the 2017 paper "Attention Is All You Need" become GPT-4, Claude and all of modern AI (Tree 1).&lt;br&gt;
Note: the transformer idea appeared in 2017. But the infrastructure for its realisation was being built from 2006 to 2020 — in parallel and independently. No one planned this as a unified programme. It simply happened that at the right moment all four trees were sufficiently grown for AI to become possible.&lt;/p&gt;

&lt;p&gt;Why This Matters&lt;br&gt;
This is precisely why genuine technological progress is so slow and expensive. Behind every visible step stand three invisible steps that had to be taken first.&lt;br&gt;
And this is precisely why when countries want to stop someone else's progress — they block not the final products, but Trees 3 and 4. They do not ban rockets or aircraft. They ban ASML lithography machines, special machine tools, precision bearings, metal alloying technologies. Whoever owns Trees 3 and 4 controls Trees 1 and 2. Always.&lt;/p&gt;

</description>
      <category>discuss</category>
      <category>learning</category>
      <category>science</category>
    </item>
    <item>
      <title>From Icon to Post-Reality: The Evolution of Ontological Synchronisation</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Sun, 22 Mar 2026 08:19:11 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/from-icon-to-post-reality-the-evolution-of-ontological-synchronisation-3k62</link>
      <guid>https://dev.to/oleg_kholin_551a551b/from-icon-to-post-reality-the-evolution-of-ontological-synchronisation-3k62</guid>
      <description>&lt;p&gt;The history of how humans engage with reality is often described as a sequence of technologies, media, or interfaces. But a more precise perspective emerges if we focus not on carriers of information, but on mechanisms of ontological synchronisation — that is, the ways in which different subjects come to share an understanding of what counts as real, existent, and actionable.&lt;/p&gt;

&lt;p&gt;From this angle, media cease to be mere channels of transmission. They become mechanisms for coordinating reality itself.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Icon as a Point of Shared Attention&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The primary level of synchronisation is associated with iconic forms — images that preserve a resemblance to what they denote. Their function is not explanatory but referential: they fix a point of shared attention.&lt;/p&gt;

&lt;p&gt;An icon does not require a shared language or a shared theory of the world. It creates a minimal common ground: the ability to point to “this” as a shared focus.&lt;/p&gt;

&lt;p&gt;However, this does not produce coinciding ontologies. Each participant retains their own interpretation. What is shared is the act of reference, not the meaning.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Interface as Operational Synchronisation&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;With the development of digital environments, iconic forms cease to be rare objects and become a universal interface layer. They no longer merely indicate; they organise action.&lt;/p&gt;

&lt;p&gt;At this point, a shift occurs from synchronisation of perception to synchronisation of operations. People begin to share not only what they see, but what they can do.&lt;/p&gt;

&lt;p&gt;Reality in an interface environment becomes operational: it is defined not by what “is”, but by what can be interacted with.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Video as Synchronisation of Time&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Video introduces a fundamentally new layer — temporal continuity. If the icon fixes an object, video fixes change.&lt;/p&gt;

&lt;p&gt;This radically intensifies the effect of reality, because it creates the illusion of a world that exists continuously over time. Yet this continuity is not a property of the world itself, but a property of the medium.&lt;/p&gt;

&lt;p&gt;Synchronisation here occurs not through objects and not through actions, but through the shared experience of a flow of events. People begin to share not only “what is”, but “how it unfolds”.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Mirror and the Splitting of the Shared Object&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;A classical mirror introduces a fundamental rupture: the same object produces different images depending on the observer’s position.&lt;/p&gt;

&lt;p&gt;This creates the first stable form of ontological divergence within apparent commonality. Everyone “sees the same thing”, yet each reconstructs a different spatial configuration.&lt;/p&gt;

&lt;p&gt;A paradox emerges: commonality is preserved at the level of the object, but breaks down at the level of experiential structure.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Algorithmic Mirror and the Personalisation of Reality&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;With the emergence of algorithmic systems, reflection ceases to be neutral. It becomes computed, adaptive, and predictive.&lt;/p&gt;

&lt;p&gt;Thus appears the controlled mirror: the system does not simply display reality but constructs its representation based on a model of the user, their behaviour, and context.&lt;/p&gt;

&lt;p&gt;This produces a shift: a shared reality fragments into personalised versions, each internally coherent and functional, but not required to coincide with others.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Virtual Reality as Coordinated Simulation&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In multi-user virtual environments, it becomes necessary to maintain consistency between participants. The world is no longer given in advance; it is computed as a shared result of interaction.&lt;/p&gt;

&lt;p&gt;However, this commonality does not imply identical perception. It implies coordinated action within a shared simulation.&lt;/p&gt;

&lt;p&gt;Reality becomes protocol-based: it exists insofar as the system maintains non-contradictory interaction among participants.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;AI as an Active Mirror and Generator of Ontology&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;With the introduction of AI, a qualitative shift occurs. The system ceases to be a passive reflector or a fixed simulator. It becomes an active generator of the environment.&lt;/p&gt;

&lt;p&gt;The AI mirror does not simply show the world — it constructs it. This construction depends on the model of the subject: the system takes into account behaviour, preferences, and responses, forming an adaptive ontology.&lt;/p&gt;

&lt;p&gt;A crucial point: a feedback loop emerges. The user does not merely perceive the world; they become an input into its generation.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Post-Reality as a Mode of Coordinated Generation&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;As a result, a condition emerges that can be described as post-reality.&lt;/p&gt;

&lt;p&gt;Its key feature is not the disappearance of reality, but the disappearance of a single external ground against which “true” and “false” could be distinguished.&lt;/p&gt;

&lt;p&gt;Reality is no longer given as an external object. It is continuously produced — through interaction, synchronisation, and computation.&lt;/p&gt;

&lt;p&gt;The criterion shifts from correspondence to the world to the stability and operability of the generated ontology within the system of interactions.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Conclusion: From Shared Object to Distributed Ontology&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The evolution of ontological synchronisation follows a последовательный shift:&lt;/p&gt;

&lt;p&gt;from a shared object that can be pointed to&lt;br&gt;
to a shared interface through which action is performed&lt;br&gt;
to a shared flow of events that can be experienced&lt;br&gt;
to personalised reflection that shapes experience&lt;br&gt;
and further to active generation of reality adapted to interaction&lt;/p&gt;

&lt;p&gt;In this perspective, reality ceases to be a stable stage. It becomes a process of coordination — a dynamic system in which ontologies do not simply coexist but are continuously produced and recalibrated.&lt;/p&gt;

&lt;p&gt;Post-reality, in this sense, is not the end of the distinction between truth and illusion, but a transition to a regime in which that distinction is no longer fundamental.&lt;/p&gt;

</description>
      <category>discuss</category>
      <category>learning</category>
      <category>systems</category>
      <category>ux</category>
    </item>
    <item>
      <title>Robots and Production: How Not to Freeze System Evolution</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Sun, 15 Mar 2026 08:08:23 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/robots-and-production-how-not-to-freeze-system-evolution-5h1l</link>
      <guid>https://dev.to/oleg_kholin_551a551b/robots-and-production-how-not-to-freeze-system-evolution-5h1l</guid>
      <description>&lt;p&gt;We live in an era of robotic hype. Humanoids perform backflips in videos, "smart" manipulators pour water, investors pour money into promises of the future. But behind all this hides a fundamental problem that few notice.&lt;br&gt;
The Investment Race, Not Engineering Necessity&lt;br&gt;
The modern race for anthropomorphic robots is not about technological progress. It's about investments. Whoever promises more gets more. A video of a robot doing a backflip or "pouring water" is not an MVP (minimum viable product), but an MVD (minimum viable demonstration) for attracting capital.&lt;br&gt;
Investors react to anthropomorphic gestures — a smile, a nod, "natural" movements — not to metrics of uptime or production cycle cost. The cost of developing a humanoid at $50-100 million is recouped not through sales (the industrial humanoid market is practically zero in 2024-2025), but through increased market capitalization of the parent company, grants for the "future of work," and selling a technological image.&lt;br&gt;
Anthropomorphism today is not a design error, but a rational strategy in a distorted market where what's measured is not functionality, but virality.&lt;br&gt;
Robot as a "Profile in Search of a Task"&lt;br&gt;
But there's a deeper problem. A robot is created as a profile, but there's no narrative explaining it. And this is because there's no context for it. The robot is being forced into something, rather than building a process around the robot.&lt;br&gt;
Here's an example. A robot can expertly insert rubber insulation into windows of new cars. Great! But who unpacks them and places them on the table? Another robot? Or a human?&lt;br&gt;
This is a perfect illustration of the point automation syndrome. They automate the "headache" — the complex operation, ignoring the "tail" tasks: logistics, preparation, cleanup. The result is a robot-island requiring human servicing at the input and output. This is not process automation, but delegation of a single movement.&lt;br&gt;
Freezing Production Dynamics&lt;br&gt;
But you're missing another critically important point. Implementing robots is freezing the dynamics and mechanics of embedded production for several years!&lt;br&gt;
With a human as the operational unit, the operational implementation of new materials, components, and technologies happened relatively quickly. A human operator is cognitively and operationally flexible. But with an implemented robot, new operations as mechanics can still be implemented, but how do you implement a new complex interaction of the robot with new materials or environments? This is simply a quiet horror for technologists!&lt;br&gt;
A human is a universal interface to uncertainty. Their cognitive and motor plasticity allows them to absorb environmental variability without changing the "hardware." New material? A human felt it with their hands — understood how to handle it (seconds or minutes). A robot requires reprogramming, testing, calibration (days or weeks). Change in part geometry? A human adapted their grip "on the fly." A robot needs a new gripper, possibly a new manipulator. A new operation in the cycle? A human learns in 1-2 shifts. A robot needs trajectory redesign, collision checking, safety protocols.&lt;br&gt;
The cost of adaptation for a human is salary and training time. For a robot — engineering hours, line downtime, possible retooling.&lt;br&gt;
When you implement a robot on a production line, you're not just automating an operation — you're fixing the process ontology. Before the robot: Material → Human → Product (flexibility at the operator level). After the robot: Material → Robot version 1.0 → Product (change = project).&lt;br&gt;
The consequences are catastrophic. The innovation paradox: the more you invest in automation, the more expensive any change becomes. This creates inertia against implementing improvements. An example from the auto industry: a plant fully robotized for Model X cannot quickly switch to Model Y without multi-million dollar investments in retooling. Meanwhile, an assembly line with humans adapts in weeks.&lt;br&gt;
The "technological cocoon" effect: the robot becomes a hostage to its own efficiency — it's perfect for the current task, but vulnerable to future changes. Robotization is insurance against current costs at the price of losing future flexibility.&lt;br&gt;
Efficiency vs. Evolution: Antagonism by Definition&lt;br&gt;
Efficiency is convergence. Direction toward one optimal point, minimization of deviations and variability, specialization, standardization, optimization for current conditions. Result — peak performance in a narrow window of conditions. Analogy — a cheetah, the fastest runner, but only on flat savanna.&lt;br&gt;
Evolution is divergence. Direction — expanding the space of possibilities, metric — preserving variability and ability to find new solutions, strategy — plasticity, redundancy, "backup" paths. Result — survival in uncertain and changing conditions. Analogy — a rat, not the fastest, not the strongest, but surviving everywhere.&lt;br&gt;
Key insight: maximum efficiency equals maximum vulnerability to changes. This is not an opinion — it's a law from systems theory and evolutionary biology. The more a system is optimized for specific conditions, the more expensive any deviation from those conditions becomes.&lt;br&gt;
The "freezing" mechanism is simple. Production with humans: Material → Human → Product, human adapts to new things in hours or days, in 6 months new materials, technologies, products — system evolution continues. Production with robots: Material → Robot version 1.0 → Product, fixed logic, trajectories, grippers, in 6 months new materials require reprogramming, STOP, project, budget, engineers, line downtime, system evolution frozen for months.&lt;br&gt;
Cost of adaptation as a barrier. New supplier with different packaging: with human plus 1 day of training, with robot plus 2 weeks of project and $50,000. New material for a part: with human plus 1 shift of adaptation, with robot plus 1 month of retooling and $200,000. New operation in the cycle: with human plus 2 days of training, with robot plus 3 months of development and $500,000.&lt;br&gt;
Result: every change becomes economically painful, creating inertia against innovations.&lt;br&gt;
Biological Metaphor: Why Dinosaurs Went Extinct&lt;br&gt;
The "efficient giant" scenario. Era of stability (100 million years): dinosaur huge, efficient, dominant, diet optimized for specific plants, environment stable climate, predictable conditions, result peak efficiency in its niche. Era of changes (meteorite): change in climate, vegetation, ecosystem, dinosaur too big, too specialized, adaptation impossible (slow reproduction, narrow diet), result extinction.&lt;br&gt;
Survivors: small mammals inefficient but plastic, diet omnivorous, adaptive, environment can live in burrows, change behavior, result evolutionary success → human.&lt;br&gt;
Transfer to production: dinosaur = fully robotized line for one model, mammal = hybrid system with human-adaptor, meteorite = new material, sanctions, market change, crisis.&lt;br&gt;
Strategy "Stable Routine + Flexible Reserve"&lt;br&gt;
But there's a way out. Let's return to the idea of atomizing the production process. Identifying rarely-changing routine — arranging boxes on shelves, size changes but shelves don't — optimizing a robot for it. In the end, we know this won't change much for another 7 years. Like in Amazon warehouses!&lt;br&gt;
A robot should not serve as a red line for mass production! It's for lowering costs through increasing work volume! If at Ferrari production a robot stopping means the entire plant stops, then at a regular plant a human can also install the sealant!&lt;br&gt;
This is a practical doctrine of reasonable robotization. Each operation is evaluated on two axes: stability (how often does this operation change?) and criticality (what happens if it stops?).&lt;br&gt;
The decision matrix for robotization looks like this. Stable operation, non-critical (there's a reserve) — ideal zone for a robot. Example: warehouses, sorting, stacking. Stable operation, critical (stoppage = line stop) — robot plus mandatory reserve (second robot or possibility of manual work). Dynamic operation, non-critical — hybrid system, robot plus human backup. Dynamic operation, critical — human or hybrid system, a robot here creates too many risks.&lt;br&gt;
Amazon example: arranging boxes on shelves — stable plus non-critical operation. Ideal zone for a robot. Even if Kiva breaks down, a human can temporarily take over the operation (slower, but without stopping). Success factors: operation stability (arranging boxes doesn't change for 10+ years), standardization (all boxes → one gripper type), scale (thousands of robots → scale effect), redundancy (if one robot breaks — neighboring one covers the zone), human factor (humans on picking and exceptions). Result: robots lower costs through volume, not through uniqueness.&lt;br&gt;
Ferrari example: robot-welder in a unique line — stable plus critical operation. Problems: operation stability low (each model is unique → robot tied to specific program), standardization low (each part is unique), scale small (10-20 thousand cars per year), redundancy minimal (robot = bottleneck), human factor (humans cannot quickly replace the robot due to specialization). Result: robot improves quality, but creates a failure point. This is justified for a premium brand, but catastrophic for mass production.&lt;br&gt;
Practical Algorithm for Robot Implementation&lt;br&gt;
Step one: decomposition of the process into atomic operations. Example: installing a sealant in a car window. Operation 1.1 — getting the sealant from the magazine (stable). Operation 1.2 — positioning the sealant (stable). Operation 1.3 — inserting into the groove (stable). Operation 1.4 — checking the fit (dynamic — depends on the batch). Operation 1.5 — correction for deviation (dynamic — exceptions).&lt;br&gt;
Step two: evaluation on two axes. Operation 1.1 — stability high, criticality low, decision robot. Operation 1.2 — stability high, criticality low, decision robot. Operation 1.3 — stability high, criticality medium, decision robot plus reserve. Operation 1.4 — stability low, criticality high, decision human. Operation 1.5 — stability low, criticality high, decision human.&lt;br&gt;
Step three: designing redundancy. Stable operation with robot: robot version 1.0 works 95% of the time, upon failure or maintenance a human temporarily takes over the operation, the line continues to work (30% slower, but without stopping).&lt;br&gt;
Step four: economic calculation. Total robotization: investment $5 million, efficiency plus 50%, flexibility 0, downtime risk $100,000 per hour, payback period 4 years. Selective robotization (our approach): investment $1.5 million, efficiency plus 30%, flexibility high, downtime risk $30,000 per hour, payback period 1.5 years.&lt;br&gt;
Selective robotization gives a smaller peak of efficiency, but greater resilience to changes. This is not a compromise — it's a strategic choice.&lt;br&gt;
Why Don't Everyone Do This? Barriers&lt;br&gt;
Barrier one: investment hype. Investors want a "fully automated plant" — this sells better than "robots where it's stable." Our approach is less viral, but more sustainable.&lt;br&gt;
Barrier two: hierarchy of engineering thinking. Engineers love "clean solutions" — a fully automated line looks more elegant than "robot here, human there." Our approach requires systemic thinking, not engineering perfectionism.&lt;br&gt;
Barrier three: organizational inertia. After implementing a robot, management doesn't want to admit that a "Plan B" with a human is needed — this seems like a retreat. In reality, it's insurance.&lt;br&gt;
Three Principles of Reasonable Robotization&lt;br&gt;
Principle one: robot where it's stable. Not "where it's complex," but "where it doesn't change." Amazon warehouses work because boxes don't change. If unique shapes arrived every day — the system would collapse.&lt;br&gt;
Principle two: a robot should not be a "red line." If a robot stopping equals the entire production stopping, you've created a failure point, not efficiency. There should always be Plan B — a human or backup module.&lt;br&gt;
Principle three: robot equals cost reduction through volume. Not through "intelligence," not through "anthropomorphism," but through scalable routine. If a robot doesn't allow increasing volume or reducing unit cost — it doesn't pay off.&lt;br&gt;
Final Formulation&lt;br&gt;
Robotization is not a war between humans and machines. It's a search for optimal role distribution: the machine takes stable routine, the human preserves flexibility. The winner is not the one who automated more, but the one who preserved the ability to evolve.&lt;br&gt;
We build systems for increasing efficiency, but lose the ability to evolve. We build "dinosaurs" and call it progress. But evolution rewards not the most efficient, but the most plastic.&lt;br&gt;
The true challenge of future engineering is not to create the perfect robot, but to design a system that can evolve without self-destruction. Robotization is not an ideology, it's a tool. And like any tool, it should be applied where it's truly useful, not where it looks impressive.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>automation</category>
      <category>discuss</category>
      <category>startup</category>
    </item>
    <item>
      <title>GitHub for Civilizational Expansion: From an Antarctic Train to the Ontology of Autonomous Infrastructure</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Fri, 13 Mar 2026 20:44:52 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/github-for-civilizational-expansion-from-an-antarctic-train-to-the-ontology-of-autonomous-1of1</link>
      <guid>https://dev.to/oleg_kholin_551a551b/github-for-civilizational-expansion-from-an-antarctic-train-to-the-ontology-of-autonomous-1of1</guid>
      <description>&lt;p&gt;I. The Problem of Environment&lt;br&gt;
Antarctica is an environment of absolute extremes. Temperatures drop to −60°C and below, wind loads reach destructive levels, and the snow-ice surface conceals cracks, cavities, and zones of loose snow beneath it. Infrastructure is virtually nonexistent, and distances between points are measured in hundreds and thousands of kilometers.&lt;br&gt;
Historically, mobility in these conditions has relied on two types of solutions: aviation and ground convoys of tracked tractors. Both approaches are fundamentally limited. Aviation depends on weather conditions and cannot carry heavy payloads. Tractor convoys depend on fuel, lack integrated scientific infrastructure, are vulnerable to snowdrift burial during extended stops, and require constant surface reconnaissance due to hidden crevasses.&lt;br&gt;
Existing solutions have reached their ceiling. A paradigm shift is required.&lt;/p&gt;

&lt;p&gt;II. The Modular Train as a New Paradigm&lt;br&gt;
The response to the limitations of existing systems is the concept of a modular autonomous train — not a vehicle, but a mobile infrastructural platform.&lt;br&gt;
The architecture rests on several principles. Each module is equipped with its own tracked undercarriage, which distributes traction and ensures fault tolerance: the failure of one module does not stop the system. Flexible couplings with longitudinal and lateral articulation allow the train to navigate uneven terrain. A system of retractable support struts with heating elements lifts the train above the surface during stops — similar to polar research buildings mounted on stilts — eliminating burial by snowdrift and freezing of the tracks to the ice.&lt;br&gt;
Inside, the train accommodates living quarters, laboratory units, communication and computing centers, storage and technical sections. The result is a mobile scientific station with an integrated transport function — not a vehicle carrying scientific equipment, but a scientific platform that moves.&lt;br&gt;
The key element is an energy module with high autonomy, providing continuous power to traction drives, heating systems, scientific equipment, and surface reconnaissance systems without dependence on external fuel supplies.&lt;/p&gt;

&lt;p&gt;III. The Technological Foundation Already Exists&lt;br&gt;
Approximately 85% of the necessary components exist right now — across different industries, dispersed, but in ready or easily adaptable form.&lt;br&gt;
Tracked platforms for extreme conditions are manufactured by the mining industry. Cold-resistant steels rated to −60°C are produced in series. Thermal insulation solutions for polar conditions have been proven on existing stations. GPR ground-penetrating radar is used in geological prospecting. Starlink already provides communications at several Antarctic stations.&lt;br&gt;
Tesla has opened patents on battery pack architecture, thermal management at low temperatures, and power electronics. SpaceX has effectively trained a generation of engineers through the transparency of its technical presentations and detailed post-mortems of every failure. The predictive control algorithms developed for Falcon 9 landings are directly applicable to platform stabilization on uneven surfaces.&lt;br&gt;
The challenge is not inventing components from scratch — it is integrating what already exists. No one has done this yet.&lt;/p&gt;

&lt;p&gt;IV. AI as the Central Nervous System&lt;br&gt;
The volume of data generated by such a platform cannot be processed manually. Ground-penetrating radars produce terabytes continuously. Pressure, temperature, wind, and vibration sensors form hundreds of parallel channels.&lt;br&gt;
The AI core does not merely process data — it predicts. A radar detects a crevasse 200 meters ahead, and the system calculates a route around it before the hazard becomes a threat. Analysis of ice surface patterns allows risk zones to be anticipated before they are reached.&lt;br&gt;
The architecture is built on the principle of majority logic: three independent sensors operate simultaneously and vote among themselves. If two out of three agree — that is the accepted value. Not a backup sensor in case of failure, but three independent sources operating in parallel as a permanent operational norm.&lt;br&gt;
A critical requirement: the system continuously explains itself to the operator even during routine operation. It does not simply act — it maintains an accurate picture of what is happening for the human in the loop. Without this, the automation paradox emerges: the better the AI performs, the less capable the operator becomes at the moment when the system does eventually make an error.&lt;/p&gt;

&lt;p&gt;V. Antifragile Interfaces&lt;br&gt;
Antifragility applied to interfaces means: the system does not merely withstand stress — it improves because of it.&lt;br&gt;
The failure of a sensor unit does not cause a system collapse — the load is redistributed to adjacent nodes. The interface does not crash; it simplifies to the minimum necessary. In normal operation — a full interface with analytics. In a critical situation — automatic narrowing to three to five key parameters. Operator cognitive overload at the moment of crisis is eliminated by architecture, not by training.&lt;br&gt;
Every abnormal situation improves the system's behavioral model. A platform that has survived a blizzard knows more about blizzards than it did before. The knowledge base accumulates directly in the field, not in a laboratory.&lt;/p&gt;

&lt;p&gt;VI. The Chrome Philosophy: Changing the Paradigm&lt;br&gt;
When Google launched Chrome in 2008, the browser seemed like a solved problem. The team reframed the foundational assumption: not "browser as application" — but "browser as operating system." Each tab is a separate process. The crash of one does not bring down the rest. Antifragility implemented architecturally.&lt;br&gt;
The same logic applies here. Not an improvement of existing convoys — a redefinition of what polar infrastructure is. Not a vehicle carrying scientific equipment — a mobile scientific corridor generating continuous longitudinal profiles of ice cover, atmosphere, and subsurface structures across thousands of kilometers.&lt;br&gt;
The development methodology follows the same philosophy: a small team with high autonomy, rapid iterations, testing to failure, analysis, next version.&lt;/p&gt;

&lt;p&gt;VII. Abstracting Every Node&lt;br&gt;
The decisive step is abstracting each functional node from the specific environment in which it operates.&lt;br&gt;
What is being built is not "a radar for ice." It is "a universal forward obstacle interface." The input can be any type of surface data: GPR radar for ice, LiDAR for rocks and sand, thermal imaging for marshland, seismic sensors for unstable ground. The output is always the same: a unified traversability map. The core does not know which sensor is active. It receives the map and makes a decision.&lt;br&gt;
What is being built is not "protection against −60°C." It is "thermal homeostasis of a module." The environmental parameters change — the logic of maintaining internal conditions remains universal.&lt;br&gt;
What is being built is not "an energy system for Antarctic conditions." It is "an abstract power source": a nuclear reactor, a diesel generator, solar panels in the Sahara, hydrogen fuel cells — the output is always a stabilized power flow.&lt;br&gt;
This is a Hardware Abstraction Layer — not for a single device, but for a physical platform in its entirety. Windows does not know what hardware is inside it — it operates through an abstract driver layer. The same principle is applied here to mobile infrastructure in any hostile environment.&lt;br&gt;
Today Antarctica. Tomorrow the Sahara. The day after, the Moon. The core of the platform does not change. The adapters do.&lt;/p&gt;

&lt;p&gt;VIII. An Ecosystem, Not a Product&lt;br&gt;
The train is one element of a broader system.&lt;br&gt;
The concept includes an autonomous base with an AI decision-making core, robotic manipulators, and 3D printers capable of producing structures from local materials. The train delivers data and mechanical constructions to the base. The base autonomously deploys and begins operation. Once sufficient resources have been accumulated, it deploys the next node. The train carries the seed kit for the new base.&lt;br&gt;
This is a replication protocol. Self-reproducing infrastructure with minimal human involvement.&lt;br&gt;
In this logic, Antarctica is not the destination — it is a proving ground with one key advantage over space: it can be reached and tested under real conditions. Every failure, every engineering decision, every adaptation is real data, not simulation.&lt;br&gt;
Earth → Antarctica (proof of concept)&lt;br&gt;
      → Moon&lt;br&gt;
      → Mars&lt;br&gt;
      → Asteroid Belt&lt;br&gt;
The logic is the same as SpaceX applied: not flying to Mars immediately, but first learning to land a rocket on a barge in the ocean. The Antarctic train is that barge. The first validated node in a long chain.&lt;/p&gt;

&lt;p&gt;IX. The GitHub Precedent&lt;br&gt;
Before GitHub, code existed — but it was closed, fragmented, and incompatible across teams and organizations. GitHub did not invent git. What was created was a social infrastructure around an existing tool. The result: any team anywhere on the planet takes any code, adapts it, and returns the improvement to the system.&lt;br&gt;
The same logic applies to physical platforms. A train is not being invented. A social and technical infrastructure is being created around the ontology of autonomous expansion.&lt;br&gt;
A verified module enters the open standard. Another team takes it, adapts it to their environment, and returns the improvement. The community verifies it. The solution enters the main branch.&lt;br&gt;
Currently, NASA is solving the problem of autonomous deployment. ESA is solving the same problem independently. Other organizations are solving the same problem, independently, expending resources in parallel. A shared platform means the problem is solved once, verified under real conditions, and made available to all subsequent projects.&lt;br&gt;
The network effect operates here exactly as it does in software: more verified modules make the platform more valuable, which attracts more teams, which raises the quality of modules. The threshold beyond which major organizations can no longer ignore the standard is reached organically.&lt;/p&gt;

&lt;p&gt;X. A Repository for Physical Infrastructure&lt;br&gt;
GitHub changed the speed at which knowledge accumulates in software. Every solution found once stopped being lost — it remained in the system and became available to everyone who came after.&lt;br&gt;
The same architecture is applicable to physical infrastructure for hostile environments. Every expedition. Every failure. Every engineering solution. Every adaptation of a module to new conditions — remains in the system and informs every subsequent project.&lt;br&gt;
Antarctica is the first commit. Not because it is a compelling metaphor, but because it is where real verified data on the operation of autonomous self-replicating infrastructure in an extreme environment will be obtained for the first time. Everything that follows is built on that foundation.&lt;/p&gt;

</description>
      <category>automation</category>
      <category>github</category>
      <category>opensource</category>
      <category>science</category>
    </item>
    <item>
      <title>The Economics of Limits: From Expansion to the Cycle. The Financialisation of Recycling as a New Growth Paradigm</title>
      <dc:creator>oleg kholin</dc:creator>
      <pubDate>Wed, 11 Mar 2026 18:58:40 +0000</pubDate>
      <link>https://dev.to/oleg_kholin_551a551b/the-economics-of-limits-from-expansion-to-the-cycle-the-financialisation-of-recycling-as-a-new-fk0</link>
      <guid>https://dev.to/oleg_kholin_551a551b/the-economics-of-limits-from-expansion-to-the-cycle-the-financialisation-of-recycling-as-a-new-fk0</guid>
      <description>&lt;p&gt;I. The End of the Frontier&lt;br&gt;
The historical logic of capitalism was built on expansion. Three successive waves drove the system forward: geographical expansion opened new markets, demographic growth delivered new consumers, and technological revolutions created entirely new categories of need. Each time one wave subsided, the system found the next.&lt;br&gt;
The transfer of manufacturing to Asia throughout the 1980s and 1990s was not merely about cheap labour — it was a mechanism for financing the computer revolution and the rollout of the internet. Low-cost production subsidised expensive innovation. Billions of people were absorbed into the global economy simultaneously as producers and as consumers. This is the classical mechanics of frontier expansion.&lt;br&gt;
Today, that mechanism is spent. The territory has been settled, basic needs in developed economies are saturated, and demographic growth is concentrated precisely where purchasing power is weakest. The system has run into its own ceiling. And it is here that economics becomes genuinely interesting.&lt;/p&gt;

&lt;p&gt;II. The Monetisation of Repeatability&lt;br&gt;
When a market can no longer expand outward, it begins expanding inward. This does not mean creating new needs — it means fragmenting existing ones into billable microevents. Hence the logic of the subscription economy, hence hardware-as-a-service, hence the printer that locks up without a proprietary cartridge.&lt;br&gt;
A kettle on a subscription model is neither absurd nor a caricature. It is the next logical step along a path already travelled by smartphones, automobiles, and software. The device becomes cheaper; control over its usage cycle remains with the manufacturer. A heating element designed to last precisely fifty boiling cycles is not a design flaw. It is a predictable cash flow engineered directly into the hardware.&lt;br&gt;
The operative word here is repeatability. Not product quality, but the reproducibility of the consumption process. The system values not how good the kettle is, but how predictably it will be replaced, updated, and repurchased. This is precisely why downloadable "floral tea brewing profiles" represent not irony but a perfectly viable monetisation layer built atop a physically stagnant product. The experience economy superimposes itself upon the object economy once the object has exhausted its growth potential.&lt;br&gt;
Yet this model carries its own ceiling. Total subscription dependency breeds existential resistance in the consumer. When everything becomes a rental, the question arises: what does a person actually own? This is not philosophical abstraction — it is a genuine social fracture, already visible in the growing appetite for repairable, autonomous, offline objects. The counter-movement of "Kettle 1.0" is not nostalgia; it is a rational response to fatigue with managed consumption.&lt;/p&gt;

&lt;p&gt;III. The Asymmetry of Crisis&lt;br&gt;
The conventional view of crisis as a "reset of Maslow's hierarchy" is an appealing but imprecise metaphor. War, epidemic, and imperial collapse have historically zeroed out the level of needs and launched new cycles of growth. Schumpeter's creative destruction functioned in a world of relatively low systemic interconnection.&lt;br&gt;
Today, crisis does not reset the system uniformly. It stratifies it.&lt;br&gt;
A man drinking coffee on the Cannes waterfront, watching the glow of fighting above the Libyan horizon over the Mediterranean, is not a metaphor for cynicism. It is an accurate description of the architecture of modern risk. The periphery absorbs the blow. The centre continues to consume. Capital flows toward safe jurisdictions. Destruction does not destroy the system — it redistributes its resources.&lt;br&gt;
At the same time, crisis generates differentiated demand for security: the poor purchase window bars and reinforced doors, the middle classes expand their insurance coverage, and the wealthy invest in private security and asset diversification. Demand does not disappear — it transforms according to purchasing power. The security market grows precisely when the threat grows.&lt;br&gt;
The Yugoslav example illustrates this mechanism in its most extreme form: the destruction of legitimate distribution channels immediately spawns shadow markets where pricing is determined by scarcity rather than production. Donor organs, medicines, foodstuffs — all become "war assets" wherever the state ceases to control the market. The economic boundary of permissibility runs precisely where resources cease to be universally accessible and become instruments of power.&lt;/p&gt;

&lt;p&gt;IV. Geopolitics as Logistics&lt;br&gt;
Strikes against Iran, the new trade corridor from India, proxy conflicts in Pakistan and Afghanistan — all of this reads not as ideological confrontation but as a struggle for control over supply chains. The global system orients itself around macro-resources and logistics; local human consequences are classified as collateral damage, unaccounted for in the market price of the primary asset.&lt;br&gt;
The new India–Middle East–Europe trade route is not simply an alternative logistics arrangement. It is an attempt to escape direct confrontation — to establish a corridor independent of instability zones, to secure supply predictability without betting on high-risk regions. This is precisely why anything threatening this corridor comes under pressure: not for ideological reasons, but for purely economic ones.&lt;br&gt;
China understands this. Its response is unpredictable for precisely that reason — it recognises that the new route is not merely a competitor's logistics play, but an attempt to architect global trade either without China or alongside it on unfavourable terms. The India-Pakistan escalation, with Chinese air defence systems deployed on Pakistan's behalf, is no longer a regional conflict. It is a signal marking where the real lines of strategic interest lie.&lt;br&gt;
Yet it is here that Western strategy carries an internal contradiction. Simultaneously financing new corridors and sustaining military pressure is a race that exhausts faster than it pays back. Competition with China at a global scale does not kill through a single blow, but through the accumulated burden on public finances, defence budgets, and technology investment. Every new route requires security guarantees. Every security guarantee requires resources. The circle closes upon itself.&lt;/p&gt;

&lt;p&gt;V. The Financialisation of Recycling as a Response to the Limit&lt;br&gt;
If expansion is no longer possible, and crisis merely redistributes rather than resets — where is the next source of growth to be found?&lt;br&gt;
The answer that emerges from this line of reasoning sounds counterintuitive: in recycling. Not in the sense of waste processing as an environmental project, but in the financialisation of an object's entire lifecycle from beginning to end.&lt;br&gt;
The IBM PC example is instructive here. The chassis and motherboard remain constant — everything else changes. This is not nostalgia for modular architecture. It is the description of an economic model in which value is created not through the production of the new, but through the controlled replacement of components at a predictable rhythm. Each replacement is a transaction. Each transaction is revenue. The entire lifecycle of an object becomes a financial instrument.&lt;br&gt;
This is fundamentally distinct from mere "quality." Repeatability outweighs perfection. A system capable of reproducing the replacement cycle with predictable precision is worth more than one that produces a flawless product a single time. This is precisely why the financialisation of recycling is not an environmental agenda. It is the next growth model in a world of saturated markets.&lt;br&gt;
The transfer of manufacturing to Asia financed the computer revolution. The financialisation of recycling may finance the next one — whatever form it takes. Closed-loop cycles, modular products, subscriptions for physical objects, digital lifecycle management of individual components — all of this already exists in fragments. The task is to assemble it into a coherent systemic model.&lt;br&gt;
China, still thinking in terms of linear export, risks falling behind here. The advantage will belong to those who first build the infrastructure of closed cycles with financial control at every node — not as a by-product of environmental policy, but as the primary business model.&lt;/p&gt;

&lt;p&gt;VI. Conclusion: Economics Without a Frontier&lt;br&gt;
We are moving into a world where growth no longer means "more." It means "deeper," "more precise," "more repeatable." Insurance transforms into predictive risk management. Manufacturing transforms into lifecycle service. War transforms into logistical pressure. Crisis transforms into a permanent background condition against which some monetise anxiety, others monetise security, and others still monetise the wreckage.&lt;br&gt;
This is neither pessimism nor an apologia for cynicism. It is a description of a system that has exhausted extensive growth and is compelled to invent intensive growth instead. History demonstrates that such transitions are painful, uneven, and invariably produce new winners — those who understood earliest in which direction value now lies.&lt;br&gt;
Today, that direction points inward along the cycle, not beyond its edges.&lt;/p&gt;

</description>
      <category>discuss</category>
      <category>learning</category>
      <category>science</category>
    </item>
  </channel>
</rss>
