<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Sergii Miasoiedov</title>
    <description>The latest articles on DEV Community by Sergii Miasoiedov (@mova).</description>
    <link>https://dev.to/mova</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3825766%2F65ce5b99-0266-44c5-8113-de24a07dc7c2.png</url>
      <title>DEV Community: Sergii Miasoiedov</title>
      <link>https://dev.to/mova</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/mova"/>
    <language>en</language>
    <item>
      <title>https://medium.com/@s.myasoedov81/ai-executes-no-one-signed-d2f143ef6aee</title>
      <dc:creator>Sergii Miasoiedov</dc:creator>
      <pubDate>Thu, 19 Mar 2026 12:05:17 +0000</pubDate>
      <link>https://dev.to/mova/httpsmediumcomsmyasoedov81ai-executes-no-one-signed-d2f143ef6aee-526g</link>
      <guid>https://dev.to/mova/httpsmediumcomsmyasoedov81ai-executes-no-one-signed-d2f143ef6aee-526g</guid>
      <description>&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
      &lt;div class="c-embed__body flex items-center justify-between"&gt;
        &lt;a href="https://medium.com/@s.myasoedov81/ai-executes-no-one-signed-d2f143ef6aee" rel="noopener noreferrer" class="c-link fw-bold flex items-center"&gt;
          &lt;span class="mr-2"&gt;medium.com&lt;/span&gt;
          

        &lt;/a&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
    </item>
    <item>
      <title>Decision Stack</title>
      <dc:creator>Sergii Miasoiedov</dc:creator>
      <pubDate>Wed, 18 Mar 2026 06:35:52 +0000</pubDate>
      <link>https://dev.to/mova/decision-stack-4dhe</link>
      <guid>https://dev.to/mova/decision-stack-4dhe</guid>
      <description>&lt;p&gt;✖ ✖ ✖ ✖ ✖ ✖ ✖&lt;/p&gt;

&lt;p&gt;topic: decision_stack&lt;/p&gt;

&lt;p&gt;Lately I’ve been following Cristiano Messina’s “Decision Stack” series.&lt;/p&gt;

&lt;p&gt;A very interesting and professional perspective&lt;br&gt;
on how modern systems should be designed.&lt;/p&gt;

&lt;p&gt;The model is simple but powerful:&lt;/p&gt;

&lt;p&gt;signals → context → objectives → constraints → execution → feedback&lt;/p&gt;

&lt;p&gt;Where:&lt;/p&gt;

&lt;p&gt;signals detect change&lt;br&gt;&lt;br&gt;
context defines meaning&lt;br&gt;&lt;br&gt;
objectives define what matters&lt;br&gt;&lt;br&gt;
constraints define boundaries&lt;br&gt;&lt;br&gt;
execution performs actions&lt;br&gt;&lt;br&gt;
feedback drives learning  &lt;/p&gt;

&lt;p&gt;✖&lt;/p&gt;

&lt;p&gt;A question to practitioners:&lt;/p&gt;

&lt;p&gt;Which of these layers&lt;br&gt;
are already explicitly implemented&lt;br&gt;
in your systems?&lt;/p&gt;

&lt;p&gt;And in what form?&lt;/p&gt;

&lt;p&gt;Are they:&lt;/p&gt;

&lt;p&gt;centralized&lt;br&gt;&lt;br&gt;
scattered&lt;br&gt;&lt;br&gt;
implicit&lt;br&gt;&lt;br&gt;
formalized  &lt;/p&gt;

&lt;p&gt;✖&lt;/p&gt;

&lt;p&gt;Curious how others structure this today.&lt;/p&gt;

&lt;p&gt;✖ ✖ ✖ ✖ ✖ ✖ ✖&lt;/p&gt;

</description>
      <category>ai</category>
      <category>devops</category>
      <category>devto</category>
      <category>automation</category>
    </item>
    <item>
      <title>Decision Stack</title>
      <dc:creator>Sergii Miasoiedov</dc:creator>
      <pubDate>Wed, 18 Mar 2026 06:15:05 +0000</pubDate>
      <link>https://dev.to/mova/decision-stack-4gj8</link>
      <guid>https://dev.to/mova/decision-stack-4gj8</guid>
      <description>&lt;p&gt;✖ ✖ ✖ ✖ ✖ ✖ ✖&lt;/p&gt;

&lt;p&gt;topic: decision_stack&lt;/p&gt;

&lt;p&gt;Lately I’ve been following Cristiano Messina’s “Decision Stack” series.&lt;/p&gt;

&lt;p&gt;A very interesting and professional perspective&lt;br&gt;
on how modern systems should be designed.&lt;/p&gt;

&lt;p&gt;The model is simple but powerful:&lt;/p&gt;

&lt;p&gt;signals → context → objectives → constraints → execution → feedback&lt;/p&gt;

&lt;p&gt;Where:&lt;/p&gt;

&lt;p&gt;signals detect change&lt;br&gt;&lt;br&gt;
context defines meaning&lt;br&gt;&lt;br&gt;
objectives define what matters&lt;br&gt;&lt;br&gt;
constraints define boundaries&lt;br&gt;&lt;br&gt;
execution performs actions&lt;br&gt;&lt;br&gt;
feedback drives learning  &lt;/p&gt;

&lt;p&gt;✖&lt;/p&gt;

&lt;p&gt;A question to practitioners:&lt;/p&gt;

&lt;p&gt;Which of these layers&lt;br&gt;
are already explicitly implemented&lt;br&gt;
in your systems?&lt;/p&gt;

&lt;p&gt;And in what form?&lt;/p&gt;

&lt;p&gt;Are they:&lt;/p&gt;

&lt;p&gt;centralized&lt;br&gt;&lt;br&gt;
scattered&lt;br&gt;&lt;br&gt;
implicit&lt;br&gt;&lt;br&gt;
formalized  &lt;/p&gt;

&lt;p&gt;✖&lt;/p&gt;

&lt;p&gt;Curious how others structure this today.&lt;/p&gt;

&lt;p&gt;✖ ✖ ✖ ✖ ✖ ✖ ✖&lt;/p&gt;

</description>
      <category>ai</category>
      <category>programming</category>
      <category>devops</category>
      <category>security</category>
    </item>
    <item>
      <title>From Prompts to Contracts: Why I Built the MOVA Contract Agent</title>
      <dc:creator>Sergii Miasoiedov</dc:creator>
      <pubDate>Mon, 16 Mar 2026 17:49:36 +0000</pubDate>
      <link>https://dev.to/mova/from-prompts-to-contracts-why-i-built-the-mova-contract-agent-2n0m</link>
      <guid>https://dev.to/mova/from-prompts-to-contracts-why-i-built-the-mova-contract-agent-2n0m</guid>
      <description>&lt;h2&gt;
  
  
  The Problem
&lt;/h2&gt;

&lt;p&gt;Over the past year, AI agents have become incredibly powerful.&lt;br&gt;&lt;br&gt;
Tools like large language models, coding assistants, and browser agents can already perform complex tasks.&lt;/p&gt;

&lt;p&gt;But there is still a fundamental problem.&lt;/p&gt;

&lt;p&gt;Most AI interactions today are based on prompts.&lt;br&gt;&lt;br&gt;
A user writes instructions in natural language and hopes the model interprets them correctly.&lt;/p&gt;

&lt;p&gt;This works for experimentation, but it becomes fragile when applied to real business processes.&lt;/p&gt;

&lt;p&gt;A business task is not just a conversation.&lt;br&gt;&lt;br&gt;
It has:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;rules
&lt;/li&gt;
&lt;li&gt;policies
&lt;/li&gt;
&lt;li&gt;verification
&lt;/li&gt;
&lt;li&gt;responsibility for the result
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without structure, AI execution becomes unpredictable.&lt;/p&gt;

&lt;p&gt;That question led me to explore a different approach.&lt;/p&gt;

&lt;p&gt;What if business tasks were not executed through prompts, but through contracts?&lt;/p&gt;




&lt;h2&gt;
  
  
  The Idea
&lt;/h2&gt;

&lt;p&gt;Instead of asking an AI agent to “solve a problem,” the system should first transform the request into a structured contract.&lt;/p&gt;

&lt;p&gt;A contract defines:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;what the user intends to achieve
&lt;/li&gt;
&lt;li&gt;which policies apply
&lt;/li&gt;
&lt;li&gt;what actions are allowed
&lt;/li&gt;
&lt;li&gt;how success will be verified
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Only after the contract is created does the agent execute the task.&lt;/p&gt;

&lt;p&gt;This changes the role of AI.&lt;/p&gt;

&lt;p&gt;The agent is no longer improvising a solution.&lt;br&gt;&lt;br&gt;
It is executing a defined agreement.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Architecture
&lt;/h2&gt;

&lt;p&gt;The architecture of the system reflects this idea.&lt;/p&gt;

&lt;p&gt;A user request first enters the system as a simple description of a problem.&lt;/p&gt;

&lt;p&gt;The system then performs three steps:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Intent Detection – understanding what the user is trying to accomplish.
&lt;/li&gt;
&lt;li&gt;Contract Creation – generating a structured contract that defines the task and the execution rules.
&lt;/li&gt;
&lt;li&gt;Execution – an agent performs the task according to the contract and produces a verifiable result.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This architecture separates intention, policy, and execution into clearly defined layers.&lt;/p&gt;

&lt;p&gt;It allows AI-driven processes to become traceable, auditable, and repeatable.&lt;/p&gt;

&lt;p&gt;In the next part of the demo, I show how this architecture works in practice.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Demo Scenario
&lt;/h2&gt;

&lt;p&gt;For the hackathon demonstration I chose a simple but realistic use case:&lt;br&gt;&lt;br&gt;
customer support tickets.&lt;/p&gt;

&lt;p&gt;A user sends a ticket describing a problem and attaches a screenshot.&lt;/p&gt;

&lt;p&gt;The system:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;analyzes the request
&lt;/li&gt;
&lt;li&gt;identifies the problem category
&lt;/li&gt;
&lt;li&gt;selects the corresponding solution contract
&lt;/li&gt;
&lt;li&gt;executes the resolution workflow
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;If the issue cannot be resolved automatically, the system escalates the ticket to a human operator.&lt;/p&gt;

&lt;p&gt;This shows how everyday business tasks can be handled through contract-based AI execution.&lt;/p&gt;




&lt;h2&gt;
  
  
  What I Learned
&lt;/h2&gt;

&lt;p&gt;Building this project reinforced an important idea.&lt;/p&gt;

&lt;p&gt;AI systems become far more reliable when they operate inside clear operational structures.&lt;/p&gt;

&lt;p&gt;Instead of asking AI to decide everything dynamically, we can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;define contracts
&lt;/li&gt;
&lt;li&gt;apply policies
&lt;/li&gt;
&lt;li&gt;verify outcomes
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This approach turns AI from an unpredictable assistant into a controlled execution system.&lt;/p&gt;




&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;The goal of this project is not to create another chatbot.&lt;/p&gt;

&lt;p&gt;It is to demonstrate a different way of thinking about AI-driven workflows.&lt;/p&gt;

&lt;p&gt;By introducing contracts as the core execution unit, business processes can become safer, more transparent, and easier to automate.&lt;/p&gt;

&lt;p&gt;This project is a small step toward that direction.&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>llm</category>
      <category>showdev</category>
    </item>
    <item>
      <title>From Prompts to Contracts: What Is Required for Businesses to Reliably Adopt Agentic AI</title>
      <dc:creator>Sergii Miasoiedov</dc:creator>
      <pubDate>Sun, 15 Mar 2026 18:42:23 +0000</pubDate>
      <link>https://dev.to/mova/from-prompts-to-contracts-what-is-required-for-businesses-to-reliably-adopt-agentic-ai-16f1</link>
      <guid>https://dev.to/mova/from-prompts-to-contracts-what-is-required-for-businesses-to-reliably-adopt-agentic-ai-16f1</guid>
      <description>&lt;ol&gt;
&lt;li&gt;The Question&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;What is required for businesses to confidently adopt agentic AI in their processes and use it to solve real business tasks?&lt;/p&gt;

&lt;p&gt;At first glance, the answer seems obvious:&lt;br&gt;
you need a strong model, access to data, and good automation.&lt;/p&gt;

&lt;p&gt;But when the conversation moves from demonstrations to real deployment, it quickly becomes clear that this is not enough.&lt;/p&gt;

&lt;p&gt;Businesses care not only about the capabilities of a system, but about its properties as an operational tool.&lt;/p&gt;

&lt;p&gt;They need to understand:&lt;br&gt;
    • what exactly the system is doing;&lt;br&gt;
    • what result counts as successful;&lt;br&gt;
    • why a particular decision was made;&lt;br&gt;
    • whether the result can be reproduced;&lt;br&gt;
    • and how much it costs to execute a task.&lt;/p&gt;

&lt;p&gt;The last point is particularly important.&lt;/p&gt;

&lt;p&gt;For a business, it is not enough to know how many tokens a model consumes on average.&lt;br&gt;
What matters is the cost of a unit of useful work:&lt;br&gt;
    • how much it costs to resolve a support ticket;&lt;br&gt;
    • how much it costs to validate a scenario;&lt;br&gt;
    • how much it costs to execute a task.&lt;/p&gt;

&lt;p&gt;Without this level of measurement, agentic AI is difficult to integrate into real operational economics.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Current Architecture and Its Limitations&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Today, most AI agents are built around the same basic scheme:&lt;/p&gt;

&lt;p&gt;prompt → model → result&lt;/p&gt;

&lt;p&gt;When this proves insufficient, developers start adding additional layers around the model:&lt;br&gt;
    • workflows&lt;br&gt;
    • orchestrators&lt;br&gt;
    • observability&lt;br&gt;
    • logs&lt;br&gt;
    • traces&lt;br&gt;
    • policies&lt;br&gt;
    • external checks&lt;/p&gt;

&lt;p&gt;But the underlying logic remains the same:&lt;br&gt;
the system receives a textual description of a task, and the model attempts to interpret and execute it.&lt;/p&gt;

&lt;p&gt;This approach has several fundamental limitations.&lt;/p&gt;

&lt;p&gt;2.1 Prompts Are Not a Formal Unit of Control&lt;/p&gt;

&lt;p&gt;A prompt is simply a description of a task in natural language.&lt;/p&gt;

&lt;p&gt;While convenient for experimentation, it is poorly suited for managing complex systems because:&lt;br&gt;
    • it cannot be validated before execution;&lt;br&gt;
    • it is difficult to formally analyze;&lt;br&gt;
    • it is hard to reproduce without drift;&lt;br&gt;
    • it provides weak boundaries for allowed actions.&lt;/p&gt;

&lt;p&gt;2.2 Workflows Do Not Fully Solve the Problem&lt;/p&gt;

&lt;p&gt;Many systems attempt to compensate for this by introducing workflow architectures, where the model becomes just one step in a predefined pipeline.&lt;/p&gt;

&lt;p&gt;But in this case the AI acts merely as a “smart function” inside a rigid process.&lt;/p&gt;

&lt;p&gt;This limits its usefulness in situations where the system needs to:&lt;br&gt;
    • analyze complex situations,&lt;br&gt;
    • compare alternatives,&lt;br&gt;
    • choose a path of action,&lt;br&gt;
    • operate under incomplete information.&lt;/p&gt;

&lt;p&gt;2.3 Observability Shows Activity but Not Meaning&lt;/p&gt;

&lt;p&gt;Even when systems collect detailed observability signals, we often still cannot interpret them with confidence.&lt;/p&gt;

&lt;p&gt;Why?&lt;/p&gt;

&lt;p&gt;Because observability shows:&lt;br&gt;
    • what happened,&lt;br&gt;
    • which calls were made,&lt;br&gt;
    • which steps were executed.&lt;/p&gt;

&lt;p&gt;But it cannot answer the most important question:&lt;/p&gt;

&lt;p&gt;Was this supposed to happen within the context of the task?&lt;/p&gt;

&lt;p&gt;To answer that question, the system needs two additional elements:&lt;br&gt;
    • a clearly defined intent,&lt;br&gt;
    • and a formal execution contract.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;What Businesses Currently Lack&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Strictly speaking, businesses do not primarily need another model or another pipeline.&lt;/p&gt;

&lt;p&gt;What they lack is an architecture where agentic AI becomes:&lt;br&gt;
    • understandable&lt;br&gt;
    • controllable&lt;br&gt;
    • verifiable&lt;br&gt;
    • measurable&lt;br&gt;
    • reproducible&lt;/p&gt;

&lt;p&gt;In other words, businesses need more than a model output.&lt;/p&gt;

&lt;p&gt;They need a system where it is possible to determine in advance:&lt;br&gt;
    • what outcome is expected,&lt;br&gt;
    • what the agent is allowed to do,&lt;br&gt;
    • how the result will be verified,&lt;br&gt;
    • how the execution will be recorded,&lt;br&gt;
    • and how the cost of the task will be calculated.&lt;/p&gt;

&lt;p&gt;Without these properties, deployment remains an experiment rather than a production system.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Proposed Solution&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;My answer to this problem is a shift toward a different architecture.&lt;/p&gt;

&lt;p&gt;Instead of:&lt;/p&gt;

&lt;p&gt;prompt → model → result&lt;/p&gt;

&lt;p&gt;I propose the following structure:&lt;/p&gt;

&lt;p&gt;intent → contract → validation → execution → verification → episode&lt;/p&gt;

&lt;p&gt;In this architecture, the key element is not the prompt but the intent.&lt;/p&gt;

&lt;p&gt;4.1 Intent&lt;/p&gt;

&lt;p&gt;Intent defines:&lt;br&gt;
    • what task we are actually trying to solve,&lt;br&gt;
    • what result is considered successful,&lt;br&gt;
    • what boundaries are acceptable,&lt;br&gt;
    • what constitutes completion.&lt;/p&gt;

&lt;p&gt;4.2 Contract&lt;/p&gt;

&lt;p&gt;Once the intent is defined, the task can be described as a contract.&lt;/p&gt;

&lt;p&gt;A contract is a structured description of the allowed solution space.&lt;/p&gt;

&lt;p&gt;It defines:&lt;br&gt;
    • inputs,&lt;br&gt;
    • expected outputs,&lt;br&gt;
    • allowed actions,&lt;br&gt;
    • constraints,&lt;br&gt;
    • execution rules,&lt;br&gt;
    • and success criteria.&lt;/p&gt;

&lt;p&gt;4.3 Validation&lt;/p&gt;

&lt;p&gt;Because the contract is formally described, it can be checked before execution begins.&lt;/p&gt;

&lt;p&gt;This becomes the first safety layer.&lt;/p&gt;

&lt;p&gt;When contracts are defined using JSON Schema, they can be validated structurally — including types, constraints, and required fields — before the agent performs any action.&lt;/p&gt;

&lt;p&gt;4.4 Execution&lt;/p&gt;

&lt;p&gt;After validation, the contract is executed with the assistance of the model.&lt;/p&gt;

&lt;p&gt;However, the model is no longer operating in an open-ended space.&lt;br&gt;
It operates within formally defined boundaries.&lt;/p&gt;

&lt;p&gt;4.5 Verification&lt;/p&gt;

&lt;p&gt;After execution, the result must be verified.&lt;/p&gt;

&lt;p&gt;The system checks whether:&lt;br&gt;
    • the contract was actually fulfilled,&lt;br&gt;
    • the result matches expectations,&lt;br&gt;
    • the execution remained within allowed boundaries.&lt;/p&gt;

&lt;p&gt;4.6 Episode&lt;/p&gt;

&lt;p&gt;Finally, the execution is recorded as an episode.&lt;/p&gt;

&lt;p&gt;An episode captures:&lt;br&gt;
    • the input signals,&lt;br&gt;
    • the selected contract,&lt;br&gt;
    • the execution process,&lt;br&gt;
    • and the resulting outcome.&lt;/p&gt;

&lt;p&gt;The cycle can then repeat.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The Execution Cycle&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The overall architecture looks like this:&lt;/p&gt;

&lt;p&gt;intent&lt;br&gt;
   ↓&lt;br&gt;
contract&lt;br&gt;
   ↓&lt;br&gt;
validation&lt;br&gt;
   ↓&lt;br&gt;
execution&lt;br&gt;
   ↓&lt;br&gt;
verification&lt;br&gt;
   ↓&lt;br&gt;
episode&lt;br&gt;
   ↓&lt;br&gt;
next cycle&lt;/p&gt;

&lt;p&gt;This is an important shift.&lt;/p&gt;

&lt;p&gt;In this architecture, AI is no longer a component that simply produces answers from text.&lt;br&gt;
Instead, it becomes part of a closed and controlled execution loop.&lt;/p&gt;

&lt;p&gt;The system can now:&lt;br&gt;
    • formalize a task,&lt;br&gt;
    • validate it,&lt;br&gt;
    • execute it,&lt;br&gt;
    • verify the result,&lt;br&gt;
    • and record the experience.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Advantages of This Approach&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Control&lt;/p&gt;

&lt;p&gt;Contracts define the allowed boundaries of agent behavior.&lt;br&gt;
The system no longer operates purely through open-ended reasoning.&lt;/p&gt;

&lt;p&gt;Verifiability&lt;/p&gt;

&lt;p&gt;Two layers of checks appear:&lt;br&gt;
    • validation before execution&lt;br&gt;
    • verification after execution&lt;/p&gt;

&lt;p&gt;Reproducibility&lt;/p&gt;

&lt;p&gt;If tasks are defined through contracts and results are recorded as episodes, processes can be reproduced and compared.&lt;/p&gt;

&lt;p&gt;Meaningful Observability&lt;/p&gt;

&lt;p&gt;Observability becomes interpretable because it now has context:&lt;br&gt;
    • intent&lt;br&gt;
    • contract&lt;br&gt;
    • expected outcome&lt;/p&gt;

&lt;p&gt;Execution Economics&lt;/p&gt;

&lt;p&gt;Contracts make tasks measurable.&lt;/p&gt;

&lt;p&gt;Instead of counting abstract token usage, systems can measure:&lt;br&gt;
    • the cost of executing a contract,&lt;br&gt;
    • the cost per unit of work,&lt;br&gt;
    • the cost of recurring episodes,&lt;br&gt;
    • the cost difference between reasoning and deterministic execution.&lt;/p&gt;

&lt;p&gt;This allows businesses to reason about the real economic cost of tasks.&lt;/p&gt;

&lt;p&gt;A Path Toward Deterministic Execution&lt;/p&gt;

&lt;p&gt;As episodes accumulate, recurring cases can be recognized and handled without repeated reasoning.&lt;/p&gt;

&lt;p&gt;Over time, part of the workload can move into a deterministic execution layer.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;What Has Already Been Developed&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Over the past year and a half, I have developed a language for formally describing processes.&lt;/p&gt;

&lt;p&gt;This language allows different types of system entities and workflows to be described, including:&lt;br&gt;
    • data structures&lt;br&gt;
    • specifications&lt;br&gt;
    • pipelines&lt;br&gt;
    • workflows&lt;br&gt;
    • agent cycles&lt;br&gt;
    • contracts&lt;br&gt;
    • execution episodes&lt;/p&gt;

&lt;p&gt;Contracts are only one application of this language.&lt;/p&gt;

&lt;p&gt;The system includes multiple representational layers, such as:&lt;br&gt;
    • schemas&lt;br&gt;
    • envelopes&lt;br&gt;
    • global context definitions&lt;br&gt;
    • episode recording&lt;br&gt;
    • semantic relationships between layers.&lt;/p&gt;

&lt;p&gt;In other words, this is not a one-off format for a specific use case.&lt;/p&gt;

&lt;p&gt;It is a broader system that allows processes to be:&lt;br&gt;
    • described&lt;br&gt;
    • validated&lt;br&gt;
    • executed&lt;br&gt;
    • observed&lt;br&gt;
    • analyzed.&lt;/p&gt;

&lt;p&gt;An execution engine for contracts and a mechanism for recording episodes have already been implemented, meaning the architecture is not only theoretical but supported by working tools.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Why This Makes Real Adoption Possible&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Businesses do not adopt AI simply because models are intelligent.&lt;/p&gt;

&lt;p&gt;They adopt AI when it can be integrated into systems of responsibility, control, measurement, and repeatability.&lt;/p&gt;

&lt;p&gt;Contract-based architectures make this possible.&lt;/p&gt;

&lt;p&gt;They allow organizations to:&lt;br&gt;
    • formalize tasks,&lt;br&gt;
    • validate them before execution,&lt;br&gt;
    • execute them in a controlled way,&lt;br&gt;
    • verify results,&lt;br&gt;
    • record episodes,&lt;br&gt;
    • and improve future cycles.&lt;/p&gt;

&lt;p&gt;This is the transition from experimental model usage to real deployment of agentic AI in business processes.&lt;/p&gt;

&lt;p&gt;⸻&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Conclusion&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;If businesses want to reliably deploy agentic AI, prompts, workflows, and observability alone are not enough.&lt;/p&gt;

&lt;p&gt;What is needed is an architecture built around:&lt;br&gt;
    • intent,&lt;br&gt;
    • contracts,&lt;br&gt;
    • validation,&lt;br&gt;
    • execution,&lt;br&gt;
    • verification,&lt;br&gt;
    • episodes,&lt;br&gt;
    • and measurable task cost.&lt;/p&gt;

&lt;p&gt;Moving from prompts to contracts is not just an interface change.&lt;/p&gt;

&lt;p&gt;It is a shift toward an architecture where agentic AI becomes:&lt;br&gt;
    • controllable,&lt;br&gt;
    • verifiable,&lt;br&gt;
    • reproducible,&lt;br&gt;
    • and economically measurable.&lt;/p&gt;

&lt;p&gt;These are precisely the properties required when AI systems move from experimentation into real operational environments.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
