- The Question
What is required for businesses to confidently adopt agentic AI in their processes and use it to solve real business tasks?
At first glance, the answer seems obvious:
you need a strong model, access to data, and good automation.
But when the conversation moves from demonstrations to real deployment, it quickly becomes clear that this is not enough.
Businesses care not only about the capabilities of a system, but about its properties as an operational tool.
They need to understand:
• what exactly the system is doing;
• what result counts as successful;
• why a particular decision was made;
• whether the result can be reproduced;
• and how much it costs to execute a task.
The last point is particularly important.
For a business, it is not enough to know how many tokens a model consumes on average.
What matters is the cost of a unit of useful work:
• how much it costs to resolve a support ticket;
• how much it costs to validate a scenario;
• how much it costs to execute a task.
Without this level of measurement, agentic AI is difficult to integrate into real operational economics.
⸻
- The Current Architecture and Its Limitations
Today, most AI agents are built around the same basic scheme:
prompt → model → result
When this proves insufficient, developers start adding additional layers around the model:
• workflows
• orchestrators
• observability
• logs
• traces
• policies
• external checks
But the underlying logic remains the same:
the system receives a textual description of a task, and the model attempts to interpret and execute it.
This approach has several fundamental limitations.
2.1 Prompts Are Not a Formal Unit of Control
A prompt is simply a description of a task in natural language.
While convenient for experimentation, it is poorly suited for managing complex systems because:
• it cannot be validated before execution;
• it is difficult to formally analyze;
• it is hard to reproduce without drift;
• it provides weak boundaries for allowed actions.
2.2 Workflows Do Not Fully Solve the Problem
Many systems attempt to compensate for this by introducing workflow architectures, where the model becomes just one step in a predefined pipeline.
But in this case the AI acts merely as a “smart function” inside a rigid process.
This limits its usefulness in situations where the system needs to:
• analyze complex situations,
• compare alternatives,
• choose a path of action,
• operate under incomplete information.
2.3 Observability Shows Activity but Not Meaning
Even when systems collect detailed observability signals, we often still cannot interpret them with confidence.
Why?
Because observability shows:
• what happened,
• which calls were made,
• which steps were executed.
But it cannot answer the most important question:
Was this supposed to happen within the context of the task?
To answer that question, the system needs two additional elements:
• a clearly defined intent,
• and a formal execution contract.
⸻
- What Businesses Currently Lack
Strictly speaking, businesses do not primarily need another model or another pipeline.
What they lack is an architecture where agentic AI becomes:
• understandable
• controllable
• verifiable
• measurable
• reproducible
In other words, businesses need more than a model output.
They need a system where it is possible to determine in advance:
• what outcome is expected,
• what the agent is allowed to do,
• how the result will be verified,
• how the execution will be recorded,
• and how the cost of the task will be calculated.
Without these properties, deployment remains an experiment rather than a production system.
⸻
- The Proposed Solution
My answer to this problem is a shift toward a different architecture.
Instead of:
prompt → model → result
I propose the following structure:
intent → contract → validation → execution → verification → episode
In this architecture, the key element is not the prompt but the intent.
4.1 Intent
Intent defines:
• what task we are actually trying to solve,
• what result is considered successful,
• what boundaries are acceptable,
• what constitutes completion.
4.2 Contract
Once the intent is defined, the task can be described as a contract.
A contract is a structured description of the allowed solution space.
It defines:
• inputs,
• expected outputs,
• allowed actions,
• constraints,
• execution rules,
• and success criteria.
4.3 Validation
Because the contract is formally described, it can be checked before execution begins.
This becomes the first safety layer.
When contracts are defined using JSON Schema, they can be validated structurally — including types, constraints, and required fields — before the agent performs any action.
4.4 Execution
After validation, the contract is executed with the assistance of the model.
However, the model is no longer operating in an open-ended space.
It operates within formally defined boundaries.
4.5 Verification
After execution, the result must be verified.
The system checks whether:
• the contract was actually fulfilled,
• the result matches expectations,
• the execution remained within allowed boundaries.
4.6 Episode
Finally, the execution is recorded as an episode.
An episode captures:
• the input signals,
• the selected contract,
• the execution process,
• and the resulting outcome.
The cycle can then repeat.
⸻
- The Execution Cycle
The overall architecture looks like this:
intent
↓
contract
↓
validation
↓
execution
↓
verification
↓
episode
↓
next cycle
This is an important shift.
In this architecture, AI is no longer a component that simply produces answers from text.
Instead, it becomes part of a closed and controlled execution loop.
The system can now:
• formalize a task,
• validate it,
• execute it,
• verify the result,
• and record the experience.
⸻
- Advantages of This Approach
Control
Contracts define the allowed boundaries of agent behavior.
The system no longer operates purely through open-ended reasoning.
Verifiability
Two layers of checks appear:
• validation before execution
• verification after execution
Reproducibility
If tasks are defined through contracts and results are recorded as episodes, processes can be reproduced and compared.
Meaningful Observability
Observability becomes interpretable because it now has context:
• intent
• contract
• expected outcome
Execution Economics
Contracts make tasks measurable.
Instead of counting abstract token usage, systems can measure:
• the cost of executing a contract,
• the cost per unit of work,
• the cost of recurring episodes,
• the cost difference between reasoning and deterministic execution.
This allows businesses to reason about the real economic cost of tasks.
A Path Toward Deterministic Execution
As episodes accumulate, recurring cases can be recognized and handled without repeated reasoning.
Over time, part of the workload can move into a deterministic execution layer.
⸻
- What Has Already Been Developed
Over the past year and a half, I have developed a language for formally describing processes.
This language allows different types of system entities and workflows to be described, including:
• data structures
• specifications
• pipelines
• workflows
• agent cycles
• contracts
• execution episodes
Contracts are only one application of this language.
The system includes multiple representational layers, such as:
• schemas
• envelopes
• global context definitions
• episode recording
• semantic relationships between layers.
In other words, this is not a one-off format for a specific use case.
It is a broader system that allows processes to be:
• described
• validated
• executed
• observed
• analyzed.
An execution engine for contracts and a mechanism for recording episodes have already been implemented, meaning the architecture is not only theoretical but supported by working tools.
⸻
- Why This Makes Real Adoption Possible
Businesses do not adopt AI simply because models are intelligent.
They adopt AI when it can be integrated into systems of responsibility, control, measurement, and repeatability.
Contract-based architectures make this possible.
They allow organizations to:
• formalize tasks,
• validate them before execution,
• execute them in a controlled way,
• verify results,
• record episodes,
• and improve future cycles.
This is the transition from experimental model usage to real deployment of agentic AI in business processes.
⸻
- Conclusion
If businesses want to reliably deploy agentic AI, prompts, workflows, and observability alone are not enough.
What is needed is an architecture built around:
• intent,
• contracts,
• validation,
• execution,
• verification,
• episodes,
• and measurable task cost.
Moving from prompts to contracts is not just an interface change.
It is a shift toward an architecture where agentic AI becomes:
• controllable,
• verifiable,
• reproducible,
• and economically measurable.
These are precisely the properties required when AI systems move from experimentation into real operational environments.
Top comments (1)
Some comments may only be visible to logged-in visitors. Sign in to view all comments.