Natural language is flexible by nature.
That flexibility is exactly why it fails as a control mechanism.
Language models are trained to continue, complete, and smooth over gaps.
They are not trained to pause, refuse, or wait for permission.
When we ask a model to “strictly follow steps” using language alone, we are creating a paradox:
We are asking a generative system to restrict itself using the very medium it generates.
This leads to predictable failure modes:
Missing steps are silently filled in
Unconfirmed assumptions are treated as facts
Execution continues even when inputs are incomplete
This is not misbehavior.
It is correct behavior under the wrong responsibility assignment.
Language is excellent for expression.
It is disastrous for enforcement.
Any system that relies on prompts alone to maintain execution boundaries will eventually break—quietly and convincingly.
Top comments (0)