There’s a growing pattern among developers using AI.
When the output is wrong, shallow, or unusable, the conclusion comes fast:
“The AI messed up...
For further actions, you may consider blocking this person and/or reporting abuse
You can enter the clearest prompt ever, and "AI" will still come back with results that reference APIs that don't exist.
You could say that the developer thinking is now no longer about creating code but double- and triple-checking what an LLM suggests, but... AI isn't a mirror for your thinking. It's a mess.
That frustration is completely valid, and you’re right to call it out. LLMs can confidently reference APIs or behaviors that don’t exist, even when the prompt is very clear. That’s not a failure of prompting, it’s a limitation of how these models work.
Where I’d slightly reframe it is this: AI isn’t a mirror of thinking, and it shouldn’t be treated as an authority. It’s closer to a proposal generator. The developer’s role hasn’t shifted to “trusting” AI, but to designing guardrails, verification steps, and feedback loops so bad suggestions are caught early. I appreciate you raising this openly, skepticism like this is healthy and necessary.
That’s a very honest way to describe it, and I think many people feel the same tension. AI gives speed and momentum, but without deliberate pauses, it’s easy to slip into copying instead of understanding. That discipline, slowing down to assimilate, question, and decide, is what keeps the work intentional and sustainable. You’re right that this challenge applies to everyone, not just newcomers.
For working on AI, I tend to treat it as if they do not know anything and has to be specific as possible. It is a notion to think that AI is smarter than you and assume to know what you are saying, although AI doesn't even know you at all. Being specific is key.
Good work!
That’s a very grounded way to approach it. Treating AI as having no context unless you explicitly provide it leads to far more reliable outcomes. Specificity forces clarity in both the instruction and the thinking behind it, rather than relying on assumptions the system can’t actually make. I appreciate you sharing this perspective, and thank you for the kind words.
Thanks for the article, in my view, what is extracted from the AI tool, by the developer, is more of a reflection on the developer's technical and reasoning skills- rather than the AI tool itself. The AI tool is an aid, not the main actor.. the whole AI code generation narrative is inflated beyond reason.
Thank you for sharing that perspective, I agree with the framing. What comes out of an AI tool often reflects the developer’s reasoning, clarity, and technical judgment more than the tool itself. AI works best as an aid that supports thinking and execution, not as the main actor. Keeping that distinction clear helps cut through a lot of the hype around AI-driven code generation.
Developers are fast to adopt AI but slow to upgrade the knowledge.
This really hits home. The part about AI “reflecting ambiguity back at us” is spot on. I’ve noticed the same thing — when the output feels wrong, it’s usually because my own thinking wasn’t fully formed yet. AI doesn’t hide gaps the way traditional systems do, and that’s uncomfortable but honestly useful. Great perspective.
Thank you for sharing that, you’ve articulated the experience very clearly. When AI reflects ambiguity back at us, it can feel uncomfortable, but it’s often a signal that our own thinking needs refinement. Unlike traditional systems, it doesn’t quietly mask those gaps. Used intentionally, that feedback becomes a powerful tool for clarity rather than something to avoid. I appreciate you adding this thoughtful perspective.
Spot on. We’re moving from an era of coding by intuition to coding by specification.
Most 'AI failures' are actually just the model reflecting back a developer's unresolved ambiguity. In traditional dev, we rely on 'common sense' or 'fixing it later' to bridge the gap between a vague idea and working code. AI doesn't have common sense—it only has your instructions.
It’s a 'thinking stress test.' If the output is shallow, it’s usually because the constraints were implicit rather than explicit. The real skill shift isn't learning to prompt; it's learning to externalize the 'invisible logic' we've been carrying in our heads for years.
AI isn't lowering the bar for engineers; it's raising the floor for how rigorously we have to think.
This is an excellent articulation of the shift. You’re absolutely right , AI acts as a thinking stress test by forcing implicit assumptions into the open. What used to be patched over with “common sense” or deferred fixes now has to be made explicit upfront. That isn’t a lowering of standards; it’s a demand for clearer reasoning and better specification. I appreciate how you framed this as externalizing invisible logic, that’s exactly where the real skill shift is happening.
Self education is very important now to meet the demand of fast changing world.
When intuition isn’t translated into structure, AI output feels random, even though it’s not.