DEV Community

Cover image for The Skill AI Adoption Actually Requires: Leadership
Daniel Sogl
Daniel Sogl

Posted on

The Skill AI Adoption Actually Requires: Leadership

AI-assisted software development doesn't just change how fast we write code — it fundamentally changes who we need to be while developing.

Over the past months, I've had countless conversations with developers and decision-makers in my role as a consultant and conference speaker. They almost always revolve around the same topics: How much speed does AI actually bring? What data privacy challenges does it create? How secure is AI-generated code?

All fair questions. But they miss the point that matters most.

The challenge isn't technical — it's a challenge of mental models. While developers today still primarily write code, the core competency of tomorrow will be leading agents. Not real employees. But the parallels to team leadership are closer than most people think.

What Social Media Gets Wrong

AI coding tools are improving at a pace that's hard to keep up with — weekly model updates, new agent frameworks, multi-agent workflows becoming the norm. The productivity numbers sound impressive.

The data tells a different story: A study by METR (2025) with experienced open-source developers found that in a controlled RCT, developers using AI tools were on average 19% slower than without — even though they believed they were 20% faster. METR's February 2026 follow-up notes that models have improved since — and that developers in the follow-up study increasingly refused to work without AI. The tools are deeply embedded in the workflow, whether productive or not.

METR Study: Forecasted vs. Actual Developer Productivity with AI Tools

A DX analysis (February 2026) based on 4.2 million developers sharpens the picture: 92.6% use AI coding assistants, nearly 27% of production code is already AI-written — and yet measurable productivity gains sit at just ~10%. The reason: developers spend only 20–40% of their time actually coding. The rest is problem analysis, product strategy, reviews, communication. A speedup in coding alone barely moves the needle.

The trust problem compounds this. The Stack Overflow Developer Survey 2025 (49,000+ developers, 166 countries): More developers actively distrust AI output (46%) than trust it (33%), with only 3% saying they highly trust AI results. Positive sentiment dropped from 70%+ (2023/2024) to 60% — despite rising adoption. 66% struggle with solutions that are almost right — just wrong enough to make debugging expensive.

Stack Overflow 2025: AI Tool Sentiment and Usage

None of this is an argument against AI coding. It's an argument for doing it right.

The Leadership Analogy

Think about what a team lead actually does.

Their job isn't to make every decision or review every line of code. A good lead enables their team to work autonomously: to make decisions, learn from mistakes, try new approaches. Each role brings specific knowledge. The team moves toward a shared goal.

Micro-management inverts this — and makes teams slow, demotivated, and dependent.

This is exactly the pattern many developers unconsciously apply to their AI agents today. Monitor every output, correct every step, distrust by default. Understandable — but it doesn't scale.

Research on software teams shows: psychological safety is the strongest predictor of team performance across all four DORA metrics. Teams with decision autonomy and a healthy error culture deliver better results. Working with agents is no different.

What Agent Leadership Means

The shift developers need to make has three dimensions:

1. From Writing to Directing

The Stack Overflow Survey 2025: 75% of developers would still ask a human when they don't trust the AI's output — even in a future where AI handles most coding tasks. The human isn't an optional backup. They're the critical filter.

An agent isn't an autopilot. It's a junior developer with broad knowledge but poor judgment in complex situations. The best coding agents today only solve ~23% of realistic software tasks correctly on their own.

2. Calibrating Trust — Neither Blind Faith Nor Blind Rejection

Google DeepMind describes a concrete paradox in a recent paper on Intelligent AI Delegation: when AI takes over too many tasks, humans lose the ability to intervene in critical situations. The fix: deliberately built-in checkpoints where humans retain control.

Not a weakness in the system — that's design.

3. Establishing a Culture of Failure

Agents make mistakes. Always. The question isn't whether, but how quickly you catch and correct them. DX data shows the difference in practice: companies with strong AI governance experience 50% fewer customer-facing incidents — poorly structured teams see twice as many. Governance isn't bureaucracy. It's the difference between AI as a force multiplier and AI as a liability.

Building Trust Through Structure

Automated tests are the equivalent of structured processes in a team.

A team lead doesn't hand over freedom and hope for the best. They create frameworks: clear goals, reviews, feedback loops. For AI agents, automated tests play exactly this role — the safety net that makes autonomy possible.

The Stanford study by Perry et al. (2023): developers using AI assistants write significantly less secure code — while being convinced they wrote secure code. The Stack Overflow Survey 2025 confirms this at scale: 45% say debugging AI-generated code takes longer than writing it themselves. Without tests, you're handing an agent full authority without accountability.

The New Skillset That Actually Matters

In a workshop last week, I asked a group of experienced developers to delegate a real task to a coding agent — the way they'd explain it to a new teammate. Most wrote a one-line prompt and waited. The agent returned code that compiled but solved the wrong problem.

That's not an agent problem. That's a delegation problem.

A new developer needs context: Which architectural decisions apply here? What's in scope? What should they not touch? An agent needs the same information — only more precisely stated, because it won't ask when something's unclear. It interprets and acts.

Once you internalize that, your approach shifts: tasks get specified more clearly. Acceptance criteria are defined before work starts, not after the first failure. Tests become the language through which you tell an agent what "done" means — not as a post-check, but as a briefing.

That's the competency shift: from executing to specifying. From writing to leading. Whoever delegates well today — to humans or agents — has a structural advantage that grows with every model update, not shrinks.

Conclusion

The conversation around AI-assisted development focuses too much on productivity and too little on the human side of this transformation. Agents are getting better, more autonomous, faster. But the human in this system isn't becoming redundant — they're changing roles.

From someone who writes code, to someone who sets direction, ensures quality, and enables a system of agents to work together toward a goal.

Leadership has always been one of the hardest disciplines in software development. In a world of agentic AI, it becomes the core competency.


Daniel is Principal Consultant for Generative AI at Thinktecture AG, Microsoft MVP for Developer Technologies, and a regular speaker at conferences across Germany and Europe.

Top comments (0)