<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Artem Koltunov</title>
    <description>The latest articles on DEV Community by Artem Koltunov (@artem-koltunov).</description>
    <link>https://dev.to/artem-koltunov</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3836488%2F8adebb7a-7494-467f-bedc-25a1bf112dcd.jpg</url>
      <title>DEV Community: Artem Koltunov</title>
      <link>https://dev.to/artem-koltunov</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/artem-koltunov"/>
    <language>en</language>
    <item>
      <title>AI Coding Tools in Practice: What a 25-40% Productivity Gain Really Looks Like</title>
      <dc:creator>Artem Koltunov</dc:creator>
      <pubDate>Sat, 25 Apr 2026 17:29:40 +0000</pubDate>
      <link>https://dev.to/artem-koltunov/ai-coding-tools-in-practice-what-a-25-40-productivity-gain-really-looks-like-2eo8</link>
      <guid>https://dev.to/artem-koltunov/ai-coding-tools-in-practice-what-a-25-40-productivity-gain-really-looks-like-2eo8</guid>
      <description>&lt;p&gt;&lt;em&gt;Our JavaScript team tested AI-assisted development on production code. Here's what we measured, what surprised us, and why we think the real gain is 25-40% -- not the 10x you keep hearing about.&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;Over the past year, AI coding tools have been surrounded by bold claims: "Develop twice as fast." "10x developer productivity." "Code that practically writes itself."&lt;/p&gt;

&lt;p&gt;We decided to test these claims on real work -- not demo projects, but production code. The kind of long-lived repositories that power SDKs and developer platforms, systems that must be maintained, reviewed, and understood years after the code is written.&lt;/p&gt;

&lt;h2&gt;
  
  
  What We Tested
&lt;/h2&gt;

&lt;p&gt;Our JavaScript team works with AI models like GPT Codex, GPT-5.2, Opus 4.5, and Gemini 3.5 through IDE plugins -- specifically GitHub Copilot Chat in WebStorm and IntelliJ IDEA.&lt;/p&gt;

&lt;p&gt;Recently, we also got access to &lt;a href="https://cursor.com" rel="noopener noreferrer"&gt;Cursor&lt;/a&gt;, an IDE with deeply integrated AI that can operate across an entire project. Unlike traditional AI plugins where you manually select files and copy code into prompts, Cursor sees the whole codebase, creates files in the right locations, and applies changes directly.&lt;/p&gt;

&lt;p&gt;The biggest immediate impact wasn't smarter code generation -- it was the disappearance of small mechanical tasks. Less time copying code, managing context, and stitching pieces together. That alone produced an early productivity improvement of roughly &lt;strong&gt;20%&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;To see where this advantage held up -- and where it didn't -- we ran three experiments on active codebases.&lt;/p&gt;

&lt;h2&gt;
  
  
  Three Experiments
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Important note:&lt;/strong&gt; The first two experiments used GitHub Copilot Chat inside WebStorm, our usual IDE. The third introduced Cursor, which gave us a chance to compare a traditional AI plugin approach with a full-project AI environment.&lt;/p&gt;

&lt;h3&gt;
  
  
  Experiment 1: Extending a Production SDK
&lt;/h3&gt;

&lt;p&gt;We added new AI-related functionality to an existing JavaScript SDK: AI Summarize (generating summaries from ~1000 chat messages) and AI Gateway (recognizing text in images and generating descriptions). The task included API integration, SDK adaptation, tests, and usage examples.&lt;/p&gt;

&lt;p&gt;For this task we used GitHub Copilot Chat inside WebStorm. The AI could generate useful code, but we still had to gather context manually -- selecting files, pasting snippets, and explaining how modules interact -- before integrating whatever came back.&lt;/p&gt;

&lt;p&gt;Even with that overhead, AI assistance made a noticeable difference.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Result:&lt;/strong&gt; ~18 hours with AI vs. 24+ hours without. A gain of &lt;strong&gt;30-35%&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;What sped things up wasn't deep architectural insight. It was the smaller tasks: generating scaffolding, following existing patterns, and wiring pieces together faster than a human would type them.&lt;/p&gt;

&lt;h3&gt;
  
  
  Experiment 2: Untangling Long-Lived Branches
&lt;/h3&gt;

&lt;p&gt;Several parallel branches had been evolving separately since &lt;strong&gt;2021&lt;/strong&gt;. They contained overlapping logic, slightly different implementations, and subtle behavioral differences.&lt;/p&gt;

&lt;p&gt;Normally, merging something like this is slow and mentally draining. It requires reading a lot of unfamiliar code and carefully comparing approaches.&lt;/p&gt;

&lt;p&gt;Using Copilot Chat, we could feed sections of each branch to the model, ask it to highlight overlaps and divergences, and get explanations of unfamiliar code. That made it much easier to focus on the important part of the job -- deciding which implementation actually made sense.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Result:&lt;/strong&gt; ~1.5 days with AI vs. ~1 week without. Acceleration was &lt;strong&gt;several times&lt;/strong&gt; for tasks involving analysis and comparison of large codebases.&lt;/p&gt;

&lt;p&gt;The biggest advantage here wasn't generating code at all. It was simply making large amounts of existing code easier to understand.&lt;/p&gt;

&lt;h3&gt;
  
  
  Experiment 3: Integrating an SDK Into a Product (with Cursor)
&lt;/h3&gt;

&lt;p&gt;This experiment used Cursor. Two developers worked in parallel using different AI models (GPT-5.2 Codex and Opus 4.5). We created a complete Redux environment, connected Figma, generated layouts, and integrated business logic.&lt;/p&gt;

&lt;p&gt;At first, the results looked impressive.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Result:&lt;/strong&gt; ~20 hours with Cursor vs. ~40 hours without. Getting to working code &lt;strong&gt;2x faster&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;But this experiment also exposed a limitation that didn't show up in the earlier tasks.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Hidden Problem With AI-Generated Code
&lt;/h2&gt;

&lt;p&gt;The AI-generated code from Experiment 3 compiled, the interface behaved correctly, and the basic tests passed. If we had stopped there, we would have considered the integration complete.&lt;/p&gt;

&lt;p&gt;But during code review, one of the developers noticed something odd.&lt;/p&gt;

&lt;p&gt;An image identifier already existed inside one of the objects being passed through the system. Logically, the code should have simply reused that ID. Instead, the generated implementation took a much longer route: it fetched the ID, downloaded the associated blob, created a new file from it, uploaded that file back to the server, and then returned a new identifier.&lt;/p&gt;

&lt;p&gt;From the outside, nothing was broken. Internally, the process was doing far more work than necessary. Each time the logic ran, it duplicated data, added network calls, and quietly increased resource usage.&lt;/p&gt;

&lt;p&gt;We discovered this only because we opened the code and read it carefully.&lt;/p&gt;

&lt;p&gt;This turned out to be a pattern we started noticing more often with AI-generated code. The output usually &lt;em&gt;works&lt;/em&gt;, but the logic behind it doesn't always match the architecture of the system it's being added to. In shared components like SDKs, such inefficiencies can spread quietly through every product that depends on them.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Industry Research Shows
&lt;/h2&gt;

&lt;p&gt;While we were running these experiments, we studied key industry research. Our experience aligned closely with what independent analysts are measuring.&lt;/p&gt;

&lt;h3&gt;
  
  
  Productivity and Code Quality
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.gitclear.com/ai_assistant_code_quality_2025_research" rel="noopener noreferrer"&gt;GitClear's 2025 analysis&lt;/a&gt; found that AI tools can increase development speed by 20-55%, but the amount of "sustainable code" -- code that stays in the codebase without being rewritten -- grows by only about 10%. Developers produce code faster, but a noticeable portion still ends up being revised or refactored later. &lt;a href="https://gitclear-public.s3.us-west-2.amazonaws.com/GitClear-AI-Copilot-Code-Quality-2025.pdf" rel="noopener noreferrer"&gt;Full PDF report&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;A &lt;a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/" rel="noopener noreferrer"&gt;randomized controlled study by METR (July 2025)&lt;/a&gt; produced a striking result: experienced developers working on their own mature projects actually spent &lt;strong&gt;19% more time&lt;/strong&gt; with AI tools, while &lt;strong&gt;subjectively estimating&lt;/strong&gt; a 20% speedup. The key takeaway: perceived speed and actual speed are different things. Full data on &lt;a href="https://arxiv.org/abs/2507.09089" rel="noopener noreferrer"&gt;arXiv&lt;/a&gt; and &lt;a href="https://github.com/METR/Measuring-Early-2025-AI-on-Exp-OSS-Devs" rel="noopener noreferrer"&gt;GitHub&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Cost of Reviewing AI Code
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.sonarsource.com/blog/state-of-code-developer-survey-report-the-current-reality-of-ai-coding/" rel="noopener noreferrer"&gt;Sonar's State of AI in Code report (January 2026)&lt;/a&gt; found that &lt;strong&gt;95% of developers&lt;/strong&gt; spend significant effort checking AI-generated code, and &lt;strong&gt;38% consider it harder&lt;/strong&gt; to review than human-written code. Developers read and verify code far more slowly than AI generates it, which creates a natural ceiling on productivity gains. &lt;a href="https://www.sonarsource.com/state-of-code-developer-survey-report.pdf" rel="noopener noreferrer"&gt;Full PDF&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Architectural Limitations of AI-Generated Code
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.ox.security/resource-category/whitepapers-and-reports/army-of-juniors/" rel="noopener noreferrer"&gt;Ox Security's "Army of Juniors" report (October 2025)&lt;/a&gt; describes AI-generated code as "highly functional but systematically lacking architectural thinking." This explains why the code works but accumulates hidden problems. &lt;a href="https://www.ox.security/wp-content/uploads/2025/10/Army-of-Juniors-The-AI-Code-Security-Crisis.pdf" rel="noopener noreferrer"&gt;Report PDF&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Technical Debt
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.hfsresearch.com/press-release/ai-wont-save-enterprises-from-tech-debt-unless-they-change-the-architecture-first/" rel="noopener noreferrer"&gt;HFS Research + Unqork (November 2025)&lt;/a&gt; surveyed 123 respondents from Global 2000 organizations: while 84% expect AI to reduce costs, &lt;strong&gt;43% admit that AI creates new technical debt&lt;/strong&gt;. Opinions on long-term impact are split almost evenly -- 55% expect debt reduction, 45% expect increase.&lt;/p&gt;

&lt;p&gt;Forrester predicts that by 2026, 75% of tech leaders will face moderate or serious technical debt, with AI code generation without engineering discipline being a key factor.&lt;/p&gt;

&lt;h3&gt;
  
  
  Impact on Delivery Stability
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://dora.dev/research/2024/dora-report/" rel="noopener noreferrer"&gt;Google DORA Report 2024&lt;/a&gt; found a critical correlation: a &lt;strong&gt;25% increase in AI usage leads to a 7.2% decrease in delivery stability&lt;/strong&gt;. There's a 2.1% productivity gain and 2.6% job satisfaction increase -- but at the cost of 1.5% throughput decrease and 7.2% stability decrease. &lt;a href="https://services.google.com/fh/files/misc/2024_final_dora_report.pdf" rel="noopener noreferrer"&gt;Full PDF&lt;/a&gt;. The &lt;a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-the-2025-dora-report" rel="noopener noreferrer"&gt;2025 DORA Report&lt;/a&gt; confirms these findings.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why the Real Gain Is 25-40%
&lt;/h2&gt;

&lt;p&gt;Looking across both our experiments and the broader research, the same pattern keeps appearing.&lt;/p&gt;

&lt;p&gt;AI tools clearly speed up certain parts of development: reducing boilerplate, navigating large codebases, scaffolding new functionality, and accelerating the path to a working implementation.&lt;/p&gt;

&lt;p&gt;But those gains come with a counterweight. The code still needs to be understood, reviewed, and integrated into an existing system. Developers reason about code far more slowly than AI can generate it.&lt;/p&gt;

&lt;p&gt;Without proper review, teams accumulate what we call "AI legacy code" -- code that works but nobody on the team truly understands. Over time, it becomes easier to regenerate than to modify. But regeneration means spending time and resources on problems that were already solved. In high-debt environments, losses reach 30-40% of the change budget and 10-20% of system operation costs.&lt;/p&gt;

&lt;p&gt;This situation can develop within months after active AI code adoption without full developer involvement.&lt;/p&gt;

&lt;p&gt;That's why the dramatic claims about "10x productivity" rarely hold up in real engineering environments. In practice, the gains stabilize in the &lt;strong&gt;25-40% range&lt;/strong&gt; -- meaningful enough to matter, but not so large that engineering judgment becomes unnecessary.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;AI coding tools are most useful when treated as assistants rather than replacements for engineering judgment.&lt;/p&gt;

&lt;p&gt;They excel at analyzing and comparing large volumes of code -- tasks that take humans significant time, AI handles very quickly. They reduce friction in everyday development and can meaningfully accelerate time-to-working-code.&lt;/p&gt;

&lt;p&gt;At the same time, tasks requiring deep understanding of business logic and architectural optimization are often solved by AI in suboptimal ways. The resulting code works but is redundant. The system functions correctly on the surface, but hidden problems related to performance, resource usage, and maintainability can form inside.&lt;/p&gt;

&lt;p&gt;Architectural decisions, quality control, and responsibility for results must stay with the team. With this discipline in place, AI tools deliver a real, measurable, and sustainable productivity boost.&lt;/p&gt;




&lt;h2&gt;
  
  
  References
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Productivity and Code Quality
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.gitclear.com/ai_assistant_code_quality_2025_research" rel="noopener noreferrer"&gt;GitClear: AI Assistant Code Quality 2025&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.gitclear.com/research/ai_tool_impact_on_developer_productive_output_from_2022_to_2025" rel="noopener noreferrer"&gt;GitClear: AI Tool Impact on Developer Output 2022-2025&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arxiv.org/abs/2507.09089" rel="noopener noreferrer"&gt;METR: Measuring AI Impact on Experienced Developers (arXiv)&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://github.com/METR/Measuring-Early-2025-AI-on-Exp-OSS-Devs" rel="noopener noreferrer"&gt;METR: Study Data and Code (GitHub)&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  AI Code Review and Security
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.sonarsource.com/resources/developer-survey-report/" rel="noopener noreferrer"&gt;Sonar: State of AI in Code Report 2026&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.ox.security/resource-category/whitepapers-and-reports/army-of-juniors/" rel="noopener noreferrer"&gt;Ox Security: Army of Juniors Report&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Technical Debt
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.hfsresearch.com/press-release/ai-wont-save-enterprises-from-tech-debt-unless-they-change-the-architecture-first/" rel="noopener noreferrer"&gt;HFS Research: AI and Tech Debt&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.ciodive.com/news/enterprises-split-ai-affect-tech-debt/806208/" rel="noopener noreferrer"&gt;CIO Dive: How AI Affects Tech Debt&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  DevOps Metrics
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://dora.dev/research/2024/dora-report/" rel="noopener noreferrer"&gt;Google DORA Report 2024&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-the-2025-dora-report" rel="noopener noreferrer"&gt;Google DORA Report 2025&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.gitclear.com/research/google_dora_2024_summary_ai_impact" rel="noopener noreferrer"&gt;GitClear: DORA 2024 Summary -- AI Impact&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Independent Reviews
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://simonwillison.net/2025/Jul/12/ai-open-source-productivity/" rel="noopener noreferrer"&gt;Simon Willison: AI Open Source Productivity&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.infoq.com/news/2025/11/ai-code-technical-debt/" rel="noopener noreferrer"&gt;InfoQ: AI Code Technical Debt&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.theregister.com/2026/01/09/devs_ai_code/" rel="noopener noreferrer"&gt;The Register: Devs and AI Code&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>productivity</category>
      <category>programming</category>
      <category>webdev</category>
    </item>
  </channel>
</rss>
