MG Software.
HomeAboutServicesPortfolioBlog
Contact Us
All blogs

The AI Coding Paradox: Why Developers Are 19% Slower With AI (And Think They're Faster)

A landmark METR study found experienced developers are 19% slower with AI tools — while believing they're 20% faster. We break down why, what it means for your team, and how to actually benefit from AI-assisted development.

Jordan Munk
Jordan Munk18 Feb 2026 · 9 min read
The AI Coding Paradox: Why Developers Are 19% Slower With AI (And Think They're Faster)

Introduction

What if the tool you rely on every day is actually making you slower — and you have no idea? That is exactly what a landmark study by METR, a respected AI research organization, found when they put AI coding tools to a rigorous scientific test.

The headline result: experienced open-source developers completed tasks <strong>19% slower</strong> when using AI tools like Cursor and GitHub Copilot. The kicker? Those same developers <strong>believed they were 20% faster</strong>. That is a 39-point gap between perception and reality — and it has massive implications for how teams adopt AI tooling.

At MG Software, we use AI coding tools daily. We build with Cursor, deploy with AI-assisted workflows, and recommend these tools to clients. So when a study says "AI makes you slower," we take it seriously. Here is what the data actually shows — and what it means for your business.

What METR Actually Tested

METR recruited 16 experienced developers from major open-source repositories — projects averaging 22,000+ GitHub stars and over a million lines of code. These were not beginners. They had an average of 5 years working on their specific projects.

Each developer completed 246 real tasks from their own repositories, randomly assigned to either allow or disallow AI tools. When AI was permitted, developers primarily used Cursor Pro with Claude 3.5 and 3.7 Sonnet — frontier models at the time of the study (February through June 2025).

This matters because the study tested AI tools in their strongest possible context: experienced developers on familiar codebases using the best available AI models. If AI tools were going to shine anywhere, it would be here.

The 39-Point Perception Gap

The most striking finding is not the slowdown itself — it is that nobody noticed it. Before the study, developers predicted AI would reduce their completion time by 24%. After using the tools, they estimated they had been 20% faster. The actual measurement showed they were 19% slower.

Why the disconnect? The researchers identified several factors. AI tools provide a sense of progress — you see code appearing on screen faster, which creates an illusion of velocity. But that initial speed comes at a cost: more time spent reviewing, debugging, and correcting AI-generated suggestions. The net effect is negative.

There is also a sunk-cost dynamic. Once you have spent time crafting a prompt and waiting for a response, you are psychologically invested in using the output — even when rewriting from scratch would be faster. We have caught ourselves doing this at MG Software, and recognizing the pattern was the first step to fixing it.

Why Experienced Developers Slow Down Most

Here is the counterintuitive insight: the more you know a codebase, the less AI helps you navigate it. Experienced developers already have a mental model of the architecture, the conventions, and the edge cases. AI suggestions often conflict with that knowledge, requiring extra cognitive effort to evaluate and correct.

For tasks requiring deep architectural understanding — refactoring complex modules, fixing subtle race conditions, designing new APIs — AI-generated code frequently misses the context that a seasoned developer carries implicitly. The developer then spends time fixing AI suggestions instead of just writing the code themselves.

This does <strong>not</strong> mean AI tools are useless. It means their value proposition differs by context. For unfamiliar codebases, boilerplate generation, and exploratory prototyping, AI excels. For deep work on code you already understand intimately, it can be a distraction.

What This Means for Your Engineering Team

If you are a CTO or engineering manager considering AI tool adoption, the METR study does not say "do not use AI." It says "be strategic about when and how you use it." Here is what we recommend based on both the research and our own experience at MG Software.

First, stop measuring AI ROI by "lines of code generated." That metric is worse than useless — it actively misleads. Measure by feature completion time, bug introduction rate, and developer satisfaction instead.

Second, create explicit guidelines for when to use AI and when to go manual. Greenfield development, writing tests, generating boilerplate, and exploring unfamiliar APIs — these are high-value AI use cases. Complex refactors, security-critical code, and performance optimization — these often benefit from unassisted focus.

Third, invest in AI tool training. The developers who benefited most from AI tools in various studies were not the ones who accepted every suggestion — they were the ones who knew when to reject suggestions quickly and move on. That skill requires practice.

How We Use AI at MG Software

We are not abandoning AI tools. Far from it. But the METR study validated something we had already noticed: AI works best as an accelerator for specific tasks, not as a general-purpose speed boost.

Our internal workflow uses Cursor extensively for scaffolding new components, writing unit tests, and generating API client code. For these tasks, we consistently see 30-50% time savings. But for architectural decisions, code review, and debugging production issues, we rely on human expertise.

The key insight is that AI tools amplify your existing workflow — they do not replace it. If your development process is well-structured, AI makes it faster. If it is chaotic, AI makes it chaotically faster, which is worse. Invest in your engineering practices first, then layer AI on top.

If you are looking to adopt AI tools effectively in your development team, let us talk. We have learned these lessons the hard way so you do not have to.

Conclusion

The METR study is a wake-up call, but not the kind most headlines suggest. It does not prove AI coding tools are bad — it proves we are bad at measuring their impact. The 39-point perception gap should concern every engineering leader who is making investment decisions based on developer self-reports.

The real question is not "should we use AI tools?" — it is "are we using them in the right contexts?" At MG Software, we believe the answer requires both data and experience. We are committed to sharing what works and what does not as the technology evolves.

Share this post

Jordan Munk

Jordan Munk

Co-Founder

Related posts

GitHub Agentic Workflows: AI Agents That Review Your Pull Requests, Fix CI, and Triage Issues
AI & automation

GitHub Agentic Workflows: AI Agents That Review Your Pull Requests, Fix CI, and Triage Issues

GitHub's new Agentic Workflows let AI agents automatically review PRs, investigate CI failures, and triage issues. We break down how it works, the security architecture, and what this means for development teams.

Jordan Munk
Jordan Munk22 Feb 2026 · 8 min read
OpenClaw: The Open-Source AI Assistant That Took Over GitHub in Weeks
AI & automation

OpenClaw: The Open-Source AI Assistant That Took Over GitHub in Weeks

OpenClaw currently has 170K+ GitHub stars — one of the fastest-growing open-source projects ever. We analyze the hype, security risks, and what it means for businesses adopting AI agents.

Jordan Munk
Jordan Munk13 Feb 2026 · 8 min read
How AI Accelerates Custom Software Development
AI & automation

How AI Accelerates Custom Software Development

How MG Software uses AI tools to deliver projects faster and at higher quality, and what this means for businesses investing in custom software.

Jordan
Jordan12 Jan 2026 · 7 min read
Chatbots: Hype or Real Value
AI & automation

Chatbots: Hype or Real Value

Chatbots are everywhere, but do they actually deliver value? We analyze when a chatbot makes sense, when it does not, and how to get it right.

Jordan
Jordan16 Sept 2025 · 7 min read
e-bloom
Fitr
Fenicks
HollandsLof
Ipse
Bloominess
Bloemenwinkel.nl
Plus
VCA
Saga Driehuis
Sportief BV
White & Green Home
One Flora Group
e-bloom
Fitr
Fenicks
HollandsLof
Ipse
Bloominess
Bloemenwinkel.nl
Plus
VCA
Saga Driehuis
Sportief BV
White & Green Home
One Flora Group

Ready to build your
digital future?

Get in touch and discover how MG Software can transform your ideas into working software.

Contact usView our projects
MG Software
MG Software
MG Software.

MG Software builds custom software, websites and AI solutions that help businesses grow.

© 2026 MG Software B.V. All rights reserved.

NavigationServicesPortfolioAbout UsContactBlog
ResourcesKnowledge BaseComparisonsExamplesToolsRefront
LocationsHaarlemAmsterdamThe HagueEindhovenBredaAmersfoortAll locations
IndustriesLegalEnergyHealthcareE-commerceLogisticsAll industries