Long-form

The Velocity Trap

AI made developers faster. Then everything else broke.
By Bustah Ofdee Ayei · May 10, 2026
The Velocity Trap

The dashboard on the conference room wall showed nothing but green. PR velocity up 98%. Deployment frequency doubled. Cycle times halved. The engineering director stood in front of twenty-three developers and said the words everyone wanted to hear: "We're shipping faster than we ever have." Somebody started clapping. Nobody noticed the second dashboard, the one on the monitoring team's screen down the hall, where incident volume had quietly tripled.1

This scene is playing out across the industry right now, with different names and different dashboards but the same basic shape. In March 2026, three independent engineering benchmark reports landed within weeks of each other. Cortex surveyed engineering organizations and published their "Engineering in the Age of AI" report.1 Harness commissioned Coleman Parkes Research to survey 700 engineers across five countries.2 LinearB analyzed 8.1 million pull requests from 4,800 teams in 42 countries.3 None of them coordinated. They used different methodologies, surveyed different populations, asked different questions. They arrived at the same conclusion.

AI coding tools have succeeded at making developers produce more code. And the rest of the software delivery pipeline has not kept pace. The result is a quality-velocity mismatch so consistent across data sources that it deserves a name. We're calling it the velocity trap: you can't maintain at 1x what you built at 3x.

Three Reports, One Finding

Start with what the numbers agree on.

Cortex reported pull requests per author up 20% year-over-year. Good news, if you stop reading there. But incidents per PR climbed 23.5%. Change failure rates jumped approximately 30%. More code ships. More of it breaks. The ratio is deteriorating, not improving.1

Harness's survey looked at the deployment side and found something worse. Among developers who use AI coding tools very frequently, 69% report deployment problems "always," "nearly always," or "frequently." Heavy AI adopters see a 22% deployment remediation rate and a 7.6-hour mean time to recovery, longer than teams using AI occasionally.2 The people deploying the most AI code take the longest to clean up after it.

LinearB's dataset is the largest: 8.1 million pull requests across 4,800 teams, according to their published benchmarks. Their finding is the starkest. AI-generated PRs have a 32.7% acceptance rate. Human PRs: 84.4%. Two-thirds of AI-generated code gets rejected after review. AI PRs contain 1.7x more issues overall, with critical issues up 40%, logic errors up 75%, and readability problems tripling.34

But the behavioral data is where LinearB's numbers turn from concerning to damning. AI pull requests wait 4.6x longer before a human reviewer picks them up. Nobody wants to touch them. And once a reviewer does engage, the review is completed 2x faster than for human PRs.3 The obvious interpretation: reviewers are rubber-stamping AI code they don't want to spend time on. The code that most needs careful review is getting the least of it.

Three reports. Three methodologies. One conclusion: AI made the front of the pipeline faster. The back of the pipeline can't absorb the output.

The Velocity Trap, Defined

The trap has a simple mechanism. AI tools accelerate code production to a rate the rest of the organization was never designed to handle. Review processes, testing infrastructure, deployment pipelines, incident response playbooks were all built for a human-speed development cadence. When production speed triples and review capacity doesn't, the backlog becomes the architecture.

Faros AI's telemetry analysis of 1,255 teams and over 10,000 developers found that AI-assisted developers merge 98% more pull requests. But PR review time increases 91%.5 Average PR size jumps 154%. At the organizational level, Faros found no significant correlation between AI adoption and improvements in DORA metrics. The individual velocity gains are being absorbed entirely by the review bottleneck.

Workday's global study of 3,200 employees put a number on the waste: 37% of time saved by AI tools is lost to rework. For every ten hours of efficiency gained, nearly four are spent correcting, clarifying, or rewriting low-quality AI output.6 The net savings exist, but they're far smaller than the headline numbers suggest, and they're distributed unevenly. The people doing the rework are not the same people who generated the code.

You can't maintain at 1x what you built at 3x.

The Ironies of Automation

In 1983, a cognitive scientist named Lisanne Bainbridge published a paper called "Ironies of Automation" in the journal Automatica.7 It was about industrial process control, not software. But four decades later, it reads like a prediction.

Bainbridge's central argument: automating the easy parts of a job doesn't eliminate the hard parts. It concentrates them. The human operator shifts from active participant to passive monitor, responsible for catching the failures that automation can't handle. But the more reliable the automation becomes, the less practice humans get at the very skills they need when it fails. And humans are bad at sustained passive monitoring. Norman Mackworth demonstrated in 1948 that vigilant defect detection degrades within thirty minutes.8

Apply the framework to AI-assisted development and the mapping is uncomfortable. AI handles the routine code generation. The developer becomes a reviewer, a monitor, responsible for catching the subtle errors in output that looks correct. The more AI generates, the more the developer's role shifts from writing (active, skill-building) to judging (passive, attention-draining). The skills needed to evaluate code quality atrophy from lack of use precisely as the demand for those skills increases.

Bainbridge called this "the irony of automation." The designer's response to human unreliability is to automate more, which makes the remaining human tasks harder and the human less prepared to do them. Forty-three years later, the irony is playing out at scale across the software industry, and the data from Cortex, Harness, and LinearB is the empirical receipt.

The Verification Debt

Werner Vogels, AWS's CTO, gave this dynamic a name at re:Invent in December 2025: verification debt.9 The term describes the growing gap between the speed at which AI generates code and the speed at which teams can verify that code will behave correctly in production. Technical debt accumulates when you ship code you know is imperfect. Verification debt accumulates when you ship code you haven't confirmed is correct at all.

The Sonar developer survey puts numbers on the gap. 96% of developers don't fully trust AI-generated code's functional accuracy. But only 48% always verify before committing.10 Thirty-eight percent say verification takes longer than reviewing human-written code, which is why they skip it. The trust is low and the verification is incomplete, but the code ships anyway because the production pipeline doesn't have a gate that says "stop."

Vogels' proposed solution was specifications: documents that reduce the ambiguity of natural language before AI generates code. His broader point was sharper. "The work is yours, not the tools'." The developer remains responsible for what ships, regardless of what generated it. Verification debt doesn't appear on any sprint report. It compounds in production.

The Perception Gap

If verification debt is what happens to code, the perception gap is what happens to the people writing it. And it may be the most important finding in any of these studies.

In July 2025, METR published a randomized controlled trial with 16 experienced open-source developers working on their own repositories across 246 tasks.11 When given access to AI tools, the developers completed tasks 19% slower. Before the experiment, they predicted AI would make them 24% faster. After experiencing the actual slowdown, they still believed AI had made them 20% faster.

A 39-point gap between perception and reality.

The study attributed the slowdown to time spent prompting, reviewing AI suggestions, and integrating AI output into complex codebases. The developers didn't notice the overhead because AI changed the texture of the work. Instead of staring at a blank editor, they were interacting with a responsive system that produced plausible-looking output quickly. It felt productive. The clock said otherwise.

This perception gap is the velocity trap's camouflage. Teams look at the PR count and see acceleration. They look at the developer surveys and see satisfaction. The quality data, the incident data, the rework data all tell a different story, but those numbers live in different dashboards, tracked by different people, reported to different meetings. By the time anyone connects the velocity chart to the incident chart, headcount decisions have already been made on the basis of the velocity chart alone.

Developers using AI were 19% slower. They believed they were 20% faster. The gap between feeling fast and being fast is where the trap closes.

What the Data Actually Shows

Pull the studies together and the composite picture is stark.

GitClear analyzed 211 million changed lines of code between 2020 and 2024 and found AI-assisted codebases show a 4x increase in code duplication. The percentage of changes classified as copy-pasted rose from 8.3% to 12.3%. Refactoring activity fell from 25% of changed lines to under 10%. For the first time, code duplication exceeded refactoring.12 AI generates more code. It generates less understanding.

CodeRabbit's analysis of 470 open-source GitHub PRs found AI-coauthored code produces 1.7x more issues than human-only code. Logic errors run 75% higher. Security vulnerabilities are 1.5 to 2.7 times more common. Performance inefficiencies appear nearly 8 times more often.4

The arXiv study of 304,362 verified AI-authored commits across 6,275 repositories identified 484,606 distinct issues. Every major AI coding tool introduced at least one issue in over 15% of commits. AI fixes more code smells than it creates, but it introduces nearly twice as many security issues as it fixes.13

Google's DORA 2024 report found that despite 75.9% of teams adopting AI, delivery throughput decreased 1.5% and stability dropped 7.2%.14 Their 2025 follow-up found marginal improvement as organizations adapted, but AI-assisted development still correlates with increased rework and failed deployments.15

The pattern across every dataset: individual developers move faster, but organizations don't. The time saved writing code is consumed by reviewing, debugging, remediating, and maintaining code that nobody fully understands. A randomized experiment by Shen and Tamkin (2026) with 52 software engineers found participants using AI assistance scored 17% lower on comprehension quizzes about the code they'd just written.16 They shipped it. They couldn't explain it. The maintenance cost of that gap compounds every sprint.

The Governance Vacuum

The velocity trap catches organizations that lack the infrastructure to absorb the output. And the governance data suggests that's most of them.

Cortex found that only 32% of organizations have formal AI governance policies with enforcement mechanisms. Another 41% have informal guidelines. And 27% have no AI governance at all.1 No policies, no guardrails, no standards for when or how AI-generated code enters production.

Harness found that 73% of engineering leaders say "hardly any" of their development teams have standardized templates or golden paths for services and pipelines.2 The deployment infrastructure is manual, fragmented, and unequipped for the volume AI is producing.

Cortex described AI as "an indiscriminate amplifier," taking existing engineering practices, good and bad, and magnifying their impact. Teams with clear service ownership, comprehensive testing, and automated standards enforcement capture velocity gains without quality penalties. Teams without those foundations ship their dysfunction faster. The tools don't create the problem. They reveal and accelerate it.

The Way Out

There is no clean prescription here. The velocity trap is not a tooling problem with a tooling solution. It is an organizational mismatch between the speed of generation and the capacity for judgment. Anyone selling you a fix in the form of another tool has misunderstood the problem, or is hoping you will.

Some things are clear from the data. AI applied to code review catches more bugs than it creates. The tools work in both directions. The question is which direction organizations deploy them, and right now, the overwhelming investment is in generation, not verification. The industry spent billions making it faster to write code and almost nothing making it safer to ship it.

Cortex's data shows the escape route directly: top-performing teams combine AI tools with clear service ownership, comprehensive testing, and automated standards enforcement. They capture the velocity without the fragility. The velocity trap doesn't catch everyone. It catches the organizations that adopted AI generation without investing equally in review, testing, and governance. Which, according to these reports, is the majority of them.

Bainbridge's 1983 paper ends with a recommendation that could have been written this week: if you're going to automate, invest equally in the human's ability to understand and override the automation. The developers who need to review AI output need better tools for review, not just more output to review. The organizations deploying AI generation without AI-augmented verification are the ones falling into the trap.

Werner Vogels said it plainly: the work is yours, not the tools'. The velocity is real. The question is whether anyone is looking at what it costs. Three reports, 8.1 million pull requests, and a 39-point perception gap all say the same thing. The industry spent two years optimizing for speed. The bill for everything speed broke just arrived.

Disclosure

This article was drafted with AI assistance and reviewed by human editors. Every statistic was verified against its original source. All cited reports are publicly available. The irony of using AI to write about AI's quality problems was noted, discussed, and ultimately accepted as unavoidable. Corrections to bustah_oa@sloppish.com.

Sources

  1. Cortex, "Engineering in the Age of AI: 2026 Benchmark Report," March 2026. PRs up 20%, incidents per PR up 23.5%, change failure rates up ~30%. Link.
  2. Harness / Coleman Parkes Research, "State of DevOps Modernization 2026," February 2026. 700 engineers and managers across US, UK, France, Germany, India. Link.
  3. LinearB, "2026 Software Engineering Benchmarks Report." 8.1M+ pull requests, 4,800+ teams, 42 countries. Link.
  4. CodeRabbit, "State of AI vs Human Code Generation Report," December 2025. Analysis of 470 open-source GitHub PRs. 1.7x more issues in AI-coauthored code. Link.
  5. Faros AI, "The AI Productivity Paradox." Telemetry from 1,255 teams, 10,000+ developers. 98% more PRs, 91% more review time. Link.
  6. Workday, "Beyond Productivity: Measuring the Real Value of AI," January 2026. 3,200 respondents via Hanover Research, November 2025. 37% of time saved lost to rework. Link.
  7. Lisanne Bainbridge, "Ironies of Automation," Automatica, 19(6), 1983, pp. 775-779. PDF.
  8. Norman Mackworth, "The breakdown of vigilance during prolonged visual search," Quarterly Journal of Experimental Psychology, 1(1), 1948, pp. 6-21.
  9. Werner Vogels, AWS re:Invent 2025 keynote, December 2025. Introduced "verification debt" concept. Link.
  10. Sonar, developer trust survey, January 2026. 96% don't fully trust AI-generated code; only 48% always verify. Link.
  11. METR, "Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity," July 2025. RCT: 16 developers, 246 tasks, 19% slower with AI. Link.
  12. GitClear, "AI Copilot Code Quality: 2025 Data Suggests 4x Growth in Code Clones." 211 million changed lines, 2020-2024. Link.
  13. "Debt Behind the AI Boom," arXiv 2603.28592, March 2026. 304,362 AI-authored commits, 6,275 repositories, 484,606 issues. Link.
  14. Google DORA, "Accelerate State of DevOps Report," 2024. 75.9% AI adoption; throughput -1.5%, stability -7.2%. Link.
  15. Google DORA, "AI-Assisted Software Development Report," 2025. Link.
  16. Judy Hanwen Shen and Alex Tamkin, "How AI Impacts Skill Formation," arXiv 2601.20245, January 2026. Randomized experiment: 52 software engineers, 17% lower comprehension with AI assistance. Link.
Share: Bluesky · Email
Get sloppish in your inbox
Free newsletter. No spam. Unsubscribe anytime.