More Dev Interrupted episodes

Why AI-assisted PRs merge at half the rate of human code | LinearBs 2026 Benchmarks thumbnail

Why AI-assisted PRs merge at half the rate of human code | LinearBs 2026 Benchmarks

Published 24 Mar 2026

Duration: 2370

The 2026 Engineering Benchmark Report reveals that while 88.3% of developers use AI regularly, AI-generated pull requests face low merge rates (32.7%), larger sizes, and prolonged reviews due to systemic issues like poor data quality, inadequate policies, and organizational gaps, emphasizing the need for governance, smaller focused PRs, and foundational practices to optimize AI's potential in engineering workflows.

Episode Description

Over 88% of developers use AI regularly, but AI-assisted pull requests merge at less than half the rate of human-authored code. In this episode, Dan L...

Overview

The 2026 Engineering Benchmark Report, the largest and most comprehensive to date, analyzed 8.1 million poll requests and 4,800+ engineering teams across 42 countries, with a new focus on AI productivity. Key findings highlight rapid AI adoption, with 88.3% of developers using AI regularly, though challenges persist. AI-generated pull requests (PRs) face significant hurdles, merging at just 32.7% of the rate of human-authored PRs due to larger size, slower reviews, and governance bottlenecks. The report categorizes PRs into three types: unassisted (fully human-authored, with high merge rates), AI-assisted (human-led with AI support, resulting in larger PRs and prolonged reviews), and agentic (fully AI-created, with the lowest merge rates). These differences underscore inefficiencies in AI integration, including scope creep, reviewer hesitation, and trust issues.

Qualitative insights reveal that while AI is pervasive in software development, its real-world impact remains limited, often exacerbating issues like technical debt and code quality. Engineering leaders emphasized the need for systemic improvements beyond code generation, such as refining workflows, governance policies, and data quality. Organizations with weak foundational practices (e.g., poor version control, unclear AI policies) struggle with AI adoption, as suboptimal data and context lead to unreliable outputs. The report also highlights that AI currently focuses on generating new code rather than refactoring or addressing legacy technical debt. Recommendations include prioritizing smaller, focused PRs, ensuring human oversight, and improving organizational readiness through clear policies and context engineering to align AI with existing systems and priorities.

Recent Episodes of Dev Interrupted

More Dev Interrupted episodes