This page is organized into four categories: Adoption Data, Output and Productivity Data, Quality Data, and Developer Sentiment Data. Each data point includes the source name, publication date, and a link to the original research. Where data points conflict, both are presented with context.
This is a living reference. Data points are added as credible research is published and removed when superseded by more recent findings from the same source.
These data points measure how widely AI coding tools are used, what percentage of code they generate, and how adoption is distributed across organizations.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| AI-generated code share in high-adoption organizations | 30-70% | Larridin internal benchmark | 2025-2026 | -- |
| Projected AI-generated code share by 2028 | >50% of all new code | Industry consensus | 2025 | -- |
AI code share -- the percentage of committed code generated or substantially drafted by AI tools -- varies significantly by organization, team, and task type. The 30-70% range reflects production data from organizations with mature AI coding workflows, not experimental or pilot programs. Teams working on greenfield features tend toward the higher end; teams maintaining complex legacy systems trend lower.
For a deeper exploration of this metric, see What Is AI Code Share? Measuring the Percentage of AI-Generated Code.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Organizations where all IT work will involve AI by 2030 | Projected majority | Gartner CIO Survey | Oct 2025 | Gartner |
| Organizations struggling to demonstrate AI value | 72% breaking even or losing money | Gartner CIO Survey | Oct 2025 | Gartner |
The gap between adoption and demonstrated value is the defining challenge of AI coding in 2026. Organizations are adopting AI tools rapidly, but most cannot yet prove that the investment is paying off -- largely because their measurement systems were designed for pre-AI workflows.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Developers who have used AI coding tools | 92% (among surveyed developers) | GitHub & Wakefield Research | Jun 2023 | GitHub Blog |
Note: The 92% figure is from a 2023 survey and reflects usage at any level, including trial and experimentation. It does not indicate sustained, daily adoption.
These data points measure the impact of AI coding tools on developer output -- speed, volume, and task completion.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Task completion speed increase (controlled experiment) | 55.8% faster | GitHub Copilot research | Feb 2023 | GitHub Blog |
GitHub's controlled experiment measured developers completing an HTTP server task with and without Copilot. Developers with Copilot completed the task 55.8% faster on average. This is the most frequently cited productivity figure in AI coding research.
Context and limitations: The experiment used a well-defined, self-contained coding task. Real-world development involves requirements ambiguity, architectural decisions, code review, and integration work that may not benefit as directly from AI assistance. The 55.8% figure should be understood as a ceiling for the type of tasks where AI assistance is most effective, not as a general productivity multiplier.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Increase in total code output (lines changed) with AI tools | Significant increase observed | GitClear | 2024-2025 | GitClear 2024, GitClear 2025 |
| PR volume increase in high-adoption teams | 3-5x | Larridin internal benchmark | 2025 | -- |
PR volume and code output increases are real, but they are output metrics, not outcome metrics. A 3-5x increase in PRs merged does not necessarily translate to a 3-5x increase in value delivered. See Quality Data below for the counterbalancing signals.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Time developers spend on "boilerplate" or repetitive tasks (reduced by AI) | Meaningful reduction reported | GitHub Copilot research | 2023 | GitHub Blog |
Developers consistently report that AI tools reduce time spent on repetitive, low-complexity coding tasks. This time reallocation is one of the clearest benefits of AI coding tools, but it is difficult to quantify precisely because "boilerplate" is subjective and varies by codebase.
These data points measure the impact of AI coding tools on code quality, durability, and maintainability. Quality data is where the AI coding narrative becomes more nuanced.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Pre-AI code churn baseline | ~3.3% | GitClear | Pre-2023 | GitClear |
| Code churn rate (2024) | ~5.7% | GitClear | 2024 | GitClear |
| Code churn rate (2025) | ~7.1% | GitClear | 2025 | GitClear |
| Change from baseline to 2025 | +115% (doubled) | GitClear | 2025 | GitClear |
Code churn -- the rate at which recently written code is modified or deleted -- has more than doubled since AI coding tools went mainstream. This is the single most important quality signal in AI coding data. More code is being written, and more of that code is being thrown away shortly after creation.
For a detailed analysis of what is driving the increase and how to address it, see Code Churn in the AI Era: Why It's Doubled and What to Do.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| AI-generated code turnover vs. human-written | 1.8-2.5x higher | Larridin internal benchmark | 2025 | -- |
| Healthy AI-to-human turnover ratio target | Below 1.5x | Larridin internal benchmark | 2025 | -- |
| Problem threshold for AI-to-human turnover ratio | Above 2.0x | Larridin internal benchmark | 2025 | -- |
Code Turnover Rate measures the percentage of merged code that is reverted, deleted, or substantially rewritten within 30 or 90 days. It is more specific than churn because it isolates rework from healthy refactoring. For full methodology, see What Is Code Turnover Rate? The AI Code Quality Metric.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Increase in moved code | Notable rise since AI adoption | GitClear | 2024-2025 | GitClear |
| Increase in copy-pasted code | Notable rise since AI adoption | GitClear | 2024-2025 | GitClear |
GitClear's research identifies specific categories of code change that have grown disproportionately with AI adoption. Moved code (code placed in the wrong location and later relocated) and copy-pasted code (duplicated implementations) are both signatures of AI-generated code that lacks codebase awareness.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| AI-generated code passes review more easily due to surface quality | Consistently observed | GitClear, Larridin internal benchmark | 2024-2025 | GitClear |
| Increase in review volume per reviewer in high-adoption teams | Proportional to code output increase | Larridin internal benchmark | 2025 | -- |
AI-generated code is typically well-formatted, syntactically clean, and properly commented -- qualities that make it appear higher quality during review. However, the issues that drive churn and turnover -- architectural misalignment, unnecessary duplication, missing edge case handling -- are precisely the issues that are hardest to catch during review and easiest to overlook when the code looks polished on the surface.
The review volume problem compounds this. When AI tools increase code output by 30-70%, review workload increases proportionally, but review capacity does not scale at the same rate. Reviewers face more code per day and are more likely to approve AI-generated PRs quickly because they "look right." This dynamic creates a quality gap between what passes review and what actually survives in the codebase long-term.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Time spent on maintenance vs. new features (industry average) | Varies widely; AI has not yet measurably reduced maintenance burden | Multiple sources | 2025 | -- |
One of the promises of AI coding tools is that by accelerating routine coding, developers will spend more time on innovation and less on maintenance. The data on whether this is actually happening is mixed. While AI tools clearly accelerate task completion, the increase in code churn and turnover introduces new maintenance demands -- debugging AI-generated code, resolving duplication, and reworking implementations that were architecturally misaligned. Whether AI is a net positive or negative for maintenance burden likely depends on how well the organization manages AI code quality through practices like test-driven development and rigorous review.
These data points measure how developers feel about AI coding tools -- satisfaction, concerns, and perceived impact.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Developers reporting greater fulfillment with AI tools | 75% | GitHub & Wakefield Research | Jun 2023 | GitHub Blog |
| Developers reporting AI helps them focus on more satisfying work | 73% | GitHub & Wakefield Research | Jun 2023 | GitHub Blog |
Developer sentiment toward AI coding tools is broadly positive. The most consistent finding across surveys is that developers feel AI tools reduce the tedious parts of their work and allow them to focus on more interesting problems.
| Metric | Value | Source | Date | Link |
|---|---|---|---|---|
| Developers concerned about AI code quality | Significant minority | Multiple surveys | 2024-2025 | -- |
| Engineering leaders unable to quantify AI ROI | 72% (breaking even or losing money) | Gartner CIO Survey | Oct 2025 | Gartner |
The sentiment data reveals a tension: developers like AI coding tools and find them useful, but engineering leaders struggle to translate that satisfaction into measurable business outcomes. This gap is fundamentally a measurement problem -- one that the Developer AI Impact Framework is designed to address.
The evidence is clear that AI coding tools accelerate individual task completion and increase code output. The GitHub Copilot research showing 55.8% faster task completion is well-designed and widely replicated directionally. Developers write more code, faster.
Speed gains are accompanied by measurable quality costs. Code churn has doubled. AI-generated code turns over at nearly twice the rate of human-written code. More code is being produced, and more of it is being discarded. The net productivity gain -- speed minus rework -- is positive for most teams, but smaller than raw output numbers suggest.
The most striking data point in this collection is not about code quality or developer speed. It is that 72% of organizations cannot demonstrate positive ROI from their AI investments. The technology is not the problem. The inability to measure its impact is.
Traditional metrics -- lines of code, PRs merged, deployment frequency -- are inflated by AI output and no longer reliably indicate productivity. New metrics that account for AI's contribution, code quality, and net value delivered are essential. The Developer AI Impact Framework provides a measurement methodology designed for this reality.
As AI handles more of the code generation, the developer's role shifts toward orchestration, specification, and verification. This shift has implications for how productivity is measured, how teams are structured, and what skills matter most. This evolution is explored in From Coding to Verification: How Developer Roles Are Changing.
This page follows strict sourcing standards:
The 55.8% figure comes from a controlled experiment conducted by GitHub in which developers completed an HTTP server coding task with and without GitHub Copilot. It measures task completion speed on a well-defined, self-contained task. It is widely cited but should be understood as specific to the task type studied, not as a universal productivity multiplier.
GitClear measures code churn as the percentage of recently added lines of code that are modified, moved, or deleted within a short time window after creation. Their dataset spans 211 million lines of changed code across a large number of repositories, providing a broad industry view rather than a single-company sample.
According to the Gartner CIO Survey (October 2025), most organizations are measuring AI's impact with metrics designed for pre-AI workflows. Lines of code, PRs merged, and deployment frequency are all inflated by AI output, making it difficult to isolate the genuine productivity gain. The measurement gap -- not the technology itself -- is the primary obstacle to demonstrating ROI.
Larridin internal benchmarks are data points derived from production engineering data collected through Larridin's developer productivity measurement platform. They reflect real-world AI-assisted development workflows across multiple engineering organizations. These benchmarks are labeled separately from third-party research to maintain transparency about sourcing.
This page is updated as credible new research is published. The last_updated date in the page metadata reflects the most recent revision. If you are citing data from this page, check the last_updated date to ensure you are referencing the most current version.