Skip to main content
March 19, 2026

Last updated: March 1, 2026


TL;DR

  • The SPACE framework was a genuine contribution to developer productivity measurement -- multi-dimensional, developer-experience aware, and backed by research. Its core principle -- that productivity cannot be reduced to a single metric -- remains correct.
  • SPACE's Activity dimension (commits, PRs, lines of code) directly rewards volume. When AI coding tools inflate volume 3-5x, Activity becomes noise, not signal.
  • SPACE has no dimension for AI attribution or code durability. It cannot distinguish AI-generated code from human-written code, and its Performance dimension does not capture whether code survives beyond the first sprint.
  • The fix is evolution, not abandonment. Keep Satisfaction and Communication. Evolve Performance and Efficiency. Replace Activity with Complexity-Adjusted Throughput. Add AI Attribution as a new dimension.

What SPACE Gets Right

Before dissecting what needs to change, it is worth pausing on what SPACE got right -- because it got a lot right.

In 2021, Nicole Forsgren, Margaret-Anne Storey, Chandra Maddila, Thomas Zimmermann, Brian Houck, and Jenna Butler published "The SPACE of Developer Productivity", proposing that productivity should be understood across five dimensions: Satisfaction and well-being, Performance, Activity, Communication and collaboration, and Efficiency and flow.

This was a meaningful advance. Before SPACE, the industry was stuck in a loop: count lines of code, count commits, count story points, argue about whether any of those numbers meant anything, and repeat. SPACE broke that loop by insisting on three things that remain true today:

Productivity is multi-dimensional. No single metric captures developer productivity. An engineer who ships ten PRs but is burning out is not productive. A team with fast cycle times but poor code quality is not performing. SPACE forced leaders to hold multiple dimensions in view simultaneously. This was correct then. It is correct now.

Developer experience matters. SPACE's Satisfaction dimension -- encompassing job satisfaction, developer tool satisfaction, and retention -- was radical for an industry that had spent decades treating developer happiness as a soft, unmeasurable concern. The research was clear: satisfied developers are more productive, and satisfaction is measurable through well-designed surveys. This insight remains foundational. Any framework that ignores developer experience is incomplete.

Collaboration is invisible but critical. The Communication dimension acknowledged that much of a developer's most valuable work -- code review, mentoring, architectural guidance, unblocking teammates -- produces no artifacts that traditional metrics can count. A senior engineer who spends a week reviewing complex PRs and mentoring junior developers may have zero commits to show for it. SPACE said that was fine. That work counts.

These three principles are not artifacts of a pre-AI world. They are durable truths about how software gets built. Any framework that replaces SPACE should preserve them.


Where SPACE Falls Short in the AI Era

SPACE was published in 2021, before AI-assisted development existed at scale. It was designed for a world where humans wrote all the code, humans reviewed all the code, and the primary throughput constraint was the speed at which humans could type, think, and collaborate. That world no longer exists.

The framework's limitations are not evenly distributed. Some dimensions hold up well. Others are actively misleading in an AI-native context. Here is where the gaps are.

Activity Rewards Volume, Not Value

SPACE's Activity dimension includes metrics like the number of commits, pull requests, code reviews completed, and lines of code written. The framework's authors were careful to note that activity metrics should not be used in isolation. But activity metrics are the easiest to collect, the simplest to put on a dashboard, and the most tempting for leaders who want a quick read on output.

Before AI, activity metrics were imperfect but directionally useful. A developer who shipped five PRs per week was probably doing more than one who shipped one -- not always, but often enough that the signal had value.

AI destroyed that signal.

A developer using Cursor, GitHub Copilot, or Claude Code can generate five PRs in the time it used to take to write one. Lines of code per day can increase 3-5x. Commit frequency rises because AI-assisted workflows naturally produce more incremental changes. None of this necessarily means more value is being delivered.

Consider a concrete week for an AI-assisted developer:

  • Monday: AI generates a complete CRUD API with tests -- four files, 800 lines, three PRs
  • Tuesday: AI scaffolds a new microservice from a template -- twelve files, 2,000 lines, one PR
  • Wednesday: Developer spends the entire day on a complex distributed systems problem, writes 40 lines of careful coordination logic -- one PR
  • Thursday: AI generates migration scripts and documentation updates -- six files, 400 lines, two PRs
  • Friday: AI-assisted refactoring of legacy module -- eight files, 600 lines, two PRs

By SPACE's Activity metrics, this developer's best day was Tuesday (2,000 lines, twelve files). By any reasonable assessment of engineering value, it was Wednesday -- the 40 lines of distributed systems logic that required deep expertise and judgment.

SPACE's Activity dimension cannot make this distinction. It counts volume. When AI inflates volume, it counts inflated volume.

No AI Attribution

SPACE has no mechanism to distinguish AI-generated code from human-written code. This was not an oversight -- it was not a relevant question in 2021. In 2026, it is arguably the most important question in developer productivity measurement.

Without AI attribution, every SPACE dimension loses interpretive power:

  • Activity cannot separate AI-generated throughput from human effort
  • Performance cannot assess whether code quality differs by authorship source
  • Efficiency cannot determine whether cycle time improvements come from genuine process improvement or from AI accelerating the easy parts

When a team reports a 3x increase in Activity metrics after adopting AI tools, the natural follow-up question is: "How much of that increase reflects genuinely expanded engineering capacity, and how much is AI-generated boilerplate?" SPACE provides no way to answer this question.

The absence of AI attribution is not a missing feature. It is a structural gap that undermines the framework's ability to function in its intended role -- helping leaders understand developer productivity.

No Code Durability Signal

SPACE's Performance dimension includes metrics like code quality, reliability, and the absence of bugs. These are important. But they are trailing indicators -- they tell you about code that has already failed in production. What they do not tell you is whether code is durable: whether it survives contact with the broader codebase or gets quietly rewritten within days of being merged.

This gap matters more than ever. GitClear's research across over one billion lines of code shows that code churn -- the percentage of code rewritten or reverted within two weeks of being committed -- has roughly doubled since AI coding tools gained widespread adoption, rising from a pre-AI baseline of approximately 3.3% to between 5.7% and 7.1% by 2024.

Code that gets rewritten within two weeks is engineering waste. It consumed review cycles, CI/CD resources, and cognitive load -- and then it was thrown away. SPACE's Performance dimension does not capture this. A team could have excellent reliability scores, zero production incidents, and still be losing 7% of their engineering output to code that does not last.

Code Turnover Rate -- the percentage of code surviving 14 or 30 days without substantial modification -- is the quality signal that SPACE is missing. It is a leading indicator: it catches fragile, duplicative, or architecturally inconsistent code before it causes production failures, not after.

Efficiency Misses the New Bottleneck

SPACE's Efficiency dimension focuses on flow -- the ability of developers to complete work with minimal interruptions, handoffs, and delays. Metrics include cycle time, time in flow states, and the ratio of productive time to overhead.

Before AI, the primary bottleneck in the development workflow was writing code. Developers spent hours or days translating requirements into working software. Reviews, while sometimes slow, were proportional to the volume of human-written code.

AI shifted the bottleneck. Writing code is no longer the constraint. Reviewing it is.

When AI generates a complete pull request -- code, tests, documentation -- in minutes, the PR enters the review queue almost immediately. But reviewing AI-generated code takes longer than reviewing human-written code for several reasons:

  • The reviewer did not write the code and must build a mental model from scratch
  • AI-generated code often works but follows patterns the team does not use elsewhere
  • Subtle architectural mismatches are harder to spot in code the reviewer did not author
  • The sheer volume of AI-generated PRs overwhelms review capacity

Teams adopting AI coding tools report a consistent pattern: lead time drops because code is generated faster, but the percentage of that lead time spent waiting for review grows from roughly 20% to over 50%. The bottleneck has moved downstream, and SPACE's Efficiency dimension -- focused on overall flow and developer time-in-state -- does not specifically surface the review bottleneck as the critical constraint.

A framework designed for 2026 needs to measure review throughput, review depth, and review queue time as first-class efficiency signals -- not just overall cycle time.


The Data: Volume Up, Durability Down

The claim that AI inflates volume metrics while degrading code durability is not theoretical. The data is clear and growing.

AI's impact on developer productivity signals
Metric Pre-AI Baseline Post-AI (2024-2025) Source
Code churn rate ~3.3% 5.7-7.1% GitClear, 2024
Lines of code per developer per day 100-150 300-500+ Directional; based on GitHub Copilot research and aggregated team data
PRs merged per developer per week 3-5 8-15 Directional; aggregated engineering team data
Commit frequency 5-8/day 12-20/day Directional; aggregated engineering team data
Time spent in code review (% of lead time) ~20% >50% Directional; aggregated engineering team data
Moved/deleted/copy-pasted lines (% of changes) ~35% ~45% GitClear, 2024

Every metric in SPACE's Activity dimension -- commits, PRs, lines of code -- shows dramatic inflation. Meanwhile, the quality signals that SPACE does not measure -- code churn, code durability, review bottleneck severity -- are moving in the wrong direction.

This is not a subtle problem. SPACE's Activity dimension is producing numbers that look like productivity gains and are, in part, measurement artifacts of AI-generated volume. Teams that rely on Activity metrics to assess performance are making decisions on distorted data.


How to Evolve SPACE for AI-Native Teams

SPACE does not need to be discarded. It needs to be evolved. The framework's core insight -- that productivity is multi-dimensional -- remains correct. What needs to change are the specific dimensions and the metrics within them.

Keep: Satisfaction and Communication

Satisfaction and well-being is more important than ever. AI tools change how developers work, and not all of those changes are positive. Developers report AI-related frustration when AI-generated code introduces bugs they did not write but must debug, when AI suggestions interrupt their flow state, or when they feel their expertise is being devalued. Measuring satisfaction through well-designed developer experience surveys remains essential -- but the survey instruments themselves need to be updated to capture AI-specific concerns.

Communication and collaboration retains its value. AI has not replaced the need for code review conversations, architectural discussions, or cross-team coordination. If anything, collaboration is more important when a significant share of code is AI-generated -- review quality, knowledge sharing about AI tool usage patterns, and alignment on when to use (and not use) AI assistance all depend on effective communication.

Evolve: Performance, Activity, and Efficiency

Performance should add Code Turnover Rate. SPACE's existing Performance metrics -- reliability, absence of bugs, code quality -- remain relevant but insufficient. Code Turnover Rate fills the gap by measuring the percentage of code surviving 14 or 30 days without being rewritten or reverted. This metric directly captures the durability problem that AI-generated code introduces. A team with a Code Turnover Rate under 3% is producing durable code. A team at 7% or higher is producing a significant amount of engineering waste -- regardless of how their production reliability looks.

Activity should be replaced with Complexity-Adjusted Throughput. The Activity dimension in its current form -- counting commits, PRs, and lines of code -- is no longer interpretable when AI generates a large share of those artifacts. Complexity-Adjusted Throughput (CAT) replaces raw volume with difficulty-weighted output. Each pull request is scored by complexity: Easy (1 point), Medium (3 points), Hard (8 points). An engineer who ships one Hard PR and two Easy PRs scores 10 points. An engineer who ships ten AI-generated Easy PRs also scores 10 points. The metric reflects cognitive effort, not line count.

CAT is resistant to AI inflation by design. AI excels at Easy work -- boilerplate, configuration, scaffolding. It struggles with Hard work -- architectural decisions, complex algorithms, cross-system integrations. Weighting by complexity means that AI-inflated volume does not distort the signal.

Efficiency should measure the review bottleneck explicitly. SPACE's Efficiency dimension should retain its focus on flow states and developer time, but it needs to add metrics that surface the review bottleneck:

  • Review queue time: How long do PRs wait before a reviewer picks them up?
  • Review depth: Are reviews substantive (comments on architecture, logic, edge cases) or superficial (approved without comment)?
  • Review load distribution: Is review work concentrated on a few senior engineers, or distributed across the team?

These metrics acknowledge the reality that writing code is no longer the constraint. When AI generates code in minutes and PRs wait hours or days for review, efficiency must be measured at the bottleneck, not averaged across the pipeline.

Add: AI Attribution

SPACE needs a sixth dimension -- or, more precisely, an attribution layer that cuts across all existing dimensions.

AI Attribution answers a foundational question: for any given metric, what share of the underlying work was AI-generated versus human-written? Without this layer, every dimension is ambiguous:

  • Is Activity high because the team is productive, or because AI is generating boilerplate?
  • Is Performance strong because code quality is high, or because AI-generated code happens to pass tests while introducing long-term maintenance burden?
  • Is Efficiency improving because the team's processes are better, or because AI is accelerating the easy parts while review capacity remains static?

AI Attribution is not a single metric. It is a segmentation that should be applied across every other dimension. When a team reports their CAT score, they should report it split by AI-assisted versus human-only work. When they measure Code Turnover Rate, they should know the turnover rate for AI-generated code versus human-written code separately.

This segmentation transforms SPACE from a descriptive framework ("here is how productive we are") into a diagnostic one ("here is how productive we are, and here is exactly how AI is affecting that productivity").

SPACE dimensions: original versus AI-era evolution
SPACE Dimension Original Metrics AI-Era Evolution Why It Changes
Satisfaction Job satisfaction, tool satisfaction, retention Keep; add AI-specific survey items AI introduces new satisfaction drivers and frustrations
Performance Reliability, absence of bugs, quality Add Code Turnover Rate Quality must include durability, not just correctness
Activity Commits, PRs, LOC, code reviews Replace with Complexity-Adjusted Throughput Volume metrics inflate 3-5x with AI; no longer interpretable
Communication Review discussions, knowledge sharing, mentoring Keep; add review depth metrics AI increases review volume; quality of review matters more
Efficiency Cycle time, flow state time, interruptions Add review queue time, review load distribution Bottleneck shifted from writing to reviewing
AI Attribution (new) N/A AI code share segmented across all dimensions Without attribution, every metric is ambiguous

The Developer AI Impact Framework

The evolution of SPACE described above is not theoretical. The Developer AI Impact Framework, developed by Larridin, operationalizes these changes across five pillars: AI Adoption, AI Code Share, Complexity-Adjusted Throughput, Code Quality (measured by turnover rate), and Cost and ROI.

Where SPACE provides conceptual dimensions, the Developer AI Impact Framework provides specific metrics, measurement approaches, and benchmarks. It preserves SPACE's multi-dimensional philosophy while solving the framework's three critical gaps: AI attribution, code durability, and complexity-weighted throughput.

For teams that have built their engineering measurement practice around SPACE, the transition is straightforward. SPACE's Satisfaction and Communication dimensions map directly. Performance gains Code Turnover Rate. Activity becomes CAT. Efficiency adds review-specific signals. And AI Attribution becomes the segmentation layer that makes every other metric interpretable.

This is not a framework war. It is the natural evolution of developer productivity measurement for a world where AI writes a significant share of the code. SPACE was the right framework for 2021. The question is whether your measurement practice has evolved with the technology -- or whether you are still measuring a world that no longer exists.

For the full framework, metrics, and implementation guidance, see: The Developer AI Impact Framework.


Frequently Asked Questions

Does the SPACE framework still work in 2026?

SPACE's core principle -- that developer productivity is multi-dimensional and cannot be reduced to a single number -- remains correct. Its Satisfaction and Communication dimensions are as relevant now as they were in 2021. However, the Activity dimension has become unreliable because AI coding tools inflate every volume-based metric it includes (commits, PRs, lines of code). The Performance dimension lacks code durability signals, and the Efficiency dimension does not surface the review bottleneck that has become the primary constraint in AI-assisted workflows. Teams using SPACE in 2026 need to evolve three of its five dimensions to get accurate productivity signals.

What is wrong with measuring developer activity?

The problem is not with measuring activity in principle -- it is with using volume-based proxies (commits, PRs, lines of code) as activity metrics when AI inflates all of them. Before AI, a developer who shipped five PRs per week was probably doing more work than one who shipped one. That correlation was imperfect but directionally useful. In 2026, a developer can generate five PRs in an hour using AI tools -- boilerplate, scaffolding, test files, configuration changes. Activity metrics now measure "how much code was produced" rather than "how much valuable engineering work was done." The fix is to replace raw volume with Complexity-Adjusted Throughput, which weights output by difficulty rather than counting artifacts.

How should you measure developer productivity instead of SPACE?

Start with SPACE's multi-dimensional philosophy, then update the specifics. Keep Satisfaction (with AI-specific survey items) and Communication (with review depth metrics). Replace Activity with Complexity-Adjusted Throughput, which weights pull requests by difficulty (Easy=1, Medium=3, Hard=8) instead of counting them. Add Code Turnover Rate to the Performance dimension to measure whether code survives beyond the first sprint. Evolve Efficiency to measure the review bottleneck -- queue time, depth, and load distribution. Finally, add AI Attribution as a segmentation layer across all dimensions so you can distinguish AI-generated output from human-written output. The Developer AI Impact Framework, developed by Larridin, operationalizes this evolution into specific metrics and benchmarks.

Is SPACE compatible with AI-native metrics?

Partially. SPACE's dimensional structure -- Satisfaction, Performance, Activity, Communication, Efficiency -- is flexible enough to accommodate new metrics within each dimension. You can add Code Turnover Rate to Performance, for example, without breaking the framework's logic. The harder problem is Activity. SPACE defines Activity using volume-based metrics (commits, PRs, lines of code, design documents produced). These metrics are structurally compromised by AI-generated output. Replacing them with complexity-weighted measures requires redefining what the Activity dimension means -- at which point you are no longer using SPACE as designed but rather evolving it into something new. The other structural gap is AI Attribution, which has no natural home in SPACE's five dimensions and must be added as a cross-cutting layer.

What replaced SPACE for AI-native teams?

No single framework has achieved the universal adoption that SPACE enjoyed in 2021-2023. The most comprehensive alternative is the Developer AI Impact Framework, developed by Larridin, which preserves SPACE's multi-dimensional philosophy while adding AI-specific measurement capabilities: AI Adoption and code share tracking, Complexity-Adjusted Throughput instead of volume-based activity metrics, Code Turnover Rate for durability measurement, and cost/ROI analysis for AI tool investment. Other teams have extended SPACE informally by adding AI-specific metrics within the existing dimensions. The common thread across all approaches is the recognition that measuring developer productivity in 2026 requires AI attribution, code durability signals, and complexity-weighted throughput -- none of which SPACE provides natively.


Related Resources

  • The Developer AI Impact Framework -- the comprehensive framework for measuring developer productivity when AI writes the code, developed by Larridin
  • Why DORA Metrics Break in the AI Era -- companion analysis of how DORA's four metrics are affected by AI-assisted development
  • What Is Complexity-Adjusted Throughput? -- the metric that replaces SPACE's Activity dimension
  • Code Turnover Rate: The AI Quality Metric -- the durability signal SPACE's Performance dimension is missing
  • The SPACE Framework Explained (coming soon) -- a neutral reference guide to the original SPACE framework and its five dimensions
  • Developer Experience Surveys for AI-Native Teams (coming soon) -- how to update SPACE's Satisfaction dimension for AI-era concerns

Explore More from Larridin