Skip to main content

TL;DR

  • The AI Value Realization Score (AVRS) is a composite metric that combines AI adoption, AI code share, perceived developer time savings, and code quality into a single 0-100 score for executive reporting on AI engineering ROI.
  • The formula weights quality and perceived impact more heavily than adoption and volume: AVRS = (WAU x 0.2) + (AI Code Rate x 0.2) + (Perceived Time Savings x 0.3) + (Quality Score x 0.3). Quality and perceived savings each contribute 30% because adoption without quality is waste, and volume without perceived benefit signals tool friction.
  • Score ranges: 0-30 Early, 30-50 Developing, 50-70 Maturing, 70-85 Advanced, 85-100 Elite. Most organizations in early 2026 fall in the 25-45 range -- past initial adoption, still building quality and workflow integration.
  • AVRS solves a specific executive problem: the VP of Engineering who has four dashboards, twelve metrics, and no single answer to "Is our AI investment working?" AVRS compresses complexity into a single trackable number without hiding the components.

Why a Composite Score?

Engineering leaders measuring AI impact face a fragmentation problem. They have adoption dashboards showing Weekly Active Users. They have AI code share data showing what percentage of code AI produces. They have developer survey results showing perceived time savings. They have quality metrics showing code turnover rate and defect density. Each metric tells part of the story. None of them answers the question the CFO is actually asking: is this investment working?

The fragmentation creates two failure modes.

Cherry-picking. Leaders select whichever metric looks best this quarter. Adoption is up 15%? Lead with that. Code share is flat but quality improved? Lead with quality. This is not dishonest -- each metric is real -- but it produces a shifting narrative that erodes confidence in the measurement program. Executives learn to distrust metric-of-the-month reporting.

Metric paralysis. Leaders present all twelve metrics on a single slide. The executive audience absorbs none of them. Too many numbers with too little synthesis produces the same outcome as no numbers: decisions get made on gut feeling, anecdote, and vendor marketing.

The AI Value Realization Score addresses both failure modes by compressing four dimensions into a single number. It is not a replacement for the underlying metrics -- each component should still be tracked and analyzed independently. It is a summary layer that gives executives a single trend line and engineering leaders a single target to optimize.

Composite scores carry risks. They can obscure problems by averaging a strong dimension with a weak one. They can create Goodhart's Law dynamics where teams optimize the score rather than the underlying behaviors. AVRS mitigates these risks through transparent component weights and explicit quality emphasis -- but the risks are real, and the underlying component metrics should always be accessible alongside the composite.

The Formula

AVRS = (WAU x 0.2) + (AI Code Rate x 0.2) + (Perceived Time Savings x 0.3) + (Quality Score x 0.3)

Each component is normalized to a 0-100 scale before weighting. The result is a single number between 0 and 100.

Component 1: Weekly Active Users (WAU) -- Weight: 0.2

What it measures: The percentage of licensed developers who actively used an AI coding tool at least once during the measurement week. "Active use" means generating, accepting, or interacting with at least one AI suggestion -- not merely having the tool installed.

How to normalize: WAU is already a percentage (0-100), so it maps directly to the 0-100 scale. A team with 65% WAU contributes 65 x 0.2 = 13 points to the composite.

Data sources: Copilot Metrics API (active user counts), Cursor license telemetry, Claude Code session logs, SSO-integrated tool usage dashboards.

Why 0.2 weight: Adoption is a prerequisite, not an outcome. A team cannot realize AI value without developers actually using the tools. But adoption alone proves nothing about impact -- a team can have 90% WAU and negligible productivity or quality improvement. Adoption gets the lowest weight (tied with code share) because it is the input, not the result.

Component 2: AI Code Rate -- Weight: 0.2

What it measures: The percentage of committed code that was AI-generated or AI-assisted, measured at the line level. This is the AI Code Share metric -- how much of your codebase AI actually produces.

How to normalize: AI Code Rate is a percentage. However, because even elite teams rarely exceed 75% AI-assisted lines, apply a cap-adjusted normalization: divide the raw percentage by 0.75, cap at 100. A team with 45% AI-assisted lines scores (45/0.75) = 60 on this component, contributing 60 x 0.2 = 12 points.

Data sources: Editor telemetry (Copilot Metrics API, Cursor analytics), commit metadata, Claude Code OTEL traces.

Why 0.2 weight: Like adoption, AI code share is a volume metric, not a value metric. High AI code share without quality is a vanity metric. A team generating 60% AI-assisted code that churns at 25% within 30 days is not realizing value -- they are generating rework. Code rate gets the same weight as adoption because it measures composition, not impact.

Component 3: Perceived Time Savings -- Weight: 0.3

What it measures: Developer-reported time savings from AI tools, collected via periodic survey. The survey asks a specific question: "Compared to working without AI coding tools, how much time do you estimate AI saves you in a typical work week?" Responses are captured as a percentage (0-100%) or converted from hours saved relative to total coding hours.

How to normalize: Survey responses are averaged across the measured population and expressed as a percentage. A team whose developers report an average of 25% time savings scores 25 on this component, contributing 25 x 0.3 = 7.5 points.

Data sources: Quarterly or monthly developer surveys. GitHub's Copilot research reports that developers feel 55% more productive with AI tools1, though self-reported time savings tend to be more conservative than self-reported productivity gains. Block reports that their AI-assisted developers experience significant perceived productivity improvements2.

Why 0.3 weight: Perceived time savings captures something telemetry cannot: whether developers actually experience AI tools as valuable in their daily workflow. A tool can have high adoption (people use it because it is mandated), high code share (it generates lots of code), and poor perceived value (developers feel it slows them down with bad suggestions they have to filter). The perceived savings component catches this disconnect. It gets a higher weight than adoption or code share because perceived value correlates more strongly with sustained adoption, effective usage patterns, and ultimately with engineering outcomes that show up in throughput and quality metrics. Developer experience is the leading indicator for AI tool ROI.

Component 4: Quality Score -- Weight: 0.3

What it measures: A derived score that penalizes low code quality in AI-generated output. Quality Score is not a single raw metric but a computed value based on two quality indicators:

Quality Score = 100 - (Turnover Rate x 2) - (Post-Acceptance Edit Rate x 1.5)
  • Turnover Rate is the 30-day code turnover rate for AI-generated code, expressed as a percentage.
  • Post-Acceptance Edit Rate is the percentage of accepted AI suggestions that are substantially modified before commit.

A team with 10% AI code turnover and 20% post-acceptance edit rate scores: 100 - (10 x 2) - (20 x 1.5) = 100 - 20 - 30 = 50. This 50 contributes 50 x 0.3 = 15 points to the composite.

The multipliers (2x for turnover, 1.5x for post-acceptance edits) reflect the relative severity of each quality signal. Turnover is penalized more heavily because it represents code that shipped to the codebase and then failed -- a more expensive failure than code that was edited before commit. Post-acceptance edits are penalized less because editing before commit is a sign of appropriate human review, even though high rates suggest the AI's initial output was not production-ready.

Data sources: Git analysis (code turnover), editor telemetry (post-acceptance edit rate), CI/CD pipeline metrics.

Why 0.3 weight: Quality is the guardrail that prevents AVRS from rewarding adoption theater. Without the quality component, a team could score well by simply adopting AI tools widely and generating high volumes of AI code -- regardless of whether that code was durable. The 0.3 weight ensures that declining quality drags the composite score down even if adoption, code share, and perceived savings are all strong. This is by design: an organization that generates a lot of AI code, reports high perceived savings, but has rising code turnover is not realizing value -- it is deferring cost.

Why the Weights Are 0.2 / 0.2 / 0.3 / 0.3

The asymmetric weighting is deliberate. Adoption and code share are inputs. Perceived savings and quality are outcomes. AVRS weights outcomes more heavily than inputs because the purpose of the score is to measure value realization, not activity.

An alternative weighting of 0.25 / 0.25 / 0.25 / 0.25 would treat all dimensions equally. This fails because it allows high adoption and high code share (both easy to achieve) to compensate for poor quality or low perceived value (the dimensions that actually matter). A team with 90% WAU, 60% AI code share, 10% perceived savings, and a Quality Score of 30 should not score well -- but under equal weights, they would score 47.5. Under AVRS weights, they score 90(0.2) + 60(0.2) + 10(0.3) + 30(0.3) = 18 + 12 + 3 + 9 = 42. The gap between 47.5 and 42 matters: it is the difference between "Developing" and "Early-to-Developing" -- a more accurate reflection of a team that has adopted AI broadly but has not yet realized value from it.

Score Ranges

Range Level Description
0-30 Early AI tools are deployed but not yet generating measurable value. Adoption may be low, quality metrics are not yet tracked, or developer perception is neutral to negative. Most organizations land here within the first 3-6 months of AI tool deployment.
30-50 Developing Adoption is established and some value signals are emerging. Developers report moderate time savings. Code quality may still be uneven. This is where most organizations sat in late 2025 through early 2026.
50-70 Maturing AI tools are integrated into daily workflows with measurable quality and productivity outcomes. Code turnover for AI-generated code is controlled. Developers report meaningful time savings. The organization is ready to scale AI-native practices.
70-85 Advanced Strong results across all four dimensions. AI is deeply embedded in engineering workflows, code quality is stable or improving, and developer experience with AI tools is consistently positive. Few organizations reach this level without deliberate investment in enablement, prompt engineering, and quality measurement.
85-100 Elite Exceptional AI value realization. High adoption, high-quality AI code, strong perceived savings, and low rework rates. This level requires mature measurement infrastructure, continuous improvement practices, and organizational alignment between engineering and leadership on AI strategy.

Most engineering organizations in early 2026 score between 25 and 45 (Larridin internal benchmark). They have cleared the initial adoption hurdle but have not yet built the quality measurement and enablement infrastructure needed to move into the Maturing range. The most common blocker is Component 4 (Quality Score): organizations track adoption and code share but have not instrumented code turnover or post-acceptance edit rate, leaving the quality component either unmeasured or estimated.

Worked Example

Consider an engineering organization with the following raw metrics:

  • WAU: 72% of licensed developers are active weekly users of AI coding tools.
  • AI Code Rate: 35% of committed lines are AI-assisted.
  • Perceived Time Savings: Developer survey reports an average of 28% time savings.
  • AI Code Turnover (30D): 12% of AI-generated code is rewritten or deleted within 30 days.
  • Post-Acceptance Edit Rate: 22% of accepted suggestions are substantially modified before commit.

Step 1: Normalize each component.

  • WAU = 72 (already 0-100)
  • AI Code Rate = 35 / 0.75 = 46.7 (cap-adjusted)
  • Perceived Time Savings = 28 (already 0-100)
  • Quality Score = 100 - (12 x 2) - (22 x 1.5) = 100 - 24 - 33 = 43

Step 2: Apply weights.

  • WAU contribution: 72 x 0.2 = 14.4
  • AI Code Rate contribution: 46.7 x 0.2 = 9.3
  • Perceived Time Savings contribution: 28 x 0.3 = 8.4
  • Quality Score contribution: 43 x 0.3 = 12.9

Step 3: Sum.

AVRS = 14.4 + 9.3 + 8.4 + 12.9 = 45.0

This organization scores 45 -- solidly in the Developing range, approaching Maturing. The adoption metrics (WAU and code rate) are healthy. The perceived savings are moderate. The quality score is the primary drag: a 12% AI code turnover rate and 22% post-acceptance edit rate indicate that AI-generated code is functional but requires non-trivial rework. The path to Maturing (50+) runs through quality improvement: better prompt engineering, stricter review standards for AI-generated PRs, and codebase context enrichment for AI tools.

How to Implement AVRS

Data collection

Each component requires a specific data pipeline:

  1. WAU: Pull from AI tool vendor dashboards or APIs. Copilot Metrics API provides this directly. For multi-tool environments, union the active user sets across tools, deduplicating by developer identity.

  2. AI Code Rate: Pull from editor telemetry and commit metadata. If precise attribution is not available, use PR-level AI tagging and estimate line-level share from the AI-assisted PR ratio. Even approximate data is better than no data.

  3. Perceived Time Savings: Run a quarterly or monthly developer survey with a single quantitative question: "What percentage of your coding time do AI tools save you in a typical week?" Supplement with an optional free-text field for context. Keep the survey short -- one to three questions -- to maximize response rates.

  4. Quality Score: Compute from git analysis (code turnover) and editor telemetry (post-acceptance edit rate). If post-acceptance edit rate is not instrumented, use code turnover alone with an adjusted formula: Quality Score = 100 - (Turnover Rate x 3). This places more weight on turnover to compensate for the missing edit rate signal.

Reporting cadence

  • Monthly: Compute and report AVRS at the organizational level. Share the composite and all four components.
  • Quarterly: Conduct deeper analysis of component trends. Identify which components are improving, which are flat, and which are declining. Present to engineering leadership with specific recommendations.
  • Per team (optional): Compute AVRS at the team level for internal benchmarking. This enables teams to compare their AI value realization trajectory and learn from high-performing peers.

Avoiding Goodhart's Law

The biggest risk of any composite score is that teams optimize for the number rather than the behaviors the number is designed to measure. Three safeguards:

  1. Always present components alongside the composite. AVRS is a summary, not a replacement. Any executive presentation that shows AVRS should also show the four component values.

  2. Do not set AVRS targets without quality floors. If you set a goal of "AVRS > 60 by Q4," pair it with a constraint: "with Quality Score > 40." This prevents teams from inflating adoption and perceived savings while ignoring quality.

  3. Weight adjustments are a leadership conversation. The 0.2/0.2/0.3/0.3 weights reflect a default philosophy: outcomes matter more than inputs. If your organization's strategic priority is adoption breadth (e.g., during initial rollout), temporarily adjusting WAU to 0.3 and reducing another component is legitimate -- as long as the adjustment is explicit and time-bounded.

How AVRS Fits the Developer AI Impact Framework

The AI Value Realization Score is a cross-pillar composite within Larridin's Developer AI Impact Framework. It draws one metric from each of the framework's five pillars:

AVRS Component Framework Pillar Key Metric
WAU Pillar 1: AI Adoption Weekly Active Users %
AI Code Rate Pillar 2: AI Code Share AI-Assisted Lines %
Perceived Time Savings Pillar 3: Throughput (subjective) Developer survey
Quality Score Pillar 4: Quality Code Turnover Rate, Post-Acceptance Edit Rate

AVRS does not cover Pillar 5 (Cost & ROI) directly. Cost metrics -- license spend per developer, cost per AI-assisted complexity point -- are deliberately excluded from the composite because they vary too much by organization size, tool selection, and contract structure to be meaningfully normalized. Organizations that need a cost-inclusive view should report AVRS alongside cost-per-point-of-AVRS: total AI tool spend divided by AVRS, producing a dollar figure per unit of realized value.

Read the full Developer AI Impact Framework -->

Frequently Asked Questions

What is the AI Value Realization Score?

The AI Value Realization Score (AVRS) is a composite 0-100 metric that measures how much value an engineering organization is extracting from its AI coding tool investment. It combines four components -- Weekly Active Users (adoption), AI Code Rate (volume), Perceived Time Savings (developer experience), and Quality Score (code durability) -- into a single number for executive reporting. The formula weights quality and perceived impact at 0.3 each, and adoption and code volume at 0.2 each, reflecting the principle that outcomes matter more than inputs.

How do you measure AI ROI for engineering teams?

AI ROI for engineering teams is best measured through a composite that balances adoption, output, developer experience, and code quality -- not through any single metric. Measuring adoption alone (how many developers use AI tools) misses whether the tools deliver value. Measuring output alone (how much AI code is generated) misses whether that code is durable. AVRS combines these dimensions with perceived time savings and a quality score derived from code turnover rate and post-acceptance edit rate. For dollar-denominated ROI, pair AVRS with cost-per-point analysis: total AI tool spend divided by AVRS score.

Why are quality and perceived savings weighted higher than adoption?

Quality and perceived savings are weighted at 0.3 each (vs. 0.2 for adoption and code share) because they measure outcomes, not inputs. High adoption and high AI code share are easy to achieve -- mandate tool installation and encourage aggressive use, and both numbers rise. But those numbers mean nothing if the resulting code churns at 25% within a month or if developers report that AI tools create more friction than they resolve. The asymmetric weighting ensures that AVRS cannot be inflated by high adoption alone. An organization that adopts widely but delivers low quality will score poorly, as it should.

What is a good AI Value Realization Score?

Most engineering organizations in early 2026 score between 25 and 45 (Developing range), with the target for mature organizations being 50-70 (Maturing). Scores below 30 indicate early-stage adoption where AI tools are deployed but not yet generating measurable value. Scores above 70 (Advanced) require mature measurement infrastructure, strong enablement programs, and sustained quality discipline. Elite scores above 85 are rare and typically require organizational alignment between engineering, leadership, and finance on AI strategy and measurement (Larridin internal benchmark).

Can AVRS be gamed?

Yes, like any composite metric, AVRS can be gamed -- which is why the underlying components should always be reported alongside the composite. The most common gaming vectors are inflating perceived time savings through optimistic survey responses and boosting WAU by counting passive tool installation as active use. The quality component is harder to game because it is derived from objective git analysis. Safeguards include: always presenting the four components alongside the composite, setting quality floor constraints alongside AVRS targets, and validating perceived savings against objective telemetry where possible.

Footnotes

Data sources and methodology:

Related Resources


  1. GitHub, "Research: Quantifying GitHub Copilot's Impact in the Enterprise with Accenture" (2024). Reports that developers using Copilot feel 55% more productive, with measurable improvements in task completion time. 

  2. Block, "AI-Assisted Development at Block". Reports approximately 95% of engineers regularly using AI to assist development, with significant perceived productivity improvements among intensive users.