Skip to main content

TL;DR

  • AI shifted the PR bottleneck from writing code to reviewing it. When code generation takes minutes instead of hours, the constraint on delivery speed moves downstream to the review queue. Time-to-first-review is now the dominant factor in PR cycle time for AI-native teams.
  • AI-assisted PRs show shorter coding phases but longer review waits. The PR is created faster, but it sits longer before a reviewer picks it up -- and the review itself takes longer because the volume and density of AI-generated code demands more scrutiny.
  • The key diagnostic is segmenting PR cycle time by AI-assisted vs. human-only PRs. Without this segmentation, aggregated cycle time can look stable or even improved while a growing review bottleneck hides beneath the average.
  • Time-to-first-review is the metric that matters most in 2026. Teams that optimize for coding speed without addressing review capacity are accelerating into a wall.

The Bottleneck Shift

Before AI coding tools, PR cycle time decomposed roughly like this in a typical product engineering team:

Phase Pre-AI Share of Cycle Time
Coding (branch to PR open) 50-65%
Time-to-first-review 15-20%
Review + iteration 15-25%
Post-approval merge 2-5%

Coding was the bottleneck. An engineer spent the majority of the cycle writing, testing, and refining the code before opening a PR. Time-to-first-review was a secondary delay -- usually a matter of hours, sometimes a day. Review itself was proportional to PR size, which was naturally constrained by how much code a human could write in a reasonable work session.

AI tools compressed the coding phase. With AI-assisted development, the time from branch creation to PR open dropped significantly -- in some teams by 40-60%, depending on task complexity and AI tool maturity. Block reports that AI-assisted developers experience substantial acceleration in initial code creation1. GitHub's research with Accenture found measurable reductions in task completion time for Copilot users2.

But the downstream phases did not compress proportionally. In many teams, they expanded.

Phase Post-AI Share of Cycle Time
Coding (branch to PR open) 25-40%
Time-to-first-review 25-35%
Review + iteration 25-35%
Post-approval merge 2-5%

The coding phase shrank. Time-to-first-review grew -- both in absolute terms (more PRs competing for reviewer attention) and as a share of total cycle time. The review phase itself grew because AI-generated PRs tend to be larger, denser, and require more careful scrutiny than human-written PRs of equivalent scope.

The result: total PR cycle time may be flat or only modestly improved, despite a dramatic acceleration in the coding phase. The bottleneck moved. Teams that do not measure where time is spent within the PR lifecycle will not see this shift.

Why Review Becomes the Bottleneck

Three dynamics drive the bottleneck shift.

More PRs per developer

When coding takes less time, developers produce more PRs per week. A developer who previously opened 3-4 PRs per week might open 5-7 with AI assistance. But the team's review capacity did not increase by the same factor. Reviewer count is fixed in the short term. Each reviewer now has more PRs in their queue, and the wait time for first review extends.

This is a classic queuing theory problem. When the arrival rate of work items (PRs) increases without a proportional increase in service capacity (reviewer bandwidth), queue length and wait time grow non-linearly. A 50% increase in PR volume can produce a 100%+ increase in review wait time if reviewers were already near capacity.

AI-generated PRs demand more review effort

Not all PRs are equal in review difficulty. AI-generated code presents specific review challenges:

  • Larger diffs. AI tools can generate substantial amounts of code quickly, producing PRs that are bigger than what a developer would typically create in a single work session. Larger diffs take longer to review.
  • Unfamiliar patterns. AI-generated code may use patterns, libraries, or approaches that the author did not choose deliberately. The reviewer must evaluate not just correctness but whether the AI's design choices align with the team's architectural conventions.
  • Subtle quality issues. GitClear's longitudinal analysis shows code churn rising from 3.3% to 5.7-7.1% coinciding with AI tool adoption34. AI-generated code can look syntactically clean and well-structured while containing duplicative logic, inappropriate abstractions, or subtle coupling. These issues are harder to spot and take more reviewer time per line than obvious bugs.
  • Review fatigue. When a reviewer knows a PR is AI-generated, there is a cognitive tension: the code looks polished and well-formatted (AI is good at surface-level quality), which creates a temptation to skim rather than scrutinize. Some reviewers resist this temptation and spend extra time; others succumb to it. Both responses affect cycle time -- the first by extending review duration, the second by allowing quality issues through that later manifest as code turnover.

Review capacity is not scaling

Most teams have not adjusted their review processes or staffing for the increased PR volume that AI tools generate. Common review models -- round-robin assignment, code ownership-based routing, ad-hoc reviewer selection -- were designed for pre-AI PR volumes. They break down when PR volume increases 40-60% without additional reviewers.

Some organizations have experimented with AI-assisted code review to address the capacity gap. These tools can flag common issues, check for style consistency, and identify potential bugs -- but they do not replace human review for architectural alignment, business logic correctness, or integration concerns. AI-assisted review reduces the effort per review but does not eliminate the bottleneck.

How to Measure PR Cycle Time in the AI Era

Decompose the cycle

Total PR cycle time -- from branch creation to merge -- is too coarse to be diagnostic. Break it into four phases:

  1. Coding time: From branch creation (or first commit) to PR open. This measures how long the author spent writing code before requesting review.

  2. Time-to-first-review: From PR open to the first substantive reviewer comment or approval. This measures how long the PR waited in the review queue. "Substantive" excludes automated bot comments and CI status updates.

  3. Review duration: From first review comment to final approval. This measures the back-and-forth between author and reviewer -- review comments, requested changes, follow-up commits, and re-reviews.

  4. Merge time: From final approval to merge. This captures process overhead -- merge queue delays, CI re-runs, manual merge steps.

Each phase has different dynamics in AI-assisted workflows, and optimizing the wrong phase is worse than not optimizing at all.

Segment by AI attribution

The single most important analytical step is segmenting all cycle time metrics by AI attribution:

  • AI-assisted PRs: PRs where the author used AI coding tools during development, identified via editor telemetry, commit metadata, or PR labels.
  • Human-only PRs: PRs with no detected AI tool involvement.

This segmentation reveals the bottleneck shift. In aggregated data, AI-assisted PRs (faster coding phase) and human-only PRs (slower coding phase) average together, producing a total cycle time that looks stable. Segmented data shows the divergence: AI-assisted PRs have shorter coding phases but longer review waits.

Without this segmentation, a team might see their overall median PR cycle time hold steady at 28 hours and conclude that AI tools have had no negative impact on their delivery pipeline. Segmented analysis might reveal that AI-assisted PRs have a 16-hour median (fast coding, long review wait) and human-only PRs have a 34-hour median (normal coding, normal review wait) -- and the aggregate is stable only because the mix of PR types has not changed enough to shift the average.

Benchmarks

These benchmarks reflect patterns across product engineering teams in early 2026. They represent medians; individual team performance varies based on team size, PR size norms, review culture, and timezone distribution.

Phase Human-Only PRs (Median) AI-Assisted PRs (Median) Healthy Target
Coding time 6-10 hours 2-5 hours Not a target -- varies by task
Time-to-first-review 3-6 hours 5-10 hours <4 hours
Review duration 4-8 hours 6-12 hours <8 hours
Merge time 0.5-2 hours 0.5-2 hours <1 hour
Total cycle time 16-26 hours 16-29 hours <24 hours

The pattern is visible in the medians: AI-assisted PRs gain 4-5 hours in coding time but lose 2-4 hours in review wait and 2-4 hours in review duration. The net effect on total cycle time ranges from modest improvement to slight degradation -- which is why aggregate cycle time alone does not capture the bottleneck shift.

The healthy target for time-to-first-review is under 4 hours regardless of PR type. Teams that achieve this consistently have either invested in review capacity (more reviewers, dedicated review time blocks), optimized PR routing (automated assignment based on expertise and availability), or reduced PR size norms to make reviews faster and less daunting.

Interventions: Unblocking the Review Bottleneck

Reduce PR size for AI-assisted work

The most effective intervention is also the simplest: smaller PRs. AI tools make it easy to generate large PRs quickly. But a 500-line AI-generated PR that took 30 minutes to create will take a reviewer an hour to evaluate -- and the review quality will be poor because no one reviews 500 lines carefully.

Set explicit PR size guidelines for AI-assisted work. A reasonable norm: AI-assisted PRs should not exceed the same size guidelines as human-written PRs. If your team targets PRs under 200 lines of diff, that target applies regardless of whether AI generated the code in 5 minutes or a human wrote it in 5 hours. The constraint is reviewer capacity, not author effort.

Dedicate review time blocks

When PR volume rises, treating review as an interrupt-driven activity breaks down. Engineers context-switch into review mode, spend 10 minutes on a PR, get pulled back to their own work, and the PR sits until they remember to return. The result is long time-to-first-review and fragmented review sessions that miss subtle issues.

Dedicated review blocks -- scheduled time where engineers focus exclusively on reviewing -- reduce both wait time and review duration. Common patterns include morning review hours (first 60-90 minutes of the day dedicated to clearing the review queue) and post-standup review sprints (30-45 minutes after daily standup). These patterns work because they convert review from an interruptible background task to a focused foreground task.

Improve PR routing

Automated PR routing based on code ownership, reviewer expertise, and current workload reduces time-to-first-review by matching PRs to the right reviewer immediately rather than relying on ad-hoc assignment or round-robin distribution. Workload-aware routing is particularly important in AI-heavy environments where PR volume is uneven across team members.

Pair AI-generated PRs with enhanced context

AI-generated PRs are harder to review in part because the reviewer did not watch the code evolve. With human-written PRs, reviewers often have informal context: they saw the Slack discussion, they know the author's approach, they can guess the design intent. AI-generated code arrives without this context.

Mitigate this by requiring enhanced PR descriptions for AI-assisted work: what the AI generated, what the author modified, why specific design choices were made, and what the author specifically wants the reviewer to scrutinize. This front-loads context that reduces review friction and duration.

How PR Cycle Time Fits the Developer AI Impact Framework

PR cycle time maps to Pillar 3 (Throughput) of Larridin's Developer AI Impact Framework. It measures how quickly completed work moves through the delivery pipeline -- complementing Complexity-Adjusted Throughput, which measures how much complex work gets done.

The relationship between the two metrics is important. A team can have high Complexity-Adjusted Throughput (shipping lots of complex work) with long PR cycle times (that work takes a long time to move from branch to merge). This pattern indicates a review bottleneck: the team is productive at coding but constrained by review capacity. Conversely, a team can have fast cycle times with low throughput -- they merge quickly but do not ship much meaningful work.

The AI-era insight is that both of these metrics need to be segmented by AI attribution to reveal the bottleneck shift. Without segmentation, the metrics can look healthy in aggregate while a growing review bottleneck silently extends lead time for AI-assisted work.

PR cycle time also connects to Pillar 4 (Quality). When review wait times increase, two quality risks emerge:

  1. Reviewers rush. Facing a growing queue, reviewers spend less time per PR, and quality issues that would have been caught in thorough review slip through. These show up later as elevated code turnover rate.

  2. Context decay. The longer a PR waits for review, the staler the author's context becomes. If a reviewer requests changes three days after the PR was opened, the author must re-load context they have already swapped out of working memory. This produces lower-quality revisions and extends the review iteration cycle.

Read the full Developer AI Impact Framework -->

Frequently Asked Questions

What is a good PR cycle time in 2026?

A healthy PR cycle time target for product engineering teams in 2026 is under 24 hours from branch creation to merge, with time-to-first-review under 4 hours. These targets apply to both AI-assisted and human-only PRs. The more diagnostic metric is the decomposition: if total cycle time is under 24 hours but time-to-first-review is 10+ hours, the team has a review bottleneck that will worsen as AI adoption increases PR volume. Track the phases independently rather than optimizing only for the total.

How has AI changed PR cycle time?

AI has compressed the coding phase of the PR lifecycle by 40-60% while expanding the review phase by 30-50% in many teams. The net effect on total PR cycle time ranges from modest improvement to slight degradation, depending on whether the team has adjusted its review capacity for the increased PR volume that AI tools generate. The bottleneck has shifted from writing code (the author's time) to reviewing code (the reviewer's time). Teams that measure only total cycle time will miss this shift because the faster coding phase offsets the slower review phase in the aggregate.

Why is code review slower for AI-generated PRs?

AI-generated PRs tend to be larger, use patterns the reviewer did not choose, and require scrutiny for subtle quality issues that automated tests do not catch. AI tools can generate substantial amounts of code quickly, producing larger diffs that take more time to review. The code may use unfamiliar libraries, duplicate existing logic, or introduce coupling that is not visible in a surface-level read. Reviewers also lack the informal context they get with human-written PRs -- they did not observe the design process unfold. All of these factors increase review effort per PR. GitClear's analysis of code churn trends34 suggests that the quality concerns driving more thorough review are warranted, not overcautious.

How do you reduce time-to-first-review?

The three most effective interventions are dedicated review time blocks, automated PR routing based on expertise and workload, and explicit PR size limits for AI-assisted work. Dedicated review blocks (60-90 minutes at the start of the day or after standup) convert review from an interrupt-driven background task to a focused foreground task. Workload-aware routing ensures PRs reach available reviewers immediately rather than accumulating in overloaded queues. PR size limits prevent AI tools from generating large diffs that sit in the queue because no reviewer wants to tackle them. These interventions address the most common causes of review delay without requiring additional headcount.

Should AI-assisted PRs have different cycle time targets?

No -- the cycle time target should be the same for AI-assisted and human-only PRs, but the intervention strategies differ. Applying a different (looser) target to AI-assisted PRs creates a two-tier system where AI-generated code gets slower review, which is the opposite of what quality requires. AI-generated code should be reviewed with at least the same rigor as human-written code. The way to achieve consistent cycle times is not to relax the target but to address the review capacity bottleneck: smaller PRs, more review bandwidth, better PR routing, and enhanced context in PR descriptions for AI-assisted work.

Footnotes

Data sources and methodology:

  • Larridin. "The Developer AI Impact Framework." Positions PR cycle time as a Pillar 3 (Throughput) metric, with the AI-era emphasis on time-to-first-review as the critical phase. Benchmark data derived from aggregated engineering data across organizations of varying size and sector (Larridin internal benchmark).

Related Resources


  1. Block, "AI-Assisted Development at Block". Reports on AI-assisted development practices across Block's engineering organization, including acceleration of initial code creation and adoption patterns among engineers. 

  2. GitHub, "Research: Quantifying GitHub Copilot's Impact in the Enterprise with Accenture" (2024). Measurable reductions in task completion time for Copilot users in enterprise environments. 

  3. GitClear, "Coding on Copilot: 2023 Data Suggests Downward Pressure on Code Quality" (2024). Analysis of 211M+ changed lines showing code churn rising from 3.3% baseline to 5.7-7.1% as AI coding tools gained adoption. 

  4. GitClear, "AI Copilot Code Quality: 2025 Data" (2025). Updated longitudinal analysis of AI-assisted code quality trends.