Skip to Content
DocsMetricsTeam Flow Metrics
DRAFT — This document is under development and not yet reviewed.

Team Flow Metrics

This framework defines metrics for measuring and managing the flow of stories through the team’s delivery system. It gives teams visibility into how work moves through their sprint cycle — where it flows freely, where it stalls, and how predictably the team delivers against its sprint commitments.


Measurement Scope

What We Measure

Two measurement domains:

Flow Metrics — How stories move through the team kanban: time per stage, WIP accumulation, throughput, and distribution across story types.

Predictability Metrics — How reliably the team delivers against its sprint goal and sprint commitment.

What We Do Not Measure

Individual performance. Flow metrics describe the behavior of the team’s delivery system — not how individual developers perform. Using team flow data for individual assessment undermines the psychological safety that makes honest measurement possible.

Quality outcomes. Defect rates, test coverage, and pipeline health are tracked through separate technical quality indicators. DORA metrics — particularly change failure rate and MTTR — complement team flow metrics from the technical delivery perspective.


What is a Story?

A Story is the team’s primary unit of flow — a specific, testable behavior the team designs, builds, tests, and integrates within a single sprint.

Story types:

TypeDescriptionTypical distribution
Functional StoryDelivers a user-visible behavior70–80%
Enabler StoryInvests in technical capability — infrastructure, refactoring, test coverage, security15–25%
SpikeTime-boxed investigation into an unknown; produces knowledge, not software<10%

Spikes that appear routinely — rather than occasionally — are a signal worth investigating. They may indicate insufficient feature refinement upstream or a team working in a domain it does not yet understand well enough.


Story-Level Metrics

Flow Time Metrics

All flow time metrics use calendar days.

1. Story Cycle Time Elapsed time from entering In Progress to Done. Date entered Done − Date entered In Progress

This is the most useful story-level flow metric. It reflects the team’s actual execution speed — the time from active work starting to the story being integrated and verified.

2. Story Lead Time Elapsed time from entering Sprint Backlog to Done. Date entered Done − Date entered Sprint Backlog

Includes wait time before work starts. The gap between Lead Time and Cycle Time reveals how much time stories spend waiting in the sprint before being picked up — an indicator of overloaded sprints or uneven pull behavior.

3. Flow Time — Refinement to Done Elapsed time from entering Refinement to Done. A broader view that captures how long stories spend in the preparatory stages before active work begins.


WIP Age

WIP Age measures how long a story has been in its current stage. Used to identify stalled items early — before they threaten the sprint goal.

A story that has been In Progress for more than 2–3 days without movement is worth examining. Multiple stalled stories are a system signal — not individual problems.


Sprint-Level Metrics

Flow Metrics (3)

1. Sprint Velocity Stories (or story points) completed per sprint. The primary measure of team throughput capacity.

Velocity should be tracked as a trend across multiple sprints — a single sprint’s velocity is not meaningful in isolation. A stable trend over 4–6 sprints gives a reliable planning baseline. High variance in velocity is itself a signal: it indicates the system is absorbing disruption that is not yet visible in other metrics.

2. Flow Load (WIP) Total stories in active work at any point during the sprint, across all active stages (In Progress + Verification). Directly observable on the team’s sprint board.

High WIP relative to team capacity causes longer cycle times and increases the probability of stories not finishing within the sprint. A team with 8 developers carrying 14 stories In Progress is not working faster — it is splitting attention and slowing every story down (Little’s Law).

Healthy WIP limits: typically one story per developer in active work, with one or two in Verification.

3. Flow Distribution The balance between Functional Stories, Enabler Stories, and Spikes in the sprint.

A sprint heavily skewed toward Enablers may indicate the team is investing in capability at the expense of near-term feature delivery — acceptable when deliberate, worth examining when not. A sprint with no Enablers over several consecutive sprints is a signal that technical debt may be accumulating.

Predictability Metrics (2)

4. Sprint Goal Achievement Whether the sprint goal was confirmed at Sprint Review — binary per sprint, tracked as a trend.

Sprint Goal Achievement is the primary predictability metric at team level. It directly answers whether the team delivered what it committed to delivering — in terms of value, not volume.

Target: Sprint goal achieved in 80%+ of sprints.

Achievement rateInterpretation
80–100%Healthy — team understands its capacity and sets realistic goals
60–79%Improvement needed — review goal-setting, commitment sizing, or unplanned work
<60%Systemic issues — overcommitment, dependency failures, or unplanned work overload

5. Story Commitment Reliability (Stories completed as committed at Sprint Planning / Total stories committed) × 100

A secondary predictability metric. Lower reliability is normal and acceptable when the sprint goal is still achieved — teams may replace or descope stories while delivering the goal’s intent. High reliability with a failed sprint goal is the more problematic pattern.

Target range: 70–85%. Above 90% consistently may indicate undercommitment.


Metric Interpretation

Cycle Time Reference Points

These are starting points for calibration — not universal targets. Each team should establish its own baseline over 4–6 sprints before setting improvement targets.

Story complexityCycle time reference
Small (well-understood, single concern)1–3 days
Medium (typical story)3–5 days
Large (near-sprint-sized)5–10 days

Stories consistently exceeding these ranges are candidates for decomposition review in refinement.

Reading the Metrics Together

Flow metrics and predictability metrics tell different parts of the same story.

Healthy pattern: stable velocity trend, cycle times within range, sprint goal achieved consistently, WIP stays manageable across the sprint.

Signals worth investigating:

  • High WIP + long cycle times → team is overloaded; reduce sprint commitment
  • Low velocity + sprint goal achieved → team may be undercommitting; calibrate sizing
  • High story reliability + sprint goal missed → stories are delivered but they don’t add up to the goal’s value; review goal formulation
  • Velocity highly variable → look for unplanned work, dependencies, or team instability
  • Spike ratio rising → refinement quality is declining upstream

Daily (Standup) — WIP Age check: are any stories stalling? Surface blockers before they compound.

End of sprint (Sprint Review + Retrospective) — Sprint Goal Achievement, velocity, WIP Load, commitment reliability. Retrospective uses this data to identify improvement hypotheses.

Every 3–4 sprints — Cycle time trend, flow distribution review. Are patterns shifting? Is technical debt accumulating?

PI Planning — Use velocity trend and cycle time data to size sprint commitments realistically. Historical data is the best input to sustainable capacity planning.


Sources

  • Lean / Little’s Law — WIP and cycle time relationship; the principle that reducing WIP is the most reliable way to reduce lead times
  • Kanban — WIP limits, flow visualization, and cycle time measurement as continuous improvement tools
  • Scrum — Sprint velocity and sprint goal as the team-level planning and measurement backbone
  • DORA metrics — Complementary technical flow indicators; see DORA Metrics for deployment frequency and lead time as pipeline health signals