Developer Productivity Measurement: Frameworks for the AI Era
SPACE, DORA, DevEx — the major developer productivity frameworks explained, what they miss, and an updated framework for measuring engineering in 2026.
Every few years, the industry produces a new framework for measuring developer productivity. SPACE. DORA. DevEx. Each one moves the conversation forward. Each one captures something real. And each one has a gap that the next framework tries to fill.
I've studied all of them — first as an engineer being measured by them, then as an investor watching portfolio companies try to implement them. Here's what each gets right, what's missing, and what a complete measurement approach looks like in 2026.
The SPACE Framework
SPACE was introduced in 2021 by Nicole Forsgren, Margaret-Anne Storey, Chandra Maddila, Thomas Zimmermann, Brian Houck, and Jenna Butler. It stands for Satisfaction and well-being, Performance, Activity, Communication and collaboration, and Efficiency and flow.
What It Gets Right
SPACE's biggest contribution is arguing that productivity is multidimensional. You can't capture it with a single metric. Any measurement system that relies on one number — velocity, cycle time, lines of code — is missing the picture.
The framework is also explicit that you should measure across at least three dimensions and at multiple levels (individual, team, organization). That's sound advice. A team with high activity but low satisfaction is heading for burnout. A team with high satisfaction but low performance has a different problem.
SPACE also introduced the important idea that developer satisfaction matters as a first-class metric. How engineers feel about their tools, their workflow, and their work isn't soft data — it's a leading indicator of retention, quality, and velocity.
What's Missing
SPACE is a framework, not a measurement system. It tells you what categories to measure but not how to measure them. The Performance dimension, for example, is defined as "the outcome of a system or process." That's correct but unhelpfully abstract. What is the outcome? How do you quantify it?
In practice, most SPACE implementations fall back on the same old proxies: story points for performance, commit counts for activity, survey scores for satisfaction. The framework is more sophisticated than its implementations.
The most significant gap: SPACE doesn't address how to measure the quality or complexity of engineering output. Activity (the A) tracks how much engineers do. Performance (the P) gestures at outcomes but doesn't specify how to assess the actual code that ships. In 2021, that was understandable — there was no scalable way to evaluate code complexity. In 2026, there is.
DORA Metrics
DORA — Deployment Frequency, Lead Time for Changes, Change Failure Rate, and Mean Time to Restore — emerged from the Accelerate research by Nicole Forsgren, Jez Humble, and Gene Kim. It's the most widely adopted engineering measurement framework in the industry.
What It Gets Right
DORA measures something real and valuable: delivery process health. A team that deploys frequently, with short lead times, low failure rates, and fast recovery, has a healthy engineering pipeline. That matters. Slow, fragile, risky deployments are a genuine engineering problem, and DORA gives you the data to identify and fix it.
The research backing is also strong. The Accelerate findings are based on years of data across thousands of organizations. The correlation between DORA performance and business outcomes is well-established.
And DORA is actionable. If your deployment frequency is low, you can identify why (manual approvals, flaky tests, monolithic releases) and fix it. The metrics point to specific process improvements.
What's Missing
I've written about this in detail, but the core issue is straightforward: DORA measures how fast your pipeline runs, not what's moving through it.
A team deploying trivial config changes ten times a day is "elite" on DORA. A team shipping one transformative feature per week with careful rollout is not. DORA can't distinguish between the two because it doesn't examine the content of what's being deployed.
This isn't a criticism of DORA — it was never designed to measure output. It was designed to measure process. The problem is when organizations treat DORA as a complete measurement system. It's not. It's one essential layer.
The DevEx Framework
DevEx, proposed by Abi Noda, Margaret-Anne Storey, and Nicole Forsgren, focuses on developer experience across three dimensions: feedback loops, cognitive load, and flow state. It's measured primarily through developer surveys.
What It Gets Right
DevEx captures something that neither SPACE nor DORA addresses directly: the subjective experience of doing engineering work. Long build times, confusing tooling, excessive meetings, unclear requirements — these are real productivity killers that don't show up in git metrics or deployment data.
The framework is also grounded in research on cognitive load and flow states, which gives it a theoretical foundation beyond "ask developers if they're happy."
Survey data provides context that quantitative metrics can't. If your cycle time just doubled, DORA tells you that it happened. A DevEx survey might tell you why — maybe a new compliance requirement added two approval steps, or maybe a key dependency became unreliable.
What's Missing
Surveys measure perception, not reality. Developers might report high satisfaction while shipping very little. Or report frustration while doing their best work. Perception data is valuable context but a poor primary measurement.
Surveys are also point-in-time and relatively low-frequency (quarterly at best). They can't tell you what's happening this week. And survey fatigue is real — response rates decline over time if people don't see the data leading to concrete changes.
The Missing Piece: Direct Output Measurement
Here's the pattern across all three frameworks: none of them directly measure the complexity or quality of what engineers ship.
- SPACE categorizes productivity dimensions but doesn't specify how to evaluate the actual code artifact.
- DORA measures how efficiently code moves through the pipeline but not the substance of what moves through it.
- DevEx measures how engineers feel about their work but not the output of that work.
It's like having frameworks for measuring a restaurant's kitchen efficiency (DORA), chef satisfaction (DevEx), and multidimensional performance categories (SPACE) — while never actually tasting the food.
Until recently, this gap was understandable. There was no scalable way to evaluate code complexity. Humans could do it (code reviews are essentially complexity assessments), but not at the speed and consistency needed for measurement. You can't have a senior engineer review every merged PR across your organization and assign a complexity score.
AI changed that. Large language models can read code diffs and evaluate complexity across multiple dimensions — scope, architecture, implementation sophistication, risk, quality — with consistency that matches or exceeds human estimation. This is the capability that makes direct output measurement possible at scale.
An Updated Framework for 2026
Given where we are today — with AI both writing more of our code and enabling new forms of measurement — here's what a complete engineering measurement system looks like.
Layer 1: Process Health (DORA)
Track the four DORA metrics. They tell you whether your delivery pipeline is healthy. Low deployment frequency, long lead times, high failure rates, or slow recovery indicate process problems that need fixing regardless of everything else.
Tools: Sleuth, Swarmia, LinearB, or your CI/CD provider's built-in metrics.
Layer 2: Output Quality (Velocity Scoring)
Score the actual code that ships. Evaluate every merged PR for engineering complexity across multiple dimensions. This tells you what your team is actually producing — not how busy they look, but how substantial their shipped work is.
Tools: GitVelocity (free, AI-powered).
Layer 3: Developer Satisfaction (Surveys)
Run structured developer experience surveys quarterly. Measure cognitive load, flow state disruptions, tool satisfaction, and friction points. This provides the why behind what the quantitative metrics show.
Tools: DX, or structured internal surveys following the DevEx framework.
Layer 4: AI Adoption (Tooling Metrics)
This is the new layer that none of the existing frameworks anticipated. Track how AI tools are being adopted and what impact they're having. Which engineers are using AI effectively? How does AI-assisted output compare to non-assisted output? Where is AI adoption lagging?
Tools: Editor telemetry for adoption data, GitVelocity for AI impact measurement (compare output scores before and after AI tool adoption).
How the Layers Work Together
Each layer answers a different question:
| Layer | Question | Data Type |
|---|---|---|
| Process Health | Is our pipeline efficient? | Quantitative, automated |
| Output Quality | What are we actually shipping? | Quantitative, AI-assessed |
| Developer Satisfaction | How does our team feel? | Qualitative, survey-based |
| AI Adoption | How is AI changing our output? | Quantitative, comparative |
No single layer is sufficient. A team with great DORA metrics but low output scores is deploying frequently but not shipping substance. A team with high output scores but low satisfaction is performing but burning out. A team with strong output but no AI adoption might be leaving productivity gains on the table.
The complete picture requires all four.
Implementing Without Creating Surveillance Culture
Every measurement discussion eventually hits the same concern: "Won't this turn into developer surveillance?" It's a legitimate worry, and it has killed more measurement programs than technical challenges ever have.
Here's what I've learned about implementing measurement that engineers actually accept.
Transparency above everything. If you're scoring code, show the rubric. Show the breakdown. Let engineers see exactly why their PR got the score it got. GitVelocity shows the full six-dimension breakdown for every score. No black boxes.
Measure output, not activity. Engineers revolt against tools that track their keystrokes, screen time, or Slack presence. They generally accept tools that evaluate the work they ship — because that's what they want to be judged on. There's a fundamental difference between "we're watching what you do" and "we're evaluating what you ship."
Use data for support, not punishment. If an engineer's output scores are declining, the right response is "what's blocking you?" not "why aren't you performing?" The same data that identifies problems should drive support conversations, not punitive ones.
Start with teams, then individuals. Roll out team-level visibility first. Let people get comfortable with the data. Then make individual data available — ideally to the individual first, then to their manager. Engineers who see their own data and find it fair become advocates for the system.
Acknowledge the history. Engineers have been burned by bad metrics. Story points used as performance targets. Commit counts on performance reviews. Lines of code compared across teams. Acknowledge that history. Explain what's different this time. Be honest about trade-offs.
Where to Start
If your team has no measurement framework today, don't try to implement all four layers at once. Start with:
DORA basics. Most CI/CD tools provide deployment frequency and lead time data out of the box. Set up those dashboards first.
Output scoring. Connect GitVelocity to your repositories. It takes minutes to set up, it's free, and it immediately gives you data you've never had: what your team is actually shipping, quantified by complexity.
A simple survey. Even five questions quarterly provides useful signal. Ask about tool satisfaction, biggest blockers, and whether engineers feel their work is recognized.
AI adoption tracking. If your team is adopting AI tools, start tracking before-and-after output data now. You'll want the baseline.
The frameworks in this article are maps, not territories. Use them to orient your thinking, then adapt to what your team actually needs. The goal isn't to implement SPACE or DORA or DevEx perfectly — it's to answer the questions that matter with data you can trust.
GitVelocity provides the output quality layer that every productivity framework is missing. Score every merged PR on six dimensions of complexity — free, transparent, and AI-powered.
Conrad is CTO and Partner at Headline, where he leads data-driven investment across early stage and growth funds with over $4B in AUM. Before becoming an investor, he founded Munchery (raised $130M+) and held engineering and product leadership roles at IAC and Convio (IPO 2010). He and the Headline engineering team built GitVelocity to help engineering organizations roll out agentic coding and measure its impact.