What Sets High-Velocity Engineering Teams Apart
After studying engineering teams across 30+ portfolio companies, five patterns separate high-output teams from everyone else.
Working at Headline, a VC firm, I have an unusual vantage point on engineering teams. I don't just see one team's practices — I see patterns across thirty-plus portfolio companies, from five-person startups to hundred-engineer organizations. I see what's working, what's not, and what the highest-output teams do that the average ones don't.
The patterns are surprisingly consistent. The specific practices vary, but the principles behind them show up again and again. Here are the five that matter most.
They Measure Output, Not Activity
This is the single biggest differentiator, and it's the one that takes the most courage to implement.
Average engineering teams track activity metrics: tickets closed, PRs merged, commits pushed, hours logged. These numbers feel productive to report and are easy to collect. They're also meaningless. An engineer who merges fifteen trivial PRs looks more "productive" than one who ships a single complex architectural improvement that will save the team hundreds of hours over the next year.
The best teams I've studied have moved past this. They measure what was actually built — the complexity, quality, and significance of the code that got shipped. Not how many times the engineer typed "git push."
This shift is harder than it sounds. Activity metrics are comforting because they always go up. There's always something to count. Output metrics are uncomfortable because they sometimes reveal that a lot of activity produced very little substance. But that discomfort is exactly the point — it surfaces problems that activity metrics hide.
One portfolio company switched from tracking PR count to tracking PR complexity scores. Within a month, they discovered that two engineers who appeared highly productive by activity metrics were actually shipping almost exclusively trivial changes — config updates, copy tweaks, minor refactors that could have been batched. Meanwhile, a quieter engineer who merged one or two PRs per week was consistently shipping the most architecturally significant work on the team. Activity metrics had inverted reality.
They Standardize AI Tool Adoption
The worst approach to AI coding tools I've seen is: "We bought Cursor licenses for everyone, use it if you want."
This sounds progressive and non-prescriptive. In practice, it creates a bimodal distribution. A few early adopters figure out effective workflows and see massive productivity gains. Everyone else uses the tool occasionally for autocomplete suggestions and wonders what the fuss is about. The team average barely moves, leadership questions the ROI, and the investment gets scrutinized.
The highest-output teams treat AI adoption as an engineering initiative, not an individual choice. They identify which workflows benefit most from AI assistance. They develop shared prompting patterns and configurations. They measure adoption and output impact to understand who's leveraging the tools effectively and who needs support.
At Headline, when we standardized our own AI tool adoption in mid-2025, the results were striking. Team productivity nearly doubled between August and November. But the most interesting finding was that junior engineers started outperforming seniors on complexity scores. The juniors, who had fewer ingrained habits to unlearn, adopted AI-assisted workflows faster and more completely.
That insight only came because we were measuring output, not activity. If we'd been tracking commit counts, the AI adoption story would have been invisible.
They Allocate Review Effort by Complexity
I made this mistake personally for years — treating every PR with identical review ceremony. I've written about it in detail in The Code Review Habit That Was Costing My Team, but the pattern shows up everywhere.
Teams that review every change with the same process end up under-reviewing the changes that matter and over-reviewing the ones that don't. Reviewers burn out on trivial PRs and then rubber-stamp the complex ones because they've exhausted their attention.
The best teams route review effort based on what's actually in the PR. A config change gets a quick scan. A database migration gets dedicated senior reviewer time and an architectural discussion. A straightforward feature implementation gets a standard review.
This requires knowing the complexity of a PR before review starts — not just the line count, but the actual risk and architectural significance. Teams that have adopted AI-powered PR scoring use it to triage their review queues. The highest-scoring PRs get the most review attention. The lowest-scoring ones get fast-tracked.
The result is better review quality where it matters and faster throughput everywhere else. Both halves of that equation improve velocity.
They Create Visibility Without Surveillance
Every high-performing team I've studied has some form of individual-level visibility into engineering output. And every one of them was careful about how they implemented it.
The difference between visibility and surveillance comes down to three things: transparency, direction, and purpose.
Transparency means engineers can see their own data. Not just managers — engineers. When an engineer can see their own complexity scores, track their own trends, and understand how their output compares, they have information they can act on. When only managers can see the data, it feels like being watched.
Direction means the data flows in both directions. It's not just a top-down management view. Engineers use the data to advocate for themselves — "Look at the complexity I shipped this quarter" is a much stronger argument in a promotion discussion than "I feel like I did a lot."
Purpose means the data is used for development, not punishment. The teams that use output data primarily for identifying growth opportunities, resource allocation, and process improvement get buy-in. The teams that use it primarily for performance reviews and stack ranking get resistance.
One company in our portfolio implemented individual scoring and made it visible to everyone on the team. The initial reaction was cautious. Within two months, it had transformed the culture. Engineers started competing on complexity scores the way salespeople compete on quota attainment. Not because management mandated it — because the engineers themselves found it motivating to see their work quantified and recognized.
The engineers who were already high performers loved it. They'd been invisible for years, doing the hard work while louder colleagues got the credit. Finally, their output spoke for itself.
They Make Data the Basis for Conversations
The average engineering manager has development conversations based on impressions. "I feel like you've been doing great work" or "I think you could be more impactful." These conversations are well-intentioned but vague, and engineers know it.
The best teams ground every conversation in data. Not data as a weapon — data as a shared reference point.
"Your complexity scores have been trending up over the last quarter — what's driving that?" is a much better conversation starter than "I think you're improving." The engineer might say they've been taking on more architecturally significant work, or that they figured out a new workflow, or that they changed how they scope PRs. Any of those answers leads to a more productive discussion than vague praise.
Similarly, "I noticed your output complexity dropped over the last three weeks — is everything okay?" opens a genuine dialogue. Maybe they're onboarding onto an unfamiliar part of the codebase. Maybe they're doing important mentoring work that doesn't show up in PR scores. Maybe they're struggling and need help. The data doesn't replace the conversation — it starts it.
This works at the team level too. When you can see output patterns across the whole team, you can have informed discussions about resource allocation. If one squad is producing high-complexity work while another produces mostly trivial changes, that's a data point worth exploring. Maybe the second squad is working on genuinely simpler problems. Or maybe they need different tooling, different staffing, or a different approach.
The Unifying Principle
All five patterns point to the same underlying principle: high-performing teams make decisions based on what's actually happening in their codebase, not on proxies, estimates, or impressions.
They measure the real output, not the activity around it. They standardize their tools based on measured impact, not assumptions. They allocate review effort based on actual PR complexity, not arbitrary process. They create visibility based on real data, not surveillance. And they have development conversations grounded in evidence, not vibes.
The reason these patterns are becoming more common now is that the tools to support them finally exist. You couldn't measure code complexity at scale five years ago. You couldn't score individual PR contributions automatically. You couldn't compare output before and after AI adoption with any precision.
Now you can. And the teams that are using these capabilities are pulling ahead.
GitVelocity measures engineering velocity by scoring every merged PR using AI. It gives engineering teams the output-level visibility that separates the highest-performing teams from everyone else.
Conrad is CTO and Partner at Headline, where he leads data-driven investment across early stage and growth funds with over $4B in AUM. Before becoming an investor, he founded Munchery (raised $130M+) and held engineering and product leadership roles at IAC and Convio (IPO 2010). He and the Headline engineering team built GitVelocity to help engineering organizations roll out agentic coding and measure its impact.