Control Planes for AI Engineering
$16M ARR Amplitude for AI code quality
The key shift is that AI coding has turned engineering analytics from a backward looking dashboard into a live measurement system for AI spend, code quality, and workflow control. Once teams start paying for Copilot, Cursor, or Claude Code, they need to see which repos use AI most, whether PRs move faster, and whether defects or review time rise with AI generated code.
-
This market sits on top of the older DORA playbook. LinearB already pulled data from GitHub, CI, and incident tools to measure deploy speed and failures. AI adds a new layer, tracking code attribution, acceptance, review depth, and downstream bugs, which creates a new budget line rather than just another dashboard tab.
-
The field is splitting into three product shapes. LinearB and Jellyfish are retrofitting existing engineering intelligence suites with AI impact modules. Newer players like Weave and Span start with AI code detection and ROI analytics. GitHub and GitLab are bundling native usage and value stream metrics into the systems where code already lives.
-
The practical buyer problem is simple. A CTO wants proof that 500 Copilot seats are speeding up delivery without flooding the codebase with low quality PRs. That is why products now track AI code ratios, cycle time, acceptance, review patterns, and defect signals together instead of reporting raw usage alone.
This category is heading toward control planes for AI engineering, not just scoreboards. The winning products will combine measurement with action, recommending reviewers, enforcing code policies, and routing work based on where AI helps versus where it creates rework. That pushes the market closer to core dev workflow infrastructure, and closer to direct competition with GitHub, GitLab, and Atlassian.