
Engineering teams are racing to adopt AI, but most still see only modest gains. Pilots stall, adoption plateaus, and impact remains hard to measure – leaving leaders asking, “Is any of this actually working?”
Over the past six months, Plandek worked with engineering leaders across Europe to understand why some teams are achieving real, measurable outcomes while others remain stuck in experimentation mode.
The result is the RACER Framework – a practical operating model for teams transitioning from traditional software delivery to AI-augmented engineering. RACER isn’t theory. It reflects what the highest-performing teams are actually doing right now to operationalize AI at scale.
Why traditional engineering operating models are failing in the AI era
Engineering teams didn’t hit a wall because of technology. They hit a wall because AI changes the nature of work – and most organizations haven’t updated the way they plan, measure, and manage delivery.
Common patterns emerged in our interviews:
AI tools are adopted by individuals, but not embedded into team workflows.
Leaders struggle to quantify whether AI is improving quality, flow, or delivery speed.
Constraints like unclear requirements, unstable environments, or weak documentation quickly offset any productivity gain.
Teams lack a shared model for tracking AI adoption, impact, and bottlenecks.
RACER was designed to solve these gaps.
Introducing the RACER Framework
RACER is a five-pillar model that gives engineering leaders a clear path to scale AI-augmented engineering.
R – Real Adoption
Track how AI tools are actually being used across teams, roles, and repositories. High performers don’t rely on anecdotes. They measure usage, patterns, and consistency of AI-driven workflows to establish a true baseline.
A – Adoption Impact
Measure the outcomes that matter: delivery speed, predictability, flow efficiency, and quality. AI is only valuable if it improves core engineering metrics. RACER connects usage to measurable delivery uplift.
C – Constraints
Every team has bottlenecks – missing documentation, unclear tickets, slow PR reviews, brittle environments. AI doesn’t remove these constraints. In many cases, it exposes them. High-performing teams use RACER to identify and systematically remove the blockers that limit AI’s effectiveness.
E – Engineering Impact
This is where AI’s value becomes visible. Leaders track how AI is changing the end-to-end delivery pipeline: faster cycle times, reduced toil, improved review throughput, lower rework, tighter epic predictability. RACER brings these improvements into a unified view.
R – Re-imagined Operating Model
AI-augmented engineering isn’t a tooling project – it’s an operating-model transformation. High-performers rethink workflows, team structures, processes, and cultural norms to unlock compounding gains.
What we learned from top engineering organizations
Across interviews with engineering and technology leaders, a consistent pattern emerged: teams that improved delivery didn’t simply “add AI.” They rebuilt their delivery systems around it.
Top performers:
Measured adoption instead of assuming it
Removed constraints before scaling use cases
Focused on flow efficiency and predictability, not just code generation
Embedded AI into rituals like planning, refinement, and retros
Created transparent dashboards to show impact up to the C-suite
The takeaway is clear: AI tools create leverage, but only when supported by the right fundamentals.
How Plandek supports the RACER Framework
Plandek is purpose-built for engineering leaders adopting AI. The platform provides:
End-to-end visibility across delivery, flow, quality, and predictability
Metrics that quantify AI impact across teams and workflows
Detection of constraints and bottlenecks limiting AI-driven productivity
Dashboards that connect engineering performance to business results
And with Dekka, our AI copilot for engineering analytics, leaders can ask natural-language questions, monitor adoption, and receive proactive insights and risks – all powered by the data already flowing through Plandek.
The bottom line
AI-augmented engineering is not a future trend. It’s the new competitive advantage. The organizations that learn to operationalize AI – and measure its true impact – will ship faster, build more predictably, and outperform their peers.
RACER gives engineering leaders a practical blueprint to get there.
Click here to view the full framework
Written by
Charlie Ponsonby
Co-founder & CEO
Charlie started his career as an economist working on trade policy in the developing world, before moving to Accenture in London. He joined the Operating Board of Selfridges, before moving to Open Interactive TV and then Sky where he was Marketing Director until leaving to found Simplifydigital in 2007. Simplifydigital was three times in the Sunday Times Tech Track 100 and grew to become the UK’s largest TV, broadband and home phone comparison service, powering clients including Dixons-Carphone, uSwitch and Comparethemarket. It was acquired by Dixons Carphone plc in April 2016. He co-founded Plandek with Dan Lee in 2018. Charlie was educated at Cambridge University. He lives in London and is married with three children.
See how your engineering efforts translate into measurable business impact
Measure delivery performance, AI impact, and engineering productivity with hundreds of metrics, OOTB dashboards and custom configurations.
Contact us
LONDON - HQ
Unit 313 The Print Rooms, 164-180
Union St, London SE1 0LH












