
AI has reached software engineering faster than any enterprise technology before it. In less than two years, tools like GitHub Copilot, Cursor, ChatGPT, and internal LLMs have gone from experiments to everyday companions for many engineers. By mid-2025, AI-assisted development is no longer a novelty. It’s normal.
And yet, for all this activity, most engineering leaders are still asking the same question:
Is any of this actually working?
Despite confident claims of step-change productivity and “10x engineers,” measurable improvements in delivery speed, predictability, and quality remain elusive for most organizations. Pilots stall. Adoption plateaus. Impact is hard to quantify. Leadership teams are left with anecdotes, not evidence.
This gap between promise and reality is not because AI doesn’t work. It’s because adopting AI fundamentally changes how engineering work happens, and most organizations are trying to absorb that change without updating their operating models.
The AI Gold Rush Problem
The current phase of AI adoption looks strikingly similar to past technology gold rushes. There is rapid uptake, intense experimentation, and no shortage of bold claims. But there is very little shared understanding of what “good” actually looks like at scale.
Across conversations with engineering leaders, a consistent set of patterns emerges:
AI adoption is happening faster than any previous developer tool
Usage is driven bottom-up by individuals, not top-down by teams
Benefits are uneven and highly context-dependent
Governance, security, and quality concerns are rising
Very few teams can clearly articulate the business impact
Some teams are genuinely pulling ahead. Most are still experimenting without a clear path forward.
Why Hyper-Productivity Is Context Dependent
One of the most important insights from engineering leaders is that AI productivity gains are not universal. The same tools produce dramatically different outcomes depending on the environment they are introduced into.
High-impact teams typically share a few characteristics:
Clear, well-defined work and acceptance criteria
Modular, understandable codebases
Strong documentation and institutional knowledge
Mature CI/CD pipelines and testing practices
Time and psychological safety to experiment
Low-impact teams, by contrast, often struggle with:
Poor or missing documentation
Monolithic or brittle architectures
Unclear requirements and frequent rework
Slow review and deployment cycles
Limited visibility into delivery performance
AI does not erase these differences. In many cases, it amplifies them. Where the system is healthy, AI compounds productivity. Where it is fragile, AI simply exposes problems faster.
The Real Bottleneck Is Not the Tools
A recurring theme from engineering leaders is that tooling itself is rarely the constraint. Most teams already have access to capable AI tools. The limiting factor is everything around them.
Common constraints include:
Outdated or missing documentation that limits AI usefulness
Ambiguous tickets that AI cannot meaningfully clarify
Fragile environments that slow iteration
Long-running PR reviews that cap throughput
Weak governance around usage and risk
No shared metrics for adoption or impact
AI accelerates execution. It does not magically fix broken systems. In fact, it often surfaces those weaknesses more quickly, creating the illusion that AI “isn’t working” when the real issue lies elsewhere.
What Winning Teams Do Differently
Teams seeing sustained gains did not treat AI adoption as a tooling rollout. They treated it as a delivery transformation.
One example discussed during the webinar involved leadership explicitly framing AI adoption as a strategic initiative. Rather than encouraging ad-hoc experimentation, they:
Defined clear goals for AI usage
Established baseline metrics for delivery performance
Identified constraints limiting AI effectiveness
Removed those constraints systematically
Measured outcomes continuously, not just usage
The result was not just faster coding. It improved flow efficiency, better predictability, and reduced delivery risk.
The lesson was clear: AI impact comes from fixing the system, not just deploying tools.
Introducing a Practical Model for AI-Augmented Engineering
To turn these insights into something actionable, Plandek distilled the findings into the RACER Framework. RACER is a practical operating model designed to help engineering leaders move from experimentation to sustained, measurable impact.
RACER focuses on five pillars:
Rollout
Deploy AI tools deliberately and establish visibility into who is using them, where, and how consistently. Winning teams measure adoption rather than assuming it.
Approach
Define how AI is used across workflows. This includes assisted development, supervised agents, and early autonomous patterns, aligned with team maturity and risk tolerance.
Constraints
Identify and remove bottlenecks that limit AI’s effectiveness. These may be technical, procedural, or cultural. AI exposes constraints. High performers eliminate them.
Engineering Impact
Measure what actually changes across delivery speed, predictability, flow efficiency, quality, and rework. Usage alone is not impact.
Results
Connect engineering improvements to business outcomes such as time to market, delivery confidence, and efficiency. This is where AI investment becomes defensible.
Teams that progress through these stages intentionally see compounding gains over time. Those that skip steps tend to stall.
Measurement Is the Missing Link
A striking number of organizations adopt AI without the ability to answer basic questions:
Which teams are using AI consistently?
Where is it helping and where is it not?
Are delivery outcomes improving, or just activity levels?
High-performing teams track AI impact against the same delivery metrics they already care about. They focus on trends over time, not one-off wins or vanity usage statistics.
This shift from anecdote to evidence is what allows leaders to scale AI with confidence.
From Experimentation to Competitive Advantage
AI-augmented engineering is no longer a future concept. It is already creating a widening gap between teams that can operationalize AI and those that cannot.
The teams pulling ahead do three things consistently:
They measure adoption instead of assuming it
They remove constraints before scaling usage
They tie engineering performance directly to business outcomes
AI creates leverage, but only when supported by the right fundamentals.
How Plandek Supports This Transition
Plandek is built to support engineering leaders navigating this shift. The platform provides end-to-end visibility across delivery, flow, quality, and predictability, helping teams understand where AI is having real impact and where constraints remain.
With Dekka, Plandek’s AI copilot for engineering analytics, leaders can ask natural-language questions, monitor adoption patterns, surface risks, and receive proactive insights using the data already flowing through their engineering systems.
The Bottom Line
AI tools alone do not create high-performing engineering teams. Systems do.
The organizations that win will be those that operationalize AI deliberately, measure what matters, and continuously remove the constraints holding them back. The RACER Framework provides a practical blueprint to get there. Click here to see the full framework.
Written by
Charlie Ponsonby
Co-founder & CEO
Charlie started his career as an economist working on trade policy in the developing world, before moving to Accenture in London. He joined the Operating Board of Selfridges, before moving to Open Interactive TV and then Sky where he was Marketing Director until leaving to found Simplifydigital in 2007. Simplifydigital was three times in the Sunday Times Tech Track 100 and grew to become the UK’s largest TV, broadband and home phone comparison service, powering clients including Dixons-Carphone, uSwitch and Comparethemarket. It was acquired by Dixons Carphone plc in April 2016. He co-founded Plandek with Dan Lee in 2018. Charlie was educated at Cambridge University. He lives in London and is married with three children.
See how your engineering efforts translate into measurable business impact
Measure delivery performance, AI impact, and engineering productivity with hundreds of metrics, OOTB dashboards and custom configurations.
Contact us
LONDON - HQ
Unit 313 The Print Rooms, 164-180
Union St, London SE1 0LH












