There is a conversation happening in nearly every boardroom right now. The technology team has run several AI pilots. Some showed promise. One or two generated enough excitement to present to leadership. But when the question arrives — "so what is our AI strategy?" — the room goes quiet.

This is the AI inflection point. And most organisations are handling it poorly.

The problem is not a lack of ambition, investment, or talent. It is a sequencing problem. Organisations have moved directly from awareness to experimentation without passing through the most important stage: strategic intent. And when experimentation precedes strategy, what you get is a portfolio of interesting pilots, a growing AI budget, and very little enterprise-scale value.

The numbers back this up. McKinsey's 2024 State of AI report found that only 28% of organisations have scaled AI beyond the pilot stage. That figure has barely moved in three years. Gartner projects that 30% of generative AI projects will be abandoned after proof-of-concept by the end of 2025. These are not fringe experiments — these are well-funded initiatives inside serious companies, dying quiet deaths because nobody connected them to a strategy worth executing.

The Pilot Trap

Running AI pilots is not a strategy — it is exploration. Exploration is valuable, but only when it is purposeful. The organisations pulling ahead with AI are not the ones running the most pilots. They are the ones that asked a prior question and actually answered it: what would it mean for us, specifically, to win with AI?

The pilot trap is seductive because it feels like progress. Teams are learning, vendors are engaged, leadership is energised. But without a north star, pilots accumulate rather than compound. They create complexity, spread investment thin, and eventually produce a portfolio of proofs-of-concept that nobody is quite sure what to do with.

We see this pattern repeat with striking consistency. A department head attends a conference, gets excited about a use case, and spins up a pilot. Three months later, another team does the same thing independently. The data science group builds something different again. Before long, you have a dozen pilots running on different infrastructure, different data sets, and different success criteria — none of them connected to each other or to the company's actual strategic priorities.

"The organisations winning with AI aren't running the most pilots. They're the ones who decided what winning looks like before they started."

The Pilot Graveyard: A Case Study

Last year, we worked with a mid-market Canadian fintech company — roughly $80 million in annual recurring revenue — that had become a textbook example of this pattern. Over 18 months, they had launched 14 separate AI pilots across three departments. Customer support had built a chatbot. Risk had prototyped a fraud detection model. Marketing was testing AI-generated content and a predictive churn tool. Operations had three different automation experiments underway.

Not one of these pilots had scaled into production.

The leadership team was frustrated. They had spent real money. They had hired data engineers. They had brought in vendors. And yet, when the CEO asked what the company's AI capability actually was, the honest answer was: 14 experiments and zero enterprise outcomes.

Our diagnostic revealed the root cause quickly. Every pilot was solving a department-level problem. None were solving enterprise-level ones. There was no shared data infrastructure — each team had built its own pipelines, sometimes pulling from the same source systems but transforming data differently. There was no governance framework defining how models should be evaluated, monitored, or retired. And critically, there was no strategic intent connecting any of the 14 initiatives to what actually mattered: the company's competitive position in a market that was consolidating fast.

We helped them collapse the 14 pilots into three strategically aligned initiatives built on a shared data layer. Three. That was the hard conversation — telling teams that their pilot, which they had spent months on and were emotionally invested in, was not going to continue. But the alternative was continuing to spread resources across 14 small bets that would never compound.

Within eight months, one of the three surviving initiatives — a unified customer intelligence model that combined signals from support, product usage, and billing data — was generating measurable margin improvement. Not because the AI was better than what they had before. Because for the first time, the data, the infrastructure, and the strategic intent were all pointed in the same direction.

Why Pilots Fail to Scale

The fintech case is not unusual. In our experience, pilots fail to scale for a small number of recurring reasons, and technical capability is almost never one of them.

No shared data foundation

Each pilot team builds its own data pipeline. When it comes time to scale, there is no common infrastructure to build on, and the cost of rebuilding is prohibitive. The organisation ends up with a dozen small data projects instead of one strategic data asset.

No governance before deployment

Model accuracy is tested, but nobody has defined who owns the model in production, how it will be monitored for drift, what happens when it produces a bad output, or how it interacts with regulatory obligations. These are the questions that kill a pilot the moment it touches real customers or real money.

No connection to competitive strategy

The pilot solves a real problem, but it is not a problem that moves the company's strategic position. It produces a local efficiency gain that, even at scale, would not change the trajectory of the business. Leadership looks at the ROI case and cannot justify the investment to operationalise it.

No change management plan

The model works in testing. But the people who would need to change their workflows to use it were never consulted, never trained, and never given a reason to trust it. Adoption stalls. The pilot sits on a shelf.

Research from MIT Sloan Management Review reinforces this point from the other direction: companies that tightly couple AI initiatives with business strategy outperform their peers by roughly two times on revenue growth. The difference is not spending more on AI. It is spending on AI that is connected to something that matters.

Three Layers of a Real AI Strategy

A real AI strategy works at three layers. Most organisations are only working on one.

Layer One: Strategic Intent

Where can AI actually differentiate this organisation? Not where AI is interesting — where it can shift competitive position. This requires an honest assessment of where the business creates value, where competitors are vulnerable, and where data and workflow conditions are right for AI to make a meaningful difference.

Strategic intent also requires decisions about what AI will not do. Resources are finite. Governance bandwidth is finite. Pursuing AI across every function simultaneously is a recipe for mediocrity everywhere and excellence nowhere. The fintech company we worked with learned this the hard way — 14 pilots meant 14 sets of vendor relationships, 14 sets of data requirements, and zero concentrated impact.

The discipline is in choosing. A strong AI strategy typically identifies two to four high-conviction bets where the intersection of data readiness, business impact, and organisational capability creates genuine potential. Everything else goes on a backlog — not abandoned, but deliberately deferred.

Layer Two: Foundational Capability

AI strategy lives or dies on data. Before an organisation can deploy AI at scale, it needs honest answers to three questions: What data do we have, and is it fit for purpose? What data infrastructure do we need, and what is the realistic path to build it? And what talent and governance capabilities must we develop internally — versus what can we access externally?

Many organisations skip this layer in their enthusiasm to deploy. The result is AI systems that underperform because the foundational data and operational conditions were never right to begin with. We have seen companies spend six figures on a machine learning model only to discover that the training data had systemic gaps that made the model unreliable for the exact use case it was built for. The fix was not a better model. It was better data governance — work that should have happened before the model was ever commissioned.

Foundational capability also includes the human layer: do you have people who can evaluate vendor claims critically? Can your engineering team maintain a model in production, or are you dependent on the consulting firm that built it? These are not technical questions. They are strategic ones.

Layer Three: Deployment Architecture

Once intent is clear and foundations are in place, the question becomes how to deploy AI in a way that creates durable value — not short-term efficiency gains that erode as competitors catch up. This means thinking carefully about the build-versus-buy-versus-partner decision for each use case, designing for change management from the outset, and building measurement frameworks that track business outcomes rather than technical metrics.

A model's F1 score does not matter to the board. What matters is whether the AI initiative moved a number that the business cares about — customer acquisition cost, margin, retention, time-to-close, compliance cost. If you cannot draw a direct line from the AI deployment to a business KPI, either the measurement framework is wrong or the initiative is not strategic enough to warrant the investment.

The Governance Imperative

One of the most consistent failure modes we observe is AI governance added as an afterthought. Organisations move quickly to deploy, encounter problems — a biased model, a data privacy issue, an output that causes reputational harm — and then scramble to build governance frameworks in response to incidents rather than in anticipation of them.

In regulated industries like fintech and financial services, this approach carries serious risk. But even in less regulated environments, the reputational and operational costs of AI failures can be significant. The organisations building durable AI capabilities are treating governance as something that gives them an edge — not just a compliance cost.

"Governance is not the brake on your AI strategy. It is the steering. Without it, speed just means you hit the wall faster."

Good governance answers a set of practical questions before deployment, not after: Who owns this model in production? What are the failure modes, and what happens when they occur? How do we detect drift? What data is this model allowed to access, and under what conditions? Who has authority to retrain, retune, or retire it? How do we explain its outputs to a regulator, a customer, or a journalist?

These questions are not bureaucratic overhead. They are the difference between an AI deployment that survives contact with reality and one that becomes a liability.

What Good Looks Like

The most advanced AI adopters we work with share several characteristics. They have a C-suite sponsor who owns AI strategy — not a technology committee. They have made deliberate choices about where AI investment will concentrate in the next 18 months. They have built or are building a data infrastructure that is fit for purpose. And they are measuring AI performance against business KPIs — revenue, margin, customer retention — not technical proxies.

Critically, they have also accepted that AI strategy is not a one-time exercise. It requires active stewardship as the technology evolves, as competitive dynamics shift, and as the organisation learns what works in its specific context. The companies treating AI strategy as a document that gets written once and filed away are the same ones wondering, 18 months later, why their pilots never scaled.

There is also a pattern in how these organisations handle failure. They expect some initiatives to underperform or fail entirely. But because those initiatives were chosen strategically and built on shared infrastructure, the learning transfers. A failed initiative on a shared data layer still improves the data layer. A failed initiative in isolation just burns budget.

The Strategic Question Leaders Should Be Asking

If your organisation is in the pilot phase, the most valuable thing you can do is pause and ask a harder question than "which pilot should we scale?" The harder question is: If we do nothing different in our AI approach over the next two years, what does our competitive position look like?

In most industries, the honest answer to that question is uncomfortable. And that discomfort is precisely the motivation needed to stop experimenting and start building.

But here is the question I find myself asking more often now, the one that keeps coming up in conversations with leadership teams across sectors: What if the real risk is not that your AI strategy fails — but that you execute a perfectly competent AI strategy 18 months too late? The gap between the companies that have figured this out and the companies that are still running pilots is widening every quarter. By 2028, I expect that gap will be the primary predictor of which mid-market companies get acquired and which ones do the acquiring.