I spend most of my working hours inside data rooms. Financial models, cap tables, IP schedules, vendor agreements, employment contracts — the accumulated documentation of a company's life laid out for inspection. After enough transactions, you develop a sense for what's actually there versus what the sellers want you to see. The gap between those two things is where deals go wrong.
Technology M&A has always been difficult. You're acquiring intangible assets — software, data, intellectual property, and engineering talent — that resist conventional valuation and punish sloppy diligence. The technical debt buried in a codebase, the customer concentration risk hiding in a revenue model, the culture mismatch that only becomes visible ninety days post-close: these are the realities that have always made technology transactions among the most unforgiving in any deal environment. Bain & Company's research puts it starkly: roughly 60% of tech M&A deals fail to create their projected value. That's not a rounding error. That's a structural problem with how these deals get evaluated.
AI has made it worse. Not because AI introduced entirely new categories of risk — though it has introduced some — but because it has dramatically amplified existing risks and created blind spots in due diligence processes that haven't kept pace with how technology companies are actually being built today. PwC reports that AI-related M&A deal volume increased approximately 70% between 2021 and 2024. The money is flowing. The diligence frameworks have not kept up.
The AI Due Diligence Gap
Most technology due diligence processes were designed for a world where a software company's primary asset was its codebase. The review examined code quality, technical debt, scalability, security posture, and intellectual property ownership. These remain important. But they are no longer sufficient.
In 2025, a growing proportion of technology companies — particularly in fintech, SaaS, healthtech, and emerging technology categories — are building their core value proposition on AI-enabled capabilities. Their pitch decks lead with model performance metrics. Their valuations are predicated on defensible AI moats. And most acquirers are still performing due diligence as if the underlying technology were conventional software.
The result is a systematic underassessment of AI-specific risk: the quality and provenance of training data, the reliability and explainability of models in production, the exposure to AI regulatory frameworks that are moving faster than most legal teams realize, and the degree to which AI capabilities are genuinely proprietary versus assembled from open-source components that create dependency risks no one has mapped.
"The acquirers paying a premium for AI capability need to know exactly what they're buying — because often, it's not what the pitch deck implies."
The Model That Looked Great
A case from our recent work illustrates the problem. We were engaged by a North American SaaS acquirer in late-stage diligence on a $45 million AI-native target. On paper, everything checked out. The target had strong recurring revenue, impressive customer retention, and model performance metrics that justified the premium the acquirer was paying. The management presentation was polished. The financial model was clean.
Our technical diligence went deeper than the standard review. When we examined the core ML model's training pipeline — not just the outputs, but the inputs and the agreements governing them — we found that the model had been trained on data the target didn't have clear licensing rights to. The issue was buried in a third-party data agreement that had been executed early in the company's life, before anyone anticipated an acquisition scenario. The language was ambiguous on derivative works. The data provider's terms had been updated twice since the original agreement, and neither update had been acknowledged or countersigned.
This wasn't fraud. It was the kind of oversight that happens when a fast-moving startup builds first and worries about legal architecture later. But for the acquirer, it represented material liability. If the data provider chose to enforce, the acquirer would be left holding an AI system whose core training data was legally compromised — and rebuilding from clean data would mean months of retraining, revalidation, and potential performance degradation.
We quantified the exposure and presented it to both parties. The deal was restructured to account for the liability, saving the acquirer approximately $8 million in post-close exposure. The transaction still closed. But it closed at a price that reflected what was actually being acquired, not what the pitch deck said was being acquired.
None of this would have surfaced in a standard diligence process. The financial model was accurate. The code was well-written. The team was strong. The risk was in a layer that most acquirers never examine.
The Five Hidden Risks We Most Frequently Encounter
Across dozens of technology transactions, we see the same categories of AI-specific risk surface repeatedly. Each one is capable of destroying value post-close, and each one is routinely missed by conventional diligence processes.
1. Data Liability
AI systems are only as valuable as the data they're trained on — and data carries legal, ethical, and strategic risk that is frequently invisible in standard due diligence. The questions that matter: Where did the training data originate? Were data subjects' rights managed in a way that will survive regulatory scrutiny, not just today but under the regulatory frameworks that are coming? Is there consent documentation, and does it cover the specific use cases the model is deployed for? Are there licensing arrangements around third-party data that create ongoing cost obligations, usage restrictions, or — as in our $45 million case — ambiguous derivative-works provisions?
Beyond the legal dimension, acquirers need to understand data dependency. If a model depends on a continuous feed from a third-party data provider, what happens when that contract comes up for renewal post-close? We've seen cases where data providers, sensing an acquisition, have used the leverage to renegotiate terms at multiples of the original price. If your model can't function without that data, you've acquired a business with a supplier concentration risk that wasn't in the financial model.
Diligence should answer: What percentage of training data is proprietary versus licensed versus sourced from public datasets with terms that could change? What is the total annual cost of data access, and what does the renewal landscape look like? Has the company ever received a takedown request, a cease-and-desist, or a data subject access request that could signal broader exposure?
2. Model Degradation Risk
This is the risk that receives the least attention and causes the most post-close damage. AI models are not static assets. They degrade. Distribution drift, changes in user behavior, shifts in the underlying data environment — all of these erode model performance over time. A model that performs at 94% accuracy today may perform at 78% in eighteen months if it isn't actively maintained.
"A trained model is not like software you ship and maintain. It degrades. And nobody in the data room is going to volunteer that information."
Due diligence rarely examines the operational infrastructure required to sustain an AI system's performance over time. Most acquirers look at the model's current metrics and assume continuity. They don't ask how often the model is retrained, what the retraining pipeline looks like, what monitoring is in place to detect performance decay, or what the cost of maintaining model quality actually is on an annualized basis.
We've seen acquirers discover post-close that a model touted as production-grade was, in practice, being manually adjusted by a single data scientist on a quarterly basis. No automated retraining pipeline. No drift detection. No documented performance thresholds that trigger intervention. The model worked — until it didn't, and nobody had built the infrastructure to catch the decline before customers did.
Diligence should answer: What is the model retraining cadence, and is it automated or manual? What monitoring is in place for performance drift? What is the annualized cost of model maintenance, including compute, data refresh, and personnel? Has model performance declined at any point in the last 24 months, and if so, what was the root cause and remediation?
3. Regulatory Exposure
AI regulation is moving fast and unevenly. The EU AI Act imposes classification requirements and compliance obligations that many AI-native companies have not fully assessed. North American regulatory frameworks are evolving at both the federal and state level. Sector-specific guidance in financial services, healthcare, and insurance is creating compliance obligations that vary by jurisdiction and by use case.
For acquirers in regulated industries, the regulatory exposure embedded in an acquisition target's AI systems can be material — and it is almost always underpriced. The reason is straightforward: most AI-native targets were built by teams focused on product and growth, not regulatory architecture. They may not have conducted a systematic assessment of which regulatory frameworks apply to their AI systems, what compliance gaps exist, or what the remediation timeline and cost would look like.
The risk isn't just fines. It's operational disruption. If a regulator determines that a model deployed in a regulated context doesn't meet explainability requirements, the acquirer may be forced to pull the product from market while remediation occurs. That's revenue loss, customer churn, and reputational damage — none of which appeared in the acquisition model.
Diligence should answer: Which regulatory frameworks apply to each AI system in production, by jurisdiction? Has the target conducted a formal AI risk assessment? What is the explainability posture of each model — can the company demonstrate why any given output was produced? Are there pending or anticipated regulatory actions in any jurisdiction where the product is deployed?
4. Technical Debt in AI Infrastructure
Traditional technical debt — in codebases, databases, and infrastructure — is well understood in diligence. AI technical debt is less well understood but equally consequential, and Deloitte's research suggests that post-merger integration costs typically run 2-3x higher than budgeted even before AI-specific debt enters the picture.
AI technical debt takes forms that conventional code review doesn't catch. Model architectures that were state-of-the-art when originally built may now be difficult to extend, retrain, or integrate with modern tooling. Data pipelines are often fragile, poorly documented, and held together by scripts written by people who are no longer at the company. Feature engineering logic may live in notebooks rather than production code. ML infrastructure may be tightly coupled to a specific cloud vendor, a specific framework version, or a specific hardware configuration that won't survive integration into the acquirer's environment.
The cost of remediating AI technical debt post-close is routinely underestimated because the people doing the estimating — usually the target's own engineering leadership — have normalized the debt. They've been living with it. They know the workarounds. They don't experience it as debt. The acquirer's integration team, encountering the system for the first time, experiences it very differently.
Diligence should answer: What is the full dependency map of the ML infrastructure, including cloud services, frameworks, and hardware? How much of the ML pipeline is documented and reproducible versus tribal knowledge? What would it cost and how long would it take to retrain the core models from scratch if the current infrastructure were unavailable? Are there single points of failure in the data or model pipeline that would cause production outages?
5. Talent Concentration
AI capability is, more often than not, talent that walks out the door at five o'clock. This is true of technology companies generally, but it is acutely true of AI-native companies where the core intellectual property often exists less in the codebase and more in the judgment, intuition, and institutional knowledge of a small number of individuals.
The most important thing an acquirer can do is understand precisely who built the AI capability, who sustains it, and what the retention risk looks like post-close. In many AI-native acquisitions, the core capability is embodied in fewer people than the headcount suggests. You might see a "data science team" of twelve, but when you trace who actually architected the models, who makes the critical decisions about retraining and feature selection, and who gets called when something breaks at 2 AM, it's often two or three people. And those people frequently have options — competing offers, the ability to start their own company, or simply the financial independence to walk away if the post-acquisition environment isn't what they want.
Standard retention packages — earnouts, golden handcuffs, vesting schedules — help, but they are not sufficient if the integration experience drives key talent away before the retention period ends. We have seen acquirers structure generous three-year retention packages only to watch their most critical AI talent leave at month eight because the integration destroyed the working environment that kept them engaged.
Diligence should answer: Who are the three to five individuals without whom the AI capability cannot be maintained or extended? What are their personal circumstances, career motivations, and flight risk indicators? What does their compensation and equity position look like relative to market, and what would a competitive retention package require? What integration decisions — reporting structure, tooling choices, project autonomy — are most likely to drive their departure?
A Better Approach to Tech M&A Due Diligence
The acquisitions that generate value are the ones where the acquirer actually knows what they're buying. This sounds obvious. In practice, it requires a significant expansion of the traditional due diligence scope — and the willingness to slow down a process that investment bankers are invariably trying to accelerate.
We recommend that any technology acquisition where AI is a material component of the value thesis include: a structured AI capability assessment conducted by practitioners who have built and deployed models in production, not generalists running a checklist; explicit data provenance and licensing review that goes beyond the IP schedule to examine the actual agreements governing training data; regulatory risk mapping across every jurisdiction where the AI system operates or is likely to operate post-close; a technical debt assessment that specifically addresses ML infrastructure, not just application code; and a talent dependency analysis that identifies the critical individuals and pressure-tests the retention plan against realistic integration scenarios.
The integration planning process should start in due diligence, not after close. The question of how AI capabilities will be maintained, extended, and eventually integrated into the acquirer's technology environment is not a post-close problem — it is a diligence-stage problem, because the answers materially affect valuation, deal structure, and the likelihood that the acquisition will generate the returns the board approved.
The Bottom Line on Tech M&A Risk
Technology M&A has a long history of destroying value post-close, and AI is creating new ways to do so. But the risk is manageable — with the right diligence process, the right expertise, and the discipline to ask hard questions even when the deal momentum is pushing in the opposite direction.
The acquirers who will generate sustainable value from technology M&A in the AI era are the ones who treat diligence as the most important phase of the deal, not a hurdle on the way to close. Every deal has a story the sellers want to tell. Your job is to find the one they don't.