← Insights
·4 min read

Clinical AI Adoption: Why Most Pilots Don't Survive Contact With the Real World

G

Ger Perdisatt

Founder, Acuity AI Advisory

Clinical AI pilots fail for predictable reasons. Understanding the failure modes in advance is the difference between a pilot that scales and one that gets quietly discontinued.

Clinical AI pilots have a low survival rate. A system that performs well in controlled evaluation, earns support from the clinical champion who proposed it, and passes procurement with strong references from other health systems frequently fails to embed in routine clinical practice. The technology continues to work. The deployment doesn't.

The failure modes are predictable. Understanding them in advance doesn't guarantee success, but it dramatically changes the odds.

Governance gaps

The most common failure is not clinical or technical. It's governance. The pilot launches without clear answers to the questions that determine whether it can be sustained: who owns this system, who monitors its performance, who decides when it needs to be reviewed or suspended, and how are errors and near-misses recorded and escalated?

Without those answers, clinical AI pilots operate in a governance vacuum. They work when the champion is engaged, and drift when the champion is on leave, reassigned, or loses interest. The system has no institutional home — it's a project, not an infrastructure — and projects end.

This gap is addressable before launch. It requires thirty to sixty days of governance design work: defining the oversight structure, assigning accountability, establishing performance monitoring, and building the error reporting pathway. It is unglamorous, it delays go-live, and it is the most consistently neglected step in clinical AI adoption.

Workflow mismatch

Clinical AI systems are typically designed and evaluated in conditions that don't reflect the time pressure, interruption frequency, and cognitive load of actual clinical work. A diagnostic support tool that requires three clicks and thirty seconds to consult, and that produces output that takes another thirty seconds to interpret, will be abandoned in a high-acuity clinical environment where the clinician has four other decisions competing for attention.

This is not a technology problem. It's an integration problem. The tool wasn't designed into the clinical workflow — it was placed alongside it.

Avoiding workflow mismatch requires genuine co-design with the clinical staff who will use the system in their actual working environment. Not a demo with a selected group of enthusiasts, but observation and design work with the full range of staff who will encounter the tool on a typical shift. The friction points that will drive abandonment are visible at this stage. They are rarely visible in a vendor demonstration.

Clinician resistance

Clinician resistance to AI is often attributed to conservatism or technophobia. This is too simple and usually wrong.

Experienced clinicians resist clinical AI for several well-founded reasons. They've seen technology implementations that created work without reducing it. They're concerned about accountability: if the AI recommends X and they follow it and it turns out to be wrong, where does responsibility sit? They're sceptical about performance claims that were validated in a population different from their patient population. And they're aware that AI systems can fail in ways that aren't immediately visible — confidently producing a wrong answer without flagging uncertainty.

These are not unreasonable concerns. Addressing them requires transparency about the AI system's validation data, its known limitations, its failure modes, and its performance on the specific patient population it will be used for. It requires a clear answer on accountability: the clinician retains clinical responsibility, the AI is a decision support tool, and there is a documented human oversight requirement. And it requires time for clinical staff to build calibrated trust through experience with the system — not blind trust in vendor performance claims.

Data quality

Clinical AI systems trained and validated on external data perform differently on local data. Coding practices vary across health systems. Population characteristics differ. Equipment configurations affect imaging outputs. Workflow differences affect what data gets captured and when.

A system that performed well in a UK or US validation study will need to be validated on Irish patient data, in the Irish clinical workflow, before performance claims can be trusted in that context. Local validation is not a bureaucratic requirement. It's the step that ensures the system you've purchased is actually the system you're deploying.

What a successful clinical AI adoption programme requires

Successful clinical AI adoption combines four elements that are rarely present simultaneously in failed pilots.

Clear governance with named accountability, established before deployment. Co-design of the clinical workflow integration, led by the staff who will use the system. Transparent communication with clinical staff about capabilities, limitations, and accountability. And local validation or, at minimum, a structured monitoring programme in the initial deployment period that would surface performance divergence from validation benchmarks.

None of these elements are technically complex. All of them require time and organisational commitment that is often underestimated at the point of procurement enthusiasm.

The pilot that survives contact with the real world is rarely the one with the best technology. It's the one that was introduced into an organisation that had done the governance and workflow work before the technology arrived.

healthcare