top of page

Why Hospital “AI Agents” Fail the Moment Work Gets Real

  • Jan 29
  • 6 min read

The global mood around AI in healthcare is confident and ambitious, but quietly uneasy.

At the World Economic Forum in Davos, political and business leaders warned that artificial intelligence is moving faster than institutions’ ability to govern it, with trust, not intelligence, emerging as the real constraint on adoption. In the Gulf, that caution sits alongside aggressive investment. Dubai-based companies are launching AI-driven platforms in drug discovery and diagnostics, positioning the region as a serious healthcare AI hub. Meanwhile, WHX Dubai (formerly Arab Health) is expanding its 2026 programme to focus explicitly on AI, digital infrastructure, and system-level healthcare transformation, a clear signal that the region wants to move beyond pilots.


Against this backdrop, a quieter pattern continues to repeat inside hospitals and clinical laboratories: The demos work. The pilots impress. The systems don’t hold.


I’ve spent the last few years trying to make agentic AI function inside real hospital networks and labs. Not as slideware, not as proof-of-concepts, but as systems that genuinely reduce workload and risk.


And I've seen this again and again. What failed wasn’t intelligence, it was the foundations.


When promise collides with practice


The pitch is now familiar. A confident Copilot interface. A promise to “connect it to the EHR.” Reassurance that a larger model can be swapped in if accuracy dips. Everyone moves quickly. Almost no one asks the critical question:


"What happens when the workflow doesn’t behave?"


Because healthcare workflows never behave. They run on interruptions, missing fields, informal workarounds, tacit judgement, and handovers that only make sense to the people carrying them. Many AI systems are built for an organisation that exists on paper, not the one delivering care at 2am.


What is consistently underestimated is that workflow is not just people and steps. It is also the integration estate: the web of systems, identifiers, queues, interfaces, and local variations that every “simple” process quietly depends on. If you don’t surface that estate early, you discover it late; when costs rise, timelines slip, and confidence erodes.


How failure actually shows up



These systems rarely fail at once in a big bang. There is no siren with a big red flashing sign that shouts "SYSTEM FAILURE!". They simply fail by becoming work.


Over time, prompts get longer. Retrieval scopes widen “just to be safe.” Outputs sound fluent but grow inconsistent precisely where stakes are highest. Humans begin checking, correcting, explaining, and compensating.


At some point, the realisation lands: operational load hasn’t decreased. A new operational layer has been created: one that costs money, consumes attention, and quietly eats margin.

This is why so many healthcare AI initiatives stall after pilots. Not because the technology is incapable, but because the organisation has not built the conditions for it to operate safely and predictably.


You don't need a "better model"


It is tempting to believe this is an intelligence problem that can be fixed with more parameters, better reasoning, and improved tool calling. Larger models may sound more confident, but confidence that outpaces reliability is not progress. It is risk.

Especially in healthcare, that assumption is wrong. What breaks first is foundation:


  1. the workflow truth: built for the ideal, not the real.

  2. the trustworthiness of signals: good data vs usable data

  3. the system’s ability to remember correctly (and prove what it did)

  4. the orchestration and platform layer that makes multiple modules behave like one system


If any of those are weak, the rest turns into expensive improvisation.


Let me explain.


Larger models may sound more confident, but confidence that outpaces reliability is not progress. It is risk.


1) Built for the ideal, not the real


Most agentic builds begin with process maps, SOPs, and policy decks. Useful.. but incomplete.


Real work includes undocumented exceptions, alerts people ignore for good reason, and informal escalation paths that keep patients safe. Critically, it includes how systems are actually used - which records are trusted, which screens are bypassed, and where human judgement fills the gaps.


The only reliable source of workflow truth is frontline behaviour: shadowing teams, capturing exceptions, and observing what happens under strain. If this feels slow, it should. Skipping it is how organisations automate an imaginary version of themselves.


2) “Good data” is not the same as “usable data”


Healthcare systems can pass every data quality check and still feed the wrong signal into an AI system.


The same metric can mean different things depending on timing, patient context, system of record versus convenience, or downstream failures. Clinicians compensate instinctively. Models do not, unless trust rules are explicitly designed.


The question is not whether data is accurate, but which signals humans rely on in that moment, and which they discount. Until that distinction is encoded, systems will continue to produce outputs that are technically correct and operationally wrong.


3) Memory and orchestration: the missing middle


Treating memory as a document store is a category error. Agentic systems need experience: a structured record of past decisions, constraints, outcomes, and safe fallbacks that can be reused consistently. The builds that started behaving were the ones where memory was treated like a first-class design surface:


  • What decision did we make last time this exception happened?

  • What constraints mattered?

  • What counts as “done”?

  • What is the safe fallback when a dependency fails?

  • What evidence is acceptable in this workflow?


That’s not a PDF.


The thing they confused (and it costed them): chatbot vs modules vs orchestration.

In hospital and lab networks, people kept using the words interchangeably:


  • “chatbot”

  • “agent”

  • “multi-agent”

  • “automation”


But underneath, we were often building a bundle of workflow automations with a chat UI on top, without a real orchestration layer to manage state, routing, tool selection, fallbacks, guardrails, and audit. That muddied everything. "Is this a chatbot answering questions?", "Is it a set of task modules?", "Is it an agentic system coordinating work across steps?"

When you don’t decide that at an infrastructure level, you don’t get clarity. You get a platform that behaves inconsistently and gets “babysat” by humans.


The economics of uncertainty


Unstable systems cost more in every direction: more context, more retries, more human oversight. Bigger models increase token spend while leaving the underlying uncertainty intact.


In clinical environments, auditability draws the line between a tool and a liability. If you cannot trace what the system believed it was doing, what evidence it used, what assumptions it made, and where it deferred, trust collapses, quickly and permanently.


Where value actually lives


Healthcare organisations that succeed with AI do not start by asking which agent to build. They start with a bleeding workflow, a measurable outcome, and a tight operational loop. Then treat the result as a platform, not a one-off.


The most promising early use cases are repetitive, exception-heavy, costly when they fail, and already handled in an “agent-like” way by humans. These are not chat problems. They are workflow integrity problems.


Solve them properly once, and the patterns can be reused many times.


Auditability is the difference between a tool and a liability


In healthcare and labs, its not as simple as just getting an answer. You need to know:


  • what workflow step it thought it was in

  • what memory it relied on

  • what evidence it pulled

  • what assumptions it made

  • what it couldn’t verify

  • what module acted, and what orchestrator rule allowed it to act


If you can’t trace that cleanly, you haven’t built an operational system. You’ve built a conversational layer that creates new accountability problems.


Frontline teams sense that immediately, even if they can’t explain it in technical terms.

They’ll simply stop trusting it.


The uncomfortable truth


AI in healthcare will not fail because we lack intelligence. It will fail because organisations lack capacity: the operational, organisational, and human ability to deploy AI responsibly at scale.


The Gulf’s ambition is real, and the investment is significant. But moving from pilots to systems requires more than technology. It requires building AI capacity at every level of the organisation: executives who understand risk and economics, operators who understand workflows, technical teams who design memory and orchestration, and frontline staff who trust what they are given.


In my next article, I'll explain how the healthcare sector should approach building AI agents that are trustworthy and deliver impactful outcomes.


This article was written by Tim Daines, Programme Director of the Cambridge Labs, a capacity-building lab that works with leaders and frontline teams to develop the skills, structures, and operating models required for AI to function reliably in complex environments such as healthcare.


 
 
 

Comments


bottom of page