In January 2026, OpenAI launched ChatGPT Health. Anthropic released Claude for Healthcare. Google shipped MedGemma 1.5. Three foundation model companies entered healthcare in the same week.
The demos were incredible. Accurate differential diagnoses. Ambient note generation that captures a full patient encounter. Drug interaction checks that outperform the average physician on standardized tests.
But the FDA has now cleared over 650 AI-enabled medical devices, roughly 69% of them in radiology. The vast majority sit unused or underused in clinical settings. The gap between “the AI works” and “the AI works inside the healthcare system” is where companies go to die.
Here’s why.
Why demos are easy
A healthcare AI demo runs under conditions that will never exist in production.
The data is clean. It’s a curated dataset, often from a single institution, with consistent formatting and complete records. The audience is cooperative. They’re investors or conference attendees, not a burned-out hospitalist on hour 11 of a 12-hour shift. There’s no compliance layer. No HIPAA review. No security assessment. No IT integration ticket that sits in a queue for 14 weeks.
And the evaluation metric is controlled. “The model got 92% accuracy on this benchmark” is a fundamentally different claim than “the model improved patient outcomes across 47 hospital sites over 18 months.”
Every successful demo is an existence proof. It proves the AI can work. It says nothing about whether it will work, at scale, inside the system.
Failure mode 1: the EHR walled garden
Electronic health records are the operating system of American healthcare. Epic alone holds roughly 42% market share among U.S. hospitals, with Oracle Health (formerly Cerner) second. Together they cover more than half of all hospital beds.
Getting data out of these systems is the first hard problem. The 21st Century Cures Act mandates interoperability and prohibits information blocking, but mandates and reality are different things. FHIR (Fast Healthcare Interoperability Resources) is the standard, but implementation varies wildly across sites.
The practical result: integrating a new AI tool with a hospital’s EHR takes 6-18 months, depending on the vendor, the institution, and whether the APIs you need actually exist. That’s 6-18 months before the product delivers value, while your burn rate stays the same.
IBM Watson Health learned this lesson. After years of development and a high-profile partnership with MD Anderson Cancer Center, the system couldn’t reliably integrate with the hospital’s existing records infrastructure. The project was paused after spending an estimated $62 million. IBM eventually sold most of Watson Health to Francisco Partners in 2022.
Failure mode 2: the regulatory surface area
HIPAA is the floor, not the ceiling.
Healthcare data breaches cost an average of $10.93 million per incident, the highest of any industry. Over 700 major breaches were reported to HHS in 2023 alone. But breach risk is only one layer.
If your AI product touches clinical decisions, the FDA may require clearance as a medical device. The 510(k) pathway has a 90-day review goal and costs $26,067 in FY2026 user fees ($6,517 for small businesses). If there’s no predicate device, you’re looking at the De Novo pathway: 150-day review goal, $173,782 in fees. PMA for high-risk devices runs $579,272.
Then there’s the state layer. In 2024 alone, 31 states enacted some form of AI legislation. Washington’s My Health My Data Act applies to any company collecting health data, not just HIPAA-covered entities — including app developers and fitness platforms. A startup shipping a product nationally navigates federal rules plus a patchwork of state-level requirements that don’t align.
Babylon Health is the cautionary tale. At peak, the UK-based telehealth company was valued at $4.2 billion after its SPAC merger in 2021. Its AI triage chatbot made bold accuracy claims. But scaling into the U.S. market meant navigating state-by-state licensure, insurance contracting, and regulatory compliance that the demo never had to face. By 2023, Babylon had filed for bankruptcy, having burned through over $1 billion.
Failure mode 3: the workflow problem
Clinicians don’t resist AI because they don’t trust the technology. They resist it because it doesn’t fit how they work.
A 2024 meta-analysis of 16 studies found that physicians override 90% of drug-drug interaction alerts. A broader scoping review across 34 studies found override rates ranging from 55% to 98%. In ICU settings, AHRQ reports that 66 beds can generate over 2 million alerts monthly — roughly 187 warnings per patient per day. Adding another AI-generated alert to an already overwhelmed system doesn’t help. It makes things worse.
The ambient scribe space is the clearest proof that workflow integration matters more than raw AI capability. Nuance’s DAX Copilot (acquired by Microsoft for $19.7 billion in 2022) works because it listens passively during the encounter and generates notes afterward. The clinician’s workflow doesn’t change. They talk to the patient the way they always have. The documentation happens in the background.
Compare that to AI products that require clinicians to open a new app, enter data in a specific format, or interpret results from an unfamiliar interface. Those products work in demos. They fail at the point of care.
The ambient scribe market validated this approach fast. Abridge raised a $250 million Series D in February 2025 and is now deployed in over 100 of the largest U.S. health systems. Nabla, Suki, and others followed the same playbook. The common thread: none of them make clinical decisions. They remove documentation burden. That constraint is what makes them deployable.
Failure mode 4: accuracy at the tail
Healthcare AI faces a problem that most industries don’t: the cost of being wrong can be catastrophic, and the cases where AI is most likely to be wrong are the cases where getting it right matters most.
Rare conditions, atypical presentations, patients with multiple comorbidities: these are the long tail where AI models trained on common cases break down. A model that’s 95% accurate on the 50 most common diagnoses may be functionally useless for the 500 rare conditions it’s never seen enough training data to learn.
And liability is unresolved. If an AI system misses a cancer diagnosis, who’s responsible? The physician who relied on it? The hospital that deployed it? The company that built it? The AMA has called for clarity but no definitive legal framework exists yet. Until it does, risk-averse health systems will hesitate to deploy AI in clinical decision-making, no matter how good the demo looks.
Failure mode 5: nobody knows who pays
This might be the most underrated problem. A healthcare AI product can work, be safe, be integrated, and still fail because the business model doesn’t map to how healthcare money flows.
There are limited CPT codes for AI-assisted services. CMS has added a few for AI-enabled radiology analysis, but coverage is narrow. If a payer won’t reimburse for the service the AI provides, the provider can’t justify paying for it.
The result is that most healthcare AI startups sell to the institution, not the payer. That means enterprise sales cycles of 12-24 months, procurement committees, pilot programs, and the ever-present risk that a champion leaves and the deal dies.
Tempus is one of the few companies that figured this out. Instead of selling AI directly, they built a data business underneath it. Genomic sequencing, structured clinical data, and analytics that hospitals pay for regardless of AI adoption. The AI sits on top of a revenue model that works without it. Tempus went public on Nasdaq in June 2024 and hit $1.27 billion in FY2025 revenue — 83% year-over-year growth — connecting to roughly 60% of U.S. academic medical centers.
What actually works
The companies crossing the demo-to-production gap share a few traits:
They solve a workflow problem, not a technology problem. Nuance DAX doesn’t ask clinicians to change behavior. Tempus doesn’t require hospitals to rethink their data strategy. The AI is invisible infrastructure, not a new tool to learn.
They own the data layer. Whoever controls the data pipeline controls the value chain. Tempus built this from scratch. Epic is building it from within. Startups that depend on someone else’s data access are structurally fragile.
They price for how healthcare buys. Per-encounter pricing, risk-sharing models, or bundling with existing workflows. Not SaaS subscriptions that don’t map to clinical economics.
They treat compliance as a product feature, not an afterthought. The companies that embed HIPAA, state privacy laws, and FDA requirements into the product design ship faster than the ones that bolt it on later.
The real question
The demo-to-production gap in healthcare AI isn’t about whether the models are good enough. They are. GPT-4 passes the USMLE. Ambient scribes generate notes that clinicians prefer to their own. Radiology AI catches findings that humans miss.
The models work. The question is whether they work inside a system designed 30 years ago for paper records, built on siloed data, governed by overlapping regulations, and paid for through a reimbursement structure that doesn’t have a line item for “AI.”
That’s the hard problem. The companies that solve it won’t be the ones with the best model. They’ll be the ones that understand how the system actually works, and build for that reality instead of the demo.
Sources
- FDA AI/ML-Enabled Medical Devices
- IBM Cost of a Data Breach Report 2023
- HHS OCR Breach Portal
- Becker’s: Epic EHR Market Share (KLAS 2024)
- 21st Century Cures Act: Information Blocking
- STAT News: IBM Watson + MD Anderson
- IBM: Francisco Partners to Acquire Watson Health Assets (2022)
- CNBC: Babylon Health $4.2B SPAC Deal (2021)
- Forbes: Babylon Health Files for Bankruptcy (2023)
- DDI Alert Override Meta-Analysis 2024 (PubMed)
- DDI Alert Override Scoping Review 2022 (PubMed)
- AHRQ PSNet: Alert Fatigue Primer
- Microsoft Completes Acquisition of Nuance ($19.7B)
- Abridge $250M Series D
- Tempus Revenue FY2025
- NCSL: AI 2024 Legislation
- Washington My Health My Data Act (RCW 19.373)
- FDA 510(k) Premarket Notification
- FDA De Novo Classification
- FDA MDUFA Fees FY2026
- AMA: Augmented Intelligence in Medicine
- AMA: CPT and AI
- GPT-4 on Medical Challenge Problems (Microsoft Research)