Back to Blog
·5 min read

MSK's AI System Reviews Patient Safety Incidents 29x Faster

Memorial Sloan Kettering's new AI-ILS platform uses LLMs to analyze healthcare incidents with 88% expert accuracy, transforming patient safety workflows.

AI in HealthcarePatient SafetyLLM ApplicationsMedical AI

When I first read about Memorial Sloan Kettering Cancer Center's new AI-based incident analysis system, I was struck by how elegantly it addresses a persistent problem in healthcare: the bottleneck in learning from safety incidents. Published in npj Digital Medicine just this week, this research represents a meaningful step forward in applying large language models to high-stakes clinical environments.

A researcher working in an MSK lab
A researcher working in an MSK lab

The Problem with Traditional Incident Review

Healthcare institutions generate thousands of incident reports annually, from near-misses to adverse events. Each report holds potential lessons for improving patient safety, but the traditional review process is painfully slow. Human experts must manually classify each incident, identify root causes, and extract actionable insights. This creates a fundamental tension: thorough analysis takes time, but delayed learning means preventable errors can recur.

Dr. Jean Moran, Director of Division of Radiotherapy Physics at MSK, summarized it well: "AI-assisted review and classification of incidents can accelerate learning to improve patient safety, allowing teams to shift focus to designing safer systems in support of patients."

How AI-ILS Works

The Artificial Intelligence-based Incident Analysis and Learning System (AI-ILS) was developed by a team led by medical physics resident Dr. Abbas Jinia, under the supervision of Drs. Jean Moran and Anyi Li. The system employs an approach borrowed from aviation safety: the Human Factors Analysis and Classification System (HFACS).

HFACS provides a structured framework for categorizing incidents across four levels: unsafe acts, preconditions for unsafe acts, unsafe supervision, and organizational influences. This methodology has proven effective in aviation for decades, and MSK's team adapted it specifically for healthcare contexts.

The AI-ILS model was trained on over 1,500 synthetic incident reports that had been expertly curated and categorized. The team then validated performance against 350 real-world clinical incidents. The results are impressive:

  • 29x faster than traditional human review
  • 88% concordance with expert classifications
  • 0.92 AUROC (area under the receiver operating characteristic curve)
  • 79% overall accuracy in cause classification

Why This Matters for Healthcare AI

What distinguishes this work from many AI healthcare applications is its focus on transparency. The tool promotes an interactive experience where reviewers can interrogate and understand the AI's reasoning. This is not a black-box system that generates unexplainable outputs. Instead, clinicians can examine why the model classified an incident in a particular way.

This design choice reflects a mature understanding of AI deployment in high-stakes settings. Trust is not built by accuracy metrics alone. Healthcare professionals need to understand, verify, and occasionally override AI recommendations. The MSK team built this capability directly into their system.

Implications for Healthcare Institutions

For hospital administrators and patient safety officers, this research signals a shift in how incident learning can scale. The traditional approach forces a choice between thoroughness and timeliness. AI-ILS suggests a third option: automated initial classification that maintains expert-level accuracy while dramatically reducing review time.

This does not eliminate the need for human oversight. Rather, it redirects human expertise toward higher-value activities. Instead of spending hours on initial classification, safety teams can focus on designing system-level interventions, analyzing trends across multiple incidents, and implementing preventive measures.

Practical Considerations for Implementation

Organizations considering similar systems should note several factors:

  • Training data quality: MSK used expertly curated incidents, not raw unreviewed reports
  • Framework selection: HFACS provided structure, but other taxonomies might suit different contexts
  • Validation requirements: Testing against real-world cases (not just synthetic data) proved essential
  • Explainability features: Interactive reasoning capabilities built trust with clinical users

The Broader Trend

This MSK research fits within a larger pattern I have been observing: LLMs are finding their most valuable applications not in replacing human judgment, but in handling high-volume, structured classification tasks that currently bottleneck expert attention.

Healthcare incident review is a perfect example. The task requires domain knowledge (which the model learns from training data), follows established taxonomies (HFACS), and benefits enormously from speed improvements. These characteristics make it well-suited for LLM augmentation.

Similar patterns are emerging in legal document review, financial compliance monitoring, and scientific literature triage. The common thread is that LLMs excel at processing large volumes of text against established frameworks, freeing human experts to focus on edge cases and strategic decisions.

Looking Forward

The MSK team's work demonstrates that LLMs can be deployed responsibly in healthcare settings where patient safety is paramount. The key is thoughtful system design: transparent reasoning, appropriate validation, and clear integration with existing human workflows.

As healthcare institutions worldwide grapple with information overload and limited expert time, solutions like AI-ILS point toward a more sustainable approach to continuous safety improvement. The 29x speed improvement is striking, but the real value lies in enabling faster organizational learning cycles that ultimately protect patients.

For AI practitioners in the Gulf region and beyond, this research offers a compelling template for deploying LLMs in high-stakes domains: start with well-defined taxonomies, prioritize explainability, validate rigorously, and design for human-AI collaboration rather than replacement.

Book a Consultation

Business Inquiry