Intracranial Hemorrhage
2025
NPJ Digital Medicine
Real-World Performance Evaluation of a Commercial Deep Learning Model for Intracranial Hemorrhage Detection
Emory University · Published in NPJ Digital Medicine (Nature)
Key Findings
Sensitivity was notably lower for subacute bleeds (45.5%) and chronic bleeds (54.8%) than for acute presentations. Overall sensitivity in outpatient settings was 72.2%. The authors note this is lower than figures reported in controlled validation studies, and attribute the gap in part to the more heterogeneous patient mix and imaging conditions in routine clinical practice.
Read the study
Intracranial Hemorrhage
2024
AJR Online
Prospective Evaluation of Artificial Intelligence Triage of Intracranial Hemorrhage on Noncontrast Head CT
University of Alabama at Birmingham (UAB) · Published in AJR Online
Key Findings
Radiologists using the AI showed no statistically significant difference in diagnostic accuracy compared to reading without it. Specificity was higher without the AI (99.8% vs. 99.3%), suggesting the AI added false positives without a measurable offsetting gain in sensitivity in this cohort.
Read the study
Pulmonary Embolism
2023
Radiology (RSNA)
Prospective Evaluation of AI Triage of Pulmonary Emboli on CT Pulmonary Angiograms
University of Alabama at Birmingham (UAB) · Published in Radiology
Key Findings
The study found no statistically significant improvement in radiologist accuracy, miss rate, or report turnaround time when the AI was used. The tool did reprioritize positive scans in the worklist, but this did not translate into a measurable difference in diagnostic outcomes in this prospective evaluation.
Read the study
Intracranial Hemorrhage
2026
PubMed
Head-to-Head Comparison of Two AI Triage Solutions for Detecting Intracranial Hemorrhage
Baylor / Texas Medical Center Network · Published on PubMed
Key Findings
The study documented false-negative rates of approximately 6% and nearly 100 false positives within the study cohort. Error analysis found that motion artifacts and scanner hardware variations were common contributing factors — sources of variability that are routine in multi-vendor hospital environments.
Read the study
C-Spine Fractures
2021
AJNR
Diagnostic Accuracy and Failure Mode Analysis of a Deep Learning Algorithm for Cervical Spine Fracture Detection
Lahey Hospital & Medical Center · Published in American Journal of Neuroradiology
Key Findings
Sensitivity was 54.9% with a Positive Predictive Value of 38.7%. Failure mode analysis found that degenerative disc changes were a common source of false positive classifications. The authors noted the algorithm had received "little or no external validation" prior to clinical use, and called for more rigorous independent validation before deployment of similar tools.
Read the study
Large Vessel Occlusion
2025
Stroke (AHA)
A Retrospective Analysis Comparing AIDoc and RAPIDAI in the Detection of Large Vessel Occlusions
Academic Medical Center · Published in Stroke: Vascular and Interventional Neurology
Key Findings
A 22% false negative rate for Aidoc in large vessel occlusion detection within this cohort — meaning roughly 1 in 5 cases were missed. The study concluded that neither major platform was superior and that both required extreme caution given the miss rates observed. The authors stopped short of recommending clinical reliance on either tool without further validation.
Read the study
Pediatric ICH
2026
PubMed
Performance Evaluation of a Commercial Deep Learning Software for Detecting ICH in a Pediatric Population
Independent Researchers · Published on PubMed
Key Findings
The study found elevated false positive rates in pediatric patients, driven by normal anatomical features in children — choroid plexus calcifications and hyperdense venous sinuses — that the algorithm, trained predominantly on adult data, classified as hemorrhage. The authors conclude that population-specific validation is necessary before applying adult-trained AI tools to pediatric imaging.
Read the study