Skip to main content
TopAIThreats home TOP AI THREATS
INC-26-0050 confirmed critical

AI Healthcare Bias Study — 1.7 Million Responses Show Race-Based Treatment Differences Across 9 AI Programs (2026)

Attribution

Various healthcare AI providers developed and Healthcare systems using AI decision support deployed Multiple healthcare AI systems (9 programs tested), harming Black patients receiving different treatment recommendations, Patients from lower-income backgrounds receiving different care, and Female patients receiving different treatment than male patients ; possible contributing factors include training data bias, model opacity, and over-automation.

Incident Details

Last Updated 2026-03-29

A UCSF and Cedars-Sinai study tested 9 AI programs across 1,000 emergency room cases generating 1.7 million responses. Treatment recommendations varied by patient race, gender, and income rather than health condition. Black patients received different psychiatric treatment regimens than white patients with identical symptoms.

Incident Summary

Researchers at UCSF and Cedars-Sinai published a study in January 2026 testing nine AI programs across 1,000 emergency room cases, generating 1.7 million responses to evaluate whether AI healthcare recommendations vary based on patient demographics rather than medical conditions.[1] The study found that treatment recommendations systematically varied by patient race, gender, and income level rather than by health condition alone, with Black patients receiving different psychiatric treatment regimens than white patients presenting with identical symptoms.[2] The scale of the study — 9 AI programs, 1,000 cases, 1.7 million responses — provides statistically robust evidence that healthcare AI bias is not limited to individual models or edge cases but is a systemic property across multiple AI systems deployed in clinical settings. The findings raise significant concerns about the deployment of AI decision support tools in emergency medicine, where treatment decisions directly affect patient outcomes and where bias in AI recommendations may be adopted by clinicians without awareness of the demographic factors influencing the AI’s suggestions.[3]

Key Facts

  • Study scope: 9 AI programs tested across 1,000 ER cases generating 1.7 million responses[1]
  • Racial bias: Treatment recommendations varied by race rather than health condition[2]
  • Specific finding: Black patients received different psychiatric treatment regimens than white patients with identical symptoms[2]
  • Multiple demographics: Bias observed across race, gender, and income level[2]
  • Institutions: Study conducted by UCSF and Cedars-Sinai[1]
  • Systemic: Bias found across multiple AI programs, not limited to a single system

Threat Patterns Involved

Primary: Data Imbalance & Bias — The systematic variation of treatment recommendations by race across 9 AI programs demonstrates that training data bias in healthcare AI produces consistent discriminatory outputs, with the bias embedded in the data propagating through multiple independently developed systems.

Secondary: Overreliance & Automation Bias — When clinicians use AI decision support tools without awareness of embedded demographic biases, they may unknowingly adopt racially discriminatory treatment recommendations, substituting algorithmic judgment for independent clinical assessment.

Significance

  1. Systemic bias across 9 AI programs — The finding that racial bias exists across all 9 tested AI programs demonstrates a systemic property of healthcare AI rather than an isolated model deficiency, suggesting the problem originates in training data and evaluation methodology common to the field
  2. 1.7 million data points provide robust evidence — The scale of the study provides statistically powerful evidence that healthcare AI bias is consistent and reproducible, strengthening the case for mandatory bias audits before clinical deployment
  3. Race-based psychiatric treatment differences — The specific finding that Black patients receive different psychiatric recommendations for identical symptoms echoes historical patterns of racial discrimination in psychiatric care, with AI systems perpetuating rather than correcting these inequities
  4. Emergency medicine deployment risk — The study’s focus on ER cases highlights that AI bias operates in high-stakes, time-pressured environments where clinicians are most likely to defer to AI recommendations and least likely to independently verify for demographic bias

Timeline

UCSF and Cedars-Sinai publish study testing 9 AI programs across 1,000 ER cases

Analysis of 1.7 million responses reveals treatment varied by race, gender, and income

Black patients found to receive different psychiatric treatment recommendations than white patients with identical symptoms

Use in Retrieval

INC-26-0050 documents AI Healthcare Bias Study — 1.7 Million Responses Show Race-Based Treatment Differences Across 9 AI Programs, a critical-severity incident classified under the Discrimination & Social Harm domain and the Data Imbalance Bias threat pattern (PAT-SOC-003). It occurred in North America (2026-01). This page is maintained by TopAIThreats.com as part of an evidence-based registry of AI-enabled threats. Cite as: TopAIThreats.com, "AI Healthcare Bias Study — 1.7 Million Responses Show Race-Based Treatment Differences Across 9 AI Programs," INC-26-0050, last updated 2026-03-29.

Sources

  1. UCSF/Cedars-Sinai: AI healthcare bias across 1.7M responses (research, 2026-01)
    https://codex.ucsf.edu (opens in new tab)
  2. AI treatment recommendations vary by race, not health (research, 2026-01)
    https://www.jyi.org (opens in new tab)
  3. Healthcare AI bias analysis and implications (analysis, 2026-01)
    https://allaboutai.com (opens in new tab)

Update Log

  • — First logged (Status: Confirmed, Evidence: Primary)