Racial differences in medical testing could introduce bias to AI models

But tweaking the models could help overcome biased data sets

5:00 AM

Author | Derek Smith

floating AI-type images in red and blues and yellow on blue background
Justine Ross, Michigan Medicine

Black patients are less likely than white patients to receive certain medical tests that doctors use to diagnose severe disease such as sepsis, researchers at the University of Michigan have shown.

Because of the differences in testing rates, some sick Black patients are assumed to be healthy in data used to train AI, and the resulting models likely underestimate illness in Black patients. But that doesn’t mean the data is unusable—the same group developed a way to correct for this bias in data sets used to train AI.

These new insights are reported in a pair of studies: one published today in PLOS Global Public Health, and the other was presented at the International Conference on Machine Learning in Vienna, Austria, in July 2024.

In the PLOS study, the researchers found that medical testing rates for white patients are up to 4.5% higher than for Black patients with the same age, sex, medical complaints and emergency department triage score, a measure of the urgency of a patient’s medical needs. 

The difference is partially explained by hospital admission rates, as white patients were more likely to be assessed as ill and admitted to the hospital than Black patients.

“If there are subgroups of patients who are systematically undertested, then you are baking this bias into your model,” said Jenna Wiens, U-M associate professor of computer science and engineering and corresponding author of the study.

“Adjusting for such confounding factors is a standard statistical technique, but it’s typically not done prior to training AI models. When training AI, it’s really important to acknowledge flaws in the available data and think about their downstream implications.”

The researchers found this bias in medical testing records from two locations: Michigan Medicine in Ann Arbor, Michigan, and one of the most widely used clinical datasets for training AI, the Medical Information Mart for Intensive Care

The dataset contains the records of patients visiting the emergency room in the Beth Israel Deaconess Medical Center in Boston.

“This research highlights the risks of using health data to train AI models without a comprehensive understanding of the data,” said Michael Sjoding, M.D., associate professor of pulmonary and critical care medicine at Michigan Medicine.

“Because of these apparent testing difference, an AI model might infer that black patients are less sick than white patients and make predictions that are potentially biased.”

Computer scientists need to account for these biases so that AI can make accurate and equitable predictions of patient illness. 

One option is to train the AI model with a less biased dataset, such as one that only includes records for patients that have received diagnostic medical tests. 

A model trained on such data might be inaccurate for less ill patients, however.

To correct the bias without omitting patient records, the researchers developed a computer algorithm that identifies whether untested patients were likely ill based on their race and vital signs, such as blood pressure. 

The algorithm accounts for race because the recorded health statuses of patients identified as Black are more likely to be affected by the testing bias.

The researchers tested the algorithm with simulated data, in which they introduced a known bias by relabeling patients identified as ill as “untested and healthy.” 

The researchers then used this dataset to train a machine learning model, the results of which were presented at the International Conference on Machine Learning. 

When the researcher-imposed bias was corrected with the algorithm, a textbook machine-learning model could accurately differentiate between patients with and without sepsis around 60% of the time. 

Without the algorithm, the biased data made the model’s performance worse than random.

The improved accuracy was on par with a textbook model that was trained on unbiased, simulated data in which everyone was equitably tested. 

Such unbiased datasets are unlikely to exist in the real world, but the researcher’s approach allowed the AI to work about as accurately as the idealized scenario despite being stuck with biased data.

“Approaches that account for systematic bias in data are an important step towards correcting some inequities in healthcare delivery, especially as more clinics turn toward AI-based solutions,” said Trenton Chang, a doctoral student in computer science and engineering and the first author of both studies.

Additional authors: Mark Nuppnau, Ying He, Keith E. Kocher and Thomas S. Valley.

Funding/disclosures: This work was supported by the National Heart, Lung, and Blood Institute NHLBI R01 HL158626.

Paper cited: "Racial differences in laboratory testing as a potential mechanism for bias in AI: A matched cohort analysis in emergency department visits," PLOS Glob Public Health. DOI: 10.1371/journal.pgph.0003555

Sign up for Health Lab newsletters todayGet medical tips from top experts and learn about new scientific discoveries every week.

Sign up for the Health Lab PodcastAdd us wherever you listen to your favorite shows. 


More Articles About: Demographics Emergency & Trauma Care Emerging Technologies Health Care Delivery, Policy and Economics Lab Tests
Health Lab word mark overlaying blue cells
Health Lab

Explore a variety of health care news & stories by visiting the Health Lab home page for more articles.

Media Contact Public Relations

Department of Communication at Michigan Medicine

[email protected]

734-764-2220

In This Story
SJODING_Michael4x5.jpg Michael William Sjoding, MD, MSc

Associate Professor

Stay Informed

Want top health & research news weekly? Sign up for Health Lab’s newsletters today!

Subscribe
Featured News & Stories white sneaker feet in dark blue and light blue scrubs rushing a hospital bed into medical area
Health Lab
Firearm hospitalizations dropped before the pandemic, then shot up
Hospital stays related to gun injuries dipped then rose sharply, and paralleled gun sales trends, with disparities by age, race and health insurance status.
paperwork with white and blue and red and says medicare john smith hospital part a medical part b 09-01 09-0
Health Lab
How did health insurance coverage changes affect older adults?
Two University of Michigan studies show how past policy decisions have affected older Americans with modest or low incomes.
woman walking on treadmill picking intensity on a chart that reads from easier to harder
Health Lab
Higher costs limit attendance for life changing cardiac rehab
Despite the success cardiac rehabilitation has shown at reducing heart-related deaths and hospital readmissions, higher out-of-pocket costs may prevent patients from participating in the program, a University of Michigan study suggests. 
person talking to older couple on couch in living room
Health Lab
85% of Mexican Americans with dementia unaware of diagnosis, outpacing overall rate
More than three-quarters of older adults with dementia may be unaware of their diagnosis, a University of Michigan study finds. 
hospital beds in hallway
Health Lab
Using data to drive sepsis care
Michigan Medicine expert, Hallie Prescott, M.D., discusses successful statewide efforts to improve sepsis treatment–and setting the bar for change at the national level
drawing of doctor with question mark about head with patient questioning and stressed over paperwork in exam room
Health Lab
People find medical test results hard to understand, increasing overall worry
In a published research letter in JAMA, researchers tested whether people could understand standard pathology reports and whether a patient-centered report might improve understanding.