Artificial intelligence (AI) technology in medical imaging has the potential to revolutionize healthcare but also carries potential pitfalls. Researchers from the University of Maryland School of Medicine found that most AI algorithms used in medical imaging did not include patient demographics, and the algorithms did not evaluate inherent biases. This lack of demographic information and biased evaluation could exacerbate existing health disparities. The researchers encourage future competitions to require fairness among different groups of people. It is essential to identify and address potential biases that may arise in the use of AI models in medical imaging.
Artificial intelligence (AI) technology has the potential to revolutionize the medical field. From automating diagnoses to decreasing physician workload and bringing specialized healthcare to remote or underdeveloped areas, the possibilities are endless. However, with these possibilities come potential pitfalls, as highlighted by recent research from the University of Maryland School of Medicine (UMSOM).
Analyzing crowd-sourced sets of data used to create AI algorithms from medical images, UMSOM researchers found that most did not include patient demographics, such as age, sex, or race. The algorithms also did not evaluate for inherent biases. This lack of demographic information and bias evaluation means that these algorithms have no way of knowing whether the images contain representative samples of the population, including underrepresented groups such as Blacks, Asians, and Indigenous Americans.
The researchers warn that much of medicine in the United States is already fraught with partiality toward certain races, genders, ages, or sexual orientations. Small biases in individual sets of data could be amplified greatly when hundreds or thousands of these datasets are combined in these algorithms. As a result, there is a risk that AI in medical imaging could exacerbate existing health disparities.
“These deep learning models can diagnose things physicians can’t see, such as when a person might die or detect Alzheimer’s disease seven years earlier than our known tests — superhuman tasks,” said senior investigator Paul Yi, MD, Assistant Professor of Diagnostic Radiology and Nuclear Medicine at UMSOM. He is also the Director of the University of Maryland Medical Intelligent Imaging (UM2ii) Center. “Because these AI machine learning techniques are so good at finding needles in a haystack, they can also define sex, gender, and age, which means these models can then use those features to make biased decisions.”
Much of the data collected in large studies tends to be from people of means who have relatively easy access to healthcare. In the U.S., this means the data tends to be skewed toward men versus women and toward people who are white rather than other races. As the U.S. tends to perform more imaging than the rest of the world, this data gets compiled into algorithms that have the potential to slant outcomes worldwide.
The researchers evaluated the datasets used in data science competitions in which computer scientists and physicians crowdsource data from around the world and try to develop the best, most accurate algorithm. These competitions tend to have leaderboards that rank each algorithm and provide a cash prize, motivating people to create the best one. Specifically, the researchers investigated medical imaging algorithms, such as those that evaluate CT scans to diagnose brain tumors or blood clots in the lungs. Of the 23 data competitions analyzed, 61 percent did not include demographic data such as age, sex, or race. None of the competitions had evaluations for biases against underrepresented or disadvantaged groups.
“We hope that by bringing awareness to this issue in these data competitions – and if applied appropriately – that there is tremendous potential to solve these biases,” said lead author Sean Garin, Program Coordinator at the UM2ii Center. The study’s authors also encourage future competitions to require not only high accuracy but also fairness among different groups of people.
“As AI models become more prevalent in medical imaging and other fields of medicine, it is important to identify and address potential biases that may exacerbate existing health inequities in clinical care – an essential priority for every academic medical institution,” said UMSOM Dean Mark T. Gladwin, MD, Vice President for Medical Affairs, University of Maryland, Baltimore, and the John Z. and Akiko K. Bowers Distinguished Professor.