TY - JOUR
T1 - Ability of artificial intelligence to identify self-reported race in chest x-ray using pixel intensity counts
AU - Burns, John Lee
AU - Zaiman, Zachary
AU - Vanschaik, Jack
AU - Luo, Gaoxiang
AU - Peng, Le
AU - Price, Brandon
AU - Mathias, Garric
AU - Mittal, Vijay
AU - Sagane, Akshay
AU - Tignanelli, Christopher
AU - Chakraborty, Sunandan
AU - Gichoya, Judy Wawira
AU - Purkayastha, Saptarshi
N1 - Publisher Copyright:
© 2023 Society of Photo-Optical Instrumentation Engineers (SPIE).
PY - 2023/11/1
Y1 - 2023/11/1
N2 - Purpose: Prior studies show convolutional neural networks predicting self-reported race using x-rays of chest, hand and spine, chest computed tomography, and mammogram. We seek an understanding of the mechanism that reveals race within x-ray images, investigating the possibility that race is not predicted using the physical structure in x-ray images but is embedded in the grayscale pixel intensities. Approach: Retrospective full year 2021, 298,827 AP/PA chest x-ray images from 3 academic health centers across the United States and MIMIC-CXR, labeled by self-reported race, were used in this study. The image structure is removed by summing the number of each grayscale value and scaling to percent per image (PPI). The resulting data are tested using multivariate analysis of variance (MANOVA) with Bonferroni multiple-comparison adjustment and class-balanced MANOVA. Machine learning (ML) feed-forward networks (FFN) and decision trees were built to predict race (binary Black or White and binary Black or other) using only grayscale value counts. Stratified analysis by body mass index, age, sex, gender, patient type, make/ model of scanner, exposure, and kilovoltage peak setting was run to study the impact of these factors on race prediction following the same methodology. Results: MANOVA rejects the null hypothesis that classes are the same with 95% confidence (F 7.38, P < 0.0001) and balanced MANOVA (F 2.02, P < 0.0001). The best FFN performance is limited [area under the receiver operating characteristic (AUROC) of 69.18%]. Gradient boosted trees predict self-reported race using grayscale PPI (AUROC 77.24%). Conclusions: Within chest x-rays, pixel intensity value counts alone are statistically significant indicators and enough for ML classification tasks of patient self-reported race.
AB - Purpose: Prior studies show convolutional neural networks predicting self-reported race using x-rays of chest, hand and spine, chest computed tomography, and mammogram. We seek an understanding of the mechanism that reveals race within x-ray images, investigating the possibility that race is not predicted using the physical structure in x-ray images but is embedded in the grayscale pixel intensities. Approach: Retrospective full year 2021, 298,827 AP/PA chest x-ray images from 3 academic health centers across the United States and MIMIC-CXR, labeled by self-reported race, were used in this study. The image structure is removed by summing the number of each grayscale value and scaling to percent per image (PPI). The resulting data are tested using multivariate analysis of variance (MANOVA) with Bonferroni multiple-comparison adjustment and class-balanced MANOVA. Machine learning (ML) feed-forward networks (FFN) and decision trees were built to predict race (binary Black or White and binary Black or other) using only grayscale value counts. Stratified analysis by body mass index, age, sex, gender, patient type, make/ model of scanner, exposure, and kilovoltage peak setting was run to study the impact of these factors on race prediction following the same methodology. Results: MANOVA rejects the null hypothesis that classes are the same with 95% confidence (F 7.38, P < 0.0001) and balanced MANOVA (F 2.02, P < 0.0001). The best FFN performance is limited [area under the receiver operating characteristic (AUROC) of 69.18%]. Gradient boosted trees predict self-reported race using grayscale PPI (AUROC 77.24%). Conclusions: Within chest x-rays, pixel intensity value counts alone are statistically significant indicators and enough for ML classification tasks of patient self-reported race.
KW - bias
KW - machine learning
KW - population imaging
KW - x-ray
UR - http://www.scopus.com/inward/record.url?scp=85177733083&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85177733083&partnerID=8YFLogxK
U2 - 10.1117/1.JMI.10.6.061106
DO - 10.1117/1.JMI.10.6.061106
M3 - Article
C2 - 37545750
AN - SCOPUS:85177733083
SN - 2329-4302
VL - 10
JO - Journal of Medical Imaging
JF - Journal of Medical Imaging
IS - 6
M1 - 061106
ER -