Ability of artificial intelligence to identify self-reported race in chest x-ray using pixel intensity counts

John Lee Burns, Zachary Zaiman, Jack Vanschaik, Gaoxiang Luo, Le Peng, Brandon Price, Garric Mathias, Vijay Mittal, Akshay Sagane, Christopher Tignanelli, Sunandan Chakraborty, Judy Wawira Gichoya, Saptarshi Purkayastha

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Purpose: Prior studies show convolutional neural networks predicting self-reported race using x-rays of chest, hand and spine, chest computed tomography, and mammogram. We seek an understanding of the mechanism that reveals race within x-ray images, investigating the possibility that race is not predicted using the physical structure in x-ray images but is embedded in the grayscale pixel intensities. Approach: Retrospective full year 2021, 298,827 AP/PA chest x-ray images from 3 academic health centers across the United States and MIMIC-CXR, labeled by self-reported race, were used in this study. The image structure is removed by summing the number of each grayscale value and scaling to percent per image (PPI). The resulting data are tested using multivariate analysis of variance (MANOVA) with Bonferroni multiple-comparison adjustment and class-balanced MANOVA. Machine learning (ML) feed-forward networks (FFN) and decision trees were built to predict race (binary Black or White and binary Black or other) using only grayscale value counts. Stratified analysis by body mass index, age, sex, gender, patient type, make/ model of scanner, exposure, and kilovoltage peak setting was run to study the impact of these factors on race prediction following the same methodology. Results: MANOVA rejects the null hypothesis that classes are the same with 95% confidence (F 7.38, P < 0.0001) and balanced MANOVA (F 2.02, P < 0.0001). The best FFN performance is limited [area under the receiver operating characteristic (AUROC) of 69.18%]. Gradient boosted trees predict self-reported race using grayscale PPI (AUROC 77.24%). Conclusions: Within chest x-rays, pixel intensity value counts alone are statistically significant indicators and enough for ML classification tasks of patient self-reported race.

Original languageEnglish (US)
Article number061106
JournalJournal of Medical Imaging
Volume10
Issue number6
DOIs
StatePublished - Nov 1 2023

Bibliographical note

Publisher Copyright:
© 2023 Society of Photo-Optical Instrumentation Engineers (SPIE).

Keywords

  • bias
  • machine learning
  • population imaging
  • x-ray

Fingerprint

Dive into the research topics of 'Ability of artificial intelligence to identify self-reported race in chest x-ray using pixel intensity counts'. Together they form a unique fingerprint.

Cite this