Loading…

Ability of artificial intelligence to identify self-reported race in chest x-ray using pixel intensity counts

Prior studies show convolutional neural networks predicting self-reported race using x-rays of chest, hand and spine, chest computed tomography, and mammogram. We seek an understanding of the mechanism that reveals race within x-ray images, investigating the possibility that race is not predicted us...

Full description

Saved in:
Bibliographic Details
Published in:Journal of medical imaging (Bellingham, Wash.) Wash.), 2023-11, Vol.10 (6), p.061106
Main Authors: Burns, John Lee, Zaiman, Zachary, Vanschaik, Jack, Luo, Gaoxiang, Peng, Le, Price, Brandon, Mathias, Garric, Mittal, Vijay, Sagane, Akshay, Tignanelli, Christopher, Chakraborty, Sunandan, Gichoya, Judy Wawira, Purkayastha, Saptarshi
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Prior studies show convolutional neural networks predicting self-reported race using x-rays of chest, hand and spine, chest computed tomography, and mammogram. We seek an understanding of the mechanism that reveals race within x-ray images, investigating the possibility that race is not predicted using the physical structure in x-ray images but is embedded in the grayscale pixel intensities. Retrospective full year 2021, 298,827 AP/PA chest x-ray images from 3 academic health centers across the United States and MIMIC-CXR, labeled by self-reported race, were used in this study. The image structure is removed by summing the number of each grayscale value and scaling to percent per image (PPI). The resulting data are tested using multivariate analysis of variance (MANOVA) with Bonferroni multiple-comparison adjustment and class-balanced MANOVA. Machine learning (ML) feed-forward networks (FFN) and decision trees were built to predict race (binary Black or White and binary Black or other) using only grayscale value counts. Stratified analysis by body mass index, age, sex, gender, patient type, make/model of scanner, exposure, and kilovoltage peak setting was run to study the impact of these factors on race prediction following the same methodology. MANOVA rejects the null hypothesis that classes are the same with 95% confidence ( 7.38, ) and balanced MANOVA ( 2.02, ). The best FFN performance is limited [area under the receiver operating characteristic (AUROC) of 69.18%]. Gradient boosted trees predict self-reported race using grayscale PPI (AUROC 77.24%). Within chest x-rays, pixel intensity value counts alone are statistically significant indicators and enough for ML classification tasks of patient self-reported race.
ISSN:2329-4302
2329-4310
DOI:10.1117/1.JMI.10.6.061106