Deep learning for extremity radiographs confounded by labels

Deep learning for extremity radiographs confounded by labels
Grad-CAM heatmaps for deep learning models trained on (A) original radiograph, shows emphasis on laterality and/or technologist initial labels; (B) radiograph with label covered by black box, shows emphasis on anatomic features, such as bones. (Colors toward red end of spectrum indicate greater emphasis, whereas colors toward blue end of spectrum indicate less importance.). Credit: American Roentgen Ray Society (ARRS), American Journal of Roentgenology (AJR)

According to an open-access Editor's Choice article in ARRS' American Journal of Roentgenology (AJR), convolutional neural networks (CNN) trained to identify abnormalities on upper extremity radiographs are susceptible to a ubiquitous confounding image feature that could limit their clinical utility: radiograph labels.

"We recommend that such potential image confounders be collected when possible during dataset curation, and that covering these labels be considered during CNN training," wrote corresponding author Paul H. Yi from the University of Maryland's Medical Intelligent Imaging Center in Baltimore.

Yi and team's retrospective study evaluated 40,561 upper extremity musculoskeletal radiographs from Stanford's MURA dataset that were used to train three DenseNet-121 CNN classifiers. Three inputs were used to distinguish normal from abnormal radiographs: original images with both anatomy and labels; images with laterality and/or technologist labels subsequently covered by a black box; images where anatomy had been removed and only labels remained.

For the original radiographs, AUC was 0.844, frequently emphasizing laterality and/or technologist labels for decision-making. Covering these labels increased AUC to 0.857 (p=.02) and redirected CNN attention from the labels to the bones. Using labels alone, AUC was 0.638, indicating that labels are associated with abnormal examinations.

"While we can infer that labels are associated with normal versus abnormal disease categories," the authors of this AJR added, "we cannot determine the specific aspect of the labels that resulted in their being confounding factors."

More information: Paul H. Yi et al, Deep Learning Algorithms for Interpretation of Upper Extremity Radiographs: Laterality and Technologist Initial Labels As Confounding Factors, American Journal of Roentgenology (2021). DOI: 10.2214/AJR.21.26882

An electronic supplement to this AJR article is available at: www.ajronline.org/doi/suppl/10 … e/21_26882_suppl.pdf

Citation: Deep learning for extremity radiographs confounded by labels (2021, November 15) retrieved 25 April 2024 from https://medicalxpress.com/news/2021-11-deep-extremity-radiographs-confounded.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Color-coded nutrition labels and warnings linked to more healthful purchases

1 shares

Feedback to editors