Fusion of electronic health records and radiographic images for a multimodal deep learning prediction model of atypical femur fractures.

Atypical femoral fractures Deep learning Fusion Multimodal

Journal

Computers in biology and medicine
ISSN: 1879-0534
Titre abrégé: Comput Biol Med
Pays: United States
ID NLM: 1250250

Informations de publication

Date de publication:
13 Nov 2023
Historique:
received: 18 07 2023
revised: 15 10 2023
accepted: 07 11 2023
medline: 20 11 2023
pubmed: 20 11 2023
entrez: 19 11 2023
Statut: aheadofprint

Résumé

Atypical femur fractures (AFF) represent a very rare type of fracture that can be difficult to discriminate radiologically from normal femur fractures (NFF). AFFs are associated with drugs that are administered to prevent osteoporosis-related fragility fractures, which are highly prevalent in the elderly population. Given that these fractures are rare and the radiologic changes are subtle currently only 7% of AFFs are correctly identified, which hinders adequate treatment for most patients with AFF. Deep learning models could be trained to classify automatically a fracture as AFF or NFF, thereby assisting radiologists in detecting these rare fractures. Historically, for this classification task, only imaging data have been used, using convolutional neural networks (CNN) or vision transformers applied to radiographs. However, to mimic situations in which all available data are used to arrive at a diagnosis, we adopted an approach of deep learning that is based on the integration of image data and tabular data (from electronic health records) for 159 patients with AFF and 914 patients with NFF. We hypothesized that the combinatorial data, compiled from all the radiology departments of 72 hospitals in Sweden and the Swedish National Patient Register, would improve classification accuracy, as compared to using only one modality. At the patient level, the area under the ROC curve (AUC) increased from 0.966 to 0.987 when using the integrated set of imaging data and seven pre-selected variables, as compared to only using imaging data. More importantly, the sensitivity increased from 0.796 to 0.903. We found a greater impact of data fusion when only a randomly selected subset of available images was used to make the image and tabular data more balanced for each patient. The AUC then increased from 0.949 to 0.984, and the sensitivity increased from 0.727 to 0.849. These AUC improvements are not large, mainly because of the already excellent performance of the CNN (AUC of 0.966) when only images are used. However, the improvement is clinically highly relevant considering the importance of accuracy in medical diagnostics. We expect an even greater effect when imaging data from a clinical workflow, comprising a more diverse set of diagnostic images, are used.

Identifiants

pubmed: 37980797
pii: S0010-4825(23)01169-1
doi: 10.1016/j.compbiomed.2023.107704
pii:
doi:

Types de publication

Journal Article

Langues

eng

Sous-ensembles de citation

IM

Pagination

107704

Informations de copyright

Copyright © 2023. Published by Elsevier Ltd.

Déclaration de conflit d'intérêts

Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Auteurs

Jörg Schilcher (J)

Department of Orthopedics and Experimental and Clinical Medicine, Faculty of Health Science, Linköping University, Linköping, Sweden; Wallenberg Centre for Molecular Medicine, Linköping University, Linköping, Sweden; Center for Medical Image Science and Visualization (CMIV), Linköping University, Linköping, Sweden.

Alva Nilsson (A)

Department of Biomedical Engineering, Linköping University, Linköping, Sweden.

Oliver Andlid (O)

Department of Biomedical Engineering, Linköping University, Linköping, Sweden.

Anders Eklund (A)

Department of Biomedical Engineering, Linköping University, Linköping, Sweden; Division of Statistics and Machine Learning, Department of Computer and Information Science, Linköping University, Linköping, Sweden; Center for Medical Image Science and Visualization (CMIV), Linköping University, Linköping, Sweden. Electronic address: anders.eklund@liu.se.

Classifications MeSH