A scoping review of interpretability and explainability concerning artificial intelligence methods in medical imaging.

Détails

Ressource 1Télécharger: 37976760.pdf (2319.01 [Ko])
Etat: Public
Version: Final published version
Licence: CC BY 4.0
ID Serval
serval:BIB_142C291FAF8F
Type
Article: article d'un périodique ou d'un magazine.
Collection
Publications
Institution
Titre
A scoping review of interpretability and explainability concerning artificial intelligence methods in medical imaging.
Périodique
European journal of radiology
Auteur⸱e⸱s
Champendal M., Müller H., Prior J.O., Dos Reis C.S.
ISSN
1872-7727 (Electronic)
ISSN-L
0720-048X
Statut éditorial
Publié
Date de publication
12/2023
Peer-reviewed
Oui
Volume
169
Pages
111159
Langue
anglais
Notes
Publication types: Journal Article ; Review
Publication Status: ppublish
Résumé
To review eXplainable Artificial Intelligence/(XAI) methods available for medical imaging/(MI).
A scoping review was conducted following the Joanna Briggs Institute's methodology. The search was performed on Pubmed, Embase, Cinhal, Web of Science, BioRxiv, MedRxiv, and Google Scholar. Studies published in French and English after 2017 were included. Keyword combinations and descriptors related to explainability, and MI modalities were employed. Two independent reviewers screened abstracts, titles and full text, resolving differences through discussion.
228 studies met the criteria. XAI publications are increasing, targeting MRI (n = 73), radiography (n = 47), CT (n = 46). Lung (n = 82) and brain (n = 74) pathologies, Covid-19 (n = 48), Alzheimer's disease (n = 25), brain tumors (n = 15) are the main pathologies explained. Explanations are presented visually (n = 186), numerically (n = 67), rule-based (n = 11), textually (n = 11), and example-based (n = 6). Commonly explained tasks include classification (n = 89), prediction (n = 47), diagnosis (n = 39), detection (n = 29), segmentation (n = 13), and image quality improvement (n = 6). The most frequently provided explanations were local (78.1 %), 5.7 % were global, and 16.2 % combined both local and global approaches. Post-hoc approaches were predominantly employed. The used terminology varied, sometimes indistinctively using explainable (n = 207), interpretable (n = 187), understandable (n = 112), transparent (n = 61), reliable (n = 31), and intelligible (n = 3).
The number of XAI publications in medical imaging is increasing, primarily focusing on applying XAI techniques to MRI, CT, and radiography for classifying and predicting lung and brain pathologies. Visual and numerical output formats are predominantly used. Terminology standardisation remains a challenge, as terms like "explainable" and "interpretable" are sometimes being used indistinctively. Future XAI development should consider user needs and perspectives.
Mots-clé
Humans, Artificial Intelligence, Radiography, Alzheimer Disease, Brain/diagnostic imaging, Brain Neoplasms, Computed Tomography, Deep Learning, Explainability, Machine Learning, Magnetic Resonance Imaging (MRI), Medical Imaging, Transparency
Pubmed
Web of science
Open Access
Oui
Création de la notice
23/11/2023 16:43
Dernière modification de la notice
15/02/2024 8:15
Données d'usage