Survey of explainable artificial intelligence techniques for biomedical imaging with Deep Neural Networks

Sajid Nazir*, Diane M. Dickson, Muhammad Usman Akram

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

127 Citations (Scopus)
1048 Downloads (Pure)

Abstract

Artificial Intelligence (AI) techniques of deep learning have revolutionized the disease diagnosis with their outstanding image classification performance. In spite of the outstanding results, the widespread adoption of these techniques in clinical practice is still taking place at a moderate pace. One of the major hindrance is that a trained Deep Neural Networks (DNN) model provides a prediction, but questions about why and how that prediction was made remain unanswered. This linkage is of utmost importance for the regulated healthcare domain to increase the trust in the automated diagnosis system by the practitioners, patients and other stakeholders. The application of deep learning for medical imaging has to be interpreted with caution due to the health and safety concerns similar to blame attribution in the case of an accident involving autonomous cars. The consequences of both a false positive and false negative cases are far reaching for patients' welfare and cannot be ignored. This is exacerbated by the fact that the state-of-the-art deep learning algorithms comprise of complex interconnected structures, millions of parameters, and a ‘black box’ nature, offering little understanding of their inner working unlike the traditional machine learning algorithms. Explainable AI (XAI) techniques help to understand model predictions which help develop trust in the system, accelerate the disease diagnosis, and meet adherence to regulatory requirements. This survey provides a comprehensive review of the promising field of XAI for biomedical imaging diagnostics. We also provide a categorization of the XAI techniques, discuss the open challenges, and provide future directions for XAI which would be of interest to clinicians, regulators and model developers.

Original languageEnglish
Article number106668
Number of pages29
JournalComputers in Biology and Medicine
Volume156
Early online date28 Feb 2023
DOIs
Publication statusPublished - Apr 2023

Keywords

  • Interpretable AI
  • Blackbox
  • Features
  • Supervised learning
  • Predictive models
  • Neural networks
  • Diagnostic imaging
  • Backpropagation

ASJC Scopus subject areas

  • Health Informatics
  • Computer Science Applications

Fingerprint

Dive into the research topics of 'Survey of explainable artificial intelligence techniques for biomedical imaging with Deep Neural Networks'. Together they form a unique fingerprint.

Cite this