Abstract
Early detection of brain tumors is vital for improving patient survival rates, yet the manual analysis of the extensive 3D MRI images can be error-prone and time-consuming. This study introduces the Deep Explainable Brain Tumor Deep Network (DeepEBTDNet), a novel deep learning model for binary classification of brain MRIs as tumorous or normal. Employing sub-image dualistic histogram equalization (DSIHE) for enhanced image quality, DeepEBTDNet utilizes 12 convolutional layers with leaky ReLU (LReLU) activation for feature extraction, followed by a fully connected classification layer. Transparency and interpretability are emphasized through the application of the Local Interpretable Model-Agnostic Explanations (LIME) method to explain model predictions. Results demonstrate DeepEBTDNet's efficacy in brain tumor detection, even across datasets, achieving a validation accuracy of 98.96% and testing accuracy of 94.0%. This study underscores the importance of explainable AI in healthcare, facilitating precise diagnoses and transparent decision-making for early brain tumor identification and improved patient outcomes.
| Original language | English |
|---|---|
| Article number | e23012 |
| Journal | International Journal of Imaging Systems and Technology |
| Volume | 34 |
| Issue number | 1 |
| DOIs | |
| State | Published - Jan 2024 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© 2023 Wiley Periodicals LLC.
Keywords
- LIME
- MRI
- brain-tumor detection
- deep learning
- explainable AI
ASJC Scopus subject areas
- Biomedical Engineering
- Radiology Nuclear Medicine and imaging
- Computer Vision and Pattern Recognition
- Computer Science Applications
- Health Informatics