An Explainable Deep Learning Model for Classification and Analysis of Alzheimer's Disease for Clinical Trust

Authors

  • H. C. Bharath Bapuji Institute of Engineering and Technology, Davangere, Affiliated to Visvesvaraya Technological University, Belagavi-590018, India
  • N. Pradeep Bapuji Institute of Engineering and Technology, Davangere, Affiliated to Visvesvaraya Technological University, Belagavi-590018, India
  • R. Shashidhar Department of Electronics and Communication Engineering, JSS Science and Technology University, Mysuru-570006, India
  • Pavitha Nooji Department of Artificial Intelligence, Faculty of Science and Technology, Vishwakarma University, Pune-411048, India
  • J. Meghana Department of Computer Applications, JSS Science and Technology University, Mysuru-570006, India
Volume: 15 | Issue: 6 | Pages: 28762-28768 | December 2025 | https://doi.org/10.48084/etasr.12997

Abstract

Alzheimer’s Disease (AD), a progressive neurodegenerative disorder, presents significant challenges in early diagnosis and treatment. While deep learning models have demonstrated high accuracy in analyzing medical imaging data, their lack of interpretability limits clinical adoption. This study proposes an explainable deep learning framework that integrates a Convolutional Neural Network (CNN) with the Local Interpretable Model-agnostic Explanations (LIME) technique to enhance transparency and clinical trust in AD classification. The model highlights key biomarkers, such as hippocampal atrophy and amyloid plaque density, that contribute to its predictions. The approach addresses the need for consistent evaluation metrics and the integration of domain expertise in AI-driven diagnosis. Experiments conducted on the OASIS benchmark dataset achieved a classification accuracy of 97%, a precision of 97.2%, a recall of 96.8%, and an AUC of 0.97, with LIME providing localized, interpretable insights that align with clinical understanding. By combining predictive performance with explainability, the framework addresses ethical concerns by fostering transparency, accountability, trustworthiness, and practical AI-assisted diagnostic tools for precision healthcare.

Keywords:

Alzheimer's Disease (AD), deep learning, Explainable Artificial Intelligence (XAI), Local Interpretable Model-agnostic Explanations (LIME), decision support systems

Downloads

Download data is not yet available.

References

E. M. McDade, "Alzheimer Disease," Continuum, vol. 28, no. 3, pp. 648–675, Jun. 2022. DOI: https://doi.org/10.1212/CON.0000000000001131

S. Gauthier, C. Webster, S. Sarvaes, J. Morais, and P. Rosa-Neto, World Alzheimer Report 2022: Life After Diagnosis: Navigating Treatment, Care and Support. Lincolnshire, IL, USA: Alzheimer’s Disease International., 2022.

N. Shaffi, F. Hajamohideen, M. Mahmud, A. Abdesselam, K. Subramanian, and A. A. Sariri, "Triplet-Loss Based Siamese Convolutional Neural Network for 4-Way Classification of Alzheimer’s Disease," in Brain Informatics, vol. 13406, M. Mahmud, J. He, S. Vassanelli, A. Van Zundert, and N. Zhong, Eds. Cham: Springer International Publishing, 2022, pp. 277–287.

S. A. Tatulian, "Challenges and Hopes for Alzheimer’s disease," Drug Discovery Today, vol. 27, no. 4, pp. 1027–1043, Apr. 2022. DOI: https://doi.org/10.1016/j.drudis.2022.01.016

A. Rai, "Explainable AI: from Black Box to Glass Box," Journal of the Academy of Marketing Science, vol. 48, no. 1, pp. 137–141, Jan. 2020. DOI: https://doi.org/10.1007/s11747-019-00710-5

G. Yang, Q. Ye, and J. Xia, "Unbox the Black-box for the Medical Explainable AI via Multi-modal and Multi-centre Data Fusion: A Mini-review, Two Showcases and Beyond," Information Fusion, vol. 77, pp. 29–52, Jan. 2022. DOI: https://doi.org/10.1016/j.inffus.2021.07.016

V. Vimbi, N. Shaffi, and M. Mahmud, "Interpreting Artificial Intelligence Models: A Systematic Review on the Application of LIME and SHAP in Alzheimer’s Disease Detection," Brain Informatics, vol. 11, no. 1, Dec. 2024, Art. no. 10. DOI: https://doi.org/10.1186/s40708-024-00222-1

B. Dubois, G. Picard, and M. Sarazin, "Early Detection of Alzheimer’s Disease: New Diagnostic Criteria," Dialogues in Clinical Neuroscience, vol. 11, no. 2, pp. 135–139, Jun. 2009. DOI: https://doi.org/10.31887/DCNS.2009.11.2/bdubois

J. Wanner, L.-V. Herm, K. Heinrich, and C. Janiesch, "Stop Ordering Machine Learning Algorithms by Their Explainability! An Empirical Investigation of the Tradeoff Between Performance and Explainability," in Responsible AI and Analytics for an Ethical and Inclusive Digitized Society, vol. 12896, D. Dennehy, A. Griva, N. Pouloudi, Y. K. Dwivedi, I. Pappas, and M. Mäntymäki, Eds. Cham: Springer International Publishing, 2021, pp. 245–258. DOI: https://doi.org/10.1007/978-3-030-85447-8_22

B. K. Raghupathy, M. R. Reddy, Prasad Theeda, E. Balasubramanian, R. K. Namachivayam, and M. Ganesan, "Harnessing Explainable Artificial Intelligence (XAI) based SHAPLEY Values and Ensemble Techniques for Accurate Alzheimer’s Disease Diagnosis," Engineering, Technology & Applied Science Research, vol. 15, no. 2, pp. 20743–20747, Apr. 2025. DOI: https://doi.org/10.48084/etasr.9619

W. Feng et al., "Automated MRI-Based Deep Learning Model for Detection of Alzheimer’s Disease Process," International Journal of Neural Systems, vol. 30, no. 06, Jun. 2020, Art. no. 2050032. DOI: https://doi.org/10.1142/S012906572050032X

S. C. Emerald and T. Vengattaraman, "Explainable Artificial Intelligence with Single Layer Feedforward Neural Network and Improved Crowned Porcupine Optimization Algorithm for Classification Problems," Engineering, Technology & Applied Science Research, vol. 15, no. 2, pp. 21593–21598, Apr. 2025. DOI: https://doi.org/10.48084/etasr.10070

M. T. Ribeiro, S. Singh, and C. Guestrin, "‘Why Should I Trust You?’: Explaining the Predictions of Any Classifier," in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco California USA, Aug. 2016, pp. 1135–1144. DOI: https://doi.org/10.1145/2939672.2939778

M. Liu, D. Cheng, K. Wang, and Y. Wang, "Multi-Modality Cascaded Convolutional Neural Networks for Alzheimer’s Disease Diagnosis," Neuroinformatics, vol. 16, no. 3–4, pp. 295–308, Oct. 2018. DOI: https://doi.org/10.1007/s12021-018-9370-4

A. S. Alatrany, W. Khan, A. Hussain, H. Kolivand, and D. Al-Jumeily, "An Explainable Machine Learning Approach for Alzheimer’s Disease Classification," Scientific Reports, vol. 14, no. 1, Feb. 2024, Art. no. 2637. DOI: https://doi.org/10.1038/s41598-024-51985-w

N. Amoroso, S. Quarto, M. La Rocca, S. Tangaro, A. Monaco, and R. Bellotti, "An eXplainability Artificial Intelligence Approach to Brain Connectivity in Alzheimer’s Disease," Frontiers in Aging Neuroscience, vol. 15, Aug. 2023, Art. no. 1238065. DOI: https://doi.org/10.3389/fnagi.2023.1238065

A. Mehmood et al., "A Transfer Learning Approach for Early Diagnosis of Alzheimer’s Disease on MRI Images," Neuroscience, vol. 460, pp. 43–52, Apr. 2021. DOI: https://doi.org/10.1016/j.neuroscience.2021.01.002

Z. Fang et al., "GFE-Mamba: Mamba-based AD Multi-modal Progression Assessment via Generative Feature Extraction from MCI." arXiv, Jan. 29, 2025.

E. Nigri, N. Ziviani, F. Cappabianco, A. Antunes, and A. Veloso, "Explainable Deep CNNs for MRI-Based Diagnosis of Alzheimer’s Disease," in 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, United Kingdom, Jul. 2020, pp. 1–8. DOI: https://doi.org/10.1109/IJCNN48605.2020.9206837

D. S. Marcus, T. H. Wang, J. Parker, J. G. Csernansky, J. C. Morris, and R. L. Buckner, "Open Access Series of Imaging Studies (OASIS): Cross-sectional MRI Data in Young, Middle Aged, Nondemented, and Demented Older Adults," Journal of Cognitive Neuroscience, vol. 19, no. 9, pp. 1498–1507, Sep. 2007. DOI: https://doi.org/10.1162/jocn.2007.19.9.1498

"Alzheimer’s Disease Association," Understanding Alzheimer’s Disease and Dementia, 2024. https://www.alz.org/.

T. Mahmud, K. Barua, S. U. Habiba, N. Sharmen, M. S. Hossain, and K. Andersson, "An Explainable AI Paradigm for Alzheimer’s Diagnosis Using Deep Transfer Learning," Diagnostics, vol. 14, no. 3, Feb. 2024, Art. no. 345. DOI: https://doi.org/10.3390/diagnostics14030345

Downloads

How to Cite

[1]
H. C. Bharath, N. Pradeep, R. Shashidhar, P. Nooji, and J. Meghana, “An Explainable Deep Learning Model for Classification and Analysis of Alzheimer’s Disease for Clinical Trust”, Eng. Technol. Appl. Sci. Res., vol. 15, no. 6, pp. 28762–28768, Dec. 2025.

Metrics

Abstract Views: 386
PDF Downloads: 279

Metrics Information