Towards Explainable Deep Learning in Computational Neuroscience: Visual and Clinical Applications
Abstract
1. Introduction
- Brain signal decoding: The DL models like RNNs and transformers have shown significant performance in decoding signals, including cognitive states, motor intentions, and sensory inputs [18], e.g., EEG, MEG, ECoG, and fMRI.
- Image segmentation: The CNNs have always played an important role in segmenting high-dimensional images that depict the neural structures. This facilitates tasks used for connectomics, i.e., mapping complicated networks of synapses and neurons [22]. This also facilitates brain atlasing, i.e., understanding the structure and function of the brain [23].
- Disease classification & biomarker discovery: DL techniques have enhanced the diagnosis of neurological and psychiatric conditions by analyzing neuroimaging and electrophysiological data [24].
- Neural encoding & predictive modeling: Models like CNNs and transformers are being utilized to predict the brain responses to the expected environmental stimuli. Interpreting the behavior of response and stimuli through the use of neural encoding and predictions helps in understanding and advancing the sensory processing that mimics the human brain.
- Generative modeling of brain activity: Generative models, such as GANs and VAEs, are also known to be vital. The key applications are synthesizing the neural data, reconstructing both images and speech in neuroscience [28].
- RQ1: How do data modalities differ in computational neuroscience in various resolutions, what challenges need to be addressed caused by the limitations in DL models, how can explainability be utilized in neuroscience, and where is our review positioned in advancing the field compared to existing literature surveys?
- RQ2: What modeling techniques are mostly used in visual neuroscience, how does XDL play its role in making models interpretable and transparent, what are the attributes of state-of-the-art XDL approaches in terms of vision tasks solved, methodologies applied, and their respective outcomes achieved?
- RQ3: What XDL techniques are utilizable in clinical neuroscience, how are the techniques leveraged to diagnose, prognose, and treat the neurological and psychiatric conditions, what are the pipeline components in clinics, how do components impact diagnostic and prognostic outcomes, and how do the XDL techniques build trust and transparency through supportive applications?
- RQ4: What are the key challenges in the state-of-the-art research relevant to visual and clinical computational neuroscience, and how are the biological priors integrable?
- RQ5: What future research directions need to be followed for advancing the explainability required for building interpretable and transparent systems in visual and clinical neuroscience environments, and how can the biological priors, standards, and human-involving designs be integrated to address the key challenges?
2. Methodology
2.1. Databases Searched
- PubMed indexes peer-reviewed biomedical and clinical research, including studies involving neuroimaging, cognitive neuroscience, and brain disorders. It was essential for retrieving medically relevant studies using DL models in clinical neuroscience. Coverage ensured comprehensive literature retrieval.
- Scopus provided multidisciplinary coverage, including computational modeling, neuroscience, and machine learning. Its citation tools aided in identifying influential papers and landmark studies.
- Web of Science offered access to high-impact journals in both science and technology, helping identify interdisciplinary contributions in XDL research.
- IEEE Xplore offers access to high-impact technical literature in AI, machine learning, and signal processing. It was crucial for capturing studies on DL model architectures, BCI, and signal decoding from EEG or MEG.
- SpringerLink provides a list of prestigious journals such as Cognitive Neurodynamics, Computational visual media, BioMed Central Medical Ethics, BioData Mining, BioMed Central Psychiatry, and Journal of neuroengineering and rehabilitation. It supports diverse neuroscience research dissemination worldwide.
- ACM Digital Library offers a variety of disciplinary journals related to evolutionary learning and optimization.
- MDPI also offers a list of high-ranked journals, including Mathematics, Diagnostics, Sensors which gave us the knowledge about clinical and visual applications relevant to multidisciplinary domains, including IoT, medical, and mathematics.
- arXiv provides early-access and research. These preprints are not peer-reviewed yet but serve as early dissemination of recent in-progress research work. It facilitates rapid sharing among neuroscience researchers.
- Books/Other Sources provide knowledge on theoretical concepts, standardized methods, and motivational aspects of the research.
2.2. Search Strategy and Keywords
Core concepts | Neuroscience focus |
|
|
Applications | Modalities |
|
|
Techniques | |
|
- AND (e.g., “explainable deep learning AND fMRI”)
- OR (e.g., “brain decoding OR neural signal interpretation”)
- (“Explainable deep learning” OR “Interpretable deep neural networks”OR “Transparent AI” OR “Black-box models” OR “Uninterpretability”OR “Model interpretability”)
- (“Brain decoding” OR “Visual neuroscience” OR “Clinical neuroimaging”OR “Neurobiologically plausible models” OR “Functional brain connectivity”OR “Electrophysiological signal modeling” OR “Spatio-temporal neural dynamics”)
- (“Cognitive state prediction” OR “Motor intention decoding”OR “Neurological disease diagnosis” OR “Psychiatric biomarker extraction”OR “Brain-computer interface explainability” OR “Neural signal reconstruction”)
- (“EEG” OR “MEG” OR “fMRI” OR “ECoG”OR “Multimodal neuroimaging” OR “Time-series neural data”)
- (“Feature attribution” OR “SHAP” OR “LRP” OR “Saliency maps”OR “Concept-based explainability” OR “TCAV” OR “ACE” OR “Causal explanation”OR “Transformer interpretability” OR “Post-hoc explainable methods”OR “Benchmarking interpretability metrics”)
- (“Kappa observer agreement for categorical data”)
2.3. Inclusion and Exclusion Criteria
- One study is from 1988, while most are from 2020 to 2025
- Two datasets are from 2000 to 2008, while most are from 2019 to 2025
- Peer-reviewed journal articles or conference proceedings
- Articles focused on XDL, neuroscience modeling, neuroimaging, or clinical diagnosis
- Research using brain data modalities (EEG, fMRI, MEG, MRI, ECoG)
- Papers on interpretable/explainable brain and cognitive modeling
- Non-English publications and preprints without peer review
- Editorials, white papers, and non-peer-reviewed articles
- Theoretical AI papers not involving neuroscience/brain data
- Duplicates or articles without full-text access
- Reviews without methodological description or reproducible implementation
2.4. Screening Procedure
3. Background and Foundations
3.1. Computational Neuroscience: A Brief Overview
3.2. Deep Learning Models in Neuroscience
3.3. Explainable Deep Learning: Concepts and Techniques
3.4. Comparison with Existing Surveys
4. XDL in Visual Computational Neuroscience
4.1. Modeling Visual Attention and Categorization
4.2. XDL for Interpreting Visual Models
5. XDL in Clinical Computational Neuroscience
5.1. Clinical Use Cases of DL Models
5.2. Enhancing Trust in Clinical Models with XDL
5.3. Salient, Concept, and Causal-Based Validation in Clinical XDL
6. Discussion and Future Directions
6.1. Key Challenges in XDL for Neuroscience
- Balancing accuracy and clarity: Some of the most accurate models, like deep CNNs and transformers, and are often difficult to interpret. On the other hand, models that are easier to explain typically do not perform as well as the complex models. Therefore, trading off between these two goals is a major challenge in computational neuroscience, especially when considering the critical clinical decisions. Beyond individual strengths, these XDL methods have some limitations that need to be addressed. These limitations include instability, reproducibility, and clinical adaptability. Computational cost is another big challenge, as high-dimensional neuroscientific imaging data requires improved models for producing better results. The techniques mentioned in this survey are domain-dependent, making the current explainable systems unable to solve a problem in a generalized way as the real-world demands.
- Missing evaluation standards: Currently, there is no widely accepted approach or method for measuring or validating the ability of a model to explain its outcome. Tools vary in what they aim to show significantly, some focus on pixel-level importance, others on broader concepts, making it hard to evaluate results or reproduce findings across studies.
- Issues with generalization and real-time use: Many models work well on specific datasets but do not hold up when tested across different groups, imaging types, or use cases. Real-time performance is also a problem, especially in clinical areas like brain-computer interfaces. On top of that, we still do not fully understand how robust these models are to noise or unexpected input changes. Recent studies also show that hybrid anomaly detection strategies can improve robustness by the removal of noise in data and by the regularization of data. This work highlights the importance of resilience in clinical XDL systems [152].
- Dataset characteristics can introduce biases. These properties can affect the XDL model’s interpretability and generalization. The most common issues are class imbalance, local-specific variations, and demographic disparities. To tackle these challenges, various harmonization strategies could be applied. Preprocessing normalization, domain adaptation, class balancing, and multimodal feature standardization are a few strategies that can be applied. Being aware of such strategies and reporting them are both crucial. These steps need to be taken to ensure reliable and generalizable explanations in computational neuroscience. Recent multimodal studies still demand enhancements in several aspects for addressing clinical challenges for alzheimer (MRI and PET) [70], epilepsy (EEG and fMRI) [74], brain tumors (MRI and genomics) [86], parkinson disease (MRI and clinical scores) [128], population densities (MRI and genetics) [83] problem domains. Identified gaps of each multimodal research work mentioned here are discussed, along with their proposed solutions, in Section 6.3.
6.2. Future Research Directions
- Integration of biological priors into model architectures: Embedding anatomical, functional, or physiological knowledge (sEMG-based muscle fatigue detection [155]) into DL models can enhance alignment with neural mechanisms and improve both interpretability and generalization. Another innovative advancement is the integration of ultra-high-frequency biosignal retrieval [156], which enhances the systems to show real-time interpretable DL. Further details on biological priors are depicted in Section 6.2.1.
- Standardized frameworks for XDL evaluation: Developing rigorous benchmarks, grounded in neuroscientific and clinical relevance, will facilitate objective assessment of explainability methods and model comparisons across studies. Recent works that utilize transformer-based clinical emotion classification emphasize the need to incorporate evaluation pipelines to enable trustworthy deployment [157].
- Human-in-the-loop and neuro-symbolic models: Combining human domain expertise with automated learning systems can improve interpretability and trust. Neuro-symbolic approaches, which integrate symbolic reasoning with neural representations, offer a promising hybrid paradigm.
- Collaborative and interdisciplinary research models: Progress in XDL will require active collaboration between AI researchers, neuroscientists, clinicians, and ethicists. Such partnerships can accelerate development while ensuring ethical deployment and practical utility.
6.2.1. Integration of Biological Priors in XDL
- Architectural constraints: The biological constraints are a key part in modeling neuroscientific models. A recent research focuses on the necessity of integration of biological priors into current DL model designs. Authors propose a brain-constrained modeling approach to incorporate the principles of brain structure internals like synaptic plasticity, local and long-range connectivity, and area-specific compositions into the design of the model [158]. This can lead to significant improvements in current DL techniques to improve the accuracy and explanatory abilities of such models for neuroscience applications.Another study investigates DL based networks in which hidden layers correspond to and mimic the structure of a biological entity, i.e., genes, pathways, and their connections, to encode the relationships of brain-mimicking nodes [159].Another study emphasizes leveraging the biological implausibility in backpropagation. The updating of errors while training these models is inspired by the actual human behavior when correcting mistakes [160].
- Regularization and loss functions: The enhancements in regularization and loss functions are a necessity for minimizing the prediction errors considering the integration of biological priors in both visual and clinical neuroscience [161]. Experiments in this work show that utilizing the structured loss can improve the MCTs like incremental learning, few-shot learning, and long-tailed classification.Another study proposes a model named bio-RNN that is trained with multiple loss terms [162], e.g., neuron-level loss, trial-matching loss, and regularization loss. These losses jointly optimize the updating weights of the model network.
- Data augmentation with priors: Data augmentation improves and extends the existing dataset, and is much needed, as the clinical datasets, specifically, are very limited in size. A study on one such augmentation technique called phylogenetic is applied, which is homologous sequence from related species [163]. This is particularly the best technique for small datasets.
- Architectural Hybrid neuro-symbolic models: Neuro-symbolic models are enabled with the ability to reason [164]. This property is embedded into models by combining NN components with symbolic reasoning components. Reasoning components include logic, rules, planning, high-level intents in hierarchical form.Another study presents a hybrid architecture to combine deep neural components. This work emphasizes the integration of pattern recognition extraction components with symbolic reasoning, leveraging the medical ontologies, rule-based inference, and statistics and probabilities [165].Another study on enforcing structural priors involves the embedding of symbolic modules in the form of known anatomical circuits and functional hierarchies into the model architecture [166].
6.2.2. Strategic Directions for Advancing XDL
6.3. Research Gaps and Proposed Approaches
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
ACE | Automated Concept-based Explanation |
ACER | Average Classification Error Rate |
ADNI | Alzheimer’s Disease Neuroimaging Initiative |
APCER | Attack Presentation Classification Error Rate |
BCI | Brain-Computer Interface |
BMS | Bayesian Model Selection |
BPCER | Bona Fide Presentation Classification Error Rate |
BraTS | Brain Tumor Segmentation |
BSMMCH | Bangabandhu Sheikh Mujib Medical College Hospital |
CAM | Class Activation Map |
CAMCAN | Cambridge Centre for Ageing and Neuroscience |
CI | Confidence Interval |
CNN | Convolutional Neural Network |
CORnet-S | Core Object Recognition Network Shallow |
DBS | Deep Brain Stimulation |
DL | Deep Learning |
DR | Diabetic Retinopathy |
DTI | Diffusion Tensor Imaging |
DVA | Deep Visual Analytics |
DWI | Diffusion-Weighted Imaging |
ECoG | Electrocorticography |
ECSSD | Extended Complex Scene Saliency Dataset |
EDF | European Data Format |
EEG | Electroencephalography |
EHR | Electronic Health Records |
EM | Electron Microscopy |
EOG | Electrooculography |
F-M | F-measure |
fMRI | Functional Magnetic Resonance Imaging |
fNIRS | Functional Near-Infrared Spectroscopy |
GANs | Generative Adversarial Networks |
GBVS | Graph-Based Visual Saliency |
HCP | Human Connectome Project |
HGC | High-Grade Condition |
HKUIS | Hong Kong University Image Saliency |
HMA | Hard Exudate Microaneurysms |
iCoSeg | Interactive Co-segmentation |
IoT | Internet of Things |
IP | Image Processing |
IT | Inferior Temporal cortex |
IVW | Inverse-Variance Weighted |
LGC | Low-Grade Condition |
LRP | Layer-wise Relevance Propagation |
MA | Exudate Microaneurysms |
MAE | Mean Absolute Error |
MCT | Machine-challenging Task |
MDD | Major Depressive Disorder |
MEG | Magnetoencephalography |
MR | Mendelian Randomization |
MRI | Magnetic Resonance Imaging |
MSRA-B | Microsoft Research Asia (version B) |
NLP | Natural Language Processing |
NMM | Normalized Mean value under the annotated Mask |
NN | Neural Network |
NST | Non-Specific Tumor |
NTL | Normal Tissue Label |
OASIS | Open Access Series of Imaging Studies |
OCTNet | Optical Coherence Tomography Network |
PAD | Presentation Attack Detection |
PRISMA | Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
PET | Positron Emission Tomography |
REST | Resting-State |
PPG | Photoplethysmogram |
RNN | Recurrent Neural Network |
RSA | Representational Similarity Analysis |
SAM | Segment Anything Model |
SaMD | Software as a Medical Device |
SED | Social Event Detection |
sEMG | Surface Electromyography |
SNP | Single Nucleotide Polymorphism |
SOD | Salient Object Detection |
SSM | State Space Model |
TCAV | Testing with Concept Activation Vectors |
TCGA | The Cancer Genome Atlas |
TUH | Temple University Hospital |
V1/V4 | Visual Cortex Area 1/Visual Cortex Area 4 |
ViT | Vision Transformer |
VAE | Variational Autoencoder |
VGG | Visual Geometry Group |
XAI | Explainable Artificial Intelligence |
XDL | Explainable Deep Learning |
References
- Guimarães, K.; Duarte, A. A Network-Level Stochastic Model for Pacemaker GABAergic Neurons in Substantia Nigra Pars Reticulata. Mathematics 2023, 11, 3778. [Google Scholar] [CrossRef]
- Iwama, S.; Tsuchimoto, S.; Mizuguchi, N.; Ushiba, J. EEG decoding with spatiotemporal convolutional neural network for visualization and closed-loop control of sensorimotor activities: A simultaneous EEG-fMRI study. Hum. Brain Mapp. 2024, 45, e26767. [Google Scholar] [CrossRef]
- Huang, Y.; Huan, Y.; Zou, Z.; Wang, Y.; Gao, X.; Zheng, L. Data-driven natural computational psychophysiology in class. Cogn. Neurodynamics 2024, 18, 3477–3489. [Google Scholar] [CrossRef]
- Elmoznino, E.; Bonner, M.F. High-performing neural network models of visual cortex benefit from high latent dimensionality. PLoS Comput. Biol. 2024, 20, e1011792. [Google Scholar] [CrossRef]
- Guo, M.H.; Xu, T.X.; Liu, J.J.; Liu, Z.N.; Jiang, P.T.; Mu, T.J.; Zhang, S.H.; Martin, R.R.; Cheng, M.M.; Hu, S.M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
- Watanabe, N.; Miyoshi, K.; Jimura, K.; Shimane, D.; Keerativittayayut, R.; Nakahara, K.; Takeda, M. Multimodal deep neural decoding reveals highly resolved spatiotemporal profile of visual object representation in humans. NeuroImage 2023, 275, 120164. [Google Scholar] [CrossRef] [PubMed]
- Farzmahdi, A.; Zarco, W.; Freiwald, W.A.; Kriegeskorte, N.; Golan, T. Emergence of brain-like mirror-symmetric viewpoint tuning in convolutional neural networks. eLife 2024, 13, e90256. [Google Scholar] [CrossRef] [PubMed]
- Salahuddin, Z.; Woodruff, H.C.; Chatterjee, A.; Lambin, P. Transparency of deep neural networks for medical image analysis: A review of interpretability methods. Comput. Biol. Med. 2022, 140, 105111. [Google Scholar] [CrossRef]
- Danks, D.; Davis, I. Causal inference in cognitive neuroscience. Wiley Interdiscip. Rev. Cogn. Sci. 2023, 14, e1650. [Google Scholar] [CrossRef]
- Žlahtič, B.; Završnik, J.; Kokol, P.; Blažun Vošner, H.; Sobotkiewicz, N.; Antolinc Schaubach, B.; Kirbiš, S. Trusting AI made decisions in healthcare by making them explainable. Sci. Prog. 2024, 107, 00368504241266573. [Google Scholar] [CrossRef] [PubMed]
- Freyer, N.; Groß, D.; Lipprandt, M. The ethical requirement of explainability for AI-DSS in healthcare: A systematic review of reasons. BMC Med. Ethics 2024, 25, 104. [Google Scholar] [CrossRef]
- Brima, Y.; Atemkeng, M. Saliency-driven explainable deep learning in medical imaging: Bridging visual explainability and statistical quantitative analysis. BioData Min. 2024, 17, 18. [Google Scholar] [CrossRef] [PubMed]
- Yang, L.; Qiao, C.; Zhou, H.; Calhoun, V.D.; Stephen, J.M.; Wilson, T.W.; Wang, Y. Explainable multimodal deep dictionary learning to capture developmental differences from three fMRI paradigms. IEEE Trans. Biomed. Eng. 2023, 70, 2404–2415. [Google Scholar] [CrossRef]
- Sattiraju, A.; Ellis, C.A.; Miller, R.L.; Calhoun, V.D. An explainable and robust deep learning approach for automated electroencephalography-based schizophrenia diagnosis. In Proceedings of the 2023 IEEE 23rd International Conference on Bioinformatics and Bioengineering (BIBE), Dayton, OH, USA, 4–6 December 2023; pp. 255–259. [Google Scholar]
- Rodríguez Mallma, M.J.; Zuloaga-Rotta, L.; Borja-Rosales, R.; Rodríguez Mallma, J.R.; Vilca-Aguilar, M.; Salas-Ojeda, M.; Mauricio, D. Explainable Machine Learning Models for Brain Diseases: Insights from a Systematic Review. Neurol. Int. 2024, 16, 1285–1307. [Google Scholar] [CrossRef]
- Hussain, T.; Shouno, H. Explainable deep learning approach for multi-class brain magnetic resonance imaging tumor classification and localization using gradient-weighted class activation mapping. Information 2023, 14, 642. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar] [CrossRef]
- Li, Y.; Lou, A.; Xu, Z.; Zhang, S.; Wang, S.; Englot, D.; Kolouri, S.; Moyer, D.; Bayrak, R.; Chang, C. Neurobolt: Resting-state eeg-to-fmri synthesis with multi-dimensional feature mapping. Adv. Neural Inf. Process. Syst. 2024, 37, 23378–23405. [Google Scholar]
- Song, L.; Ren, Y.; Xu, S.; Hou, Y.; He, X. A hybrid spatiotemporal deep belief network and sparse representation-based framework reveals multilevel core functional components in decoding multitask fMRI signals. Netw. Neurosci. 2023, 7, 1513–1532. [Google Scholar] [CrossRef] [PubMed]
- Hu, W.; Meng, X.; Bai, Y.; Zhang, A.; Qu, G.; Cai, B.; Zhang, G.; Wilson, T.W.; Stephen, J.M.; Calhoun, V.D.; et al. Interpretable multimodal fusion networks reveal mechanisms of brain cognition. IEEE Trans. Med. Imaging 2021, 40, 1474–1483. [Google Scholar] [CrossRef] [PubMed]
- Ji, Y.; Yang, C.; Liang, Y. A multiview deep learning method for brain functional connectivity classification. Comput. Intell. Neurosci. 2022, 2022, 5782569. [Google Scholar] [CrossRef]
- Germann, J.; Zadeh, G.; Mansouri, A.; Kucharczyk, W.; Lozano, A.M.; Boutet, A. Untapped neuroimaging tools for neuro-oncology: Connectomics and spatial transcriptomics. Cancers 2022, 14, 464. [Google Scholar] [CrossRef]
- Luo, X.; Zeng, Z.; Zheng, S.; Chen, J.; Jannin, P. Statistical Multiscore Functional Atlas Creation for Image-Guided Deep Brain Stimulation. IEEE Trans. Neural Syst. Rehabil. Eng. 2025, 33, 818–828. [Google Scholar] [CrossRef]
- Li, R. Integrative diagnosis of psychiatric conditions using ChatGPT and fMRI data. BMC Psychiatry 2025, 25, 145. [Google Scholar] [CrossRef]
- Khan, S.; Kallis, L.; Mee, H.; El Hadwe, S.; Barone, D.; Hutchinson, P.; Kolias, A. Invasive Brain–Computer Interface for Communication: A Scoping Review. Brain Sci. 2025, 15, 336. [Google Scholar] [CrossRef]
- Baniqued, P.D.E.; Stanyer, E.C.; Awais, M.; Alazmani, A.; Jackson, A.E.; Mon-Williams, M.A.; Mushtaq, F.; Holt, R.J. Brain–computer interface robotics for hand rehabilitation after stroke: A systematic review. J. Neuroeng. Rehabil. 2021, 18, 1–25. [Google Scholar] [CrossRef]
- Zhang, X.; Zhang, T.; Jiang, Y.; Zhang, W.; Lu, Z.; Wang, Y.; Tao, Q. A novel brain-controlled prosthetic hand method integrating AR-SSVEP augmentation, asynchronous control, and machine vision assistance. Heliyon 2024, 10, e26521. [Google Scholar] [CrossRef] [PubMed]
- Berezutskaya, J.; Freudenburg, Z.V.; Vansteensel, M.J.; Aarnoutse, E.J.; Ramsey, N.F.; van Gerven, M.A. Direct speech reconstruction from sensorimotor brain activity with optimized deep learning models. J. Neural Eng. 2023, 20, 056010. [Google Scholar] [CrossRef]
- Lai, T. Interpretable medical imagery diagnosis with self-attentive transformers: A review of explainable AI for health care. BioMedInformatics 2024, 4, 113–126. [Google Scholar] [CrossRef]
- Sinz, F.H.; Pitkow, X.; Reimer, J.; Bethge, M.; Tolias, A.S. Engineering a less artificial intelligence. Neuron 2019, 103, 967–979. [Google Scholar] [CrossRef] [PubMed]
- Stringer, C.; Pachitariu, M.; Steinmetz, N.; Carandini, M.; Harris, K.D. High-dimensional geometry of population responses in visual cortex. Nature 2019, 571, 361–365. [Google Scholar] [CrossRef]
- Räuker, T.; Ho, A.; Casper, S.; Hadfield-Menell, D. Toward transparent ai: A survey on interpreting the inner structures of deep neural networks. In Proceedings of the 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SATML), Raleigh, NC, USA, 8–10 February 2023; pp. 464–483. [Google Scholar]
- Retzlaff, C.O.; Angerschmid, A.; Saranti, A.; Schneeberger, D.; Roettger, R.; Mueller, H.; Holzinger, A. Post-hoc vs ante-hoc explanations: xAI design guidelines for data scientists. Cogn. Syst. Res. 2024, 86, 101243. [Google Scholar] [CrossRef]
- Salih, A.M.; Galazzo, I.B.; Gkontra, P.; Rauseo, E.; Lee, A.M.; Lekadir, K.; Radeva, P.; Petersen, S.E.; Menegaz, G. A review of evaluation approaches for explainable AI with applications in cardiology. Artif. Intell. Rev. 2024, 57, 240. [Google Scholar] [CrossRef]
- Liu, Y.; Ge, E.; Kang, Z.; Qiang, N.; Liu, T.; Ge, B. Spatial-temporal convolutional attention for discovering and characterizing functional brain networks in task fMRI. NeuroImage 2024, 287, 120519. [Google Scholar] [CrossRef]
- El Shawi, R. Conceptglassbox: Guided concept-based explanation for deep neural networks. Cogn. Comput. 2024, 16, 2660–2673. [Google Scholar] [CrossRef]
- Schwalbe, G.; Finzel, B. A comprehensive taxonomy for explainable artificial intelligence: A systematic survey of surveys on methods and concepts. Data Min. Knowl. Discov. 2024, 38, 3043–3101. [Google Scholar] [CrossRef]
- Wang, Z.; Huang, C.; Li, Y.; Yao, X. Multi-objective feature attribution explanation for explainable machine learning. ACM Trans. Evol. Learn. Optim. 2024, 4, 1–32. [Google Scholar] [CrossRef]
- Lee, S.; Lee, K.S. Predictive and Explainable Artificial Intelligence for Neuroimaging Applications. Diagnostics 2024, 14, 2394. [Google Scholar] [CrossRef] [PubMed]
- Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G.; PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. Int. J. Surg. 2010, 8, 336–341. [Google Scholar] [CrossRef] [PubMed]
- Kemp, B.; Zwinderman, A.H.; Tuk, B.; Kamphuisen, H.A.; Oberye, J.J. Analysis of a sleep-dependent neuronal feedback loop: The slow-wave microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 2000, 47, 1185–1194. [Google Scholar] [CrossRef]
- LaMontagne, P.J.; Benzinger, T.L.; Morris, J.C.; Keefe, S.; Hornbeck, R.; Xiong, C.; Grant, E.; Hassenstab, J.; Moulder, K.; Vlassenko, A.G.; et al. OASIS-3: Longitudinal neuroimaging, clinical, and cognitive dataset for normal aging and Alzheimer disease. medRxiv 2019. [Google Scholar] [CrossRef]
- Markiewicz, C.J.; Gorgolewski, K.J.; Feingold, F.; Blair, R.; Halchenko, Y.O.; Miller, E.; Hardcastle, N.; Wexler, J.; Esteban, O.; Goncavles, M.; et al. The OpenNeuro resource for sharing of neuroscience data. eLife 2021, 10, e71774. [Google Scholar] [CrossRef]
- Mehmood, A.; Ko, J.; Kim, H.; Kim, J. Optimizing Image Enhancement: Feature Engineering for Improved Classification in AI-Assisted Artificial Retinas. Sensors 2024, 24, 2678. [Google Scholar] [CrossRef] [PubMed]
- Landis, J.R.; Koch, G.G. The measurement of observer agreement for categorical data. Biometrics 1977, 159–174. [Google Scholar] [CrossRef]
- Lai-Dang, Q.V.; Kang, T.; Son, S. Adaptive Transformer Programs: Bridging the Gap Between Performance and Interpretability in Transformers. In Proceedings of the The Thirteenth International Conference on Learning Representations, Singapore, 24–28 April 2025. [Google Scholar]
- Dayan, P.; Abbott, L.F. Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems; MIT Press: Cambridge, MA, USA, 2005; Volume 15, pp. 563–577. [Google Scholar]
- Koch, C.; Segev, I. Methods in Neuronal Modeling: From Synapses to Networks; Computational Neuroscience; MIT Press: Cambridge, MA, USA, 1988; p. 524. [Google Scholar]
- Moshruba, A.; Poursiami, H.; Parsa, M. Izhikevich-Inspired Temporal Dynamics for Enhancing Privacy, Efficiency, and Transferability in Spiking Neural Networks. arXiv 2025, arXiv:2505.04034. [Google Scholar] [CrossRef]
- Wojcik, G.M.; Masiak, J.; Kawiak, A.; Kwasniewicz, L.; Schneider, P.; Polak, N.; Gajos-Balinska, A. Mapping the human brain in frequency band analysis of brain cortex electroencephalographic activity for selected psychiatric disorders. Front. Neuroinform. 2018, 12, 73. [Google Scholar] [CrossRef]
- Yang, Y.; Luo, S.; Wang, W.; Gao, X.; Yao, X.; Wu, T. From bench to bedside: Overview of magnetoencephalography in basic principle, signal processing, source localization and clinical applications. Neuroimage Clin. 2024, 42, 103608. [Google Scholar] [CrossRef]
- Kaiju, T.; Doi, K.; Yokota, M.; Watanabe, K.; Inoue, M.; Ando, H.; Takahashi, K.; Yoshida, F.; Hirata, M.; Suzuki, T. High spatiotemporal resolution ECoG recording of somatosensory evoked potentials with flexible micro-electrode arrays. Front. Neural Circuits 2017, 11, 20. [Google Scholar] [CrossRef]
- Sadad, T.; Aurangzeb, R.A.; Safran, M.; Imran; Alfarhood, S.; Kim, J. Classification of highly divergent viruses from DNA/RNA sequence using transformer-based models. Biomedicines 2023, 11, 1323. [Google Scholar] [CrossRef]
- Abrol, A.; Fu, Z.; Salman, M.; Silva, R.; Du, Y.; Plis, S.; Calhoun, V. Deep learning encodes robust discriminative neuroimaging representations to outperform standard machine learning. Nat. Commun. 2021, 12, 353. [Google Scholar] [CrossRef] [PubMed]
- Storrs, K.R.; Kriegeskorte, N. Deep learning for cognitive neuroscience. arXiv 2019, arXiv:1903.01458. [Google Scholar] [CrossRef]
- Livezey, J.A.; Glaser, J.I. Deep learning approaches for neural decoding across architectures and recording modalities. Briefings Bioinform. 2021, 22, 1577–1591. [Google Scholar] [CrossRef]
- Jack, C.R., Jr.; Bernstein, M.A.; Fox, N.C.; Thompson, P.; Alexander, G.; Harvey, D.; Borowski, B.; Britson, P.J.; L. Whitwell, J.; Ward, C.; et al. The Alzheimer’s disease neuroimaging initiative (ADNI): MRI methods. J. Magn. Reson. Imaging Off. J. Int. Soc. Magn. Reson. Med. 2008, 27, 685–691. [Google Scholar] [CrossRef]
- Mehmood, F.; Mehmood, A.; Whangbo, T.K. Alzheimer’s Disease Detection in Various Brain Anatomies Based on Optimized Vision Transformer. Mathematics 2025, 13, 1927. [Google Scholar] [CrossRef]
- Rehman, A.; Yi, M.K.; Majeed, A.; Hwang, S.O. Early diagnosis of alzheimer’s disease using 18f-fdg pet with soften latent representation. IEEE Access 2024, 12, 87923–87933. [Google Scholar] [CrossRef]
- Makropoulos, A.; Robinson, E.C.; Schuh, A.; Wright, R.; Fitzgibbon, S.; Bozek, J.; Counsell, S.J.; Steinweg, J.; Vecchiato, K.; Passerat-Palmbach, J.; et al. The developing human connectome project: A minimal processing pipeline for neonatal cortical surface reconstruction. Neuroimage 2018, 173, 88–112. [Google Scholar] [CrossRef]
- Liao, W.; Wang, W. Eegencoder: Advancing bci with transformer-based motor imagery classification. arXiv 2024, arXiv:2404.14869. [Google Scholar] [CrossRef]
- Shahwar, T.; Rehman, A.U. Automated detection of brain disease using quantum machine learning. In Brain-Computer Interfaces; Elsevier: Amsterdam, The Netherlands, 2025; pp. 91–114. [Google Scholar]
- Ishfaq, Q.U.A.; Bibi, R.; Ali, A.; Jamil, F.; Saeed, Y.; Alnashwan, R.O.; Chelloug, S.A.; Muthanna, M.S.A. Automatic smart brain tumor classification and prediction system using deep learning. Sci. Rep. 2025, 15, 14876. [Google Scholar] [CrossRef] [PubMed]
- Hashmi, S.; Lugo, J.; Elsayed, A.; Saggurthi, D.; Elseiagy, M.; Nurkamal, A.; Walia, J.; Maani, F.A.; Yaqub, M. Optimizing brain tumor segmentation with mednext: Brats 2024 ssa and pediatrics. arXiv 2024, arXiv:2411.15872. [Google Scholar]
- Abbas, K.; Khan, P.W.; Ahmed, K.T.; Song, W.C. Automatic brain tumor detection in medical imaging using machine learning. In Proceedings of the 2019 International Conference on Information and Communication Technology Convergence (ICTC), Jeju, Republic of Korea, 16–18 October 2019; pp. 531–536. [Google Scholar]
- Bhutta, M.A.M.; Kim, S.; Kim, Y.J. Dual Photoplethysmography Pulse Wave Velocity-Based Blood Pressure Prediction for Wearable Healthcare Internet of Things. IEEE Internet Things J. 2025, 12, 20773–20786. [Google Scholar] [CrossRef]
- Ali, S.T.A.; Kim, S.; Kim, Y.J. Towards Reliable ECG Analysis: Addressing Validation Gaps in the Electrocardiographic R-Peak Detection. Appl. Sci. 2024, 14, 10078. [Google Scholar] [CrossRef]
- Bycroft, C.; Freeman, C.; Petkova, D.; Band, G.; Elliott, L.T.; Sharp, K.; Motyer, A.; Vukcevic, D.; Delaneau, O.; O’Connell, J.; et al. The UK Biobank resource with deep phenotyping and genomic data. Nature 2018, 562, 203–209. [Google Scholar] [CrossRef]
- Landau, S.M.; Harrison, T.M.; Baker, S.L.; Boswell, M.S.; Lee, J.; Taggett, J.; Ward, T.J.; Chadwick, T.; Murphy, A.; DeCarli, C.; et al. Positron emission tomography harmonization in the Alzheimer’s Disease Neuroimaging Initiative: A scalable and rigorous approach to multisite amyloid and tau quantification. Alzheimers Dement. 2025, 21, e14378. [Google Scholar] [CrossRef]
- Abdelaziz, M.; Wang, T.; Anwaar, W.; Elazab, A. Multi-scale multimodal deep learning framework for Alzheimer’s disease diagnosis. Comput. Biol. Med. 2025, 184, 109438. [Google Scholar] [CrossRef]
- Matsushita, S.; Tatekawa, H.; Ueda, D.; Takita, H.; Horiuchi, D.; Tsukamoto, T.; Shimono, T.; Miki, Y. The Association of Metabolic Brain MRI, amyloid PET, and clinical factors: A study of Alzheimer’s disease and Normal controls from the open access series of imaging studies dataset. J. Magn. Reson. Imaging 2024, 59, 1341–1348. [Google Scholar] [CrossRef]
- Van Essen, D.C.; Smith, S.M.; Barch, D.M.; Behrens, T.E.; Yacoub, E.; Ugurbil, K.; WU-Minn HCP Consortium. The WU-Minn human connectome project: An overview. Neuroimage 2013, 80, 62–79. [Google Scholar] [CrossRef]
- Germani, E.; Fromont, E.; Maurel, P.; Maumet, C. The HCP multi-pipeline dataset: An opportunity to investigate analytical variability in fMRI data analysis. arXiv 2023, arXiv:2312.14493. [Google Scholar]
- Warbrick, T. Simultaneous EEG-fMRI: What have we learned and what does the future hold? Sensors 2022, 22, 2262. [Google Scholar] [CrossRef]
- Shafto, M.A.; Tyler, L.K.; Dixon, M.; Taylor, J.R.; Rowe, J.B.; Cusack, R.; Calder, A.J.; Marslen-Wilson, W.D.; Duncan, J.; Dalgleish, T.; et al. The Cambridge Centre for Ageing and Neuroscience (Cam-CAN) study protocol: A cross-sectional, lifespan, multidisciplinary examination of healthy cognitive ageing. BMC Neurol. 2014, 14, 204. [Google Scholar] [CrossRef]
- Taylor, J.R.; Williams, N.; Cusack, R.; Auer, T.; Shafto, M.A.; Dixon, M.; Tyler, L.K.; Henson, R.N. The Cambridge Centre for Ageing and Neuroscience (Cam-CAN) data repository: Structural and functional MRI, MEG, and cognitive data from a cross-sectional adult lifespan sample. Neuroimage 2017, 144, 262–269. [Google Scholar] [CrossRef] [PubMed]
- Tangermann, M.; Müller, K.R.; Aertsen, A.; Birbaumer, N.; Braun, C.; Brunner, C.; Leeb, R.; Mehring, C.; Miller, K.J.; Müller-Putz, G.R.; et al. Review of the BCI competition IV. Front. Neurosci. 2012, 6, 55. [Google Scholar] [CrossRef] [PubMed]
- Zhang, H.; Guan, C.; Ang, K.K.; Wang, C.; Chin, Z.Y. BCI competition IV–data set I: Learning discriminative patterns for self-paced EEG-based motor imagery detection. Front. Neurosci. 2012, 6, 7. [Google Scholar]
- Obeid, I.; Picone, J. The temple university hospital EEG data corpus. Front. Neurosci. 2016, 10, 196. [Google Scholar] [CrossRef]
- Mellot, A.; Collas, A.; Rodrigues, P.L.; Engemann, D.; Gramfort, A. Harmonizing and aligning M/EEG datasets with covariance-based techniques to enhance predictive regression modeling. Imaging Neurosci. 2023, 1, 1–23. [Google Scholar] [CrossRef]
- Ujma, P.P.; Dresler, M.; Bódizs, R. Comparing Manual and Automatic Artifact Detection in Sleep EEG Recordings. Psychophysiology 2025, 62, e70016. [Google Scholar] [CrossRef] [PubMed]
- Sudlow, C.; Gallacher, J.; Allen, N.; Beral, V.; Burton, P.; Danesh, J.; Downey, P.; Elliott, P.; Green, J.; Landray, M.; et al. UK biobank: An open access resource for identifying the causes of a wide range of complex diseases of middle and old age. PLoS Med. 2015, 12, e1001779. [Google Scholar] [CrossRef] [PubMed]
- Mirabnahrazam, G.; Ma, D.; Lee, S.; Popuri, K.; Lee, H.; Cao, J.; Wang, L.; Galvin, J.E.; Beg, M.F.; Initiative, A.D.N. Machine learning based multimodal neuroimaging genomics dementia score for predicting future conversion to alzheimer’s disease. J. Alzheimers Dis. 2022, 87, 1345–1365. [Google Scholar] [CrossRef]
- Starck, S.; Sideri-Lampretsa, V.; Ritter, J.J.; Zimmer, V.A.; Braren, R.; Mueller, T.T.; Rueckert, D. Using uk biobank data to establish population-specific atlases from whole body mri. Commun. Med. 2024, 4, 237. [Google Scholar] [CrossRef]
- Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
- Lyu, Q.; Costen, F. A twin-tower model using MRI and gene for prediction on brain tumor patients’ response to therapy. Bioinform. Adv. 2025, 5, vbaf041. [Google Scholar] [CrossRef] [PubMed]
- Sidume, F.; Soula, O.; Wacira, J.M.; Zhu, Y.; Muhammad, A.R.; Zeraii, A.; Kalejaye, O.; Ibrahim, H.; Gaddour, O.; Halubanza, B.; et al. Resource-Efficient Glioma Segmentation on Sub-Saharan MRI. arXiv 2025, arXiv:2509.09469. [Google Scholar]
- Bhati, D.; Neha, F.; Amiruzzaman, M.; Guercio, A.; Shukla, D.K.; Ward, B. Neural network interpretability with layer-wise relevance propagation: Novel techniques for neuron selection and visualization. In Proceedings of the 2025 IEEE 15th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 6–8 January 2025; pp. 441–447. [Google Scholar]
- Nicolson, A.; Schut, L.; Noble, J.A.; Gal, Y. Explaining Explainability: Recommendations for Effective Use of Concept Activation Vectors. arXiv 2024, arXiv:2404.03713. [Google Scholar]
- Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K.R. Evaluating the visualization of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2660–2673. [Google Scholar] [CrossRef]
- Ramachandram, D.; Joshi, H.; Zhu, J.; Gandhi, D.; Hartman, L.; Raval, A. Transparent AI: The Case for Interpretability and Explainability. arXiv 2025, arXiv:2507.23535. [Google Scholar] [CrossRef]
- Mehmood, F.; Rehman, S.U.; Choi, A. Vision-AQ: Explainable Multi-Modal Deep Learning for Air Pollution Classification in Smart Cities. Mathematics 2025, 13, 3017. [Google Scholar] [CrossRef]
- Rehman, S.U.; Yasin, A.U.; Ul Haq, E.; Ali, M.; Kim, J.; Mehmood, A. Enhancing human activity recognition through integrated multimodal analysis: A focus on RGB imaging, skeletal tracking, and pose estimation. Sensors 2024, 24, 4646. [Google Scholar] [CrossRef] [PubMed]
- Park, E.S.; Kim, S.; Mujtaba, G.; Ryu, E.S. Detection of Power Transmission Equipment in Image using Guided Grad-CAM. In Proceedings of the 2020 Summer Conference of the Korean Society of Broadcast and Media Engineers (KIBME), Online, 13–15 July 2020; pp. 709–713. [Google Scholar]
- Mi, X.; Zou, B.; Zou, F.; Hu, J. Permutation-based identification of important biomarkers for complex diseases via machine learning models. Nat. Commun. 2021, 12, 3008. [Google Scholar] [CrossRef]
- Ramaswamy, V.V.; Kim, S.S.; Fong, R.; Russakovsky, O. Overlooked factors in concept-based explanations: Dataset choice, concept learnability, and human capability. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10932–10941. [Google Scholar]
- Dai, Z.; Gifford, D.K. Ablation based counterfactuals. arXiv 2024, arXiv:2406.07908. [Google Scholar] [CrossRef]
- Saranya, A.; Subhashini, R. A systematic review of Explainable Artificial Intelligence models and applications: Recent developments and future trends. Decis. Anal. J. 2023, 7, 100230. [Google Scholar] [CrossRef]
- Fellous, J.M.; Sapiro, G.; Rossi, A.; Mayberg, H.; Ferrante, M. Explainable artificial intelligence for neuroscience: Behavioral neurostimulation. Front. Neurosci. 2019, 13, 1346. [Google Scholar] [CrossRef]
- Mahmood, S.; Teo, C.; Sim, J.; Zhang, W.; Muyun, J.; Bhuvana, R.; Teo, K.; Yeo, T.T.; Lu, J.; Gulyas, B.; et al. The application of eXplainable artificial intelligence in studying cognition: A scoping review. Ibrain 2024, 10, 245–265. [Google Scholar] [CrossRef]
- Sathyan, A.; Weinberg, A.I.; Cohen, K. Interpretable AI for bio-medical applications. Complex Eng. Syst. 2022, 2, 18. [Google Scholar] [CrossRef]
- Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
- Warren, S.L.; Moustafa, A.A. Functional magnetic resonance imaging, deep learning, and Alzheimer’s disease: A systematic review. J. Neuroimaging 2023, 33, 5–18. [Google Scholar] [CrossRef]
- Avberšek, L.K.; Repovš, G. Deep learning in neuroimaging data analysis: Applications, challenges, and solutions. Front. Neuroimaging 2022, 1, 981642. [Google Scholar] [CrossRef] [PubMed]
- Feng, B.; Yu, T.; Wang, H.; Liu, K.; Wu, W.; Long, W. Machine learning and deep learning in biomedical signal analysis. Front. Hum. Neurosci. 2023, 17, 1183840. [Google Scholar] [CrossRef] [PubMed]
- Cui, J.; Yuan, L.; Wang, Z.; Li, R.; Jiang, T. Towards best practice of interpreting deep learning models for EEG-based brain computer interfaces. Front. Comput. Neurosci. 2023, 17, 1232925. [Google Scholar] [CrossRef]
- Mathi, S.; Deb, D.; Chaudhuri, A.K.; Joseph, L.; Biswas, S. A Review Of Convolutional Neural Networks For Medical Image Analysis: Trends And Future Directions. J. Neonatal Surg. 2025, 14. [Google Scholar] [CrossRef]
- Chang, S.J.; Lee, J.Y.; Kim, S.I.; Lee, S.Y.; Lee, D.W. Simple Wearable Eye Tracker With Mini-Infrared Point Sensors. IEEE Access 2024, 12, 41019–41026. [Google Scholar] [CrossRef]
- Khaki, A.M.Z.; Choi, A. A real-time integrated eye tracker with in-pixel image processing in 0.18-μm CMOS technology. Integration 2025, 105, 102526. [Google Scholar] [CrossRef]
- Kuen, J.; Wang, Z.; Wang, G. Recurrent Attentional Networks for Saliency Detection. arXiv 2016, arXiv:1604.03227. [Google Scholar] [CrossRef]
- Li, G.; Yu, Y. Visual Saliency Based on Multiscale Deep Features. arXiv 2015, arXiv:1503.08663. [Google Scholar] [CrossRef]
- Boyd, A.; Bowyer, K.; Czajka, A. Human-Aided Saliency Maps Improve Generalization of Deep Learning. arXiv 2021, arXiv:2105.03492. [Google Scholar] [CrossRef]
- He, S.; Tavakoli, H.R.; Borji, A.; Mi, Y.; Pugeault, N. Understanding and Visualizing Deep Visual Saliency Models. arXiv 2019, arXiv:1903.02501. [Google Scholar] [CrossRef]
- Dobs, K.; Martinez, J.; Kell, A.J.; Kanwisher, N. Brain-like functional specialization emerges spontaneously in deep neural networks. Sci. Adv. 2022, 8, eabl8913. [Google Scholar] [CrossRef]
- Chefer, H.; Gur, S.; Wolf, L. Transformer interpretability beyond attention visualization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 782–791. [Google Scholar]
- Golan, T.; Raju, P.C.; Kriegeskorte, N. Controversial stimuli: Pitting neural networks against each other as models of human cognition. Proc. Natl. Acad. Sci. USA 2020, 117, 29330–29337. [Google Scholar] [CrossRef] [PubMed]
- Khosla, M.; Ngo, G.; Jamison, K.; Kuceyeski, A.; Sabuncu, M. Neural encoding with visual attention. Adv. Neural Inf. Process. Syst. 2020, 33, 15942–15953. [Google Scholar]
- Kubilius, J.; Schrimpf, M.; Kar, K.; Rajalingham, R.; Hong, H.; Majaj, N.; Issa, E.; Bashivan, P.; Prescott-Roy, J.; Schmidt, K.; et al. Brain-like object recognition with high-performing shallow recurrent ANNs. arXiv 2019, arXiv:1909.06161. [Google Scholar] [CrossRef]
- Shahriar, T. Comparative Analysis of Lightweight Deep Learning Models for Memory-Constrained Devices. arXiv 2025, arXiv:2505.03303. [Google Scholar] [CrossRef]
- Cichy, R.M.; Khosla, A.; Pantazis, D.; Torralba, A.; Oliva, A. Deep neural networks predict hierarchical spatio-temporal cortical dynamics of human visual object recognition. arXiv 2016, arXiv:1601.02970. [Google Scholar] [CrossRef]
- Alain, G.; Bengio, Y. Understanding intermediate layers using linear classifier probes. arXiv 2016, arXiv:1610.01644. [Google Scholar]
- Iqbal, S.; Qureshi, A.N.; Aurangzeb, K.; Alhussein, M.; Wang, S.; Anwar, M.S.; Khan, F. Hybrid parallel fuzzy CNN paradigm: Unmasking intricacies for accurate brain MRI insights. IEEE Trans. Fuzzy Syst. 2024, 32, 5533–5544. [Google Scholar] [CrossRef]
- Berrich, Y.; Guennoun, Z. EEG-based epilepsy detection using CNN-SVM and DNN-SVM with feature dimensionality reduction by PCA. Sci. Rep. 2025, 15, 14313. [Google Scholar] [CrossRef]
- Jafarbeglou, F.; Ahmadieh, H.; Soleimani, F.; Karimi, A.; Daftarian, N.; Fekri, S.; Motevasseli, T.; Naderan, M.; Kamali Doust Azad, B.; Sheikhtaheri, A.; et al. A deep learning model for diagnosis of inherited retinal diseases. Sci. Rep. 2025, 15, 22523. [Google Scholar] [CrossRef]
- Noda, Y.; Sakaue, K.; Wada, M.; Takano, M.; Nakajima, S. Development of artificial intelligence for determining major depressive disorder based on resting-state EEG and single-pulse transcranial magnetic stimulation-evoked EEG indices. J. Pers. Med. 2024, 14, 101. [Google Scholar] [CrossRef]
- Liaqat, H.; Parveen, A.; Kim, S.Y. Neuroprotective natural products’ regulatory effects on depression via gut–brain axis targeting tryptophan. Nutrients 2022, 14, 3270. [Google Scholar] [CrossRef]
- Gallucci, A.; Varoli, E.; Del Mauro, L.; Hassan, G.; Rovida, M.; Comanducci, A.; Casarotto, S.; Lo Re, V.; Romero Lauro, L.J. Multimodal approaches supporting the diagnosis, prognosis and investigation of neural correlates of disorders of consciousness: A systematic review. Eur. J. Neurosci. 2024, 59, 874–933. [Google Scholar] [CrossRef]
- Dentamaro, V.; Impedovo, D.; Musti, L.; Pirlo, G.; Taurisano, P. Enhancing early Parkinson’s disease detection through multimodal deep learning and explainable AI: Insights from the PPMI database. Sci. Rep. 2024, 14, 20941. [Google Scholar] [CrossRef]
- Rahman, A.U.; Ali, S.; Saqia, B.; Halim, Z.; Al-Khasawneh, M.; AlHammadi, D.A.; Khan, M.Z.; Ullah, I.; Alharbi, M. Alzheimer’s disease prediction using 3D-CNNs: Intelligent processing of neuroimaging data. SLAS Technol. 2025, 32, 100265. [Google Scholar] [CrossRef] [PubMed]
- Huh, Y.J.; Park, J.H.; Kim, Y.J.; Kim, K.G. Ensemble Learning-Based Alzheimer’s Disease Classification Using Electroencephalogram Signals and Clock Drawing Test Images. Sensors 2025, 25, 2881. [Google Scholar] [CrossRef] [PubMed]
- Yazdan, S.A.; Ahmad, R.; Iqbal, N.; Rizwan, A.; Khan, A.N.; Kim, D.H. An efficient multi-scale convolutional neural network based multi-class brain MRI classification for SaMD. Tomography 2022, 8, 1905–1927. [Google Scholar] [CrossRef] [PubMed]
- Shen, M.; Yang, F.; Wen, P.; Song, B.; Li, Y. A real-time epilepsy seizure detection approach based on EEG using short-time Fourier transform and Google-Net convolutional neural network. Heliyon 2024, 10, e31827. [Google Scholar] [CrossRef]
- Conte, G.M.; Weston, A.D.; Vogelsang, D.C.; Philbrick, K.A.; Cai, J.C.; Barbera, M.; Sanvito, F.; Lachance, D.H.; Jenkins, R.B.; Tobin, W.O.; et al. Generative adversarial networks to synthesize missing T1 and FLAIR MRI sequences for use in a multisequence brain tumor segmentation model. Radiology 2021, 299, 313–323. [Google Scholar] [CrossRef] [PubMed]
- Khan, M.A.; Jang, J.H.; Iqbal, N.; Jamil, H.; Naqvi, S.S.A.; Khan, S.; Kim, J.C.; Kim, D.H. Enhancing patient rehabilitation predictions with a hybrid anomaly detection model: Density-based clustering and interquartile range methods. CAAI Trans. Intell. Technol. 2025, 10, 983–1006. [Google Scholar] [CrossRef]
- Paek, S.H.; Cho, Z.H.; Son, Y.D.; Kwon, D.H.; PARK, S.B. Magnetic Resonance Imaging-Based Alternating Magnetic Field Treatment System. U.S. Patent Application No. 19/052776, 5 June 2025. [Google Scholar]
- Liu, Z.; Si, L.; Shi, S.; Li, J.; Zhu, J.; Lee, W.H.; Lo, S.L.; Yan, X.; Chen, B.; Fu, F.; et al. Classification of three anesthesia stages based on near-infrared spectroscopy signals. IEEE J. Biomed. Health Inform. 2024, 28, 5270–5279. [Google Scholar] [CrossRef]
- Mehmood, F.; Mumtaz, N.; Mehmood, A. Next-Generation Tools for Patient Care and Rehabilitation: A Review of Modern Innovations. Actuators 2025, 14, 133. [Google Scholar] [CrossRef]
- Ahmed, M.; Dar, A.R.; Helfert, M.; Khan, A.; Kim, J. Data provenance in healthcare: Approaches, challenges, and future directions. Sensors 2023, 23, 6495. [Google Scholar] [CrossRef] [PubMed]
- Khalil, I.; Mehmood, A.; Kim, H.; Kim, J. OCTNet: A modified multi-scale attention feature fusion network with InceptionV3 for retinal OCT image classification. Mathematics 2024, 12, 3003. [Google Scholar] [CrossRef]
- Zhu, M.; Quan, Y.; He, X. The classification of brain network for major depressive disorder patients based on deep graph convolutional neural network. Front. Hum. Neurosci. 2023, 17, 1094592. [Google Scholar] [CrossRef]
- Jiang, H.; Xu, J.; Shi, R.; Yang, K.; Zhang, D.; Gao, M.; Ma, H.; Qian, W. A multi-label deep learning model with interpretable Grad-CAM for diabetic retinopathy classification. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 1560–1563. [Google Scholar]
- Yang, J.; Wang, G.; Xiao, X.; Bao, M.; Tian, G. Explainable ensemble learning method for OCT detection with transfer learning. PLoS ONE 2024, 19, e0296175. [Google Scholar] [CrossRef] [PubMed]
- Lundervold, A.S.; Lundervold, A. An overview of deep learning in medical imaging focusing on MRI. Z. Med. Phys. 2019, 29, 102–127. [Google Scholar] [CrossRef]
- Yu, A.C.; Mohajer, B.; Eng, J. External validation of deep learning algorithms for radiologic diagnosis: A systematic review. Radiol. Artif. Intell. 2022, 4, e210064. [Google Scholar] [CrossRef] [PubMed]
- Faghani, S.; Moassefi, M.; Rouzrokh, P.; Khosravi, B.; Baffour, F.I.; Ringler, M.D.; Erickson, B.J. Quantifying uncertainty in deep learning of radiologic images. Radiology 2023, 308, e222217. [Google Scholar] [CrossRef]
- Roy, Y.; Banville, H.; Albuquerque, I.; Gramfort, A.; Falk, T.H.; Faubert, J. Deep learning-based electroencephalography analysis: A systematic review. J. Neural Eng. 2019, 16, 051001. [Google Scholar] [CrossRef]
- Ahmed, M.M.; Hossain, M.M.; Islam, M.R.; Ali, M.S.; Nafi, A.A.N.; Ahmed, M.F.; Ahmed, K.M.; Miah, M.S.; Rahman, M.M.; Niu, M.; et al. Brain tumor detection and classification in MRI using hybrid ViT and GRU model with explainable AI in Southern Bangladesh. Sci. Rep. 2024, 14, 22797. [Google Scholar] [CrossRef]
- Jiang, R.; Yin, X.; Yang, P.; Cheng, L.; Hu, J.; Yang, J.; Wang, Y.; Fu, X.; Shang, L.; Li, L.; et al. A transformer-based weakly supervised computational pathology method for clinical-grade diagnosis and molecular marker discovery of gliomas. Nat. Mach. Intell. 2024, 6, 876–891. [Google Scholar] [CrossRef]
- Evans, I.D.; Sharpley, C.F.; Bitsika, V.; Vessey, K.A.; Williams, R.J.; Jesulola, E.; Agnew, L.L. Differences in EEG Functional Connectivity in the Dorsal and Ventral Attentional and Salience Networks Across Multiple Subtypes of Depression. Appl. Sci. 2025, 15, 1459. [Google Scholar] [CrossRef]
- Kim, B.; Wattenberg, M.; Gilmer, J.; Cai, C.; Wexler, J.; Viegas, F.; Sayres, R. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In Proceedings of the International Conference on Machine Learning. PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 2668–2677. [Google Scholar]
- Li, L.; Li, J.; Chang, M.; Wu, X.; Yin, Z.; Chen, Z.; Wang, Z. The Causal Relationship Between Dietary Factors and the Risk of Intracranial Aneurysms: A Mendelian Randomization Analysis. Biomedicines 2025, 13, 533. [Google Scholar] [CrossRef] [PubMed]
- Khan, M.A.; Iqbal, N.; Jamil, H.; Qayyum, F.; Jang, J.H.; Khan, S.; Kim, J.C.; Kim, D.H. Enhanced abnormal data detection hybrid strategy based on heuristic and stochastic approaches for efficient patients rehabilitation. Future Gener. Comput. Syst. 2024, 154, 101–122. [Google Scholar] [CrossRef]
- Schug, D.; Yerramreddy, S.; Caruana, R.; Greenberg, C.; Zwolak, J. Explainable classification techniques for quantum dot device measurements. In Proceedings of the XAI4Sci: Explainable Machine Learning for Sciences Workshop (AAAI 2024), Vancouver, CA, Canada, 26 February 2024; pp. 1–6. [Google Scholar]
- Agarwal, C.; Queen, O.; Lakkaraju, H.; Zitnik, M. Evaluating explainability for graph neural networks. Sci. Data 2023, 10, 144. [Google Scholar] [CrossRef] [PubMed]
- Ou, J.; Li, N.; He, H.; He, J.; Zhang, L.; Jiang, N. Detecting muscle fatigue among community-dwelling senior adults with shape features of the probability density function of sEMG. J. Neuroeng. Rehabil. 2024, 21, 196. [Google Scholar] [CrossRef]
- Aïssa, B.; Sinopoli, A.; Ali, A.; Zakaria, Y.; Zekri, A.; Helal, M.; Nedil, M.; Rosei, F.; Mansour, S.; Mahmoud, K. Nanoelectromagnetic of a highly conductive 2D transition metal carbide (MXene)/Graphene nanoplatelets composite in the EHF M-band frequency. Carbon 2021, 173, 528–539. [Google Scholar] [CrossRef]
- Kumar, M.; Khan, L.; Choi, A. RAMHA: A Hybrid Social Text-Based Transformer with Adapter for Mental Health Emotion Classification. Mathematics 2025, 13, 2918. [Google Scholar] [CrossRef]
- Pulvermüller, F.; Tomasello, R.; Henningsen-Schomers, M.R.; Wennekers, T. Biological constraints on neural network models of cognitive function. Nat. Rev. Neurosci. 2021, 22, 488–502. [Google Scholar] [CrossRef]
- Esser-Skala, W.; Fortelny, N. Reliable interpretability of biology-inspired deep neural networks. NPJ Syst. Biol. Appl. 2023, 9, 50. [Google Scholar] [CrossRef]
- Ororbia, A.; Mali, A.; Kohan, A.; Millidge, B.; Salvatori, T. A review of neuroscience-inspired machine learning. arXiv 2024, arXiv:2403.18929. [Google Scholar] [CrossRef]
- Gütlin, D.; Auksztulewicz, R. Predictive Coding algorithms induce brain-like responses in Artificial Neural Networks. bioRxiv 2025. [Google Scholar] [CrossRef]
- Sourmpis, C.; Petersen, C.C.; Gerstner, W.; Bellec, G. Biologically informed cortical models predict optogenetic perturbations. bioRxiv 2024. [Google Scholar] [CrossRef]
- Duncan, A.G.; Mitchell, J.A.; Moses, A.M. Improving the performance of supervised deep learning for regulatory genomics using phylogenetic augmentation. Bioinformatics 2024, 40, btae190. [Google Scholar] [CrossRef]
- Oltramari, A.; Francis, J.; Henson, C.; Ma, K.; Wickramarachchi, R. Neuro-symbolic architectures for context understanding. arXiv 2020, arXiv:2003.04707. [Google Scholar] [CrossRef]
- Gandhirajan, S. Neurosymbolic Approaches for Knowledge-Infused Clinical Reasoning in Explainable Artificial Intelligence for Differential Diagnosis. J. Asian Sci. Res. (JASR) 2025, 15, 24–30. [Google Scholar]
- Kiruluta, A. A Novel Architecture for Symbolic Reasoning with Decision Trees and LLM Agents. arXiv 2025, arXiv:2508.05311. [Google Scholar] [CrossRef]
- Chevallier, S.; Carrara, I.; Aristimunha, B.; Guetschel, P.; Sedlar, S.; Lopes, B.; Velut, S.; Khazem, S.; Moreau, T. The largest EEG-based BCI reproducibility study for open science: The MOABB benchmark. arXiv 2024, arXiv:2404.15319. [Google Scholar] [CrossRef]
- Mumuni, F.; Mumuni, A. Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning. Cogn. Syst. Res. 2024, 84, 101188. [Google Scholar] [CrossRef]
- Nasarian, E.; Alizadehsani, R.; Acharya, U.R.; Tsui, K.L. Designing interpretable ML system to enhance trust in healthcare: A systematic review to proposed responsible clinician-AI-collaboration framework. Inf. Fusion 2024, 108, 102412. [Google Scholar] [CrossRef]
- Li, W.; Zhou, H.; Yu, J.; Song, Z.; Yang, W. Coupled mamba: Enhanced multimodal fusion with coupled state space model. Adv. Neural Inf. Process. Syst. 2024, 37, 59808–59832. [Google Scholar]
- Mehmood, F.; Ahmad, S.; Whangbo, T.K. Object detection based on deep learning techniques in resource-constrained environment for healthcare industry. In Proceedings of the 2022 International Conference on Electronics, Information, and Communication (ICEIC), Jeju, Republic of Korea, 6–9 February 2022; pp. 1–5. [Google Scholar]
- Lundervold, A.J. Precision neuropsychology in the area of AI. Front. Psychol. 2025, 16, 1537368. [Google Scholar] [CrossRef]
- Shmuel, A.; Glickman, O.; Lazebnik, T. A comprehensive benchmark of machine and deep learning across diverse tabular datasets. arXiv 2024, arXiv:2408.14817. [Google Scholar] [CrossRef]
- Cantero, D.; Esnaola-Gonzalez, I.; Miguel-Alonso, J.; Jauregi, E. Benchmarking object detection deep learning models in embedded devices. Sensors 2022, 22, 4205. [Google Scholar] [CrossRef]
- Jiao, M.; Xian, X.; Wang, B.; Zhang, Y.; Yang, S.; Chen, S.; Sun, H.; Liu, F. XDL-ESI: Electrophysiological sources imaging via explainable deep learning framework with validation on simultaneous EEG and IEEG. Neuroimage 2024, 299, 120802. [Google Scholar] [CrossRef] [PubMed]
- Canha, D.; Kubler, S.; Främling, K.; Fagherazzi, G. A Functionally-Grounded Benchmark Framework for XAI Methods: Insights and Foundations from a Systematic Literature Review. ACM Comput. Surv. 2025, 57, 1–40. [Google Scholar] [CrossRef]
- Pochelu, P. Deep learning inference frameworks benchmark. arXiv 2022, arXiv:2210.04323. [Google Scholar] [CrossRef]
Factor | Description |
---|---|
Uninterpretable architecture | Problem: Deep models (e.g., CNNs, transformers) consist of many non-linear layers, making it difficult to trace how inputs are transformed into outputs [10,11]. Possible solution: Leveraging XDL can make the learning process more transparent. A possible solution could be to include embedding explainable attention layers into the model. |
Lack of biological grounding | Problem: Most DL architectures are not inherently constrained by known neurobiological mechanisms, limiting their plausibility in neuroscience [30]. Possible solution: Tailoring the modeling phase utilizing the brain structures or pathways could guide the model toward more plausible interpretations. |
High-dimensional feature spaces | Problem: DL models often operate in complex and abstract feature spaces that are difficult to interpret intuitively. This requires significant efforts to align such models with the known cognitive functions [31]. Possible solution: Concept attribution techniques allows the linkage of abstract model features to more meaningful brain functions or behaviors. |
Overparameterization | Problem: Deep networks tend to have millions of parameters, increasing model complexity and obscuring the contribution of individual features or layers [32]. Possible solution: Tools like LRP enable model analysis to identify which parts actually matter the most, providing the opportunity to simplify the network. |
Non-transparent decision boundaries | Problem: Unlike rule-based or linear models, DL models do not offer clear decision logic, making it hard to validate predictions or diagnoses [8]. Possible solution: Applying concepts like rule-based models can act like interpretable stand-ins as they depict the logic behind outcomes. |
Limited explainability tools integration | Problem: Explainability techniques (e.g., saliency maps) are often added post hoc, and may not be tightly integrated into the model training process [33]. Possible solution: Including explainability in the training process can produce reliable and aligned results. |
Inadequate evaluation metrics | Problem: There is a lack of standardized benchmarks to quantitatively assess interpretability, making it difficult to compare or trust explanations [34]. Possible solution: Standards to evaluate the training performance must be established, as they will reshape the comparison criteria to be fair. It will also impose on the models to be responsible for how well various models are explainable in their outcomes. |
Modality | Temporal Resolution | Spatial Resolution | Invasiveness | Suitability for DL |
---|---|---|---|---|
EEG | High (1–5 ms) | Low (~11 mm) | Non-invasive | High |
MEG | High (<1 ms) | Medium (2–4 mm) | Non-invasive | Medium–High |
fMRI | Low (500–2000 ms) | High (1.5–3.5 mm) | Non-invasive | High |
ECoG | High (~1 ms) | High (1–4 mm) | Invasive (minimal) | High |
fNIRS | Medium (>100 ms) | Low (~10 mm) | Non-invasive | Medium |
Dataset (Modality) | Application Domain | Sample Size | General Remarks & Bias/Harmonization Notes |
---|---|---|---|
ADNI (MRI, PET) [57] | Alzheimer’s Disease diagnosis | 1000 subjects | Widely used, multi-modal neuroimaging.
Multi-site variability and intensitynormalization applied [69]. Recent works leverage MRI and PET for Alzheimer diagnosis [70]. |
OASIS-3 (MRI) [42] | Aging and dementia studies | 1098 subjects | Longitudinal, includes cognitive scores. Demographic imbalance (age/sex) andpreprocessing harmonized across scans [71]. |
HCP (fMRI, DWI, MEG) [60,72] | Brain connectivityand functional mapping | 1200 subjects | High-quality, useful for deep encoding. Scanner differences mitigated and standardized preprocessing pipelines [73]. Multimodal combinations (e.g., EEG and fMRI) explored in epilepsy studies [74]. |
CamCAN(MEG, MRI) [75] | Lifespan cognitive aging | 650 subjects | Broad age range, resting-state and task. Age distribution bias and artifact removal and normalization applied [76]. |
OpenNeuro(fMRI, EEG) [43] | Cognitive tasks and perception | 23 subjects (median)928 subjects (largest) | Open-access, highly diverse tasks. Small sample bias for some tasks and task standardization recommended [43]. |
BCI Competition IV(EEG) [61,77] | Motor imagery | 9 subjects | Standard benchmark for BCI modeling. Very small sample, class imbalance, and cross-subject normalization suggested [78]. |
TUH EEG Corpus [79] | Epilepsy detection | 24,000 records | Largest public clinical EEG dataset. Multi-center recordings, site-specific variations, and preprocessing standardization applied [80]. |
Sleep-EDF Expanded(EEG, EOG) [41] | Sleep staging | 197 records | Used for sleep research and BCI studies. Limited sample, demographic bias, artifact correction applied [81]. |
UK Biobank(MRI Genetics) [68,82] | Population-level analysis | 500,000 records | Imaging data with extensive phenotypic, genetic, and clinical metadata. Multimodal neuroimaging-genomics used for dementia prediction [83]. Site differences, large demographic variability, harmonized preprocessing and QC pipeline [84]. |
BraTS (MRI) [64,85] | Brain tumor segmentation | 800 cases | Challenge dataset, used in DL segmentation. Recent multimodal MRI+genomics frameworks extend brain tumor applications [86]. Multi-institutional variability, intensity normalization, and augmentation applied [87]. |
Category | Techniques | Interpretable | Application Areas | Limitations |
---|---|---|---|---|
Saliency [12] | LRP, Grad-CAM [94] | Medium | Visualizing model attention | Highlights correlation only, not causal |
Perbutation [95] | LIME, SHAP | High | Feature importance analysis | Computationally expensive, post-hoc |
Concept [96] | TCAV, ACE | High | Understanding concepts | Requires semantic definition |
Causal validation [97] | MR analysis, ablation, counterfactuals, interventions | High | Testing causal contribution of features | Complex to design, limited scalability |
Intrinsic XDL [98] | Attention mechanisms, interpretable architectures | High | Model transparency | Architectural limits accuracy |
Survey Focus | Methodology | Applications | Explainable? | Datasets | Novelty | Limitations | Our Review |
---|---|---|---|---|---|---|---|
Application of explainability in neurostimulation [99] | Literature-based conceptual overview, and no formal/PRISMA analysis | DBS, computational psychiatry, broader potential in other medical fields | Yes (mechanistic, functional) | neural, behavioral, imaging | Framework for integrating Explainability. Discussion on closed-loop behavioral modulation | Lacks structured evidence synthesis and fewer technical insights. Mainly focused on what is needed rather than solutions. | Provides structured and technical evidence, and covers visual & clinical domains. Addresses explainability trade-offs & need for benchmarks |
Focuses on explainability applied to cognitive functions and dysfunctions [100] | Joanna Briggs and PRISMA | Normal cognition and impaired cognition | Yes (good) | Anatomical datasets. Emphasis on correlation finding | Qualitative framework to evaluate explainability in cognitive neuroscience | Limited number of studies (12). Oversimplification | Larger evidence studies (177), discussed key issues, explainability trade-offs, and biological priors integration |
Broad application of DL in medical imaging [102] | Narrative review of 300+ studies | Cancer, radiology, pathology | No | Imaging datasets-MRI-CT-PET | Comprehensive early review of DL in medicine | Lack of interpretability. Minimal focus on neuroscience | Extends DL towards neuroscience with explainability |
DL in fMRI Neuroimaging [103] | Systematic DL function survey of neuroimaging | Cognitive neuroscience, brain state decoding | Limited | fMRI datasets | Highlights DL potential in functional brain mapping | Does not cover explainable AI approaches | Integrates multimodal datasets and explainable tools |
DL for structural and functional brain MRI [104] | Summarized DL pipelines | Alzheimer’s, Multiple sclerosis, tumor segmentation | No | MRI datasets | Overview of DL pipelines in neuroimaging | No interpretability focus. Clinical translation briefly discussed | Focuses on explainable clinical trust and pipelines |
DL across biomedical signals and health data [105] | Review over EEG, MRI, EHR, multimodal | EEG, electronic health records, wearable signals | Minimal | EEG, MRI, multimodal data | Broad biomedical scope including signals | Not neuroscience-specific. Lacks transparency discussion | Centers on neuroscience with interpretability emphasis |
DL applied to EEG data [106] | Focused review on DL models | BCI, seizure prediction | Limited | EEG datasets | Detailed DL coverage in EEG domain | No significant focus on XDL techniques | Part of broader multimodal explainable DL perspective |
CNN applications in neural imaging [107] | Overview of CNN methods | Object recognition, vision tasks, neuropathology | No | MRI, histology | Early CNN-centric neuroimaging focus | Excludes transformer and XDL methods | Covers CNN, RNN, Transformers with explainability |
Interpretability and explainability in healthcare AI [101] | Narrative and taxonomy of LIME, SHAP | Healthcare broadly, not neuroscience-specific | Yes | Clinical datasets (EHR, imaging) | Defines interpretability frameworks in AI | Not neuroscience-centric | Bridges XDL specifically into visual and clinical neuroscience |
Reference | DL Model Type | Visual Task | XDL Method | Outcome/Neuroscientific Insight |
---|---|---|---|---|
Kubilius [118] | CNN (CORnet-S) | Object recognition | Saliency maps, CAM | Layer-wise features map to ventral stream hierarchy. |
Selection criteria: Anatomical mapping and recurrence | ||||
Brain-data alignment: Composite Brain-Score (V4, IT, behavior) | ||||
Dobs [114] | CNN | Face detectionObject detection | Lesionexperiments | Networks spontaneously segregate face and object processing similar to brain functional specialization. Dual-task networks resemble human visual behavior. |
Selection criteria: Chosen to test dual-task demands and emergence of specialized subsystems | ||||
Brain-data alignment: Aligned CNN representations with fMRI/behavioral face–object patterns | ||||
Chefer [115] | Vision transformer | Imageclassification | Attentionvisualization | Attention heads correlate with semantic segment boundaries and objects. This aids in transformer interpretability. |
Selection criteria: Selected for benchmark performance and interpretability | ||||
Brain-data alignment: Aligned attention maps with cortical object-selective activations | ||||
Shahriar [119] | CNN (lightweight) | Imageclassification | Performancebenchmarking | Transfer learning enhances CNNs for memory-limited devices via accuracy-efficiency trade-offs. |
Selection criteria: Selected for low parameters and efficiency | ||||
Brain-data alignment: Feature similarity with early visual cortex | ||||
Cichy [120] | CNN (AlexNet) | Brain–DLmodel alignment | Layer-wise featureattribution | Temporal dynamics of brain activity matched CNN layer activation. |
Selection criteria: Pre-trained CNN on image recognition benchmarks | ||||
Brain-data alignment: RSA between CNN layers and MEG data | ||||
Golan [116] | CNN (VGG-16) | Visualcategorization | Integratedgradients | Feature importance maps show overlap with IT cortex response patterns. |
Selection criteria: High-performing CNN trained on object recognition | ||||
Brain-data alignment: Correlation of saliency maps with primate IT activity | ||||
Khosla [117] | Vision transformer | Image saliencyprediction | Self-attentionvisualization | Predictive saliency regions correspond with human fixation maps. |
Selection criteria: Alignment with human eye-tracking datasets | ||||
Brain-data alignment: Correlation with fixation distributions | ||||
Alain [121] | Shallow networks | Featurevisualization (class) | Featureimportance | Reveals shallow network focus on high-frequency, edge-related features. |
Selection criteria: Early convolutional layers with clear feature localization | ||||
Brain-data alignment: Feature activations compared with early visual cortex (V1). | ||||
Ramaswamy [96] | CNN and TCAV | Concept learningin vision tasks | Conceptactivation vectors | Extracted abstract concepts matched labeled categories, e.g., striped patterns. |
Selection criteria: Predefined visual concepts with clear semantic labels | ||||
Brain-data alignment: Learned concept vectors compared with labeled categories |
Reference | Clinical Domain | DL Model | XDL Method | Interpretability Outcome |
---|---|---|---|---|
Lundervold [143] | Alzheimer | CNN (3D-MRI) | Saliency map, LRP | Identified hippocampal and cortical atrophy as key-markers. |
External Validation: Reported performance drop on external datasets [144] | ||||
Calibration: Matched predicted with actual outcome rates [145] | ||||
Uncertainty Quantification: Aleatoric and epistemic uncertainties estimated [145] | ||||
Roy [146] | Epilepsy, EEG | RNN, 1D-CNN | Featureattribution | Revealed seizure-indicative frequency-bands. |
External Validation: Not performed | ||||
Calibration: Not specifically reported | ||||
Uncertainty Quantification: Not quantified | ||||
Zhang [141] | Retinal disease | RetinaNet, ResNet | Grad-CAM | Highlighted retinal areas most associated with diabetic retinopathy through visual mapping. |
External Validation: Tested on external datasets from EyePACS and Messidor-2 | ||||
Calibration: Not explicitly reported | ||||
Uncertainty Quantification: Not performed | ||||
He. [140] | Depression, fMRI | Deepgraph CNN | SortPoolinglayer | Achieved 72.1% classification accuracy on large multi-site MDD dataset. Improved interpretability by identifying salient brain regions through graph vertex sorting and functional connectivity patterns. |
External Validation: 10-fold cross-validation on large multi-site REST-meta-MDD dataset | ||||
Calibration: Not explicitly reported | ||||
Uncertainty Quantification: Not performed | ||||
Ahmed [147] | Brain tumor, MRI | ViT with GRU | Extracted essential features via ViT and captured feature relationships via GRU for classification. | |
External Validation: Validated on Brain Tumor Kaggle Dataset alongside primary MRI data from BSMMCH, Faridpur, Bangladesh | ||||
Calibration: Calibration achieved with cross-entropy loss and performance curves | ||||
Uncertainty Quantification: Uncertainty assessed through confidence scores and model robustness checks | ||||
Jiang [148] | Glioma Diagnosis, Molecular Marker Discovery | Transformer | Attention maps, Saliency maps, Feature attribution | Visual explanations of tumor regions and molecular marker associations. |
External Validation: Yes, tested on independent datasets (TCGA and Xiangya) | ||||
Calibration: Not explicitly reported | ||||
Uncertainty Quantification: Not explicitly reported |
Exposure | SNPs | Odds Ratio (95%CI) | Odds Ratio (95%CI) | p Value | Significant |
---|---|---|---|---|---|
Alcoholic drinks per week | 34 | 1.057 (0.788∼1.42) | 0.711 | No | |
Alcoholic intake frequency | 96 | 1.084 (0.909∼1.29) | 0.369 | No | |
Processed meat intake | 23 | 1.354 (0.55664∼3.29) | 0.504 | No | |
Poultry intake | 7 | 1.06 (0.157∼7.19) | 0.949 | No | |
Beef intake | 14 | 0.5945 (0.114∼3.1) | 0.537 | No | |
Non-oily fish intake | 11 | 0.72101 (0.156∼3.34) | 0.676 | No | |
Oily fish intake | 61 | 0.815 (0.4796∼1.39) | 0.45 | No | |
Pork intake | 14 | 4.375 (0.717∼26.7) | 0.11 | No | |
Lamb/mutton intake | 31 | 2.831 (1.0304∼7.78) | 0.044 | Yes | |
Bread intake | 30 | 1.097 (0.518∼2.32) | 0.81 | No | |
Cheese intake | 61 | 1.21 (0.697∼2.1) | 0.499 | No | |
Cooked vegetable intake | 17 | 0.66 (0.18357∼2.37) | 0.524 | No | |
Tea intake | 39 | 1.086 (0.66∼1.79) | 0.745 | No | |
Fresh fruit intake | 53 | 0.209 (0.0893∼0.491) | 0.000327 | Yes | |
Cereal intake | 38 | 0.591 (0.3001∼1.17) | 0.129 | No | |
Salad/raw vegetable intake | 19 | 0.807 (0.1841∼3.54) | 0.776 | No | |
Coffee intake | 38 | 1.149 (0.575∼2.3) | 0.694 | No | |
Dried fruit intake | 40 | 0.734 (0.3367∼1.6) | 0.437 | No | |
Research Direction | Objective | Primary Appl. Domain | Key Challenges |
---|---|---|---|
Integration ofbiological priors | Enhance model fidelity by embedding anatomical, physiological, or functional constraints into DL architectures | Visual attention modeling, brain-signal decoding, neural encoding | Balancing realism and flexibility. Limited availability of detailed neural priors |
Standardized frameworksfor XDL evaluation | Establish consistent benchmarks for assessing interpretability and trustworthiness of XDL models | Clinical diagnostics, regulatory validation, cross-study comparisons | Metric inconsistency. Data heterogeneity. Lack of community consensus |
Human-in-the-loop andneuro-symbolic systems | Combine expert knowledge with machine learning to enhance interpretability and domain relevance | Interactive BCIs, explainable clinicaldecision-support tools | Real-time integration, hybrid model design complexity |
Collaborativeinterdisciplinaryresearch models | Foster synergy among AI, neuroscience, clinical, and ethical domains for responsible innovation | Translational neuroscience, XDL healthcare deployment | Communication barriers. Differing research cultures and standards |
Identified Gap | Proposed Approach (References) | Potential Impact |
---|---|---|
Lack of standardized evaluation metrics for XDL (EEG-based BCI) in neuroscience. | Development of unified benchmark datasets and evaluation pipelines, acceleration of benchmarking practice adoption across the field [167]. | Ensures reproducibility and comparability across XDL pipelines. Standard reporting evaluation. metrics for BCI/EEG based XDL studies. |
Limited integration of biological priors into DL models. | Exploration of knowledge-informed and neurobiological-constrained cognitive systems (e.g., brain-inspired models, neural priors) [168]. | Improves biological plausibility and interpretability of models. |
Insufficient trust in clinical adoption of XDL models. | Human-in-the-loop systems, clinician-centered interpretability studies, regulatory frameworks to achieve safety and transparency [169]. | Builds trust, facilitates clinical integration, and ensures ethical deployment. |
Limited real-world applicability (multimodal) [70]. | Thorough real-world clinical evaluation with large multi-site longitudinal datasets. | Enhanced clinical decision-making reliability. |
Optimal integration challenge [74] (multimodal). | Advanced algorithm development. | Enhanced clinical insights. |
Limited interpretability transparency [86] (multimodal). | Integrate explainable AI techniques. | Improves clinical trust and decision-making. |
Limited clinical interpretability [128] (multimodal). | Integrate advanced explainable AI techniques for clearer clinical insights. | Improved diagnostic trust and clinical decision-making. |
Limited clinical explainability [83] (multimodal). | Integrate tailored imputation and nuanced interpretation | Enhance personalized and effective clinical decisions |
Lack of real-time, scalable XDL methods for large datasets. | Efficient model architectures for multimodal fusion (coupled SSM model) hardware-aware XDL techniques. This approach uses an inter-modal hidden states transition scheme for efficiency, and formulates a global convolution kernel for enabling parallelism [170]. | Enables deployment in time-sensitive neuroscience and clinical applications for provisioning real-time and scalable XDL models. Another study demonstrates deployable XDL for portable clinical neuroscience with limited resources, much needed for real-time and scalable systems [171]. |
Fragmented research between visual and clinical neuroscience. | Interdisciplinary frameworks combining cognitive neuroscience, clinical practice, personalized treatment through and XDL [172]. | Encourages holistic, cross-domain advancements for the XDL the neuroscience. |
Component | Purpose | Illustrative Examples |
---|---|---|
Task | Define evaluation objectives | Classification, prognosis. Benchmarks supervised learning tasks (classification, prognosis) using clinical/biomedical tabular datasets [173]. Evaluates AI benchmarking for predictive modeling (object detection/classification), applicable to clinical differentiation tasks [174]. |
Dataset | Ensure multimodal coverage | ADNI, UK Biobank, OpenNeuro. Demonstrates use of diverse biomedical datasets for benchmarking and generalization [173]. Uses electrophysiological datasets from multiple sources to ensure multimodal, multi-site evaluations [175]. |
Metrics | Quantify interpretability | Faithfulness, sensitivity, calibration, and clinical utility. Proposes a benchmarking framework for XAI, focusing on interpretability, faithfulness, and clinical relevance [176]. Benchmarks frameworks with a focus on accuracy, reliability, robustness, including calibration and sensitivity [177]. |
Governance | Promote reproducibility | Open code/data, reporting standards. Advocates open data/code, reproducible evaluation protocols, and reporting standards in benchmarking studies [173]. Discusses transparency, standardized evaluation, and community-driven checklists for reproducibility and governance [176]. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mehmood, A.; Mehmood, F.; Kim, J. Towards Explainable Deep Learning in Computational Neuroscience: Visual and Clinical Applications. Mathematics 2025, 13, 3286. https://doi.org/10.3390/math13203286
Mehmood A, Mehmood F, Kim J. Towards Explainable Deep Learning in Computational Neuroscience: Visual and Clinical Applications. Mathematics. 2025; 13(20):3286. https://doi.org/10.3390/math13203286
Chicago/Turabian StyleMehmood, Asif, Faisal Mehmood, and Jungsuk Kim. 2025. "Towards Explainable Deep Learning in Computational Neuroscience: Visual and Clinical Applications" Mathematics 13, no. 20: 3286. https://doi.org/10.3390/math13203286
APA StyleMehmood, A., Mehmood, F., & Kim, J. (2025). Towards Explainable Deep Learning in Computational Neuroscience: Visual and Clinical Applications. Mathematics, 13(20), 3286. https://doi.org/10.3390/math13203286