Next Article in Journal
Gamma-Polyglutamic Acid Reduces Heavy Metal Uptake and Stabilize Microbial Biosafety in Edible Mushroom Cultures
Previous Article in Journal
Influence of Tiered Cyclic Shear Stress on Shear Friction and Instability Behavior of Marble Specimens with the Fractures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3D Convolutional Neural Network Model for Detection of Major Depressive Disorder from Grey Matter Images

1
Department of Computer Science and Engineering, Sri Jayachamarajendra College of Engineering, JSS Science and Technology University, Mysuru 570006, Karnataka, India
2
Department of Information Science and Engineering, Sri Jayachamarajendra College of Engineering, JSS Science and Technology University, Mysuru 570006, Karnataka, India
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(19), 10312; https://doi.org/10.3390/app151910312
Submission received: 26 August 2025 / Revised: 16 September 2025 / Accepted: 19 September 2025 / Published: 23 September 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

Major depressive disorder is a mental health condition characterized by ongoing feelings of sadness, trouble focusing or making decisions, and a frequent sense of fatigue or hopelessness that lasts for a prolonged period. If left undiagnosed, it can have serious consequences, including suicide. This study proposes a 3D convolutional neural network model to detect major depressive disorder using 3D grey matter images from magnetic resonance imaging. The proposed 3D convolutional architecture comprises multiple hierarchical convolutional and pooling layers, designed to automatically learn spatial patterns from magnetic resonance imaging data. The model was optimized via Bayesian hyperparameter tuning, achieving an accuracy of 72.26%, an area under the receiver operating characteristic curve of 0.80, and an area under the precision–recall curve of 0.81 on a large multisite dataset comprising 1276 patients and 1104 healthy controls. Gradient-weighted class activation mapping is utilized to find brain regions associated with major depressive disorder. From this study, six regions were identified, namely, the frontal lobe, parietal lobe, temporal lobe, thalamus, insular cortex and corpus callosum which may be affected by major depressive disorder.

1. Introduction

Mental health is a vital part of overall well-being, influencing how one handles stress, relates to others, and makes choices in life. Major depressive disorder (MDD) is a serious mental health condition characterized by a consistently low mood, less interest in usual activities, and disturbances in daily functioning, such as altered sleep patterns, appetite changes, and reduced energy levels. MDD may affect anyone at any age, but it is usually prevalent during early adulthood. MDD affects women about twice as often as it does men and one in six people will experience it at some point in their lives [1]. Research shows that MDD is one of the main risk factors for suicide and about 60% of people who lose their life have a history of MDD [2]. Additionally, MDD is commonly co-morbid with physical illnesses like cancer, stroke, and acute coronary syndrome, which can negatively impact quality of life and increase mortality rates [3]. Given the association of MDD with high rates of mortality and comorbidity, it is important to detect it as early as possible. MDD diagnosis typically involves clinical evaluation, including interviews and questionnaires. The gold standard for diagnosing MDD is clinical interviews and rating scales like the Hamilton Depression Scale [4] and use of the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition (DSM-V) [5]. But these rely on subjective reports and clinician judgment, which can cause variability or delayed diagnoses. Alternatively, magnetic resonance imaging (MRI) can be used to distinguish affected individuals from healthy controls (HCs) [6]. Grey matter (GM), white matter (WM), and cerebrospinal fluid (CSF) are three essential elements of the brain that can be identified and extracted from MRI scans [7,8]. GM primarily consists of neural cell bodies, dendrites, and synapses, forming the outer layer of the brain. WM is made up of groups of nerve fibers covered in a fatty layer called myelin, which makes it look white. CSF is a clear liquid that encloses and protects the brain and spinal cord, filling the spaces within and around the central nervous system to act as a protective cushion. On MRI, these structures are differentiated based on their signal intensities [9]. Several MRI studies have found significant and region-specific changes in grey matter volume in MDD patients which correlate with MDD symptoms [10,11,12]. These structural changes serve as features for detection of MDD using 3D grey matter images.
Recent research employing machine learning and deep learning techniques has shown encouraging results in the identification of MDD. Deep learning models are increasingly used for classifying diseases based on MRI, as they can automatically extract features rather than relying on manually extracted features [13,14,15]. In traditional machine learning approaches like support vector machines and random forests, manually extracted features are used as input for classification. These approaches have demonstrated effectiveness in MDD classification but might be limited by their dependence on predefined feature extraction methods. A study by Wang et al. employed 3D DenseNet for classification of MDD and HCs, obtaining an accuracy of 77.42% [16]. Lin et al. in their late-life depression study on functional magnetic resonance imaging with 77 participants, using a 3D convolutional neural network (3D CNN), achieved an accuracy of 85% [17]. Li et al. utilized fMRI scans and employed a deep neural network to extract features, which were then classified using a kernel extreme learning machine, obtained a classification accuracy of 89% [18]. Chemin and colleagues used a 3D convolutional neural network on 83 resting state fMRI images to forecast suicidal tendencies in elderly individuals suffering from depression with an accuracy of 78.5% [19]. Belov et al. classified MDD from MRI data using a support vector machine algorithm and achieved an accuracy of 62% [20]. A study conducted by Alotaibi et al. employed a multi-atlas ensemble graph neural network for classification of MDD using functional MRI, obtaining an accuracy of 75.80% [21]. A study by Wu et al. employed a multi-task learning approach to simultaneously predict patient survival outcomes and performed semi-supervised segmentation of gliomas from brain MRI scans [22]. Yan et al. used deep learning signatures with diffusion tensor imaging data for improved glioma survival prediction [23]. Qu et al. carried out a study on MDD classification using MRI data, where they applied an attention-guided unified deep convolutional neural network and achieved an accuracy of 76.54% [24]. Qianqian et al. classified MDD by employing an adaptive multimodal neuroimage integration framework for classification using MRI and functional MRI images and obtained an accuracy of 65% [25].
A review of the literature shows that the majority of studies have employed two-dimensional CNN models for the classification of MDD. Since MRI is a 3D imaging modality, this study proposes an MDD-Net 3D convolutional neural network model to detect major depressive disorder from grey matter images. The 3D CNN model is capable of capturing spatial information from an image, allowing it to learn complex features relevant for classification. The proposed MDD-Net was trained on a large-scale multisite dataset comprising a total of 2380 subjects obtained from the REST-meta-MDD project of the DIRECT consortium [26]. Further, gradient-weighted class activation mapping was employed to identify the brain regions affected by major depressive disorder.
This paper is structured as follows: Section 2 provides a detailed description of the dataset and the proposed MDD-Net model developed specifically for the classification of major depressive disorder. Section 3 presents the experimental results and discusses the main findings. The main conclusions derived from the study are presented in Section 4.

2. Materials and Methods

2.1. Dataset

The dataset used in this study is obtained from the REST-meta-MDD project of the Depression Imaging REsearch ConsorTium (DIRECT) [26,27]. The dataset consists of a total of 2380 subjects from 17 hospitals, which include 1104 healthy controls and 1276 individuals with major depressive disorder. Table 1 presents the demographic information of the dataset used in this study.

2.2. Data Preprocessing

This study utilizes preprocessed data that are already available in the REST-meta-MDD project [27]. The project adopted a standardized preprocessing pipeline at all participating sites to minimize heterogeneity in preprocessing methods. An overview of the MRI preprocessing pipeline is shown in Figure 1. Structural T1-weighted MRI scans were acquired using standard imaging protocols and preprocessed with Processing Assistant for Resting-State fMRI, Statistical Parametric Mapping version 8 and the Voxel-Based Morphometry toolbox version 8. Each T1 image was segmented into grey matter, white matter and cerebrospinal fluid, followed by spatial normalization to the Montreal Neurological Institute (MNI) space using the Diffeomorphic Anatomical Registration Through Exponentiated Lie Algebra (DARTEL) algorithm. The normalized images were then smoothed with an 8 mm full-width at half-maximum (FWHM) Gaussian kernel to reduce noise and account for inter-individual variability. Quality control was conducted through assessment of sample homogeneity via covariance checks and visual inspection of single slices across all subjects. A total of 2380 preprocessed 3D grey matter images were obtained. These 3D grey matter images were then used as input to the proposed MDD-Net model for detecting major depressive disorder.

2.3. MDD-Net for Detection of Major Depressive Disorder

The overall framework for the detection of major depressive disorder is shown in Figure 2. Initially, 3D grey matter images are used as input to a 3D convolutional neural network model for the classification of MDD and HCs. The model consists of six convolutional blocks designed for hierarchical feature extraction, a 3D max-pooling layer for spatial downsampling, and a fully connected layer which integrates features captured across the entire input to produce the final classification. Feature maps from the trained model are used to visually highlight the specific brain regions using the gradient-weighted class activation mapping technique.

2.3.1. MDD-Net Model

The proposed MDD-Net model is fed with a 3D grey matter MRI image of dimension 121 × 145 × 121 (depth × height × width), allowing it to learn important features directly from the image. The model comprises six sequential convolutional blocks, each designed for hierarchical feature extraction. Each block is made up of two consecutive 3D convolutional layers, both using a 3 × 3 × 3 kernel with padding set to 1, and each layer is directly followed by a ReLU (Rectified Linear Unit) activation function. After these convolutional layers, a 3D max-pooling layer with a 2 × 2 × 2 kernel and a stride of 2 is applied to perform spatial downsampling, which reduces computational complexity and increases the receptive field in deeper layers. The network starts with 8 filters in its first convolutional layer and then doubles the number of filters in each new block, so the layers use 8, 16, 32, 64, 128, and finally, 256 filters. In the last two convolutional layers, 128 filters are used in block 5 and 256 filters are used in block 6 for experimental purposes. This gradual increase allows the network to capture more intricate feature representations as the depth of the model grows. Following the fourth convolutional block, the resulting 3D feature maps are flattened into a 28,224-size one-dimensional vector. This vector is passed into the classifier module, beginning with a fully connected layer that decreases the feature dimension to 128. A one-dimensional batch normalization layer is applied, followed by a ReLU activation and a dropout layer with a 0.7 dropout rate, to help prevent overfitting. After a fully connected layer reduces the feature size from 128 to 64, a ReLU activation function is employed to incorporate non-linearity into the model, enabling the network to more effectively learn and identify key patterns within the data. The proposed 3D CNN model is as shown in Figure 3. A detailed summary of each convolutional layer’s operations, output dimensions, filter counts and parameter counts is provided in Table 2. All parameters in the 3D CNN model are trainable and are updated during backpropagation.

2.3.2. Loss Function and Convergence

The classifier terminates in a fully connected layer that yields two raw output scores (logits) z = ( z 1 , z 2 ) for the two output classes. For training, the standard approach involves using the c { 1 , 2 } multi-class cross-entropy loss function, which typically combines a log-softmax operation with the negative log-likelihood loss. For a single instance with true class c (where c { 1 , 2 } ) and logits z, this loss is commonly formulated as:
L CE ( z , c ) = log e z c j = 1 2 e z j
where z c is the logit corresponding to the true (actual) class, e z c gives the exponential score for the true class and j = 1 2 e z j gives the sum of the exponentials for all classes.
With this direct formulation, training consistently stalled at L 0.69 . To restore learning, the logits were first transformed to class probabilities
p i = e z i j = 1 2 e z j , i = 1 , 2
where e z i is the exponential transformation of the logit for class, and j = 1 2 e z j is the normalization term (sum over exponentials of all logits), ensuring all output probabilities sum to 1. These probabilities are then passed to the CrossEntropyLoss,
L custom ( z , c ) = CrossEntropy ( Softmax ( z ) , c ) .
Empirically, utilizing this custom loss formulation enabled the training process to successfully converge, leading to the results reported in this study. While the standard cross entropy loss is generally preferred for its numerical stability and direct application to logits, our findings indicate that in this specific experimental context, the standard approach failed, whereas the custom loss formulation provided a path to effective model training. The CrossEntropyLoss internally performs a log-softmax, supplying the explicit soft-maxed outputs, resulting in reliable convergence.

2.3.3. Hyperparameter Optimization

To systematically determine the optimal model configurations and key hyperparameters, a hyperparameter optimization (HPO) based on Bayesian optimization is employed. Bayesian optimization is a probabilistic optimization technique designed to efficiently find the minimum or maximum of expensive, black-box functions, especially when evaluations are costly or noisy. It works by building a surrogate probabilistic model of the objective function and using an acquisition function to select the most promising next point to evaluate, balancing exploration and exploitation [28]. The main goal of HPO is to reduce the loss measured on the validation dataset. During HPO, the study explored several important hyperparameters, including model depth (comparing architectures with either four or six convolutional blocks), optimizer choice (evaluating both Adam and AdamW), initial learning rate (testing values of 0.01, 0.001, and 0.0001), and loss function formulation (contrasting the standard cross-entropy loss with a custom two-step loss, as described in Section 2.3.2). To maintain consistency and reliability throughout the experiments, specific parameters were kept constant in all runs. The original sequence is preserved across the convolutional layers by applying the rectified linear unit (ReLU) activation function and Kaiming initialization is used as a weight initialization method. A batch size of 16 is maintained for both training and validation, and all models were trained for a fixed duration of 50 epochs. Furthermore, to reduce the risk of overfitting, dropout layers are included in the fully connected head, with a dropout rate of 0.7 determined from the initial experimental results. This experimental setup provided a robust and reproducible framework for evaluating the performance of various 3D CNN configurations. Table 3 provides information about the hyperparameters used for training the final model.

3. Results

3.1. Experimental Results

The experiments were conducted on a large multisite dataset from the REST-meta-MDD project of the DIRECT consortium, consisting of 2380 subjects, including 1104 healthy controls and 1276 individuals with major depressive disorder. All experiments were conducted using PyTorch version 1.13.1 as the deep learning framework. The model training was performed on a computer equipped with a 13th Generation Intel Core i7-13620H processor (2.40 GHz), 16.0 GB of RAM, and an NVIDIA RTX 4050 GPU 6 GB VRAM. The dataset was divided into three separate subsets, namely, blind test, validation and training sets. A proportion of 10% was set aside as a blind test set, while 90% of the data were assigned at random for training. In our experiments, 10-fold cross-validation without stratification was used. The validation set plays a crucial role in the tuning of hyperparameters, helping to reduce the likelihood of overfitting and improving the overall performance of the model. After model training and hyperparameter tuning, the final model was evaluated using a blind test set that was kept separate from both the training and validation stages. This approach ensured an unbiased assessment of the model’s generalization capability. Model performance on the blind test set was quantified using standard metrics, such as accuracy, precision and recall, thereby providing a comprehensive evaluation of the model’s effectiveness on previously unseen data. The blind test set, which was separated from the original dataset prior to model training, was utilized exclusively for the evaluation of the final model. A variety of evaluation metrics were employed to rigorously assess the performance and robustness of the proposed model. Advanced performance measurements, including the area under the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AUPRC), were computed in addition to standard classification metrics like accuracy, sensitivity, specificity, and the F1-score. The models are designed to output non-normalized prediction scores (logits) for the two target classes, providing a direct measure of their discriminative capability. The loss function acted as the main optimization metric, indicating how effectively the model could differentiate between classes using its outputs.
A comprehensive comparison of selected model configurations, focusing on the effects of model depth and loss function choice, is presented in Table 4. Config-A achieves high sensitivity (0.93), meaning it is very effective at identifying positive cases, but low specificity (0.36), so it misclassifies negatives as positives, resulting in an accuracy of 66.95%. Config-B offers a much better balance, whereby it maintains a good sensitivity (0.83), improved specificity (0.59) and achieves the highest accuracy of 72.26%, along with the highest AUROC (0.80) and AUPRC (0.81) values, making it the most reliable configuration. Config-C, in contrast, has extremely high specificity (0.99) but very low sensitivity (0.13) and an accuracy of 52.38%, indicating it rarely identifies positives correctly and mostly predicts negative outcomes. Config-D provides high sensitivity (0.75) and specificity (0.67) as well as an accuracy of 71.43%. Config-E is similar to Config-D, with good specificity (0.76), lower sensitivity (0.67) and accuracy of 72.00%. Config-F, much like Config-C, has high specificity (0.98) but low sensitivity (0.13) and ends up with the lowest accuracies (53.78%). Among all the configurations, Config-B is found to be the best choice with AUROC of 0.80 (95% Confidence Interval: 0.748–0.843) and AUPRC of 0.81 (95% Confidence Interval: 0.761–0.861), as it offers the most balanced performance across key classification metrics, as shown in Table 5. These results underscore the model’s generalization ability and effectiveness in accurately classifying MDD subjects and healthy controls.
Additional experiments were conducted to compare the proposed 3D convolutional neural network model with the 3D-ResNet 50 model pretrained on medical images. During training, 3D-ResNet 50 exhibited significant convergence issues, with the loss function consistently plateauing around 0.693 across multiple epochs despite extensive hyperparameter tuning. Due to these convergence challenges, when adapting the pretrained model to our specific dataset characteristics, direct performance comparison was not feasible. However, the classification performance of our proposed 3D convolutional neural network model was evaluated against results from previous studies and a comparative summary is provided in Table 6. Yuqi et al. in their MDD classification study proposed a model dual-expert fMRI harmonization (DFH) and obtained an accuracy of 57.7% [29]. Wang et al. utilized 128 MRI images for classification of MDD using a multimodal fusion model that incorporates a CNN (high-frequency encoder based on SFCN), Transformer (low-frequency encoder) and multilayer perceptron along with ensemble learning, obtained an accuracy of 72.4% [30]. Lin et al. conducted a study involving 77 fMRI subjects to distinguish late-life depression from HCs, achieving an accuracy of 85% using a 3D CNN model [17]. Hong et al. utilized a dataset of 68 MRI scans to classify MDD using a 3D feature map reconstruction network built upon the ResNet architecture (3D FRN-ResNet), achieving an accuracy of 86.7% [31]. From Table 6, it is observed that the previous studies have used a lower number of subjects, which may limit the generalizability of the models. The proposed 3D convolutional neural network model was specifically trained from scratch (not fine-tuned from pretrained weights) on a large multisite dataset of 2380 subjects, allowing it to learn features directly relevant to distinguishing major depressive disorder patients from healthy controls. This dedicated training on MDD specific grey matter patterns, combined with the custom loss function that enabled successful convergence, allows the model to achieve its 72.26% classification accuracy. The proposed MDD-Net provides a generalizable solution for major depressive disorder classification, offering clear methodological and sample size advantages over previous studies.

3.2. Identification of Brain Regions Affected by Major Depressive Disorder

Gradient-weighted class activation mapping (Grad-CAM) is an explainable AI technique for visualizing class-specific activation maps by leveraging gradients to highlight important regions in the input image. This approach helps in explaining convolutional neural network model predictions [32,33]. Grad-CAM offers insights into how deep learning models process visual information and make decisions. It enhances trust and accountability in artificial intelligence systems by making their predictions more interpretable [34,35]. In this study, Grad-CAM is utilized on the final convolutional layer of the model to identify the brain regions associated with MDD. It overlays a heatmap on the image, using warm colors (reds and yellows) to mark the regions that contributed most to the model’s decision. Identification of these regions is accomplished by estimating the target prediction’s gradients with respect to the models final layer’s feature mappings. The feature maps are weighted according to the gradients, then combined and subjected to a ReLU activation that highlights only positive effects. The MDD brain regions identified by using Grad-CAM are the frontal lobe, temporal lobe, parietal lobe, thalamus, corpus callosum and insular cortex. Figure 4 and Figure 5 display the regions highlighted through the Grad-CAM analysis, revealing the regions that contributed most to the model’s decision. Figure 4a shows the parietal lobe and insular cortex regions and Figure 4b shows the temporal lobe. Figure 5a shows the thalamus and corpus callosum regions and Figure 5b shows the frontal lobe. Table 7 lists the identified brain regions affected by major depressive disorder along with their corresponding functions.
The frontal lobe regions identified from this study are responsible for for a wide range of essential brain functions, particularly those related to higher-order cognition, behavior, and motor control [36]. Numerous studies show that the prefrontal cortex, including regions like the dorsolateral prefrontal cortex, orbitofrontal cortex, and subgenual prefrontal cortex, exhibit a reduction in grey matter volume among patients suffering from MDD [37,38]. These reductions are associated with impaired emotional regulation, decision-making, and motivation. The temporal lobe regions identified in this study are responsible for processing auditory information, language comprehension, memory formation and retrieval, emotion regulation and visual object recognition [39]. Several MRI studies have indicated significant reductions in grey matter volume in the temporal lobes, particularly the hippocampus and amygdala regions which are often associated with mood regulation [40,41]. The parietal lobe identified in this study is primarily responsible for processing sensory information and supporting spatial awareness, coordination, math skills, and language comprehension [42]. Chen et al. reported a reduced magnetization transfer ratio in the left superior parietal lobule of MDD patients, potentially linked to cognitive and attentional impairments [43]. According to a review by Zhang et al., MDD causes structural and functional alterations in the parietal lobe, which hampers learning and information processing [38]. Kili et al. reported reduced parietal lobe volume in MDD patients compared to healthy individuals, linking it to cognitive deficits like poor attention and working memory common in mood disorders [44].
The thalamus region identified in this study plays a key role in relaying motor and sensory signals to the cortex and is involved in regulating sleep, alertness, emotion, memory and attention [45,46]. Several studies have examined thalamic volume deficits using MRI and found that thalamic atrophy could be a structural marker in MDD [47,48]. Further, the findings highlighted altered thalamic activity and connections with other brain regions involved in mood regulation. The corpus callosum identified as a key region in this study is responsible for linking the left and right hemispheres of the brain, facilitating communication between them and coordinating motor, sensory, and cognitive activities [49]. A study by Lee et al. indicated that the corpus callosum, a significant brain structure linking the two hemispheres, exhibits changes in individuals with MDD, which may reflect brain dysfunction associated with the disorder [50]. Piras et al. reported a study that identified structural alterations in the corpus callosum in individuals with MDD, indicating notable thinning or atrophy, especially among those with chronic forms of the condition [51]. The insular cortex region identified in this study integrates sensory, emotional and autonomic information, which is crucial for self-awareness, cognitive control and autonomic regulation [52]. Multiple studies have identified reduced functional connectivity between the left insula and the amygdala, which may disrupt mood regulation [53,54]. According to a study by Takahashi et al. changes in the insular cortex’s general structure may increase the likelihood that MDD patients develop neurological defects and might also make their symptoms more severe [55].
Grad-CAM analysis was also conducted to provide a direct visual comparison of brain regions between MDD and healthy controls. The healthy control brain regions identified by using Grad-CAM are the parietal lobe, and the thalamocapsular and insular cortex regions. Figure 6 shows the brain regions highlighted using Grad-CAM for healthy controls. Figure 6a shows the parietal lobe and thalamocapsular regions. Figure 6b shows the parietal lobe and insular cortex regions. The parietal lobe in healthy controls is mainly responsible for integrating sensory information, supporting spatial awareness and movement coordination, while the thalamocapsular region functions as a relay center for sensory and motor signals, which plays a role in maintaining alertness and consciousness. The insular cortex region is responsible for bodily sensations, emotional awareness and social experiences in healthy controls.
The findings from the study indicate that the brain regions affected in MDD include the frontal lobe, temporal lobe, parietal lobe, thalamus, corpus callosum and insular cortex. These brain regions play a role in thinking, emotional regulation, memory and how different parts of the brain work together. Problems in the frontal lobe can affect decision-making and mood control, while changes in the parietal lobe can negatively affect attention and planning. The temporal lobe is linked to memory and feelings, and the thalamus helps pass information around the brain. The corpus callosum connects the two halves of the brain and the insular cortex helps a person feel emotions. Identifying brain regions which are affected in major depressive disorder enables clinicians to diagnose the condition more confidently, anticipate its progression and provide effective treatments.

4. Conclusions

The MDD-Net is proposed in this study to classify individuals with major depressive disorder from healthy controls using magnetic resonance imaging data. The model is trained on 3D grey matter MRI images for classification of major depressive disorder, achieving an accuracy of 72.26%. In addition, the proposed 3D convolutional neural network model attained an area under the receiver operating characteristic curve of 0.80 (95% Confidence Interval: 0.748–0.843), indicating strong discriminative ability and area under the precision–recall curve of 0.81 (95% Confidence Interval: 0.761–0.861), further highlighting the model’s reliability in identifying true positive cases of major depressive disorder. By employing gradient-weighted class activation mapping, the study identified six brain regions, namely, the frontal lobe, temporal lobe, parietal lobe, thalamus, corpus callosum and insular cortex which may be affected by major depressive disorder. The functions of these brain regions including self-awareness, emotional regulation and cognitive processing are mainly associated with major depressive disorder. The findings from the study demonstrate the effectiveness of the proposed 3D convolutional neural network model for major depressive disorder detection and highlight its potential in clinical decision support.

Author Contributions

Conceptualization, B.S.M. and B.A.R.; methodology, B.A.R. and A.A.; resources, DIRECT Consortium; data analysis, B.A.R. and A.A.; writing—original draft preparation, B.A.R. and A.A.; writing—review and editing, B.S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available at https://rfmri.org/REST-meta-MDD (accessed on 21 January 2020).

Acknowledgments

The authors are thankful to the DIRECT consortium and the REST-meta-MDD project for providing the data. The authors are indebted to Jain Anubha Rakesh and Aditya Shirsat, Department of Radiology, Bharati Vidyapeeth Medical College, Sangli, Maharashtra for validating the brain regions identified in this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MRIMagnetic resonance imaging
MDDMajor depressive disorder
HCsHealthy controls
3D CNN3D convolutional neural network
GMGrey matter
Grad-CAMGradient-weighted class activation mapping

References

  1. Otte, C.; Gold, S.M.; Penninx, B.W.; Pariante, C.M.; Etkin, A.; Fava, M.; Mohr, D.C.; Schatzberg, A.F. Major depressive disorder. Nat. Rev. Dis. Prim. 2016, 2, 16065. [Google Scholar] [CrossRef]
  2. Ren, Y.; Li, M.; Yang, C.; Jiang, W.; Wu, H.; Pan, R.; Yang, Z.; Wang, X.; Wang, W.; Wang, W.; et al. Suicidal risk is associated with hyper-connections in the frontal-parietal network in patients with depression. Transl. Psychiatry 2025, 15, 49. [Google Scholar] [CrossRef]
  3. Arnaud, A.; Brister, T.; Duckworth, K.; Foxworth, P.; Fulwider, T.; Suthoff, E.; Werneburg, B.; Aleksanderek, I.; Reinhart, M. Impact of Major Depressive Disorder on Comorbidities: A Systematic Literature Review. J. Clin. Psychiatry 2022, 83, 21r14328. [Google Scholar] [CrossRef]
  4. Reynolds, W.M.; Kobak, K.A. Reliability and validity of the Hamilton Depression Inventory: A paper-and-pencil version of the Hamilton Depression Rating Scale clinical interview. Psychol. Assess. 1995, 7, 472–483. [Google Scholar] [CrossRef]
  5. First, M.B.; Gaebel, W.; Maj, M.; Stein, D.J.; Kogan, C.S.; Saunders, J.B.; Poznyak, V.B.; Gureje, O.; Lewis-Fernández, R.; Maercker, A.; et al. An organization-and category-level comparison of diagnostic requirements for mental disorders in ICD-11 and DSM-5. World Psychiatry 2021, 20, 34–51. [Google Scholar] [CrossRef]
  6. Qiu, H.; Li, J. Major depressive disorder and magnetic resonance imaging: A mini-review of recent progress. Curr. Pharm. Des. 2018, 24, 2524–2529. [Google Scholar] [CrossRef] [PubMed]
  7. Thacker, N.; Jackson, A. Mathematical segmentation of grey matter, white matter and cerebral spinal fluid from MR image pairs. Br. J. Radiol. 2001, 74, 234–242. [Google Scholar] [CrossRef]
  8. Bedell, B.J.; Narayana, P.A. Volumetric analysis of white matter, gray matter, and CSF using fractional volume analysis. Magn. Reson. Med. 1998, 39, 961–969. [Google Scholar] [CrossRef]
  9. Girbau-Massana, D.; Garcia-Marti, G.; Marti-Bonmati, L.; Schwartz, R.G. Gray–white matter and cerebrospinal fluid volume differences in children with specific language impairment and/or reading disability. Neuropsychologia 2014, 56, 90–100. [Google Scholar] [CrossRef] [PubMed]
  10. Zheng, R.; Zhang, Y.; Yang, Z.; Han, S.; Cheng, J. Reduced Brain Gray Matter Volume in Patients with First-Episode Major Depressive Disorder: A Quantitative Meta-Analysis. Front. Psychiatry 2021, 12, 671348. [Google Scholar] [CrossRef]
  11. Shad, M.U.; Muddasani, S.; Rao, U. Gray matter differences between healthy and depressed adolescents: A voxel-based morphometry study. J. Child Adolesc. Psychopharmacol. 2012, 22, 190–197. [Google Scholar] [CrossRef]
  12. Romeo, Z.; Biondi, M.; Oltedal, L.; Spironelli, C. The Dark and Gloomy Brain: Grey Matter Volume Alterations in Major Depressive Disorder–Fine-Grained Meta-Analyses. Depress. Anxiety 2024, 2024, 6673522. [Google Scholar] [CrossRef]
  13. Plis, S.; Hjelm, D.; Salakhutdinov, R.; Allen, E.; Bockholt, H.; Long, J.; Johnson, H.; Paulsen, J.; Turner, J.; Calhoun, V. Deep learning for neuroimaging: A validation study. Front. Neurosci. 2014, 8, 229. [Google Scholar] [CrossRef]
  14. Avberšek, L.; Repovš, G. Deep learning in neuroimaging data analysis: Applications, challenges, and solutions. Front. Neuroimaging 2022, 1, 981642. [Google Scholar] [CrossRef]
  15. Smucny, J.; Shi, G.; Davidson, I. Deep Learning in Neuroimaging: Overcoming Challenges with Emerging Approaches. Front. Psychiatry 2022, 13, 912600. [Google Scholar] [CrossRef]
  16. Wang, Y.; Gong, N.; Fu, C. Major depression disorder diagnosis and analysis based on structural magnetic resonance imaging and deep learning. J. Integr. Neurosci. 2021, 20, 977–984. [Google Scholar] [CrossRef]
  17. Lin, C.; Lee, S.H.; Huang, C.M.; Chen, G.Y.; Chang, W.; Liu, H.L.; Ng, S.H.; Lee, T.M.C.; Wu, S.C. Automatic diagnosis of late-life depression by 3D convolutional neural networks and cross-sample Entropy analysis from resting-state fMRI. Brain Imaging Behav. 2023, 17, 125–135. [Google Scholar] [CrossRef]
  18. Li, M.; Liu, M.; Kang, J.; Zhang, W.; Lu, S. Depression recognition method based on regional homogeneity features from emotional response fMRI using deep convolutional neural network. In Proceedings of the 2021 3rd International Conference on Intelligent Medicine and Image Processing, Tianjin, China, 23–26 April 2021; pp. 45–49. [Google Scholar]
  19. Lin, C.; Huang, C.; Chang, W.; Chang, Y.; Liu, H.; Ng, S.; Lin, H.; Lee, T.; Lee, S.; Wu, S. Predicting suicidality in late-life depression by 3D convolutional neural network and cross-sample entropy analysis of resting-state fMRI. Brain Behav. 2024, 14, e3348. [Google Scholar] [CrossRef]
  20. Belov, V.; Erwin-Grabner, T.; Aghajani, M.; Aleman, A.; Amod, A.R.; Basgoze, Z.; Benedetti, F.; Besteher, B.; Bülow, R.; Ching, C.R.; et al. Multi-site benchmark classification of major depressive disorder using machine learning on cortical and subcortical measures. Sci. Rep. 2024, 14, 1084. [Google Scholar] [CrossRef]
  21. Alotaibi, N.M.; Alhothali, A.M.; Ali, M.S. Multi-atlas ensemble graph neural network model for major depressive disorder detection using functional MRI data. Front. Comput. Neurosci. 2025, 19, 1537284. [Google Scholar] [CrossRef]
  22. Wu, W.; Yan, J.; Zhao, Y.; Sun, Q.; Zhang, H.; Cheng, J.; Liang, D.; Chen, Y.; Zhang, Z.; Li, Z.C. Multi-task learning for concurrent survival prediction and semi-supervised segmentation of gliomas in brain MRI. Displays 2023, 78, 102402. [Google Scholar] [CrossRef]
  23. Yan, J.; Zhao, Y.; Chen, Y.; Wang, W.; Duan, W.; Wang, L.; Zhang, S.; Ding, T.; Liu, L.; Sun, Q.; et al. Deep learning features from diffusion tensor imaging improve glioma stratification and identify risk groups with distinct molecular pathway activities. eBioMedicine 2021, 72, 103583. [Google Scholar] [CrossRef]
  24. Qu, X.; Xiong, Y.; Zhai, K.; Yang, X.; Yang, J. An Efficient Attention-Based Network for Screening Major Depressive Disorder with sMRI. In Proceedings of the 2023 29th International Conference on Mechatronics and Machine Vision in Practice (M2VIP), Queenstown, New Zealand, 21–24 November 2023; pp. 1–6. [Google Scholar]
  25. Wang, Q.; Li, L.; Qiao, L.; Liu, M. Adaptive multimodal neuroimage integration for major depression disorder detection. Front. Neuroinform. 2022, 16, 856175. [Google Scholar] [CrossRef]
  26. Yan, C.G.; Chen, X.; Li, L.; Castellanos, F.X.; Bai, T.J.; Bo, Q.J.; Cao, J.; Chen, G.M.; Chen, N.X.; Chen, W.; et al. Reduced default mode network functional connectivity in patients with recurrent major depressive disorder. Proc. Natl. Acad. Sci. USA 2019, 116, 9078–9083. [Google Scholar] [CrossRef]
  27. Chen, X.; Lu, B.; Li, H.X.; Li, X.Y.; Wang, Y.W.; Castellanos, F.X.; Cao, L.P.; Chen, N.X.; Chen, W.; Cheng, Y.Q.; et al. The DIRECT consortium and the REST-meta-MDD project: Towards neuroimaging biomarkers of major depressive disorder. Psychoradiology 2022, 2, 32–42. [Google Scholar] [CrossRef]
  28. Shakya, A.; Biswas, M.; Pal, M. Classification of Radar data using Bayesian optimized two-dimensional Convolutional Neural Network. In Radar Remote Sensing; Elsevier: Amsterdam, The Netherlands, 2022; pp. 175–186. [Google Scholar]
  29. Fang, Y.; Potter, G.G.; Wu, D.; Zhu, H.; Liu, M. Addressing multi-site functional MRI heterogeneity through dual-expert collaborative learning for brain disease identification. Hum. Brain Mapp. 2023, 44, 4256–4271. [Google Scholar] [CrossRef]
  30. Wang, J.; Li, T.; Sun, Q.; Guo, Y.; Yu, J.; Yao, Z.; Hou, N.; Hu, B. Automatic Diagnosis of Major Depressive Disorder Using a High-and Low-Frequency Feature Fusion Framework. Brain Sci. 2023, 13, 1590. [Google Scholar] [CrossRef]
  31. Hong, J.; Huang, Y.; Ye, J.; Wang, J.; Xu, X.; Wu, Y.; Li, Y.; Zhao, J.; Li, R.; Kang, J.; et al. 3D FRN-ResNet: An Automated Major Depressive Disorder Structural Magnetic Resonance Imaging Data Identification Framework. Front. Aging Neurosci. 2022, 14, 912283. [Google Scholar] [CrossRef]
  32. Ramprasaath, R.; Selvaraju, M.; Das, A. Visual Explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 618–626. [Google Scholar]
  33. Samek, W.; Montavon, G.; Vedaldi, A.; Hansen, L.K.; Müller, K.R. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning; Springer Nature: Berlin/Heidelberg, Germany, 2019; Volume 11700. [Google Scholar]
  34. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Computer Vision–ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part I 13; Springer: Berlin/Heidelberg, Germany, 2014; pp. 818–833. [Google Scholar]
  35. Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K.R.; Samek, W. Layer-wise relevance propagation for neural networks with local renormalization layers. In Artificial Neural Networks and Machine Learning–ICANN 2016, Proceedings of the 25th International Conference on Artificial Neural Networks, Barcelona, Spain, 6–9 September 2016; Proceedings, Part II 25; Springer: Berlin/Heidelberg, Germany, 2016; pp. 63–71. [Google Scholar]
  36. Chayer, C.; Freedman, M. Frontal lobe functions. Curr. Neurol. Neurosci. Rep. 2001, 1, 547–552. [Google Scholar] [CrossRef]
  37. Zhuo, C.; Li, G.; Lin, X.; Jiang, D.; Xu, Y.; Tian, H.; Wang, W.; Song, X. The rise and fall of MRI studies in major depressive disorder. Transl. Psychiatry 2019, 9, 335. [Google Scholar] [CrossRef]
  38. Zhang, F.F.; Peng, W.; Sweeney, J.A.; Jia, Z.Y.; Gong, Q.Y. Brain structure alterations in depression: Psychoradiological evidence. CNS Neurosci. Ther. 2018, 24, 994–1003. [Google Scholar] [CrossRef]
  39. Milner, B. Intellectual function of the temporal lobes. Psychol. Bull. 1954, 51, 42. [Google Scholar] [CrossRef]
  40. Lai, C.H. Gray matter volume in major depressive disorder: A meta-analysis of voxel-based morphometry studies. Psychiatry Res. Neuroimaging 2013, 211, 37–46. [Google Scholar] [CrossRef]
  41. Lorenzetti, V.; Allen, N.B.; Fornito, A.; Yücel, M. Structural brain abnormalities in major depressive disorder: A selective review of recent MRI studies. J. Affect. Disord. 2009, 117, 1–17. [Google Scholar] [CrossRef] [PubMed]
  42. Coslett, H.B.; Schwartz, M.F. The parietal lobe and language. Handb. Clin. Neurol. 2018, 151, 365–375. [Google Scholar]
  43. Chen, Z.; Peng, W.; Sun, H.; Kuang, W.; Li, W.; Jia, Z.; Gong, Q. High-field magnetic resonance imaging of structural alterations in first-episode, drug-naive patients with major depressive disorder. Transl. Psychiatry 2016, 6, e942. [Google Scholar] [CrossRef] [PubMed]
  44. Kilic, F.; Kartal, F.; Erbay, M.F.; Karlidağ, R. Parietal Cortex Volume and Functions in Major Depression and Bipolar Disorder: A Cloud-Based Magnetic Resonans Imaging Study. Arch. Neuropsychiatry 2025, 61, 47. [Google Scholar]
  45. Klostermann, F.; Krugel, L.; Ehlen, F. Functional roles of the thalamus for language capacities. Front. Syst. Neurosci. 2013, 7, 32. [Google Scholar] [CrossRef]
  46. Ward, L.M. The thalamus: Gateway to the mind. Wiley Interdiscip. Rev. Cogn. Sci. 2013, 4, 609–622. [Google Scholar] [CrossRef]
  47. Zhang, K.; Zhu, Y.; Zhu, Y.; Wu, S.; Liu, H.; Zhang, W.; Xu, C.; Zhang, H.; Hayashi, T.; Tian, M. Molecular, functional, and structural imaging of major depressive disorder. Neurosci. Bull. 2016, 32, 273–285. [Google Scholar] [CrossRef]
  48. Nugent, A.C.; Davis, R.M.; Zarate, C.A., Jr.; Drevets, W.C. Reduced thalamic volumes in major depressive disorder. Psychiatry Res. Neuroimaging 2013, 213, 179–185. [Google Scholar] [CrossRef]
  49. Hinkley, L.; Marco, E.; Findlay, A.; Honma, S.; Jeremy, R.; Strominger, Z.; Bukshpun, P.; Wakahiro, M.; Brown, W.; Paul, L.; et al. The role of corpus callosum development in functional connectivity and cognitive processing. PLoS ONE 2012, 7, e39804. [Google Scholar] [CrossRef]
  50. Lee, S.; Pyun, S.; Choi, K.; Tae, W. Shape and Volumetric Differences in the Corpus Callosum Between Patients with Major Depressive Disorder and Healthy Controls. Psychiatry Investig. 2020, 17, 941–950. [Google Scholar] [CrossRef]
  51. Piras, F.; Vecchio, D.; Kurth, F.; Piras, F.; Banaj, N.; Ciullo, V.; Luders, E.; Spalletta, G. Corpus callosum morphology in major mental disorders: A magnetic resonance imaging study. Brain Commun. 2021, 3, fcab100. [Google Scholar] [CrossRef]
  52. Shura, R.D.; Hurley, R.A.; Taber, K.H. Insular cortex: Structural and functional neuroanatomy. J. Neuropsychiatry Clin. Neurosci. 2014, 26, 276–282. [Google Scholar] [CrossRef]
  53. Sliz, D.; Hayley, S. Major depressive disorder and alterations in insular cortical activity: A review of current functional magnetic imaging research. Front. Hum. Neurosci. 2012, 6, 323. [Google Scholar] [CrossRef]
  54. Peng, X.; Lin, P.; Wu, X.; Gong, R.; Yang, R.; Wang, J. Insular subdivisions functional connectivity dysfunction within major depressive disorder. J. Affect. Disord. 2018, 227, 280–288. [Google Scholar] [CrossRef]
  55. Takahashi, T.; Sasabayashi, D.; Yücel, M.; Whittle, S.; Suzuki, M.; Pantelis, C.; Malhi, G.S.; Allen, N.B. Gross anatomical features of the insular cortex in affective disorders. Front. Psychiatry 2024, 15, 1482990. [Google Scholar] [CrossRef]
Figure 1. MRI preprocessing pipeline.
Figure 1. MRI preprocessing pipeline.
Applsci 15 10312 g001
Figure 2. Overall framework for the detection of major depressive disorder.
Figure 2. Overall framework for the detection of major depressive disorder.
Applsci 15 10312 g002
Figure 3. Proposed MDD-Net model.
Figure 3. Proposed MDD-Net model.
Applsci 15 10312 g003
Figure 4. Grad-CAM identified brain regions - Parietal lobe, insular cortex and temporal lobe.
Figure 4. Grad-CAM identified brain regions - Parietal lobe, insular cortex and temporal lobe.
Applsci 15 10312 g004
Figure 5. Grad-CAM identified brain regions - Thalamus, corpus callosum and frontal lobe.
Figure 5. Grad-CAM identified brain regions - Thalamus, corpus callosum and frontal lobe.
Applsci 15 10312 g005
Figure 6. Grad-CAM identified brain regions for healthy controls - Parietal lobe, thalamocapsular and insular cortex.
Figure 6. Grad-CAM identified brain regions for healthy controls - Parietal lobe, thalamocapsular and insular cortex.
Applsci 15 10312 g006
Table 1. Demographic information of the dataset used in the study.
Table 1. Demographic information of the dataset used in the study.
MDDHealthy Controls
No. of Subjects12761104
Male463462
Female813642
 Average Age36 years
Range (14–80)
36 years
Range (12–82)
Table 2. The parameters used in the MDD-Net model.
Table 2. The parameters used in the MDD-Net model.
Layer (Type)Output ShapeParameters
Input Layer(B, 1, 121, 145, 121)0
Block 1
Conv3D (8 filters, 3 × 3 × 3)(B, 8, 121, 145, 121)224 (216 + 8)
Conv3D (8 filters, 3 × 3 × 3)(B, 8, 121, 145, 121)1736 (1728 + 8)
MaxPool3D (2 × 2 × 2)(B, 8, 60, 72, 60)0
Block 2
Conv3D (16 filters, 3 × 3 × 3)(B, 16, 60, 72, 60)3472 (3456 + 16)
Conv3D (16 filters, 3 × 3 × 3)(B, 16, 60, 72, 60)6928 (6912 + 16)
MaxPool3D (2 × 2 × 2)(B, 16, 30, 36, 30)0
Block 3
Conv3D (32 filters, 3 × 3 × 3)(B, 32, 30, 36, 30)13,856 (13,824 + 32)
Conv3D (32 filters, 3 × 3 × 3)(B, 32, 30, 36, 30)27,680 (27,648 + 32)
MaxPool3D (2 × 2 × 2)(B, 32, 15, 18, 15)0
Block 4
Conv3D (64 filters, 3 × 3 × 3)(B, 64, 15, 18, 15)55,360 (55,296 + 64)
Conv3D (64 filters, 3 × 3 × 3)(B, 64, 15, 18, 15)110,656 (110,592 + 64)
MaxPool3D (2 × 2 × 2)(B, 64, 7, 9, 7)0
Block 5
Conv3D (128 filters, 3 × 3 × 3)(B, 128, 7, 9, 7)221,312 (221,184 + 128)
Conv3D (128 filters, 3 × 3 × 3)(B, 128, 7, 9, 7)442,496 (442,368 + 128)
MaxPool3D (2 × 2 × 2)(B, 128, 3, 4, 3)0
Block 6
Conv3D (256 filters, 3 × 3 × 3)(B, 256, 3, 4, 3)884,992 (884,736 + 256)
Conv3D (256 filters, 3 × 3 × 3)(B, 256, 3, 4, 3)1,769,728 (1,769,472 + 256)
MaxPool3D (2 × 2 × 2)(B, 256, 1, 2, 1)0
Fully Connected Layers
Fully connected 1 (512 → 128)(B, 128)65,664 (65,536 + 128)
BatchNorm1D (128)(B, 128)256 (128 + 128)
Dropout(B, 128)0
Fully connected 2 (128 → 64)(B, 64)8256 (8192 + 64)
Fully connected 3 (64 → 2)(B, 2)130 (128 + 2)
Table 3. Hyperparameters used in the MDD-Net model.
Table 3. Hyperparameters used in the MDD-Net model.
Activation functionRectified linear unit
Weight initializationKaiming initialization
OptimizerAdamW
Model depth4
Dropout rate0.7
No. of epoch50
Learning Rate0.001
Loss functionCustom Categorical Cross-Entropy Loss
Batch size16
Table 4. Configuration of models from hyperparameter tuning.
Table 4. Configuration of models from hyperparameter tuning.
  Config   Model Depth   OptimizerLearning Rate
( η )
Loss Function
Formulation
Config-A4Adam0.01 L c u s t o m
Config-B4AdamW0.001 L c u s t o m
Config-C4AdamW0.001 L s t a n d a r d
Config-D6AdamW0.001 L c u s t o m
Config-E6Adam0.0001 L c u s t o m
Config-F6AdamW0.001 L s t a n d a r d
Table 5. Classification performance MDD-Net model.
Table 5. Classification performance MDD-Net model.
ConfigAccuracySensitivitySpecificityF1 ScoreAUROCAUPRC
Config-A66.95%0.930.360.750.740.78
Config-B72.26%0.830.590.760.800.81
Config-C52.38%0.130.990.220.770.77
Config-D71.43%0.750.670.740.760.76
Config-E72.00%0.670.760.720.800.81
Config-F53.78%0.130.980.240.750.77
Table 6. Performance comparison.
Table 6. Performance comparison.
SI No.AuthorsSample SizeModel UsedAccuracy
1.Yuqi et al. [29]MDD = 441, HC = 395DFH57.7%
2.Wang et al. [30]MDD = 54, HC = 62Ensemble model72.4%
3.Lin et al. [17]MDD = 49, HC = 283D CNN85%
4.Hong et al. [31]MDD = 34, HC = 343D FRN-ResNet86.7%
5.ProposedMDD = 1276, HC = 11043D CNN72.26%
Table 7. Identified brain regions affected by major depressive disorder and their functions.
Table 7. Identified brain regions affected by major depressive disorder and their functions.
SI
No.
Brain RegionsFunctionalities
1.Frontal LobePlanning, decision-making, problem-solving, voluntary
motor control, behavior, emotional regulation, speech and hearing
2.Parietal LobeIntegration of somatosensory information like touch, pain,
temperature, proprioception, spatial orientation, perception,
aspects of language and mathematical processing
3.Temporal LobePrimary auditory processing, language comprehension,
encoding and retrieval of declarative memory, visual object
and face recognition and emotional processing
4.ThalamusRegulation of consciousness, memory, emotion, alertness,
sleep-wake cycles, relaying sensory and motor signals to the
cerebral cortex
5.Insular CortexInvolved in interceptive awareness, taste processing,
automatic control, pain perception and emotional regulation
6.Corpus CallosumCommunication and coordination of motor, sensory and
cognitive functions
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

A. R., B.; Adiga, A.; Mahanand, B.S.; DIRECT Consortium. 3D Convolutional Neural Network Model for Detection of Major Depressive Disorder from Grey Matter Images. Appl. Sci. 2025, 15, 10312. https://doi.org/10.3390/app151910312

AMA Style

A. R. B, Adiga A, Mahanand BS, DIRECT Consortium. 3D Convolutional Neural Network Model for Detection of Major Depressive Disorder from Grey Matter Images. Applied Sciences. 2025; 15(19):10312. https://doi.org/10.3390/app151910312

Chicago/Turabian Style

A. R., Bindiya, Aditya Adiga, B. S. Mahanand, and DIRECT Consortium. 2025. "3D Convolutional Neural Network Model for Detection of Major Depressive Disorder from Grey Matter Images" Applied Sciences 15, no. 19: 10312. https://doi.org/10.3390/app151910312

APA Style

A. R., B., Adiga, A., Mahanand, B. S., & DIRECT Consortium. (2025). 3D Convolutional Neural Network Model for Detection of Major Depressive Disorder from Grey Matter Images. Applied Sciences, 15(19), 10312. https://doi.org/10.3390/app151910312

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop